/linux/include/linux/ |
A D | pagevec.h | 26 void __pagevec_release(struct pagevec *pvec); 27 void __pagevec_lru_add(struct pagevec *pvec); 50 pvec->nr = 0; in pagevec_init() 51 pvec->percpu_pvec_drained = false; in pagevec_init() 56 pvec->nr = 0; in pagevec_reinit() 61 return pvec->nr; in pagevec_count() 66 return PAGEVEC_SIZE - pvec->nr; in pagevec_space() 74 pvec->pages[pvec->nr++] = page; in pagevec_add() 75 return pagevec_space(pvec); in pagevec_add() 80 if (pagevec_count(pvec)) in pagevec_release() [all …]
|
/linux/mm/ |
A D | swap.c | 52 struct pagevec pvec; member 215 release_pages(pvec->pages, pvec->nr); in pagevec_lru_move_fn() 216 pagevec_reinit(pvec); in pagevec_lru_move_fn() 254 struct pagevec *pvec; in folio_rotate_reclaimable() local 259 pvec = this_cpu_ptr(&lru_rotate.pvec); in folio_rotate_reclaimable() 382 struct pagevec *pvec; in __lru_cache_activate_folio() local 615 pvec = &per_cpu(lru_rotate.pvec, cpu); in lru_add_drain_cpu() 1000 release_pages(pvec->pages, pagevec_count(pvec)); in __pagevec_release() 1075 release_pages(pvec->pages, pvec->nr); in __pagevec_lru_add() 1097 pvec->nr = j; in pagevec_remove_exceptionals() [all …]
|
A D | truncate.c | 90 pvec->pages[j++] = page; in truncate_exceptional_pvec_entries() 108 pvec->nr = j; in truncate_exceptional_pvec_entries() 299 struct pagevec pvec; in truncate_inode_pages_range() local 328 pagevec_init(&pvec); in truncate_inode_pages_range() 331 &pvec, indices)) { in truncate_inode_pages_range() 339 pagevec_release(&pvec); in truncate_inode_pages_range() 411 pagevec_release(&pvec); in truncate_inode_pages_range() 482 struct pagevec pvec; in __invalidate_mapping_pages() local 488 pagevec_init(&pvec); in __invalidate_mapping_pages() 627 struct pagevec pvec; in invalidate_inode_pages2_range() local [all …]
|
A D | mlock.c | 223 pagevec_add(pvec, page); in __putback_lru_fast_prepare() 246 __pagevec_lru_add(pvec); in __putback_lru_fast() 263 int nr = pagevec_count(pvec); in __munlock_pagevec() 273 struct page *page = pvec->pages[i]; in __munlock_pagevec() 298 pvec->pages[i] = NULL; in __munlock_pagevec() 312 struct page *page = pvec->pages[i]; in __munlock_pagevec() 397 if (pagevec_add(pvec, page) == 0) in __munlock_pagevec_fill() 431 struct pagevec pvec; in munlock_vma_pages_range() local 434 pagevec_init(&pvec); in munlock_vma_pages_range() 465 pagevec_add(&pvec, page); in munlock_vma_pages_range() [all …]
|
A D | filemap.c | 343 if (!pagevec_count(pvec)) in delete_from_page_cache_batch() 525 struct pagevec pvec; in __filemap_fdatawait_range() local 531 pagevec_init(&pvec); in __filemap_fdatawait_range() 546 pagevec_release(&pvec); in __filemap_fdatawait_range() 2071 pvec->nr = ret; in find_get_entries() 2483 struct pagevec *pvec) in filemap_create_page() argument 2537 struct pagevec *pvec) in filemap_get_pages() argument 2570 page = pvec->pages[pagevec_count(pvec) - 1]; in filemap_get_pages() 2588 if (likely(--pvec->nr)) in filemap_get_pages() 2615 struct pagevec pvec; in filemap_read() local [all …]
|
A D | shmem.c | 867 struct pagevec pvec; in shmem_unlock_mapping() local 870 pagevec_init(&pvec); in shmem_unlock_mapping() 878 pagevec_release(&pvec); in shmem_unlock_mapping() 922 struct pagevec pvec; in shmem_undo_range() local 934 pagevec_init(&pvec); in shmem_undo_range() 937 &pvec, indices)) { in shmem_undo_range() 957 pagevec_release(&pvec); in shmem_undo_range() 1044 pagevec_release(&pvec); in shmem_undo_range() 1249 struct pagevec pvec; in shmem_unuse_inode() local 1254 pagevec_init(&pvec); in shmem_unuse_inode() [all …]
|
A D | page-writeback.c | 2184 struct pagevec pvec; in write_cache_pages() local 2192 pagevec_init(&pvec); in write_cache_pages() 2212 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages() 2218 struct page *page = pvec.pages[i]; in write_cache_pages() 2294 pagevec_release(&pvec); in write_cache_pages()
|
/linux/fs/nilfs2/ |
A D | page.c | 243 struct pagevec pvec; in nilfs_copy_dirty_pages() local 248 pagevec_init(&pvec); in nilfs_copy_dirty_pages() 278 pagevec_release(&pvec); in nilfs_copy_dirty_pages() 297 struct pagevec pvec; in nilfs_copy_back_pages() local 301 pagevec_init(&pvec); in nilfs_copy_back_pages() 360 struct pagevec pvec; in nilfs_clear_dirty_pages() local 364 pagevec_init(&pvec); in nilfs_clear_dirty_pages() 498 struct pagevec pvec; in nilfs_find_uncommitted_extent() local 507 pagevec_init(&pvec); in nilfs_find_uncommitted_extent() 511 pvec.pages); in nilfs_find_uncommitted_extent() [all …]
|
A D | segment.c | 683 struct pagevec pvec; in nilfs_lookup_dirty_data_buffers() local 697 pagevec_init(&pvec); in nilfs_lookup_dirty_data_buffers() 704 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_lookup_dirty_data_buffers() 706 struct page *page = pvec.pages[i]; in nilfs_lookup_dirty_data_buffers() 721 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers() 727 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers() 737 struct pagevec pvec; in nilfs_lookup_dirty_node_buffers() local 742 pagevec_init(&pvec); in nilfs_lookup_dirty_node_buffers() 746 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_lookup_dirty_node_buffers() 747 bh = head = page_buffers(pvec.pages[i]); in nilfs_lookup_dirty_node_buffers() [all …]
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_userptr.c | 115 pvec = obj->userptr.pvec; in i915_gem_object_userptr_drop_ref() 120 if (pvec) { in i915_gem_object_userptr_drop_ref() 124 kvfree(pvec); in i915_gem_object_userptr_drop_ref() 134 struct page **pvec; in i915_gem_userptr_get_pages() local 147 pvec = obj->userptr.pvec; in i915_gem_userptr_get_pages() 263 struct page **pvec; in i915_gem_object_userptr_submit_init() local 288 if (!pvec) in i915_gem_object_userptr_submit_init() 318 obj->userptr.pvec = pvec; in i915_gem_object_userptr_submit_init() 320 pvec = NULL; in i915_gem_object_userptr_submit_init() 330 if (pvec) { in i915_gem_object_userptr_submit_init() [all …]
|
A D | i915_gem_shmem.c | 23 check_move_unevictable_pages(pvec); in check_release_pagevec() 24 __pagevec_release(pvec); in check_release_pagevec() 197 struct pagevec pvec; in shmem_get_pages() local 199 pagevec_init(&pvec); in shmem_get_pages() 204 if (pagevec_count(&pvec)) in shmem_get_pages() 205 check_release_pagevec(&pvec); in shmem_get_pages() 317 struct pagevec pvec; in i915_gem_object_put_pages_shmem() local 330 pagevec_init(&pvec); in i915_gem_object_put_pages_shmem() 338 if (!pagevec_add(&pvec, page)) in i915_gem_object_put_pages_shmem() 341 if (pagevec_count(&pvec)) in i915_gem_object_put_pages_shmem() [all …]
|
A D | i915_gem_object_types.h | 591 struct page **pvec; member
|
/linux/drivers/gpu/drm/ |
A D | drm_gem.c | 510 check_move_unevictable_pages(pvec); in drm_gem_check_release_pagevec() 511 __pagevec_release(pvec); in drm_gem_check_release_pagevec() 544 struct pagevec pvec; in drm_gem_get_pages() local 587 pagevec_init(&pvec); in drm_gem_get_pages() 589 if (!pagevec_add(&pvec, pages[i])) in drm_gem_get_pages() 592 if (pagevec_count(&pvec)) in drm_gem_get_pages() 593 drm_gem_check_release_pagevec(&pvec); in drm_gem_get_pages() 612 struct pagevec pvec; in drm_gem_put_pages() local 625 pagevec_init(&pvec); in drm_gem_put_pages() 637 if (!pagevec_add(&pvec, pages[i])) in drm_gem_put_pages() [all …]
|
/linux/fs/gfs2/ |
A D | aops.c | 235 struct pagevec *pvec, in gfs2_write_jdata_pagevec() argument 250 struct page *page = pvec->pages[i]; in gfs2_write_jdata_pagevec() 333 struct pagevec pvec; in gfs2_write_cache_jdata() local 343 pagevec_init(&pvec); in gfs2_write_cache_jdata() 369 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in gfs2_write_cache_jdata() 374 ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, &done_index); in gfs2_write_cache_jdata() 379 pagevec_release(&pvec); in gfs2_write_cache_jdata()
|
/linux/drivers/gpu/drm/etnaviv/ |
A D | etnaviv_gem.c | 662 struct page **pvec = NULL; in etnaviv_gem_userptr_get_pages() local 671 pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_userptr_get_pages() 672 if (!pvec) in etnaviv_gem_userptr_get_pages() 678 struct page **pages = pvec + pinned; in etnaviv_gem_userptr_get_pages() 684 unpin_user_pages(pvec, pinned); in etnaviv_gem_userptr_get_pages() 685 kvfree(pvec); in etnaviv_gem_userptr_get_pages() 693 etnaviv_obj->pages = pvec; in etnaviv_gem_userptr_get_pages()
|
/linux/fs/ceph/ |
A D | addr.c | 716 struct pagevec pvec; in ceph_writepages_start() local 740 pagevec_init(&pvec); in ceph_writepages_start() 804 page = pvec.pages[i]; in ceph_writepages_start() 924 pvec.pages[i] = NULL; in ceph_writepages_start() 936 if (!pvec.pages[j]) in ceph_writepages_start() 939 pvec.pages[n] = pvec.pages[j]; in ceph_writepages_start() 942 pvec.nr = n; in ceph_writepages_start() 947 pagevec_release(&pvec); in ceph_writepages_start() 1073 pvec.nr ? pvec.pages[0] : NULL); in ceph_writepages_start() 1074 pagevec_release(&pvec); in ceph_writepages_start() [all …]
|
/linux/fs/hugetlbfs/ |
A D | inode.c | 111 static void huge_pagevec_release(struct pagevec *pvec) in huge_pagevec_release() argument 115 for (i = 0; i < pagevec_count(pvec); ++i) in huge_pagevec_release() 116 put_page(pvec->pages[i]); in huge_pagevec_release() 118 pagevec_reinit(pvec); in huge_pagevec_release() 471 struct pagevec pvec; in remove_inode_hugepages() local 476 pagevec_init(&pvec); in remove_inode_hugepages() 482 if (!pagevec_lookup_range(&pvec, mapping, &next, end - 1)) in remove_inode_hugepages() 485 for (i = 0; i < pagevec_count(&pvec); ++i) { in remove_inode_hugepages() 486 struct page *page = pvec.pages[i]; in remove_inode_hugepages() 545 huge_pagevec_release(&pvec); in remove_inode_hugepages()
|
/linux/fs/afs/ |
A D | write.c | 423 struct pagevec pvec; in afs_extend_writeback() local 434 pagevec_init(&pvec); in afs_extend_writeback() 491 if (!pagevec_add(&pvec, &folio->page)) in afs_extend_writeback() 504 if (!pagevec_count(&pvec)) in afs_extend_writeback() 507 for (i = 0; i < pagevec_count(&pvec); i++) { in afs_extend_writeback() 508 folio = page_folio(pvec.pages[i]); in afs_extend_writeback() 520 pagevec_release(&pvec); in afs_extend_writeback()
|
/linux/fs/f2fs/ |
A D | node.c | 1508 struct pagevec pvec; in last_fsync_dnode() local 1512 pagevec_init(&pvec); in last_fsync_dnode() 1555 pagevec_release(&pvec); in last_fsync_dnode() 1721 struct pagevec pvec; in f2fs_fsync_node_pages() local 1735 pagevec_init(&pvec); in f2fs_fsync_node_pages() 1813 pagevec_release(&pvec); in f2fs_fsync_node_pages() 1878 struct pagevec pvec; in f2fs_flush_inline_data() local 1881 pagevec_init(&pvec); in f2fs_flush_inline_data() 1915 pagevec_release(&pvec); in f2fs_flush_inline_data() 1925 struct pagevec pvec; in f2fs_sync_node_pages() local [all …]
|
A D | compress.c | 885 bool f2fs_all_cluster_page_loaded(struct compress_ctx *cc, struct pagevec *pvec, in f2fs_all_cluster_page_loaded() argument 894 pgidx = pvec->pages[index]->index; in f2fs_all_cluster_page_loaded() 897 if (pvec->pages[index + i]->index != pgidx + i) in f2fs_all_cluster_page_loaded() 1856 struct pagevec pvec; in f2fs_invalidate_compress_pages() local 1863 pagevec_init(&pvec); in f2fs_invalidate_compress_pages() 1869 nr_pages = pagevec_lookup_range(&pvec, mapping, in f2fs_invalidate_compress_pages() 1875 struct page *page = pvec.pages[i]; in f2fs_invalidate_compress_pages() 1894 pagevec_release(&pvec); in f2fs_invalidate_compress_pages()
|
A D | checkpoint.c | 375 struct pagevec pvec; in f2fs_sync_meta_pages() local 383 pagevec_init(&pvec); in f2fs_sync_meta_pages() 387 while ((nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, in f2fs_sync_meta_pages() 392 struct page *page = pvec.pages[i]; in f2fs_sync_meta_pages() 397 pagevec_release(&pvec); in f2fs_sync_meta_pages() 427 pagevec_release(&pvec); in f2fs_sync_meta_pages()
|
/linux/arch/x86/kvm/mmu/ |
A D | mmu.c | 1811 for (i=0; i < pvec->nr; i++) in mmu_pages_add() 1812 if (pvec->page[i].sp == sp) in mmu_pages_add() 1815 pvec->page[pvec->nr].sp = sp; in mmu_pages_add() 1816 pvec->page[pvec->nr].idx = idx; in mmu_pages_add() 1817 pvec->nr++; in mmu_pages_add() 1829 struct kvm_mmu_pages *pvec) in __mmu_unsync_walk() argument 1870 struct kvm_mmu_pages *pvec) in mmu_unsync_walk() argument 1872 pvec->nr = 0; in mmu_unsync_walk() 1954 i < pvec.nr && ({ sp = pvec.page[i].sp; 1;}); \ 1984 if (pvec->nr == 0) in mmu_pages_first() [all …]
|
/linux/fs/fscache/ |
A D | page.c | 1213 struct pagevec pvec; in __fscache_uncache_all_inode_pages() local 1224 pagevec_init(&pvec); in __fscache_uncache_all_inode_pages() 1227 if (!pagevec_lookup(&pvec, mapping, &next)) in __fscache_uncache_all_inode_pages() 1229 for (i = 0; i < pagevec_count(&pvec); i++) { in __fscache_uncache_all_inode_pages() 1230 struct page *page = pvec.pages[i]; in __fscache_uncache_all_inode_pages() 1236 pagevec_release(&pvec); in __fscache_uncache_all_inode_pages()
|
/linux/fs/ext4/ |
A D | inode.c | 1547 struct pagevec pvec; in mpage_release_unused_pages() local 1565 pagevec_init(&pvec); in mpage_release_unused_pages() 1583 pagevec_release(&pvec); in mpage_release_unused_pages() 2309 struct pagevec pvec; in mpage_map_and_submit_buffers() local 2324 pagevec_init(&pvec); in mpage_map_and_submit_buffers() 2347 pagevec_release(&pvec); in mpage_map_and_submit_buffers() 2354 pagevec_release(&pvec); in mpage_map_and_submit_buffers() 2552 struct pagevec pvec; in mpage_prepare_extent_to_map() local 2568 pagevec_init(&pvec); in mpage_prepare_extent_to_map() 2627 pagevec_release(&pvec); in mpage_prepare_extent_to_map() [all …]
|
/linux/drivers/mtd/lpddr/ |
A D | lpddr_cmds.c | 404 unsigned long adr, const struct kvec **pvec, in do_write_buffer() argument 436 vec = *pvec; in do_write_buffer() 466 *pvec = vec; in do_write_buffer()
|