/linux/fs/hfsplus/ |
A D | bnode.c | 146 kunmap(*src_page); in hfs_bnode_copy() 153 kunmap(*src_page); in hfs_bnode_copy() 174 kunmap(*src_page); in hfs_bnode_copy() 180 src_page++; in hfs_bnode_copy() 206 kunmap(*src_page); in hfs_bnode_move() 211 src_page--; in hfs_bnode_move() 217 kunmap(*src_page); in hfs_bnode_move() 237 kunmap(*src_page); in hfs_bnode_move() 243 src_page--; in hfs_bnode_move() 256 kunmap(*src_page); in hfs_bnode_move() [all …]
|
/linux/fs/ |
A D | remap_range.c | 198 struct page *src_page; in vfs_dedupe_file_range_compare() local 215 src_page = vfs_dedupe_get_page(src, srcoff); in vfs_dedupe_file_range_compare() 216 if (IS_ERR(src_page)) { in vfs_dedupe_file_range_compare() 217 error = PTR_ERR(src_page); in vfs_dedupe_file_range_compare() 223 put_page(src_page); in vfs_dedupe_file_range_compare() 227 vfs_lock_two_pages(src_page, dest_page); in vfs_dedupe_file_range_compare() 235 src_page->mapping != src->i_mapping || in vfs_dedupe_file_range_compare() 241 src_addr = kmap_atomic(src_page); in vfs_dedupe_file_range_compare() 244 flush_dcache_page(src_page); in vfs_dedupe_file_range_compare() 253 vfs_unlock_two_pages(src_page, dest_page); in vfs_dedupe_file_range_compare() [all …]
|
/linux/fs/verity/ |
A D | enable.c | 72 struct page *src_page; in build_merkle_tree_level() local 80 src_page = read_file_data_page(filp, i, &ra, in build_merkle_tree_level() 82 if (IS_ERR(src_page)) { in build_merkle_tree_level() 83 err = PTR_ERR(src_page); in build_merkle_tree_level() 95 src_page = vops->read_merkle_tree_page(inode, in build_merkle_tree_level() 98 if (IS_ERR(src_page)) { in build_merkle_tree_level() 99 err = PTR_ERR(src_page); in build_merkle_tree_level() 107 err = fsverity_hash_page(params, inode, req, src_page, in build_merkle_tree_level() 109 put_page(src_page); in build_merkle_tree_level()
|
/linux/arch/arm/lib/ |
A D | uaccess_with_memcpy.c | 237 struct page *src_page, *dst_page; 243 src_page = alloc_page(GFP_KERNEL); 244 if (!src_page) 249 kernel_ptr = page_address(src_page); 282 put_page(src_page);
|
/linux/include/linux/ |
A D | highmem.h | 302 struct page *src_page, size_t src_off, in memcpy_page() argument 306 char *src = kmap_local_page(src_page); in memcpy_page() 315 struct page *src_page, size_t src_off, in memmove_page() argument 319 char *src = kmap_local_page(src_page); in memmove_page()
|
/linux/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_blit.c | 367 u32 src_page = src_offset >> PAGE_SHIFT; in vmw_bo_cpu_blit_line() local 371 bool unmap_src = d->src_addr && (src_page != d->mapped_src || in vmw_bo_cpu_blit_line() 401 if (WARN_ON_ONCE(src_page >= d->src_num_pages)) in vmw_bo_cpu_blit_line() 405 kmap_atomic_prot(d->src_pages[src_page], in vmw_bo_cpu_blit_line() 410 d->mapped_src = src_page; in vmw_bo_cpu_blit_line()
|
/linux/arch/x86/kernel/cpu/sgx/ |
A D | ioctl.c | 237 struct page *src_page; in __sgx_encl_add_page() local 248 ret = get_user_pages(src, 1, 0, &src_page, NULL); in __sgx_encl_add_page() 255 pginfo.contents = (unsigned long)kmap_atomic(src_page); in __sgx_encl_add_page() 260 put_page(src_page); in __sgx_encl_add_page()
|
/linux/mm/ |
A D | khugepaged.c | 745 struct page *src_page, *tmp; in __collapse_huge_page_copy() local 767 src_page = pte_page(pteval); in __collapse_huge_page_copy() 768 copy_user_highpage(page, src_page, address, vma); in __collapse_huge_page_copy() 769 if (!PageCompound(src_page)) in __collapse_huge_page_copy() 770 release_pte_page(src_page); in __collapse_huge_page_copy() 782 page_remove_rmap(src_page, false); in __collapse_huge_page_copy() 784 free_page_and_swap_cache(src_page); in __collapse_huge_page_copy() 788 list_for_each_entry_safe(src_page, tmp, compound_pagelist, lru) { in __collapse_huge_page_copy() 789 list_del(&src_page->lru); in __collapse_huge_page_copy() 790 release_pte_page(src_page); in __collapse_huge_page_copy()
|
A D | huge_memory.c | 1032 struct page *src_page; in copy_huge_pmd() local 1097 src_page = pmd_page(pmd); in copy_huge_pmd() 1098 VM_BUG_ON_PAGE(!PageHead(src_page), src_page); in copy_huge_pmd() 1107 if (unlikely(page_needs_cow_for_dma(src_vma, src_page))) { in copy_huge_pmd() 1115 get_page(src_page); in copy_huge_pmd() 1116 page_dup_rmap(src_page, true); in copy_huge_pmd()
|
/linux/fs/hfs/ |
A D | bnode.c | 116 struct page *src_page, *dst_page; in hfs_bnode_copy() local 123 src_page = src_node->page[0]; in hfs_bnode_copy() 126 memcpy(kmap(dst_page) + dst, kmap(src_page) + src, len); in hfs_bnode_copy() 127 kunmap(src_page); in hfs_bnode_copy()
|
/linux/fs/crypto/ |
A D | crypto.c | 94 u64 lblk_num, struct page *src_page, in fscrypt_crypt_block() argument 124 sg_set_page(&src, src_page, len, offs); in fscrypt_crypt_block()
|
A D | fscrypt_private.h | 269 u64 lblk_num, struct page *src_page,
|
/linux/fs/f2fs/ |
A D | node.c | 139 struct page *src_page; in get_next_nat_page() local 149 src_page = get_current_nat_page(sbi, nid); in get_next_nat_page() 150 if (IS_ERR(src_page)) in get_next_nat_page() 151 return src_page; in get_next_nat_page() 153 f2fs_bug_on(sbi, PageDirty(src_page)); in get_next_nat_page() 155 src_addr = page_address(src_page); in get_next_nat_page() 159 f2fs_put_page(src_page, 1); in get_next_nat_page()
|
/linux/fs/ecryptfs/ |
A D | crypto.c | 376 struct page *src_page, in crypt_extent() argument 379 pgoff_t page_index = op == ENCRYPT ? src_page->index : dst_page->index; in crypt_extent() 399 sg_set_page(&src_sg, src_page, extent_size, in crypt_extent()
|
/linux/fs/btrfs/ |
A D | extent_io.c | 7164 static void copy_pages(struct page *dst_page, struct page *src_page, in copy_pages() argument 7172 if (dst_page != src_page) { in copy_pages() 7173 src_kaddr = page_address(src_page); in copy_pages()
|