Searched refs:hpage (Results 1 – 13 of 13) sorted by relevance
/linux/fs/verity/ |
A D | verify.c | 46 void *virt = kmap_atomic(hpage); in extract_hash() 114 struct page *hpage; in verify_page() local 123 if (IS_ERR(hpage)) { in verify_page() 124 err = PTR_ERR(hpage); in verify_page() 131 if (PageChecked(hpage)) { in verify_page() 134 put_page(hpage); in verify_page() 141 hpages[level] = hpage; in verify_page() 151 struct page *hpage = hpages[level - 1]; in verify_page() local 160 SetPageChecked(hpage); in verify_page() 161 extract_hash(hpage, hoffset, hsize, _want_hash); in verify_page() [all …]
|
/linux/mm/ |
A D | khugepaged.c | 869 *hpage = NULL; in khugepaged_prealloc_page() 873 *hpage = NULL; in khugepaged_prealloc_page() 882 VM_BUG_ON_PAGE(*hpage, *hpage); in khugepaged_alloc_page() 893 return *hpage; in khugepaged_alloc_page() 918 if (!hpage) { in khugepaged_alloc_hugepage() 929 return hpage; in khugepaged_alloc_hugepage() 941 if (*hpage && page_count(*hpage) > 1) { in khugepaged_prealloc_page() 946 if (!*hpage) in khugepaged_prealloc_page() 1209 *hpage = NULL; in collapse_huge_page() 1459 if (!hpage) in collapse_pte_mapped_thp() [all …]
|
A D | memory-failure.c | 1032 if (!PageHuge(hpage)) in me_huge_page() 1038 unlock_page(hpage); in me_huge_page() 1041 unlock_page(hpage); in me_huge_page() 1047 if (PageAnon(hpage)) in me_huge_page() 1048 put_page(hpage); in me_huge_page() 1319 if (!page_mapped(hpage)) in hwpoison_user_mappings() 1343 SetPageDirty(hpage); in hwpoison_user_mappings() 1363 if (!PageHuge(hpage)) { in hwpoison_user_mappings() 1366 if (!PageAnon(hpage)) { in hwpoison_user_mappings() 1395 shake_page(hpage); in hwpoison_user_mappings() [all …]
|
A D | hwpoison-inject.c | 18 struct page *hpage; in hwpoison_inject() local 28 hpage = compound_head(p); in hwpoison_inject() 33 shake_page(hpage); in hwpoison_inject() 37 if (!PageLRU(hpage) && !PageHuge(p)) in hwpoison_inject() 45 err = hwpoison_filter(hpage); in hwpoison_inject()
|
A D | migrate.c | 1293 if (page_count(hpage) == 1) { in unmap_and_move_huge_page() 1303 if (!trylock_page(hpage)) { in unmap_and_move_huge_page() 1313 lock_page(hpage); in unmap_and_move_huge_page() 1321 if (hugetlb_page_subpool(hpage) && !page_mapping(hpage)) { in unmap_and_move_huge_page() 1326 if (PageAnon(hpage)) in unmap_and_move_huge_page() 1332 if (page_mapped(hpage)) { in unmap_and_move_huge_page() 1336 if (!PageAnon(hpage)) { in unmap_and_move_huge_page() 1351 try_to_migrate(hpage, ttu); in unmap_and_move_huge_page() 1358 if (!page_mapped(hpage)) in unmap_and_move_huge_page() 1362 remove_migration_ptes(hpage, in unmap_and_move_huge_page() [all …]
|
A D | hugetlb.c | 1865 struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage) in hugetlb_page_mapping_lock_write() argument 1867 struct address_space *mapping = page_mapping(hpage); in hugetlb_page_mapping_lock_write()
|
/linux/drivers/dma-buf/ |
A D | udmabuf.c | 173 struct page *page, *hpage = NULL; in udmabuf_create() local 227 if (!hpage) { in udmabuf_create() 228 hpage = find_get_page_flags(mapping, pgoff, in udmabuf_create() 230 if (!hpage) { in udmabuf_create() 235 page = hpage + subpgoff; in udmabuf_create() 239 put_page(hpage); in udmabuf_create() 240 hpage = NULL; in udmabuf_create() 256 if (hpage) { in udmabuf_create() 257 put_page(hpage); in udmabuf_create() 258 hpage = NULL; in udmabuf_create()
|
/linux/include/linux/ |
A D | hugetlb.h | 182 struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage); 233 struct page *hpage) in hugetlb_page_mapping_lock_write() argument 669 static inline struct hugepage_subpool *hugetlb_page_subpool(struct page *hpage) in hugetlb_page_subpool() argument 671 return (void *)page_private(hpage + SUBPAGE_INDEX_SUBPOOL); in hugetlb_page_subpool() 674 static inline void hugetlb_set_page_subpool(struct page *hpage, in hugetlb_set_page_subpool() argument 677 set_page_private(hpage + SUBPAGE_INDEX_SUBPOOL, (unsigned long)subpool); in hugetlb_set_page_subpool() 916 static inline struct hugepage_subpool *hugetlb_page_subpool(struct page *hpage)
|
/linux/arch/powerpc/kvm/ |
A D | book3s_pr.c | 642 struct page *hpage; in kvmppc_patch_dcbz() local 647 hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); in kvmppc_patch_dcbz() 648 if (is_error_page(hpage)) in kvmppc_patch_dcbz() 655 get_page(hpage); in kvmppc_patch_dcbz() 656 page = kmap_atomic(hpage); in kvmppc_patch_dcbz() 664 put_page(hpage); in kvmppc_patch_dcbz()
|
/linux/Documentation/virt/kvm/ |
A D | running-nested-guests.rst | 169 .. note:: On s390x, the kernel parameter ``hpage`` is mutually exclusive 171 ``nested``, the ``hpage`` parameter *must* be disabled.
|
A D | api.rst | 6606 :Returns: 0 on success, -EINVAL if hpage module parameter was not set 6614 hpage module parameter is not set to 1, -EINVAL is returned.
|
/linux/arch/s390/kvm/ |
A D | kvm-s390.c | 191 static int hpage; variable 192 module_param(hpage, int, 0444); 193 MODULE_PARM_DESC(hpage, "1m huge page backing support"); 574 if (hpage && !kvm_is_ucontrol(kvm)) in kvm_vm_ioctl_check_extension() 790 else if (!hpage || kvm->arch.use_cmma || kvm_is_ucontrol(kvm)) in kvm_vm_ioctl_enable_cap() 5101 if (nested && hpage) { in kvm_s390_init()
|
/linux/fs/ |
A D | io_uring.c | 8967 int nr_pages, struct page *hpage) in headpage_already_acct() argument 8975 if (compound_head(pages[i]) == hpage) in headpage_already_acct() 8986 if (compound_head(imu->bvec[j].bv_page) == hpage) in headpage_already_acct() 9005 struct page *hpage; in io_buffer_account_pin() local 9007 hpage = compound_head(pages[i]); in io_buffer_account_pin() 9008 if (hpage == *last_hpage) in io_buffer_account_pin() 9010 *last_hpage = hpage; in io_buffer_account_pin() 9011 if (headpage_already_acct(ctx, pages, i, hpage)) in io_buffer_account_pin() 9013 imu->acct_pages += page_size(hpage) >> PAGE_SHIFT; in io_buffer_account_pin()
|
Completed in 93 milliseconds