/linux/mm/ |
A D | pgtable-generic.c | 113 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_set_access_flags() 127 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_clear_flush_young() 141 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_huge_clear_flush() 198 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate() 218 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_collapse_flush()
|
A D | huge_memory.c | 328 return sysfs_emit(buf, "%lu\n", HPAGE_PMD_SIZE); in hpage_pmd_size_show() 1510 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in madvise_free_huge_pmd() 1541 if (next - addr != HPAGE_PMD_SIZE) { in madvise_free_huge_pmd() 1586 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in zap_huge_pmd() 1637 tlb_remove_page_size(tlb, page, HPAGE_PMD_SIZE); in zap_huge_pmd() 1971 VM_BUG_ON_VMA(vma->vm_end < haddr + HPAGE_PMD_SIZE, vma); in __split_huge_pmd_locked() 2155 (address & HPAGE_PMD_MASK) + HPAGE_PMD_SIZE); in __split_huge_pmd() 2255 if (!IS_ALIGNED(address, HPAGE_PMD_SIZE) && in split_huge_pmd_if_needed() 2256 range_in_vma(vma, ALIGN_DOWN(address, HPAGE_PMD_SIZE), in split_huge_pmd_if_needed() 2257 ALIGN(address, HPAGE_PMD_SIZE))) in split_huge_pmd_if_needed() [all …]
|
A D | khugepaged.c | 986 if (address < hstart || address + HPAGE_PMD_SIZE > hend) in hugepage_vma_revalidate() 1141 address, address + HPAGE_PMD_SIZE); in collapse_huge_page() 1445 !range_in_vma(vma, haddr, haddr + HPAGE_PMD_SIZE)) in collapse_pte_mapped_thp() 1583 if (vma->vm_end < addr + HPAGE_PMD_SIZE) in retract_page_tables() 2145 khugepaged_scan.address + HPAGE_PMD_SIZE > in khugepaged_scan_mm_slot() 2162 khugepaged_scan.address += HPAGE_PMD_SIZE; in khugepaged_scan_mm_slot()
|
A D | shmem.c | 578 round_up(inode->i_size, HPAGE_PMD_SIZE)) { in shmem_unused_huge_shrink() 1076 stat->blksize = HPAGE_PMD_SIZE; in shmem_getattr() 2150 if (len < HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2182 offset = (pgoff << PAGE_SHIFT) & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2183 if (offset && offset + len < 2 * HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2185 if ((addr & (HPAGE_PMD_SIZE-1)) == offset) in shmem_get_unmapped_area() 2188 inflated_len = len + HPAGE_PMD_SIZE - PAGE_SIZE; in shmem_get_unmapped_area() 2200 inflated_offset = inflated_addr & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2203 inflated_addr += HPAGE_PMD_SIZE; in shmem_get_unmapped_area()
|
A D | madvise.c | 330 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in madvise_cold_or_pageout_pte_range() 351 if (next - addr != HPAGE_PMD_SIZE) { in madvise_cold_or_pageout_pte_range()
|
A D | mprotect.c | 261 if (next - addr != HPAGE_PMD_SIZE) { in change_pmd_range()
|
A D | mremap.c | 537 if (extent == HPAGE_PMD_SIZE && in move_page_tables()
|
A D | memory.c | 1148 VM_BUG_ON_VMA(next-addr != HPAGE_PMD_SIZE, src_vma); in copy_pmd_range() 1441 if (next - addr != HPAGE_PMD_SIZE) in zap_pmd_range() 1448 next - addr == HPAGE_PMD_SIZE && pmd_none(*pmd)) { in zap_pmd_range()
|
/linux/include/linux/ |
A D | huge_mm.h | 109 #define HPAGE_PMD_SIZE ((1UL) << HPAGE_PMD_SHIFT) macro 110 #define HPAGE_PMD_MASK (~(HPAGE_PMD_SIZE - 1)) 128 if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) in transhuge_vma_suitable() 324 #define HPAGE_PMD_SIZE ({ BUILD_BUG(); 0; }) macro
|
A D | mmu_notifier.h | 609 ___haddr + HPAGE_PMD_SIZE); \
|
/linux/arch/powerpc/mm/book3s64/ |
A D | pgtable.c | 119 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate() 137 flush_pmd_tlb_range(vma, addr, addr + HPAGE_PMD_SIZE); in pmdp_huge_get_and_clear_full()
|
A D | radix_tlb.c | 1410 end = addr + HPAGE_PMD_SIZE; in radix__flush_tlb_collapsed_pmd()
|
A D | hash_utils.c | 1868 max_hpte_count = HPAGE_PMD_SIZE >> shift; in flush_hash_hugepage()
|
/linux/arch/arc/mm/ |
A D | tlb.c | 675 mmu->s_pg_sz_m != TO_MB(HPAGE_PMD_SIZE)) in arc_mmu_init() 677 (unsigned long)TO_MB(HPAGE_PMD_SIZE)); in arc_mmu_init()
|
/linux/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_thp.c | 90 align_pages = (HPAGE_PMD_SIZE >> PAGE_SHIFT); in vmw_thp_get_node()
|
/linux/drivers/gpu/drm/ |
A D | drm_file.c | 1020 if (len < HPAGE_PMD_SIZE) in drm_get_unmapped_area() 1033 HPAGE_PMD_SIZE); in drm_get_unmapped_area()
|
/linux/arch/sparc/mm/ |
A D | tlb.c | 250 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate()
|
/linux/fs/proc/ |
A D | task_mmu.c | 562 mss->anonymous_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 564 mss->shmem_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 568 mss->file_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 1810 HPAGE_PMD_SIZE/PAGE_SIZE); in gather_pte_stats()
|
/linux/include/asm-generic/ |
A D | tlb.h | 585 tlb_flush_pmd_range(tlb, address, HPAGE_PMD_SIZE); \
|
/linux/drivers/nvdimm/ |
A D | pfn_devs.c | 101 alignments[1] = HPAGE_PMD_SIZE; in nd_pfn_supported_alignments() 116 return HPAGE_PMD_SIZE; in nd_pfn_default_alignment()
|
/linux/arch/x86/mm/ |
A D | pgtable.c | 607 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_clear_flush_young()
|