/linux/arch/powerpc/mm/book3s64/ |
A D | hash_64k.c | 44 unsigned long old_pte, new_pte, subpg_pte; in __hash_page_4K() local 68 new_pte |= _PAGE_DIRTY; in __hash_page_4K() 74 subpg_pte = new_pte & ~subpg_prot; in __hash_page_4K() 107 new_pte &= ~H_PAGE_HASHPTE; in __hash_page_4K() 129 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K() 217 new_pte |= H_PAGE_HASHPTE; in __hash_page_4K() 219 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K() 230 unsigned long old_pte, new_pte; in __hash_page_64K() local 260 new_pte |= _PAGE_DIRTY; in __hash_page_64K() 328 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; in __hash_page_64K() [all …]
|
A D | hash_4k.c | 26 unsigned long old_pte, new_pte; in __hash_page_4K() local 48 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; in __hash_page_4K() 50 new_pte |= _PAGE_DIRTY; in __hash_page_4K() 51 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_4K() 57 rflags = htab_convert_pte_flags(new_pte, flags); in __hash_page_4K() 119 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; in __hash_page_4K() 120 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); in __hash_page_4K() 122 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K()
|
A D | hash_hugetlbpage.c | 25 unsigned long old_pte, new_pte; in __hash_page_huge() local 60 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; in __hash_page_huge() 62 new_pte |= _PAGE_DIRTY; in __hash_page_huge() 63 } while(!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_huge() 69 rflags = htab_convert_pte_flags(new_pte, flags); in __hash_page_huge() 100 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; in __hash_page_huge() 116 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, offset); in __hash_page_huge() 122 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_huge()
|
A D | radix_pgtable.c | 1049 unsigned long old_pte, new_pte; in radix__ptep_set_access_flags() local 1055 new_pte = old_pte | set; in radix__ptep_set_access_flags() 1057 __radix_pte_update(ptep, _PAGE_INVALID, new_pte); in radix__ptep_set_access_flags()
|
/linux/arch/arm/mm/ |
A D | pgd.c | 36 pte_t *new_pte, *init_pte; in pgd_alloc() local 103 new_pte = pte_alloc_map(mm, new_pmd, 0); in pgd_alloc() 104 if (!new_pte) in pgd_alloc() 121 set_pte_ext(new_pte + 0, init_pte[0], 0); in pgd_alloc() 122 set_pte_ext(new_pte + 1, init_pte[1], 0); in pgd_alloc() 124 pte_unmap(new_pte); in pgd_alloc()
|
/linux/arch/powerpc/include/asm/ |
A D | kvm_book3s_64.h | 436 pte_t old_pte, new_pte = __pte(0); in kvmppc_read_update_linux_pte() local 454 new_pte = pte_mkyoung(old_pte); in kvmppc_read_update_linux_pte() 456 new_pte = pte_mkdirty(new_pte); in kvmppc_read_update_linux_pte() 458 if (pte_xchg(ptep, old_pte, new_pte)) in kvmppc_read_update_linux_pte() 461 return new_pte; in kvmppc_read_update_linux_pte()
|
/linux/arch/riscv/kvm/ |
A D | mmu.c | 175 gpa_t addr, const pte_t *new_pte) in stage2_set_pte() argument 204 *ptep = *new_pte; in stage2_set_pte() 219 pte_t new_pte; in stage2_map_page() local 250 new_pte = pfn_pte(PFN_DOWN(hpa), prot); in stage2_map_page() 251 new_pte = pte_mkdirty(new_pte); in stage2_map_page() 253 return stage2_set_pte(kvm, level, pcache, gpa, &new_pte); in stage2_map_page()
|
/linux/arch/x86/mm/ |
A D | mem_encrypt.c | 296 pte_t new_pte; in __set_clr_pte_enc() local 329 new_pte = pfn_pte(pfn, new_prot); in __set_clr_pte_enc() 330 set_pte_atomic(kpte, new_pte); in __set_clr_pte_enc()
|
A D | init_64.c | 291 static void __set_pte_vaddr(pud_t *pud, unsigned long vaddr, pte_t new_pte) in __set_pte_vaddr() argument 296 set_pte(pte, new_pte); in __set_pte_vaddr() 305 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_p4d() argument 310 __set_pte_vaddr(pud, vaddr, new_pte); in set_pte_vaddr_p4d() 313 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_pud() argument 317 __set_pte_vaddr(pud, vaddr, new_pte); in set_pte_vaddr_pud()
|
/linux/arch/x86/include/asm/ |
A D | pgtable_64.h | 62 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte); 63 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte);
|
/linux/arch/mips/kvm/ |
A D | mmu.c | 131 pte_t *new_pte; in kvm_mips_walk_pgd() local 135 new_pte = kvm_mmu_memory_cache_alloc(cache); in kvm_mips_walk_pgd() 136 clear_page(new_pte); in kvm_mips_walk_pgd() 137 pmd_populate_kernel(NULL, pmd, new_pte); in kvm_mips_walk_pgd()
|
/linux/mm/ |
A D | mremap.c | 140 pte_t *old_pte, *new_pte, pte; in move_ptes() local 171 new_pte = pte_offset_map(new_pmd, new_addr); in move_ptes() 179 new_pte++, new_addr += PAGE_SIZE) { in move_ptes() 199 set_pte_at(mm, new_addr, new_pte, pte); in move_ptes() 207 pte_unmap(new_pte - 1); in move_ptes()
|
A D | hugetlb.c | 5388 pte_t new_pte; in hugetlb_no_page() local 5515 new_pte = make_huge_pte(vma, page, ((vma->vm_flags & VM_WRITE) in hugetlb_no_page() 5517 set_huge_pte_at(mm, haddr, ptep, new_pte); in hugetlb_no_page()
|
/linux/arch/x86/mm/pat/ |
A D | set_memory.c | 781 pte_t new_pte, *tmp; in __should_split_large_page() local 912 new_pte = pfn_pte(old_pfn, new_prot); in __should_split_large_page() 913 __set_pmd_pte(kpte, address, new_pte); in __should_split_large_page() 1538 pte_t new_pte; in __change_page_attr() local 1557 new_pte = pfn_pte(pfn, new_prot); in __change_page_attr() 1562 if (pte_val(old_pte) != pte_val(new_pte)) { in __change_page_attr() 1563 set_pte_atomic(kpte, new_pte); in __change_page_attr()
|
/linux/arch/powerpc/include/asm/book3s/64/ |
A D | hugetlb.h | 63 pte_t old_pte, pte_t new_pte);
|
/linux/arch/x86/kvm/mmu/ |
A D | paging_tmpl.h | 148 pt_element_t orig_pte, pt_element_t new_pte) in FNAME() 158 ret = CMPXCHG(&table[index], orig_pte, new_pte); in FNAME() 181 ret = CMPXCHG(&table[index], orig_pte, new_pte); in FNAME()
|
/linux/tools/testing/selftests/kvm/lib/x86_64/ |
A D | processor.c | 371 struct pageTableEntry *new_pte = _vm_get_page_table_entry(vm, vcpuid, in vm_set_page_table_entry() local 374 *(uint64_t *)new_pte = pte; in vm_set_page_table_entry()
|