Searched refs:pte_lockptr (Results 1 – 9 of 9) sorted by relevance
/linux/mm/ |
A D | page_vma_mapped.c | 51 pvmw->ptl = pte_lockptr(pvmw->vma->vm_mm, pvmw->pmd); in map_pte() 283 pvmw->ptl = pte_lockptr(mm, pvmw->pmd); in page_vma_mapped_walk() 289 pvmw->ptl = pte_lockptr(mm, pvmw->pmd); in page_vma_mapped_walk()
|
A D | mremap.c | 172 new_ptl = pte_lockptr(mm, new_pmd); in move_ptes()
|
A D | memory.c | 1030 src_ptl = pte_lockptr(src_mm, src_pmd); in copy_pte_range() 2718 spinlock_t *ptl = pte_lockptr(vmf->vma->vm_mm, vmf->pmd); in pte_unmap_same() 4337 vmf->ptl = pte_lockptr(vma->vm_mm, vmf->pmd); in do_numa_page() 4560 vmf->ptl = pte_lockptr(vmf->vma->vm_mm, vmf->pmd); in handle_pte_fault()
|
A D | khugepaged.c | 1145 pte_ptl = pte_lockptr(mm, pmd); in collapse_huge_page()
|
A D | migrate.c | 325 spinlock_t *ptl = pte_lockptr(mm, pmd); in migration_entry_wait()
|
/linux/arch/arm/mm/ |
A D | fault-armv.c | 120 ptl = pte_lockptr(vma->vm_mm, pmd); in adjust_pte()
|
/linux/arch/powerpc/mm/ |
A D | pgtable.c | 321 assert_spin_locked(pte_lockptr(mm, pmd)); in assert_pte_locked()
|
/linux/Documentation/vm/ |
A D | split_page_table_lock.rst | 26 - pte_lockptr()
|
/linux/include/linux/ |
A D | mm.h | 2317 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd) in pte_lockptr() function 2342 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd) in pte_lockptr() function 2375 spinlock_t *__ptl = pte_lockptr(mm, pmd); \
|
Completed in 33 milliseconds