/linux/mm/damon/ |
A D | vaddr-test.h | 26 vmas[i].vm_next = &vmas[i + 1]; in __link_vmas() 28 vmas[i].vm_rb.rb_left = NULL; in __link_vmas() 29 vmas[i].vm_rb.rb_right = &vmas[i + 1].vm_rb; in __link_vmas() 35 gap = vmas[j].vm_start - vmas[j - 1].vm_end; in __link_vmas() 39 vmas[i].rb_subtree_gap = largest_gap; in __link_vmas() 41 vmas[i].vm_next = NULL; in __link_vmas() 42 vmas[i].vm_rb.rb_right = NULL; in __link_vmas() 43 vmas[i].rb_subtree_gap = 0; in __link_vmas() 77 struct vm_area_struct vmas[] = { in damon_test_three_regions_in_vmas() local 86 __link_vmas(vmas, 6); in damon_test_three_regions_in_vmas() [all …]
|
/linux/mm/ |
A D | vmacache.c | 38 current->vmacache.vmas[VMACACHE_HASH(addr)] = newvma; in vmacache_update() 72 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find() 105 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find_exact()
|
A D | gup.c | 1192 if (vmas) { in __get_user_pages() 1193 vmas[i] = vma; in __get_user_pages() 1313 struct vm_area_struct **vmas, in __get_user_pages_locked() argument 1322 BUG_ON(vmas); in __get_user_pages_locked() 1346 vmas, locked); in __get_user_pages_locked() 1649 if (vmas) in __get_user_pages_locked() 1650 vmas[i] = vma; in __get_user_pages_locked() 2063 pages, vmas, locked); in get_user_pages_remote() 2103 struct vm_area_struct **vmas) in get_user_pages() argument 3062 pages, vmas, locked); in pin_user_pages_remote() [all …]
|
A D | hugetlb.c | 5922 struct vm_area_struct **vmas) in record_subpages_vmas() argument 5929 if (vmas) in record_subpages_vmas() 5930 vmas[nr] = vma; in record_subpages_vmas() 5935 struct page **pages, struct vm_area_struct **vmas, in follow_hugetlb_page() argument 6053 if (!pages && !vmas && !pfn_offset && in follow_hugetlb_page() 6067 if (pages || vmas) in follow_hugetlb_page() 6071 vmas ? vmas + i : NULL); in follow_hugetlb_page()
|
A D | nommu.c | 626 if (curr->vmacache.vmas[i] == vma) { in delete_vma_from_mm()
|
/linux/include/linux/ |
A D | vmacache.h | 10 memset(tsk->vmacache.vmas, 0, sizeof(tsk->vmacache.vmas)); in vmacache_flush()
|
A D | mm_types_task.h | 36 struct vm_area_struct *vmas[VMACACHE_SIZE]; member
|
A D | mm.h | 1939 struct vm_area_struct **vmas, int *locked); 1943 struct vm_area_struct **vmas, int *locked); 1946 struct vm_area_struct **vmas); 1949 struct vm_area_struct **vmas);
|
A D | hugetlb.h | 253 struct vm_area_struct **vmas, unsigned long *position, in follow_hugetlb_page() argument
|
/linux/drivers/gpu/drm/msm/ |
A D | msm_gem.c | 312 list_add_tail(&vma->list, &msm_obj->vmas); in add_vma() 325 list_for_each_entry(vma, &msm_obj->vmas, list) { in lookup_vma() 356 list_for_each_entry(vma, &msm_obj->vmas, list) { in put_iova_spaces() 374 list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) { in put_iova_vmas() 929 if (!list_empty(&msm_obj->vmas)) { in msm_gem_describe() 933 list_for_each_entry(vma, &msm_obj->vmas, list) { in msm_gem_describe() 1136 INIT_LIST_HEAD(&msm_obj->vmas); in msm_gem_new_impl()
|
A D | msm_gem.h | 95 struct list_head vmas; /* list of msm_gem_vma */ member
|
/linux/drivers/video/fbdev/vermilion/ |
A D | vermilion.h | 210 atomic_t vmas; member
|
/linux/kernel/debug/ |
A D | debug_core.c | 289 if (!current->vmacache.vmas[i]) in kgdb_flush_swbreak_addr() 291 flush_cache_range(current->vmacache.vmas[i], in kgdb_flush_swbreak_addr()
|
/linux/Documentation/locking/ |
A D | robust-futexes.rst | 58 FUTEX_RECOVER. At do_exit() time, all vmas are searched to see whether 69 microsecond on Linux, but with thousands (or tens of thousands) of vmas
|
/linux/Documentation/admin-guide/mm/ |
A D | userfaultfd.rst | 35 operations never involve heavyweight structures like vmas (in fact the 40 Terabytes. Too many vmas would be needed for that.
|
/linux/fs/ |
A D | io_uring.c | 9031 struct vm_area_struct **vmas = NULL; in io_sqe_buffer_register() local 9054 vmas = kvmalloc_array(nr_pages, sizeof(struct vm_area_struct *), in io_sqe_buffer_register() 9056 if (!vmas) in io_sqe_buffer_register() 9066 pages, vmas); in io_sqe_buffer_register() 9070 struct vm_area_struct *vma = vmas[i]; in io_sqe_buffer_register() 9122 kvfree(vmas); in io_sqe_buffer_register()
|