/linux/mm/ |
A D | nommu.c | 467 if (region->vm_start < pregion->vm_start) in add_nommu_region() 469 else if (region->vm_start > pregion->vm_start) in add_nommu_region() 583 if (vma->vm_start < pvma->vm_start) in add_vma_to_mm() 585 else if (vma->vm_start > pvma->vm_start) { in add_vma_to_mm() 1014 vma->vm_start = region->vm_start; in do_mmap_private() 1040 region->vm_start = vma->vm_start = 0; in do_mmap_private() 1205 vma->vm_start = region->vm_start = addr; in do_mmap() 1372 region->vm_start = new->vm_start = addr; in split_vma() 1383 vma->vm_region->vm_start = vma->vm_start = addr; in split_vma() 1413 vma->vm_start = to; in shrink_vma() [all …]
|
A D | mmap.c | 343 vma->vm_start, prev); in browse_rb() 366 prev = vma->vm_start; in browse_rb() 626 if (vma->vm_start > end) in count_vma_pages_range() 879 vma->vm_start = start; in __vma_adjust() 1267 b->vm_pgoff == a->vm_pgoff + ((b->vm_start - a->vm_start) >> PAGE_SHIFT); in anon_vma_compatible() 1772 vma->vm_start = addr; in mmap_region() 1799 addr = vma->vm_start; in mmap_region() 2599 start = vma->vm_start; in find_extend_vma() 2720 new->vm_start = addr; in __split_vma() 3073 vma->vm_start = addr; in do_brk_flags() [all …]
|
A D | mlock.c | 512 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); in mlock_fixup() 521 if (start != vma->vm_start) { in mlock_fixup() 575 if (!vma || vma->vm_start > start) in apply_vma_lock_flags() 579 if (start > vma->vm_start) in apply_vma_lock_flags() 601 if (!vma || vma->vm_start != nstart) { in apply_vma_lock_flags() 632 if (start + len <= vma->vm_start) in count_mm_mlocked_page_nr() 635 if (start > vma->vm_start) in count_mm_mlocked_page_nr() 636 count -= (start - vma->vm_start); in count_mm_mlocked_page_nr() 638 count += start + len - vma->vm_start; in count_mm_mlocked_page_nr() 641 count += vma->vm_end - vma->vm_start; in count_mm_mlocked_page_nr() [all …]
|
A D | mremap.c | 595 if (vma->vm_start != old_addr) in move_vma() 660 excess = vma->vm_end - vma->vm_start - old_len; in move_vma() 661 if (old_addr > vma->vm_start && in move_vma() 690 if (new_vma != vma && vma->vm_start == old_addr && in move_vma() 758 pgoff = (addr - vma->vm_start) >> PAGE_SHIFT; in vma_to_resize() 853 ((addr - vma->vm_start) >> PAGE_SHIFT), in mremap_to() 874 if (vma->vm_next && vma->vm_next->vm_start < end) /* intersection */ in vma_expandable() 876 if (get_unmapped_area(NULL, vma->vm_start, end - vma->vm_start, in vma_expandable() 946 if (!vma || vma->vm_start > addr) { in SYSCALL_DEFINE5() 1022 if (vma_adjust(vma, vma->vm_start, addr + new_len, in SYSCALL_DEFINE5() [all …]
|
A D | msync.c | 73 if (start < vma->vm_start) { in SYSCALL_DEFINE3() 76 start = vma->vm_start; in SYSCALL_DEFINE3() 88 fstart = (start - vma->vm_start) + in SYSCALL_DEFINE3()
|
A D | madvise.c | 138 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); in madvise_behavior() 149 if (start != vma->vm_start) { in madvise_behavior() 299 offset = (loff_t)(start - vma->vm_start) in madvise_willneed() 719 range.start = max(vma->vm_start, start_addr); in madvise_free_single_vma() 723 if (range.end <= vma->vm_start) in madvise_free_single_vma() 787 if (start < vma->vm_start) { in madvise_dontneed_free() 847 if (!vma || start < vma->vm_start) in madvise_populate() 911 offset = (loff_t)(start - vma->vm_start) in madvise_remove() 1178 if (vma && start > vma->vm_start) in do_madvise() 1189 if (start < vma->vm_start) { in do_madvise() [all …]
|
A D | internal.h | 400 munlock_vma_pages_range(vma, vma->vm_start, vma->vm_end); in munlock_vma_pages_all() 439 address = vma->vm_start + in vma_address() 442 if (address < vma->vm_start || address >= vma->vm_end) in vma_address() 447 address = vma->vm_start; in vma_address() 467 address = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); in vma_address_end() 469 if (address < vma->vm_start || address > vma->vm_end) in vma_address_end()
|
A D | pagewalk.c | 453 } else if (start < vma->vm_start) { /* outside vma */ in walk_page_range() 455 next = min(end, vma->vm_start); in walk_page_range() 525 err = walk_page_test(vma->vm_start, vma->vm_end, &walk); in walk_page_vma() 530 return __walk_page_range(vma->vm_start, vma->vm_end, &walk); in walk_page_vma() 587 start_addr = ((cba - vba) << PAGE_SHIFT) + vma->vm_start; in walk_page_mapping() 588 end_addr = ((cea - vba) << PAGE_SHIFT) + vma->vm_start; in walk_page_mapping() 595 err = walk_page_test(vma->vm_start, vma->vm_end, &walk); in walk_page_mapping()
|
A D | mprotect.c | 464 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); in mprotect_fixup() 476 if (start != vma->vm_start) { in mprotect_fixup() 568 if (vma->vm_start >= end) in do_mprotect_pkey() 570 start = vma->vm_start; in do_mprotect_pkey() 575 if (vma->vm_start > start) in do_mprotect_pkey() 585 if (start > vma->vm_start) in do_mprotect_pkey() 651 if (!vma || vma->vm_start != nstart) { in do_mprotect_pkey()
|
/linux/drivers/gpu/drm/ |
A D | drm_vm.c | 79 vma->vm_start)) in drm_io_prot() 210 offset = vmf->address - vma->vm_start; in drm_vm_shm_fault() 240 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_shm_close() 314 offset = vmf->address - vma->vm_start; in drm_vm_dma_fault() 351 offset = vmf->address - vma->vm_start; in drm_vm_sg_fault() 395 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_open_locked() 421 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_close_locked() 559 if (map->size < vma->vm_end - vma->vm_start) in drm_mmap_locked() 598 if (io_remap_pfn_range(vma, vma->vm_start, in drm_mmap_locked() 600 vma->vm_end - vma->vm_start, in drm_mmap_locked() [all …]
|
/linux/mm/damon/ |
A D | vaddr-test.h | 35 gap = vmas[j].vm_start - vmas[j - 1].vm_end; in __link_vmas() 78 (struct vm_area_struct) {.vm_start = 10, .vm_end = 20}, in damon_test_three_regions_in_vmas() 79 (struct vm_area_struct) {.vm_start = 20, .vm_end = 25}, in damon_test_three_regions_in_vmas() 80 (struct vm_area_struct) {.vm_start = 200, .vm_end = 210}, in damon_test_three_regions_in_vmas() 81 (struct vm_area_struct) {.vm_start = 210, .vm_end = 220}, in damon_test_three_regions_in_vmas() 82 (struct vm_area_struct) {.vm_start = 300, .vm_end = 305}, in damon_test_three_regions_in_vmas() 83 (struct vm_area_struct) {.vm_start = 307, .vm_end = 330}, in damon_test_three_regions_in_vmas()
|
/linux/arch/parisc/mm/ |
A D | fault.c | 128 if (tree->vm_start > addr) { 134 if (prev->vm_next->vm_start > addr) 255 vma->vm_start, vma->vm_end); in show_signal_msg() 289 if (!vma || address < vma->vm_start) in do_page_fault() 359 address < vma->vm_start || address >= vma->vm_end) { in do_page_fault()
|
/linux/fs/proc/ |
A D | task_nommu.c | 37 size += region->vm_end - region->vm_start; in task_mem() 39 size = vma->vm_end - vma->vm_start; in task_mem() 92 vsize += vma->vm_end - vma->vm_start; in task_vsize() 114 size += region->vm_end - region->vm_start; in task_statm() 138 return vma->vm_start <= mm->start_stack && in is_stack() 167 vma->vm_start, in nommu_vma_show()
|
/linux/drivers/media/v4l2-core/ |
A D | videobuf-vmalloc.c | 56 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_open() 68 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_close() 227 buf->baddr = vma->vm_start; in __videobuf_mmap_mapper() 233 pages = PAGE_ALIGN(vma->vm_end - vma->vm_start); in __videobuf_mmap_mapper() 254 map, q, vma->vm_start, vma->vm_end, in __videobuf_mmap_mapper()
|
/linux/drivers/char/ |
A D | mspec.c | 75 unsigned long vm_start; /* Original (unsplit) base. */ member 114 last_index = (vdata->vm_end - vdata->vm_start) >> PAGE_SHIFT; in mspec_close() 202 vdata->vm_start = vma->vm_start; in mspec_mmap()
|
/linux/scripts/coccinelle/api/ |
A D | vma_pages.cocci | 22 * (vma->vm_end - vma->vm_start) >> PAGE_SHIFT 32 - ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) 44 (vma->vm_end@p - vma->vm_start) >> PAGE_SHIFT
|
/linux/arch/ia64/include/asm/ |
A D | fb.h | 13 if (efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) in fb_pgprotect()
|
/linux/arch/csky/include/asm/ |
A D | tlb.h | 11 flush_cache_range(vma, (vma)->vm_start, (vma)->vm_end); \ 17 flush_tlb_range(vma, (vma)->vm_start, (vma)->vm_end); \
|
/linux/drivers/soc/qcom/ |
A D | rmtfs_mem.c | 136 if (vma->vm_end - vma->vm_start > rmtfs_mem->size) { in qcom_rmtfs_mem_mmap() 139 vma->vm_end, vma->vm_start, in qcom_rmtfs_mem_mmap() 140 (vma->vm_end - vma->vm_start), &rmtfs_mem->size); in qcom_rmtfs_mem_mmap() 146 vma->vm_start, in qcom_rmtfs_mem_mmap() 148 vma->vm_end - vma->vm_start, in qcom_rmtfs_mem_mmap()
|
/linux/arch/arm64/kernel/ |
A D | vdso.c | 84 current->mm->context.vdso = (void *)new_vma->vm_start; in vdso_mremap() 143 unsigned long size = vma->vm_end - vma->vm_start; in vdso_join_timens() 146 zap_page_range(vma, vma->vm_start, size); in vdso_join_timens() 149 zap_page_range(vma, vma->vm_start, size); in vdso_join_timens() 277 current->mm->context.sigpage = (void *)new_vma->vm_start; in aarch32_sigpage_mremap()
|
/linux/arch/x86/um/ |
A D | mem_32.c | 17 gate_vma.vm_start = FIXADDR_USER_START; in gate_vma_init() 49 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
|
/linux/arch/arc/kernel/ |
A D | arc_hostlink.c | 22 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in arc_hl_mmap() 23 vma->vm_end - vma->vm_start, in arc_hl_mmap()
|
/linux/arch/powerpc/platforms/book3s/ |
A D | vas-api.c | 364 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { in coproc_mmap() 366 (vma->vm_end - vma->vm_start), PAGE_SIZE); in coproc_mmap() 395 rc = remap_pfn_range(vma, vma->vm_start, pfn + vma->vm_pgoff, in coproc_mmap() 396 vma->vm_end - vma->vm_start, prot); in coproc_mmap() 399 paste_addr, vma->vm_start, rc); in coproc_mmap()
|
/linux/arch/riscv/kvm/ |
A D | mmu.c | 517 hva_t vm_start, vm_end; in kvm_arch_prepare_memory_region() local 519 if (!vma || vma->vm_start >= reg_end) in kvm_arch_prepare_memory_region() 532 vm_start = max(hva, vma->vm_start); in kvm_arch_prepare_memory_region() 537 (vm_start - mem->userspace_addr); in kvm_arch_prepare_memory_region() 541 pa += vm_start - vma->vm_start; in kvm_arch_prepare_memory_region() 550 vm_end - vm_start, writable); in kvm_arch_prepare_memory_region()
|
/linux/fs/ |
A D | userfaultfd.c | 1332 if (vma->vm_start >= end) in userfaultfd_register() 1379 end > cur->vm_start) { in userfaultfd_register() 1411 if (vma->vm_start < start) in userfaultfd_register() 1431 if (vma->vm_start > start) in userfaultfd_register() 1432 start = vma->vm_start; in userfaultfd_register() 1444 if (vma->vm_start < start) { in userfaultfd_register() 1538 if (vma->vm_start >= end) in userfaultfd_unregister() 1577 if (vma->vm_start < start) in userfaultfd_unregister() 1595 if (vma->vm_start > start) in userfaultfd_unregister() 1596 start = vma->vm_start; in userfaultfd_unregister() [all …]
|