Lines Matching refs:va_start
765 return (va->va_end - va->va_start); in va_size()
813 if (tmp->va_start <= addr) in find_vmap_area_exceed_addr()
832 if (addr < va->va_start) in __find_vmap_area()
882 if (va->va_start < tmp_va->va_end && in find_va_links()
883 va->va_end <= tmp_va->va_start) in find_va_links()
885 else if (va->va_end > tmp_va->va_start && in find_va_links()
886 va->va_start >= tmp_va->va_end) in find_va_links()
890 va->va_start, va->va_end, tmp_va->va_start, tmp_va->va_end); in find_va_links()
1107 if (sibling->va_start == va->va_end) { in merge_or_add_vmap_area()
1108 sibling->va_start = va->va_start; in merge_or_add_vmap_area()
1128 if (sibling->va_end == va->va_start) { in merge_or_add_vmap_area()
1174 if (va->va_start > vstart) in is_within_this_va()
1175 nva_start_addr = ALIGN(va->va_start, align); in is_within_this_va()
1206 vstart < va->va_start) { in find_vmap_lowest_match()
1234 vstart <= va->va_start) { in find_vmap_lowest_match()
1241 vstart = va->va_start + 1; in find_vmap_lowest_match()
1305 if (nva_start_addr < va->va_start || in classify_va_fit_type()
1310 if (va->va_start == nva_start_addr) { in classify_va_fit_type()
1349 va->va_start += size; in adjust_va_to_fit_type()
1402 lva->va_start = va->va_start; in adjust_va_to_fit_type()
1408 va->va_start = nva_start_addr + size; in adjust_va_to_fit_type()
1441 if (va->va_start > vstart) in __alloc_vmap_area()
1442 nva_start_addr = ALIGN(va->va_start, align); in __alloc_vmap_area()
1557 va->va_start = addr; in alloc_vmap_area()
1565 BUG_ON(!IS_ALIGNED(va->va_start, align)); in alloc_vmap_area()
1566 BUG_ON(va->va_start < vstart); in alloc_vmap_area()
1681 struct vmap_area, list)->va_start); in __purge_vmap_area_lazy()
1692 unsigned long nr = (va->va_end - va->va_start) >> PAGE_SHIFT; in __purge_vmap_area_lazy()
1693 unsigned long orig_start = va->va_start; in __purge_vmap_area_lazy()
1709 va->va_start, va->va_end); in __purge_vmap_area_lazy()
1756 nr_lazy = atomic_long_add_return((va->va_end - va->va_start) >> in free_vmap_area_noflush()
1777 flush_cache_vunmap(va->va_start, va->va_end); in free_unmap_vmap_area()
1778 vunmap_range_noflush(va->va_start, va->va_end); in free_unmap_vmap_area()
1780 flush_tlb_kernel_range(va->va_start, va->va_end); in free_unmap_vmap_area()
1865 static void *vmap_block_vaddr(unsigned long va_start, unsigned long pages_off) in vmap_block_vaddr() argument
1869 addr = va_start + (pages_off << PAGE_SHIFT); in vmap_block_vaddr()
1870 BUG_ON(addr_to_vb_idx(addr) != addr_to_vb_idx(va_start)); in vmap_block_vaddr()
1906 vaddr = vmap_block_vaddr(va->va_start, 0); in new_vmap_block()
1917 vb_idx = addr_to_vb_idx(va->va_start); in new_vmap_block()
1938 tmp = xa_erase(&vmap_blocks, addr_to_vb_idx(vb->va->va_start)); in free_vmap_block()
2019 vaddr = vmap_block_vaddr(vb->va->va_start, pages_off); in vb_alloc()
2093 unsigned long va_start = vb->va->va_start; in _vm_unmap_aliases() local
2096 s = va_start + (vb->dirty_min << PAGE_SHIFT); in _vm_unmap_aliases()
2097 e = va_start + (vb->dirty_max << PAGE_SHIFT); in _vm_unmap_aliases()
2165 debug_check_no_locks_freed((void *)va->va_start, in vm_unmap_ram()
2166 (va->va_end - va->va_start)); in vm_unmap_ram()
2203 addr = va->va_start; in vm_map_ram()
2310 if (busy->va_start - vmap_start > 0) { in vmap_init_free_space()
2313 free->va_start = vmap_start; in vmap_init_free_space()
2314 free->va_end = busy->va_start; in vmap_init_free_space()
2328 free->va_start = vmap_start; in vmap_init_free_space()
2367 va->va_start = (unsigned long)tmp->addr; in vmalloc_init()
2368 va->va_end = va->va_start + tmp->size; in vmalloc_init()
2384 vm->addr = (void *)va->va_start; in setup_vmalloc_vm_locked()
2385 vm->size = va->va_end - va->va_start; in setup_vmalloc_vm_locked()
2440 kasan_unpoison_vmalloc((void *)va->va_start, requested_size); in __get_vm_area_node()
3374 if ((unsigned long)addr + count <= va->va_start) in vread()
3540 if (tmp->va_start <= addr) { in pvm_find_va_enclose_addr()
3574 if ((*va)->va_start < addr) in pvm_determine_end_from_reverse()
3696 if (base + start < va->va_start) { in pcpu_get_vm_areas()
3739 va->va_start = start; in pcpu_get_vm_areas()
3747 if (kasan_populate_vmalloc(vas[area]->va_start, sizes[area])) in pcpu_get_vm_areas()
3750 kasan_unpoison_vmalloc((void *)vas[area]->va_start, in pcpu_get_vm_areas()
3775 orig_start = vas[area]->va_start; in pcpu_get_vm_areas()
3781 va->va_start, va->va_end); in pcpu_get_vm_areas()
3825 orig_start = vas[area]->va_start; in pcpu_get_vm_areas()
3831 va->va_start, va->va_end); in pcpu_get_vm_areas()
3928 (void *)va->va_start, (void *)va->va_end, in show_purge_info()
3929 va->va_end - va->va_start); in show_purge_info()
3947 (void *)va->va_start, (void *)va->va_end, in s_show()
3948 va->va_end - va->va_start); in s_show()