Searched refs:IS_ALIGNED (Results 1 – 15 of 15) sorted by relevance
/xen/xen/arch/x86/ |
A D | xen.lds.S | 370 ASSERT(IS_ALIGNED(_start, MB(2)), "_start misaligned") 372 ASSERT(IS_ALIGNED(__2M_text_end, SECTION_ALIGN), "__2M_text_end misaligned") 374 ASSERT(IS_ALIGNED(__2M_rodata_end, SECTION_ALIGN), "__2M_rodata_end misaligned") 376 ASSERT(IS_ALIGNED(__2M_init_end, SECTION_ALIGN), "__2M_init_end misaligned") 380 ASSERT(IS_ALIGNED(cpu0_stack, STACK_SIZE), "cpu0_stack misaligned") 382 ASSERT(IS_ALIGNED(__init_begin, PAGE_SIZE), "__init_begin misaligned") 383 ASSERT(IS_ALIGNED(__init_end, PAGE_SIZE), "__init_end misaligned") 385 ASSERT(IS_ALIGNED(trampoline_start, 4), "trampoline_start misaligned") 386 ASSERT(IS_ALIGNED(trampoline_end, 4), "trampoline_end misaligned") 387 ASSERT(IS_ALIGNED(__bss_start, 8), "__bss_start misaligned") [all …]
|
A D | mm.c | 5617 ASSERT(IS_ALIGNED(s, PAGE_SIZE)); in modify_xen_mappings() 5618 ASSERT(IS_ALIGNED(e, PAGE_SIZE)); in modify_xen_mappings() 6151 ASSERT(IS_ALIGNED(_p, PAGE_SIZE)); in __memguard_change_range() 6152 ASSERT(IS_ALIGNED(_l, PAGE_SIZE)); in __memguard_change_range()
|
/xen/xen/include/xen/ |
A D | bitmap.h | 248 IS_ALIGNED(start, BITMAP_MEM_ALIGNMENT) && in bitmap_set() 250 IS_ALIGNED(nbits, BITMAP_MEM_ALIGNMENT)) in bitmap_set() 262 IS_ALIGNED(start, BITMAP_MEM_ALIGNMENT) && in bitmap_clear() 264 IS_ALIGNED(nbits, BITMAP_MEM_ALIGNMENT)) in bitmap_clear()
|
A D | config.h | 85 #define IS_ALIGNED(val, align) (((val) & ((align) - 1)) == 0) macro
|
/xen/xen/arch/arm/ |
A D | xen.lds.S | 254 ASSERT(IS_ALIGNED(__init_begin, 4), "__init_begin is misaligned") 255 ASSERT(IS_ALIGNED(__init_end, 4), "__init_end is misaligned")
|
A D | mm.c | 1263 if ( !IS_ALIGNED(virt, PAGE_SIZE) ) in xen_pt_update()
|
/xen/xen/arch/x86/pv/ |
A D | grant_table.c | 68 if ( !IS_ALIGNED(addr, sizeof(nl1e)) ) in create_grant_pv_mapping() 225 if ( !IS_ALIGNED(addr, sizeof(nl1e)) ) in replace_grant_pv_mapping()
|
A D | descriptor-tables.c | 198 if ( !IS_ALIGNED(gaddr, sizeof(d)) || !check_descriptor(currd, &d) ) in do_update_descriptor()
|
/xen/xen/drivers/passthrough/ |
A D | iommu.c | 253 ASSERT(IS_ALIGNED(dfn_x(dfn), (1ul << page_order))); in iommu_map() 254 ASSERT(IS_ALIGNED(mfn_x(mfn), (1ul << page_order))); in iommu_map() 313 ASSERT(IS_ALIGNED(dfn_x(dfn), (1ul << page_order))); in iommu_unmap()
|
/xen/xen/arch/x86/hvm/ |
A D | dom0_build.c | 128 if ( IS_ALIGNED(start, orders[j].align) && in pvh_populate_memory_range() 516 ASSERT(IS_ALIGNED(start, PAGE_SIZE) && IS_ALIGNED(end, PAGE_SIZE)); in find_memory()
|
/xen/xen/common/ubsan/ |
A D | ubsan.c | 332 else if (data->alignment && !IS_ALIGNED(ptr, data->alignment)) in __ubsan_handle_type_mismatch()
|
/xen/xen/arch/arm/vgic/ |
A D | vgic-mmio.c | 523 if ( (region->access_flags & flags) && IS_ALIGNED(addr, len) ) in check_region()
|
/xen/xen/common/ |
A D | argo.c | 2109 BUILD_BUG_ON(!IS_ALIGNED(XEN_ARGO_MAX_RING_SIZE, PAGE_SIZE)); in do_argo_op()
|
/xen/xen/drivers/passthrough/vtd/ |
A D | iommu.c | 608 dfn_eq(dfn, INVALID_DFN) || !IS_ALIGNED(dfn_x(dfn), page_count) ) in iommu_flush_iotlb()
|
/xen/xen/arch/x86/hvm/vmx/ |
A D | vvmx.c | 805 if ( !IS_ALIGNED(gpa, PAGE_SIZE) ) in _map_msr_bitmap()
|
Completed in 45 milliseconds