Home
last modified time | relevance | path

Searched refs:order (Results 1 – 25 of 131) sorted by relevance

123456

/xen/tools/libvchan/
A Dinit.c77 int pages_left = ctrl->read.order >= PAGE_SHIFT ? 1 << (ctrl->read.order - PAGE_SHIFT) : 0; in init_gnt_srv()
100 switch (ctrl->read.order) { in init_gnt_srv()
114 switch (ctrl->write.order) { in init_gnt_srv()
137 ctrl->write.order = ctrl->read.order = 0; in init_gnt_srv()
157 if (ctrl->write.order < SMALL_RING_SHIFT || ctrl->write.order > MAX_RING_SHIFT) in init_gnt_cli()
159 if (ctrl->read.order < SMALL_RING_SHIFT || ctrl->read.order > MAX_RING_SHIFT) in init_gnt_cli()
161 if (ctrl->read.order == ctrl->write.order && ctrl->read.order < PAGE_SHIFT) in init_gnt_cli()
166 switch (ctrl->write.order) { in init_gnt_cli()
184 switch (ctrl->read.order) { in init_gnt_cli()
211 ctrl->write.order = ctrl->read.order = 0; in init_gnt_cli()
[all …]
A Dio.c86 return (1 << ctrl->write.order); in wr_ring_size()
91 return (1 << ctrl->read.order); in rd_ring_size()
362 if (ctrl->read.order >= PAGE_SHIFT) in libxenvchan_close()
363 munmap(ctrl->read.buffer, 1 << ctrl->read.order); in libxenvchan_close()
364 if (ctrl->write.order >= PAGE_SHIFT) in libxenvchan_close()
365 munmap(ctrl->write.buffer, 1 << ctrl->write.order); in libxenvchan_close()
/xen/xen/arch/x86/mm/
A Dp2m-pod.c71 __func__, mfn_x(mfn), order, ((1UL << order) - 1)); in p2m_pod_cache_add()
112 switch ( order ) in p2m_pod_cache_add()
168 switch ( order ) in p2m_pod_cache_get()
209 int order; in p2m_pod_set_cache_target() local
250 unsigned int order; in p2m_pod_set_cache_target() local
566 ret = 1UL << order; in p2m_pod_decrease_reservation()
580 if ( steal_for_cache && order < SUPERPAGE_ORDER && ram == (1UL << order) && in p2m_pod_decrease_reservation()
583 pod = 1UL << order; in p2m_pod_decrease_reservation()
826 t.order = 9; in p2m_pod_zero_check_superpage()
1127 gfn_t gfn_aligned = _gfn((gfn_x(gfn) >> order) << order); in p2m_pod_demand_populate()
[all …]
A Dp2m.c622 unsigned int order; in p2m_set_entry() local
641 order = 0; in p2m_set_entry()
650 todo -= 1ul << order; in p2m_set_entry()
1279 gfn_lock(p2m, gfn, order); in set_typed_p2m_entry()
1281 if ( cur_order < order ) in set_typed_p2m_entry()
1405 gfn_lock(p2m, gfn, order); in clear_mmio_p2m_entry()
1407 if ( cur_order < order ) in clear_mmio_p2m_entry()
1899 unsigned int iter, order; in map_mmio_regions() local
1909 order = ret - 1 ) in map_mmio_regions()
1931 unsigned int iter, order; in unmap_mmio_regions() local
[all …]
A Dp2m-ept.c675 unsigned int i, target = order / EPT_TABLE_ORDER; in ept_set_entry()
698 if ( (fn_mask & ((1UL << order) - 1)) || in ept_set_entry()
700 (order % EPT_TABLE_ORDER) ) in ept_set_entry()
832 (gfn + (1UL << order) - 1 > p2m->max_mapped_pfn) ) in ept_set_entry()
834 p2m->max_mapped_pfn = gfn + (1UL << order) - 1; in ept_set_entry()
846 rc = iommu_iotlb_flush(d, _dfn(gfn), (1u << order), in ept_set_entry()
853 iommu_legacy_unmap(d, _dfn(gfn), order); in ept_set_entry()
1316 int order; in ept_dump_p2m_table() local
1352 order = i * EPT_TABLE_ORDER; in ept_dump_p2m_table()
1353 ept_entry = table + (gfn_remainder >> order); in ept_dump_p2m_table()
[all …]
/xen/xen/drivers/passthrough/amd/
A Diommu_cmd.c107 ASSERT ( order == 0 || order == 9 || order == 18 ); in invalidate_iommu_pages()
110 if ( order || (io_addr == INV_IOMMU_ALL_PAGES_ADDRESS ) ) in invalidate_iommu_pages()
118 if ( order ) in invalidate_iommu_pages()
120 u64 mask = 1ULL << (order - 1 + PAGE_SHIFT); in invalidate_iommu_pages()
164 ASSERT ( order == 0 || order == 9 || order == 18 ); in invalidate_iotlb_pages()
166 if ( order || (io_addr == INV_IOMMU_ALL_PAGES_ADDRESS ) ) in invalidate_iotlb_pages()
171 if ( order ) in invalidate_iotlb_pages()
173 u64 mask = 1ULL << (order - 1 + PAGE_SHIFT); in invalidate_iotlb_pages()
313 unsigned int order) in amd_iommu_flush_all_iotlbs() argument
350 amd_iommu_flush_all_iotlbs(d, daddr, order); in _amd_iommu_flush_pages()
[all …]
A Diommu.h255 unsigned int order);
257 uint64_t gaddr, unsigned int order);
375 static inline void *__alloc_amd_iommu_tables(unsigned int order) in __alloc_amd_iommu_tables() argument
377 return alloc_xenheap_pages(order, 0); in __alloc_amd_iommu_tables()
380 static inline void __free_amd_iommu_tables(void *table, unsigned int order) in __free_amd_iommu_tables() argument
382 free_xenheap_pages(table, order); in __free_amd_iommu_tables()
/xen/xen/arch/arm/arm32/lib/
A Dlib1funcs.S107 rsb \order, \order, #31
118 addhs \order, \order, #8
122 addhs \order, \order, #4
125 addhi \order, \order, #3
126 addls \order, \order, \divisor, lsr #1
139 sub \order, \order, \spare
144 mov \order, #0
153 addlo \order, \order, #4
161 addlo \order, \order, #1
168 subs \order, \order, #3 @ yes, 3 is intended here
[all …]
/xen/xen/common/
A Dpage_alloc.c450 #define heap(node, zone, order) ((*_heap[node])[zone][order]) argument
652 unsigned int order; in setup_low_mem_virq() local
691 order--; in setup_low_mem_virq()
739 PFN_ORDER(pg) = order; in page_list_add_scrub()
1366 } while ( order-- != 0 ); in scrub_free_pages()
1386 ASSERT(order <= MAX_ORDER); in free_heap_pages()
1453 while ( order < MAX_ORDER ) in free_heap_pages()
1455 mask = 1UL << order; in free_heap_pages()
1502 order++; in free_heap_pages()
2255 unsigned int order, in assign_pages() argument
[all …]
A Dbitmap.c270 int bitmap_find_free_region(unsigned long *bitmap, int bits, int order) in bitmap_find_free_region() argument
273 int pages = 1 << order; in bitmap_find_free_region()
306 void bitmap_release_region(unsigned long *bitmap, int pos, int order) in bitmap_release_region() argument
308 int pages = 1 << order; in bitmap_release_region()
317 int bitmap_allocate_region(unsigned long *bitmap, int pos, int order) in bitmap_allocate_region() argument
319 int pages = 1 << order; in bitmap_allocate_region()
/xen/xen/arch/x86/guest/hyperv/
A Dtlb.c39 unsigned int order) in fill_gva_list() argument
43 unsigned long end = cur + (PAGE_SIZE << order); in fill_gva_list()
77 unsigned int max_gvas, order = (flags - 1) & FLUSH_ORDER_MASK; in flush_tlb_ex() local
107 if ( !va || (PAGE_SIZE << order) / HV_TLB_FLUSH_UNIT > max_gvas ) in flush_tlb_ex()
119 fill_gva_list(gva_list, va, order), in flush_tlb_ex()
131 unsigned int order = (flags - 1) & FLUSH_ORDER_MASK; in hyperv_flush_tlb() local
186 if ( !va || (PAGE_SIZE << order) / HV_TLB_FLUSH_UNIT > MAX_GVAS ) in hyperv_flush_tlb()
191 fill_gva_list(flush->gva_list, va, order), in hyperv_flush_tlb()
/xen/xen/include/xen/
A Dbitops.h133 int order; in get_bitmask_order() local
135 order = fls(count); in get_bitmask_order()
136 return order; /* We could be slightly more clever with -1 here... */ in get_bitmask_order()
141 int order; in get_count_order() local
143 order = fls(count) - 1; in get_count_order()
145 order++; in get_count_order()
146 return order; in get_count_order()
A Dmm.h165 void *alloc_xenheap_pages(unsigned int order, unsigned int memflags);
166 void free_xenheap_pages(void *v, unsigned int order);
201 struct domain *d, unsigned int order, unsigned int memflags);
202 void free_domheap_pages(struct page_info *pg, unsigned int order);
218 unsigned int order,
568 unsigned int order; in get_order_from_bytes() local
571 for ( order = 0; size; order++ ) in get_order_from_bytes()
574 return order; in get_order_from_bytes()
579 unsigned int order; in get_order_from_pages() local
582 for ( order = 0; nr_pages; order++ ) in get_order_from_pages()
[all …]
/xen/xen/include/asm-x86/pv/
A Dshim.h42 void pv_shim_online_memory(unsigned int nr, unsigned int order);
43 void pv_shim_offline_memory(unsigned int nr, unsigned int order);
79 static inline void pv_shim_online_memory(unsigned int nr, unsigned int order) in pv_shim_online_memory() argument
83 static inline void pv_shim_offline_memory(unsigned int nr, unsigned int order) in pv_shim_offline_memory() argument
/xen/xen/arch/x86/hvm/
A Dmtrr.c202 uint64_t mask = -(uint64_t)PAGE_SIZE << order; in get_mtrr_type()
217 if ( order > 4 ) in get_mtrr_type()
225 if ( order > 2 ) in get_mtrr_type()
235 if ( order ) in get_mtrr_type()
255 if ( unlikely(m->overlapped) || order ) in get_mtrr_type()
277 if ( order ) in get_mtrr_type()
559 unsigned int order) in hvm_get_mem_pinned_cacheattr() argument
562 uint64_t mask = ~(uint64_t)0 << order; in hvm_get_mem_pinned_cacheattr()
822 if ( order ) in epte_get_entry_emt()
840 for ( i = 0; i < (1ul << order); i++ ) in epte_get_entry_emt()
[all …]
A Ddom0_build.c109 unsigned int order; in pvh_populate_memory_range() member
123 unsigned int order, j; in pvh_populate_memory_range() local
155 order = min(order ? order - 1 : 0, max_order); in pvh_populate_memory_range()
157 order = min(order, start ? find_first_set_bit(start) : MAX_ORDER); in pvh_populate_memory_range()
161 if ( order == 0 && dom0_memflags ) in pvh_populate_memory_range()
168 if ( order == 0 ) in pvh_populate_memory_range()
173 max_order = order - 1; in pvh_populate_memory_range()
178 order); in pvh_populate_memory_range()
185 start += 1UL << order; in pvh_populate_memory_range()
186 nr_pages -= 1UL << order; in pvh_populate_memory_range()
[all …]
A Dnestedhvm.c135 unsigned int i, order = get_order_from_pages(nr); in nestedhvm_setup() local
152 shadow_io_bitmap[i] = alloc_xenheap_pages(order, 0); in nestedhvm_setup()
157 free_xenheap_pages(shadow_io_bitmap[i], order); in nestedhvm_setup()
/xen/xen/arch/x86/pv/
A Ddom0_build.c241 if ( order > last_order ) in alloc_chunk()
242 order = last_order; in alloc_chunk()
244 --order; in alloc_chunk()
247 if ( order-- == 0 ) in alloc_chunk()
250 last_order = order; in alloc_chunk()
266 for ( free_order = order; !dom0_memflags && page && order--; ) in alloc_chunk()
277 free_order = order; in alloc_chunk()
280 free_domheap_pages(pg2, order); in alloc_chunk()
291 int i, rc, compatible, order, machine; in dom0_construct_pv() local
493 order = get_order_from_bytes(count); in dom0_construct_pv()
[all …]
A Dshim.c896 .extent_order = order, in batch_memory_op()
920 void pv_shim_online_memory(unsigned int nr, unsigned int order) in pv_shim_online_memory() argument
929 if ( page->v.free.order != order ) in pv_shim_online_memory()
942 nr, order); in pv_shim_online_memory()
944 nr = batch_memory_op(XENMEM_populate_physmap, order, &list); in pv_shim_online_memory()
948 free_domheap_pages(page, order); in pv_shim_online_memory()
961 void pv_shim_offline_memory(unsigned int nr, unsigned int order) in pv_shim_offline_memory() argument
968 page = alloc_domheap_pages(NULL, order, 0); in pv_shim_offline_memory()
973 page->v.free.order = order; in pv_shim_offline_memory()
979 nr + 1, order); in pv_shim_offline_memory()
[all …]
/xen/xen/arch/x86/
A Dflushtlb.c191 unsigned int order = (flags - 1) & FLUSH_ORDER_MASK; in flush_area_local() local
195 if ( order == 0 ) in flush_area_local()
239 if ( order < (BITS_PER_LONG - PAGE_SHIFT) ) in flush_area_local()
240 sz = 1UL << (order + PAGE_SHIFT); in flush_area_local()
/xen/xen/arch/arm/
A Ddomain_build.c117 size = pfn_to_paddr(1UL << order); in insert_11_bank()
124 order); in insert_11_bank()
197 free_domheap_pages(pg, order); in insert_11_bank()
286 while ( order >= min_low_order ) in allocate_memory_11()
299 order--; in allocate_memory_11()
320 pg = alloc_domheap_pages(d, order, in allocate_memory_11()
324 order --; in allocate_memory_11()
333 if ( order >= min_order ) in allocate_memory_11()
401 order = min(max_order, order); in allocate_bank_memory()
410 if ( !order ) in allocate_bank_memory()
[all …]
A Dp2m.c1090 unsigned long order; in p2m_set_entry() local
1104 order = THIRD_ORDER; in p2m_set_entry()
1106 order = FIRST_ORDER; in p2m_set_entry()
1108 order = SECOND_ORDER; in p2m_set_entry()
1110 order = THIRD_ORDER; in p2m_set_entry()
1116 sgfn = gfn_add(sgfn, (1 << order)); in p2m_set_entry()
1118 smfn = mfn_add(smfn, (1 << order)); in p2m_set_entry()
1120 nr -= (1 << order); in p2m_set_entry()
1573 unsigned int order; in relinquish_p2m_mapping() local
1582 start = gfn_next_boundary(start, order) ) in relinquish_p2m_mapping()
[all …]
A Dmem_access.c360 unsigned int order; in p2m_set_mem_access() local
406 gfn = gfn_next_boundary(gfn, order) ) in p2m_set_mem_access()
409 mfn_t mfn = p2m_get_entry(p2m, gfn, &t, NULL, &order, NULL); in p2m_set_mem_access()
414 order = 0; in p2m_set_mem_access()
420 start += gfn_x(gfn_next_boundary(gfn, order)) - gfn_x(gfn); in p2m_set_mem_access()
/xen/docs/misc/
A Dpvh.pandoc4 native entry point, a `ELFNOTE` is used in order to tell the domain
40 Other relevant information needed in order to boot a guest kernel
50 The following VCPU hypercalls can be used in order to bring up secondary vCPUs:
81 done in order to take advantage of new hardware assisted virtualization
83 devices will also have the necessary interrupt controllers in order to manage
102 this will have to be revisited when pci-passthrough is added to PVH in order to
/xen/xen/drivers/char/
A Dconsole.c429 unsigned int order; in dump_console_ring_key() local
436 order = get_order_from_bytes(conring_size + 1); in dump_console_ring_key()
437 buf = alloc_xenheap_pages(order, 0); in dump_console_ring_key()
460 free_xenheap_pages(buf, order); in dump_console_ring_key()
1026 unsigned int i, order, memflags; in console_init_ring() local
1032 order = get_order_from_bytes(max(opt_conring_size, conring_size)); in console_init_ring()
1034 while ( (ring = alloc_xenheap_pages(order, memflags)) == NULL ) in console_init_ring()
1036 BUG_ON(order == 0); in console_init_ring()
1037 order--; in console_init_ring()
1039 opt_conring_size = PAGE_SIZE << order; in console_init_ring()

Completed in 55 milliseconds

123456