/xen/tools/libxc/ |
A D | xc_resource.c | 26 DECLARE_NAMED_HYPERCALL_BOUNCE(entries, op->entries, in xc_resource_op_one() 30 if ( xc_hypercall_bounce_pre(xch, entries) ) in xc_resource_op_one() 36 set_xen_guest_handle(platform_op.u.resource_op.entries, entries); in xc_resource_op_one() 41 xc_hypercall_bounce_post(xch, entries); in xc_resource_op_one() 76 DECLARE_HYPERCALL_BUFFER(xc_resource_entry_t, entries); in xc_resource_op_multi() 90 entries, entries_size); in xc_resource_op_multi() 91 if ( !entries) in xc_resource_op_multi() 96 memcpy(entries, op->entries, entries_size); in xc_resource_op_multi() 106 set_xen_guest_handle(platform_op->u.resource_op.entries, entries); in xc_resource_op_multi() 121 entries, entries_size); in xc_resource_op_multi() [all …]
|
A D | xc_psr.c | 166 xc_resource_entry_t entries[3]; in xc_psr_cmt_get_data() local 189 entries[nr].rsvd = 0; in xc_psr_cmt_get_data() 193 entries[nr].idx = MSR_IA32_CMT_CTR; in xc_psr_cmt_get_data() 194 entries[nr].val = 0; in xc_psr_cmt_get_data() 195 entries[nr].rsvd = 0; in xc_psr_cmt_get_data() 200 tsc_entry = &entries[nr]; in xc_psr_cmt_get_data() 203 entries[nr].val = 0; in xc_psr_cmt_get_data() 204 entries[nr].rsvd = 0; in xc_psr_cmt_get_data() 208 assert(nr <= ARRAY_SIZE(entries)); in xc_psr_cmt_get_data() 212 op.entries = entries; in xc_psr_cmt_get_data() [all …]
|
A D | xc_offline_page.c | 44 struct pte_backup_entry *entries; member 185 backup->entries = realloc(backup->entries, in backup_ptes() 187 if (backup->entries == NULL) in backup_ptes() 403 privcmd_mmap_entry_t *entries; in xc_map_m2p() local 427 if (entries == NULL) in xc_map_m2p() 434 entries[i].mfn = extent_start[i]; in xc_map_m2p() 438 entries, m2p_chunks); in xc_map_m2p() 446 *mfn0 = entries[0].mfn; in xc_map_m2p() 449 free(entries); in xc_map_m2p() 553 if (!old_ptes.entries) in xc_exchange_page() [all …]
|
A D | xc_dom_hvmloader.c | 189 privcmd_mmap_entry_t *entries = NULL; in xc_dom_load_hvm_kernel() local 195 entries = calloc(pages, sizeof(privcmd_mmap_entry_t)); in xc_dom_load_hvm_kernel() 196 if ( entries == NULL ) in xc_dom_load_hvm_kernel() 200 entries[i].mfn = (elf->pstart >> PAGE_SHIFT) + i; in xc_dom_load_hvm_kernel() 205 entries, pages); in xc_dom_load_hvm_kernel() 233 free(entries); in xc_dom_load_hvm_kernel() 238 free(entries); in xc_dom_load_hvm_kernel()
|
A D | xc_sr_common_x86_pv.c | 101 privcmd_mmap_entry_t *entries = NULL; in x86_pv_map_m2p() local 129 entries = malloc(m2p_chunks * sizeof(privcmd_mmap_entry_t)); in x86_pv_map_m2p() 130 if ( !entries ) in x86_pv_map_m2p() 138 entries[i].mfn = extents_start[i]; in x86_pv_map_m2p() 142 M2P_CHUNK_SIZE, entries, m2p_chunks); in x86_pv_map_m2p() 154 ctx->x86.pv.compat_m2p_mfn0 = entries[0].mfn; in x86_pv_map_m2p() 179 free(entries); in x86_pv_map_m2p()
|
A D | xc_sr_save_x86_hvm.c | 82 struct xc_sr_rec_hvm_params_entry entries[ARRAY_SIZE(params)]; in write_hvm_params() local 108 entries[hdr.count].index = index; in write_hvm_params() 109 entries[hdr.count].value = value; in write_hvm_params() 118 rc = write_split_record(ctx, &rec, entries, hdr.count * sizeof(*entries)); in write_hvm_params()
|
A D | xc_foreign_memory.c | 59 privcmd_mmap_entry_t entries[], in xc_map_foreign_ranges() argument 77 arr[i * num_per_entry + j] = entries[i].mfn + j; in xc_map_foreign_ranges()
|
A D | xc_dom_boot.c | 133 privcmd_mmap_entry_t *entries; in xc_dom_boot_domU_map() local 138 entries = xc_dom_malloc(dom, count * sizeof(privcmd_mmap_entry_t)); in xc_dom_boot_domU_map() 139 if ( entries == NULL ) in xc_dom_boot_domU_map() 148 entries[i].mfn = xc_dom_p2m(dom, pfn + i); in xc_dom_boot_domU_map() 152 entries, count); in xc_dom_boot_domU_map()
|
/xen/xen/arch/x86/pv/ |
A D | descriptor-tables.c | 85 unsigned int entries) in pv_set_gdt() argument 89 unsigned int i, nr_frames = DIV_ROUND_UP(entries, 512); in pv_set_gdt() 93 if ( entries > FIRST_RESERVED_GDT_ENTRY ) in pv_set_gdt() 110 v->arch.pv.gdt_ents = entries; in pv_set_gdt() 128 unsigned int entries) in do_set_gdt() argument 130 unsigned int nr_frames = DIV_ROUND_UP(entries, 512); in do_set_gdt() 136 if ( entries > FIRST_RESERVED_GDT_ENTRY ) in do_set_gdt() 144 if ( (ret = pv_set_gdt(curr, frames, entries)) == 0 ) in do_set_gdt() 153 unsigned int entries) in compat_set_gdt() argument 161 if ( entries > FIRST_RESERVED_GDT_ENTRY ) in compat_set_gdt() [all …]
|
/xen/xen/arch/x86/ |
A D | domain_page.c | 101 ASSERT(idx < dcache->entries); in map_domain_page() 121 idx = find_next_zero_bit(dcache->inuse, dcache->entries, dcache->cursor); in map_domain_page() 122 if ( unlikely(idx >= dcache->entries) ) in map_domain_page() 127 for ( i = 0; i < BITS_TO_LONGS(dcache->entries); i++ ) in map_domain_page() 134 if ( accum | (prev & BITMAP_LAST_WORD_MASK(dcache->entries)) ) in map_domain_page() 135 idx = find_first_zero_bit(dcache->inuse, dcache->entries); in map_domain_page() 155 BUG_ON(idx >= dcache->entries); in map_domain_page() 272 if ( ents > dcache->entries ) in mapcache_vcpu_init() 289 dcache->entries = ents; in mapcache_vcpu_init()
|
A D | platform_hypercall.c | 45 xenpf_resource_entry_t *entries; member 88 xenpf_resource_entry_t *entry = ra->entries + i; in check_resource_access() 132 xenpf_resource_entry_t *entry = ra->entries + i; in resource_access() 733 if ( !ra.entries ) in do_platform_op() 741 if ( copy_from_guest(ra.entries, guest_entries, ra.nr_entries) ) in do_platform_op() 743 xfree(ra.entries); in do_platform_op() 753 if ( __copy_to_guest(guest_entries, ra.entries, 1) ) in do_platform_op() 756 xfree(ra.entries); in do_platform_op() 763 xfree(ra.entries); in do_platform_op() 773 if ( __copy_to_guest(guest_entries, ra.entries, in do_platform_op() [all …]
|
/xen/xen/drivers/vpci/ |
A D | msix.c | 26 #define VMSIX_SIZE(num) offsetof(struct vpci_msix, entries[num]) 99 if ( msix->entries[i].masked || !msix->entries[i].updated || in control_write() 100 update_entry(&msix->entries[i], pdev, i) ) in control_write() 103 msix->entries[i].updated = false; in control_write() 115 rc = vpci_msix_arch_disable_entry(&msix->entries[i], pdev); in control_write() 130 msix->entries[i].updated = true; in control_write() 194 return &msix->entries[(addr - start) / PCI_MSIX_ENTRY_SIZE]; in get_entry() 478 pdev->vpci->msix->entries[i].masked = true; in init_msix() 479 vpci_msix_arch_init_entry(&pdev->vpci->msix->entries[i]); in init_msix()
|
/xen/xen/drivers/acpi/apei/ |
A D | apei-base.c | 58 u32 entries) in apei_exec_ctx_init() argument 66 ctx->entries = entries; in apei_exec_ctx_init() 174 for (i = 0; i < ctx->entries; i++) { in __apei_exec_run() 215 for (i = 0; i < ctx->entries; i++) { in apei_exec_for_each_entry() 255 ctx_unmap.entries = end; in apei_exec_pre_map_gars()
|
A D | apei-internal.h | 31 u32 entries; member 38 u32 entries);
|
/xen/xen/include/asm-x86/pv/ |
A D | mm.h | 29 unsigned int entries); 48 unsigned int entries) in pv_set_gdt() argument
|
/xen/xen/common/ |
A D | kimage.c | 434 kimage_entry_t *entries; in kimage_add_entry() local 444 entries = __map_domain_page(image->entry_page); in kimage_add_entry() 445 entries[image->next_entry] = page_to_maddr(page) | IND_INDIRECTION; in kimage_add_entry() 446 unmap_domain_page(entries); in kimage_add_entry() 452 entries = __map_domain_page(image->entry_page); in kimage_add_entry() 453 entries[image->next_entry] = entry; in kimage_add_entry() 455 unmap_domain_page(entries); in kimage_add_entry() 481 kimage_entry_t *entries; in kimage_terminate() local 483 entries = __map_domain_page(image->entry_page); in kimage_terminate() 484 entries[image->next_entry] = IND_DONE; in kimage_terminate() [all …]
|
/xen/docs/misc/ |
A D | grant-tables.txt | 38 1. Shared grant entries 39 2. Active grant entries 46 Shared grant entries 50 entries. The guest writes into these entries to create grant references. The 58 Active grant entries 62 entries for safety, and to reference count mappings. 93 active entries 110 Active entries are obtained by calling active_entry_acquire(gt, ref). 127 Active entries cannot be acquired while holding the maptrack lock. 128 Multiple active entries can be acquired while holding the grant table [all …]
|
/xen/docs/man/ |
A D | xenhypfs.1.pod | 12 It can be used to show the available entries, to show their contents and 21 List the available entries below I<path>. 36 Show all the entries of the file system as a tree.
|
/xen/xen/drivers/passthrough/amd/ |
A D | iommu-defs.h | 35 #define PTE_PER_TABLE_ALIGN(entries) \ argument 36 (((entries) + PTE_PER_TABLE_SIZE - 1) & PTE_PER_TABLE_MASK) 37 #define PTE_PER_TABLE_ALLOC(entries) \ argument 38 PAGE_SIZE * (PTE_PER_TABLE_ALIGN(entries) >> PTE_PER_TABLE_SHIFT)
|
/xen/tools/examples/ |
A D | xlexample.pvhlinux | 35 # A list of 'vifspec' entries as described in 40 # A list of `diskspec' entries as described in
|
A D | xlexample.pvlinux | 37 # A list of 'vifspec' entries as described in 42 # A list of `diskspec' entries as described in
|
A D | xlexample.hvm | 36 # A list of 'vifspec' entries as described in 41 # A list of `diskspec' entries as described in
|
/xen/xen/include/asm-x86/ |
A D | hypercall.h | 69 unsigned int entries); 174 XEN_GUEST_HANDLE_PARAM(uint) frame_list, unsigned int entries);
|
/xen/xen/tools/kconfig/ |
A D | mconf.c | 313 struct list_head entries; member 330 list_for_each_entry(sp, &trail, entries) { in set_subtitle() 369 list_for_each_entry(pos, data->head, entries) { in update_text() 429 list_add_tail(&stpart.entries, &trail); in search_conf() 456 list_for_each_entry_safe(pos, tmp, &head, entries) in search_conf() 656 list_add_tail(&stpart.entries, &trail); in conf()
|
/xen/xen/include/xen/ |
A D | vpci.h | 137 } entries[]; member 210 return entry - msix->entries; in vmsix_entry_nr()
|