| /xen/xen/include/asm-x86/ |
| A D | mpspec.h | 36 unsigned long mask[PHYSID_ARRAY_SIZE]; member 41 #define physid_set(physid, map) set_bit(physid, (map).mask) 46 #define first_physid(map) find_first_bit((map).mask, \ 48 #define next_physid(id, map) find_next_bit((map).mask, \ 51 const unsigned long *mask = (map).mask; \ 53 for (id = find_first_bit(mask, MAX_APICS); id < MAX_APICS; \ 54 id = find_next_bit(mask, MAX_APICS, (id) + 1)) \ 59 #define physids_and(dst, src1, src2) bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS) 60 #define physids_or(dst, src1, src2) bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS) 62 #define physids_complement(dst, src) bitmap_complement((dst).mask,(src).mask, MAX_APICS) [all …]
|
| A D | flushtlb.h | 77 static inline void tlbflush_filter(cpumask_t *mask, uint32_t page_timestamp) in tlbflush_filter() argument 85 for_each_cpu ( cpu, mask ) in tlbflush_filter() 87 __cpumask_clear_cpu(cpu, mask); in tlbflush_filter() 146 #define flush_mask(mask, flags) flush_area_mask(mask, NULL, flags) argument 159 #define flush_tlb_mask(mask) \ argument 160 flush_mask(mask, FLUSH_TLB) 161 #define flush_tlb_one_mask(mask,v) \ argument 162 flush_area_mask(mask, (const void *)(v), FLUSH_TLB|FLUSH_ORDER(0)) 169 #define arch_flush_tlb_mask(mask) flush_mask(mask, FLUSH_TLB | FLUSH_FORCE_IPI) argument 201 void guest_flush_tlb_mask(const struct domain *d, const cpumask_t *mask);
|
| A D | vpmu.h | 84 static inline void vpmu_set(struct vpmu_struct *vpmu, const u32 mask) in vpmu_set() argument 86 vpmu->flags |= mask; in vpmu_set() 88 static inline void vpmu_reset(struct vpmu_struct *vpmu, const u32 mask) in vpmu_reset() argument 90 vpmu->flags &= ~mask; in vpmu_reset() 97 static inline bool_t vpmu_is_set(const struct vpmu_struct *vpmu, const u32 mask) in vpmu_is_set() argument 99 return !!(vpmu->flags & mask); in vpmu_is_set() 102 const u32 mask) in vpmu_are_all_set() argument 104 return !!((vpmu->flags & mask) == mask); in vpmu_are_all_set()
|
| /xen/xen/arch/x86/ |
| A D | smp.c | 34 static unsigned int prepare_ICR2(unsigned int mask) in prepare_ICR2() argument 36 return SET_xAPIC_DEST_FIELD(mask); in prepare_ICR2() 174 mask &= cpumask_bits(&cpu_online_map)[0]; in send_IPI_mask_flat() 175 mask &= ~(1UL << smp_processor_id()); in send_IPI_mask_flat() 176 if ( mask == 0 ) in send_IPI_mask_flat() 189 cfg = prepare_ICR2(mask); in send_IPI_mask_flat() 212 for_each_cpu ( query_cpu, mask ) in send_IPI_mask_phys() 266 cpumask_test_cpu(cpu, mask) ) in flush_area_mask() 270 !cpumask_subset(mask, cpumask_of(cpu)) ) in flush_area_mask() 307 send_IPI_mask(mask, EVENT_CHECK_VECTOR); in smp_send_event_check_mask() [all …]
|
| A D | hypercall.c | 180 unsigned int mask, ...) in hypercall_xlat_continuation() argument 190 ASSERT(!(mask >> nr)); in hypercall_xlat_continuation() 192 ASSERT(!id || !(mask & (1U << *id))); in hypercall_xlat_continuation() 194 va_start(args, mask); in hypercall_xlat_continuation() 204 for ( i = 0; i < nr; ++i, mask >>= 1 ) in hypercall_xlat_continuation() 206 if ( mask & 1 ) in hypercall_xlat_continuation() 211 mask &= ~1U; in hypercall_xlat_continuation() 232 for ( i = 0; i < nr; ++i, mask >>= 1 ) in hypercall_xlat_continuation() 246 if ( (mask & 1) ) in hypercall_xlat_continuation() 251 mask &= ~1U; in hypercall_xlat_continuation() [all …]
|
| A D | xstate.c | 293 uint32_t hmask = mask >> 32; in xsave() 294 uint32_t lmask = mask; in xsave() 361 if ( mask & X86_XCR0_FP ) in xsave() 367 uint32_t hmask = mask >> 32; in xrstor() 368 uint32_t lmask = mask; in xrstor() 698 u64 mask; in handle_xsetbv() local 738 if ( mask ) in handle_xsetbv() 748 mask |= XSTATE_LAZY; in handle_xsetbv() 753 xrstor(curr, mask); in handle_xsetbv() 807 if ( (~xcr0 & mask) && !set_xcr0(xcr0 | mask) ) in xstate_set_init() [all …]
|
| A D | i8259.c | 98 unsigned int mask = 1 << irq; in _disable_8259A_irq() local 102 cached_irq_mask |= mask; in _disable_8259A_irq() 118 unsigned int mask = ~(1 << desc->irq); in enable_8259A_irq() local 122 cached_irq_mask &= mask; in enable_8259A_irq() 133 unsigned int mask = 1<<irq; in i8259A_irq_pending() local 139 ret = inb(0x20) & mask; in i8259A_irq_pending() 141 ret = inb(0xA0) & (mask >> 8); in i8259A_irq_pending()
|
| /xen/xen/arch/arm/arm64/ |
| A D | insn.c | 43 u32 mask; in aarch64_get_imm_shift_mask() local 48 mask = BIT(26, UL) - 1; in aarch64_get_imm_shift_mask() 52 mask = BIT(19, UL) - 1; in aarch64_get_imm_shift_mask() 56 mask = BIT(16, UL) - 1; in aarch64_get_imm_shift_mask() 60 mask = BIT(14, UL) - 1; in aarch64_get_imm_shift_mask() 68 mask = BIT(9, UL) - 1; in aarch64_get_imm_shift_mask() 72 mask = BIT(7, UL) - 1; in aarch64_get_imm_shift_mask() 77 mask = BIT(6, UL) - 1; in aarch64_get_imm_shift_mask() 81 mask = BIT(6, UL) - 1; in aarch64_get_imm_shift_mask() 88 *maskp = mask; in aarch64_get_imm_shift_mask() [all …]
|
| /xen/xen/arch/arm/ |
| A D | guest_walk.c | 38 paddr_t mask, paddr; in guest_walk_sd() local 46 if ( n == 0 || !(gva & mask) ) in guest_walk_sd() 91 mask = GENMASK(31, (14 - n)); in guest_walk_sd() 92 paddr = (ttbr & mask); in guest_walk_sd() 94 mask = GENMASK((31 - n), 20); in guest_walk_sd() 95 paddr |= (gva & mask) >> 18; in guest_walk_sd() 114 mask = GENMASK(19, 12); in guest_walk_sd() 165 *ipa = gva & mask; in guest_walk_sd() 369 paddr_t mask, paddr; in guest_walk_ld() local 477 if ( ((t1_sz == 0) && !ttbr) || (t1_sz && (gva & mask) == mask) ) in guest_walk_ld() [all …]
|
| A D | smp.c | 8 void arch_flush_tlb_mask(const cpumask_t *mask) in arch_flush_tlb_mask() argument 14 void smp_send_event_check_mask(const cpumask_t *mask) in smp_send_event_check_mask() argument 16 send_SGI_mask(mask, GIC_SGI_EVENT_CHECK); in smp_send_event_check_mask() 19 void smp_send_call_function_mask(const cpumask_t *mask) in smp_send_call_function_mask() argument 23 cpumask_andnot(&target_mask, mask, cpumask_of(smp_processor_id())); in smp_send_call_function_mask() 27 if ( cpumask_test_cpu(smp_processor_id(), mask) ) in smp_send_call_function_mask()
|
| /xen/xen/include/xen/ |
| A D | cpumask.h | 316 return *mask != NULL; in alloc_cpumask_var() 321 if (*mask == NULL) in cond_alloc_cpumask_var() 323 return *mask != NULL; in cond_alloc_cpumask_var() 329 return *mask != NULL; in zalloc_cpumask_var() 334 if (*mask == NULL) in cond_zalloc_cpumask_var() 337 cpumask_clear(*mask); in cond_zalloc_cpumask_var() 338 return *mask != NULL; in cond_zalloc_cpumask_var() 343 xfree(mask); in free_cpumask_var() 359 cpumask_clear(*mask); in zalloc_cpumask_var() 372 #define for_each_cpu(cpu, mask) \ argument [all …]
|
| /xen/tools/libxc/ |
| A D | xc_pagetab.c | 33 uint64_t paddr, mask, pte = 0; in xc_translate_foreign_address() local 72 mask = 0x0000ff8000000000ull; in xc_translate_foreign_address() 75 mask = 0x0000007fc0000000ull; in xc_translate_foreign_address() 78 mask = 0x00000000ffc00000ull; in xc_translate_foreign_address() 84 paddr += ((virt & mask) >> (xc_ffs64(mask) - 1)) * size; in xc_translate_foreign_address() 97 mask = ((mask ^ ~-mask) >> 1); /* All bits below first set bit */ in xc_translate_foreign_address() 98 return ((paddr & ~mask) | (virt & mask)) >> PAGE_SHIFT; in xc_translate_foreign_address() 100 mask >>= (pt_levels == 2 ? 10 : 9); in xc_translate_foreign_address()
|
| A D | xc_tbuf.c | 117 int xc_tbuf_set_cpu_mask(xc_interface *xch, xc_cpumap_t mask) in xc_tbuf_set_cpu_mask() argument 120 DECLARE_HYPERCALL_BOUNCE(mask, 0, XC_HYPERCALL_BUFFER_BOUNCE_IN); in xc_tbuf_set_cpu_mask() 131 HYPERCALL_BOUNCE_SET_SIZE(mask, cpusize); in xc_tbuf_set_cpu_mask() 140 if ( xc_hypercall_bounce_pre(xch, mask) ) in xc_tbuf_set_cpu_mask() 150 set_xen_guest_handle(sysctl.u.tbuf_op.cpu_mask.bitmap, mask); in xc_tbuf_set_cpu_mask() 155 xc_hypercall_bounce_post(xch, mask); in xc_tbuf_set_cpu_mask() 161 int xc_tbuf_set_evt_mask(xc_interface *xch, uint32_t mask) in xc_tbuf_set_evt_mask() argument 168 sysctl.u.tbuf_op.evt_mask = mask; in xc_tbuf_set_evt_mask()
|
| /xen/xen/include/asm-arm/ |
| A D | bitops.h | 54 void clear_mask16(uint16_t mask, volatile void *p); 74 bool clear_mask16_timeout(uint16_t mask, volatile void *p, 88 unsigned int mask = BITOP_MASK(nr); in __test_and_set_bit() local 93 *p = old | mask; in __test_and_set_bit() 94 return (old & mask) != 0; in __test_and_set_bit() 108 unsigned int mask = BITOP_MASK(nr); in __test_and_clear_bit() local 113 *p = old & ~mask; in __test_and_clear_bit() 114 return (old & mask) != 0; in __test_and_clear_bit() 121 unsigned int mask = BITOP_MASK(nr); in __test_and_change_bit() local 126 *p = old ^ mask; in __test_and_change_bit() [all …]
|
| A D | vreg.h | 136 unsigned long mask = VREG_REG_MASK(size); in vreg_reg_update() local 139 *reg &= ~(mask << shift); in vreg_reg_update() 140 *reg |= ((unsigned long)val & mask) << shift; in vreg_reg_update() 147 unsigned long mask = VREG_REG_MASK(size); in vreg_reg_setbits() local 150 *reg |= ((unsigned long)bits & mask) << shift; in vreg_reg_setbits() 157 unsigned long mask = VREG_REG_MASK(size); in vreg_reg_clearbits() local 160 *reg &= ~(((unsigned long)bits & mask) << shift); in vreg_reg_clearbits()
|
| /xen/xen/arch/arm/arm64/lib/ |
| A D | bitops.c | 36 const uint32_t mask = BITOP_MASK((unsigned int)nr); \ 46 : "r" (mask)); \ 72 const uint32_t mask = BITOP_MASK(bit); \ 83 : "r" (mask), "r" (bit) \ 121 static always_inline bool int_clear_mask16(uint16_t mask, volatile uint16_t *p, in bitop() 133 : "r" (mask)); in bitop() 142 void clear_mask16(uint16_t mask, volatile void *p) in clear_mask16() argument 144 if ( !int_clear_mask16(mask, p, false, 0) ) in clear_mask16() 148 bool clear_mask16_timeout(uint16_t mask, volatile void *p, in clear_mask16_timeout() argument 151 return int_clear_mask16(mask, p, true, max_try); in clear_mask16_timeout()
|
| A D | memcmp.S | 58 mask .req x13 label 93 mov mask, #~0 94 CPU_BE( lsr mask, mask, limit ) 95 CPU_LE( lsl mask, mask, limit ) 96 bic data1, data1, mask 97 bic data2, data2, mask 99 orr diff, diff, mask
|
| /xen/xen/common/ |
| A D | pdx.c | 45 static u64 __init fill_mask(u64 mask) in fill_mask() argument 47 while (mask & (mask + 1)) in fill_mask() 48 mask |= mask + 1; in fill_mask() 50 return mask; in fill_mask() 76 void __init pfn_pdx_hole_setup(unsigned long mask) in pfn_pdx_hole_setup() argument 91 i = find_next_zero_bit(&mask, BITS_PER_LONG, j + 1); in pfn_pdx_hole_setup() 94 j = find_next_bit(&mask, BITS_PER_LONG, i + 1); in pfn_pdx_hole_setup()
|
| A D | bitmap.c | 272 unsigned long mask; in bitmap_find_free_region() local 280 mask = (1ul << (pages - 1)); in bitmap_find_free_region() 281 mask += mask - 1; in bitmap_find_free_region() 289 bitmap[index] |= (mask << offset); in bitmap_find_free_region() 309 unsigned long mask = (1ul << (pages - 1)); in bitmap_release_region() local 312 mask += mask - 1; in bitmap_release_region() 313 bitmap[index] &= ~(mask << offset); in bitmap_release_region() 320 unsigned long mask = (1ul << (pages - 1)); in bitmap_allocate_region() local 329 mask += mask - 1; in bitmap_allocate_region() 330 if (bitmap[index] & (mask << offset)) in bitmap_allocate_region() [all …]
|
| A D | softirq.c | 83 void cpumask_raise_softirq(const cpumask_t *mask, unsigned int nr) in cpumask_raise_softirq() argument 96 for_each_cpu(cpu, mask) in cpumask_raise_softirq() 129 cpumask_t *mask = &per_cpu(batch_mask, this_cpu); in cpu_raise_softirq_batch_finish() local 132 for_each_cpu ( cpu, mask ) in cpu_raise_softirq_batch_finish() 134 __cpumask_clear_cpu(cpu, mask); in cpu_raise_softirq_batch_finish() 135 smp_send_event_check_mask(mask); in cpu_raise_softirq_batch_finish() 136 cpumask_clear(mask); in cpu_raise_softirq_batch_finish()
|
| /xen/xen/arch/x86/hvm/ |
| A D | vpic.c | 54 if ( mask == 0 ) in vpic_get_priority() 67 uint8_t mask; in vpic_get_highest_priority_irq() local 84 mask = vpic->isr; in vpic_get_highest_priority_irq() 86 mask &= ~(1 << 2); in vpic_get_highest_priority_irq() 88 mask &= ~vpic->imr; in vpic_get_highest_priority_irq() 140 uint8_t mask = 1 << irq; in __vpic_intack() local 146 if ( !(vpic->elcr & mask) ) in __vpic_intack() 147 vpic->irr &= ~mask; in __vpic_intack() 150 vpic->isr |= mask; in __vpic_intack() 462 vpic->irr |= mask; in vpic_irq_positive_edge() [all …]
|
| /xen/xen/arch/arm/arm32/lib/ |
| A D | bitops.c | 37 const uint32_t mask = BITOP_MASK((unsigned int)nr); \ 50 : "r" (mask)); \ 76 const uint32_t mask = BITOP_MASK(bit); \ 92 : "r" (mask), "r" (bit)); \ 129 static always_inline bool int_clear_mask16(uint16_t mask, volatile uint16_t *p, in bitop() 143 : "r" (mask)); in bitop() 152 void clear_mask16(uint16_t mask, volatile void *p) in clear_mask16() argument 154 if ( !int_clear_mask16(mask, p, false, 0) ) in clear_mask16() 158 bool clear_mask16_timeout(uint16_t mask, volatile void *p, in clear_mask16_timeout() argument 161 return int_clear_mask16(mask, p, true, max_try); in clear_mask16_timeout()
|
| /xen/xen/arch/x86/cpu/mtrr/ |
| A D | generic.c | 238 unsigned long mask = smp_changes_mask; in mtrr_state_warn() local 242 if (!mask) in mtrr_state_warn() 244 if (mask & MTRR_CHANGE_MASK_FIXED) in mtrr_state_warn() 246 if (mask & MTRR_CHANGE_MASK_VARIABLE) in mtrr_state_warn() 248 if (mask & MTRR_CHANGE_MASK_DEFTYPE) in mtrr_state_warn() 383 mask_lo = (uint32_t)vr->mask; in set_mtrr_var_ranges() 384 mask_hi = (uint32_t)(vr->mask >> 32); in set_mtrr_var_ranges() 504 unsigned long mask, count; in generic_set_all() local 512 mask = set_mtrr_state(); in generic_set_all() 519 if (mask & 0x01) in generic_set_all() [all …]
|
| /xen/xen/arch/x86/guest/hyperv/ |
| A D | tlb.c | 72 static uint64_t flush_tlb_ex(const cpumask_t *mask, const void *va, in flush_tlb_ex() argument 94 nr_banks = cpumask_to_vpset(&flush->hv_vp_set, mask); in flush_tlb_ex() 126 int hyperv_flush_tlb(const cpumask_t *mask, const void *va, in hyperv_flush_tlb() argument 134 if ( !flush || cpumask_empty(mask) ) in hyperv_flush_tlb() 150 if ( cpumask_equal(mask, &cpu_online_map) ) in hyperv_flush_tlb() 162 if ( hv_vp_index(cpumask_last(mask)) >= 64 ) in hyperv_flush_tlb() 165 for_each_cpu ( cpu, mask ) in hyperv_flush_tlb() 196 ret = flush_tlb_ex(mask, va, flags); in hyperv_flush_tlb()
|
| /xen/tools/firmware/hvmloader/ |
| A D | 32bitbios_support.c | 58 uint32_t mask = shdr[i].sh_addralign - 1; in relocate_32bitbios() local 59 reloc_off = (reloc_off + mask) & ~mask; in relocate_32bitbios() 81 uint32_t mask = shdr[i].sh_addralign - 1; in relocate_32bitbios() local 88 reloc_off = (reloc_off + mask) & ~mask; in relocate_32bitbios()
|