Home
last modified time | relevance | path

Searched refs:mask (Results 1 – 25 of 187) sorted by relevance

12345678

/xen/xen/include/asm-x86/
A Dmpspec.h36 unsigned long mask[PHYSID_ARRAY_SIZE]; member
41 #define physid_set(physid, map) set_bit(physid, (map).mask)
46 #define first_physid(map) find_first_bit((map).mask, \
48 #define next_physid(id, map) find_next_bit((map).mask, \
51 const unsigned long *mask = (map).mask; \
53 for (id = find_first_bit(mask, MAX_APICS); id < MAX_APICS; \
54 id = find_next_bit(mask, MAX_APICS, (id) + 1)) \
59 #define physids_and(dst, src1, src2) bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
60 #define physids_or(dst, src1, src2) bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
62 #define physids_complement(dst, src) bitmap_complement((dst).mask,(src).mask, MAX_APICS)
[all …]
A Dflushtlb.h77 static inline void tlbflush_filter(cpumask_t *mask, uint32_t page_timestamp) in tlbflush_filter() argument
85 for_each_cpu ( cpu, mask ) in tlbflush_filter()
87 __cpumask_clear_cpu(cpu, mask); in tlbflush_filter()
146 #define flush_mask(mask, flags) flush_area_mask(mask, NULL, flags) argument
159 #define flush_tlb_mask(mask) \ argument
160 flush_mask(mask, FLUSH_TLB)
161 #define flush_tlb_one_mask(mask,v) \ argument
162 flush_area_mask(mask, (const void *)(v), FLUSH_TLB|FLUSH_ORDER(0))
169 #define arch_flush_tlb_mask(mask) flush_mask(mask, FLUSH_TLB | FLUSH_FORCE_IPI) argument
201 void guest_flush_tlb_mask(const struct domain *d, const cpumask_t *mask);
A Dvpmu.h84 static inline void vpmu_set(struct vpmu_struct *vpmu, const u32 mask) in vpmu_set() argument
86 vpmu->flags |= mask; in vpmu_set()
88 static inline void vpmu_reset(struct vpmu_struct *vpmu, const u32 mask) in vpmu_reset() argument
90 vpmu->flags &= ~mask; in vpmu_reset()
97 static inline bool_t vpmu_is_set(const struct vpmu_struct *vpmu, const u32 mask) in vpmu_is_set() argument
99 return !!(vpmu->flags & mask); in vpmu_is_set()
102 const u32 mask) in vpmu_are_all_set() argument
104 return !!((vpmu->flags & mask) == mask); in vpmu_are_all_set()
/xen/xen/arch/x86/
A Dsmp.c34 static unsigned int prepare_ICR2(unsigned int mask) in prepare_ICR2() argument
36 return SET_xAPIC_DEST_FIELD(mask); in prepare_ICR2()
174 mask &= cpumask_bits(&cpu_online_map)[0]; in send_IPI_mask_flat()
175 mask &= ~(1UL << smp_processor_id()); in send_IPI_mask_flat()
176 if ( mask == 0 ) in send_IPI_mask_flat()
189 cfg = prepare_ICR2(mask); in send_IPI_mask_flat()
212 for_each_cpu ( query_cpu, mask ) in send_IPI_mask_phys()
266 cpumask_test_cpu(cpu, mask) ) in flush_area_mask()
270 !cpumask_subset(mask, cpumask_of(cpu)) ) in flush_area_mask()
307 send_IPI_mask(mask, EVENT_CHECK_VECTOR); in smp_send_event_check_mask()
[all …]
A Dhypercall.c180 unsigned int mask, ...) in hypercall_xlat_continuation() argument
190 ASSERT(!(mask >> nr)); in hypercall_xlat_continuation()
192 ASSERT(!id || !(mask & (1U << *id))); in hypercall_xlat_continuation()
194 va_start(args, mask); in hypercall_xlat_continuation()
204 for ( i = 0; i < nr; ++i, mask >>= 1 ) in hypercall_xlat_continuation()
206 if ( mask & 1 ) in hypercall_xlat_continuation()
211 mask &= ~1U; in hypercall_xlat_continuation()
232 for ( i = 0; i < nr; ++i, mask >>= 1 ) in hypercall_xlat_continuation()
246 if ( (mask & 1) ) in hypercall_xlat_continuation()
251 mask &= ~1U; in hypercall_xlat_continuation()
[all …]
A Dxstate.c293 uint32_t hmask = mask >> 32; in xsave()
294 uint32_t lmask = mask; in xsave()
361 if ( mask & X86_XCR0_FP ) in xsave()
367 uint32_t hmask = mask >> 32; in xrstor()
368 uint32_t lmask = mask; in xrstor()
698 u64 mask; in handle_xsetbv() local
738 if ( mask ) in handle_xsetbv()
748 mask |= XSTATE_LAZY; in handle_xsetbv()
753 xrstor(curr, mask); in handle_xsetbv()
807 if ( (~xcr0 & mask) && !set_xcr0(xcr0 | mask) ) in xstate_set_init()
[all …]
A Di8259.c98 unsigned int mask = 1 << irq; in _disable_8259A_irq() local
102 cached_irq_mask |= mask; in _disable_8259A_irq()
118 unsigned int mask = ~(1 << desc->irq); in enable_8259A_irq() local
122 cached_irq_mask &= mask; in enable_8259A_irq()
133 unsigned int mask = 1<<irq; in i8259A_irq_pending() local
139 ret = inb(0x20) & mask; in i8259A_irq_pending()
141 ret = inb(0xA0) & (mask >> 8); in i8259A_irq_pending()
/xen/xen/arch/arm/arm64/
A Dinsn.c43 u32 mask; in aarch64_get_imm_shift_mask() local
48 mask = BIT(26, UL) - 1; in aarch64_get_imm_shift_mask()
52 mask = BIT(19, UL) - 1; in aarch64_get_imm_shift_mask()
56 mask = BIT(16, UL) - 1; in aarch64_get_imm_shift_mask()
60 mask = BIT(14, UL) - 1; in aarch64_get_imm_shift_mask()
68 mask = BIT(9, UL) - 1; in aarch64_get_imm_shift_mask()
72 mask = BIT(7, UL) - 1; in aarch64_get_imm_shift_mask()
77 mask = BIT(6, UL) - 1; in aarch64_get_imm_shift_mask()
81 mask = BIT(6, UL) - 1; in aarch64_get_imm_shift_mask()
88 *maskp = mask; in aarch64_get_imm_shift_mask()
[all …]
/xen/xen/arch/arm/
A Dguest_walk.c38 paddr_t mask, paddr; in guest_walk_sd() local
46 if ( n == 0 || !(gva & mask) ) in guest_walk_sd()
91 mask = GENMASK(31, (14 - n)); in guest_walk_sd()
92 paddr = (ttbr & mask); in guest_walk_sd()
94 mask = GENMASK((31 - n), 20); in guest_walk_sd()
95 paddr |= (gva & mask) >> 18; in guest_walk_sd()
114 mask = GENMASK(19, 12); in guest_walk_sd()
165 *ipa = gva & mask; in guest_walk_sd()
369 paddr_t mask, paddr; in guest_walk_ld() local
477 if ( ((t1_sz == 0) && !ttbr) || (t1_sz && (gva & mask) == mask) ) in guest_walk_ld()
[all …]
A Dsmp.c8 void arch_flush_tlb_mask(const cpumask_t *mask) in arch_flush_tlb_mask() argument
14 void smp_send_event_check_mask(const cpumask_t *mask) in smp_send_event_check_mask() argument
16 send_SGI_mask(mask, GIC_SGI_EVENT_CHECK); in smp_send_event_check_mask()
19 void smp_send_call_function_mask(const cpumask_t *mask) in smp_send_call_function_mask() argument
23 cpumask_andnot(&target_mask, mask, cpumask_of(smp_processor_id())); in smp_send_call_function_mask()
27 if ( cpumask_test_cpu(smp_processor_id(), mask) ) in smp_send_call_function_mask()
/xen/xen/include/xen/
A Dcpumask.h316 return *mask != NULL; in alloc_cpumask_var()
321 if (*mask == NULL) in cond_alloc_cpumask_var()
323 return *mask != NULL; in cond_alloc_cpumask_var()
329 return *mask != NULL; in zalloc_cpumask_var()
334 if (*mask == NULL) in cond_zalloc_cpumask_var()
337 cpumask_clear(*mask); in cond_zalloc_cpumask_var()
338 return *mask != NULL; in cond_zalloc_cpumask_var()
343 xfree(mask); in free_cpumask_var()
359 cpumask_clear(*mask); in zalloc_cpumask_var()
372 #define for_each_cpu(cpu, mask) \ argument
[all …]
/xen/tools/libxc/
A Dxc_pagetab.c33 uint64_t paddr, mask, pte = 0; in xc_translate_foreign_address() local
72 mask = 0x0000ff8000000000ull; in xc_translate_foreign_address()
75 mask = 0x0000007fc0000000ull; in xc_translate_foreign_address()
78 mask = 0x00000000ffc00000ull; in xc_translate_foreign_address()
84 paddr += ((virt & mask) >> (xc_ffs64(mask) - 1)) * size; in xc_translate_foreign_address()
97 mask = ((mask ^ ~-mask) >> 1); /* All bits below first set bit */ in xc_translate_foreign_address()
98 return ((paddr & ~mask) | (virt & mask)) >> PAGE_SHIFT; in xc_translate_foreign_address()
100 mask >>= (pt_levels == 2 ? 10 : 9); in xc_translate_foreign_address()
A Dxc_tbuf.c117 int xc_tbuf_set_cpu_mask(xc_interface *xch, xc_cpumap_t mask) in xc_tbuf_set_cpu_mask() argument
120 DECLARE_HYPERCALL_BOUNCE(mask, 0, XC_HYPERCALL_BUFFER_BOUNCE_IN); in xc_tbuf_set_cpu_mask()
131 HYPERCALL_BOUNCE_SET_SIZE(mask, cpusize); in xc_tbuf_set_cpu_mask()
140 if ( xc_hypercall_bounce_pre(xch, mask) ) in xc_tbuf_set_cpu_mask()
150 set_xen_guest_handle(sysctl.u.tbuf_op.cpu_mask.bitmap, mask); in xc_tbuf_set_cpu_mask()
155 xc_hypercall_bounce_post(xch, mask); in xc_tbuf_set_cpu_mask()
161 int xc_tbuf_set_evt_mask(xc_interface *xch, uint32_t mask) in xc_tbuf_set_evt_mask() argument
168 sysctl.u.tbuf_op.evt_mask = mask; in xc_tbuf_set_evt_mask()
/xen/xen/include/asm-arm/
A Dbitops.h54 void clear_mask16(uint16_t mask, volatile void *p);
74 bool clear_mask16_timeout(uint16_t mask, volatile void *p,
88 unsigned int mask = BITOP_MASK(nr); in __test_and_set_bit() local
93 *p = old | mask; in __test_and_set_bit()
94 return (old & mask) != 0; in __test_and_set_bit()
108 unsigned int mask = BITOP_MASK(nr); in __test_and_clear_bit() local
113 *p = old & ~mask; in __test_and_clear_bit()
114 return (old & mask) != 0; in __test_and_clear_bit()
121 unsigned int mask = BITOP_MASK(nr); in __test_and_change_bit() local
126 *p = old ^ mask; in __test_and_change_bit()
[all …]
A Dvreg.h136 unsigned long mask = VREG_REG_MASK(size); in vreg_reg_update() local
139 *reg &= ~(mask << shift); in vreg_reg_update()
140 *reg |= ((unsigned long)val & mask) << shift; in vreg_reg_update()
147 unsigned long mask = VREG_REG_MASK(size); in vreg_reg_setbits() local
150 *reg |= ((unsigned long)bits & mask) << shift; in vreg_reg_setbits()
157 unsigned long mask = VREG_REG_MASK(size); in vreg_reg_clearbits() local
160 *reg &= ~(((unsigned long)bits & mask) << shift); in vreg_reg_clearbits()
/xen/xen/arch/arm/arm64/lib/
A Dbitops.c36 const uint32_t mask = BITOP_MASK((unsigned int)nr); \
46 : "r" (mask)); \
72 const uint32_t mask = BITOP_MASK(bit); \
83 : "r" (mask), "r" (bit) \
121 static always_inline bool int_clear_mask16(uint16_t mask, volatile uint16_t *p, in bitop()
133 : "r" (mask)); in bitop()
142 void clear_mask16(uint16_t mask, volatile void *p) in clear_mask16() argument
144 if ( !int_clear_mask16(mask, p, false, 0) ) in clear_mask16()
148 bool clear_mask16_timeout(uint16_t mask, volatile void *p, in clear_mask16_timeout() argument
151 return int_clear_mask16(mask, p, true, max_try); in clear_mask16_timeout()
A Dmemcmp.S58 mask .req x13 label
93 mov mask, #~0
94 CPU_BE( lsr mask, mask, limit )
95 CPU_LE( lsl mask, mask, limit )
96 bic data1, data1, mask
97 bic data2, data2, mask
99 orr diff, diff, mask
/xen/xen/common/
A Dpdx.c45 static u64 __init fill_mask(u64 mask) in fill_mask() argument
47 while (mask & (mask + 1)) in fill_mask()
48 mask |= mask + 1; in fill_mask()
50 return mask; in fill_mask()
76 void __init pfn_pdx_hole_setup(unsigned long mask) in pfn_pdx_hole_setup() argument
91 i = find_next_zero_bit(&mask, BITS_PER_LONG, j + 1); in pfn_pdx_hole_setup()
94 j = find_next_bit(&mask, BITS_PER_LONG, i + 1); in pfn_pdx_hole_setup()
A Dbitmap.c272 unsigned long mask; in bitmap_find_free_region() local
280 mask = (1ul << (pages - 1)); in bitmap_find_free_region()
281 mask += mask - 1; in bitmap_find_free_region()
289 bitmap[index] |= (mask << offset); in bitmap_find_free_region()
309 unsigned long mask = (1ul << (pages - 1)); in bitmap_release_region() local
312 mask += mask - 1; in bitmap_release_region()
313 bitmap[index] &= ~(mask << offset); in bitmap_release_region()
320 unsigned long mask = (1ul << (pages - 1)); in bitmap_allocate_region() local
329 mask += mask - 1; in bitmap_allocate_region()
330 if (bitmap[index] & (mask << offset)) in bitmap_allocate_region()
[all …]
A Dsoftirq.c83 void cpumask_raise_softirq(const cpumask_t *mask, unsigned int nr) in cpumask_raise_softirq() argument
96 for_each_cpu(cpu, mask) in cpumask_raise_softirq()
129 cpumask_t *mask = &per_cpu(batch_mask, this_cpu); in cpu_raise_softirq_batch_finish() local
132 for_each_cpu ( cpu, mask ) in cpu_raise_softirq_batch_finish()
134 __cpumask_clear_cpu(cpu, mask); in cpu_raise_softirq_batch_finish()
135 smp_send_event_check_mask(mask); in cpu_raise_softirq_batch_finish()
136 cpumask_clear(mask); in cpu_raise_softirq_batch_finish()
/xen/xen/arch/x86/hvm/
A Dvpic.c54 if ( mask == 0 ) in vpic_get_priority()
67 uint8_t mask; in vpic_get_highest_priority_irq() local
84 mask = vpic->isr; in vpic_get_highest_priority_irq()
86 mask &= ~(1 << 2); in vpic_get_highest_priority_irq()
88 mask &= ~vpic->imr; in vpic_get_highest_priority_irq()
140 uint8_t mask = 1 << irq; in __vpic_intack() local
146 if ( !(vpic->elcr & mask) ) in __vpic_intack()
147 vpic->irr &= ~mask; in __vpic_intack()
150 vpic->isr |= mask; in __vpic_intack()
462 vpic->irr |= mask; in vpic_irq_positive_edge()
[all …]
/xen/xen/arch/arm/arm32/lib/
A Dbitops.c37 const uint32_t mask = BITOP_MASK((unsigned int)nr); \
50 : "r" (mask)); \
76 const uint32_t mask = BITOP_MASK(bit); \
92 : "r" (mask), "r" (bit)); \
129 static always_inline bool int_clear_mask16(uint16_t mask, volatile uint16_t *p, in bitop()
143 : "r" (mask)); in bitop()
152 void clear_mask16(uint16_t mask, volatile void *p) in clear_mask16() argument
154 if ( !int_clear_mask16(mask, p, false, 0) ) in clear_mask16()
158 bool clear_mask16_timeout(uint16_t mask, volatile void *p, in clear_mask16_timeout() argument
161 return int_clear_mask16(mask, p, true, max_try); in clear_mask16_timeout()
/xen/xen/arch/x86/cpu/mtrr/
A Dgeneric.c238 unsigned long mask = smp_changes_mask; in mtrr_state_warn() local
242 if (!mask) in mtrr_state_warn()
244 if (mask & MTRR_CHANGE_MASK_FIXED) in mtrr_state_warn()
246 if (mask & MTRR_CHANGE_MASK_VARIABLE) in mtrr_state_warn()
248 if (mask & MTRR_CHANGE_MASK_DEFTYPE) in mtrr_state_warn()
383 mask_lo = (uint32_t)vr->mask; in set_mtrr_var_ranges()
384 mask_hi = (uint32_t)(vr->mask >> 32); in set_mtrr_var_ranges()
504 unsigned long mask, count; in generic_set_all() local
512 mask = set_mtrr_state(); in generic_set_all()
519 if (mask & 0x01) in generic_set_all()
[all …]
/xen/xen/arch/x86/guest/hyperv/
A Dtlb.c72 static uint64_t flush_tlb_ex(const cpumask_t *mask, const void *va, in flush_tlb_ex() argument
94 nr_banks = cpumask_to_vpset(&flush->hv_vp_set, mask); in flush_tlb_ex()
126 int hyperv_flush_tlb(const cpumask_t *mask, const void *va, in hyperv_flush_tlb() argument
134 if ( !flush || cpumask_empty(mask) ) in hyperv_flush_tlb()
150 if ( cpumask_equal(mask, &cpu_online_map) ) in hyperv_flush_tlb()
162 if ( hv_vp_index(cpumask_last(mask)) >= 64 ) in hyperv_flush_tlb()
165 for_each_cpu ( cpu, mask ) in hyperv_flush_tlb()
196 ret = flush_tlb_ex(mask, va, flags); in hyperv_flush_tlb()
/xen/tools/firmware/hvmloader/
A D32bitbios_support.c58 uint32_t mask = shdr[i].sh_addralign - 1; in relocate_32bitbios() local
59 reloc_off = (reloc_off + mask) & ~mask; in relocate_32bitbios()
81 uint32_t mask = shdr[i].sh_addralign - 1; in relocate_32bitbios() local
88 reloc_off = (reloc_off + mask) & ~mask; in relocate_32bitbios()

Completed in 45 milliseconds

12345678