/xen/xen/include/xen/ |
A D | cpumask.h | 142 static inline void cpumask_and(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_and() 148 static inline void cpumask_or(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_or() 149 const cpumask_t *src2p) in cpumask_or() 154 static inline void cpumask_xor(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_xor() 160 static inline void cpumask_andnot(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_andnot() 161 const cpumask_t *src2p) in cpumask_andnot() 166 static inline void cpumask_complement(cpumask_t *dstp, const cpumask_t *srcp) in cpumask_complement() 172 const cpumask_t *src2p) in cpumask_equal() 178 const cpumask_t *src2p) in cpumask_intersects() 184 const cpumask_t *src2p) in cpumask_subset() [all …]
|
A D | smp.h | 11 extern void smp_send_event_check_mask(const cpumask_t *mask); 39 const cpumask_t *selected, 66 void smp_send_call_function_mask(const cpumask_t *mask);
|
A D | irq.h | 68 void (*set_affinity)(struct irq_desc *, const cpumask_t *); 161 extern void pirq_set_affinity(struct domain *d, int irq, const cpumask_t *); 167 unsigned int set_desc_affinity(struct irq_desc *, const cpumask_t *);
|
A D | softirq.h | 29 void cpumask_raise_softirq(const cpumask_t *, unsigned int nr);
|
/xen/xen/include/asm-x86/ |
A D | genapic.h | 36 const cpumask_t *(*vector_allocation_cpumask)(int cpu); 37 unsigned int (*cpu_mask_to_apicid)(const cpumask_t *cpumask); 38 void (*send_IPI_mask)(const cpumask_t *mask, int vector); 58 unsigned int cpu_mask_to_apicid_flat(const cpumask_t *cpumask); 59 void send_IPI_mask_flat(const cpumask_t *mask, int vector); 60 const cpumask_t *vector_allocation_cpumask_flat(int cpu); 73 unsigned int cpu_mask_to_apicid_phys(const cpumask_t *cpumask); 74 void send_IPI_mask_phys(const cpumask_t *mask, int vector); 75 const cpumask_t *vector_allocation_cpumask_phys(int cpu);
|
A D | flushtlb.h | 77 static inline void tlbflush_filter(cpumask_t *mask, uint32_t page_timestamp) in tlbflush_filter() 145 void flush_area_mask(const cpumask_t *, const void *va, unsigned int flags); 201 void guest_flush_tlb_mask(const struct domain *d, const cpumask_t *mask);
|
A D | irq.h | 160 void fixup_irqs(const cpumask_t *mask, bool verbose); 174 int assign_irq_vector(int irq, const cpumask_t *); 188 int bind_irq_vector(int irq, int vector, const cpumask_t *); 191 void irq_set_affinity(struct irq_desc *, const cpumask_t *mask);
|
A D | smp.h | 35 void send_IPI_mask(const cpumask_t *, int vector); 76 extern cpumask_t **socket_cpumask;
|
/xen/xen/arch/arm/ |
A D | smp.c | 8 void arch_flush_tlb_mask(const cpumask_t *mask) in arch_flush_tlb_mask() 14 void smp_send_event_check_mask(const cpumask_t *mask) in smp_send_event_check_mask() 19 void smp_send_call_function_mask(const cpumask_t *mask) in smp_send_call_function_mask() 21 cpumask_t target_mask; in smp_send_call_function_mask()
|
A D | smpboot.c | 40 cpumask_t cpu_online_map; 41 cpumask_t cpu_present_map; 42 cpumask_t cpu_possible_map;
|
/xen/xen/arch/x86/genapic/ |
A D | x2apic.c | 33 static DEFINE_PER_CPU_READ_MOSTLY(cpumask_t *, cluster_cpus); 34 static cpumask_t *cluster_cpus_spare; 72 static const cpumask_t *vector_allocation_cpumask_x2apic_cluster(int cpu) in vector_allocation_cpumask_x2apic_cluster() 77 static unsigned int cpu_mask_to_apicid_x2apic_cluster(const cpumask_t *cpumask) in cpu_mask_to_apicid_x2apic_cluster() 81 const cpumask_t *cluster_cpus = per_cpu(cluster_cpus, cpu); in cpu_mask_to_apicid_x2apic_cluster() 95 static void send_IPI_mask_x2apic_phys(const cpumask_t *cpumask, int vector) in send_IPI_mask_x2apic_phys() 128 static void send_IPI_mask_x2apic_cluster(const cpumask_t *cpumask, int vector) in send_IPI_mask_x2apic_cluster() 131 cpumask_t *ipimask = per_cpu(scratch_mask, cpu); in send_IPI_mask_x2apic_cluster() 132 const cpumask_t *cluster_cpus; in send_IPI_mask_x2apic_cluster() 197 cluster_cpus_spare = xzalloc(cpumask_t); in update_clusterinfo()
|
A D | delivery.c | 27 const cpumask_t *vector_allocation_cpumask_flat(int cpu) in vector_allocation_cpumask_flat() 32 unsigned int cpu_mask_to_apicid_flat(const cpumask_t *cpumask) in cpu_mask_to_apicid_flat() 51 const cpumask_t *vector_allocation_cpumask_phys(int cpu) in vector_allocation_cpumask_phys() 56 unsigned int cpu_mask_to_apicid_phys(const cpumask_t *cpumask) in cpu_mask_to_apicid_phys()
|
/xen/xen/arch/x86/ |
A D | smp.c | 67 void send_IPI_mask(const cpumask_t *mask, int vector) in send_IPI_mask() 70 cpumask_t *scratch = this_cpu(send_ipi_cpumask); in send_IPI_mask() 168 void send_IPI_mask_flat(const cpumask_t *cpumask, int vector) in send_IPI_mask_flat() 205 void send_IPI_mask_phys(const cpumask_t *mask, int vector) in send_IPI_mask_phys() 243 static cpumask_t flush_cpumask; 259 void flush_area_mask(const cpumask_t *mask, const void *va, unsigned int flags) in flush_area_mask() 293 cpumask_t allbutself; in new_tlbflush_clock_period() 305 void smp_send_event_check_mask(const cpumask_t *mask) in smp_send_event_check_mask() 310 void smp_send_call_function_mask(const cpumask_t *mask) in smp_send_call_function_mask()
|
A D | smpboot.c | 59 static cpumask_t scratch_cpu0mask; 62 static cpumask_t send_ipi_cpu0mask; 64 cpumask_t cpu_online_map __read_mostly; 70 cpumask_t **__read_mostly socket_cpumask; 71 static cpumask_t *secondary_socket_cpumask; 123 static cpumask_t tsc_sync_cpu_mask; 236 static cpumask_t cpu_sibling_setup_map; 534 cpumask_t tmp_map; in alloc_cpu_id() 1058 (secondary_socket_cpumask = xzalloc(cpumask_t)) == NULL ) in cpu_smpboot_alloc() 1139 socket_cpumask = xzalloc_array(cpumask_t *, nr_sockets); in smp_prepare_cpus() [all …]
|
/xen/xen/include/asm-x86/guest/ |
A D | hypervisor.h | 38 int (*flush_tlb)(const cpumask_t *mask, const void *va, unsigned int flags); 54 int hypervisor_flush_tlb(const cpumask_t *mask, const void *va, 67 static inline int hypervisor_flush_tlb(const cpumask_t *mask, const void *va, in hypervisor_flush_tlb()
|
/xen/xen/common/ |
A D | softirq.c | 25 static DEFINE_PER_CPU(cpumask_t, batch_mask); 83 void cpumask_raise_softirq(const cpumask_t *mask, unsigned int nr) in cpumask_raise_softirq() 86 cpumask_t send_mask, *raise_mask; in cpumask_raise_softirq() 129 cpumask_t *mask = &per_cpu(batch_mask, this_cpu); in cpu_raise_softirq_batch_finish()
|
A D | smp.c | 32 cpumask_t selected; 40 cpumask_t allbutself; in smp_call_function() 48 const cpumask_t *selected, in on_selected_cpus()
|
A D | rwlock.c | 106 static DEFINE_PER_CPU(cpumask_t, percpu_rwlock_readers); 112 cpumask_t *rwlock_readers = &this_cpu(percpu_rwlock_readers); in _percpu_write_lock()
|
/xen/xen/arch/x86/guest/hyperv/ |
A D | private.h | 38 int hyperv_flush_tlb(const cpumask_t *mask, const void *va, 42 int cpumask_to_vpset(struct hv_vpset *vpset, const cpumask_t *mask);
|
A D | tlb.c | 72 static uint64_t flush_tlb_ex(const cpumask_t *mask, const void *va, in flush_tlb_ex() 126 int hyperv_flush_tlb(const cpumask_t *mask, const void *va, in hyperv_flush_tlb()
|
/xen/xen/include/asm-arm/ |
A D | flushtlb.h | 11 static inline void tlbflush_filter(cpumask_t *mask, uint32_t page_timestamp) {} in tlbflush_filter() 29 void arch_flush_tlb_mask(const cpumask_t *mask);
|
/xen/xen/common/sched/ |
A D | private.h | 16 extern cpumask_t cpupool_free_cpus; 24 extern cpumask_t sched_res_mask; 205 DECLARE_PER_CPU(cpumask_t, cpumask_scratch); 468 const cpumask_t *hard, in sched_adjust_affinity() 469 const cpumask_t *soft) in sched_adjust_affinity() 519 static inline cpumask_t *cpupool_domain_master_cpumask(const struct domain *d) in cpupool_domain_master_cpumask() 580 cpumask_t *mask) in affinity_balance_cpumask() 594 const cpumask_t *sched_get_opt_cpumask(enum sched_gran opt, unsigned int cpu);
|
A D | cpupool.c | 30 cpumask_t cpupool_free_cpus; /* cpus not in any cpupool */ 36 static cpumask_t cpupool_locked_cpus; 383 const cpumask_t *cpus; in cpupool_assign_cpu_locked() 420 const cpumask_t *cpus; in cpupool_unassign_cpu_finish() 472 const cpumask_t *cpus; in cpupool_unassign_cpu_start() 639 const cpumask_t *cpus; in cpupool_cpu_add() 693 cpumask_t *cpus; in cpupool_cpu_remove_prologue() 810 const cpumask_t *cpus; in cpupool_do_sysctl() 923 const cpumask_t *cpupool_valid_cpus(const struct cpupool *pool) in cpupool_valid_cpus()
|
A D | core.c | 68 cpumask_t sched_res_mask; 85 DEFINE_PER_CPU(cpumask_t, cpumask_scratch); 99 struct sched_unit *unit, const cpumask_t *hard, const cpumask_t *soft); 1281 struct sched_unit *unit, const cpumask_t *hard, const cpumask_t *soft) in sched_set_affinity() 1299 struct vcpu *v, const cpumask_t *affinity, const cpumask_t *which) in vcpu_set_affinity() 1342 cpumask_t *online; in vcpu_set_hard_affinity() 1736 cpumask_t *dom_affinity; in domain_update_node_affinity() 1737 const cpumask_t *online; in domain_update_node_affinity() 2844 const cpumask_t *mask; in sched_get_opt_cpumask() 3039 const cpumask_t *mask; in schedule_cpu_add() [all …]
|
A D | rt.c | 193 cpumask_t tickled; /* cpus been tickled */ 311 cpumask_t *cpupool_mask, *mask; in rt_dump_unit() 642 cpumask_t *cpus = cpumask_scratch_cpu(locked_cpu); in rt_res_pick_locked() 643 const cpumask_t *online; in rt_res_pick_locked() 763 cpumask_t *online = get_sched_res(cpu)->cpupool->res_valid; in rt_deinit_pdata() 994 runq_pick(const struct scheduler *ops, const cpumask_t *mask, unsigned int cpu) in runq_pick() 1000 cpumask_t *cpu_common = cpumask_scratch_cpu(cpu); in runq_pick() 1001 const cpumask_t *online; in runq_pick() 1177 cpumask_t *not_tickled = cpumask_scratch_cpu(smp_processor_id()); in runq_tickle() 1178 const cpumask_t *online; in runq_tickle()
|