Home
last modified time | relevance | path

Searched refs:cpumask_t (Results 1 – 25 of 83) sorted by relevance

1234

/xen/xen/include/xen/
A Dcpumask.h142 static inline void cpumask_and(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_and()
148 static inline void cpumask_or(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_or()
149 const cpumask_t *src2p) in cpumask_or()
154 static inline void cpumask_xor(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_xor()
160 static inline void cpumask_andnot(cpumask_t *dstp, const cpumask_t *src1p, in cpumask_andnot()
161 const cpumask_t *src2p) in cpumask_andnot()
166 static inline void cpumask_complement(cpumask_t *dstp, const cpumask_t *srcp) in cpumask_complement()
172 const cpumask_t *src2p) in cpumask_equal()
178 const cpumask_t *src2p) in cpumask_intersects()
184 const cpumask_t *src2p) in cpumask_subset()
[all …]
A Dsmp.h11 extern void smp_send_event_check_mask(const cpumask_t *mask);
39 const cpumask_t *selected,
66 void smp_send_call_function_mask(const cpumask_t *mask);
A Dirq.h68 void (*set_affinity)(struct irq_desc *, const cpumask_t *);
161 extern void pirq_set_affinity(struct domain *d, int irq, const cpumask_t *);
167 unsigned int set_desc_affinity(struct irq_desc *, const cpumask_t *);
A Dsoftirq.h29 void cpumask_raise_softirq(const cpumask_t *, unsigned int nr);
/xen/xen/include/asm-x86/
A Dgenapic.h36 const cpumask_t *(*vector_allocation_cpumask)(int cpu);
37 unsigned int (*cpu_mask_to_apicid)(const cpumask_t *cpumask);
38 void (*send_IPI_mask)(const cpumask_t *mask, int vector);
58 unsigned int cpu_mask_to_apicid_flat(const cpumask_t *cpumask);
59 void send_IPI_mask_flat(const cpumask_t *mask, int vector);
60 const cpumask_t *vector_allocation_cpumask_flat(int cpu);
73 unsigned int cpu_mask_to_apicid_phys(const cpumask_t *cpumask);
74 void send_IPI_mask_phys(const cpumask_t *mask, int vector);
75 const cpumask_t *vector_allocation_cpumask_phys(int cpu);
A Dflushtlb.h77 static inline void tlbflush_filter(cpumask_t *mask, uint32_t page_timestamp) in tlbflush_filter()
145 void flush_area_mask(const cpumask_t *, const void *va, unsigned int flags);
201 void guest_flush_tlb_mask(const struct domain *d, const cpumask_t *mask);
A Dirq.h160 void fixup_irqs(const cpumask_t *mask, bool verbose);
174 int assign_irq_vector(int irq, const cpumask_t *);
188 int bind_irq_vector(int irq, int vector, const cpumask_t *);
191 void irq_set_affinity(struct irq_desc *, const cpumask_t *mask);
A Dsmp.h35 void send_IPI_mask(const cpumask_t *, int vector);
76 extern cpumask_t **socket_cpumask;
/xen/xen/arch/arm/
A Dsmp.c8 void arch_flush_tlb_mask(const cpumask_t *mask) in arch_flush_tlb_mask()
14 void smp_send_event_check_mask(const cpumask_t *mask) in smp_send_event_check_mask()
19 void smp_send_call_function_mask(const cpumask_t *mask) in smp_send_call_function_mask()
21 cpumask_t target_mask; in smp_send_call_function_mask()
A Dsmpboot.c40 cpumask_t cpu_online_map;
41 cpumask_t cpu_present_map;
42 cpumask_t cpu_possible_map;
/xen/xen/arch/x86/genapic/
A Dx2apic.c33 static DEFINE_PER_CPU_READ_MOSTLY(cpumask_t *, cluster_cpus);
34 static cpumask_t *cluster_cpus_spare;
72 static const cpumask_t *vector_allocation_cpumask_x2apic_cluster(int cpu) in vector_allocation_cpumask_x2apic_cluster()
77 static unsigned int cpu_mask_to_apicid_x2apic_cluster(const cpumask_t *cpumask) in cpu_mask_to_apicid_x2apic_cluster()
81 const cpumask_t *cluster_cpus = per_cpu(cluster_cpus, cpu); in cpu_mask_to_apicid_x2apic_cluster()
95 static void send_IPI_mask_x2apic_phys(const cpumask_t *cpumask, int vector) in send_IPI_mask_x2apic_phys()
128 static void send_IPI_mask_x2apic_cluster(const cpumask_t *cpumask, int vector) in send_IPI_mask_x2apic_cluster()
131 cpumask_t *ipimask = per_cpu(scratch_mask, cpu); in send_IPI_mask_x2apic_cluster()
132 const cpumask_t *cluster_cpus; in send_IPI_mask_x2apic_cluster()
197 cluster_cpus_spare = xzalloc(cpumask_t); in update_clusterinfo()
A Ddelivery.c27 const cpumask_t *vector_allocation_cpumask_flat(int cpu) in vector_allocation_cpumask_flat()
32 unsigned int cpu_mask_to_apicid_flat(const cpumask_t *cpumask) in cpu_mask_to_apicid_flat()
51 const cpumask_t *vector_allocation_cpumask_phys(int cpu) in vector_allocation_cpumask_phys()
56 unsigned int cpu_mask_to_apicid_phys(const cpumask_t *cpumask) in cpu_mask_to_apicid_phys()
/xen/xen/arch/x86/
A Dsmp.c67 void send_IPI_mask(const cpumask_t *mask, int vector) in send_IPI_mask()
70 cpumask_t *scratch = this_cpu(send_ipi_cpumask); in send_IPI_mask()
168 void send_IPI_mask_flat(const cpumask_t *cpumask, int vector) in send_IPI_mask_flat()
205 void send_IPI_mask_phys(const cpumask_t *mask, int vector) in send_IPI_mask_phys()
243 static cpumask_t flush_cpumask;
259 void flush_area_mask(const cpumask_t *mask, const void *va, unsigned int flags) in flush_area_mask()
293 cpumask_t allbutself; in new_tlbflush_clock_period()
305 void smp_send_event_check_mask(const cpumask_t *mask) in smp_send_event_check_mask()
310 void smp_send_call_function_mask(const cpumask_t *mask) in smp_send_call_function_mask()
A Dsmpboot.c59 static cpumask_t scratch_cpu0mask;
62 static cpumask_t send_ipi_cpu0mask;
64 cpumask_t cpu_online_map __read_mostly;
70 cpumask_t **__read_mostly socket_cpumask;
71 static cpumask_t *secondary_socket_cpumask;
123 static cpumask_t tsc_sync_cpu_mask;
236 static cpumask_t cpu_sibling_setup_map;
534 cpumask_t tmp_map; in alloc_cpu_id()
1058 (secondary_socket_cpumask = xzalloc(cpumask_t)) == NULL ) in cpu_smpboot_alloc()
1139 socket_cpumask = xzalloc_array(cpumask_t *, nr_sockets); in smp_prepare_cpus()
[all …]
/xen/xen/include/asm-x86/guest/
A Dhypervisor.h38 int (*flush_tlb)(const cpumask_t *mask, const void *va, unsigned int flags);
54 int hypervisor_flush_tlb(const cpumask_t *mask, const void *va,
67 static inline int hypervisor_flush_tlb(const cpumask_t *mask, const void *va, in hypervisor_flush_tlb()
/xen/xen/common/
A Dsoftirq.c25 static DEFINE_PER_CPU(cpumask_t, batch_mask);
83 void cpumask_raise_softirq(const cpumask_t *mask, unsigned int nr) in cpumask_raise_softirq()
86 cpumask_t send_mask, *raise_mask; in cpumask_raise_softirq()
129 cpumask_t *mask = &per_cpu(batch_mask, this_cpu); in cpu_raise_softirq_batch_finish()
A Dsmp.c32 cpumask_t selected;
40 cpumask_t allbutself; in smp_call_function()
48 const cpumask_t *selected, in on_selected_cpus()
A Drwlock.c106 static DEFINE_PER_CPU(cpumask_t, percpu_rwlock_readers);
112 cpumask_t *rwlock_readers = &this_cpu(percpu_rwlock_readers); in _percpu_write_lock()
/xen/xen/arch/x86/guest/hyperv/
A Dprivate.h38 int hyperv_flush_tlb(const cpumask_t *mask, const void *va,
42 int cpumask_to_vpset(struct hv_vpset *vpset, const cpumask_t *mask);
A Dtlb.c72 static uint64_t flush_tlb_ex(const cpumask_t *mask, const void *va, in flush_tlb_ex()
126 int hyperv_flush_tlb(const cpumask_t *mask, const void *va, in hyperv_flush_tlb()
/xen/xen/include/asm-arm/
A Dflushtlb.h11 static inline void tlbflush_filter(cpumask_t *mask, uint32_t page_timestamp) {} in tlbflush_filter()
29 void arch_flush_tlb_mask(const cpumask_t *mask);
/xen/xen/common/sched/
A Dprivate.h16 extern cpumask_t cpupool_free_cpus;
24 extern cpumask_t sched_res_mask;
205 DECLARE_PER_CPU(cpumask_t, cpumask_scratch);
468 const cpumask_t *hard, in sched_adjust_affinity()
469 const cpumask_t *soft) in sched_adjust_affinity()
519 static inline cpumask_t *cpupool_domain_master_cpumask(const struct domain *d) in cpupool_domain_master_cpumask()
580 cpumask_t *mask) in affinity_balance_cpumask()
594 const cpumask_t *sched_get_opt_cpumask(enum sched_gran opt, unsigned int cpu);
A Dcpupool.c30 cpumask_t cpupool_free_cpus; /* cpus not in any cpupool */
36 static cpumask_t cpupool_locked_cpus;
383 const cpumask_t *cpus; in cpupool_assign_cpu_locked()
420 const cpumask_t *cpus; in cpupool_unassign_cpu_finish()
472 const cpumask_t *cpus; in cpupool_unassign_cpu_start()
639 const cpumask_t *cpus; in cpupool_cpu_add()
693 cpumask_t *cpus; in cpupool_cpu_remove_prologue()
810 const cpumask_t *cpus; in cpupool_do_sysctl()
923 const cpumask_t *cpupool_valid_cpus(const struct cpupool *pool) in cpupool_valid_cpus()
A Dcore.c68 cpumask_t sched_res_mask;
85 DEFINE_PER_CPU(cpumask_t, cpumask_scratch);
99 struct sched_unit *unit, const cpumask_t *hard, const cpumask_t *soft);
1281 struct sched_unit *unit, const cpumask_t *hard, const cpumask_t *soft) in sched_set_affinity()
1299 struct vcpu *v, const cpumask_t *affinity, const cpumask_t *which) in vcpu_set_affinity()
1342 cpumask_t *online; in vcpu_set_hard_affinity()
1736 cpumask_t *dom_affinity; in domain_update_node_affinity()
1737 const cpumask_t *online; in domain_update_node_affinity()
2844 const cpumask_t *mask; in sched_get_opt_cpumask()
3039 const cpumask_t *mask; in schedule_cpu_add()
[all …]
A Drt.c193 cpumask_t tickled; /* cpus been tickled */
311 cpumask_t *cpupool_mask, *mask; in rt_dump_unit()
642 cpumask_t *cpus = cpumask_scratch_cpu(locked_cpu); in rt_res_pick_locked()
643 const cpumask_t *online; in rt_res_pick_locked()
763 cpumask_t *online = get_sched_res(cpu)->cpupool->res_valid; in rt_deinit_pdata()
994 runq_pick(const struct scheduler *ops, const cpumask_t *mask, unsigned int cpu) in runq_pick()
1000 cpumask_t *cpu_common = cpumask_scratch_cpu(cpu); in runq_pick()
1001 const cpumask_t *online; in runq_pick()
1177 cpumask_t *not_tickled = cpumask_scratch_cpu(smp_processor_id()); in runq_tickle()
1178 const cpumask_t *online; in runq_tickle()

Completed in 31 milliseconds

1234