/linux/arch/powerpc/platforms/pseries/ |
A D | hotplug-cpu.c | 153 cpumask_var_t *cpu_mask) in find_cpu_id_range() argument 162 cpumask_clear(*cpu_mask); in find_cpu_id_range() 164 cpumask_set_cpu(cpu, *cpu_mask); in find_cpu_id_range() 188 while (!cpumask_empty(*cpu_mask)) { in find_cpu_id_range() 192 cpumask_shift_left(*cpu_mask, *cpu_mask, nthreads); in find_cpu_id_range() 195 if (!cpumask_empty(*cpu_mask)) in find_cpu_id_range() 214 cpumask_var_t cpu_mask; in pseries_add_processor() local 254 for_each_cpu(cpu, cpu_mask) { in pseries_add_processor() 269 cpu = cpumask_first(cpu_mask); in pseries_add_processor() 277 cpu_mask); in pseries_add_processor() [all …]
|
/linux/scripts/gdb/linux/ |
A D | cpus.py | 56 cpu_mask = {} variable 60 global cpu_mask 61 cpu_mask = {} 68 global cpu_mask 70 if mask_name in cpu_mask: 71 mask = cpu_mask[mask_name] 75 cpu_mask[mask_name] = mask
|
/linux/arch/x86/kernel/cpu/resctrl/ |
A D | rdtgroup.c | 353 cpumask_or(&prgrp->cpu_mask, &prgrp->cpu_mask, tmpmask); in cpus_mon_write() 367 cpumask_andnot(&crgrp->cpu_mask, &crgrp->cpu_mask, in cpus_mon_write() 383 cpumask_andnot(&r->cpu_mask, &r->cpu_mask, m); in cpumask_rdtgrp_clear() 386 cpumask_and(&crgrp->cpu_mask, &r->cpu_mask, &crgrp->cpu_mask); in cpumask_rdtgrp_clear() 436 cpumask_and(tmpmask, &rdtgrp->cpu_mask, &crgrp->cpu_mask); in cpus_ctrl_write() 1840 cpumask_var_t cpu_mask; in set_cache_qos_cfg() local 1862 cpumask_set_cpu(cpumask_any(&d->cpu_mask), cpu_mask); in set_cache_qos_cfg() 2310 cpumask_var_t cpu_mask; in reset_all_ctrls() local 2328 cpumask_set_cpu(cpumask_any(&d->cpu_mask), cpu_mask); in reset_all_ctrls() 2425 &rdtgroup_default.cpu_mask, &rdtgrp->cpu_mask); in rmdir_all_sub() [all …]
|
A D | ctrlmondata.c | 264 cpumask_var_t cpu_mask, bool mba_sc) in apply_config() argument 270 cpumask_set_cpu(cpumask_any(&dom->cpu_mask), cpu_mask); in apply_config() 285 cpumask_var_t cpu_mask; in resctrl_arch_update_domains() local 291 if (!zalloc_cpumask_var(&cpu_mask, GFP_KERNEL)) in resctrl_arch_update_domains() 304 if (!apply_config(hw_dom, cfg, idx, cpu_mask, mba_sc)) in resctrl_arch_update_domains() 322 if (cpumask_empty(cpu_mask) || mba_sc) in resctrl_arch_update_domains() 326 if (cpumask_test_cpu(cpu, cpu_mask)) in resctrl_arch_update_domains() 329 smp_call_function_many(cpu_mask, rdt_ctrl_update, &msr_param, 1); in resctrl_arch_update_domains() 333 free_cpumask_var(cpu_mask); in resctrl_arch_update_domains() 515 smp_call_function_any(&d->cpu_mask, mon_event_count, rr, 1); in mon_event_read()
|
A D | core.c | 342 if (cpumask_test_cpu(cpu, &d->cpu_mask)) in get_domain_from_cpu() 513 cpumask_set_cpu(cpu, &d->cpu_mask); in domain_add_cpu() 525 cpumask_set_cpu(cpu, &d->cpu_mask); in domain_add_cpu() 564 cpumask_clear_cpu(cpu, &d->cpu_mask); in domain_remove_cpu() 565 if (cpumask_empty(&d->cpu_mask)) { in domain_remove_cpu() 636 cpumask_set_cpu(cpu, &rdtgroup_default.cpu_mask); in resctrl_online_cpu() 648 if (cpumask_test_and_clear_cpu(cpu, &cr->cpu_mask)) { in clear_childcpus() 663 if (cpumask_test_and_clear_cpu(cpu, &rdtgrp->cpu_mask)) { in resctrl_offline_cpu()
|
A D | pseudo_lock.c | 210 for_each_cpu(cpu, &plr->d->cpu_mask) { in pseudo_lock_cstates_constrain() 286 plr->cpu = cpumask_first(&plr->d->cpu_mask); in pseudo_lock_region_init() 717 if (!cpumask_empty(&rdtgrp->cpu_mask)) { in rdtgroup_locksetup_enter() 842 &d_i->cpu_mask); in rdtgroup_pseudo_locked_in_hierarchy() 850 if (cpumask_intersects(&d->cpu_mask, cpu_with_psl)) in rdtgroup_pseudo_locked_in_hierarchy() 1178 cpu = cpumask_first(&plr->d->cpu_mask); in pseudo_lock_measure_cycles() 1508 if (!cpumask_subset(current->cpus_ptr, &plr->d->cpu_mask)) { in pseudo_lock_dev_mmap()
|
/linux/arch/sparc/mm/ |
A D | srmmu.c | 1658 cpumask_t cpu_mask; in smp_flush_cache_mm() local 1661 if (!cpumask_empty(&cpu_mask)) in smp_flush_cache_mm() 1670 cpumask_t cpu_mask; in smp_flush_tlb_mm() local 1690 cpumask_t cpu_mask; in smp_flush_cache_range() local 1693 if (!cpumask_empty(&cpu_mask)) in smp_flush_cache_range() 1707 cpumask_t cpu_mask; in smp_flush_tlb_range() local 1710 if (!cpumask_empty(&cpu_mask)) in smp_flush_tlb_range() 1722 cpumask_t cpu_mask; in smp_flush_cache_page() local 1737 cpumask_t cpu_mask; in smp_flush_tlb_page() local 1763 cpumask_t cpu_mask; in smp_flush_sig_insns() local [all …]
|
/linux/kernel/irq/ |
A D | affinity.c | 130 const struct cpumask *cpu_mask, in alloc_nodes_vectors() argument 145 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); in alloc_nodes_vectors() 251 const struct cpumask *cpu_mask, in __irq_build_affinity_masks() argument 261 if (!cpumask_weight(cpu_mask)) in __irq_build_affinity_masks() 264 nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk); in __irq_build_affinity_masks() 287 alloc_nodes_vectors(numvecs, node_to_cpumask, cpu_mask, in __irq_build_affinity_masks() 298 cpumask_and(nmsk, cpu_mask, node_to_cpumask[nv->id]); in __irq_build_affinity_masks()
|
/linux/arch/x86/events/amd/ |
A D | power.c | 41 static cpumask_t cpu_mask; variable 149 return cpumap_print_to_pagebuf(true, buf, &cpu_mask); in get_attr_cpumask() 223 if (!cpumask_test_and_clear_cpu(cpu, &cpu_mask)) in power_cpu_exit() 233 cpumask_set_cpu(target, &cpu_mask); in power_cpu_exit() 254 cpumask_set_cpu(cpu, &cpu_mask); in power_cpu_init()
|
/linux/kernel/trace/ |
A D | preemptirq_delay_test.c | 122 struct cpumask cpu_mask; in preemptirq_delay_run() local 125 cpumask_clear(&cpu_mask); in preemptirq_delay_run() 126 cpumask_set_cpu(cpu_affinity, &cpu_mask); in preemptirq_delay_run() 127 if (set_cpus_allowed_ptr(current, &cpu_mask)) in preemptirq_delay_run()
|
/linux/drivers/irqchip/ |
A D | irq-vf610-mscm-ir.c | 43 u16 cpu_mask; member 95 WARN_ON(irsprc & ~chip_data->cpu_mask); in vf610_mscm_ir_enable() 97 writew_relaxed(chip_data->cpu_mask, in vf610_mscm_ir_enable() 210 mscm_ir_data->cpu_mask = 0x1 << cpuid; in vf610_mscm_ir_of_init()
|
A D | irq-hip04.c | 272 unsigned int cpu_mask, cpu = smp_processor_id(); in hip04_irq_cpu_init() local 279 cpu_mask = hip04_get_cpumask(intc); in hip04_irq_cpu_init() 280 hip04_cpu_map[cpu] = cpu_mask; in hip04_irq_cpu_init() 288 hip04_cpu_map[i] &= ~cpu_mask; in hip04_irq_cpu_init()
|
/linux/arch/arm/mach-omap2/ |
A D | clkt2xxx_virt_prcm_set.c | 80 if (!(ptr->flags & cpu_mask)) in omap2_round_to_table_rate() 104 if (!(prcm->flags & cpu_mask)) in omap2_select_table_rate() 177 if (!(prcm->flags & cpu_mask)) in omap2xxx_clkt_vps_check_bootloader_rates()
|
/linux/tools/perf/tests/ |
A D | perf-record.c | 54 cpu_set_t cpu_mask; in test__PERF_RECORD() local 55 size_t cpu_mask_size = sizeof(cpu_mask); in test__PERF_RECORD() 114 err = sched__get_first_possible_cpu(evlist->workload.pid, &cpu_mask); in test__PERF_RECORD() 126 if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, &cpu_mask) < 0) { in test__PERF_RECORD()
|
/linux/tools/testing/selftests/bpf/prog_tests/ |
A D | cpu_mask.c | 33 const char *cpu_mask; member 65 err = parse_cpu_mask_str(test_cases[i].cpu_mask, &mask, &n); in test_cpu_mask()
|
/linux/drivers/soc/ti/ |
A D | knav_qmss_acc.c | 200 struct cpumask *cpu_mask; in knav_range_setup_acc_irq() local 207 cpu_mask = range->irqs[0].cpu_mask; in knav_range_setup_acc_irq() 211 cpu_mask = range->irqs[queue].cpu_mask; in knav_range_setup_acc_irq() 234 if (!ret && cpu_mask) { in knav_range_setup_acc_irq() 235 ret = irq_set_affinity_hint(irq, cpu_mask); in knav_range_setup_acc_irq()
|
/linux/arch/arm/mm/ |
A D | proc-arm9tdmi.S | 71 .macro arm9tdmi_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req 75 .long \cpu_mask
|
A D | proc-arm7tdmi.S | 77 .macro arm7tdmi_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req, \ 82 .long \cpu_mask
|
A D | proc-arm720.S | 177 .macro arm720_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req, cpu_flush:req 181 .long \cpu_mask
|
/linux/arch/arm/mach-omap1/ |
A D | clock_data.c | 761 u32 cpu_mask; variable 789 cpu_mask = 0; in omap1_clk_init() 791 cpu_mask |= CK_1710; in omap1_clk_init() 793 cpu_mask |= CK_16XX; in omap1_clk_init() 795 cpu_mask |= CK_1510; in omap1_clk_init() 797 cpu_mask |= CK_7XX; in omap1_clk_init() 799 cpu_mask |= CK_310; in omap1_clk_init() 802 if (c->cpu & cpu_mask) { in omap1_clk_init()
|
/linux/arch/x86/hyperv/ |
A D | hv_apic.c | 196 ipi_arg.cpu_mask = 0; in __send_ipi_mask() 212 __set_bit(vcpu, (unsigned long *)&ipi_arg.cpu_mask); in __send_ipi_mask() 216 ipi_arg.cpu_mask); in __send_ipi_mask()
|
/linux/fs/ |
A D | io-wq.c | 105 cpumask_var_t cpu_mask; member 712 set_cpus_allowed_ptr(tsk, wqe->cpu_mask); in io_init_new_worker() 1158 if (!alloc_cpumask_var(&wqe->cpu_mask, GFP_KERNEL)) in io_wq_create() 1160 cpumask_copy(wqe->cpu_mask, cpumask_of_node(node)); in io_wq_create() 1191 free_cpumask_var(wq->wqes[node]->cpu_mask); in io_wq_create() 1267 free_cpumask_var(wqe->cpu_mask); in io_wq_destroy() 1292 cpumask_set_cpu(od->cpu, worker->wqe->cpu_mask); in io_wq_worker_affinity() 1294 cpumask_clear_cpu(od->cpu, worker->wqe->cpu_mask); in io_wq_worker_affinity() 1336 cpumask_copy(wqe->cpu_mask, mask); in io_wq_cpu_affinity() 1338 cpumask_copy(wqe->cpu_mask, cpumask_of_node(i)); in io_wq_cpu_affinity()
|
/linux/tools/power/x86/intel_pstate_tracer/ |
A D | intel_pstate_tracer.py | 331 if cpu_mask[cpu_int] == 0: 352 if cpu_mask[int(index)] != 0: 513 cpu_mask = zeros((MAX_CPUS,), dtype=int) variable 546 cpu_mask[int(p)] = 1 549 cpu_mask[i] = 1
|
/linux/arch/ia64/kernel/ |
A D | msi_ia64.c | 17 const cpumask_t *cpu_mask, bool force) in ia64_set_msi_irq_affinity() argument 21 int cpu = cpumask_first_and(cpu_mask, cpu_online_mask); in ia64_set_msi_irq_affinity()
|
/linux/arch/arm/include/asm/ |
A D | procinfo.h | 28 unsigned int cpu_mask; member
|