/linux/arch/powerpc/kernel/ |
A D | watchdog.c | 159 if (cpumask_test_cpu(cpu, &wd_smp_cpus_pending)) in watchdog_smp_panic() 207 if (!cpumask_test_cpu(cpu, &wd_smp_cpus_pending)) { in wd_smp_clear_cpu_pending() 208 if (unlikely(cpumask_test_cpu(cpu, &wd_smp_cpus_stuck))) { in wd_smp_clear_cpu_pending() 263 if (!cpumask_test_cpu(cpu, &wd_cpus_enabled)) in DEFINE_INTERRUPT_HANDLER_NMI() 271 if (cpumask_test_cpu(cpu, &wd_smp_cpus_stuck)) { in DEFINE_INTERRUPT_HANDLER_NMI() 307 if (!cpumask_test_cpu(cpu, &watchdog_cpumask)) in watchdog_timer_fn() 336 if (cpumask_test_cpu(cpu, &wd_cpus_enabled)) { in start_watchdog() 344 if (!cpumask_test_cpu(cpu, &watchdog_cpumask)) in start_watchdog() 374 if (!cpumask_test_cpu(cpu, &wd_cpus_enabled)) in stop_watchdog()
|
/linux/arch/alpha/kernel/ |
A D | sys_titan.c | 87 if (!cpumask_test_cpu(0, &cpm)) dim0 = &dummy; in titan_update_irq_hw() 88 if (!cpumask_test_cpu(1, &cpm)) dim1 = &dummy; in titan_update_irq_hw() 89 if (!cpumask_test_cpu(2, &cpm)) dim2 = &dummy; in titan_update_irq_hw() 90 if (!cpumask_test_cpu(3, &cpm)) dim3 = &dummy; in titan_update_irq_hw() 140 if (cpumask_test_cpu(cpu, &affinity)) in titan_cpu_set_irq_affinity()
|
/linux/arch/sparc/kernel/ |
A D | sun4m_smp.c | 66 while (!cpumask_test_cpu(cpuid, &smp_commenced_mask)) in sun4m_cpu_pre_online() 197 if (cpumask_test_cpu(i, &mask)) { in sun4m_cross_call() 213 if (!cpumask_test_cpu(i, &mask)) in sun4m_cross_call() 221 if (!cpumask_test_cpu(i, &mask)) in sun4m_cross_call()
|
A D | sun4d_smp.c | 103 while (!cpumask_test_cpu(cpuid, &smp_commenced_mask)) in sun4d_cpu_pre_online() 321 if (cpumask_test_cpu(i, &mask)) { in sun4d_cross_call() 334 if (!cpumask_test_cpu(i, &mask)) in sun4d_cross_call() 342 if (!cpumask_test_cpu(i, &mask)) in sun4d_cross_call()
|
A D | leon_smp.c | 98 while (!cpumask_test_cpu(cpuid, &smp_commenced_mask)) in leon_cpu_pre_online() 409 if (cpumask_test_cpu(i, &mask)) { in leon_cross_call() 423 if (!cpumask_test_cpu(i, &mask)) in leon_cross_call() 432 if (!cpumask_test_cpu(i, &mask)) in leon_cross_call()
|
/linux/kernel/irq/ |
A D | cpuhotplug.c | 50 return cpumask_test_cpu(cpu, m); in irq_needs_fixup() 186 return cpumask_test_cpu(cpu, hk_mask); in hk_should_isolate() 195 !irq_data_get_irq_chip(data) || !cpumask_test_cpu(cpu, affinity)) in irq_restore_affinity_of_irq()
|
A D | ipi.c | 170 if (!cpumask_test_cpu(cpu, ipimask)) in ipi_get_hwirq() 204 if (!cpumask_test_cpu(cpu, ipimask)) in ipi_send_verify()
|
/linux/kernel/time/ |
A D | tick-broadcast.c | 283 if (!cpumask_test_cpu(cpu, tick_broadcast_on)) in tick_device_uses_broadcast() 317 ret = cpumask_test_cpu(cpu, tick_broadcast_mask); in tick_device_uses_broadcast() 354 if (cpumask_test_cpu(cpu, mask)) { in tick_do_broadcast() 575 return cpumask_test_cpu(smp_processor_id(), tick_broadcast_mask); in tick_resume_check_broadcast() 627 return cpumask_test_cpu(smp_processor_id(), tick_broadcast_force_mask); in tick_check_broadcast_expired() 667 if (cpumask_test_cpu(smp_processor_id(), tick_broadcast_oneshot_mask)) { in tick_check_oneshot_broadcast_this_cpu() 826 WARN_ON_ONCE(cpumask_test_cpu(cpu, tick_broadcast_pending_mask)); in ___tick_broadcast_oneshot_control() 841 if (cpumask_test_cpu(cpu, tick_broadcast_force_mask)) { in ___tick_broadcast_oneshot_control()
|
/linux/drivers/perf/ |
A D | arm_pmu.c | 333 if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus)) in armpmu_add() 510 !cpumask_test_cpu(event->cpu, &armpmu->supported_cpus)) in armpmu_event_init() 530 if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus)) in armpmu_enable() 542 if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus)) in armpmu_disable() 559 ret = cpumask_test_cpu(cpu, &armpmu->supported_cpus); in armpmu_filter_match() 711 if (!cpumask_test_cpu(cpu, &pmu->supported_cpus)) in arm_perf_starting_cpu() 730 if (!cpumask_test_cpu(cpu, &pmu->supported_cpus)) in arm_perf_teardown_cpu() 790 if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus)) in cpu_pm_pmu_notify()
|
A D | arm_dsu_pmu.c | 255 if (WARN_ON(!cpumask_test_cpu(smp_processor_id(), in dsu_pmu_read_counter() 281 if (WARN_ON(!cpumask_test_cpu(smp_processor_id(), in dsu_pmu_write_counter() 447 if (WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(), in dsu_pmu_add() 567 if (!cpumask_test_cpu(event->cpu, &dsu_pmu->associated_cpus)) { in dsu_pmu_event_init() 817 if (!cpumask_test_cpu(cpu, &dsu_pmu->associated_cpus)) in dsu_pmu_cpu_online()
|
/linux/arch/ia64/kernel/ |
A D | numa.c | 30 if (cpumask_test_cpu(cpu, &node_to_cpu_mask[oldnid])) { in map_cpu_to_node() 44 WARN_ON(!cpumask_test_cpu(cpu, &node_to_cpu_mask[nid])); in unmap_cpu_from_node()
|
A D | irq_ia64.c | 240 if (!cpumask_test_cpu(cpu, &irq_cfg[irq].domain)) in __setup_vector_irq() 271 if (cpumask_test_cpu(cpu, &cfg->domain)) in __irq_prepare_move() 305 if (unlikely(cpumask_test_cpu(smp_processor_id(), &cfg->old_domain))) in irq_complete_move() 336 if (!cpumask_test_cpu(me, &cfg->old_domain)) in smp_irq_move_cleanup_interrupt()
|
/linux/lib/ |
A D | nmi_backtrace.c | 61 if (cpumask_test_cpu(this_cpu, to_cpumask(backtrace_mask))) in nmi_trigger_cpumask_backtrace() 97 if (cpumask_test_cpu(cpu, to_cpumask(backtrace_mask))) { in nmi_cpu_backtrace()
|
/linux/include/linux/ |
A D | cpumask.h | 342 static inline int cpumask_test_cpu(int cpu, const struct cpumask *cpumask) in cpumask_test_cpu() function 895 return cpumask_test_cpu(cpu, cpu_online_mask); in cpu_online() 900 return cpumask_test_cpu(cpu, cpu_possible_mask); in cpu_possible() 905 return cpumask_test_cpu(cpu, cpu_present_mask); in cpu_present() 910 return cpumask_test_cpu(cpu, cpu_active_mask); in cpu_active() 915 return cpumask_test_cpu(cpu, cpu_dying_mask); in cpu_dying()
|
A D | mmu_context.h | 28 # define task_cpu_possible(cpu, p) cpumask_test_cpu((cpu), task_cpu_possible_mask(p))
|
/linux/arch/arm/include/asm/ |
A D | cacheflush.h | 220 if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) in vivt_flush_cache_mm() 229 if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) in vivt_flush_cache_range() 239 if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) { in vivt_flush_cache_page()
|
/linux/arch/arm/mach-tegra/ |
A D | platsmp.c | 93 if (cpumask_test_cpu(cpu, &tegra_cpu_init_mask)) { in tegra30_boot_secondary() 135 if (cpumask_test_cpu(cpu, &tegra_cpu_init_mask)) { in tegra114_boot_secondary()
|
/linux/kernel/trace/ |
A D | ring_buffer.c | 917 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_wait() 1020 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_poll_wait() 3830 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_write() 4019 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_record_disable_cpu() 4039 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_record_enable_cpu() 4072 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_oldest_event_ts() 4103 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_bytes_cpu() 4122 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_entries_cpu() 4142 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_overrun_cpu() 4165 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_commit_overrun_cpu() [all …]
|
/linux/drivers/hwtracing/coresight/ |
A D | coresight-trbe.c | 1055 if (!cpumask_test_cpu(cpu, &drvdata->supported_cpus)) in is_perf_trbe() 1299 if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) in arm_trbe_probe_coresight() 1301 if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) in arm_trbe_probe_coresight() 1328 if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) { in arm_trbe_cpu_startup() 1336 if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) in arm_trbe_cpu_startup() 1338 if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) in arm_trbe_cpu_startup() 1351 if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) { in arm_trbe_cpu_teardown()
|
/linux/drivers/cpuidle/ |
A D | coupled.c | 429 if (!cpumask_test_cpu(cpu, &cpuidle_coupled_poke_pending)) in cpuidle_coupled_clear_pokes() 433 while (cpumask_test_cpu(cpu, &cpuidle_coupled_poke_pending)) in cpuidle_coupled_clear_pokes() 520 !cpumask_test_cpu(dev->cpu, &cpuidle_coupled_poked)) { in cpuidle_enter_state_coupled()
|
/linux/drivers/sh/intc/ |
A D | chip.c | 25 if (!cpumask_test_cpu(cpu, irq_data_get_affinity_mask(data))) in _intc_enable() 53 if (!cpumask_test_cpu(cpu, irq_data_get_affinity_mask(data))) in intc_disable()
|
/linux/drivers/virt/nitro_enclaves/ |
A D | ne_misc_dev.c | 242 if (cpumask_test_cpu(0, cpu_pool)) { in ne_setup_cpu_pool() 251 if (cpumask_test_cpu(cpu_sibling, cpu_pool)) { in ne_setup_cpu_pool() 268 if (!cpumask_test_cpu(cpu_sibling, cpu_pool)) { in ne_setup_cpu_pool() 471 if (cpumask_test_cpu(cpu, ne_enclave->vcpu_ids)) in ne_donated_cpu() 621 if (cpumask_test_cpu(vcpu_id, ne_cpu_pool.avail_threads_per_core[i])) { in ne_get_vcpu_core_from_cpu_pool() 659 if (cpumask_test_cpu(vcpu_id, ne_enclave->threads_per_core[i])) in ne_check_cpu_in_cpu_pool() 1051 if (!cpumask_test_cpu(cpu, ne_enclave->vcpu_ids)) { in ne_start_enclave_ioctl()
|
/linux/arch/arm/mach-omap2/ |
A D | cpuidle44xx.c | 133 if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) { in omap_enter_idle_coupled() 191 if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) { in omap_enter_idle_coupled()
|
/linux/arch/s390/kernel/ |
A D | topology.c | 70 if (!cpumask_test_cpu(cpu, &cpu_setup_mask)) in cpu_group_map() 76 if (cpumask_test_cpu(cpu, &info->mask)) { in cpu_group_map() 102 if (!cpumask_test_cpu(cpu, &cpu_setup_mask)) in cpu_thread_map() 109 if (cpumask_test_cpu(cpu + i, &cpu_setup_mask)) in cpu_thread_map()
|
/linux/drivers/net/wireguard/ |
A D | queueing.h | 112 !cpumask_test_cpu(cpu, cpu_online_mask))) { in wg_cpumask_choose_online() 133 while (unlikely(!cpumask_test_cpu(cpu, cpu_online_mask))) in wg_cpumask_next_online()
|