/linux/include/linux/ |
A D | rcu_node_tree.h | 54 #if NR_CPUS <= RCU_FANOUT_1 61 #elif NR_CPUS <= RCU_FANOUT_2 64 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 69 #elif NR_CPUS <= RCU_FANOUT_3 72 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2) 73 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 78 #elif NR_CPUS <= RCU_FANOUT_4 81 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_3) 82 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2) 83 # define NUM_RCU_LVL_3 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1)
|
A D | cpumask.h | 36 #if NR_CPUS == 1 47 #define nr_cpumask_bits ((unsigned int)NR_CPUS) 119 #if NR_CPUS == 1 296 [0 ... BITS_TO_LONGS(NR_CPUS)-1] = 0UL \ 876 #if NR_CPUS > 1 954 #if NR_CPUS <= BITS_PER_LONG 957 [BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \ 965 [BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \ 1025 #if NR_CPUS <= BITS_PER_LONG 1028 [BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \ [all …]
|
A D | blockgroup_lock.h | 14 #define NR_BG_LOCKS (4 << ilog2(NR_CPUS < 32 ? NR_CPUS : 32))
|
/linux/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
A D | preempt.c | 26 struct lock_impl cpu_preemption_locks[NR_CPUS] = { 28 #if (NR_CPUS - 1) & 1 31 #if (NR_CPUS - 1) & 2 34 #if (NR_CPUS - 1) & 4 37 #if (NR_CPUS - 1) & 8 40 #if (NR_CPUS - 1) & 16 43 #if (NR_CPUS - 1) & 32 67 assume(thread_cpu_id < NR_CPUS); in preempt_disable()
|
/linux/arch/powerpc/include/asm/ |
A D | irq.h | 44 extern void *critirq_ctx[NR_CPUS]; 45 extern void *dbgirq_ctx[NR_CPUS]; 46 extern void *mcheckirq_ctx[NR_CPUS]; 52 extern void *hardirq_ctx[NR_CPUS]; 53 extern void *softirq_ctx[NR_CPUS];
|
/linux/tools/testing/selftests/fpu/ |
A D | run_test_fpu.sh | 23 NR_CPUS=$(getconf _NPROCESSORS_ONLN) 24 if [ ! $NR_CPUS ]; then 25 NR_CPUS=1 41 for c in $(seq 1 $NR_CPUS); do
|
/linux/arch/ia64/kernel/ |
A D | err_inject.c | 43 static u64 call_start[NR_CPUS]; 44 static u64 phys_addr[NR_CPUS]; 45 static u64 err_type_info[NR_CPUS]; 46 static u64 err_struct_info[NR_CPUS]; 51 } __attribute__((__aligned__(16))) err_data_buffer[NR_CPUS]; 52 static s64 status[NR_CPUS]; 53 static u64 capabilities[NR_CPUS]; 54 static u64 resources[NR_CPUS];
|
/linux/arch/riscv/kernel/ |
A D | cpu_ops.c | 16 const struct cpu_operations *cpu_ops[NR_CPUS] __ro_after_init; 18 void *__cpu_up_stack_pointer[NR_CPUS] __section(".data"); 19 void *__cpu_up_task_pointer[NR_CPUS] __section(".data");
|
A D | smp.c | 35 unsigned long __cpuid_to_hartid_map[NR_CPUS] __ro_after_init = { 36 [0 ... NR_CPUS-1] = INVALID_HARTID 48 } ipi_data[NR_CPUS] __cacheline_aligned; 54 for (i = 0; i < NR_CPUS; i++) in riscv_hartid_to_cpuid()
|
/linux/arch/sparc/include/asm/ |
A D | topology_64.h | 53 extern cpumask_t cpu_core_map[NR_CPUS]; 54 extern cpumask_t cpu_core_sib_map[NR_CPUS]; 55 extern cpumask_t cpu_core_sib_cache_map[NR_CPUS];
|
/linux/arch/ia64/include/asm/ |
A D | smp.h | 50 int cpu_phys_id[NR_CPUS]; 55 extern cpumask_t cpu_core_map[NR_CPUS]; 74 for (i = 0; i < NR_CPUS; ++i) in cpu_logical_id()
|
/linux/arch/sparc/kernel/ |
A D | leon_smp.c | 54 extern volatile unsigned long cpu_callin_map[NR_CPUS]; 167 (unsigned int)nrcpu, (unsigned int)NR_CPUS, in leon_boot_cpus() 237 for (i = 0; i < NR_CPUS; i++) { in leon_smp_done() 368 unsigned long processors_in[NR_CPUS]; /* Set when ipi entered. */ 369 unsigned long processors_out[NR_CPUS]; /* Set when ipi exited. */ 380 register int high = NR_CPUS - 1; in leon_cross_call()
|
A D | smp_32.c | 42 volatile unsigned long cpu_callin_map[NR_CPUS] = {0,}; 190 if (cpuid >= NR_CPUS) in smp_prepare_cpus() 234 if (mid < NR_CPUS) { in smp_setup_cpu_possible_map() 246 if (cpuid >= NR_CPUS) { in smp_prepare_boot_cpu()
|
/linux/arch/arm/mach-shmobile/ |
A D | headsmp.S | 118 cmp r1, #NR_CPUS 141 .space NR_CPUS * 4 144 .space NR_CPUS * 4 147 .space NR_CPUS * 4
|
/linux/arch/sh/kernel/ |
A D | irq.c | 66 static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly; 67 static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly; 69 static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss; 70 static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
|
/linux/arch/ia64/include/asm/native/ |
A D | irq.h | 14 #if (NR_VECTORS + 32 * NR_CPUS) < 1024 15 #define IA64_NATIVE_NR_IRQS (NR_VECTORS + 32 * NR_CPUS)
|
/linux/arch/alpha/kernel/ |
A D | smp.c | 54 struct cpuinfo_alpha cpu_data[NR_CPUS]; 60 } ipi_data[NR_CPUS] __cacheline_aligned; 246 for (i = 0; i < NR_CPUS; i++) { in recv_secondary_console_msg() 489 for(cpu = 0; cpu < NR_CPUS; cpu++) in smp_cpus_done() 653 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_tlb_mm() 700 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_tlb_page() 754 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_icache_user_page()
|
/linux/arch/arm64/kvm/hyp/nvhe/ |
A D | hyp-smp.c | 17 u64 __ro_after_init hyp_cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = INVALID_HWID };
|
/linux/tools/testing/selftests/arm64/fp/ |
A D | sve-stress | 8 NR_CPUS=`nproc` 42 for x in `seq 0 $((NR_CPUS * 4))`; do
|
A D | fpsimd-stress | 8 NR_CPUS=`nproc` 43 for x in `seq 0 $((NR_CPUS * 4))`; do
|
/linux/arch/x86/kernel/ |
A D | cpuid.c | 151 if (__register_chrdev(CPUID_MAJOR, 0, NR_CPUS, in cpuid_init() 175 __unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid"); in cpuid_init() 184 __unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid"); in cpuid_exit()
|
/linux/arch/powerpc/lib/ |
A D | locks.c | 29 BUG_ON(holder_cpu >= NR_CPUS); in splpar_spin_yield() 55 BUG_ON(holder_cpu >= NR_CPUS); in splpar_rw_yield()
|
/linux/arch/ia64/mm/ |
A D | numa.c | 29 struct node_cpuid_s node_cpuid[NR_CPUS] = 30 { [0 ... NR_CPUS-1] = { .phys_id = 0, .nid = NUMA_NO_NODE } };
|
/linux/drivers/acpi/ |
A D | acpi_pad.c | 82 static unsigned long cpu_weight[NR_CPUS]; 83 static int tsk_in_cpu[NR_CPUS] = {[0 ... NR_CPUS-1] = -1}; 84 static DECLARE_BITMAP(pad_busy_cpus_bits, NR_CPUS); 207 static struct task_struct *ps_tsks[NR_CPUS];
|
/linux/arch/mips/kernel/ |
A D | smp-cps.c | 28 static DECLARE_BITMAP(core_power, NR_CPUS); 74 for (v = 0; v < min_t(int, core_vpes, NR_CPUS - nvpes); v++) { in cps_smp_setup() 88 for (v = 0; v < min_t(unsigned, nvpes, NR_CPUS); v++) { in cps_smp_setup() 332 for (remote = 0; remote < NR_CPUS; remote++) { in cps_boot_secondary() 338 if (remote >= NR_CPUS) { in cps_boot_secondary()
|