Searched refs:hv_context (Results 1 – 4 of 4) sorted by relevance
26 struct hv_context hv_context; variable35 hv_context.cpu_context = alloc_percpu(struct hv_per_cpu_context); in hv_init()36 if (!hv_context.cpu_context) in hv_init()93 hv_cpu = get_cpu_ptr(hv_context.cpu_context); in hv_post_message()129 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()135 if (hv_context.hv_numa_map == NULL) { in hv_synic_alloc()141 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()189 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_free()196 kfree(hv_context.hv_numa_map); in hv_synic_free()209 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_enable_regs()[all …]
138 struct hv_context { struct153 extern struct hv_context hv_context; argument432 cpumask_set_cpu(cpu, &hv_context.hv_numa_map[cpu_to_node(cpu)]); in hv_set_alloced_cpu()439 cpumask_clear_cpu(cpu, &hv_context.hv_numa_map[cpu_to_node(cpu)]); in hv_clear_alloced_cpu()
763 alloced_mask = &hv_context.hv_numa_map[numa_node]; in init_vp_index()827 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_wait_for_unload()865 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_wait_for_unload()
1336 = this_cpu_ptr(hv_context.cpu_context); in vmbus_isr()2764 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_exit()
Completed in 13 milliseconds