Home
last modified time | relevance | path

Searched refs:mmu_lock (Results 1 – 25 of 32) sorted by relevance

12

/linux/virt/kvm/
A Dmmu_lock.h14 #define KVM_MMU_LOCK_INIT(kvm) rwlock_init(&(kvm)->mmu_lock)
15 #define KVM_MMU_LOCK(kvm) write_lock(&(kvm)->mmu_lock)
16 #define KVM_MMU_UNLOCK(kvm) write_unlock(&(kvm)->mmu_lock)
18 #define KVM_MMU_LOCK_INIT(kvm) spin_lock_init(&(kvm)->mmu_lock)
19 #define KVM_MMU_LOCK(kvm) spin_lock(&(kvm)->mmu_lock)
20 #define KVM_MMU_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock)
/linux/arch/riscv/kvm/
A Dmmu.c374 spin_lock(&kvm->mmu_lock); in stage2_wp_memory_region()
376 spin_unlock(&kvm->mmu_lock); in stage2_wp_memory_region()
404 spin_lock(&kvm->mmu_lock); in stage2_ioremap()
406 spin_unlock(&kvm->mmu_lock); in stage2_ioremap()
459 spin_lock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()
461 spin_unlock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()
560 spin_lock(&kvm->mmu_lock); in kvm_arch_prepare_memory_region()
564 spin_unlock(&kvm->mmu_lock); in kvm_arch_prepare_memory_region()
709 spin_lock(&kvm->mmu_lock); in kvm_riscv_stage2_map()
728 spin_unlock(&kvm->mmu_lock); in kvm_riscv_stage2_map()
[all …]
/linux/arch/powerpc/kvm/
A Dbook3s_hv_nested.c724 spin_lock(&kvm->mmu_lock); in kvmhv_remove_nested()
754 spin_lock(&kvm->mmu_lock); in kvmhv_release_all_nested()
783 spin_lock(&kvm->mmu_lock); in kvmhv_flush_nested()
801 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested()
813 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested()
839 spin_lock(&kvm->mmu_lock); in kvmhv_put_nested()
1036 spin_lock(&kvm->mmu_lock); in kvmhv_invalidate_shadow_pte()
1159 spin_lock(&kvm->mmu_lock); in kvmhv_emulate_tlbie_all_lpid()
1433 spin_lock(&kvm->mmu_lock); in kvmhv_handle_nested_set_rc()
1583 spin_lock(&kvm->mmu_lock); in __kvmhv_nested_page_fault()
[all …]
A Dbook3s_mmu_hpte.c63 spin_lock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map()
92 spin_unlock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map()
110 spin_lock(&vcpu3s->mmu_lock); in invalidate_pte()
114 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte()
127 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte()
369 spin_lock_init(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_init()
A Dbook3s_64_mmu_radix.c635 spin_lock(&kvm->mmu_lock); in kvmppc_create_pte()
771 spin_unlock(&kvm->mmu_lock); in kvmppc_create_pte()
859 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page()
864 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page()
994 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault()
998 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault()
1104 spin_lock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()
1116 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()
1131 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()
1176 spin_lock(&kvm->mmu_lock); in kvmppc_radix_flush_memslot()
[all …]
A Dbook3s_hv_rm_mmu.c254 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()
269 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()
283 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()
944 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()
956 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()
972 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
987 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
A Dbook3s_64_mmu_host.c153 spin_lock(&kvm->mmu_lock); in kvmppc_mmu_map_page()
205 spin_unlock(&kvm->mmu_lock); in kvmppc_mmu_map_page()
A Dbook3s_64_vio_hv.c543 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()
588 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()
A Dbook3s_64_mmu_hv.c609 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault()
614 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault()
743 spin_lock(&kvm->mmu_lock); in kvmppc_rmap_reset()
750 spin_unlock(&kvm->mmu_lock); in kvmppc_rmap_reset()
1383 spin_lock(&kvm->mmu_lock); in resize_hpt_pivot()
1390 spin_unlock(&kvm->mmu_lock); in resize_hpt_pivot()
A De500_mmu_host.c462 spin_lock(&kvm->mmu_lock); in kvmppc_e500_shadow_map()
502 spin_unlock(&kvm->mmu_lock); in kvmppc_e500_shadow_map()
/linux/arch/arm64/kvm/
A Dmmu.c216 spin_lock(&kvm->mmu_lock); in stage2_flush_vm()
222 spin_unlock(&kvm->mmu_lock); in stage2_flush_vm()
602 spin_lock(&kvm->mmu_lock); in stage2_unmap_vm()
608 spin_unlock(&kvm->mmu_lock); in stage2_unmap_vm()
618 spin_lock(&kvm->mmu_lock); in kvm_free_stage2_pgd()
625 spin_unlock(&kvm->mmu_lock); in kvm_free_stage2_pgd()
662 spin_lock(&kvm->mmu_lock); in kvm_phys_addr_ioremap()
713 spin_lock(&kvm->mmu_lock); in kvm_mmu_wp_memory_region()
715 spin_unlock(&kvm->mmu_lock); in kvm_mmu_wp_memory_region()
1091 spin_lock(&kvm->mmu_lock); in user_mem_abort()
[all …]
/linux/arch/x86/kvm/mmu/
A Dtdp_mmu.c36 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()
38 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()
199 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_get_vcpu_root_hpa()
287 lockdep_assert_held_write(&kvm->mmu_lock); in tdp_mmu_unlink_page()
507 lockdep_assert_held_read(&kvm->mmu_lock); in tdp_mmu_set_spte_atomic()
582 lockdep_assert_held_write(&kvm->mmu_lock); in __tdp_mmu_set_spte()
833 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_tdp_mmu_zap_invalidated_roots()
887 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_invalidate_all_roots()
1222 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_tdp_mmu_wrprot_slot()
1292 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_tdp_mmu_clear_dirty_slot()
[all …]
A Dmmu.c2518 write_lock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()
2540 write_lock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()
3281 write_lock(&kvm->mmu_lock); in kvm_mmu_free_roots()
5665 write_lock(&kvm->mmu_lock); in kvm_mmu_zap_all_fast()
5701 read_lock(&kvm->mmu_lock); in kvm_mmu_zap_all_fast()
5778 write_lock(&kvm->mmu_lock); in kvm_zap_gfn_range()
5821 read_lock(&kvm->mmu_lock); in kvm_mmu_slot_remove_write_access()
5897 read_lock(&kvm->mmu_lock); in kvm_mmu_zap_collapsible_sptes()
5934 read_lock(&kvm->mmu_lock); in kvm_mmu_slot_leaf_clear_dirty()
5955 write_lock(&kvm->mmu_lock); in kvm_mmu_zap_all()
[all …]
A Dpage_track.c225 write_lock(&kvm->mmu_lock); in kvm_page_track_register_notifier()
227 write_unlock(&kvm->mmu_lock); in kvm_page_track_register_notifier()
243 write_lock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier()
245 write_unlock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier()
A Dtdp_mmu.h42 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_zap_sp()
A Dpaging_tmpl.h913 write_lock(&vcpu->kvm->mmu_lock); in FNAME()
926 write_unlock(&vcpu->kvm->mmu_lock); in FNAME()
964 write_lock(&vcpu->kvm->mmu_lock); in FNAME()
999 write_unlock(&vcpu->kvm->mmu_lock); in FNAME()
/linux/arch/mips/kvm/
A Dmmu.c522 spin_lock(&kvm->mmu_lock); in _kvm_mips_map_page_fast()
557 spin_unlock(&kvm->mmu_lock); in _kvm_mips_map_page_fast()
639 spin_lock(&kvm->mmu_lock); in kvm_mips_map_page()
647 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page()
677 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page()
A Dmips.c227 spin_lock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()
232 spin_unlock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()
267 spin_lock(&kvm->mmu_lock); in kvm_arch_commit_memory_region()
273 spin_unlock(&kvm->mmu_lock); in kvm_arch_commit_memory_region()
/linux/arch/x86/kvm/
A Ddebugfs.c110 write_lock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show()
130 write_unlock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show()
/linux/drivers/misc/habanalabs/common/
A Dcommand_buffer.c65 mutex_lock(&ctx->mmu_lock); in cb_map_mem()
85 mutex_unlock(&ctx->mmu_lock); in cb_map_mem()
102 mutex_unlock(&ctx->mmu_lock); in cb_map_mem()
119 mutex_lock(&ctx->mmu_lock); in cb_unmap_mem()
131 mutex_unlock(&ctx->mmu_lock); in cb_unmap_mem()
A Dmemory.c1194 mutex_lock(&ctx->mmu_lock); in map_device_va()
1198 mutex_unlock(&ctx->mmu_lock); in map_device_va()
1207 mutex_unlock(&ctx->mmu_lock); in map_device_va()
1342 mutex_lock(&ctx->mmu_lock); in unmap_device_va()
1356 mutex_unlock(&ctx->mmu_lock); in unmap_device_va()
2633 mutex_lock(&ctx->mmu_lock); in hl_vm_ctx_fini()
2639 mutex_unlock(&ctx->mmu_lock); in hl_vm_ctx_fini()
/linux/drivers/misc/habanalabs/common/mmu/
A Dmmu.c84 mutex_init(&ctx->mmu_lock); in hl_mmu_ctx_init()
121 mutex_destroy(&ctx->mmu_lock); in hl_mmu_ctx_fini()
577 mutex_lock(&ctx->mmu_lock); in hl_mmu_get_tlb_info()
586 mutex_unlock(&ctx->mmu_lock); in hl_mmu_get_tlb_info()
/linux/Documentation/virt/kvm/
A Dlocking.rst34 - kvm->arch.mmu_lock is an rwlock. kvm->arch.tdp_mmu_pages_lock and
35 kvm->arch.mmu_unsync_pages_lock are taken inside kvm->arch.mmu_lock, and
36 cannot be taken without already holding kvm->arch.mmu_lock (typically with
232 :Name: kvm->mmu_lock
/linux/arch/powerpc/include/asm/
A Dkvm_book3s_64.h666 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_secondary_pte()
678 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_host_pte()
/linux/drivers/gpu/drm/i915/gvt/
A Dkvmgt.c1809 write_lock(&kvm->mmu_lock); in kvmgt_page_track_add()
1818 write_unlock(&kvm->mmu_lock); in kvmgt_page_track_add()
1843 write_lock(&kvm->mmu_lock); in kvmgt_page_track_remove()
1852 write_unlock(&kvm->mmu_lock); in kvmgt_page_track_remove()
1878 write_lock(&kvm->mmu_lock); in kvmgt_page_track_flush_slot()
1887 write_unlock(&kvm->mmu_lock); in kvmgt_page_track_flush_slot()

Completed in 224 milliseconds

12