/linux/virt/kvm/ |
A D | mmu_lock.h | 14 #define KVM_MMU_LOCK_INIT(kvm) rwlock_init(&(kvm)->mmu_lock) 15 #define KVM_MMU_LOCK(kvm) write_lock(&(kvm)->mmu_lock) 16 #define KVM_MMU_UNLOCK(kvm) write_unlock(&(kvm)->mmu_lock) 18 #define KVM_MMU_LOCK_INIT(kvm) spin_lock_init(&(kvm)->mmu_lock) 19 #define KVM_MMU_LOCK(kvm) spin_lock(&(kvm)->mmu_lock) 20 #define KVM_MMU_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock)
|
/linux/arch/riscv/kvm/ |
A D | mmu.c | 374 spin_lock(&kvm->mmu_lock); in stage2_wp_memory_region() 376 spin_unlock(&kvm->mmu_lock); in stage2_wp_memory_region() 404 spin_lock(&kvm->mmu_lock); in stage2_ioremap() 406 spin_unlock(&kvm->mmu_lock); in stage2_ioremap() 459 spin_lock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot() 461 spin_unlock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot() 560 spin_lock(&kvm->mmu_lock); in kvm_arch_prepare_memory_region() 564 spin_unlock(&kvm->mmu_lock); in kvm_arch_prepare_memory_region() 709 spin_lock(&kvm->mmu_lock); in kvm_riscv_stage2_map() 728 spin_unlock(&kvm->mmu_lock); in kvm_riscv_stage2_map() [all …]
|
/linux/arch/powerpc/kvm/ |
A D | book3s_hv_nested.c | 724 spin_lock(&kvm->mmu_lock); in kvmhv_remove_nested() 754 spin_lock(&kvm->mmu_lock); in kvmhv_release_all_nested() 783 spin_lock(&kvm->mmu_lock); in kvmhv_flush_nested() 801 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested() 813 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested() 839 spin_lock(&kvm->mmu_lock); in kvmhv_put_nested() 1036 spin_lock(&kvm->mmu_lock); in kvmhv_invalidate_shadow_pte() 1159 spin_lock(&kvm->mmu_lock); in kvmhv_emulate_tlbie_all_lpid() 1433 spin_lock(&kvm->mmu_lock); in kvmhv_handle_nested_set_rc() 1583 spin_lock(&kvm->mmu_lock); in __kvmhv_nested_page_fault() [all …]
|
A D | book3s_mmu_hpte.c | 63 spin_lock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map() 92 spin_unlock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map() 110 spin_lock(&vcpu3s->mmu_lock); in invalidate_pte() 114 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte() 127 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte() 369 spin_lock_init(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_init()
|
A D | book3s_64_mmu_radix.c | 635 spin_lock(&kvm->mmu_lock); in kvmppc_create_pte() 771 spin_unlock(&kvm->mmu_lock); in kvmppc_create_pte() 859 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page() 864 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page() 994 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault() 998 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault() 1104 spin_lock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty() 1116 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty() 1131 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty() 1176 spin_lock(&kvm->mmu_lock); in kvmppc_radix_flush_memslot() [all …]
|
A D | book3s_hv_rm_mmu.c | 254 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter() 269 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter() 283 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter() 944 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero() 956 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero() 972 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy() 987 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
|
A D | book3s_64_mmu_host.c | 153 spin_lock(&kvm->mmu_lock); in kvmppc_mmu_map_page() 205 spin_unlock(&kvm->mmu_lock); in kvmppc_mmu_map_page()
|
A D | book3s_64_vio_hv.c | 543 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect() 588 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()
|
A D | book3s_64_mmu_hv.c | 609 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault() 614 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault() 743 spin_lock(&kvm->mmu_lock); in kvmppc_rmap_reset() 750 spin_unlock(&kvm->mmu_lock); in kvmppc_rmap_reset() 1383 spin_lock(&kvm->mmu_lock); in resize_hpt_pivot() 1390 spin_unlock(&kvm->mmu_lock); in resize_hpt_pivot()
|
A D | e500_mmu_host.c | 462 spin_lock(&kvm->mmu_lock); in kvmppc_e500_shadow_map() 502 spin_unlock(&kvm->mmu_lock); in kvmppc_e500_shadow_map()
|
/linux/arch/arm64/kvm/ |
A D | mmu.c | 216 spin_lock(&kvm->mmu_lock); in stage2_flush_vm() 222 spin_unlock(&kvm->mmu_lock); in stage2_flush_vm() 602 spin_lock(&kvm->mmu_lock); in stage2_unmap_vm() 608 spin_unlock(&kvm->mmu_lock); in stage2_unmap_vm() 618 spin_lock(&kvm->mmu_lock); in kvm_free_stage2_pgd() 625 spin_unlock(&kvm->mmu_lock); in kvm_free_stage2_pgd() 662 spin_lock(&kvm->mmu_lock); in kvm_phys_addr_ioremap() 713 spin_lock(&kvm->mmu_lock); in kvm_mmu_wp_memory_region() 715 spin_unlock(&kvm->mmu_lock); in kvm_mmu_wp_memory_region() 1091 spin_lock(&kvm->mmu_lock); in user_mem_abort() [all …]
|
/linux/arch/x86/kvm/mmu/ |
A D | tdp_mmu.c | 36 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held() 38 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held() 199 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_get_vcpu_root_hpa() 287 lockdep_assert_held_write(&kvm->mmu_lock); in tdp_mmu_unlink_page() 507 lockdep_assert_held_read(&kvm->mmu_lock); in tdp_mmu_set_spte_atomic() 582 lockdep_assert_held_write(&kvm->mmu_lock); in __tdp_mmu_set_spte() 833 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_tdp_mmu_zap_invalidated_roots() 887 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_invalidate_all_roots() 1222 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_tdp_mmu_wrprot_slot() 1292 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_tdp_mmu_clear_dirty_slot() [all …]
|
A D | mmu.c | 2518 write_lock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages() 2540 write_lock(&kvm->mmu_lock); in kvm_mmu_unprotect_page() 3281 write_lock(&kvm->mmu_lock); in kvm_mmu_free_roots() 5665 write_lock(&kvm->mmu_lock); in kvm_mmu_zap_all_fast() 5701 read_lock(&kvm->mmu_lock); in kvm_mmu_zap_all_fast() 5778 write_lock(&kvm->mmu_lock); in kvm_zap_gfn_range() 5821 read_lock(&kvm->mmu_lock); in kvm_mmu_slot_remove_write_access() 5897 read_lock(&kvm->mmu_lock); in kvm_mmu_zap_collapsible_sptes() 5934 read_lock(&kvm->mmu_lock); in kvm_mmu_slot_leaf_clear_dirty() 5955 write_lock(&kvm->mmu_lock); in kvm_mmu_zap_all() [all …]
|
A D | page_track.c | 225 write_lock(&kvm->mmu_lock); in kvm_page_track_register_notifier() 227 write_unlock(&kvm->mmu_lock); in kvm_page_track_register_notifier() 243 write_lock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier() 245 write_unlock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier()
|
A D | tdp_mmu.h | 42 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_zap_sp()
|
A D | paging_tmpl.h | 913 write_lock(&vcpu->kvm->mmu_lock); in FNAME() 926 write_unlock(&vcpu->kvm->mmu_lock); in FNAME() 964 write_lock(&vcpu->kvm->mmu_lock); in FNAME() 999 write_unlock(&vcpu->kvm->mmu_lock); in FNAME()
|
/linux/arch/mips/kvm/ |
A D | mmu.c | 522 spin_lock(&kvm->mmu_lock); in _kvm_mips_map_page_fast() 557 spin_unlock(&kvm->mmu_lock); in _kvm_mips_map_page_fast() 639 spin_lock(&kvm->mmu_lock); in kvm_mips_map_page() 647 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page() 677 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page()
|
A D | mips.c | 227 spin_lock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot() 232 spin_unlock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot() 267 spin_lock(&kvm->mmu_lock); in kvm_arch_commit_memory_region() 273 spin_unlock(&kvm->mmu_lock); in kvm_arch_commit_memory_region()
|
/linux/arch/x86/kvm/ |
A D | debugfs.c | 110 write_lock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show() 130 write_unlock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show()
|
/linux/drivers/misc/habanalabs/common/ |
A D | command_buffer.c | 65 mutex_lock(&ctx->mmu_lock); in cb_map_mem() 85 mutex_unlock(&ctx->mmu_lock); in cb_map_mem() 102 mutex_unlock(&ctx->mmu_lock); in cb_map_mem() 119 mutex_lock(&ctx->mmu_lock); in cb_unmap_mem() 131 mutex_unlock(&ctx->mmu_lock); in cb_unmap_mem()
|
A D | memory.c | 1194 mutex_lock(&ctx->mmu_lock); in map_device_va() 1198 mutex_unlock(&ctx->mmu_lock); in map_device_va() 1207 mutex_unlock(&ctx->mmu_lock); in map_device_va() 1342 mutex_lock(&ctx->mmu_lock); in unmap_device_va() 1356 mutex_unlock(&ctx->mmu_lock); in unmap_device_va() 2633 mutex_lock(&ctx->mmu_lock); in hl_vm_ctx_fini() 2639 mutex_unlock(&ctx->mmu_lock); in hl_vm_ctx_fini()
|
/linux/drivers/misc/habanalabs/common/mmu/ |
A D | mmu.c | 84 mutex_init(&ctx->mmu_lock); in hl_mmu_ctx_init() 121 mutex_destroy(&ctx->mmu_lock); in hl_mmu_ctx_fini() 577 mutex_lock(&ctx->mmu_lock); in hl_mmu_get_tlb_info() 586 mutex_unlock(&ctx->mmu_lock); in hl_mmu_get_tlb_info()
|
/linux/Documentation/virt/kvm/ |
A D | locking.rst | 34 - kvm->arch.mmu_lock is an rwlock. kvm->arch.tdp_mmu_pages_lock and 35 kvm->arch.mmu_unsync_pages_lock are taken inside kvm->arch.mmu_lock, and 36 cannot be taken without already holding kvm->arch.mmu_lock (typically with 232 :Name: kvm->mmu_lock
|
/linux/arch/powerpc/include/asm/ |
A D | kvm_book3s_64.h | 666 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_secondary_pte() 678 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_host_pte()
|
/linux/drivers/gpu/drm/i915/gvt/ |
A D | kvmgt.c | 1809 write_lock(&kvm->mmu_lock); in kvmgt_page_track_add() 1818 write_unlock(&kvm->mmu_lock); in kvmgt_page_track_add() 1843 write_lock(&kvm->mmu_lock); in kvmgt_page_track_remove() 1852 write_unlock(&kvm->mmu_lock); in kvmgt_page_track_remove() 1878 write_lock(&kvm->mmu_lock); in kvmgt_page_track_flush_slot() 1887 write_unlock(&kvm->mmu_lock); in kvmgt_page_track_flush_slot()
|