Lines Matching refs:host_kvm

24 struct host_kvm host_kvm;  variable
72 host_kvm.mm_ops = (struct kvm_pgtable_mm_ops) { in prepare_s2_pool()
93 host_kvm.arch.vtcr = kvm_get_vtcr(id_aa64mmfr0_el1_sys_val, in prepare_host_vtcr()
101 struct kvm_s2_mmu *mmu = &host_kvm.arch.mmu; in kvm_host_prepare_stage2()
105 hyp_spin_lock_init(&host_kvm.lock); in kvm_host_prepare_stage2()
111 ret = __kvm_pgtable_stage2_init(&host_kvm.pgt, &host_kvm.arch, in kvm_host_prepare_stage2()
112 &host_kvm.mm_ops, KVM_HOST_S2_FLAGS, in kvm_host_prepare_stage2()
117 mmu->pgd_phys = __hyp_pa(host_kvm.pgt.pgd); in kvm_host_prepare_stage2()
118 mmu->arch = &host_kvm.arch; in kvm_host_prepare_stage2()
119 mmu->pgt = &host_kvm.pgt; in kvm_host_prepare_stage2()
128 struct kvm_s2_mmu *mmu = &host_kvm.arch.mmu; in __pkvm_prot_finalize()
135 params->vtcr = host_kvm.arch.vtcr; in __pkvm_prot_finalize()
140 __load_stage2(&host_kvm.arch.mmu, &host_kvm.arch); in __pkvm_prot_finalize()
158 struct kvm_pgtable *pgt = &host_kvm.pgt; in host_stage2_unmap_dev_all()
233 return kvm_pgtable_stage2_map(&host_kvm.pgt, start, end - start, start, in __host_stage2_idmap()
246 hyp_assert_lock_held(&host_kvm.lock); \
269 hyp_assert_lock_held(&host_kvm.lock); in host_stage2_adjust_range()
270 ret = kvm_pgtable_get_leaf(&host_kvm.pgt, addr, &pte, &level); in host_stage2_adjust_range()
297 hyp_assert_lock_held(&host_kvm.lock); in host_stage2_idmap_locked()
304 hyp_assert_lock_held(&host_kvm.lock); in host_stage2_set_owner_locked()
306 return host_stage2_try(kvm_pgtable_stage2_set_owner, &host_kvm.pgt, in host_stage2_set_owner_locked()
341 hyp_spin_lock(&host_kvm.lock); in host_stage2_idmap()
348 hyp_spin_unlock(&host_kvm.lock); in host_stage2_idmap()
372 hyp_spin_lock(&host_kvm.lock); in __pkvm_host_share_hyp()
375 ret = kvm_pgtable_get_leaf(&host_kvm.pgt, addr, &pte, NULL); in __pkvm_host_share_hyp()
436 hyp_spin_unlock(&host_kvm.lock); in __pkvm_host_share_hyp()