Lines Matching refs:hptep

340 	__be64 *hptep;  in kvmppc_mmu_book3s_64_hv_xlate()  local
366 hptep = (__be64 *)(kvm->arch.hpt.virt + (index << 4)); in kvmppc_mmu_book3s_64_hv_xlate()
367 v = orig_v = be64_to_cpu(hptep[0]) & ~HPTE_V_HVLOCK; in kvmppc_mmu_book3s_64_hv_xlate()
369 v = hpte_new_to_old_v(v, be64_to_cpu(hptep[1])); in kvmppc_mmu_book3s_64_hv_xlate()
372 unlock_hpte(hptep, orig_v); in kvmppc_mmu_book3s_64_hv_xlate()
489 __be64 *hptep; in kvmppc_book3s_hv_page_fault() local
531 hptep = (__be64 *)(kvm->arch.hpt.virt + (index << 4)); in kvmppc_book3s_hv_page_fault()
534 while (!try_lock_hpte(hptep, HPTE_V_HVLOCK)) in kvmppc_book3s_hv_page_fault()
536 hpte[0] = be64_to_cpu(hptep[0]) & ~HPTE_V_HVLOCK; in kvmppc_book3s_hv_page_fault()
537 hpte[1] = be64_to_cpu(hptep[1]); in kvmppc_book3s_hv_page_fault()
539 unlock_hpte(hptep, hpte[0]); in kvmppc_book3s_hv_page_fault()
658 while (!try_lock_hpte(hptep, HPTE_V_HVLOCK)) in kvmppc_book3s_hv_page_fault()
660 hnow_v = be64_to_cpu(hptep[0]); in kvmppc_book3s_hv_page_fault()
661 hnow_r = be64_to_cpu(hptep[1]); in kvmppc_book3s_hv_page_fault()
697 if (be64_to_cpu(hptep[0]) & HPTE_V_VALID) { in kvmppc_book3s_hv_page_fault()
700 hptep[0] |= cpu_to_be64(HPTE_V_ABSENT); in kvmppc_book3s_hv_page_fault()
701 kvmppc_invalidate_hpte(kvm, hptep, index); in kvmppc_book3s_hv_page_fault()
703 r |= be64_to_cpu(hptep[1]) & (HPTE_R_R | HPTE_R_C); in kvmppc_book3s_hv_page_fault()
712 hptep[1] = cpu_to_be64(r); in kvmppc_book3s_hv_page_fault()
714 __unlock_hpte(hptep, hpte[0]); in kvmppc_book3s_hv_page_fault()
728 __unlock_hpte(hptep, be64_to_cpu(hptep[0])); in kvmppc_book3s_hv_page_fault()
760 __be64 *hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); in kvmppc_unmap_hpte() local
780 psize = kvmppc_actual_pgsz(be64_to_cpu(hptep[0]), ptel); in kvmppc_unmap_hpte()
781 if ((be64_to_cpu(hptep[0]) & HPTE_V_VALID) && in kvmppc_unmap_hpte()
783 hptep[0] |= cpu_to_be64(HPTE_V_ABSENT); in kvmppc_unmap_hpte()
784 kvmppc_invalidate_hpte(kvm, hptep, i); in kvmppc_unmap_hpte()
785 hptep[1] &= ~cpu_to_be64(HPTE_R_KEY_HI | HPTE_R_KEY_LO); in kvmppc_unmap_hpte()
787 rcbits = be64_to_cpu(hptep[1]) & (HPTE_R_R | HPTE_R_C); in kvmppc_unmap_hpte()
802 __be64 *hptep; in kvm_unmap_rmapp() local
819 hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); in kvm_unmap_rmapp()
820 if (!try_lock_hpte(hptep, HPTE_V_HVLOCK)) { in kvm_unmap_rmapp()
823 while (be64_to_cpu(hptep[0]) & HPTE_V_HVLOCK) in kvm_unmap_rmapp()
830 __unlock_hpte(hptep, be64_to_cpu(hptep[0])); in kvm_unmap_rmapp()
881 __be64 *hptep; in kvm_age_rmapp() local
899 hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); in kvm_age_rmapp()
903 if (!(be64_to_cpu(hptep[1]) & HPTE_R_R)) in kvm_age_rmapp()
906 if (!try_lock_hpte(hptep, HPTE_V_HVLOCK)) { in kvm_age_rmapp()
909 while (be64_to_cpu(hptep[0]) & HPTE_V_HVLOCK) in kvm_age_rmapp()
915 if ((be64_to_cpu(hptep[0]) & HPTE_V_VALID) && in kvm_age_rmapp()
916 (be64_to_cpu(hptep[1]) & HPTE_R_R)) { in kvm_age_rmapp()
917 kvmppc_clear_ref_hpte(kvm, hptep, i); in kvm_age_rmapp()
924 __unlock_hpte(hptep, be64_to_cpu(hptep[0])); in kvm_age_rmapp()
1017 __be64 *hptep; in kvm_test_clear_dirty_npages() local
1030 hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); in kvm_test_clear_dirty_npages()
1047 hptep1 = be64_to_cpu(hptep[1]); in kvm_test_clear_dirty_npages()
1052 if (!try_lock_hpte(hptep, HPTE_V_HVLOCK)) { in kvm_test_clear_dirty_npages()
1055 while (hptep[0] & cpu_to_be64(HPTE_V_HVLOCK)) in kvm_test_clear_dirty_npages()
1061 if (!(hptep[0] & cpu_to_be64(HPTE_V_VALID))) { in kvm_test_clear_dirty_npages()
1062 __unlock_hpte(hptep, be64_to_cpu(hptep[0])); in kvm_test_clear_dirty_npages()
1067 hptep[0] |= cpu_to_be64(HPTE_V_ABSENT); in kvm_test_clear_dirty_npages()
1068 kvmppc_invalidate_hpte(kvm, hptep, i); in kvm_test_clear_dirty_npages()
1069 v = be64_to_cpu(hptep[0]); in kvm_test_clear_dirty_npages()
1070 r = be64_to_cpu(hptep[1]); in kvm_test_clear_dirty_npages()
1072 hptep[1] = cpu_to_be64(r & ~HPTE_R_C); in kvm_test_clear_dirty_npages()
1085 __unlock_hpte(hptep, v); in kvm_test_clear_dirty_npages()
1211 __be64 *hptep, *new_hptep; in resize_hpt_rehash_hpte() local
1219 hptep = (__be64 *)(old->virt + (idx << 4)); in resize_hpt_rehash_hpte()
1224 vpte = be64_to_cpu(hptep[0]); in resize_hpt_rehash_hpte()
1228 while (!try_lock_hpte(hptep, HPTE_V_HVLOCK)) in resize_hpt_rehash_hpte()
1231 vpte = be64_to_cpu(hptep[0]); in resize_hpt_rehash_hpte()
1239 rpte = be64_to_cpu(hptep[1]); in resize_hpt_rehash_hpte()
1271 vpte = be64_to_cpu(hptep[0]); in resize_hpt_rehash_hpte()
1279 rpte = be64_to_cpu(hptep[1]); in resize_hpt_rehash_hpte()
1354 unlock_hpte(hptep, vpte); in resize_hpt_rehash_hpte()