Home
last modified time | relevance | path

Searched refs:pgt (Results 1 – 17 of 17) sorted by relevance

/linux/arch/arm64/kvm/hyp/
A Dpgtable.c257 struct kvm_pgtable *pgt = data->pgt; in _kvm_pgtable_walk() local
263 if (!pgt->pgd) in _kvm_pgtable_walk()
281 .pgt = pgt, in kvm_pgtable_walk()
469 if (!pgt->pgd) in kvm_pgtable_hyp_init()
498 WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker)); in kvm_pgtable_hyp_destroy()
499 pgt->mm_ops->put_page(pgt->pgd); in kvm_pgtable_hyp_destroy()
664 struct kvm_pgtable *pgt = data->mmu->pgt; in stage2_map_walker_try_leaf() local
942 .arg = pgt, in kvm_pgtable_stage2_unmap()
1109 .arg = pgt, in kvm_pgtable_stage2_flush()
1132 if (!pgt->pgd) in __kvm_pgtable_stage2_init()
[all …]
/linux/arch/arm64/include/asm/
A Dkvm_pgtable.h220 int kvm_pgtable_hyp_init(struct kvm_pgtable *pgt, u32 va_bits,
230 void kvm_pgtable_hyp_destroy(struct kvm_pgtable *pgt);
287 #define kvm_pgtable_stage2_init(pgt, arch, mm_ops) \ argument
288 __kvm_pgtable_stage2_init(pgt, arch, mm_ops, 0, NULL)
297 void kvm_pgtable_stage2_destroy(struct kvm_pgtable *pgt);
327 int kvm_pgtable_stage2_map(struct kvm_pgtable *pgt, u64 addr, u64 size,
418 kvm_pte_t kvm_pgtable_stage2_mkold(struct kvm_pgtable *pgt, u64 addr);
437 int kvm_pgtable_stage2_relax_perms(struct kvm_pgtable *pgt, u64 addr,
450 bool kvm_pgtable_stage2_is_young(struct kvm_pgtable *pgt, u64 addr);
488 int kvm_pgtable_walk(struct kvm_pgtable *pgt, u64 addr, u64 size,
[all …]
A Dkvm_host.h94 struct kvm_pgtable *pgt; member
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
A Dvmm.c31 if (pgt) { in nvkm_vmm_pt_del()
32 kvfree(pgt->pde); in nvkm_vmm_pt_del()
33 kfree(pgt); in nvkm_vmm_pt_del()
56 if (!(pgt = kzalloc(sizeof(*pgt) + lpte, GFP_KERNEL))) in nvkm_vmm_pt_new()
62 pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); in nvkm_vmm_pt_new()
63 if (!pgt->pde) { in nvkm_vmm_pt_new()
64 kfree(pgt); in nvkm_vmm_pt_new()
69 return pgt; in nvkm_vmm_pt_new()
216 if (!pgt->refs[0]) in nvkm_vmm_unref_sptes()
280 if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) in nvkm_vmm_unref_ptes()
[all …]
A Dvmmgp100.c238 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd0_pde() local
242 if (pgt->pt[0] && !gp100_vmm_pde(pgt->pt[0], &data[0])) in gp100_vmm_pd0_pde()
244 if (pgt->pt[1] && !gp100_vmm_pde(pgt->pt[1], &data[1])) in gp100_vmm_pd0_pde()
365 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd1_pde() local
369 if (!gp100_vmm_pde(pgt->pt[0], &data)) in gp100_vmm_pd1_pde()
A Dvmmnv50.c106 nv50_vmm_pde(struct nvkm_vmm *vmm, struct nvkm_vmm_pt *pgt, u64 *pdata) in nv50_vmm_pde() argument
110 if (pgt && (pt = pgt->pt[0])) { in nv50_vmm_pde()
111 switch (pgt->page) { in nv50_vmm_pde()
A Dvmmgf100.c108 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gf100_vmm_pgd_pde() local
113 if ((pt = pgt->pt[0])) { in gf100_vmm_pgd_pde()
127 if ((pt = pgt->pt[1])) { in gf100_vmm_pgd_pde()
/linux/arch/arm64/kvm/
A Dmmu.c51 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in stage2_apply_range() local
52 if (!pgt) in stage2_apply_range()
515 pgt = kzalloc(sizeof(*pgt), GFP_KERNEL_ACCOUNT); in kvm_init_stage2_mmu()
516 if (!pgt) in kvm_init_stage2_mmu()
533 mmu->pgt = pgt; in kvm_init_stage2_mmu()
541 kfree(pgt); in kvm_init_stage2_mmu()
619 pgt = mmu->pgt; in kvm_free_stage2_pgd()
620 if (pgt) { in kvm_free_stage2_pgd()
627 if (pgt) { in kvm_free_stage2_pgd()
648 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in kvm_phys_addr_ioremap() local
[all …]
/linux/arch/arm64/kvm/hyp/nvhe/
A Dmem_protect.c111 ret = __kvm_pgtable_stage2_init(&host_kvm.pgt, &host_kvm.arch, in kvm_host_prepare_stage2()
117 mmu->pgd_phys = __hyp_pa(host_kvm.pgt.pgd); in kvm_host_prepare_stage2()
119 mmu->pgt = &host_kvm.pgt; in kvm_host_prepare_stage2()
158 struct kvm_pgtable *pgt = &host_kvm.pgt; in host_stage2_unmap_dev_all() local
166 ret = kvm_pgtable_stage2_unmap(pgt, addr, reg->base - addr); in host_stage2_unmap_dev_all()
170 return kvm_pgtable_stage2_unmap(pgt, addr, BIT(pgt->ia_bits) - addr); in host_stage2_unmap_dev_all()
233 return kvm_pgtable_stage2_map(&host_kvm.pgt, start, end - start, start, in __host_stage2_idmap()
270 ret = kvm_pgtable_get_leaf(&host_kvm.pgt, addr, &pte, &level); in host_stage2_adjust_range()
306 return host_stage2_try(kvm_pgtable_stage2_set_owner, &host_kvm.pgt, in host_stage2_set_owner_locked()
375 ret = kvm_pgtable_get_leaf(&host_kvm.pgt, addr, &pte, NULL); in __pkvm_host_share_hyp()
/linux/drivers/gpu/drm/nouveau/nvkm/engine/dma/
A Dusernv04.c52 struct nvkm_memory *pgt = in nv04_dmaobj_bind() local
55 return nvkm_gpuobj_wrap(pgt, pgpuobj); in nv04_dmaobj_bind()
56 nvkm_kmap(pgt); in nv04_dmaobj_bind()
57 offset = nvkm_ro32(pgt, 8 + (offset >> 10)); in nv04_dmaobj_bind()
59 nvkm_done(pgt); in nv04_dmaobj_bind()
/linux/arch/s390/kvm/
A Dgaccess.c997 unsigned long *pgt, int *dat_protection, in kvm_s390_shadow_tables() argument
1051 *pgt = ptr + vaddr.rfx * 8; in kvm_s390_shadow_tables()
1078 *pgt = ptr + vaddr.rsx * 8; in kvm_s390_shadow_tables()
1106 *pgt = ptr + vaddr.rtx * 8; in kvm_s390_shadow_tables()
1143 *pgt = ptr + vaddr.sx * 8; in kvm_s390_shadow_tables()
1169 *pgt = ptr; in kvm_s390_shadow_tables()
1192 unsigned long pgt = 0; in kvm_s390_shadow_fault() local
1211 pte.val = pgt + vaddr.px * PAGE_SIZE; in kvm_s390_shadow_fault()
1220 pgt |= PEI_NOT_PTE; in kvm_s390_shadow_fault()
1223 pgt += vaddr.px * 8; in kvm_s390_shadow_fault()
[all …]
/linux/arch/s390/mm/
A Dgmap.c1313 unsigned long *pgt) in __gmap_unshadow_pgt() argument
1319 pgt[i] = _PAGE_INVALID; in __gmap_unshadow_pgt()
1331 unsigned long sto, *ste, *pgt; in gmap_unshadow_pgt() local
1343 __gmap_unshadow_pgt(sg, raddr, pgt); in gmap_unshadow_pgt()
1345 page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT); in gmap_unshadow_pgt()
1361 unsigned long *pgt; in __gmap_unshadow_sgt() local
1371 __gmap_unshadow_pgt(sg, raddr, pgt); in __gmap_unshadow_sgt()
1373 page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT); in __gmap_unshadow_sgt()
1993 unsigned long *pgt, int *dat_protection, in gmap_shadow_pgt_lookup() argument
2006 *pgt = page->index & ~GMAP_SHADOW_FAKE_TABLE; in gmap_shadow_pgt_lookup()
[all …]
/linux/arch/s390/include/asm/
A Dgmap.h135 int gmap_shadow_pgt(struct gmap *sg, unsigned long saddr, unsigned long pgt,
138 unsigned long *pgt, int *dat_protection, int *fake);
/linux/arch/powerpc/kvm/
A Dbook3s_64_mmu_radix.c1292 pgd_t *pgt; in debugfs_radix_read() local
1330 pgt = NULL; in debugfs_radix_read()
1334 pgt = NULL; in debugfs_radix_read()
1344 if (!pgt) { in debugfs_radix_read()
1346 pgt = kvm->arch.pgtable; in debugfs_radix_read()
1353 pgt = nested->shadow_pgtable; in debugfs_radix_read()
1362 "pgdir: %lx\n", (unsigned long)pgt); in debugfs_radix_read()
1367 pgdp = pgt + pgd_index(gpa); in debugfs_radix_read()
/linux/arch/arm64/kvm/hyp/include/nvhe/
A Dmem_protect.h43 struct kvm_pgtable pgt; member
/linux/arch/x86/events/intel/
A Duncore_nhmex.c877 DEFINE_UNCORE_FORMAT_ATTR(pgt, pgt, "config1:0-31");
/linux/drivers/net/ethernet/broadcom/bnx2x/
A Dbnx2x_self_test.c78 static int pgt(struct st_pred_args *args) in pgt() function
395 NA, 1, 0, pgt,

Completed in 50 milliseconds