/linux/arch/x86/kvm/mmu/ |
A D | tdp_iter.h | 34 int root_level; member 60 #define for_each_tdp_pte_min_level(iter, root, root_level, min_level, start, end) \ argument 61 for (tdp_iter_start(&iter, root, root_level, min_level, start); \ 65 #define for_each_tdp_pte(iter, root, root_level, start, end) \ argument 66 for_each_tdp_pte_min_level(iter, root, root_level, PG_LEVEL_4K, start, end) 70 void tdp_iter_start(struct tdp_iter *iter, u64 *root_pt, int root_level,
|
A D | tdp_iter.c | 31 iter->level = iter->root_level; in tdp_iter_restart() 43 void tdp_iter_start(struct tdp_iter *iter, u64 *root_pt, int root_level, in tdp_iter_start() argument 46 WARN_ON(root_level < 1); in tdp_iter_start() 47 WARN_ON(root_level > PT64_ROOT_MAX_LEVEL); in tdp_iter_start() 50 iter->root_level = root_level; in tdp_iter_start() 52 iter->pt_path[iter->root_level - 1] = (tdp_ptep_t)root_pt; in tdp_iter_start() 136 if (iter->level == iter->root_level) in try_step_up()
|
A D | mmu_audit.c | 62 if (vcpu->arch.mmu->root_level >= PT64_ROOT_4LEVEL) { in mmu_spte_walk() 66 __mmu_spte_walk(vcpu, sp, fn, vcpu->arch.mmu->root_level); in mmu_spte_walk() 122 "ent %llxn", vcpu->arch.mmu->root_level, pfn, in audit_mappings()
|
A D | tdp_mmu.h | 85 int *root_level);
|
A D | mmu.c | 2181 vcpu->arch.mmu->root_level < PT64_ROOT_4LEVEL && in shadow_walk_init_using_root() 3488 if (mmu->root_level == PT32E_ROOT_LEVEL) { in mmu_alloc_shadow_roots() 3512 if (mmu->root_level >= PT64_ROOT_4LEVEL) { in mmu_alloc_shadow_roots() 3551 if (mmu->root_level == PT32E_ROOT_LEVEL) { in mmu_alloc_shadow_roots() 3690 if (vcpu->arch.mmu->root_level >= PT64_ROOT_4LEVEL) { in kvm_mmu_sync_roots() 3782 *root_level = iterator.level; in get_walk() 4189 mmu->root_level >= PT64_ROOT_4LEVEL) in fast_pgd_switch() 4383 context->root_level, is_efer_nx(context), in reset_rsvds_bits_mask() 4854 context->root_level = role_regs_to_root_level(regs); in shadow_mmu_init_context() 4951 context->root_level = level; in kvm_init_shadow_ept_mmu() [all …]
|
A D | paging_tmpl.h | 364 walker->level = mmu->root_level; in FNAME() 670 top_level = vcpu->arch.mmu->root_level; in FNAME()
|
A D | tdp_mmu.c | 1486 int *root_level) in kvm_tdp_mmu_get_walk() argument 1493 *root_level = vcpu->arch.mmu->shadow_root_level; in kvm_tdp_mmu_get_walk()
|
/linux/fs/btrfs/ |
A D | qgroup.c | 1849 if (root_level == 0) in adjust_slots_upwards() 1852 while (level <= root_level) { in adjust_slots_upwards() 1863 if (level != root_level) { in adjust_slots_upwards() 1946 int cur_level = root_level; in qgroup_trace_extent_swap() 1968 src_path->slots[root_level] = dst_path->slots[root_level]; in qgroup_trace_extent_swap() 2077 root_level < 0 || root_level >= BTRFS_MAX_LEVEL - 1 || in qgroup_trace_new_subtree_blocks() 2097 __func__, root_level, root_level, cur_level); in qgroup_trace_new_subtree_blocks() 2229 BUG_ON(root_level < 0 || root_level >= BTRFS_MAX_LEVEL); in btrfs_qgroup_trace_subtree() 2241 if (root_level == 0) { in btrfs_qgroup_trace_subtree() 2261 path->slots[root_level] = 0; in btrfs_qgroup_trace_subtree() [all …]
|
A D | qgroup.h | 340 u64 root_gen, int root_level);
|
A D | backref.c | 544 int root_level; in resolve_indirect_ref() local 577 root_level = btrfs_header_level(root->commit_root); in resolve_indirect_ref() 579 root_level = btrfs_header_level(root->node); in resolve_indirect_ref() 581 root_level = btrfs_old_root_level(root, time_seq); in resolve_indirect_ref() 583 if (root_level + 1 == level) in resolve_indirect_ref()
|
A D | ctree.h | 259 u8 root_level; member 2229 BTRFS_SETGET_STACK_FUNCS(root_level, struct btrfs_root_item, level, 8); 2410 root_level, 8);
|
A D | send.c | 6733 int *level, int root_level) in tree_move_next_or_upnext() argument 6742 if (*level == root_level) in tree_move_next_or_upnext() 6763 int *level, int root_level, in tree_advance() argument 6771 ret = tree_move_next_or_upnext(path, level, root_level); in tree_advance()
|
A D | transaction.c | 1829 super->root_level = root_item->level; in update_super_roots()
|
/linux/fs/xfs/scrub/ |
A D | btree.c | 482 unsigned int root_level = cur->bc_nlevels - 1; in xchk_btree_check_minrecs() local 502 root_block = xfs_btree_get_block(cur, root_level, &root_bp); in xchk_btree_check_minrecs() 503 root_maxrecs = cur->bc_ops->get_dmaxrecs(cur, root_level); in xchk_btree_check_minrecs() 515 if (level < root_level) in xchk_btree_check_minrecs()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_vm.c | 195 adev->vm_manager.root_level); in amdgpu_vm_num_entries() 197 if (level == adev->vm_manager.root_level) in amdgpu_vm_num_entries() 221 shift = amdgpu_vm_level_shift(adev, adev->vm_manager.root_level); in amdgpu_vm_num_ats_entries() 237 if (level <= adev->vm_manager.root_level) in amdgpu_vm_entries_mask() 441 cursor->level = adev->vm_manager.root_level; in amdgpu_vm_pt_start() 797 unsigned level = adev->vm_manager.root_level; in amdgpu_vm_clear_bo() 1351 level += params->adev->vm_manager.root_level; in amdgpu_vm_update_pde() 2865 adev->vm_manager.root_level = AMDGPU_VM_PDB2; in amdgpu_vm_adjust_size() 2868 adev->vm_manager.root_level = AMDGPU_VM_PDB1; in amdgpu_vm_adjust_size() 2871 adev->vm_manager.root_level = AMDGPU_VM_PDB0; in amdgpu_vm_adjust_size() [all …]
|
A D | amdgpu_vm.h | 340 enum amdgpu_vm_level root_level; member
|
/linux/arch/x86/kvm/vmx/ |
A D | vmx.h | 380 u64 construct_eptp(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
|
A D | vmx.c | 3083 u64 construct_eptp(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level) in construct_eptp() argument 3087 eptp |= (root_level == 5) ? VMX_EPTP_PWL_5 : VMX_EPTP_PWL_4; in construct_eptp() 3098 int root_level) in vmx_load_mmu_pgd() argument 3106 eptp = construct_eptp(vcpu, root_hpa, root_level); in vmx_load_mmu_pgd()
|
/linux/arch/x86/include/asm/ |
A D | kvm_host.h | 433 u8 root_level; member 1409 int root_level);
|
/linux/arch/x86/kvm/svm/ |
A D | svm.c | 3988 int root_level) in svm_load_mmu_pgd() argument
|