/linux/arch/mips/kvm/ |
A D | mmu.c | 512 bool write_fault, in _kvm_mips_map_page_fast() argument 538 if (write_fault && !pte_dirty(*ptep)) { in _kvm_mips_map_page_fast() 588 bool write_fault, in kvm_mips_map_page() argument 603 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page() 633 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable); in kvm_mips_map_page() 659 if (write_fault) { in kvm_mips_map_page() 687 bool write_fault) in kvm_mips_handle_vz_root_tlb_fault() argument 691 ret = kvm_mips_map_page(vcpu, badvaddr, write_fault, NULL, NULL); in kvm_mips_handle_vz_root_tlb_fault()
|
/linux/arch/arm64/kvm/ |
A D | mmu.c | 947 bool write_fault, writable, force_pte = false; in user_mem_abort() local 965 write_fault = kvm_is_write_fault(vcpu); in user_mem_abort() 967 VM_BUG_ON(write_fault && exec_fault); in user_mem_abort() 969 if (fault_status == FSC_PERM && !write_fault && !exec_fault) { in user_mem_abort() 1036 if (fault_status != FSC_PERM || (logging_active && write_fault)) { in user_mem_abort() 1060 write_fault, &writable, NULL); in user_mem_abort() 1080 } else if (logging_active && !write_fault) { in user_mem_abort() 1192 bool is_iabt, write_fault, writable; in kvm_handle_guest_abort() local 1231 write_fault = kvm_is_write_fault(vcpu); in kvm_handle_guest_abort() 1232 if (kvm_is_error_hva(hva) || (write_fault && !writable)) { in kvm_handle_guest_abort()
|
/linux/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_svm.h | 178 unsigned int pasid, uint64_t addr, bool write_fault); 214 bool write_fault) in svm_range_restore_pages() argument
|
A D | kfd_svm.c | 2565 svm_fault_allowed(struct vm_area_struct *vma, bool write_fault) in svm_fault_allowed() argument 2569 if (write_fault) in svm_fault_allowed() 2579 uint64_t addr, bool write_fault) in svm_range_restore_pages() argument 2684 if (!svm_fault_allowed(vma, write_fault)) { in svm_range_restore_pages() 2686 write_fault ? "write" : "read"); in svm_range_restore_pages()
|
/linux/arch/x86/kvm/mmu/ |
A D | paging_tmpl.h | 238 gpa_t addr, int write_fault) in FNAME() 259 if (level == walker->level && write_fault && in FNAME() 355 const int write_fault = access & PFERR_WRITE_MASK; in FNAME() local 476 if (!write_fault) in FNAME() 489 addr, write_fault); in FNAME() 502 errcode |= write_fault | user_fault; in FNAME() 526 if (write_fault) in FNAME()
|
A D | mmu.c | 2699 bool write_fault = fault && fault->write; in mmu_set_spte() local 2702 *sptep, write_fault, gfn); in mmu_set_spte() 2741 if (write_fault) in mmu_set_spte()
|
/linux/fs/xfs/ |
A D | xfs_file.c | 1313 bool write_fault) in __xfs_filemap_fault() argument 1319 trace_xfs_filemap_fault(ip, pe_size, write_fault); in __xfs_filemap_fault() 1321 if (write_fault) { in __xfs_filemap_fault() 1331 (write_fault && !vmf->cow_page) ? in __xfs_filemap_fault() 1338 if (write_fault) { in __xfs_filemap_fault() 1348 if (write_fault) in __xfs_filemap_fault()
|
A D | xfs_trace.h | 800 bool write_fault), 801 TP_ARGS(ip, pe_size, write_fault), 806 __field(bool, write_fault) 812 __entry->write_fault = write_fault; 821 __entry->write_fault)
|
/linux/virt/kvm/ |
A D | kvm_main.c | 2258 if (!(write_fault || writable)) in hva_to_pfn_fast() 2286 *writable = write_fault; in hva_to_pfn_slow() 2288 if (write_fault) in hva_to_pfn_slow() 2298 if (unlikely(!write_fault) && writable) { in hva_to_pfn_slow() 2331 bool write_fault, bool *writable, in hva_to_pfn_remapped() argument 2347 (write_fault ? FAULT_FLAG_WRITE : 0), in hva_to_pfn_remapped() 2359 if (write_fault && !pte_write(*ptep)) { in hva_to_pfn_remapped() 2410 bool write_fault, bool *writable) in hva_to_pfn() argument 2448 if (async && vma_is_valid(vma, write_fault)) in hva_to_pfn() 2484 return hva_to_pfn(addr, atomic, async, write_fault, in __gfn_to_pfn_memslot() [all …]
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_vm.h | 451 uint64_t addr, bool write_fault);
|
A D | gmc_v10_0.c | 96 bool write_fault = !!(entry->src_data[1] & 0x20); in gmc_v10_0_process_interrupt() local 125 if (amdgpu_vm_handle_fault(adev, entry->pasid, addr, write_fault)) in gmc_v10_0_process_interrupt()
|
A D | gmc_v9_0.c | 510 bool write_fault = !!(entry->src_data[1] & 0x20); in gmc_v9_0_process_interrupt() local 541 if (amdgpu_vm_handle_fault(adev, entry->pasid, addr, write_fault)) in gmc_v9_0_process_interrupt()
|
A D | amdgpu_vm.c | 3366 uint64_t addr, bool write_fault) in amdgpu_vm_handle_fault() argument 3391 !svm_range_restore_pages(adev, pasid, addr, write_fault)) { in amdgpu_vm_handle_fault()
|
/linux/include/linux/ |
A D | kvm_host.h | 864 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, 869 bool atomic, bool *async, bool write_fault,
|
/linux/drivers/vfio/ |
A D | vfio_iommu_type1.c | 507 bool write_fault) in follow_fault_pfn() argument 519 (write_fault ? FAULT_FLAG_WRITE : 0), in follow_fault_pfn() 532 if (write_fault && !pte_write(*ptep)) in follow_fault_pfn()
|
/linux/arch/mips/include/asm/ |
A D | kvm_host.h | 791 struct kvm_vcpu *vcpu, bool write_fault);
|