/linux-6.6.21/arch/mips/kvm/ |
D | mmu.c | 511 bool write_fault, in _kvm_mips_map_page_fast() argument 537 if (write_fault && !pte_dirty(*ptep)) { in _kvm_mips_map_page_fast() 587 bool write_fault, in kvm_mips_map_page() argument 602 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page() 632 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable); in kvm_mips_map_page() 658 if (write_fault) { in kvm_mips_map_page() 685 bool write_fault) in kvm_mips_handle_vz_root_tlb_fault() argument 689 ret = kvm_mips_map_page(vcpu, badvaddr, write_fault, NULL, NULL); in kvm_mips_handle_vz_root_tlb_fault()
|
/linux-6.6.21/fs/xfs/ |
D | xfs_file.c | 1328 bool write_fault, in xfs_dax_fault() argument 1332 (write_fault && !vmf->cow_page) ? in xfs_dax_fault() 1341 bool write_fault, in xfs_dax_fault() argument 1363 bool write_fault) in __xfs_filemap_fault() argument 1370 trace_xfs_filemap_fault(ip, order, write_fault); in __xfs_filemap_fault() 1372 if (write_fault) { in __xfs_filemap_fault() 1377 if (IS_DAX(inode) || write_fault) in __xfs_filemap_fault() 1383 ret = xfs_dax_fault(vmf, order, write_fault, &pfn); in __xfs_filemap_fault() 1386 } else if (write_fault) { in __xfs_filemap_fault() 1395 if (write_fault) in __xfs_filemap_fault()
|
D | xfs_trace.h | 812 TP_PROTO(struct xfs_inode *ip, unsigned int order, bool write_fault), 813 TP_ARGS(ip, order, write_fault), 818 __field(bool, write_fault) 824 __entry->write_fault = write_fault; 830 __entry->write_fault)
|
/linux-6.6.21/arch/x86/kvm/mmu/ |
D | paging_tmpl.h | 200 gpa_t addr, int write_fault) in FNAME() 221 if (level == walker->level && write_fault && in FNAME() 317 const int write_fault = access & PFERR_WRITE_MASK; in FNAME() local 454 if (!write_fault) in FNAME() 467 addr, write_fault); in FNAME() 477 errcode |= write_fault | user_fault; in FNAME() 502 if (write_fault) in FNAME()
|
D | mmu.c | 2915 bool write_fault = fault && fault->write; in mmu_set_spte() local 2953 if (write_fault) in mmu_set_spte()
|
/linux-6.6.21/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_svm.h | 176 bool write_fault); 226 uint64_t addr, bool write_fault) in svm_range_restore_pages() argument
|
D | kfd_smi_events.h | 33 unsigned long address, bool write_fault,
|
D | kfd_smi_events.c | 254 unsigned long address, bool write_fault, in kfd_smi_event_page_fault_start() argument 259 address, node->id, write_fault ? 'W' : 'R'); in kfd_smi_event_page_fault_start()
|
D | kfd_svm.c | 2874 svm_fault_allowed(struct vm_area_struct *vma, bool write_fault) in svm_fault_allowed() argument 2878 if (write_fault) in svm_fault_allowed() 2889 uint64_t addr, bool write_fault) in svm_range_restore_pages() argument 3004 if (!svm_fault_allowed(vma, write_fault)) { in svm_range_restore_pages() 3006 write_fault ? "write" : "read"); in svm_range_restore_pages() 3024 write_fault, timestamp); in svm_range_restore_pages()
|
/linux-6.6.21/virt/kvm/ |
D | kvm_mm.h | 24 bool *async, bool write_fault, bool *writable);
|
D | kvm_main.c | 2515 static bool hva_to_pfn_fast(unsigned long addr, bool write_fault, in hva_to_pfn_fast() argument 2525 if (!(write_fault || writable)) in hva_to_pfn_fast() 2543 static int hva_to_pfn_slow(unsigned long addr, bool *async, bool write_fault, in hva_to_pfn_slow() argument 2564 *writable = write_fault; in hva_to_pfn_slow() 2566 if (write_fault) in hva_to_pfn_slow() 2578 if (unlikely(!write_fault) && writable) { in hva_to_pfn_slow() 2591 static bool vma_is_valid(struct vm_area_struct *vma, bool write_fault) in vma_is_valid() argument 2596 if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE)))) in vma_is_valid() 2613 unsigned long addr, bool write_fault, in hva_to_pfn_remapped() argument 2630 (write_fault ? FAULT_FLAG_WRITE : 0), in hva_to_pfn_remapped() [all …]
|
/linux-6.6.21/arch/arm64/kvm/ |
D | mmu.c | 1399 bool write_fault, writable, force_pte = false; in user_mem_abort() local 1416 write_fault = kvm_is_write_fault(vcpu); in user_mem_abort() 1418 VM_BUG_ON(write_fault && exec_fault); in user_mem_abort() 1420 if (fault_status == ESR_ELx_FSC_PERM && !write_fault && !exec_fault) { in user_mem_abort() 1432 (logging_active && write_fault)) { in user_mem_abort() 1508 write_fault, &writable, NULL); in user_mem_abort() 1528 } else if (logging_active && !write_fault) { in user_mem_abort() 1644 bool is_iabt, write_fault, writable; in kvm_handle_guest_abort() local 1703 write_fault = kvm_is_write_fault(vcpu); in kvm_handle_guest_abort() 1704 if (kvm_is_error_hva(hva) || (write_fault && !writable)) { in kvm_handle_guest_abort()
|
/linux-6.6.21/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vm.h | 473 bool write_fault);
|
D | gmc_v10_0.c | 109 bool write_fault = !!(entry->src_data[1] & 0x20); in gmc_v10_0_process_interrupt() local 137 if (amdgpu_vm_handle_fault(adev, entry->pasid, 0, 0, addr, write_fault)) in gmc_v10_0_process_interrupt()
|
D | gmc_v9_0.c | 551 bool write_fault = !!(entry->src_data[1] & 0x20); in gmc_v9_0_process_interrupt() local 597 addr, write_fault); in gmc_v9_0_process_interrupt() 620 addr, write_fault)) in gmc_v9_0_process_interrupt()
|
D | amdgpu_vm.c | 2554 bool write_fault) in amdgpu_vm_handle_fault() argument 2579 node_id, addr, write_fault)) { in amdgpu_vm_handle_fault()
|
/linux-6.6.21/tools/testing/selftests/user_events/ |
D | ftrace_test.c | 389 TEST_F(user, write_fault) { in TEST_F() argument
|
/linux-6.6.21/drivers/vfio/ |
D | vfio_iommu_type1.c | 514 bool write_fault) in follow_fault_pfn() argument 527 (write_fault ? FAULT_FLAG_WRITE : 0), in follow_fault_pfn() 542 if (write_fault && !pte_write(pte)) in follow_fault_pfn()
|
/linux-6.6.21/arch/mips/include/asm/ |
D | kvm_host.h | 790 struct kvm_vcpu *vcpu, bool write_fault);
|
/linux-6.6.21/include/linux/ |
D | kvm_host.h | 1181 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, 1187 bool write_fault, bool *writable, hva_t *hva);
|