fault 358 arch/alpha/include/asm/core_titan.h u64 fault; /* 0x28 */ fault 620 arch/alpha/kernel/err_titan.c emchk->fault, fault 1207 arch/alpha/kernel/osf_sys.c goto fault; fault 1216 arch/alpha/kernel/osf_sys.c goto fault; fault 1220 arch/alpha/kernel/osf_sys.c fault: fault 91 arch/alpha/mm/fault.c vm_fault_t fault; fault 151 arch/alpha/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 153 arch/alpha/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 156 arch/alpha/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 157 arch/alpha/mm/fault.c if (fault & VM_FAULT_OOM) fault 159 arch/alpha/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 161 arch/alpha/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 167 arch/alpha/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 171 arch/alpha/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 87 arch/arc/include/asm/disasm.h int fault, write; fault 40 arch/arc/kernel/disasm.c goto fault; fault 53 arch/arc/kernel/disasm.c goto fault; fault 429 arch/arc/kernel/disasm.c fault: state->fault = 1; fault 70 arch/arc/kernel/process.c goto fault; fault 77 arch/arc/kernel/process.c goto fault; fault 85 arch/arc/kernel/process.c fault: fault 50 arch/arc/kernel/unaligned.c goto fault; \ fault 65 arch/arc/kernel/unaligned.c goto fault; \ fault 92 arch/arc/kernel/unaligned.c goto fault; \ fault 125 arch/arc/kernel/unaligned.c goto fault; \ fault 159 arch/arc/kernel/unaligned.c fault: state->fault = 1; fault 179 arch/arc/kernel/unaligned.c goto fault; fault 191 arch/arc/kernel/unaligned.c fault: state->fault = 1; fault 225 arch/arc/kernel/unaligned.c if (state.fault) fault 226 arch/arc/kernel/unaligned.c goto fault; fault 230 arch/arc/kernel/unaligned.c goto fault; fault 237 arch/arc/kernel/unaligned.c if (state.fault) fault 238 arch/arc/kernel/unaligned.c goto fault; fault 257 arch/arc/kernel/unaligned.c fault: fault 68 arch/arc/mm/fault.c vm_fault_t fault = VM_FAULT_SIGSEGV; /* handle_mm_fault() output */ fault 128 arch/arc/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 133 arch/arc/mm/fault.c if (unlikely(fault & VM_FAULT_RETRY)) { fault 164 arch/arc/mm/fault.c if (likely(!(fault & VM_FAULT_ERROR))) { fault 165 arch/arc/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 182 arch/arc/mm/fault.c if (fault & VM_FAULT_OOM) { fault 187 arch/arc/mm/fault.c if (fault & VM_FAULT_SIGBUS) { fault 159 arch/arm/include/asm/kvm_emulate.h return vcpu->arch.fault.hsr; fault 174 arch/arm/include/asm/kvm_emulate.h return vcpu->arch.fault.hxfar; fault 179 arch/arm/include/asm/kvm_emulate.h return ((phys_addr_t)vcpu->arch.fault.hpfar & HPFAR_MASK) << 8; fault 175 arch/arm/include/asm/kvm_host.h struct kvm_vcpu_fault_info fault; fault 102 arch/arm/kvm/hyp/switch.c vcpu->arch.fault.hsr = hsr; fault 141 arch/arm/kvm/hyp/switch.c vcpu->arch.fault.hxfar = far; fault 142 arch/arm/kvm/hyp/switch.c vcpu->arch.fault.hpfar = hpfar; fault 222 arch/arm/mm/alignment.c goto fault; \ fault 243 arch/arm/mm/alignment.c goto fault; \ fault 275 arch/arm/mm/alignment.c goto fault; \ fault 317 arch/arm/mm/alignment.c goto fault; \ fault 384 arch/arm/mm/alignment.c fault: fault 446 arch/arm/mm/alignment.c fault: fault 482 arch/arm/mm/alignment.c fault: fault 582 arch/arm/mm/alignment.c fault: fault 772 arch/arm/mm/alignment.c int fault; fault 775 arch/arm/mm/alignment.c fault = get_user(instr, ip); fault 777 arch/arm/mm/alignment.c fault = probe_kernel_address(ip, instr); fault 781 arch/arm/mm/alignment.c return fault; fault 787 arch/arm/mm/alignment.c int fault; fault 790 arch/arm/mm/alignment.c fault = get_user(instr, ip); fault 792 arch/arm/mm/alignment.c fault = probe_kernel_address(ip, instr); fault 796 arch/arm/mm/alignment.c return fault; fault 810 arch/arm/mm/alignment.c int fault; fault 820 arch/arm/mm/alignment.c fault = alignment_get_thumb(regs, ptr, &tinstr); fault 821 arch/arm/mm/alignment.c if (!fault) { fault 826 arch/arm/mm/alignment.c fault = alignment_get_thumb(regs, ptr + 1, &tinst2); fault 835 arch/arm/mm/alignment.c fault = alignment_get_arm(regs, (void *)instrptr, &instr); fault 838 arch/arm/mm/alignment.c if (fault) { fault 207 arch/arm/mm/fault.c vm_fault_t fault; fault 210 arch/arm/mm/fault.c fault = VM_FAULT_BADMAP; fault 222 arch/arm/mm/fault.c fault = VM_FAULT_BADACCESS; fault 234 arch/arm/mm/fault.c return fault; fault 243 arch/arm/mm/fault.c vm_fault_t fault; fault 292 arch/arm/mm/fault.c fault = __do_page_fault(mm, addr, fsr, flags, tsk); fault 298 arch/arm/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) { fault 311 arch/arm/mm/fault.c if (!(fault & VM_FAULT_ERROR) && flags & FAULT_FLAG_ALLOW_RETRY) { fault 312 arch/arm/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 321 arch/arm/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 335 arch/arm/mm/fault.c if (likely(!(fault & (VM_FAULT_ERROR | VM_FAULT_BADMAP | VM_FAULT_BADACCESS)))) fault 345 arch/arm/mm/fault.c if (fault & VM_FAULT_OOM) { fault 355 arch/arm/mm/fault.c if (fault & VM_FAULT_SIGBUS) { fault 368 arch/arm/mm/fault.c code = fault == VM_FAULT_BADACCESS ? fault 249 arch/arm64/include/asm/kvm_emulate.h return vcpu->arch.fault.esr_el2; fault 264 arch/arm64/include/asm/kvm_emulate.h return vcpu->arch.fault.far_el2; fault 269 arch/arm64/include/asm/kvm_emulate.h return ((phys_addr_t)vcpu->arch.fault.hpfar_el2 & HPFAR_MASK) << 8; fault 274 arch/arm64/include/asm/kvm_emulate.h return vcpu->arch.fault.disr_el1; fault 254 arch/arm64/include/asm/kvm_host.h struct kvm_vcpu_fault_info fault; fault 410 arch/arm64/kernel/armv8_deprecated.c goto fault; fault 415 arch/arm64/kernel/armv8_deprecated.c goto fault; fault 431 arch/arm64/kernel/armv8_deprecated.c fault: fault 93 arch/arm64/kernel/asm-offsets.c DEFINE(VCPU_FAULT_DISR, offsetof(struct kvm_vcpu, arch.fault.disr_el1)); fault 130 arch/arm64/kvm/handle_exit.c run->debug.arch.far = vcpu->arch.fault.far_el2; fault 280 arch/arm64/kvm/hyp/switch.c esr = vcpu->arch.fault.esr_el2; fault 308 arch/arm64/kvm/hyp/switch.c vcpu->arch.fault.far_el2 = far; fault 309 arch/arm64/kvm/hyp/switch.c vcpu->arch.fault.hpfar_el2 = hpfar; fault 459 arch/arm64/kvm/hyp/switch.c vcpu->arch.fault.esr_el2 = read_sysreg_el2(SYS_ESR); fault 456 arch/arm64/mm/fault.c vm_fault_t fault, major = 0; fault 522 arch/arm64/mm/fault.c fault = __do_page_fault(mm, addr, mm_flags, vm_flags); fault 523 arch/arm64/mm/fault.c major |= fault & VM_FAULT_MAJOR; fault 525 arch/arm64/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 553 arch/arm64/mm/fault.c if (likely(!(fault & (VM_FAULT_ERROR | VM_FAULT_BADMAP | fault 581 arch/arm64/mm/fault.c if (fault & VM_FAULT_OOM) { fault 593 arch/arm64/mm/fault.c if (fault & VM_FAULT_SIGBUS) { fault 600 arch/arm64/mm/fault.c } else if (fault & (VM_FAULT_HWPOISON_LARGE | VM_FAULT_HWPOISON)) { fault 604 arch/arm64/mm/fault.c if (fault & VM_FAULT_HWPOISON_LARGE) fault 605 arch/arm64/mm/fault.c lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); fault 615 arch/arm64/mm/fault.c fault == VM_FAULT_BADACCESS ? SEGV_ACCERR : SEGV_MAPERR, fault 22 arch/csky/abiv2/fpu.c int fault; fault 32 arch/csky/abiv2/fpu.c fault = __get_user(instr_low, (u16 *)instrptr); fault 33 arch/csky/abiv2/fpu.c if (fault) fault 36 arch/csky/abiv2/fpu.c fault = __get_user(instr_hi, (u16 *)(instrptr + 2)); fault 37 arch/csky/abiv2/fpu.c if (fault) fault 53 arch/csky/mm/fault.c int fault; fault 149 arch/csky/mm/fault.c fault = handle_mm_fault(vma, address, write ? FAULT_FLAG_WRITE : 0); fault 150 arch/csky/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 151 arch/csky/mm/fault.c if (fault & VM_FAULT_OOM) fault 153 arch/csky/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 155 arch/csky/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 159 arch/csky/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 42 arch/hexagon/mm/vm_fault.c vm_fault_t fault; fault 92 arch/hexagon/mm/vm_fault.c fault = handle_mm_fault(vma, address, flags); fault 94 arch/hexagon/mm/vm_fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 98 arch/hexagon/mm/vm_fault.c if (likely(!(fault & VM_FAULT_ERROR))) { fault 100 arch/hexagon/mm/vm_fault.c if (fault & VM_FAULT_MAJOR) fault 104 arch/hexagon/mm/vm_fault.c if (fault & VM_FAULT_RETRY) { fault 121 arch/hexagon/mm/vm_fault.c if (fault & VM_FAULT_OOM) { fault 129 arch/hexagon/mm/vm_fault.c if (fault & VM_FAULT_SIGBUS) { fault 76 arch/ia64/kernel/minstate.h (pUStk) lfetch.fault.excl.nt1 [r22]; \ fault 88 arch/ia64/kernel/minstate.h lfetch.fault.excl.nt1 [r17],L1_CACHE_BYTES; \ fault 91 arch/ia64/kernel/minstate.h lfetch.fault.excl.nt1 [r17]; \ fault 67 arch/ia64/mm/fault.c vm_fault_t fault; fault 142 arch/ia64/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 144 arch/ia64/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 147 arch/ia64/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 153 arch/ia64/mm/fault.c if (fault & VM_FAULT_OOM) { fault 155 arch/ia64/mm/fault.c } else if (fault & VM_FAULT_SIGSEGV) { fault 157 arch/ia64/mm/fault.c } else if (fault & VM_FAULT_SIGBUS) { fault 165 arch/ia64/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 169 arch/ia64/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 19600 arch/m68k/ifpsp060/src/fpsp.S # if an immediate data access fails, the resulting fault status # fault 73 arch/m68k/mm/fault.c vm_fault_t fault; fault 138 arch/m68k/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 139 arch/m68k/mm/fault.c pr_debug("handle_mm_fault returns %x\n", fault); fault 141 arch/m68k/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 144 arch/m68k/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 145 arch/m68k/mm/fault.c if (fault & VM_FAULT_OOM) fault 147 arch/m68k/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 149 arch/m68k/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 160 arch/m68k/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 164 arch/m68k/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 93 arch/microblaze/mm/fault.c vm_fault_t fault; fault 218 arch/microblaze/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 220 arch/microblaze/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 223 arch/microblaze/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 224 arch/microblaze/mm/fault.c if (fault & VM_FAULT_OOM) fault 226 arch/microblaze/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 228 arch/microblaze/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 234 arch/microblaze/mm/fault.c if (unlikely(fault & VM_FAULT_MAJOR)) fault 238 arch/microblaze/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 944 arch/mips/kernel/unaligned.c goto fault; fault 953 arch/mips/kernel/unaligned.c goto fault; fault 979 arch/mips/kernel/unaligned.c goto fault; fault 992 arch/mips/kernel/unaligned.c goto fault; fault 1005 arch/mips/kernel/unaligned.c goto fault; fault 1020 arch/mips/kernel/unaligned.c goto fault; fault 1033 arch/mips/kernel/unaligned.c goto fault; fault 1058 arch/mips/kernel/unaligned.c goto fault; fault 1077 arch/mips/kernel/unaligned.c goto fault; fault 1096 arch/mips/kernel/unaligned.c goto fault; fault 1115 arch/mips/kernel/unaligned.c goto fault; fault 1138 arch/mips/kernel/unaligned.c goto fault; fault 1164 arch/mips/kernel/unaligned.c goto fault; fault 1184 arch/mips/kernel/unaligned.c goto fault; fault 1203 arch/mips/kernel/unaligned.c goto fault; fault 1276 arch/mips/kernel/unaligned.c goto fault; fault 1309 arch/mips/kernel/unaligned.c goto fault; fault 1359 arch/mips/kernel/unaligned.c fault: fault 1430 arch/mips/kernel/unaligned.c goto fault; fault 1437 arch/mips/kernel/unaligned.c goto fault; fault 1472 arch/mips/kernel/unaligned.c goto fault; fault 1477 arch/mips/kernel/unaligned.c goto fault; fault 1492 arch/mips/kernel/unaligned.c goto fault; fault 1497 arch/mips/kernel/unaligned.c goto fault; fault 1511 arch/mips/kernel/unaligned.c goto fault; fault 1516 arch/mips/kernel/unaligned.c goto fault; fault 1535 arch/mips/kernel/unaligned.c goto fault; fault 1540 arch/mips/kernel/unaligned.c goto fault; fault 1563 arch/mips/kernel/unaligned.c goto fault; fault 1570 arch/mips/kernel/unaligned.c goto fault; fault 1577 arch/mips/kernel/unaligned.c goto fault; fault 1600 arch/mips/kernel/unaligned.c goto fault; fault 1607 arch/mips/kernel/unaligned.c goto fault; fault 1614 arch/mips/kernel/unaligned.c goto fault; fault 1637 arch/mips/kernel/unaligned.c goto fault; fault 1644 arch/mips/kernel/unaligned.c goto fault; fault 1651 arch/mips/kernel/unaligned.c goto fault; fault 1679 arch/mips/kernel/unaligned.c goto fault; fault 1686 arch/mips/kernel/unaligned.c goto fault; fault 1693 arch/mips/kernel/unaligned.c goto fault; fault 1794 arch/mips/kernel/unaligned.c goto fault; fault 1800 arch/mips/kernel/unaligned.c goto fault; fault 1815 arch/mips/kernel/unaligned.c goto fault; fault 1821 arch/mips/kernel/unaligned.c goto fault; fault 1867 arch/mips/kernel/unaligned.c goto fault; fault 1877 arch/mips/kernel/unaligned.c goto fault; fault 1887 arch/mips/kernel/unaligned.c goto fault; fault 1905 arch/mips/kernel/unaligned.c goto fault; fault 1927 arch/mips/kernel/unaligned.c goto fault; fault 1942 arch/mips/kernel/unaligned.c goto fault; fault 1952 arch/mips/kernel/unaligned.c goto fault; fault 1970 arch/mips/kernel/unaligned.c goto fault; fault 1985 arch/mips/kernel/unaligned.c fault: fault 2127 arch/mips/kernel/unaligned.c goto fault; fault 2138 arch/mips/kernel/unaligned.c goto fault; fault 2151 arch/mips/kernel/unaligned.c goto fault; fault 2170 arch/mips/kernel/unaligned.c goto fault; fault 2194 arch/mips/kernel/unaligned.c goto fault; fault 2211 arch/mips/kernel/unaligned.c goto fault; fault 2224 arch/mips/kernel/unaligned.c goto fault; fault 2244 arch/mips/kernel/unaligned.c goto fault; fault 2265 arch/mips/kernel/unaligned.c fault: fault 46 arch/mips/mm/fault.c vm_fault_t fault; fault 155 arch/mips/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 157 arch/mips/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 161 arch/mips/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 162 arch/mips/mm/fault.c if (fault & VM_FAULT_OOM) fault 164 arch/mips/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 166 arch/mips/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 171 arch/mips/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 180 arch/mips/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 58 arch/nds32/mm/alignment.c goto fault; \ fault 74 arch/nds32/mm/alignment.c goto fault; \ fault 107 arch/nds32/mm/alignment.c goto fault; \ fault 139 arch/nds32/mm/alignment.c goto fault; \ fault 309 arch/nds32/mm/alignment.c fault: fault 507 arch/nds32/mm/alignment.c fault: fault 77 arch/nds32/mm/fault.c vm_fault_t fault; fault 206 arch/nds32/mm/fault.c fault = handle_mm_fault(vma, addr, flags); fault 213 arch/nds32/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) { fault 219 arch/nds32/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 220 arch/nds32/mm/fault.c if (fault & VM_FAULT_OOM) fault 222 arch/nds32/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 235 arch/nds32/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 244 arch/nds32/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 72 arch/nios2/kernel/misaligned.c unsigned int fault; fault 85 arch/nios2/kernel/misaligned.c fault = 0; fault 98 arch/nios2/kernel/misaligned.c fault |= __get_user(d0, (u8 *)(addr+0)); fault 99 arch/nios2/kernel/misaligned.c fault |= __get_user(d1, (u8 *)(addr+1)); fault 111 arch/nios2/kernel/misaligned.c fault |= __put_user(d0, (u8 *)(addr+0)); fault 112 arch/nios2/kernel/misaligned.c fault |= __put_user(d1, (u8 *)(addr+1)); fault 116 arch/nios2/kernel/misaligned.c fault |= __get_user(d0, (u8 *)(addr+0)); fault 117 arch/nios2/kernel/misaligned.c fault |= __get_user(d1, (u8 *)(addr+1)); fault 133 arch/nios2/kernel/misaligned.c fault |= __put_user(d0, (u8 *)(addr+0)); fault 134 arch/nios2/kernel/misaligned.c fault |= __put_user(d1, (u8 *)(addr+1)); fault 135 arch/nios2/kernel/misaligned.c fault |= __put_user(d2, (u8 *)(addr+2)); fault 136 arch/nios2/kernel/misaligned.c fault |= __put_user(d3, (u8 *)(addr+3)); fault 140 arch/nios2/kernel/misaligned.c fault |= __get_user(d0, (u8 *)(addr+0)); fault 141 arch/nios2/kernel/misaligned.c fault |= __get_user(d1, (u8 *)(addr+1)); fault 142 arch/nios2/kernel/misaligned.c fault |= __get_user(d2, (u8 *)(addr+2)); fault 143 arch/nios2/kernel/misaligned.c fault |= __get_user(d3, (u8 *)(addr+3)); fault 153 arch/nios2/kernel/misaligned.c if (fault) { fault 49 arch/nios2/mm/fault.c vm_fault_t fault; fault 134 arch/nios2/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 136 arch/nios2/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 139 arch/nios2/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 140 arch/nios2/mm/fault.c if (fault & VM_FAULT_OOM) fault 142 arch/nios2/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 144 arch/nios2/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 155 arch/nios2/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 159 arch/nios2/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 52 arch/openrisc/mm/fault.c vm_fault_t fault; fault 162 arch/openrisc/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 164 arch/openrisc/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 167 arch/openrisc/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 168 arch/openrisc/mm/fault.c if (fault & VM_FAULT_OOM) fault 170 arch/openrisc/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 172 arch/openrisc/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 179 arch/openrisc/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 183 arch/openrisc/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 266 arch/parisc/mm/fault.c vm_fault_t fault = 0; fault 305 arch/parisc/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 307 arch/parisc/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 310 arch/parisc/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 316 arch/parisc/mm/fault.c if (fault & VM_FAULT_OOM) fault 318 arch/parisc/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 320 arch/parisc/mm/fault.c else if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| fault 326 arch/parisc/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 330 arch/parisc/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 391 arch/parisc/mm/fault.c if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { fault 401 arch/parisc/mm/fault.c if (fault & VM_FAULT_HWPOISON_LARGE) fault 402 arch/parisc/mm/fault.c lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); fault 403 arch/parisc/mm/fault.c else if (fault & VM_FAULT_HWPOISON) fault 229 arch/powerpc/kvm/book3s_64_vio.c .fault = kvm_spapr_tce_fault, fault 277 arch/powerpc/kvm/book3s_xive_native.c .fault = xive_native_esb_fault, fault 298 arch/powerpc/kvm/book3s_xive_native.c .fault = xive_native_tima_fault, fault 143 arch/powerpc/mm/fault.c vm_fault_t fault) fault 150 arch/powerpc/mm/fault.c if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { fault 156 arch/powerpc/mm/fault.c if (fault & VM_FAULT_HWPOISON_LARGE) fault 157 arch/powerpc/mm/fault.c lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); fault 158 arch/powerpc/mm/fault.c if (fault & VM_FAULT_HWPOISON) fault 171 arch/powerpc/mm/fault.c vm_fault_t fault) fault 181 arch/powerpc/mm/fault.c if (fault & VM_FAULT_OOM) { fault 190 arch/powerpc/mm/fault.c if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| fault 192 arch/powerpc/mm/fault.c return do_sigbus(regs, addr, fault); fault 193 arch/powerpc/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 445 arch/powerpc/mm/fault.c vm_fault_t fault, major = 0; fault 570 arch/powerpc/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 577 arch/powerpc/mm/fault.c if (unlikely(fault & VM_FAULT_SIGSEGV) && fault 587 arch/powerpc/mm/fault.c major |= fault & VM_FAULT_MAJOR; fault 593 arch/powerpc/mm/fault.c if (unlikely(fault & VM_FAULT_RETRY)) { fault 615 arch/powerpc/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) fault 616 arch/powerpc/mm/fault.c return mm_fault_error(regs, address, fault); fault 278 arch/powerpc/platforms/cell/spufs/file.c .fault = spufs_mem_mmap_fault, fault 366 arch/powerpc/platforms/cell/spufs/file.c .fault = spufs_cntl_mmap_fault, fault 1049 arch/powerpc/platforms/cell/spufs/file.c .fault = spufs_signal1_mmap_fault, fault 1187 arch/powerpc/platforms/cell/spufs/file.c .fault = spufs_signal2_mmap_fault, fault 1307 arch/powerpc/platforms/cell/spufs/file.c .fault = spufs_mss_mmap_fault, fault 1369 arch/powerpc/platforms/cell/spufs/file.c .fault = spufs_psmap_mmap_fault, fault 1429 arch/powerpc/platforms/cell/spufs/file.c .fault = spufs_mfc_mmap_fault, fault 35 arch/riscv/mm/fault.c vm_fault_t fault; fault 113 arch/riscv/mm/fault.c fault = handle_mm_fault(vma, addr, flags); fault 120 arch/riscv/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(tsk)) fault 123 arch/riscv/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 124 arch/riscv/mm/fault.c if (fault & VM_FAULT_OOM) fault 126 arch/riscv/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 137 arch/riscv/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 146 arch/riscv/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 96 arch/s390/kernel/vdso.c .fault = vdso_fault, fault 322 arch/s390/mm/fault.c vm_fault_t fault) fault 326 arch/s390/mm/fault.c switch (fault) { fault 335 arch/s390/mm/fault.c si_code = (fault == VM_FAULT_BADMAP) ? fault 351 arch/s390/mm/fault.c if (fault & VM_FAULT_OOM) { fault 356 arch/s390/mm/fault.c } else if (fault & VM_FAULT_SIGSEGV) { fault 362 arch/s390/mm/fault.c } else if (fault & VM_FAULT_SIGBUS) { fault 395 arch/s390/mm/fault.c vm_fault_t fault; fault 415 arch/s390/mm/fault.c fault = VM_FAULT_BADCONTEXT; fault 421 arch/s390/mm/fault.c fault = VM_FAULT_BADMAP; fault 447 arch/s390/mm/fault.c fault = VM_FAULT_BADMAP; fault 455 arch/s390/mm/fault.c fault = VM_FAULT_BADMAP; fault 471 arch/s390/mm/fault.c fault = VM_FAULT_BADACCESS; fault 482 arch/s390/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 484 arch/s390/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) { fault 485 arch/s390/mm/fault.c fault = VM_FAULT_SIGNAL; fault 490 arch/s390/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) fault 499 arch/s390/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 508 arch/s390/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 514 arch/s390/mm/fault.c fault = VM_FAULT_PFAULT; fault 530 arch/s390/mm/fault.c fault = VM_FAULT_BADMAP; fault 534 arch/s390/mm/fault.c fault = VM_FAULT_OOM; fault 538 arch/s390/mm/fault.c fault = 0; fault 542 arch/s390/mm/fault.c return fault; fault 549 arch/s390/mm/fault.c vm_fault_t fault; fault 572 arch/s390/mm/fault.c fault = VM_FAULT_BADACCESS; fault 575 arch/s390/mm/fault.c fault = do_exception(regs, access); fault 577 arch/s390/mm/fault.c if (unlikely(fault)) fault 578 arch/s390/mm/fault.c do_fault_error(regs, access, fault); fault 585 arch/s390/mm/fault.c vm_fault_t fault; fault 588 arch/s390/mm/fault.c fault = do_exception(regs, access); fault 589 arch/s390/mm/fault.c if (unlikely(fault)) fault 590 arch/s390/mm/fault.c do_fault_error(regs, access, fault); fault 299 arch/sh/mm/fault.c unsigned long address, vm_fault_t fault) fault 306 arch/sh/mm/fault.c if (!(fault & VM_FAULT_RETRY)) fault 313 arch/sh/mm/fault.c if (!(fault & VM_FAULT_ERROR)) fault 316 arch/sh/mm/fault.c if (fault & VM_FAULT_OOM) { fault 332 arch/sh/mm/fault.c if (fault & VM_FAULT_SIGBUS) fault 334 arch/sh/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 382 arch/sh/mm/fault.c vm_fault_t fault; fault 467 arch/sh/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 469 arch/sh/mm/fault.c if (unlikely(fault & (VM_FAULT_RETRY | VM_FAULT_ERROR))) fault 470 arch/sh/mm/fault.c if (mm_fault_error(regs, error_code, address, fault)) fault 474 arch/sh/mm/fault.c if (fault & VM_FAULT_MAJOR) { fault 483 arch/sh/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 170 arch/sparc/mm/fault_32.c vm_fault_t fault; fault 238 arch/sparc/mm/fault_32.c fault = handle_mm_fault(vma, address, flags); fault 240 arch/sparc/mm/fault_32.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 243 arch/sparc/mm/fault_32.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 244 arch/sparc/mm/fault_32.c if (fault & VM_FAULT_OOM) fault 246 arch/sparc/mm/fault_32.c else if (fault & VM_FAULT_SIGSEGV) fault 248 arch/sparc/mm/fault_32.c else if (fault & VM_FAULT_SIGBUS) fault 254 arch/sparc/mm/fault_32.c if (fault & VM_FAULT_MAJOR) { fault 263 arch/sparc/mm/fault_32.c if (fault & VM_FAULT_RETRY) { fault 268 arch/sparc/mm/fault_64.c vm_fault_t fault; fault 422 arch/sparc/mm/fault_64.c fault = handle_mm_fault(vma, address, flags); fault 424 arch/sparc/mm/fault_64.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 427 arch/sparc/mm/fault_64.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 428 arch/sparc/mm/fault_64.c if (fault & VM_FAULT_OOM) fault 430 arch/sparc/mm/fault_64.c else if (fault & VM_FAULT_SIGSEGV) fault 432 arch/sparc/mm/fault_64.c else if (fault & VM_FAULT_SIGBUS) fault 438 arch/sparc/mm/fault_64.c if (fault & VM_FAULT_MAJOR) { fault 447 arch/sparc/mm/fault_64.c if (fault & VM_FAULT_RETRY) { fault 75 arch/um/kernel/trap.c vm_fault_t fault; fault 77 arch/um/kernel/trap.c fault = handle_mm_fault(vma, address, flags); fault 79 arch/um/kernel/trap.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 82 arch/um/kernel/trap.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 83 arch/um/kernel/trap.c if (fault & VM_FAULT_OOM) { fault 85 arch/um/kernel/trap.c } else if (fault & VM_FAULT_SIGSEGV) { fault 87 arch/um/kernel/trap.c } else if (fault & VM_FAULT_SIGBUS) { fault 94 arch/um/kernel/trap.c if (fault & VM_FAULT_MAJOR) fault 98 arch/um/kernel/trap.c if (fault & VM_FAULT_RETRY) { fault 106 arch/unicore32/mm/alignment.c goto fault; \ fault 130 arch/unicore32/mm/alignment.c goto fault; \ fault 160 arch/unicore32/mm/alignment.c goto fault; \ fault 174 arch/unicore32/mm/alignment.c goto fault; \ fault 191 arch/unicore32/mm/alignment.c goto fault; \ fault 242 arch/unicore32/mm/alignment.c fault: fault 268 arch/unicore32/mm/alignment.c fault: fault 344 arch/unicore32/mm/alignment.c fault: fault 503 arch/unicore32/mm/alignment.c fault: fault 166 arch/unicore32/mm/fault.c vm_fault_t fault; fault 169 arch/unicore32/mm/fault.c fault = VM_FAULT_BADMAP; fault 181 arch/unicore32/mm/fault.c fault = VM_FAULT_BADACCESS; fault 189 arch/unicore32/mm/fault.c fault = handle_mm_fault(vma, addr & PAGE_MASK, flags); fault 190 arch/unicore32/mm/fault.c return fault; fault 196 arch/unicore32/mm/fault.c return fault; fault 204 arch/unicore32/mm/fault.c vm_fault_t fault; fault 247 arch/unicore32/mm/fault.c fault = __do_pf(mm, addr, fsr, flags, tsk); fault 253 arch/unicore32/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 256 arch/unicore32/mm/fault.c if (!(fault & VM_FAULT_ERROR) && (flags & FAULT_FLAG_ALLOW_RETRY)) { fault 257 arch/unicore32/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 261 arch/unicore32/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 274 arch/unicore32/mm/fault.c if (likely(!(fault & fault 285 arch/unicore32/mm/fault.c if (fault & VM_FAULT_OOM) { fault 295 arch/unicore32/mm/fault.c if (fault & VM_FAULT_SIGBUS) { fault 308 arch/unicore32/mm/fault.c code = fault == VM_FAULT_BADACCESS ? SEGV_ACCERR : SEGV_MAPERR; fault 133 arch/x86/entry/vdso/vma.c .fault = vdso_fault, fault 138 arch/x86/entry/vdso/vma.c .fault = vvar_fault, fault 29 arch/x86/hyperv/nested.c goto fault; fault 40 arch/x86/hyperv/nested.c goto fault; fault 53 arch/x86/hyperv/nested.c fault: fault 101 arch/x86/hyperv/nested.c goto fault; fault 111 arch/x86/hyperv/nested.c goto fault; fault 120 arch/x86/hyperv/nested.c goto fault; fault 132 arch/x86/hyperv/nested.c fault: fault 115 arch/x86/include/asm/kvm_emulate.h struct x86_exception *fault, bool system); fault 137 arch/x86/include/asm/kvm_emulate.h struct x86_exception *fault, bool system); fault 147 arch/x86/include/asm/kvm_emulate.h struct x86_exception *fault); fault 157 arch/x86/include/asm/kvm_emulate.h struct x86_exception *fault); fault 169 arch/x86/include/asm/kvm_emulate.h struct x86_exception *fault); fault 184 arch/x86/include/asm/kvm_emulate.h struct x86_exception *fault); fault 386 arch/x86/include/asm/kvm_host.h struct x86_exception *fault); fault 1404 arch/x86/include/asm/kvm_host.h void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault); fault 152 arch/x86/include/asm/msr.h : "c" (msr), [fault] "i" (-EIO)); fault 182 arch/x86/include/asm/msr.h [fault] "i" (-EIO) fault 4458 arch/x86/kvm/mmu.c struct x86_exception *fault) fault 4460 arch/x86/kvm/mmu.c vcpu->arch.mmu->inject_page_fault(vcpu, fault); fault 179 arch/x86/kvm/mmu.h bool fault = (mmu->permissions[index] >> pte_access) & 1; fault 200 arch/x86/kvm/mmu.h fault |= (pkru_bits != 0); fault 203 arch/x86/kvm/mmu.h return -(u32)fault & errcode; fault 96 arch/x86/kvm/paging_tmpl.h struct x86_exception fault; fault 365 arch/x86/kvm/paging_tmpl.h &walker->fault); fault 427 arch/x86/kvm/paging_tmpl.h real_gpa = mmu->translate_gpa(vcpu, gfn_to_gpa(gfn), access, &walker->fault); fault 462 arch/x86/kvm/paging_tmpl.h walker->fault.vector = PF_VECTOR; fault 463 arch/x86/kvm/paging_tmpl.h walker->fault.error_code_valid = true; fault 464 arch/x86/kvm/paging_tmpl.h walker->fault.error_code = errcode; fault 491 arch/x86/kvm/paging_tmpl.h walker->fault.address = addr; fault 492 arch/x86/kvm/paging_tmpl.h walker->fault.nested_page_fault = mmu != vcpu->arch.walk_mmu; fault 494 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_walker_error(walker->fault.error_code); fault 806 arch/x86/kvm/paging_tmpl.h inject_page_fault(vcpu, &walker.fault); fault 962 arch/x86/kvm/paging_tmpl.h *exception = walker.fault; fault 988 arch/x86/kvm/paging_tmpl.h *exception = walker.fault; fault 3000 arch/x86/kvm/svm.c struct x86_exception *fault) fault 3012 arch/x86/kvm/svm.c svm->vmcb->control.exit_info_2 = fault->address; fault 3016 arch/x86/kvm/svm.c svm->vmcb->control.exit_info_1 |= fault->error_code; fault 326 arch/x86/kvm/vmx/nested.c struct x86_exception *fault) fault 337 arch/x86/kvm/vmx/nested.c } else if (fault->error_code & PFERR_RSVD_MASK) fault 343 arch/x86/kvm/vmx/nested.c vmcs12->guest_physical_address = fault->address; fault 422 arch/x86/kvm/vmx/nested.c struct x86_exception *fault) fault 428 arch/x86/kvm/vmx/nested.c if (nested_vmx_is_page_fault_vmexit(vmcs12, fault->error_code) && fault 430 arch/x86/kvm/vmx/nested.c vmcs12->vm_exit_intr_error_code = fault->error_code; fault 434 arch/x86/kvm/vmx/nested.c fault->address); fault 436 arch/x86/kvm/vmx/nested.c kvm_inject_page_fault(vcpu, fault); fault 15 arch/x86/kvm/vmx/ops.h asmlinkage void vmread_error(unsigned long field, bool fault); fault 17 arch/x86/kvm/vmx/ops.h bool fault); fault 145 arch/x86/kvm/vmx/ops.h _ASM_EXTABLE(1b, %l[fault]) \ fault 146 arch/x86/kvm/vmx/ops.h : : op1 : "cc" : error, fault); \ fault 151 arch/x86/kvm/vmx/ops.h fault: \ fault 160 arch/x86/kvm/vmx/ops.h _ASM_EXTABLE(1b, %l[fault]) \ fault 161 arch/x86/kvm/vmx/ops.h : : op1, op2 : "cc" : error, fault); \ fault 166 arch/x86/kvm/vmx/ops.h fault: \ fault 357 arch/x86/kvm/vmx/vmx.c asmlinkage void vmread_error(unsigned long field, bool fault) fault 359 arch/x86/kvm/vmx/vmx.c if (fault) fault 591 arch/x86/kvm/x86.c void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault) fault 595 arch/x86/kvm/x86.c is_guest_mode(vcpu) && fault->async_page_fault; fault 597 arch/x86/kvm/x86.c vcpu->arch.apf.nested_apf_token = fault->address; fault 598 arch/x86/kvm/x86.c kvm_queue_exception_e(vcpu, PF_VECTOR, fault->error_code); fault 600 arch/x86/kvm/x86.c kvm_queue_exception_e_p(vcpu, PF_VECTOR, fault->error_code, fault 601 arch/x86/kvm/x86.c fault->address); fault 606 arch/x86/kvm/x86.c static bool kvm_propagate_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault) fault 608 arch/x86/kvm/x86.c if (mmu_is_nested(vcpu) && !fault->nested_page_fault) fault 609 arch/x86/kvm/x86.c vcpu->arch.nested_mmu.inject_page_fault(vcpu, fault); fault 611 arch/x86/kvm/x86.c vcpu->arch.mmu->inject_page_fault(vcpu, fault); fault 613 arch/x86/kvm/x86.c return fault->nested_page_fault; fault 10190 arch/x86/kvm/x86.c struct x86_exception fault; fault 10197 arch/x86/kvm/x86.c fault.vector = PF_VECTOR; fault 10198 arch/x86/kvm/x86.c fault.error_code_valid = true; fault 10199 arch/x86/kvm/x86.c fault.error_code = 0; fault 10200 arch/x86/kvm/x86.c fault.nested_page_fault = false; fault 10201 arch/x86/kvm/x86.c fault.address = work->arch.token; fault 10202 arch/x86/kvm/x86.c fault.async_page_fault = true; fault 10203 arch/x86/kvm/x86.c kvm_inject_page_fault(vcpu, &fault); fault 10220 arch/x86/kvm/x86.c struct x86_exception fault; fault 10243 arch/x86/kvm/x86.c fault.vector = PF_VECTOR; fault 10244 arch/x86/kvm/x86.c fault.error_code_valid = true; fault 10245 arch/x86/kvm/x86.c fault.error_code = 0; fault 10246 arch/x86/kvm/x86.c fault.nested_page_fault = false; fault 10247 arch/x86/kvm/x86.c fault.address = work->arch.token; fault 10248 arch/x86/kvm/x86.c fault.async_page_fault = true; fault 10249 arch/x86/kvm/x86.c kvm_inject_page_fault(vcpu, &fault); fault 1021 arch/x86/mm/fault.c vm_fault_t fault) fault 1036 arch/x86/mm/fault.c if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { fault 1043 arch/x86/mm/fault.c if (fault & VM_FAULT_HWPOISON_LARGE) fault 1044 arch/x86/mm/fault.c lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); fault 1045 arch/x86/mm/fault.c if (fault & VM_FAULT_HWPOISON) fault 1056 arch/x86/mm/fault.c unsigned long address, vm_fault_t fault) fault 1063 arch/x86/mm/fault.c if (fault & VM_FAULT_OOM) { fault 1078 arch/x86/mm/fault.c if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| fault 1080 arch/x86/mm/fault.c do_sigbus(regs, error_code, address, fault); fault 1081 arch/x86/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 1311 arch/x86/mm/fault.c vm_fault_t fault, major = 0; fault 1463 arch/x86/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 1464 arch/x86/mm/fault.c major |= fault & VM_FAULT_MAJOR; fault 1471 arch/x86/mm/fault.c if (unlikely(fault & VM_FAULT_RETRY)) { fault 1490 arch/x86/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 1491 arch/x86/mm/fault.c mm_fault_error(regs, hw_error_code, address, fault); fault 41 arch/xtensa/include/asm/futex.h [fault] "I" (-EFAULT) \ fault 65 arch/xtensa/include/asm/futex.h [fault] "I" (-EFAULT) \ fault 155 arch/xtensa/include/asm/futex.h [fault] "I" (-EFAULT) fault 45 arch/xtensa/mm/fault.c vm_fault_t fault; fault 111 arch/xtensa/mm/fault.c fault = handle_mm_fault(vma, address, flags); fault 113 arch/xtensa/mm/fault.c if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) fault 116 arch/xtensa/mm/fault.c if (unlikely(fault & VM_FAULT_ERROR)) { fault 117 arch/xtensa/mm/fault.c if (fault & VM_FAULT_OOM) fault 119 arch/xtensa/mm/fault.c else if (fault & VM_FAULT_SIGSEGV) fault 121 arch/xtensa/mm/fault.c else if (fault & VM_FAULT_SIGBUS) fault 126 arch/xtensa/mm/fault.c if (fault & VM_FAULT_MAJOR) fault 130 arch/xtensa/mm/fault.c if (fault & VM_FAULT_RETRY) { fault 5164 drivers/android/binder.c .fault = binder_vm_fault, fault 42 drivers/char/agp/alpha-agp.c .fault = alpha_core_agp_vm_fault, fault 171 drivers/char/mspec.c .fault = mspec_fault, fault 300 drivers/dax/device.c .fault = dev_dax_fault, fault 34 drivers/dma-buf/udmabuf.c .fault = udmabuf_vm_fault, fault 86 drivers/gpio/gpio-max3191x.c unsigned long *fault; fault 180 drivers/gpio/gpio-max3191x.c __assign_bit(i, max3191x->fault, val); fault 194 drivers/gpio/gpio-max3191x.c if (!max3191x->ignore_uv && test_bit(chipnum, max3191x->fault)) fault 378 drivers/gpio/gpio-max3191x.c max3191x->fault = devm_kcalloc(dev, n, sizeof(long), GFP_KERNEL); fault 383 drivers/gpio/gpio-max3191x.c !max3191x->fault || !max3191x->xfer.rx_buf) fault 272 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c struct amdgpu_gmc_fault *fault; fault 283 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault = &gmc->fault_ring[gmc->fault_hash[hash].idx]; fault 284 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c while (fault->timestamp >= stamp) { fault 287 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c if (fault->key == key) fault 290 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c tmp = fault->timestamp; fault 291 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault = &gmc->fault_ring[fault->next]; fault 294 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c if (fault->timestamp >= tmp) fault 299 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault = &gmc->fault_ring[gmc->last_fault]; fault 300 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault->key = key; fault 301 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault->timestamp = timestamp; fault 304 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault->next = gmc->fault_hash[hash].idx; fault 29 drivers/gpu/drm/armada/armada_gem.c .fault = armada_gem_vm_fault, fault 509 drivers/gpu/drm/drm_gem_shmem_helper.c .fault = drm_gem_shmem_fault, fault 371 drivers/gpu/drm/drm_vm.c .fault = drm_vm_fault, fault 378 drivers/gpu/drm/drm_vm.c .fault = drm_vm_shm_fault, fault 385 drivers/gpu/drm/drm_vm.c .fault = drm_vm_dma_fault, fault 392 drivers/gpu/drm/drm_vm.c .fault = drm_vm_sg_fault, fault 486 drivers/gpu/drm/etnaviv/etnaviv_drv.c .fault = etnaviv_gem_fault, fault 79 drivers/gpu/drm/exynos/exynos_drm_drv.c .fault = exynos_drm_gem_fault, fault 141 drivers/gpu/drm/gma500/framebuffer.c .fault = psbfb_vm_fault, fault 451 drivers/gpu/drm/gma500/psb_drv.c .fault = psb_gem_fault, fault 101 drivers/gpu/drm/i915/gt/intel_gt.c u32 fault; fault 104 drivers/gpu/drm/i915/gt/intel_gt.c fault = GEN6_RING_FAULT_REG_READ(engine); fault 105 drivers/gpu/drm/i915/gt/intel_gt.c if (fault & RING_FAULT_VALID) { fault 111 drivers/gpu/drm/i915/gt/intel_gt.c fault & PAGE_MASK, fault 112 drivers/gpu/drm/i915/gt/intel_gt.c fault & RING_FAULT_GTTSEL_MASK ? fault 114 drivers/gpu/drm/i915/gt/intel_gt.c RING_FAULT_SRCID(fault), fault 115 drivers/gpu/drm/i915/gt/intel_gt.c RING_FAULT_FAULT_TYPE(fault)); fault 124 drivers/gpu/drm/i915/gt/intel_gt.c u32 fault; fault 136 drivers/gpu/drm/i915/gt/intel_gt.c fault = intel_uncore_read(uncore, fault_reg); fault 137 drivers/gpu/drm/i915/gt/intel_gt.c if (fault & RING_FAULT_VALID) { fault 156 drivers/gpu/drm/i915/gt/intel_gt.c GEN8_RING_FAULT_ENGINE_ID(fault), fault 157 drivers/gpu/drm/i915/gt/intel_gt.c RING_FAULT_SRCID(fault), fault 158 drivers/gpu/drm/i915/gt/intel_gt.c RING_FAULT_FAULT_TYPE(fault)); fault 2775 drivers/gpu/drm/i915/i915_drv.c .fault = i915_gem_fault, fault 107 drivers/gpu/drm/lima/lima_gem.c .fault = lima_gem_fault, fault 42 drivers/gpu/drm/lima/lima_mmu.c u32 fault = mmu_read(LIMA_MMU_PAGE_FAULT_ADDR); fault 45 drivers/gpu/drm/lima/lima_mmu.c fault, LIMA_MMU_STATUS_BUS_ID(status), fault 978 drivers/gpu/drm/msm/msm_drv.c .fault = msm_gem_fault, fault 137 drivers/gpu/drm/nouveau/include/nvkm/core/device.h struct nvkm_fault *fault; fault 210 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*fault )(struct nvkm_device *, int idx, struct nvkm_fault **); fault 17 drivers/gpu/drm/nouveau/include/nvkm/subdev/top.h enum nvkm_devidx nvkm_top_fault(struct nvkm_device *, int fault); fault 63 drivers/gpu/drm/nouveau/nouveau_svm.c u8 fault; fault 65 drivers/gpu/drm/nouveau/nouveau_svm.c } **fault; fault 412 drivers/gpu/drm/nouveau/nouveau_svm.c struct nouveau_svm_fault *fault) fault 414 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_svm_fault_cancel(svm, fault->inst, fault 415 drivers/gpu/drm/nouveau/nouveau_svm.c fault->hub, fault 416 drivers/gpu/drm/nouveau/nouveau_svm.c fault->gpc, fault 417 drivers/gpu/drm/nouveau/nouveau_svm.c fault->client); fault 452 drivers/gpu/drm/nouveau/nouveau_svm.c struct nouveau_svm_fault *fault; fault 460 drivers/gpu/drm/nouveau/nouveau_svm.c if (!buffer->fault[buffer->fault_nr]) { fault 461 drivers/gpu/drm/nouveau/nouveau_svm.c fault = kmalloc(sizeof(*fault), GFP_KERNEL); fault 462 drivers/gpu/drm/nouveau/nouveau_svm.c if (WARN_ON(!fault)) { fault 466 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[buffer->fault_nr] = fault; fault 469 drivers/gpu/drm/nouveau/nouveau_svm.c fault = buffer->fault[buffer->fault_nr++]; fault 470 drivers/gpu/drm/nouveau/nouveau_svm.c fault->inst = inst; fault 471 drivers/gpu/drm/nouveau/nouveau_svm.c fault->addr = (u64)addrhi << 32 | addrlo; fault 472 drivers/gpu/drm/nouveau/nouveau_svm.c fault->time = (u64)timehi << 32 | timelo; fault 473 drivers/gpu/drm/nouveau/nouveau_svm.c fault->engine = engine; fault 474 drivers/gpu/drm/nouveau/nouveau_svm.c fault->gpc = gpc; fault 475 drivers/gpu/drm/nouveau/nouveau_svm.c fault->hub = hub; fault 476 drivers/gpu/drm/nouveau/nouveau_svm.c fault->access = (info & 0x000f0000) >> 16; fault 477 drivers/gpu/drm/nouveau/nouveau_svm.c fault->client = client; fault 478 drivers/gpu/drm/nouveau/nouveau_svm.c fault->fault = (info & 0x0000001f); fault 481 drivers/gpu/drm/nouveau/nouveau_svm.c fault->inst, fault->addr, fault->access); fault 571 drivers/gpu/drm/nouveau/nouveau_svm.c sort(buffer->fault, buffer->fault_nr, sizeof(*buffer->fault), fault 577 drivers/gpu/drm/nouveau/nouveau_svm.c if (!svmm || buffer->fault[fi]->inst != inst) { fault 579 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_ivmm_find(svm, buffer->fault[fi]->inst); fault 581 drivers/gpu/drm/nouveau/nouveau_svm.c inst = buffer->fault[fi]->inst; fault 584 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fi]->svmm = svmm; fault 597 drivers/gpu/drm/nouveau/nouveau_svm.c if (!(svmm = buffer->fault[fi]->svmm)) { fault 598 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]); fault 601 drivers/gpu/drm/nouveau/nouveau_svm.c SVMM_DBG(svmm, "addr %016llx", buffer->fault[fi]->addr); fault 606 drivers/gpu/drm/nouveau/nouveau_svm.c start = buffer->fault[fi]->addr; fault 623 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]); fault 630 drivers/gpu/drm/nouveau/nouveau_svm.c if (buffer->fault[fi]->addr != start) { fault 631 drivers/gpu/drm/nouveau/nouveau_svm.c SVMM_ERR(svmm, "addr %016llx", buffer->fault[fi]->addr); fault 633 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]); fault 648 drivers/gpu/drm/nouveau/nouveau_svm.c if (buffer->fault[fn]->access != 0 /* READ. */ && fault 649 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fn]->access != 3 /* PREFETCH. */) { fault 665 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fn]->svmm == svmm && fault 666 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fn ]->addr == fault 667 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fn - 1]->addr); fault 673 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fn]->svmm != svmm || fault 674 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fn]->addr >= limit) fault 678 drivers/gpu/drm/nouveau/nouveau_svm.c fill = (buffer->fault[fn ]->addr - fault 679 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault[fn - 1]->addr) >> PAGE_SHIFT; fault 722 drivers/gpu/drm/nouveau/nouveau_svm.c struct nouveau_svm_fault *fault = buffer->fault[fi++]; fault 723 drivers/gpu/drm/nouveau/nouveau_svm.c pi = (fault->addr - range.start) >> PAGE_SHIFT; fault 727 drivers/gpu/drm/nouveau/nouveau_svm.c fault->access != 0 && fault->access != 3)) { fault 728 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_svm_fault_cancel_fault(svm, fault); fault 765 drivers/gpu/drm/nouveau/nouveau_svm.c if (buffer->fault) { fault 766 drivers/gpu/drm/nouveau/nouveau_svm.c for (i = 0; buffer->fault[i] && i < buffer->entries; i++) fault 767 drivers/gpu/drm/nouveau/nouveau_svm.c kfree(buffer->fault[i]); fault 768 drivers/gpu/drm/nouveau/nouveau_svm.c kvfree(buffer->fault); fault 806 drivers/gpu/drm/nouveau/nouveau_svm.c buffer->fault = kvzalloc(sizeof(*buffer->fault) * buffer->entries, GFP_KERNEL); fault 807 drivers/gpu/drm/nouveau/nouveau_svm.c if (!buffer->fault) fault 2164 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gp100_fault_new, fault 2200 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gp100_fault_new, fault 2236 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gp100_fault_new, fault 2272 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gp100_fault_new, fault 2308 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gp100_fault_new, fault 2344 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gp100_fault_new, fault 2378 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gp100_fault_new, fault 2404 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = gv100_fault_new, fault 2445 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = tu102_fault_new, fault 2480 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = tu102_fault_new, fault 2515 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = tu102_fault_new, fault 2550 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = tu102_fault_new, fault 2585 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .fault = tu102_fault_new, fault 2646 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c _(FAULT , device->fault , &device->fault->subdev); fault 3152 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c _(NVKM_SUBDEV_FAULT , fault); fault 372 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c else if (device->fault && index-- == 0) fault 373 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c sclass = &device->fault->user; fault 62 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c return fifo->func->fault(fifo, info); fault 675 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c .fault = gf100_fifo_fault, fault 427 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c const struct nvkm_enum *en = fifo->func->fault.engine; fault 477 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c er = nvkm_enum_find(fifo->func->fault.reason, info->reason); fault 478 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c ee = nvkm_enum_find(fifo->func->fault.engine, info->engine); fault 480 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c ec = nvkm_enum_find(fifo->func->fault.hubclient, info->client); fault 482 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c ec = nvkm_enum_find(fifo->func->fault.gpcclient, info->client); fault 485 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c ea = nvkm_enum_find(fifo->func->fault.access, info->access); fault 827 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c fifo->func->intr.fault(&fifo->base, unit); fault 1035 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c .fault = gk104_fifo_fault, fault 1182 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c .intr.fault = gf100_fifo_intr_fault, fault 1184 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c .fault.access = gk104_fifo_fault_access, fault 1185 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c .fault.engine = gk104_fifo_fault_engine, fault 1186 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c .fault.reason = gk104_fifo_fault_reason, fault 1187 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 1188 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 49 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.h void (*fault)(struct nvkm_fifo *, int unit); fault 64 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.h } fault; fault 51 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c .intr.fault = gf100_fifo_intr_fault, fault 53 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c .fault.access = gk104_fifo_fault_access, fault 54 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c .fault.engine = gk104_fifo_fault_engine, fault 55 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c .fault.reason = gk104_fifo_fault_reason, fault 56 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 57 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 48 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk208.c .intr.fault = gf100_fifo_intr_fault, fault 50 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk208.c .fault.access = gk104_fifo_fault_access, fault 51 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk208.c .fault.engine = gk104_fifo_fault_engine, fault 52 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk208.c .fault.reason = gk104_fifo_fault_reason, fault 53 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk208.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 54 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk208.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 29 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk20a.c .intr.fault = gf100_fifo_intr_fault, fault 31 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk20a.c .fault.access = gk104_fifo_fault_access, fault 32 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk20a.c .fault.engine = gk104_fifo_fault_engine, fault 33 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk20a.c .fault.reason = gk104_fifo_fault_reason, fault 34 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk20a.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 35 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk20a.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 97 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c .intr.fault = gm107_fifo_intr_fault, fault 99 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c .fault.access = gk104_fifo_fault_access, fault 100 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c .fault.engine = gm107_fifo_fault_engine, fault 101 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c .fault.reason = gk104_fifo_fault_reason, fault 102 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 103 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 45 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm200.c .intr.fault = gm107_fifo_intr_fault, fault 47 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm200.c .fault.access = gk104_fifo_fault_access, fault 48 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm200.c .fault.engine = gm107_fifo_fault_engine, fault 49 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm200.c .fault.reason = gk104_fifo_fault_reason, fault 50 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm200.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 51 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm200.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 29 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm20b.c .intr.fault = gm107_fifo_intr_fault, fault 31 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm20b.c .fault.access = gk104_fifo_fault_access, fault 32 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm20b.c .fault.engine = gm107_fifo_fault_engine, fault 33 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm20b.c .fault.reason = gk104_fifo_fault_reason, fault 34 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm20b.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 35 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm20b.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c .intr.fault = gp100_fifo_intr_fault, fault 83 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c .fault.access = gk104_fifo_fault_access, fault 84 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c .fault.engine = gp100_fifo_fault_engine, fault 85 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c .fault.reason = gk104_fifo_fault_reason, fault 86 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 87 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 29 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp10b.c .intr.fault = gp100_fifo_intr_fault, fault 31 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp10b.c .fault.access = gk104_fifo_fault_access, fault 32 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp10b.c .fault.engine = gp100_fifo_fault_engine, fault 33 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp10b.c .fault.reason = gk104_fifo_fault_reason, fault 34 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp10b.c .fault.hubclient = gk104_fifo_fault_hubclient, fault 35 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp10b.c .fault.gpcclient = gk104_fifo_fault_gpcclient, fault 292 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c .fault.access = gv100_fifo_fault_access, fault 293 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c .fault.engine = gv100_fifo_fault_engine, fault 294 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c .fault.reason = gv100_fifo_fault_reason, fault 295 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c .fault.hubclient = gv100_fifo_fault_hubclient, fault 296 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c .fault.gpcclient = gv100_fifo_fault_gpcclient, fault 25 drivers/gpu/drm/nouveau/nvkm/engine/fifo/priv.h void (*fault)(struct nvkm_fifo *, struct nvkm_fault_data *); fault 101 drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c .fault.access = gv100_fifo_fault_access, fault 102 drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c .fault.engine = tu102_fifo_fault_engine, fault 103 drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c .fault.reason = gv100_fifo_fault_reason, fault 104 drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c .fault.hubclient = gv100_fifo_fault_hubclient, fault 105 drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c .fault.gpcclient = gv100_fifo_fault_gpcclient, fault 30 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault = container_of(event, typeof(*fault), event); fault 31 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->func->buffer.intr(fault->buffer[index], false); fault 37 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault = container_of(event, typeof(*fault), event); fault 38 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->func->buffer.intr(fault->buffer[index], true); fault 65 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault = nvkm_fault(subdev); fault 66 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c return fault->func->intr(fault); fault 72 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault = nvkm_fault(subdev); fault 73 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c if (fault->func->fini) fault 74 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->func->fini(fault); fault 81 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault = nvkm_fault(subdev); fault 82 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c if (fault->func->init) fault 83 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->func->init(fault); fault 88 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c nvkm_fault_oneinit_buffer(struct nvkm_fault *fault, int id) fault 90 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_subdev *subdev = &fault->subdev; fault 97 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c buffer->fault = fault; fault 99 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->func->buffer.info(buffer); fault 100 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->buffer[id] = buffer; fault 105 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->func->buffer.entry_size, 0x1000, true, fault 121 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault = nvkm_fault(subdev); fault 124 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c for (i = 0; i < ARRAY_SIZE(fault->buffer); i++) { fault 125 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c if (i < fault->func->buffer.nr) { fault 126 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c ret = nvkm_fault_oneinit_buffer(fault, i); fault 129 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->buffer_nr = i + 1; fault 133 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c ret = nvkm_event_init(&nvkm_fault_ntfy, 1, fault->buffer_nr, fault 134 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c &fault->event); fault 138 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c if (fault->func->oneinit) fault 139 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c ret = fault->func->oneinit(fault); fault 146 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault = nvkm_fault(subdev); fault 149 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c nvkm_notify_fini(&fault->nrpfb); fault 150 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c nvkm_event_fini(&fault->event); fault 152 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c for (i = 0; i < fault->buffer_nr; i++) { fault 153 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c if (fault->buffer[i]) { fault 154 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c nvkm_memory_unref(&fault->buffer[i]->mem); fault 155 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c kfree(fault->buffer[i]); fault 159 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c return fault; fault 175 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault *fault; fault 176 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c if (!(fault = *pfault = kzalloc(sizeof(*fault), GFP_KERNEL))) fault 178 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c nvkm_subdev_ctor(&nvkm_fault, device, index, &fault->subdev); fault 179 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->func = func; fault 180 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->user.ctor = nvkm_ufault_new; fault 181 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->user.base = func->user.base; fault 31 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 38 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 45 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 54 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c buffer->entries = nvkm_rd32(buffer->fault->subdev.device, 0x002a78); fault 60 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c gp100_fault_intr(struct nvkm_fault *fault) fault 62 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c nvkm_event_send(&fault->event, 1, 0, NULL, 0); fault 33 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 42 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c const u32 base = get * buffer->fault->func->buffer.entry_size; fault 76 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 87 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 95 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 107 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_device *device = buffer->fault->subdev.device; fault 120 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_fault *fault = container_of(notify, typeof(*fault), nrpfb); fault 121 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c gv100_fault_buffer_process(fault->buffer[0]); fault 126 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c gv100_fault_intr_fault(struct nvkm_fault *fault) fault 128 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_subdev *subdev = &fault->subdev; fault 152 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c gv100_fault_intr(struct nvkm_fault *fault) fault 154 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c struct nvkm_subdev *subdev = &fault->subdev; fault 159 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c gv100_fault_intr_fault(fault); fault 165 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c if (fault->buffer[0]) { fault 166 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c nvkm_event_send(&fault->event, 1, 0, NULL, 0); fault 172 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c if (fault->buffer[1]) { fault 173 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c nvkm_event_send(&fault->event, 1, 1, NULL, 0); fault 184 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c gv100_fault_fini(struct nvkm_fault *fault) fault 186 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c nvkm_notify_put(&fault->nrpfb); fault 187 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c if (fault->buffer[0]) fault 188 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c fault->func->buffer.fini(fault->buffer[0]); fault 189 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c nvkm_mask(fault->subdev.device, 0x100a34, 0x80000000, 0x80000000); fault 193 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c gv100_fault_init(struct nvkm_fault *fault) fault 195 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c nvkm_mask(fault->subdev.device, 0x100a2c, 0x80000000, 0x80000000); fault 196 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c fault->func->buffer.init(fault->buffer[0]); fault 197 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c nvkm_notify_get(&fault->nrpfb); fault 201 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c gv100_fault_oneinit(struct nvkm_fault *fault) fault 203 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c return nvkm_notify_init(&fault->buffer[0]->object, &fault->event, fault 205 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c &fault->nrpfb); fault 12 drivers/gpu/drm/nouveau/nvkm/subdev/fault/priv.h struct nvkm_fault *fault; fault 42 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c struct nvkm_device *device = buffer->fault->subdev.device; fault 50 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c struct nvkm_device *device = buffer->fault->subdev.device; fault 62 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c struct nvkm_device *device = buffer->fault->subdev.device; fault 73 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c tu102_fault_intr_fault(struct nvkm_fault *fault) fault 75 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c struct nvkm_subdev *subdev = &fault->subdev; fault 99 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c tu102_fault_intr(struct nvkm_fault *fault) fault 101 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c struct nvkm_subdev *subdev = &fault->subdev; fault 106 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c tu102_fault_intr_fault(fault); fault 112 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c if (fault->buffer[0]) { fault 113 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c nvkm_event_send(&fault->event, 1, 0, NULL, 0); fault 120 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c if (fault->buffer[1]) { fault 121 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c nvkm_event_send(&fault->event, 1, 1, NULL, 0); fault 132 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c tu102_fault_fini(struct nvkm_fault *fault) fault 134 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c nvkm_notify_put(&fault->nrpfb); fault 135 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c if (fault->buffer[0]) fault 136 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c fault->func->buffer.fini(fault->buffer[0]); fault 141 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c tu102_fault_init(struct nvkm_fault *fault) fault 144 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c fault->func->buffer.init(fault->buffer[0]); fault 145 drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c nvkm_notify_get(&fault->nrpfb); fault 35 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c struct nvkm_device *device = buffer->fault->subdev.device; fault 48 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c *pevent = &buffer->fault->event; fault 58 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c buffer->fault->func->buffer.fini(buffer); fault 66 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c buffer->fault->func->buffer.init(buffer); fault 92 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c struct nvkm_fault *fault = device->fault; fault 93 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c struct nvkm_fault_buffer *buffer = fault->buffer[fault->func->user.rp]; fault 33 drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c info->fault = -1; fault 121 drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c if (info->index == devidx && info->fault >= 0) fault 122 drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c return info->fault; fault 129 drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c nvkm_top_fault(struct nvkm_device *device, int fault) fault 135 drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c if (info->fault == fault) fault 52 drivers/gpu/drm/nouveau/nvkm/subdev/top/gk104.c info->fault = (data & 0x000003f8) >> 3; fault 103 drivers/gpu/drm/nouveau/nvkm/subdev/top/gk104.c info->addr, info->fault, info->engine, info->runlist, fault 17 drivers/gpu/drm/nouveau/nvkm/subdev/top/priv.h int fault; fault 504 drivers/gpu/drm/omapdrm/omap_drv.c .fault = omap_gem_fault, fault 62 drivers/gpu/drm/qxl/qxl_ttm.c ret = ttm_vm_ops->fault(vmf); fault 86 drivers/gpu/drm/qxl/qxl_ttm.c qxl_ttm_vm_ops.fault = &qxl_ttm_fault; fault 898 drivers/gpu/drm/radeon/radeon_ttm.c ret = ttm_vm_ops->fault(vmf); fault 919 drivers/gpu/drm/radeon/radeon_ttm.c radeon_ttm_vm_ops.fault = &radeon_ttm_fault; fault 443 drivers/gpu/drm/tegra/gem.c .fault = tegra_bo_fault, fault 397 drivers/gpu/drm/ttm/ttm_bo_vm.c .fault = ttm_bo_vm_fault, fault 36 drivers/gpu/drm/udl/udl_drv.c .fault = udl_gem_fault, fault 144 drivers/gpu/drm/vc4/vc4_drv.c .fault = vc4_fault, fault 126 drivers/gpu/drm/vgem/vgem_drv.c .fault = vgem_gem_fault, fault 53 drivers/gpu/drm/vkms/vkms_drv.c .fault = vkms_gem_fault, fault 444 drivers/hid/usbhid/hiddev.c goto fault; fault 447 drivers/hid/usbhid/hiddev.c goto fault; fault 471 drivers/hid/usbhid/hiddev.c goto fault; fault 524 drivers/hid/usbhid/hiddev.c goto fault; fault 541 drivers/hid/usbhid/hiddev.c goto fault; fault 553 drivers/hid/usbhid/hiddev.c fault: fault 1101 drivers/hsi/clients/cmt_speech.c .fault = cs_char_vma_fault, fault 207 drivers/hwmon/emc2103.c bool fault = (data->temp[nr].degrees == -128); fault 208 drivers/hwmon/emc2103.c return sprintf(buf, "%d\n", fault ? 1 : 0); fault 408 drivers/hwmon/emc2103.c bool fault = ((data->fan_tach & 0x1fe0) == 0x1fe0); fault 409 drivers/hwmon/emc2103.c return sprintf(buf, "%d\n", fault ? 1 : 0); fault 325 drivers/hwmon/ltc2945.c unsigned int fault; fault 328 drivers/hwmon/ltc2945.c ret = regmap_read(regmap, LTC2945_FAULT, &fault); fault 332 drivers/hwmon/ltc2945.c fault &= attr->index; fault 333 drivers/hwmon/ltc2945.c if (fault) /* Clear reported faults in chip register */ fault 336 drivers/hwmon/ltc2945.c return snprintf(buf, PAGE_SIZE, "%d\n", !!fault); fault 105 drivers/hwmon/ltc4222.c unsigned int fault; fault 108 drivers/hwmon/ltc4222.c ret = regmap_read(regmap, attr->nr, &fault); fault 111 drivers/hwmon/ltc4222.c fault &= attr->index; fault 112 drivers/hwmon/ltc4222.c if (fault) /* Clear reported faults in chip register */ fault 115 drivers/hwmon/ltc4222.c return snprintf(buf, PAGE_SIZE, "%d\n", !!fault); fault 90 drivers/hwmon/ltc4260.c unsigned int fault; fault 93 drivers/hwmon/ltc4260.c ret = regmap_read(regmap, LTC4260_FAULT, &fault); fault 97 drivers/hwmon/ltc4260.c fault &= attr->index; fault 98 drivers/hwmon/ltc4260.c if (fault) /* Clear reported faults in chip register */ fault 101 drivers/hwmon/ltc4260.c return snprintf(buf, PAGE_SIZE, "%d\n", !!fault); fault 141 drivers/hwmon/ltc4261.c u8 fault; fault 146 drivers/hwmon/ltc4261.c fault = data->regs[LTC4261_FAULT] & attr->index; fault 147 drivers/hwmon/ltc4261.c if (fault) /* Clear reported faults in chip register */ fault 148 drivers/hwmon/ltc4261.c i2c_smbus_write_byte_data(data->client, LTC4261_FAULT, ~fault); fault 150 drivers/hwmon/ltc4261.c return snprintf(buf, PAGE_SIZE, "%d\n", fault ? 1 : 0); fault 96 drivers/hwmon/max16065.c int fault[2]; fault 165 drivers/hwmon/max16065.c data->fault[i] fault 180 drivers/hwmon/max16065.c int val = data->fault[attr2->nr]; fault 1599 drivers/hwtracing/intel_th/msu.c .fault = msc_mmap_fault, fault 204 drivers/iio/dac/ad5421.c unsigned int fault; fault 208 drivers/iio/dac/ad5421.c fault = ad5421_read(indio_dev, AD5421_REG_FAULT); fault 209 drivers/iio/dac/ad5421.c if (!fault) fault 231 drivers/iio/dac/ad5421.c if (fault == 0xffff) fault 232 drivers/iio/dac/ad5421.c fault = 0; fault 235 drivers/iio/dac/ad5421.c events = (old_fault ^ fault) & fault; fault 265 drivers/iio/dac/ad5421.c old_fault = fault; fault 266 drivers/iio/dac/ad5421.c fault = ad5421_read(indio_dev, AD5421_REG_FAULT); fault 269 drivers/iio/dac/ad5421.c if (fault & AD5421_FAULT_TRIGGER_IRQ) fault 272 drivers/iio/dac/ad5421.c } while (fault & AD5421_FAULT_TRIGGER_IRQ); fault 227 drivers/iio/temperature/max31856.c bool fault; fault 233 drivers/iio/temperature/max31856.c fault = reg_val & faultbit; fault 235 drivers/iio/temperature/max31856.c return sprintf(buf, "%d\n", fault); fault 929 drivers/infiniband/core/uverbs_main.c .fault = rdma_umap_fault, fault 111 drivers/infiniband/hw/hfi1/fault.c if (!ibd->fault->n_rxfaults[i] && !ibd->fault->n_txfaults[i]) fault 116 drivers/infiniband/hw/hfi1/fault.c (unsigned long long)ibd->fault->n_rxfaults[i], fault 117 drivers/infiniband/hw/hfi1/fault.c (unsigned long long)ibd->fault->n_txfaults[i]); fault 138 drivers/infiniband/hw/hfi1/fault.c struct fault *fault = file->private_data; fault 180 drivers/infiniband/hw/hfi1/fault.c bitmap_zero(fault->opcodes, sizeof(fault->opcodes) * fault 190 drivers/infiniband/hw/hfi1/fault.c clear_bit(i, fault->opcodes); fault 192 drivers/infiniband/hw/hfi1/fault.c set_bit(i, fault->opcodes); fault 212 drivers/infiniband/hw/hfi1/fault.c struct fault *fault = file->private_data; fault 213 drivers/infiniband/hw/hfi1/fault.c size_t bitsize = sizeof(fault->opcodes) * BITS_PER_BYTE; fault 221 drivers/infiniband/hw/hfi1/fault.c bit = find_first_bit(fault->opcodes, bitsize); fault 223 drivers/infiniband/hw/hfi1/fault.c zero = find_next_zero_bit(fault->opcodes, bitsize, bit); fault 232 drivers/infiniband/hw/hfi1/fault.c bit = find_next_bit(fault->opcodes, bitsize, zero); fault 253 drivers/infiniband/hw/hfi1/fault.c if (ibd->fault) fault 254 drivers/infiniband/hw/hfi1/fault.c debugfs_remove_recursive(ibd->fault->dir); fault 255 drivers/infiniband/hw/hfi1/fault.c kfree(ibd->fault); fault 256 drivers/infiniband/hw/hfi1/fault.c ibd->fault = NULL; fault 264 drivers/infiniband/hw/hfi1/fault.c ibd->fault = kzalloc(sizeof(*ibd->fault), GFP_KERNEL); fault 265 drivers/infiniband/hw/hfi1/fault.c if (!ibd->fault) fault 268 drivers/infiniband/hw/hfi1/fault.c ibd->fault->attr.interval = 1; fault 269 drivers/infiniband/hw/hfi1/fault.c ibd->fault->attr.require_end = ULONG_MAX; fault 270 drivers/infiniband/hw/hfi1/fault.c ibd->fault->attr.stacktrace_depth = 32; fault 271 drivers/infiniband/hw/hfi1/fault.c ibd->fault->attr.dname = NULL; fault 272 drivers/infiniband/hw/hfi1/fault.c ibd->fault->attr.verbose = 0; fault 273 drivers/infiniband/hw/hfi1/fault.c ibd->fault->enable = false; fault 274 drivers/infiniband/hw/hfi1/fault.c ibd->fault->opcode = false; fault 275 drivers/infiniband/hw/hfi1/fault.c ibd->fault->fault_skip = 0; fault 276 drivers/infiniband/hw/hfi1/fault.c ibd->fault->skip = 0; fault 277 drivers/infiniband/hw/hfi1/fault.c ibd->fault->direction = HFI1_FAULT_DIR_TXRX; fault 278 drivers/infiniband/hw/hfi1/fault.c ibd->fault->suppress_err = false; fault 279 drivers/infiniband/hw/hfi1/fault.c bitmap_zero(ibd->fault->opcodes, fault 280 drivers/infiniband/hw/hfi1/fault.c sizeof(ibd->fault->opcodes) * BITS_PER_BYTE); fault 283 drivers/infiniband/hw/hfi1/fault.c fault_create_debugfs_attr("fault", parent, &ibd->fault->attr); fault 285 drivers/infiniband/hw/hfi1/fault.c kfree(ibd->fault); fault 286 drivers/infiniband/hw/hfi1/fault.c ibd->fault = NULL; fault 289 drivers/infiniband/hw/hfi1/fault.c ibd->fault->dir = fault_dir; fault 293 drivers/infiniband/hw/hfi1/fault.c debugfs_create_bool("enable", 0600, fault_dir, &ibd->fault->enable); fault 295 drivers/infiniband/hw/hfi1/fault.c &ibd->fault->suppress_err); fault 297 drivers/infiniband/hw/hfi1/fault.c &ibd->fault->opcode); fault 298 drivers/infiniband/hw/hfi1/fault.c debugfs_create_file("opcodes", 0600, fault_dir, ibd->fault, fault 301 drivers/infiniband/hw/hfi1/fault.c &ibd->fault->fault_skip); fault 303 drivers/infiniband/hw/hfi1/fault.c &ibd->fault->fault_skip_usec); fault 304 drivers/infiniband/hw/hfi1/fault.c debugfs_create_u8("direction", 0600, fault_dir, &ibd->fault->direction); fault 311 drivers/infiniband/hw/hfi1/fault.c if (ibd->fault) fault 312 drivers/infiniband/hw/hfi1/fault.c return ibd->fault->suppress_err; fault 321 drivers/infiniband/hw/hfi1/fault.c if (!ibd->fault || !ibd->fault->enable) fault 323 drivers/infiniband/hw/hfi1/fault.c if (!(ibd->fault->direction & direction)) fault 325 drivers/infiniband/hw/hfi1/fault.c if (ibd->fault->opcode) { fault 326 drivers/infiniband/hw/hfi1/fault.c if (bitmap_empty(ibd->fault->opcodes, fault 327 drivers/infiniband/hw/hfi1/fault.c (sizeof(ibd->fault->opcodes) * fault 330 drivers/infiniband/hw/hfi1/fault.c if (!(test_bit(opcode, ibd->fault->opcodes))) fault 333 drivers/infiniband/hw/hfi1/fault.c if (ibd->fault->fault_skip_usec && fault 334 drivers/infiniband/hw/hfi1/fault.c time_before(jiffies, ibd->fault->skip_usec)) fault 336 drivers/infiniband/hw/hfi1/fault.c if (ibd->fault->fault_skip && ibd->fault->skip) { fault 337 drivers/infiniband/hw/hfi1/fault.c ibd->fault->skip--; fault 340 drivers/infiniband/hw/hfi1/fault.c ret = should_fail(&ibd->fault->attr, 1); fault 342 drivers/infiniband/hw/hfi1/fault.c ibd->fault->skip = ibd->fault->fault_skip; fault 343 drivers/infiniband/hw/hfi1/fault.c ibd->fault->skip_usec = jiffies + fault 344 drivers/infiniband/hw/hfi1/fault.c usecs_to_jiffies(ibd->fault->fault_skip_usec); fault 355 drivers/infiniband/hw/hfi1/fault.c ibd->fault->n_txfaults[opcode]++; fault 367 drivers/infiniband/hw/hfi1/fault.c ibd->fault->n_rxfaults[packet->opcode]++; fault 129 drivers/infiniband/hw/hfi1/file_ops.c .fault = vma_fault, fault 302 drivers/infiniband/hw/hfi1/verbs.h struct fault *fault; fault 890 drivers/infiniband/hw/qib/qib_file_ops.c .fault = qib_file_vma_fault, fault 2359 drivers/input/touchscreen/atmel_mxt_ts.c goto fault; fault 2375 drivers/input/touchscreen/atmel_mxt_ts.c goto fault; fault 2381 drivers/input/touchscreen/atmel_mxt_ts.c fault: fault 658 drivers/iommu/amd_iommu.c struct amd_iommu_fault fault; fault 665 drivers/iommu/amd_iommu.c fault.address = raw[1]; fault 666 drivers/iommu/amd_iommu.c fault.pasid = PPR_PASID(raw[0]); fault 667 drivers/iommu/amd_iommu.c fault.device_id = PPR_DEVID(raw[0]); fault 668 drivers/iommu/amd_iommu.c fault.tag = PPR_TAG(raw[0]); fault 669 drivers/iommu/amd_iommu.c fault.flags = PPR_FLAGS(raw[0]); fault 671 drivers/iommu/amd_iommu.c atomic_notifier_call_chain(&ppr_notifier, 0, &fault); fault 428 drivers/iommu/amd_iommu_v2.c static void handle_fault_error(struct fault *fault) fault 432 drivers/iommu/amd_iommu_v2.c if (!fault->dev_state->inv_ppr_cb) { fault 433 drivers/iommu/amd_iommu_v2.c set_pri_tag_status(fault->state, fault->tag, PPR_INVALID); fault 437 drivers/iommu/amd_iommu_v2.c status = fault->dev_state->inv_ppr_cb(fault->dev_state->pdev, fault 438 drivers/iommu/amd_iommu_v2.c fault->pasid, fault 439 drivers/iommu/amd_iommu_v2.c fault->address, fault 440 drivers/iommu/amd_iommu_v2.c fault->flags); fault 443 drivers/iommu/amd_iommu_v2.c set_pri_tag_status(fault->state, fault->tag, PPR_SUCCESS); fault 446 drivers/iommu/amd_iommu_v2.c set_pri_tag_status(fault->state, fault->tag, PPR_INVALID); fault 449 drivers/iommu/amd_iommu_v2.c set_pri_tag_status(fault->state, fault->tag, PPR_FAILURE); fault 456 drivers/iommu/amd_iommu_v2.c static bool access_error(struct vm_area_struct *vma, struct fault *fault) fault 460 drivers/iommu/amd_iommu_v2.c if (fault->flags & PPR_FAULT_EXEC) fault 463 drivers/iommu/amd_iommu_v2.c if (fault->flags & PPR_FAULT_READ) fault 466 drivers/iommu/amd_iommu_v2.c if (fault->flags & PPR_FAULT_WRITE) fault 474 drivers/iommu/amd_iommu_v2.c struct fault *fault = container_of(work, struct fault, work); fault 481 drivers/iommu/amd_iommu_v2.c mm = fault->state->mm; fault 482 drivers/iommu/amd_iommu_v2.c address = fault->address; fault 484 drivers/iommu/amd_iommu_v2.c if (fault->flags & PPR_FAULT_USER) fault 486 drivers/iommu/amd_iommu_v2.c if (fault->flags & PPR_FAULT_WRITE) fault 497 drivers/iommu/amd_iommu_v2.c if (access_error(vma, fault)) fault 506 drivers/iommu/amd_iommu_v2.c handle_fault_error(fault); fault 508 drivers/iommu/amd_iommu_v2.c finish_pri_tag(fault->dev_state, fault->state, fault->tag); fault 510 drivers/iommu/amd_iommu_v2.c put_pasid_state(fault->state); fault 512 drivers/iommu/amd_iommu_v2.c kfree(fault); fault 521 drivers/iommu/amd_iommu_v2.c struct fault *fault; fault 566 drivers/iommu/amd_iommu_v2.c fault = kzalloc(sizeof(*fault), GFP_ATOMIC); fault 567 drivers/iommu/amd_iommu_v2.c if (fault == NULL) { fault 573 drivers/iommu/amd_iommu_v2.c fault->dev_state = dev_state; fault 574 drivers/iommu/amd_iommu_v2.c fault->address = iommu_fault->address; fault 575 drivers/iommu/amd_iommu_v2.c fault->state = pasid_state; fault 576 drivers/iommu/amd_iommu_v2.c fault->tag = tag; fault 577 drivers/iommu/amd_iommu_v2.c fault->finish = finish; fault 578 drivers/iommu/amd_iommu_v2.c fault->pasid = iommu_fault->pasid; fault 579 drivers/iommu/amd_iommu_v2.c fault->flags = iommu_fault->flags; fault 580 drivers/iommu/amd_iommu_v2.c INIT_WORK(&fault->work, do_fault); fault 582 drivers/iommu/amd_iommu_v2.c queue_work(iommu_wq, &fault->work); fault 1158 drivers/iommu/dmar.c u32 fault; fault 1167 drivers/iommu/dmar.c fault = readl(iommu->reg + DMAR_FSTS_REG); fault 1174 drivers/iommu/dmar.c if (fault & DMA_FSTS_IQE) { fault 1198 drivers/iommu/dmar.c if (fault & DMA_FSTS_ITE) { fault 1217 drivers/iommu/dmar.c if (fault & DMA_FSTS_ICE) fault 1056 drivers/iommu/iommu.c if (evt->fault.type == IOMMU_FAULT_PAGE_REQ && fault 1057 drivers/iommu/iommu.c (evt->fault.prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE)) { fault 1069 drivers/iommu/iommu.c ret = fparam->handler(&evt->fault, fparam->data); fault 1113 drivers/iommu/iommu.c prm = &evt->fault.prm; fault 96 drivers/iommu/virtio-iommu.c struct virtio_iommu_fault fault; fault 524 drivers/iommu/virtio-iommu.c struct virtio_iommu_fault *fault) fault 528 drivers/iommu/virtio-iommu.c u8 reason = fault->reason; fault 529 drivers/iommu/virtio-iommu.c u32 flags = le32_to_cpu(fault->flags); fault 530 drivers/iommu/virtio-iommu.c u32 endpoint = le32_to_cpu(fault->endpoint); fault 531 drivers/iommu/virtio-iommu.c u64 address = le64_to_cpu(fault->address); fault 573 drivers/iommu/virtio-iommu.c viommu_fault_handler(viommu, &evt->fault); fault 193 drivers/leds/led-class-flash.c u32 fault, mask = 0x1; fault 197 drivers/leds/led-class-flash.c ret = led_get_flash_fault(fled_cdev, &fault); fault 204 drivers/leds/led-class-flash.c if (fault & mask) { fault 356 drivers/leds/led-class-flash.c int led_get_flash_fault(struct led_classdev_flash *fled_cdev, u32 *fault) fault 358 drivers/leds/led-class-flash.c return call_flash_op(fled_cdev, fault_get, fault); fault 251 drivers/leds/leds-as3645a.c static int as3645a_get_fault(struct led_classdev_flash *fled, u32 *fault) fault 262 drivers/leds/leds-as3645a.c *fault |= LED_FAULT_OVER_CURRENT; fault 265 drivers/leds/leds-as3645a.c *fault |= LED_FAULT_INDICATOR; fault 271 drivers/leds/leds-as3645a.c *fault |= LED_FAULT_TIMEOUT; fault 274 drivers/leds/leds-as3645a.c *fault |= LED_FAULT_OVER_TEMPERATURE; fault 277 drivers/leds/leds-as3645a.c *fault |= LED_FAULT_OVER_CURRENT; fault 280 drivers/leds/leds-as3645a.c *fault |= LED_FAULT_INPUT_VOLTAGE; fault 383 drivers/leds/leds-as3645a.c u32 fault = 0; fault 406 drivers/leds/leds-as3645a.c rval = as3645a_get_fault(&flash->fled, &fault); fault 303 drivers/leds/leds-lm3601x.c u32 *fault) fault 309 drivers/leds/leds-lm3601x.c *fault = led->last_flag; fault 182 drivers/leds/leds-lp8860.c int ret, fault; fault 189 drivers/leds/leds-lp8860.c fault = read_buf; fault 195 drivers/leds/leds-lp8860.c fault |= read_buf; fault 198 drivers/leds/leds-lp8860.c if (fault) fault 551 drivers/leds/leds-max77693.c u32 *fault) fault 555 drivers/leds/leds-max77693.c *fault = sub_led->flash_faults; fault 80 drivers/media/i2c/adp1653.c int fault; fault 83 drivers/media/i2c/adp1653.c fault = i2c_smbus_read_byte_data(client, ADP1653_REG_FAULT); fault 84 drivers/media/i2c/adp1653.c if (fault < 0) fault 85 drivers/media/i2c/adp1653.c return fault; fault 87 drivers/media/i2c/adp1653.c flash->fault |= fault; fault 89 drivers/media/i2c/adp1653.c if (!flash->fault) fault 103 drivers/media/i2c/adp1653.c return flash->fault; fault 152 drivers/media/i2c/adp1653.c if (flash->fault & ADP1653_REG_FAULT_FLT_SCP) fault 154 drivers/media/i2c/adp1653.c if (flash->fault & ADP1653_REG_FAULT_FLT_OT) fault 156 drivers/media/i2c/adp1653.c if (flash->fault & ADP1653_REG_FAULT_FLT_TMR) fault 158 drivers/media/i2c/adp1653.c if (flash->fault & ADP1653_REG_FAULT_FLT_OV) fault 161 drivers/media/i2c/adp1653.c flash->fault = 0; fault 200 drivers/media/i2c/adp1653.c struct v4l2_ctrl *fault; fault 241 drivers/media/i2c/adp1653.c fault = v4l2_ctrl_new_std(&flash->ctrls, &adp1653_ctrl_ops, fault 250 drivers/media/i2c/adp1653.c fault->flags |= V4L2_CTRL_FLAG_VOLATILE; fault 275 drivers/media/i2c/adp1653.c flash->fault = 0; fault 168 drivers/media/i2c/lm3560.c s32 fault = 0; fault 174 drivers/media/i2c/lm3560.c fault |= V4L2_FLASH_FAULT_SHORT_CIRCUIT; fault 176 drivers/media/i2c/lm3560.c fault |= V4L2_FLASH_FAULT_OVER_TEMPERATURE; fault 178 drivers/media/i2c/lm3560.c fault |= V4L2_FLASH_FAULT_TIMEOUT; fault 179 drivers/media/i2c/lm3560.c ctrl->cur.val = fault; fault 281 drivers/media/i2c/lm3560.c struct v4l2_ctrl *fault; fault 323 drivers/media/i2c/lm3560.c fault = v4l2_ctrl_new_std(hdl, ops, V4L2_CID_FLASH_FAULT, 0, fault 328 drivers/media/i2c/lm3560.c if (fault != NULL) fault 329 drivers/media/i2c/lm3560.c fault->flags |= V4L2_CTRL_FLAG_VOLATILE; fault 205 drivers/media/i2c/lm3646.c struct v4l2_ctrl *fault; fault 245 drivers/media/i2c/lm3646.c fault = v4l2_ctrl_new_std(hdl, ops, V4L2_CID_FLASH_FAULT, 0, fault 250 drivers/media/i2c/lm3646.c if (fault != NULL) fault 251 drivers/media/i2c/lm3646.c fault->flags |= V4L2_CTRL_FLAG_VOLATILE; fault 458 drivers/media/v4l2-core/videobuf-dma-sg.c .fault = videobuf_vm_fault, fault 176 drivers/misc/cxl/context.c .fault = cxl_mmap_fault, fault 491 drivers/misc/cxl/file.c event.fault.addr = ctx->fault_addr; fault 492 drivers/misc/cxl/file.c event.fault.dsisr = ctx->fault_dsisr; fault 482 drivers/misc/enclosure.c return snprintf(buf, 40, "%d\n", ecomp->fault); fault 642 drivers/misc/enclosure.c static DEVICE_ATTR(fault, S_IRUGO | S_IWUSR, get_component_fault, fault 156 drivers/misc/ocxl/context.c .fault = ocxl_mmap_fault, fault 133 drivers/misc/ocxl/link.c struct xsl_fault *fault = container_of(fault_work, struct xsl_fault, fault 135 drivers/misc/ocxl/link.c struct spa *spa = container_of(fault, struct spa, xsl_fault); fault 143 drivers/misc/ocxl/link.c rc = copro_handle_mm_fault(fault->pe_data.mm, fault->dar, fault->dsisr, fault 147 drivers/misc/ocxl/link.c if (fault->pe_data.xsl_err_cb) { fault 148 drivers/misc/ocxl/link.c fault->pe_data.xsl_err_cb( fault 149 drivers/misc/ocxl/link.c fault->pe_data.xsl_err_data, fault 150 drivers/misc/ocxl/link.c fault->dar, fault->dsisr); fault 163 drivers/misc/ocxl/link.c if (fault->dsisr & SPA_XSL_S) fault 166 drivers/misc/ocxl/link.c if (get_region_id(fault->dar) != USER_REGION_ID) fault 170 drivers/misc/ocxl/link.c hash_page_mm(fault->pe_data.mm, fault->dar, access, 0x300, fault 176 drivers/misc/ocxl/link.c mmput(fault->pe_data.mm); fault 89 drivers/misc/ocxl/sysfs.c .fault = global_mmio_fault, fault 594 drivers/misc/sgi-gru/grufile.c .fault = gru_fault, fault 1218 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_fault *fault) fault 1227 drivers/net/ethernet/intel/fm10k/fm10k_pci.c switch (fault->type) { fault 1241 drivers/net/ethernet/intel/fm10k/fm10k_pci.c switch (fault->type) { fault 1250 drivers/net/ethernet/intel/fm10k/fm10k_pci.c switch (fault->type) { fault 1274 drivers/net/ethernet/intel/fm10k/fm10k_pci.c error, fault->address, fault->specinfo, fault 1275 drivers/net/ethernet/intel/fm10k/fm10k_pci.c PCI_SLOT(fault->func), PCI_FUNC(fault->func)); fault 1286 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (fault->func && iov_data) { fault 1287 drivers/net/ethernet/intel/fm10k/fm10k_pci.c int vf = fault->func - 1; fault 1305 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_fault fault = { 0 }; fault 1316 drivers/net/ethernet/intel/fm10k/fm10k_pci.c err = hw->mac.ops.get_fault(hw, type, &fault); fault 1323 drivers/net/ethernet/intel/fm10k/fm10k_pci.c fm10k_handle_fault(interface, type, &fault); fault 1546 drivers/net/ethernet/intel/fm10k/fm10k_pf.c struct fm10k_fault *fault) fault 1566 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fault->address = fm10k_read_reg(hw, type + FM10K_FAULT_ADDR_HI); fault 1567 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fault->address <<= 32; fault 1568 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fault->address |= fm10k_read_reg(hw, type + FM10K_FAULT_ADDR_LO); fault 1569 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fault->specinfo = fm10k_read_reg(hw, type + FM10K_FAULT_SPECINFO); fault 1576 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fault->func = 0; fault 1578 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fault->func = 1 + ((func & FM10K_FAULT_FUNC_VF_MASK) >> fault 1582 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fault->type = func & FM10K_FAULT_FUNC_TYPE_MASK; fault 149 drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c bool fault; fault 157 drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c mlxsw_reg_fore_unpack(fore_pl, mlwsw_hwmon_attr->type_index, &fault); fault 159 drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c return sprintf(buf, "%u\n", fault); fault 239 drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c u8 module, fault; fault 260 drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c fault = 1; fault 266 drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c fault = 0; fault 270 drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c return sprintf(buf, "%u\n", fault); fault 8146 drivers/net/ethernet/mellanox/mlxsw/reg.h bool *fault) fault 8150 drivers/net/ethernet/mellanox/mlxsw/reg.h if (fault) { fault 8152 drivers/net/ethernet/mellanox/mlxsw/reg.h *fault = limit & BIT(tacho); fault 83 drivers/platform/chrome/cros_usbpd_logger.c const char *fault, *role, *chg_type; fault 132 drivers/platform/chrome/cros_usbpd_logger.c fault = r->data < ARRAY_SIZE(fault_names) ? fault_names[r->data] fault 134 drivers/platform/chrome/cros_usbpd_logger.c len += append_str(buf, len, "Power supply fault: %s", fault); fault 78 drivers/power/supply/bq24257_charger.c u8 fault; fault 340 drivers/power/supply/bq24257_charger.c switch (state.fault) { fault 436 drivers/power/supply/bq24257_charger.c state->fault = ret; fault 449 drivers/power/supply/bq24257_charger.c switch (state->fault) { fault 469 drivers/power/supply/bq24257_charger.c bq->state.fault != new_state->fault || fault 637 drivers/power/supply/bq24257_charger.c } else if (new_state->fault == FAULT_NO_BAT) { fault 639 drivers/power/supply/bq24257_charger.c } else if (new_state->fault == FAULT_TIMER) { fault 663 drivers/power/supply/bq24257_charger.c state.status, state.fault, state.power_good); fault 729 drivers/power/supply/bq24257_charger.c else if (state.fault != FAULT_NO_BAT) fault 107 drivers/power/supply/da9030_battery.c uint8_t fault; fault 225 drivers/power/supply/da9030_battery.c da903x_read(charger->master, DA9030_FAULT_LOG, &charger->fault); fault 330 drivers/power/supply/da9030_battery.c if (charger->fault & DA9030_FAULT_LOG_OVER_TEMP) fault 332 drivers/power/supply/da9030_battery.c else if (charger->fault & DA9030_FAULT_LOG_VBAT_OVER) fault 25 drivers/power/supply/max8903_charger.c bool fault; fault 61 drivers/power/supply/max8903_charger.c if (data->fault) fault 155 drivers/power/supply/max8903_charger.c bool fault; fault 157 drivers/power/supply/max8903_charger.c fault = gpio_get_value(pdata->flt) ? false : true; fault 159 drivers/power/supply/max8903_charger.c if (fault == data->fault) fault 162 drivers/power/supply/max8903_charger.c data->fault = fault; fault 164 drivers/power/supply/max8903_charger.c if (fault) fault 326 drivers/power/supply/max8903_charger.c data->fault = false; fault 572 drivers/regulator/wm8350-regulator.c u16 stop, u16 fault) fault 613 drivers/regulator/wm8350-regulator.c (fault << WM8350_DC1_ERRACT_SHIFT)); fault 1077 drivers/scsi/cxlflash/ocxl_hw.c event.fault.addr = ctx->fault_addr; fault 1078 drivers/scsi/cxlflash/ocxl_hw.c event.fault.dsisr = ctx->fault_dsisr; fault 1152 drivers/scsi/cxlflash/ocxl_hw.c .fault = ocxlflash_mmap_fault, fault 1134 drivers/scsi/cxlflash/superpipe.c rc = ctxi->cxl_mmap_vmops->fault(vmf); fault 1166 drivers/scsi/cxlflash/superpipe.c .fault = cxlflash_mmap_fault, fault 1656 drivers/scsi/libiscsi.c goto fault; fault 1698 drivers/scsi/libiscsi.c goto fault; fault 1705 drivers/scsi/libiscsi.c goto fault; fault 1711 drivers/scsi/libiscsi.c goto fault; fault 1766 drivers/scsi/libiscsi.c fault: fault 340 drivers/scsi/lpfc/lpfc.h uint8_t fault; fault 973 drivers/scsi/lpfc/lpfc_attr.c trunk_errmsg[link.link0.fault]); fault 980 drivers/scsi/lpfc/lpfc_attr.c trunk_errmsg[link.link1.fault]); fault 987 drivers/scsi/lpfc/lpfc_attr.c trunk_errmsg[link.link2.fault]); fault 994 drivers/scsi/lpfc/lpfc_attr.c trunk_errmsg[link.link3.fault]); fault 4927 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.link_state.fault = fault 4942 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.link_state.fault); fault 5076 drivers/scsi/lpfc/lpfc_init.c phba->trunk_link.link0.fault = port_fault & 0x1 ? err : 0; fault 5082 drivers/scsi/lpfc/lpfc_init.c phba->trunk_link.link1.fault = port_fault & 0x2 ? err : 0; fault 5088 drivers/scsi/lpfc/lpfc_init.c phba->trunk_link.link2.fault = port_fault & 0x4 ? err : 0; fault 5094 drivers/scsi/lpfc/lpfc_init.c phba->trunk_link.link3.fault = port_fault & 0x8 ? err : 0; fault 5166 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.link_state.fault = fault 5186 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.link_state.fault); fault 270 drivers/scsi/lpfc/lpfc_sli4.h uint8_t fault; fault 199 drivers/scsi/ses.c ecomp->fault = 0; fault 204 drivers/scsi/ses.c ecomp->fault = (desc[3] & 0x60) >> 4; fault 1260 drivers/scsi/sg.c .fault = sg_vma_fault, fault 102 drivers/spi/spi-tle62x0.c unsigned long fault = 0; fault 115 drivers/spi/spi-tle62x0.c fault <<= 8; fault 116 drivers/spi/spi-tle62x0.c fault |= ((unsigned long)buff[ptr]); fault 122 drivers/spi/spi-tle62x0.c bp += sprintf(bp, "%s ", decode_fault(fault >> (ptr * 2))); fault 703 drivers/staging/greybus/light.c u32 *fault) fault 726 drivers/staging/greybus/light.c *fault = le32_to_cpu(resp.fault); fault 533 drivers/staging/iio/resolver/ad2s1210.c static IIO_DEVICE_ATTR(fault, 0644, fault 1564 drivers/target/target_core_user.c .fault = tcmu_vma_fault, fault 709 drivers/uio/uio.c .fault = uio_vma_fault, fault 1264 drivers/usb/mon/mon_bin.c .fault = mon_bin_vma_fault, fault 136 drivers/vfio/pci/vfio_pci_nvlink2.c .fault = vfio_pci_nvgpu_mmap_fault, fault 146 drivers/video/backlight/sky81452-backlight.c static DEVICE_ATTR(fault, S_IRUGO, sky81452_bl_show_fault, NULL); fault 150 drivers/video/fbdev/core/fb_defio.c .fault = fb_deferred_io_fault, fault 129 drivers/xen/privcmd-buf.c .fault = privcmd_buf_vma_fault, fault 923 drivers/xen/privcmd.c .fault = privcmd_fault fault 634 fs/9p/vfs_file.c .fault = filemap_fault, fault 641 fs/9p/vfs_file.c .fault = filemap_fault, fault 62 fs/afs/file.c .fault = filemap_fault, fault 359 fs/aio.c .fault = filemap_fault, fault 2242 fs/btrfs/file.c .fault = filemap_fault, fault 1794 fs/ceph/addr.c .fault = ceph_filemap_fault, fault 4049 fs/cifs/file.c .fault = filemap_fault, fault 112 fs/ext2/file.c .fault = ext2_dax_fault, fault 356 fs/ext4/file.c .fault = ext4_dax_fault, fault 366 fs/ext4/file.c .fault = ext4_filemap_fault, fault 132 fs/f2fs/file.c .fault = f2fs_filemap_fault, fault 2294 fs/fuse/file.c .fault = filemap_fault, fault 555 fs/gfs2/file.c .fault = filemap_fault, fault 363 fs/kernfs/file.c if (of->vm_ops->fault) fault 364 fs/kernfs/file.c ret = of->vm_ops->fault(vmf); fault 460 fs/kernfs/file.c .fault = kernfs_vma_fault, fault 578 fs/nfs/file.c .fault = filemap_fault, fault 115 fs/nilfs2/file.c .fault = filemap_fault, fault 160 fs/ocfs2/mmap.c .fault = ocfs2_fault, fault 453 fs/orangefs/file.c .fault = orangefs_fault, fault 445 fs/proc/vmcore.c .fault = mmap_vmcore_fault, fault 1597 fs/ubifs/file.c .fault = filemap_fault, fault 1219 fs/xfs/xfs_file.c .fault = xfs_filemap_fault, fault 21 include/asm-generic/termios-base.h goto fault; fault 25 include/asm-generic/termios-base.h goto fault; fault 29 include/asm-generic/termios-base.h goto fault; fault 33 include/asm-generic/termios-base.h goto fault; fault 37 include/asm-generic/termios-base.h goto fault; fault 40 include/asm-generic/termios-base.h goto fault; fault 44 include/asm-generic/termios-base.h fault: fault 26 include/asm-generic/termios.h goto fault; fault 30 include/asm-generic/termios.h goto fault; fault 34 include/asm-generic/termios.h goto fault; fault 38 include/asm-generic/termios.h goto fault; fault 42 include/asm-generic/termios.h goto fault; fault 45 include/asm-generic/termios.h goto fault; fault 49 include/asm-generic/termios.h fault: fault 135 include/linux/cdrom.h __u8 fault : 1; fault 144 include/linux/cdrom.h __u8 fault : 1; fault 41 include/linux/ceph/messenger.h void (*fault) (struct ceph_connection *con); fault 90 include/linux/enclosure.h int fault; fault 1781 include/linux/greybus/greybus_protocols.h __le32 fault; fault 337 include/linux/iommu.h struct iommu_fault fault; fault 47 include/linux/led-class-flash.h int (*fault_get)(struct led_classdev_flash *fled_cdev, u32 *fault); fault 195 include/linux/led-class-flash.h u32 *fault); fault 759 include/linux/mfd/wm8350/pmic.h u16 stop, u16 fault); fault 472 include/linux/mm.h vm_fault_t (*fault)(struct vm_fault *vmf); fault 732 include/linux/mm_types.h vm_fault_t (*fault)(const struct vm_special_mapping *sm, fault 111 include/media/i2c/adp1653.h int fault; fault 150 include/uapi/misc/cxl.h struct cxl_event_data_storage fault; fault 56 init/do_mounts_md.c int minor, level, factor, fault, partitioned = 0; fault 87 init/do_mounts_md.c get_option(&str, &fault) != 2) { fault 434 ipc/shm.c return sfd->vm_ops->fault(vmf); fault 508 ipc/shm.c WARN_ON(!sfd->vm_ops->fault); fault 584 ipc/shm.c .fault = shm_fault, fault 2392 kernel/bpf/syscall.c bool fault; fault 2403 kernel/bpf/syscall.c fault = copy_to_user(uinsns, insns_sanitized, ulen); fault 2405 kernel/bpf/syscall.c if (fault) fault 5700 kernel/events/core.c .fault = perf_mmap_fault, fault 1504 kernel/events/uprobes.c area->xol_mapping.fault = NULL; fault 64 kernel/relay.c .fault = relay_buf_fault, fault 2698 mm/filemap.c .fault = filemap_fault, fault 134 mm/gup.c if ((flags & FOLL_DUMP) && (!vma->vm_ops || !vma->vm_ops->fault)) fault 284 mm/hmm.c bool fault, bool write_fault, fault 300 mm/hmm.c if (fault || write_fault) { fault 310 mm/hmm.c return (fault || write_fault) ? -EBUSY : 0; fault 315 mm/hmm.c bool *fault, bool *write_fault) fault 342 mm/hmm.c *fault = true; fault 348 mm/hmm.c *fault = !(cpu_flags & range->flags[HMM_PFN_VALID]); fault 353 mm/hmm.c *fault = true; fault 359 mm/hmm.c uint64_t cpu_flags, bool *fault, fault 365 mm/hmm.c *fault = *write_fault = false; fault 369 mm/hmm.c *fault = *write_fault = false; fault 372 mm/hmm.c fault, write_fault); fault 383 mm/hmm.c bool fault, write_fault; fault 391 mm/hmm.c 0, &fault, &write_fault); fault 392 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); fault 411 mm/hmm.c bool fault, write_fault; fault 417 mm/hmm.c &fault, &write_fault); fault 419 mm/hmm.c if (pmd_protnone(pmd) || fault || write_fault) fault 420 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); fault 460 mm/hmm.c bool fault, write_fault; fault 466 mm/hmm.c fault = write_fault = false; fault 470 mm/hmm.c &fault, &write_fault); fault 471 mm/hmm.c if (fault || write_fault) fault 472 mm/hmm.c goto fault; fault 482 mm/hmm.c &fault, &write_fault); fault 483 mm/hmm.c if (fault || write_fault) fault 484 mm/hmm.c goto fault; fault 498 mm/hmm.c &fault, &write_fault); fault 499 mm/hmm.c if (fault || write_fault) fault 500 mm/hmm.c goto fault; fault 508 mm/hmm.c if (fault || write_fault) { fault 523 mm/hmm.c &fault, &write_fault); fault 526 mm/hmm.c if (fault || write_fault) fault 527 mm/hmm.c goto fault; fault 542 mm/hmm.c fault: fault 549 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); fault 570 mm/hmm.c bool fault, write_fault; fault 579 mm/hmm.c 0, &fault, &write_fault); fault 580 mm/hmm.c if (fault || write_fault) { fault 674 mm/hmm.c bool fault, write_fault; fault 685 mm/hmm.c cpu_flags, &fault, &write_fault); fault 686 mm/hmm.c if (fault || write_fault) fault 687 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, fault 735 mm/hmm.c bool fault, write_fault; fault 747 mm/hmm.c fault = write_fault = false; fault 749 mm/hmm.c &fault, &write_fault); fault 750 mm/hmm.c if (fault || write_fault) { fault 765 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); fault 3357 mm/hugetlb.c .fault = hugetlb_vm_op_fault, fault 525 mm/memory.c vma->vm_ops ? vma->vm_ops->fault : NULL, fault 3108 mm/memory.c ret = vma->vm_ops->fault(vmf); fault 3607 mm/memory.c if (!vma->vm_ops->fault) { fault 3353 mm/mmap.c .fault = special_mapping_fault, fault 3360 mm/mmap.c .fault = special_mapping_fault, fault 3374 mm/mmap.c if (sm->fault) fault 3375 mm/mmap.c return sm->fault(sm, vmf->vma, vmf); fault 3864 mm/shmem.c .fault = shmem_fault, fault 737 mm/slub.c u8 *fault; fault 741 mm/slub.c fault = memchr_inv(start, value, bytes); fault 743 mm/slub.c if (!fault) fault 747 mm/slub.c while (end > fault && end[-1] == value) fault 752 mm/slub.c fault, end - 1, fault[0], value); fault 755 mm/slub.c restore_bytes(s, what, value, fault, end); fault 822 mm/slub.c u8 *fault; fault 840 mm/slub.c fault = memchr_inv(pad, POISON_INUSE, remainder); fault 842 mm/slub.c if (!fault) fault 844 mm/slub.c while (end > fault && end[-1] == POISON_INUSE) fault 847 mm/slub.c slab_err(s, page, "Padding overwritten. 0x%p-0x%p", fault, end - 1); fault 850 mm/slub.c restore_bytes(s, "slab padding", POISON_INUSE, fault, end); fault 770 net/bluetooth/hidp/core.c goto fault; fault 806 net/bluetooth/hidp/core.c fault: fault 2910 net/ceph/messenger.c if (con->ops->fault) fault 2911 net/ceph/messenger.c con->ops->fault(con); fault 2921 net/ceph/messenger.c bool fault; fault 2927 net/ceph/messenger.c if ((fault = con_sock_closed(con))) { fault 2955 net/ceph/messenger.c fault = true; fault 2965 net/ceph/messenger.c fault = true; fault 2970 net/ceph/messenger.c if (fault) fault 2974 net/ceph/messenger.c if (fault) fault 1379 net/ceph/mon_client.c .fault = mon_fault, fault 5671 net/ceph/osd_client.c .fault = osd_fault, fault 476 net/core/datagram.c goto fault; fault 491 net/core/datagram.c fault: fault 497 net/core/datagram.c goto fault; fault 564 net/core/datagram.c goto fault; fault 587 net/core/datagram.c goto fault; fault 608 net/core/datagram.c goto fault; fault 618 net/core/datagram.c fault: fault 732 net/core/datagram.c goto fault; fault 737 net/core/datagram.c goto fault; fault 749 net/core/datagram.c fault: fault 2195 net/core/skbuff.c goto fault; fault 2249 net/core/skbuff.c goto fault; fault 2261 net/core/skbuff.c fault: fault 2557 net/core/skbuff.c goto fault; fault 2611 net/core/skbuff.c goto fault; fault 2622 net/core/skbuff.c fault: fault 765 samples/vfio-mdev/mbochs.c .fault = mbochs_region_vm_fault, fault 800 samples/vfio-mdev/mbochs.c .fault = mbochs_dmabuf_vm_fault, fault 477 security/selinux/selinuxfs.c .fault = sel_mmap_policy_fault, fault 3254 sound/core/pcm_native.c .fault = snd_pcm_mmap_status_fault, fault 3290 sound/core/pcm_native.c .fault = snd_pcm_mmap_control_fault, fault 3398 sound/core/pcm_native.c .fault = snd_pcm_mmap_data_fault, fault 4319 sound/soc/codecs/wm_adsp.c unsigned int fault[6]; fault 4330 sound/soc/codecs/wm_adsp.c fault); fault 4337 sound/soc/codecs/wm_adsp.c *fault & HALO_AHBM_FLAGS_ERR_MASK, fault 4338 sound/soc/codecs/wm_adsp.c (*fault & HALO_AHBM_CORE_ERR_ADDR_MASK) >> fault 4342 sound/soc/codecs/wm_adsp.c fault); fault 4348 sound/soc/codecs/wm_adsp.c adsp_warn(dsp, "AHB: SYS_ADDR: 0x%x\n", *fault); fault 4351 sound/soc/codecs/wm_adsp.c fault, ARRAY_SIZE(fault)); fault 4357 sound/soc/codecs/wm_adsp.c adsp_warn(dsp, "XM: STATUS:0x%x ADDR:0x%x\n", fault[1], fault[0]); fault 4358 sound/soc/codecs/wm_adsp.c adsp_warn(dsp, "YM: STATUS:0x%x ADDR:0x%x\n", fault[3], fault[2]); fault 4359 sound/soc/codecs/wm_adsp.c adsp_warn(dsp, "PM: STATUS:0x%x ADDR:0x%x\n", fault[5], fault[4]); fault 174 sound/usb/usx2y/us122l.c .fault = usb_stream_hwdep_vm_fault, fault 44 sound/usb/usx2y/usX2Yhwdep.c .fault = snd_us428ctls_vm_fault, fault 661 sound/usb/usx2y/usx2yhwdeppcm.c .fault = snd_usX2Y_hwdep_pcm_vm_fault, fault 2734 virt/kvm/kvm_main.c .fault = kvm_vcpu_fault,