Lines Matching refs:lockup_detected

2011 static u32 lockup_detected(struct ctlr_info *h)  in lockup_detected()  function
2014 u32 rc, *lockup_detected; in lockup_detected() local
2017 lockup_detected = per_cpu_ptr(h->lockup_detected, cpu); in lockup_detected()
2018 rc = *lockup_detected; in lockup_detected()
2027 if (unlikely(lockup_detected(h))) in hpsa_scsi_do_simple_cmd_core_if_no_lockup()
4105 if (unlikely(lockup_detected(h))) { in hpsa_scsi_queue_command()
4115 if (unlikely(lockup_detected(h))) { in hpsa_scsi_queue_command()
4176 if (unlikely(lockup_detected(h))) in hpsa_scan_start()
4195 if (unlikely(lockup_detected(h))) in hpsa_scan_start()
4354 if (lockup_detected(h)) in hpsa_eh_device_reset_handler()
4580 if (lockup_detected(h)) in hpsa_eh_abort_handler()
6638 u32 *lockup_detected; in set_lockup_detected_for_all_cpus() local
6639 lockup_detected = per_cpu_ptr(h->lockup_detected, cpu); in set_lockup_detected_for_all_cpus()
6640 *lockup_detected = value; in set_lockup_detected_for_all_cpus()
6648 u32 lockup_detected; in controller_lockup_detected() local
6652 lockup_detected = readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in controller_lockup_detected()
6653 if (!lockup_detected) { in controller_lockup_detected()
6657 lockup_detected = 0xffffffff; in controller_lockup_detected()
6659 set_lockup_detected_for_all_cpus(h, lockup_detected); in controller_lockup_detected()
6662 lockup_detected); in controller_lockup_detected()
6818 if (lockup_detected(h)) in hpsa_monitor_ctlr_worker()
6895 h->lockup_detected = alloc_percpu(u32); in hpsa_init_one()
6896 if (!h->lockup_detected) { in hpsa_init_one()
7033 if (h->lockup_detected) in hpsa_init_one()
7034 free_percpu(h->lockup_detected); in hpsa_init_one()
7045 if (unlikely(lockup_detected(h))) in hpsa_flush_cache()
7132 free_percpu(h->lockup_detected); in hpsa_remove_one()