hrrq 696 drivers/scsi/ipr.c struct ipr_cmnd *__ipr_get_free_ipr_cmnd(struct ipr_hrr_queue *hrrq) hrrq 700 drivers/scsi/ipr.c if (likely(!list_empty(&hrrq->hrrq_free_q))) { hrrq 701 drivers/scsi/ipr.c ipr_cmd = list_entry(hrrq->hrrq_free_q.next, hrrq 721 drivers/scsi/ipr.c __ipr_get_free_ipr_cmnd(&ioa_cfg->hrrq[IPR_INIT_HRRQ]); hrrq 745 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 746 drivers/scsi/ipr.c ioa_cfg->hrrq[i].allow_interrupts = 0; hrrq 747 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 829 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 844 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; hrrq 847 drivers/scsi/ipr.c spin_lock_irqsave(&hrrq->_lock, hrrq_flags); hrrq 849 drivers/scsi/ipr.c spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); hrrq 872 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 888 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; hrrq 890 drivers/scsi/ipr.c spin_lock_irqsave(&hrrq->_lock, hrrq_flags); hrrq 892 drivers/scsi/ipr.c spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); hrrq 907 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 910 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 911 drivers/scsi/ipr.c spin_lock(&hrrq->_lock); hrrq 913 drivers/scsi/ipr.c temp, &hrrq->hrrq_pending_q, queue) { hrrq 931 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 982 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); hrrq 1084 drivers/scsi/ipr.c unsigned int hrrq; hrrq 1087 drivers/scsi/ipr.c hrrq = 0; hrrq 1089 drivers/scsi/ipr.c hrrq = atomic_add_return(1, &ioa_cfg->hrrq_index); hrrq 1090 drivers/scsi/ipr.c hrrq = (hrrq % (ioa_cfg->hrrq_num - 1)) + 1; hrrq 1092 drivers/scsi/ipr.c return hrrq; hrrq 1114 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) { hrrq 1116 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); hrrq 1505 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 2649 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 3340 drivers/scsi/ipr.c if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) { hrrq 3649 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) hrrq 3681 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead && hrrq 3684 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 3685 drivers/scsi/ipr.c ioa_cfg->hrrq[i].ioa_is_dead = 0; hrrq 3686 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 3809 drivers/scsi/ipr.c irq_poll_disable(&ioa_cfg->hrrq[i].iopoll); hrrq 3816 drivers/scsi/ipr.c irq_poll_init(&ioa_cfg->hrrq[i].iopoll, hrrq 4390 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead && !ioa_cfg->dump_taken) { hrrq 5072 drivers/scsi/ipr.c list_for_each_entry(loop_cmd, &ipr_cmd->hrrq->hrrq_free_q, queue) { hrrq 5112 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 5120 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 5121 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, flags); hrrq 5122 drivers/scsi/ipr.c for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { hrrq 5131 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, flags); hrrq 5140 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 5141 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, flags); hrrq 5142 drivers/scsi/ipr.c for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { hrrq 5151 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, flags); hrrq 5176 drivers/scsi/ipr.c if (!ioa_cfg->in_reset_reload && !ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { hrrq 5191 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { hrrq 5246 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 5326 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 5339 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) hrrq 5342 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 5343 drivers/scsi/ipr.c spin_lock(&hrrq->_lock); hrrq 5344 drivers/scsi/ipr.c for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { hrrq 5360 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 5437 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 5499 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 5510 drivers/scsi/ipr.c ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) hrrq 5525 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 5526 drivers/scsi/ipr.c spin_lock(&hrrq->_lock); hrrq 5527 drivers/scsi/ipr.c for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { hrrq 5535 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 5562 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 5584 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead || ioa_cfg->scan_done) hrrq 5766 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 5772 drivers/scsi/ipr.c hrrq = container_of(iop, struct ipr_hrr_queue, iopoll); hrrq 5773 drivers/scsi/ipr.c ioa_cfg = hrrq->ioa_cfg; hrrq 5775 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, hrrq_flags); hrrq 5776 drivers/scsi/ipr.c completed_ops = ipr_process_hrrq(hrrq, budget, &doneq); hrrq 5780 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 5801 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp; hrrq 5802 drivers/scsi/ipr.c struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg; hrrq 5811 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, hrrq_flags); hrrq 5813 drivers/scsi/ipr.c if (!hrrq->allow_interrupts) { hrrq 5814 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 5819 drivers/scsi/ipr.c if (ipr_process_hrrq(hrrq, -1, &doneq)) { hrrq 5850 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 5869 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp; hrrq 5870 drivers/scsi/ipr.c struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg; hrrq 5876 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, hrrq_flags); hrrq 5879 drivers/scsi/ipr.c if (!hrrq->allow_interrupts) { hrrq 5880 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 5885 drivers/scsi/ipr.c if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) == hrrq 5886 drivers/scsi/ipr.c hrrq->toggle_bit) { hrrq 5887 drivers/scsi/ipr.c irq_poll_sched(&hrrq->iopoll); hrrq 5888 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 5892 drivers/scsi/ipr.c if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) == hrrq 5893 drivers/scsi/ipr.c hrrq->toggle_bit) hrrq 5895 drivers/scsi/ipr.c if (ipr_process_hrrq(hrrq, -1, &doneq)) hrrq 5899 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6056 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 6071 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; hrrq 6074 drivers/scsi/ipr.c spin_lock_irqsave(&hrrq->_lock, hrrq_flags); hrrq 6076 drivers/scsi/ipr.c spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); hrrq 6158 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; hrrq 6161 drivers/scsi/ipr.c spin_lock_irqsave(&hrrq->_lock, hrrq_flags); hrrq 6163 drivers/scsi/ipr.c spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); hrrq 6494 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 6519 drivers/scsi/ipr.c spin_lock_irqsave(ipr_cmd->hrrq->lock, lock_flags); hrrq 6523 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 6524 drivers/scsi/ipr.c spin_unlock_irqrestore(ipr_cmd->hrrq->lock, lock_flags); hrrq 6527 drivers/scsi/ipr.c spin_lock(&ipr_cmd->hrrq->_lock); hrrq 6529 drivers/scsi/ipr.c spin_unlock(&ipr_cmd->hrrq->_lock); hrrq 6555 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 6571 drivers/scsi/ipr.c hrrq = &ioa_cfg->hrrq[hrrq_id]; hrrq 6573 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, hrrq_flags); hrrq 6579 drivers/scsi/ipr.c if (unlikely(!hrrq->allow_cmds && !hrrq->ioa_is_dead && !hrrq->removing_ioa)) { hrrq 6580 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6588 drivers/scsi/ipr.c if (unlikely(hrrq->ioa_is_dead || hrrq->removing_ioa || !res)) { hrrq 6589 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6593 drivers/scsi/ipr.c ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq); hrrq 6595 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6598 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6643 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, hrrq_flags); hrrq 6644 drivers/scsi/ipr.c if (unlikely(rc || (!hrrq->allow_cmds && !hrrq->ioa_is_dead))) { hrrq 6645 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q); hrrq 6646 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6652 drivers/scsi/ipr.c if (unlikely(hrrq->ioa_is_dead)) { hrrq 6653 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q); hrrq 6654 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6664 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_pending_q); hrrq 6667 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6671 drivers/scsi/ipr.c spin_lock_irqsave(hrrq->lock, hrrq_flags); hrrq 6675 drivers/scsi/ipr.c spin_unlock_irqrestore(hrrq->lock, hrrq_flags); hrrq 6773 drivers/scsi/ipr.c if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) hrrq 6804 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 6814 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 6815 drivers/scsi/ipr.c spin_lock(&hrrq->_lock); hrrq 6816 drivers/scsi/ipr.c list_for_each_entry(ipr_cmd, &hrrq->hrrq_pending_q, queue) { hrrq 6822 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 6871 drivers/scsi/ipr.c spin_lock(&ipr_cmd->hrrq->_lock); hrrq 6887 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 6888 drivers/scsi/ipr.c spin_unlock(&ipr_cmd->hrrq->_lock); hrrq 6996 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 7000 drivers/scsi/ipr.c hrrq = &ioa_cfg->hrrq[hrrq_id]; hrrq 7003 drivers/scsi/ipr.c spin_lock(&hrrq->_lock); hrrq 7004 drivers/scsi/ipr.c if (unlikely(hrrq->ioa_is_dead)) { hrrq 7005 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 7009 drivers/scsi/ipr.c if (unlikely(!hrrq->allow_cmds)) { hrrq 7010 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 7014 drivers/scsi/ipr.c ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq); hrrq 7016 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 7021 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 7050 drivers/scsi/ipr.c spin_lock(&ipr_cmd->hrrq->_lock); hrrq 7051 drivers/scsi/ipr.c if (unlikely(!ipr_cmd->hrrq->allow_cmds || hrrq 7052 drivers/scsi/ipr.c ipr_cmd->hrrq->ioa_is_dead)) { hrrq 7053 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 7054 drivers/scsi/ipr.c spin_unlock(&ipr_cmd->hrrq->_lock); hrrq 7070 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); hrrq 7110 drivers/scsi/ipr.c spin_unlock(&ipr_cmd->hrrq->_lock); hrrq 7115 drivers/scsi/ipr.c spin_unlock(&ipr_cmd->hrrq->_lock); hrrq 7224 drivers/scsi/ipr.c if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { hrrq 7233 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 7234 drivers/scsi/ipr.c ioa_cfg->hrrq[i].ioa_is_dead = 1; hrrq 7235 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 7239 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 7266 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[j]._lock); hrrq 7267 drivers/scsi/ipr.c ioa_cfg->hrrq[j].allow_cmds = 1; hrrq 7268 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[j]._lock); hrrq 7298 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 7633 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 8141 drivers/scsi/ipr.c &ioa_cfg->hrrq->hrrq_free_q); hrrq 8194 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 8202 drivers/scsi/ipr.c hrrq = &ioa_cfg->hrrq[ioa_cfg->identify_hrrq_index]; hrrq 8217 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma >> 24) & 0xff; hrrq 8219 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma >> 16) & 0xff; hrrq 8221 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma >> 8) & 0xff; hrrq 8223 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma) & 0xff; hrrq 8225 drivers/scsi/ipr.c ((sizeof(u32) * hrrq->size) >> 8) & 0xff; hrrq 8227 drivers/scsi/ipr.c (sizeof(u32) * hrrq->size) & 0xff; hrrq 8235 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma >> 56) & 0xff; hrrq 8237 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma >> 48) & 0xff; hrrq 8239 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma >> 40) & 0xff; hrrq 8241 drivers/scsi/ipr.c ((u64) hrrq->host_rrq_dma >> 32) & 0xff; hrrq 8310 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); hrrq 8327 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 8329 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 8330 drivers/scsi/ipr.c spin_lock(&hrrq->_lock); hrrq 8331 drivers/scsi/ipr.c memset(hrrq->host_rrq, 0, sizeof(u32) * hrrq->size); hrrq 8334 drivers/scsi/ipr.c hrrq->hrrq_start = hrrq->host_rrq; hrrq 8335 drivers/scsi/ipr.c hrrq->hrrq_end = &hrrq->host_rrq[hrrq->size - 1]; hrrq 8336 drivers/scsi/ipr.c hrrq->hrrq_curr = hrrq->hrrq_start; hrrq 8337 drivers/scsi/ipr.c hrrq->toggle_bit = 1; hrrq 8338 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 8403 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); hrrq 8430 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 8431 drivers/scsi/ipr.c ioa_cfg->hrrq[i].allow_interrupts = 1; hrrq 8432 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 8472 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); hrrq 8994 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq; hrrq 9001 drivers/scsi/ipr.c for_each_hrrq(hrrq, ioa_cfg) { hrrq 9002 drivers/scsi/ipr.c spin_lock(&hrrq->_lock); hrrq 9003 drivers/scsi/ipr.c list_for_each_entry(loop_cmd, &hrrq->hrrq_pending_q, queue) { hrrq 9006 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 9010 drivers/scsi/ipr.c spin_unlock(&hrrq->_lock); hrrq 9035 drivers/scsi/ipr.c struct ipr_hrr_queue *hrrq = &ioa_cfg->hrrq[IPR_INIT_HRRQ]; hrrq 9040 drivers/scsi/ipr.c if (!hrrq->ioa_is_dead) { hrrq 9042 drivers/scsi/ipr.c list_for_each_entry(hcam_cmd, &hrrq->hrrq_pending_q, queue) { hrrq 9161 drivers/scsi/ipr.c !ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { hrrq 9210 drivers/scsi/ipr.c &ipr_cmd->hrrq->hrrq_free_q); hrrq 9249 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 9250 drivers/scsi/ipr.c ioa_cfg->hrrq[i].allow_cmds = 0; hrrq 9251 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 9254 drivers/scsi/ipr.c if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { hrrq 9285 drivers/scsi/ipr.c if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) hrrq 9301 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 9302 drivers/scsi/ipr.c ioa_cfg->hrrq[i].ioa_is_dead = 1; hrrq 9303 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 9313 drivers/scsi/ipr.c if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { hrrq 9343 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 9344 drivers/scsi/ipr.c ioa_cfg->hrrq[i].allow_interrupts = 0; hrrq 9345 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 9348 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); hrrq 9437 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 9438 drivers/scsi/ipr.c ioa_cfg->hrrq[i].allow_cmds = 0; hrrq 9439 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 9557 drivers/scsi/ipr.c sizeof(u32) * ioa_cfg->hrrq[i].size, hrrq 9558 drivers/scsi/ipr.c ioa_cfg->hrrq[i].host_rrq, hrrq 9559 drivers/scsi/ipr.c ioa_cfg->hrrq[i].host_rrq_dma); hrrq 9591 drivers/scsi/ipr.c free_irq(pci_irq_vector(pdev, i), &ioa_cfg->hrrq[i]); hrrq 9653 drivers/scsi/ipr.c ioa_cfg->hrrq[i].min_cmd_id = 0; hrrq 9654 drivers/scsi/ipr.c ioa_cfg->hrrq[i].max_cmd_id = hrrq 9660 drivers/scsi/ipr.c ioa_cfg->hrrq[i].min_cmd_id = hrrq 9663 drivers/scsi/ipr.c ioa_cfg->hrrq[i].max_cmd_id = hrrq 9669 drivers/scsi/ipr.c ioa_cfg->hrrq[i].min_cmd_id = 0; hrrq 9670 drivers/scsi/ipr.c ioa_cfg->hrrq[i].max_cmd_id = (entries_each_hrrq - 1); hrrq 9672 drivers/scsi/ipr.c ioa_cfg->hrrq[i].size = entries_each_hrrq; hrrq 9678 drivers/scsi/ipr.c ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id - 1; hrrq 9680 drivers/scsi/ipr.c ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].size += i; hrrq 9681 drivers/scsi/ipr.c ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id += i; hrrq 9723 drivers/scsi/ipr.c ipr_cmd->hrrq = &ioa_cfg->hrrq[hrrq_id]; hrrq 9724 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 9725 drivers/scsi/ipr.c if (i >= ioa_cfg->hrrq[hrrq_id].max_cmd_id) hrrq 9769 drivers/scsi/ipr.c ioa_cfg->hrrq[i].host_rrq = dma_alloc_coherent(&pdev->dev, hrrq 9770 drivers/scsi/ipr.c sizeof(u32) * ioa_cfg->hrrq[i].size, hrrq 9771 drivers/scsi/ipr.c &ioa_cfg->hrrq[i].host_rrq_dma, hrrq 9774 drivers/scsi/ipr.c if (!ioa_cfg->hrrq[i].host_rrq) { hrrq 9777 drivers/scsi/ipr.c sizeof(u32) * ioa_cfg->hrrq[i].size, hrrq 9778 drivers/scsi/ipr.c ioa_cfg->hrrq[i].host_rrq, hrrq 9779 drivers/scsi/ipr.c ioa_cfg->hrrq[i].host_rrq_dma); hrrq 9782 drivers/scsi/ipr.c ioa_cfg->hrrq[i].ioa_cfg = ioa_cfg; hrrq 9831 drivers/scsi/ipr.c sizeof(u32) * ioa_cfg->hrrq[i].size, hrrq 9832 drivers/scsi/ipr.c ioa_cfg->hrrq[i].host_rrq, hrrq 9833 drivers/scsi/ipr.c ioa_cfg->hrrq[i].host_rrq_dma); hrrq 9974 drivers/scsi/ipr.c for (i = 0; i < ARRAY_SIZE(ioa_cfg->hrrq); i++) { hrrq 9975 drivers/scsi/ipr.c INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_free_q); hrrq 9976 drivers/scsi/ipr.c INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_pending_q); hrrq 9977 drivers/scsi/ipr.c spin_lock_init(&ioa_cfg->hrrq[i]._lock); hrrq 9979 drivers/scsi/ipr.c ioa_cfg->hrrq[i].lock = ioa_cfg->host->host_lock; hrrq 9981 drivers/scsi/ipr.c ioa_cfg->hrrq[i].lock = &ioa_cfg->hrrq[i]._lock; hrrq 10046 drivers/scsi/ipr.c &ioa_cfg->hrrq[i]); hrrq 10050 drivers/scsi/ipr.c &ioa_cfg->hrrq[i]); hrrq 10365 drivers/scsi/ipr.c &ioa_cfg->hrrq[0]); hrrq 10371 drivers/scsi/ipr.c IPR_NAME, &ioa_cfg->hrrq[0]); hrrq 10472 drivers/scsi/ipr.c spin_lock(&ioa_cfg->hrrq[i]._lock); hrrq 10473 drivers/scsi/ipr.c ioa_cfg->hrrq[i].removing_ioa = 1; hrrq 10474 drivers/scsi/ipr.c spin_unlock(&ioa_cfg->hrrq[i]._lock); hrrq 10603 drivers/scsi/ipr.c irq_poll_init(&ioa_cfg->hrrq[i].iopoll, hrrq 10634 drivers/scsi/ipr.c irq_poll_disable(&ioa_cfg->hrrq[i].iopoll); hrrq 10792 drivers/scsi/ipr.c list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); hrrq 10814 drivers/scsi/ipr.c if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds || hrrq 1102 drivers/scsi/ipr.h #define for_each_hrrq(hrrq, ioa_cfg) \ hrrq 1103 drivers/scsi/ipr.h for (hrrq = (ioa_cfg)->hrrq; \ hrrq 1104 drivers/scsi/ipr.h hrrq < ((ioa_cfg)->hrrq + (ioa_cfg)->hrrq_num); hrrq++) hrrq 1538 drivers/scsi/ipr.h struct ipr_hrr_queue hrrq[IPR_MAX_HRRQ_NUM]; hrrq 1638 drivers/scsi/ipr.h struct ipr_hrr_queue *hrrq;