hrrq              696 drivers/scsi/ipr.c struct ipr_cmnd *__ipr_get_free_ipr_cmnd(struct ipr_hrr_queue *hrrq)
hrrq              700 drivers/scsi/ipr.c 	if (likely(!list_empty(&hrrq->hrrq_free_q))) {
hrrq              701 drivers/scsi/ipr.c 		ipr_cmd = list_entry(hrrq->hrrq_free_q.next,
hrrq              721 drivers/scsi/ipr.c 		__ipr_get_free_ipr_cmnd(&ioa_cfg->hrrq[IPR_INIT_HRRQ]);
hrrq              745 drivers/scsi/ipr.c 		spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq              746 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].allow_interrupts = 0;
hrrq              747 drivers/scsi/ipr.c 		spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq              829 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq              844 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq;
hrrq              847 drivers/scsi/ipr.c 	spin_lock_irqsave(&hrrq->_lock, hrrq_flags);
hrrq              849 drivers/scsi/ipr.c 	spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags);
hrrq              872 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq              888 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq;
hrrq              890 drivers/scsi/ipr.c 	spin_lock_irqsave(&hrrq->_lock, hrrq_flags);
hrrq              892 drivers/scsi/ipr.c 	spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags);
hrrq              907 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq              910 drivers/scsi/ipr.c 	for_each_hrrq(hrrq, ioa_cfg) {
hrrq              911 drivers/scsi/ipr.c 		spin_lock(&hrrq->_lock);
hrrq              913 drivers/scsi/ipr.c 					temp, &hrrq->hrrq_pending_q, queue) {
hrrq              931 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq              982 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
hrrq             1084 drivers/scsi/ipr.c 	unsigned int hrrq;
hrrq             1087 drivers/scsi/ipr.c 		hrrq = 0;
hrrq             1089 drivers/scsi/ipr.c 		hrrq = atomic_add_return(1, &ioa_cfg->hrrq_index);
hrrq             1090 drivers/scsi/ipr.c 		hrrq = (hrrq % (ioa_cfg->hrrq_num - 1)) + 1;
hrrq             1092 drivers/scsi/ipr.c 	return hrrq;
hrrq             1114 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) {
hrrq             1116 drivers/scsi/ipr.c 		list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
hrrq             1505 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             2649 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             3340 drivers/scsi/ipr.c 		if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) {
hrrq             3649 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead)
hrrq             3681 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead &&
hrrq             3684 drivers/scsi/ipr.c 			spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             3685 drivers/scsi/ipr.c 			ioa_cfg->hrrq[i].ioa_is_dead = 0;
hrrq             3686 drivers/scsi/ipr.c 			spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             3809 drivers/scsi/ipr.c 			irq_poll_disable(&ioa_cfg->hrrq[i].iopoll);
hrrq             3816 drivers/scsi/ipr.c 			irq_poll_init(&ioa_cfg->hrrq[i].iopoll,
hrrq             4390 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead && !ioa_cfg->dump_taken) {
hrrq             5072 drivers/scsi/ipr.c 	list_for_each_entry(loop_cmd, &ipr_cmd->hrrq->hrrq_free_q, queue) {
hrrq             5112 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             5120 drivers/scsi/ipr.c 		for_each_hrrq(hrrq, ioa_cfg) {
hrrq             5121 drivers/scsi/ipr.c 			spin_lock_irqsave(hrrq->lock, flags);
hrrq             5122 drivers/scsi/ipr.c 			for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) {
hrrq             5131 drivers/scsi/ipr.c 			spin_unlock_irqrestore(hrrq->lock, flags);
hrrq             5140 drivers/scsi/ipr.c 				for_each_hrrq(hrrq, ioa_cfg) {
hrrq             5141 drivers/scsi/ipr.c 					spin_lock_irqsave(hrrq->lock, flags);
hrrq             5142 drivers/scsi/ipr.c 					for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) {
hrrq             5151 drivers/scsi/ipr.c 					spin_unlock_irqrestore(hrrq->lock, flags);
hrrq             5176 drivers/scsi/ipr.c 	if (!ioa_cfg->in_reset_reload && !ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) {
hrrq             5191 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) {
hrrq             5246 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             5326 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             5339 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead)
hrrq             5342 drivers/scsi/ipr.c 	for_each_hrrq(hrrq, ioa_cfg) {
hrrq             5343 drivers/scsi/ipr.c 		spin_lock(&hrrq->_lock);
hrrq             5344 drivers/scsi/ipr.c 		for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) {
hrrq             5360 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             5437 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             5499 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             5510 drivers/scsi/ipr.c 	    ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead)
hrrq             5525 drivers/scsi/ipr.c 	for_each_hrrq(hrrq, ioa_cfg) {
hrrq             5526 drivers/scsi/ipr.c 		spin_lock(&hrrq->_lock);
hrrq             5527 drivers/scsi/ipr.c 		for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) {
hrrq             5535 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             5562 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             5584 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead || ioa_cfg->scan_done)
hrrq             5766 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             5772 drivers/scsi/ipr.c 	hrrq = container_of(iop, struct ipr_hrr_queue, iopoll);
hrrq             5773 drivers/scsi/ipr.c 	ioa_cfg = hrrq->ioa_cfg;
hrrq             5775 drivers/scsi/ipr.c 	spin_lock_irqsave(hrrq->lock, hrrq_flags);
hrrq             5776 drivers/scsi/ipr.c 	completed_ops = ipr_process_hrrq(hrrq, budget, &doneq);
hrrq             5780 drivers/scsi/ipr.c 	spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             5801 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp;
hrrq             5802 drivers/scsi/ipr.c 	struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg;
hrrq             5811 drivers/scsi/ipr.c 	spin_lock_irqsave(hrrq->lock, hrrq_flags);
hrrq             5813 drivers/scsi/ipr.c 	if (!hrrq->allow_interrupts) {
hrrq             5814 drivers/scsi/ipr.c 		spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             5819 drivers/scsi/ipr.c 		if (ipr_process_hrrq(hrrq, -1, &doneq)) {
hrrq             5850 drivers/scsi/ipr.c 	spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             5869 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp;
hrrq             5870 drivers/scsi/ipr.c 	struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg;
hrrq             5876 drivers/scsi/ipr.c 	spin_lock_irqsave(hrrq->lock, hrrq_flags);
hrrq             5879 drivers/scsi/ipr.c 	if (!hrrq->allow_interrupts) {
hrrq             5880 drivers/scsi/ipr.c 		spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             5885 drivers/scsi/ipr.c 		if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) ==
hrrq             5886 drivers/scsi/ipr.c 		       hrrq->toggle_bit) {
hrrq             5887 drivers/scsi/ipr.c 			irq_poll_sched(&hrrq->iopoll);
hrrq             5888 drivers/scsi/ipr.c 			spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             5892 drivers/scsi/ipr.c 		if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) ==
hrrq             5893 drivers/scsi/ipr.c 			hrrq->toggle_bit)
hrrq             5895 drivers/scsi/ipr.c 			if (ipr_process_hrrq(hrrq, -1, &doneq))
hrrq             5899 drivers/scsi/ipr.c 	spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6056 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             6071 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq;
hrrq             6074 drivers/scsi/ipr.c 	spin_lock_irqsave(&hrrq->_lock, hrrq_flags);
hrrq             6076 drivers/scsi/ipr.c 	spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags);
hrrq             6158 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq;
hrrq             6161 drivers/scsi/ipr.c 	spin_lock_irqsave(&hrrq->_lock, hrrq_flags);
hrrq             6163 drivers/scsi/ipr.c 	spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags);
hrrq             6494 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             6519 drivers/scsi/ipr.c 		spin_lock_irqsave(ipr_cmd->hrrq->lock, lock_flags);
hrrq             6523 drivers/scsi/ipr.c 		list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             6524 drivers/scsi/ipr.c 		spin_unlock_irqrestore(ipr_cmd->hrrq->lock, lock_flags);
hrrq             6527 drivers/scsi/ipr.c 		spin_lock(&ipr_cmd->hrrq->_lock);
hrrq             6529 drivers/scsi/ipr.c 		spin_unlock(&ipr_cmd->hrrq->_lock);
hrrq             6555 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             6571 drivers/scsi/ipr.c 	hrrq = &ioa_cfg->hrrq[hrrq_id];
hrrq             6573 drivers/scsi/ipr.c 	spin_lock_irqsave(hrrq->lock, hrrq_flags);
hrrq             6579 drivers/scsi/ipr.c 	if (unlikely(!hrrq->allow_cmds && !hrrq->ioa_is_dead && !hrrq->removing_ioa)) {
hrrq             6580 drivers/scsi/ipr.c 		spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6588 drivers/scsi/ipr.c 	if (unlikely(hrrq->ioa_is_dead || hrrq->removing_ioa || !res)) {
hrrq             6589 drivers/scsi/ipr.c 		spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6593 drivers/scsi/ipr.c 	ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq);
hrrq             6595 drivers/scsi/ipr.c 		spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6598 drivers/scsi/ipr.c 	spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6643 drivers/scsi/ipr.c 	spin_lock_irqsave(hrrq->lock, hrrq_flags);
hrrq             6644 drivers/scsi/ipr.c 	if (unlikely(rc || (!hrrq->allow_cmds && !hrrq->ioa_is_dead))) {
hrrq             6645 drivers/scsi/ipr.c 		list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q);
hrrq             6646 drivers/scsi/ipr.c 		spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6652 drivers/scsi/ipr.c 	if (unlikely(hrrq->ioa_is_dead)) {
hrrq             6653 drivers/scsi/ipr.c 		list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q);
hrrq             6654 drivers/scsi/ipr.c 		spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6664 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_pending_q);
hrrq             6667 drivers/scsi/ipr.c 	spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6671 drivers/scsi/ipr.c 	spin_lock_irqsave(hrrq->lock, hrrq_flags);
hrrq             6675 drivers/scsi/ipr.c 	spin_unlock_irqrestore(hrrq->lock, hrrq_flags);
hrrq             6773 drivers/scsi/ipr.c 	if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds)
hrrq             6804 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             6814 drivers/scsi/ipr.c 	for_each_hrrq(hrrq, ioa_cfg) {
hrrq             6815 drivers/scsi/ipr.c 		spin_lock(&hrrq->_lock);
hrrq             6816 drivers/scsi/ipr.c 		list_for_each_entry(ipr_cmd, &hrrq->hrrq_pending_q, queue) {
hrrq             6822 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             6871 drivers/scsi/ipr.c 	spin_lock(&ipr_cmd->hrrq->_lock);
hrrq             6887 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             6888 drivers/scsi/ipr.c 	spin_unlock(&ipr_cmd->hrrq->_lock);
hrrq             6996 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             7000 drivers/scsi/ipr.c 	hrrq = &ioa_cfg->hrrq[hrrq_id];
hrrq             7003 drivers/scsi/ipr.c 	spin_lock(&hrrq->_lock);
hrrq             7004 drivers/scsi/ipr.c 	if (unlikely(hrrq->ioa_is_dead)) {
hrrq             7005 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             7009 drivers/scsi/ipr.c 	if (unlikely(!hrrq->allow_cmds)) {
hrrq             7010 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             7014 drivers/scsi/ipr.c 	ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq);
hrrq             7016 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             7021 drivers/scsi/ipr.c 	spin_unlock(&hrrq->_lock);
hrrq             7050 drivers/scsi/ipr.c 	spin_lock(&ipr_cmd->hrrq->_lock);
hrrq             7051 drivers/scsi/ipr.c 	if (unlikely(!ipr_cmd->hrrq->allow_cmds ||
hrrq             7052 drivers/scsi/ipr.c 			ipr_cmd->hrrq->ioa_is_dead)) {
hrrq             7053 drivers/scsi/ipr.c 		list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             7054 drivers/scsi/ipr.c 		spin_unlock(&ipr_cmd->hrrq->_lock);
hrrq             7070 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
hrrq             7110 drivers/scsi/ipr.c 		spin_unlock(&ipr_cmd->hrrq->_lock);
hrrq             7115 drivers/scsi/ipr.c 	spin_unlock(&ipr_cmd->hrrq->_lock);
hrrq             7224 drivers/scsi/ipr.c 	if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) {
hrrq             7233 drivers/scsi/ipr.c 		spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             7234 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].ioa_is_dead = 1;
hrrq             7235 drivers/scsi/ipr.c 		spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             7239 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             7266 drivers/scsi/ipr.c 		spin_lock(&ioa_cfg->hrrq[j]._lock);
hrrq             7267 drivers/scsi/ipr.c 		ioa_cfg->hrrq[j].allow_cmds = 1;
hrrq             7268 drivers/scsi/ipr.c 		spin_unlock(&ioa_cfg->hrrq[j]._lock);
hrrq             7298 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             7633 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             8141 drivers/scsi/ipr.c 					&ioa_cfg->hrrq->hrrq_free_q);
hrrq             8194 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             8202 drivers/scsi/ipr.c 		hrrq = &ioa_cfg->hrrq[ioa_cfg->identify_hrrq_index];
hrrq             8217 drivers/scsi/ipr.c 			((u64) hrrq->host_rrq_dma >> 24) & 0xff;
hrrq             8219 drivers/scsi/ipr.c 			((u64) hrrq->host_rrq_dma >> 16) & 0xff;
hrrq             8221 drivers/scsi/ipr.c 			((u64) hrrq->host_rrq_dma >> 8) & 0xff;
hrrq             8223 drivers/scsi/ipr.c 			((u64) hrrq->host_rrq_dma) & 0xff;
hrrq             8225 drivers/scsi/ipr.c 			((sizeof(u32) * hrrq->size) >> 8) & 0xff;
hrrq             8227 drivers/scsi/ipr.c 			(sizeof(u32) * hrrq->size) & 0xff;
hrrq             8235 drivers/scsi/ipr.c 				((u64) hrrq->host_rrq_dma >> 56) & 0xff;
hrrq             8237 drivers/scsi/ipr.c 				((u64) hrrq->host_rrq_dma >> 48) & 0xff;
hrrq             8239 drivers/scsi/ipr.c 				((u64) hrrq->host_rrq_dma >> 40) & 0xff;
hrrq             8241 drivers/scsi/ipr.c 				((u64) hrrq->host_rrq_dma >> 32) & 0xff;
hrrq             8310 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
hrrq             8327 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             8329 drivers/scsi/ipr.c 	for_each_hrrq(hrrq, ioa_cfg) {
hrrq             8330 drivers/scsi/ipr.c 		spin_lock(&hrrq->_lock);
hrrq             8331 drivers/scsi/ipr.c 		memset(hrrq->host_rrq, 0, sizeof(u32) * hrrq->size);
hrrq             8334 drivers/scsi/ipr.c 		hrrq->hrrq_start = hrrq->host_rrq;
hrrq             8335 drivers/scsi/ipr.c 		hrrq->hrrq_end = &hrrq->host_rrq[hrrq->size - 1];
hrrq             8336 drivers/scsi/ipr.c 		hrrq->hrrq_curr = hrrq->hrrq_start;
hrrq             8337 drivers/scsi/ipr.c 		hrrq->toggle_bit = 1;
hrrq             8338 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             8403 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
hrrq             8430 drivers/scsi/ipr.c 		spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             8431 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].allow_interrupts = 1;
hrrq             8432 drivers/scsi/ipr.c 		spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             8472 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
hrrq             8994 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq;
hrrq             9001 drivers/scsi/ipr.c 	for_each_hrrq(hrrq, ioa_cfg) {
hrrq             9002 drivers/scsi/ipr.c 		spin_lock(&hrrq->_lock);
hrrq             9003 drivers/scsi/ipr.c 		list_for_each_entry(loop_cmd, &hrrq->hrrq_pending_q, queue) {
hrrq             9006 drivers/scsi/ipr.c 			list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             9010 drivers/scsi/ipr.c 		spin_unlock(&hrrq->_lock);
hrrq             9035 drivers/scsi/ipr.c 	struct ipr_hrr_queue *hrrq = &ioa_cfg->hrrq[IPR_INIT_HRRQ];
hrrq             9040 drivers/scsi/ipr.c 	if (!hrrq->ioa_is_dead) {
hrrq             9042 drivers/scsi/ipr.c 			list_for_each_entry(hcam_cmd, &hrrq->hrrq_pending_q, queue) {
hrrq             9161 drivers/scsi/ipr.c 			!ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) {
hrrq             9210 drivers/scsi/ipr.c 					&ipr_cmd->hrrq->hrrq_free_q);
hrrq             9249 drivers/scsi/ipr.c 		spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             9250 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].allow_cmds = 0;
hrrq             9251 drivers/scsi/ipr.c 		spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             9254 drivers/scsi/ipr.c 	if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) {
hrrq             9285 drivers/scsi/ipr.c 	if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead)
hrrq             9301 drivers/scsi/ipr.c 			spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             9302 drivers/scsi/ipr.c 			ioa_cfg->hrrq[i].ioa_is_dead = 1;
hrrq             9303 drivers/scsi/ipr.c 			spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             9313 drivers/scsi/ipr.c 			if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) {
hrrq             9343 drivers/scsi/ipr.c 		spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             9344 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].allow_interrupts = 0;
hrrq             9345 drivers/scsi/ipr.c 		spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             9348 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
hrrq             9437 drivers/scsi/ipr.c 			spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             9438 drivers/scsi/ipr.c 			ioa_cfg->hrrq[i].allow_cmds = 0;
hrrq             9439 drivers/scsi/ipr.c 			spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             9557 drivers/scsi/ipr.c 				  sizeof(u32) * ioa_cfg->hrrq[i].size,
hrrq             9558 drivers/scsi/ipr.c 				  ioa_cfg->hrrq[i].host_rrq,
hrrq             9559 drivers/scsi/ipr.c 				  ioa_cfg->hrrq[i].host_rrq_dma);
hrrq             9591 drivers/scsi/ipr.c 		free_irq(pci_irq_vector(pdev, i), &ioa_cfg->hrrq[i]);
hrrq             9653 drivers/scsi/ipr.c 				ioa_cfg->hrrq[i].min_cmd_id = 0;
hrrq             9654 drivers/scsi/ipr.c 				ioa_cfg->hrrq[i].max_cmd_id =
hrrq             9660 drivers/scsi/ipr.c 				ioa_cfg->hrrq[i].min_cmd_id =
hrrq             9663 drivers/scsi/ipr.c 				ioa_cfg->hrrq[i].max_cmd_id =
hrrq             9669 drivers/scsi/ipr.c 			ioa_cfg->hrrq[i].min_cmd_id = 0;
hrrq             9670 drivers/scsi/ipr.c 			ioa_cfg->hrrq[i].max_cmd_id = (entries_each_hrrq - 1);
hrrq             9672 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].size = entries_each_hrrq;
hrrq             9678 drivers/scsi/ipr.c 		ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id - 1;
hrrq             9680 drivers/scsi/ipr.c 		ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].size += i;
hrrq             9681 drivers/scsi/ipr.c 		ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id += i;
hrrq             9723 drivers/scsi/ipr.c 		ipr_cmd->hrrq = &ioa_cfg->hrrq[hrrq_id];
hrrq             9724 drivers/scsi/ipr.c 		list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             9725 drivers/scsi/ipr.c 		if (i >= ioa_cfg->hrrq[hrrq_id].max_cmd_id)
hrrq             9769 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].host_rrq = dma_alloc_coherent(&pdev->dev,
hrrq             9770 drivers/scsi/ipr.c 					sizeof(u32) * ioa_cfg->hrrq[i].size,
hrrq             9771 drivers/scsi/ipr.c 					&ioa_cfg->hrrq[i].host_rrq_dma,
hrrq             9774 drivers/scsi/ipr.c 		if (!ioa_cfg->hrrq[i].host_rrq)  {
hrrq             9777 drivers/scsi/ipr.c 					sizeof(u32) * ioa_cfg->hrrq[i].size,
hrrq             9778 drivers/scsi/ipr.c 					ioa_cfg->hrrq[i].host_rrq,
hrrq             9779 drivers/scsi/ipr.c 					ioa_cfg->hrrq[i].host_rrq_dma);
hrrq             9782 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].ioa_cfg = ioa_cfg;
hrrq             9831 drivers/scsi/ipr.c 				  sizeof(u32) * ioa_cfg->hrrq[i].size,
hrrq             9832 drivers/scsi/ipr.c 				  ioa_cfg->hrrq[i].host_rrq,
hrrq             9833 drivers/scsi/ipr.c 				  ioa_cfg->hrrq[i].host_rrq_dma);
hrrq             9974 drivers/scsi/ipr.c 	for (i = 0; i < ARRAY_SIZE(ioa_cfg->hrrq); i++) {
hrrq             9975 drivers/scsi/ipr.c 		INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_free_q);
hrrq             9976 drivers/scsi/ipr.c 		INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_pending_q);
hrrq             9977 drivers/scsi/ipr.c 		spin_lock_init(&ioa_cfg->hrrq[i]._lock);
hrrq             9979 drivers/scsi/ipr.c 			ioa_cfg->hrrq[i].lock = ioa_cfg->host->host_lock;
hrrq             9981 drivers/scsi/ipr.c 			ioa_cfg->hrrq[i].lock = &ioa_cfg->hrrq[i]._lock;
hrrq             10046 drivers/scsi/ipr.c 			&ioa_cfg->hrrq[i]);
hrrq             10050 drivers/scsi/ipr.c 					&ioa_cfg->hrrq[i]);
hrrq             10365 drivers/scsi/ipr.c 			&ioa_cfg->hrrq[0]);
hrrq             10371 drivers/scsi/ipr.c 			 IPR_NAME, &ioa_cfg->hrrq[0]);
hrrq             10472 drivers/scsi/ipr.c 		spin_lock(&ioa_cfg->hrrq[i]._lock);
hrrq             10473 drivers/scsi/ipr.c 		ioa_cfg->hrrq[i].removing_ioa = 1;
hrrq             10474 drivers/scsi/ipr.c 		spin_unlock(&ioa_cfg->hrrq[i]._lock);
hrrq             10603 drivers/scsi/ipr.c 			irq_poll_init(&ioa_cfg->hrrq[i].iopoll,
hrrq             10634 drivers/scsi/ipr.c 			irq_poll_disable(&ioa_cfg->hrrq[i].iopoll);
hrrq             10792 drivers/scsi/ipr.c 	list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
hrrq             10814 drivers/scsi/ipr.c 		if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds ||
hrrq             1102 drivers/scsi/ipr.h #define for_each_hrrq(hrrq, ioa_cfg) \
hrrq             1103 drivers/scsi/ipr.h 		for (hrrq = (ioa_cfg)->hrrq; \
hrrq             1104 drivers/scsi/ipr.h 			hrrq < ((ioa_cfg)->hrrq + (ioa_cfg)->hrrq_num); hrrq++)
hrrq             1538 drivers/scsi/ipr.h 	struct ipr_hrr_queue hrrq[IPR_MAX_HRRQ_NUM];
hrrq             1638 drivers/scsi/ipr.h 	struct ipr_hrr_queue *hrrq;