Lines Matching refs:evt

153 static void ibmvfc_trc_start(struct ibmvfc_event *evt)  in ibmvfc_trc_start()  argument
155 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_trc_start()
156 struct ibmvfc_cmd *vfc_cmd = &evt->iu.cmd; in ibmvfc_trc_start()
157 struct ibmvfc_mad_common *mad = &evt->iu.mad_common; in ibmvfc_trc_start()
161 entry->evt = evt; in ibmvfc_trc_start()
163 entry->fmt = evt->crq.format; in ibmvfc_trc_start()
187 static void ibmvfc_trc_end(struct ibmvfc_event *evt) in ibmvfc_trc_end() argument
189 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_trc_end()
190 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; in ibmvfc_trc_end()
191 struct ibmvfc_mad_common *mad = &evt->xfer_iu->mad_common; in ibmvfc_trc_end()
194 entry->evt = evt; in ibmvfc_trc_end()
196 entry->fmt = evt->crq.format; in ibmvfc_trc_end()
222 #define ibmvfc_trc_start(evt) do { } while (0) argument
223 #define ibmvfc_trc_end(evt) do { } while (0) argument
749 struct ibmvfc_event *evt) in ibmvfc_valid_event() argument
751 int index = evt - pool->events; in ibmvfc_valid_event()
754 if (evt != pool->events + index) /* unaligned */ in ibmvfc_valid_event()
764 static void ibmvfc_free_event(struct ibmvfc_event *evt) in ibmvfc_free_event() argument
766 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_free_event()
769 BUG_ON(!ibmvfc_valid_event(pool, evt)); in ibmvfc_free_event()
770 BUG_ON(atomic_inc_return(&evt->free) != 1); in ibmvfc_free_event()
771 list_add_tail(&evt->queue, &vhost->free); in ibmvfc_free_event()
781 static void ibmvfc_scsi_eh_done(struct ibmvfc_event *evt) in ibmvfc_scsi_eh_done() argument
783 struct scsi_cmnd *cmnd = evt->cmnd; in ibmvfc_scsi_eh_done()
790 if (evt->eh_comp) in ibmvfc_scsi_eh_done()
791 complete(evt->eh_comp); in ibmvfc_scsi_eh_done()
793 ibmvfc_free_event(evt); in ibmvfc_scsi_eh_done()
804 static void ibmvfc_fail_request(struct ibmvfc_event *evt, int error_code) in ibmvfc_fail_request() argument
806 if (evt->cmnd) { in ibmvfc_fail_request()
807 evt->cmnd->result = (error_code << 16); in ibmvfc_fail_request()
808 evt->done = ibmvfc_scsi_eh_done; in ibmvfc_fail_request()
810 evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_DRIVER_FAILED); in ibmvfc_fail_request()
812 list_del(&evt->queue); in ibmvfc_fail_request()
813 del_timer(&evt->timer); in ibmvfc_fail_request()
814 ibmvfc_trc_end(evt); in ibmvfc_fail_request()
815 evt->done(evt); in ibmvfc_fail_request()
828 struct ibmvfc_event *evt, *pos; in ibmvfc_purge_requests() local
831 list_for_each_entry_safe(evt, pos, &vhost->sent, queue) in ibmvfc_purge_requests()
832 ibmvfc_fail_request(evt, error_code); in ibmvfc_purge_requests()
1225 struct ibmvfc_event *evt = &pool->events[i]; in ibmvfc_init_event_pool() local
1226 atomic_set(&evt->free, 1); in ibmvfc_init_event_pool()
1227 evt->crq.valid = 0x80; in ibmvfc_init_event_pool()
1228 evt->crq.ioba = cpu_to_be64(pool->iu_token + (sizeof(*evt->xfer_iu) * i)); in ibmvfc_init_event_pool()
1229 evt->xfer_iu = pool->iu_storage + i; in ibmvfc_init_event_pool()
1230 evt->vhost = vhost; in ibmvfc_init_event_pool()
1231 evt->ext_list = NULL; in ibmvfc_init_event_pool()
1232 list_add_tail(&evt->queue, &vhost->free); in ibmvfc_init_event_pool()
1274 struct ibmvfc_event *evt; in ibmvfc_get_event() local
1277 evt = list_entry(vhost->free.next, struct ibmvfc_event, queue); in ibmvfc_get_event()
1278 atomic_set(&evt->free, 0); in ibmvfc_get_event()
1279 list_del(&evt->queue); in ibmvfc_get_event()
1280 return evt; in ibmvfc_get_event()
1290 static void ibmvfc_init_event(struct ibmvfc_event *evt, in ibmvfc_init_event() argument
1293 evt->cmnd = NULL; in ibmvfc_init_event()
1294 evt->sync_iu = NULL; in ibmvfc_init_event()
1295 evt->crq.format = format; in ibmvfc_init_event()
1296 evt->done = done; in ibmvfc_init_event()
1297 evt->eh_comp = NULL; in ibmvfc_init_event()
1330 struct ibmvfc_event *evt, in ibmvfc_map_sg_data() argument
1363 if (!evt->ext_list) { in ibmvfc_map_sg_data()
1364 evt->ext_list = dma_pool_alloc(vhost->sg_pool, GFP_ATOMIC, in ibmvfc_map_sg_data()
1365 &evt->ext_list_token); in ibmvfc_map_sg_data()
1367 if (!evt->ext_list) { in ibmvfc_map_sg_data()
1375 ibmvfc_map_sg_list(scmd, sg_mapped, evt->ext_list); in ibmvfc_map_sg_data()
1377 data->va = cpu_to_be64(evt->ext_list_token); in ibmvfc_map_sg_data()
1389 static void ibmvfc_timeout(struct ibmvfc_event *evt) in ibmvfc_timeout() argument
1391 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_timeout()
1392 dev_err(vhost->dev, "Command timed out (%p). Resetting connection\n", evt); in ibmvfc_timeout()
1404 static int ibmvfc_send_event(struct ibmvfc_event *evt, in ibmvfc_send_event() argument
1407 __be64 *crq_as_u64 = (__be64 *) &evt->crq; in ibmvfc_send_event()
1411 *evt->xfer_iu = evt->iu; in ibmvfc_send_event()
1412 if (evt->crq.format == IBMVFC_CMD_FORMAT) in ibmvfc_send_event()
1413 evt->xfer_iu->cmd.tag = cpu_to_be64((u64)evt); in ibmvfc_send_event()
1414 else if (evt->crq.format == IBMVFC_MAD_FORMAT) in ibmvfc_send_event()
1415 evt->xfer_iu->mad_common.tag = cpu_to_be64((u64)evt); in ibmvfc_send_event()
1419 list_add_tail(&evt->queue, &vhost->sent); in ibmvfc_send_event()
1420 init_timer(&evt->timer); in ibmvfc_send_event()
1423 evt->timer.data = (unsigned long) evt; in ibmvfc_send_event()
1424 evt->timer.expires = jiffies + (timeout * HZ); in ibmvfc_send_event()
1425 evt->timer.function = (void (*)(unsigned long))ibmvfc_timeout; in ibmvfc_send_event()
1426 add_timer(&evt->timer); in ibmvfc_send_event()
1433 list_del(&evt->queue); in ibmvfc_send_event()
1434 del_timer(&evt->timer); in ibmvfc_send_event()
1444 if (evt->cmnd) in ibmvfc_send_event()
1445 scsi_dma_unmap(evt->cmnd); in ibmvfc_send_event()
1446 ibmvfc_free_event(evt); in ibmvfc_send_event()
1451 if (evt->cmnd) { in ibmvfc_send_event()
1452 evt->cmnd->result = DID_ERROR << 16; in ibmvfc_send_event()
1453 evt->done = ibmvfc_scsi_eh_done; in ibmvfc_send_event()
1455 evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_CRQ_ERROR); in ibmvfc_send_event()
1457 evt->done(evt); in ibmvfc_send_event()
1459 ibmvfc_trc_start(evt); in ibmvfc_send_event()
1469 static void ibmvfc_log_error(struct ibmvfc_event *evt) in ibmvfc_log_error() argument
1471 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; in ibmvfc_log_error()
1472 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_log_error()
1474 struct scsi_cmnd *cmnd = evt->cmnd; in ibmvfc_log_error()
1524 static void ibmvfc_scsi_done(struct ibmvfc_event *evt) in ibmvfc_scsi_done() argument
1526 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; in ibmvfc_scsi_done()
1528 struct scsi_cmnd *cmnd = evt->cmnd; in ibmvfc_scsi_done()
1556 ibmvfc_log_error(evt); in ibmvfc_scsi_done()
1567 if (evt->eh_comp) in ibmvfc_scsi_done()
1568 complete(evt->eh_comp); in ibmvfc_scsi_done()
1570 ibmvfc_free_event(evt); in ibmvfc_scsi_done()
1617 struct ibmvfc_event *evt; in ibmvfc_queuecommand_lck() local
1628 evt = ibmvfc_get_event(vhost); in ibmvfc_queuecommand_lck()
1629 ibmvfc_init_event(evt, ibmvfc_scsi_done, IBMVFC_CMD_FORMAT); in ibmvfc_queuecommand_lck()
1630 evt->cmnd = cmnd; in ibmvfc_queuecommand_lck()
1632 vfc_cmd = &evt->iu.cmd; in ibmvfc_queuecommand_lck()
1634 vfc_cmd->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); in ibmvfc_queuecommand_lck()
1650 if (likely(!(rc = ibmvfc_map_sg_data(cmnd, evt, vfc_cmd, vhost->dev)))) in ibmvfc_queuecommand_lck()
1651 return ibmvfc_send_event(evt, vhost, 0); in ibmvfc_queuecommand_lck()
1653 ibmvfc_free_event(evt); in ibmvfc_queuecommand_lck()
1673 static void ibmvfc_sync_completion(struct ibmvfc_event *evt) in DEF_SCSI_QCMD()
1676 if (evt->sync_iu) in DEF_SCSI_QCMD()
1677 *evt->sync_iu = *evt->xfer_iu; in DEF_SCSI_QCMD()
1679 complete(&evt->comp); in DEF_SCSI_QCMD()
1687 static void ibmvfc_bsg_timeout_done(struct ibmvfc_event *evt) in ibmvfc_bsg_timeout_done() argument
1689 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_bsg_timeout_done()
1691 ibmvfc_free_event(evt); in ibmvfc_bsg_timeout_done()
1707 struct ibmvfc_event *evt; in ibmvfc_bsg_timeout() local
1721 evt = ibmvfc_get_event(vhost); in ibmvfc_bsg_timeout()
1722 ibmvfc_init_event(evt, ibmvfc_bsg_timeout_done, IBMVFC_MAD_FORMAT); in ibmvfc_bsg_timeout()
1724 tmf = &evt->iu.tmf; in ibmvfc_bsg_timeout()
1732 rc = ibmvfc_send_event(evt, vhost, default_timeout); in ibmvfc_bsg_timeout()
1760 struct ibmvfc_event *evt; in ibmvfc_bsg_plogi() local
1779 evt = ibmvfc_get_event(vhost); in ibmvfc_bsg_plogi()
1780 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); in ibmvfc_bsg_plogi()
1781 plogi = &evt->iu.plogi; in ibmvfc_bsg_plogi()
1787 evt->sync_iu = &rsp_iu; in ibmvfc_bsg_plogi()
1788 init_completion(&evt->comp); in ibmvfc_bsg_plogi()
1790 rc = ibmvfc_send_event(evt, vhost, default_timeout); in ibmvfc_bsg_plogi()
1796 wait_for_completion(&evt->comp); in ibmvfc_bsg_plogi()
1802 ibmvfc_free_event(evt); in ibmvfc_bsg_plogi()
1821 struct ibmvfc_event *evt; in ibmvfc_bsg_request() local
1893 evt = ibmvfc_get_event(vhost); in ibmvfc_bsg_request()
1894 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); in ibmvfc_bsg_request()
1895 mad = &evt->iu.passthru; in ibmvfc_bsg_request()
1902 mad->cmd_ioba.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + in ibmvfc_bsg_request()
1916 mad->iu.tag = cpu_to_be64((u64)evt); in ibmvfc_bsg_request()
1919 evt->sync_iu = &rsp_iu; in ibmvfc_bsg_request()
1920 init_completion(&evt->comp); in ibmvfc_bsg_request()
1921 rc = ibmvfc_send_event(evt, vhost, 0); in ibmvfc_bsg_request()
1929 wait_for_completion(&evt->comp); in ibmvfc_bsg_request()
1937 ibmvfc_free_event(evt); in ibmvfc_bsg_request()
1966 struct ibmvfc_event *evt = NULL; in ibmvfc_reset_device() local
1975 evt = ibmvfc_get_event(vhost); in ibmvfc_reset_device()
1976 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); in ibmvfc_reset_device()
1978 tmf = &evt->iu.cmd; in ibmvfc_reset_device()
1980 tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); in ibmvfc_reset_device()
1990 evt->sync_iu = &rsp_iu; in ibmvfc_reset_device()
1992 init_completion(&evt->comp); in ibmvfc_reset_device()
1993 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); in ibmvfc_reset_device()
2004 wait_for_completion(&evt->comp); in ibmvfc_reset_device()
2023 ibmvfc_free_event(evt); in ibmvfc_reset_device()
2036 static int ibmvfc_match_rport(struct ibmvfc_event *evt, void *rport) in ibmvfc_match_rport() argument
2040 if (evt->cmnd) { in ibmvfc_match_rport()
2041 cmd_rport = starget_to_rport(scsi_target(evt->cmnd->device)); in ibmvfc_match_rport()
2056 static int ibmvfc_match_target(struct ibmvfc_event *evt, void *device) in ibmvfc_match_target() argument
2058 if (evt->cmnd && scsi_target(evt->cmnd->device) == device) in ibmvfc_match_target()
2071 static int ibmvfc_match_lun(struct ibmvfc_event *evt, void *device) in ibmvfc_match_lun() argument
2073 if (evt->cmnd && evt->cmnd->device == device) in ibmvfc_match_lun()
2090 struct ibmvfc_event *evt; in ibmvfc_wait_for_ops() local
2100 list_for_each_entry(evt, &vhost->sent, queue) { in ibmvfc_wait_for_ops()
2101 if (match(evt, device)) { in ibmvfc_wait_for_ops()
2102 evt->eh_comp = &comp; in ibmvfc_wait_for_ops()
2114 list_for_each_entry(evt, &vhost->sent, queue) { in ibmvfc_wait_for_ops()
2115 if (match(evt, device)) { in ibmvfc_wait_for_ops()
2116 evt->eh_comp = NULL; in ibmvfc_wait_for_ops()
2150 struct ibmvfc_event *evt, *found_evt; in ibmvfc_cancel_all() local
2159 list_for_each_entry(evt, &vhost->sent, queue) { in ibmvfc_cancel_all()
2160 if (evt->cmnd && evt->cmnd->device == sdev) { in ibmvfc_cancel_all()
2161 found_evt = evt; in ibmvfc_cancel_all()
2174 evt = ibmvfc_get_event(vhost); in ibmvfc_cancel_all()
2175 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); in ibmvfc_cancel_all()
2177 tmf = &evt->iu.tmf; in ibmvfc_cancel_all()
2193 evt->sync_iu = &rsp; in ibmvfc_cancel_all()
2194 init_completion(&evt->comp); in ibmvfc_cancel_all()
2195 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); in ibmvfc_cancel_all()
2210 wait_for_completion(&evt->comp); in ibmvfc_cancel_all()
2213 ibmvfc_free_event(evt); in ibmvfc_cancel_all()
2241 static int ibmvfc_match_key(struct ibmvfc_event *evt, void *key) in ibmvfc_match_key() argument
2245 if (evt->crq.format == IBMVFC_CMD_FORMAT && in ibmvfc_match_key()
2246 be32_to_cpu(evt->iu.cmd.cancel_key) == cancel_key) in ibmvfc_match_key()
2259 static int ibmvfc_match_evt(struct ibmvfc_event *evt, void *match) in ibmvfc_match_evt() argument
2261 if (evt == match) in ibmvfc_match_evt()
2281 struct ibmvfc_event *evt, *found_evt; in ibmvfc_abort_task_set() local
2290 list_for_each_entry(evt, &vhost->sent, queue) { in ibmvfc_abort_task_set()
2291 if (evt->cmnd && evt->cmnd->device == sdev) { in ibmvfc_abort_task_set()
2292 found_evt = evt; in ibmvfc_abort_task_set()
2305 evt = ibmvfc_get_event(vhost); in ibmvfc_abort_task_set()
2306 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); in ibmvfc_abort_task_set()
2308 tmf = &evt->iu.cmd; in ibmvfc_abort_task_set()
2310 tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); in ibmvfc_abort_task_set()
2320 evt->sync_iu = &rsp_iu; in ibmvfc_abort_task_set()
2322 init_completion(&evt->comp); in ibmvfc_abort_task_set()
2323 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); in ibmvfc_abort_task_set()
2334 timeout = wait_for_completion_timeout(&evt->comp, timeout); in ibmvfc_abort_task_set()
2353 rc = ibmvfc_wait_for_ops(vhost, evt, ibmvfc_match_evt); in ibmvfc_abort_task_set()
2383 ibmvfc_free_event(evt); in ibmvfc_abort_task_set()
2724 struct ibmvfc_event *evt = (struct ibmvfc_event *)be64_to_cpu(crq->ioba); in ibmvfc_handle_crq() local
2778 if (unlikely(!ibmvfc_valid_event(&vhost->pool, evt))) { in ibmvfc_handle_crq()
2784 if (unlikely(atomic_read(&evt->free))) { in ibmvfc_handle_crq()
2790 del_timer(&evt->timer); in ibmvfc_handle_crq()
2791 list_del(&evt->queue); in ibmvfc_handle_crq()
2792 ibmvfc_trc_end(evt); in ibmvfc_handle_crq()
2793 evt->done(evt); in ibmvfc_handle_crq()
3290 static void ibmvfc_tgt_prli_done(struct ibmvfc_event *evt) in ibmvfc_tgt_prli_done() argument
3292 struct ibmvfc_target *tgt = evt->tgt; in ibmvfc_tgt_prli_done()
3293 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_tgt_prli_done()
3294 struct ibmvfc_process_login *rsp = &evt->xfer_iu->prli; in ibmvfc_tgt_prli_done()
3350 ibmvfc_free_event(evt); in ibmvfc_tgt_prli_done()
3363 struct ibmvfc_event *evt; in ibmvfc_tgt_send_prli() local
3369 evt = ibmvfc_get_event(vhost); in ibmvfc_tgt_send_prli()
3371 ibmvfc_init_event(evt, ibmvfc_tgt_prli_done, IBMVFC_MAD_FORMAT); in ibmvfc_tgt_send_prli()
3372 evt->tgt = tgt; in ibmvfc_tgt_send_prli()
3373 prli = &evt->iu.prli; in ibmvfc_tgt_send_prli()
3385 if (ibmvfc_send_event(evt, vhost, default_timeout)) { in ibmvfc_tgt_send_prli()
3398 static void ibmvfc_tgt_plogi_done(struct ibmvfc_event *evt) in ibmvfc_tgt_plogi_done() argument
3400 struct ibmvfc_target *tgt = evt->tgt; in ibmvfc_tgt_plogi_done()
3401 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_tgt_plogi_done()
3402 struct ibmvfc_port_login *rsp = &evt->xfer_iu->plogi; in ibmvfc_tgt_plogi_done()
3446 ibmvfc_free_event(evt); in ibmvfc_tgt_plogi_done()
3459 struct ibmvfc_event *evt; in ibmvfc_tgt_send_plogi() local
3466 evt = ibmvfc_get_event(vhost); in ibmvfc_tgt_send_plogi()
3469 ibmvfc_init_event(evt, ibmvfc_tgt_plogi_done, IBMVFC_MAD_FORMAT); in ibmvfc_tgt_send_plogi()
3470 evt->tgt = tgt; in ibmvfc_tgt_send_plogi()
3471 plogi = &evt->iu.plogi; in ibmvfc_tgt_send_plogi()
3478 if (ibmvfc_send_event(evt, vhost, default_timeout)) { in ibmvfc_tgt_send_plogi()
3491 static void ibmvfc_tgt_implicit_logout_done(struct ibmvfc_event *evt) in ibmvfc_tgt_implicit_logout_done() argument
3493 struct ibmvfc_target *tgt = evt->tgt; in ibmvfc_tgt_implicit_logout_done()
3494 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_tgt_implicit_logout_done()
3495 struct ibmvfc_implicit_logout *rsp = &evt->xfer_iu->implicit_logout; in ibmvfc_tgt_implicit_logout_done()
3499 ibmvfc_free_event(evt); in ibmvfc_tgt_implicit_logout_done()
3534 struct ibmvfc_event *evt; in ibmvfc_tgt_implicit_logout() local
3540 evt = ibmvfc_get_event(vhost); in ibmvfc_tgt_implicit_logout()
3542 ibmvfc_init_event(evt, ibmvfc_tgt_implicit_logout_done, IBMVFC_MAD_FORMAT); in ibmvfc_tgt_implicit_logout()
3543 evt->tgt = tgt; in ibmvfc_tgt_implicit_logout()
3544 mad = &evt->iu.implicit_logout; in ibmvfc_tgt_implicit_logout()
3552 if (ibmvfc_send_event(evt, vhost, default_timeout)) { in ibmvfc_tgt_implicit_logout()
3587 static void ibmvfc_tgt_adisc_done(struct ibmvfc_event *evt) in ibmvfc_tgt_adisc_done() argument
3589 struct ibmvfc_target *tgt = evt->tgt; in ibmvfc_tgt_adisc_done()
3590 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_tgt_adisc_done()
3591 struct ibmvfc_passthru_mad *mad = &evt->xfer_iu->passthru; in ibmvfc_tgt_adisc_done()
3621 ibmvfc_free_event(evt); in ibmvfc_tgt_adisc_done()
3630 static void ibmvfc_init_passthru(struct ibmvfc_event *evt) in ibmvfc_init_passthru() argument
3632 struct ibmvfc_passthru_mad *mad = &evt->iu.passthru; in ibmvfc_init_passthru()
3638 mad->cmd_ioba.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + in ibmvfc_init_passthru()
3643 mad->iu.cmd.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + in ibmvfc_init_passthru()
3647 mad->iu.rsp.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + in ibmvfc_init_passthru()
3663 static void ibmvfc_tgt_adisc_cancel_done(struct ibmvfc_event *evt) in ibmvfc_tgt_adisc_cancel_done() argument
3665 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_tgt_adisc_cancel_done()
3666 struct ibmvfc_target *tgt = evt->tgt; in ibmvfc_tgt_adisc_cancel_done()
3670 ibmvfc_free_event(evt); in ibmvfc_tgt_adisc_cancel_done()
3686 struct ibmvfc_event *evt; in ibmvfc_adisc_timeout() local
3703 evt = ibmvfc_get_event(vhost); in ibmvfc_adisc_timeout()
3704 ibmvfc_init_event(evt, ibmvfc_tgt_adisc_cancel_done, IBMVFC_MAD_FORMAT); in ibmvfc_adisc_timeout()
3706 evt->tgt = tgt; in ibmvfc_adisc_timeout()
3707 tmf = &evt->iu.tmf; in ibmvfc_adisc_timeout()
3715 rc = ibmvfc_send_event(evt, vhost, default_timeout); in ibmvfc_adisc_timeout()
3742 struct ibmvfc_event *evt; in ibmvfc_tgt_adisc() local
3748 evt = ibmvfc_get_event(vhost); in ibmvfc_tgt_adisc()
3750 ibmvfc_init_event(evt, ibmvfc_tgt_adisc_done, IBMVFC_MAD_FORMAT); in ibmvfc_tgt_adisc()
3751 evt->tgt = tgt; in ibmvfc_tgt_adisc()
3753 ibmvfc_init_passthru(evt); in ibmvfc_tgt_adisc()
3754 mad = &evt->iu.passthru; in ibmvfc_tgt_adisc()
3776 if (ibmvfc_send_event(evt, vhost, IBMVFC_ADISC_PLUS_CANCEL_TIMEOUT)) { in ibmvfc_tgt_adisc()
3790 static void ibmvfc_tgt_query_target_done(struct ibmvfc_event *evt) in ibmvfc_tgt_query_target_done() argument
3792 struct ibmvfc_target *tgt = evt->tgt; in ibmvfc_tgt_query_target_done()
3793 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_tgt_query_target_done()
3794 struct ibmvfc_query_tgt *rsp = &evt->xfer_iu->query_tgt; in ibmvfc_tgt_query_target_done()
3834 ibmvfc_free_event(evt); in ibmvfc_tgt_query_target_done()
3847 struct ibmvfc_event *evt; in ibmvfc_tgt_query_target() local
3853 evt = ibmvfc_get_event(vhost); in ibmvfc_tgt_query_target()
3855 evt->tgt = tgt; in ibmvfc_tgt_query_target()
3856 ibmvfc_init_event(evt, ibmvfc_tgt_query_target_done, IBMVFC_MAD_FORMAT); in ibmvfc_tgt_query_target()
3857 query_tgt = &evt->iu.query_tgt; in ibmvfc_tgt_query_target()
3865 if (ibmvfc_send_event(evt, vhost, default_timeout)) { in ibmvfc_tgt_query_target()
3944 static void ibmvfc_discover_targets_done(struct ibmvfc_event *evt) in ibmvfc_discover_targets_done() argument
3946 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_discover_targets_done()
3947 struct ibmvfc_discover_targets *rsp = &evt->xfer_iu->discover_targets; in ibmvfc_discover_targets_done()
3971 ibmvfc_free_event(evt); in ibmvfc_discover_targets_done()
3983 struct ibmvfc_event *evt = ibmvfc_get_event(vhost); in ibmvfc_discover_targets() local
3985 ibmvfc_init_event(evt, ibmvfc_discover_targets_done, IBMVFC_MAD_FORMAT); in ibmvfc_discover_targets()
3986 mad = &evt->iu.discover_targets; in ibmvfc_discover_targets()
3996 if (!ibmvfc_send_event(evt, vhost, default_timeout)) in ibmvfc_discover_targets()
4007 static void ibmvfc_npiv_login_done(struct ibmvfc_event *evt) in ibmvfc_npiv_login_done() argument
4009 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_npiv_login_done()
4010 u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_login.common.status); in ibmvfc_npiv_login_done()
4017 ibmvfc_free_event(evt); in ibmvfc_npiv_login_done()
4027 ibmvfc_free_event(evt); in ibmvfc_npiv_login_done()
4032 ibmvfc_free_event(evt); in ibmvfc_npiv_login_done()
4037 ibmvfc_free_event(evt); in ibmvfc_npiv_login_done()
4094 struct ibmvfc_event *evt = ibmvfc_get_event(vhost); in ibmvfc_npiv_login() local
4098 ibmvfc_init_event(evt, ibmvfc_npiv_login_done, IBMVFC_MAD_FORMAT); in ibmvfc_npiv_login()
4101 mad = &evt->iu.npiv_login; in ibmvfc_npiv_login()
4111 if (!ibmvfc_send_event(evt, vhost, default_timeout)) in ibmvfc_npiv_login()
4122 static void ibmvfc_npiv_logout_done(struct ibmvfc_event *evt) in ibmvfc_npiv_logout_done() argument
4124 struct ibmvfc_host *vhost = evt->vhost; in ibmvfc_npiv_logout_done()
4125 u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_logout.common.status); in ibmvfc_npiv_logout_done()
4127 ibmvfc_free_event(evt); in ibmvfc_npiv_logout_done()
4157 struct ibmvfc_event *evt; in ibmvfc_npiv_logout() local
4159 evt = ibmvfc_get_event(vhost); in ibmvfc_npiv_logout()
4160 ibmvfc_init_event(evt, ibmvfc_npiv_logout_done, IBMVFC_MAD_FORMAT); in ibmvfc_npiv_logout()
4162 mad = &evt->iu.npiv_logout; in ibmvfc_npiv_logout()
4170 if (!ibmvfc_send_event(evt, vhost, default_timeout)) in ibmvfc_npiv_logout()