tgtport            36 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport		*tgtport;
tgtport            79 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport		*tgtport;
tgtport           105 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport		*tgtport;
tgtport           142 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport		*tgtport;
tgtport           153 drivers/nvme/target/fc.c 	return (iodptr - iodptr->tgtport->iod);
tgtport           225 drivers/nvme/target/fc.c static void nvmet_fc_tgtport_put(struct nvmet_fc_tgtport *tgtport);
tgtport           226 drivers/nvme/target/fc.c static int nvmet_fc_tgtport_get(struct nvmet_fc_tgtport *tgtport);
tgtport           227 drivers/nvme/target/fc.c static void nvmet_fc_handle_fcp_rqst(struct nvmet_fc_tgtport *tgtport,
tgtport           325 drivers/nvme/target/fc.c nvmet_fc_alloc_ls_iodlist(struct nvmet_fc_tgtport *tgtport)
tgtport           335 drivers/nvme/target/fc.c 	tgtport->iod = iod;
tgtport           339 drivers/nvme/target/fc.c 		iod->tgtport = tgtport;
tgtport           340 drivers/nvme/target/fc.c 		list_add_tail(&iod->ls_list, &tgtport->ls_list);
tgtport           349 drivers/nvme/target/fc.c 		iod->rspdma = fc_dma_map_single(tgtport->dev, iod->rspbuf,
tgtport           352 drivers/nvme/target/fc.c 		if (fc_dma_mapping_error(tgtport->dev, iod->rspdma))
tgtport           362 drivers/nvme/target/fc.c 		fc_dma_unmap_single(tgtport->dev, iod->rspdma,
tgtport           374 drivers/nvme/target/fc.c nvmet_fc_free_ls_iodlist(struct nvmet_fc_tgtport *tgtport)
tgtport           376 drivers/nvme/target/fc.c 	struct nvmet_fc_ls_iod *iod = tgtport->iod;
tgtport           380 drivers/nvme/target/fc.c 		fc_dma_unmap_single(tgtport->dev,
tgtport           386 drivers/nvme/target/fc.c 	kfree(tgtport->iod);
tgtport           390 drivers/nvme/target/fc.c nvmet_fc_alloc_ls_iod(struct nvmet_fc_tgtport *tgtport)
tgtport           395 drivers/nvme/target/fc.c 	spin_lock_irqsave(&tgtport->lock, flags);
tgtport           396 drivers/nvme/target/fc.c 	iod = list_first_entry_or_null(&tgtport->ls_list,
tgtport           399 drivers/nvme/target/fc.c 		list_move_tail(&iod->ls_list, &tgtport->ls_busylist);
tgtport           400 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           406 drivers/nvme/target/fc.c nvmet_fc_free_ls_iod(struct nvmet_fc_tgtport *tgtport,
tgtport           411 drivers/nvme/target/fc.c 	spin_lock_irqsave(&tgtport->lock, flags);
tgtport           412 drivers/nvme/target/fc.c 	list_move(&iod->ls_list, &tgtport->ls_list);
tgtport           413 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           417 drivers/nvme/target/fc.c nvmet_fc_prep_fcp_iodlist(struct nvmet_fc_tgtport *tgtport,
tgtport           425 drivers/nvme/target/fc.c 		fod->tgtport = tgtport;
tgtport           434 drivers/nvme/target/fc.c 		fod->rspdma = fc_dma_map_single(tgtport->dev, &fod->rspiubuf,
tgtport           436 drivers/nvme/target/fc.c 		if (fc_dma_mapping_error(tgtport->dev, fod->rspdma)) {
tgtport           439 drivers/nvme/target/fc.c 				fc_dma_unmap_single(tgtport->dev, fod->rspdma,
tgtport           452 drivers/nvme/target/fc.c nvmet_fc_destroy_fcp_iodlist(struct nvmet_fc_tgtport *tgtport,
tgtport           460 drivers/nvme/target/fc.c 			fc_dma_unmap_single(tgtport->dev, fod->rspdma,
tgtport           488 drivers/nvme/target/fc.c nvmet_fc_queue_fcp_req(struct nvmet_fc_tgtport *tgtport,
tgtport           499 drivers/nvme/target/fc.c 			((queue->qid - 1) % tgtport->ops->max_hw_queues) : 0;
tgtport           501 drivers/nvme/target/fc.c 	nvmet_fc_handle_fcp_rqst(tgtport, fod);
tgtport           511 drivers/nvme/target/fc.c 	nvmet_fc_queue_fcp_req(fod->tgtport, fod->queue, fod->fcpreq);
tgtport           520 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = fod->tgtport;
tgtport           524 drivers/nvme/target/fc.c 	fc_dma_sync_single_for_cpu(tgtport->dev, fod->rspdma,
tgtport           535 drivers/nvme/target/fc.c 	tgtport->ops->fcp_req_release(&tgtport->fc_target_port, fcpreq);
tgtport           570 drivers/nvme/target/fc.c 	tgtport->ops->defer_rcv(&tgtport->fc_target_port, fcpreq);
tgtport           599 drivers/nvme/target/fc.c 				assoc->tgtport->fc_target_port.port_num,
tgtport           617 drivers/nvme/target/fc.c 	nvmet_fc_prep_fcp_iodlist(assoc->tgtport, queue);
tgtport           624 drivers/nvme/target/fc.c 	spin_lock_irqsave(&assoc->tgtport->lock, flags);
tgtport           626 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&assoc->tgtport->lock, flags);
tgtport           631 drivers/nvme/target/fc.c 	nvmet_fc_destroy_fcp_iodlist(assoc->tgtport, queue);
tgtport           648 drivers/nvme/target/fc.c 	spin_lock_irqsave(&queue->assoc->tgtport->lock, flags);
tgtport           650 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&queue->assoc->tgtport->lock, flags);
tgtport           652 drivers/nvme/target/fc.c 	nvmet_fc_destroy_fcp_iodlist(queue->assoc->tgtport, queue);
tgtport           677 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = queue->assoc->tgtport;
tgtport           703 drivers/nvme/target/fc.c 				tgtport->ops->fcp_abort(
tgtport           704 drivers/nvme/target/fc.c 					&tgtport->fc_target_port, fod->fcpreq);
tgtport           725 drivers/nvme/target/fc.c 		tgtport->ops->defer_rcv(&tgtport->fc_target_port,
tgtport           728 drivers/nvme/target/fc.c 		tgtport->ops->fcp_abort(&tgtport->fc_target_port,
tgtport           731 drivers/nvme/target/fc.c 		tgtport->ops->fcp_req_release(&tgtport->fc_target_port,
tgtport           752 drivers/nvme/target/fc.c nvmet_fc_find_target_queue(struct nvmet_fc_tgtport *tgtport,
tgtport           764 drivers/nvme/target/fc.c 	spin_lock_irqsave(&tgtport->lock, flags);
tgtport           765 drivers/nvme/target/fc.c 	list_for_each_entry(assoc, &tgtport->assoc_list, a_list) {
tgtport           772 drivers/nvme/target/fc.c 			spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           776 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           791 drivers/nvme/target/fc.c nvmet_fc_alloc_target_assoc(struct nvmet_fc_tgtport *tgtport)
tgtport           803 drivers/nvme/target/fc.c 	idx = ida_simple_get(&tgtport->assoc_cnt, 0, 0, GFP_KERNEL);
tgtport           807 drivers/nvme/target/fc.c 	if (!nvmet_fc_tgtport_get(tgtport))
tgtport           810 drivers/nvme/target/fc.c 	assoc->tgtport = tgtport;
tgtport           820 drivers/nvme/target/fc.c 		spin_lock_irqsave(&tgtport->lock, flags);
tgtport           822 drivers/nvme/target/fc.c 		list_for_each_entry(tmpassoc, &tgtport->assoc_list, a_list)
tgtport           829 drivers/nvme/target/fc.c 			list_add_tail(&assoc->a_list, &tgtport->assoc_list);
tgtport           831 drivers/nvme/target/fc.c 		spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           837 drivers/nvme/target/fc.c 	ida_simple_remove(&tgtport->assoc_cnt, idx);
tgtport           848 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = assoc->tgtport;
tgtport           851 drivers/nvme/target/fc.c 	spin_lock_irqsave(&tgtport->lock, flags);
tgtport           853 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           854 drivers/nvme/target/fc.c 	ida_simple_remove(&tgtport->assoc_cnt, assoc->a_id);
tgtport           856 drivers/nvme/target/fc.c 	nvmet_fc_tgtport_put(tgtport);
tgtport           874 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = assoc->tgtport;
tgtport           879 drivers/nvme/target/fc.c 	spin_lock_irqsave(&tgtport->lock, flags);
tgtport           885 drivers/nvme/target/fc.c 			spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           888 drivers/nvme/target/fc.c 			spin_lock_irqsave(&tgtport->lock, flags);
tgtport           891 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           897 drivers/nvme/target/fc.c nvmet_fc_find_target_assoc(struct nvmet_fc_tgtport *tgtport,
tgtport           904 drivers/nvme/target/fc.c 	spin_lock_irqsave(&tgtport->lock, flags);
tgtport           905 drivers/nvme/target/fc.c 	list_for_each_entry(assoc, &tgtport->assoc_list, a_list) {
tgtport           912 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport           918 drivers/nvme/target/fc.c nvmet_fc_portentry_bind(struct nvmet_fc_tgtport *tgtport,
tgtport           924 drivers/nvme/target/fc.c 	pe->tgtport = tgtport;
tgtport           925 drivers/nvme/target/fc.c 	tgtport->pe = pe;
tgtport           930 drivers/nvme/target/fc.c 	pe->node_name = tgtport->fc_target_port.node_name;
tgtport           931 drivers/nvme/target/fc.c 	pe->port_name = tgtport->fc_target_port.port_name;
tgtport           943 drivers/nvme/target/fc.c 	if (pe->tgtport)
tgtport           944 drivers/nvme/target/fc.c 		pe->tgtport->pe = NULL;
tgtport           955 drivers/nvme/target/fc.c nvmet_fc_portentry_unbind_tgt(struct nvmet_fc_tgtport *tgtport)
tgtport           961 drivers/nvme/target/fc.c 	pe = tgtport->pe;
tgtport           963 drivers/nvme/target/fc.c 		pe->tgtport = NULL;
tgtport           964 drivers/nvme/target/fc.c 	tgtport->pe = NULL;
tgtport           977 drivers/nvme/target/fc.c nvmet_fc_portentry_rebind_tgt(struct nvmet_fc_tgtport *tgtport)
tgtport           984 drivers/nvme/target/fc.c 		if (tgtport->fc_target_port.node_name == pe->node_name &&
tgtport           985 drivers/nvme/target/fc.c 		    tgtport->fc_target_port.port_name == pe->port_name) {
tgtport           986 drivers/nvme/target/fc.c 			WARN_ON(pe->tgtport);
tgtport           987 drivers/nvme/target/fc.c 			tgtport->pe = pe;
tgtport           988 drivers/nvme/target/fc.c 			pe->tgtport = tgtport;
tgtport          1096 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport =
tgtport          1098 drivers/nvme/target/fc.c 	struct device *dev = tgtport->dev;
tgtport          1102 drivers/nvme/target/fc.c 	list_del(&tgtport->tgt_list);
tgtport          1105 drivers/nvme/target/fc.c 	nvmet_fc_free_ls_iodlist(tgtport);
tgtport          1108 drivers/nvme/target/fc.c 	tgtport->ops->targetport_delete(&tgtport->fc_target_port);
tgtport          1111 drivers/nvme/target/fc.c 			tgtport->fc_target_port.port_num);
tgtport          1113 drivers/nvme/target/fc.c 	ida_destroy(&tgtport->assoc_cnt);
tgtport          1115 drivers/nvme/target/fc.c 	kfree(tgtport);
tgtport          1121 drivers/nvme/target/fc.c nvmet_fc_tgtport_put(struct nvmet_fc_tgtport *tgtport)
tgtport          1123 drivers/nvme/target/fc.c 	kref_put(&tgtport->ref, nvmet_fc_free_tgtport);
tgtport          1127 drivers/nvme/target/fc.c nvmet_fc_tgtport_get(struct nvmet_fc_tgtport *tgtport)
tgtport          1129 drivers/nvme/target/fc.c 	return kref_get_unless_zero(&tgtport->ref);
tgtport          1133 drivers/nvme/target/fc.c __nvmet_fc_free_assocs(struct nvmet_fc_tgtport *tgtport)
tgtport          1138 drivers/nvme/target/fc.c 	spin_lock_irqsave(&tgtport->lock, flags);
tgtport          1140 drivers/nvme/target/fc.c 				&tgtport->assoc_list, a_list) {
tgtport          1146 drivers/nvme/target/fc.c 	spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport          1155 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport, *next;
tgtport          1163 drivers/nvme/target/fc.c 	list_for_each_entry_safe(tgtport, next, &nvmet_fc_target_list,
tgtport          1165 drivers/nvme/target/fc.c 		if (!nvmet_fc_tgtport_get(tgtport))
tgtport          1169 drivers/nvme/target/fc.c 		spin_lock_irqsave(&tgtport->lock, flags);
tgtport          1170 drivers/nvme/target/fc.c 		list_for_each_entry(assoc, &tgtport->assoc_list, a_list) {
tgtport          1178 drivers/nvme/target/fc.c 		spin_unlock_irqrestore(&tgtport->lock, flags);
tgtport          1180 drivers/nvme/target/fc.c 		nvmet_fc_tgtport_put(tgtport);
tgtport          1207 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = targetport_to_tgtport(target_port);
tgtport          1209 drivers/nvme/target/fc.c 	nvmet_fc_portentry_unbind_tgt(tgtport);
tgtport          1212 drivers/nvme/target/fc.c 	__nvmet_fc_free_assocs(tgtport);
tgtport          1214 drivers/nvme/target/fc.c 	nvmet_fc_tgtport_put(tgtport);
tgtport          1321 drivers/nvme/target/fc.c nvmet_fc_ls_create_association(struct nvmet_fc_tgtport *tgtport,
tgtport          1359 drivers/nvme/target/fc.c 		iod->assoc = nvmet_fc_alloc_target_assoc(tgtport);
tgtport          1371 drivers/nvme/target/fc.c 		dev_err(tgtport->dev,
tgtport          1407 drivers/nvme/target/fc.c nvmet_fc_ls_create_connection(struct nvmet_fc_tgtport *tgtport,
tgtport          1445 drivers/nvme/target/fc.c 		iod->assoc = nvmet_fc_find_target_assoc(tgtport,
tgtport          1462 drivers/nvme/target/fc.c 		dev_err(tgtport->dev,
tgtport          1495 drivers/nvme/target/fc.c nvmet_fc_ls_disconnect(struct nvmet_fc_tgtport *tgtport,
tgtport          1531 drivers/nvme/target/fc.c 		assoc = nvmet_fc_find_target_assoc(tgtport,
tgtport          1539 drivers/nvme/target/fc.c 		dev_err(tgtport->dev,
tgtport          1580 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = iod->tgtport;
tgtport          1582 drivers/nvme/target/fc.c 	fc_dma_sync_single_for_cpu(tgtport->dev, iod->rspdma,
tgtport          1584 drivers/nvme/target/fc.c 	nvmet_fc_free_ls_iod(tgtport, iod);
tgtport          1585 drivers/nvme/target/fc.c 	nvmet_fc_tgtport_put(tgtport);
tgtport          1589 drivers/nvme/target/fc.c nvmet_fc_xmt_ls_rsp(struct nvmet_fc_tgtport *tgtport,
tgtport          1594 drivers/nvme/target/fc.c 	fc_dma_sync_single_for_device(tgtport->dev, iod->rspdma,
tgtport          1597 drivers/nvme/target/fc.c 	ret = tgtport->ops->xmt_ls_rsp(&tgtport->fc_target_port, iod->lsreq);
tgtport          1606 drivers/nvme/target/fc.c nvmet_fc_handle_ls_rqst(struct nvmet_fc_tgtport *tgtport,
tgtport          1629 drivers/nvme/target/fc.c 		nvmet_fc_ls_create_association(tgtport, iod);
tgtport          1633 drivers/nvme/target/fc.c 		nvmet_fc_ls_create_connection(tgtport, iod);
tgtport          1637 drivers/nvme/target/fc.c 		nvmet_fc_ls_disconnect(tgtport, iod);
tgtport          1645 drivers/nvme/target/fc.c 	nvmet_fc_xmt_ls_rsp(tgtport, iod);
tgtport          1656 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = iod->tgtport;
tgtport          1658 drivers/nvme/target/fc.c 	nvmet_fc_handle_ls_rqst(tgtport, iod);
tgtport          1684 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = targetport_to_tgtport(target_port);
tgtport          1690 drivers/nvme/target/fc.c 	if (!nvmet_fc_tgtport_get(tgtport))
tgtport          1693 drivers/nvme/target/fc.c 	iod = nvmet_fc_alloc_ls_iod(tgtport);
tgtport          1695 drivers/nvme/target/fc.c 		nvmet_fc_tgtport_put(tgtport);
tgtport          1729 drivers/nvme/target/fc.c 	fod->data_sg_cnt = fc_dma_map_sg(fod->tgtport->dev, sg, nent,
tgtport          1747 drivers/nvme/target/fc.c 	fc_dma_unmap_sg(fod->tgtport->dev, fod->data_sg, fod->data_sg_cnt,
tgtport          1773 drivers/nvme/target/fc.c nvmet_fc_prep_fcp_rsp(struct nvmet_fc_tgtport *tgtport,
tgtport          1831 drivers/nvme/target/fc.c 	fc_dma_sync_single_for_device(tgtport->dev, fod->rspdma,
tgtport          1838 drivers/nvme/target/fc.c nvmet_fc_abort_op(struct nvmet_fc_tgtport *tgtport,
tgtport          1852 drivers/nvme/target/fc.c 		tgtport->ops->fcp_abort(&tgtport->fc_target_port, fcpreq);
tgtport          1858 drivers/nvme/target/fc.c nvmet_fc_xmt_fcp_rsp(struct nvmet_fc_tgtport *tgtport,
tgtport          1866 drivers/nvme/target/fc.c 	nvmet_fc_prep_fcp_rsp(tgtport, fod);
tgtport          1868 drivers/nvme/target/fc.c 	ret = tgtport->ops->fcp_op(&tgtport->fc_target_port, fod->fcpreq);
tgtport          1870 drivers/nvme/target/fc.c 		nvmet_fc_abort_op(tgtport, fod);
tgtport          1874 drivers/nvme/target/fc.c nvmet_fc_transfer_fcp_data(struct nvmet_fc_tgtport *tgtport,
tgtport          1900 drivers/nvme/target/fc.c 	       fcpreq->sg_cnt < tgtport->max_sg_cnt &&
tgtport          1927 drivers/nvme/target/fc.c 	    (tgtport->ops->target_features & NVMET_FCTGTFEAT_READDATA_RSP)) {
tgtport          1929 drivers/nvme/target/fc.c 		nvmet_fc_prep_fcp_rsp(tgtport, fod);
tgtport          1932 drivers/nvme/target/fc.c 	ret = tgtport->ops->fcp_op(&tgtport->fc_target_port, fod->fcpreq);
tgtport          1958 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = fod->tgtport;
tgtport          1967 drivers/nvme/target/fc.c 		nvmet_fc_abort_op(tgtport, fod);
tgtport          1981 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = fod->tgtport;
tgtport          2012 drivers/nvme/target/fc.c 			nvmet_fc_transfer_fcp_data(tgtport, fod,
tgtport          2027 drivers/nvme/target/fc.c 			nvmet_fc_abort_op(tgtport, fod);
tgtport          2043 drivers/nvme/target/fc.c 			nvmet_fc_transfer_fcp_data(tgtport, fod,
tgtport          2053 drivers/nvme/target/fc.c 		nvmet_fc_xmt_fcp_rsp(tgtport, fod);
tgtport          2080 drivers/nvme/target/fc.c __nvmet_fc_fcp_nvme_cmd_done(struct nvmet_fc_tgtport *tgtport,
tgtport          2097 drivers/nvme/target/fc.c 		nvmet_fc_abort_op(tgtport, fod);
tgtport          2118 drivers/nvme/target/fc.c 			nvmet_fc_transfer_fcp_data(tgtport, fod,
tgtport          2129 drivers/nvme/target/fc.c 	nvmet_fc_xmt_fcp_rsp(tgtport, fod);
tgtport          2137 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = fod->tgtport;
tgtport          2139 drivers/nvme/target/fc.c 	__nvmet_fc_fcp_nvme_cmd_done(tgtport, fod, 0);
tgtport          2147 drivers/nvme/target/fc.c nvmet_fc_handle_fcp_rqst(struct nvmet_fc_tgtport *tgtport,
tgtport          2158 drivers/nvme/target/fc.c 	if (!tgtport->pe)
tgtport          2188 drivers/nvme/target/fc.c 	fod->req.port = tgtport->pe->port;
tgtport          2224 drivers/nvme/target/fc.c 		nvmet_fc_transfer_fcp_data(tgtport, fod, NVMET_FCOP_WRITEDATA);
tgtport          2238 drivers/nvme/target/fc.c 	nvmet_fc_abort_op(tgtport, fod);
tgtport          2293 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = targetport_to_tgtport(target_port);
tgtport          2307 drivers/nvme/target/fc.c 	queue = nvmet_fc_find_target_queue(tgtport,
tgtport          2330 drivers/nvme/target/fc.c 		nvmet_fc_queue_fcp_req(tgtport, queue, fcpreq);
tgtport          2335 drivers/nvme/target/fc.c 	if (!tgtport->ops->defer_rcv) {
tgtport          2498 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport;
tgtport          2522 drivers/nvme/target/fc.c 	list_for_each_entry(tgtport, &nvmet_fc_target_list, tgt_list) {
tgtport          2523 drivers/nvme/target/fc.c 		if ((tgtport->fc_target_port.node_name == traddr.nn) &&
tgtport          2524 drivers/nvme/target/fc.c 		    (tgtport->fc_target_port.port_name == traddr.pn)) {
tgtport          2526 drivers/nvme/target/fc.c 			if (!tgtport->pe) {
tgtport          2527 drivers/nvme/target/fc.c 				nvmet_fc_portentry_bind(tgtport, pe, port);
tgtport          2556 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport = pe->tgtport;
tgtport          2558 drivers/nvme/target/fc.c 	if (tgtport && tgtport->ops->discovery_event)
tgtport          2559 drivers/nvme/target/fc.c 		tgtport->ops->discovery_event(&tgtport->fc_target_port);
tgtport           373 drivers/nvme/target/fcloop.c fcloop_tgt_discovery_evt(struct nvmet_fc_target_port *tgtport)
tgtport           381 drivers/nvme/target/fcloop.c 	tgt_rscn->tport = tgtport->private;
tgtport           615 drivers/nvme/target/fcloop.c fcloop_fcp_op(struct nvmet_fc_target_port *tgtport,
tgtport           708 drivers/nvme/target/fcloop.c fcloop_tgt_fcp_abort(struct nvmet_fc_target_port *tgtport,
tgtport           732 drivers/nvme/target/fcloop.c fcloop_fcp_req_release(struct nvmet_fc_target_port *tgtport,
tgtport           840 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_ls_rsp(struct nvmet_fc_target_port *tgtport,
tgtport           849 drivers/scsi/lpfc/lpfc_nvmet.c 	struct lpfc_nvmet_tgtport *nvmep = tgtport->private;
tgtport           930 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_fcp_op(struct nvmet_fc_target_port *tgtport,
tgtport           933 drivers/scsi/lpfc/lpfc_nvmet.c 	struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private;
tgtport          1064 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_fcp_abort(struct nvmet_fc_target_port *tgtport,
tgtport          1067 drivers/scsi/lpfc/lpfc_nvmet.c 	struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private;
tgtport          1126 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_fcp_release(struct nvmet_fc_target_port *tgtport,
tgtport          1129 drivers/scsi/lpfc/lpfc_nvmet.c 	struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private;
tgtport          1169 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_defer_rcv(struct nvmet_fc_target_port *tgtport,
tgtport          1203 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_discovery_event(struct nvmet_fc_target_port *tgtport)
tgtport          1209 drivers/scsi/lpfc/lpfc_nvmet.c 	tgtp = tgtport->private;
tgtport           831 include/linux/nvme-fc-driver.h 	void (*targetport_delete)(struct nvmet_fc_target_port *tgtport);
tgtport           832 include/linux/nvme-fc-driver.h 	int (*xmt_ls_rsp)(struct nvmet_fc_target_port *tgtport,
tgtport           834 include/linux/nvme-fc-driver.h 	int (*fcp_op)(struct nvmet_fc_target_port *tgtport,
tgtport           836 include/linux/nvme-fc-driver.h 	void (*fcp_abort)(struct nvmet_fc_target_port *tgtport,
tgtport           838 include/linux/nvme-fc-driver.h 	void (*fcp_req_release)(struct nvmet_fc_target_port *tgtport,
tgtport           840 include/linux/nvme-fc-driver.h 	void (*defer_rcv)(struct nvmet_fc_target_port *tgtport,
tgtport           842 include/linux/nvme-fc-driver.h 	void (*discovery_event)(struct nvmet_fc_target_port *tgtport);
tgtport           860 include/linux/nvme-fc-driver.h int nvmet_fc_unregister_targetport(struct nvmet_fc_target_port *tgtport);
tgtport           862 include/linux/nvme-fc-driver.h int nvmet_fc_rcv_ls_req(struct nvmet_fc_target_port *tgtport,
tgtport           866 include/linux/nvme-fc-driver.h int nvmet_fc_rcv_fcp_req(struct nvmet_fc_target_port *tgtport,
tgtport           870 include/linux/nvme-fc-driver.h void nvmet_fc_rcv_fcp_abort(struct nvmet_fc_target_port *tgtport,