1#ifndef TARGET_CORE_FABRIC_H
2#define TARGET_CORE_FABRIC_H
3
4struct target_core_fabric_ops {
5	struct module *module;
6	const char *name;
7	size_t node_acl_size;
8	/*
9	 * Limits number of scatterlist entries per SCF_SCSI_DATA_CDB payload.
10	 * Setting this value tells target-core to enforce this limit, and
11	 * report as INQUIRY EVPD=b0 MAXIMUM TRANSFER LENGTH.
12	 *
13	 * target-core will currently reset se_cmd->data_length to this
14	 * maximum size, and set UNDERFLOW residual count if length exceeds
15	 * this limit.
16	 *
17	 * XXX: Not all initiator hosts honor this block-limit EVPD
18	 * XXX: Currently assumes single PAGE_SIZE per scatterlist entry
19	 */
20	u32 max_data_sg_nents;
21	char *(*get_fabric_name)(void);
22	char *(*tpg_get_wwn)(struct se_portal_group *);
23	u16 (*tpg_get_tag)(struct se_portal_group *);
24	u32 (*tpg_get_default_depth)(struct se_portal_group *);
25	int (*tpg_check_demo_mode)(struct se_portal_group *);
26	int (*tpg_check_demo_mode_cache)(struct se_portal_group *);
27	int (*tpg_check_demo_mode_write_protect)(struct se_portal_group *);
28	int (*tpg_check_prod_mode_write_protect)(struct se_portal_group *);
29	/*
30	 * Optionally used by fabrics to allow demo-mode login, but not
31	 * expose any TPG LUNs, and return 'not connected' in standard
32	 * inquiry response
33	 */
34	int (*tpg_check_demo_mode_login_only)(struct se_portal_group *);
35	/*
36	 * Optionally used as a configfs tunable to determine when
37	 * target-core should signal the PROTECT=1 feature bit for
38	 * backends that don't support T10-PI, so that either fabric
39	 * HW offload or target-core emulation performs the associated
40	 * WRITE_STRIP and READ_INSERT operations.
41	 */
42	int (*tpg_check_prot_fabric_only)(struct se_portal_group *);
43	u32 (*tpg_get_inst_index)(struct se_portal_group *);
44	/*
45	 * Optional to release struct se_cmd and fabric dependent allocated
46	 * I/O descriptor in transport_cmd_check_stop().
47	 *
48	 * Returning 1 will signal a descriptor has been released.
49	 * Returning 0 will signal a descriptor has not been released.
50	 */
51	int (*check_stop_free)(struct se_cmd *);
52	void (*release_cmd)(struct se_cmd *);
53	/*
54	 * Called with spin_lock_bh(struct se_portal_group->session_lock held.
55	 */
56	int (*shutdown_session)(struct se_session *);
57	void (*close_session)(struct se_session *);
58	u32 (*sess_get_index)(struct se_session *);
59	/*
60	 * Used only for SCSI fabrics that contain multi-value TransportIDs
61	 * (like iSCSI).  All other SCSI fabrics should set this to NULL.
62	 */
63	u32 (*sess_get_initiator_sid)(struct se_session *,
64				      unsigned char *, u32);
65	int (*write_pending)(struct se_cmd *);
66	int (*write_pending_status)(struct se_cmd *);
67	void (*set_default_node_attributes)(struct se_node_acl *);
68	int (*get_cmd_state)(struct se_cmd *);
69	int (*queue_data_in)(struct se_cmd *);
70	int (*queue_status)(struct se_cmd *);
71	void (*queue_tm_rsp)(struct se_cmd *);
72	void (*aborted_task)(struct se_cmd *);
73	/*
74	 * fabric module calls for target_core_fabric_configfs.c
75	 */
76	struct se_wwn *(*fabric_make_wwn)(struct target_fabric_configfs *,
77				struct config_group *, const char *);
78	void (*fabric_drop_wwn)(struct se_wwn *);
79	struct se_portal_group *(*fabric_make_tpg)(struct se_wwn *,
80				struct config_group *, const char *);
81	void (*fabric_drop_tpg)(struct se_portal_group *);
82	int (*fabric_post_link)(struct se_portal_group *,
83				struct se_lun *);
84	void (*fabric_pre_unlink)(struct se_portal_group *,
85				struct se_lun *);
86	struct se_tpg_np *(*fabric_make_np)(struct se_portal_group *,
87				struct config_group *, const char *);
88	void (*fabric_drop_np)(struct se_tpg_np *);
89	int (*fabric_init_nodeacl)(struct se_node_acl *, const char *);
90	void (*fabric_cleanup_nodeacl)(struct se_node_acl *);
91
92	struct configfs_attribute **tfc_discovery_attrs;
93	struct configfs_attribute **tfc_wwn_attrs;
94	struct configfs_attribute **tfc_tpg_base_attrs;
95	struct configfs_attribute **tfc_tpg_np_base_attrs;
96	struct configfs_attribute **tfc_tpg_attrib_attrs;
97	struct configfs_attribute **tfc_tpg_auth_attrs;
98	struct configfs_attribute **tfc_tpg_param_attrs;
99	struct configfs_attribute **tfc_tpg_nacl_base_attrs;
100	struct configfs_attribute **tfc_tpg_nacl_attrib_attrs;
101	struct configfs_attribute **tfc_tpg_nacl_auth_attrs;
102	struct configfs_attribute **tfc_tpg_nacl_param_attrs;
103};
104
105int target_register_template(const struct target_core_fabric_ops *fo);
106void target_unregister_template(const struct target_core_fabric_ops *fo);
107
108int target_depend_item(struct config_item *item);
109void target_undepend_item(struct config_item *item);
110
111struct se_session *transport_init_session(enum target_prot_op);
112int transport_alloc_session_tags(struct se_session *, unsigned int,
113		unsigned int);
114struct se_session *transport_init_session_tags(unsigned int, unsigned int,
115		enum target_prot_op);
116void	__transport_register_session(struct se_portal_group *,
117		struct se_node_acl *, struct se_session *, void *);
118void	transport_register_session(struct se_portal_group *,
119		struct se_node_acl *, struct se_session *, void *);
120void	target_get_session(struct se_session *);
121void	target_put_session(struct se_session *);
122ssize_t	target_show_dynamic_sessions(struct se_portal_group *, char *);
123void	transport_free_session(struct se_session *);
124void	target_put_nacl(struct se_node_acl *);
125void	transport_deregister_session_configfs(struct se_session *);
126void	transport_deregister_session(struct se_session *);
127
128
129void	transport_init_se_cmd(struct se_cmd *,
130		const struct target_core_fabric_ops *,
131		struct se_session *, u32, int, int, unsigned char *);
132sense_reason_t transport_lookup_cmd_lun(struct se_cmd *, u64);
133sense_reason_t target_setup_cmd_from_cdb(struct se_cmd *, unsigned char *);
134int	target_submit_cmd_map_sgls(struct se_cmd *, struct se_session *,
135		unsigned char *, unsigned char *, u64, u32, int, int, int,
136		struct scatterlist *, u32, struct scatterlist *, u32,
137		struct scatterlist *, u32);
138int	target_submit_cmd(struct se_cmd *, struct se_session *, unsigned char *,
139		unsigned char *, u64, u32, int, int, int);
140int	target_submit_tmr(struct se_cmd *se_cmd, struct se_session *se_sess,
141		unsigned char *sense, u64 unpacked_lun,
142		void *fabric_tmr_ptr, unsigned char tm_type,
143		gfp_t, unsigned int, int);
144int	transport_handle_cdb_direct(struct se_cmd *);
145sense_reason_t	transport_generic_new_cmd(struct se_cmd *);
146
147void	target_execute_cmd(struct se_cmd *cmd);
148
149int	transport_generic_free_cmd(struct se_cmd *, int);
150
151bool	transport_wait_for_tasks(struct se_cmd *);
152int	transport_check_aborted_status(struct se_cmd *, int);
153int	transport_send_check_condition_and_sense(struct se_cmd *,
154		sense_reason_t, int);
155int	target_get_sess_cmd(struct se_cmd *, bool);
156int	target_put_sess_cmd(struct se_cmd *);
157void	target_sess_cmd_list_set_waiting(struct se_session *);
158void	target_wait_for_sess_cmds(struct se_session *);
159
160int	core_alua_check_nonop_delay(struct se_cmd *);
161
162int	core_tmr_alloc_req(struct se_cmd *, void *, u8, gfp_t);
163void	core_tmr_release_req(struct se_tmr_req *);
164int	transport_generic_handle_tmr(struct se_cmd *);
165void	transport_generic_request_failure(struct se_cmd *, sense_reason_t);
166void	__target_execute_cmd(struct se_cmd *);
167int	transport_lookup_tmr_lun(struct se_cmd *, u64);
168void	core_allocate_nexus_loss_ua(struct se_node_acl *acl);
169
170struct se_node_acl *core_tpg_get_initiator_node_acl(struct se_portal_group *tpg,
171		unsigned char *);
172struct se_node_acl *core_tpg_check_initiator_node_acl(struct se_portal_group *,
173		unsigned char *);
174int	core_tpg_set_initiator_node_queue_depth(struct se_portal_group *,
175		unsigned char *, u32, int);
176int	core_tpg_set_initiator_node_tag(struct se_portal_group *,
177		struct se_node_acl *, const char *);
178int	core_tpg_register(struct se_wwn *, struct se_portal_group *, int);
179int	core_tpg_deregister(struct se_portal_group *);
180
181/*
182 * The LIO target core uses DMA_TO_DEVICE to mean that data is going
183 * to the target (eg handling a WRITE) and DMA_FROM_DEVICE to mean
184 * that data is coming from the target (eg handling a READ).  However,
185 * this is just the opposite of what we have to tell the DMA mapping
186 * layer -- eg when handling a READ, the HBA will have to DMA the data
187 * out of memory so it can send it to the initiator, which means we
188 * need to use DMA_TO_DEVICE when we map the data.
189 */
190static inline enum dma_data_direction
191target_reverse_dma_direction(struct se_cmd *se_cmd)
192{
193	if (se_cmd->se_cmd_flags & SCF_BIDI)
194		return DMA_BIDIRECTIONAL;
195
196	switch (se_cmd->data_direction) {
197	case DMA_TO_DEVICE:
198		return DMA_FROM_DEVICE;
199	case DMA_FROM_DEVICE:
200		return DMA_TO_DEVICE;
201	case DMA_NONE:
202	default:
203		return DMA_NONE;
204	}
205}
206
207#endif /* TARGET_CORE_FABRICH */
208