1
2
3
4
5
6
7
8
9
10 #ifndef _HPTIOP_H_
11 #define _HPTIOP_H_
12
13 struct hpt_iopmu_itl {
14 __le32 resrved0[4];
15 __le32 inbound_msgaddr0;
16 __le32 inbound_msgaddr1;
17 __le32 outbound_msgaddr0;
18 __le32 outbound_msgaddr1;
19 __le32 inbound_doorbell;
20 __le32 inbound_intstatus;
21 __le32 inbound_intmask;
22 __le32 outbound_doorbell;
23 __le32 outbound_intstatus;
24 __le32 outbound_intmask;
25 __le32 reserved1[2];
26 __le32 inbound_queue;
27 __le32 outbound_queue;
28 };
29
30 #define IOPMU_QUEUE_EMPTY 0xffffffff
31 #define IOPMU_QUEUE_MASK_HOST_BITS 0xf0000000
32 #define IOPMU_QUEUE_ADDR_HOST_BIT 0x80000000
33 #define IOPMU_QUEUE_REQUEST_SIZE_BIT 0x40000000
34 #define IOPMU_QUEUE_REQUEST_RESULT_BIT 0x40000000
35
36 #define IOPMU_OUTBOUND_INT_MSG0 1
37 #define IOPMU_OUTBOUND_INT_MSG1 2
38 #define IOPMU_OUTBOUND_INT_DOORBELL 4
39 #define IOPMU_OUTBOUND_INT_POSTQUEUE 8
40 #define IOPMU_OUTBOUND_INT_PCI 0x10
41
42 #define IOPMU_INBOUND_INT_MSG0 1
43 #define IOPMU_INBOUND_INT_MSG1 2
44 #define IOPMU_INBOUND_INT_DOORBELL 4
45 #define IOPMU_INBOUND_INT_ERROR 8
46 #define IOPMU_INBOUND_INT_POSTQUEUE 0x10
47
48 #define MVIOP_QUEUE_LEN 512
49
50 struct hpt_iopmu_mv {
51 __le32 inbound_head;
52 __le32 inbound_tail;
53 __le32 outbound_head;
54 __le32 outbound_tail;
55 __le32 inbound_msg;
56 __le32 outbound_msg;
57 __le32 reserve[10];
58 __le64 inbound_q[MVIOP_QUEUE_LEN];
59 __le64 outbound_q[MVIOP_QUEUE_LEN];
60 };
61
62 struct hpt_iopmv_regs {
63 __le32 reserved[0x20400 / 4];
64 __le32 inbound_doorbell;
65 __le32 inbound_intmask;
66 __le32 outbound_doorbell;
67 __le32 outbound_intmask;
68 };
69
70 #pragma pack(1)
71 struct hpt_iopmu_mvfrey {
72 __le32 reserved0[(0x4000 - 0) / 4];
73 __le32 inbound_base;
74 __le32 inbound_base_high;
75 __le32 reserved1[(0x4018 - 0x4008) / 4];
76 __le32 inbound_write_ptr;
77 __le32 reserved2[(0x402c - 0x401c) / 4];
78 __le32 inbound_conf_ctl;
79 __le32 reserved3[(0x4050 - 0x4030) / 4];
80 __le32 outbound_base;
81 __le32 outbound_base_high;
82 __le32 outbound_shadow_base;
83 __le32 outbound_shadow_base_high;
84 __le32 reserved4[(0x4088 - 0x4060) / 4];
85 __le32 isr_cause;
86 __le32 isr_enable;
87 __le32 reserved5[(0x1020c - 0x4090) / 4];
88 __le32 pcie_f0_int_enable;
89 __le32 reserved6[(0x10400 - 0x10210) / 4];
90 __le32 f0_to_cpu_msg_a;
91 __le32 reserved7[(0x10420 - 0x10404) / 4];
92 __le32 cpu_to_f0_msg_a;
93 __le32 reserved8[(0x10480 - 0x10424) / 4];
94 __le32 f0_doorbell;
95 __le32 f0_doorbell_enable;
96 };
97
98 struct mvfrey_inlist_entry {
99 dma_addr_t addr;
100 __le32 intrfc_len;
101 __le32 reserved;
102 };
103
104 struct mvfrey_outlist_entry {
105 __le32 val;
106 };
107 #pragma pack()
108
109 #define MVIOP_MU_QUEUE_ADDR_HOST_MASK (~(0x1full))
110 #define MVIOP_MU_QUEUE_ADDR_HOST_BIT 4
111
112 #define MVIOP_MU_QUEUE_ADDR_IOP_HIGH32 0xffffffff
113 #define MVIOP_MU_QUEUE_REQUEST_RESULT_BIT 1
114 #define MVIOP_MU_QUEUE_REQUEST_RETURN_CONTEXT 2
115
116 #define MVIOP_MU_INBOUND_INT_MSG 1
117 #define MVIOP_MU_INBOUND_INT_POSTQUEUE 2
118 #define MVIOP_MU_OUTBOUND_INT_MSG 1
119 #define MVIOP_MU_OUTBOUND_INT_POSTQUEUE 2
120
121 #define CL_POINTER_TOGGLE 0x00004000
122 #define CPU_TO_F0_DRBL_MSG_BIT 0x02000000
123
124 enum hpt_iopmu_message {
125
126 IOPMU_INBOUND_MSG0_NOP = 0,
127 IOPMU_INBOUND_MSG0_RESET,
128 IOPMU_INBOUND_MSG0_FLUSH,
129 IOPMU_INBOUND_MSG0_SHUTDOWN,
130 IOPMU_INBOUND_MSG0_STOP_BACKGROUND_TASK,
131 IOPMU_INBOUND_MSG0_START_BACKGROUND_TASK,
132 IOPMU_INBOUND_MSG0_RESET_COMM,
133 IOPMU_INBOUND_MSG0_MAX = 0xff,
134
135 IOPMU_OUTBOUND_MSG0_REGISTER_DEVICE_0 = 0x100,
136 IOPMU_OUTBOUND_MSG0_REGISTER_DEVICE_MAX = 0x1ff,
137 IOPMU_OUTBOUND_MSG0_UNREGISTER_DEVICE_0 = 0x200,
138 IOPMU_OUTBOUND_MSG0_UNREGISTER_DEVICE_MAX = 0x2ff,
139 IOPMU_OUTBOUND_MSG0_REVALIDATE_DEVICE_0 = 0x300,
140 IOPMU_OUTBOUND_MSG0_REVALIDATE_DEVICE_MAX = 0x3ff,
141 };
142
143 struct hpt_iop_request_header {
144 __le32 size;
145 __le32 type;
146 __le32 flags;
147 __le32 result;
148 __le32 context;
149 __le32 context_hi32;
150 };
151
152 #define IOP_REQUEST_FLAG_SYNC_REQUEST 1
153 #define IOP_REQUEST_FLAG_BIST_REQUEST 2
154 #define IOP_REQUEST_FLAG_REMAPPED 4
155 #define IOP_REQUEST_FLAG_OUTPUT_CONTEXT 8
156 #define IOP_REQUEST_FLAG_ADDR_BITS 0x40
157
158 enum hpt_iop_request_type {
159 IOP_REQUEST_TYPE_GET_CONFIG = 0,
160 IOP_REQUEST_TYPE_SET_CONFIG,
161 IOP_REQUEST_TYPE_BLOCK_COMMAND,
162 IOP_REQUEST_TYPE_SCSI_COMMAND,
163 IOP_REQUEST_TYPE_IOCTL_COMMAND,
164 IOP_REQUEST_TYPE_MAX
165 };
166
167 enum hpt_iop_result_type {
168 IOP_RESULT_PENDING = 0,
169 IOP_RESULT_SUCCESS,
170 IOP_RESULT_FAIL,
171 IOP_RESULT_BUSY,
172 IOP_RESULT_RESET,
173 IOP_RESULT_INVALID_REQUEST,
174 IOP_RESULT_BAD_TARGET,
175 IOP_RESULT_CHECK_CONDITION,
176 };
177
178 struct hpt_iop_request_get_config {
179 struct hpt_iop_request_header header;
180 __le32 interface_version;
181 __le32 firmware_version;
182 __le32 max_requests;
183 __le32 request_size;
184 __le32 max_sg_count;
185 __le32 data_transfer_length;
186 __le32 alignment_mask;
187 __le32 max_devices;
188 __le32 sdram_size;
189 };
190
191 struct hpt_iop_request_set_config {
192 struct hpt_iop_request_header header;
193 __le32 iop_id;
194 __le16 vbus_id;
195 __le16 max_host_request_size;
196 __le32 reserve[6];
197 };
198
199 struct hpt_iopsg {
200 __le32 size;
201 __le32 eot;
202 __le64 pci_address;
203 };
204
205 struct hpt_iop_request_block_command {
206 struct hpt_iop_request_header header;
207 u8 channel;
208 u8 target;
209 u8 lun;
210 u8 pad1;
211 __le16 command;
212 __le16 sectors;
213 __le64 lba;
214 struct hpt_iopsg sg_list[1];
215 };
216
217 #define IOP_BLOCK_COMMAND_READ 1
218 #define IOP_BLOCK_COMMAND_WRITE 2
219 #define IOP_BLOCK_COMMAND_VERIFY 3
220 #define IOP_BLOCK_COMMAND_FLUSH 4
221 #define IOP_BLOCK_COMMAND_SHUTDOWN 5
222
223 struct hpt_iop_request_scsi_command {
224 struct hpt_iop_request_header header;
225 u8 channel;
226 u8 target;
227 u8 lun;
228 u8 pad1;
229 u8 cdb[16];
230 __le32 dataxfer_length;
231 struct hpt_iopsg sg_list[1];
232 };
233
234 struct hpt_iop_request_ioctl_command {
235 struct hpt_iop_request_header header;
236 __le32 ioctl_code;
237 __le32 inbuf_size;
238 __le32 outbuf_size;
239 __le32 bytes_returned;
240 u8 buf[1];
241
242 };
243
244 #define HPTIOP_MAX_REQUESTS 256u
245
246 struct hptiop_request {
247 struct hptiop_request *next;
248 void *req_virt;
249 u32 req_shifted_phy;
250 struct scsi_cmnd *scp;
251 int index;
252 };
253
254 struct hpt_scsi_pointer {
255 int mapped;
256 int sgcnt;
257 dma_addr_t dma_handle;
258 };
259
260 #define HPT_SCP(scp) ((struct hpt_scsi_pointer *)&(scp)->SCp)
261
262 enum hptiop_family {
263 UNKNOWN_BASED_IOP,
264 INTEL_BASED_IOP,
265 MV_BASED_IOP,
266 MVFREY_BASED_IOP
267 } ;
268
269 struct hptiop_hba {
270 struct hptiop_adapter_ops *ops;
271 union {
272 struct {
273 struct hpt_iopmu_itl __iomem *iop;
274 void __iomem *plx;
275 } itl;
276 struct {
277 struct hpt_iopmv_regs *regs;
278 struct hpt_iopmu_mv __iomem *mu;
279 void *internal_req;
280 dma_addr_t internal_req_phy;
281 } mv;
282 struct {
283 struct hpt_iop_request_get_config __iomem *config;
284 struct hpt_iopmu_mvfrey __iomem *mu;
285
286 int internal_mem_size;
287 struct hptiop_request internal_req;
288 int list_count;
289 struct mvfrey_inlist_entry *inlist;
290 dma_addr_t inlist_phy;
291 __le32 inlist_wptr;
292 struct mvfrey_outlist_entry *outlist;
293 dma_addr_t outlist_phy;
294 __le32 *outlist_cptr;
295 dma_addr_t outlist_cptr_phy;
296 __le32 outlist_rptr;
297 } mvfrey;
298 } u;
299
300 struct Scsi_Host *host;
301 struct pci_dev *pcidev;
302
303
304 u32 interface_version;
305 u32 firmware_version;
306 u32 sdram_size;
307 u32 max_devices;
308 u32 max_requests;
309 u32 max_request_size;
310 u32 max_sg_descriptors;
311
312 u32 req_size;
313
314 u32 iopintf_v2: 1;
315 u32 initialized: 1;
316 u32 msg_done: 1;
317
318 struct hptiop_request * req_list;
319 struct hptiop_request reqs[HPTIOP_MAX_REQUESTS];
320
321
322 void *dma_coherent[HPTIOP_MAX_REQUESTS];
323 dma_addr_t dma_coherent_handle[HPTIOP_MAX_REQUESTS];
324
325 atomic_t reset_count;
326 atomic_t resetting;
327
328 wait_queue_head_t reset_wq;
329 wait_queue_head_t ioctl_wq;
330 };
331
332 struct hpt_ioctl_k {
333 struct hptiop_hba * hba;
334 u32 ioctl_code;
335 u32 inbuf_size;
336 u32 outbuf_size;
337 void *inbuf;
338 void *outbuf;
339 u32 *bytes_returned;
340 void (*done)(struct hpt_ioctl_k *);
341 int result;
342 };
343
344 struct hptiop_adapter_ops {
345 enum hptiop_family family;
346 int (*iop_wait_ready)(struct hptiop_hba *hba, u32 millisec);
347 int (*internal_memalloc)(struct hptiop_hba *hba);
348 int (*internal_memfree)(struct hptiop_hba *hba);
349 int (*map_pci_bar)(struct hptiop_hba *hba);
350 void (*unmap_pci_bar)(struct hptiop_hba *hba);
351 void (*enable_intr)(struct hptiop_hba *hba);
352 void (*disable_intr)(struct hptiop_hba *hba);
353 int (*get_config)(struct hptiop_hba *hba,
354 struct hpt_iop_request_get_config *config);
355 int (*set_config)(struct hptiop_hba *hba,
356 struct hpt_iop_request_set_config *config);
357 int (*iop_intr)(struct hptiop_hba *hba);
358 void (*post_msg)(struct hptiop_hba *hba, u32 msg);
359 void (*post_req)(struct hptiop_hba *hba, struct hptiop_request *_req);
360 int hw_dma_bit_mask;
361 int (*reset_comm)(struct hptiop_hba *hba);
362 __le64 host_phy_flag;
363 };
364
365 #define HPT_IOCTL_RESULT_OK 0
366 #define HPT_IOCTL_RESULT_FAILED (-1)
367
368 #if 0
369 #define dprintk(fmt, args...) do { printk(fmt, ##args); } while(0)
370 #else
371 #define dprintk(fmt, args...)
372 #endif
373
374 #endif