]>
Commit | Line | Data |
---|---|---|
1da177e4 | 1 | /* |
00bab910 | 2 | * zfcp device driver |
1da177e4 | 3 | * |
00bab910 | 4 | * Setup and helper functions to access QDIO. |
1da177e4 | 5 | * |
00bab910 | 6 | * Copyright IBM Corporation 2002, 2008 |
1da177e4 LT |
7 | */ |
8 | ||
1da177e4 LT |
9 | #include "zfcp_ext.h" |
10 | ||
00bab910 SS |
11 | /* FIXME(tune): free space should be one max. SBAL chain plus what? */ |
12 | #define ZFCP_QDIO_PCI_INTERVAL (QDIO_MAX_BUFFERS_PER_Q \ | |
c41f8cbd | 13 | - (FSF_MAX_SBALS_PER_REQ + 4)) |
5d4e2262 | 14 | #define QBUFF_PER_PAGE (PAGE_SIZE / sizeof(struct qdio_buffer)) |
1da177e4 | 15 | |
00bab910 | 16 | static int zfcp_qdio_buffers_enqueue(struct qdio_buffer **sbal) |
1da177e4 | 17 | { |
b4e44590 | 18 | int pos; |
1da177e4 | 19 | |
b4e44590 | 20 | for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos += QBUFF_PER_PAGE) { |
00bab910 SS |
21 | sbal[pos] = (struct qdio_buffer *) get_zeroed_page(GFP_KERNEL); |
22 | if (!sbal[pos]) | |
b4e44590 | 23 | return -ENOMEM; |
b4e44590 SS |
24 | } |
25 | for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos++) | |
26 | if (pos % QBUFF_PER_PAGE) | |
00bab910 | 27 | sbal[pos] = sbal[pos - 1] + 1; |
b4e44590 | 28 | return 0; |
1da177e4 LT |
29 | } |
30 | ||
00bab910 SS |
31 | static volatile struct qdio_buffer_element * |
32 | zfcp_qdio_sbale(struct zfcp_qdio_queue *q, int sbal_idx, int sbale_idx) | |
1da177e4 | 33 | { |
00bab910 | 34 | return &q->sbal[sbal_idx]->element[sbale_idx]; |
1da177e4 LT |
35 | } |
36 | ||
00bab910 SS |
37 | /** |
38 | * zfcp_qdio_free - free memory used by request- and resposne queue | |
39 | * @adapter: pointer to the zfcp_adapter structure | |
40 | */ | |
41 | void zfcp_qdio_free(struct zfcp_adapter *adapter) | |
1da177e4 | 42 | { |
00bab910 SS |
43 | struct qdio_buffer **sbal_req, **sbal_resp; |
44 | int p; | |
1da177e4 | 45 | |
00bab910 SS |
46 | if (adapter->ccw_device) |
47 | qdio_free(adapter->ccw_device); | |
1da177e4 | 48 | |
00bab910 SS |
49 | sbal_req = adapter->req_q.sbal; |
50 | sbal_resp = adapter->resp_q.sbal; | |
1da177e4 | 51 | |
00bab910 SS |
52 | for (p = 0; p < QDIO_MAX_BUFFERS_PER_Q; p += QBUFF_PER_PAGE) { |
53 | free_page((unsigned long) sbal_req[p]); | |
54 | free_page((unsigned long) sbal_resp[p]); | |
55 | } | |
1da177e4 LT |
56 | } |
57 | ||
00bab910 | 58 | static void zfcp_qdio_handler_error(struct zfcp_adapter *adapter, u8 id) |
1da177e4 | 59 | { |
00bab910 | 60 | dev_warn(&adapter->ccw_device->dev, "QDIO problem occurred.\n"); |
1da177e4 | 61 | |
00bab910 SS |
62 | zfcp_erp_adapter_reopen(adapter, |
63 | ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED | | |
64 | ZFCP_STATUS_COMMON_ERP_FAILED, id, NULL); | |
1da177e4 LT |
65 | } |
66 | ||
5d4e2262 CS |
67 | static void zfcp_qdio_zero_sbals(struct qdio_buffer *sbal[], int first, int cnt) |
68 | { | |
69 | int i, sbal_idx; | |
70 | ||
71 | for (i = first; i < first + cnt; i++) { | |
72 | sbal_idx = i % QDIO_MAX_BUFFERS_PER_Q; | |
73 | memset(sbal[sbal_idx], 0, sizeof(struct qdio_buffer)); | |
74 | } | |
75 | } | |
76 | ||
779e6e1c JG |
77 | static void zfcp_qdio_int_req(struct ccw_device *cdev, unsigned int qdio_err, |
78 | int queue_no, int first, int count, | |
00bab910 | 79 | unsigned long parm) |
1da177e4 | 80 | { |
00bab910 SS |
81 | struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm; |
82 | struct zfcp_qdio_queue *queue = &adapter->req_q; | |
1da177e4 | 83 | |
779e6e1c JG |
84 | if (unlikely(qdio_err)) { |
85 | zfcp_hba_dbf_event_qdio(adapter, qdio_err, first, count); | |
00bab910 SS |
86 | zfcp_qdio_handler_error(adapter, 140); |
87 | return; | |
88 | } | |
1da177e4 LT |
89 | |
90 | /* cleanup all SBALs being program-owned now */ | |
00bab910 | 91 | zfcp_qdio_zero_sbals(queue->sbal, first, count); |
1da177e4 | 92 | |
00bab910 | 93 | atomic_add(count, &queue->count); |
1da177e4 | 94 | wake_up(&adapter->request_wq); |
1da177e4 LT |
95 | } |
96 | ||
b03670e5 | 97 | static void zfcp_qdio_reqid_check(struct zfcp_adapter *adapter, |
00bab910 | 98 | unsigned long req_id, int sbal_idx) |
fea9d6c7 VS |
99 | { |
100 | struct zfcp_fsf_req *fsf_req; | |
101 | unsigned long flags; | |
102 | ||
fea9d6c7 | 103 | spin_lock_irqsave(&adapter->req_list_lock, flags); |
ca2d02c2 | 104 | fsf_req = zfcp_reqlist_find(adapter, req_id); |
fea9d6c7 | 105 | |
ca2d02c2 HC |
106 | if (!fsf_req) |
107 | /* | |
108 | * Unknown request means that we have potentially memory | |
109 | * corruption and must stop the machine immediatly. | |
110 | */ | |
00bab910 | 111 | panic("error: unknown request id (%lx) on adapter %s.\n", |
ca2d02c2 | 112 | req_id, zfcp_get_busid_by_adapter(adapter)); |
fea9d6c7 | 113 | |
ca2d02c2 | 114 | zfcp_reqlist_remove(adapter, fsf_req); |
fea9d6c7 VS |
115 | spin_unlock_irqrestore(&adapter->req_list_lock, flags); |
116 | ||
00bab910 | 117 | fsf_req->sbal_response = sbal_idx; |
fea9d6c7 | 118 | zfcp_fsf_req_complete(fsf_req); |
fea9d6c7 VS |
119 | } |
120 | ||
00bab910 | 121 | static void zfcp_qdio_resp_put_back(struct zfcp_adapter *adapter, int processed) |
1da177e4 | 122 | { |
00bab910 SS |
123 | struct zfcp_qdio_queue *queue = &adapter->resp_q; |
124 | struct ccw_device *cdev = adapter->ccw_device; | |
125 | u8 count, start = queue->first; | |
126 | unsigned int retval; | |
1da177e4 | 127 | |
00bab910 SS |
128 | count = atomic_read(&queue->count) + processed; |
129 | ||
779e6e1c | 130 | retval = do_QDIO(cdev, QDIO_FLAG_SYNC_INPUT, 0, start, count); |
00bab910 SS |
131 | |
132 | if (unlikely(retval)) { | |
133 | atomic_set(&queue->count, count); | |
134 | /* FIXME: Recover this with an adapter reopen? */ | |
135 | } else { | |
136 | queue->first += count; | |
137 | queue->first %= QDIO_MAX_BUFFERS_PER_Q; | |
138 | atomic_set(&queue->count, 0); | |
139 | } | |
140 | } | |
141 | ||
779e6e1c JG |
142 | static void zfcp_qdio_int_resp(struct ccw_device *cdev, unsigned int qdio_err, |
143 | int queue_no, int first, int count, | |
00bab910 SS |
144 | unsigned long parm) |
145 | { | |
146 | struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm; | |
147 | struct zfcp_qdio_queue *queue = &adapter->resp_q; | |
148 | volatile struct qdio_buffer_element *sbale; | |
149 | int sbal_idx, sbale_idx, sbal_no; | |
150 | ||
779e6e1c JG |
151 | if (unlikely(qdio_err)) { |
152 | zfcp_hba_dbf_event_qdio(adapter, qdio_err, first, count); | |
00bab910 SS |
153 | zfcp_qdio_handler_error(adapter, 147); |
154 | return; | |
155 | } | |
1da177e4 | 156 | |
1da177e4 LT |
157 | /* |
158 | * go through all SBALs from input queue currently | |
159 | * returned by QDIO layer | |
160 | */ | |
00bab910 SS |
161 | for (sbal_no = 0; sbal_no < count; sbal_no++) { |
162 | sbal_idx = (first + sbal_no) % QDIO_MAX_BUFFERS_PER_Q; | |
1da177e4 LT |
163 | |
164 | /* go through all SBALEs of SBAL */ | |
00bab910 SS |
165 | for (sbale_idx = 0; sbale_idx < QDIO_MAX_ELEMENTS_PER_BUFFER; |
166 | sbale_idx++) { | |
167 | sbale = zfcp_qdio_sbale(queue, sbal_idx, sbale_idx); | |
b03670e5 | 168 | zfcp_qdio_reqid_check(adapter, |
00bab910 SS |
169 | (unsigned long) sbale->addr, |
170 | sbal_idx); | |
171 | if (likely(sbale->flags & SBAL_FLAGS_LAST_ENTRY)) | |
1da177e4 LT |
172 | break; |
173 | }; | |
174 | ||
00bab910 SS |
175 | if (unlikely(!(sbale->flags & SBAL_FLAGS_LAST_ENTRY))) |
176 | dev_warn(&adapter->ccw_device->dev, | |
177 | "Protocol violation by adapter. " | |
178 | "Continuing operations.\n"); | |
1da177e4 LT |
179 | } |
180 | ||
181 | /* | |
182 | * put range of SBALs back to response queue | |
183 | * (including SBALs which have already been free before) | |
184 | */ | |
00bab910 | 185 | zfcp_qdio_resp_put_back(adapter, count); |
1da177e4 LT |
186 | } |
187 | ||
1da177e4 | 188 | /** |
00bab910 SS |
189 | * zfcp_qdio_sbale_req - return ptr to SBALE of req_q for a struct zfcp_fsf_req |
190 | * @fsf_req: pointer to struct fsf_req | |
191 | * Returns: pointer to qdio_buffer_element (SBALE) structure | |
1da177e4 | 192 | */ |
4d284cac | 193 | volatile struct qdio_buffer_element * |
00bab910 | 194 | zfcp_qdio_sbale_req(struct zfcp_fsf_req *req) |
1da177e4 | 195 | { |
00bab910 | 196 | return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last, 0); |
1da177e4 LT |
197 | } |
198 | ||
199 | /** | |
00bab910 SS |
200 | * zfcp_qdio_sbale_curr - return curr SBALE on req_q for a struct zfcp_fsf_req |
201 | * @fsf_req: pointer to struct fsf_req | |
202 | * Returns: pointer to qdio_buffer_element (SBALE) structure | |
1da177e4 | 203 | */ |
4d284cac | 204 | volatile struct qdio_buffer_element * |
00bab910 | 205 | zfcp_qdio_sbale_curr(struct zfcp_fsf_req *req) |
1da177e4 | 206 | { |
00bab910 SS |
207 | return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last, |
208 | req->sbale_curr); | |
1da177e4 LT |
209 | } |
210 | ||
00bab910 | 211 | static void zfcp_qdio_sbal_limit(struct zfcp_fsf_req *fsf_req, int max_sbals) |
1da177e4 | 212 | { |
00bab910 | 213 | int count = atomic_read(&fsf_req->adapter->req_q.count); |
1da177e4 | 214 | count = min(count, max_sbals); |
00bab910 SS |
215 | fsf_req->sbal_limit = (fsf_req->sbal_first + count - 1) |
216 | % QDIO_MAX_BUFFERS_PER_Q; | |
1da177e4 LT |
217 | } |
218 | ||
4d284cac | 219 | static volatile struct qdio_buffer_element * |
1da177e4 LT |
220 | zfcp_qdio_sbal_chain(struct zfcp_fsf_req *fsf_req, unsigned long sbtype) |
221 | { | |
222 | volatile struct qdio_buffer_element *sbale; | |
223 | ||
224 | /* set last entry flag in current SBALE of current SBAL */ | |
225 | sbale = zfcp_qdio_sbale_curr(fsf_req); | |
226 | sbale->flags |= SBAL_FLAGS_LAST_ENTRY; | |
227 | ||
228 | /* don't exceed last allowed SBAL */ | |
e891bffe | 229 | if (fsf_req->sbal_last == fsf_req->sbal_limit) |
1da177e4 LT |
230 | return NULL; |
231 | ||
232 | /* set chaining flag in first SBALE of current SBAL */ | |
00bab910 | 233 | sbale = zfcp_qdio_sbale_req(fsf_req); |
1da177e4 LT |
234 | sbale->flags |= SBAL_FLAGS0_MORE_SBALS; |
235 | ||
236 | /* calculate index of next SBAL */ | |
e891bffe MP |
237 | fsf_req->sbal_last++; |
238 | fsf_req->sbal_last %= QDIO_MAX_BUFFERS_PER_Q; | |
1da177e4 LT |
239 | |
240 | /* keep this requests number of SBALs up-to-date */ | |
241 | fsf_req->sbal_number++; | |
242 | ||
243 | /* start at first SBALE of new SBAL */ | |
244 | fsf_req->sbale_curr = 0; | |
245 | ||
246 | /* set storage-block type for new SBAL */ | |
247 | sbale = zfcp_qdio_sbale_curr(fsf_req); | |
248 | sbale->flags |= sbtype; | |
249 | ||
250 | return sbale; | |
251 | } | |
252 | ||
4d284cac | 253 | static volatile struct qdio_buffer_element * |
1da177e4 LT |
254 | zfcp_qdio_sbale_next(struct zfcp_fsf_req *fsf_req, unsigned long sbtype) |
255 | { | |
256 | if (fsf_req->sbale_curr == ZFCP_LAST_SBALE_PER_SBAL) | |
257 | return zfcp_qdio_sbal_chain(fsf_req, sbtype); | |
1da177e4 | 258 | fsf_req->sbale_curr++; |
1da177e4 LT |
259 | return zfcp_qdio_sbale_curr(fsf_req); |
260 | } | |
261 | ||
00bab910 | 262 | static void zfcp_qdio_undo_sbals(struct zfcp_fsf_req *fsf_req) |
1da177e4 | 263 | { |
00bab910 SS |
264 | struct qdio_buffer **sbal = fsf_req->adapter->req_q.sbal; |
265 | int first = fsf_req->sbal_first; | |
266 | int last = fsf_req->sbal_last; | |
267 | int count = (last - first + QDIO_MAX_BUFFERS_PER_Q) % | |
268 | QDIO_MAX_BUFFERS_PER_Q + 1; | |
269 | zfcp_qdio_zero_sbals(sbal, first, count); | |
1da177e4 LT |
270 | } |
271 | ||
00bab910 SS |
272 | static int zfcp_qdio_fill_sbals(struct zfcp_fsf_req *fsf_req, |
273 | unsigned int sbtype, void *start_addr, | |
274 | unsigned int total_length) | |
1da177e4 LT |
275 | { |
276 | volatile struct qdio_buffer_element *sbale; | |
1da177e4 LT |
277 | unsigned long remaining, length; |
278 | void *addr; | |
279 | ||
00bab910 | 280 | /* split segment up */ |
1da177e4 LT |
281 | for (addr = start_addr, remaining = total_length; remaining > 0; |
282 | addr += length, remaining -= length) { | |
00bab910 SS |
283 | sbale = zfcp_qdio_sbale_next(fsf_req, sbtype); |
284 | if (!sbale) { | |
2450d3e7 | 285 | atomic_inc(&fsf_req->adapter->qdio_outb_full); |
00bab910 | 286 | zfcp_qdio_undo_sbals(fsf_req); |
1da177e4 LT |
287 | return -EINVAL; |
288 | } | |
00bab910 SS |
289 | |
290 | /* new piece must not exceed next page boundary */ | |
1da177e4 | 291 | length = min(remaining, |
00bab910 | 292 | (PAGE_SIZE - ((unsigned long)addr & |
1da177e4 | 293 | (PAGE_SIZE - 1)))); |
00bab910 SS |
294 | sbale->addr = addr; |
295 | sbale->length = length; | |
1da177e4 | 296 | } |
00bab910 | 297 | return 0; |
1da177e4 LT |
298 | } |
299 | ||
1da177e4 LT |
300 | /** |
301 | * zfcp_qdio_sbals_from_sg - fill SBALs from scatter-gather list | |
302 | * @fsf_req: request to be processed | |
303 | * @sbtype: SBALE flags | |
304 | * @sg: scatter-gather list | |
1da177e4 | 305 | * @max_sbals: upper bound for number of SBALs to be used |
00bab910 | 306 | * Returns: number of bytes, or error (negativ) |
1da177e4 | 307 | */ |
00bab910 SS |
308 | int zfcp_qdio_sbals_from_sg(struct zfcp_fsf_req *fsf_req, unsigned long sbtype, |
309 | struct scatterlist *sg, int max_sbals) | |
1da177e4 | 310 | { |
1da177e4 | 311 | volatile struct qdio_buffer_element *sbale; |
00bab910 | 312 | int retval, bytes = 0; |
1da177e4 LT |
313 | |
314 | /* figure out last allowed SBAL */ | |
315 | zfcp_qdio_sbal_limit(fsf_req, max_sbals); | |
316 | ||
00bab910 SS |
317 | /* set storage-block type for this request */ |
318 | sbale = zfcp_qdio_sbale_req(fsf_req); | |
1da177e4 LT |
319 | sbale->flags |= sbtype; |
320 | ||
00bab910 SS |
321 | for (; sg; sg = sg_next(sg)) { |
322 | retval = zfcp_qdio_fill_sbals(fsf_req, sbtype, sg_virt(sg), | |
323 | sg->length); | |
324 | if (retval < 0) | |
325 | return retval; | |
326 | bytes += sg->length; | |
1da177e4 | 327 | } |
00bab910 | 328 | |
1da177e4 LT |
329 | /* assume that no other SBALEs are to follow in the same SBAL */ |
330 | sbale = zfcp_qdio_sbale_curr(fsf_req); | |
331 | sbale->flags |= SBAL_FLAGS_LAST_ENTRY; | |
00bab910 | 332 | |
1da177e4 LT |
333 | return bytes; |
334 | } | |
335 | ||
1da177e4 | 336 | /** |
00bab910 SS |
337 | * zfcp_qdio_send - set PCI flag in first SBALE and send req to QDIO |
338 | * @fsf_req: pointer to struct zfcp_fsf_req | |
339 | * Returns: 0 on success, error otherwise | |
1da177e4 | 340 | */ |
00bab910 | 341 | int zfcp_qdio_send(struct zfcp_fsf_req *fsf_req) |
1da177e4 | 342 | { |
00bab910 SS |
343 | struct zfcp_adapter *adapter = fsf_req->adapter; |
344 | struct zfcp_qdio_queue *req_q = &adapter->req_q; | |
345 | int first = fsf_req->sbal_first; | |
346 | int count = fsf_req->sbal_number; | |
347 | int retval, pci, pci_batch; | |
348 | volatile struct qdio_buffer_element *sbale; | |
349 | ||
350 | /* acknowledgements for transferred buffers */ | |
351 | pci_batch = req_q->pci_batch + count; | |
352 | if (unlikely(pci_batch >= ZFCP_QDIO_PCI_INTERVAL)) { | |
353 | pci_batch %= ZFCP_QDIO_PCI_INTERVAL; | |
354 | pci = first + count - (pci_batch + 1); | |
355 | pci %= QDIO_MAX_BUFFERS_PER_Q; | |
356 | sbale = zfcp_qdio_sbale(req_q, pci, 0); | |
357 | sbale->flags |= SBAL_FLAGS0_PCI; | |
358 | } | |
359 | ||
360 | retval = do_QDIO(adapter->ccw_device, QDIO_FLAG_SYNC_OUTPUT, 0, first, | |
779e6e1c | 361 | count); |
00bab910 SS |
362 | if (unlikely(retval)) { |
363 | zfcp_qdio_zero_sbals(req_q->sbal, first, count); | |
364 | return retval; | |
365 | } | |
366 | ||
367 | /* account for transferred buffers */ | |
368 | atomic_sub(count, &req_q->count); | |
369 | req_q->first += count; | |
370 | req_q->first %= QDIO_MAX_BUFFERS_PER_Q; | |
371 | req_q->pci_batch = pci_batch; | |
372 | return 0; | |
1da177e4 LT |
373 | } |
374 | ||
00bab910 SS |
375 | /** |
376 | * zfcp_qdio_allocate - allocate queue memory and initialize QDIO data | |
377 | * @adapter: pointer to struct zfcp_adapter | |
378 | * Returns: -ENOMEM on memory allocation error or return value from | |
379 | * qdio_allocate | |
380 | */ | |
381 | int zfcp_qdio_allocate(struct zfcp_adapter *adapter) | |
382 | { | |
383 | struct qdio_initialize *init_data; | |
384 | ||
385 | if (zfcp_qdio_buffers_enqueue(adapter->req_q.sbal) || | |
386 | zfcp_qdio_buffers_enqueue(adapter->resp_q.sbal)) | |
387 | return -ENOMEM; | |
388 | ||
389 | init_data = &adapter->qdio_init_data; | |
390 | ||
391 | init_data->cdev = adapter->ccw_device; | |
392 | init_data->q_format = QDIO_ZFCP_QFMT; | |
393 | memcpy(init_data->adapter_name, zfcp_get_busid_by_adapter(adapter), 8); | |
394 | ASCEBC(init_data->adapter_name, 8); | |
395 | init_data->qib_param_field_format = 0; | |
396 | init_data->qib_param_field = NULL; | |
397 | init_data->input_slib_elements = NULL; | |
398 | init_data->output_slib_elements = NULL; | |
00bab910 SS |
399 | init_data->no_input_qs = 1; |
400 | init_data->no_output_qs = 1; | |
401 | init_data->input_handler = zfcp_qdio_int_resp; | |
402 | init_data->output_handler = zfcp_qdio_int_req; | |
403 | init_data->int_parm = (unsigned long) adapter; | |
404 | init_data->flags = QDIO_INBOUND_0COPY_SBALS | | |
405 | QDIO_OUTBOUND_0COPY_SBALS | QDIO_USE_OUTBOUND_PCIS; | |
406 | init_data->input_sbal_addr_array = | |
407 | (void **) (adapter->resp_q.sbal); | |
408 | init_data->output_sbal_addr_array = | |
409 | (void **) (adapter->req_q.sbal); | |
410 | ||
411 | return qdio_allocate(init_data); | |
412 | } | |
413 | ||
414 | /** | |
415 | * zfcp_close_qdio - close qdio queues for an adapter | |
1da177e4 | 416 | */ |
00bab910 | 417 | void zfcp_qdio_close(struct zfcp_adapter *adapter) |
1da177e4 | 418 | { |
00bab910 SS |
419 | struct zfcp_qdio_queue *req_q; |
420 | int first, count; | |
421 | ||
422 | if (!atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status)) | |
423 | return; | |
424 | ||
425 | /* clear QDIOUP flag, thus do_QDIO is not called during qdio_shutdown */ | |
426 | req_q = &adapter->req_q; | |
d4538817 | 427 | spin_lock_bh(&req_q->lock); |
00bab910 | 428 | atomic_clear_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status); |
d4538817 | 429 | spin_unlock_bh(&req_q->lock); |
00bab910 | 430 | |
779e6e1c | 431 | qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR); |
00bab910 SS |
432 | |
433 | /* cleanup used outbound sbals */ | |
434 | count = atomic_read(&req_q->count); | |
435 | if (count < QDIO_MAX_BUFFERS_PER_Q) { | |
436 | first = (req_q->first + count) % QDIO_MAX_BUFFERS_PER_Q; | |
437 | count = QDIO_MAX_BUFFERS_PER_Q - count; | |
438 | zfcp_qdio_zero_sbals(req_q->sbal, first, count); | |
1da177e4 | 439 | } |
00bab910 SS |
440 | req_q->first = 0; |
441 | atomic_set(&req_q->count, 0); | |
442 | req_q->pci_batch = 0; | |
443 | adapter->resp_q.first = 0; | |
444 | atomic_set(&adapter->resp_q.count, 0); | |
1da177e4 LT |
445 | } |
446 | ||
00bab910 SS |
447 | /** |
448 | * zfcp_qdio_open - prepare and initialize response queue | |
449 | * @adapter: pointer to struct zfcp_adapter | |
450 | * Returns: 0 on success, otherwise -EIO | |
451 | */ | |
452 | int zfcp_qdio_open(struct zfcp_adapter *adapter) | |
453 | { | |
454 | volatile struct qdio_buffer_element *sbale; | |
455 | int cc; | |
456 | ||
457 | if (atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status)) | |
458 | return -EIO; | |
459 | ||
460 | if (qdio_establish(&adapter->qdio_init_data)) { | |
461 | dev_err(&adapter->ccw_device->dev, | |
462 | "Establish of QDIO queues failed.\n"); | |
463 | return -EIO; | |
464 | } | |
465 | ||
779e6e1c | 466 | if (qdio_activate(adapter->ccw_device)) { |
00bab910 SS |
467 | dev_err(&adapter->ccw_device->dev, |
468 | "Activate of QDIO queues failed.\n"); | |
469 | goto failed_qdio; | |
470 | } | |
471 | ||
472 | for (cc = 0; cc < QDIO_MAX_BUFFERS_PER_Q; cc++) { | |
473 | sbale = &(adapter->resp_q.sbal[cc]->element[0]); | |
474 | sbale->length = 0; | |
475 | sbale->flags = SBAL_FLAGS_LAST_ENTRY; | |
476 | sbale->addr = NULL; | |
477 | } | |
478 | ||
479 | if (do_QDIO(adapter->ccw_device, QDIO_FLAG_SYNC_INPUT, 0, 0, | |
779e6e1c | 480 | QDIO_MAX_BUFFERS_PER_Q)) { |
00bab910 SS |
481 | dev_err(&adapter->ccw_device->dev, |
482 | "Init of QDIO response queue failed.\n"); | |
483 | goto failed_qdio; | |
484 | } | |
485 | ||
486 | /* set index of first avalable SBALS / number of available SBALS */ | |
487 | adapter->req_q.first = 0; | |
488 | atomic_set(&adapter->req_q.count, QDIO_MAX_BUFFERS_PER_Q); | |
489 | adapter->req_q.pci_batch = 0; | |
490 | ||
491 | return 0; | |
492 | ||
493 | failed_qdio: | |
779e6e1c | 494 | qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR); |
00bab910 SS |
495 | return -EIO; |
496 | } |