]> git.proxmox.com Git - mirror_ubuntu-kernels.git/blame - drivers/s390/scsi/zfcp_qdio.c
[SCSI] zfcp: add queue_full sysfs attribute
[mirror_ubuntu-kernels.git] / drivers / s390 / scsi / zfcp_qdio.c
CommitLineData
1da177e4 1/*
00bab910 2 * zfcp device driver
1da177e4 3 *
00bab910 4 * Setup and helper functions to access QDIO.
1da177e4 5 *
00bab910 6 * Copyright IBM Corporation 2002, 2008
1da177e4
LT
7 */
8
1da177e4
LT
9#include "zfcp_ext.h"
10
00bab910
SS
11/* FIXME(tune): free space should be one max. SBAL chain plus what? */
12#define ZFCP_QDIO_PCI_INTERVAL (QDIO_MAX_BUFFERS_PER_Q \
c41f8cbd 13 - (FSF_MAX_SBALS_PER_REQ + 4))
5d4e2262 14#define QBUFF_PER_PAGE (PAGE_SIZE / sizeof(struct qdio_buffer))
1da177e4 15
00bab910 16static int zfcp_qdio_buffers_enqueue(struct qdio_buffer **sbal)
1da177e4 17{
b4e44590 18 int pos;
1da177e4 19
b4e44590 20 for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos += QBUFF_PER_PAGE) {
00bab910
SS
21 sbal[pos] = (struct qdio_buffer *) get_zeroed_page(GFP_KERNEL);
22 if (!sbal[pos])
b4e44590 23 return -ENOMEM;
b4e44590
SS
24 }
25 for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos++)
26 if (pos % QBUFF_PER_PAGE)
00bab910 27 sbal[pos] = sbal[pos - 1] + 1;
b4e44590 28 return 0;
1da177e4
LT
29}
30
00bab910
SS
31static volatile struct qdio_buffer_element *
32zfcp_qdio_sbale(struct zfcp_qdio_queue *q, int sbal_idx, int sbale_idx)
1da177e4 33{
00bab910 34 return &q->sbal[sbal_idx]->element[sbale_idx];
1da177e4
LT
35}
36
00bab910
SS
37/**
38 * zfcp_qdio_free - free memory used by request- and resposne queue
39 * @adapter: pointer to the zfcp_adapter structure
40 */
41void zfcp_qdio_free(struct zfcp_adapter *adapter)
1da177e4 42{
00bab910
SS
43 struct qdio_buffer **sbal_req, **sbal_resp;
44 int p;
1da177e4 45
00bab910
SS
46 if (adapter->ccw_device)
47 qdio_free(adapter->ccw_device);
1da177e4 48
00bab910
SS
49 sbal_req = adapter->req_q.sbal;
50 sbal_resp = adapter->resp_q.sbal;
1da177e4 51
00bab910
SS
52 for (p = 0; p < QDIO_MAX_BUFFERS_PER_Q; p += QBUFF_PER_PAGE) {
53 free_page((unsigned long) sbal_req[p]);
54 free_page((unsigned long) sbal_resp[p]);
55 }
1da177e4
LT
56}
57
00bab910 58static void zfcp_qdio_handler_error(struct zfcp_adapter *adapter, u8 id)
1da177e4 59{
00bab910 60 dev_warn(&adapter->ccw_device->dev, "QDIO problem occurred.\n");
1da177e4 61
00bab910
SS
62 zfcp_erp_adapter_reopen(adapter,
63 ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED |
64 ZFCP_STATUS_COMMON_ERP_FAILED, id, NULL);
1da177e4
LT
65}
66
5d4e2262
CS
67static void zfcp_qdio_zero_sbals(struct qdio_buffer *sbal[], int first, int cnt)
68{
69 int i, sbal_idx;
70
71 for (i = first; i < first + cnt; i++) {
72 sbal_idx = i % QDIO_MAX_BUFFERS_PER_Q;
73 memset(sbal[sbal_idx], 0, sizeof(struct qdio_buffer));
74 }
75}
76
779e6e1c
JG
77static void zfcp_qdio_int_req(struct ccw_device *cdev, unsigned int qdio_err,
78 int queue_no, int first, int count,
00bab910 79 unsigned long parm)
1da177e4 80{
00bab910
SS
81 struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm;
82 struct zfcp_qdio_queue *queue = &adapter->req_q;
1da177e4 83
779e6e1c
JG
84 if (unlikely(qdio_err)) {
85 zfcp_hba_dbf_event_qdio(adapter, qdio_err, first, count);
00bab910
SS
86 zfcp_qdio_handler_error(adapter, 140);
87 return;
88 }
1da177e4
LT
89
90 /* cleanup all SBALs being program-owned now */
00bab910 91 zfcp_qdio_zero_sbals(queue->sbal, first, count);
1da177e4 92
00bab910 93 atomic_add(count, &queue->count);
1da177e4 94 wake_up(&adapter->request_wq);
1da177e4
LT
95}
96
b03670e5 97static void zfcp_qdio_reqid_check(struct zfcp_adapter *adapter,
00bab910 98 unsigned long req_id, int sbal_idx)
fea9d6c7
VS
99{
100 struct zfcp_fsf_req *fsf_req;
101 unsigned long flags;
102
fea9d6c7 103 spin_lock_irqsave(&adapter->req_list_lock, flags);
ca2d02c2 104 fsf_req = zfcp_reqlist_find(adapter, req_id);
fea9d6c7 105
ca2d02c2
HC
106 if (!fsf_req)
107 /*
108 * Unknown request means that we have potentially memory
109 * corruption and must stop the machine immediatly.
110 */
00bab910 111 panic("error: unknown request id (%lx) on adapter %s.\n",
ca2d02c2 112 req_id, zfcp_get_busid_by_adapter(adapter));
fea9d6c7 113
ca2d02c2 114 zfcp_reqlist_remove(adapter, fsf_req);
fea9d6c7
VS
115 spin_unlock_irqrestore(&adapter->req_list_lock, flags);
116
00bab910 117 fsf_req->sbal_response = sbal_idx;
fea9d6c7 118 zfcp_fsf_req_complete(fsf_req);
fea9d6c7
VS
119}
120
00bab910 121static void zfcp_qdio_resp_put_back(struct zfcp_adapter *adapter, int processed)
1da177e4 122{
00bab910
SS
123 struct zfcp_qdio_queue *queue = &adapter->resp_q;
124 struct ccw_device *cdev = adapter->ccw_device;
125 u8 count, start = queue->first;
126 unsigned int retval;
1da177e4 127
00bab910
SS
128 count = atomic_read(&queue->count) + processed;
129
779e6e1c 130 retval = do_QDIO(cdev, QDIO_FLAG_SYNC_INPUT, 0, start, count);
00bab910
SS
131
132 if (unlikely(retval)) {
133 atomic_set(&queue->count, count);
134 /* FIXME: Recover this with an adapter reopen? */
135 } else {
136 queue->first += count;
137 queue->first %= QDIO_MAX_BUFFERS_PER_Q;
138 atomic_set(&queue->count, 0);
139 }
140}
141
779e6e1c
JG
142static void zfcp_qdio_int_resp(struct ccw_device *cdev, unsigned int qdio_err,
143 int queue_no, int first, int count,
00bab910
SS
144 unsigned long parm)
145{
146 struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm;
147 struct zfcp_qdio_queue *queue = &adapter->resp_q;
148 volatile struct qdio_buffer_element *sbale;
149 int sbal_idx, sbale_idx, sbal_no;
150
779e6e1c
JG
151 if (unlikely(qdio_err)) {
152 zfcp_hba_dbf_event_qdio(adapter, qdio_err, first, count);
00bab910
SS
153 zfcp_qdio_handler_error(adapter, 147);
154 return;
155 }
1da177e4 156
1da177e4
LT
157 /*
158 * go through all SBALs from input queue currently
159 * returned by QDIO layer
160 */
00bab910
SS
161 for (sbal_no = 0; sbal_no < count; sbal_no++) {
162 sbal_idx = (first + sbal_no) % QDIO_MAX_BUFFERS_PER_Q;
1da177e4
LT
163
164 /* go through all SBALEs of SBAL */
00bab910
SS
165 for (sbale_idx = 0; sbale_idx < QDIO_MAX_ELEMENTS_PER_BUFFER;
166 sbale_idx++) {
167 sbale = zfcp_qdio_sbale(queue, sbal_idx, sbale_idx);
b03670e5 168 zfcp_qdio_reqid_check(adapter,
00bab910
SS
169 (unsigned long) sbale->addr,
170 sbal_idx);
171 if (likely(sbale->flags & SBAL_FLAGS_LAST_ENTRY))
1da177e4
LT
172 break;
173 };
174
00bab910
SS
175 if (unlikely(!(sbale->flags & SBAL_FLAGS_LAST_ENTRY)))
176 dev_warn(&adapter->ccw_device->dev,
177 "Protocol violation by adapter. "
178 "Continuing operations.\n");
1da177e4
LT
179 }
180
181 /*
182 * put range of SBALs back to response queue
183 * (including SBALs which have already been free before)
184 */
00bab910 185 zfcp_qdio_resp_put_back(adapter, count);
1da177e4
LT
186}
187
1da177e4 188/**
00bab910
SS
189 * zfcp_qdio_sbale_req - return ptr to SBALE of req_q for a struct zfcp_fsf_req
190 * @fsf_req: pointer to struct fsf_req
191 * Returns: pointer to qdio_buffer_element (SBALE) structure
1da177e4 192 */
4d284cac 193volatile struct qdio_buffer_element *
00bab910 194zfcp_qdio_sbale_req(struct zfcp_fsf_req *req)
1da177e4 195{
00bab910 196 return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last, 0);
1da177e4
LT
197}
198
199/**
00bab910
SS
200 * zfcp_qdio_sbale_curr - return curr SBALE on req_q for a struct zfcp_fsf_req
201 * @fsf_req: pointer to struct fsf_req
202 * Returns: pointer to qdio_buffer_element (SBALE) structure
1da177e4 203 */
4d284cac 204volatile struct qdio_buffer_element *
00bab910 205zfcp_qdio_sbale_curr(struct zfcp_fsf_req *req)
1da177e4 206{
00bab910
SS
207 return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last,
208 req->sbale_curr);
1da177e4
LT
209}
210
00bab910 211static void zfcp_qdio_sbal_limit(struct zfcp_fsf_req *fsf_req, int max_sbals)
1da177e4 212{
00bab910 213 int count = atomic_read(&fsf_req->adapter->req_q.count);
1da177e4 214 count = min(count, max_sbals);
00bab910
SS
215 fsf_req->sbal_limit = (fsf_req->sbal_first + count - 1)
216 % QDIO_MAX_BUFFERS_PER_Q;
1da177e4
LT
217}
218
4d284cac 219static volatile struct qdio_buffer_element *
1da177e4
LT
220zfcp_qdio_sbal_chain(struct zfcp_fsf_req *fsf_req, unsigned long sbtype)
221{
222 volatile struct qdio_buffer_element *sbale;
223
224 /* set last entry flag in current SBALE of current SBAL */
225 sbale = zfcp_qdio_sbale_curr(fsf_req);
226 sbale->flags |= SBAL_FLAGS_LAST_ENTRY;
227
228 /* don't exceed last allowed SBAL */
e891bffe 229 if (fsf_req->sbal_last == fsf_req->sbal_limit)
1da177e4
LT
230 return NULL;
231
232 /* set chaining flag in first SBALE of current SBAL */
00bab910 233 sbale = zfcp_qdio_sbale_req(fsf_req);
1da177e4
LT
234 sbale->flags |= SBAL_FLAGS0_MORE_SBALS;
235
236 /* calculate index of next SBAL */
e891bffe
MP
237 fsf_req->sbal_last++;
238 fsf_req->sbal_last %= QDIO_MAX_BUFFERS_PER_Q;
1da177e4
LT
239
240 /* keep this requests number of SBALs up-to-date */
241 fsf_req->sbal_number++;
242
243 /* start at first SBALE of new SBAL */
244 fsf_req->sbale_curr = 0;
245
246 /* set storage-block type for new SBAL */
247 sbale = zfcp_qdio_sbale_curr(fsf_req);
248 sbale->flags |= sbtype;
249
250 return sbale;
251}
252
4d284cac 253static volatile struct qdio_buffer_element *
1da177e4
LT
254zfcp_qdio_sbale_next(struct zfcp_fsf_req *fsf_req, unsigned long sbtype)
255{
256 if (fsf_req->sbale_curr == ZFCP_LAST_SBALE_PER_SBAL)
257 return zfcp_qdio_sbal_chain(fsf_req, sbtype);
1da177e4 258 fsf_req->sbale_curr++;
1da177e4
LT
259 return zfcp_qdio_sbale_curr(fsf_req);
260}
261
00bab910 262static void zfcp_qdio_undo_sbals(struct zfcp_fsf_req *fsf_req)
1da177e4 263{
00bab910
SS
264 struct qdio_buffer **sbal = fsf_req->adapter->req_q.sbal;
265 int first = fsf_req->sbal_first;
266 int last = fsf_req->sbal_last;
267 int count = (last - first + QDIO_MAX_BUFFERS_PER_Q) %
268 QDIO_MAX_BUFFERS_PER_Q + 1;
269 zfcp_qdio_zero_sbals(sbal, first, count);
1da177e4
LT
270}
271
00bab910
SS
272static int zfcp_qdio_fill_sbals(struct zfcp_fsf_req *fsf_req,
273 unsigned int sbtype, void *start_addr,
274 unsigned int total_length)
1da177e4
LT
275{
276 volatile struct qdio_buffer_element *sbale;
1da177e4
LT
277 unsigned long remaining, length;
278 void *addr;
279
00bab910 280 /* split segment up */
1da177e4
LT
281 for (addr = start_addr, remaining = total_length; remaining > 0;
282 addr += length, remaining -= length) {
00bab910
SS
283 sbale = zfcp_qdio_sbale_next(fsf_req, sbtype);
284 if (!sbale) {
2450d3e7 285 atomic_inc(&fsf_req->adapter->qdio_outb_full);
00bab910 286 zfcp_qdio_undo_sbals(fsf_req);
1da177e4
LT
287 return -EINVAL;
288 }
00bab910
SS
289
290 /* new piece must not exceed next page boundary */
1da177e4 291 length = min(remaining,
00bab910 292 (PAGE_SIZE - ((unsigned long)addr &
1da177e4 293 (PAGE_SIZE - 1))));
00bab910
SS
294 sbale->addr = addr;
295 sbale->length = length;
1da177e4 296 }
00bab910 297 return 0;
1da177e4
LT
298}
299
1da177e4
LT
300/**
301 * zfcp_qdio_sbals_from_sg - fill SBALs from scatter-gather list
302 * @fsf_req: request to be processed
303 * @sbtype: SBALE flags
304 * @sg: scatter-gather list
1da177e4 305 * @max_sbals: upper bound for number of SBALs to be used
00bab910 306 * Returns: number of bytes, or error (negativ)
1da177e4 307 */
00bab910
SS
308int zfcp_qdio_sbals_from_sg(struct zfcp_fsf_req *fsf_req, unsigned long sbtype,
309 struct scatterlist *sg, int max_sbals)
1da177e4 310{
1da177e4 311 volatile struct qdio_buffer_element *sbale;
00bab910 312 int retval, bytes = 0;
1da177e4
LT
313
314 /* figure out last allowed SBAL */
315 zfcp_qdio_sbal_limit(fsf_req, max_sbals);
316
00bab910
SS
317 /* set storage-block type for this request */
318 sbale = zfcp_qdio_sbale_req(fsf_req);
1da177e4
LT
319 sbale->flags |= sbtype;
320
00bab910
SS
321 for (; sg; sg = sg_next(sg)) {
322 retval = zfcp_qdio_fill_sbals(fsf_req, sbtype, sg_virt(sg),
323 sg->length);
324 if (retval < 0)
325 return retval;
326 bytes += sg->length;
1da177e4 327 }
00bab910 328
1da177e4
LT
329 /* assume that no other SBALEs are to follow in the same SBAL */
330 sbale = zfcp_qdio_sbale_curr(fsf_req);
331 sbale->flags |= SBAL_FLAGS_LAST_ENTRY;
00bab910 332
1da177e4
LT
333 return bytes;
334}
335
1da177e4 336/**
00bab910
SS
337 * zfcp_qdio_send - set PCI flag in first SBALE and send req to QDIO
338 * @fsf_req: pointer to struct zfcp_fsf_req
339 * Returns: 0 on success, error otherwise
1da177e4 340 */
00bab910 341int zfcp_qdio_send(struct zfcp_fsf_req *fsf_req)
1da177e4 342{
00bab910
SS
343 struct zfcp_adapter *adapter = fsf_req->adapter;
344 struct zfcp_qdio_queue *req_q = &adapter->req_q;
345 int first = fsf_req->sbal_first;
346 int count = fsf_req->sbal_number;
347 int retval, pci, pci_batch;
348 volatile struct qdio_buffer_element *sbale;
349
350 /* acknowledgements for transferred buffers */
351 pci_batch = req_q->pci_batch + count;
352 if (unlikely(pci_batch >= ZFCP_QDIO_PCI_INTERVAL)) {
353 pci_batch %= ZFCP_QDIO_PCI_INTERVAL;
354 pci = first + count - (pci_batch + 1);
355 pci %= QDIO_MAX_BUFFERS_PER_Q;
356 sbale = zfcp_qdio_sbale(req_q, pci, 0);
357 sbale->flags |= SBAL_FLAGS0_PCI;
358 }
359
360 retval = do_QDIO(adapter->ccw_device, QDIO_FLAG_SYNC_OUTPUT, 0, first,
779e6e1c 361 count);
00bab910
SS
362 if (unlikely(retval)) {
363 zfcp_qdio_zero_sbals(req_q->sbal, first, count);
364 return retval;
365 }
366
367 /* account for transferred buffers */
368 atomic_sub(count, &req_q->count);
369 req_q->first += count;
370 req_q->first %= QDIO_MAX_BUFFERS_PER_Q;
371 req_q->pci_batch = pci_batch;
372 return 0;
1da177e4
LT
373}
374
00bab910
SS
375/**
376 * zfcp_qdio_allocate - allocate queue memory and initialize QDIO data
377 * @adapter: pointer to struct zfcp_adapter
378 * Returns: -ENOMEM on memory allocation error or return value from
379 * qdio_allocate
380 */
381int zfcp_qdio_allocate(struct zfcp_adapter *adapter)
382{
383 struct qdio_initialize *init_data;
384
385 if (zfcp_qdio_buffers_enqueue(adapter->req_q.sbal) ||
386 zfcp_qdio_buffers_enqueue(adapter->resp_q.sbal))
387 return -ENOMEM;
388
389 init_data = &adapter->qdio_init_data;
390
391 init_data->cdev = adapter->ccw_device;
392 init_data->q_format = QDIO_ZFCP_QFMT;
393 memcpy(init_data->adapter_name, zfcp_get_busid_by_adapter(adapter), 8);
394 ASCEBC(init_data->adapter_name, 8);
395 init_data->qib_param_field_format = 0;
396 init_data->qib_param_field = NULL;
397 init_data->input_slib_elements = NULL;
398 init_data->output_slib_elements = NULL;
00bab910
SS
399 init_data->no_input_qs = 1;
400 init_data->no_output_qs = 1;
401 init_data->input_handler = zfcp_qdio_int_resp;
402 init_data->output_handler = zfcp_qdio_int_req;
403 init_data->int_parm = (unsigned long) adapter;
404 init_data->flags = QDIO_INBOUND_0COPY_SBALS |
405 QDIO_OUTBOUND_0COPY_SBALS | QDIO_USE_OUTBOUND_PCIS;
406 init_data->input_sbal_addr_array =
407 (void **) (adapter->resp_q.sbal);
408 init_data->output_sbal_addr_array =
409 (void **) (adapter->req_q.sbal);
410
411 return qdio_allocate(init_data);
412}
413
414/**
415 * zfcp_close_qdio - close qdio queues for an adapter
1da177e4 416 */
00bab910 417void zfcp_qdio_close(struct zfcp_adapter *adapter)
1da177e4 418{
00bab910
SS
419 struct zfcp_qdio_queue *req_q;
420 int first, count;
421
422 if (!atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status))
423 return;
424
425 /* clear QDIOUP flag, thus do_QDIO is not called during qdio_shutdown */
426 req_q = &adapter->req_q;
d4538817 427 spin_lock_bh(&req_q->lock);
00bab910 428 atomic_clear_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status);
d4538817 429 spin_unlock_bh(&req_q->lock);
00bab910 430
779e6e1c 431 qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR);
00bab910
SS
432
433 /* cleanup used outbound sbals */
434 count = atomic_read(&req_q->count);
435 if (count < QDIO_MAX_BUFFERS_PER_Q) {
436 first = (req_q->first + count) % QDIO_MAX_BUFFERS_PER_Q;
437 count = QDIO_MAX_BUFFERS_PER_Q - count;
438 zfcp_qdio_zero_sbals(req_q->sbal, first, count);
1da177e4 439 }
00bab910
SS
440 req_q->first = 0;
441 atomic_set(&req_q->count, 0);
442 req_q->pci_batch = 0;
443 adapter->resp_q.first = 0;
444 atomic_set(&adapter->resp_q.count, 0);
1da177e4
LT
445}
446
00bab910
SS
447/**
448 * zfcp_qdio_open - prepare and initialize response queue
449 * @adapter: pointer to struct zfcp_adapter
450 * Returns: 0 on success, otherwise -EIO
451 */
452int zfcp_qdio_open(struct zfcp_adapter *adapter)
453{
454 volatile struct qdio_buffer_element *sbale;
455 int cc;
456
457 if (atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status))
458 return -EIO;
459
460 if (qdio_establish(&adapter->qdio_init_data)) {
461 dev_err(&adapter->ccw_device->dev,
462 "Establish of QDIO queues failed.\n");
463 return -EIO;
464 }
465
779e6e1c 466 if (qdio_activate(adapter->ccw_device)) {
00bab910
SS
467 dev_err(&adapter->ccw_device->dev,
468 "Activate of QDIO queues failed.\n");
469 goto failed_qdio;
470 }
471
472 for (cc = 0; cc < QDIO_MAX_BUFFERS_PER_Q; cc++) {
473 sbale = &(adapter->resp_q.sbal[cc]->element[0]);
474 sbale->length = 0;
475 sbale->flags = SBAL_FLAGS_LAST_ENTRY;
476 sbale->addr = NULL;
477 }
478
479 if (do_QDIO(adapter->ccw_device, QDIO_FLAG_SYNC_INPUT, 0, 0,
779e6e1c 480 QDIO_MAX_BUFFERS_PER_Q)) {
00bab910
SS
481 dev_err(&adapter->ccw_device->dev,
482 "Init of QDIO response queue failed.\n");
483 goto failed_qdio;
484 }
485
486 /* set index of first avalable SBALS / number of available SBALS */
487 adapter->req_q.first = 0;
488 atomic_set(&adapter->req_q.count, QDIO_MAX_BUFFERS_PER_Q);
489 adapter->req_q.pci_batch = 0;
490
491 return 0;
492
493failed_qdio:
779e6e1c 494 qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR);
00bab910
SS
495 return -EIO;
496}