]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * S/390 common I/O routines -- channel subsystem call | |
3 | * | |
4 | * Copyright IBM Corp. 1999,2012 | |
5 | * Author(s): Ingo Adlung (adlung@de.ibm.com) | |
6 | * Cornelia Huck (cornelia.huck@de.ibm.com) | |
7 | * Arnd Bergmann (arndb@de.ibm.com) | |
8 | */ | |
9 | ||
10 | #define KMSG_COMPONENT "cio" | |
11 | #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt | |
12 | ||
13 | #include <linux/module.h> | |
14 | #include <linux/slab.h> | |
15 | #include <linux/init.h> | |
16 | #include <linux/device.h> | |
17 | #include <linux/mutex.h> | |
18 | #include <linux/pci.h> | |
19 | ||
20 | #include <asm/cio.h> | |
21 | #include <asm/chpid.h> | |
22 | #include <asm/chsc.h> | |
23 | #include <asm/crw.h> | |
24 | #include <asm/isc.h> | |
25 | #include <asm/ebcdic.h> | |
26 | ||
27 | #include "css.h" | |
28 | #include "cio.h" | |
29 | #include "cio_debug.h" | |
30 | #include "ioasm.h" | |
31 | #include "chp.h" | |
32 | #include "chsc.h" | |
33 | ||
34 | static void *sei_page; | |
35 | static void *chsc_page; | |
36 | static DEFINE_SPINLOCK(chsc_page_lock); | |
37 | ||
38 | /** | |
39 | * chsc_error_from_response() - convert a chsc response to an error | |
40 | * @response: chsc response code | |
41 | * | |
42 | * Returns an appropriate Linux error code for @response. | |
43 | */ | |
44 | int chsc_error_from_response(int response) | |
45 | { | |
46 | switch (response) { | |
47 | case 0x0001: | |
48 | return 0; | |
49 | case 0x0002: | |
50 | case 0x0003: | |
51 | case 0x0006: | |
52 | case 0x0007: | |
53 | case 0x0008: | |
54 | case 0x000a: | |
55 | case 0x0104: | |
56 | return -EINVAL; | |
57 | case 0x0004: | |
58 | return -EOPNOTSUPP; | |
59 | case 0x000b: | |
60 | case 0x0107: /* "Channel busy" for the op 0x003d */ | |
61 | return -EBUSY; | |
62 | case 0x0100: | |
63 | case 0x0102: | |
64 | return -ENOMEM; | |
65 | default: | |
66 | return -EIO; | |
67 | } | |
68 | } | |
69 | EXPORT_SYMBOL_GPL(chsc_error_from_response); | |
70 | ||
71 | struct chsc_ssd_area { | |
72 | struct chsc_header request; | |
73 | u16 :10; | |
74 | u16 ssid:2; | |
75 | u16 :4; | |
76 | u16 f_sch; /* first subchannel */ | |
77 | u16 :16; | |
78 | u16 l_sch; /* last subchannel */ | |
79 | u32 :32; | |
80 | struct chsc_header response; | |
81 | u32 :32; | |
82 | u8 sch_valid : 1; | |
83 | u8 dev_valid : 1; | |
84 | u8 st : 3; /* subchannel type */ | |
85 | u8 zeroes : 3; | |
86 | u8 unit_addr; /* unit address */ | |
87 | u16 devno; /* device number */ | |
88 | u8 path_mask; | |
89 | u8 fla_valid_mask; | |
90 | u16 sch; /* subchannel */ | |
91 | u8 chpid[8]; /* chpids 0-7 */ | |
92 | u16 fla[8]; /* full link addresses 0-7 */ | |
93 | } __attribute__ ((packed)); | |
94 | ||
95 | int chsc_get_ssd_info(struct subchannel_id schid, struct chsc_ssd_info *ssd) | |
96 | { | |
97 | struct chsc_ssd_area *ssd_area; | |
98 | int ccode; | |
99 | int ret; | |
100 | int i; | |
101 | int mask; | |
102 | ||
103 | spin_lock_irq(&chsc_page_lock); | |
104 | memset(chsc_page, 0, PAGE_SIZE); | |
105 | ssd_area = chsc_page; | |
106 | ssd_area->request.length = 0x0010; | |
107 | ssd_area->request.code = 0x0004; | |
108 | ssd_area->ssid = schid.ssid; | |
109 | ssd_area->f_sch = schid.sch_no; | |
110 | ssd_area->l_sch = schid.sch_no; | |
111 | ||
112 | ccode = chsc(ssd_area); | |
113 | /* Check response. */ | |
114 | if (ccode > 0) { | |
115 | ret = (ccode == 3) ? -ENODEV : -EBUSY; | |
116 | goto out; | |
117 | } | |
118 | ret = chsc_error_from_response(ssd_area->response.code); | |
119 | if (ret != 0) { | |
120 | CIO_MSG_EVENT(2, "chsc: ssd failed for 0.%x.%04x (rc=%04x)\n", | |
121 | schid.ssid, schid.sch_no, | |
122 | ssd_area->response.code); | |
123 | goto out; | |
124 | } | |
125 | if (!ssd_area->sch_valid) { | |
126 | ret = -ENODEV; | |
127 | goto out; | |
128 | } | |
129 | /* Copy data */ | |
130 | ret = 0; | |
131 | memset(ssd, 0, sizeof(struct chsc_ssd_info)); | |
132 | if ((ssd_area->st != SUBCHANNEL_TYPE_IO) && | |
133 | (ssd_area->st != SUBCHANNEL_TYPE_MSG)) | |
134 | goto out; | |
135 | ssd->path_mask = ssd_area->path_mask; | |
136 | ssd->fla_valid_mask = ssd_area->fla_valid_mask; | |
137 | for (i = 0; i < 8; i++) { | |
138 | mask = 0x80 >> i; | |
139 | if (ssd_area->path_mask & mask) { | |
140 | chp_id_init(&ssd->chpid[i]); | |
141 | ssd->chpid[i].id = ssd_area->chpid[i]; | |
142 | } | |
143 | if (ssd_area->fla_valid_mask & mask) | |
144 | ssd->fla[i] = ssd_area->fla[i]; | |
145 | } | |
146 | out: | |
147 | spin_unlock_irq(&chsc_page_lock); | |
148 | return ret; | |
149 | } | |
150 | ||
151 | /** | |
152 | * chsc_ssqd() - store subchannel QDIO data (SSQD) | |
153 | * @schid: id of the subchannel on which SSQD is performed | |
154 | * @ssqd: request and response block for SSQD | |
155 | * | |
156 | * Returns 0 on success. | |
157 | */ | |
158 | int chsc_ssqd(struct subchannel_id schid, struct chsc_ssqd_area *ssqd) | |
159 | { | |
160 | memset(ssqd, 0, sizeof(*ssqd)); | |
161 | ssqd->request.length = 0x0010; | |
162 | ssqd->request.code = 0x0024; | |
163 | ssqd->first_sch = schid.sch_no; | |
164 | ssqd->last_sch = schid.sch_no; | |
165 | ssqd->ssid = schid.ssid; | |
166 | ||
167 | if (chsc(ssqd)) | |
168 | return -EIO; | |
169 | ||
170 | return chsc_error_from_response(ssqd->response.code); | |
171 | } | |
172 | EXPORT_SYMBOL_GPL(chsc_ssqd); | |
173 | ||
174 | /** | |
175 | * chsc_sadc() - set adapter device controls (SADC) | |
176 | * @schid: id of the subchannel on which SADC is performed | |
177 | * @scssc: request and response block for SADC | |
178 | * @summary_indicator_addr: summary indicator address | |
179 | * @subchannel_indicator_addr: subchannel indicator address | |
180 | * | |
181 | * Returns 0 on success. | |
182 | */ | |
183 | int chsc_sadc(struct subchannel_id schid, struct chsc_scssc_area *scssc, | |
184 | u64 summary_indicator_addr, u64 subchannel_indicator_addr) | |
185 | { | |
186 | memset(scssc, 0, sizeof(*scssc)); | |
187 | scssc->request.length = 0x0fe0; | |
188 | scssc->request.code = 0x0021; | |
189 | scssc->operation_code = 0; | |
190 | ||
191 | scssc->summary_indicator_addr = summary_indicator_addr; | |
192 | scssc->subchannel_indicator_addr = subchannel_indicator_addr; | |
193 | ||
194 | scssc->ks = PAGE_DEFAULT_KEY >> 4; | |
195 | scssc->kc = PAGE_DEFAULT_KEY >> 4; | |
196 | scssc->isc = QDIO_AIRQ_ISC; | |
197 | scssc->schid = schid; | |
198 | ||
199 | /* enable the time delay disablement facility */ | |
200 | if (css_general_characteristics.aif_tdd) | |
201 | scssc->word_with_d_bit = 0x10000000; | |
202 | ||
203 | if (chsc(scssc)) | |
204 | return -EIO; | |
205 | ||
206 | return chsc_error_from_response(scssc->response.code); | |
207 | } | |
208 | EXPORT_SYMBOL_GPL(chsc_sadc); | |
209 | ||
210 | static int s390_subchannel_remove_chpid(struct subchannel *sch, void *data) | |
211 | { | |
212 | spin_lock_irq(sch->lock); | |
213 | if (sch->driver && sch->driver->chp_event) | |
214 | if (sch->driver->chp_event(sch, data, CHP_OFFLINE) != 0) | |
215 | goto out_unreg; | |
216 | spin_unlock_irq(sch->lock); | |
217 | return 0; | |
218 | ||
219 | out_unreg: | |
220 | sch->lpm = 0; | |
221 | spin_unlock_irq(sch->lock); | |
222 | css_schedule_eval(sch->schid); | |
223 | return 0; | |
224 | } | |
225 | ||
226 | void chsc_chp_offline(struct chp_id chpid) | |
227 | { | |
228 | struct channel_path *chp = chpid_to_chp(chpid); | |
229 | struct chp_link link; | |
230 | char dbf_txt[15]; | |
231 | ||
232 | sprintf(dbf_txt, "chpr%x.%02x", chpid.cssid, chpid.id); | |
233 | CIO_TRACE_EVENT(2, dbf_txt); | |
234 | ||
235 | if (chp_get_status(chpid) <= 0) | |
236 | return; | |
237 | memset(&link, 0, sizeof(struct chp_link)); | |
238 | link.chpid = chpid; | |
239 | /* Wait until previous actions have settled. */ | |
240 | css_wait_for_slow_path(); | |
241 | ||
242 | mutex_lock(&chp->lock); | |
243 | chp_update_desc(chp); | |
244 | mutex_unlock(&chp->lock); | |
245 | ||
246 | for_each_subchannel_staged(s390_subchannel_remove_chpid, NULL, &link); | |
247 | } | |
248 | ||
249 | static int __s390_process_res_acc(struct subchannel *sch, void *data) | |
250 | { | |
251 | spin_lock_irq(sch->lock); | |
252 | if (sch->driver && sch->driver->chp_event) | |
253 | sch->driver->chp_event(sch, data, CHP_ONLINE); | |
254 | spin_unlock_irq(sch->lock); | |
255 | ||
256 | return 0; | |
257 | } | |
258 | ||
259 | static void s390_process_res_acc(struct chp_link *link) | |
260 | { | |
261 | char dbf_txt[15]; | |
262 | ||
263 | sprintf(dbf_txt, "accpr%x.%02x", link->chpid.cssid, | |
264 | link->chpid.id); | |
265 | CIO_TRACE_EVENT( 2, dbf_txt); | |
266 | if (link->fla != 0) { | |
267 | sprintf(dbf_txt, "fla%x", link->fla); | |
268 | CIO_TRACE_EVENT( 2, dbf_txt); | |
269 | } | |
270 | /* Wait until previous actions have settled. */ | |
271 | css_wait_for_slow_path(); | |
272 | /* | |
273 | * I/O resources may have become accessible. | |
274 | * Scan through all subchannels that may be concerned and | |
275 | * do a validation on those. | |
276 | * The more information we have (info), the less scanning | |
277 | * will we have to do. | |
278 | */ | |
279 | for_each_subchannel_staged(__s390_process_res_acc, NULL, link); | |
280 | css_schedule_reprobe(); | |
281 | } | |
282 | ||
283 | struct chsc_sei_nt0_area { | |
284 | u8 flags; | |
285 | u8 vf; /* validity flags */ | |
286 | u8 rs; /* reporting source */ | |
287 | u8 cc; /* content code */ | |
288 | u16 fla; /* full link address */ | |
289 | u16 rsid; /* reporting source id */ | |
290 | u32 reserved1; | |
291 | u32 reserved2; | |
292 | /* ccdf has to be big enough for a link-incident record */ | |
293 | u8 ccdf[PAGE_SIZE - 24 - 16]; /* content-code dependent field */ | |
294 | } __packed; | |
295 | ||
296 | struct chsc_sei_nt2_area { | |
297 | u8 flags; /* p and v bit */ | |
298 | u8 reserved1; | |
299 | u8 reserved2; | |
300 | u8 cc; /* content code */ | |
301 | u32 reserved3[13]; | |
302 | u8 ccdf[PAGE_SIZE - 24 - 56]; /* content-code dependent field */ | |
303 | } __packed; | |
304 | ||
305 | #define CHSC_SEI_NT0 (1ULL << 63) | |
306 | #define CHSC_SEI_NT2 (1ULL << 61) | |
307 | ||
308 | struct chsc_sei { | |
309 | struct chsc_header request; | |
310 | u32 reserved1; | |
311 | u64 ntsm; /* notification type mask */ | |
312 | struct chsc_header response; | |
313 | u32 :24; | |
314 | u8 nt; | |
315 | union { | |
316 | struct chsc_sei_nt0_area nt0_area; | |
317 | struct chsc_sei_nt2_area nt2_area; | |
318 | u8 nt_area[PAGE_SIZE - 24]; | |
319 | } u; | |
320 | } __packed; | |
321 | ||
322 | /* | |
323 | * Node Descriptor as defined in SA22-7204, "Common I/O-Device Commands" | |
324 | */ | |
325 | ||
326 | #define ND_VALIDITY_VALID 0 | |
327 | #define ND_VALIDITY_OUTDATED 1 | |
328 | #define ND_VALIDITY_INVALID 2 | |
329 | ||
330 | struct node_descriptor { | |
331 | /* Flags. */ | |
332 | union { | |
333 | struct { | |
334 | u32 validity:3; | |
335 | u32 reserved:5; | |
336 | } __packed; | |
337 | u8 byte0; | |
338 | } __packed; | |
339 | ||
340 | /* Node parameters. */ | |
341 | u32 params:24; | |
342 | ||
343 | /* Node ID. */ | |
344 | char type[6]; | |
345 | char model[3]; | |
346 | char manufacturer[3]; | |
347 | char plant[2]; | |
348 | char seq[12]; | |
349 | u16 tag; | |
350 | } __packed; | |
351 | ||
352 | /* | |
353 | * Link Incident Record as defined in SA22-7202, "ESCON I/O Interface" | |
354 | */ | |
355 | ||
356 | #define LIR_IQ_CLASS_INFO 0 | |
357 | #define LIR_IQ_CLASS_DEGRADED 1 | |
358 | #define LIR_IQ_CLASS_NOT_OPERATIONAL 2 | |
359 | ||
360 | struct lir { | |
361 | struct { | |
362 | u32 null:1; | |
363 | u32 reserved:3; | |
364 | u32 class:2; | |
365 | u32 reserved2:2; | |
366 | } __packed iq; | |
367 | u32 ic:8; | |
368 | u32 reserved:16; | |
369 | struct node_descriptor incident_node; | |
370 | struct node_descriptor attached_node; | |
371 | u8 reserved2[32]; | |
372 | } __packed; | |
373 | ||
374 | #define PARAMS_LEN 10 /* PARAMS=xx,xxxxxx */ | |
375 | #define NODEID_LEN 35 /* NODEID=tttttt/mdl,mmm.ppssssssssssss,xxxx */ | |
376 | ||
377 | /* Copy EBCIDC text, convert to ASCII and optionally add delimiter. */ | |
378 | static char *store_ebcdic(char *dest, const char *src, unsigned long len, | |
379 | char delim) | |
380 | { | |
381 | memcpy(dest, src, len); | |
382 | EBCASC(dest, len); | |
383 | ||
384 | if (delim) | |
385 | dest[len++] = delim; | |
386 | ||
387 | return dest + len; | |
388 | } | |
389 | ||
390 | /* Format node ID and parameters for output in LIR log message. */ | |
391 | static void format_node_data(char *params, char *id, struct node_descriptor *nd) | |
392 | { | |
393 | memset(params, 0, PARAMS_LEN); | |
394 | memset(id, 0, NODEID_LEN); | |
395 | ||
396 | if (nd->validity != ND_VALIDITY_VALID) { | |
397 | strncpy(params, "n/a", PARAMS_LEN - 1); | |
398 | strncpy(id, "n/a", NODEID_LEN - 1); | |
399 | return; | |
400 | } | |
401 | ||
402 | /* PARAMS=xx,xxxxxx */ | |
403 | snprintf(params, PARAMS_LEN, "%02x,%06x", nd->byte0, nd->params); | |
404 | /* NODEID=tttttt/mdl,mmm.ppssssssssssss,xxxx */ | |
405 | id = store_ebcdic(id, nd->type, sizeof(nd->type), '/'); | |
406 | id = store_ebcdic(id, nd->model, sizeof(nd->model), ','); | |
407 | id = store_ebcdic(id, nd->manufacturer, sizeof(nd->manufacturer), '.'); | |
408 | id = store_ebcdic(id, nd->plant, sizeof(nd->plant), 0); | |
409 | id = store_ebcdic(id, nd->seq, sizeof(nd->seq), ','); | |
410 | sprintf(id, "%04X", nd->tag); | |
411 | } | |
412 | ||
413 | static void chsc_process_sei_link_incident(struct chsc_sei_nt0_area *sei_area) | |
414 | { | |
415 | struct lir *lir = (struct lir *) &sei_area->ccdf; | |
416 | char iuparams[PARAMS_LEN], iunodeid[NODEID_LEN], auparams[PARAMS_LEN], | |
417 | aunodeid[NODEID_LEN]; | |
418 | ||
419 | CIO_CRW_EVENT(4, "chsc: link incident (rs=%02x, rs_id=%04x, iq=%02x)\n", | |
420 | sei_area->rs, sei_area->rsid, sei_area->ccdf[0]); | |
421 | ||
422 | /* Ignore NULL Link Incident Records. */ | |
423 | if (lir->iq.null) | |
424 | return; | |
425 | ||
426 | /* Inform user that a link requires maintenance actions because it has | |
427 | * become degraded or not operational. Note that this log message is | |
428 | * the primary intention behind a Link Incident Record. */ | |
429 | ||
430 | format_node_data(iuparams, iunodeid, &lir->incident_node); | |
431 | format_node_data(auparams, aunodeid, &lir->attached_node); | |
432 | ||
433 | switch (lir->iq.class) { | |
434 | case LIR_IQ_CLASS_DEGRADED: | |
435 | pr_warn("Link degraded: RS=%02x RSID=%04x IC=%02x " | |
436 | "IUPARAMS=%s IUNODEID=%s AUPARAMS=%s AUNODEID=%s\n", | |
437 | sei_area->rs, sei_area->rsid, lir->ic, iuparams, | |
438 | iunodeid, auparams, aunodeid); | |
439 | break; | |
440 | case LIR_IQ_CLASS_NOT_OPERATIONAL: | |
441 | pr_err("Link stopped: RS=%02x RSID=%04x IC=%02x " | |
442 | "IUPARAMS=%s IUNODEID=%s AUPARAMS=%s AUNODEID=%s\n", | |
443 | sei_area->rs, sei_area->rsid, lir->ic, iuparams, | |
444 | iunodeid, auparams, aunodeid); | |
445 | break; | |
446 | default: | |
447 | break; | |
448 | } | |
449 | } | |
450 | ||
451 | static void chsc_process_sei_res_acc(struct chsc_sei_nt0_area *sei_area) | |
452 | { | |
453 | struct chp_link link; | |
454 | struct chp_id chpid; | |
455 | int status; | |
456 | ||
457 | CIO_CRW_EVENT(4, "chsc: resource accessibility event (rs=%02x, " | |
458 | "rs_id=%04x)\n", sei_area->rs, sei_area->rsid); | |
459 | if (sei_area->rs != 4) | |
460 | return; | |
461 | chp_id_init(&chpid); | |
462 | chpid.id = sei_area->rsid; | |
463 | /* allocate a new channel path structure, if needed */ | |
464 | status = chp_get_status(chpid); | |
465 | if (status < 0) | |
466 | chp_new(chpid); | |
467 | else if (!status) | |
468 | return; | |
469 | memset(&link, 0, sizeof(struct chp_link)); | |
470 | link.chpid = chpid; | |
471 | if ((sei_area->vf & 0xc0) != 0) { | |
472 | link.fla = sei_area->fla; | |
473 | if ((sei_area->vf & 0xc0) == 0xc0) | |
474 | /* full link address */ | |
475 | link.fla_mask = 0xffff; | |
476 | else | |
477 | /* link address */ | |
478 | link.fla_mask = 0xff00; | |
479 | } | |
480 | s390_process_res_acc(&link); | |
481 | } | |
482 | ||
483 | static void chsc_process_sei_chp_avail(struct chsc_sei_nt0_area *sei_area) | |
484 | { | |
485 | struct channel_path *chp; | |
486 | struct chp_id chpid; | |
487 | u8 *data; | |
488 | int num; | |
489 | ||
490 | CIO_CRW_EVENT(4, "chsc: channel path availability information\n"); | |
491 | if (sei_area->rs != 0) | |
492 | return; | |
493 | data = sei_area->ccdf; | |
494 | chp_id_init(&chpid); | |
495 | for (num = 0; num <= __MAX_CHPID; num++) { | |
496 | if (!chp_test_bit(data, num)) | |
497 | continue; | |
498 | chpid.id = num; | |
499 | ||
500 | CIO_CRW_EVENT(4, "Update information for channel path " | |
501 | "%x.%02x\n", chpid.cssid, chpid.id); | |
502 | chp = chpid_to_chp(chpid); | |
503 | if (!chp) { | |
504 | chp_new(chpid); | |
505 | continue; | |
506 | } | |
507 | mutex_lock(&chp->lock); | |
508 | chp_update_desc(chp); | |
509 | mutex_unlock(&chp->lock); | |
510 | } | |
511 | } | |
512 | ||
513 | struct chp_config_data { | |
514 | u8 map[32]; | |
515 | u8 op; | |
516 | u8 pc; | |
517 | }; | |
518 | ||
519 | static void chsc_process_sei_chp_config(struct chsc_sei_nt0_area *sei_area) | |
520 | { | |
521 | struct chp_config_data *data; | |
522 | struct chp_id chpid; | |
523 | int num; | |
524 | char *events[3] = {"configure", "deconfigure", "cancel deconfigure"}; | |
525 | ||
526 | CIO_CRW_EVENT(4, "chsc: channel-path-configuration notification\n"); | |
527 | if (sei_area->rs != 0) | |
528 | return; | |
529 | data = (struct chp_config_data *) &(sei_area->ccdf); | |
530 | chp_id_init(&chpid); | |
531 | for (num = 0; num <= __MAX_CHPID; num++) { | |
532 | if (!chp_test_bit(data->map, num)) | |
533 | continue; | |
534 | chpid.id = num; | |
535 | pr_notice("Processing %s for channel path %x.%02x\n", | |
536 | events[data->op], chpid.cssid, chpid.id); | |
537 | switch (data->op) { | |
538 | case 0: | |
539 | chp_cfg_schedule(chpid, 1); | |
540 | break; | |
541 | case 1: | |
542 | chp_cfg_schedule(chpid, 0); | |
543 | break; | |
544 | case 2: | |
545 | chp_cfg_cancel_deconfigure(chpid); | |
546 | break; | |
547 | } | |
548 | } | |
549 | } | |
550 | ||
551 | static void chsc_process_sei_scm_change(struct chsc_sei_nt0_area *sei_area) | |
552 | { | |
553 | int ret; | |
554 | ||
555 | CIO_CRW_EVENT(4, "chsc: scm change notification\n"); | |
556 | if (sei_area->rs != 7) | |
557 | return; | |
558 | ||
559 | ret = scm_update_information(); | |
560 | if (ret) | |
561 | CIO_CRW_EVENT(0, "chsc: updating change notification" | |
562 | " failed (rc=%d).\n", ret); | |
563 | } | |
564 | ||
565 | static void chsc_process_sei_scm_avail(struct chsc_sei_nt0_area *sei_area) | |
566 | { | |
567 | int ret; | |
568 | ||
569 | CIO_CRW_EVENT(4, "chsc: scm available information\n"); | |
570 | if (sei_area->rs != 7) | |
571 | return; | |
572 | ||
573 | ret = scm_process_availability_information(); | |
574 | if (ret) | |
575 | CIO_CRW_EVENT(0, "chsc: process availability information" | |
576 | " failed (rc=%d).\n", ret); | |
577 | } | |
578 | ||
579 | static void chsc_process_sei_nt2(struct chsc_sei_nt2_area *sei_area) | |
580 | { | |
581 | switch (sei_area->cc) { | |
582 | case 1: | |
583 | zpci_event_error(sei_area->ccdf); | |
584 | break; | |
585 | case 2: | |
586 | zpci_event_availability(sei_area->ccdf); | |
587 | break; | |
588 | default: | |
589 | CIO_CRW_EVENT(2, "chsc: sei nt2 unhandled cc=%d\n", | |
590 | sei_area->cc); | |
591 | break; | |
592 | } | |
593 | } | |
594 | ||
595 | static void chsc_process_sei_nt0(struct chsc_sei_nt0_area *sei_area) | |
596 | { | |
597 | /* which kind of information was stored? */ | |
598 | switch (sei_area->cc) { | |
599 | case 1: /* link incident*/ | |
600 | chsc_process_sei_link_incident(sei_area); | |
601 | break; | |
602 | case 2: /* i/o resource accessibility */ | |
603 | chsc_process_sei_res_acc(sei_area); | |
604 | break; | |
605 | case 7: /* channel-path-availability information */ | |
606 | chsc_process_sei_chp_avail(sei_area); | |
607 | break; | |
608 | case 8: /* channel-path-configuration notification */ | |
609 | chsc_process_sei_chp_config(sei_area); | |
610 | break; | |
611 | case 12: /* scm change notification */ | |
612 | chsc_process_sei_scm_change(sei_area); | |
613 | break; | |
614 | case 14: /* scm available notification */ | |
615 | chsc_process_sei_scm_avail(sei_area); | |
616 | break; | |
617 | default: /* other stuff */ | |
618 | CIO_CRW_EVENT(2, "chsc: sei nt0 unhandled cc=%d\n", | |
619 | sei_area->cc); | |
620 | break; | |
621 | } | |
622 | ||
623 | /* Check if we might have lost some information. */ | |
624 | if (sei_area->flags & 0x40) { | |
625 | CIO_CRW_EVENT(2, "chsc: event overflow\n"); | |
626 | css_schedule_eval_all(); | |
627 | } | |
628 | } | |
629 | ||
630 | static void chsc_process_event_information(struct chsc_sei *sei, u64 ntsm) | |
631 | { | |
632 | static int ntsm_unsupported; | |
633 | ||
634 | while (true) { | |
635 | memset(sei, 0, sizeof(*sei)); | |
636 | sei->request.length = 0x0010; | |
637 | sei->request.code = 0x000e; | |
638 | if (!ntsm_unsupported) | |
639 | sei->ntsm = ntsm; | |
640 | ||
641 | if (chsc(sei)) | |
642 | break; | |
643 | ||
644 | if (sei->response.code != 0x0001) { | |
645 | CIO_CRW_EVENT(2, "chsc: sei failed (rc=%04x, ntsm=%llx)\n", | |
646 | sei->response.code, sei->ntsm); | |
647 | ||
648 | if (sei->response.code == 3 && sei->ntsm) { | |
649 | /* Fallback for old firmware. */ | |
650 | ntsm_unsupported = 1; | |
651 | continue; | |
652 | } | |
653 | break; | |
654 | } | |
655 | ||
656 | CIO_CRW_EVENT(2, "chsc: sei successful (nt=%d)\n", sei->nt); | |
657 | switch (sei->nt) { | |
658 | case 0: | |
659 | chsc_process_sei_nt0(&sei->u.nt0_area); | |
660 | break; | |
661 | case 2: | |
662 | chsc_process_sei_nt2(&sei->u.nt2_area); | |
663 | break; | |
664 | default: | |
665 | CIO_CRW_EVENT(2, "chsc: unhandled nt: %d\n", sei->nt); | |
666 | break; | |
667 | } | |
668 | ||
669 | if (!(sei->u.nt0_area.flags & 0x80)) | |
670 | break; | |
671 | } | |
672 | } | |
673 | ||
674 | /* | |
675 | * Handle channel subsystem related CRWs. | |
676 | * Use store event information to find out what's going on. | |
677 | * | |
678 | * Note: Access to sei_page is serialized through machine check handler | |
679 | * thread, so no need for locking. | |
680 | */ | |
681 | static void chsc_process_crw(struct crw *crw0, struct crw *crw1, int overflow) | |
682 | { | |
683 | struct chsc_sei *sei = sei_page; | |
684 | ||
685 | if (overflow) { | |
686 | css_schedule_eval_all(); | |
687 | return; | |
688 | } | |
689 | CIO_CRW_EVENT(2, "CRW reports slct=%d, oflw=%d, " | |
690 | "chn=%d, rsc=%X, anc=%d, erc=%X, rsid=%X\n", | |
691 | crw0->slct, crw0->oflw, crw0->chn, crw0->rsc, crw0->anc, | |
692 | crw0->erc, crw0->rsid); | |
693 | ||
694 | CIO_TRACE_EVENT(2, "prcss"); | |
695 | chsc_process_event_information(sei, CHSC_SEI_NT0 | CHSC_SEI_NT2); | |
696 | } | |
697 | ||
698 | void chsc_chp_online(struct chp_id chpid) | |
699 | { | |
700 | struct channel_path *chp = chpid_to_chp(chpid); | |
701 | struct chp_link link; | |
702 | char dbf_txt[15]; | |
703 | ||
704 | sprintf(dbf_txt, "cadd%x.%02x", chpid.cssid, chpid.id); | |
705 | CIO_TRACE_EVENT(2, dbf_txt); | |
706 | ||
707 | if (chp_get_status(chpid) != 0) { | |
708 | memset(&link, 0, sizeof(struct chp_link)); | |
709 | link.chpid = chpid; | |
710 | /* Wait until previous actions have settled. */ | |
711 | css_wait_for_slow_path(); | |
712 | ||
713 | mutex_lock(&chp->lock); | |
714 | chp_update_desc(chp); | |
715 | mutex_unlock(&chp->lock); | |
716 | ||
717 | for_each_subchannel_staged(__s390_process_res_acc, NULL, | |
718 | &link); | |
719 | css_schedule_reprobe(); | |
720 | } | |
721 | } | |
722 | ||
723 | static void __s390_subchannel_vary_chpid(struct subchannel *sch, | |
724 | struct chp_id chpid, int on) | |
725 | { | |
726 | unsigned long flags; | |
727 | struct chp_link link; | |
728 | ||
729 | memset(&link, 0, sizeof(struct chp_link)); | |
730 | link.chpid = chpid; | |
731 | spin_lock_irqsave(sch->lock, flags); | |
732 | if (sch->driver && sch->driver->chp_event) | |
733 | sch->driver->chp_event(sch, &link, | |
734 | on ? CHP_VARY_ON : CHP_VARY_OFF); | |
735 | spin_unlock_irqrestore(sch->lock, flags); | |
736 | } | |
737 | ||
738 | static int s390_subchannel_vary_chpid_off(struct subchannel *sch, void *data) | |
739 | { | |
740 | struct chp_id *chpid = data; | |
741 | ||
742 | __s390_subchannel_vary_chpid(sch, *chpid, 0); | |
743 | return 0; | |
744 | } | |
745 | ||
746 | static int s390_subchannel_vary_chpid_on(struct subchannel *sch, void *data) | |
747 | { | |
748 | struct chp_id *chpid = data; | |
749 | ||
750 | __s390_subchannel_vary_chpid(sch, *chpid, 1); | |
751 | return 0; | |
752 | } | |
753 | ||
754 | /** | |
755 | * chsc_chp_vary - propagate channel-path vary operation to subchannels | |
756 | * @chpid: channl-path ID | |
757 | * @on: non-zero for vary online, zero for vary offline | |
758 | */ | |
759 | int chsc_chp_vary(struct chp_id chpid, int on) | |
760 | { | |
761 | struct channel_path *chp = chpid_to_chp(chpid); | |
762 | ||
763 | /* Wait until previous actions have settled. */ | |
764 | css_wait_for_slow_path(); | |
765 | /* | |
766 | * Redo PathVerification on the devices the chpid connects to | |
767 | */ | |
768 | if (on) { | |
769 | /* Try to update the channel path description. */ | |
770 | chp_update_desc(chp); | |
771 | for_each_subchannel_staged(s390_subchannel_vary_chpid_on, | |
772 | NULL, &chpid); | |
773 | css_schedule_reprobe(); | |
774 | } else | |
775 | for_each_subchannel_staged(s390_subchannel_vary_chpid_off, | |
776 | NULL, &chpid); | |
777 | ||
778 | return 0; | |
779 | } | |
780 | ||
781 | static void | |
782 | chsc_remove_cmg_attr(struct channel_subsystem *css) | |
783 | { | |
784 | int i; | |
785 | ||
786 | for (i = 0; i <= __MAX_CHPID; i++) { | |
787 | if (!css->chps[i]) | |
788 | continue; | |
789 | chp_remove_cmg_attr(css->chps[i]); | |
790 | } | |
791 | } | |
792 | ||
793 | static int | |
794 | chsc_add_cmg_attr(struct channel_subsystem *css) | |
795 | { | |
796 | int i, ret; | |
797 | ||
798 | ret = 0; | |
799 | for (i = 0; i <= __MAX_CHPID; i++) { | |
800 | if (!css->chps[i]) | |
801 | continue; | |
802 | ret = chp_add_cmg_attr(css->chps[i]); | |
803 | if (ret) | |
804 | goto cleanup; | |
805 | } | |
806 | return ret; | |
807 | cleanup: | |
808 | for (--i; i >= 0; i--) { | |
809 | if (!css->chps[i]) | |
810 | continue; | |
811 | chp_remove_cmg_attr(css->chps[i]); | |
812 | } | |
813 | return ret; | |
814 | } | |
815 | ||
816 | int __chsc_do_secm(struct channel_subsystem *css, int enable) | |
817 | { | |
818 | struct { | |
819 | struct chsc_header request; | |
820 | u32 operation_code : 2; | |
821 | u32 : 30; | |
822 | u32 key : 4; | |
823 | u32 : 28; | |
824 | u32 zeroes1; | |
825 | u32 cub_addr1; | |
826 | u32 zeroes2; | |
827 | u32 cub_addr2; | |
828 | u32 reserved[13]; | |
829 | struct chsc_header response; | |
830 | u32 status : 8; | |
831 | u32 : 4; | |
832 | u32 fmt : 4; | |
833 | u32 : 16; | |
834 | } __attribute__ ((packed)) *secm_area; | |
835 | int ret, ccode; | |
836 | ||
837 | spin_lock_irq(&chsc_page_lock); | |
838 | memset(chsc_page, 0, PAGE_SIZE); | |
839 | secm_area = chsc_page; | |
840 | secm_area->request.length = 0x0050; | |
841 | secm_area->request.code = 0x0016; | |
842 | ||
843 | secm_area->key = PAGE_DEFAULT_KEY >> 4; | |
844 | secm_area->cub_addr1 = (u64)(unsigned long)css->cub_addr1; | |
845 | secm_area->cub_addr2 = (u64)(unsigned long)css->cub_addr2; | |
846 | ||
847 | secm_area->operation_code = enable ? 0 : 1; | |
848 | ||
849 | ccode = chsc(secm_area); | |
850 | if (ccode > 0) { | |
851 | ret = (ccode == 3) ? -ENODEV : -EBUSY; | |
852 | goto out; | |
853 | } | |
854 | ||
855 | switch (secm_area->response.code) { | |
856 | case 0x0102: | |
857 | case 0x0103: | |
858 | ret = -EINVAL; | |
859 | break; | |
860 | default: | |
861 | ret = chsc_error_from_response(secm_area->response.code); | |
862 | } | |
863 | if (ret != 0) | |
864 | CIO_CRW_EVENT(2, "chsc: secm failed (rc=%04x)\n", | |
865 | secm_area->response.code); | |
866 | out: | |
867 | spin_unlock_irq(&chsc_page_lock); | |
868 | return ret; | |
869 | } | |
870 | ||
871 | int | |
872 | chsc_secm(struct channel_subsystem *css, int enable) | |
873 | { | |
874 | int ret; | |
875 | ||
876 | if (enable && !css->cm_enabled) { | |
877 | css->cub_addr1 = (void *)get_zeroed_page(GFP_KERNEL | GFP_DMA); | |
878 | css->cub_addr2 = (void *)get_zeroed_page(GFP_KERNEL | GFP_DMA); | |
879 | if (!css->cub_addr1 || !css->cub_addr2) { | |
880 | free_page((unsigned long)css->cub_addr1); | |
881 | free_page((unsigned long)css->cub_addr2); | |
882 | return -ENOMEM; | |
883 | } | |
884 | } | |
885 | ret = __chsc_do_secm(css, enable); | |
886 | if (!ret) { | |
887 | css->cm_enabled = enable; | |
888 | if (css->cm_enabled) { | |
889 | ret = chsc_add_cmg_attr(css); | |
890 | if (ret) { | |
891 | __chsc_do_secm(css, 0); | |
892 | css->cm_enabled = 0; | |
893 | } | |
894 | } else | |
895 | chsc_remove_cmg_attr(css); | |
896 | } | |
897 | if (!css->cm_enabled) { | |
898 | free_page((unsigned long)css->cub_addr1); | |
899 | free_page((unsigned long)css->cub_addr2); | |
900 | } | |
901 | return ret; | |
902 | } | |
903 | ||
904 | int chsc_determine_channel_path_desc(struct chp_id chpid, int fmt, int rfmt, | |
905 | int c, int m, void *page) | |
906 | { | |
907 | struct chsc_scpd *scpd_area; | |
908 | int ccode, ret; | |
909 | ||
910 | if ((rfmt == 1) && !css_general_characteristics.fcs) | |
911 | return -EINVAL; | |
912 | if ((rfmt == 2) && !css_general_characteristics.cib) | |
913 | return -EINVAL; | |
914 | ||
915 | memset(page, 0, PAGE_SIZE); | |
916 | scpd_area = page; | |
917 | scpd_area->request.length = 0x0010; | |
918 | scpd_area->request.code = 0x0002; | |
919 | scpd_area->cssid = chpid.cssid; | |
920 | scpd_area->first_chpid = chpid.id; | |
921 | scpd_area->last_chpid = chpid.id; | |
922 | scpd_area->m = m; | |
923 | scpd_area->c = c; | |
924 | scpd_area->fmt = fmt; | |
925 | scpd_area->rfmt = rfmt; | |
926 | ||
927 | ccode = chsc(scpd_area); | |
928 | if (ccode > 0) | |
929 | return (ccode == 3) ? -ENODEV : -EBUSY; | |
930 | ||
931 | ret = chsc_error_from_response(scpd_area->response.code); | |
932 | if (ret) | |
933 | CIO_CRW_EVENT(2, "chsc: scpd failed (rc=%04x)\n", | |
934 | scpd_area->response.code); | |
935 | return ret; | |
936 | } | |
937 | EXPORT_SYMBOL_GPL(chsc_determine_channel_path_desc); | |
938 | ||
939 | int chsc_determine_base_channel_path_desc(struct chp_id chpid, | |
940 | struct channel_path_desc *desc) | |
941 | { | |
942 | struct chsc_response_struct *chsc_resp; | |
943 | struct chsc_scpd *scpd_area; | |
944 | unsigned long flags; | |
945 | int ret; | |
946 | ||
947 | spin_lock_irqsave(&chsc_page_lock, flags); | |
948 | scpd_area = chsc_page; | |
949 | ret = chsc_determine_channel_path_desc(chpid, 0, 0, 0, 0, scpd_area); | |
950 | if (ret) | |
951 | goto out; | |
952 | chsc_resp = (void *)&scpd_area->response; | |
953 | memcpy(desc, &chsc_resp->data, sizeof(*desc)); | |
954 | out: | |
955 | spin_unlock_irqrestore(&chsc_page_lock, flags); | |
956 | return ret; | |
957 | } | |
958 | ||
959 | int chsc_determine_fmt1_channel_path_desc(struct chp_id chpid, | |
960 | struct channel_path_desc_fmt1 *desc) | |
961 | { | |
962 | struct chsc_response_struct *chsc_resp; | |
963 | struct chsc_scpd *scpd_area; | |
964 | unsigned long flags; | |
965 | int ret; | |
966 | ||
967 | spin_lock_irqsave(&chsc_page_lock, flags); | |
968 | scpd_area = chsc_page; | |
969 | ret = chsc_determine_channel_path_desc(chpid, 0, 0, 1, 0, scpd_area); | |
970 | if (ret) | |
971 | goto out; | |
972 | chsc_resp = (void *)&scpd_area->response; | |
973 | memcpy(desc, &chsc_resp->data, sizeof(*desc)); | |
974 | out: | |
975 | spin_unlock_irqrestore(&chsc_page_lock, flags); | |
976 | return ret; | |
977 | } | |
978 | ||
979 | static void | |
980 | chsc_initialize_cmg_chars(struct channel_path *chp, u8 cmcv, | |
981 | struct cmg_chars *chars) | |
982 | { | |
983 | int i, mask; | |
984 | ||
985 | for (i = 0; i < NR_MEASUREMENT_CHARS; i++) { | |
986 | mask = 0x80 >> (i + 3); | |
987 | if (cmcv & mask) | |
988 | chp->cmg_chars.values[i] = chars->values[i]; | |
989 | else | |
990 | chp->cmg_chars.values[i] = 0; | |
991 | } | |
992 | } | |
993 | ||
994 | int chsc_get_channel_measurement_chars(struct channel_path *chp) | |
995 | { | |
996 | int ccode, ret; | |
997 | ||
998 | struct { | |
999 | struct chsc_header request; | |
1000 | u32 : 24; | |
1001 | u32 first_chpid : 8; | |
1002 | u32 : 24; | |
1003 | u32 last_chpid : 8; | |
1004 | u32 zeroes1; | |
1005 | struct chsc_header response; | |
1006 | u32 zeroes2; | |
1007 | u32 not_valid : 1; | |
1008 | u32 shared : 1; | |
1009 | u32 : 22; | |
1010 | u32 chpid : 8; | |
1011 | u32 cmcv : 5; | |
1012 | u32 : 11; | |
1013 | u32 cmgq : 8; | |
1014 | u32 cmg : 8; | |
1015 | u32 zeroes3; | |
1016 | u32 data[NR_MEASUREMENT_CHARS]; | |
1017 | } __attribute__ ((packed)) *scmc_area; | |
1018 | ||
1019 | chp->shared = -1; | |
1020 | chp->cmg = -1; | |
1021 | ||
1022 | if (!css_chsc_characteristics.scmc || !css_chsc_characteristics.secm) | |
1023 | return -EINVAL; | |
1024 | ||
1025 | spin_lock_irq(&chsc_page_lock); | |
1026 | memset(chsc_page, 0, PAGE_SIZE); | |
1027 | scmc_area = chsc_page; | |
1028 | scmc_area->request.length = 0x0010; | |
1029 | scmc_area->request.code = 0x0022; | |
1030 | scmc_area->first_chpid = chp->chpid.id; | |
1031 | scmc_area->last_chpid = chp->chpid.id; | |
1032 | ||
1033 | ccode = chsc(scmc_area); | |
1034 | if (ccode > 0) { | |
1035 | ret = (ccode == 3) ? -ENODEV : -EBUSY; | |
1036 | goto out; | |
1037 | } | |
1038 | ||
1039 | ret = chsc_error_from_response(scmc_area->response.code); | |
1040 | if (ret) { | |
1041 | CIO_CRW_EVENT(2, "chsc: scmc failed (rc=%04x)\n", | |
1042 | scmc_area->response.code); | |
1043 | goto out; | |
1044 | } | |
1045 | if (scmc_area->not_valid) | |
1046 | goto out; | |
1047 | ||
1048 | chp->cmg = scmc_area->cmg; | |
1049 | chp->shared = scmc_area->shared; | |
1050 | if (chp->cmg != 2 && chp->cmg != 3) { | |
1051 | /* No cmg-dependent data. */ | |
1052 | goto out; | |
1053 | } | |
1054 | chsc_initialize_cmg_chars(chp, scmc_area->cmcv, | |
1055 | (struct cmg_chars *) &scmc_area->data); | |
1056 | out: | |
1057 | spin_unlock_irq(&chsc_page_lock); | |
1058 | return ret; | |
1059 | } | |
1060 | ||
1061 | int __init chsc_init(void) | |
1062 | { | |
1063 | int ret; | |
1064 | ||
1065 | sei_page = (void *)get_zeroed_page(GFP_KERNEL | GFP_DMA); | |
1066 | chsc_page = (void *)get_zeroed_page(GFP_KERNEL | GFP_DMA); | |
1067 | if (!sei_page || !chsc_page) { | |
1068 | ret = -ENOMEM; | |
1069 | goto out_err; | |
1070 | } | |
1071 | ret = crw_register_handler(CRW_RSC_CSS, chsc_process_crw); | |
1072 | if (ret) | |
1073 | goto out_err; | |
1074 | return ret; | |
1075 | out_err: | |
1076 | free_page((unsigned long)chsc_page); | |
1077 | free_page((unsigned long)sei_page); | |
1078 | return ret; | |
1079 | } | |
1080 | ||
1081 | void __init chsc_init_cleanup(void) | |
1082 | { | |
1083 | crw_unregister_handler(CRW_RSC_CSS); | |
1084 | free_page((unsigned long)chsc_page); | |
1085 | free_page((unsigned long)sei_page); | |
1086 | } | |
1087 | ||
1088 | int __chsc_enable_facility(struct chsc_sda_area *sda_area, int operation_code) | |
1089 | { | |
1090 | int ret; | |
1091 | ||
1092 | sda_area->request.length = 0x0400; | |
1093 | sda_area->request.code = 0x0031; | |
1094 | sda_area->operation_code = operation_code; | |
1095 | ||
1096 | ret = chsc(sda_area); | |
1097 | if (ret > 0) { | |
1098 | ret = (ret == 3) ? -ENODEV : -EBUSY; | |
1099 | goto out; | |
1100 | } | |
1101 | ||
1102 | switch (sda_area->response.code) { | |
1103 | case 0x0101: | |
1104 | ret = -EOPNOTSUPP; | |
1105 | break; | |
1106 | default: | |
1107 | ret = chsc_error_from_response(sda_area->response.code); | |
1108 | } | |
1109 | out: | |
1110 | return ret; | |
1111 | } | |
1112 | ||
1113 | int chsc_enable_facility(int operation_code) | |
1114 | { | |
1115 | struct chsc_sda_area *sda_area; | |
1116 | unsigned long flags; | |
1117 | int ret; | |
1118 | ||
1119 | spin_lock_irqsave(&chsc_page_lock, flags); | |
1120 | memset(chsc_page, 0, PAGE_SIZE); | |
1121 | sda_area = chsc_page; | |
1122 | ||
1123 | ret = __chsc_enable_facility(sda_area, operation_code); | |
1124 | if (ret != 0) | |
1125 | CIO_CRW_EVENT(2, "chsc: sda (oc=%x) failed (rc=%04x)\n", | |
1126 | operation_code, sda_area->response.code); | |
1127 | ||
1128 | spin_unlock_irqrestore(&chsc_page_lock, flags); | |
1129 | return ret; | |
1130 | } | |
1131 | ||
1132 | struct css_general_char css_general_characteristics; | |
1133 | struct css_chsc_char css_chsc_characteristics; | |
1134 | ||
1135 | int __init | |
1136 | chsc_determine_css_characteristics(void) | |
1137 | { | |
1138 | int result; | |
1139 | struct { | |
1140 | struct chsc_header request; | |
1141 | u32 reserved1; | |
1142 | u32 reserved2; | |
1143 | u32 reserved3; | |
1144 | struct chsc_header response; | |
1145 | u32 reserved4; | |
1146 | u32 general_char[510]; | |
1147 | u32 chsc_char[508]; | |
1148 | } __attribute__ ((packed)) *scsc_area; | |
1149 | ||
1150 | spin_lock_irq(&chsc_page_lock); | |
1151 | memset(chsc_page, 0, PAGE_SIZE); | |
1152 | scsc_area = chsc_page; | |
1153 | scsc_area->request.length = 0x0010; | |
1154 | scsc_area->request.code = 0x0010; | |
1155 | ||
1156 | result = chsc(scsc_area); | |
1157 | if (result) { | |
1158 | result = (result == 3) ? -ENODEV : -EBUSY; | |
1159 | goto exit; | |
1160 | } | |
1161 | ||
1162 | result = chsc_error_from_response(scsc_area->response.code); | |
1163 | if (result == 0) { | |
1164 | memcpy(&css_general_characteristics, scsc_area->general_char, | |
1165 | sizeof(css_general_characteristics)); | |
1166 | memcpy(&css_chsc_characteristics, scsc_area->chsc_char, | |
1167 | sizeof(css_chsc_characteristics)); | |
1168 | } else | |
1169 | CIO_CRW_EVENT(2, "chsc: scsc failed (rc=%04x)\n", | |
1170 | scsc_area->response.code); | |
1171 | exit: | |
1172 | spin_unlock_irq(&chsc_page_lock); | |
1173 | return result; | |
1174 | } | |
1175 | ||
1176 | EXPORT_SYMBOL_GPL(css_general_characteristics); | |
1177 | EXPORT_SYMBOL_GPL(css_chsc_characteristics); | |
1178 | ||
1179 | int chsc_sstpc(void *page, unsigned int op, u16 ctrl, u64 *clock_delta) | |
1180 | { | |
1181 | struct { | |
1182 | struct chsc_header request; | |
1183 | unsigned int rsvd0; | |
1184 | unsigned int op : 8; | |
1185 | unsigned int rsvd1 : 8; | |
1186 | unsigned int ctrl : 16; | |
1187 | unsigned int rsvd2[5]; | |
1188 | struct chsc_header response; | |
1189 | unsigned int rsvd3[3]; | |
1190 | u64 clock_delta; | |
1191 | unsigned int rsvd4[2]; | |
1192 | } __attribute__ ((packed)) *rr; | |
1193 | int rc; | |
1194 | ||
1195 | memset(page, 0, PAGE_SIZE); | |
1196 | rr = page; | |
1197 | rr->request.length = 0x0020; | |
1198 | rr->request.code = 0x0033; | |
1199 | rr->op = op; | |
1200 | rr->ctrl = ctrl; | |
1201 | rc = chsc(rr); | |
1202 | if (rc) | |
1203 | return -EIO; | |
1204 | rc = (rr->response.code == 0x0001) ? 0 : -EIO; | |
1205 | if (clock_delta) | |
1206 | *clock_delta = rr->clock_delta; | |
1207 | return rc; | |
1208 | } | |
1209 | ||
1210 | int chsc_sstpi(void *page, void *result, size_t size) | |
1211 | { | |
1212 | struct { | |
1213 | struct chsc_header request; | |
1214 | unsigned int rsvd0[3]; | |
1215 | struct chsc_header response; | |
1216 | char data[size]; | |
1217 | } __attribute__ ((packed)) *rr; | |
1218 | int rc; | |
1219 | ||
1220 | memset(page, 0, PAGE_SIZE); | |
1221 | rr = page; | |
1222 | rr->request.length = 0x0010; | |
1223 | rr->request.code = 0x0038; | |
1224 | rc = chsc(rr); | |
1225 | if (rc) | |
1226 | return -EIO; | |
1227 | memcpy(result, &rr->data, size); | |
1228 | return (rr->response.code == 0x0001) ? 0 : -EIO; | |
1229 | } | |
1230 | ||
1231 | int chsc_siosl(struct subchannel_id schid) | |
1232 | { | |
1233 | struct { | |
1234 | struct chsc_header request; | |
1235 | u32 word1; | |
1236 | struct subchannel_id sid; | |
1237 | u32 word3; | |
1238 | struct chsc_header response; | |
1239 | u32 word[11]; | |
1240 | } __attribute__ ((packed)) *siosl_area; | |
1241 | unsigned long flags; | |
1242 | int ccode; | |
1243 | int rc; | |
1244 | ||
1245 | spin_lock_irqsave(&chsc_page_lock, flags); | |
1246 | memset(chsc_page, 0, PAGE_SIZE); | |
1247 | siosl_area = chsc_page; | |
1248 | siosl_area->request.length = 0x0010; | |
1249 | siosl_area->request.code = 0x0046; | |
1250 | siosl_area->word1 = 0x80000000; | |
1251 | siosl_area->sid = schid; | |
1252 | ||
1253 | ccode = chsc(siosl_area); | |
1254 | if (ccode > 0) { | |
1255 | if (ccode == 3) | |
1256 | rc = -ENODEV; | |
1257 | else | |
1258 | rc = -EBUSY; | |
1259 | CIO_MSG_EVENT(2, "chsc: chsc failed for 0.%x.%04x (ccode=%d)\n", | |
1260 | schid.ssid, schid.sch_no, ccode); | |
1261 | goto out; | |
1262 | } | |
1263 | rc = chsc_error_from_response(siosl_area->response.code); | |
1264 | if (rc) | |
1265 | CIO_MSG_EVENT(2, "chsc: siosl failed for 0.%x.%04x (rc=%04x)\n", | |
1266 | schid.ssid, schid.sch_no, | |
1267 | siosl_area->response.code); | |
1268 | else | |
1269 | CIO_MSG_EVENT(4, "chsc: siosl succeeded for 0.%x.%04x\n", | |
1270 | schid.ssid, schid.sch_no); | |
1271 | out: | |
1272 | spin_unlock_irqrestore(&chsc_page_lock, flags); | |
1273 | return rc; | |
1274 | } | |
1275 | EXPORT_SYMBOL_GPL(chsc_siosl); | |
1276 | ||
1277 | /** | |
1278 | * chsc_scm_info() - store SCM information (SSI) | |
1279 | * @scm_area: request and response block for SSI | |
1280 | * @token: continuation token | |
1281 | * | |
1282 | * Returns 0 on success. | |
1283 | */ | |
1284 | int chsc_scm_info(struct chsc_scm_info *scm_area, u64 token) | |
1285 | { | |
1286 | int ccode, ret; | |
1287 | ||
1288 | memset(scm_area, 0, sizeof(*scm_area)); | |
1289 | scm_area->request.length = 0x0020; | |
1290 | scm_area->request.code = 0x004C; | |
1291 | scm_area->reqtok = token; | |
1292 | ||
1293 | ccode = chsc(scm_area); | |
1294 | if (ccode > 0) { | |
1295 | ret = (ccode == 3) ? -ENODEV : -EBUSY; | |
1296 | goto out; | |
1297 | } | |
1298 | ret = chsc_error_from_response(scm_area->response.code); | |
1299 | if (ret != 0) | |
1300 | CIO_MSG_EVENT(2, "chsc: scm info failed (rc=%04x)\n", | |
1301 | scm_area->response.code); | |
1302 | out: | |
1303 | return ret; | |
1304 | } | |
1305 | EXPORT_SYMBOL_GPL(chsc_scm_info); | |
1306 | ||
1307 | /** | |
1308 | * chsc_pnso_brinfo() - Perform Network-Subchannel Operation, Bridge Info. | |
1309 | * @schid: id of the subchannel on which PNSO is performed | |
1310 | * @brinfo_area: request and response block for the operation | |
1311 | * @resume_token: resume token for multiblock response | |
1312 | * @cnc: Boolean change-notification control | |
1313 | * | |
1314 | * brinfo_area must be allocated by the caller with get_zeroed_page(GFP_KERNEL) | |
1315 | * | |
1316 | * Returns 0 on success. | |
1317 | */ | |
1318 | int chsc_pnso_brinfo(struct subchannel_id schid, | |
1319 | struct chsc_pnso_area *brinfo_area, | |
1320 | struct chsc_brinfo_resume_token resume_token, | |
1321 | int cnc) | |
1322 | { | |
1323 | memset(brinfo_area, 0, sizeof(*brinfo_area)); | |
1324 | brinfo_area->request.length = 0x0030; | |
1325 | brinfo_area->request.code = 0x003d; /* network-subchannel operation */ | |
1326 | brinfo_area->m = schid.m; | |
1327 | brinfo_area->ssid = schid.ssid; | |
1328 | brinfo_area->sch = schid.sch_no; | |
1329 | brinfo_area->cssid = schid.cssid; | |
1330 | brinfo_area->oc = 0; /* Store-network-bridging-information list */ | |
1331 | brinfo_area->resume_token = resume_token; | |
1332 | brinfo_area->n = (cnc != 0); | |
1333 | if (chsc(brinfo_area)) | |
1334 | return -EIO; | |
1335 | return chsc_error_from_response(brinfo_area->response.code); | |
1336 | } | |
1337 | EXPORT_SYMBOL_GPL(chsc_pnso_brinfo); |