]>
Commit | Line | Data |
---|---|---|
9cc6fc50 DH |
1 | /* Handle fileserver selection and rotation. |
2 | * | |
3 | * Copyright (C) 2017 Red Hat, Inc. All Rights Reserved. | |
4 | * Written by David Howells (dhowells@redhat.com) | |
5 | * | |
6 | * This program is free software; you can redistribute it and/or | |
7 | * modify it under the terms of the GNU General Public Licence | |
8 | * as published by the Free Software Foundation; either version | |
9 | * 2 of the Licence, or (at your option) any later version. | |
10 | */ | |
11 | ||
12 | #include <linux/kernel.h> | |
13 | #include <linux/slab.h> | |
d2ddc776 DH |
14 | #include <linux/fs.h> |
15 | #include <linux/sched.h> | |
16 | #include <linux/delay.h> | |
17 | #include <linux/sched/signal.h> | |
9cc6fc50 | 18 | #include "internal.h" |
d2ddc776 | 19 | #include "afs_fs.h" |
9cc6fc50 DH |
20 | |
21 | /* | |
22 | * Initialise a filesystem server cursor for iterating over FS servers. | |
23 | */ | |
24 | void afs_init_fs_cursor(struct afs_fs_cursor *fc, struct afs_vnode *vnode) | |
25 | { | |
26 | memset(fc, 0, sizeof(*fc)); | |
27 | } | |
28 | ||
d2ddc776 DH |
29 | /* |
30 | * Begin an operation on the fileserver. | |
31 | * | |
32 | * Fileserver operations are serialised on the server by vnode, so we serialise | |
33 | * them here also using the io_lock. | |
34 | */ | |
35 | bool afs_begin_vnode_operation(struct afs_fs_cursor *fc, struct afs_vnode *vnode, | |
36 | struct key *key) | |
37 | { | |
38 | afs_init_fs_cursor(fc, vnode); | |
39 | fc->vnode = vnode; | |
40 | fc->key = key; | |
41 | fc->ac.error = SHRT_MAX; | |
42 | ||
43 | if (mutex_lock_interruptible(&vnode->io_lock) < 0) { | |
44 | fc->ac.error = -EINTR; | |
45 | fc->flags |= AFS_FS_CURSOR_STOP; | |
46 | return false; | |
47 | } | |
48 | ||
0fafdc9f | 49 | if (vnode->lock_state != AFS_VNODE_LOCK_NONE) |
d2ddc776 DH |
50 | fc->flags |= AFS_FS_CURSOR_CUR_ONLY; |
51 | return true; | |
52 | } | |
53 | ||
54 | /* | |
55 | * Begin iteration through a server list, starting with the vnode's last used | |
56 | * server if possible, or the last recorded good server if not. | |
57 | */ | |
58 | static bool afs_start_fs_iteration(struct afs_fs_cursor *fc, | |
59 | struct afs_vnode *vnode) | |
60 | { | |
61 | struct afs_cb_interest *cbi; | |
62 | int i; | |
63 | ||
64 | read_lock(&vnode->volume->servers_lock); | |
65 | fc->server_list = afs_get_serverlist(vnode->volume->servers); | |
66 | read_unlock(&vnode->volume->servers_lock); | |
67 | ||
68 | cbi = vnode->cb_interest; | |
69 | if (cbi) { | |
70 | /* See if the vnode's preferred record is still available */ | |
71 | for (i = 0; i < fc->server_list->nr_servers; i++) { | |
72 | if (fc->server_list->servers[i].cb_interest == cbi) { | |
73 | fc->start = i; | |
74 | goto found_interest; | |
75 | } | |
76 | } | |
77 | ||
78 | /* If we have a lock outstanding on a server that's no longer | |
79 | * serving this vnode, then we can't switch to another server | |
80 | * and have to return an error. | |
81 | */ | |
82 | if (fc->flags & AFS_FS_CURSOR_CUR_ONLY) { | |
83 | fc->ac.error = -ESTALE; | |
84 | return false; | |
85 | } | |
86 | ||
87 | /* Note that the callback promise is effectively broken */ | |
88 | write_seqlock(&vnode->cb_lock); | |
89 | ASSERTCMP(cbi, ==, vnode->cb_interest); | |
90 | vnode->cb_interest = NULL; | |
91 | if (test_and_clear_bit(AFS_VNODE_CB_PROMISED, &vnode->flags)) | |
92 | vnode->cb_break++; | |
93 | write_sequnlock(&vnode->cb_lock); | |
94 | ||
95 | afs_put_cb_interest(afs_v2net(vnode), cbi); | |
96 | cbi = NULL; | |
97 | } else { | |
98 | fc->start = READ_ONCE(fc->server_list->index); | |
99 | } | |
100 | ||
101 | found_interest: | |
102 | fc->index = fc->start; | |
103 | return true; | |
104 | } | |
105 | ||
106 | /* | |
107 | * Post volume busy note. | |
108 | */ | |
109 | static void afs_busy(struct afs_volume *volume, u32 abort_code) | |
110 | { | |
111 | const char *m; | |
112 | ||
113 | switch (abort_code) { | |
114 | case VOFFLINE: m = "offline"; break; | |
115 | case VRESTARTING: m = "restarting"; break; | |
116 | case VSALVAGING: m = "being salvaged"; break; | |
117 | default: m = "busy"; break; | |
118 | } | |
0fafdc9f | 119 | |
d2ddc776 DH |
120 | pr_notice("kAFS: Volume %u '%s' is %s\n", volume->vid, volume->name, m); |
121 | } | |
122 | ||
123 | /* | |
124 | * Sleep and retry the operation to the same fileserver. | |
125 | */ | |
126 | static bool afs_sleep_and_retry(struct afs_fs_cursor *fc) | |
127 | { | |
128 | msleep_interruptible(1000); | |
129 | if (signal_pending(current)) { | |
130 | fc->ac.error = -ERESTARTSYS; | |
131 | return false; | |
132 | } | |
133 | ||
134 | return true; | |
135 | } | |
136 | ||
137 | /* | |
138 | * Select the fileserver to use. May be called multiple times to rotate | |
139 | * through the fileservers. | |
140 | */ | |
141 | bool afs_select_fileserver(struct afs_fs_cursor *fc) | |
142 | { | |
143 | struct afs_addr_list *alist; | |
144 | struct afs_server *server; | |
145 | struct afs_vnode *vnode = fc->vnode; | |
146 | ||
147 | _enter("%u/%u,%u/%u,%d,%d", | |
148 | fc->index, fc->start, | |
149 | fc->ac.index, fc->ac.start, | |
150 | fc->ac.error, fc->ac.abort_code); | |
151 | ||
152 | if (fc->flags & AFS_FS_CURSOR_STOP) { | |
153 | _leave(" = f [stopped]"); | |
154 | return false; | |
155 | } | |
156 | ||
157 | /* Evaluate the result of the previous operation, if there was one. */ | |
158 | switch (fc->ac.error) { | |
159 | case SHRT_MAX: | |
160 | goto start; | |
161 | ||
162 | case 0: | |
163 | default: | |
164 | /* Success or local failure. Stop. */ | |
165 | fc->flags |= AFS_FS_CURSOR_STOP; | |
166 | _leave(" = f [okay/local %d]", fc->ac.error); | |
167 | return false; | |
168 | ||
169 | case -ECONNABORTED: | |
170 | /* The far side rejected the operation on some grounds. This | |
171 | * might involve the server being busy or the volume having been moved. | |
172 | */ | |
173 | switch (fc->ac.abort_code) { | |
174 | case VNOVOL: | |
175 | /* This fileserver doesn't know about the volume. | |
176 | * - May indicate that the VL is wrong - retry once and compare | |
177 | * the results. | |
178 | * - May indicate that the fileserver couldn't attach to the vol. | |
179 | */ | |
180 | if (fc->flags & AFS_FS_CURSOR_VNOVOL) { | |
181 | fc->ac.error = -EREMOTEIO; | |
182 | goto failed; | |
183 | } | |
184 | ||
185 | write_lock(&vnode->volume->servers_lock); | |
186 | fc->server_list->vnovol_mask |= 1 << fc->index; | |
187 | write_unlock(&vnode->volume->servers_lock); | |
188 | ||
189 | set_bit(AFS_VOLUME_NEEDS_UPDATE, &vnode->volume->flags); | |
190 | fc->ac.error = afs_check_volume_status(vnode->volume, fc->key); | |
191 | if (fc->ac.error < 0) | |
192 | goto failed; | |
193 | ||
194 | if (test_bit(AFS_VOLUME_DELETED, &vnode->volume->flags)) { | |
195 | fc->ac.error = -ENOMEDIUM; | |
196 | goto failed; | |
197 | } | |
198 | ||
199 | /* If the server list didn't change, then assume that | |
200 | * it's the fileserver having trouble. | |
201 | */ | |
202 | if (vnode->volume->servers == fc->server_list) { | |
203 | fc->ac.error = -EREMOTEIO; | |
204 | goto failed; | |
205 | } | |
206 | ||
207 | /* Try again */ | |
208 | fc->flags |= AFS_FS_CURSOR_VNOVOL; | |
209 | _leave(" = t [vnovol]"); | |
210 | return true; | |
211 | ||
212 | case VSALVAGE: /* TODO: Should this return an error or iterate? */ | |
213 | case VVOLEXISTS: | |
214 | case VNOSERVICE: | |
215 | case VONLINE: | |
216 | case VDISKFULL: | |
217 | case VOVERQUOTA: | |
218 | fc->ac.error = afs_abort_to_error(fc->ac.abort_code); | |
219 | goto next_server; | |
220 | ||
221 | case VOFFLINE: | |
222 | if (!test_and_set_bit(AFS_VOLUME_OFFLINE, &vnode->volume->flags)) { | |
223 | afs_busy(vnode->volume, fc->ac.abort_code); | |
224 | clear_bit(AFS_VOLUME_BUSY, &vnode->volume->flags); | |
225 | } | |
226 | if (fc->flags & AFS_FS_CURSOR_NO_VSLEEP) { | |
227 | fc->ac.error = -EADV; | |
228 | goto failed; | |
229 | } | |
230 | if (fc->flags & AFS_FS_CURSOR_CUR_ONLY) { | |
231 | fc->ac.error = -ESTALE; | |
232 | goto failed; | |
233 | } | |
234 | goto busy; | |
235 | ||
236 | case VSALVAGING: | |
237 | case VRESTARTING: | |
238 | case VBUSY: | |
239 | /* Retry after going round all the servers unless we | |
240 | * have a file lock we need to maintain. | |
241 | */ | |
242 | if (fc->flags & AFS_FS_CURSOR_NO_VSLEEP) { | |
243 | fc->ac.error = -EBUSY; | |
244 | goto failed; | |
245 | } | |
246 | if (!test_and_set_bit(AFS_VOLUME_BUSY, &vnode->volume->flags)) { | |
247 | afs_busy(vnode->volume, fc->ac.abort_code); | |
248 | clear_bit(AFS_VOLUME_OFFLINE, &vnode->volume->flags); | |
249 | } | |
250 | busy: | |
251 | if (fc->flags & AFS_FS_CURSOR_CUR_ONLY) { | |
252 | if (!afs_sleep_and_retry(fc)) | |
253 | goto failed; | |
254 | ||
255 | /* Retry with same server & address */ | |
256 | _leave(" = t [vbusy]"); | |
257 | return true; | |
258 | } | |
259 | ||
260 | fc->flags |= AFS_FS_CURSOR_VBUSY; | |
261 | goto next_server; | |
262 | ||
263 | case VMOVED: | |
264 | /* The volume migrated to another server. We consider | |
265 | * consider all locks and callbacks broken and request | |
266 | * an update from the VLDB. | |
267 | * | |
268 | * We also limit the number of VMOVED hops we will | |
269 | * honour, just in case someone sets up a loop. | |
270 | */ | |
271 | if (fc->flags & AFS_FS_CURSOR_VMOVED) { | |
272 | fc->ac.error = -EREMOTEIO; | |
273 | goto failed; | |
274 | } | |
275 | fc->flags |= AFS_FS_CURSOR_VMOVED; | |
276 | ||
277 | set_bit(AFS_VOLUME_WAIT, &vnode->volume->flags); | |
278 | set_bit(AFS_VOLUME_NEEDS_UPDATE, &vnode->volume->flags); | |
279 | fc->ac.error = afs_check_volume_status(vnode->volume, fc->key); | |
280 | if (fc->ac.error < 0) | |
281 | goto failed; | |
282 | ||
283 | /* If the server list didn't change, then the VLDB is | |
284 | * out of sync with the fileservers. This is hopefully | |
285 | * a temporary condition, however, so we don't want to | |
286 | * permanently block access to the file. | |
287 | * | |
288 | * TODO: Try other fileservers if we can. | |
289 | * | |
290 | * TODO: Retry a few times with sleeps. | |
291 | */ | |
292 | if (vnode->volume->servers == fc->server_list) { | |
293 | fc->ac.error = -ENOMEDIUM; | |
294 | goto failed; | |
295 | } | |
296 | ||
297 | goto restart_from_beginning; | |
298 | ||
299 | default: | |
300 | clear_bit(AFS_VOLUME_OFFLINE, &vnode->volume->flags); | |
301 | clear_bit(AFS_VOLUME_BUSY, &vnode->volume->flags); | |
302 | fc->ac.error = afs_abort_to_error(fc->ac.abort_code); | |
303 | goto failed; | |
304 | } | |
305 | ||
306 | case -ENETUNREACH: | |
307 | case -EHOSTUNREACH: | |
308 | case -ECONNREFUSED: | |
309 | case -ETIMEDOUT: | |
310 | case -ETIME: | |
311 | _debug("no conn"); | |
312 | goto iterate_address; | |
313 | } | |
314 | ||
315 | restart_from_beginning: | |
316 | _debug("restart"); | |
317 | afs_end_cursor(&fc->ac); | |
318 | afs_put_cb_interest(afs_v2net(vnode), fc->cbi); | |
319 | fc->cbi = NULL; | |
320 | afs_put_serverlist(afs_v2net(vnode), fc->server_list); | |
321 | fc->server_list = NULL; | |
322 | start: | |
323 | _debug("start"); | |
324 | /* See if we need to do an update of the volume record. Note that the | |
325 | * volume may have moved or even have been deleted. | |
326 | */ | |
327 | fc->ac.error = afs_check_volume_status(vnode->volume, fc->key); | |
328 | if (fc->ac.error < 0) | |
329 | goto failed; | |
330 | ||
331 | if (!afs_start_fs_iteration(fc, vnode)) | |
332 | goto failed; | |
333 | goto use_server; | |
334 | ||
335 | next_server: | |
336 | _debug("next"); | |
337 | afs_put_cb_interest(afs_v2net(vnode), fc->cbi); | |
338 | fc->cbi = NULL; | |
339 | fc->index++; | |
340 | if (fc->index >= fc->server_list->nr_servers) | |
341 | fc->index = 0; | |
342 | if (fc->index != fc->start) | |
343 | goto use_server; | |
344 | ||
345 | /* That's all the servers poked to no good effect. Try again if some | |
346 | * of them were busy. | |
347 | */ | |
348 | if (fc->flags & AFS_FS_CURSOR_VBUSY) | |
349 | goto restart_from_beginning; | |
350 | ||
351 | fc->ac.error = -EDESTADDRREQ; | |
352 | goto failed; | |
353 | ||
354 | use_server: | |
355 | _debug("use"); | |
356 | /* We're starting on a different fileserver from the list. We need to | |
357 | * check it, create a callback intercept, find its address list and | |
358 | * probe its capabilities before we use it. | |
359 | */ | |
360 | ASSERTCMP(fc->ac.alist, ==, NULL); | |
361 | server = fc->server_list->servers[fc->index].server; | |
362 | ||
363 | if (!afs_check_server_record(fc, server)) | |
364 | goto failed; | |
365 | ||
366 | _debug("USING SERVER: %pU", &server->uuid); | |
367 | ||
368 | /* Make sure we've got a callback interest record for this server. We | |
369 | * have to link it in before we send the request as we can be sent a | |
370 | * break request before we've finished decoding the reply and | |
371 | * installing the vnode. | |
372 | */ | |
373 | fc->ac.error = afs_register_server_cb_interest( | |
374 | vnode, &fc->server_list->servers[fc->index]); | |
375 | if (fc->ac.error < 0) | |
376 | goto failed; | |
377 | ||
378 | fc->cbi = afs_get_cb_interest(vnode->cb_interest); | |
379 | ||
380 | read_lock(&server->fs_lock); | |
381 | alist = rcu_dereference_protected(server->addresses, | |
382 | lockdep_is_held(&server->fs_lock)); | |
383 | afs_get_addrlist(alist); | |
384 | read_unlock(&server->fs_lock); | |
385 | ||
57ad16ab | 386 | memset(&fc->ac, 0, sizeof(fc->ac)); |
d2ddc776 DH |
387 | |
388 | /* Probe the current fileserver if we haven't done so yet. */ | |
389 | if (!test_bit(AFS_SERVER_FL_PROBED, &server->flags)) { | |
390 | fc->ac.alist = afs_get_addrlist(alist); | |
391 | ||
392 | if (!afs_probe_fileserver(fc)) | |
393 | goto failed; | |
394 | } | |
395 | ||
396 | if (!fc->ac.alist) | |
397 | fc->ac.alist = alist; | |
398 | else | |
399 | afs_put_addrlist(alist); | |
400 | ||
d2ddc776 DH |
401 | fc->ac.start = READ_ONCE(alist->index); |
402 | fc->ac.index = fc->ac.start; | |
d2ddc776 DH |
403 | goto iterate_address; |
404 | ||
405 | iterate_address: | |
406 | ASSERT(fc->ac.alist); | |
407 | _debug("iterate %d/%d", fc->ac.index, fc->ac.alist->nr_addrs); | |
408 | /* Iterate over the current server's address list to try and find an | |
409 | * address on which it will respond to us. | |
410 | */ | |
411 | if (afs_iterate_addresses(&fc->ac)) { | |
412 | _leave(" = t"); | |
413 | return true; | |
414 | } | |
415 | ||
416 | afs_end_cursor(&fc->ac); | |
417 | goto next_server; | |
418 | ||
419 | failed: | |
420 | fc->flags |= AFS_FS_CURSOR_STOP; | |
421 | _leave(" = f [failed %d]", fc->ac.error); | |
422 | return false; | |
423 | } | |
424 | ||
425 | /* | |
426 | * Select the same fileserver we used for a vnode before and only that | |
427 | * fileserver. We use this when we have a lock on that file, which is backed | |
428 | * only by the fileserver we obtained it from. | |
429 | */ | |
430 | bool afs_select_current_fileserver(struct afs_fs_cursor *fc) | |
431 | { | |
432 | struct afs_vnode *vnode = fc->vnode; | |
433 | struct afs_cb_interest *cbi = vnode->cb_interest; | |
434 | struct afs_addr_list *alist; | |
435 | ||
436 | _enter(""); | |
437 | ||
0fafdc9f DH |
438 | switch (fc->ac.error) { |
439 | case SHRT_MAX: | |
440 | if (!cbi) { | |
441 | fc->ac.error = -ESTALE; | |
442 | fc->flags |= AFS_FS_CURSOR_STOP; | |
443 | return false; | |
444 | } | |
445 | ||
446 | fc->cbi = afs_get_cb_interest(vnode->cb_interest); | |
447 | ||
448 | read_lock(&cbi->server->fs_lock); | |
449 | alist = rcu_dereference_protected(cbi->server->addresses, | |
450 | lockdep_is_held(&cbi->server->fs_lock)); | |
451 | afs_get_addrlist(alist); | |
452 | read_unlock(&cbi->server->fs_lock); | |
453 | if (!alist) { | |
454 | fc->ac.error = -ESTALE; | |
455 | fc->flags |= AFS_FS_CURSOR_STOP; | |
456 | return false; | |
457 | } | |
458 | ||
57ad16ab | 459 | memset(&fc->ac, 0, sizeof(fc->ac)); |
0fafdc9f | 460 | fc->ac.alist = alist; |
0fafdc9f DH |
461 | fc->ac.start = READ_ONCE(alist->index); |
462 | fc->ac.index = fc->ac.start; | |
0fafdc9f DH |
463 | goto iterate_address; |
464 | ||
465 | case 0: | |
466 | default: | |
467 | /* Success or local failure. Stop. */ | |
d2ddc776 | 468 | fc->flags |= AFS_FS_CURSOR_STOP; |
0fafdc9f | 469 | _leave(" = f [okay/local %d]", fc->ac.error); |
d2ddc776 | 470 | return false; |
d2ddc776 | 471 | |
0fafdc9f | 472 | case -ECONNABORTED: |
d2ddc776 | 473 | fc->flags |= AFS_FS_CURSOR_STOP; |
0fafdc9f | 474 | _leave(" = f [abort]"); |
d2ddc776 | 475 | return false; |
0fafdc9f DH |
476 | |
477 | case -ENETUNREACH: | |
478 | case -EHOSTUNREACH: | |
479 | case -ECONNREFUSED: | |
480 | case -ETIMEDOUT: | |
481 | case -ETIME: | |
482 | _debug("no conn"); | |
483 | goto iterate_address; | |
d2ddc776 DH |
484 | } |
485 | ||
0fafdc9f DH |
486 | iterate_address: |
487 | /* Iterate over the current server's address list to try and find an | |
488 | * address on which it will respond to us. | |
489 | */ | |
490 | if (afs_iterate_addresses(&fc->ac)) { | |
491 | _leave(" = t"); | |
492 | return true; | |
493 | } | |
494 | ||
495 | afs_end_cursor(&fc->ac); | |
496 | return false; | |
d2ddc776 DH |
497 | } |
498 | ||
499 | /* | |
500 | * Tidy up a filesystem cursor and unlock the vnode. | |
501 | */ | |
502 | int afs_end_vnode_operation(struct afs_fs_cursor *fc) | |
503 | { | |
504 | struct afs_net *net = afs_v2net(fc->vnode); | |
505 | int ret; | |
506 | ||
507 | mutex_unlock(&fc->vnode->io_lock); | |
508 | ||
509 | afs_end_cursor(&fc->ac); | |
510 | afs_put_cb_interest(net, fc->cbi); | |
511 | afs_put_serverlist(net, fc->server_list); | |
512 | ||
513 | ret = fc->ac.error; | |
514 | if (ret == -ECONNABORTED) | |
515 | afs_abort_to_error(fc->ac.abort_code); | |
516 | ||
517 | return fc->ac.error; | |
518 | } | |
519 | ||
520 | #if 0 | |
9cc6fc50 DH |
521 | /* |
522 | * Set a filesystem server cursor for using a specific FS server. | |
523 | */ | |
524 | int afs_set_fs_cursor(struct afs_fs_cursor *fc, struct afs_vnode *vnode) | |
525 | { | |
526 | afs_init_fs_cursor(fc, vnode); | |
527 | ||
528 | read_seqlock_excl(&vnode->cb_lock); | |
529 | if (vnode->cb_interest) { | |
530 | if (vnode->cb_interest->server->fs_state == 0) | |
531 | fc->server = afs_get_server(vnode->cb_interest->server); | |
532 | else | |
533 | fc->ac.error = vnode->cb_interest->server->fs_state; | |
534 | } else { | |
535 | fc->ac.error = -ESTALE; | |
536 | } | |
537 | read_sequnlock_excl(&vnode->cb_lock); | |
538 | ||
539 | return fc->ac.error; | |
540 | } | |
541 | ||
542 | /* | |
543 | * pick a server to use to try accessing this volume | |
544 | * - returns with an elevated usage count on the server chosen | |
545 | */ | |
546 | bool afs_volume_pick_fileserver(struct afs_fs_cursor *fc, struct afs_vnode *vnode) | |
547 | { | |
548 | struct afs_volume *volume = vnode->volume; | |
549 | struct afs_server *server; | |
550 | int ret, state, loop; | |
551 | ||
552 | _enter("%s", volume->vlocation->vldb.name); | |
553 | ||
554 | /* stick with the server we're already using if we can */ | |
555 | if (vnode->cb_interest && vnode->cb_interest->server->fs_state == 0) { | |
556 | fc->server = afs_get_server(vnode->cb_interest->server); | |
557 | goto set_server; | |
558 | } | |
559 | ||
560 | down_read(&volume->server_sem); | |
561 | ||
562 | /* handle the no-server case */ | |
563 | if (volume->nservers == 0) { | |
564 | fc->ac.error = volume->rjservers ? -ENOMEDIUM : -ESTALE; | |
565 | up_read(&volume->server_sem); | |
566 | _leave(" = f [no servers %d]", fc->ac.error); | |
567 | return false; | |
568 | } | |
569 | ||
570 | /* basically, just search the list for the first live server and use | |
571 | * that */ | |
572 | ret = 0; | |
573 | for (loop = 0; loop < volume->nservers; loop++) { | |
574 | server = volume->servers[loop]; | |
575 | state = server->fs_state; | |
576 | ||
577 | _debug("consider %d [%d]", loop, state); | |
578 | ||
579 | switch (state) { | |
580 | case 0: | |
581 | goto picked_server; | |
582 | ||
583 | case -ENETUNREACH: | |
584 | if (ret == 0) | |
585 | ret = state; | |
586 | break; | |
587 | ||
588 | case -EHOSTUNREACH: | |
589 | if (ret == 0 || | |
590 | ret == -ENETUNREACH) | |
591 | ret = state; | |
592 | break; | |
593 | ||
594 | case -ECONNREFUSED: | |
595 | if (ret == 0 || | |
596 | ret == -ENETUNREACH || | |
597 | ret == -EHOSTUNREACH) | |
598 | ret = state; | |
599 | break; | |
600 | ||
601 | default: | |
602 | case -EREMOTEIO: | |
603 | if (ret == 0 || | |
604 | ret == -ENETUNREACH || | |
605 | ret == -EHOSTUNREACH || | |
606 | ret == -ECONNREFUSED) | |
607 | ret = state; | |
608 | break; | |
609 | } | |
610 | } | |
611 | ||
612 | error: | |
613 | fc->ac.error = ret; | |
614 | ||
615 | /* no available servers | |
616 | * - TODO: handle the no active servers case better | |
617 | */ | |
618 | up_read(&volume->server_sem); | |
619 | _leave(" = f [%d]", fc->ac.error); | |
620 | return false; | |
621 | ||
622 | picked_server: | |
623 | /* Found an apparently healthy server. We need to register an interest | |
624 | * in receiving callbacks before we talk to it. | |
625 | */ | |
626 | ret = afs_register_server_cb_interest(vnode, | |
627 | &volume->cb_interests[loop], server); | |
628 | if (ret < 0) | |
629 | goto error; | |
630 | ||
631 | fc->server = afs_get_server(server); | |
632 | up_read(&volume->server_sem); | |
633 | set_server: | |
634 | fc->ac.alist = afs_get_addrlist(fc->server->addrs); | |
635 | fc->ac.addr = &fc->ac.alist->addrs[0]; | |
636 | _debug("USING SERVER: %pIS\n", &fc->ac.addr->transport); | |
637 | _leave(" = t (picked %pIS)", &fc->ac.addr->transport); | |
638 | return true; | |
639 | } | |
640 | ||
641 | /* | |
642 | * release a server after use | |
643 | * - releases the ref on the server struct that was acquired by picking | |
644 | * - records result of using a particular server to access a volume | |
645 | * - return true to try again, false if okay or to issue error | |
646 | * - the caller must release the server struct if result was false | |
647 | */ | |
648 | bool afs_iterate_fs_cursor(struct afs_fs_cursor *fc, | |
649 | struct afs_vnode *vnode) | |
650 | { | |
651 | struct afs_volume *volume = vnode->volume; | |
652 | struct afs_server *server = fc->server; | |
653 | unsigned loop; | |
654 | ||
655 | _enter("%s,%pIS,%d", | |
656 | volume->vlocation->vldb.name, &fc->ac.addr->transport, | |
657 | fc->ac.error); | |
658 | ||
659 | switch (fc->ac.error) { | |
660 | /* success */ | |
661 | case 0: | |
662 | server->fs_state = 0; | |
663 | _leave(" = f"); | |
664 | return false; | |
665 | ||
666 | /* the fileserver denied all knowledge of the volume */ | |
667 | case -ENOMEDIUM: | |
668 | down_write(&volume->server_sem); | |
669 | ||
670 | /* firstly, find where the server is in the active list (if it | |
671 | * is) */ | |
672 | for (loop = 0; loop < volume->nservers; loop++) | |
673 | if (volume->servers[loop] == server) | |
674 | goto present; | |
675 | ||
676 | /* no longer there - may have been discarded by another op */ | |
677 | goto try_next_server_upw; | |
678 | ||
679 | present: | |
680 | volume->nservers--; | |
681 | memmove(&volume->servers[loop], | |
682 | &volume->servers[loop + 1], | |
683 | sizeof(volume->servers[loop]) * | |
684 | (volume->nservers - loop)); | |
685 | volume->servers[volume->nservers] = NULL; | |
686 | afs_put_server(afs_v2net(vnode), server); | |
687 | volume->rjservers++; | |
688 | ||
689 | if (volume->nservers > 0) | |
690 | /* another server might acknowledge its existence */ | |
691 | goto try_next_server_upw; | |
692 | ||
693 | /* handle the case where all the fileservers have rejected the | |
694 | * volume | |
695 | * - TODO: try asking the fileservers for volume information | |
696 | * - TODO: contact the VL server again to see if the volume is | |
697 | * no longer registered | |
698 | */ | |
699 | up_write(&volume->server_sem); | |
700 | afs_put_server(afs_v2net(vnode), server); | |
701 | fc->server = NULL; | |
702 | _leave(" = f [completely rejected]"); | |
703 | return false; | |
704 | ||
705 | /* problem reaching the server */ | |
706 | case -ENETUNREACH: | |
707 | case -EHOSTUNREACH: | |
708 | case -ECONNREFUSED: | |
709 | case -ETIME: | |
710 | case -ETIMEDOUT: | |
711 | case -EREMOTEIO: | |
712 | /* mark the server as dead | |
713 | * TODO: vary dead timeout depending on error | |
714 | */ | |
715 | spin_lock(&server->fs_lock); | |
716 | if (!server->fs_state) { | |
717 | server->fs_state = fc->ac.error; | |
718 | printk("kAFS: SERVER DEAD state=%d\n", fc->ac.error); | |
719 | } | |
720 | spin_unlock(&server->fs_lock); | |
721 | goto try_next_server; | |
722 | ||
723 | /* miscellaneous error */ | |
724 | default: | |
725 | case -ENOMEM: | |
726 | case -ENONET: | |
727 | /* tell the caller to accept the result */ | |
728 | afs_put_server(afs_v2net(vnode), server); | |
729 | fc->server = NULL; | |
730 | _leave(" = f [local failure]"); | |
731 | return false; | |
732 | } | |
733 | ||
734 | /* tell the caller to loop around and try the next server */ | |
735 | try_next_server_upw: | |
736 | up_write(&volume->server_sem); | |
737 | try_next_server: | |
738 | afs_put_server(afs_v2net(vnode), server); | |
739 | _leave(" = t [try next server]"); | |
740 | return true; | |
741 | } | |
742 | ||
743 | /* | |
744 | * Clean up a fileserver cursor. | |
745 | */ | |
746 | int afs_end_fs_cursor(struct afs_fs_cursor *fc, struct afs_net *net) | |
747 | { | |
748 | afs_end_cursor(&fc->ac); | |
749 | afs_put_server(net, fc->server); | |
750 | return fc->ac.error; | |
751 | } | |
d2ddc776 DH |
752 | |
753 | #endif |