]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blob - net/ceph/crush/mapper.c
crush: remove an obsolete comment
[mirror_ubuntu-bionic-kernel.git] / net / ceph / crush / mapper.c
1 /*
2 * Ceph - scalable distributed file system
3 *
4 * Copyright (C) 2015 Intel Corporation All Rights Reserved
5 *
6 * This is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License version 2.1, as published by the Free Software
9 * Foundation. See file COPYING.
10 *
11 */
12
13 #ifdef __KERNEL__
14 # include <linux/string.h>
15 # include <linux/slab.h>
16 # include <linux/bug.h>
17 # include <linux/kernel.h>
18 # include <linux/crush/crush.h>
19 # include <linux/crush/hash.h>
20 # include <linux/crush/mapper.h>
21 #else
22 # include "crush_compat.h"
23 # include "crush.h"
24 # include "hash.h"
25 # include "mapper.h"
26 #endif
27 #include "crush_ln_table.h"
28
29 #define dprintk(args...) /* printf(args) */
30
31 /*
32 * Implement the core CRUSH mapping algorithm.
33 */
34
35 /**
36 * crush_find_rule - find a crush_rule id for a given ruleset, type, and size.
37 * @map: the crush_map
38 * @ruleset: the storage ruleset id (user defined)
39 * @type: storage ruleset type (user defined)
40 * @size: output set size
41 */
42 int crush_find_rule(const struct crush_map *map, int ruleset, int type, int size)
43 {
44 __u32 i;
45
46 for (i = 0; i < map->max_rules; i++) {
47 if (map->rules[i] &&
48 map->rules[i]->mask.ruleset == ruleset &&
49 map->rules[i]->mask.type == type &&
50 map->rules[i]->mask.min_size <= size &&
51 map->rules[i]->mask.max_size >= size)
52 return i;
53 }
54 return -1;
55 }
56
57 /*
58 * bucket choose methods
59 *
60 * For each bucket algorithm, we have a "choose" method that, given a
61 * crush input @x and replica position (usually, position in output set) @r,
62 * will produce an item in the bucket.
63 */
64
65 /*
66 * Choose based on a random permutation of the bucket.
67 *
68 * We used to use some prime number arithmetic to do this, but it
69 * wasn't very random, and had some other bad behaviors. Instead, we
70 * calculate an actual random permutation of the bucket members.
71 * Since this is expensive, we optimize for the r=0 case, which
72 * captures the vast majority of calls.
73 */
74 static int bucket_perm_choose(const struct crush_bucket *bucket,
75 struct crush_work_bucket *work,
76 int x, int r)
77 {
78 unsigned int pr = r % bucket->size;
79 unsigned int i, s;
80
81 /* start a new permutation if @x has changed */
82 if (work->perm_x != (__u32)x || work->perm_n == 0) {
83 dprintk("bucket %d new x=%d\n", bucket->id, x);
84 work->perm_x = x;
85
86 /* optimize common r=0 case */
87 if (pr == 0) {
88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) %
89 bucket->size;
90 work->perm[0] = s;
91 work->perm_n = 0xffff; /* magic value, see below */
92 goto out;
93 }
94
95 for (i = 0; i < bucket->size; i++)
96 work->perm[i] = i;
97 work->perm_n = 0;
98 } else if (work->perm_n == 0xffff) {
99 /* clean up after the r=0 case above */
100 for (i = 1; i < bucket->size; i++)
101 work->perm[i] = i;
102 work->perm[work->perm[0]] = 0;
103 work->perm_n = 1;
104 }
105
106 /* calculate permutation up to pr */
107 for (i = 0; i < work->perm_n; i++)
108 dprintk(" perm_choose have %d: %d\n", i, work->perm[i]);
109 while (work->perm_n <= pr) {
110 unsigned int p = work->perm_n;
111 /* no point in swapping the final entry */
112 if (p < bucket->size - 1) {
113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) %
114 (bucket->size - p);
115 if (i) {
116 unsigned int t = work->perm[p + i];
117 work->perm[p + i] = work->perm[p];
118 work->perm[p] = t;
119 }
120 dprintk(" perm_choose swap %d with %d\n", p, p+i);
121 }
122 work->perm_n++;
123 }
124 for (i = 0; i < bucket->size; i++)
125 dprintk(" perm_choose %d: %d\n", i, work->perm[i]);
126
127 s = work->perm[pr];
128 out:
129 dprintk(" perm_choose %d sz=%d x=%d r=%d (%d) s=%d\n", bucket->id,
130 bucket->size, x, r, pr, s);
131 return bucket->items[s];
132 }
133
134 /* uniform */
135 static int bucket_uniform_choose(const struct crush_bucket_uniform *bucket,
136 struct crush_work_bucket *work, int x, int r)
137 {
138 return bucket_perm_choose(&bucket->h, work, x, r);
139 }
140
141 /* list */
142 static int bucket_list_choose(const struct crush_bucket_list *bucket,
143 int x, int r)
144 {
145 int i;
146
147 for (i = bucket->h.size-1; i >= 0; i--) {
148 __u64 w = crush_hash32_4(bucket->h.hash, x, bucket->h.items[i],
149 r, bucket->h.id);
150 w &= 0xffff;
151 dprintk("list_choose i=%d x=%d r=%d item %d weight %x "
152 "sw %x rand %llx",
153 i, x, r, bucket->h.items[i], bucket->item_weights[i],
154 bucket->sum_weights[i], w);
155 w *= bucket->sum_weights[i];
156 w = w >> 16;
157 /*dprintk(" scaled %llx\n", w);*/
158 if (w < bucket->item_weights[i]) {
159 return bucket->h.items[i];
160 }
161 }
162
163 dprintk("bad list sums for bucket %d\n", bucket->h.id);
164 return bucket->h.items[0];
165 }
166
167
168 /* (binary) tree */
169 static int height(int n)
170 {
171 int h = 0;
172 while ((n & 1) == 0) {
173 h++;
174 n = n >> 1;
175 }
176 return h;
177 }
178
179 static int left(int x)
180 {
181 int h = height(x);
182 return x - (1 << (h-1));
183 }
184
185 static int right(int x)
186 {
187 int h = height(x);
188 return x + (1 << (h-1));
189 }
190
191 static int terminal(int x)
192 {
193 return x & 1;
194 }
195
196 static int bucket_tree_choose(const struct crush_bucket_tree *bucket,
197 int x, int r)
198 {
199 int n;
200 __u32 w;
201 __u64 t;
202
203 /* start at root */
204 n = bucket->num_nodes >> 1;
205
206 while (!terminal(n)) {
207 int l;
208 /* pick point in [0, w) */
209 w = bucket->node_weights[n];
210 t = (__u64)crush_hash32_4(bucket->h.hash, x, n, r,
211 bucket->h.id) * (__u64)w;
212 t = t >> 32;
213
214 /* descend to the left or right? */
215 l = left(n);
216 if (t < bucket->node_weights[l])
217 n = l;
218 else
219 n = right(n);
220 }
221
222 return bucket->h.items[n >> 1];
223 }
224
225
226 /* straw */
227
228 static int bucket_straw_choose(const struct crush_bucket_straw *bucket,
229 int x, int r)
230 {
231 __u32 i;
232 int high = 0;
233 __u64 high_draw = 0;
234 __u64 draw;
235
236 for (i = 0; i < bucket->h.size; i++) {
237 draw = crush_hash32_3(bucket->h.hash, x, bucket->h.items[i], r);
238 draw &= 0xffff;
239 draw *= bucket->straws[i];
240 if (i == 0 || draw > high_draw) {
241 high = i;
242 high_draw = draw;
243 }
244 }
245 return bucket->h.items[high];
246 }
247
248 /* compute 2^44*log2(input+1) */
249 static __u64 crush_ln(unsigned int xin)
250 {
251 unsigned int x = xin;
252 int iexpon, index1, index2;
253 __u64 RH, LH, LL, xl64, result;
254
255 x++;
256
257 /* normalize input */
258 iexpon = 15;
259
260 /*
261 * figure out number of bits we need to shift and
262 * do it in one step instead of iteratively
263 */
264 if (!(x & 0x18000)) {
265 int bits = __builtin_clz(x & 0x1FFFF) - 16;
266 x <<= bits;
267 iexpon = 15 - bits;
268 }
269
270 index1 = (x >> 8) << 1;
271 /* RH ~ 2^56/index1 */
272 RH = __RH_LH_tbl[index1 - 256];
273 /* LH ~ 2^48 * log2(index1/256) */
274 LH = __RH_LH_tbl[index1 + 1 - 256];
275
276 /* RH*x ~ 2^48 * (2^15 + xf), xf<2^8 */
277 xl64 = (__s64)x * RH;
278 xl64 >>= 48;
279
280 result = iexpon;
281 result <<= (12 + 32);
282
283 index2 = xl64 & 0xff;
284 /* LL ~ 2^48*log2(1.0+index2/2^15) */
285 LL = __LL_tbl[index2];
286
287 LH = LH + LL;
288
289 LH >>= (48 - 12 - 32);
290 result += LH;
291
292 return result;
293 }
294
295
296 /*
297 * straw2
298 *
299 * for reference, see:
300 *
301 * http://en.wikipedia.org/wiki/Exponential_distribution#Distribution_of_the_minimum_of_exponential_random_variables
302 *
303 */
304
305 static __u32 *get_choose_arg_weights(const struct crush_bucket_straw2 *bucket,
306 const struct crush_choose_arg *arg,
307 int position)
308 {
309 if (!arg || !arg->weight_set || arg->weight_set_size == 0)
310 return bucket->item_weights;
311
312 if (position >= arg->weight_set_size)
313 position = arg->weight_set_size - 1;
314 return arg->weight_set[position].weights;
315 }
316
317 static __s32 *get_choose_arg_ids(const struct crush_bucket_straw2 *bucket,
318 const struct crush_choose_arg *arg)
319 {
320 if (!arg || !arg->ids)
321 return bucket->h.items;
322
323 return arg->ids;
324 }
325
326 static int bucket_straw2_choose(const struct crush_bucket_straw2 *bucket,
327 int x, int r,
328 const struct crush_choose_arg *arg,
329 int position)
330 {
331 unsigned int i, high = 0;
332 unsigned int u;
333 __s64 ln, draw, high_draw = 0;
334 __u32 *weights = get_choose_arg_weights(bucket, arg, position);
335 __s32 *ids = get_choose_arg_ids(bucket, arg);
336
337 for (i = 0; i < bucket->h.size; i++) {
338 dprintk("weight 0x%x item %d\n", weights[i], ids[i]);
339 if (weights[i]) {
340 u = crush_hash32_3(bucket->h.hash, x, ids[i], r);
341 u &= 0xffff;
342
343 /*
344 * for some reason slightly less than 0x10000 produces
345 * a slightly more accurate distribution... probably a
346 * rounding effect.
347 *
348 * the natural log lookup table maps [0,0xffff]
349 * (corresponding to real numbers [1/0x10000, 1] to
350 * [0, 0xffffffffffff] (corresponding to real numbers
351 * [-11.090355,0]).
352 */
353 ln = crush_ln(u) - 0x1000000000000ll;
354
355 /*
356 * divide by 16.16 fixed-point weight. note
357 * that the ln value is negative, so a larger
358 * weight means a larger (less negative) value
359 * for draw.
360 */
361 draw = div64_s64(ln, weights[i]);
362 } else {
363 draw = S64_MIN;
364 }
365
366 if (i == 0 || draw > high_draw) {
367 high = i;
368 high_draw = draw;
369 }
370 }
371
372 return bucket->h.items[high];
373 }
374
375
376 static int crush_bucket_choose(const struct crush_bucket *in,
377 struct crush_work_bucket *work,
378 int x, int r,
379 const struct crush_choose_arg *arg,
380 int position)
381 {
382 dprintk(" crush_bucket_choose %d x=%d r=%d\n", in->id, x, r);
383 BUG_ON(in->size == 0);
384 switch (in->alg) {
385 case CRUSH_BUCKET_UNIFORM:
386 return bucket_uniform_choose(
387 (const struct crush_bucket_uniform *)in,
388 work, x, r);
389 case CRUSH_BUCKET_LIST:
390 return bucket_list_choose((const struct crush_bucket_list *)in,
391 x, r);
392 case CRUSH_BUCKET_TREE:
393 return bucket_tree_choose((const struct crush_bucket_tree *)in,
394 x, r);
395 case CRUSH_BUCKET_STRAW:
396 return bucket_straw_choose(
397 (const struct crush_bucket_straw *)in,
398 x, r);
399 case CRUSH_BUCKET_STRAW2:
400 return bucket_straw2_choose(
401 (const struct crush_bucket_straw2 *)in,
402 x, r, arg, position);
403 default:
404 dprintk("unknown bucket %d alg %d\n", in->id, in->alg);
405 return in->items[0];
406 }
407 }
408
409 /*
410 * true if device is marked "out" (failed, fully offloaded)
411 * of the cluster
412 */
413 static int is_out(const struct crush_map *map,
414 const __u32 *weight, int weight_max,
415 int item, int x)
416 {
417 if (item >= weight_max)
418 return 1;
419 if (weight[item] >= 0x10000)
420 return 0;
421 if (weight[item] == 0)
422 return 1;
423 if ((crush_hash32_2(CRUSH_HASH_RJENKINS1, x, item) & 0xffff)
424 < weight[item])
425 return 0;
426 return 1;
427 }
428
429 /**
430 * crush_choose_firstn - choose numrep distinct items of given type
431 * @map: the crush_map
432 * @bucket: the bucket we are choose an item from
433 * @x: crush input value
434 * @numrep: the number of items to choose
435 * @type: the type of item to choose
436 * @out: pointer to output vector
437 * @outpos: our position in that vector
438 * @out_size: size of the out vector
439 * @tries: number of attempts to make
440 * @recurse_tries: number of attempts to have recursive chooseleaf make
441 * @local_retries: localized retries
442 * @local_fallback_retries: localized fallback retries
443 * @recurse_to_leaf: true if we want one device under each item of given type (chooseleaf instead of choose)
444 * @stable: stable mode starts rep=0 in the recursive call for all replicas
445 * @vary_r: pass r to recursive calls
446 * @out2: second output vector for leaf items (if @recurse_to_leaf)
447 * @parent_r: r value passed from the parent
448 */
449 static int crush_choose_firstn(const struct crush_map *map,
450 struct crush_work *work,
451 const struct crush_bucket *bucket,
452 const __u32 *weight, int weight_max,
453 int x, int numrep, int type,
454 int *out, int outpos,
455 int out_size,
456 unsigned int tries,
457 unsigned int recurse_tries,
458 unsigned int local_retries,
459 unsigned int local_fallback_retries,
460 int recurse_to_leaf,
461 unsigned int vary_r,
462 unsigned int stable,
463 int *out2,
464 int parent_r,
465 const struct crush_choose_arg *choose_args)
466 {
467 int rep;
468 unsigned int ftotal, flocal;
469 int retry_descent, retry_bucket, skip_rep;
470 const struct crush_bucket *in = bucket;
471 int r;
472 int i;
473 int item = 0;
474 int itemtype;
475 int collide, reject;
476 int count = out_size;
477
478 dprintk("CHOOSE%s bucket %d x %d outpos %d numrep %d tries %d recurse_tries %d local_retries %d local_fallback_retries %d parent_r %d stable %d\n",
479 recurse_to_leaf ? "_LEAF" : "",
480 bucket->id, x, outpos, numrep,
481 tries, recurse_tries, local_retries, local_fallback_retries,
482 parent_r, stable);
483
484 for (rep = stable ? 0 : outpos; rep < numrep && count > 0 ; rep++) {
485 /* keep trying until we get a non-out, non-colliding item */
486 ftotal = 0;
487 skip_rep = 0;
488 do {
489 retry_descent = 0;
490 in = bucket; /* initial bucket */
491
492 /* choose through intervening buckets */
493 flocal = 0;
494 do {
495 collide = 0;
496 retry_bucket = 0;
497 r = rep + parent_r;
498 /* r' = r + f_total */
499 r += ftotal;
500
501 /* bucket choose */
502 if (in->size == 0) {
503 reject = 1;
504 goto reject;
505 }
506 if (local_fallback_retries > 0 &&
507 flocal >= (in->size>>1) &&
508 flocal > local_fallback_retries)
509 item = bucket_perm_choose(
510 in, work->work[-1-in->id],
511 x, r);
512 else
513 item = crush_bucket_choose(
514 in, work->work[-1-in->id],
515 x, r,
516 (choose_args ?
517 &choose_args[-1-in->id] : 0),
518 outpos);
519 if (item >= map->max_devices) {
520 dprintk(" bad item %d\n", item);
521 skip_rep = 1;
522 break;
523 }
524
525 /* desired type? */
526 if (item < 0)
527 itemtype = map->buckets[-1-item]->type;
528 else
529 itemtype = 0;
530 dprintk(" item %d type %d\n", item, itemtype);
531
532 /* keep going? */
533 if (itemtype != type) {
534 if (item >= 0 ||
535 (-1-item) >= map->max_buckets) {
536 dprintk(" bad item type %d\n", type);
537 skip_rep = 1;
538 break;
539 }
540 in = map->buckets[-1-item];
541 retry_bucket = 1;
542 continue;
543 }
544
545 /* collision? */
546 for (i = 0; i < outpos; i++) {
547 if (out[i] == item) {
548 collide = 1;
549 break;
550 }
551 }
552
553 reject = 0;
554 if (!collide && recurse_to_leaf) {
555 if (item < 0) {
556 int sub_r;
557 if (vary_r)
558 sub_r = r >> (vary_r-1);
559 else
560 sub_r = 0;
561 if (crush_choose_firstn(
562 map,
563 work,
564 map->buckets[-1-item],
565 weight, weight_max,
566 x, stable ? 1 : outpos+1, 0,
567 out2, outpos, count,
568 recurse_tries, 0,
569 local_retries,
570 local_fallback_retries,
571 0,
572 vary_r,
573 stable,
574 NULL,
575 sub_r,
576 choose_args) <= outpos)
577 /* didn't get leaf */
578 reject = 1;
579 } else {
580 /* we already have a leaf! */
581 out2[outpos] = item;
582 }
583 }
584
585 if (!reject && !collide) {
586 /* out? */
587 if (itemtype == 0)
588 reject = is_out(map, weight,
589 weight_max,
590 item, x);
591 }
592
593 reject:
594 if (reject || collide) {
595 ftotal++;
596 flocal++;
597
598 if (collide && flocal <= local_retries)
599 /* retry locally a few times */
600 retry_bucket = 1;
601 else if (local_fallback_retries > 0 &&
602 flocal <= in->size + local_fallback_retries)
603 /* exhaustive bucket search */
604 retry_bucket = 1;
605 else if (ftotal < tries)
606 /* then retry descent */
607 retry_descent = 1;
608 else
609 /* else give up */
610 skip_rep = 1;
611 dprintk(" reject %d collide %d "
612 "ftotal %u flocal %u\n",
613 reject, collide, ftotal,
614 flocal);
615 }
616 } while (retry_bucket);
617 } while (retry_descent);
618
619 if (skip_rep) {
620 dprintk("skip rep\n");
621 continue;
622 }
623
624 dprintk("CHOOSE got %d\n", item);
625 out[outpos] = item;
626 outpos++;
627 count--;
628 #ifndef __KERNEL__
629 if (map->choose_tries && ftotal <= map->choose_total_tries)
630 map->choose_tries[ftotal]++;
631 #endif
632 }
633
634 dprintk("CHOOSE returns %d\n", outpos);
635 return outpos;
636 }
637
638
639 /**
640 * crush_choose_indep: alternative breadth-first positionally stable mapping
641 *
642 */
643 static void crush_choose_indep(const struct crush_map *map,
644 struct crush_work *work,
645 const struct crush_bucket *bucket,
646 const __u32 *weight, int weight_max,
647 int x, int left, int numrep, int type,
648 int *out, int outpos,
649 unsigned int tries,
650 unsigned int recurse_tries,
651 int recurse_to_leaf,
652 int *out2,
653 int parent_r,
654 const struct crush_choose_arg *choose_args)
655 {
656 const struct crush_bucket *in = bucket;
657 int endpos = outpos + left;
658 int rep;
659 unsigned int ftotal;
660 int r;
661 int i;
662 int item = 0;
663 int itemtype;
664 int collide;
665
666 dprintk("CHOOSE%s INDEP bucket %d x %d outpos %d numrep %d\n", recurse_to_leaf ? "_LEAF" : "",
667 bucket->id, x, outpos, numrep);
668
669 /* initially my result is undefined */
670 for (rep = outpos; rep < endpos; rep++) {
671 out[rep] = CRUSH_ITEM_UNDEF;
672 if (out2)
673 out2[rep] = CRUSH_ITEM_UNDEF;
674 }
675
676 for (ftotal = 0; left > 0 && ftotal < tries; ftotal++) {
677 #ifdef DEBUG_INDEP
678 if (out2 && ftotal) {
679 dprintk("%u %d a: ", ftotal, left);
680 for (rep = outpos; rep < endpos; rep++) {
681 dprintk(" %d", out[rep]);
682 }
683 dprintk("\n");
684 dprintk("%u %d b: ", ftotal, left);
685 for (rep = outpos; rep < endpos; rep++) {
686 dprintk(" %d", out2[rep]);
687 }
688 dprintk("\n");
689 }
690 #endif
691 for (rep = outpos; rep < endpos; rep++) {
692 if (out[rep] != CRUSH_ITEM_UNDEF)
693 continue;
694
695 in = bucket; /* initial bucket */
696
697 /* choose through intervening buckets */
698 for (;;) {
699 /* note: we base the choice on the position
700 * even in the nested call. that means that
701 * if the first layer chooses the same bucket
702 * in a different position, we will tend to
703 * choose a different item in that bucket.
704 * this will involve more devices in data
705 * movement and tend to distribute the load.
706 */
707 r = rep + parent_r;
708
709 /* be careful */
710 if (in->alg == CRUSH_BUCKET_UNIFORM &&
711 in->size % numrep == 0)
712 /* r'=r+(n+1)*f_total */
713 r += (numrep+1) * ftotal;
714 else
715 /* r' = r + n*f_total */
716 r += numrep * ftotal;
717
718 /* bucket choose */
719 if (in->size == 0) {
720 dprintk(" empty bucket\n");
721 break;
722 }
723
724 item = crush_bucket_choose(
725 in, work->work[-1-in->id],
726 x, r,
727 (choose_args ?
728 &choose_args[-1-in->id] : 0),
729 outpos);
730 if (item >= map->max_devices) {
731 dprintk(" bad item %d\n", item);
732 out[rep] = CRUSH_ITEM_NONE;
733 if (out2)
734 out2[rep] = CRUSH_ITEM_NONE;
735 left--;
736 break;
737 }
738
739 /* desired type? */
740 if (item < 0)
741 itemtype = map->buckets[-1-item]->type;
742 else
743 itemtype = 0;
744 dprintk(" item %d type %d\n", item, itemtype);
745
746 /* keep going? */
747 if (itemtype != type) {
748 if (item >= 0 ||
749 (-1-item) >= map->max_buckets) {
750 dprintk(" bad item type %d\n", type);
751 out[rep] = CRUSH_ITEM_NONE;
752 if (out2)
753 out2[rep] =
754 CRUSH_ITEM_NONE;
755 left--;
756 break;
757 }
758 in = map->buckets[-1-item];
759 continue;
760 }
761
762 /* collision? */
763 collide = 0;
764 for (i = outpos; i < endpos; i++) {
765 if (out[i] == item) {
766 collide = 1;
767 break;
768 }
769 }
770 if (collide)
771 break;
772
773 if (recurse_to_leaf) {
774 if (item < 0) {
775 crush_choose_indep(
776 map,
777 work,
778 map->buckets[-1-item],
779 weight, weight_max,
780 x, 1, numrep, 0,
781 out2, rep,
782 recurse_tries, 0,
783 0, NULL, r,
784 choose_args);
785 if (out2[rep] == CRUSH_ITEM_NONE) {
786 /* placed nothing; no leaf */
787 break;
788 }
789 } else {
790 /* we already have a leaf! */
791 out2[rep] = item;
792 }
793 }
794
795 /* out? */
796 if (itemtype == 0 &&
797 is_out(map, weight, weight_max, item, x))
798 break;
799
800 /* yay! */
801 out[rep] = item;
802 left--;
803 break;
804 }
805 }
806 }
807 for (rep = outpos; rep < endpos; rep++) {
808 if (out[rep] == CRUSH_ITEM_UNDEF) {
809 out[rep] = CRUSH_ITEM_NONE;
810 }
811 if (out2 && out2[rep] == CRUSH_ITEM_UNDEF) {
812 out2[rep] = CRUSH_ITEM_NONE;
813 }
814 }
815 #ifndef __KERNEL__
816 if (map->choose_tries && ftotal <= map->choose_total_tries)
817 map->choose_tries[ftotal]++;
818 #endif
819 #ifdef DEBUG_INDEP
820 if (out2) {
821 dprintk("%u %d a: ", ftotal, left);
822 for (rep = outpos; rep < endpos; rep++) {
823 dprintk(" %d", out[rep]);
824 }
825 dprintk("\n");
826 dprintk("%u %d b: ", ftotal, left);
827 for (rep = outpos; rep < endpos; rep++) {
828 dprintk(" %d", out2[rep]);
829 }
830 dprintk("\n");
831 }
832 #endif
833 }
834
835
836 /*
837 * This takes a chunk of memory and sets it up to be a shiny new
838 * working area for a CRUSH placement computation. It must be called
839 * on any newly allocated memory before passing it in to
840 * crush_do_rule. It may be used repeatedly after that, so long as the
841 * map has not changed. If the map /has/ changed, you must make sure
842 * the working size is no smaller than what was allocated and re-run
843 * crush_init_workspace.
844 *
845 * If you do retain the working space between calls to crush, make it
846 * thread-local.
847 */
848 void crush_init_workspace(const struct crush_map *map, void *v)
849 {
850 struct crush_work *w = v;
851 __s32 b;
852
853 /*
854 * We work by moving through the available space and setting
855 * values and pointers as we go.
856 *
857 * It's a bit like Forth's use of the 'allot' word since we
858 * set the pointer first and then reserve the space for it to
859 * point to by incrementing the point.
860 */
861 v += sizeof(struct crush_work);
862 w->work = v;
863 v += map->max_buckets * sizeof(struct crush_work_bucket *);
864 for (b = 0; b < map->max_buckets; ++b) {
865 if (!map->buckets[b])
866 continue;
867
868 w->work[b] = v;
869 switch (map->buckets[b]->alg) {
870 default:
871 v += sizeof(struct crush_work_bucket);
872 break;
873 }
874 w->work[b]->perm_x = 0;
875 w->work[b]->perm_n = 0;
876 w->work[b]->perm = v;
877 v += map->buckets[b]->size * sizeof(__u32);
878 }
879 BUG_ON(v - (void *)w != map->working_size);
880 }
881
882 /**
883 * crush_do_rule - calculate a mapping with the given input and rule
884 * @map: the crush_map
885 * @ruleno: the rule id
886 * @x: hash input
887 * @result: pointer to result vector
888 * @result_max: maximum result size
889 * @weight: weight vector (for map leaves)
890 * @weight_max: size of weight vector
891 * @cwin: pointer to at least crush_work_size() bytes of memory
892 * @choose_args: weights and ids for each known bucket
893 */
894 int crush_do_rule(const struct crush_map *map,
895 int ruleno, int x, int *result, int result_max,
896 const __u32 *weight, int weight_max,
897 void *cwin, const struct crush_choose_arg *choose_args)
898 {
899 int result_len;
900 struct crush_work *cw = cwin;
901 int *a = cwin + map->working_size;
902 int *b = a + result_max;
903 int *c = b + result_max;
904 int *w = a;
905 int *o = b;
906 int recurse_to_leaf;
907 int wsize = 0;
908 int osize;
909 int *tmp;
910 const struct crush_rule *rule;
911 __u32 step;
912 int i, j;
913 int numrep;
914 int out_size;
915 /*
916 * the original choose_total_tries value was off by one (it
917 * counted "retries" and not "tries"). add one.
918 */
919 int choose_tries = map->choose_total_tries + 1;
920 int choose_leaf_tries = 0;
921 /*
922 * the local tries values were counted as "retries", though,
923 * and need no adjustment
924 */
925 int choose_local_retries = map->choose_local_tries;
926 int choose_local_fallback_retries = map->choose_local_fallback_tries;
927
928 int vary_r = map->chooseleaf_vary_r;
929 int stable = map->chooseleaf_stable;
930
931 if ((__u32)ruleno >= map->max_rules) {
932 dprintk(" bad ruleno %d\n", ruleno);
933 return 0;
934 }
935
936 rule = map->rules[ruleno];
937 result_len = 0;
938
939 for (step = 0; step < rule->len; step++) {
940 int firstn = 0;
941 const struct crush_rule_step *curstep = &rule->steps[step];
942
943 switch (curstep->op) {
944 case CRUSH_RULE_TAKE:
945 if ((curstep->arg1 >= 0 &&
946 curstep->arg1 < map->max_devices) ||
947 (-1-curstep->arg1 >= 0 &&
948 -1-curstep->arg1 < map->max_buckets &&
949 map->buckets[-1-curstep->arg1])) {
950 w[0] = curstep->arg1;
951 wsize = 1;
952 } else {
953 dprintk(" bad take value %d\n", curstep->arg1);
954 }
955 break;
956
957 case CRUSH_RULE_SET_CHOOSE_TRIES:
958 if (curstep->arg1 > 0)
959 choose_tries = curstep->arg1;
960 break;
961
962 case CRUSH_RULE_SET_CHOOSELEAF_TRIES:
963 if (curstep->arg1 > 0)
964 choose_leaf_tries = curstep->arg1;
965 break;
966
967 case CRUSH_RULE_SET_CHOOSE_LOCAL_TRIES:
968 if (curstep->arg1 >= 0)
969 choose_local_retries = curstep->arg1;
970 break;
971
972 case CRUSH_RULE_SET_CHOOSE_LOCAL_FALLBACK_TRIES:
973 if (curstep->arg1 >= 0)
974 choose_local_fallback_retries = curstep->arg1;
975 break;
976
977 case CRUSH_RULE_SET_CHOOSELEAF_VARY_R:
978 if (curstep->arg1 >= 0)
979 vary_r = curstep->arg1;
980 break;
981
982 case CRUSH_RULE_SET_CHOOSELEAF_STABLE:
983 if (curstep->arg1 >= 0)
984 stable = curstep->arg1;
985 break;
986
987 case CRUSH_RULE_CHOOSELEAF_FIRSTN:
988 case CRUSH_RULE_CHOOSE_FIRSTN:
989 firstn = 1;
990 /* fall through */
991 case CRUSH_RULE_CHOOSELEAF_INDEP:
992 case CRUSH_RULE_CHOOSE_INDEP:
993 if (wsize == 0)
994 break;
995
996 recurse_to_leaf =
997 curstep->op ==
998 CRUSH_RULE_CHOOSELEAF_FIRSTN ||
999 curstep->op ==
1000 CRUSH_RULE_CHOOSELEAF_INDEP;
1001
1002 /* reset output */
1003 osize = 0;
1004
1005 for (i = 0; i < wsize; i++) {
1006 int bno;
1007 numrep = curstep->arg1;
1008 if (numrep <= 0) {
1009 numrep += result_max;
1010 if (numrep <= 0)
1011 continue;
1012 }
1013 j = 0;
1014 /* make sure bucket id is valid */
1015 bno = -1 - w[i];
1016 if (bno < 0 || bno >= map->max_buckets) {
1017 /* w[i] is probably CRUSH_ITEM_NONE */
1018 dprintk(" bad w[i] %d\n", w[i]);
1019 continue;
1020 }
1021 if (firstn) {
1022 int recurse_tries;
1023 if (choose_leaf_tries)
1024 recurse_tries =
1025 choose_leaf_tries;
1026 else if (map->chooseleaf_descend_once)
1027 recurse_tries = 1;
1028 else
1029 recurse_tries = choose_tries;
1030 osize += crush_choose_firstn(
1031 map,
1032 cw,
1033 map->buckets[bno],
1034 weight, weight_max,
1035 x, numrep,
1036 curstep->arg2,
1037 o+osize, j,
1038 result_max-osize,
1039 choose_tries,
1040 recurse_tries,
1041 choose_local_retries,
1042 choose_local_fallback_retries,
1043 recurse_to_leaf,
1044 vary_r,
1045 stable,
1046 c+osize,
1047 0,
1048 choose_args);
1049 } else {
1050 out_size = ((numrep < (result_max-osize)) ?
1051 numrep : (result_max-osize));
1052 crush_choose_indep(
1053 map,
1054 cw,
1055 map->buckets[bno],
1056 weight, weight_max,
1057 x, out_size, numrep,
1058 curstep->arg2,
1059 o+osize, j,
1060 choose_tries,
1061 choose_leaf_tries ?
1062 choose_leaf_tries : 1,
1063 recurse_to_leaf,
1064 c+osize,
1065 0,
1066 choose_args);
1067 osize += out_size;
1068 }
1069 }
1070
1071 if (recurse_to_leaf)
1072 /* copy final _leaf_ values to output set */
1073 memcpy(o, c, osize*sizeof(*o));
1074
1075 /* swap o and w arrays */
1076 tmp = o;
1077 o = w;
1078 w = tmp;
1079 wsize = osize;
1080 break;
1081
1082
1083 case CRUSH_RULE_EMIT:
1084 for (i = 0; i < wsize && result_len < result_max; i++) {
1085 result[result_len] = w[i];
1086 result_len++;
1087 }
1088 wsize = 0;
1089 break;
1090
1091 default:
1092 dprintk(" unknown op %d at step %d\n",
1093 curstep->op, step);
1094 break;
1095 }
1096 }
1097
1098 return result_len;
1099 }