]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blob - drivers/nvdimm/label.c
Merge branch 'for-linus' of git://git.armlinux.org.uk/~rmk/linux-arm
[mirror_ubuntu-bionic-kernel.git] / drivers / nvdimm / label.c
1 /*
2 * Copyright(c) 2013-2015 Intel Corporation. All rights reserved.
3 *
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of version 2 of the GNU General Public License as
6 * published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope that it will be useful, but
9 * WITHOUT ANY WARRANTY; without even the implied warranty of
10 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
11 * General Public License for more details.
12 */
13 #include <linux/device.h>
14 #include <linux/ndctl.h>
15 #include <linux/uuid.h>
16 #include <linux/slab.h>
17 #include <linux/io.h>
18 #include <linux/nd.h>
19 #include "nd-core.h"
20 #include "label.h"
21 #include "nd.h"
22
23 static guid_t nvdimm_btt_guid;
24 static guid_t nvdimm_btt2_guid;
25 static guid_t nvdimm_pfn_guid;
26 static guid_t nvdimm_dax_guid;
27
28 static u32 best_seq(u32 a, u32 b)
29 {
30 a &= NSINDEX_SEQ_MASK;
31 b &= NSINDEX_SEQ_MASK;
32
33 if (a == 0 || a == b)
34 return b;
35 else if (b == 0)
36 return a;
37 else if (nd_inc_seq(a) == b)
38 return b;
39 else
40 return a;
41 }
42
43 unsigned sizeof_namespace_label(struct nvdimm_drvdata *ndd)
44 {
45 return ndd->nslabel_size;
46 }
47
48 int nvdimm_num_label_slots(struct nvdimm_drvdata *ndd)
49 {
50 return ndd->nsarea.config_size / (sizeof_namespace_label(ndd) + 1);
51 }
52
53 size_t sizeof_namespace_index(struct nvdimm_drvdata *ndd)
54 {
55 u32 nslot, space, size;
56
57 /*
58 * The minimum index space is 512 bytes, with that amount of
59 * index we can describe ~1400 labels which is less than a byte
60 * of overhead per label. Round up to a byte of overhead per
61 * label and determine the size of the index region. Yes, this
62 * starts to waste space at larger config_sizes, but it's
63 * unlikely we'll ever see anything but 128K.
64 */
65 nslot = nvdimm_num_label_slots(ndd);
66 space = ndd->nsarea.config_size - nslot * sizeof_namespace_label(ndd);
67 size = ALIGN(sizeof(struct nd_namespace_index) + DIV_ROUND_UP(nslot, 8),
68 NSINDEX_ALIGN) * 2;
69 if (size <= space)
70 return size / 2;
71
72 dev_err(ndd->dev, "label area (%d) too small to host (%d byte) labels\n",
73 ndd->nsarea.config_size, sizeof_namespace_label(ndd));
74 return 0;
75 }
76
77 static int __nd_label_validate(struct nvdimm_drvdata *ndd)
78 {
79 /*
80 * On media label format consists of two index blocks followed
81 * by an array of labels. None of these structures are ever
82 * updated in place. A sequence number tracks the current
83 * active index and the next one to write, while labels are
84 * written to free slots.
85 *
86 * +------------+
87 * | |
88 * | nsindex0 |
89 * | |
90 * +------------+
91 * | |
92 * | nsindex1 |
93 * | |
94 * +------------+
95 * | label0 |
96 * +------------+
97 * | label1 |
98 * +------------+
99 * | |
100 * ....nslot...
101 * | |
102 * +------------+
103 * | labelN |
104 * +------------+
105 */
106 struct nd_namespace_index *nsindex[] = {
107 to_namespace_index(ndd, 0),
108 to_namespace_index(ndd, 1),
109 };
110 const int num_index = ARRAY_SIZE(nsindex);
111 struct device *dev = ndd->dev;
112 bool valid[2] = { 0 };
113 int i, num_valid = 0;
114 u32 seq;
115
116 for (i = 0; i < num_index; i++) {
117 u32 nslot;
118 u8 sig[NSINDEX_SIG_LEN];
119 u64 sum_save, sum, size;
120 unsigned int version, labelsize;
121
122 memcpy(sig, nsindex[i]->sig, NSINDEX_SIG_LEN);
123 if (memcmp(sig, NSINDEX_SIGNATURE, NSINDEX_SIG_LEN) != 0) {
124 dev_dbg(dev, "%s: nsindex%d signature invalid\n",
125 __func__, i);
126 continue;
127 }
128
129 /* label sizes larger than 128 arrived with v1.2 */
130 version = __le16_to_cpu(nsindex[i]->major) * 100
131 + __le16_to_cpu(nsindex[i]->minor);
132 if (version >= 102)
133 labelsize = 1 << (7 + nsindex[i]->labelsize);
134 else
135 labelsize = 128;
136
137 if (labelsize != sizeof_namespace_label(ndd)) {
138 dev_dbg(dev, "%s: nsindex%d labelsize %d invalid\n",
139 __func__, i, nsindex[i]->labelsize);
140 continue;
141 }
142
143 sum_save = __le64_to_cpu(nsindex[i]->checksum);
144 nsindex[i]->checksum = __cpu_to_le64(0);
145 sum = nd_fletcher64(nsindex[i], sizeof_namespace_index(ndd), 1);
146 nsindex[i]->checksum = __cpu_to_le64(sum_save);
147 if (sum != sum_save) {
148 dev_dbg(dev, "%s: nsindex%d checksum invalid\n",
149 __func__, i);
150 continue;
151 }
152
153 seq = __le32_to_cpu(nsindex[i]->seq);
154 if ((seq & NSINDEX_SEQ_MASK) == 0) {
155 dev_dbg(dev, "%s: nsindex%d sequence: %#x invalid\n",
156 __func__, i, seq);
157 continue;
158 }
159
160 /* sanity check the index against expected values */
161 if (__le64_to_cpu(nsindex[i]->myoff)
162 != i * sizeof_namespace_index(ndd)) {
163 dev_dbg(dev, "%s: nsindex%d myoff: %#llx invalid\n",
164 __func__, i, (unsigned long long)
165 __le64_to_cpu(nsindex[i]->myoff));
166 continue;
167 }
168 if (__le64_to_cpu(nsindex[i]->otheroff)
169 != (!i) * sizeof_namespace_index(ndd)) {
170 dev_dbg(dev, "%s: nsindex%d otheroff: %#llx invalid\n",
171 __func__, i, (unsigned long long)
172 __le64_to_cpu(nsindex[i]->otheroff));
173 continue;
174 }
175
176 size = __le64_to_cpu(nsindex[i]->mysize);
177 if (size > sizeof_namespace_index(ndd)
178 || size < sizeof(struct nd_namespace_index)) {
179 dev_dbg(dev, "%s: nsindex%d mysize: %#llx invalid\n",
180 __func__, i, size);
181 continue;
182 }
183
184 nslot = __le32_to_cpu(nsindex[i]->nslot);
185 if (nslot * sizeof_namespace_label(ndd)
186 + 2 * sizeof_namespace_index(ndd)
187 > ndd->nsarea.config_size) {
188 dev_dbg(dev, "%s: nsindex%d nslot: %u invalid, config_size: %#x\n",
189 __func__, i, nslot,
190 ndd->nsarea.config_size);
191 continue;
192 }
193 valid[i] = true;
194 num_valid++;
195 }
196
197 switch (num_valid) {
198 case 0:
199 break;
200 case 1:
201 for (i = 0; i < num_index; i++)
202 if (valid[i])
203 return i;
204 /* can't have num_valid > 0 but valid[] = { false, false } */
205 WARN_ON(1);
206 break;
207 default:
208 /* pick the best index... */
209 seq = best_seq(__le32_to_cpu(nsindex[0]->seq),
210 __le32_to_cpu(nsindex[1]->seq));
211 if (seq == (__le32_to_cpu(nsindex[1]->seq) & NSINDEX_SEQ_MASK))
212 return 1;
213 else
214 return 0;
215 break;
216 }
217
218 return -1;
219 }
220
221 int nd_label_validate(struct nvdimm_drvdata *ndd)
222 {
223 /*
224 * In order to probe for and validate namespace index blocks we
225 * need to know the size of the labels, and we can't trust the
226 * size of the labels until we validate the index blocks.
227 * Resolve this dependency loop by probing for known label
228 * sizes, but default to v1.2 256-byte namespace labels if
229 * discovery fails.
230 */
231 int label_size[] = { 128, 256 };
232 int i, rc;
233
234 for (i = 0; i < ARRAY_SIZE(label_size); i++) {
235 ndd->nslabel_size = label_size[i];
236 rc = __nd_label_validate(ndd);
237 if (rc >= 0)
238 return rc;
239 }
240
241 return -1;
242 }
243
244 void nd_label_copy(struct nvdimm_drvdata *ndd, struct nd_namespace_index *dst,
245 struct nd_namespace_index *src)
246 {
247 if (dst && src)
248 /* pass */;
249 else
250 return;
251
252 memcpy(dst, src, sizeof_namespace_index(ndd));
253 }
254
255 static struct nd_namespace_label *nd_label_base(struct nvdimm_drvdata *ndd)
256 {
257 void *base = to_namespace_index(ndd, 0);
258
259 return base + 2 * sizeof_namespace_index(ndd);
260 }
261
262 static int to_slot(struct nvdimm_drvdata *ndd,
263 struct nd_namespace_label *nd_label)
264 {
265 unsigned long label, base;
266
267 label = (unsigned long) nd_label;
268 base = (unsigned long) nd_label_base(ndd);
269
270 return (label - base) / sizeof_namespace_label(ndd);
271 }
272
273 static struct nd_namespace_label *to_label(struct nvdimm_drvdata *ndd, int slot)
274 {
275 unsigned long label, base;
276
277 base = (unsigned long) nd_label_base(ndd);
278 label = base + sizeof_namespace_label(ndd) * slot;
279
280 return (struct nd_namespace_label *) label;
281 }
282
283 #define for_each_clear_bit_le(bit, addr, size) \
284 for ((bit) = find_next_zero_bit_le((addr), (size), 0); \
285 (bit) < (size); \
286 (bit) = find_next_zero_bit_le((addr), (size), (bit) + 1))
287
288 /**
289 * preamble_index - common variable initialization for nd_label_* routines
290 * @ndd: dimm container for the relevant label set
291 * @idx: namespace_index index
292 * @nsindex_out: on return set to the currently active namespace index
293 * @free: on return set to the free label bitmap in the index
294 * @nslot: on return set to the number of slots in the label space
295 */
296 static bool preamble_index(struct nvdimm_drvdata *ndd, int idx,
297 struct nd_namespace_index **nsindex_out,
298 unsigned long **free, u32 *nslot)
299 {
300 struct nd_namespace_index *nsindex;
301
302 nsindex = to_namespace_index(ndd, idx);
303 if (nsindex == NULL)
304 return false;
305
306 *free = (unsigned long *) nsindex->free;
307 *nslot = __le32_to_cpu(nsindex->nslot);
308 *nsindex_out = nsindex;
309
310 return true;
311 }
312
313 char *nd_label_gen_id(struct nd_label_id *label_id, u8 *uuid, u32 flags)
314 {
315 if (!label_id || !uuid)
316 return NULL;
317 snprintf(label_id->id, ND_LABEL_ID_SIZE, "%s-%pUb",
318 flags & NSLABEL_FLAG_LOCAL ? "blk" : "pmem", uuid);
319 return label_id->id;
320 }
321
322 static bool preamble_current(struct nvdimm_drvdata *ndd,
323 struct nd_namespace_index **nsindex,
324 unsigned long **free, u32 *nslot)
325 {
326 return preamble_index(ndd, ndd->ns_current, nsindex,
327 free, nslot);
328 }
329
330 static bool preamble_next(struct nvdimm_drvdata *ndd,
331 struct nd_namespace_index **nsindex,
332 unsigned long **free, u32 *nslot)
333 {
334 return preamble_index(ndd, ndd->ns_next, nsindex,
335 free, nslot);
336 }
337
338 static bool slot_valid(struct nvdimm_drvdata *ndd,
339 struct nd_namespace_label *nd_label, u32 slot)
340 {
341 /* check that we are written where we expect to be written */
342 if (slot != __le32_to_cpu(nd_label->slot))
343 return false;
344
345 /* check that DPA allocations are page aligned */
346 if ((__le64_to_cpu(nd_label->dpa)
347 | __le64_to_cpu(nd_label->rawsize)) % SZ_4K)
348 return false;
349
350 /* check checksum */
351 if (namespace_label_has(ndd, checksum)) {
352 u64 sum, sum_save;
353
354 sum_save = __le64_to_cpu(nd_label->checksum);
355 nd_label->checksum = __cpu_to_le64(0);
356 sum = nd_fletcher64(nd_label, sizeof_namespace_label(ndd), 1);
357 nd_label->checksum = __cpu_to_le64(sum_save);
358 if (sum != sum_save) {
359 dev_dbg(ndd->dev, "%s fail checksum. slot: %d expect: %#llx\n",
360 __func__, slot, sum);
361 return false;
362 }
363 }
364
365 return true;
366 }
367
368 int nd_label_reserve_dpa(struct nvdimm_drvdata *ndd)
369 {
370 struct nd_namespace_index *nsindex;
371 unsigned long *free;
372 u32 nslot, slot;
373
374 if (!preamble_current(ndd, &nsindex, &free, &nslot))
375 return 0; /* no label, nothing to reserve */
376
377 for_each_clear_bit_le(slot, free, nslot) {
378 struct nd_namespace_label *nd_label;
379 struct nd_region *nd_region = NULL;
380 u8 label_uuid[NSLABEL_UUID_LEN];
381 struct nd_label_id label_id;
382 struct resource *res;
383 u32 flags;
384
385 nd_label = to_label(ndd, slot);
386
387 if (!slot_valid(ndd, nd_label, slot))
388 continue;
389
390 memcpy(label_uuid, nd_label->uuid, NSLABEL_UUID_LEN);
391 flags = __le32_to_cpu(nd_label->flags);
392 nd_label_gen_id(&label_id, label_uuid, flags);
393 res = nvdimm_allocate_dpa(ndd, &label_id,
394 __le64_to_cpu(nd_label->dpa),
395 __le64_to_cpu(nd_label->rawsize));
396 nd_dbg_dpa(nd_region, ndd, res, "reserve\n");
397 if (!res)
398 return -EBUSY;
399 }
400
401 return 0;
402 }
403
404 int nd_label_active_count(struct nvdimm_drvdata *ndd)
405 {
406 struct nd_namespace_index *nsindex;
407 unsigned long *free;
408 u32 nslot, slot;
409 int count = 0;
410
411 if (!preamble_current(ndd, &nsindex, &free, &nslot))
412 return 0;
413
414 for_each_clear_bit_le(slot, free, nslot) {
415 struct nd_namespace_label *nd_label;
416
417 nd_label = to_label(ndd, slot);
418
419 if (!slot_valid(ndd, nd_label, slot)) {
420 u32 label_slot = __le32_to_cpu(nd_label->slot);
421 u64 size = __le64_to_cpu(nd_label->rawsize);
422 u64 dpa = __le64_to_cpu(nd_label->dpa);
423
424 dev_dbg(ndd->dev,
425 "%s: slot%d invalid slot: %d dpa: %llx size: %llx\n",
426 __func__, slot, label_slot, dpa, size);
427 continue;
428 }
429 count++;
430 }
431 return count;
432 }
433
434 struct nd_namespace_label *nd_label_active(struct nvdimm_drvdata *ndd, int n)
435 {
436 struct nd_namespace_index *nsindex;
437 unsigned long *free;
438 u32 nslot, slot;
439
440 if (!preamble_current(ndd, &nsindex, &free, &nslot))
441 return NULL;
442
443 for_each_clear_bit_le(slot, free, nslot) {
444 struct nd_namespace_label *nd_label;
445
446 nd_label = to_label(ndd, slot);
447 if (!slot_valid(ndd, nd_label, slot))
448 continue;
449
450 if (n-- == 0)
451 return to_label(ndd, slot);
452 }
453
454 return NULL;
455 }
456
457 u32 nd_label_alloc_slot(struct nvdimm_drvdata *ndd)
458 {
459 struct nd_namespace_index *nsindex;
460 unsigned long *free;
461 u32 nslot, slot;
462
463 if (!preamble_next(ndd, &nsindex, &free, &nslot))
464 return UINT_MAX;
465
466 WARN_ON(!is_nvdimm_bus_locked(ndd->dev));
467
468 slot = find_next_bit_le(free, nslot, 0);
469 if (slot == nslot)
470 return UINT_MAX;
471
472 clear_bit_le(slot, free);
473
474 return slot;
475 }
476
477 bool nd_label_free_slot(struct nvdimm_drvdata *ndd, u32 slot)
478 {
479 struct nd_namespace_index *nsindex;
480 unsigned long *free;
481 u32 nslot;
482
483 if (!preamble_next(ndd, &nsindex, &free, &nslot))
484 return false;
485
486 WARN_ON(!is_nvdimm_bus_locked(ndd->dev));
487
488 if (slot < nslot)
489 return !test_and_set_bit_le(slot, free);
490 return false;
491 }
492
493 u32 nd_label_nfree(struct nvdimm_drvdata *ndd)
494 {
495 struct nd_namespace_index *nsindex;
496 unsigned long *free;
497 u32 nslot;
498
499 WARN_ON(!is_nvdimm_bus_locked(ndd->dev));
500
501 if (!preamble_next(ndd, &nsindex, &free, &nslot))
502 return nvdimm_num_label_slots(ndd);
503
504 return bitmap_weight(free, nslot);
505 }
506
507 static int nd_label_write_index(struct nvdimm_drvdata *ndd, int index, u32 seq,
508 unsigned long flags)
509 {
510 struct nd_namespace_index *nsindex;
511 unsigned long offset;
512 u64 checksum;
513 u32 nslot;
514 int rc;
515
516 nsindex = to_namespace_index(ndd, index);
517 if (flags & ND_NSINDEX_INIT)
518 nslot = nvdimm_num_label_slots(ndd);
519 else
520 nslot = __le32_to_cpu(nsindex->nslot);
521
522 memcpy(nsindex->sig, NSINDEX_SIGNATURE, NSINDEX_SIG_LEN);
523 memset(&nsindex->flags, 0, 3);
524 nsindex->labelsize = sizeof_namespace_label(ndd) >> 8;
525 nsindex->seq = __cpu_to_le32(seq);
526 offset = (unsigned long) nsindex
527 - (unsigned long) to_namespace_index(ndd, 0);
528 nsindex->myoff = __cpu_to_le64(offset);
529 nsindex->mysize = __cpu_to_le64(sizeof_namespace_index(ndd));
530 offset = (unsigned long) to_namespace_index(ndd,
531 nd_label_next_nsindex(index))
532 - (unsigned long) to_namespace_index(ndd, 0);
533 nsindex->otheroff = __cpu_to_le64(offset);
534 offset = (unsigned long) nd_label_base(ndd)
535 - (unsigned long) to_namespace_index(ndd, 0);
536 nsindex->labeloff = __cpu_to_le64(offset);
537 nsindex->nslot = __cpu_to_le32(nslot);
538 nsindex->major = __cpu_to_le16(1);
539 if (sizeof_namespace_label(ndd) < 256)
540 nsindex->minor = __cpu_to_le16(1);
541 else
542 nsindex->minor = __cpu_to_le16(2);
543 nsindex->checksum = __cpu_to_le64(0);
544 if (flags & ND_NSINDEX_INIT) {
545 unsigned long *free = (unsigned long *) nsindex->free;
546 u32 nfree = ALIGN(nslot, BITS_PER_LONG);
547 int last_bits, i;
548
549 memset(nsindex->free, 0xff, nfree / 8);
550 for (i = 0, last_bits = nfree - nslot; i < last_bits; i++)
551 clear_bit_le(nslot + i, free);
552 }
553 checksum = nd_fletcher64(nsindex, sizeof_namespace_index(ndd), 1);
554 nsindex->checksum = __cpu_to_le64(checksum);
555 rc = nvdimm_set_config_data(ndd, __le64_to_cpu(nsindex->myoff),
556 nsindex, sizeof_namespace_index(ndd));
557 if (rc < 0)
558 return rc;
559
560 if (flags & ND_NSINDEX_INIT)
561 return 0;
562
563 /* copy the index we just wrote to the new 'next' */
564 WARN_ON(index != ndd->ns_next);
565 nd_label_copy(ndd, to_current_namespace_index(ndd), nsindex);
566 ndd->ns_current = nd_label_next_nsindex(ndd->ns_current);
567 ndd->ns_next = nd_label_next_nsindex(ndd->ns_next);
568 WARN_ON(ndd->ns_current == ndd->ns_next);
569
570 return 0;
571 }
572
573 static unsigned long nd_label_offset(struct nvdimm_drvdata *ndd,
574 struct nd_namespace_label *nd_label)
575 {
576 return (unsigned long) nd_label
577 - (unsigned long) to_namespace_index(ndd, 0);
578 }
579
580 enum nvdimm_claim_class to_nvdimm_cclass(guid_t *guid)
581 {
582 if (guid_equal(guid, &nvdimm_btt_guid))
583 return NVDIMM_CCLASS_BTT;
584 else if (guid_equal(guid, &nvdimm_btt2_guid))
585 return NVDIMM_CCLASS_BTT2;
586 else if (guid_equal(guid, &nvdimm_pfn_guid))
587 return NVDIMM_CCLASS_PFN;
588 else if (guid_equal(guid, &nvdimm_dax_guid))
589 return NVDIMM_CCLASS_DAX;
590 else if (guid_equal(guid, &guid_null))
591 return NVDIMM_CCLASS_NONE;
592
593 return NVDIMM_CCLASS_UNKNOWN;
594 }
595
596 static const guid_t *to_abstraction_guid(enum nvdimm_claim_class claim_class,
597 guid_t *target)
598 {
599 if (claim_class == NVDIMM_CCLASS_BTT)
600 return &nvdimm_btt_guid;
601 else if (claim_class == NVDIMM_CCLASS_BTT2)
602 return &nvdimm_btt2_guid;
603 else if (claim_class == NVDIMM_CCLASS_PFN)
604 return &nvdimm_pfn_guid;
605 else if (claim_class == NVDIMM_CCLASS_DAX)
606 return &nvdimm_dax_guid;
607 else if (claim_class == NVDIMM_CCLASS_UNKNOWN) {
608 /*
609 * If we're modifying a namespace for which we don't
610 * know the claim_class, don't touch the existing guid.
611 */
612 return target;
613 } else
614 return &guid_null;
615 }
616
617 static int __pmem_label_update(struct nd_region *nd_region,
618 struct nd_mapping *nd_mapping, struct nd_namespace_pmem *nspm,
619 int pos)
620 {
621 struct nd_namespace_common *ndns = &nspm->nsio.common;
622 struct nd_interleave_set *nd_set = nd_region->nd_set;
623 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
624 struct nd_label_ent *label_ent, *victim = NULL;
625 struct nd_namespace_label *nd_label;
626 struct nd_namespace_index *nsindex;
627 struct nd_label_id label_id;
628 struct resource *res;
629 unsigned long *free;
630 u32 nslot, slot;
631 size_t offset;
632 u64 cookie;
633 int rc;
634
635 if (!preamble_next(ndd, &nsindex, &free, &nslot))
636 return -ENXIO;
637
638 cookie = nd_region_interleave_set_cookie(nd_region, nsindex);
639 nd_label_gen_id(&label_id, nspm->uuid, 0);
640 for_each_dpa_resource(ndd, res)
641 if (strcmp(res->name, label_id.id) == 0)
642 break;
643
644 if (!res) {
645 WARN_ON_ONCE(1);
646 return -ENXIO;
647 }
648
649 /* allocate and write the label to the staging (next) index */
650 slot = nd_label_alloc_slot(ndd);
651 if (slot == UINT_MAX)
652 return -ENXIO;
653 dev_dbg(ndd->dev, "%s: allocated: %d\n", __func__, slot);
654
655 nd_label = to_label(ndd, slot);
656 memset(nd_label, 0, sizeof_namespace_label(ndd));
657 memcpy(nd_label->uuid, nspm->uuid, NSLABEL_UUID_LEN);
658 if (nspm->alt_name)
659 memcpy(nd_label->name, nspm->alt_name, NSLABEL_NAME_LEN);
660 nd_label->flags = __cpu_to_le32(NSLABEL_FLAG_UPDATING);
661 nd_label->nlabel = __cpu_to_le16(nd_region->ndr_mappings);
662 nd_label->position = __cpu_to_le16(pos);
663 nd_label->isetcookie = __cpu_to_le64(cookie);
664 nd_label->rawsize = __cpu_to_le64(resource_size(res));
665 nd_label->lbasize = __cpu_to_le64(nspm->lbasize);
666 nd_label->dpa = __cpu_to_le64(res->start);
667 nd_label->slot = __cpu_to_le32(slot);
668 if (namespace_label_has(ndd, type_guid))
669 guid_copy(&nd_label->type_guid, &nd_set->type_guid);
670 if (namespace_label_has(ndd, abstraction_guid))
671 guid_copy(&nd_label->abstraction_guid,
672 to_abstraction_guid(ndns->claim_class,
673 &nd_label->abstraction_guid));
674 if (namespace_label_has(ndd, checksum)) {
675 u64 sum;
676
677 nd_label->checksum = __cpu_to_le64(0);
678 sum = nd_fletcher64(nd_label, sizeof_namespace_label(ndd), 1);
679 nd_label->checksum = __cpu_to_le64(sum);
680 }
681 nd_dbg_dpa(nd_region, ndd, res, "%s\n", __func__);
682
683 /* update label */
684 offset = nd_label_offset(ndd, nd_label);
685 rc = nvdimm_set_config_data(ndd, offset, nd_label,
686 sizeof_namespace_label(ndd));
687 if (rc < 0)
688 return rc;
689
690 /* Garbage collect the previous label */
691 mutex_lock(&nd_mapping->lock);
692 list_for_each_entry(label_ent, &nd_mapping->labels, list) {
693 if (!label_ent->label)
694 continue;
695 if (memcmp(nspm->uuid, label_ent->label->uuid,
696 NSLABEL_UUID_LEN) != 0)
697 continue;
698 victim = label_ent;
699 list_move_tail(&victim->list, &nd_mapping->labels);
700 break;
701 }
702 if (victim) {
703 dev_dbg(ndd->dev, "%s: free: %d\n", __func__, slot);
704 slot = to_slot(ndd, victim->label);
705 nd_label_free_slot(ndd, slot);
706 victim->label = NULL;
707 }
708
709 /* update index */
710 rc = nd_label_write_index(ndd, ndd->ns_next,
711 nd_inc_seq(__le32_to_cpu(nsindex->seq)), 0);
712 if (rc == 0) {
713 list_for_each_entry(label_ent, &nd_mapping->labels, list)
714 if (!label_ent->label) {
715 label_ent->label = nd_label;
716 nd_label = NULL;
717 break;
718 }
719 dev_WARN_ONCE(&nspm->nsio.common.dev, nd_label,
720 "failed to track label: %d\n",
721 to_slot(ndd, nd_label));
722 if (nd_label)
723 rc = -ENXIO;
724 }
725 mutex_unlock(&nd_mapping->lock);
726
727 return rc;
728 }
729
730 static bool is_old_resource(struct resource *res, struct resource **list, int n)
731 {
732 int i;
733
734 if (res->flags & DPA_RESOURCE_ADJUSTED)
735 return false;
736 for (i = 0; i < n; i++)
737 if (res == list[i])
738 return true;
739 return false;
740 }
741
742 static struct resource *to_resource(struct nvdimm_drvdata *ndd,
743 struct nd_namespace_label *nd_label)
744 {
745 struct resource *res;
746
747 for_each_dpa_resource(ndd, res) {
748 if (res->start != __le64_to_cpu(nd_label->dpa))
749 continue;
750 if (resource_size(res) != __le64_to_cpu(nd_label->rawsize))
751 continue;
752 return res;
753 }
754
755 return NULL;
756 }
757
758 /*
759 * 1/ Account all the labels that can be freed after this update
760 * 2/ Allocate and write the label to the staging (next) index
761 * 3/ Record the resources in the namespace device
762 */
763 static int __blk_label_update(struct nd_region *nd_region,
764 struct nd_mapping *nd_mapping, struct nd_namespace_blk *nsblk,
765 int num_labels)
766 {
767 int i, alloc, victims, nfree, old_num_resources, nlabel, rc = -ENXIO;
768 struct nd_interleave_set *nd_set = nd_region->nd_set;
769 struct nd_namespace_common *ndns = &nsblk->common;
770 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
771 struct nd_namespace_label *nd_label;
772 struct nd_label_ent *label_ent, *e;
773 struct nd_namespace_index *nsindex;
774 unsigned long *free, *victim_map = NULL;
775 struct resource *res, **old_res_list;
776 struct nd_label_id label_id;
777 u8 uuid[NSLABEL_UUID_LEN];
778 int min_dpa_idx = 0;
779 LIST_HEAD(list);
780 u32 nslot, slot;
781
782 if (!preamble_next(ndd, &nsindex, &free, &nslot))
783 return -ENXIO;
784
785 old_res_list = nsblk->res;
786 nfree = nd_label_nfree(ndd);
787 old_num_resources = nsblk->num_resources;
788 nd_label_gen_id(&label_id, nsblk->uuid, NSLABEL_FLAG_LOCAL);
789
790 /*
791 * We need to loop over the old resources a few times, which seems a
792 * bit inefficient, but we need to know that we have the label
793 * space before we start mutating the tracking structures.
794 * Otherwise the recovery method of last resort for userspace is
795 * disable and re-enable the parent region.
796 */
797 alloc = 0;
798 for_each_dpa_resource(ndd, res) {
799 if (strcmp(res->name, label_id.id) != 0)
800 continue;
801 if (!is_old_resource(res, old_res_list, old_num_resources))
802 alloc++;
803 }
804
805 victims = 0;
806 if (old_num_resources) {
807 /* convert old local-label-map to dimm-slot victim-map */
808 victim_map = kcalloc(BITS_TO_LONGS(nslot), sizeof(long),
809 GFP_KERNEL);
810 if (!victim_map)
811 return -ENOMEM;
812
813 /* mark unused labels for garbage collection */
814 for_each_clear_bit_le(slot, free, nslot) {
815 nd_label = to_label(ndd, slot);
816 memcpy(uuid, nd_label->uuid, NSLABEL_UUID_LEN);
817 if (memcmp(uuid, nsblk->uuid, NSLABEL_UUID_LEN) != 0)
818 continue;
819 res = to_resource(ndd, nd_label);
820 if (res && is_old_resource(res, old_res_list,
821 old_num_resources))
822 continue;
823 slot = to_slot(ndd, nd_label);
824 set_bit(slot, victim_map);
825 victims++;
826 }
827 }
828
829 /* don't allow updates that consume the last label */
830 if (nfree - alloc < 0 || nfree - alloc + victims < 1) {
831 dev_info(&nsblk->common.dev, "insufficient label space\n");
832 kfree(victim_map);
833 return -ENOSPC;
834 }
835 /* from here on we need to abort on error */
836
837
838 /* assign all resources to the namespace before writing the labels */
839 nsblk->res = NULL;
840 nsblk->num_resources = 0;
841 for_each_dpa_resource(ndd, res) {
842 if (strcmp(res->name, label_id.id) != 0)
843 continue;
844 if (!nsblk_add_resource(nd_region, ndd, nsblk, res->start)) {
845 rc = -ENOMEM;
846 goto abort;
847 }
848 }
849
850 /*
851 * Find the resource associated with the first label in the set
852 * per the v1.2 namespace specification.
853 */
854 for (i = 0; i < nsblk->num_resources; i++) {
855 struct resource *min = nsblk->res[min_dpa_idx];
856
857 res = nsblk->res[i];
858 if (res->start < min->start)
859 min_dpa_idx = i;
860 }
861
862 for (i = 0; i < nsblk->num_resources; i++) {
863 size_t offset;
864
865 res = nsblk->res[i];
866 if (is_old_resource(res, old_res_list, old_num_resources))
867 continue; /* carry-over */
868 slot = nd_label_alloc_slot(ndd);
869 if (slot == UINT_MAX)
870 goto abort;
871 dev_dbg(ndd->dev, "%s: allocated: %d\n", __func__, slot);
872
873 nd_label = to_label(ndd, slot);
874 memset(nd_label, 0, sizeof_namespace_label(ndd));
875 memcpy(nd_label->uuid, nsblk->uuid, NSLABEL_UUID_LEN);
876 if (nsblk->alt_name)
877 memcpy(nd_label->name, nsblk->alt_name,
878 NSLABEL_NAME_LEN);
879 nd_label->flags = __cpu_to_le32(NSLABEL_FLAG_LOCAL);
880
881 /*
882 * Use the presence of the type_guid as a flag to
883 * determine isetcookie usage and nlabel + position
884 * policy for blk-aperture namespaces.
885 */
886 if (namespace_label_has(ndd, type_guid)) {
887 if (i == min_dpa_idx) {
888 nd_label->nlabel = __cpu_to_le16(nsblk->num_resources);
889 nd_label->position = __cpu_to_le16(0);
890 } else {
891 nd_label->nlabel = __cpu_to_le16(0xffff);
892 nd_label->position = __cpu_to_le16(0xffff);
893 }
894 nd_label->isetcookie = __cpu_to_le64(nd_set->cookie2);
895 } else {
896 nd_label->nlabel = __cpu_to_le16(0); /* N/A */
897 nd_label->position = __cpu_to_le16(0); /* N/A */
898 nd_label->isetcookie = __cpu_to_le64(0); /* N/A */
899 }
900
901 nd_label->dpa = __cpu_to_le64(res->start);
902 nd_label->rawsize = __cpu_to_le64(resource_size(res));
903 nd_label->lbasize = __cpu_to_le64(nsblk->lbasize);
904 nd_label->slot = __cpu_to_le32(slot);
905 if (namespace_label_has(ndd, type_guid))
906 guid_copy(&nd_label->type_guid, &nd_set->type_guid);
907 if (namespace_label_has(ndd, abstraction_guid))
908 guid_copy(&nd_label->abstraction_guid,
909 to_abstraction_guid(ndns->claim_class,
910 &nd_label->abstraction_guid));
911
912 if (namespace_label_has(ndd, checksum)) {
913 u64 sum;
914
915 nd_label->checksum = __cpu_to_le64(0);
916 sum = nd_fletcher64(nd_label,
917 sizeof_namespace_label(ndd), 1);
918 nd_label->checksum = __cpu_to_le64(sum);
919 }
920
921 /* update label */
922 offset = nd_label_offset(ndd, nd_label);
923 rc = nvdimm_set_config_data(ndd, offset, nd_label,
924 sizeof_namespace_label(ndd));
925 if (rc < 0)
926 goto abort;
927 }
928
929 /* free up now unused slots in the new index */
930 for_each_set_bit(slot, victim_map, victim_map ? nslot : 0) {
931 dev_dbg(ndd->dev, "%s: free: %d\n", __func__, slot);
932 nd_label_free_slot(ndd, slot);
933 }
934
935 /* update index */
936 rc = nd_label_write_index(ndd, ndd->ns_next,
937 nd_inc_seq(__le32_to_cpu(nsindex->seq)), 0);
938 if (rc)
939 goto abort;
940
941 /*
942 * Now that the on-dimm labels are up to date, fix up the tracking
943 * entries in nd_mapping->labels
944 */
945 nlabel = 0;
946 mutex_lock(&nd_mapping->lock);
947 list_for_each_entry_safe(label_ent, e, &nd_mapping->labels, list) {
948 nd_label = label_ent->label;
949 if (!nd_label)
950 continue;
951 nlabel++;
952 memcpy(uuid, nd_label->uuid, NSLABEL_UUID_LEN);
953 if (memcmp(uuid, nsblk->uuid, NSLABEL_UUID_LEN) != 0)
954 continue;
955 nlabel--;
956 list_move(&label_ent->list, &list);
957 label_ent->label = NULL;
958 }
959 list_splice_tail_init(&list, &nd_mapping->labels);
960 mutex_unlock(&nd_mapping->lock);
961
962 if (nlabel + nsblk->num_resources > num_labels) {
963 /*
964 * Bug, we can't end up with more resources than
965 * available labels
966 */
967 WARN_ON_ONCE(1);
968 rc = -ENXIO;
969 goto out;
970 }
971
972 mutex_lock(&nd_mapping->lock);
973 label_ent = list_first_entry_or_null(&nd_mapping->labels,
974 typeof(*label_ent), list);
975 if (!label_ent) {
976 WARN_ON(1);
977 mutex_unlock(&nd_mapping->lock);
978 rc = -ENXIO;
979 goto out;
980 }
981 for_each_clear_bit_le(slot, free, nslot) {
982 nd_label = to_label(ndd, slot);
983 memcpy(uuid, nd_label->uuid, NSLABEL_UUID_LEN);
984 if (memcmp(uuid, nsblk->uuid, NSLABEL_UUID_LEN) != 0)
985 continue;
986 res = to_resource(ndd, nd_label);
987 res->flags &= ~DPA_RESOURCE_ADJUSTED;
988 dev_vdbg(&nsblk->common.dev, "assign label slot: %d\n", slot);
989 list_for_each_entry_from(label_ent, &nd_mapping->labels, list) {
990 if (label_ent->label)
991 continue;
992 label_ent->label = nd_label;
993 nd_label = NULL;
994 break;
995 }
996 if (nd_label)
997 dev_WARN(&nsblk->common.dev,
998 "failed to track label slot%d\n", slot);
999 }
1000 mutex_unlock(&nd_mapping->lock);
1001
1002 out:
1003 kfree(old_res_list);
1004 kfree(victim_map);
1005 return rc;
1006
1007 abort:
1008 /*
1009 * 1/ repair the allocated label bitmap in the index
1010 * 2/ restore the resource list
1011 */
1012 nd_label_copy(ndd, nsindex, to_current_namespace_index(ndd));
1013 kfree(nsblk->res);
1014 nsblk->res = old_res_list;
1015 nsblk->num_resources = old_num_resources;
1016 old_res_list = NULL;
1017 goto out;
1018 }
1019
1020 static int init_labels(struct nd_mapping *nd_mapping, int num_labels)
1021 {
1022 int i, old_num_labels = 0;
1023 struct nd_label_ent *label_ent;
1024 struct nd_namespace_index *nsindex;
1025 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1026
1027 mutex_lock(&nd_mapping->lock);
1028 list_for_each_entry(label_ent, &nd_mapping->labels, list)
1029 old_num_labels++;
1030 mutex_unlock(&nd_mapping->lock);
1031
1032 /*
1033 * We need to preserve all the old labels for the mapping so
1034 * they can be garbage collected after writing the new labels.
1035 */
1036 for (i = old_num_labels; i < num_labels; i++) {
1037 label_ent = kzalloc(sizeof(*label_ent), GFP_KERNEL);
1038 if (!label_ent)
1039 return -ENOMEM;
1040 mutex_lock(&nd_mapping->lock);
1041 list_add_tail(&label_ent->list, &nd_mapping->labels);
1042 mutex_unlock(&nd_mapping->lock);
1043 }
1044
1045 if (ndd->ns_current == -1 || ndd->ns_next == -1)
1046 /* pass */;
1047 else
1048 return max(num_labels, old_num_labels);
1049
1050 nsindex = to_namespace_index(ndd, 0);
1051 memset(nsindex, 0, ndd->nsarea.config_size);
1052 for (i = 0; i < 2; i++) {
1053 int rc = nd_label_write_index(ndd, i, i*2, ND_NSINDEX_INIT);
1054
1055 if (rc)
1056 return rc;
1057 }
1058 ndd->ns_next = 1;
1059 ndd->ns_current = 0;
1060
1061 return max(num_labels, old_num_labels);
1062 }
1063
1064 static int del_labels(struct nd_mapping *nd_mapping, u8 *uuid)
1065 {
1066 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1067 struct nd_label_ent *label_ent, *e;
1068 struct nd_namespace_index *nsindex;
1069 u8 label_uuid[NSLABEL_UUID_LEN];
1070 unsigned long *free;
1071 LIST_HEAD(list);
1072 u32 nslot, slot;
1073 int active = 0;
1074
1075 if (!uuid)
1076 return 0;
1077
1078 /* no index || no labels == nothing to delete */
1079 if (!preamble_next(ndd, &nsindex, &free, &nslot))
1080 return 0;
1081
1082 mutex_lock(&nd_mapping->lock);
1083 list_for_each_entry_safe(label_ent, e, &nd_mapping->labels, list) {
1084 struct nd_namespace_label *nd_label = label_ent->label;
1085
1086 if (!nd_label)
1087 continue;
1088 active++;
1089 memcpy(label_uuid, nd_label->uuid, NSLABEL_UUID_LEN);
1090 if (memcmp(label_uuid, uuid, NSLABEL_UUID_LEN) != 0)
1091 continue;
1092 active--;
1093 slot = to_slot(ndd, nd_label);
1094 nd_label_free_slot(ndd, slot);
1095 dev_dbg(ndd->dev, "%s: free: %d\n", __func__, slot);
1096 list_move_tail(&label_ent->list, &list);
1097 label_ent->label = NULL;
1098 }
1099 list_splice_tail_init(&list, &nd_mapping->labels);
1100
1101 if (active == 0) {
1102 nd_mapping_free_labels(nd_mapping);
1103 dev_dbg(ndd->dev, "%s: no more active labels\n", __func__);
1104 }
1105 mutex_unlock(&nd_mapping->lock);
1106
1107 return nd_label_write_index(ndd, ndd->ns_next,
1108 nd_inc_seq(__le32_to_cpu(nsindex->seq)), 0);
1109 }
1110
1111 int nd_pmem_namespace_label_update(struct nd_region *nd_region,
1112 struct nd_namespace_pmem *nspm, resource_size_t size)
1113 {
1114 int i;
1115
1116 for (i = 0; i < nd_region->ndr_mappings; i++) {
1117 struct nd_mapping *nd_mapping = &nd_region->mapping[i];
1118 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1119 struct resource *res;
1120 int rc, count = 0;
1121
1122 if (size == 0) {
1123 rc = del_labels(nd_mapping, nspm->uuid);
1124 if (rc)
1125 return rc;
1126 continue;
1127 }
1128
1129 for_each_dpa_resource(ndd, res)
1130 if (strncmp(res->name, "pmem", 4) == 0)
1131 count++;
1132 WARN_ON_ONCE(!count);
1133
1134 rc = init_labels(nd_mapping, count);
1135 if (rc < 0)
1136 return rc;
1137
1138 rc = __pmem_label_update(nd_region, nd_mapping, nspm, i);
1139 if (rc)
1140 return rc;
1141 }
1142
1143 return 0;
1144 }
1145
1146 int nd_blk_namespace_label_update(struct nd_region *nd_region,
1147 struct nd_namespace_blk *nsblk, resource_size_t size)
1148 {
1149 struct nd_mapping *nd_mapping = &nd_region->mapping[0];
1150 struct resource *res;
1151 int count = 0;
1152
1153 if (size == 0)
1154 return del_labels(nd_mapping, nsblk->uuid);
1155
1156 for_each_dpa_resource(to_ndd(nd_mapping), res)
1157 count++;
1158
1159 count = init_labels(nd_mapping, count);
1160 if (count < 0)
1161 return count;
1162
1163 return __blk_label_update(nd_region, nd_mapping, nsblk, count);
1164 }
1165
1166 int __init nd_label_init(void)
1167 {
1168 WARN_ON(guid_parse(NVDIMM_BTT_GUID, &nvdimm_btt_guid));
1169 WARN_ON(guid_parse(NVDIMM_BTT2_GUID, &nvdimm_btt2_guid));
1170 WARN_ON(guid_parse(NVDIMM_PFN_GUID, &nvdimm_pfn_guid));
1171 WARN_ON(guid_parse(NVDIMM_DAX_GUID, &nvdimm_dax_guid));
1172
1173 return 0;
1174 }