]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * Simple MTD partitioning layer | |
3 | * | |
4 | * (C) 2000 Nicolas Pitre <nico@cam.org> | |
5 | * | |
6 | * This code is GPL | |
7 | * | |
1da177e4 LT |
8 | * 02-21-2002 Thomas Gleixner <gleixner@autronix.de> |
9 | * added support for read_oob, write_oob | |
97894cda | 10 | */ |
1da177e4 LT |
11 | |
12 | #include <linux/module.h> | |
13 | #include <linux/types.h> | |
14 | #include <linux/kernel.h> | |
15 | #include <linux/slab.h> | |
16 | #include <linux/list.h> | |
1da177e4 LT |
17 | #include <linux/kmod.h> |
18 | #include <linux/mtd/mtd.h> | |
19 | #include <linux/mtd/partitions.h> | |
20 | #include <linux/mtd/compatmac.h> | |
21 | ||
22 | /* Our partition linked list */ | |
23 | static LIST_HEAD(mtd_partitions); | |
24 | ||
25 | /* Our partition node structure */ | |
26 | struct mtd_part { | |
27 | struct mtd_info mtd; | |
28 | struct mtd_info *master; | |
69423d99 | 29 | uint64_t offset; |
1da177e4 LT |
30 | int index; |
31 | struct list_head list; | |
32 | int registered; | |
33 | }; | |
34 | ||
35 | /* | |
36 | * Given a pointer to the MTD object in the mtd_part structure, we can retrieve | |
37 | * the pointer to that structure with this macro. | |
38 | */ | |
39 | #define PART(x) ((struct mtd_part *)(x)) | |
40 | ||
97894cda TG |
41 | |
42 | /* | |
1da177e4 LT |
43 | * MTD methods which simply translate the effective address and pass through |
44 | * to the _real_ device. | |
45 | */ | |
46 | ||
b33a2887 AN |
47 | static int part_read(struct mtd_info *mtd, loff_t from, size_t len, |
48 | size_t *retlen, u_char *buf) | |
1da177e4 LT |
49 | { |
50 | struct mtd_part *part = PART(mtd); | |
f1a28c02 TG |
51 | int res; |
52 | ||
1da177e4 LT |
53 | if (from >= mtd->size) |
54 | len = 0; | |
55 | else if (from + len > mtd->size) | |
56 | len = mtd->size - from; | |
b33a2887 | 57 | res = part->master->read(part->master, from + part->offset, |
9223a456 | 58 | len, retlen, buf); |
f1a28c02 TG |
59 | if (unlikely(res)) { |
60 | if (res == -EUCLEAN) | |
61 | mtd->ecc_stats.corrected++; | |
62 | if (res == -EBADMSG) | |
63 | mtd->ecc_stats.failed++; | |
64 | } | |
65 | return res; | |
1da177e4 LT |
66 | } |
67 | ||
b33a2887 AN |
68 | static int part_point(struct mtd_info *mtd, loff_t from, size_t len, |
69 | size_t *retlen, void **virt, resource_size_t *phys) | |
1da177e4 LT |
70 | { |
71 | struct mtd_part *part = PART(mtd); | |
72 | if (from >= mtd->size) | |
73 | len = 0; | |
74 | else if (from + len > mtd->size) | |
75 | len = mtd->size - from; | |
97894cda | 76 | return part->master->point (part->master, from + part->offset, |
a98889f3 | 77 | len, retlen, virt, phys); |
1da177e4 | 78 | } |
9223a456 | 79 | |
a98889f3 | 80 | static void part_unpoint(struct mtd_info *mtd, loff_t from, size_t len) |
1da177e4 LT |
81 | { |
82 | struct mtd_part *part = PART(mtd); | |
83 | ||
a98889f3 | 84 | part->master->unpoint(part->master, from + part->offset, len); |
1da177e4 LT |
85 | } |
86 | ||
402d3265 DH |
87 | static unsigned long part_get_unmapped_area(struct mtd_info *mtd, |
88 | unsigned long len, | |
89 | unsigned long offset, | |
90 | unsigned long flags) | |
91 | { | |
92 | struct mtd_part *part = PART(mtd); | |
93 | ||
94 | offset += part->offset; | |
95 | return part->master->get_unmapped_area(part->master, len, offset, | |
96 | flags); | |
97 | } | |
98 | ||
8593fbc6 | 99 | static int part_read_oob(struct mtd_info *mtd, loff_t from, |
b33a2887 | 100 | struct mtd_oob_ops *ops) |
1da177e4 LT |
101 | { |
102 | struct mtd_part *part = PART(mtd); | |
f1a28c02 | 103 | int res; |
8593fbc6 | 104 | |
1da177e4 | 105 | if (from >= mtd->size) |
8593fbc6 | 106 | return -EINVAL; |
7014568b | 107 | if (ops->datbuf && from + ops->len > mtd->size) |
8593fbc6 | 108 | return -EINVAL; |
f1a28c02 TG |
109 | res = part->master->read_oob(part->master, from + part->offset, ops); |
110 | ||
111 | if (unlikely(res)) { | |
112 | if (res == -EUCLEAN) | |
113 | mtd->ecc_stats.corrected++; | |
114 | if (res == -EBADMSG) | |
115 | mtd->ecc_stats.failed++; | |
116 | } | |
117 | return res; | |
1da177e4 LT |
118 | } |
119 | ||
b33a2887 AN |
120 | static int part_read_user_prot_reg(struct mtd_info *mtd, loff_t from, |
121 | size_t len, size_t *retlen, u_char *buf) | |
1da177e4 LT |
122 | { |
123 | struct mtd_part *part = PART(mtd); | |
b33a2887 | 124 | return part->master->read_user_prot_reg(part->master, from, |
1da177e4 LT |
125 | len, retlen, buf); |
126 | } | |
127 | ||
b33a2887 AN |
128 | static int part_get_user_prot_info(struct mtd_info *mtd, |
129 | struct otp_info *buf, size_t len) | |
f77814dd NP |
130 | { |
131 | struct mtd_part *part = PART(mtd); | |
b33a2887 | 132 | return part->master->get_user_prot_info(part->master, buf, len); |
f77814dd NP |
133 | } |
134 | ||
b33a2887 AN |
135 | static int part_read_fact_prot_reg(struct mtd_info *mtd, loff_t from, |
136 | size_t len, size_t *retlen, u_char *buf) | |
1da177e4 LT |
137 | { |
138 | struct mtd_part *part = PART(mtd); | |
b33a2887 | 139 | return part->master->read_fact_prot_reg(part->master, from, |
1da177e4 LT |
140 | len, retlen, buf); |
141 | } | |
142 | ||
b33a2887 AN |
143 | static int part_get_fact_prot_info(struct mtd_info *mtd, struct otp_info *buf, |
144 | size_t len) | |
f77814dd NP |
145 | { |
146 | struct mtd_part *part = PART(mtd); | |
b33a2887 | 147 | return part->master->get_fact_prot_info(part->master, buf, len); |
f77814dd NP |
148 | } |
149 | ||
b33a2887 AN |
150 | static int part_write(struct mtd_info *mtd, loff_t to, size_t len, |
151 | size_t *retlen, const u_char *buf) | |
1da177e4 LT |
152 | { |
153 | struct mtd_part *part = PART(mtd); | |
154 | if (!(mtd->flags & MTD_WRITEABLE)) | |
155 | return -EROFS; | |
156 | if (to >= mtd->size) | |
157 | len = 0; | |
158 | else if (to + len > mtd->size) | |
159 | len = mtd->size - to; | |
b33a2887 | 160 | return part->master->write(part->master, to + part->offset, |
9223a456 | 161 | len, retlen, buf); |
1da177e4 LT |
162 | } |
163 | ||
b33a2887 AN |
164 | static int part_panic_write(struct mtd_info *mtd, loff_t to, size_t len, |
165 | size_t *retlen, const u_char *buf) | |
388bbb09 RP |
166 | { |
167 | struct mtd_part *part = PART(mtd); | |
168 | if (!(mtd->flags & MTD_WRITEABLE)) | |
169 | return -EROFS; | |
170 | if (to >= mtd->size) | |
171 | len = 0; | |
172 | else if (to + len > mtd->size) | |
173 | len = mtd->size - to; | |
b33a2887 | 174 | return part->master->panic_write(part->master, to + part->offset, |
388bbb09 RP |
175 | len, retlen, buf); |
176 | } | |
177 | ||
8593fbc6 | 178 | static int part_write_oob(struct mtd_info *mtd, loff_t to, |
b33a2887 | 179 | struct mtd_oob_ops *ops) |
1da177e4 LT |
180 | { |
181 | struct mtd_part *part = PART(mtd); | |
8593fbc6 | 182 | |
1da177e4 LT |
183 | if (!(mtd->flags & MTD_WRITEABLE)) |
184 | return -EROFS; | |
8593fbc6 | 185 | |
1da177e4 | 186 | if (to >= mtd->size) |
8593fbc6 | 187 | return -EINVAL; |
7014568b | 188 | if (ops->datbuf && to + ops->len > mtd->size) |
8593fbc6 TG |
189 | return -EINVAL; |
190 | return part->master->write_oob(part->master, to + part->offset, ops); | |
1da177e4 LT |
191 | } |
192 | ||
b33a2887 AN |
193 | static int part_write_user_prot_reg(struct mtd_info *mtd, loff_t from, |
194 | size_t len, size_t *retlen, u_char *buf) | |
1da177e4 LT |
195 | { |
196 | struct mtd_part *part = PART(mtd); | |
b33a2887 | 197 | return part->master->write_user_prot_reg(part->master, from, |
1da177e4 LT |
198 | len, retlen, buf); |
199 | } | |
200 | ||
b33a2887 AN |
201 | static int part_lock_user_prot_reg(struct mtd_info *mtd, loff_t from, |
202 | size_t len) | |
f77814dd NP |
203 | { |
204 | struct mtd_part *part = PART(mtd); | |
b33a2887 | 205 | return part->master->lock_user_prot_reg(part->master, from, len); |
f77814dd NP |
206 | } |
207 | ||
b33a2887 AN |
208 | static int part_writev(struct mtd_info *mtd, const struct kvec *vecs, |
209 | unsigned long count, loff_t to, size_t *retlen) | |
1da177e4 LT |
210 | { |
211 | struct mtd_part *part = PART(mtd); | |
212 | if (!(mtd->flags & MTD_WRITEABLE)) | |
213 | return -EROFS; | |
b33a2887 | 214 | return part->master->writev(part->master, vecs, count, |
1da177e4 | 215 | to + part->offset, retlen); |
1da177e4 LT |
216 | } |
217 | ||
b33a2887 | 218 | static int part_erase(struct mtd_info *mtd, struct erase_info *instr) |
1da177e4 LT |
219 | { |
220 | struct mtd_part *part = PART(mtd); | |
221 | int ret; | |
222 | if (!(mtd->flags & MTD_WRITEABLE)) | |
223 | return -EROFS; | |
224 | if (instr->addr >= mtd->size) | |
225 | return -EINVAL; | |
226 | instr->addr += part->offset; | |
227 | ret = part->master->erase(part->master, instr); | |
74641d75 | 228 | if (ret) { |
bb0eb217 | 229 | if (instr->fail_addr != MTD_FAIL_ADDR_UNKNOWN) |
74641d75 AH |
230 | instr->fail_addr -= part->offset; |
231 | instr->addr -= part->offset; | |
232 | } | |
1da177e4 LT |
233 | return ret; |
234 | } | |
235 | ||
236 | void mtd_erase_callback(struct erase_info *instr) | |
237 | { | |
238 | if (instr->mtd->erase == part_erase) { | |
239 | struct mtd_part *part = PART(instr->mtd); | |
240 | ||
bb0eb217 | 241 | if (instr->fail_addr != MTD_FAIL_ADDR_UNKNOWN) |
1da177e4 LT |
242 | instr->fail_addr -= part->offset; |
243 | instr->addr -= part->offset; | |
244 | } | |
245 | if (instr->callback) | |
246 | instr->callback(instr); | |
247 | } | |
248 | EXPORT_SYMBOL_GPL(mtd_erase_callback); | |
249 | ||
69423d99 | 250 | static int part_lock(struct mtd_info *mtd, loff_t ofs, uint64_t len) |
1da177e4 LT |
251 | { |
252 | struct mtd_part *part = PART(mtd); | |
97894cda | 253 | if ((len + ofs) > mtd->size) |
1da177e4 LT |
254 | return -EINVAL; |
255 | return part->master->lock(part->master, ofs + part->offset, len); | |
256 | } | |
257 | ||
69423d99 | 258 | static int part_unlock(struct mtd_info *mtd, loff_t ofs, uint64_t len) |
1da177e4 LT |
259 | { |
260 | struct mtd_part *part = PART(mtd); | |
97894cda | 261 | if ((len + ofs) > mtd->size) |
1da177e4 LT |
262 | return -EINVAL; |
263 | return part->master->unlock(part->master, ofs + part->offset, len); | |
264 | } | |
265 | ||
266 | static void part_sync(struct mtd_info *mtd) | |
267 | { | |
268 | struct mtd_part *part = PART(mtd); | |
269 | part->master->sync(part->master); | |
270 | } | |
271 | ||
272 | static int part_suspend(struct mtd_info *mtd) | |
273 | { | |
274 | struct mtd_part *part = PART(mtd); | |
275 | return part->master->suspend(part->master); | |
276 | } | |
277 | ||
278 | static void part_resume(struct mtd_info *mtd) | |
279 | { | |
280 | struct mtd_part *part = PART(mtd); | |
281 | part->master->resume(part->master); | |
282 | } | |
283 | ||
b33a2887 | 284 | static int part_block_isbad(struct mtd_info *mtd, loff_t ofs) |
1da177e4 LT |
285 | { |
286 | struct mtd_part *part = PART(mtd); | |
287 | if (ofs >= mtd->size) | |
288 | return -EINVAL; | |
289 | ofs += part->offset; | |
290 | return part->master->block_isbad(part->master, ofs); | |
291 | } | |
292 | ||
b33a2887 | 293 | static int part_block_markbad(struct mtd_info *mtd, loff_t ofs) |
1da177e4 LT |
294 | { |
295 | struct mtd_part *part = PART(mtd); | |
f1a28c02 TG |
296 | int res; |
297 | ||
1da177e4 LT |
298 | if (!(mtd->flags & MTD_WRITEABLE)) |
299 | return -EROFS; | |
300 | if (ofs >= mtd->size) | |
301 | return -EINVAL; | |
302 | ofs += part->offset; | |
f1a28c02 TG |
303 | res = part->master->block_markbad(part->master, ofs); |
304 | if (!res) | |
305 | mtd->ecc_stats.badblocks++; | |
306 | return res; | |
1da177e4 LT |
307 | } |
308 | ||
97894cda TG |
309 | /* |
310 | * This function unregisters and destroy all slave MTD objects which are | |
1da177e4 LT |
311 | * attached to the given master MTD object. |
312 | */ | |
313 | ||
314 | int del_mtd_partitions(struct mtd_info *master) | |
315 | { | |
71a928c0 | 316 | struct mtd_part *slave, *next; |
1da177e4 | 317 | |
71a928c0 | 318 | list_for_each_entry_safe(slave, next, &mtd_partitions, list) |
1da177e4 | 319 | if (slave->master == master) { |
71a928c0 | 320 | list_del(&slave->list); |
b33a2887 | 321 | if (slave->registered) |
1da177e4 LT |
322 | del_mtd_device(&slave->mtd); |
323 | kfree(slave); | |
1da177e4 | 324 | } |
1da177e4 LT |
325 | |
326 | return 0; | |
327 | } | |
b33a2887 | 328 | EXPORT_SYMBOL(del_mtd_partitions); |
1da177e4 | 329 | |
7788ba71 AN |
330 | static struct mtd_part *add_one_partition(struct mtd_info *master, |
331 | const struct mtd_partition *part, int partno, | |
69423d99 | 332 | uint64_t cur_offset) |
7788ba71 AN |
333 | { |
334 | struct mtd_part *slave; | |
335 | ||
336 | /* allocate the partition structure */ | |
b33a2887 | 337 | slave = kzalloc(sizeof(*slave), GFP_KERNEL); |
7788ba71 | 338 | if (!slave) { |
b33a2887 | 339 | printk(KERN_ERR"memory allocation error while creating partitions for \"%s\"\n", |
7788ba71 AN |
340 | master->name); |
341 | del_mtd_partitions(master); | |
342 | return NULL; | |
343 | } | |
344 | list_add(&slave->list, &mtd_partitions); | |
345 | ||
346 | /* set up the MTD object for this partition */ | |
347 | slave->mtd.type = master->type; | |
348 | slave->mtd.flags = master->flags & ~part->mask_flags; | |
349 | slave->mtd.size = part->size; | |
350 | slave->mtd.writesize = master->writesize; | |
351 | slave->mtd.oobsize = master->oobsize; | |
352 | slave->mtd.oobavail = master->oobavail; | |
353 | slave->mtd.subpage_sft = master->subpage_sft; | |
354 | ||
355 | slave->mtd.name = part->name; | |
356 | slave->mtd.owner = master->owner; | |
402d3265 | 357 | slave->mtd.backing_dev_info = master->backing_dev_info; |
7788ba71 AN |
358 | |
359 | slave->mtd.read = part_read; | |
360 | slave->mtd.write = part_write; | |
361 | ||
362 | if (master->panic_write) | |
363 | slave->mtd.panic_write = part_panic_write; | |
364 | ||
b33a2887 | 365 | if (master->point && master->unpoint) { |
7788ba71 AN |
366 | slave->mtd.point = part_point; |
367 | slave->mtd.unpoint = part_unpoint; | |
368 | } | |
369 | ||
402d3265 DH |
370 | if (master->get_unmapped_area) |
371 | slave->mtd.get_unmapped_area = part_get_unmapped_area; | |
7788ba71 AN |
372 | if (master->read_oob) |
373 | slave->mtd.read_oob = part_read_oob; | |
374 | if (master->write_oob) | |
375 | slave->mtd.write_oob = part_write_oob; | |
b33a2887 | 376 | if (master->read_user_prot_reg) |
7788ba71 | 377 | slave->mtd.read_user_prot_reg = part_read_user_prot_reg; |
b33a2887 | 378 | if (master->read_fact_prot_reg) |
7788ba71 | 379 | slave->mtd.read_fact_prot_reg = part_read_fact_prot_reg; |
b33a2887 | 380 | if (master->write_user_prot_reg) |
7788ba71 | 381 | slave->mtd.write_user_prot_reg = part_write_user_prot_reg; |
b33a2887 | 382 | if (master->lock_user_prot_reg) |
7788ba71 | 383 | slave->mtd.lock_user_prot_reg = part_lock_user_prot_reg; |
b33a2887 | 384 | if (master->get_user_prot_info) |
7788ba71 | 385 | slave->mtd.get_user_prot_info = part_get_user_prot_info; |
b33a2887 | 386 | if (master->get_fact_prot_info) |
7788ba71 AN |
387 | slave->mtd.get_fact_prot_info = part_get_fact_prot_info; |
388 | if (master->sync) | |
389 | slave->mtd.sync = part_sync; | |
390 | if (!partno && master->suspend && master->resume) { | |
391 | slave->mtd.suspend = part_suspend; | |
392 | slave->mtd.resume = part_resume; | |
393 | } | |
394 | if (master->writev) | |
395 | slave->mtd.writev = part_writev; | |
396 | if (master->lock) | |
397 | slave->mtd.lock = part_lock; | |
398 | if (master->unlock) | |
399 | slave->mtd.unlock = part_unlock; | |
400 | if (master->block_isbad) | |
401 | slave->mtd.block_isbad = part_block_isbad; | |
402 | if (master->block_markbad) | |
403 | slave->mtd.block_markbad = part_block_markbad; | |
404 | slave->mtd.erase = part_erase; | |
405 | slave->master = master; | |
406 | slave->offset = part->offset; | |
407 | slave->index = partno; | |
408 | ||
409 | if (slave->offset == MTDPART_OFS_APPEND) | |
410 | slave->offset = cur_offset; | |
411 | if (slave->offset == MTDPART_OFS_NXTBLK) { | |
412 | slave->offset = cur_offset; | |
69423d99 | 413 | if (mtd_mod_by_eb(cur_offset, master) != 0) { |
7788ba71 | 414 | /* Round up to next erasesize */ |
69423d99 | 415 | slave->offset = (mtd_div_by_eb(cur_offset, master) + 1) * master->erasesize; |
7788ba71 | 416 | printk(KERN_NOTICE "Moving partition %d: " |
69423d99 AH |
417 | "0x%012llx -> 0x%012llx\n", partno, |
418 | (unsigned long long)cur_offset, (unsigned long long)slave->offset); | |
7788ba71 AN |
419 | } |
420 | } | |
421 | if (slave->mtd.size == MTDPART_SIZ_FULL) | |
422 | slave->mtd.size = master->size - slave->offset; | |
423 | ||
69423d99 AH |
424 | printk(KERN_NOTICE "0x%012llx-0x%012llx : \"%s\"\n", (unsigned long long)slave->offset, |
425 | (unsigned long long)(slave->offset + slave->mtd.size), slave->mtd.name); | |
7788ba71 AN |
426 | |
427 | /* let's do some sanity checks */ | |
428 | if (slave->offset >= master->size) { | |
f636ffb4 | 429 | /* let's register it anyway to preserve ordering */ |
7788ba71 AN |
430 | slave->offset = 0; |
431 | slave->mtd.size = 0; | |
b33a2887 | 432 | printk(KERN_ERR"mtd: partition \"%s\" is out of reach -- disabled\n", |
7788ba71 | 433 | part->name); |
f636ffb4 | 434 | goto out_register; |
7788ba71 AN |
435 | } |
436 | if (slave->offset + slave->mtd.size > master->size) { | |
437 | slave->mtd.size = master->size - slave->offset; | |
69423d99 AH |
438 | printk(KERN_WARNING"mtd: partition \"%s\" extends beyond the end of device \"%s\" -- size truncated to %#llx\n", |
439 | part->name, master->name, (unsigned long long)slave->mtd.size); | |
7788ba71 | 440 | } |
b33a2887 | 441 | if (master->numeraseregions > 1) { |
7788ba71 | 442 | /* Deal with variable erase size stuff */ |
6910c136 | 443 | int i, max = master->numeraseregions; |
69423d99 | 444 | u64 end = slave->offset + slave->mtd.size; |
7788ba71 AN |
445 | struct mtd_erase_region_info *regions = master->eraseregions; |
446 | ||
6910c136 AN |
447 | /* Find the first erase regions which is part of this |
448 | * partition. */ | |
449 | for (i = 0; i < max && regions[i].offset <= slave->offset; i++) | |
7788ba71 | 450 | ; |
6910c136 AN |
451 | /* The loop searched for the region _behind_ the first one */ |
452 | i--; | |
7788ba71 | 453 | |
6910c136 AN |
454 | /* Pick biggest erasesize */ |
455 | for (; i < max && regions[i].offset < end; i++) { | |
7788ba71 AN |
456 | if (slave->mtd.erasesize < regions[i].erasesize) { |
457 | slave->mtd.erasesize = regions[i].erasesize; | |
458 | } | |
459 | } | |
6910c136 | 460 | BUG_ON(slave->mtd.erasesize == 0); |
7788ba71 AN |
461 | } else { |
462 | /* Single erase size */ | |
463 | slave->mtd.erasesize = master->erasesize; | |
464 | } | |
465 | ||
466 | if ((slave->mtd.flags & MTD_WRITEABLE) && | |
69423d99 | 467 | mtd_mod_by_eb(slave->offset, &slave->mtd)) { |
7788ba71 | 468 | /* Doesn't start on a boundary of major erase size */ |
b33a2887 AN |
469 | /* FIXME: Let it be writable if it is on a boundary of |
470 | * _minor_ erase size though */ | |
7788ba71 | 471 | slave->mtd.flags &= ~MTD_WRITEABLE; |
b33a2887 | 472 | printk(KERN_WARNING"mtd: partition \"%s\" doesn't start on an erase block boundary -- force read-only\n", |
7788ba71 AN |
473 | part->name); |
474 | } | |
475 | if ((slave->mtd.flags & MTD_WRITEABLE) && | |
69423d99 | 476 | mtd_mod_by_eb(slave->mtd.size, &slave->mtd)) { |
7788ba71 | 477 | slave->mtd.flags &= ~MTD_WRITEABLE; |
b33a2887 | 478 | printk(KERN_WARNING"mtd: partition \"%s\" doesn't end on an erase block -- force read-only\n", |
7788ba71 AN |
479 | part->name); |
480 | } | |
481 | ||
482 | slave->mtd.ecclayout = master->ecclayout; | |
483 | if (master->block_isbad) { | |
69423d99 | 484 | uint64_t offs = 0; |
7788ba71 | 485 | |
b33a2887 | 486 | while (offs < slave->mtd.size) { |
7788ba71 AN |
487 | if (master->block_isbad(master, |
488 | offs + slave->offset)) | |
489 | slave->mtd.ecc_stats.badblocks++; | |
490 | offs += slave->mtd.erasesize; | |
491 | } | |
492 | } | |
493 | ||
f636ffb4 | 494 | out_register: |
b33a2887 AN |
495 | if (part->mtdp) { |
496 | /* store the object pointer (caller may or may not register it*/ | |
7788ba71 AN |
497 | *part->mtdp = &slave->mtd; |
498 | slave->registered = 0; | |
499 | } else { | |
500 | /* register our partition */ | |
501 | add_mtd_device(&slave->mtd); | |
502 | slave->registered = 1; | |
503 | } | |
504 | return slave; | |
505 | } | |
506 | ||
1da177e4 LT |
507 | /* |
508 | * This function, given a master MTD object and a partition table, creates | |
509 | * and registers slave MTD objects which are bound to the master according to | |
510 | * the partition definitions. | |
511 | * (Q: should we register the master MTD object as well?) | |
512 | */ | |
513 | ||
97894cda | 514 | int add_mtd_partitions(struct mtd_info *master, |
1da177e4 LT |
515 | const struct mtd_partition *parts, |
516 | int nbparts) | |
517 | { | |
518 | struct mtd_part *slave; | |
69423d99 | 519 | uint64_t cur_offset = 0; |
1da177e4 LT |
520 | int i; |
521 | ||
b33a2887 | 522 | printk(KERN_NOTICE "Creating %d MTD partitions on \"%s\":\n", nbparts, master->name); |
1da177e4 LT |
523 | |
524 | for (i = 0; i < nbparts; i++) { | |
7788ba71 AN |
525 | slave = add_one_partition(master, parts + i, i, cur_offset); |
526 | if (!slave) | |
1da177e4 | 527 | return -ENOMEM; |
1da177e4 | 528 | cur_offset = slave->offset + slave->mtd.size; |
1da177e4 LT |
529 | } |
530 | ||
531 | return 0; | |
532 | } | |
1da177e4 | 533 | EXPORT_SYMBOL(add_mtd_partitions); |
1da177e4 LT |
534 | |
535 | static DEFINE_SPINLOCK(part_parser_lock); | |
536 | static LIST_HEAD(part_parsers); | |
537 | ||
538 | static struct mtd_part_parser *get_partition_parser(const char *name) | |
539 | { | |
71a928c0 | 540 | struct mtd_part_parser *p, *ret = NULL; |
1da177e4 | 541 | |
71a928c0 | 542 | spin_lock(&part_parser_lock); |
1da177e4 | 543 | |
71a928c0 | 544 | list_for_each_entry(p, &part_parsers, list) |
1da177e4 LT |
545 | if (!strcmp(p->name, name) && try_module_get(p->owner)) { |
546 | ret = p; | |
547 | break; | |
548 | } | |
71a928c0 | 549 | |
1da177e4 LT |
550 | spin_unlock(&part_parser_lock); |
551 | ||
552 | return ret; | |
553 | } | |
554 | ||
555 | int register_mtd_parser(struct mtd_part_parser *p) | |
556 | { | |
557 | spin_lock(&part_parser_lock); | |
558 | list_add(&p->list, &part_parsers); | |
559 | spin_unlock(&part_parser_lock); | |
560 | ||
561 | return 0; | |
562 | } | |
b33a2887 | 563 | EXPORT_SYMBOL_GPL(register_mtd_parser); |
1da177e4 LT |
564 | |
565 | int deregister_mtd_parser(struct mtd_part_parser *p) | |
566 | { | |
567 | spin_lock(&part_parser_lock); | |
568 | list_del(&p->list); | |
569 | spin_unlock(&part_parser_lock); | |
570 | return 0; | |
571 | } | |
b33a2887 | 572 | EXPORT_SYMBOL_GPL(deregister_mtd_parser); |
1da177e4 | 573 | |
97894cda | 574 | int parse_mtd_partitions(struct mtd_info *master, const char **types, |
1da177e4 LT |
575 | struct mtd_partition **pparts, unsigned long origin) |
576 | { | |
577 | struct mtd_part_parser *parser; | |
578 | int ret = 0; | |
97894cda | 579 | |
1da177e4 LT |
580 | for ( ; ret <= 0 && *types; types++) { |
581 | parser = get_partition_parser(*types); | |
1da177e4 LT |
582 | if (!parser && !request_module("%s", *types)) |
583 | parser = get_partition_parser(*types); | |
1da177e4 LT |
584 | if (!parser) { |
585 | printk(KERN_NOTICE "%s partition parsing not available\n", | |
586 | *types); | |
587 | continue; | |
588 | } | |
589 | ret = (*parser->parse_fn)(master, pparts, origin); | |
590 | if (ret > 0) { | |
97894cda | 591 | printk(KERN_NOTICE "%d %s partitions found on MTD device %s\n", |
1da177e4 LT |
592 | ret, parser->name, master->name); |
593 | } | |
594 | put_partition_parser(parser); | |
595 | } | |
596 | return ret; | |
597 | } | |
1da177e4 | 598 | EXPORT_SYMBOL_GPL(parse_mtd_partitions); |