]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blame - fs/jffs2/wbuf.c
[MTD] NAND modularize write function
[mirror_ubuntu-artful-kernel.git] / fs / jffs2 / wbuf.c
CommitLineData
1da177e4
LT
1/*
2 * JFFS2 -- Journalling Flash File System, Version 2.
3 *
4 * Copyright (C) 2001-2003 Red Hat, Inc.
5 * Copyright (C) 2004 Thomas Gleixner <tglx@linutronix.de>
6 *
7 * Created by David Woodhouse <dwmw2@infradead.org>
8 * Modified debugged and enhanced by Thomas Gleixner <tglx@linutronix.de>
9 *
10 * For licensing information, see the file 'LICENCE' in this directory.
11 *
daba5cc4 12 * $Id: wbuf.c,v 1.100 2005/09/30 13:59:13 dedekind Exp $
1da177e4
LT
13 *
14 */
15
16#include <linux/kernel.h>
17#include <linux/slab.h>
18#include <linux/mtd/mtd.h>
19#include <linux/crc32.h>
20#include <linux/mtd/nand.h>
4e57b681
TS
21#include <linux/jiffies.h>
22
1da177e4
LT
23#include "nodelist.h"
24
25/* For testing write failures */
26#undef BREAKME
27#undef BREAKMEHEADER
28
29#ifdef BREAKME
30static unsigned char *brokenbuf;
31#endif
32
daba5cc4
AB
33#define PAGE_DIV(x) ( ((unsigned long)(x) / (unsigned long)(c->wbuf_pagesize)) * (unsigned long)(c->wbuf_pagesize) )
34#define PAGE_MOD(x) ( (unsigned long)(x) % (unsigned long)(c->wbuf_pagesize) )
35
1da177e4
LT
36/* max. erase failures before we mark a block bad */
37#define MAX_ERASE_FAILURES 2
38
1da177e4
LT
39struct jffs2_inodirty {
40 uint32_t ino;
41 struct jffs2_inodirty *next;
42};
43
44static struct jffs2_inodirty inodirty_nomem;
45
46static int jffs2_wbuf_pending_for_ino(struct jffs2_sb_info *c, uint32_t ino)
47{
48 struct jffs2_inodirty *this = c->wbuf_inodes;
49
50 /* If a malloc failed, consider _everything_ dirty */
51 if (this == &inodirty_nomem)
52 return 1;
53
54 /* If ino == 0, _any_ non-GC writes mean 'yes' */
55 if (this && !ino)
56 return 1;
57
58 /* Look to see if the inode in question is pending in the wbuf */
59 while (this) {
60 if (this->ino == ino)
61 return 1;
62 this = this->next;
63 }
64 return 0;
65}
66
67static void jffs2_clear_wbuf_ino_list(struct jffs2_sb_info *c)
68{
69 struct jffs2_inodirty *this;
70
71 this = c->wbuf_inodes;
72
73 if (this != &inodirty_nomem) {
74 while (this) {
75 struct jffs2_inodirty *next = this->next;
76 kfree(this);
77 this = next;
78 }
79 }
80 c->wbuf_inodes = NULL;
81}
82
83static void jffs2_wbuf_dirties_inode(struct jffs2_sb_info *c, uint32_t ino)
84{
85 struct jffs2_inodirty *new;
86
87 /* Mark the superblock dirty so that kupdated will flush... */
4d952709 88 jffs2_erase_pending_trigger(c);
1da177e4
LT
89
90 if (jffs2_wbuf_pending_for_ino(c, ino))
91 return;
92
93 new = kmalloc(sizeof(*new), GFP_KERNEL);
94 if (!new) {
95 D1(printk(KERN_DEBUG "No memory to allocate inodirty. Fallback to all considered dirty\n"));
96 jffs2_clear_wbuf_ino_list(c);
97 c->wbuf_inodes = &inodirty_nomem;
98 return;
99 }
100 new->ino = ino;
101 new->next = c->wbuf_inodes;
102 c->wbuf_inodes = new;
103 return;
104}
105
106static inline void jffs2_refile_wbuf_blocks(struct jffs2_sb_info *c)
107{
108 struct list_head *this, *next;
109 static int n;
110
111 if (list_empty(&c->erasable_pending_wbuf_list))
112 return;
113
114 list_for_each_safe(this, next, &c->erasable_pending_wbuf_list) {
115 struct jffs2_eraseblock *jeb = list_entry(this, struct jffs2_eraseblock, list);
116
117 D1(printk(KERN_DEBUG "Removing eraseblock at 0x%08x from erasable_pending_wbuf_list...\n", jeb->offset));
118 list_del(this);
119 if ((jiffies + (n++)) & 127) {
120 /* Most of the time, we just erase it immediately. Otherwise we
121 spend ages scanning it on mount, etc. */
122 D1(printk(KERN_DEBUG "...and adding to erase_pending_list\n"));
123 list_add_tail(&jeb->list, &c->erase_pending_list);
124 c->nr_erasing_blocks++;
125 jffs2_erase_pending_trigger(c);
126 } else {
127 /* Sometimes, however, we leave it elsewhere so it doesn't get
128 immediately reused, and we spread the load a bit. */
129 D1(printk(KERN_DEBUG "...and adding to erasable_list\n"));
130 list_add_tail(&jeb->list, &c->erasable_list);
131 }
132 }
133}
134
7f716cf3
EH
135#define REFILE_NOTEMPTY 0
136#define REFILE_ANYWAY 1
137
138static void jffs2_block_refile(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, int allow_empty)
1da177e4
LT
139{
140 D1(printk("About to refile bad block at %08x\n", jeb->offset));
141
1da177e4
LT
142 /* File the existing block on the bad_used_list.... */
143 if (c->nextblock == jeb)
144 c->nextblock = NULL;
145 else /* Not sure this should ever happen... need more coffee */
146 list_del(&jeb->list);
147 if (jeb->first_node) {
148 D1(printk("Refiling block at %08x to bad_used_list\n", jeb->offset));
149 list_add(&jeb->list, &c->bad_used_list);
150 } else {
9b88f473 151 BUG_ON(allow_empty == REFILE_NOTEMPTY);
1da177e4
LT
152 /* It has to have had some nodes or we couldn't be here */
153 D1(printk("Refiling block at %08x to erase_pending_list\n", jeb->offset));
154 list_add(&jeb->list, &c->erase_pending_list);
155 c->nr_erasing_blocks++;
156 jffs2_erase_pending_trigger(c);
157 }
1da177e4
LT
158
159 /* Adjust its size counts accordingly */
160 c->wasted_size += jeb->free_size;
161 c->free_size -= jeb->free_size;
162 jeb->wasted_size += jeb->free_size;
163 jeb->free_size = 0;
164
e0c8e42f
AB
165 jffs2_dbg_dump_block_lists_nolock(c);
166 jffs2_dbg_acct_sanity_check_nolock(c,jeb);
167 jffs2_dbg_acct_paranoia_check_nolock(c, jeb);
1da177e4
LT
168}
169
170/* Recover from failure to write wbuf. Recover the nodes up to the
171 * wbuf, not the one which we were starting to try to write. */
172
173static void jffs2_wbuf_recover(struct jffs2_sb_info *c)
174{
175 struct jffs2_eraseblock *jeb, *new_jeb;
176 struct jffs2_raw_node_ref **first_raw, **raw;
177 size_t retlen;
178 int ret;
179 unsigned char *buf;
180 uint32_t start, end, ofs, len;
181
046b8b98
DW
182 jeb = &c->blocks[c->wbuf_ofs / c->sector_size];
183
184 if (jffs2_prealloc_raw_node_refs(c, jeb, c->reserved_refs + 1))
2f785402
DW
185 return;
186
1da177e4
LT
187 spin_lock(&c->erase_completion_lock);
188
7f716cf3 189 jffs2_block_refile(c, jeb, REFILE_NOTEMPTY);
1da177e4
LT
190
191 /* Find the first node to be recovered, by skipping over every
192 node which ends before the wbuf starts, or which is obsolete. */
193 first_raw = &jeb->first_node;
182ec4ee 194 while (*first_raw &&
1da177e4
LT
195 (ref_obsolete(*first_raw) ||
196 (ref_offset(*first_raw)+ref_totlen(c, jeb, *first_raw)) < c->wbuf_ofs)) {
197 D1(printk(KERN_DEBUG "Skipping node at 0x%08x(%d)-0x%08x which is either before 0x%08x or obsolete\n",
198 ref_offset(*first_raw), ref_flags(*first_raw),
199 (ref_offset(*first_raw) + ref_totlen(c, jeb, *first_raw)),
200 c->wbuf_ofs));
201 first_raw = &(*first_raw)->next_phys;
202 }
203
204 if (!*first_raw) {
205 /* All nodes were obsolete. Nothing to recover. */
206 D1(printk(KERN_DEBUG "No non-obsolete nodes to be recovered. Just filing block bad\n"));
207 spin_unlock(&c->erase_completion_lock);
208 return;
209 }
210
211 start = ref_offset(*first_raw);
212 end = ref_offset(*first_raw) + ref_totlen(c, jeb, *first_raw);
213
214 /* Find the last node to be recovered */
215 raw = first_raw;
216 while ((*raw)) {
217 if (!ref_obsolete(*raw))
218 end = ref_offset(*raw) + ref_totlen(c, jeb, *raw);
219
220 raw = &(*raw)->next_phys;
221 }
222 spin_unlock(&c->erase_completion_lock);
223
224 D1(printk(KERN_DEBUG "wbuf recover %08x-%08x\n", start, end));
225
226 buf = NULL;
227 if (start < c->wbuf_ofs) {
228 /* First affected node was already partially written.
229 * Attempt to reread the old data into our buffer. */
230
231 buf = kmalloc(end - start, GFP_KERNEL);
232 if (!buf) {
233 printk(KERN_CRIT "Malloc failure in wbuf recovery. Data loss ensues.\n");
234
235 goto read_failed;
236 }
237
238 /* Do the read... */
9223a456 239 ret = c->mtd->read(c->mtd, start, c->wbuf_ofs - start, &retlen, buf);
182ec4ee 240
1da177e4
LT
241 if (ret == -EBADMSG && retlen == c->wbuf_ofs - start) {
242 /* ECC recovered */
243 ret = 0;
244 }
245 if (ret || retlen != c->wbuf_ofs - start) {
246 printk(KERN_CRIT "Old data are already lost in wbuf recovery. Data loss ensues.\n");
247
248 kfree(buf);
249 buf = NULL;
250 read_failed:
251 first_raw = &(*first_raw)->next_phys;
252 /* If this was the only node to be recovered, give up */
253 if (!(*first_raw))
254 return;
255
256 /* It wasn't. Go on and try to recover nodes complete in the wbuf */
257 start = ref_offset(*first_raw);
258 } else {
259 /* Read succeeded. Copy the remaining data from the wbuf */
260 memcpy(buf + (c->wbuf_ofs - start), c->wbuf, end - c->wbuf_ofs);
261 }
262 }
263 /* OK... we're to rewrite (end-start) bytes of data from first_raw onwards.
264 Either 'buf' contains the data, or we find it in the wbuf */
265
266
267 /* ... and get an allocation of space from a shiny new block instead */
9fe4854c 268 ret = jffs2_reserve_space_gc(c, end-start, &len, JFFS2_SUMMARY_NOSUM_SIZE);
1da177e4
LT
269 if (ret) {
270 printk(KERN_WARNING "Failed to allocate space for wbuf recovery. Data loss ensues.\n");
9b88f473 271 kfree(buf);
1da177e4
LT
272 return;
273 }
9fe4854c
DW
274 ofs = write_ofs(c);
275
1da177e4 276 if (end-start >= c->wbuf_pagesize) {
7f716cf3 277 /* Need to do another write immediately, but it's possible
9b88f473 278 that this is just because the wbuf itself is completely
182ec4ee
TG
279 full, and there's nothing earlier read back from the
280 flash. Hence 'buf' isn't necessarily what we're writing
9b88f473 281 from. */
7f716cf3 282 unsigned char *rewrite_buf = buf?:c->wbuf;
1da177e4
LT
283 uint32_t towrite = (end-start) - ((end-start)%c->wbuf_pagesize);
284
285 D1(printk(KERN_DEBUG "Write 0x%x bytes at 0x%08x in wbuf recover\n",
286 towrite, ofs));
182ec4ee 287
1da177e4
LT
288#ifdef BREAKMEHEADER
289 static int breakme;
290 if (breakme++ == 20) {
291 printk(KERN_NOTICE "Faking write error at 0x%08x\n", ofs);
292 breakme = 0;
9223a456
TG
293 c->mtd->write(c->mtd, ofs, towrite, &retlen,
294 brokenbuf);
1da177e4
LT
295 ret = -EIO;
296 } else
297#endif
9223a456
TG
298 ret = c->mtd->write(c->mtd, ofs, towrite, &retlen,
299 rewrite_buf);
1da177e4
LT
300
301 if (ret || retlen != towrite) {
302 /* Argh. We tried. Really we did. */
303 printk(KERN_CRIT "Recovery of wbuf failed due to a second write error\n");
9b88f473 304 kfree(buf);
1da177e4 305
2f785402
DW
306 if (retlen)
307 jffs2_add_physical_node_ref(c, ofs | REF_OBSOLETE, ref_totlen(c, jeb, *first_raw), NULL);
1da177e4 308
1da177e4
LT
309 return;
310 }
311 printk(KERN_NOTICE "Recovery of wbuf succeeded to %08x\n", ofs);
312
313 c->wbuf_len = (end - start) - towrite;
314 c->wbuf_ofs = ofs + towrite;
7f716cf3 315 memmove(c->wbuf, rewrite_buf + towrite, c->wbuf_len);
1da177e4 316 /* Don't muck about with c->wbuf_inodes. False positives are harmless. */
f99d49ad 317 kfree(buf);
1da177e4
LT
318 } else {
319 /* OK, now we're left with the dregs in whichever buffer we're using */
320 if (buf) {
321 memcpy(c->wbuf, buf, end-start);
322 kfree(buf);
323 } else {
324 memmove(c->wbuf, c->wbuf + (start - c->wbuf_ofs), end - start);
325 }
326 c->wbuf_ofs = ofs;
327 c->wbuf_len = end - start;
328 }
329
330 /* Now sort out the jffs2_raw_node_refs, moving them from the old to the next block */
331 new_jeb = &c->blocks[ofs / c->sector_size];
332
333 spin_lock(&c->erase_completion_lock);
334 if (new_jeb->first_node) {
335 /* Odd, but possible with ST flash later maybe */
336 new_jeb->last_node->next_phys = *first_raw;
337 } else {
338 new_jeb->first_node = *first_raw;
339 }
340
341 raw = first_raw;
342 while (*raw) {
343 uint32_t rawlen = ref_totlen(c, jeb, *raw);
344
345 D1(printk(KERN_DEBUG "Refiling block of %08x at %08x(%d) to %08x\n",
346 rawlen, ref_offset(*raw), ref_flags(*raw), ofs));
347
348 if (ref_obsolete(*raw)) {
349 /* Shouldn't really happen much */
350 new_jeb->dirty_size += rawlen;
351 new_jeb->free_size -= rawlen;
352 c->dirty_size += rawlen;
353 } else {
354 new_jeb->used_size += rawlen;
355 new_jeb->free_size -= rawlen;
356 jeb->dirty_size += rawlen;
357 jeb->used_size -= rawlen;
358 c->dirty_size += rawlen;
359 }
360 c->free_size -= rawlen;
361 (*raw)->flash_offset = ofs | ref_flags(*raw);
362 ofs += rawlen;
363 new_jeb->last_node = *raw;
364
365 raw = &(*raw)->next_phys;
366 }
367
368 /* Fix up the original jeb now it's on the bad_list */
369 *first_raw = NULL;
370 if (first_raw == &jeb->first_node) {
371 jeb->last_node = NULL;
372 D1(printk(KERN_DEBUG "Failing block at %08x is now empty. Moving to erase_pending_list\n", jeb->offset));
373 list_del(&jeb->list);
374 list_add(&jeb->list, &c->erase_pending_list);
375 c->nr_erasing_blocks++;
376 jffs2_erase_pending_trigger(c);
377 }
378 else
379 jeb->last_node = container_of(first_raw, struct jffs2_raw_node_ref, next_phys);
380
e0c8e42f
AB
381 jffs2_dbg_acct_sanity_check_nolock(c, jeb);
382 jffs2_dbg_acct_paranoia_check_nolock(c, jeb);
1da177e4 383
e0c8e42f
AB
384 jffs2_dbg_acct_sanity_check_nolock(c, new_jeb);
385 jffs2_dbg_acct_paranoia_check_nolock(c, new_jeb);
1da177e4
LT
386
387 spin_unlock(&c->erase_completion_lock);
388
389 D1(printk(KERN_DEBUG "wbuf recovery completed OK\n"));
390}
391
392/* Meaning of pad argument:
393 0: Do not pad. Probably pointless - we only ever use this when we can't pad anyway.
394 1: Pad, do not adjust nextblock free_size
395 2: Pad, adjust nextblock free_size
396*/
397#define NOPAD 0
398#define PAD_NOACCOUNT 1
399#define PAD_ACCOUNTING 2
400
401static int __jffs2_flush_wbuf(struct jffs2_sb_info *c, int pad)
402{
403 int ret;
404 size_t retlen;
405
3be36675 406 /* Nothing to do if not write-buffering the flash. In particular, we shouldn't
1da177e4 407 del_timer() the timer we never initialised. */
3be36675 408 if (!jffs2_is_writebuffered(c))
1da177e4
LT
409 return 0;
410
411 if (!down_trylock(&c->alloc_sem)) {
412 up(&c->alloc_sem);
413 printk(KERN_CRIT "jffs2_flush_wbuf() called with alloc_sem not locked!\n");
414 BUG();
415 }
416
3be36675 417 if (!c->wbuf_len) /* already checked c->wbuf above */
1da177e4
LT
418 return 0;
419
046b8b98 420 if (jffs2_prealloc_raw_node_refs(c, c->nextblock, c->reserved_refs + 1))
2f785402
DW
421 return -ENOMEM;
422
1da177e4
LT
423 /* claim remaining space on the page
424 this happens, if we have a change to a new block,
425 or if fsync forces us to flush the writebuffer.
426 if we have a switch to next page, we will not have
182ec4ee 427 enough remaining space for this.
1da177e4 428 */
daba5cc4 429 if (pad ) {
1da177e4
LT
430 c->wbuf_len = PAD(c->wbuf_len);
431
432 /* Pad with JFFS2_DIRTY_BITMASK initially. this helps out ECC'd NOR
433 with 8 byte page size */
434 memset(c->wbuf + c->wbuf_len, 0, c->wbuf_pagesize - c->wbuf_len);
182ec4ee 435
1da177e4
LT
436 if ( c->wbuf_len + sizeof(struct jffs2_unknown_node) < c->wbuf_pagesize) {
437 struct jffs2_unknown_node *padnode = (void *)(c->wbuf + c->wbuf_len);
438 padnode->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
439 padnode->nodetype = cpu_to_je16(JFFS2_NODETYPE_PADDING);
440 padnode->totlen = cpu_to_je32(c->wbuf_pagesize - c->wbuf_len);
441 padnode->hdr_crc = cpu_to_je32(crc32(0, padnode, sizeof(*padnode)-4));
442 }
443 }
444 /* else jffs2_flash_writev has actually filled in the rest of the
445 buffer for us, and will deal with the node refs etc. later. */
182ec4ee 446
1da177e4
LT
447#ifdef BREAKME
448 static int breakme;
449 if (breakme++ == 20) {
450 printk(KERN_NOTICE "Faking write error at 0x%08x\n", c->wbuf_ofs);
451 breakme = 0;
9223a456
TG
452 c->mtd->write(c->mtd, c->wbuf_ofs, c->wbuf_pagesize, &retlen,
453 brokenbuf);
1da177e4 454 ret = -EIO;
182ec4ee 455 } else
1da177e4 456#endif
182ec4ee 457
1da177e4
LT
458 ret = c->mtd->write(c->mtd, c->wbuf_ofs, c->wbuf_pagesize, &retlen, c->wbuf);
459
460 if (ret || retlen != c->wbuf_pagesize) {
461 if (ret)
462 printk(KERN_WARNING "jffs2_flush_wbuf(): Write failed with %d\n",ret);
463 else {
464 printk(KERN_WARNING "jffs2_flush_wbuf(): Write was short: %zd instead of %d\n",
465 retlen, c->wbuf_pagesize);
466 ret = -EIO;
467 }
468
469 jffs2_wbuf_recover(c);
470
471 return ret;
472 }
473
1da177e4 474 /* Adjust free size of the block if we padded. */
daba5cc4 475 if (pad) {
1da177e4 476 struct jffs2_eraseblock *jeb;
0bcc099d 477 uint32_t waste = c->wbuf_pagesize - c->wbuf_len;
1da177e4
LT
478
479 jeb = &c->blocks[c->wbuf_ofs / c->sector_size];
480
481 D1(printk(KERN_DEBUG "jffs2_flush_wbuf() adjusting free_size of %sblock at %08x\n",
482 (jeb==c->nextblock)?"next":"", jeb->offset));
483
182ec4ee 484 /* wbuf_pagesize - wbuf_len is the amount of space that's to be
1da177e4
LT
485 padded. If there is less free space in the block than that,
486 something screwed up */
0bcc099d 487 if (jeb->free_size < waste) {
1da177e4 488 printk(KERN_CRIT "jffs2_flush_wbuf(): Accounting error. wbuf at 0x%08x has 0x%03x bytes, 0x%03x left.\n",
0bcc099d 489 c->wbuf_ofs, c->wbuf_len, waste);
1da177e4
LT
490 printk(KERN_CRIT "jffs2_flush_wbuf(): But free_size for block at 0x%08x is only 0x%08x\n",
491 jeb->offset, jeb->free_size);
492 BUG();
493 }
0bcc099d
DW
494
495 spin_lock(&c->erase_completion_lock);
496
2f785402 497 jffs2_link_node_ref(c, jeb, (c->wbuf_ofs + c->wbuf_len) | REF_OBSOLETE, waste, NULL);
0bcc099d
DW
498 /* FIXME: that made it count as dirty. Convert to wasted */
499 jeb->dirty_size -= waste;
500 c->dirty_size -= waste;
501 jeb->wasted_size += waste;
502 c->wasted_size += waste;
503 } else
504 spin_lock(&c->erase_completion_lock);
1da177e4
LT
505
506 /* Stick any now-obsoleted blocks on the erase_pending_list */
507 jffs2_refile_wbuf_blocks(c);
508 jffs2_clear_wbuf_ino_list(c);
509 spin_unlock(&c->erase_completion_lock);
510
511 memset(c->wbuf,0xff,c->wbuf_pagesize);
512 /* adjust write buffer offset, else we get a non contiguous write bug */
513 c->wbuf_ofs += c->wbuf_pagesize;
514 c->wbuf_len = 0;
515 return 0;
516}
517
182ec4ee 518/* Trigger garbage collection to flush the write-buffer.
1da177e4 519 If ino arg is zero, do it if _any_ real (i.e. not GC) writes are
182ec4ee 520 outstanding. If ino arg non-zero, do it only if a write for the
1da177e4
LT
521 given inode is outstanding. */
522int jffs2_flush_wbuf_gc(struct jffs2_sb_info *c, uint32_t ino)
523{
524 uint32_t old_wbuf_ofs;
525 uint32_t old_wbuf_len;
526 int ret = 0;
527
528 D1(printk(KERN_DEBUG "jffs2_flush_wbuf_gc() called for ino #%u...\n", ino));
529
8aee6ac1
DW
530 if (!c->wbuf)
531 return 0;
532
1da177e4
LT
533 down(&c->alloc_sem);
534 if (!jffs2_wbuf_pending_for_ino(c, ino)) {
535 D1(printk(KERN_DEBUG "Ino #%d not pending in wbuf. Returning\n", ino));
536 up(&c->alloc_sem);
537 return 0;
538 }
539
540 old_wbuf_ofs = c->wbuf_ofs;
541 old_wbuf_len = c->wbuf_len;
542
543 if (c->unchecked_size) {
544 /* GC won't make any progress for a while */
545 D1(printk(KERN_DEBUG "jffs2_flush_wbuf_gc() padding. Not finished checking\n"));
546 down_write(&c->wbuf_sem);
547 ret = __jffs2_flush_wbuf(c, PAD_ACCOUNTING);
7f716cf3
EH
548 /* retry flushing wbuf in case jffs2_wbuf_recover
549 left some data in the wbuf */
550 if (ret)
7f716cf3 551 ret = __jffs2_flush_wbuf(c, PAD_ACCOUNTING);
1da177e4
LT
552 up_write(&c->wbuf_sem);
553 } else while (old_wbuf_len &&
554 old_wbuf_ofs == c->wbuf_ofs) {
555
556 up(&c->alloc_sem);
557
558 D1(printk(KERN_DEBUG "jffs2_flush_wbuf_gc() calls gc pass\n"));
559
560 ret = jffs2_garbage_collect_pass(c);
561 if (ret) {
562 /* GC failed. Flush it with padding instead */
563 down(&c->alloc_sem);
564 down_write(&c->wbuf_sem);
565 ret = __jffs2_flush_wbuf(c, PAD_ACCOUNTING);
7f716cf3
EH
566 /* retry flushing wbuf in case jffs2_wbuf_recover
567 left some data in the wbuf */
568 if (ret)
7f716cf3 569 ret = __jffs2_flush_wbuf(c, PAD_ACCOUNTING);
1da177e4
LT
570 up_write(&c->wbuf_sem);
571 break;
572 }
573 down(&c->alloc_sem);
574 }
575
576 D1(printk(KERN_DEBUG "jffs2_flush_wbuf_gc() ends...\n"));
577
578 up(&c->alloc_sem);
579 return ret;
580}
581
582/* Pad write-buffer to end and write it, wasting space. */
583int jffs2_flush_wbuf_pad(struct jffs2_sb_info *c)
584{
585 int ret;
586
8aee6ac1
DW
587 if (!c->wbuf)
588 return 0;
589
1da177e4
LT
590 down_write(&c->wbuf_sem);
591 ret = __jffs2_flush_wbuf(c, PAD_NOACCOUNT);
7f716cf3
EH
592 /* retry - maybe wbuf recover left some data in wbuf. */
593 if (ret)
594 ret = __jffs2_flush_wbuf(c, PAD_NOACCOUNT);
1da177e4
LT
595 up_write(&c->wbuf_sem);
596
597 return ret;
598}
dcb09328
TG
599
600static size_t jffs2_fill_wbuf(struct jffs2_sb_info *c, const uint8_t *buf,
601 size_t len)
1da177e4 602{
dcb09328
TG
603 if (len && !c->wbuf_len && (len >= c->wbuf_pagesize))
604 return 0;
605
606 if (len > (c->wbuf_pagesize - c->wbuf_len))
607 len = c->wbuf_pagesize - c->wbuf_len;
608 memcpy(c->wbuf + c->wbuf_len, buf, len);
609 c->wbuf_len += (uint32_t) len;
610 return len;
611}
612
613int jffs2_flash_writev(struct jffs2_sb_info *c, const struct kvec *invecs,
614 unsigned long count, loff_t to, size_t *retlen,
615 uint32_t ino)
616{
617 struct jffs2_eraseblock *jeb;
618 size_t wbuf_retlen, donelen = 0;
1da177e4 619 uint32_t outvec_to = to;
dcb09328 620 int ret, invec;
1da177e4 621
dcb09328 622 /* If not writebuffered flash, don't bother */
3be36675 623 if (!jffs2_is_writebuffered(c))
1da177e4 624 return jffs2_flash_direct_writev(c, invecs, count, to, retlen);
182ec4ee 625
1da177e4
LT
626 down_write(&c->wbuf_sem);
627
628 /* If wbuf_ofs is not initialized, set it to target address */
629 if (c->wbuf_ofs == 0xFFFFFFFF) {
630 c->wbuf_ofs = PAGE_DIV(to);
182ec4ee 631 c->wbuf_len = PAGE_MOD(to);
1da177e4
LT
632 memset(c->wbuf,0xff,c->wbuf_pagesize);
633 }
634
dcb09328
TG
635 /*
636 * Sanity checks on target address. It's permitted to write
637 * at PAD(c->wbuf_len+c->wbuf_ofs), and it's permitted to
638 * write at the beginning of a new erase block. Anything else,
639 * and you die. New block starts at xxx000c (0-b = block
640 * header)
641 */
3be36675 642 if (SECTOR_ADDR(to) != SECTOR_ADDR(c->wbuf_ofs)) {
1da177e4
LT
643 /* It's a write to a new block */
644 if (c->wbuf_len) {
dcb09328
TG
645 D1(printk(KERN_DEBUG "jffs2_flash_writev() to 0x%lx "
646 "causes flush of wbuf at 0x%08x\n",
647 (unsigned long)to, c->wbuf_ofs));
1da177e4 648 ret = __jffs2_flush_wbuf(c, PAD_NOACCOUNT);
dcb09328
TG
649 if (ret)
650 goto outerr;
1da177e4
LT
651 }
652 /* set pointer to new block */
653 c->wbuf_ofs = PAGE_DIV(to);
182ec4ee
TG
654 c->wbuf_len = PAGE_MOD(to);
655 }
1da177e4
LT
656
657 if (to != PAD(c->wbuf_ofs + c->wbuf_len)) {
658 /* We're not writing immediately after the writebuffer. Bad. */
dcb09328
TG
659 printk(KERN_CRIT "jffs2_flash_writev(): Non-contiguous write "
660 "to %08lx\n", (unsigned long)to);
1da177e4
LT
661 if (c->wbuf_len)
662 printk(KERN_CRIT "wbuf was previously %08x-%08x\n",
dcb09328 663 c->wbuf_ofs, c->wbuf_ofs+c->wbuf_len);
1da177e4
LT
664 BUG();
665 }
666
dcb09328
TG
667 /* adjust alignment offset */
668 if (c->wbuf_len != PAGE_MOD(to)) {
669 c->wbuf_len = PAGE_MOD(to);
670 /* take care of alignment to next page */
671 if (!c->wbuf_len) {
672 c->wbuf_len = c->wbuf_pagesize;
673 ret = __jffs2_flush_wbuf(c, NOPAD);
674 if (ret)
675 goto outerr;
1da177e4
LT
676 }
677 }
678
dcb09328
TG
679 for (invec = 0; invec < count; invec++) {
680 int vlen = invecs[invec].iov_len;
681 uint8_t *v = invecs[invec].iov_base;
7f716cf3 682
dcb09328 683 wbuf_retlen = jffs2_fill_wbuf(c, v, vlen);
7f716cf3 684
dcb09328
TG
685 if (c->wbuf_len == c->wbuf_pagesize) {
686 ret = __jffs2_flush_wbuf(c, NOPAD);
687 if (ret)
688 goto outerr;
1da177e4 689 }
dcb09328
TG
690 vlen -= wbuf_retlen;
691 outvec_to += wbuf_retlen;
1da177e4 692 donelen += wbuf_retlen;
dcb09328
TG
693 v += wbuf_retlen;
694
695 if (vlen >= c->wbuf_pagesize) {
696 ret = c->mtd->write(c->mtd, outvec_to, PAGE_DIV(vlen),
697 &wbuf_retlen, v);
698 if (ret < 0 || wbuf_retlen != PAGE_DIV(vlen))
699 goto outfile;
700
701 vlen -= wbuf_retlen;
702 outvec_to += wbuf_retlen;
703 c->wbuf_ofs = outvec_to;
704 donelen += wbuf_retlen;
705 v += wbuf_retlen;
1da177e4
LT
706 }
707
dcb09328
TG
708 wbuf_retlen = jffs2_fill_wbuf(c, v, vlen);
709 if (c->wbuf_len == c->wbuf_pagesize) {
710 ret = __jffs2_flush_wbuf(c, NOPAD);
711 if (ret)
712 goto outerr;
713 }
1da177e4 714
dcb09328
TG
715 outvec_to += wbuf_retlen;
716 donelen += wbuf_retlen;
1da177e4 717 }
1da177e4 718
dcb09328
TG
719 /*
720 * If there's a remainder in the wbuf and it's a non-GC write,
721 * remember that the wbuf affects this ino
722 */
1da177e4
LT
723 *retlen = donelen;
724
e631ddba
FH
725 if (jffs2_sum_active()) {
726 int res = jffs2_sum_add_kvec(c, invecs, count, (uint32_t) to);
727 if (res)
728 return res;
729 }
730
1da177e4
LT
731 if (c->wbuf_len && ino)
732 jffs2_wbuf_dirties_inode(c, ino);
733
734 ret = 0;
dcb09328
TG
735 up_write(&c->wbuf_sem);
736 return ret;
737
738outfile:
739 /*
740 * At this point we have no problem, c->wbuf is empty. However
741 * refile nextblock to avoid writing again to same address.
742 */
743
744 spin_lock(&c->erase_completion_lock);
745
746 jeb = &c->blocks[outvec_to / c->sector_size];
747 jffs2_block_refile(c, jeb, REFILE_ANYWAY);
748
749 spin_unlock(&c->erase_completion_lock);
182ec4ee 750
dcb09328
TG
751outerr:
752 *retlen = 0;
1da177e4
LT
753 up_write(&c->wbuf_sem);
754 return ret;
755}
756
757/*
758 * This is the entry for flash write.
759 * Check, if we work on NAND FLASH, if so build an kvec and write it via vritev
760*/
761int jffs2_flash_write(struct jffs2_sb_info *c, loff_t ofs, size_t len, size_t *retlen, const u_char *buf)
762{
763 struct kvec vecs[1];
764
3be36675 765 if (!jffs2_is_writebuffered(c))
e631ddba 766 return jffs2_flash_direct_write(c, ofs, len, retlen, buf);
1da177e4
LT
767
768 vecs[0].iov_base = (unsigned char *) buf;
769 vecs[0].iov_len = len;
770 return jffs2_flash_writev(c, vecs, 1, ofs, retlen, 0);
771}
772
773/*
774 Handle readback from writebuffer and ECC failure return
775*/
776int jffs2_flash_read(struct jffs2_sb_info *c, loff_t ofs, size_t len, size_t *retlen, u_char *buf)
777{
778 loff_t orbf = 0, owbf = 0, lwbf = 0;
779 int ret;
780
3be36675 781 if (!jffs2_is_writebuffered(c))
1da177e4
LT
782 return c->mtd->read(c->mtd, ofs, len, retlen, buf);
783
3be36675 784 /* Read flash */
894214d1 785 down_read(&c->wbuf_sem);
9223a456 786 ret = c->mtd->read(c->mtd, ofs, len, retlen, buf);
3be36675
AV
787
788 if ( (ret == -EBADMSG) && (*retlen == len) ) {
789 printk(KERN_WARNING "mtd->read(0x%zx bytes from 0x%llx) returned ECC error\n",
790 len, ofs);
182ec4ee
TG
791 /*
792 * We have the raw data without ECC correction in the buffer, maybe
3be36675
AV
793 * we are lucky and all data or parts are correct. We check the node.
794 * If data are corrupted node check will sort it out.
795 * We keep this block, it will fail on write or erase and the we
796 * mark it bad. Or should we do that now? But we should give him a chance.
182ec4ee 797 * Maybe we had a system crash or power loss before the ecc write or
3be36675
AV
798 * a erase was completed.
799 * So we return success. :)
800 */
801 ret = 0;
182ec4ee 802 }
3be36675 803
1da177e4
LT
804 /* if no writebuffer available or write buffer empty, return */
805 if (!c->wbuf_pagesize || !c->wbuf_len)
894214d1 806 goto exit;
1da177e4
LT
807
808 /* if we read in a different block, return */
3be36675 809 if (SECTOR_ADDR(ofs) != SECTOR_ADDR(c->wbuf_ofs))
894214d1 810 goto exit;
1da177e4
LT
811
812 if (ofs >= c->wbuf_ofs) {
813 owbf = (ofs - c->wbuf_ofs); /* offset in write buffer */
814 if (owbf > c->wbuf_len) /* is read beyond write buffer ? */
815 goto exit;
816 lwbf = c->wbuf_len - owbf; /* number of bytes to copy */
182ec4ee 817 if (lwbf > len)
1da177e4 818 lwbf = len;
182ec4ee 819 } else {
1da177e4
LT
820 orbf = (c->wbuf_ofs - ofs); /* offset in read buffer */
821 if (orbf > len) /* is write beyond write buffer ? */
822 goto exit;
823 lwbf = len - orbf; /* number of bytes to copy */
182ec4ee 824 if (lwbf > c->wbuf_len)
1da177e4 825 lwbf = c->wbuf_len;
182ec4ee 826 }
1da177e4
LT
827 if (lwbf > 0)
828 memcpy(buf+orbf,c->wbuf+owbf,lwbf);
829
830exit:
831 up_read(&c->wbuf_sem);
832 return ret;
833}
834
835/*
836 * Check, if the out of band area is empty
837 */
838int jffs2_check_oob_empty( struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, int mode)
839{
840 unsigned char *buf;
841 int ret = 0;
842 int i,len,page;
843 size_t retlen;
844 int oob_size;
845
846 /* allocate a buffer for all oob data in this sector */
847 oob_size = c->mtd->oobsize;
848 len = 4 * oob_size;
849 buf = kmalloc(len, GFP_KERNEL);
850 if (!buf) {
851 printk(KERN_NOTICE "jffs2_check_oob_empty(): allocation of temporary data buffer for oob check failed\n");
852 return -ENOMEM;
853 }
182ec4ee 854 /*
1da177e4
LT
855 * if mode = 0, we scan for a total empty oob area, else we have
856 * to take care of the cleanmarker in the first page of the block
857 */
858 ret = jffs2_flash_read_oob(c, jeb->offset, len , &retlen, buf);
859 if (ret) {
860 D1(printk(KERN_WARNING "jffs2_check_oob_empty(): Read OOB failed %d for block at %08x\n", ret, jeb->offset));
861 goto out;
862 }
182ec4ee 863
1da177e4
LT
864 if (retlen < len) {
865 D1(printk(KERN_WARNING "jffs2_check_oob_empty(): Read OOB return short read "
866 "(%zd bytes not %d) for block at %08x\n", retlen, len, jeb->offset));
867 ret = -EIO;
868 goto out;
869 }
182ec4ee 870
1da177e4
LT
871 /* Special check for first page */
872 for(i = 0; i < oob_size ; i++) {
873 /* Yeah, we know about the cleanmarker. */
182ec4ee 874 if (mode && i >= c->fsdata_pos &&
1da177e4
LT
875 i < c->fsdata_pos + c->fsdata_len)
876 continue;
877
878 if (buf[i] != 0xFF) {
879 D2(printk(KERN_DEBUG "Found %02x at %x in OOB for %08x\n",
730554d9 880 buf[i], i, jeb->offset));
182ec4ee 881 ret = 1;
1da177e4
LT
882 goto out;
883 }
884 }
885
182ec4ee 886 /* we know, we are aligned :) */
1da177e4
LT
887 for (page = oob_size; page < len; page += sizeof(long)) {
888 unsigned long dat = *(unsigned long *)(&buf[page]);
889 if(dat != -1) {
182ec4ee 890 ret = 1;
1da177e4
LT
891 goto out;
892 }
893 }
894
895out:
182ec4ee
TG
896 kfree(buf);
897
1da177e4
LT
898 return ret;
899}
900
901/*
902* Scan for a valid cleanmarker and for bad blocks
903* For virtual blocks (concatenated physical blocks) check the cleanmarker
904* only in the first page of the first physical block, but scan for bad blocks in all
905* physical blocks
906*/
907int jffs2_check_nand_cleanmarker (struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb)
908{
909 struct jffs2_unknown_node n;
910 unsigned char buf[2 * NAND_MAX_OOBSIZE];
911 unsigned char *p;
912 int ret, i, cnt, retval = 0;
913 size_t retlen, offset;
914 int oob_size;
915
916 offset = jeb->offset;
917 oob_size = c->mtd->oobsize;
918
919 /* Loop through the physical blocks */
920 for (cnt = 0; cnt < (c->sector_size / c->mtd->erasesize); cnt++) {
921 /* Check first if the block is bad. */
922 if (c->mtd->block_isbad (c->mtd, offset)) {
923 D1 (printk (KERN_WARNING "jffs2_check_nand_cleanmarker(): Bad block at %08x\n", jeb->offset));
924 return 2;
925 }
926 /*
927 * We read oob data from page 0 and 1 of the block.
928 * page 0 contains cleanmarker and badblock info
929 * page 1 contains failure count of this block
930 */
931 ret = c->mtd->read_oob (c->mtd, offset, oob_size << 1, &retlen, buf);
932
933 if (ret) {
934 D1 (printk (KERN_WARNING "jffs2_check_nand_cleanmarker(): Read OOB failed %d for block at %08x\n", ret, jeb->offset));
935 return ret;
936 }
937 if (retlen < (oob_size << 1)) {
938 D1 (printk (KERN_WARNING "jffs2_check_nand_cleanmarker(): Read OOB return short read (%zd bytes not %d) for block at %08x\n", retlen, oob_size << 1, jeb->offset));
939 return -EIO;
940 }
941
942 /* Check cleanmarker only on the first physical block */
943 if (!cnt) {
944 n.magic = cpu_to_je16 (JFFS2_MAGIC_BITMASK);
945 n.nodetype = cpu_to_je16 (JFFS2_NODETYPE_CLEANMARKER);
946 n.totlen = cpu_to_je32 (8);
947 p = (unsigned char *) &n;
948
949 for (i = 0; i < c->fsdata_len; i++) {
950 if (buf[c->fsdata_pos + i] != p[i]) {
951 retval = 1;
952 }
953 }
954 D1(if (retval == 1) {
955 printk(KERN_WARNING "jffs2_check_nand_cleanmarker(): Cleanmarker node not detected in block at %08x\n", jeb->offset);
956 printk(KERN_WARNING "OOB at %08x was ", offset);
957 for (i=0; i < oob_size; i++) {
958 printk("%02x ", buf[i]);
959 }
960 printk("\n");
961 })
962 }
963 offset += c->mtd->erasesize;
964 }
965 return retval;
966}
967
968int jffs2_write_nand_cleanmarker(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb)
969{
970 struct jffs2_unknown_node n;
971 int ret;
972 size_t retlen;
973
974 n.magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
975 n.nodetype = cpu_to_je16(JFFS2_NODETYPE_CLEANMARKER);
976 n.totlen = cpu_to_je32(8);
977
978 ret = jffs2_flash_write_oob(c, jeb->offset + c->fsdata_pos, c->fsdata_len, &retlen, (unsigned char *)&n);
182ec4ee 979
1da177e4
LT
980 if (ret) {
981 D1(printk(KERN_WARNING "jffs2_write_nand_cleanmarker(): Write failed for block at %08x: error %d\n", jeb->offset, ret));
982 return ret;
983 }
984 if (retlen != c->fsdata_len) {
985 D1(printk(KERN_WARNING "jffs2_write_nand_cleanmarker(): Short write for block at %08x: %zd not %d\n", jeb->offset, retlen, c->fsdata_len));
986 return ret;
987 }
988 return 0;
989}
990
182ec4ee 991/*
1da177e4
LT
992 * On NAND we try to mark this block bad. If the block was erased more
993 * than MAX_ERASE_FAILURES we mark it finaly bad.
994 * Don't care about failures. This block remains on the erase-pending
995 * or badblock list as long as nobody manipulates the flash with
996 * a bootloader or something like that.
997 */
998
999int jffs2_write_nand_badblock(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, uint32_t bad_offset)
1000{
1001 int ret;
1002
1003 /* if the count is < max, we try to write the counter to the 2nd page oob area */
1004 if( ++jeb->bad_count < MAX_ERASE_FAILURES)
1005 return 0;
1006
1007 if (!c->mtd->block_markbad)
1008 return 1; // What else can we do?
1009
1010 D1(printk(KERN_WARNING "jffs2_write_nand_badblock(): Marking bad block at %08x\n", bad_offset));
1011 ret = c->mtd->block_markbad(c->mtd, bad_offset);
182ec4ee 1012
1da177e4
LT
1013 if (ret) {
1014 D1(printk(KERN_WARNING "jffs2_write_nand_badblock(): Write failed for block at %08x: error %d\n", jeb->offset, ret));
1015 return ret;
1016 }
1017 return 1;
1018}
1019
1020#define NAND_JFFS2_OOB16_FSDALEN 8
1021
1022static struct nand_oobinfo jffs2_oobinfo_docecc = {
1023 .useecc = MTD_NANDECC_PLACE,
1024 .eccbytes = 6,
1025 .eccpos = {0,1,2,3,4,5}
1026};
1027
1028
1029static int jffs2_nand_set_oobinfo(struct jffs2_sb_info *c)
1030{
1031 struct nand_oobinfo *oinfo = &c->mtd->oobinfo;
1032
1033 /* Do this only, if we have an oob buffer */
1034 if (!c->mtd->oobsize)
1035 return 0;
182ec4ee 1036
1da177e4
LT
1037 /* Cleanmarker is out-of-band, so inline size zero */
1038 c->cleanmarker_size = 0;
1039
1040 /* Should we use autoplacement ? */
1041 if (oinfo && oinfo->useecc == MTD_NANDECC_AUTOPLACE) {
1042 D1(printk(KERN_DEBUG "JFFS2 using autoplace on NAND\n"));
1043 /* Get the position of the free bytes */
1044 if (!oinfo->oobfree[0][1]) {
1045 printk (KERN_WARNING "jffs2_nand_set_oobinfo(): Eeep. Autoplacement selected and no empty space in oob\n");
1046 return -ENOSPC;
1047 }
1048 c->fsdata_pos = oinfo->oobfree[0][0];
1049 c->fsdata_len = oinfo->oobfree[0][1];
1050 if (c->fsdata_len > 8)
1051 c->fsdata_len = 8;
1052 } else {
1053 /* This is just a legacy fallback and should go away soon */
1054 switch(c->mtd->ecctype) {
1055 case MTD_ECC_RS_DiskOnChip:
1056 printk(KERN_WARNING "JFFS2 using DiskOnChip hardware ECC without autoplacement. Fix it!\n");
1057 c->oobinfo = &jffs2_oobinfo_docecc;
1058 c->fsdata_pos = 6;
1059 c->fsdata_len = NAND_JFFS2_OOB16_FSDALEN;
1060 c->badblock_pos = 15;
1061 break;
182ec4ee 1062
1da177e4
LT
1063 default:
1064 D1(printk(KERN_DEBUG "JFFS2 on NAND. No autoplacment info found\n"));
1065 return -EINVAL;
1066 }
1067 }
1068 return 0;
1069}
1070
1071int jffs2_nand_flash_setup(struct jffs2_sb_info *c)
1072{
1073 int res;
1074
1075 /* Initialise write buffer */
1076 init_rwsem(&c->wbuf_sem);
28318776 1077 c->wbuf_pagesize = c->mtd->writesize;
1da177e4 1078 c->wbuf_ofs = 0xFFFFFFFF;
182ec4ee 1079
1da177e4
LT
1080 c->wbuf = kmalloc(c->wbuf_pagesize, GFP_KERNEL);
1081 if (!c->wbuf)
1082 return -ENOMEM;
1083
1084 res = jffs2_nand_set_oobinfo(c);
1085
1086#ifdef BREAKME
1087 if (!brokenbuf)
1088 brokenbuf = kmalloc(c->wbuf_pagesize, GFP_KERNEL);
1089 if (!brokenbuf) {
1090 kfree(c->wbuf);
1091 return -ENOMEM;
1092 }
1093 memset(brokenbuf, 0xdb, c->wbuf_pagesize);
1094#endif
1095 return res;
1096}
1097
1098void jffs2_nand_flash_cleanup(struct jffs2_sb_info *c)
1099{
1100 kfree(c->wbuf);
1101}
1102
8f15fd55
AV
1103int jffs2_dataflash_setup(struct jffs2_sb_info *c) {
1104 c->cleanmarker_size = 0; /* No cleanmarkers needed */
182ec4ee 1105
8f15fd55
AV
1106 /* Initialize write buffer */
1107 init_rwsem(&c->wbuf_sem);
8f15fd55 1108
182ec4ee 1109
daba5cc4 1110 c->wbuf_pagesize = c->mtd->erasesize;
182ec4ee 1111
daba5cc4
AB
1112 /* Find a suitable c->sector_size
1113 * - Not too much sectors
1114 * - Sectors have to be at least 4 K + some bytes
1115 * - All known dataflashes have erase sizes of 528 or 1056
1116 * - we take at least 8 eraseblocks and want to have at least 8K size
1117 * - The concatenation should be a power of 2
1118 */
1119
1120 c->sector_size = 8 * c->mtd->erasesize;
182ec4ee 1121
daba5cc4
AB
1122 while (c->sector_size < 8192) {
1123 c->sector_size *= 2;
1124 }
182ec4ee 1125
daba5cc4
AB
1126 /* It may be necessary to adjust the flash size */
1127 c->flash_size = c->mtd->size;
8f15fd55 1128
daba5cc4
AB
1129 if ((c->flash_size % c->sector_size) != 0) {
1130 c->flash_size = (c->flash_size / c->sector_size) * c->sector_size;
1131 printk(KERN_WARNING "JFFS2 flash size adjusted to %dKiB\n", c->flash_size);
1132 };
182ec4ee 1133
daba5cc4 1134 c->wbuf_ofs = 0xFFFFFFFF;
8f15fd55
AV
1135 c->wbuf = kmalloc(c->wbuf_pagesize, GFP_KERNEL);
1136 if (!c->wbuf)
1137 return -ENOMEM;
1138
daba5cc4 1139 printk(KERN_INFO "JFFS2 write-buffering enabled buffer (%d) erasesize (%d)\n", c->wbuf_pagesize, c->sector_size);
8f15fd55
AV
1140
1141 return 0;
1142}
1143
1144void jffs2_dataflash_cleanup(struct jffs2_sb_info *c) {
1145 kfree(c->wbuf);
1146}
8f15fd55 1147
59da721a 1148int jffs2_nor_wbuf_flash_setup(struct jffs2_sb_info *c) {
c8b229de
JE
1149 /* Cleanmarker currently occupies whole programming regions,
1150 * either one or 2 for 8Byte STMicro flashes. */
1151 c->cleanmarker_size = max(16u, c->mtd->writesize);
59da721a
NP
1152
1153 /* Initialize write buffer */
1154 init_rwsem(&c->wbuf_sem);
28318776 1155 c->wbuf_pagesize = c->mtd->writesize;
59da721a
NP
1156 c->wbuf_ofs = 0xFFFFFFFF;
1157
1158 c->wbuf = kmalloc(c->wbuf_pagesize, GFP_KERNEL);
1159 if (!c->wbuf)
1160 return -ENOMEM;
1161
1162 return 0;
1163}
1164
1165void jffs2_nor_wbuf_flash_cleanup(struct jffs2_sb_info *c) {
1166 kfree(c->wbuf);
1167}