]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blame - fs/f2fs/super.c
f2fs: return wrong error number on f2fs_quota_write
[mirror_ubuntu-bionic-kernel.git] / fs / f2fs / super.c
CommitLineData
0a8165d7 1/*
aff063e2
JK
2 * fs/f2fs/super.c
3 *
4 * Copyright (c) 2012 Samsung Electronics Co., Ltd.
5 * http://www.samsung.com/
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 */
11#include <linux/module.h>
12#include <linux/init.h>
13#include <linux/fs.h>
14#include <linux/statfs.h>
aff063e2
JK
15#include <linux/buffer_head.h>
16#include <linux/backing-dev.h>
17#include <linux/kthread.h>
18#include <linux/parser.h>
19#include <linux/mount.h>
20#include <linux/seq_file.h>
5e176d54 21#include <linux/proc_fs.h>
aff063e2
JK
22#include <linux/random.h>
23#include <linux/exportfs.h>
d3ee456d 24#include <linux/blkdev.h>
0abd675e 25#include <linux/quotaops.h>
aff063e2 26#include <linux/f2fs_fs.h>
b59d0bae 27#include <linux/sysfs.h>
aff063e2
JK
28
29#include "f2fs.h"
30#include "node.h"
5ec4e49f 31#include "segment.h"
aff063e2 32#include "xattr.h"
b59d0bae 33#include "gc.h"
db9f7c1a 34#include "trace.h"
aff063e2 35
a2a4a7e4
NJ
36#define CREATE_TRACE_POINTS
37#include <trace/events/f2fs.h>
38
aff063e2
JK
39static struct kmem_cache *f2fs_inode_cachep;
40
73faec4d 41#ifdef CONFIG_F2FS_FAULT_INJECTION
2c63fead
JK
42
43char *fault_name[FAULT_MAX] = {
44 [FAULT_KMALLOC] = "kmalloc",
c41f3cc3 45 [FAULT_PAGE_ALLOC] = "page alloc",
cb78942b
JK
46 [FAULT_ALLOC_NID] = "alloc nid",
47 [FAULT_ORPHAN] = "orphan",
48 [FAULT_BLOCK] = "no more block",
49 [FAULT_DIR_DEPTH] = "too big dir depth",
53aa6bbf 50 [FAULT_EVICT_INODE] = "evict_inode fail",
14b44d23 51 [FAULT_TRUNCATE] = "truncate fail",
8b038c70 52 [FAULT_IO] = "IO error",
0f348028 53 [FAULT_CHECKPOINT] = "checkpoint error",
2c63fead 54};
08796897 55
1ecc0c5c
CY
56static void f2fs_build_fault_attr(struct f2fs_sb_info *sbi,
57 unsigned int rate)
08796897 58{
1ecc0c5c
CY
59 struct f2fs_fault_info *ffi = &sbi->fault_info;
60
08796897 61 if (rate) {
1ecc0c5c
CY
62 atomic_set(&ffi->inject_ops, 0);
63 ffi->inject_rate = rate;
64 ffi->inject_type = (1 << FAULT_MAX) - 1;
08796897 65 } else {
1ecc0c5c 66 memset(ffi, 0, sizeof(struct f2fs_fault_info));
08796897
SY
67 }
68}
73faec4d
JK
69#endif
70
2658e50d
JK
71/* f2fs-wide shrinker description */
72static struct shrinker f2fs_shrinker_info = {
73 .scan_objects = f2fs_shrink_scan,
74 .count_objects = f2fs_shrink_count,
75 .seeks = DEFAULT_SEEKS,
76};
77
aff063e2 78enum {
696c018c 79 Opt_gc_background,
aff063e2 80 Opt_disable_roll_forward,
2d834bf9 81 Opt_norecovery,
aff063e2 82 Opt_discard,
64058be9 83 Opt_nodiscard,
aff063e2 84 Opt_noheap,
7a20b8a6 85 Opt_heap,
4058c511 86 Opt_user_xattr,
aff063e2 87 Opt_nouser_xattr,
4058c511 88 Opt_acl,
aff063e2
JK
89 Opt_noacl,
90 Opt_active_logs,
91 Opt_disable_ext_identify,
444c580f 92 Opt_inline_xattr,
23cf7212 93 Opt_noinline_xattr,
8274de77 94 Opt_inline_data,
5efd3c6f 95 Opt_inline_dentry,
97c1794a 96 Opt_noinline_dentry,
6b4afdd7 97 Opt_flush_merge,
69e9e427 98 Opt_noflush_merge,
0f7b2abd 99 Opt_nobarrier,
d5053a34 100 Opt_fastboot,
89672159 101 Opt_extent_cache,
7daaea25 102 Opt_noextent_cache,
75342797 103 Opt_noinline_data,
343f40f0 104 Opt_data_flush,
36abef4e 105 Opt_mode,
ec91538d 106 Opt_io_size_bits,
73faec4d 107 Opt_fault_injection,
6d94c74a
JK
108 Opt_lazytime,
109 Opt_nolazytime,
0abd675e
CY
110 Opt_usrquota,
111 Opt_grpquota,
5c57132e 112 Opt_prjquota,
aff063e2
JK
113 Opt_err,
114};
115
116static match_table_t f2fs_tokens = {
696c018c 117 {Opt_gc_background, "background_gc=%s"},
aff063e2 118 {Opt_disable_roll_forward, "disable_roll_forward"},
2d834bf9 119 {Opt_norecovery, "norecovery"},
aff063e2 120 {Opt_discard, "discard"},
64058be9 121 {Opt_nodiscard, "nodiscard"},
aff063e2 122 {Opt_noheap, "no_heap"},
7a20b8a6 123 {Opt_heap, "heap"},
4058c511 124 {Opt_user_xattr, "user_xattr"},
aff063e2 125 {Opt_nouser_xattr, "nouser_xattr"},
4058c511 126 {Opt_acl, "acl"},
aff063e2
JK
127 {Opt_noacl, "noacl"},
128 {Opt_active_logs, "active_logs=%u"},
129 {Opt_disable_ext_identify, "disable_ext_identify"},
444c580f 130 {Opt_inline_xattr, "inline_xattr"},
23cf7212 131 {Opt_noinline_xattr, "noinline_xattr"},
8274de77 132 {Opt_inline_data, "inline_data"},
5efd3c6f 133 {Opt_inline_dentry, "inline_dentry"},
97c1794a 134 {Opt_noinline_dentry, "noinline_dentry"},
6b4afdd7 135 {Opt_flush_merge, "flush_merge"},
69e9e427 136 {Opt_noflush_merge, "noflush_merge"},
0f7b2abd 137 {Opt_nobarrier, "nobarrier"},
d5053a34 138 {Opt_fastboot, "fastboot"},
89672159 139 {Opt_extent_cache, "extent_cache"},
7daaea25 140 {Opt_noextent_cache, "noextent_cache"},
75342797 141 {Opt_noinline_data, "noinline_data"},
343f40f0 142 {Opt_data_flush, "data_flush"},
36abef4e 143 {Opt_mode, "mode=%s"},
ec91538d 144 {Opt_io_size_bits, "io_bits=%u"},
73faec4d 145 {Opt_fault_injection, "fault_injection=%u"},
6d94c74a
JK
146 {Opt_lazytime, "lazytime"},
147 {Opt_nolazytime, "nolazytime"},
0abd675e
CY
148 {Opt_usrquota, "usrquota"},
149 {Opt_grpquota, "grpquota"},
5c57132e 150 {Opt_prjquota, "prjquota"},
aff063e2
JK
151 {Opt_err, NULL},
152};
153
a07ef784
NJ
154void f2fs_msg(struct super_block *sb, const char *level, const char *fmt, ...)
155{
156 struct va_format vaf;
157 va_list args;
158
159 va_start(args, fmt);
160 vaf.fmt = fmt;
161 vaf.va = &args;
162 printk("%sF2FS-fs (%s): %pV\n", level, sb->s_id, &vaf);
163 va_end(args);
164}
165
aff063e2
JK
166static void init_once(void *foo)
167{
168 struct f2fs_inode_info *fi = (struct f2fs_inode_info *) foo;
169
aff063e2
JK
170 inode_init_once(&fi->vfs_inode);
171}
172
696c018c
NJ
173static int parse_options(struct super_block *sb, char *options)
174{
175 struct f2fs_sb_info *sbi = F2FS_SB(sb);
09d54cdd 176 struct request_queue *q;
696c018c
NJ
177 substring_t args[MAX_OPT_ARGS];
178 char *p, *name;
179 int arg = 0;
180
181 if (!options)
182 return 0;
183
184 while ((p = strsep(&options, ",")) != NULL) {
185 int token;
186 if (!*p)
187 continue;
188 /*
189 * Initialize args struct so we know whether arg was
190 * found; some options take optional arguments.
191 */
192 args[0].to = args[0].from = NULL;
193 token = match_token(p, f2fs_tokens, args);
194
195 switch (token) {
196 case Opt_gc_background:
197 name = match_strdup(&args[0]);
198
199 if (!name)
200 return -ENOMEM;
6aefd93b 201 if (strlen(name) == 2 && !strncmp(name, "on", 2)) {
696c018c 202 set_opt(sbi, BG_GC);
6aefd93b
JK
203 clear_opt(sbi, FORCE_FG_GC);
204 } else if (strlen(name) == 3 && !strncmp(name, "off", 3)) {
696c018c 205 clear_opt(sbi, BG_GC);
6aefd93b
JK
206 clear_opt(sbi, FORCE_FG_GC);
207 } else if (strlen(name) == 4 && !strncmp(name, "sync", 4)) {
208 set_opt(sbi, BG_GC);
209 set_opt(sbi, FORCE_FG_GC);
210 } else {
696c018c
NJ
211 kfree(name);
212 return -EINVAL;
213 }
214 kfree(name);
215 break;
216 case Opt_disable_roll_forward:
217 set_opt(sbi, DISABLE_ROLL_FORWARD);
218 break;
2d834bf9
JK
219 case Opt_norecovery:
220 /* this option mounts f2fs with ro */
221 set_opt(sbi, DISABLE_ROLL_FORWARD);
222 if (!f2fs_readonly(sb))
223 return -EINVAL;
224 break;
696c018c 225 case Opt_discard:
09d54cdd
CY
226 q = bdev_get_queue(sb->s_bdev);
227 if (blk_queue_discard(q)) {
228 set_opt(sbi, DISCARD);
0ab02998 229 } else if (!f2fs_sb_mounted_blkzoned(sb)) {
09d54cdd
CY
230 f2fs_msg(sb, KERN_WARNING,
231 "mounting with \"discard\" option, but "
232 "the device does not support discard");
233 }
696c018c 234 break;
64058be9 235 case Opt_nodiscard:
96ba2dec
DLM
236 if (f2fs_sb_mounted_blkzoned(sb)) {
237 f2fs_msg(sb, KERN_WARNING,
238 "discard is required for zoned block devices");
239 return -EINVAL;
240 }
64058be9 241 clear_opt(sbi, DISCARD);
487df616 242 break;
696c018c
NJ
243 case Opt_noheap:
244 set_opt(sbi, NOHEAP);
245 break;
7a20b8a6
JK
246 case Opt_heap:
247 clear_opt(sbi, NOHEAP);
248 break;
696c018c 249#ifdef CONFIG_F2FS_FS_XATTR
4058c511
KA
250 case Opt_user_xattr:
251 set_opt(sbi, XATTR_USER);
252 break;
696c018c
NJ
253 case Opt_nouser_xattr:
254 clear_opt(sbi, XATTR_USER);
255 break;
444c580f
JK
256 case Opt_inline_xattr:
257 set_opt(sbi, INLINE_XATTR);
258 break;
23cf7212
CY
259 case Opt_noinline_xattr:
260 clear_opt(sbi, INLINE_XATTR);
261 break;
696c018c 262#else
4058c511
KA
263 case Opt_user_xattr:
264 f2fs_msg(sb, KERN_INFO,
265 "user_xattr options not supported");
266 break;
696c018c
NJ
267 case Opt_nouser_xattr:
268 f2fs_msg(sb, KERN_INFO,
269 "nouser_xattr options not supported");
270 break;
444c580f
JK
271 case Opt_inline_xattr:
272 f2fs_msg(sb, KERN_INFO,
273 "inline_xattr options not supported");
274 break;
23cf7212
CY
275 case Opt_noinline_xattr:
276 f2fs_msg(sb, KERN_INFO,
277 "noinline_xattr options not supported");
278 break;
696c018c
NJ
279#endif
280#ifdef CONFIG_F2FS_FS_POSIX_ACL
4058c511
KA
281 case Opt_acl:
282 set_opt(sbi, POSIX_ACL);
283 break;
696c018c
NJ
284 case Opt_noacl:
285 clear_opt(sbi, POSIX_ACL);
286 break;
287#else
4058c511
KA
288 case Opt_acl:
289 f2fs_msg(sb, KERN_INFO, "acl options not supported");
290 break;
696c018c
NJ
291 case Opt_noacl:
292 f2fs_msg(sb, KERN_INFO, "noacl options not supported");
293 break;
294#endif
295 case Opt_active_logs:
296 if (args->from && match_int(args, &arg))
297 return -EINVAL;
298 if (arg != 2 && arg != 4 && arg != NR_CURSEG_TYPE)
299 return -EINVAL;
300 sbi->active_logs = arg;
301 break;
302 case Opt_disable_ext_identify:
303 set_opt(sbi, DISABLE_EXT_IDENTIFY);
304 break;
8274de77
HL
305 case Opt_inline_data:
306 set_opt(sbi, INLINE_DATA);
307 break;
5efd3c6f
CY
308 case Opt_inline_dentry:
309 set_opt(sbi, INLINE_DENTRY);
310 break;
97c1794a
CY
311 case Opt_noinline_dentry:
312 clear_opt(sbi, INLINE_DENTRY);
313 break;
6b4afdd7
JK
314 case Opt_flush_merge:
315 set_opt(sbi, FLUSH_MERGE);
316 break;
69e9e427
JK
317 case Opt_noflush_merge:
318 clear_opt(sbi, FLUSH_MERGE);
319 break;
0f7b2abd
JK
320 case Opt_nobarrier:
321 set_opt(sbi, NOBARRIER);
322 break;
d5053a34
JK
323 case Opt_fastboot:
324 set_opt(sbi, FASTBOOT);
325 break;
89672159
CY
326 case Opt_extent_cache:
327 set_opt(sbi, EXTENT_CACHE);
328 break;
7daaea25
JK
329 case Opt_noextent_cache:
330 clear_opt(sbi, EXTENT_CACHE);
331 break;
75342797
WL
332 case Opt_noinline_data:
333 clear_opt(sbi, INLINE_DATA);
334 break;
343f40f0
CY
335 case Opt_data_flush:
336 set_opt(sbi, DATA_FLUSH);
337 break;
36abef4e
JK
338 case Opt_mode:
339 name = match_strdup(&args[0]);
340
341 if (!name)
342 return -ENOMEM;
343 if (strlen(name) == 8 &&
344 !strncmp(name, "adaptive", 8)) {
3adc57e9
DLM
345 if (f2fs_sb_mounted_blkzoned(sb)) {
346 f2fs_msg(sb, KERN_WARNING,
347 "adaptive mode is not allowed with "
348 "zoned block device feature");
349 kfree(name);
350 return -EINVAL;
351 }
52763a4b 352 set_opt_mode(sbi, F2FS_MOUNT_ADAPTIVE);
36abef4e
JK
353 } else if (strlen(name) == 3 &&
354 !strncmp(name, "lfs", 3)) {
52763a4b 355 set_opt_mode(sbi, F2FS_MOUNT_LFS);
36abef4e
JK
356 } else {
357 kfree(name);
358 return -EINVAL;
359 }
360 kfree(name);
361 break;
ec91538d
JK
362 case Opt_io_size_bits:
363 if (args->from && match_int(args, &arg))
364 return -EINVAL;
365 if (arg > __ilog2_u32(BIO_MAX_PAGES)) {
366 f2fs_msg(sb, KERN_WARNING,
367 "Not support %d, larger than %d",
368 1 << arg, BIO_MAX_PAGES);
369 return -EINVAL;
370 }
371 sbi->write_io_size_bits = arg;
372 break;
73faec4d
JK
373 case Opt_fault_injection:
374 if (args->from && match_int(args, &arg))
375 return -EINVAL;
376#ifdef CONFIG_F2FS_FAULT_INJECTION
1ecc0c5c 377 f2fs_build_fault_attr(sbi, arg);
0cc0dec2 378 set_opt(sbi, FAULT_INJECTION);
73faec4d
JK
379#else
380 f2fs_msg(sb, KERN_INFO,
381 "FAULT_INJECTION was not selected");
382#endif
383 break;
6d94c74a
JK
384 case Opt_lazytime:
385 sb->s_flags |= MS_LAZYTIME;
386 break;
387 case Opt_nolazytime:
388 sb->s_flags &= ~MS_LAZYTIME;
389 break;
0abd675e
CY
390#ifdef CONFIG_QUOTA
391 case Opt_usrquota:
392 set_opt(sbi, USRQUOTA);
393 break;
394 case Opt_grpquota:
395 set_opt(sbi, GRPQUOTA);
396 break;
5c57132e
CY
397 case Opt_prjquota:
398 set_opt(sbi, PRJQUOTA);
399 break;
0abd675e
CY
400#else
401 case Opt_usrquota:
402 case Opt_grpquota:
5c57132e 403 case Opt_prjquota:
0abd675e
CY
404 f2fs_msg(sb, KERN_INFO,
405 "quota operations not supported");
406 break;
407#endif
696c018c
NJ
408 default:
409 f2fs_msg(sb, KERN_ERR,
410 "Unrecognized mount option \"%s\" or missing value",
411 p);
412 return -EINVAL;
413 }
414 }
ec91538d
JK
415
416 if (F2FS_IO_SIZE_BITS(sbi) && !test_opt(sbi, LFS)) {
417 f2fs_msg(sb, KERN_ERR,
418 "Should set mode=lfs with %uKB-sized IO",
419 F2FS_IO_SIZE_KB(sbi));
420 return -EINVAL;
421 }
696c018c
NJ
422 return 0;
423}
424
aff063e2
JK
425static struct inode *f2fs_alloc_inode(struct super_block *sb)
426{
427 struct f2fs_inode_info *fi;
428
a0acdfe0 429 fi = kmem_cache_alloc(f2fs_inode_cachep, GFP_F2FS_ZERO);
aff063e2
JK
430 if (!fi)
431 return NULL;
432
433 init_once((void *) fi);
434
434720fa 435 /* Initialize f2fs-specific inode info */
aff063e2 436 fi->vfs_inode.i_version = 1;
204706c7 437 atomic_set(&fi->dirty_pages, 0);
aff063e2
JK
438 fi->i_current_depth = 1;
439 fi->i_advise = 0;
d928bfbf 440 init_rwsem(&fi->i_sem);
2710fd7e 441 INIT_LIST_HEAD(&fi->dirty_list);
0f18b462 442 INIT_LIST_HEAD(&fi->gdirty_list);
88b88a66
JK
443 INIT_LIST_HEAD(&fi->inmem_pages);
444 mutex_init(&fi->inmem_lock);
82e0a5aa
CY
445 init_rwsem(&fi->dio_rwsem[READ]);
446 init_rwsem(&fi->dio_rwsem[WRITE]);
5a3a2d83 447 init_rwsem(&fi->i_mmap_sem);
aff063e2 448
0abd675e
CY
449#ifdef CONFIG_QUOTA
450 memset(&fi->i_dquot, 0, sizeof(fi->i_dquot));
451 fi->i_reserved_quota = 0;
452#endif
ab9fa662
JK
453 /* Will be used by directory only */
454 fi->i_dir_level = F2FS_SB(sb)->dir_level;
f2470371 455
aff063e2
JK
456 return &fi->vfs_inode;
457}
458
531ad7d5
JK
459static int f2fs_drop_inode(struct inode *inode)
460{
b8d96a30 461 int ret;
531ad7d5
JK
462 /*
463 * This is to avoid a deadlock condition like below.
464 * writeback_single_inode(inode)
465 * - f2fs_write_data_page
466 * - f2fs_gc -> iput -> evict
467 * - inode_wait_for_writeback(inode)
468 */
0f18b462 469 if ((!inode_unhashed(inode) && inode->i_state & I_SYNC)) {
06e1bc05 470 if (!inode->i_nlink && !is_bad_inode(inode)) {
3e72f721
JK
471 /* to avoid evict_inode call simultaneously */
472 atomic_inc(&inode->i_count);
06e1bc05
JK
473 spin_unlock(&inode->i_lock);
474
475 /* some remained atomic pages should discarded */
476 if (f2fs_is_atomic_file(inode))
29b96b54 477 drop_inmem_pages(inode);
06e1bc05 478
3e72f721
JK
479 /* should remain fi->extent_tree for writepage */
480 f2fs_destroy_extent_node(inode);
481
06e1bc05 482 sb_start_intwrite(inode->i_sb);
fc9581c8 483 f2fs_i_size_write(inode, 0);
06e1bc05
JK
484
485 if (F2FS_HAS_BLOCKS(inode))
9a449e9c 486 f2fs_truncate(inode);
06e1bc05
JK
487
488 sb_end_intwrite(inode->i_sb);
489
0b81d077 490 fscrypt_put_encryption_info(inode, NULL);
06e1bc05 491 spin_lock(&inode->i_lock);
3e72f721 492 atomic_dec(&inode->i_count);
06e1bc05 493 }
b8d96a30 494 trace_f2fs_drop_inode(inode, 0);
531ad7d5 495 return 0;
06e1bc05 496 }
b8d96a30
HP
497 ret = generic_drop_inode(inode);
498 trace_f2fs_drop_inode(inode, ret);
499 return ret;
531ad7d5
JK
500}
501
7c45729a 502int f2fs_inode_dirtied(struct inode *inode, bool sync)
b3783873 503{
0f18b462 504 struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
7c45729a 505 int ret = 0;
0f18b462 506
0f18b462
JK
507 spin_lock(&sbi->inode_lock[DIRTY_META]);
508 if (is_inode_flag_set(inode, FI_DIRTY_INODE)) {
7c45729a
JK
509 ret = 1;
510 } else {
511 set_inode_flag(inode, FI_DIRTY_INODE);
512 stat_inc_dirty_inode(sbi, DIRTY_META);
0f18b462 513 }
7c45729a
JK
514 if (sync && list_empty(&F2FS_I(inode)->gdirty_list)) {
515 list_add_tail(&F2FS_I(inode)->gdirty_list,
0f18b462 516 &sbi->inode_list[DIRTY_META]);
7c45729a
JK
517 inc_page_count(sbi, F2FS_DIRTY_IMETA);
518 }
338bbfa0 519 spin_unlock(&sbi->inode_lock[DIRTY_META]);
7c45729a 520 return ret;
0f18b462
JK
521}
522
523void f2fs_inode_synced(struct inode *inode)
524{
525 struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
526
527 spin_lock(&sbi->inode_lock[DIRTY_META]);
528 if (!is_inode_flag_set(inode, FI_DIRTY_INODE)) {
529 spin_unlock(&sbi->inode_lock[DIRTY_META]);
530 return;
531 }
7c45729a
JK
532 if (!list_empty(&F2FS_I(inode)->gdirty_list)) {
533 list_del_init(&F2FS_I(inode)->gdirty_list);
534 dec_page_count(sbi, F2FS_DIRTY_IMETA);
535 }
0f18b462 536 clear_inode_flag(inode, FI_DIRTY_INODE);
26de9b11 537 clear_inode_flag(inode, FI_AUTO_RECOVER);
0f18b462 538 stat_dec_dirty_inode(F2FS_I_SB(inode), DIRTY_META);
338bbfa0 539 spin_unlock(&sbi->inode_lock[DIRTY_META]);
b3783873
JK
540}
541
b56ab837
JK
542/*
543 * f2fs_dirty_inode() is called from __mark_inode_dirty()
544 *
545 * We should call set_dirty_inode to write the dirty inode through write_inode.
546 */
547static void f2fs_dirty_inode(struct inode *inode, int flags)
548{
549 struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
550
551 if (inode->i_ino == F2FS_NODE_INO(sbi) ||
552 inode->i_ino == F2FS_META_INO(sbi))
553 return;
554
555 if (flags == I_DIRTY_TIME)
556 return;
557
558 if (is_inode_flag_set(inode, FI_AUTO_RECOVER))
559 clear_inode_flag(inode, FI_AUTO_RECOVER);
560
7c45729a 561 f2fs_inode_dirtied(inode, false);
b56ab837
JK
562}
563
aff063e2
JK
564static void f2fs_i_callback(struct rcu_head *head)
565{
566 struct inode *inode = container_of(head, struct inode, i_rcu);
567 kmem_cache_free(f2fs_inode_cachep, F2FS_I(inode));
568}
569
25ca923b 570static void f2fs_destroy_inode(struct inode *inode)
aff063e2
JK
571{
572 call_rcu(&inode->i_rcu, f2fs_i_callback);
573}
574
523be8a6
JK
575static void destroy_percpu_info(struct f2fs_sb_info *sbi)
576{
41382ec4 577 percpu_counter_destroy(&sbi->alloc_valid_block_count);
513c5f37 578 percpu_counter_destroy(&sbi->total_valid_inode_count);
523be8a6
JK
579}
580
3c62be17
JK
581static void destroy_device_list(struct f2fs_sb_info *sbi)
582{
583 int i;
584
585 for (i = 0; i < sbi->s_ndevs; i++) {
586 blkdev_put(FDEV(i).bdev, FMODE_EXCL);
587#ifdef CONFIG_BLK_DEV_ZONED
588 kfree(FDEV(i).blkz_type);
589#endif
590 }
591 kfree(sbi->devs);
592}
593
0abd675e 594static void f2fs_quota_off_umount(struct super_block *sb);
aff063e2
JK
595static void f2fs_put_super(struct super_block *sb)
596{
597 struct f2fs_sb_info *sbi = F2FS_SB(sb);
a398101a 598 int i;
aff063e2 599
0abd675e 600 f2fs_quota_off_umount(sb);
aff063e2 601
2658e50d
JK
602 /* prevent remaining shrinker jobs */
603 mutex_lock(&sbi->umount_mutex);
604
85dc2f2c
JK
605 /*
606 * We don't need to do checkpoint when superblock is clean.
607 * But, the previous checkpoint was not done by umount, it needs to do
608 * clean checkpoint again.
609 */
caf0047e 610 if (is_sbi_flag_set(sbi, SBI_IS_DIRTY) ||
aaec2b1d 611 !is_set_ckpt_flags(sbi, CP_UMOUNT_FLAG)) {
75ab4cb8
JK
612 struct cp_control cpc = {
613 .reason = CP_UMOUNT,
614 };
615 write_checkpoint(sbi, &cpc);
616 }
aff063e2 617
4e6a8d9b 618 /* be sure to wait for any on-going discard commands */
d431413f 619 f2fs_wait_discard_bios(sbi);
4e6a8d9b 620
663f387b 621 if (f2fs_discard_en(sbi) && !sbi->discard_blks) {
1f43e2ad
CY
622 struct cp_control cpc = {
623 .reason = CP_UMOUNT | CP_TRIMMED,
624 };
625 write_checkpoint(sbi, &cpc);
626 }
627
eca616f8
JK
628 /* write_checkpoint can update stat informaion */
629 f2fs_destroy_stats(sbi);
630
cf779cab
JK
631 /*
632 * normally superblock is clean, so we need to release this.
633 * In addition, EIO will skip do checkpoint, we need this as well.
634 */
74ef9241 635 release_ino_entry(sbi, true);
6f12ac25 636
2658e50d
JK
637 f2fs_leave_shrinker(sbi);
638 mutex_unlock(&sbi->umount_mutex);
639
17c19120 640 /* our cp_error case, we can wait for any writeback page */
b9109b0e 641 f2fs_flush_merged_writes(sbi);
17c19120 642
aff063e2
JK
643 iput(sbi->node_inode);
644 iput(sbi->meta_inode);
645
646 /* destroy f2fs internal modules */
647 destroy_node_manager(sbi);
648 destroy_segment_manager(sbi);
649
650 kfree(sbi->ckpt);
a398101a 651
dc6b2055 652 f2fs_unregister_sysfs(sbi);
aff063e2
JK
653
654 sb->s_fs_info = NULL;
43b6573b
KM
655 if (sbi->s_chksum_driver)
656 crypto_free_shash(sbi->s_chksum_driver);
b39f0de2 657 kfree(sbi->raw_super);
523be8a6 658
3c62be17 659 destroy_device_list(sbi);
b6895e8f 660 mempool_destroy(sbi->write_io_dummy);
523be8a6 661 destroy_percpu_info(sbi);
a912b54d
JK
662 for (i = 0; i < NR_PAGE_TYPE; i++)
663 kfree(sbi->write_io[i]);
aff063e2
JK
664 kfree(sbi);
665}
666
667int f2fs_sync_fs(struct super_block *sb, int sync)
668{
669 struct f2fs_sb_info *sbi = F2FS_SB(sb);
c34f42e2 670 int err = 0;
aff063e2 671
a2a4a7e4
NJ
672 trace_f2fs_sync_fs(sb, sync);
673
b7473754 674 if (sync) {
d5053a34
JK
675 struct cp_control cpc;
676
119ee914
JK
677 cpc.reason = __get_cp_reason(sbi);
678
b7473754 679 mutex_lock(&sbi->gc_mutex);
c34f42e2 680 err = write_checkpoint(sbi, &cpc);
b7473754 681 mutex_unlock(&sbi->gc_mutex);
b7473754 682 }
05ca3632 683 f2fs_trace_ios(NULL, 1);
aff063e2 684
c34f42e2 685 return err;
aff063e2
JK
686}
687
d6212a5f
CL
688static int f2fs_freeze(struct super_block *sb)
689{
77888c1e 690 if (f2fs_readonly(sb))
d6212a5f
CL
691 return 0;
692
b4b9d34c
JK
693 /* IO error happened before */
694 if (unlikely(f2fs_cp_error(F2FS_SB(sb))))
695 return -EIO;
696
697 /* must be clean, since sync_filesystem() was already called */
698 if (is_sbi_flag_set(F2FS_SB(sb), SBI_IS_DIRTY))
699 return -EINVAL;
700 return 0;
d6212a5f
CL
701}
702
703static int f2fs_unfreeze(struct super_block *sb)
704{
705 return 0;
706}
707
ddc34e32
CY
708#ifdef CONFIG_QUOTA
709static int f2fs_statfs_project(struct super_block *sb,
710 kprojid_t projid, struct kstatfs *buf)
711{
712 struct kqid qid;
713 struct dquot *dquot;
714 u64 limit;
715 u64 curblock;
716
717 qid = make_kqid_projid(projid);
718 dquot = dqget(sb, qid);
719 if (IS_ERR(dquot))
720 return PTR_ERR(dquot);
721 spin_lock(&dq_data_lock);
722
723 limit = (dquot->dq_dqb.dqb_bsoftlimit ?
724 dquot->dq_dqb.dqb_bsoftlimit :
725 dquot->dq_dqb.dqb_bhardlimit) >> sb->s_blocksize_bits;
726 if (limit && buf->f_blocks > limit) {
727 curblock = dquot->dq_dqb.dqb_curspace >> sb->s_blocksize_bits;
728 buf->f_blocks = limit;
729 buf->f_bfree = buf->f_bavail =
730 (buf->f_blocks > curblock) ?
731 (buf->f_blocks - curblock) : 0;
732 }
733
734 limit = dquot->dq_dqb.dqb_isoftlimit ?
735 dquot->dq_dqb.dqb_isoftlimit :
736 dquot->dq_dqb.dqb_ihardlimit;
737 if (limit && buf->f_files > limit) {
738 buf->f_files = limit;
739 buf->f_ffree =
740 (buf->f_files > dquot->dq_dqb.dqb_curinodes) ?
741 (buf->f_files - dquot->dq_dqb.dqb_curinodes) : 0;
742 }
743
744 spin_unlock(&dq_data_lock);
745 dqput(dquot);
746 return 0;
747}
748#endif
749
aff063e2
JK
750static int f2fs_statfs(struct dentry *dentry, struct kstatfs *buf)
751{
752 struct super_block *sb = dentry->d_sb;
753 struct f2fs_sb_info *sbi = F2FS_SB(sb);
754 u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
755 block_t total_count, user_block_count, start_count, ovp_count;
0cc091d0 756 u64 avail_node_count;
aff063e2
JK
757
758 total_count = le64_to_cpu(sbi->raw_super->block_count);
759 user_block_count = sbi->user_block_count;
760 start_count = le32_to_cpu(sbi->raw_super->segment0_blkaddr);
761 ovp_count = SM_I(sbi)->ovp_segments << sbi->log_blocks_per_seg;
762 buf->f_type = F2FS_SUPER_MAGIC;
763 buf->f_bsize = sbi->blocksize;
764
765 buf->f_blocks = total_count - start_count;
3e6d0b4d 766 buf->f_bfree = user_block_count - valid_user_blocks(sbi) + ovp_count;
daeb433e
CY
767 buf->f_bavail = user_block_count - valid_user_blocks(sbi) -
768 sbi->reserved_blocks;
aff063e2 769
0cc091d0
JK
770 avail_node_count = sbi->total_node_count - F2FS_RESERVED_NODE_NUM;
771
772 if (avail_node_count > user_block_count) {
773 buf->f_files = user_block_count;
774 buf->f_ffree = buf->f_bavail;
775 } else {
776 buf->f_files = avail_node_count;
777 buf->f_ffree = min(avail_node_count - valid_node_count(sbi),
778 buf->f_bavail);
779 }
aff063e2 780
5a20d339 781 buf->f_namelen = F2FS_NAME_LEN;
aff063e2
JK
782 buf->f_fsid.val[0] = (u32)id;
783 buf->f_fsid.val[1] = (u32)(id >> 32);
784
ddc34e32
CY
785#ifdef CONFIG_QUOTA
786 if (is_inode_flag_set(dentry->d_inode, FI_PROJ_INHERIT) &&
787 sb_has_quota_limits_enabled(sb, PRJQUOTA)) {
788 f2fs_statfs_project(sb, F2FS_I(dentry->d_inode)->i_projid, buf);
789 }
790#endif
aff063e2
JK
791 return 0;
792}
793
794static int f2fs_show_options(struct seq_file *seq, struct dentry *root)
795{
796 struct f2fs_sb_info *sbi = F2FS_SB(root->d_sb);
797
6aefd93b
JK
798 if (!f2fs_readonly(sbi->sb) && test_opt(sbi, BG_GC)) {
799 if (test_opt(sbi, FORCE_FG_GC))
800 seq_printf(seq, ",background_gc=%s", "sync");
801 else
802 seq_printf(seq, ",background_gc=%s", "on");
803 } else {
696c018c 804 seq_printf(seq, ",background_gc=%s", "off");
6aefd93b 805 }
aff063e2
JK
806 if (test_opt(sbi, DISABLE_ROLL_FORWARD))
807 seq_puts(seq, ",disable_roll_forward");
808 if (test_opt(sbi, DISCARD))
809 seq_puts(seq, ",discard");
810 if (test_opt(sbi, NOHEAP))
7a20b8a6
JK
811 seq_puts(seq, ",no_heap");
812 else
813 seq_puts(seq, ",heap");
aff063e2
JK
814#ifdef CONFIG_F2FS_FS_XATTR
815 if (test_opt(sbi, XATTR_USER))
816 seq_puts(seq, ",user_xattr");
817 else
818 seq_puts(seq, ",nouser_xattr");
444c580f
JK
819 if (test_opt(sbi, INLINE_XATTR))
820 seq_puts(seq, ",inline_xattr");
23cf7212
CY
821 else
822 seq_puts(seq, ",noinline_xattr");
aff063e2
JK
823#endif
824#ifdef CONFIG_F2FS_FS_POSIX_ACL
825 if (test_opt(sbi, POSIX_ACL))
826 seq_puts(seq, ",acl");
827 else
828 seq_puts(seq, ",noacl");
829#endif
830 if (test_opt(sbi, DISABLE_EXT_IDENTIFY))
aa43507f 831 seq_puts(seq, ",disable_ext_identify");
8274de77
HL
832 if (test_opt(sbi, INLINE_DATA))
833 seq_puts(seq, ",inline_data");
75342797
WL
834 else
835 seq_puts(seq, ",noinline_data");
5efd3c6f
CY
836 if (test_opt(sbi, INLINE_DENTRY))
837 seq_puts(seq, ",inline_dentry");
97c1794a
CY
838 else
839 seq_puts(seq, ",noinline_dentry");
b270ad6f 840 if (!f2fs_readonly(sbi->sb) && test_opt(sbi, FLUSH_MERGE))
6b4afdd7 841 seq_puts(seq, ",flush_merge");
0f7b2abd
JK
842 if (test_opt(sbi, NOBARRIER))
843 seq_puts(seq, ",nobarrier");
d5053a34
JK
844 if (test_opt(sbi, FASTBOOT))
845 seq_puts(seq, ",fastboot");
89672159
CY
846 if (test_opt(sbi, EXTENT_CACHE))
847 seq_puts(seq, ",extent_cache");
7daaea25
JK
848 else
849 seq_puts(seq, ",noextent_cache");
343f40f0
CY
850 if (test_opt(sbi, DATA_FLUSH))
851 seq_puts(seq, ",data_flush");
36abef4e
JK
852
853 seq_puts(seq, ",mode=");
854 if (test_opt(sbi, ADAPTIVE))
855 seq_puts(seq, "adaptive");
856 else if (test_opt(sbi, LFS))
857 seq_puts(seq, "lfs");
aff063e2 858 seq_printf(seq, ",active_logs=%u", sbi->active_logs);
ec91538d
JK
859 if (F2FS_IO_SIZE_BITS(sbi))
860 seq_printf(seq, ",io_size=%uKB", F2FS_IO_SIZE_KB(sbi));
0cc0dec2
KX
861#ifdef CONFIG_F2FS_FAULT_INJECTION
862 if (test_opt(sbi, FAULT_INJECTION))
44529f89
CY
863 seq_printf(seq, ",fault_injection=%u",
864 sbi->fault_info.inject_rate);
0cc0dec2 865#endif
0abd675e
CY
866#ifdef CONFIG_QUOTA
867 if (test_opt(sbi, USRQUOTA))
868 seq_puts(seq, ",usrquota");
869 if (test_opt(sbi, GRPQUOTA))
870 seq_puts(seq, ",grpquota");
5c57132e
CY
871 if (test_opt(sbi, PRJQUOTA))
872 seq_puts(seq, ",prjquota");
0cc0dec2 873#endif
aff063e2
JK
874
875 return 0;
876}
877
498c5e9f
YH
878static void default_options(struct f2fs_sb_info *sbi)
879{
880 /* init some FS parameters */
881 sbi->active_logs = NR_CURSEG_TYPE;
882
883 set_opt(sbi, BG_GC);
39133a50 884 set_opt(sbi, INLINE_XATTR);
498c5e9f 885 set_opt(sbi, INLINE_DATA);
97c1794a 886 set_opt(sbi, INLINE_DENTRY);
3e72f721 887 set_opt(sbi, EXTENT_CACHE);
7a20b8a6 888 set_opt(sbi, NOHEAP);
6d94c74a 889 sbi->sb->s_flags |= MS_LAZYTIME;
69e9e427 890 set_opt(sbi, FLUSH_MERGE);
0bfd7a09 891 if (f2fs_sb_mounted_blkzoned(sbi->sb)) {
52763a4b
JK
892 set_opt_mode(sbi, F2FS_MOUNT_LFS);
893 set_opt(sbi, DISCARD);
894 } else {
895 set_opt_mode(sbi, F2FS_MOUNT_ADAPTIVE);
896 }
498c5e9f
YH
897
898#ifdef CONFIG_F2FS_FS_XATTR
899 set_opt(sbi, XATTR_USER);
900#endif
901#ifdef CONFIG_F2FS_FS_POSIX_ACL
902 set_opt(sbi, POSIX_ACL);
903#endif
36dbd328
CY
904
905#ifdef CONFIG_F2FS_FAULT_INJECTION
906 f2fs_build_fault_attr(sbi, 0);
907#endif
498c5e9f
YH
908}
909
696c018c
NJ
910static int f2fs_remount(struct super_block *sb, int *flags, char *data)
911{
912 struct f2fs_sb_info *sbi = F2FS_SB(sb);
913 struct f2fs_mount_info org_mount_opt;
0abd675e 914 unsigned long old_sb_flags;
696c018c 915 int err, active_logs;
876dc59e
GZ
916 bool need_restart_gc = false;
917 bool need_stop_gc = false;
9cd81ce3 918 bool no_extent_cache = !test_opt(sbi, EXTENT_CACHE);
2443b8b3
CY
919#ifdef CONFIG_F2FS_FAULT_INJECTION
920 struct f2fs_fault_info ffi = sbi->fault_info;
921#endif
696c018c
NJ
922
923 /*
924 * Save the old mount options in case we
925 * need to restore them.
926 */
927 org_mount_opt = sbi->mount_opt;
0abd675e 928 old_sb_flags = sb->s_flags;
696c018c
NJ
929 active_logs = sbi->active_logs;
930
df728b0f
JK
931 /* recover superblocks we couldn't write due to previous RO mount */
932 if (!(*flags & MS_RDONLY) && is_sbi_flag_set(sbi, SBI_NEED_SB_WRITE)) {
933 err = f2fs_commit_super(sbi, false);
934 f2fs_msg(sb, KERN_INFO,
935 "Try to recover all the superblocks, ret: %d", err);
936 if (!err)
937 clear_sbi_flag(sbi, SBI_NEED_SB_WRITE);
938 }
939
498c5e9f 940 default_options(sbi);
26666c8a 941
696c018c
NJ
942 /* parse mount options */
943 err = parse_options(sb, data);
944 if (err)
945 goto restore_opts;
946
947 /*
948 * Previous and new state of filesystem is RO,
876dc59e 949 * so skip checking GC and FLUSH_MERGE conditions.
696c018c 950 */
6b2920a5 951 if (f2fs_readonly(sb) && (*flags & MS_RDONLY))
696c018c
NJ
952 goto skip;
953
0abd675e
CY
954 if (!f2fs_readonly(sb) && (*flags & MS_RDONLY)) {
955 err = dquot_suspend(sb, -1);
956 if (err < 0)
957 goto restore_opts;
958 } else {
959 /* dquot_resume needs RW */
960 sb->s_flags &= ~MS_RDONLY;
961 dquot_resume(sb, -1);
962 }
963
9cd81ce3
CY
964 /* disallow enable/disable extent_cache dynamically */
965 if (no_extent_cache == !!test_opt(sbi, EXTENT_CACHE)) {
966 err = -EINVAL;
967 f2fs_msg(sbi->sb, KERN_WARNING,
968 "switch extent_cache option is not allowed");
969 goto restore_opts;
970 }
971
696c018c
NJ
972 /*
973 * We stop the GC thread if FS is mounted as RO
974 * or if background_gc = off is passed in mount
975 * option. Also sync the filesystem.
976 */
977 if ((*flags & MS_RDONLY) || !test_opt(sbi, BG_GC)) {
978 if (sbi->gc_thread) {
979 stop_gc_thread(sbi);
876dc59e 980 need_restart_gc = true;
696c018c 981 }
aba291b3 982 } else if (!sbi->gc_thread) {
696c018c
NJ
983 err = start_gc_thread(sbi);
984 if (err)
985 goto restore_opts;
876dc59e
GZ
986 need_stop_gc = true;
987 }
988
faa0e55b
JK
989 if (*flags & MS_RDONLY) {
990 writeback_inodes_sb(sb, WB_REASON_SYNC);
991 sync_inodes_sb(sb);
992
993 set_sbi_flag(sbi, SBI_IS_DIRTY);
994 set_sbi_flag(sbi, SBI_IS_CLOSE);
995 f2fs_sync_fs(sb, 1);
996 clear_sbi_flag(sbi, SBI_IS_CLOSE);
997 }
998
876dc59e
GZ
999 /*
1000 * We stop issue flush thread if FS is mounted as RO
1001 * or if flush_merge is not passed in mount option.
1002 */
1003 if ((*flags & MS_RDONLY) || !test_opt(sbi, FLUSH_MERGE)) {
5eba8c5d
JK
1004 clear_opt(sbi, FLUSH_MERGE);
1005 destroy_flush_cmd_control(sbi, false);
1006 } else {
2163d198
GZ
1007 err = create_flush_cmd_control(sbi);
1008 if (err)
a688b9d9 1009 goto restore_gc;
696c018c
NJ
1010 }
1011skip:
1012 /* Update the POSIXACL Flag */
df728b0f 1013 sb->s_flags = (sb->s_flags & ~MS_POSIXACL) |
696c018c 1014 (test_opt(sbi, POSIX_ACL) ? MS_POSIXACL : 0);
df728b0f 1015
696c018c 1016 return 0;
876dc59e
GZ
1017restore_gc:
1018 if (need_restart_gc) {
1019 if (start_gc_thread(sbi))
1020 f2fs_msg(sbi->sb, KERN_WARNING,
e1c42045 1021 "background gc thread has stopped");
876dc59e
GZ
1022 } else if (need_stop_gc) {
1023 stop_gc_thread(sbi);
1024 }
696c018c
NJ
1025restore_opts:
1026 sbi->mount_opt = org_mount_opt;
1027 sbi->active_logs = active_logs;
0abd675e 1028 sb->s_flags = old_sb_flags;
2443b8b3
CY
1029#ifdef CONFIG_F2FS_FAULT_INJECTION
1030 sbi->fault_info = ffi;
1031#endif
696c018c
NJ
1032 return err;
1033}
1034
0abd675e
CY
1035#ifdef CONFIG_QUOTA
1036/* Read data from quotafile */
1037static ssize_t f2fs_quota_read(struct super_block *sb, int type, char *data,
1038 size_t len, loff_t off)
1039{
1040 struct inode *inode = sb_dqopt(sb)->files[type];
1041 struct address_space *mapping = inode->i_mapping;
1042 block_t blkidx = F2FS_BYTES_TO_BLK(off);
1043 int offset = off & (sb->s_blocksize - 1);
1044 int tocopy;
1045 size_t toread;
1046 loff_t i_size = i_size_read(inode);
1047 struct page *page;
1048 char *kaddr;
1049
1050 if (off > i_size)
1051 return 0;
1052
1053 if (off + len > i_size)
1054 len = i_size - off;
1055 toread = len;
1056 while (toread > 0) {
1057 tocopy = min_t(unsigned long, sb->s_blocksize - offset, toread);
1058repeat:
1059 page = read_mapping_page(mapping, blkidx, NULL);
1060 if (IS_ERR(page))
1061 return PTR_ERR(page);
1062
1063 lock_page(page);
1064
1065 if (unlikely(page->mapping != mapping)) {
1066 f2fs_put_page(page, 1);
1067 goto repeat;
1068 }
1069 if (unlikely(!PageUptodate(page))) {
1070 f2fs_put_page(page, 1);
1071 return -EIO;
1072 }
1073
1074 kaddr = kmap_atomic(page);
1075 memcpy(data, kaddr + offset, tocopy);
1076 kunmap_atomic(kaddr);
1077 f2fs_put_page(page, 1);
1078
1079 offset = 0;
1080 toread -= tocopy;
1081 data += tocopy;
1082 blkidx++;
1083 }
1084 return len;
1085}
1086
1087/* Write to quotafile */
1088static ssize_t f2fs_quota_write(struct super_block *sb, int type,
1089 const char *data, size_t len, loff_t off)
1090{
1091 struct inode *inode = sb_dqopt(sb)->files[type];
1092 struct address_space *mapping = inode->i_mapping;
1093 const struct address_space_operations *a_ops = mapping->a_ops;
1094 int offset = off & (sb->s_blocksize - 1);
1095 size_t towrite = len;
1096 struct page *page;
1097 char *kaddr;
1098 int err = 0;
1099 int tocopy;
1100
1101 while (towrite > 0) {
1102 tocopy = min_t(unsigned long, sb->s_blocksize - offset,
1103 towrite);
1104
1105 err = a_ops->write_begin(NULL, mapping, off, tocopy, 0,
1106 &page, NULL);
1107 if (unlikely(err))
1108 break;
1109
1110 kaddr = kmap_atomic(page);
1111 memcpy(kaddr + offset, data, tocopy);
1112 kunmap_atomic(kaddr);
1113 flush_dcache_page(page);
1114
1115 a_ops->write_end(NULL, mapping, off, tocopy, tocopy,
1116 page, NULL);
1117 offset = 0;
1118 towrite -= tocopy;
1119 off += tocopy;
1120 data += tocopy;
1121 cond_resched();
1122 }
1123
1124 if (len == towrite)
4f31d26b 1125 return 0;
0abd675e
CY
1126 inode->i_version++;
1127 inode->i_mtime = inode->i_ctime = current_time(inode);
1128 f2fs_mark_inode_dirty_sync(inode, false);
1129 return len - towrite;
1130}
1131
1132static struct dquot **f2fs_get_dquots(struct inode *inode)
1133{
1134 return F2FS_I(inode)->i_dquot;
1135}
1136
1137static qsize_t *f2fs_get_reserved_space(struct inode *inode)
1138{
1139 return &F2FS_I(inode)->i_reserved_quota;
1140}
1141
1142static int f2fs_quota_sync(struct super_block *sb, int type)
1143{
1144 struct quota_info *dqopt = sb_dqopt(sb);
1145 int cnt;
1146 int ret;
1147
1148 ret = dquot_writeback_dquots(sb, type);
1149 if (ret)
1150 return ret;
1151
1152 /*
1153 * Now when everything is written we can discard the pagecache so
1154 * that userspace sees the changes.
1155 */
1156 for (cnt = 0; cnt < MAXQUOTAS; cnt++) {
1157 if (type != -1 && cnt != type)
1158 continue;
1159 if (!sb_has_quota_active(sb, cnt))
1160 continue;
1161
1162 ret = filemap_write_and_wait(dqopt->files[cnt]->i_mapping);
1163 if (ret)
1164 return ret;
1165
1166 inode_lock(dqopt->files[cnt]);
1167 truncate_inode_pages(&dqopt->files[cnt]->i_data, 0);
1168 inode_unlock(dqopt->files[cnt]);
1169 }
1170 return 0;
1171}
1172
1173static int f2fs_quota_on(struct super_block *sb, int type, int format_id,
1174 const struct path *path)
1175{
1176 struct inode *inode;
1177 int err;
1178
1179 err = f2fs_quota_sync(sb, -1);
1180 if (err)
1181 return err;
1182
1183 err = dquot_quota_on(sb, type, format_id, path);
1184 if (err)
1185 return err;
1186
1187 inode = d_inode(path->dentry);
1188
1189 inode_lock(inode);
1190 F2FS_I(inode)->i_flags |= FS_NOATIME_FL | FS_IMMUTABLE_FL;
1191 inode_set_flags(inode, S_NOATIME | S_IMMUTABLE,
1192 S_NOATIME | S_IMMUTABLE);
1193 inode_unlock(inode);
1194 f2fs_mark_inode_dirty_sync(inode, false);
1195
1196 return 0;
1197}
1198
1199static int f2fs_quota_off(struct super_block *sb, int type)
1200{
1201 struct inode *inode = sb_dqopt(sb)->files[type];
1202 int err;
1203
1204 if (!inode || !igrab(inode))
1205 return dquot_quota_off(sb, type);
1206
1207 f2fs_quota_sync(sb, -1);
1208
1209 err = dquot_quota_off(sb, type);
1210 if (err)
1211 goto out_put;
1212
1213 inode_lock(inode);
1214 F2FS_I(inode)->i_flags &= ~(FS_NOATIME_FL | FS_IMMUTABLE_FL);
1215 inode_set_flags(inode, 0, S_NOATIME | S_IMMUTABLE);
1216 inode_unlock(inode);
1217 f2fs_mark_inode_dirty_sync(inode, false);
1218out_put:
1219 iput(inode);
1220 return err;
1221}
1222
1223static void f2fs_quota_off_umount(struct super_block *sb)
1224{
1225 int type;
1226
1227 for (type = 0; type < MAXQUOTAS; type++)
1228 f2fs_quota_off(sb, type);
1229}
1230
5c57132e
CY
1231int f2fs_get_projid(struct inode *inode, kprojid_t *projid)
1232{
1233 *projid = F2FS_I(inode)->i_projid;
1234 return 0;
1235}
1236
0abd675e
CY
1237static const struct dquot_operations f2fs_quota_operations = {
1238 .get_reserved_space = f2fs_get_reserved_space,
1239 .write_dquot = dquot_commit,
1240 .acquire_dquot = dquot_acquire,
1241 .release_dquot = dquot_release,
1242 .mark_dirty = dquot_mark_dquot_dirty,
1243 .write_info = dquot_commit_info,
1244 .alloc_dquot = dquot_alloc,
1245 .destroy_dquot = dquot_destroy,
5c57132e 1246 .get_projid = f2fs_get_projid,
0abd675e
CY
1247 .get_next_id = dquot_get_next_id,
1248};
1249
1250static const struct quotactl_ops f2fs_quotactl_ops = {
1251 .quota_on = f2fs_quota_on,
1252 .quota_off = f2fs_quota_off,
1253 .quota_sync = f2fs_quota_sync,
1254 .get_state = dquot_get_state,
1255 .set_info = dquot_set_dqinfo,
1256 .get_dqblk = dquot_get_dqblk,
1257 .set_dqblk = dquot_set_dqblk,
1258 .get_nextdqblk = dquot_get_next_dqblk,
1259};
1260#else
1261static inline void f2fs_quota_off_umount(struct super_block *sb)
1262{
1263}
1264#endif
1265
aff063e2
JK
1266static struct super_operations f2fs_sops = {
1267 .alloc_inode = f2fs_alloc_inode,
531ad7d5 1268 .drop_inode = f2fs_drop_inode,
aff063e2
JK
1269 .destroy_inode = f2fs_destroy_inode,
1270 .write_inode = f2fs_write_inode,
b3783873 1271 .dirty_inode = f2fs_dirty_inode,
aff063e2 1272 .show_options = f2fs_show_options,
0abd675e
CY
1273#ifdef CONFIG_QUOTA
1274 .quota_read = f2fs_quota_read,
1275 .quota_write = f2fs_quota_write,
1276 .get_dquots = f2fs_get_dquots,
1277#endif
aff063e2
JK
1278 .evict_inode = f2fs_evict_inode,
1279 .put_super = f2fs_put_super,
1280 .sync_fs = f2fs_sync_fs,
d6212a5f
CL
1281 .freeze_fs = f2fs_freeze,
1282 .unfreeze_fs = f2fs_unfreeze,
aff063e2 1283 .statfs = f2fs_statfs,
696c018c 1284 .remount_fs = f2fs_remount,
aff063e2
JK
1285};
1286
0b81d077
JK
1287#ifdef CONFIG_F2FS_FS_ENCRYPTION
1288static int f2fs_get_context(struct inode *inode, void *ctx, size_t len)
1289{
1290 return f2fs_getxattr(inode, F2FS_XATTR_INDEX_ENCRYPTION,
1291 F2FS_XATTR_NAME_ENCRYPTION_CONTEXT,
1292 ctx, len, NULL);
1293}
1294
1295static int f2fs_set_context(struct inode *inode, const void *ctx, size_t len,
1296 void *fs_data)
1297{
1298 return f2fs_setxattr(inode, F2FS_XATTR_INDEX_ENCRYPTION,
1299 F2FS_XATTR_NAME_ENCRYPTION_CONTEXT,
1300 ctx, len, fs_data, XATTR_CREATE);
1301}
1302
1303static unsigned f2fs_max_namelen(struct inode *inode)
1304{
1305 return S_ISLNK(inode->i_mode) ?
1306 inode->i_sb->s_blocksize : F2FS_NAME_LEN;
1307}
1308
6f69f0ed 1309static const struct fscrypt_operations f2fs_cryptops = {
a5d431ef 1310 .key_prefix = "f2fs:",
0b81d077
JK
1311 .get_context = f2fs_get_context,
1312 .set_context = f2fs_set_context,
1313 .is_encrypted = f2fs_encrypted_inode,
1314 .empty_dir = f2fs_empty_dir,
1315 .max_namelen = f2fs_max_namelen,
1316};
1317#else
6f69f0ed 1318static const struct fscrypt_operations f2fs_cryptops = {
0b81d077
JK
1319 .is_encrypted = f2fs_encrypted_inode,
1320};
1321#endif
1322
aff063e2
JK
1323static struct inode *f2fs_nfs_get_inode(struct super_block *sb,
1324 u64 ino, u32 generation)
1325{
1326 struct f2fs_sb_info *sbi = F2FS_SB(sb);
1327 struct inode *inode;
1328
d6b7d4b3 1329 if (check_nid_range(sbi, ino))
910bb12d 1330 return ERR_PTR(-ESTALE);
aff063e2
JK
1331
1332 /*
1333 * f2fs_iget isn't quite right if the inode is currently unallocated!
1334 * However f2fs_iget currently does appropriate checks to handle stale
1335 * inodes so everything is OK.
1336 */
1337 inode = f2fs_iget(sb, ino);
1338 if (IS_ERR(inode))
1339 return ERR_CAST(inode);
6bacf52f 1340 if (unlikely(generation && inode->i_generation != generation)) {
aff063e2
JK
1341 /* we didn't find the right inode.. */
1342 iput(inode);
1343 return ERR_PTR(-ESTALE);
1344 }
1345 return inode;
1346}
1347
1348static struct dentry *f2fs_fh_to_dentry(struct super_block *sb, struct fid *fid,
1349 int fh_len, int fh_type)
1350{
1351 return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
1352 f2fs_nfs_get_inode);
1353}
1354
1355static struct dentry *f2fs_fh_to_parent(struct super_block *sb, struct fid *fid,
1356 int fh_len, int fh_type)
1357{
1358 return generic_fh_to_parent(sb, fid, fh_len, fh_type,
1359 f2fs_nfs_get_inode);
1360}
1361
1362static const struct export_operations f2fs_export_ops = {
1363 .fh_to_dentry = f2fs_fh_to_dentry,
1364 .fh_to_parent = f2fs_fh_to_parent,
1365 .get_parent = f2fs_get_parent,
1366};
1367
e0afc4d6 1368static loff_t max_file_blocks(void)
aff063e2 1369{
7a2af766 1370 loff_t result = 0;
aff063e2
JK
1371 loff_t leaf_count = ADDRS_PER_BLOCK;
1372
7a2af766
CY
1373 /*
1374 * note: previously, result is equal to (DEF_ADDRS_PER_INODE -
1375 * F2FS_INLINE_XATTR_ADDRS), but now f2fs try to reserve more
1376 * space in inode.i_addr, it will be more safe to reassign
1377 * result as zero.
1378 */
1379
aff063e2
JK
1380 /* two direct node blocks */
1381 result += (leaf_count * 2);
1382
1383 /* two indirect node blocks */
1384 leaf_count *= NIDS_PER_BLOCK;
1385 result += (leaf_count * 2);
1386
1387 /* one double indirect node block */
1388 leaf_count *= NIDS_PER_BLOCK;
1389 result += leaf_count;
1390
aff063e2
JK
1391 return result;
1392}
1393
fd694733
JK
1394static int __f2fs_commit_super(struct buffer_head *bh,
1395 struct f2fs_super_block *super)
1396{
1397 lock_buffer(bh);
1398 if (super)
1399 memcpy(bh->b_data + F2FS_SUPER_OFFSET, super, sizeof(*super));
1400 set_buffer_uptodate(bh);
1401 set_buffer_dirty(bh);
1402 unlock_buffer(bh);
1403
1404 /* it's rare case, we can do fua all the time */
3adc5fcb 1405 return __sync_dirty_buffer(bh, REQ_SYNC | REQ_PREFLUSH | REQ_FUA);
fd694733
JK
1406}
1407
df728b0f 1408static inline bool sanity_check_area_boundary(struct f2fs_sb_info *sbi,
fd694733 1409 struct buffer_head *bh)
9a59b62f 1410{
fd694733
JK
1411 struct f2fs_super_block *raw_super = (struct f2fs_super_block *)
1412 (bh->b_data + F2FS_SUPER_OFFSET);
df728b0f 1413 struct super_block *sb = sbi->sb;
9a59b62f
CY
1414 u32 segment0_blkaddr = le32_to_cpu(raw_super->segment0_blkaddr);
1415 u32 cp_blkaddr = le32_to_cpu(raw_super->cp_blkaddr);
1416 u32 sit_blkaddr = le32_to_cpu(raw_super->sit_blkaddr);
1417 u32 nat_blkaddr = le32_to_cpu(raw_super->nat_blkaddr);
1418 u32 ssa_blkaddr = le32_to_cpu(raw_super->ssa_blkaddr);
1419 u32 main_blkaddr = le32_to_cpu(raw_super->main_blkaddr);
1420 u32 segment_count_ckpt = le32_to_cpu(raw_super->segment_count_ckpt);
1421 u32 segment_count_sit = le32_to_cpu(raw_super->segment_count_sit);
1422 u32 segment_count_nat = le32_to_cpu(raw_super->segment_count_nat);
1423 u32 segment_count_ssa = le32_to_cpu(raw_super->segment_count_ssa);
1424 u32 segment_count_main = le32_to_cpu(raw_super->segment_count_main);
1425 u32 segment_count = le32_to_cpu(raw_super->segment_count);
1426 u32 log_blocks_per_seg = le32_to_cpu(raw_super->log_blocks_per_seg);
fd694733
JK
1427 u64 main_end_blkaddr = main_blkaddr +
1428 (segment_count_main << log_blocks_per_seg);
1429 u64 seg_end_blkaddr = segment0_blkaddr +
1430 (segment_count << log_blocks_per_seg);
9a59b62f
CY
1431
1432 if (segment0_blkaddr != cp_blkaddr) {
1433 f2fs_msg(sb, KERN_INFO,
1434 "Mismatch start address, segment0(%u) cp_blkaddr(%u)",
1435 segment0_blkaddr, cp_blkaddr);
1436 return true;
1437 }
1438
1439 if (cp_blkaddr + (segment_count_ckpt << log_blocks_per_seg) !=
1440 sit_blkaddr) {
1441 f2fs_msg(sb, KERN_INFO,
1442 "Wrong CP boundary, start(%u) end(%u) blocks(%u)",
1443 cp_blkaddr, sit_blkaddr,
1444 segment_count_ckpt << log_blocks_per_seg);
1445 return true;
1446 }
1447
1448 if (sit_blkaddr + (segment_count_sit << log_blocks_per_seg) !=
1449 nat_blkaddr) {
1450 f2fs_msg(sb, KERN_INFO,
1451 "Wrong SIT boundary, start(%u) end(%u) blocks(%u)",
1452 sit_blkaddr, nat_blkaddr,
1453 segment_count_sit << log_blocks_per_seg);
1454 return true;
1455 }
1456
1457 if (nat_blkaddr + (segment_count_nat << log_blocks_per_seg) !=
1458 ssa_blkaddr) {
1459 f2fs_msg(sb, KERN_INFO,
1460 "Wrong NAT boundary, start(%u) end(%u) blocks(%u)",
1461 nat_blkaddr, ssa_blkaddr,
1462 segment_count_nat << log_blocks_per_seg);
1463 return true;
1464 }
1465
1466 if (ssa_blkaddr + (segment_count_ssa << log_blocks_per_seg) !=
1467 main_blkaddr) {
1468 f2fs_msg(sb, KERN_INFO,
1469 "Wrong SSA boundary, start(%u) end(%u) blocks(%u)",
1470 ssa_blkaddr, main_blkaddr,
1471 segment_count_ssa << log_blocks_per_seg);
1472 return true;
1473 }
1474
fd694733 1475 if (main_end_blkaddr > seg_end_blkaddr) {
9a59b62f 1476 f2fs_msg(sb, KERN_INFO,
fd694733 1477 "Wrong MAIN_AREA boundary, start(%u) end(%u) block(%u)",
9a59b62f 1478 main_blkaddr,
fd694733
JK
1479 segment0_blkaddr +
1480 (segment_count << log_blocks_per_seg),
9a59b62f
CY
1481 segment_count_main << log_blocks_per_seg);
1482 return true;
fd694733
JK
1483 } else if (main_end_blkaddr < seg_end_blkaddr) {
1484 int err = 0;
1485 char *res;
1486
1487 /* fix in-memory information all the time */
1488 raw_super->segment_count = cpu_to_le32((main_end_blkaddr -
1489 segment0_blkaddr) >> log_blocks_per_seg);
1490
1491 if (f2fs_readonly(sb) || bdev_read_only(sb->s_bdev)) {
df728b0f 1492 set_sbi_flag(sbi, SBI_NEED_SB_WRITE);
fd694733
JK
1493 res = "internally";
1494 } else {
1495 err = __f2fs_commit_super(bh, NULL);
1496 res = err ? "failed" : "done";
1497 }
1498 f2fs_msg(sb, KERN_INFO,
1499 "Fix alignment : %s, start(%u) end(%u) block(%u)",
1500 res, main_blkaddr,
1501 segment0_blkaddr +
1502 (segment_count << log_blocks_per_seg),
1503 segment_count_main << log_blocks_per_seg);
1504 if (err)
1505 return true;
9a59b62f 1506 }
9a59b62f
CY
1507 return false;
1508}
1509
df728b0f 1510static int sanity_check_raw_super(struct f2fs_sb_info *sbi,
fd694733 1511 struct buffer_head *bh)
aff063e2 1512{
fd694733
JK
1513 struct f2fs_super_block *raw_super = (struct f2fs_super_block *)
1514 (bh->b_data + F2FS_SUPER_OFFSET);
df728b0f 1515 struct super_block *sb = sbi->sb;
aff063e2
JK
1516 unsigned int blocksize;
1517
a07ef784
NJ
1518 if (F2FS_SUPER_MAGIC != le32_to_cpu(raw_super->magic)) {
1519 f2fs_msg(sb, KERN_INFO,
1520 "Magic Mismatch, valid(0x%x) - read(0x%x)",
1521 F2FS_SUPER_MAGIC, le32_to_cpu(raw_super->magic));
aff063e2 1522 return 1;
a07ef784 1523 }
aff063e2 1524
5c9b4692 1525 /* Currently, support only 4KB page cache size */
09cbfeaf 1526 if (F2FS_BLKSIZE != PAGE_SIZE) {
5c9b4692 1527 f2fs_msg(sb, KERN_INFO,
14d7e9de 1528 "Invalid page_cache_size (%lu), supports only 4KB\n",
09cbfeaf 1529 PAGE_SIZE);
5c9b4692 1530 return 1;
1531 }
1532
aff063e2
JK
1533 /* Currently, support only 4KB block size */
1534 blocksize = 1 << le32_to_cpu(raw_super->log_blocksize);
5c9b4692 1535 if (blocksize != F2FS_BLKSIZE) {
a07ef784
NJ
1536 f2fs_msg(sb, KERN_INFO,
1537 "Invalid blocksize (%u), supports only 4KB\n",
1538 blocksize);
aff063e2 1539 return 1;
a07ef784 1540 }
5c9b4692 1541
9a59b62f
CY
1542 /* check log blocks per segment */
1543 if (le32_to_cpu(raw_super->log_blocks_per_seg) != 9) {
1544 f2fs_msg(sb, KERN_INFO,
1545 "Invalid log blocks per segment (%u)\n",
1546 le32_to_cpu(raw_super->log_blocks_per_seg));
1547 return 1;
1548 }
1549
55cf9cb6
CY
1550 /* Currently, support 512/1024/2048/4096 bytes sector size */
1551 if (le32_to_cpu(raw_super->log_sectorsize) >
1552 F2FS_MAX_LOG_SECTOR_SIZE ||
1553 le32_to_cpu(raw_super->log_sectorsize) <
1554 F2FS_MIN_LOG_SECTOR_SIZE) {
1555 f2fs_msg(sb, KERN_INFO, "Invalid log sectorsize (%u)",
1556 le32_to_cpu(raw_super->log_sectorsize));
aff063e2 1557 return 1;
a07ef784 1558 }
55cf9cb6
CY
1559 if (le32_to_cpu(raw_super->log_sectors_per_block) +
1560 le32_to_cpu(raw_super->log_sectorsize) !=
1561 F2FS_MAX_LOG_SECTOR_SIZE) {
1562 f2fs_msg(sb, KERN_INFO,
1563 "Invalid log sectors per block(%u) log sectorsize(%u)",
1564 le32_to_cpu(raw_super->log_sectors_per_block),
1565 le32_to_cpu(raw_super->log_sectorsize));
aff063e2 1566 return 1;
a07ef784 1567 }
9a59b62f
CY
1568
1569 /* check reserved ino info */
1570 if (le32_to_cpu(raw_super->node_ino) != 1 ||
1571 le32_to_cpu(raw_super->meta_ino) != 2 ||
1572 le32_to_cpu(raw_super->root_ino) != 3) {
1573 f2fs_msg(sb, KERN_INFO,
1574 "Invalid Fs Meta Ino: node(%u) meta(%u) root(%u)",
1575 le32_to_cpu(raw_super->node_ino),
1576 le32_to_cpu(raw_super->meta_ino),
1577 le32_to_cpu(raw_super->root_ino));
1578 return 1;
1579 }
1580
b9dd4618
JQ
1581 if (le32_to_cpu(raw_super->segment_count) > F2FS_MAX_SEGMENT) {
1582 f2fs_msg(sb, KERN_INFO,
1583 "Invalid segment count (%u)",
1584 le32_to_cpu(raw_super->segment_count));
1585 return 1;
1586 }
1587
9a59b62f 1588 /* check CP/SIT/NAT/SSA/MAIN_AREA area boundary */
df728b0f 1589 if (sanity_check_area_boundary(sbi, bh))
9a59b62f
CY
1590 return 1;
1591
aff063e2
JK
1592 return 0;
1593}
1594
984ec63c 1595int sanity_check_ckpt(struct f2fs_sb_info *sbi)
aff063e2
JK
1596{
1597 unsigned int total, fsmeta;
577e3495
JK
1598 struct f2fs_super_block *raw_super = F2FS_RAW_SUPER(sbi);
1599 struct f2fs_checkpoint *ckpt = F2FS_CKPT(sbi);
2040fce8 1600 unsigned int ovp_segments, reserved_segments;
15d3042a
JQ
1601 unsigned int main_segs, blocks_per_seg;
1602 int i;
aff063e2
JK
1603
1604 total = le32_to_cpu(raw_super->segment_count);
1605 fsmeta = le32_to_cpu(raw_super->segment_count_ckpt);
1606 fsmeta += le32_to_cpu(raw_super->segment_count_sit);
1607 fsmeta += le32_to_cpu(raw_super->segment_count_nat);
1608 fsmeta += le32_to_cpu(ckpt->rsvd_segment_count);
1609 fsmeta += le32_to_cpu(raw_super->segment_count_ssa);
1610
6bacf52f 1611 if (unlikely(fsmeta >= total))
aff063e2 1612 return 1;
577e3495 1613
2040fce8
JK
1614 ovp_segments = le32_to_cpu(ckpt->overprov_segment_count);
1615 reserved_segments = le32_to_cpu(ckpt->rsvd_segment_count);
1616
1617 if (unlikely(fsmeta < F2FS_MIN_SEGMENTS ||
1618 ovp_segments == 0 || reserved_segments == 0)) {
1619 f2fs_msg(sbi->sb, KERN_ERR,
1620 "Wrong layout: check mkfs.f2fs version");
1621 return 1;
1622 }
1623
15d3042a
JQ
1624 main_segs = le32_to_cpu(raw_super->segment_count_main);
1625 blocks_per_seg = sbi->blocks_per_seg;
1626
1627 for (i = 0; i < NR_CURSEG_NODE_TYPE; i++) {
1628 if (le32_to_cpu(ckpt->cur_node_segno[i]) >= main_segs ||
1629 le16_to_cpu(ckpt->cur_node_blkoff[i]) >= blocks_per_seg)
1630 return 1;
1631 }
1632 for (i = 0; i < NR_CURSEG_DATA_TYPE; i++) {
1633 if (le32_to_cpu(ckpt->cur_data_segno[i]) >= main_segs ||
1634 le16_to_cpu(ckpt->cur_data_blkoff[i]) >= blocks_per_seg)
1635 return 1;
1636 }
1637
1e968fdf 1638 if (unlikely(f2fs_cp_error(sbi))) {
577e3495
JK
1639 f2fs_msg(sbi->sb, KERN_ERR, "A bug case: need to run fsck");
1640 return 1;
1641 }
aff063e2
JK
1642 return 0;
1643}
1644
1645static void init_sb_info(struct f2fs_sb_info *sbi)
1646{
1647 struct f2fs_super_block *raw_super = sbi->raw_super;
e41e6d75 1648 int i, j;
aff063e2
JK
1649
1650 sbi->log_sectors_per_block =
1651 le32_to_cpu(raw_super->log_sectors_per_block);
1652 sbi->log_blocksize = le32_to_cpu(raw_super->log_blocksize);
1653 sbi->blocksize = 1 << sbi->log_blocksize;
1654 sbi->log_blocks_per_seg = le32_to_cpu(raw_super->log_blocks_per_seg);
1655 sbi->blocks_per_seg = 1 << sbi->log_blocks_per_seg;
1656 sbi->segs_per_sec = le32_to_cpu(raw_super->segs_per_sec);
1657 sbi->secs_per_zone = le32_to_cpu(raw_super->secs_per_zone);
1658 sbi->total_sections = le32_to_cpu(raw_super->section_count);
1659 sbi->total_node_count =
1660 (le32_to_cpu(raw_super->segment_count_nat) / 2)
1661 * sbi->blocks_per_seg * NAT_ENTRY_PER_BLOCK;
1662 sbi->root_ino_num = le32_to_cpu(raw_super->root_ino);
1663 sbi->node_ino_num = le32_to_cpu(raw_super->node_ino);
1664 sbi->meta_ino_num = le32_to_cpu(raw_super->meta_ino);
5ec4e49f 1665 sbi->cur_victim_sec = NULL_SECNO;
b1c57c1c 1666 sbi->max_victim_search = DEF_MAX_VICTIM_SEARCH;
aff063e2 1667
ab9fa662 1668 sbi->dir_level = DEF_DIR_LEVEL;
6beceb54 1669 sbi->interval_time[CP_TIME] = DEF_CP_INTERVAL;
d0239e1b 1670 sbi->interval_time[REQ_TIME] = DEF_IDLE_INTERVAL;
caf0047e 1671 clear_sbi_flag(sbi, SBI_NEED_FSCK);
2658e50d 1672
35782b23
JK
1673 for (i = 0; i < NR_COUNT_TYPE; i++)
1674 atomic_set(&sbi->nr_pages[i], 0);
1675
687de7f1
JK
1676 atomic_set(&sbi->wb_sync_req, 0);
1677
2658e50d
JK
1678 INIT_LIST_HEAD(&sbi->s_list);
1679 mutex_init(&sbi->umount_mutex);
e41e6d75
CY
1680 for (i = 0; i < NR_PAGE_TYPE - 1; i++)
1681 for (j = HOT; j < NR_TEMP_TYPE; j++)
1682 mutex_init(&sbi->wio_mutex[i][j]);
aaec2b1d 1683 spin_lock_init(&sbi->cp_lock);
aff063e2
JK
1684}
1685
523be8a6
JK
1686static int init_percpu_info(struct f2fs_sb_info *sbi)
1687{
35782b23 1688 int err;
41382ec4 1689
513c5f37
JK
1690 err = percpu_counter_init(&sbi->alloc_valid_block_count, 0, GFP_KERNEL);
1691 if (err)
1692 return err;
1693
1694 return percpu_counter_init(&sbi->total_valid_inode_count, 0,
41382ec4 1695 GFP_KERNEL);
523be8a6
JK
1696}
1697
178053e2 1698#ifdef CONFIG_BLK_DEV_ZONED
3c62be17 1699static int init_blkz_info(struct f2fs_sb_info *sbi, int devi)
178053e2 1700{
3c62be17 1701 struct block_device *bdev = FDEV(devi).bdev;
178053e2
DLM
1702 sector_t nr_sectors = bdev->bd_part->nr_sects;
1703 sector_t sector = 0;
1704 struct blk_zone *zones;
1705 unsigned int i, nr_zones;
1706 unsigned int n = 0;
1707 int err = -EIO;
1708
1709 if (!f2fs_sb_mounted_blkzoned(sbi->sb))
1710 return 0;
1711
3c62be17 1712 if (sbi->blocks_per_blkz && sbi->blocks_per_blkz !=
f99e8648 1713 SECTOR_TO_BLOCK(bdev_zone_sectors(bdev)))
3c62be17 1714 return -EINVAL;
f99e8648 1715 sbi->blocks_per_blkz = SECTOR_TO_BLOCK(bdev_zone_sectors(bdev));
3c62be17
JK
1716 if (sbi->log_blocks_per_blkz && sbi->log_blocks_per_blkz !=
1717 __ilog2_u32(sbi->blocks_per_blkz))
1718 return -EINVAL;
178053e2 1719 sbi->log_blocks_per_blkz = __ilog2_u32(sbi->blocks_per_blkz);
3c62be17
JK
1720 FDEV(devi).nr_blkz = SECTOR_TO_BLOCK(nr_sectors) >>
1721 sbi->log_blocks_per_blkz;
f99e8648 1722 if (nr_sectors & (bdev_zone_sectors(bdev) - 1))
3c62be17 1723 FDEV(devi).nr_blkz++;
178053e2 1724
3c62be17
JK
1725 FDEV(devi).blkz_type = kmalloc(FDEV(devi).nr_blkz, GFP_KERNEL);
1726 if (!FDEV(devi).blkz_type)
178053e2
DLM
1727 return -ENOMEM;
1728
1729#define F2FS_REPORT_NR_ZONES 4096
1730
1731 zones = kcalloc(F2FS_REPORT_NR_ZONES, sizeof(struct blk_zone),
1732 GFP_KERNEL);
1733 if (!zones)
1734 return -ENOMEM;
1735
1736 /* Get block zones type */
1737 while (zones && sector < nr_sectors) {
1738
1739 nr_zones = F2FS_REPORT_NR_ZONES;
1740 err = blkdev_report_zones(bdev, sector,
1741 zones, &nr_zones,
1742 GFP_KERNEL);
1743 if (err)
1744 break;
1745 if (!nr_zones) {
1746 err = -EIO;
1747 break;
1748 }
1749
1750 for (i = 0; i < nr_zones; i++) {
3c62be17 1751 FDEV(devi).blkz_type[n] = zones[i].type;
178053e2
DLM
1752 sector += zones[i].len;
1753 n++;
1754 }
1755 }
1756
1757 kfree(zones);
1758
1759 return err;
1760}
1761#endif
1762
9076a75f
GZ
1763/*
1764 * Read f2fs raw super block.
2b39e907
SL
1765 * Because we have two copies of super block, so read both of them
1766 * to get the first valid one. If any one of them is broken, we pass
1767 * them recovery flag back to the caller.
9076a75f 1768 */
df728b0f 1769static int read_raw_super_block(struct f2fs_sb_info *sbi,
9076a75f 1770 struct f2fs_super_block **raw_super,
e8240f65 1771 int *valid_super_block, int *recovery)
14d7e9de 1772{
df728b0f 1773 struct super_block *sb = sbi->sb;
2b39e907 1774 int block;
e8240f65 1775 struct buffer_head *bh;
fd694733 1776 struct f2fs_super_block *super;
da554e48 1777 int err = 0;
14d7e9de 1778
b39f0de2
YH
1779 super = kzalloc(sizeof(struct f2fs_super_block), GFP_KERNEL);
1780 if (!super)
1781 return -ENOMEM;
2b39e907
SL
1782
1783 for (block = 0; block < 2; block++) {
1784 bh = sb_bread(sb, block);
1785 if (!bh) {
1786 f2fs_msg(sb, KERN_ERR, "Unable to read %dth superblock",
9076a75f 1787 block + 1);
2b39e907
SL
1788 err = -EIO;
1789 continue;
1790 }
14d7e9de 1791
2b39e907 1792 /* sanity checking of raw super */
df728b0f 1793 if (sanity_check_raw_super(sbi, bh)) {
2b39e907
SL
1794 f2fs_msg(sb, KERN_ERR,
1795 "Can't find valid F2FS filesystem in %dth superblock",
1796 block + 1);
1797 err = -EINVAL;
1798 brelse(bh);
1799 continue;
1800 }
14d7e9de 1801
2b39e907 1802 if (!*raw_super) {
fd694733
JK
1803 memcpy(super, bh->b_data + F2FS_SUPER_OFFSET,
1804 sizeof(*super));
2b39e907
SL
1805 *valid_super_block = block;
1806 *raw_super = super;
1807 }
1808 brelse(bh);
da554e48 1809 }
1810
2b39e907
SL
1811 /* Fail to read any one of the superblocks*/
1812 if (err < 0)
1813 *recovery = 1;
da554e48 1814
da554e48 1815 /* No valid superblock */
2b39e907 1816 if (!*raw_super)
b39f0de2 1817 kfree(super);
2b39e907
SL
1818 else
1819 err = 0;
da554e48 1820
2b39e907 1821 return err;
14d7e9de 1822}
1823
fd694733 1824int f2fs_commit_super(struct f2fs_sb_info *sbi, bool recover)
26d815ad 1825{
5d909cdb 1826 struct buffer_head *bh;
26d815ad
JK
1827 int err;
1828
df728b0f
JK
1829 if ((recover && f2fs_readonly(sbi->sb)) ||
1830 bdev_read_only(sbi->sb->s_bdev)) {
1831 set_sbi_flag(sbi, SBI_NEED_SB_WRITE);
f2353d7b 1832 return -EROFS;
df728b0f 1833 }
f2353d7b 1834
fd694733
JK
1835 /* write back-up superblock first */
1836 bh = sb_getblk(sbi->sb, sbi->valid_super_block ? 0: 1);
5d909cdb
JK
1837 if (!bh)
1838 return -EIO;
fd694733 1839 err = __f2fs_commit_super(bh, F2FS_RAW_SUPER(sbi));
5d909cdb 1840 brelse(bh);
c5bda1c8
CY
1841
1842 /* if we are in recovery path, skip writing valid superblock */
1843 if (recover || err)
5d909cdb 1844 return err;
26d815ad
JK
1845
1846 /* write current valid superblock */
fd694733
JK
1847 bh = sb_getblk(sbi->sb, sbi->valid_super_block);
1848 if (!bh)
1849 return -EIO;
1850 err = __f2fs_commit_super(bh, F2FS_RAW_SUPER(sbi));
1851 brelse(bh);
1852 return err;
26d815ad
JK
1853}
1854
3c62be17
JK
1855static int f2fs_scan_devices(struct f2fs_sb_info *sbi)
1856{
1857 struct f2fs_super_block *raw_super = F2FS_RAW_SUPER(sbi);
7bb3a371 1858 unsigned int max_devices = MAX_DEVICES;
3c62be17
JK
1859 int i;
1860
7bb3a371
MS
1861 /* Initialize single device information */
1862 if (!RDEV(0).path[0]) {
1863 if (!bdev_is_zoned(sbi->sb->s_bdev))
3c62be17 1864 return 0;
7bb3a371
MS
1865 max_devices = 1;
1866 }
3c62be17 1867
7bb3a371
MS
1868 /*
1869 * Initialize multiple devices information, or single
1870 * zoned block device information.
1871 */
1872 sbi->devs = kcalloc(max_devices, sizeof(struct f2fs_dev_info),
1873 GFP_KERNEL);
1874 if (!sbi->devs)
1875 return -ENOMEM;
3c62be17 1876
7bb3a371 1877 for (i = 0; i < max_devices; i++) {
3c62be17 1878
7bb3a371
MS
1879 if (i > 0 && !RDEV(i).path[0])
1880 break;
1881
1882 if (max_devices == 1) {
1883 /* Single zoned block device mount */
1884 FDEV(0).bdev =
1885 blkdev_get_by_dev(sbi->sb->s_bdev->bd_dev,
3c62be17 1886 sbi->sb->s_mode, sbi->sb->s_type);
7bb3a371
MS
1887 } else {
1888 /* Multi-device mount */
1889 memcpy(FDEV(i).path, RDEV(i).path, MAX_PATH_LEN);
1890 FDEV(i).total_segments =
1891 le32_to_cpu(RDEV(i).total_segments);
1892 if (i == 0) {
1893 FDEV(i).start_blk = 0;
1894 FDEV(i).end_blk = FDEV(i).start_blk +
1895 (FDEV(i).total_segments <<
1896 sbi->log_blocks_per_seg) - 1 +
1897 le32_to_cpu(raw_super->segment0_blkaddr);
1898 } else {
1899 FDEV(i).start_blk = FDEV(i - 1).end_blk + 1;
1900 FDEV(i).end_blk = FDEV(i).start_blk +
1901 (FDEV(i).total_segments <<
1902 sbi->log_blocks_per_seg) - 1;
1903 }
1904 FDEV(i).bdev = blkdev_get_by_path(FDEV(i).path,
3c62be17 1905 sbi->sb->s_mode, sbi->sb->s_type);
7bb3a371 1906 }
3c62be17
JK
1907 if (IS_ERR(FDEV(i).bdev))
1908 return PTR_ERR(FDEV(i).bdev);
1909
1910 /* to release errored devices */
1911 sbi->s_ndevs = i + 1;
1912
1913#ifdef CONFIG_BLK_DEV_ZONED
1914 if (bdev_zoned_model(FDEV(i).bdev) == BLK_ZONED_HM &&
1915 !f2fs_sb_mounted_blkzoned(sbi->sb)) {
1916 f2fs_msg(sbi->sb, KERN_ERR,
1917 "Zoned block device feature not enabled\n");
1918 return -EINVAL;
1919 }
1920 if (bdev_zoned_model(FDEV(i).bdev) != BLK_ZONED_NONE) {
1921 if (init_blkz_info(sbi, i)) {
1922 f2fs_msg(sbi->sb, KERN_ERR,
1923 "Failed to initialize F2FS blkzone information");
1924 return -EINVAL;
1925 }
7bb3a371
MS
1926 if (max_devices == 1)
1927 break;
3c62be17
JK
1928 f2fs_msg(sbi->sb, KERN_INFO,
1929 "Mount Device [%2d]: %20s, %8u, %8x - %8x (zone: %s)",
1930 i, FDEV(i).path,
1931 FDEV(i).total_segments,
1932 FDEV(i).start_blk, FDEV(i).end_blk,
1933 bdev_zoned_model(FDEV(i).bdev) == BLK_ZONED_HA ?
1934 "Host-aware" : "Host-managed");
1935 continue;
1936 }
1937#endif
1938 f2fs_msg(sbi->sb, KERN_INFO,
1939 "Mount Device [%2d]: %20s, %8u, %8x - %8x",
1940 i, FDEV(i).path,
1941 FDEV(i).total_segments,
1942 FDEV(i).start_blk, FDEV(i).end_blk);
1943 }
0a595eba
JK
1944 f2fs_msg(sbi->sb, KERN_INFO,
1945 "IO Block Size: %8d KB", F2FS_IO_SIZE_KB(sbi));
3c62be17
JK
1946 return 0;
1947}
1948
aff063e2
JK
1949static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
1950{
1951 struct f2fs_sb_info *sbi;
da554e48 1952 struct f2fs_super_block *raw_super;
aff063e2 1953 struct inode *root;
99e3e858 1954 int err;
2adc3505 1955 bool retry = true, need_fsck = false;
dabc4a5c 1956 char *options = NULL;
e8240f65 1957 int recovery, i, valid_super_block;
8f1dbbbb 1958 struct curseg_info *seg_i;
aff063e2 1959
ed2e621a 1960try_onemore:
da554e48 1961 err = -EINVAL;
1962 raw_super = NULL;
e8240f65 1963 valid_super_block = -1;
da554e48 1964 recovery = 0;
1965
aff063e2
JK
1966 /* allocate memory for f2fs-specific super block info */
1967 sbi = kzalloc(sizeof(struct f2fs_sb_info), GFP_KERNEL);
1968 if (!sbi)
1969 return -ENOMEM;
1970
df728b0f
JK
1971 sbi->sb = sb;
1972
43b6573b
KM
1973 /* Load the checksum driver */
1974 sbi->s_chksum_driver = crypto_alloc_shash("crc32", 0, 0);
1975 if (IS_ERR(sbi->s_chksum_driver)) {
1976 f2fs_msg(sb, KERN_ERR, "Cannot load crc32 driver.");
1977 err = PTR_ERR(sbi->s_chksum_driver);
1978 sbi->s_chksum_driver = NULL;
1979 goto free_sbi;
1980 }
1981
ff9234ad 1982 /* set a block size */
6bacf52f 1983 if (unlikely(!sb_set_blocksize(sb, F2FS_BLKSIZE))) {
a07ef784 1984 f2fs_msg(sb, KERN_ERR, "unable to set blocksize");
aff063e2 1985 goto free_sbi;
a07ef784 1986 }
aff063e2 1987
df728b0f 1988 err = read_raw_super_block(sbi, &raw_super, &valid_super_block,
e8240f65 1989 &recovery);
9076a75f
GZ
1990 if (err)
1991 goto free_sbi;
1992
5fb08372 1993 sb->s_fs_info = sbi;
52763a4b
JK
1994 sbi->raw_super = raw_super;
1995
d1b959c8
DLM
1996 /*
1997 * The BLKZONED feature indicates that the drive was formatted with
1998 * zone alignment optimization. This is optional for host-aware
1999 * devices, but mandatory for host-managed zoned block devices.
2000 */
2001#ifndef CONFIG_BLK_DEV_ZONED
2002 if (f2fs_sb_mounted_blkzoned(sb)) {
2003 f2fs_msg(sb, KERN_ERR,
2004 "Zoned block device support is not enabled\n");
1727f317 2005 err = -EOPNOTSUPP;
d1b959c8
DLM
2006 goto free_sb_buf;
2007 }
d1b959c8 2008#endif
498c5e9f 2009 default_options(sbi);
aff063e2 2010 /* parse mount options */
dabc4a5c
JK
2011 options = kstrdup((const char *)data, GFP_KERNEL);
2012 if (data && !options) {
2013 err = -ENOMEM;
aff063e2 2014 goto free_sb_buf;
dabc4a5c
JK
2015 }
2016
2017 err = parse_options(sb, options);
2018 if (err)
2019 goto free_options;
aff063e2 2020
e0afc4d6
CY
2021 sbi->max_file_blocks = max_file_blocks();
2022 sb->s_maxbytes = sbi->max_file_blocks <<
2023 le32_to_cpu(raw_super->log_blocksize);
aff063e2
JK
2024 sb->s_max_links = F2FS_LINK_MAX;
2025 get_random_bytes(&sbi->s_next_generation, sizeof(u32));
2026
0abd675e
CY
2027#ifdef CONFIG_QUOTA
2028 sb->dq_op = &f2fs_quota_operations;
2029 sb->s_qcop = &f2fs_quotactl_ops;
5c57132e 2030 sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP | QTYPE_MASK_PRJ;
0abd675e
CY
2031#endif
2032
aff063e2 2033 sb->s_op = &f2fs_sops;
0b81d077 2034 sb->s_cop = &f2fs_cryptops;
aff063e2
JK
2035 sb->s_xattr = f2fs_xattr_handlers;
2036 sb->s_export_op = &f2fs_export_ops;
2037 sb->s_magic = F2FS_SUPER_MAGIC;
aff063e2
JK
2038 sb->s_time_gran = 1;
2039 sb->s_flags = (sb->s_flags & ~MS_POSIXACL) |
2040 (test_opt(sbi, POSIX_ACL) ? MS_POSIXACL : 0);
85787090 2041 memcpy(&sb->s_uuid, raw_super->uuid, sizeof(raw_super->uuid));
aff063e2
JK
2042
2043 /* init f2fs-specific super block info */
e8240f65 2044 sbi->valid_super_block = valid_super_block;
aff063e2 2045 mutex_init(&sbi->gc_mutex);
aff063e2 2046 mutex_init(&sbi->cp_mutex);
b3582c68 2047 init_rwsem(&sbi->node_write);
59c9081b 2048 init_rwsem(&sbi->node_change);
315df839
JK
2049
2050 /* disallow all the data/node/meta page writes */
2051 set_sbi_flag(sbi, SBI_POR_DOING);
aff063e2 2052 spin_lock_init(&sbi->stat_lock);
971767ca 2053
458e6197 2054 for (i = 0; i < NR_PAGE_TYPE; i++) {
a912b54d
JK
2055 int n = (i == META) ? 1: NR_TEMP_TYPE;
2056 int j;
2057
2058 sbi->write_io[i] = kmalloc(n * sizeof(struct f2fs_bio_info),
2059 GFP_KERNEL);
b63def91
CJ
2060 if (!sbi->write_io[i]) {
2061 err = -ENOMEM;
a912b54d 2062 goto free_options;
b63def91 2063 }
a912b54d
JK
2064
2065 for (j = HOT; j < n; j++) {
2066 init_rwsem(&sbi->write_io[i][j].io_rwsem);
2067 sbi->write_io[i][j].sbi = sbi;
2068 sbi->write_io[i][j].bio = NULL;
fb830fc5
CY
2069 spin_lock_init(&sbi->write_io[i][j].io_lock);
2070 INIT_LIST_HEAD(&sbi->write_io[i][j].io_list);
a912b54d 2071 }
458e6197 2072 }
971767ca 2073
b873b798 2074 init_rwsem(&sbi->cp_rwsem);
fb51b5ef 2075 init_waitqueue_head(&sbi->cp_wait);
aff063e2
JK
2076 init_sb_info(sbi);
2077
523be8a6
JK
2078 err = init_percpu_info(sbi);
2079 if (err)
2080 goto free_options;
2081
0a595eba
JK
2082 if (F2FS_IO_SIZE(sbi) > 1) {
2083 sbi->write_io_dummy =
a3ebfe4f 2084 mempool_create_page_pool(2 * (F2FS_IO_SIZE(sbi) - 1), 0);
1727f317
CY
2085 if (!sbi->write_io_dummy) {
2086 err = -ENOMEM;
0a595eba 2087 goto free_options;
1727f317 2088 }
0a595eba
JK
2089 }
2090
aff063e2
JK
2091 /* get an inode for meta space */
2092 sbi->meta_inode = f2fs_iget(sb, F2FS_META_INO(sbi));
2093 if (IS_ERR(sbi->meta_inode)) {
a07ef784 2094 f2fs_msg(sb, KERN_ERR, "Failed to read F2FS meta data inode");
aff063e2 2095 err = PTR_ERR(sbi->meta_inode);
0a595eba 2096 goto free_io_dummy;
aff063e2
JK
2097 }
2098
2099 err = get_valid_checkpoint(sbi);
a07ef784
NJ
2100 if (err) {
2101 f2fs_msg(sb, KERN_ERR, "Failed to get valid F2FS checkpoint");
aff063e2 2102 goto free_meta_inode;
a07ef784 2103 }
aff063e2 2104
3c62be17
JK
2105 /* Initialize device list */
2106 err = f2fs_scan_devices(sbi);
2107 if (err) {
2108 f2fs_msg(sb, KERN_ERR, "Failed to find devices");
2109 goto free_devices;
2110 }
2111
aff063e2
JK
2112 sbi->total_valid_node_count =
2113 le32_to_cpu(sbi->ckpt->valid_node_count);
513c5f37
JK
2114 percpu_counter_set(&sbi->total_valid_inode_count,
2115 le32_to_cpu(sbi->ckpt->valid_inode_count));
aff063e2
JK
2116 sbi->user_block_count = le64_to_cpu(sbi->ckpt->user_block_count);
2117 sbi->total_valid_block_count =
2118 le64_to_cpu(sbi->ckpt->valid_block_count);
2119 sbi->last_valid_block_count = sbi->total_valid_block_count;
daeb433e 2120 sbi->reserved_blocks = 0;
41382ec4 2121
c227f912
CY
2122 for (i = 0; i < NR_INODE_TYPE; i++) {
2123 INIT_LIST_HEAD(&sbi->inode_list[i]);
2124 spin_lock_init(&sbi->inode_lock[i]);
2125 }
aff063e2 2126
1dcc336b
CY
2127 init_extent_cache_info(sbi);
2128
6451e041 2129 init_ino_entry_info(sbi);
aff063e2
JK
2130
2131 /* setup f2fs internal modules */
2132 err = build_segment_manager(sbi);
a07ef784
NJ
2133 if (err) {
2134 f2fs_msg(sb, KERN_ERR,
2135 "Failed to initialize F2FS segment manager");
aff063e2 2136 goto free_sm;
a07ef784 2137 }
aff063e2 2138 err = build_node_manager(sbi);
a07ef784
NJ
2139 if (err) {
2140 f2fs_msg(sb, KERN_ERR,
2141 "Failed to initialize F2FS node manager");
aff063e2 2142 goto free_nm;
a07ef784 2143 }
aff063e2 2144
8f1dbbbb
SL
2145 /* For write statistics */
2146 if (sb->s_bdev->bd_part)
2147 sbi->sectors_written_start =
2148 (u64)part_stat_read(sb->s_bdev->bd_part, sectors[1]);
2149
2150 /* Read accumulated write IO statistics if exists */
2151 seg_i = CURSEG_I(sbi, CURSEG_HOT_NODE);
2152 if (__exist_node_summaries(sbi))
2153 sbi->kbytes_written =
b2dde6fc 2154 le64_to_cpu(seg_i->journal->info.kbytes_written);
8f1dbbbb 2155
aff063e2
JK
2156 build_gc_manager(sbi);
2157
2158 /* get an inode for node space */
2159 sbi->node_inode = f2fs_iget(sb, F2FS_NODE_INO(sbi));
2160 if (IS_ERR(sbi->node_inode)) {
a07ef784 2161 f2fs_msg(sb, KERN_ERR, "Failed to read node inode");
aff063e2
JK
2162 err = PTR_ERR(sbi->node_inode);
2163 goto free_nm;
2164 }
2165
2658e50d
JK
2166 f2fs_join_shrinker(sbi);
2167
aa51d08a
JK
2168 err = f2fs_build_stats(sbi);
2169 if (err)
2170 goto free_nm;
2171
aff063e2 2172 /* if there are nt orphan nodes free them */
8c14bfad
CY
2173 err = recover_orphan_inodes(sbi);
2174 if (err)
2175 goto free_node_inode;
aff063e2
JK
2176
2177 /* read root inode and dentry */
2178 root = f2fs_iget(sb, F2FS_ROOT_INO(sbi));
2179 if (IS_ERR(root)) {
a07ef784 2180 f2fs_msg(sb, KERN_ERR, "Failed to read root inode");
aff063e2
JK
2181 err = PTR_ERR(root);
2182 goto free_node_inode;
2183 }
8f99a946 2184 if (!S_ISDIR(root->i_mode) || !root->i_blocks || !root->i_size) {
9d847950 2185 iput(root);
8f99a946 2186 err = -EINVAL;
9d847950 2187 goto free_node_inode;
8f99a946 2188 }
aff063e2
JK
2189
2190 sb->s_root = d_make_root(root); /* allocate root dentry */
2191 if (!sb->s_root) {
2192 err = -ENOMEM;
2193 goto free_root_inode;
2194 }
2195
dc6b2055 2196 err = f2fs_register_sysfs(sbi);
b59d0bae 2197 if (err)
a398101a 2198 goto free_root_inode;
b59d0bae 2199
6437d1b0
JK
2200 /* recover fsynced data */
2201 if (!test_opt(sbi, DISABLE_ROLL_FORWARD)) {
081d78c2
JK
2202 /*
2203 * mount should be failed, when device has readonly mode, and
2204 * previous checkpoint was not done by clean system shutdown.
2205 */
2206 if (bdev_read_only(sb->s_bdev) &&
aaec2b1d 2207 !is_set_ckpt_flags(sbi, CP_UMOUNT_FLAG)) {
081d78c2 2208 err = -EROFS;
a398101a 2209 goto free_sysfs;
081d78c2 2210 }
2adc3505
CY
2211
2212 if (need_fsck)
2213 set_sbi_flag(sbi, SBI_NEED_FSCK);
2214
a468f0ef
JK
2215 if (!retry)
2216 goto skip_recovery;
2217
6781eabb
JK
2218 err = recover_fsync_data(sbi, false);
2219 if (err < 0) {
2adc3505 2220 need_fsck = true;
6437d1b0 2221 f2fs_msg(sb, KERN_ERR,
99e3e858 2222 "Cannot recover all fsync data errno=%d", err);
a398101a 2223 goto free_sysfs;
ed2e621a 2224 }
6781eabb
JK
2225 } else {
2226 err = recover_fsync_data(sbi, true);
2227
2228 if (!f2fs_readonly(sb) && err > 0) {
2229 err = -EINVAL;
2230 f2fs_msg(sb, KERN_ERR,
2231 "Need to recover fsync data");
a398101a 2232 goto free_sysfs;
6781eabb 2233 }
6437d1b0 2234 }
a468f0ef 2235skip_recovery:
315df839
JK
2236 /* recover_fsync_data() cleared this already */
2237 clear_sbi_flag(sbi, SBI_POR_DOING);
b59d0bae 2238
6437d1b0
JK
2239 /*
2240 * If filesystem is not mounted as read-only then
2241 * do start the gc_thread.
2242 */
6c029932 2243 if (test_opt(sbi, BG_GC) && !f2fs_readonly(sb)) {
6437d1b0
JK
2244 /* After POR, we can run background GC thread.*/
2245 err = start_gc_thread(sbi);
2246 if (err)
a398101a 2247 goto free_sysfs;
6437d1b0 2248 }
dabc4a5c 2249 kfree(options);
da554e48 2250
2251 /* recover broken superblock */
f2353d7b 2252 if (recovery) {
41214b3c
CY
2253 err = f2fs_commit_super(sbi, true);
2254 f2fs_msg(sb, KERN_INFO,
99e3e858 2255 "Try to recover %dth superblock, ret: %d",
41214b3c 2256 sbi->valid_super_block ? 1 : 2, err);
da554e48 2257 }
2258
1200abb2
JK
2259 f2fs_msg(sbi->sb, KERN_NOTICE, "Mounted with checkpoint version = %llx",
2260 cur_cp_version(F2FS_CKPT(sbi)));
6beceb54 2261 f2fs_update_time(sbi, CP_TIME);
d0239e1b 2262 f2fs_update_time(sbi, REQ_TIME);
aff063e2 2263 return 0;
6437d1b0 2264
a398101a 2265free_sysfs:
0f18b462 2266 f2fs_sync_inode_meta(sbi);
dc6b2055 2267 f2fs_unregister_sysfs(sbi);
aff063e2
JK
2268free_root_inode:
2269 dput(sb->s_root);
2270 sb->s_root = NULL;
2271free_node_inode:
bb5dada7 2272 truncate_inode_pages_final(NODE_MAPPING(sbi));
2658e50d 2273 mutex_lock(&sbi->umount_mutex);
d41065e2 2274 release_ino_entry(sbi, true);
2658e50d 2275 f2fs_leave_shrinker(sbi);
09922800
JK
2276 /*
2277 * Some dirty meta pages can be produced by recover_orphan_inodes()
2278 * failed by EIO. Then, iput(node_inode) can trigger balance_fs_bg()
2279 * followed by write_checkpoint() through f2fs_write_node_pages(), which
2280 * falls into an infinite loop in sync_meta_pages().
2281 */
2282 truncate_inode_pages_final(META_MAPPING(sbi));
aff063e2 2283 iput(sbi->node_inode);
2658e50d 2284 mutex_unlock(&sbi->umount_mutex);
aa51d08a 2285 f2fs_destroy_stats(sbi);
aff063e2
JK
2286free_nm:
2287 destroy_node_manager(sbi);
2288free_sm:
2289 destroy_segment_manager(sbi);
3c62be17
JK
2290free_devices:
2291 destroy_device_list(sbi);
aff063e2
JK
2292 kfree(sbi->ckpt);
2293free_meta_inode:
2294 make_bad_inode(sbi->meta_inode);
2295 iput(sbi->meta_inode);
0a595eba
JK
2296free_io_dummy:
2297 mempool_destroy(sbi->write_io_dummy);
dabc4a5c 2298free_options:
a912b54d
JK
2299 for (i = 0; i < NR_PAGE_TYPE; i++)
2300 kfree(sbi->write_io[i]);
523be8a6 2301 destroy_percpu_info(sbi);
dabc4a5c 2302 kfree(options);
aff063e2 2303free_sb_buf:
b39f0de2 2304 kfree(raw_super);
aff063e2 2305free_sbi:
43b6573b
KM
2306 if (sbi->s_chksum_driver)
2307 crypto_free_shash(sbi->s_chksum_driver);
aff063e2 2308 kfree(sbi);
ed2e621a
JK
2309
2310 /* give only one another chance */
2311 if (retry) {
9df47ba7 2312 retry = false;
ed2e621a
JK
2313 shrink_dcache_sb(sb);
2314 goto try_onemore;
2315 }
aff063e2
JK
2316 return err;
2317}
2318
2319static struct dentry *f2fs_mount(struct file_system_type *fs_type, int flags,
2320 const char *dev_name, void *data)
2321{
2322 return mount_bdev(fs_type, flags, dev_name, data, f2fs_fill_super);
2323}
2324
30a5537f
JK
2325static void kill_f2fs_super(struct super_block *sb)
2326{
cce13252 2327 if (sb->s_root) {
caf0047e 2328 set_sbi_flag(F2FS_SB(sb), SBI_IS_CLOSE);
cce13252
CY
2329 stop_gc_thread(F2FS_SB(sb));
2330 stop_discard_thread(F2FS_SB(sb));
2331 }
30a5537f
JK
2332 kill_block_super(sb);
2333}
2334
aff063e2
JK
2335static struct file_system_type f2fs_fs_type = {
2336 .owner = THIS_MODULE,
2337 .name = "f2fs",
2338 .mount = f2fs_mount,
30a5537f 2339 .kill_sb = kill_f2fs_super,
aff063e2
JK
2340 .fs_flags = FS_REQUIRES_DEV,
2341};
7f78e035 2342MODULE_ALIAS_FS("f2fs");
aff063e2 2343
6e6093a8 2344static int __init init_inodecache(void)
aff063e2 2345{
5d097056
VD
2346 f2fs_inode_cachep = kmem_cache_create("f2fs_inode_cache",
2347 sizeof(struct f2fs_inode_info), 0,
2348 SLAB_RECLAIM_ACCOUNT|SLAB_ACCOUNT, NULL);
6bacf52f 2349 if (!f2fs_inode_cachep)
aff063e2
JK
2350 return -ENOMEM;
2351 return 0;
2352}
2353
2354static void destroy_inodecache(void)
2355{
2356 /*
2357 * Make sure all delayed rcu free inodes are flushed before we
2358 * destroy cache.
2359 */
2360 rcu_barrier();
2361 kmem_cache_destroy(f2fs_inode_cachep);
2362}
2363
2364static int __init init_f2fs_fs(void)
2365{
2366 int err;
2367
c0508650
JK
2368 f2fs_build_trace_ios();
2369
aff063e2
JK
2370 err = init_inodecache();
2371 if (err)
2372 goto fail;
2373 err = create_node_manager_caches();
2374 if (err)
9890ff3f 2375 goto free_inodecache;
7fd9e544 2376 err = create_segment_manager_caches();
aff063e2 2377 if (err)
9890ff3f 2378 goto free_node_manager_caches;
aff063e2
JK
2379 err = create_checkpoint_caches();
2380 if (err)
06292073 2381 goto free_segment_manager_caches;
1dcc336b
CY
2382 err = create_extent_cache();
2383 if (err)
2384 goto free_checkpoint_caches;
dc6b2055 2385 err = f2fs_init_sysfs();
a398101a 2386 if (err)
1dcc336b 2387 goto free_extent_cache;
2658e50d 2388 err = register_shrinker(&f2fs_shrinker_info);
cfc4d971 2389 if (err)
a398101a 2390 goto free_sysfs;
2658e50d
JK
2391 err = register_filesystem(&f2fs_fs_type);
2392 if (err)
2393 goto free_shrinker;
787c7b8c
CY
2394 err = f2fs_create_root_stats();
2395 if (err)
2396 goto free_filesystem;
9890ff3f
ZH
2397 return 0;
2398
787c7b8c
CY
2399free_filesystem:
2400 unregister_filesystem(&f2fs_fs_type);
2658e50d
JK
2401free_shrinker:
2402 unregister_shrinker(&f2fs_shrinker_info);
a398101a 2403free_sysfs:
dc6b2055 2404 f2fs_exit_sysfs();
1dcc336b
CY
2405free_extent_cache:
2406 destroy_extent_cache();
9890ff3f
ZH
2407free_checkpoint_caches:
2408 destroy_checkpoint_caches();
7fd9e544
JK
2409free_segment_manager_caches:
2410 destroy_segment_manager_caches();
9890ff3f
ZH
2411free_node_manager_caches:
2412 destroy_node_manager_caches();
2413free_inodecache:
2414 destroy_inodecache();
aff063e2
JK
2415fail:
2416 return err;
2417}
2418
2419static void __exit exit_f2fs_fs(void)
2420{
4589d25d 2421 f2fs_destroy_root_stats();
aff063e2 2422 unregister_filesystem(&f2fs_fs_type);
b8bef79d 2423 unregister_shrinker(&f2fs_shrinker_info);
dc6b2055 2424 f2fs_exit_sysfs();
fdf6c8be 2425 destroy_extent_cache();
aff063e2 2426 destroy_checkpoint_caches();
5dcd8a71 2427 destroy_segment_manager_caches();
aff063e2
JK
2428 destroy_node_manager_caches();
2429 destroy_inodecache();
351f4fba 2430 f2fs_destroy_trace_ios();
aff063e2
JK
2431}
2432
2433module_init(init_f2fs_fs)
2434module_exit(exit_f2fs_fs)
2435
2436MODULE_AUTHOR("Samsung Electronics's Praesto Team");
2437MODULE_DESCRIPTION("Flash Friendly File System");
2438MODULE_LICENSE("GPL");
b4b9d34c 2439