]>
Commit | Line | Data |
---|---|---|
98e4da8c JK |
1 | ================================================================================ |
2 | WHAT IS Flash-Friendly File System (F2FS)? | |
3 | ================================================================================ | |
4 | ||
5 | NAND flash memory-based storage devices, such as SSD, eMMC, and SD cards, have | |
6 | been equipped on a variety systems ranging from mobile to server systems. Since | |
7 | they are known to have different characteristics from the conventional rotating | |
8 | disks, a file system, an upper layer to the storage device, should adapt to the | |
9 | changes from the sketch in the design level. | |
10 | ||
11 | F2FS is a file system exploiting NAND flash memory-based storage devices, which | |
12 | is based on Log-structured File System (LFS). The design has been focused on | |
13 | addressing the fundamental issues in LFS, which are snowball effect of wandering | |
14 | tree and high cleaning overhead. | |
15 | ||
16 | Since a NAND flash memory-based storage device shows different characteristic | |
17 | according to its internal geometry or flash memory management scheme, namely FTL, | |
18 | F2FS and its tools support various parameters not only for configuring on-disk | |
19 | layout, but also for selecting allocation and cleaning algorithms. | |
20 | ||
d51a7fba CL |
21 | The following git tree provides the file system formatting tool (mkfs.f2fs), |
22 | a consistency checking tool (fsck.f2fs), and a debugging tool (dump.f2fs). | |
5bb446a2 JK |
23 | >> git://git.kernel.org/pub/scm/linux/kernel/git/jaegeuk/f2fs-tools.git |
24 | ||
25 | For reporting bugs and sending patches, please use the following mailing list: | |
26 | >> linux-f2fs-devel@lists.sourceforge.net | |
98e4da8c JK |
27 | |
28 | ================================================================================ | |
29 | BACKGROUND AND DESIGN ISSUES | |
30 | ================================================================================ | |
31 | ||
32 | Log-structured File System (LFS) | |
33 | -------------------------------- | |
34 | "A log-structured file system writes all modifications to disk sequentially in | |
35 | a log-like structure, thereby speeding up both file writing and crash recovery. | |
36 | The log is the only structure on disk; it contains indexing information so that | |
37 | files can be read back from the log efficiently. In order to maintain large free | |
38 | areas on disk for fast writing, we divide the log into segments and use a | |
39 | segment cleaner to compress the live information from heavily fragmented | |
40 | segments." from Rosenblum, M. and Ousterhout, J. K., 1992, "The design and | |
41 | implementation of a log-structured file system", ACM Trans. Computer Systems | |
42 | 10, 1, 26–52. | |
43 | ||
44 | Wandering Tree Problem | |
45 | ---------------------- | |
46 | In LFS, when a file data is updated and written to the end of log, its direct | |
47 | pointer block is updated due to the changed location. Then the indirect pointer | |
48 | block is also updated due to the direct pointer block update. In this manner, | |
49 | the upper index structures such as inode, inode map, and checkpoint block are | |
50 | also updated recursively. This problem is called as wandering tree problem [1], | |
51 | and in order to enhance the performance, it should eliminate or relax the update | |
52 | propagation as much as possible. | |
53 | ||
54 | [1] Bityutskiy, A. 2005. JFFS3 design issues. http://www.linux-mtd.infradead.org/ | |
55 | ||
56 | Cleaning Overhead | |
57 | ----------------- | |
58 | Since LFS is based on out-of-place writes, it produces so many obsolete blocks | |
59 | scattered across the whole storage. In order to serve new empty log space, it | |
60 | needs to reclaim these obsolete blocks seamlessly to users. This job is called | |
61 | as a cleaning process. | |
62 | ||
63 | The process consists of three operations as follows. | |
64 | 1. A victim segment is selected through referencing segment usage table. | |
65 | 2. It loads parent index structures of all the data in the victim identified by | |
66 | segment summary blocks. | |
67 | 3. It checks the cross-reference between the data and its parent index structure. | |
68 | 4. It moves valid data selectively. | |
69 | ||
70 | This cleaning job may cause unexpected long delays, so the most important goal | |
71 | is to hide the latencies to users. And also definitely, it should reduce the | |
72 | amount of valid data to be moved, and move them quickly as well. | |
73 | ||
74 | ================================================================================ | |
75 | KEY FEATURES | |
76 | ================================================================================ | |
77 | ||
78 | Flash Awareness | |
79 | --------------- | |
80 | - Enlarge the random write area for better performance, but provide the high | |
81 | spatial locality | |
82 | - Align FS data structures to the operational units in FTL as best efforts | |
83 | ||
84 | Wandering Tree Problem | |
85 | ---------------------- | |
86 | - Use a term, “node”, that represents inodes as well as various pointer blocks | |
87 | - Introduce Node Address Table (NAT) containing the locations of all the “node” | |
88 | blocks; this will cut off the update propagation. | |
89 | ||
90 | Cleaning Overhead | |
91 | ----------------- | |
92 | - Support a background cleaning process | |
93 | - Support greedy and cost-benefit algorithms for victim selection policies | |
94 | - Support multi-head logs for static/dynamic hot and cold data separation | |
95 | - Introduce adaptive logging for efficient block allocation | |
96 | ||
97 | ================================================================================ | |
98 | MOUNT OPTIONS | |
99 | ================================================================================ | |
100 | ||
696c018c NJ |
101 | background_gc=%s Turn on/off cleaning operations, namely garbage |
102 | collection, triggered in background when I/O subsystem is | |
103 | idle. If background_gc=on, it will turn on the garbage | |
104 | collection and if background_gc=off, garbage collection | |
105 | will be truned off. | |
106 | Default value for this option is on. So garbage | |
107 | collection is on by default. | |
98e4da8c JK |
108 | disable_roll_forward Disable the roll-forward recovery routine |
109 | discard Issue discard/TRIM commands when a segment is cleaned. | |
110 | no_heap Disable heap-style segment allocation which finds free | |
111 | segments for data from the beginning of main area, while | |
112 | for node from the end of main area. | |
113 | nouser_xattr Disable Extended User Attributes. Note: xattr is enabled | |
114 | by default if CONFIG_F2FS_FS_XATTR is selected. | |
115 | noacl Disable POSIX Access Control List. Note: acl is enabled | |
116 | by default if CONFIG_F2FS_FS_POSIX_ACL is selected. | |
117 | active_logs=%u Support configuring the number of active logs. In the | |
118 | current design, f2fs supports only 2, 4, and 6 logs. | |
119 | Default number is 6. | |
120 | disable_ext_identify Disable the extension list configured by mkfs, so f2fs | |
121 | does not aware of cold files such as media files. | |
66e960c6 | 122 | inline_xattr Enable the inline xattrs feature. |
e4024e86 HL |
123 | inline_data Enable the inline data feature: New created small(<~3.4k) |
124 | files can be written into inode block. | |
98e4da8c JK |
125 | |
126 | ================================================================================ | |
127 | DEBUGFS ENTRIES | |
128 | ================================================================================ | |
129 | ||
130 | /sys/kernel/debug/f2fs/ contains information about all the partitions mounted as | |
131 | f2fs. Each file shows the whole f2fs information. | |
132 | ||
133 | /sys/kernel/debug/f2fs/status includes: | |
134 | - major file system information managed by f2fs currently | |
135 | - average SIT information about whole segments | |
136 | - current memory footprint consumed by f2fs. | |
137 | ||
b59d0bae NJ |
138 | ================================================================================ |
139 | SYSFS ENTRIES | |
140 | ================================================================================ | |
141 | ||
142 | Information about mounted f2f2 file systems can be found in | |
143 | /sys/fs/f2fs. Each mounted filesystem will have a directory in | |
144 | /sys/fs/f2fs based on its device name (i.e., /sys/fs/f2fs/sda). | |
145 | The files in each per-device directory are shown in table below. | |
146 | ||
147 | Files in /sys/fs/f2fs/<devname> | |
148 | (see also Documentation/ABI/testing/sysfs-fs-f2fs) | |
149 | .............................................................................. | |
150 | File Content | |
151 | ||
152 | gc_max_sleep_time This tuning parameter controls the maximum sleep | |
153 | time for the garbage collection thread. Time is | |
154 | in milliseconds. | |
155 | ||
156 | gc_min_sleep_time This tuning parameter controls the minimum sleep | |
157 | time for the garbage collection thread. Time is | |
158 | in milliseconds. | |
159 | ||
160 | gc_no_gc_sleep_time This tuning parameter controls the default sleep | |
161 | time for the garbage collection thread. Time is | |
162 | in milliseconds. | |
163 | ||
d2dc095f NJ |
164 | gc_idle This parameter controls the selection of victim |
165 | policy for garbage collection. Setting gc_idle = 0 | |
166 | (default) will disable this option. Setting | |
167 | gc_idle = 1 will select the Cost Benefit approach | |
168 | & setting gc_idle = 2 will select the greedy aproach. | |
169 | ||
ea91e9b0 JK |
170 | reclaim_segments This parameter controls the number of prefree |
171 | segments to be reclaimed. If the number of prefree | |
172 | segments is larger than this number, f2fs tries to | |
173 | conduct checkpoint to reclaim the prefree segments | |
174 | to free segments. By default, 100 segments, 200MB. | |
175 | ||
ba0697ec JK |
176 | max_small_discards This parameter controls the number of discard |
177 | commands that consist small blocks less than 2MB. | |
178 | The candidates to be discarded are cached until | |
179 | checkpoint is triggered, and issued during the | |
180 | checkpoint. By default, it is disabled with 0. | |
181 | ||
216fbd64 JK |
182 | ipu_policy This parameter controls the policy of in-place |
183 | updates in f2fs. There are five policies: | |
184 | 0: F2FS_IPU_FORCE, 1: F2FS_IPU_SSR, | |
185 | 2: F2FS_IPU_UTIL, 3: F2FS_IPU_SSR_UTIL, | |
186 | 4: F2FS_IPU_DISABLE. | |
187 | ||
188 | min_ipu_util This parameter controls the threshold to trigger | |
189 | in-place-updates. The number indicates percentage | |
190 | of the filesystem utilization, and used by | |
191 | F2FS_IPU_UTIL and F2FS_IPU_SSR_UTIL policies. | |
192 | ||
3bac380c JK |
193 | max_victim_search This parameter controls the number of trials to |
194 | find a victim segment when conducting SSR and | |
195 | cleaning operations. The default value is 4096 | |
196 | which covers 8GB block address range. | |
197 | ||
ab9fa662 JK |
198 | dir_level This parameter controls the directory level to |
199 | support large directory. If a directory has a | |
200 | number of files, it can reduce the file lookup | |
201 | latency by increasing this dir_level value. | |
202 | Otherwise, it needs to decrease this value to | |
203 | reduce the space overhead. The default value is 0. | |
204 | ||
cdfc41c1 JK |
205 | ram_thresh This parameter controls the memory footprint used |
206 | by free nids and cached nat entries. By default, | |
207 | 10 is set, which indicates 10 MB / 1 GB RAM. | |
208 | ||
98e4da8c JK |
209 | ================================================================================ |
210 | USAGE | |
211 | ================================================================================ | |
212 | ||
213 | 1. Download userland tools and compile them. | |
214 | ||
215 | 2. Skip, if f2fs was compiled statically inside kernel. | |
216 | Otherwise, insert the f2fs.ko module. | |
217 | # insmod f2fs.ko | |
218 | ||
219 | 3. Create a directory trying to mount | |
220 | # mkdir /mnt/f2fs | |
221 | ||
222 | 4. Format the block device, and then mount as f2fs | |
223 | # mkfs.f2fs -l label /dev/block_device | |
224 | # mount -t f2fs /dev/block_device /mnt/f2fs | |
225 | ||
d51a7fba CL |
226 | mkfs.f2fs |
227 | --------- | |
228 | The mkfs.f2fs is for the use of formatting a partition as the f2fs filesystem, | |
229 | which builds a basic on-disk layout. | |
230 | ||
231 | The options consist of: | |
1571f84a | 232 | -l [label] : Give a volume label, up to 512 unicode name. |
98e4da8c JK |
233 | -a [0 or 1] : Split start location of each area for heap-based allocation. |
234 | 1 is set by default, which performs this. | |
235 | -o [int] : Set overprovision ratio in percent over volume size. | |
236 | 5 is set by default. | |
237 | -s [int] : Set the number of segments per section. | |
238 | 1 is set by default. | |
239 | -z [int] : Set the number of sections per zone. | |
240 | 1 is set by default. | |
241 | -e [str] : Set basic extension list. e.g. "mp3,gif,mov" | |
1571f84a CL |
242 | -t [0 or 1] : Disable discard command or not. |
243 | 1 is set by default, which conducts discard. | |
98e4da8c | 244 | |
d51a7fba CL |
245 | fsck.f2fs |
246 | --------- | |
247 | The fsck.f2fs is a tool to check the consistency of an f2fs-formatted | |
248 | partition, which examines whether the filesystem metadata and user-made data | |
249 | are cross-referenced correctly or not. | |
250 | Note that, initial version of the tool does not fix any inconsistency. | |
251 | ||
252 | The options consist of: | |
253 | -d debug level [default:0] | |
254 | ||
255 | dump.f2fs | |
256 | --------- | |
257 | The dump.f2fs shows the information of specific inode and dumps SSA and SIT to | |
258 | file. Each file is dump_ssa and dump_sit. | |
259 | ||
260 | The dump.f2fs is used to debug on-disk data structures of the f2fs filesystem. | |
261 | It shows on-disk inode information reconized by a given inode number, and is | |
262 | able to dump all the SSA and SIT entries into predefined files, ./dump_ssa and | |
263 | ./dump_sit respectively. | |
264 | ||
265 | The options consist of: | |
266 | -d debug level [default:0] | |
267 | -i inode no (hex) | |
268 | -s [SIT dump segno from #1~#2 (decimal), for all 0~-1] | |
269 | -a [SSA dump segno from #1~#2 (decimal), for all 0~-1] | |
270 | ||
271 | Examples: | |
272 | # dump.f2fs -i [ino] /dev/sdx | |
273 | # dump.f2fs -s 0~-1 /dev/sdx (SIT dump) | |
274 | # dump.f2fs -a 0~-1 /dev/sdx (SSA dump) | |
275 | ||
98e4da8c JK |
276 | ================================================================================ |
277 | DESIGN | |
278 | ================================================================================ | |
279 | ||
280 | On-disk Layout | |
281 | -------------- | |
282 | ||
283 | F2FS divides the whole volume into a number of segments, each of which is fixed | |
284 | to 2MB in size. A section is composed of consecutive segments, and a zone | |
285 | consists of a set of sections. By default, section and zone sizes are set to one | |
286 | segment size identically, but users can easily modify the sizes by mkfs. | |
287 | ||
288 | F2FS splits the entire volume into six areas, and all the areas except superblock | |
289 | consists of multiple segments as described below. | |
290 | ||
291 | align with the zone size <-| | |
292 | |-> align with the segment size | |
293 | _________________________________________________________________________ | |
9268cc35 HL |
294 | | | | Segment | Node | Segment | | |
295 | | Superblock | Checkpoint | Info. | Address | Summary | Main | | |
296 | | (SB) | (CP) | Table (SIT) | Table (NAT) | Area (SSA) | | | |
98e4da8c JK |
297 | |____________|_____2______|______N______|______N______|______N_____|__N___| |
298 | . . | |
299 | . . | |
300 | . . | |
301 | ._________________________________________. | |
302 | |_Segment_|_..._|_Segment_|_..._|_Segment_| | |
303 | . . | |
304 | ._________._________ | |
305 | |_section_|__...__|_ | |
306 | . . | |
307 | .________. | |
308 | |__zone__| | |
309 | ||
310 | - Superblock (SB) | |
311 | : It is located at the beginning of the partition, and there exist two copies | |
312 | to avoid file system crash. It contains basic partition information and some | |
313 | default parameters of f2fs. | |
314 | ||
315 | - Checkpoint (CP) | |
316 | : It contains file system information, bitmaps for valid NAT/SIT sets, orphan | |
317 | inode lists, and summary entries of current active segments. | |
318 | ||
98e4da8c JK |
319 | - Segment Information Table (SIT) |
320 | : It contains segment information such as valid block count and bitmap for the | |
321 | validity of all the blocks. | |
322 | ||
9268cc35 HL |
323 | - Node Address Table (NAT) |
324 | : It is composed of a block address table for all the node blocks stored in | |
325 | Main area. | |
326 | ||
98e4da8c JK |
327 | - Segment Summary Area (SSA) |
328 | : It contains summary entries which contains the owner information of all the | |
329 | data and node blocks stored in Main area. | |
330 | ||
331 | - Main Area | |
332 | : It contains file and directory data including their indices. | |
333 | ||
334 | In order to avoid misalignment between file system and flash-based storage, F2FS | |
335 | aligns the start block address of CP with the segment size. Also, it aligns the | |
336 | start block address of Main area with the zone size by reserving some segments | |
337 | in SSA area. | |
338 | ||
339 | Reference the following survey for additional technical details. | |
340 | https://wiki.linaro.org/WorkingGroups/Kernel/Projects/FlashCardSurvey | |
341 | ||
342 | File System Metadata Structure | |
343 | ------------------------------ | |
344 | ||
345 | F2FS adopts the checkpointing scheme to maintain file system consistency. At | |
346 | mount time, F2FS first tries to find the last valid checkpoint data by scanning | |
347 | CP area. In order to reduce the scanning time, F2FS uses only two copies of CP. | |
348 | One of them always indicates the last valid data, which is called as shadow copy | |
349 | mechanism. In addition to CP, NAT and SIT also adopt the shadow copy mechanism. | |
350 | ||
351 | For file system consistency, each CP points to which NAT and SIT copies are | |
352 | valid, as shown as below. | |
353 | ||
354 | +--------+----------+---------+ | |
9268cc35 | 355 | | CP | SIT | NAT | |
98e4da8c JK |
356 | +--------+----------+---------+ |
357 | . . . . | |
358 | . . . . | |
359 | . . . . | |
360 | +-------+-------+--------+--------+--------+--------+ | |
9268cc35 | 361 | | CP #0 | CP #1 | SIT #0 | SIT #1 | NAT #0 | NAT #1 | |
98e4da8c JK |
362 | +-------+-------+--------+--------+--------+--------+ |
363 | | ^ ^ | |
364 | | | | | |
365 | `----------------------------------------' | |
366 | ||
367 | Index Structure | |
368 | --------------- | |
369 | ||
370 | The key data structure to manage the data locations is a "node". Similar to | |
371 | traditional file structures, F2FS has three types of node: inode, direct node, | |
d08ab08d | 372 | indirect node. F2FS assigns 4KB to an inode block which contains 923 data block |
98e4da8c JK |
373 | indices, two direct node pointers, two indirect node pointers, and one double |
374 | indirect node pointer as described below. One direct node block contains 1018 | |
375 | data blocks, and one indirect node block contains also 1018 node blocks. Thus, | |
376 | one inode block (i.e., a file) covers: | |
377 | ||
378 | 4KB * (923 + 2 * 1018 + 2 * 1018 * 1018 + 1018 * 1018 * 1018) := 3.94TB. | |
379 | ||
380 | Inode block (4KB) | |
381 | |- data (923) | |
382 | |- direct node (2) | |
383 | | `- data (1018) | |
384 | |- indirect node (2) | |
385 | | `- direct node (1018) | |
386 | | `- data (1018) | |
387 | `- double indirect node (1) | |
388 | `- indirect node (1018) | |
389 | `- direct node (1018) | |
390 | `- data (1018) | |
391 | ||
392 | Note that, all the node blocks are mapped by NAT which means the location of | |
393 | each node is translated by the NAT table. In the consideration of the wandering | |
394 | tree problem, F2FS is able to cut off the propagation of node updates caused by | |
395 | leaf data writes. | |
396 | ||
397 | Directory Structure | |
398 | ------------------- | |
399 | ||
400 | A directory entry occupies 11 bytes, which consists of the following attributes. | |
401 | ||
402 | - hash hash value of the file name | |
403 | - ino inode number | |
404 | - len the length of file name | |
405 | - type file type such as directory, symlink, etc | |
406 | ||
407 | A dentry block consists of 214 dentry slots and file names. Therein a bitmap is | |
408 | used to represent whether each dentry is valid or not. A dentry block occupies | |
409 | 4KB with the following composition. | |
410 | ||
411 | Dentry Block(4 K) = bitmap (27 bytes) + reserved (3 bytes) + | |
412 | dentries(11 * 214 bytes) + file name (8 * 214 bytes) | |
413 | ||
414 | [Bucket] | |
415 | +--------------------------------+ | |
416 | |dentry block 1 | dentry block 2 | | |
417 | +--------------------------------+ | |
418 | . . | |
419 | . . | |
420 | . [Dentry Block Structure: 4KB] . | |
421 | +--------+----------+----------+------------+ | |
422 | | bitmap | reserved | dentries | file names | | |
423 | +--------+----------+----------+------------+ | |
424 | [Dentry Block: 4KB] . . | |
425 | . . | |
426 | . . | |
427 | +------+------+-----+------+ | |
428 | | hash | ino | len | type | | |
429 | +------+------+-----+------+ | |
430 | [Dentry Structure: 11 bytes] | |
431 | ||
432 | F2FS implements multi-level hash tables for directory structure. Each level has | |
433 | a hash table with dedicated number of hash buckets as shown below. Note that | |
434 | "A(2B)" means a bucket includes 2 data blocks. | |
435 | ||
436 | ---------------------- | |
437 | A : bucket | |
438 | B : block | |
439 | N : MAX_DIR_HASH_DEPTH | |
440 | ---------------------- | |
441 | ||
442 | level #0 | A(2B) | |
443 | | | |
444 | level #1 | A(2B) - A(2B) | |
445 | | | |
446 | level #2 | A(2B) - A(2B) - A(2B) - A(2B) | |
447 | . | . . . . | |
448 | level #N/2 | A(2B) - A(2B) - A(2B) - A(2B) - A(2B) - ... - A(2B) | |
449 | . | . . . . | |
450 | level #N | A(4B) - A(4B) - A(4B) - A(4B) - A(4B) - ... - A(4B) | |
451 | ||
452 | The number of blocks and buckets are determined by, | |
453 | ||
454 | ,- 2, if n < MAX_DIR_HASH_DEPTH / 2, | |
455 | # of blocks in level #n = | | |
456 | `- 4, Otherwise | |
457 | ||
38431545 JK |
458 | ,- 2^ (n + dir_level), |
459 | | if n < MAX_DIR_HASH_DEPTH / 2, | |
98e4da8c | 460 | # of buckets in level #n = | |
38431545 JK |
461 | `- 2^((MAX_DIR_HASH_DEPTH / 2 + dir_level) - 1), |
462 | Otherwise | |
98e4da8c JK |
463 | |
464 | When F2FS finds a file name in a directory, at first a hash value of the file | |
465 | name is calculated. Then, F2FS scans the hash table in level #0 to find the | |
466 | dentry consisting of the file name and its inode number. If not found, F2FS | |
467 | scans the next hash table in level #1. In this way, F2FS scans hash tables in | |
468 | each levels incrementally from 1 to N. In each levels F2FS needs to scan only | |
469 | one bucket determined by the following equation, which shows O(log(# of files)) | |
470 | complexity. | |
471 | ||
472 | bucket number to scan in level #n = (hash value) % (# of buckets in level #n) | |
473 | ||
474 | In the case of file creation, F2FS finds empty consecutive slots that cover the | |
475 | file name. F2FS searches the empty slots in the hash tables of whole levels from | |
476 | 1 to N in the same way as the lookup operation. | |
477 | ||
478 | The following figure shows an example of two cases holding children. | |
479 | --------------> Dir <-------------- | |
480 | | | | |
481 | child child | |
482 | ||
483 | child - child [hole] - child | |
484 | ||
485 | child - child - child [hole] - [hole] - child | |
486 | ||
487 | Case 1: Case 2: | |
488 | Number of children = 6, Number of children = 3, | |
489 | File size = 7 File size = 7 | |
490 | ||
491 | Default Block Allocation | |
492 | ------------------------ | |
493 | ||
494 | At runtime, F2FS manages six active logs inside "Main" area: Hot/Warm/Cold node | |
495 | and Hot/Warm/Cold data. | |
496 | ||
497 | - Hot node contains direct node blocks of directories. | |
498 | - Warm node contains direct node blocks except hot node blocks. | |
499 | - Cold node contains indirect node blocks | |
500 | - Hot data contains dentry blocks | |
501 | - Warm data contains data blocks except hot and cold data blocks | |
502 | - Cold data contains multimedia data or migrated data blocks | |
503 | ||
504 | LFS has two schemes for free space management: threaded log and copy-and-compac- | |
505 | tion. The copy-and-compaction scheme which is known as cleaning, is well-suited | |
506 | for devices showing very good sequential write performance, since free segments | |
507 | are served all the time for writing new data. However, it suffers from cleaning | |
508 | overhead under high utilization. Contrarily, the threaded log scheme suffers | |
509 | from random writes, but no cleaning process is needed. F2FS adopts a hybrid | |
510 | scheme where the copy-and-compaction scheme is adopted by default, but the | |
511 | policy is dynamically changed to the threaded log scheme according to the file | |
512 | system status. | |
513 | ||
514 | In order to align F2FS with underlying flash-based storage, F2FS allocates a | |
515 | segment in a unit of section. F2FS expects that the section size would be the | |
516 | same as the unit size of garbage collection in FTL. Furthermore, with respect | |
517 | to the mapping granularity in FTL, F2FS allocates each section of the active | |
518 | logs from different zones as much as possible, since FTL can write the data in | |
519 | the active logs into one allocation unit according to its mapping granularity. | |
520 | ||
521 | Cleaning process | |
522 | ---------------- | |
523 | ||
524 | F2FS does cleaning both on demand and in the background. On-demand cleaning is | |
525 | triggered when there are not enough free segments to serve VFS calls. Background | |
526 | cleaner is operated by a kernel thread, and triggers the cleaning job when the | |
527 | system is idle. | |
528 | ||
529 | F2FS supports two victim selection policies: greedy and cost-benefit algorithms. | |
530 | In the greedy algorithm, F2FS selects a victim segment having the smallest number | |
531 | of valid blocks. In the cost-benefit algorithm, F2FS selects a victim segment | |
532 | according to the segment age and the number of valid blocks in order to address | |
533 | log block thrashing problem in the greedy algorithm. F2FS adopts the greedy | |
534 | algorithm for on-demand cleaner, while background cleaner adopts cost-benefit | |
535 | algorithm. | |
536 | ||
537 | In order to identify whether the data in the victim segment are valid or not, | |
538 | F2FS manages a bitmap. Each bit represents the validity of a block, and the | |
539 | bitmap is composed of a bit stream covering whole blocks in main area. |