123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579 |
- ================================================================================
- WHAT IS Flash-Friendly File System (F2FS)?
- ================================================================================
- NAND flash memory-based storage devices, such as SSD, eMMC, and SD cards, have
- been equipped on a variety systems ranging from mobile to server systems. Since
- they are known to have different characteristics from the conventional rotating
- disks, a file system, an upper layer to the storage device, should adapt to the
- changes from the sketch in the design level.
- F2FS is a file system exploiting NAND flash memory-based storage devices, which
- is based on Log-structured File System (LFS). The design has been focused on
- addressing the fundamental issues in LFS, which are snowball effect of wandering
- tree and high cleaning overhead.
- Since a NAND flash memory-based storage device shows different characteristic
- according to its internal geometry or flash memory management scheme, namely FTL,
- F2FS and its tools support various parameters not only for configuring on-disk
- layout, but also for selecting allocation and cleaning algorithms.
- The following git tree provides the file system formatting tool (mkfs.f2fs),
- a consistency checking tool (fsck.f2fs), and a debugging tool (dump.f2fs).
- >> git://git.kernel.org/pub/scm/linux/kernel/git/jaegeuk/f2fs-tools.git
- For reporting bugs and sending patches, please use the following mailing list:
- >> linux-f2fs-devel@lists.sourceforge.net
- ================================================================================
- BACKGROUND AND DESIGN ISSUES
- ================================================================================
- Log-structured File System (LFS)
- --------------------------------
- "A log-structured file system writes all modifications to disk sequentially in
- a log-like structure, thereby speeding up both file writing and crash recovery.
- The log is the only structure on disk; it contains indexing information so that
- files can be read back from the log efficiently. In order to maintain large free
- areas on disk for fast writing, we divide the log into segments and use a
- segment cleaner to compress the live information from heavily fragmented
- segments." from Rosenblum, M. and Ousterhout, J. K., 1992, "The design and
- implementation of a log-structured file system", ACM Trans. Computer Systems
- 10, 1, 26–52.
- Wandering Tree Problem
- ----------------------
- In LFS, when a file data is updated and written to the end of log, its direct
- pointer block is updated due to the changed location. Then the indirect pointer
- block is also updated due to the direct pointer block update. In this manner,
- the upper index structures such as inode, inode map, and checkpoint block are
- also updated recursively. This problem is called as wandering tree problem [1],
- and in order to enhance the performance, it should eliminate or relax the update
- propagation as much as possible.
- [1] Bityutskiy, A. 2005. JFFS3 design issues. http://www.linux-mtd.infradead.org/
- Cleaning Overhead
- -----------------
- Since LFS is based on out-of-place writes, it produces so many obsolete blocks
- scattered across the whole storage. In order to serve new empty log space, it
- needs to reclaim these obsolete blocks seamlessly to users. This job is called
- as a cleaning process.
- The process consists of three operations as follows.
- 1. A victim segment is selected through referencing segment usage table.
- 2. It loads parent index structures of all the data in the victim identified by
- segment summary blocks.
- 3. It checks the cross-reference between the data and its parent index structure.
- 4. It moves valid data selectively.
- This cleaning job may cause unexpected long delays, so the most important goal
- is to hide the latencies to users. And also definitely, it should reduce the
- amount of valid data to be moved, and move them quickly as well.
- ================================================================================
- KEY FEATURES
- ================================================================================
- Flash Awareness
- ---------------
- - Enlarge the random write area for better performance, but provide the high
- spatial locality
- - Align FS data structures to the operational units in FTL as best efforts
- Wandering Tree Problem
- ----------------------
- - Use a term, “node”, that represents inodes as well as various pointer blocks
- - Introduce Node Address Table (NAT) containing the locations of all the “node”
- blocks; this will cut off the update propagation.
- Cleaning Overhead
- -----------------
- - Support a background cleaning process
- - Support greedy and cost-benefit algorithms for victim selection policies
- - Support multi-head logs for static/dynamic hot and cold data separation
- - Introduce adaptive logging for efficient block allocation
- ================================================================================
- MOUNT OPTIONS
- ================================================================================
- background_gc=%s Turn on/off cleaning operations, namely garbage
- collection, triggered in background when I/O subsystem is
- idle. If background_gc=on, it will turn on the garbage
- collection and if background_gc=off, garbage collection
- will be truned off. If background_gc=sync, it will turn
- on synchronous garbage collection running in background.
- Default value for this option is on. So garbage
- collection is on by default.
- disable_roll_forward Disable the roll-forward recovery routine
- norecovery Disable the roll-forward recovery routine, mounted read-
- only (i.e., -o ro,disable_roll_forward)
- discard Issue discard/TRIM commands when a segment is cleaned.
- no_heap Disable heap-style segment allocation which finds free
- segments for data from the beginning of main area, while
- for node from the end of main area.
- nouser_xattr Disable Extended User Attributes. Note: xattr is enabled
- by default if CONFIG_F2FS_FS_XATTR is selected.
- noacl Disable POSIX Access Control List. Note: acl is enabled
- by default if CONFIG_F2FS_FS_POSIX_ACL is selected.
- active_logs=%u Support configuring the number of active logs. In the
- current design, f2fs supports only 2, 4, and 6 logs.
- Default number is 6.
- disable_ext_identify Disable the extension list configured by mkfs, so f2fs
- does not aware of cold files such as media files.
- inline_xattr Enable the inline xattrs feature.
- inline_data Enable the inline data feature: New created small(<~3.4k)
- files can be written into inode block.
- inline_dentry Enable the inline dir feature: data in new created
- directory entries can be written into inode block. The
- space of inode block which is used to store inline
- dentries is limited to ~3.4k.
- flush_merge Merge concurrent cache_flush commands as much as possible
- to eliminate redundant command issues. If the underlying
- device handles the cache_flush command relatively slowly,
- recommend to enable this option.
- nobarrier This option can be used if underlying storage guarantees
- its cached data should be written to the novolatile area.
- If this option is set, no cache_flush commands are issued
- but f2fs still guarantees the write ordering of all the
- data writes.
- fastboot This option is used when a system wants to reduce mount
- time as much as possible, even though normal performance
- can be sacrificed.
- extent_cache Enable an extent cache based on rb-tree, it can cache
- as many as extent which map between contiguous logical
- address and physical address per inode, resulting in
- increasing the cache hit ratio. Set by default.
- noextent_cache Diable an extent cache based on rb-tree explicitly, see
- the above extent_cache mount option.
- noinline_data Disable the inline data feature, inline data feature is
- enabled by default.
- ================================================================================
- DEBUGFS ENTRIES
- ================================================================================
- /sys/kernel/debug/f2fs/ contains information about all the partitions mounted as
- f2fs. Each file shows the whole f2fs information.
- /sys/kernel/debug/f2fs/status includes:
- - major file system information managed by f2fs currently
- - average SIT information about whole segments
- - current memory footprint consumed by f2fs.
- ================================================================================
- SYSFS ENTRIES
- ================================================================================
- Information about mounted f2f2 file systems can be found in
- /sys/fs/f2fs. Each mounted filesystem will have a directory in
- /sys/fs/f2fs based on its device name (i.e., /sys/fs/f2fs/sda).
- The files in each per-device directory are shown in table below.
- Files in /sys/fs/f2fs/<devname>
- (see also Documentation/ABI/testing/sysfs-fs-f2fs)
- ..............................................................................
- File Content
- gc_max_sleep_time This tuning parameter controls the maximum sleep
- time for the garbage collection thread. Time is
- in milliseconds.
- gc_min_sleep_time This tuning parameter controls the minimum sleep
- time for the garbage collection thread. Time is
- in milliseconds.
- gc_no_gc_sleep_time This tuning parameter controls the default sleep
- time for the garbage collection thread. Time is
- in milliseconds.
- gc_idle This parameter controls the selection of victim
- policy for garbage collection. Setting gc_idle = 0
- (default) will disable this option. Setting
- gc_idle = 1 will select the Cost Benefit approach
- & setting gc_idle = 2 will select the greedy aproach.
- reclaim_segments This parameter controls the number of prefree
- segments to be reclaimed. If the number of prefree
- segments is larger than the number of segments
- in the proportion to the percentage over total
- volume size, f2fs tries to conduct checkpoint to
- reclaim the prefree segments to free segments.
- By default, 5% over total # of segments.
- max_small_discards This parameter controls the number of discard
- commands that consist small blocks less than 2MB.
- The candidates to be discarded are cached until
- checkpoint is triggered, and issued during the
- checkpoint. By default, it is disabled with 0.
- trim_sections This parameter controls the number of sections
- to be trimmed out in batch mode when FITRIM
- conducts. 32 sections is set by default.
- ipu_policy This parameter controls the policy of in-place
- updates in f2fs. There are five policies:
- 0x01: F2FS_IPU_FORCE, 0x02: F2FS_IPU_SSR,
- 0x04: F2FS_IPU_UTIL, 0x08: F2FS_IPU_SSR_UTIL,
- 0x10: F2FS_IPU_FSYNC.
- min_ipu_util This parameter controls the threshold to trigger
- in-place-updates. The number indicates percentage
- of the filesystem utilization, and used by
- F2FS_IPU_UTIL and F2FS_IPU_SSR_UTIL policies.
- min_fsync_blocks This parameter controls the threshold to trigger
- in-place-updates when F2FS_IPU_FSYNC mode is set.
- The number indicates the number of dirty pages
- when fsync needs to flush on its call path. If
- the number is less than this value, it triggers
- in-place-updates.
- max_victim_search This parameter controls the number of trials to
- find a victim segment when conducting SSR and
- cleaning operations. The default value is 4096
- which covers 8GB block address range.
- dir_level This parameter controls the directory level to
- support large directory. If a directory has a
- number of files, it can reduce the file lookup
- latency by increasing this dir_level value.
- Otherwise, it needs to decrease this value to
- reduce the space overhead. The default value is 0.
- ram_thresh This parameter controls the memory footprint used
- by free nids and cached nat entries. By default,
- 10 is set, which indicates 10 MB / 1 GB RAM.
- ================================================================================
- USAGE
- ================================================================================
- 1. Download userland tools and compile them.
- 2. Skip, if f2fs was compiled statically inside kernel.
- Otherwise, insert the f2fs.ko module.
- # insmod f2fs.ko
- 3. Create a directory trying to mount
- # mkdir /mnt/f2fs
- 4. Format the block device, and then mount as f2fs
- # mkfs.f2fs -l label /dev/block_device
- # mount -t f2fs /dev/block_device /mnt/f2fs
- mkfs.f2fs
- ---------
- The mkfs.f2fs is for the use of formatting a partition as the f2fs filesystem,
- which builds a basic on-disk layout.
- The options consist of:
- -l [label] : Give a volume label, up to 512 unicode name.
- -a [0 or 1] : Split start location of each area for heap-based allocation.
- 1 is set by default, which performs this.
- -o [int] : Set overprovision ratio in percent over volume size.
- 5 is set by default.
- -s [int] : Set the number of segments per section.
- 1 is set by default.
- -z [int] : Set the number of sections per zone.
- 1 is set by default.
- -e [str] : Set basic extension list. e.g. "mp3,gif,mov"
- -t [0 or 1] : Disable discard command or not.
- 1 is set by default, which conducts discard.
- fsck.f2fs
- ---------
- The fsck.f2fs is a tool to check the consistency of an f2fs-formatted
- partition, which examines whether the filesystem metadata and user-made data
- are cross-referenced correctly or not.
- Note that, initial version of the tool does not fix any inconsistency.
- The options consist of:
- -d debug level [default:0]
- dump.f2fs
- ---------
- The dump.f2fs shows the information of specific inode and dumps SSA and SIT to
- file. Each file is dump_ssa and dump_sit.
- The dump.f2fs is used to debug on-disk data structures of the f2fs filesystem.
- It shows on-disk inode information reconized by a given inode number, and is
- able to dump all the SSA and SIT entries into predefined files, ./dump_ssa and
- ./dump_sit respectively.
- The options consist of:
- -d debug level [default:0]
- -i inode no (hex)
- -s [SIT dump segno from #1~#2 (decimal), for all 0~-1]
- -a [SSA dump segno from #1~#2 (decimal), for all 0~-1]
- Examples:
- # dump.f2fs -i [ino] /dev/sdx
- # dump.f2fs -s 0~-1 /dev/sdx (SIT dump)
- # dump.f2fs -a 0~-1 /dev/sdx (SSA dump)
- ================================================================================
- DESIGN
- ================================================================================
- On-disk Layout
- --------------
- F2FS divides the whole volume into a number of segments, each of which is fixed
- to 2MB in size. A section is composed of consecutive segments, and a zone
- consists of a set of sections. By default, section and zone sizes are set to one
- segment size identically, but users can easily modify the sizes by mkfs.
- F2FS splits the entire volume into six areas, and all the areas except superblock
- consists of multiple segments as described below.
- align with the zone size <-|
- |-> align with the segment size
- _________________________________________________________________________
- | | | Segment | Node | Segment | |
- | Superblock | Checkpoint | Info. | Address | Summary | Main |
- | (SB) | (CP) | Table (SIT) | Table (NAT) | Area (SSA) | |
- |____________|_____2______|______N______|______N______|______N_____|__N___|
- . .
- . .
- . .
- ._________________________________________.
- |_Segment_|_..._|_Segment_|_..._|_Segment_|
- . .
- ._________._________
- |_section_|__...__|_
- . .
- .________.
- |__zone__|
- - Superblock (SB)
- : It is located at the beginning of the partition, and there exist two copies
- to avoid file system crash. It contains basic partition information and some
- default parameters of f2fs.
- - Checkpoint (CP)
- : It contains file system information, bitmaps for valid NAT/SIT sets, orphan
- inode lists, and summary entries of current active segments.
- - Segment Information Table (SIT)
- : It contains segment information such as valid block count and bitmap for the
- validity of all the blocks.
- - Node Address Table (NAT)
- : It is composed of a block address table for all the node blocks stored in
- Main area.
- - Segment Summary Area (SSA)
- : It contains summary entries which contains the owner information of all the
- data and node blocks stored in Main area.
- - Main Area
- : It contains file and directory data including their indices.
- In order to avoid misalignment between file system and flash-based storage, F2FS
- aligns the start block address of CP with the segment size. Also, it aligns the
- start block address of Main area with the zone size by reserving some segments
- in SSA area.
- Reference the following survey for additional technical details.
- https://wiki.linaro.org/WorkingGroups/Kernel/Projects/FlashCardSurvey
- File System Metadata Structure
- ------------------------------
- F2FS adopts the checkpointing scheme to maintain file system consistency. At
- mount time, F2FS first tries to find the last valid checkpoint data by scanning
- CP area. In order to reduce the scanning time, F2FS uses only two copies of CP.
- One of them always indicates the last valid data, which is called as shadow copy
- mechanism. In addition to CP, NAT and SIT also adopt the shadow copy mechanism.
- For file system consistency, each CP points to which NAT and SIT copies are
- valid, as shown as below.
- +--------+----------+---------+
- | CP | SIT | NAT |
- +--------+----------+---------+
- . . . .
- . . . .
- . . . .
- +-------+-------+--------+--------+--------+--------+
- | CP #0 | CP #1 | SIT #0 | SIT #1 | NAT #0 | NAT #1 |
- +-------+-------+--------+--------+--------+--------+
- | ^ ^
- | | |
- `----------------------------------------'
- Index Structure
- ---------------
- The key data structure to manage the data locations is a "node". Similar to
- traditional file structures, F2FS has three types of node: inode, direct node,
- indirect node. F2FS assigns 4KB to an inode block which contains 923 data block
- indices, two direct node pointers, two indirect node pointers, and one double
- indirect node pointer as described below. One direct node block contains 1018
- data blocks, and one indirect node block contains also 1018 node blocks. Thus,
- one inode block (i.e., a file) covers:
- 4KB * (923 + 2 * 1018 + 2 * 1018 * 1018 + 1018 * 1018 * 1018) := 3.94TB.
- Inode block (4KB)
- |- data (923)
- |- direct node (2)
- | `- data (1018)
- |- indirect node (2)
- | `- direct node (1018)
- | `- data (1018)
- `- double indirect node (1)
- `- indirect node (1018)
- `- direct node (1018)
- `- data (1018)
- Note that, all the node blocks are mapped by NAT which means the location of
- each node is translated by the NAT table. In the consideration of the wandering
- tree problem, F2FS is able to cut off the propagation of node updates caused by
- leaf data writes.
- Directory Structure
- -------------------
- A directory entry occupies 11 bytes, which consists of the following attributes.
- - hash hash value of the file name
- - ino inode number
- - len the length of file name
- - type file type such as directory, symlink, etc
- A dentry block consists of 214 dentry slots and file names. Therein a bitmap is
- used to represent whether each dentry is valid or not. A dentry block occupies
- 4KB with the following composition.
- Dentry Block(4 K) = bitmap (27 bytes) + reserved (3 bytes) +
- dentries(11 * 214 bytes) + file name (8 * 214 bytes)
- [Bucket]
- +--------------------------------+
- |dentry block 1 | dentry block 2 |
- +--------------------------------+
- . .
- . .
- . [Dentry Block Structure: 4KB] .
- +--------+----------+----------+------------+
- | bitmap | reserved | dentries | file names |
- +--------+----------+----------+------------+
- [Dentry Block: 4KB] . .
- . .
- . .
- +------+------+-----+------+
- | hash | ino | len | type |
- +------+------+-----+------+
- [Dentry Structure: 11 bytes]
- F2FS implements multi-level hash tables for directory structure. Each level has
- a hash table with dedicated number of hash buckets as shown below. Note that
- "A(2B)" means a bucket includes 2 data blocks.
- ----------------------
- A : bucket
- B : block
- N : MAX_DIR_HASH_DEPTH
- ----------------------
- level #0 | A(2B)
- |
- level #1 | A(2B) - A(2B)
- |
- level #2 | A(2B) - A(2B) - A(2B) - A(2B)
- . | . . . .
- level #N/2 | A(2B) - A(2B) - A(2B) - A(2B) - A(2B) - ... - A(2B)
- . | . . . .
- level #N | A(4B) - A(4B) - A(4B) - A(4B) - A(4B) - ... - A(4B)
- The number of blocks and buckets are determined by,
- ,- 2, if n < MAX_DIR_HASH_DEPTH / 2,
- # of blocks in level #n = |
- `- 4, Otherwise
- ,- 2^(n + dir_level),
- | if n + dir_level < MAX_DIR_HASH_DEPTH / 2,
- # of buckets in level #n = |
- `- 2^((MAX_DIR_HASH_DEPTH / 2) - 1),
- Otherwise
- When F2FS finds a file name in a directory, at first a hash value of the file
- name is calculated. Then, F2FS scans the hash table in level #0 to find the
- dentry consisting of the file name and its inode number. If not found, F2FS
- scans the next hash table in level #1. In this way, F2FS scans hash tables in
- each levels incrementally from 1 to N. In each levels F2FS needs to scan only
- one bucket determined by the following equation, which shows O(log(# of files))
- complexity.
- bucket number to scan in level #n = (hash value) % (# of buckets in level #n)
- In the case of file creation, F2FS finds empty consecutive slots that cover the
- file name. F2FS searches the empty slots in the hash tables of whole levels from
- 1 to N in the same way as the lookup operation.
- The following figure shows an example of two cases holding children.
- --------------> Dir <--------------
- | |
- child child
- child - child [hole] - child
- child - child - child [hole] - [hole] - child
- Case 1: Case 2:
- Number of children = 6, Number of children = 3,
- File size = 7 File size = 7
- Default Block Allocation
- ------------------------
- At runtime, F2FS manages six active logs inside "Main" area: Hot/Warm/Cold node
- and Hot/Warm/Cold data.
- - Hot node contains direct node blocks of directories.
- - Warm node contains direct node blocks except hot node blocks.
- - Cold node contains indirect node blocks
- - Hot data contains dentry blocks
- - Warm data contains data blocks except hot and cold data blocks
- - Cold data contains multimedia data or migrated data blocks
- LFS has two schemes for free space management: threaded log and copy-and-compac-
- tion. The copy-and-compaction scheme which is known as cleaning, is well-suited
- for devices showing very good sequential write performance, since free segments
- are served all the time for writing new data. However, it suffers from cleaning
- overhead under high utilization. Contrarily, the threaded log scheme suffers
- from random writes, but no cleaning process is needed. F2FS adopts a hybrid
- scheme where the copy-and-compaction scheme is adopted by default, but the
- policy is dynamically changed to the threaded log scheme according to the file
- system status.
- In order to align F2FS with underlying flash-based storage, F2FS allocates a
- segment in a unit of section. F2FS expects that the section size would be the
- same as the unit size of garbage collection in FTL. Furthermore, with respect
- to the mapping granularity in FTL, F2FS allocates each section of the active
- logs from different zones as much as possible, since FTL can write the data in
- the active logs into one allocation unit according to its mapping granularity.
- Cleaning process
- ----------------
- F2FS does cleaning both on demand and in the background. On-demand cleaning is
- triggered when there are not enough free segments to serve VFS calls. Background
- cleaner is operated by a kernel thread, and triggers the cleaning job when the
- system is idle.
- F2FS supports two victim selection policies: greedy and cost-benefit algorithms.
- In the greedy algorithm, F2FS selects a victim segment having the smallest number
- of valid blocks. In the cost-benefit algorithm, F2FS selects a victim segment
- according to the segment age and the number of valid blocks in order to address
- log block thrashing problem in the greedy algorithm. F2FS adopts the greedy
- algorithm for on-demand cleaner, while background cleaner adopts cost-benefit
- algorithm.
- In order to identify whether the data in the victim segment are valid or not,
- F2FS manages a bitmap. Each bit represents the validity of a block, and the
- bitmap is composed of a bit stream covering whole blocks in main area.
|