mirror of https://github.com/OpenIPC/firmware.git
877 lines
25 KiB
Diff
877 lines
25 KiB
Diff
diff -drupN a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c
|
|
--- a/fs/f2fs/checkpoint.c 2018-08-06 17:23:04.000000000 +0300
|
|
+++ b/fs/f2fs/checkpoint.c 2022-06-12 05:28:14.000000000 +0300
|
|
@@ -29,9 +29,8 @@ struct kmem_cache *inode_entry_slab;
|
|
void f2fs_stop_checkpoint(struct f2fs_sb_info *sbi, bool end_io)
|
|
{
|
|
set_ckpt_flags(sbi, CP_ERROR_FLAG);
|
|
- sbi->sb->s_flags |= MS_RDONLY;
|
|
if (!end_io)
|
|
- f2fs_flush_merged_bios(sbi);
|
|
+ f2fs_flush_merged_writes(sbi);
|
|
}
|
|
|
|
/*
|
|
@@ -65,10 +64,11 @@ static struct page *__get_meta_page(stru
|
|
.sbi = sbi,
|
|
.type = META,
|
|
.op = REQ_OP_READ,
|
|
- .op_flags = READ_SYNC | REQ_META | REQ_PRIO,
|
|
+ .op_flags = REQ_META | REQ_PRIO,
|
|
.old_blkaddr = index,
|
|
.new_blkaddr = index,
|
|
.encrypted_page = NULL,
|
|
+ .is_meta = is_meta,
|
|
};
|
|
|
|
if (unlikely(!is_meta))
|
|
@@ -160,8 +160,10 @@ int ra_meta_pages(struct f2fs_sb_info *s
|
|
.sbi = sbi,
|
|
.type = META,
|
|
.op = REQ_OP_READ,
|
|
- .op_flags = sync ? (READ_SYNC | REQ_META | REQ_PRIO) : REQ_RAHEAD,
|
|
+ .op_flags = sync ? (REQ_META | REQ_PRIO) : REQ_RAHEAD,
|
|
.encrypted_page = NULL,
|
|
+ .in_list = false,
|
|
+ .is_meta = (type != META_POR),
|
|
};
|
|
struct blk_plug plug;
|
|
|
|
@@ -207,12 +209,10 @@ int ra_meta_pages(struct f2fs_sb_info *s
|
|
}
|
|
|
|
fio.page = page;
|
|
- fio.old_blkaddr = fio.new_blkaddr;
|
|
- f2fs_submit_page_mbio(&fio);
|
|
+ f2fs_submit_page_bio(&fio);
|
|
f2fs_put_page(page, 0);
|
|
}
|
|
out:
|
|
- f2fs_submit_merged_bio(sbi, META, READ);
|
|
blk_finish_plug(&plug);
|
|
return blkno - start;
|
|
}
|
|
@@ -228,33 +228,38 @@ void ra_meta_pages_cond(struct f2fs_sb_i
|
|
f2fs_put_page(page, 0);
|
|
|
|
if (readahead)
|
|
- ra_meta_pages(sbi, index, MAX_BIO_BLOCKS(sbi), META_POR, true);
|
|
+ ra_meta_pages(sbi, index, BIO_MAX_PAGES, META_POR, true);
|
|
}
|
|
|
|
-static int f2fs_write_meta_page(struct page *page,
|
|
- struct writeback_control *wbc)
|
|
+static int __f2fs_write_meta_page(struct page *page,
|
|
+ struct writeback_control *wbc,
|
|
+ enum iostat_type io_type)
|
|
{
|
|
struct f2fs_sb_info *sbi = F2FS_P_SB(page);
|
|
|
|
trace_f2fs_writepage(page, META);
|
|
|
|
+ if (unlikely(f2fs_cp_error(sbi))) {
|
|
+ dec_page_count(sbi, F2FS_DIRTY_META);
|
|
+ unlock_page(page);
|
|
+ return 0;
|
|
+ }
|
|
if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING)))
|
|
goto redirty_out;
|
|
if (wbc->for_reclaim && page->index < GET_SUM_BLOCK(sbi, 0))
|
|
goto redirty_out;
|
|
- if (unlikely(f2fs_cp_error(sbi)))
|
|
- goto redirty_out;
|
|
|
|
- write_meta_page(sbi, page);
|
|
+ write_meta_page(sbi, page, io_type);
|
|
dec_page_count(sbi, F2FS_DIRTY_META);
|
|
|
|
if (wbc->for_reclaim)
|
|
- f2fs_submit_merged_bio_cond(sbi, NULL, page, 0, META, WRITE);
|
|
+ f2fs_submit_merged_write_cond(sbi, page->mapping->host,
|
|
+ 0, page->index, META);
|
|
|
|
unlock_page(page);
|
|
|
|
if (unlikely(f2fs_cp_error(sbi)))
|
|
- f2fs_submit_merged_bio(sbi, META, WRITE);
|
|
+ f2fs_submit_merged_write(sbi, META);
|
|
|
|
return 0;
|
|
|
|
@@ -263,23 +268,33 @@ redirty_out:
|
|
return AOP_WRITEPAGE_ACTIVATE;
|
|
}
|
|
|
|
+static int f2fs_write_meta_page(struct page *page,
|
|
+ struct writeback_control *wbc)
|
|
+{
|
|
+ return __f2fs_write_meta_page(page, wbc, FS_META_IO);
|
|
+}
|
|
+
|
|
static int f2fs_write_meta_pages(struct address_space *mapping,
|
|
struct writeback_control *wbc)
|
|
{
|
|
struct f2fs_sb_info *sbi = F2FS_M_SB(mapping);
|
|
long diff, written;
|
|
|
|
+ if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING)))
|
|
+ goto skip_write;
|
|
+
|
|
/* collect a number of dirty meta pages and write together */
|
|
if (wbc->for_kupdate ||
|
|
get_pages(sbi, F2FS_DIRTY_META) < nr_pages_to_skip(sbi, META))
|
|
goto skip_write;
|
|
|
|
- trace_f2fs_writepages(mapping->host, wbc, META);
|
|
+ /* if locked failed, cp will flush dirty pages instead */
|
|
+ if (!mutex_trylock(&sbi->cp_mutex))
|
|
+ goto skip_write;
|
|
|
|
- /* if mounting is failed, skip writing node pages */
|
|
- mutex_lock(&sbi->cp_mutex);
|
|
+ trace_f2fs_writepages(mapping->host, wbc, META);
|
|
diff = nr_pages_to_write(sbi, META, wbc);
|
|
- written = sync_meta_pages(sbi, META, wbc->nr_to_write);
|
|
+ written = sync_meta_pages(sbi, META, wbc->nr_to_write, FS_META_IO);
|
|
mutex_unlock(&sbi->cp_mutex);
|
|
wbc->nr_to_write = max((long)0, wbc->nr_to_write - written - diff);
|
|
return 0;
|
|
@@ -291,7 +306,7 @@ skip_write:
|
|
}
|
|
|
|
long sync_meta_pages(struct f2fs_sb_info *sbi, enum page_type type,
|
|
- long nr_to_write)
|
|
+ long nr_to_write, enum iostat_type io_type)
|
|
{
|
|
struct address_space *mapping = META_MAPPING(sbi);
|
|
pgoff_t index = 0, end = ULONG_MAX, prev = ULONG_MAX;
|
|
@@ -342,7 +357,7 @@ continue_unlock:
|
|
if (!clear_page_dirty_for_io(page))
|
|
goto continue_unlock;
|
|
|
|
- if (mapping->a_ops->writepage(page, &wbc)) {
|
|
+ if (__f2fs_write_meta_page(page, &wbc, io_type)) {
|
|
unlock_page(page);
|
|
break;
|
|
}
|
|
@@ -356,7 +371,7 @@ continue_unlock:
|
|
}
|
|
stop:
|
|
if (nwritten)
|
|
- f2fs_submit_merged_bio(sbi, type, WRITE);
|
|
+ f2fs_submit_merged_write(sbi, type);
|
|
|
|
blk_finish_plug(&plug);
|
|
|
|
@@ -370,7 +385,7 @@ static int f2fs_set_meta_page_dirty(stru
|
|
if (!PageUptodate(page))
|
|
SetPageUptodate(page);
|
|
if (!PageDirty(page)) {
|
|
- f2fs_set_page_dirty_nobuffers(page);
|
|
+ __set_page_dirty_nobuffers(page);
|
|
inc_page_count(F2FS_P_SB(page), F2FS_DIRTY_META);
|
|
SetPagePrivate(page);
|
|
f2fs_trace_pid(page);
|
|
@@ -390,24 +405,23 @@ const struct address_space_operations f2
|
|
#endif
|
|
};
|
|
|
|
-static void __add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
|
|
+static void __add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino,
|
|
+ unsigned int devidx, int type)
|
|
{
|
|
struct inode_management *im = &sbi->im[type];
|
|
struct ino_entry *e, *tmp;
|
|
|
|
tmp = f2fs_kmem_cache_alloc(ino_entry_slab, GFP_NOFS);
|
|
-retry:
|
|
+
|
|
radix_tree_preload(GFP_NOFS | __GFP_NOFAIL);
|
|
|
|
spin_lock(&im->ino_lock);
|
|
e = radix_tree_lookup(&im->ino_root, ino);
|
|
if (!e) {
|
|
e = tmp;
|
|
- if (radix_tree_insert(&im->ino_root, ino, e)) {
|
|
- spin_unlock(&im->ino_lock);
|
|
- radix_tree_preload_end();
|
|
- goto retry;
|
|
- }
|
|
+ if (unlikely(radix_tree_insert(&im->ino_root, ino, e)))
|
|
+ f2fs_bug_on(sbi, 1);
|
|
+
|
|
memset(e, 0, sizeof(struct ino_entry));
|
|
e->ino = ino;
|
|
|
|
@@ -415,6 +429,10 @@ retry:
|
|
if (type != ORPHAN_INO)
|
|
im->ino_num++;
|
|
}
|
|
+
|
|
+ if (type == FLUSH_INO)
|
|
+ f2fs_set_bit(devidx, (char *)&e->dirty_device);
|
|
+
|
|
spin_unlock(&im->ino_lock);
|
|
radix_tree_preload_end();
|
|
|
|
@@ -443,7 +461,7 @@ static void __remove_ino_entry(struct f2
|
|
void add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
|
|
{
|
|
/* add new dirty ino entry into list */
|
|
- __add_ino_entry(sbi, ino, type);
|
|
+ __add_ino_entry(sbi, ino, 0, type);
|
|
}
|
|
|
|
void remove_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
|
|
@@ -469,7 +487,7 @@ void release_ino_entry(struct f2fs_sb_in
|
|
struct ino_entry *e, *tmp;
|
|
int i;
|
|
|
|
- for (i = all ? ORPHAN_INO: APPEND_INO; i <= UPDATE_INO; i++) {
|
|
+ for (i = all ? ORPHAN_INO : APPEND_INO; i < MAX_INO_ENTRY; i++) {
|
|
struct inode_management *im = &sbi->im[i];
|
|
|
|
spin_lock(&im->ino_lock);
|
|
@@ -483,6 +501,27 @@ void release_ino_entry(struct f2fs_sb_in
|
|
}
|
|
}
|
|
|
|
+void set_dirty_device(struct f2fs_sb_info *sbi, nid_t ino,
|
|
+ unsigned int devidx, int type)
|
|
+{
|
|
+ __add_ino_entry(sbi, ino, devidx, type);
|
|
+}
|
|
+
|
|
+bool is_dirty_device(struct f2fs_sb_info *sbi, nid_t ino,
|
|
+ unsigned int devidx, int type)
|
|
+{
|
|
+ struct inode_management *im = &sbi->im[type];
|
|
+ struct ino_entry *e;
|
|
+ bool is_dirty = false;
|
|
+
|
|
+ spin_lock(&im->ino_lock);
|
|
+ e = radix_tree_lookup(&im->ino_root, ino);
|
|
+ if (e && f2fs_test_bit(devidx, (char *)&e->dirty_device))
|
|
+ is_dirty = true;
|
|
+ spin_unlock(&im->ino_lock);
|
|
+ return is_dirty;
|
|
+}
|
|
+
|
|
int acquire_orphan_inode(struct f2fs_sb_info *sbi)
|
|
{
|
|
struct inode_management *im = &sbi->im[ORPHAN_INO];
|
|
@@ -493,6 +532,7 @@ int acquire_orphan_inode(struct f2fs_sb_
|
|
#ifdef CONFIG_F2FS_FAULT_INJECTION
|
|
if (time_to_inject(sbi, FAULT_ORPHAN)) {
|
|
spin_unlock(&im->ino_lock);
|
|
+ f2fs_show_injection_info(FAULT_ORPHAN);
|
|
return -ENOSPC;
|
|
}
|
|
#endif
|
|
@@ -518,7 +558,7 @@ void release_orphan_inode(struct f2fs_sb
|
|
void add_orphan_inode(struct inode *inode)
|
|
{
|
|
/* add new orphan ino entry into list */
|
|
- __add_ino_entry(F2FS_I_SB(inode), inode->i_ino, ORPHAN_INO);
|
|
+ __add_ino_entry(F2FS_I_SB(inode), inode->i_ino, 0, ORPHAN_INO);
|
|
update_inode_page(inode);
|
|
}
|
|
|
|
@@ -534,15 +574,10 @@ static int recover_orphan_inode(struct f
|
|
struct node_info ni;
|
|
int err = acquire_orphan_inode(sbi);
|
|
|
|
- if (err) {
|
|
- set_sbi_flag(sbi, SBI_NEED_FSCK);
|
|
- f2fs_msg(sbi->sb, KERN_WARNING,
|
|
- "%s: orphan failed (ino=%x), run fsck to fix.",
|
|
- __func__, ino);
|
|
- return err;
|
|
- }
|
|
+ if (err)
|
|
+ goto err_out;
|
|
|
|
- __add_ino_entry(sbi, ino, ORPHAN_INO);
|
|
+ __add_ino_entry(sbi, ino, 0, ORPHAN_INO);
|
|
|
|
inode = f2fs_iget_retry(sbi->sb, ino);
|
|
if (IS_ERR(inode)) {
|
|
@@ -554,6 +589,11 @@ static int recover_orphan_inode(struct f
|
|
return PTR_ERR(inode);
|
|
}
|
|
|
|
+ err = dquot_initialize(inode);
|
|
+ if (err)
|
|
+ goto err_out;
|
|
+
|
|
+ dquot_initialize(inode);
|
|
clear_nlink(inode);
|
|
|
|
/* truncate all the data during iput */
|
|
@@ -563,24 +603,45 @@ static int recover_orphan_inode(struct f
|
|
|
|
/* ENOMEM was fully retried in f2fs_evict_inode. */
|
|
if (ni.blk_addr != NULL_ADDR) {
|
|
- set_sbi_flag(sbi, SBI_NEED_FSCK);
|
|
- f2fs_msg(sbi->sb, KERN_WARNING,
|
|
- "%s: orphan failed (ino=%x), run fsck to fix.",
|
|
- __func__, ino);
|
|
- return -EIO;
|
|
+ err = -EIO;
|
|
+ goto err_out;
|
|
}
|
|
__remove_ino_entry(sbi, ino, ORPHAN_INO);
|
|
return 0;
|
|
+
|
|
+err_out:
|
|
+ set_sbi_flag(sbi, SBI_NEED_FSCK);
|
|
+ f2fs_msg(sbi->sb, KERN_WARNING,
|
|
+ "%s: orphan failed (ino=%x), run fsck to fix.",
|
|
+ __func__, ino);
|
|
+ return err;
|
|
}
|
|
|
|
int recover_orphan_inodes(struct f2fs_sb_info *sbi)
|
|
{
|
|
block_t start_blk, orphan_blocks, i, j;
|
|
- int err;
|
|
+ unsigned int s_flags = sbi->sb->s_flags;
|
|
+ int err = 0;
|
|
+#ifdef CONFIG_QUOTA
|
|
+ int quota_enabled;
|
|
+#endif
|
|
|
|
if (!is_set_ckpt_flags(sbi, CP_ORPHAN_PRESENT_FLAG))
|
|
return 0;
|
|
|
|
+ if (s_flags & MS_RDONLY) {
|
|
+ f2fs_msg(sbi->sb, KERN_INFO, "orphan cleanup on readonly fs");
|
|
+ sbi->sb->s_flags &= ~MS_RDONLY;
|
|
+ }
|
|
+
|
|
+#ifdef CONFIG_QUOTA
|
|
+ /* Needed for iput() to work correctly and not trash data */
|
|
+ sbi->sb->s_flags |= MS_ACTIVE;
|
|
+
|
|
+ /* Turn on quotas so that they are updated correctly */
|
|
+ quota_enabled = f2fs_enable_quota_files(sbi, s_flags & MS_RDONLY);
|
|
+#endif
|
|
+
|
|
start_blk = __start_cp_addr(sbi) + 1 + __cp_payload(sbi);
|
|
orphan_blocks = __start_sum_addr(sbi) - 1 - __cp_payload(sbi);
|
|
|
|
@@ -596,14 +657,22 @@ int recover_orphan_inodes(struct f2fs_sb
|
|
err = recover_orphan_inode(sbi, ino);
|
|
if (err) {
|
|
f2fs_put_page(page, 1);
|
|
- return err;
|
|
+ goto out;
|
|
}
|
|
}
|
|
f2fs_put_page(page, 1);
|
|
}
|
|
/* clear Orphan Flag */
|
|
clear_ckpt_flags(sbi, CP_ORPHAN_PRESENT_FLAG);
|
|
- return 0;
|
|
+out:
|
|
+#ifdef CONFIG_QUOTA
|
|
+ /* Turn quotas off */
|
|
+ if (quota_enabled)
|
|
+ f2fs_quota_off_umount(sbi->sb);
|
|
+#endif
|
|
+ sbi->sb->s_flags = s_flags; /* Restore MS_RDONLY status */
|
|
+
|
|
+ return err;
|
|
}
|
|
|
|
static void write_orphan_inodes(struct f2fs_sb_info *sbi, block_t start_blk)
|
|
@@ -675,14 +744,13 @@ static int get_checkpoint_version(struct
|
|
*cp_block = (struct f2fs_checkpoint *)page_address(*cp_page);
|
|
|
|
crc_offset = le32_to_cpu((*cp_block)->checksum_offset);
|
|
- if (crc_offset >= blk_size) {
|
|
+ if (crc_offset > (blk_size - sizeof(__le32))) {
|
|
f2fs_msg(sbi->sb, KERN_WARNING,
|
|
"invalid crc_offset: %zu", crc_offset);
|
|
return -EINVAL;
|
|
}
|
|
|
|
- crc = le32_to_cpu(*((__le32 *)((unsigned char *)*cp_block
|
|
- + crc_offset)));
|
|
+ crc = cur_cp_crc(*cp_block);
|
|
if (!f2fs_crc_valid(sbi, crc, *cp_block, crc_offset)) {
|
|
f2fs_msg(sbi->sb, KERN_WARNING, "invalid crc value");
|
|
return -EINVAL;
|
|
@@ -737,7 +805,7 @@ int get_valid_checkpoint(struct f2fs_sb_
|
|
block_t cp_blk_no;
|
|
int i;
|
|
|
|
- sbi->ckpt = kzalloc(cp_blks * blk_size, GFP_KERNEL);
|
|
+ sbi->ckpt = f2fs_kzalloc(sbi, cp_blks * blk_size, GFP_KERNEL);
|
|
if (!sbi->ckpt)
|
|
return -ENOMEM;
|
|
/*
|
|
@@ -770,7 +838,7 @@ int get_valid_checkpoint(struct f2fs_sb_
|
|
|
|
/* Sanity checking of checkpoint */
|
|
if (sanity_check_ckpt(sbi))
|
|
- goto fail_no_cp;
|
|
+ goto free_fail_no_cp;
|
|
|
|
if (cur_page == cp1)
|
|
sbi->cur_cp_pack = 1;
|
|
@@ -798,6 +866,9 @@ done:
|
|
f2fs_put_page(cp2, 1);
|
|
return 0;
|
|
|
|
+free_fail_no_cp:
|
|
+ f2fs_put_page(cp1, 1);
|
|
+ f2fs_put_page(cp2, 1);
|
|
fail_no_cp:
|
|
kfree(sbi->ckpt);
|
|
return -EINVAL;
|
|
@@ -812,7 +883,9 @@ static void __add_dirty_inode(struct ino
|
|
return;
|
|
|
|
set_inode_flag(inode, flag);
|
|
- list_add_tail(&F2FS_I(inode)->dirty_list, &sbi->inode_list[type]);
|
|
+ if (!f2fs_is_volatile_file(inode))
|
|
+ list_add_tail(&F2FS_I(inode)->dirty_list,
|
|
+ &sbi->inode_list[type]);
|
|
stat_inc_dirty_inode(sbi, type);
|
|
}
|
|
|
|
@@ -870,6 +943,7 @@ int sync_dirty_inodes(struct f2fs_sb_inf
|
|
struct inode *inode;
|
|
struct f2fs_inode_info *fi;
|
|
bool is_dir = (type == DIR_INODE);
|
|
+ unsigned long ino = 0;
|
|
|
|
trace_f2fs_sync_dirty_inodes_enter(sbi->sb, is_dir,
|
|
get_pages(sbi, is_dir ?
|
|
@@ -888,18 +962,34 @@ retry:
|
|
F2FS_DIRTY_DENTS : F2FS_DIRTY_DATA));
|
|
return 0;
|
|
}
|
|
- fi = list_entry(head->next, struct f2fs_inode_info, dirty_list);
|
|
+ fi = list_first_entry(head, struct f2fs_inode_info, dirty_list);
|
|
inode = igrab(&fi->vfs_inode);
|
|
spin_unlock(&sbi->inode_lock[type]);
|
|
if (inode) {
|
|
+ unsigned long cur_ino = inode->i_ino;
|
|
+
|
|
+ if (is_dir)
|
|
+ F2FS_I(inode)->cp_task = current;
|
|
+
|
|
filemap_fdatawrite(inode->i_mapping);
|
|
+
|
|
+ if (is_dir)
|
|
+ F2FS_I(inode)->cp_task = NULL;
|
|
+
|
|
iput(inode);
|
|
+ /* We need to give cpu to another writers. */
|
|
+ if (ino == cur_ino) {
|
|
+ congestion_wait(BLK_RW_ASYNC, HZ/50);
|
|
+ cond_resched();
|
|
+ } else {
|
|
+ ino = cur_ino;
|
|
+ }
|
|
} else {
|
|
/*
|
|
* We should submit bio, since it exists several
|
|
* wribacking dentry pages in the freeing inode.
|
|
*/
|
|
- f2fs_submit_merged_bio(sbi, DATA, WRITE);
|
|
+ f2fs_submit_merged_write(sbi, DATA);
|
|
cond_resched();
|
|
}
|
|
goto retry;
|
|
@@ -921,18 +1011,35 @@ int f2fs_sync_inode_meta(struct f2fs_sb_
|
|
spin_unlock(&sbi->inode_lock[DIRTY_META]);
|
|
return 0;
|
|
}
|
|
- fi = list_entry(head->next, struct f2fs_inode_info,
|
|
+ fi = list_first_entry(head, struct f2fs_inode_info,
|
|
gdirty_list);
|
|
inode = igrab(&fi->vfs_inode);
|
|
spin_unlock(&sbi->inode_lock[DIRTY_META]);
|
|
if (inode) {
|
|
- update_inode_page(inode);
|
|
+ sync_inode_metadata(inode, 0);
|
|
+
|
|
+ /* it's on eviction */
|
|
+ if (is_inode_flag_set(inode, FI_DIRTY_INODE))
|
|
+ update_inode_page(inode);
|
|
iput(inode);
|
|
}
|
|
- };
|
|
+ }
|
|
return 0;
|
|
}
|
|
|
|
+static void __prepare_cp_block(struct f2fs_sb_info *sbi)
|
|
+{
|
|
+ struct f2fs_checkpoint *ckpt = F2FS_CKPT(sbi);
|
|
+ struct f2fs_nm_info *nm_i = NM_I(sbi);
|
|
+ nid_t last_nid = nm_i->next_scan_nid;
|
|
+
|
|
+ next_free_nid(sbi, &last_nid);
|
|
+ ckpt->valid_block_count = cpu_to_le64(valid_user_blocks(sbi));
|
|
+ ckpt->valid_node_count = cpu_to_le32(valid_node_count(sbi));
|
|
+ ckpt->valid_inode_count = cpu_to_le32(valid_inode_count(sbi));
|
|
+ ckpt->next_free_nid = cpu_to_le32(last_nid);
|
|
+}
|
|
+
|
|
/*
|
|
* Freeze all the FS-operations for checkpoint.
|
|
*/
|
|
@@ -956,33 +1063,47 @@ retry_flush_dents:
|
|
err = sync_dirty_inodes(sbi, DIR_INODE);
|
|
if (err)
|
|
goto out;
|
|
+ cond_resched();
|
|
goto retry_flush_dents;
|
|
}
|
|
|
|
+ /*
|
|
+ * POR: we should ensure that there are no dirty node pages
|
|
+ * until finishing nat/sit flush. inode->i_blocks can be updated.
|
|
+ */
|
|
+ down_write(&sbi->node_change);
|
|
+
|
|
if (get_pages(sbi, F2FS_DIRTY_IMETA)) {
|
|
+ up_write(&sbi->node_change);
|
|
f2fs_unlock_all(sbi);
|
|
err = f2fs_sync_inode_meta(sbi);
|
|
if (err)
|
|
goto out;
|
|
+ cond_resched();
|
|
goto retry_flush_dents;
|
|
}
|
|
|
|
- /*
|
|
- * POR: we should ensure that there are no dirty node pages
|
|
- * until finishing nat/sit flush.
|
|
- */
|
|
retry_flush_nodes:
|
|
down_write(&sbi->node_write);
|
|
|
|
if (get_pages(sbi, F2FS_DIRTY_NODES)) {
|
|
up_write(&sbi->node_write);
|
|
- err = sync_node_pages(sbi, &wbc);
|
|
+ err = sync_node_pages(sbi, &wbc, false, FS_CP_NODE_IO);
|
|
if (err) {
|
|
+ up_write(&sbi->node_change);
|
|
f2fs_unlock_all(sbi);
|
|
goto out;
|
|
}
|
|
+ cond_resched();
|
|
goto retry_flush_nodes;
|
|
}
|
|
+
|
|
+ /*
|
|
+ * sbi->node_change is used only for AIO write_begin path which produces
|
|
+ * dirty node blocks and some checkpoint values by block allocation.
|
|
+ */
|
|
+ __prepare_cp_block(sbi);
|
|
+ up_write(&sbi->node_change);
|
|
out:
|
|
blk_finish_plug(&plug);
|
|
return err;
|
|
@@ -991,8 +1112,6 @@ out:
|
|
static void unblock_operations(struct f2fs_sb_info *sbi)
|
|
{
|
|
up_write(&sbi->node_write);
|
|
-
|
|
- build_free_nids(sbi);
|
|
f2fs_unlock_all(sbi);
|
|
}
|
|
|
|
@@ -1003,7 +1122,7 @@ static void wait_on_all_pages_writeback(
|
|
for (;;) {
|
|
prepare_to_wait(&sbi->cp_wait, &wait, TASK_UNINTERRUPTIBLE);
|
|
|
|
- if (!atomic_read(&sbi->nr_wb_bios))
|
|
+ if (!get_pages(sbi, F2FS_WB_CP_DATA))
|
|
break;
|
|
|
|
io_schedule_timeout(5*HZ);
|
|
@@ -1015,15 +1134,26 @@ static void update_ckpt_flags(struct f2f
|
|
{
|
|
unsigned long orphan_num = sbi->im[ORPHAN_INO].ino_num;
|
|
struct f2fs_checkpoint *ckpt = F2FS_CKPT(sbi);
|
|
+ unsigned long flags;
|
|
|
|
- spin_lock(&sbi->cp_lock);
|
|
+ spin_lock_irqsave(&sbi->cp_lock, flags);
|
|
|
|
- if (cpc->reason == CP_UMOUNT)
|
|
+ if ((cpc->reason & CP_UMOUNT) &&
|
|
+ le32_to_cpu(ckpt->cp_pack_total_block_count) >
|
|
+ sbi->blocks_per_seg - NM_I(sbi)->nat_bits_blocks)
|
|
+ disable_nat_bits(sbi, false);
|
|
+
|
|
+ if (cpc->reason & CP_TRIMMED)
|
|
+ __set_ckpt_flags(ckpt, CP_TRIMMED_FLAG);
|
|
+ else
|
|
+ __clear_ckpt_flags(ckpt, CP_TRIMMED_FLAG);
|
|
+
|
|
+ if (cpc->reason & CP_UMOUNT)
|
|
__set_ckpt_flags(ckpt, CP_UMOUNT_FLAG);
|
|
else
|
|
__clear_ckpt_flags(ckpt, CP_UMOUNT_FLAG);
|
|
|
|
- if (cpc->reason == CP_FASTBOOT)
|
|
+ if (cpc->reason & CP_FASTBOOT)
|
|
__set_ckpt_flags(ckpt, CP_FASTBOOT_FLAG);
|
|
else
|
|
__clear_ckpt_flags(ckpt, CP_FASTBOOT_FLAG);
|
|
@@ -1038,16 +1168,49 @@ static void update_ckpt_flags(struct f2f
|
|
|
|
/* set this flag to activate crc|cp_ver for recovery */
|
|
__set_ckpt_flags(ckpt, CP_CRC_RECOVERY_FLAG);
|
|
+ __clear_ckpt_flags(ckpt, CP_NOCRC_RECOVERY_FLAG);
|
|
|
|
- spin_unlock(&sbi->cp_lock);
|
|
+ spin_unlock_irqrestore(&sbi->cp_lock, flags);
|
|
+}
|
|
+
|
|
+static void commit_checkpoint(struct f2fs_sb_info *sbi,
|
|
+ void *src, block_t blk_addr)
|
|
+{
|
|
+ struct writeback_control wbc = {
|
|
+ .for_reclaim = 0,
|
|
+ };
|
|
+
|
|
+ /*
|
|
+ * pagevec_lookup_tag and lock_page again will take
|
|
+ * some extra time. Therefore, update_meta_pages and
|
|
+ * sync_meta_pages are combined in this function.
|
|
+ */
|
|
+ struct page *page = grab_meta_page(sbi, blk_addr);
|
|
+ int err;
|
|
+
|
|
+ memcpy(page_address(page), src, PAGE_SIZE);
|
|
+ set_page_dirty(page);
|
|
+
|
|
+ f2fs_wait_on_page_writeback(page, META, true);
|
|
+ f2fs_bug_on(sbi, PageWriteback(page));
|
|
+ if (unlikely(!clear_page_dirty_for_io(page)))
|
|
+ f2fs_bug_on(sbi, 1);
|
|
+
|
|
+ /* writeout cp pack 2 page */
|
|
+ err = __f2fs_write_meta_page(page, &wbc, FS_CP_META_IO);
|
|
+ f2fs_bug_on(sbi, err);
|
|
+
|
|
+ f2fs_put_page(page, 0);
|
|
+
|
|
+ /* submit checkpoint (with barrier if NOBARRIER is not set) */
|
|
+ f2fs_submit_merged_write(sbi, META_FLUSH);
|
|
}
|
|
|
|
static int do_checkpoint(struct f2fs_sb_info *sbi, struct cp_control *cpc)
|
|
{
|
|
struct f2fs_checkpoint *ckpt = F2FS_CKPT(sbi);
|
|
struct f2fs_nm_info *nm_i = NM_I(sbi);
|
|
- unsigned long orphan_num = sbi->im[ORPHAN_INO].ino_num;
|
|
- nid_t last_nid = nm_i->next_scan_nid;
|
|
+ unsigned long orphan_num = sbi->im[ORPHAN_INO].ino_num, flags;
|
|
block_t start_blk;
|
|
unsigned int data_sum_blocks, orphan_blocks;
|
|
__u32 crc32 = 0;
|
|
@@ -1056,22 +1219,20 @@ static int do_checkpoint(struct f2fs_sb_
|
|
struct super_block *sb = sbi->sb;
|
|
struct curseg_info *seg_i = CURSEG_I(sbi, CURSEG_HOT_NODE);
|
|
u64 kbytes_written;
|
|
+ int err;
|
|
|
|
/* Flush all the NAT/SIT pages */
|
|
while (get_pages(sbi, F2FS_DIRTY_META)) {
|
|
- sync_meta_pages(sbi, META, LONG_MAX);
|
|
+ sync_meta_pages(sbi, META, LONG_MAX, FS_CP_META_IO);
|
|
if (unlikely(f2fs_cp_error(sbi)))
|
|
return -EIO;
|
|
}
|
|
|
|
- next_free_nid(sbi, &last_nid);
|
|
-
|
|
/*
|
|
* modify checkpoint
|
|
* version number is already updated
|
|
*/
|
|
ckpt->elapsed_time = cpu_to_le64(get_mtime(sbi));
|
|
- ckpt->valid_block_count = cpu_to_le64(valid_user_blocks(sbi));
|
|
ckpt->free_segment_count = cpu_to_le32(free_segments(sbi));
|
|
for (i = 0; i < NR_CURSEG_NODE_TYPE; i++) {
|
|
ckpt->cur_node_segno[i] =
|
|
@@ -1090,18 +1251,14 @@ static int do_checkpoint(struct f2fs_sb_
|
|
curseg_alloc_type(sbi, i + CURSEG_HOT_DATA);
|
|
}
|
|
|
|
- ckpt->valid_node_count = cpu_to_le32(valid_node_count(sbi));
|
|
- ckpt->valid_inode_count = cpu_to_le32(valid_inode_count(sbi));
|
|
- ckpt->next_free_nid = cpu_to_le32(last_nid);
|
|
-
|
|
/* 2 cp + n data seg summary + orphan inode blocks */
|
|
data_sum_blocks = npages_for_summary_flush(sbi, false);
|
|
- spin_lock(&sbi->cp_lock);
|
|
+ spin_lock_irqsave(&sbi->cp_lock, flags);
|
|
if (data_sum_blocks < NR_CURSEG_DATA_TYPE)
|
|
__set_ckpt_flags(ckpt, CP_COMPACT_SUM_FLAG);
|
|
else
|
|
__clear_ckpt_flags(ckpt, CP_COMPACT_SUM_FLAG);
|
|
- spin_unlock(&sbi->cp_lock);
|
|
+ spin_unlock_irqrestore(&sbi->cp_lock, flags);
|
|
|
|
orphan_blocks = GET_ORPHAN_BLOCKS(orphan_num);
|
|
ckpt->cp_pack_start_sum = cpu_to_le32(1 + cp_payload_blks +
|
|
@@ -1130,10 +1287,26 @@ static int do_checkpoint(struct f2fs_sb_
|
|
|
|
start_blk = __start_cp_next_addr(sbi);
|
|
|
|
- /* need to wait for end_io results */
|
|
- wait_on_all_pages_writeback(sbi);
|
|
- if (unlikely(f2fs_cp_error(sbi)))
|
|
- return -EIO;
|
|
+ /* write nat bits */
|
|
+ if (enabled_nat_bits(sbi, cpc)) {
|
|
+ __u64 cp_ver = cur_cp_version(ckpt);
|
|
+ block_t blk;
|
|
+
|
|
+ cp_ver |= ((__u64)crc32 << 32);
|
|
+ *(__le64 *)nm_i->nat_bits = cpu_to_le64(cp_ver);
|
|
+
|
|
+ blk = start_blk + sbi->blocks_per_seg - nm_i->nat_bits_blocks;
|
|
+ for (i = 0; i < nm_i->nat_bits_blocks; i++)
|
|
+ update_meta_page(sbi, nm_i->nat_bits +
|
|
+ (i << F2FS_BLKSIZE_BITS), blk + i);
|
|
+
|
|
+ /* Flush all the NAT BITS pages */
|
|
+ while (get_pages(sbi, F2FS_DIRTY_META)) {
|
|
+ sync_meta_pages(sbi, META, LONG_MAX, FS_CP_META_IO);
|
|
+ if (unlikely(f2fs_cp_error(sbi)))
|
|
+ return -EIO;
|
|
+ }
|
|
+ }
|
|
|
|
/* write out checkpoint buffer at block 0 */
|
|
update_meta_page(sbi, ckpt, start_blk++);
|
|
@@ -1162,26 +1335,26 @@ static int do_checkpoint(struct f2fs_sb_
|
|
start_blk += NR_CURSEG_NODE_TYPE;
|
|
}
|
|
|
|
- /* writeout checkpoint block */
|
|
- update_meta_page(sbi, ckpt, start_blk);
|
|
+ /* update user_block_counts */
|
|
+ sbi->last_valid_block_count = sbi->total_valid_block_count;
|
|
+ percpu_counter_set(&sbi->alloc_valid_block_count, 0);
|
|
|
|
- /* wait for previous submitted node/meta pages writeback */
|
|
+ /* Here, we have one bio having CP pack except cp pack 2 page */
|
|
+ sync_meta_pages(sbi, META, LONG_MAX, FS_CP_META_IO);
|
|
+
|
|
+ /* wait for previous submitted meta pages writeback */
|
|
wait_on_all_pages_writeback(sbi);
|
|
|
|
if (unlikely(f2fs_cp_error(sbi)))
|
|
return -EIO;
|
|
|
|
- filemap_fdatawait_range(NODE_MAPPING(sbi), 0, LLONG_MAX);
|
|
- filemap_fdatawait_range(META_MAPPING(sbi), 0, LLONG_MAX);
|
|
-
|
|
- /* update user_block_counts */
|
|
- sbi->last_valid_block_count = sbi->total_valid_block_count;
|
|
- percpu_counter_set(&sbi->alloc_valid_block_count, 0);
|
|
-
|
|
- /* Here, we only have one bio having CP pack */
|
|
- sync_meta_pages(sbi, META_FLUSH, LONG_MAX);
|
|
+ /* flush all device cache */
|
|
+ err = f2fs_flush_device_cache(sbi);
|
|
+ if (err)
|
|
+ return err;
|
|
|
|
- /* wait for previous submitted meta pages writeback */
|
|
+ /* barrier and flush checkpoint cp pack 2 page if it can */
|
|
+ commit_checkpoint(sbi, ckpt, start_blk);
|
|
wait_on_all_pages_writeback(sbi);
|
|
|
|
release_ino_entry(sbi, false);
|
|
@@ -1189,7 +1362,6 @@ static int do_checkpoint(struct f2fs_sb_
|
|
if (unlikely(f2fs_cp_error(sbi)))
|
|
return -EIO;
|
|
|
|
- clear_prefree_segments(sbi, cpc);
|
|
clear_sbi_flag(sbi, SBI_IS_DIRTY);
|
|
clear_sbi_flag(sbi, SBI_NEED_CP);
|
|
__set_cp_next_pack(sbi);
|
|
@@ -1219,8 +1391,8 @@ int write_checkpoint(struct f2fs_sb_info
|
|
mutex_lock(&sbi->cp_mutex);
|
|
|
|
if (!is_sbi_flag_set(sbi, SBI_IS_DIRTY) &&
|
|
- (cpc->reason == CP_FASTBOOT || cpc->reason == CP_SYNC ||
|
|
- (cpc->reason == CP_DISCARD && !sbi->discard_blks)))
|
|
+ ((cpc->reason & CP_FASTBOOT) || (cpc->reason & CP_SYNC) ||
|
|
+ ((cpc->reason & CP_DISCARD) && !sbi->discard_blks)))
|
|
goto out;
|
|
if (unlikely(f2fs_cp_error(sbi))) {
|
|
err = -EIO;
|
|
@@ -1239,18 +1411,23 @@ int write_checkpoint(struct f2fs_sb_info
|
|
|
|
trace_f2fs_write_checkpoint(sbi->sb, cpc->reason, "finish block_ops");
|
|
|
|
- f2fs_flush_merged_bios(sbi);
|
|
+ f2fs_flush_merged_writes(sbi);
|
|
|
|
/* this is the case of multiple fstrims without any changes */
|
|
- if (cpc->reason == CP_DISCARD && !is_sbi_flag_set(sbi, SBI_IS_DIRTY)) {
|
|
- f2fs_bug_on(sbi, NM_I(sbi)->dirty_nat_cnt);
|
|
- f2fs_bug_on(sbi, SIT_I(sbi)->dirty_sentries);
|
|
- f2fs_bug_on(sbi, prefree_segments(sbi));
|
|
- flush_sit_entries(sbi, cpc);
|
|
- clear_prefree_segments(sbi, cpc);
|
|
- f2fs_wait_all_discard_bio(sbi);
|
|
- unblock_operations(sbi);
|
|
- goto out;
|
|
+ if (cpc->reason & CP_DISCARD) {
|
|
+ if (!exist_trim_candidates(sbi, cpc)) {
|
|
+ unblock_operations(sbi);
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ if (NM_I(sbi)->dirty_nat_cnt == 0 &&
|
|
+ SIT_I(sbi)->dirty_sentries == 0 &&
|
|
+ prefree_segments(sbi) == 0) {
|
|
+ flush_sit_entries(sbi, cpc);
|
|
+ clear_prefree_segments(sbi, cpc);
|
|
+ unblock_operations(sbi);
|
|
+ goto out;
|
|
+ }
|
|
}
|
|
|
|
/*
|
|
@@ -1262,18 +1439,20 @@ int write_checkpoint(struct f2fs_sb_info
|
|
ckpt->checkpoint_ver = cpu_to_le64(++ckpt_ver);
|
|
|
|
/* write cached NAT/SIT entries to NAT/SIT area */
|
|
- flush_nat_entries(sbi);
|
|
+ flush_nat_entries(sbi, cpc);
|
|
flush_sit_entries(sbi, cpc);
|
|
|
|
/* unlock all the fs_lock[] in do_checkpoint() */
|
|
err = do_checkpoint(sbi, cpc);
|
|
-
|
|
- f2fs_wait_all_discard_bio(sbi);
|
|
+ if (err)
|
|
+ release_discard_addrs(sbi);
|
|
+ else
|
|
+ clear_prefree_segments(sbi, cpc);
|
|
|
|
unblock_operations(sbi);
|
|
stat_inc_cp_count(sbi->stat_info);
|
|
|
|
- if (cpc->reason == CP_RECOVERY)
|
|
+ if (cpc->reason & CP_RECOVERY)
|
|
f2fs_msg(sbi->sb, KERN_NOTICE,
|
|
"checkpoint: version = %llx", ckpt_ver);
|
|
|