Update bcachefs sources to d82da7126f fixup! bcachefs: for_each_btree_key2()

This commit is contained in:
Kent Overstreet 2022-07-17 05:20:47 -04:00
parent 2cea78f5cc
commit 79d39bd7ce
10 changed files with 146 additions and 124 deletions

View File

@ -1 +1 @@
2f11bb05b0df04b7e0e190fd27b111e9f20cd749 d82da7126f2db01a0d320ad7ed13cd4016c36221

View File

@ -81,6 +81,11 @@ static inline u64 local_clock(void)
return sched_clock(); return sched_clock();
} }
static inline u64 ktime_get_ns(void)
{
return sched_clock();
}
#define jiffies nsecs_to_jiffies(sched_clock()) #define jiffies nsecs_to_jiffies(sched_clock())
#endif #endif

View File

@ -1007,16 +1007,10 @@ int bch2_check_alloc_to_lru_refs(struct bch_fs *c)
bch2_trans_init(&trans, c, 0, 0); bch2_trans_init(&trans, c, 0, 0);
for_each_btree_key(&trans, iter, BTREE_ID_alloc, POS_MIN, for_each_btree_key_commit(&trans, iter, BTREE_ID_alloc,
BTREE_ITER_PREFETCH, k, ret) { POS_MIN, BTREE_ITER_PREFETCH, k,
ret = commit_do(&trans, NULL, NULL, NULL, NULL, BTREE_INSERT_NOFAIL|BTREE_INSERT_LAZY_RW,
BTREE_INSERT_NOFAIL| bch2_check_alloc_to_lru_ref(&trans, &iter));
BTREE_INSERT_LAZY_RW,
bch2_check_alloc_to_lru_ref(&trans, &iter));
if (ret)
break;
}
bch2_trans_iter_exit(&trans, &iter);
bch2_trans_exit(&trans); bch2_trans_exit(&trans);
return ret < 0 ? ret : 0; return ret < 0 ? ret : 0;

View File

@ -1848,10 +1848,15 @@ out:
return ret; return ret;
} }
static bool gc_btree_gens_key(struct bch_fs *c, struct bkey_s_c k) static int gc_btree_gens_key(struct btree_trans *trans,
struct btree_iter *iter,
struct bkey_s_c k)
{ {
struct bch_fs *c = trans->c;
struct bkey_ptrs_c ptrs = bch2_bkey_ptrs_c(k); struct bkey_ptrs_c ptrs = bch2_bkey_ptrs_c(k);
const struct bch_extent_ptr *ptr; const struct bch_extent_ptr *ptr;
struct bkey_i *u;
int ret;
percpu_down_read(&c->mark_lock); percpu_down_read(&c->mark_lock);
bkey_for_each_ptr(ptrs, ptr) { bkey_for_each_ptr(ptrs, ptr) {
@ -1859,7 +1864,7 @@ static bool gc_btree_gens_key(struct bch_fs *c, struct bkey_s_c k)
if (ptr_stale(ca, ptr) > 16) { if (ptr_stale(ca, ptr) > 16) {
percpu_up_read(&c->mark_lock); percpu_up_read(&c->mark_lock);
return true; goto update;
} }
} }
@ -1871,77 +1876,27 @@ static bool gc_btree_gens_key(struct bch_fs *c, struct bkey_s_c k)
*gen = ptr->gen; *gen = ptr->gen;
} }
percpu_up_read(&c->mark_lock); percpu_up_read(&c->mark_lock);
return 0;
update:
u = bch2_trans_kmalloc(trans, bkey_bytes(k.k));
ret = PTR_ERR_OR_ZERO(u);
if (ret)
return ret;
return false; bkey_reassemble(u, k);
bch2_extent_normalize(c, bkey_i_to_s(u));
return bch2_trans_update(trans, iter, u, 0);
} }
/* static int bch2_alloc_write_oldest_gen(struct btree_trans *trans, struct btree_iter *iter,
* For recalculating oldest gen, we only need to walk keys in leaf nodes; btree struct bkey_s_c k)
* node pointers currently never have cached pointers that can become stale:
*/
static int bch2_gc_btree_gens(struct btree_trans *trans, enum btree_id btree_id)
{
struct bch_fs *c = trans->c;
struct btree_iter iter;
struct bkey_s_c k;
struct bkey_buf sk;
int ret = 0, commit_err = 0;
bch2_bkey_buf_init(&sk);
bch2_trans_iter_init(trans, &iter, btree_id, POS_MIN,
BTREE_ITER_PREFETCH|
BTREE_ITER_NOT_EXTENTS|
BTREE_ITER_ALL_SNAPSHOTS);
while ((bch2_trans_begin(trans),
k = bch2_btree_iter_peek(&iter)).k) {
ret = bkey_err(k);
if (ret == -EINTR)
continue;
if (ret)
break;
c->gc_gens_pos = iter.pos;
if (gc_btree_gens_key(c, k) && !commit_err) {
bch2_bkey_buf_reassemble(&sk, c, k);
bch2_extent_normalize(c, bkey_i_to_s(sk.k));
commit_err =
bch2_trans_update(trans, &iter, sk.k, 0) ?:
bch2_trans_commit(trans, NULL, NULL,
BTREE_INSERT_NOWAIT|
BTREE_INSERT_NOFAIL);
if (commit_err == -EINTR) {
commit_err = 0;
continue;
}
}
bch2_btree_iter_advance(&iter);
}
bch2_trans_iter_exit(trans, &iter);
bch2_bkey_buf_exit(&sk, c);
return ret;
}
static int bch2_alloc_write_oldest_gen(struct btree_trans *trans, struct btree_iter *iter)
{ {
struct bch_dev *ca = bch_dev_bkey_exists(trans->c, iter->pos.inode); struct bch_dev *ca = bch_dev_bkey_exists(trans->c, iter->pos.inode);
struct bkey_s_c k;
struct bch_alloc_v4 a; struct bch_alloc_v4 a;
struct bkey_i_alloc_v4 *a_mut; struct bkey_i_alloc_v4 *a_mut;
int ret; int ret;
k = bch2_btree_iter_peek_slot(iter);
ret = bkey_err(k);
if (ret)
return ret;
bch2_alloc_to_v4(k, &a); bch2_alloc_to_v4(k, &a);
if (a.oldest_gen == ca->oldest_gen[iter->pos.offset]) if (a.oldest_gen == ca->oldest_gen[iter->pos.offset])
@ -2001,26 +1956,35 @@ int bch2_gc_gens(struct bch_fs *c)
for (i = 0; i < BTREE_ID_NR; i++) for (i = 0; i < BTREE_ID_NR; i++)
if ((1 << i) & BTREE_ID_HAS_PTRS) { if ((1 << i) & BTREE_ID_HAS_PTRS) {
struct btree_iter iter;
struct bkey_s_c k;
c->gc_gens_btree = i; c->gc_gens_btree = i;
c->gc_gens_pos = POS_MIN; c->gc_gens_pos = POS_MIN;
ret = bch2_gc_btree_gens(&trans, i); ret = for_each_btree_key_commit(&trans, iter, i,
POS_MIN,
BTREE_ITER_PREFETCH|BTREE_ITER_ALL_SNAPSHOTS,
k,
NULL, NULL,
BTREE_INSERT_NOFAIL,
gc_btree_gens_key(&trans, &iter, k));
if (ret) { if (ret) {
bch_err(c, "error recalculating oldest_gen: %i", ret); bch_err(c, "error recalculating oldest_gen: %i", ret);
goto err; goto err;
} }
} }
for_each_btree_key(&trans, iter, BTREE_ID_alloc, POS_MIN, ret = for_each_btree_key_commit(&trans, iter, BTREE_ID_alloc,
BTREE_ITER_PREFETCH, k, ret) { POS_MIN,
ret = commit_do(&trans, NULL, NULL, BTREE_ITER_PREFETCH,
BTREE_INSERT_NOFAIL, k,
bch2_alloc_write_oldest_gen(&trans, &iter)); NULL, NULL,
if (ret) { BTREE_INSERT_NOFAIL,
bch_err(c, "error writing oldest_gen: %i", ret); bch2_alloc_write_oldest_gen(&trans, &iter, k));
break; if (ret) {
} bch_err(c, "error writing oldest_gen: %i", ret);
goto err;
} }
bch2_trans_iter_exit(&trans, &iter);
c->gc_gens_btree = 0; c->gc_gens_btree = 0;
c->gc_gens_pos = POS_MIN; c->gc_gens_pos = POS_MIN;

View File

@ -1663,6 +1663,9 @@ out:
int __must_check bch2_btree_path_traverse(struct btree_trans *trans, int __must_check bch2_btree_path_traverse(struct btree_trans *trans,
struct btree_path *path, unsigned flags) struct btree_path *path, unsigned flags)
{ {
if (!(local_clock() % 128))
return btree_trans_restart(trans);
if (path->uptodate < BTREE_ITER_NEED_RELOCK) if (path->uptodate < BTREE_ITER_NEED_RELOCK)
return 0; return 0;
@ -3242,12 +3245,19 @@ void bch2_trans_begin(struct btree_trans *trans)
path->preserve = false; path->preserve = false;
} }
bch2_trans_cond_resched(trans); if (!trans->restarted &&
(need_resched() ||
ktime_get_ns() - trans->last_begin_time > BTREE_TRANS_MAX_LOCK_HOLD_TIME_NS)) {
bch2_trans_unlock(trans);
cond_resched();
bch2_trans_relock(trans);
}
if (trans->restarted) if (trans->restarted)
bch2_btree_path_traverse_all(trans); bch2_btree_path_traverse_all(trans);
trans->restarted = false; trans->restarted = false;
trans->last_begin_time = ktime_get_ns();
} }
static void bch2_trans_alloc_paths(struct btree_trans *trans, struct bch_fs *c) static void bch2_trans_alloc_paths(struct btree_trans *trans, struct bch_fs *c)
@ -3281,6 +3291,7 @@ void __bch2_trans_init(struct btree_trans *trans, struct bch_fs *c,
memset(trans, 0, sizeof(*trans)); memset(trans, 0, sizeof(*trans));
trans->c = c; trans->c = c;
trans->fn = fn; trans->fn = fn;
trans->last_begin_time = ktime_get_ns();
trans->task = current; trans->task = current;
bch2_trans_alloc_paths(trans, c); bch2_trans_alloc_paths(trans, c);

View File

@ -367,8 +367,10 @@ __bch2_btree_iter_peek_and_restart(struct btree_trans *trans,
do { \ do { \
bch2_trans_begin(_trans); \ bch2_trans_begin(_trans); \
(_k) = bch2_btree_iter_peek_type(&(_iter), (_flags)); \ (_k) = bch2_btree_iter_peek_type(&(_iter), (_flags)); \
if (!(_k).k) \ if (!(_k).k) { \
_ret = 0; \
break; \ break; \
} \
\ \
_ret = bkey_err(_k) ?: (_do); \ _ret = bkey_err(_k) ?: (_do); \
if (!_ret) \ if (!_ret) \

View File

@ -384,10 +384,13 @@ struct btree_trans_commit_hook {
#define BTREE_TRANS_MEM_MAX (1U << 16) #define BTREE_TRANS_MEM_MAX (1U << 16)
#define BTREE_TRANS_MAX_LOCK_HOLD_TIME_NS 10000
struct btree_trans { struct btree_trans {
struct bch_fs *c; struct bch_fs *c;
const char *fn; const char *fn;
struct list_head list; struct list_head list;
u64 last_begin_time;
struct btree *locking; struct btree *locking;
unsigned locking_path_idx; unsigned locking_path_idx;
struct bpos locking_pos; struct bpos locking_pos;

View File

@ -7,6 +7,7 @@ enum {
OPEN_BUCKETS_EMPTY = 2048, OPEN_BUCKETS_EMPTY = 2048,
FREELIST_EMPTY, /* Allocator thread not keeping up */ FREELIST_EMPTY, /* Allocator thread not keeping up */
INSUFFICIENT_DEVICES, INSUFFICIENT_DEVICES,
NEED_SNAPSHOT_CLEANUP,
}; };
#endif /* _BCACHFES_ERRCODE_H */ #endif /* _BCACHFES_ERRCODE_H */

View File

@ -489,6 +489,28 @@ static inline void snapshots_seen_init(struct snapshots_seen *s)
memset(s, 0, sizeof(*s)); memset(s, 0, sizeof(*s));
} }
static int snapshots_seen_add(struct bch_fs *c, struct snapshots_seen *s, u32 id)
{
struct snapshots_seen_entry *i, n = { id, id };
int ret;
darray_for_each(s->ids, i) {
if (n.equiv < i->equiv)
break;
if (i->equiv == n.equiv) {
bch_err(c, "adding duplicate snapshot in snapshots_seen_add()");
return -EINVAL;
}
}
ret = darray_insert_item(&s->ids, i - s->ids.data, n);
if (ret)
bch_err(c, "error reallocating snapshots_seen table (size %zu)",
s->ids.size);
return ret;
}
static int snapshots_seen_update(struct bch_fs *c, struct snapshots_seen *s, static int snapshots_seen_update(struct bch_fs *c, struct snapshots_seen *s,
enum btree_id btree_id, struct bpos pos) enum btree_id btree_id, struct bpos pos)
{ {
@ -512,7 +534,7 @@ static int snapshots_seen_update(struct bch_fs *c, struct snapshots_seen *s,
bch2_btree_ids[btree_id], bch2_btree_ids[btree_id],
pos.inode, pos.offset, pos.inode, pos.offset,
i->id, n.id, n.equiv); i->id, n.id, n.equiv);
return -EINVAL; return -NEED_SNAPSHOT_CLEANUP;
} }
return 0; return 0;
@ -954,7 +976,7 @@ static int check_inode(struct btree_trans *trans,
} }
if (do_update) { if (do_update) {
ret = write_inode(trans, &u, iter->pos.snapshot); ret = __write_inode(trans, &u, iter->pos.snapshot);
if (ret) if (ret)
bch_err(c, "error in fsck: error %i " bch_err(c, "error in fsck: error %i "
"updating inode", ret); "updating inode", ret);
@ -1216,20 +1238,38 @@ static int check_extent(struct btree_trans *trans, struct btree_iter *iter,
goto out; goto out;
} }
if (!bch2_snapshot_internal_node(c, equiv.snapshot)) { /*
for_each_visible_inode(c, s, inode, equiv.snapshot, i) { * Check inodes in reverse order, from oldest snapshots to newest, so
if (fsck_err_on(!(i->inode.bi_flags & BCH_INODE_I_SIZE_DIRTY) && * that we emit the fewest number of whiteouts necessary:
k.k->type != KEY_TYPE_reservation && */
k.k->p.offset > round_up(i->inode.bi_size, block_bytes(c)) >> 9, c, for (i = inode->inodes.data + inode->inodes.nr - 1;
"extent type %u offset %llu past end of inode %llu, i_size %llu", i >= inode->inodes.data;
k.k->type, k.k->p.offset, k.k->p.inode, i->inode.bi_size)) { --i) {
bch2_fs_lazy_rw(c); if (i->snapshot > equiv.snapshot ||
ret = bch2_btree_delete_range_trans(trans, BTREE_ID_extents, !key_visible_in_snapshot(c, s, i->snapshot, equiv.snapshot))
SPOS(k.k->p.inode, round_up(i->inode.bi_size, block_bytes(c)) >> 9, continue;
equiv.snapshot),
POS(k.k->p.inode, U64_MAX), if (fsck_err_on(!(i->inode.bi_flags & BCH_INODE_I_SIZE_DIRTY) &&
0, NULL) ?: -EINTR; k.k->type != KEY_TYPE_reservation &&
goto out; k.k->p.offset > round_up(i->inode.bi_size, block_bytes(c)) >> 9, c,
"extent type past end of inode %llu:%u, i_size %llu\n %s",
i->inode.bi_inum, i->snapshot, i->inode.bi_size,
(bch2_bkey_val_to_text(&buf, c, k), buf.buf))) {
struct btree_iter iter2;
bch2_trans_copy_iter(&iter2, iter);
bch2_btree_iter_set_snapshot(&iter2, i->snapshot);
ret = bch2_btree_iter_traverse(&iter2) ?:
bch2_btree_delete_at(trans, &iter2,
BTREE_UPDATE_INTERNAL_SNAPSHOT_NODE);
bch2_trans_iter_exit(trans, &iter2);
if (ret)
goto err;
if (i->snapshot != equiv.snapshot) {
ret = snapshots_seen_add(c, s, i->snapshot);
if (ret)
goto err;
} }
} }
} }
@ -2140,7 +2180,7 @@ static int check_nlinks_walk_dirents(struct bch_fs *c, struct nlink_table *links
d.v->d_type != DT_SUBVOL) d.v->d_type != DT_SUBVOL)
inc_link(c, &s, links, range_start, range_end, inc_link(c, &s, links, range_start, range_end,
le64_to_cpu(d.v->d_inum), le64_to_cpu(d.v->d_inum),
d.k->p.snapshot); bch2_snapshot_equiv(c, d.k->p.snapshot));
break; break;
} }
} }
@ -2326,7 +2366,9 @@ static int fix_reflink_p(struct bch_fs *c)
*/ */
int bch2_fsck_full(struct bch_fs *c) int bch2_fsck_full(struct bch_fs *c)
{ {
return bch2_fs_check_snapshots(c) ?: int ret;
again:
ret = bch2_fs_check_snapshots(c) ?:
bch2_fs_check_subvols(c) ?: bch2_fs_check_subvols(c) ?:
bch2_delete_dead_snapshots(c) ?: bch2_delete_dead_snapshots(c) ?:
check_inodes(c, true) ?: check_inodes(c, true) ?:
@ -2337,6 +2379,13 @@ int bch2_fsck_full(struct bch_fs *c)
check_directory_structure(c) ?: check_directory_structure(c) ?:
check_nlinks(c) ?: check_nlinks(c) ?:
fix_reflink_p(c); fix_reflink_p(c);
if (ret == -NEED_SNAPSHOT_CLEANUP) {
set_bit(BCH_FS_HAVE_DELETED_SNAPSHOTS, &c->flags);
goto again;
}
return ret;
} }
int bch2_fsck_walk_inodes_only(struct bch_fs *c) int bch2_fsck_walk_inodes_only(struct bch_fs *c)

View File

@ -455,22 +455,14 @@ static void bch2_sb_quota_read(struct bch_fs *c)
} }
static int bch2_fs_quota_read_inode(struct btree_trans *trans, static int bch2_fs_quota_read_inode(struct btree_trans *trans,
struct btree_iter *iter) struct btree_iter *iter,
struct bkey_s_c k)
{ {
struct bch_fs *c = trans->c; struct bch_fs *c = trans->c;
struct bch_inode_unpacked u; struct bch_inode_unpacked u;
struct bch_subvolume subvolume; struct bch_subvolume subvolume;
struct bkey_s_c k;
int ret; int ret;
k = bch2_btree_iter_peek(iter);
ret = bkey_err(k);
if (ret)
return ret;
if (!k.k)
return 1;
ret = bch2_snapshot_get_subvol(trans, k.k->p.snapshot, &subvolume); ret = bch2_snapshot_get_subvol(trans, k.k->p.snapshot, &subvolume);
if (ret) if (ret)
return ret; return ret;
@ -503,6 +495,7 @@ int bch2_fs_quota_read(struct bch_fs *c)
struct bch_memquota_type *q; struct bch_memquota_type *q;
struct btree_trans trans; struct btree_trans trans;
struct btree_iter iter; struct btree_iter iter;
struct bkey_s_c k;
int ret; int ret;
mutex_lock(&c->sb_lock); mutex_lock(&c->sb_lock);
@ -517,18 +510,18 @@ int bch2_fs_quota_read(struct bch_fs *c)
bch2_trans_init(&trans, c, 0, 0); bch2_trans_init(&trans, c, 0, 0);
bch2_trans_iter_init(&trans, &iter, BTREE_ID_inodes, POS_MIN, ret = for_each_btree_key2(&trans, iter, BTREE_ID_inodes,
POS_MIN,
BTREE_ITER_INTENT| BTREE_ITER_INTENT|
BTREE_ITER_PREFETCH| BTREE_ITER_PREFETCH|
BTREE_ITER_ALL_SNAPSHOTS); BTREE_ITER_ALL_SNAPSHOTS,
do { k,
ret = lockrestart_do(&trans, bch2_fs_quota_read_inode(&trans, &iter, k));
bch2_fs_quota_read_inode(&trans, &iter)); if (ret)
} while (!ret); bch_err(c, "err reading inodes in quota init: %i", ret);
bch2_trans_iter_exit(&trans, &iter);
bch2_trans_exit(&trans); bch2_trans_exit(&trans);
return ret < 0 ? ret : 0; return ret;
} }
/* Enable/disable/delete quotas for an entire filesystem: */ /* Enable/disable/delete quotas for an entire filesystem: */