summaryrefslogtreecommitdiff
path: root/fs/bcachefs/btree_cache.h
blob: f7b9bcfe09a3a1ab621092b138b1b8dc9af9de3a (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _BCACHEFS_BTREE_CACHE_H
#define _BCACHEFS_BTREE_CACHE_H

#include "bcachefs.h"
#include "btree_types.h"
#include "extents.h"

struct btree_iter;

extern const char * const bch2_btree_ids[];

void bch2_recalc_btree_reserve(struct bch_fs *);

void bch2_btree_node_hash_remove(struct btree_cache *, struct btree *);
int __bch2_btree_node_hash_insert(struct btree_cache *, struct btree *);
int bch2_btree_node_hash_insert(struct btree_cache *, struct btree *,
				unsigned, enum btree_id);

void bch2_btree_cache_cannibalize_unlock(struct bch_fs *);
int bch2_btree_cache_cannibalize_lock(struct bch_fs *, struct closure *);

struct btree *bch2_btree_node_mem_alloc(struct bch_fs *);

struct btree *bch2_btree_node_get(struct bch_fs *, struct btree_iter *,
				  const struct bkey_i *, unsigned,
				  enum six_lock_type, bool);

struct btree *bch2_btree_node_get_sibling(struct bch_fs *, struct btree_iter *,
					  struct btree *, bool,
					  enum btree_node_sibling);

void bch2_btree_node_prefetch(struct bch_fs *, struct btree_iter *,
			      const struct bkey_i *, unsigned);

void bch2_fs_btree_cache_exit(struct bch_fs *);
int bch2_fs_btree_cache_init(struct bch_fs *);
void bch2_fs_btree_cache_init_early(struct btree_cache *);

#define PTR_HASH(_k)	(bkey_i_to_extent_c(_k)->v._data[0])

/* is btree node in hash table? */
static inline bool btree_node_hashed(struct btree *b)
{
	return bkey_extent_is_data(&b->key.k) && PTR_HASH(&b->key);
}

#define for_each_cached_btree(_b, _c, _tbl, _iter, _pos)		\
	for ((_tbl) = rht_dereference_rcu((_c)->btree_cache.table.tbl,	\
					  &(_c)->btree_cache.table),	\
	     _iter = 0;	_iter < (_tbl)->size; _iter++)			\
		rht_for_each_entry_rcu((_b), (_pos), _tbl, _iter, hash)

static inline size_t btree_bytes(struct bch_fs *c)
{
	return c->opts.btree_node_size << 9;
}

static inline size_t btree_max_u64s(struct bch_fs *c)
{
	return (btree_bytes(c) - sizeof(struct btree_node)) / sizeof(u64);
}

static inline size_t btree_page_order(struct bch_fs *c)
{
	return get_order(btree_bytes(c));
}

static inline size_t btree_pages(struct bch_fs *c)
{
	return 1 << btree_page_order(c);
}

static inline unsigned btree_blocks(struct bch_fs *c)
{
	return c->opts.btree_node_size >> c->block_bits;
}

#define BTREE_SPLIT_THRESHOLD(c)		(btree_blocks(c) * 3 / 4)

#define BTREE_FOREGROUND_MERGE_THRESHOLD(c)	(btree_max_u64s(c) * 1 / 3)
#define BTREE_FOREGROUND_MERGE_HYSTERESIS(c)			\
	(BTREE_FOREGROUND_MERGE_THRESHOLD(c) +			\
	 (BTREE_FOREGROUND_MERGE_THRESHOLD(c) << 2))

#define btree_node_root(_c, _b)	((_c)->btree_roots[(_b)->btree_id].b)

int bch2_print_btree_node(struct bch_fs *, struct btree *,
			 char *, size_t);

#endif /* _BCACHEFS_BTREE_CACHE_H */