| /* SPDX-License-Identifier: GPL-2.0 */ |
| #ifndef _BCACHEFS_BKEY_H |
| #define _BCACHEFS_BKEY_H |
| |
| #include <linux/bug.h> |
| #include "bcachefs_format.h" |
| |
| #include "util.h" |
| #include "vstructs.h" |
| |
| #if 0 |
| |
| /* |
| * compiled unpack functions are disabled, pending a new interface for |
| * dynamically allocating executable memory: |
| */ |
| |
| #ifdef CONFIG_X86_64 |
| #define HAVE_BCACHEFS_COMPILED_UNPACK 1 |
| #endif |
| #endif |
| |
| void bch2_to_binary(char *, const u64 *, unsigned); |
| |
| /* bkey with split value, const */ |
| struct bkey_s_c { |
| const struct bkey *k; |
| const struct bch_val *v; |
| }; |
| |
| /* bkey with split value */ |
| struct bkey_s { |
| union { |
| struct { |
| struct bkey *k; |
| struct bch_val *v; |
| }; |
| struct bkey_s_c s_c; |
| }; |
| }; |
| |
| #define bkey_next(_k) vstruct_next(_k) |
| |
| static inline unsigned bkey_val_u64s(const struct bkey *k) |
| { |
| return k->u64s - BKEY_U64s; |
| } |
| |
| static inline size_t bkey_val_bytes(const struct bkey *k) |
| { |
| return bkey_val_u64s(k) * sizeof(u64); |
| } |
| |
| static inline void set_bkey_val_u64s(struct bkey *k, unsigned val_u64s) |
| { |
| k->u64s = BKEY_U64s + val_u64s; |
| } |
| |
| static inline void set_bkey_val_bytes(struct bkey *k, unsigned bytes) |
| { |
| k->u64s = BKEY_U64s + DIV_ROUND_UP(bytes, sizeof(u64)); |
| } |
| |
| #define bkey_deleted(_k) ((_k)->type == KEY_TYPE_DELETED) |
| |
| #define bkey_whiteout(_k) \ |
| ((_k)->type == KEY_TYPE_DELETED || (_k)->type == KEY_TYPE_DISCARD) |
| |
| #define bkey_packed_typecheck(_k) \ |
| ({ \ |
| BUILD_BUG_ON(!type_is(_k, struct bkey *) && \ |
| !type_is(_k, struct bkey_packed *)); \ |
| type_is(_k, struct bkey_packed *); \ |
| }) |
| |
| enum bkey_lr_packed { |
| BKEY_PACKED_BOTH, |
| BKEY_PACKED_RIGHT, |
| BKEY_PACKED_LEFT, |
| BKEY_PACKED_NONE, |
| }; |
| |
| #define bkey_lr_packed_typecheck(_l, _r) \ |
| (!bkey_packed_typecheck(_l) + ((!bkey_packed_typecheck(_r)) << 1)) |
| |
| #define bkey_lr_packed(_l, _r) \ |
| ((_l)->format + ((_r)->format << 1)) |
| |
| #define bkey_copy(_dst, _src) \ |
| do { \ |
| BUILD_BUG_ON(!type_is(_dst, struct bkey_i *) && \ |
| !type_is(_dst, struct bkey_packed *)); \ |
| BUILD_BUG_ON(!type_is(_src, struct bkey_i *) && \ |
| !type_is(_src, struct bkey_packed *)); \ |
| EBUG_ON((u64 *) (_dst) > (u64 *) (_src) && \ |
| (u64 *) (_dst) < (u64 *) (_src) + \ |
| ((struct bkey *) (_src))->u64s); \ |
| \ |
| __memmove_u64s_down((_dst), (_src), \ |
| ((struct bkey *) (_src))->u64s); \ |
| } while (0) |
| |
| struct btree; |
| |
| struct bkey_format_state { |
| u64 field_min[BKEY_NR_FIELDS]; |
| u64 field_max[BKEY_NR_FIELDS]; |
| }; |
| |
| void bch2_bkey_format_init(struct bkey_format_state *); |
| void bch2_bkey_format_add_key(struct bkey_format_state *, const struct bkey *); |
| void bch2_bkey_format_add_pos(struct bkey_format_state *, struct bpos); |
| struct bkey_format bch2_bkey_format_done(struct bkey_format_state *); |
| const char *bch2_bkey_format_validate(struct bkey_format *); |
| |
| __pure |
| unsigned bch2_bkey_greatest_differing_bit(const struct btree *, |
| const struct bkey_packed *, |
| const struct bkey_packed *); |
| __pure |
| unsigned bch2_bkey_ffs(const struct btree *, const struct bkey_packed *); |
| |
| __pure |
| int __bch2_bkey_cmp_packed_format_checked(const struct bkey_packed *, |
| const struct bkey_packed *, |
| const struct btree *); |
| |
| __pure |
| int __bch2_bkey_cmp_left_packed_format_checked(const struct btree *, |
| const struct bkey_packed *, |
| const struct bpos *); |
| |
| __pure |
| int __bch2_bkey_cmp_packed(const struct bkey_packed *, |
| const struct bkey_packed *, |
| const struct btree *); |
| |
| __pure |
| int __bch2_bkey_cmp_left_packed(const struct btree *, |
| const struct bkey_packed *, |
| const struct bpos *); |
| |
| static inline __pure |
| int bkey_cmp_left_packed(const struct btree *b, |
| const struct bkey_packed *l, const struct bpos *r) |
| { |
| return __bch2_bkey_cmp_left_packed(b, l, r); |
| } |
| |
| /* |
| * we prefer to pass bpos by ref, but it's often enough terribly convenient to |
| * pass it by by val... as much as I hate c++, const ref would be nice here: |
| */ |
| __pure __flatten |
| static inline int bkey_cmp_left_packed_byval(const struct btree *b, |
| const struct bkey_packed *l, |
| struct bpos r) |
| { |
| return bkey_cmp_left_packed(b, l, &r); |
| } |
| |
| /* |
| * If @_l or @_r are struct bkey * (not bkey_packed *), uses type information to |
| * skip dispatching on k->format: |
| */ |
| #define bkey_cmp_packed(_b, _l, _r) \ |
| ({ \ |
| int _cmp; \ |
| \ |
| switch (bkey_lr_packed_typecheck(_l, _r)) { \ |
| case BKEY_PACKED_NONE: \ |
| _cmp = bkey_cmp(((struct bkey *) (_l))->p, \ |
| ((struct bkey *) (_r))->p); \ |
| break; \ |
| case BKEY_PACKED_LEFT: \ |
| _cmp = bkey_cmp_left_packed((_b), \ |
| (struct bkey_packed *) (_l), \ |
| &((struct bkey *) (_r))->p); \ |
| break; \ |
| case BKEY_PACKED_RIGHT: \ |
| _cmp = -bkey_cmp_left_packed((_b), \ |
| (struct bkey_packed *) (_r), \ |
| &((struct bkey *) (_l))->p); \ |
| break; \ |
| case BKEY_PACKED_BOTH: \ |
| _cmp = __bch2_bkey_cmp_packed((void *) (_l), \ |
| (void *) (_r), (_b)); \ |
| break; \ |
| } \ |
| _cmp; \ |
| }) |
| |
| #if 1 |
| static __always_inline int bkey_cmp(struct bpos l, struct bpos r) |
| { |
| if (l.inode != r.inode) |
| return l.inode < r.inode ? -1 : 1; |
| if (l.offset != r.offset) |
| return l.offset < r.offset ? -1 : 1; |
| if (l.snapshot != r.snapshot) |
| return l.snapshot < r.snapshot ? -1 : 1; |
| return 0; |
| } |
| #else |
| int bkey_cmp(struct bpos l, struct bpos r); |
| #endif |
| |
| static inline struct bpos bpos_min(struct bpos l, struct bpos r) |
| { |
| return bkey_cmp(l, r) < 0 ? l : r; |
| } |
| |
| void bch2_bpos_swab(struct bpos *); |
| void bch2_bkey_swab_key(const struct bkey_format *, struct bkey_packed *); |
| |
| static __always_inline int bversion_cmp(struct bversion l, struct bversion r) |
| { |
| return (l.hi > r.hi) - (l.hi < r.hi) ?: |
| (l.lo > r.lo) - (l.lo < r.lo); |
| } |
| |
| #define ZERO_VERSION ((struct bversion) { .hi = 0, .lo = 0 }) |
| #define MAX_VERSION ((struct bversion) { .hi = ~0, .lo = ~0ULL }) |
| |
| static __always_inline int bversion_zero(struct bversion v) |
| { |
| return !bversion_cmp(v, ZERO_VERSION); |
| } |
| |
| #ifdef CONFIG_BCACHEFS_DEBUG |
| /* statement expressions confusing unlikely()? */ |
| #define bkey_packed(_k) \ |
| ({ EBUG_ON((_k)->format > KEY_FORMAT_CURRENT); \ |
| (_k)->format != KEY_FORMAT_CURRENT; }) |
| #else |
| #define bkey_packed(_k) ((_k)->format != KEY_FORMAT_CURRENT) |
| #endif |
| |
| /* |
| * It's safe to treat an unpacked bkey as a packed one, but not the reverse |
| */ |
| static inline struct bkey_packed *bkey_to_packed(struct bkey_i *k) |
| { |
| return (struct bkey_packed *) k; |
| } |
| |
| static inline const struct bkey_packed *bkey_to_packed_c(const struct bkey_i *k) |
| { |
| return (const struct bkey_packed *) k; |
| } |
| |
| static inline struct bkey_i *packed_to_bkey(struct bkey_packed *k) |
| { |
| return bkey_packed(k) ? NULL : (struct bkey_i *) k; |
| } |
| |
| static inline const struct bkey *packed_to_bkey_c(const struct bkey_packed *k) |
| { |
| return bkey_packed(k) ? NULL : (const struct bkey *) k; |
| } |
| |
| static inline unsigned bkey_format_key_bits(const struct bkey_format *format) |
| { |
| return format->bits_per_field[BKEY_FIELD_INODE] + |
| format->bits_per_field[BKEY_FIELD_OFFSET] + |
| format->bits_per_field[BKEY_FIELD_SNAPSHOT]; |
| } |
| |
| static inline struct bpos bkey_successor(struct bpos p) |
| { |
| struct bpos ret = p; |
| |
| if (!++ret.offset) |
| BUG_ON(!++ret.inode); |
| |
| return ret; |
| } |
| |
| static inline struct bpos bkey_predecessor(struct bpos p) |
| { |
| struct bpos ret = p; |
| |
| if (!ret.offset--) |
| BUG_ON(!ret.inode--); |
| |
| return ret; |
| } |
| |
| static inline u64 bkey_start_offset(const struct bkey *k) |
| { |
| return k->p.offset - k->size; |
| } |
| |
| static inline struct bpos bkey_start_pos(const struct bkey *k) |
| { |
| return (struct bpos) { |
| .inode = k->p.inode, |
| .offset = bkey_start_offset(k), |
| .snapshot = k->p.snapshot, |
| }; |
| } |
| |
| /* Packed helpers */ |
| |
| static inline unsigned bkeyp_key_u64s(const struct bkey_format *format, |
| const struct bkey_packed *k) |
| { |
| unsigned ret = bkey_packed(k) ? format->key_u64s : BKEY_U64s; |
| |
| EBUG_ON(k->u64s < ret); |
| return ret; |
| } |
| |
| static inline unsigned bkeyp_key_bytes(const struct bkey_format *format, |
| const struct bkey_packed *k) |
| { |
| return bkeyp_key_u64s(format, k) * sizeof(u64); |
| } |
| |
| static inline unsigned bkeyp_val_u64s(const struct bkey_format *format, |
| const struct bkey_packed *k) |
| { |
| return k->u64s - bkeyp_key_u64s(format, k); |
| } |
| |
| static inline size_t bkeyp_val_bytes(const struct bkey_format *format, |
| const struct bkey_packed *k) |
| { |
| return bkeyp_val_u64s(format, k) * sizeof(u64); |
| } |
| |
| static inline void set_bkeyp_val_u64s(const struct bkey_format *format, |
| struct bkey_packed *k, unsigned val_u64s) |
| { |
| k->u64s = bkeyp_key_u64s(format, k) + val_u64s; |
| } |
| |
| #define bkeyp_val(_format, _k) \ |
| ((struct bch_val *) ((_k)->_data + bkeyp_key_u64s(_format, _k))) |
| |
| extern const struct bkey_format bch2_bkey_format_current; |
| |
| bool bch2_bkey_transform(const struct bkey_format *, |
| struct bkey_packed *, |
| const struct bkey_format *, |
| const struct bkey_packed *); |
| |
| struct bkey __bch2_bkey_unpack_key(const struct bkey_format *, |
| const struct bkey_packed *); |
| |
| #ifndef HAVE_BCACHEFS_COMPILED_UNPACK |
| struct bpos __bkey_unpack_pos(const struct bkey_format *, |
| const struct bkey_packed *); |
| #endif |
| |
| bool bch2_bkey_pack_key(struct bkey_packed *, const struct bkey *, |
| const struct bkey_format *); |
| |
| enum bkey_pack_pos_ret { |
| BKEY_PACK_POS_EXACT, |
| BKEY_PACK_POS_SMALLER, |
| BKEY_PACK_POS_FAIL, |
| }; |
| |
| enum bkey_pack_pos_ret bch2_bkey_pack_pos_lossy(struct bkey_packed *, struct bpos, |
| const struct btree *); |
| |
| static inline bool bkey_pack_pos(struct bkey_packed *out, struct bpos in, |
| const struct btree *b) |
| { |
| return bch2_bkey_pack_pos_lossy(out, in, b) == BKEY_PACK_POS_EXACT; |
| } |
| |
| void bch2_bkey_unpack(const struct btree *, struct bkey_i *, |
| const struct bkey_packed *); |
| bool bch2_bkey_pack(struct bkey_packed *, const struct bkey_i *, |
| const struct bkey_format *); |
| |
| static inline u64 bkey_field_max(const struct bkey_format *f, |
| enum bch_bkey_fields nr) |
| { |
| return f->bits_per_field[nr] < 64 |
| ? (le64_to_cpu(f->field_offset[nr]) + |
| ~(~0ULL << f->bits_per_field[nr])) |
| : U64_MAX; |
| } |
| |
| #ifdef HAVE_BCACHEFS_COMPILED_UNPACK |
| |
| int bch2_compile_bkey_format(const struct bkey_format *, void *); |
| |
| #else |
| |
| static inline int bch2_compile_bkey_format(const struct bkey_format *format, |
| void *out) { return 0; } |
| |
| #endif |
| |
| static inline void bkey_reassemble(struct bkey_i *dst, |
| struct bkey_s_c src) |
| { |
| BUG_ON(bkey_packed(src.k)); |
| dst->k = *src.k; |
| memcpy_u64s(&dst->v, src.v, bkey_val_u64s(src.k)); |
| } |
| |
| #define bkey_s_null ((struct bkey_s) { .k = NULL }) |
| #define bkey_s_c_null ((struct bkey_s_c) { .k = NULL }) |
| |
| #define bkey_s_err(err) ((struct bkey_s) { .k = ERR_PTR(err) }) |
| #define bkey_s_c_err(err) ((struct bkey_s_c) { .k = ERR_PTR(err) }) |
| |
| static inline struct bkey_s bkey_to_s(struct bkey *k) |
| { |
| return (struct bkey_s) { .k = k, .v = NULL }; |
| } |
| |
| static inline struct bkey_s_c bkey_to_s_c(const struct bkey *k) |
| { |
| return (struct bkey_s_c) { .k = k, .v = NULL }; |
| } |
| |
| static inline struct bkey_s bkey_i_to_s(struct bkey_i *k) |
| { |
| return (struct bkey_s) { .k = &k->k, .v = &k->v }; |
| } |
| |
| static inline struct bkey_s_c bkey_i_to_s_c(const struct bkey_i *k) |
| { |
| return (struct bkey_s_c) { .k = &k->k, .v = &k->v }; |
| } |
| |
| /* |
| * For a given type of value (e.g. struct bch_extent), generates the types for |
| * bkey + bch_extent - inline, split, split const - and also all the conversion |
| * functions, which also check that the value is of the correct type. |
| * |
| * We use anonymous unions for upcasting - e.g. converting from e.g. a |
| * bkey_i_extent to a bkey_i - since that's always safe, instead of conversion |
| * functions. |
| */ |
| #define __BKEY_VAL_ACCESSORS(name, nr, _assert) \ |
| struct bkey_s_c_##name { \ |
| union { \ |
| struct { \ |
| const struct bkey *k; \ |
| const struct bch_##name *v; \ |
| }; \ |
| struct bkey_s_c s_c; \ |
| }; \ |
| }; \ |
| \ |
| struct bkey_s_##name { \ |
| union { \ |
| struct { \ |
| struct bkey *k; \ |
| struct bch_##name *v; \ |
| }; \ |
| struct bkey_s_c_##name c; \ |
| struct bkey_s s; \ |
| struct bkey_s_c s_c; \ |
| }; \ |
| }; \ |
| \ |
| static inline struct bkey_i_##name *bkey_i_to_##name(struct bkey_i *k) \ |
| { \ |
| _assert(k->k.type, nr); \ |
| return container_of(&k->k, struct bkey_i_##name, k); \ |
| } \ |
| \ |
| static inline const struct bkey_i_##name * \ |
| bkey_i_to_##name##_c(const struct bkey_i *k) \ |
| { \ |
| _assert(k->k.type, nr); \ |
| return container_of(&k->k, struct bkey_i_##name, k); \ |
| } \ |
| \ |
| static inline struct bkey_s_##name bkey_s_to_##name(struct bkey_s k) \ |
| { \ |
| _assert(k.k->type, nr); \ |
| return (struct bkey_s_##name) { \ |
| .k = k.k, \ |
| .v = container_of(k.v, struct bch_##name, v), \ |
| }; \ |
| } \ |
| \ |
| static inline struct bkey_s_c_##name bkey_s_c_to_##name(struct bkey_s_c k)\ |
| { \ |
| _assert(k.k->type, nr); \ |
| return (struct bkey_s_c_##name) { \ |
| .k = k.k, \ |
| .v = container_of(k.v, struct bch_##name, v), \ |
| }; \ |
| } \ |
| \ |
| static inline struct bkey_s_##name name##_i_to_s(struct bkey_i_##name *k)\ |
| { \ |
| return (struct bkey_s_##name) { \ |
| .k = &k->k, \ |
| .v = &k->v, \ |
| }; \ |
| } \ |
| \ |
| static inline struct bkey_s_c_##name \ |
| name##_i_to_s_c(const struct bkey_i_##name *k) \ |
| { \ |
| return (struct bkey_s_c_##name) { \ |
| .k = &k->k, \ |
| .v = &k->v, \ |
| }; \ |
| } \ |
| \ |
| static inline struct bkey_s_##name bkey_i_to_s_##name(struct bkey_i *k) \ |
| { \ |
| _assert(k->k.type, nr); \ |
| return (struct bkey_s_##name) { \ |
| .k = &k->k, \ |
| .v = container_of(&k->v, struct bch_##name, v), \ |
| }; \ |
| } \ |
| \ |
| static inline struct bkey_s_c_##name \ |
| bkey_i_to_s_c_##name(const struct bkey_i *k) \ |
| { \ |
| _assert(k->k.type, nr); \ |
| return (struct bkey_s_c_##name) { \ |
| .k = &k->k, \ |
| .v = container_of(&k->v, struct bch_##name, v), \ |
| }; \ |
| } \ |
| \ |
| static inline struct bch_##name * \ |
| bkey_p_##name##_val(const struct bkey_format *f, \ |
| struct bkey_packed *k) \ |
| { \ |
| return container_of(bkeyp_val(f, k), struct bch_##name, v); \ |
| } \ |
| \ |
| static inline const struct bch_##name * \ |
| bkey_p_c_##name##_val(const struct bkey_format *f, \ |
| const struct bkey_packed *k) \ |
| { \ |
| return container_of(bkeyp_val(f, k), struct bch_##name, v); \ |
| } \ |
| \ |
| static inline struct bkey_i_##name *bkey_##name##_init(struct bkey_i *_k)\ |
| { \ |
| struct bkey_i_##name *k = \ |
| container_of(&_k->k, struct bkey_i_##name, k); \ |
| \ |
| bkey_init(&k->k); \ |
| memset(&k->v, 0, sizeof(k->v)); \ |
| k->k.type = nr; \ |
| set_bkey_val_bytes(&k->k, sizeof(k->v)); \ |
| \ |
| return k; \ |
| } |
| |
| #define __BKEY_VAL_ASSERT(_type, _nr) EBUG_ON(_type != _nr) |
| |
| #define BKEY_VAL_ACCESSORS(name, _nr) \ |
| static inline void __bch_##name##_assert(u8 type, u8 nr) \ |
| { \ |
| EBUG_ON(type != _nr); \ |
| } \ |
| \ |
| __BKEY_VAL_ACCESSORS(name, _nr, __bch_##name##_assert) |
| |
| BKEY_VAL_ACCESSORS(cookie, KEY_TYPE_COOKIE); |
| |
| static inline void __bch2_extent_assert(u8 type, u8 nr) |
| { |
| EBUG_ON(type != BCH_EXTENT && type != BCH_EXTENT_CACHED); |
| } |
| |
| __BKEY_VAL_ACCESSORS(extent, BCH_EXTENT, __bch2_extent_assert); |
| BKEY_VAL_ACCESSORS(reservation, BCH_RESERVATION); |
| |
| BKEY_VAL_ACCESSORS(inode, BCH_INODE_FS); |
| BKEY_VAL_ACCESSORS(inode_blockdev, BCH_INODE_BLOCKDEV); |
| BKEY_VAL_ACCESSORS(inode_generation, BCH_INODE_GENERATION); |
| |
| BKEY_VAL_ACCESSORS(dirent, BCH_DIRENT); |
| |
| BKEY_VAL_ACCESSORS(xattr, BCH_XATTR); |
| |
| BKEY_VAL_ACCESSORS(alloc, BCH_ALLOC); |
| |
| BKEY_VAL_ACCESSORS(quota, BCH_QUOTA); |
| |
| BKEY_VAL_ACCESSORS(stripe, BCH_STRIPE); |
| |
| /* byte order helpers */ |
| |
| #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ |
| |
| static inline unsigned high_word_offset(const struct bkey_format *f) |
| { |
| return f->key_u64s - 1; |
| } |
| |
| #define high_bit_offset 0 |
| #define nth_word(p, n) ((p) - (n)) |
| |
| #elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
| |
| static inline unsigned high_word_offset(const struct bkey_format *f) |
| { |
| return 0; |
| } |
| |
| #define high_bit_offset KEY_PACKED_BITS_START |
| #define nth_word(p, n) ((p) + (n)) |
| |
| #else |
| #error edit for your odd byteorder. |
| #endif |
| |
| #define high_word(f, k) ((k)->_data + high_word_offset(f)) |
| #define next_word(p) nth_word(p, 1) |
| #define prev_word(p) nth_word(p, -1) |
| |
| #ifdef CONFIG_BCACHEFS_DEBUG |
| void bch2_bkey_pack_test(void); |
| #else |
| static inline void bch2_bkey_pack_test(void) {} |
| #endif |
| |
| #endif /* _BCACHEFS_BKEY_H */ |