| /* SPDX-License-Identifier: GPL-2.0 */ |
| |
| #ifndef BTRFS_LRU_CACHE_H |
| #define BTRFS_LRU_CACHE_H |
| |
| #include <linux/maple_tree.h> |
| #include <linux/list.h> |
| |
| /* |
| * A cache entry. This is meant to be embedded in a structure of a user of |
| * this module. Similar to how struct list_head and struct rb_node are used. |
| * |
| * Note: it should be embedded as the first element in a struct (offset 0), and |
| * this module assumes it was allocated with kmalloc(), so it calls kfree() when |
| * it needs to free an entry. |
| */ |
| struct btrfs_lru_cache_entry { |
| struct list_head lru_list; |
| u64 key; |
| /* |
| * Optional generation associated to a key. Use 0 if not needed/used. |
| * Entries with the same key and different generations are stored in a |
| * linked list, so use this only for cases where there's a small number |
| * of different generations. |
| */ |
| u64 gen; |
| /* |
| * The maple tree uses unsigned long type for the keys, which is 32 bits |
| * on 32 bits systems, and 64 bits on 64 bits systems. So if we want to |
| * use something like inode numbers as keys, which are always a u64, we |
| * have to deal with this in a special way - we store the key in the |
| * entry itself, as a u64, and the values inserted into the maple tree |
| * are linked lists of entries - so in case we are on a 64 bits system, |
| * that list always has a single entry, while on 32 bits systems it |
| * may have more than one, with each entry having the same value for |
| * their lower 32 bits of the u64 key. |
| */ |
| struct list_head list; |
| }; |
| |
| struct btrfs_lru_cache { |
| struct list_head lru_list; |
| struct maple_tree entries; |
| /* Number of entries stored in the cache. */ |
| unsigned int size; |
| /* Maximum number of entries the cache can have. */ |
| unsigned int max_size; |
| }; |
| |
| #define btrfs_lru_cache_for_each_entry_safe(cache, entry, tmp) \ |
| list_for_each_entry_safe_reverse((entry), (tmp), &(cache)->lru_list, lru_list) |
| |
| static inline unsigned int btrfs_lru_cache_size(const struct btrfs_lru_cache *cache) |
| { |
| return cache->size; |
| } |
| |
| static inline bool btrfs_lru_cache_is_full(const struct btrfs_lru_cache *cache) |
| { |
| return cache->size >= cache->max_size; |
| } |
| |
| static inline struct btrfs_lru_cache_entry *btrfs_lru_cache_lru_entry( |
| struct btrfs_lru_cache *cache) |
| { |
| return list_first_entry_or_null(&cache->lru_list, |
| struct btrfs_lru_cache_entry, lru_list); |
| } |
| |
| void btrfs_lru_cache_init(struct btrfs_lru_cache *cache, unsigned int max_size); |
| struct btrfs_lru_cache_entry *btrfs_lru_cache_lookup(struct btrfs_lru_cache *cache, |
| u64 key, u64 gen); |
| int btrfs_lru_cache_store(struct btrfs_lru_cache *cache, |
| struct btrfs_lru_cache_entry *new_entry, |
| gfp_t gfp); |
| void btrfs_lru_cache_remove(struct btrfs_lru_cache *cache, |
| struct btrfs_lru_cache_entry *entry); |
| void btrfs_lru_cache_clear(struct btrfs_lru_cache *cache); |
| |
| #endif |