| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * Copyright (c) 2023-2025 Christoph Hellwig. |
| * Copyright (c) 2024-2025, Western Digital Corporation or its affiliates. |
| */ |
| #include "xfs.h" |
| #include "xfs_shared.h" |
| #include "xfs_format.h" |
| #include "xfs_log_format.h" |
| #include "xfs_error.h" |
| #include "xfs_trans_resv.h" |
| #include "xfs_mount.h" |
| #include "xfs_inode.h" |
| #include "xfs_iomap.h" |
| #include "xfs_trans.h" |
| #include "xfs_alloc.h" |
| #include "xfs_bmap.h" |
| #include "xfs_bmap_btree.h" |
| #include "xfs_trans_space.h" |
| #include "xfs_refcount.h" |
| #include "xfs_rtbitmap.h" |
| #include "xfs_rtrmap_btree.h" |
| #include "xfs_zone_alloc.h" |
| #include "xfs_zone_priv.h" |
| #include "xfs_zones.h" |
| #include "xfs_trace.h" |
| #include "xfs_mru_cache.h" |
| |
| void |
| xfs_open_zone_put( |
| struct xfs_open_zone *oz) |
| { |
| if (atomic_dec_and_test(&oz->oz_ref)) { |
| xfs_rtgroup_rele(oz->oz_rtg); |
| kfree(oz); |
| } |
| } |
| |
| static inline uint32_t |
| xfs_zone_bucket( |
| struct xfs_mount *mp, |
| uint32_t used_blocks) |
| { |
| return XFS_ZONE_USED_BUCKETS * used_blocks / |
| mp->m_groups[XG_TYPE_RTG].blocks; |
| } |
| |
| static inline void |
| xfs_zone_add_to_bucket( |
| struct xfs_zone_info *zi, |
| xfs_rgnumber_t rgno, |
| uint32_t to_bucket) |
| { |
| __set_bit(rgno, zi->zi_used_bucket_bitmap[to_bucket]); |
| zi->zi_used_bucket_entries[to_bucket]++; |
| } |
| |
| static inline void |
| xfs_zone_remove_from_bucket( |
| struct xfs_zone_info *zi, |
| xfs_rgnumber_t rgno, |
| uint32_t from_bucket) |
| { |
| __clear_bit(rgno, zi->zi_used_bucket_bitmap[from_bucket]); |
| zi->zi_used_bucket_entries[from_bucket]--; |
| } |
| |
| static void |
| xfs_zone_account_reclaimable( |
| struct xfs_rtgroup *rtg, |
| uint32_t freed) |
| { |
| struct xfs_group *xg = &rtg->rtg_group; |
| struct xfs_mount *mp = rtg_mount(rtg); |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| uint32_t used = rtg_rmap(rtg)->i_used_blocks; |
| xfs_rgnumber_t rgno = rtg_rgno(rtg); |
| uint32_t from_bucket = xfs_zone_bucket(mp, used + freed); |
| uint32_t to_bucket = xfs_zone_bucket(mp, used); |
| bool was_full = (used + freed == rtg_blocks(rtg)); |
| |
| /* |
| * This can be called from log recovery, where the zone_info structure |
| * hasn't been allocated yet. Skip all work as xfs_mount_zones will |
| * add the zones to the right buckets before the file systems becomes |
| * active. |
| */ |
| if (!zi) |
| return; |
| |
| if (!used) { |
| /* |
| * The zone is now empty, remove it from the bottom bucket and |
| * trigger a reset. |
| */ |
| trace_xfs_zone_emptied(rtg); |
| |
| if (!was_full) |
| xfs_group_clear_mark(xg, XFS_RTG_RECLAIMABLE); |
| |
| spin_lock(&zi->zi_used_buckets_lock); |
| if (!was_full) |
| xfs_zone_remove_from_bucket(zi, rgno, from_bucket); |
| spin_unlock(&zi->zi_used_buckets_lock); |
| |
| spin_lock(&zi->zi_reset_list_lock); |
| xg->xg_next_reset = zi->zi_reset_list; |
| zi->zi_reset_list = xg; |
| spin_unlock(&zi->zi_reset_list_lock); |
| |
| if (zi->zi_gc_thread) |
| wake_up_process(zi->zi_gc_thread); |
| } else if (was_full) { |
| /* |
| * The zone transitioned from full, mark it up as reclaimable |
| * and wake up GC which might be waiting for zones to reclaim. |
| */ |
| spin_lock(&zi->zi_used_buckets_lock); |
| xfs_zone_add_to_bucket(zi, rgno, to_bucket); |
| spin_unlock(&zi->zi_used_buckets_lock); |
| |
| xfs_group_set_mark(xg, XFS_RTG_RECLAIMABLE); |
| if (zi->zi_gc_thread && xfs_zoned_need_gc(mp)) |
| wake_up_process(zi->zi_gc_thread); |
| } else if (to_bucket != from_bucket) { |
| /* |
| * Move the zone to a new bucket if it dropped below the |
| * threshold. |
| */ |
| spin_lock(&zi->zi_used_buckets_lock); |
| xfs_zone_add_to_bucket(zi, rgno, to_bucket); |
| xfs_zone_remove_from_bucket(zi, rgno, from_bucket); |
| spin_unlock(&zi->zi_used_buckets_lock); |
| } |
| } |
| |
| static void |
| xfs_open_zone_mark_full( |
| struct xfs_open_zone *oz) |
| { |
| struct xfs_rtgroup *rtg = oz->oz_rtg; |
| struct xfs_mount *mp = rtg_mount(rtg); |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| uint32_t used = rtg_rmap(rtg)->i_used_blocks; |
| |
| trace_xfs_zone_full(rtg); |
| |
| WRITE_ONCE(rtg->rtg_open_zone, NULL); |
| |
| spin_lock(&zi->zi_open_zones_lock); |
| if (oz->oz_is_gc) { |
| ASSERT(current == zi->zi_gc_thread); |
| zi->zi_open_gc_zone = NULL; |
| } else { |
| zi->zi_nr_open_zones--; |
| list_del_init(&oz->oz_entry); |
| } |
| spin_unlock(&zi->zi_open_zones_lock); |
| xfs_open_zone_put(oz); |
| |
| wake_up_all(&zi->zi_zone_wait); |
| if (used < rtg_blocks(rtg)) |
| xfs_zone_account_reclaimable(rtg, rtg_blocks(rtg) - used); |
| } |
| |
| static void |
| xfs_zone_record_blocks( |
| struct xfs_trans *tp, |
| xfs_fsblock_t fsbno, |
| xfs_filblks_t len, |
| struct xfs_open_zone *oz, |
| bool used) |
| { |
| struct xfs_mount *mp = tp->t_mountp; |
| struct xfs_rtgroup *rtg = oz->oz_rtg; |
| struct xfs_inode *rmapip = rtg_rmap(rtg); |
| |
| trace_xfs_zone_record_blocks(oz, xfs_rtb_to_rgbno(mp, fsbno), len); |
| |
| xfs_rtgroup_lock(rtg, XFS_RTGLOCK_RMAP); |
| xfs_rtgroup_trans_join(tp, rtg, XFS_RTGLOCK_RMAP); |
| if (used) { |
| rmapip->i_used_blocks += len; |
| ASSERT(rmapip->i_used_blocks <= rtg_blocks(rtg)); |
| } else { |
| xfs_add_frextents(mp, len); |
| } |
| oz->oz_written += len; |
| if (oz->oz_written == rtg_blocks(rtg)) |
| xfs_open_zone_mark_full(oz); |
| xfs_trans_log_inode(tp, rmapip, XFS_ILOG_CORE); |
| } |
| |
| static int |
| xfs_zoned_map_extent( |
| struct xfs_trans *tp, |
| struct xfs_inode *ip, |
| struct xfs_bmbt_irec *new, |
| struct xfs_open_zone *oz, |
| xfs_fsblock_t old_startblock) |
| { |
| struct xfs_bmbt_irec data; |
| int nmaps = 1; |
| int error; |
| |
| /* Grab the corresponding mapping in the data fork. */ |
| error = xfs_bmapi_read(ip, new->br_startoff, new->br_blockcount, &data, |
| &nmaps, 0); |
| if (error) |
| return error; |
| |
| /* |
| * Cap the update to the existing extent in the data fork because we can |
| * only overwrite one extent at a time. |
| */ |
| ASSERT(new->br_blockcount >= data.br_blockcount); |
| new->br_blockcount = data.br_blockcount; |
| |
| /* |
| * If a data write raced with this GC write, keep the existing data in |
| * the data fork, mark our newly written GC extent as reclaimable, then |
| * move on to the next extent. |
| */ |
| if (old_startblock != NULLFSBLOCK && |
| old_startblock != data.br_startblock) |
| goto skip; |
| |
| trace_xfs_reflink_cow_remap_from(ip, new); |
| trace_xfs_reflink_cow_remap_to(ip, &data); |
| |
| error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK, |
| XFS_IEXT_REFLINK_END_COW_CNT); |
| if (error) |
| return error; |
| |
| if (data.br_startblock != HOLESTARTBLOCK) { |
| ASSERT(data.br_startblock != DELAYSTARTBLOCK); |
| ASSERT(!isnullstartblock(data.br_startblock)); |
| |
| xfs_bmap_unmap_extent(tp, ip, XFS_DATA_FORK, &data); |
| if (xfs_is_reflink_inode(ip)) { |
| xfs_refcount_decrease_extent(tp, true, &data); |
| } else { |
| error = xfs_free_extent_later(tp, data.br_startblock, |
| data.br_blockcount, NULL, |
| XFS_AG_RESV_NONE, |
| XFS_FREE_EXTENT_REALTIME); |
| if (error) |
| return error; |
| } |
| } |
| |
| xfs_zone_record_blocks(tp, new->br_startblock, new->br_blockcount, oz, |
| true); |
| |
| /* Map the new blocks into the data fork. */ |
| xfs_bmap_map_extent(tp, ip, XFS_DATA_FORK, new); |
| return 0; |
| |
| skip: |
| trace_xfs_reflink_cow_remap_skip(ip, new); |
| xfs_zone_record_blocks(tp, new->br_startblock, new->br_blockcount, oz, |
| false); |
| return 0; |
| } |
| |
| int |
| xfs_zoned_end_io( |
| struct xfs_inode *ip, |
| xfs_off_t offset, |
| xfs_off_t count, |
| xfs_daddr_t daddr, |
| struct xfs_open_zone *oz, |
| xfs_fsblock_t old_startblock) |
| { |
| struct xfs_mount *mp = ip->i_mount; |
| xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + count); |
| struct xfs_bmbt_irec new = { |
| .br_startoff = XFS_B_TO_FSBT(mp, offset), |
| .br_startblock = xfs_daddr_to_rtb(mp, daddr), |
| .br_state = XFS_EXT_NORM, |
| }; |
| unsigned int resblks = |
| XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK); |
| struct xfs_trans *tp; |
| int error; |
| |
| if (xfs_is_shutdown(mp)) |
| return -EIO; |
| |
| while (new.br_startoff < end_fsb) { |
| new.br_blockcount = end_fsb - new.br_startoff; |
| |
| error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, resblks, 0, |
| XFS_TRANS_RESERVE | XFS_TRANS_RES_FDBLKS, &tp); |
| if (error) |
| return error; |
| xfs_ilock(ip, XFS_ILOCK_EXCL); |
| xfs_trans_ijoin(tp, ip, 0); |
| |
| error = xfs_zoned_map_extent(tp, ip, &new, oz, old_startblock); |
| if (error) |
| xfs_trans_cancel(tp); |
| else |
| error = xfs_trans_commit(tp); |
| xfs_iunlock(ip, XFS_ILOCK_EXCL); |
| if (error) |
| return error; |
| |
| new.br_startoff += new.br_blockcount; |
| new.br_startblock += new.br_blockcount; |
| if (old_startblock != NULLFSBLOCK) |
| old_startblock += new.br_blockcount; |
| } |
| |
| return 0; |
| } |
| |
| /* |
| * "Free" blocks allocated in a zone. |
| * |
| * Just decrement the used blocks counter and report the space as freed. |
| */ |
| int |
| xfs_zone_free_blocks( |
| struct xfs_trans *tp, |
| struct xfs_rtgroup *rtg, |
| xfs_fsblock_t fsbno, |
| xfs_filblks_t len) |
| { |
| struct xfs_mount *mp = tp->t_mountp; |
| struct xfs_inode *rmapip = rtg_rmap(rtg); |
| |
| xfs_assert_ilocked(rmapip, XFS_ILOCK_EXCL); |
| |
| if (len > rmapip->i_used_blocks) { |
| xfs_err(mp, |
| "trying to free more blocks (%lld) than used counter (%u).", |
| len, rmapip->i_used_blocks); |
| ASSERT(len <= rmapip->i_used_blocks); |
| xfs_rtginode_mark_sick(rtg, XFS_RTGI_RMAP); |
| xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE); |
| return -EFSCORRUPTED; |
| } |
| |
| trace_xfs_zone_free_blocks(rtg, xfs_rtb_to_rgbno(mp, fsbno), len); |
| |
| rmapip->i_used_blocks -= len; |
| /* |
| * Don't add open zones to the reclaimable buckets. The I/O completion |
| * for writing the last block will take care of accounting for already |
| * unused blocks instead. |
| */ |
| if (!READ_ONCE(rtg->rtg_open_zone)) |
| xfs_zone_account_reclaimable(rtg, len); |
| xfs_add_frextents(mp, len); |
| xfs_trans_log_inode(tp, rmapip, XFS_ILOG_CORE); |
| return 0; |
| } |
| |
| /* |
| * Check if the zone containing the data just before the offset we are |
| * writing to is still open and has space. |
| */ |
| static struct xfs_open_zone * |
| xfs_last_used_zone( |
| struct iomap_ioend *ioend) |
| { |
| struct xfs_inode *ip = XFS_I(ioend->io_inode); |
| struct xfs_mount *mp = ip->i_mount; |
| xfs_fileoff_t offset_fsb = XFS_B_TO_FSB(mp, ioend->io_offset); |
| struct xfs_rtgroup *rtg = NULL; |
| struct xfs_open_zone *oz = NULL; |
| struct xfs_iext_cursor icur; |
| struct xfs_bmbt_irec got; |
| |
| xfs_ilock(ip, XFS_ILOCK_SHARED); |
| if (!xfs_iext_lookup_extent_before(ip, &ip->i_df, &offset_fsb, |
| &icur, &got)) { |
| xfs_iunlock(ip, XFS_ILOCK_SHARED); |
| return NULL; |
| } |
| xfs_iunlock(ip, XFS_ILOCK_SHARED); |
| |
| rtg = xfs_rtgroup_grab(mp, xfs_rtb_to_rgno(mp, got.br_startblock)); |
| if (!rtg) |
| return NULL; |
| |
| xfs_ilock(rtg_rmap(rtg), XFS_ILOCK_SHARED); |
| oz = READ_ONCE(rtg->rtg_open_zone); |
| if (oz && (oz->oz_is_gc || !atomic_inc_not_zero(&oz->oz_ref))) |
| oz = NULL; |
| xfs_iunlock(rtg_rmap(rtg), XFS_ILOCK_SHARED); |
| |
| xfs_rtgroup_rele(rtg); |
| return oz; |
| } |
| |
| static struct xfs_group * |
| xfs_find_free_zone( |
| struct xfs_mount *mp, |
| unsigned long start, |
| unsigned long end) |
| { |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| XA_STATE (xas, &mp->m_groups[XG_TYPE_RTG].xa, start); |
| struct xfs_group *xg; |
| |
| xas_lock(&xas); |
| xas_for_each_marked(&xas, xg, end, XFS_RTG_FREE) |
| if (atomic_inc_not_zero(&xg->xg_active_ref)) |
| goto found; |
| xas_unlock(&xas); |
| return NULL; |
| |
| found: |
| xas_clear_mark(&xas, XFS_RTG_FREE); |
| atomic_dec(&zi->zi_nr_free_zones); |
| zi->zi_free_zone_cursor = xg->xg_gno; |
| xas_unlock(&xas); |
| return xg; |
| } |
| |
| static struct xfs_open_zone * |
| xfs_init_open_zone( |
| struct xfs_rtgroup *rtg, |
| xfs_rgblock_t write_pointer, |
| enum rw_hint write_hint, |
| bool is_gc) |
| { |
| struct xfs_open_zone *oz; |
| |
| oz = kzalloc(sizeof(*oz), GFP_NOFS | __GFP_NOFAIL); |
| spin_lock_init(&oz->oz_alloc_lock); |
| atomic_set(&oz->oz_ref, 1); |
| oz->oz_rtg = rtg; |
| oz->oz_write_pointer = write_pointer; |
| oz->oz_written = write_pointer; |
| oz->oz_write_hint = write_hint; |
| oz->oz_is_gc = is_gc; |
| |
| /* |
| * All dereferences of rtg->rtg_open_zone hold the ILOCK for the rmap |
| * inode, but we don't really want to take that here because we are |
| * under the zone_list_lock. Ensure the pointer is only set for a fully |
| * initialized open zone structure so that a racy lookup finding it is |
| * fine. |
| */ |
| WRITE_ONCE(rtg->rtg_open_zone, oz); |
| return oz; |
| } |
| |
| /* |
| * Find a completely free zone, open it, and return a reference. |
| */ |
| struct xfs_open_zone * |
| xfs_open_zone( |
| struct xfs_mount *mp, |
| enum rw_hint write_hint, |
| bool is_gc) |
| { |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| struct xfs_group *xg; |
| |
| xg = xfs_find_free_zone(mp, zi->zi_free_zone_cursor, ULONG_MAX); |
| if (!xg) |
| xg = xfs_find_free_zone(mp, 0, zi->zi_free_zone_cursor); |
| if (!xg) |
| return NULL; |
| |
| set_current_state(TASK_RUNNING); |
| return xfs_init_open_zone(to_rtg(xg), 0, write_hint, is_gc); |
| } |
| |
| static struct xfs_open_zone * |
| xfs_try_open_zone( |
| struct xfs_mount *mp, |
| enum rw_hint write_hint) |
| { |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| struct xfs_open_zone *oz; |
| |
| if (zi->zi_nr_open_zones >= mp->m_max_open_zones - XFS_OPEN_GC_ZONES) |
| return NULL; |
| if (atomic_read(&zi->zi_nr_free_zones) < |
| XFS_GC_ZONES - XFS_OPEN_GC_ZONES) |
| return NULL; |
| |
| /* |
| * Increment the open zone count to reserve our slot before dropping |
| * zi_open_zones_lock. |
| */ |
| zi->zi_nr_open_zones++; |
| spin_unlock(&zi->zi_open_zones_lock); |
| oz = xfs_open_zone(mp, write_hint, false); |
| spin_lock(&zi->zi_open_zones_lock); |
| if (!oz) { |
| zi->zi_nr_open_zones--; |
| return NULL; |
| } |
| |
| atomic_inc(&oz->oz_ref); |
| list_add_tail(&oz->oz_entry, &zi->zi_open_zones); |
| |
| /* |
| * If this was the last free zone, other waiters might be waiting |
| * on us to write to it as well. |
| */ |
| wake_up_all(&zi->zi_zone_wait); |
| |
| if (xfs_zoned_need_gc(mp)) |
| wake_up_process(zi->zi_gc_thread); |
| |
| trace_xfs_zone_opened(oz->oz_rtg); |
| return oz; |
| } |
| |
| /* |
| * For data with short or medium lifetime, try to colocated it into an |
| * already open zone with a matching temperature. |
| */ |
| static bool |
| xfs_colocate_eagerly( |
| enum rw_hint file_hint) |
| { |
| switch (file_hint) { |
| case WRITE_LIFE_MEDIUM: |
| case WRITE_LIFE_SHORT: |
| case WRITE_LIFE_NONE: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| static bool |
| xfs_good_hint_match( |
| struct xfs_open_zone *oz, |
| enum rw_hint file_hint) |
| { |
| switch (oz->oz_write_hint) { |
| case WRITE_LIFE_LONG: |
| case WRITE_LIFE_EXTREME: |
| /* colocate long and extreme */ |
| if (file_hint == WRITE_LIFE_LONG || |
| file_hint == WRITE_LIFE_EXTREME) |
| return true; |
| break; |
| case WRITE_LIFE_MEDIUM: |
| /* colocate medium with medium */ |
| if (file_hint == WRITE_LIFE_MEDIUM) |
| return true; |
| break; |
| case WRITE_LIFE_SHORT: |
| case WRITE_LIFE_NONE: |
| case WRITE_LIFE_NOT_SET: |
| /* colocate short and none */ |
| if (file_hint <= WRITE_LIFE_SHORT) |
| return true; |
| break; |
| } |
| return false; |
| } |
| |
| static bool |
| xfs_try_use_zone( |
| struct xfs_zone_info *zi, |
| enum rw_hint file_hint, |
| struct xfs_open_zone *oz, |
| bool lowspace) |
| { |
| if (oz->oz_write_pointer == rtg_blocks(oz->oz_rtg)) |
| return false; |
| if (!lowspace && !xfs_good_hint_match(oz, file_hint)) |
| return false; |
| if (!atomic_inc_not_zero(&oz->oz_ref)) |
| return false; |
| |
| /* |
| * If we have a hint set for the data, use that for the zone even if |
| * some data was written already without any hint set, but don't change |
| * the temperature after that as that would make little sense without |
| * tracking per-temperature class written block counts, which is |
| * probably overkill anyway. |
| */ |
| if (file_hint != WRITE_LIFE_NOT_SET && |
| oz->oz_write_hint == WRITE_LIFE_NOT_SET) |
| oz->oz_write_hint = file_hint; |
| |
| /* |
| * If we couldn't match by inode or life time we just pick the first |
| * zone with enough space above. For that we want the least busy zone |
| * for some definition of "least" busy. For now this simple LRU |
| * algorithm that rotates every zone to the end of the list will do it, |
| * even if it isn't exactly cache friendly. |
| */ |
| if (!list_is_last(&oz->oz_entry, &zi->zi_open_zones)) |
| list_move_tail(&oz->oz_entry, &zi->zi_open_zones); |
| return true; |
| } |
| |
| static struct xfs_open_zone * |
| xfs_select_open_zone_lru( |
| struct xfs_zone_info *zi, |
| enum rw_hint file_hint, |
| bool lowspace) |
| { |
| struct xfs_open_zone *oz; |
| |
| lockdep_assert_held(&zi->zi_open_zones_lock); |
| |
| list_for_each_entry(oz, &zi->zi_open_zones, oz_entry) |
| if (xfs_try_use_zone(zi, file_hint, oz, lowspace)) |
| return oz; |
| |
| cond_resched_lock(&zi->zi_open_zones_lock); |
| return NULL; |
| } |
| |
| static struct xfs_open_zone * |
| xfs_select_open_zone_mru( |
| struct xfs_zone_info *zi, |
| enum rw_hint file_hint) |
| { |
| struct xfs_open_zone *oz; |
| |
| lockdep_assert_held(&zi->zi_open_zones_lock); |
| |
| list_for_each_entry_reverse(oz, &zi->zi_open_zones, oz_entry) |
| if (xfs_try_use_zone(zi, file_hint, oz, false)) |
| return oz; |
| |
| cond_resched_lock(&zi->zi_open_zones_lock); |
| return NULL; |
| } |
| |
| static inline enum rw_hint xfs_inode_write_hint(struct xfs_inode *ip) |
| { |
| if (xfs_has_nolifetime(ip->i_mount)) |
| return WRITE_LIFE_NOT_SET; |
| return VFS_I(ip)->i_write_hint; |
| } |
| |
| /* |
| * Try to pack inodes that are written back after they were closed tight instead |
| * of trying to open new zones for them or spread them to the least recently |
| * used zone. This optimizes the data layout for workloads that untar or copy |
| * a lot of small files. Right now this does not separate multiple such |
| * streams. |
| */ |
| static inline bool xfs_zoned_pack_tight(struct xfs_inode *ip) |
| { |
| return !inode_is_open_for_write(VFS_I(ip)) && |
| !(ip->i_diflags & XFS_DIFLAG_APPEND); |
| } |
| |
| /* |
| * Pick a new zone for writes. |
| * |
| * If we aren't using up our budget of open zones just open a new one from the |
| * freelist. Else try to find one that matches the expected data lifetime. If |
| * we don't find one that is good pick any zone that is available. |
| */ |
| static struct xfs_open_zone * |
| xfs_select_zone_nowait( |
| struct xfs_mount *mp, |
| enum rw_hint write_hint, |
| bool pack_tight) |
| { |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| struct xfs_open_zone *oz = NULL; |
| |
| if (xfs_is_shutdown(mp)) |
| return NULL; |
| |
| /* |
| * Try to fill up open zones with matching temperature if available. It |
| * is better to try to co-locate data when this is favorable, so we can |
| * activate empty zones when it is statistically better to separate |
| * data. |
| */ |
| spin_lock(&zi->zi_open_zones_lock); |
| if (xfs_colocate_eagerly(write_hint)) |
| oz = xfs_select_open_zone_lru(zi, write_hint, false); |
| else if (pack_tight) |
| oz = xfs_select_open_zone_mru(zi, write_hint); |
| if (oz) |
| goto out_unlock; |
| |
| /* |
| * See if we can open a new zone and use that. |
| */ |
| oz = xfs_try_open_zone(mp, write_hint); |
| if (oz) |
| goto out_unlock; |
| |
| /* |
| * Try to colocate cold data with other cold data if we failed to open a |
| * new zone for it. |
| */ |
| if (write_hint != WRITE_LIFE_NOT_SET && |
| !xfs_colocate_eagerly(write_hint)) |
| oz = xfs_select_open_zone_lru(zi, write_hint, false); |
| if (!oz) |
| oz = xfs_select_open_zone_lru(zi, WRITE_LIFE_NOT_SET, false); |
| if (!oz) |
| oz = xfs_select_open_zone_lru(zi, WRITE_LIFE_NOT_SET, true); |
| out_unlock: |
| spin_unlock(&zi->zi_open_zones_lock); |
| return oz; |
| } |
| |
| static struct xfs_open_zone * |
| xfs_select_zone( |
| struct xfs_mount *mp, |
| enum rw_hint write_hint, |
| bool pack_tight) |
| { |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| DEFINE_WAIT (wait); |
| struct xfs_open_zone *oz; |
| |
| oz = xfs_select_zone_nowait(mp, write_hint, pack_tight); |
| if (oz) |
| return oz; |
| |
| for (;;) { |
| prepare_to_wait(&zi->zi_zone_wait, &wait, TASK_UNINTERRUPTIBLE); |
| oz = xfs_select_zone_nowait(mp, write_hint, pack_tight); |
| if (oz) |
| break; |
| schedule(); |
| } |
| finish_wait(&zi->zi_zone_wait, &wait); |
| return oz; |
| } |
| |
| static unsigned int |
| xfs_zone_alloc_blocks( |
| struct xfs_open_zone *oz, |
| xfs_filblks_t count_fsb, |
| sector_t *sector, |
| bool *is_seq) |
| { |
| struct xfs_rtgroup *rtg = oz->oz_rtg; |
| struct xfs_mount *mp = rtg_mount(rtg); |
| xfs_rgblock_t rgbno; |
| |
| spin_lock(&oz->oz_alloc_lock); |
| count_fsb = min3(count_fsb, XFS_MAX_BMBT_EXTLEN, |
| (xfs_filblks_t)rtg_blocks(rtg) - oz->oz_write_pointer); |
| if (!count_fsb) { |
| spin_unlock(&oz->oz_alloc_lock); |
| return 0; |
| } |
| rgbno = oz->oz_write_pointer; |
| oz->oz_write_pointer += count_fsb; |
| spin_unlock(&oz->oz_alloc_lock); |
| |
| trace_xfs_zone_alloc_blocks(oz, rgbno, count_fsb); |
| |
| *sector = xfs_gbno_to_daddr(&rtg->rtg_group, 0); |
| *is_seq = bdev_zone_is_seq(mp->m_rtdev_targp->bt_bdev, *sector); |
| if (!*is_seq) |
| *sector += XFS_FSB_TO_BB(mp, rgbno); |
| return XFS_FSB_TO_B(mp, count_fsb); |
| } |
| |
| void |
| xfs_mark_rtg_boundary( |
| struct iomap_ioend *ioend) |
| { |
| struct xfs_mount *mp = XFS_I(ioend->io_inode)->i_mount; |
| sector_t sector = ioend->io_bio.bi_iter.bi_sector; |
| |
| if (xfs_rtb_to_rgbno(mp, xfs_daddr_to_rtb(mp, sector)) == 0) |
| ioend->io_flags |= IOMAP_IOEND_BOUNDARY; |
| } |
| |
| static void |
| xfs_submit_zoned_bio( |
| struct iomap_ioend *ioend, |
| struct xfs_open_zone *oz, |
| bool is_seq) |
| { |
| ioend->io_bio.bi_iter.bi_sector = ioend->io_sector; |
| ioend->io_private = oz; |
| atomic_inc(&oz->oz_ref); /* for xfs_zoned_end_io */ |
| |
| if (is_seq) { |
| ioend->io_bio.bi_opf &= ~REQ_OP_WRITE; |
| ioend->io_bio.bi_opf |= REQ_OP_ZONE_APPEND; |
| } else { |
| xfs_mark_rtg_boundary(ioend); |
| } |
| |
| submit_bio(&ioend->io_bio); |
| } |
| |
| /* |
| * Cache the last zone written to for an inode so that it is considered first |
| * for subsequent writes. |
| */ |
| struct xfs_zone_cache_item { |
| struct xfs_mru_cache_elem mru; |
| struct xfs_open_zone *oz; |
| }; |
| |
| static inline struct xfs_zone_cache_item * |
| xfs_zone_cache_item(struct xfs_mru_cache_elem *mru) |
| { |
| return container_of(mru, struct xfs_zone_cache_item, mru); |
| } |
| |
| static void |
| xfs_zone_cache_free_func( |
| void *data, |
| struct xfs_mru_cache_elem *mru) |
| { |
| struct xfs_zone_cache_item *item = xfs_zone_cache_item(mru); |
| |
| xfs_open_zone_put(item->oz); |
| kfree(item); |
| } |
| |
| /* |
| * Check if we have a cached last open zone available for the inode and |
| * if yes return a reference to it. |
| */ |
| static struct xfs_open_zone * |
| xfs_cached_zone( |
| struct xfs_mount *mp, |
| struct xfs_inode *ip) |
| { |
| struct xfs_mru_cache_elem *mru; |
| struct xfs_open_zone *oz; |
| |
| mru = xfs_mru_cache_lookup(mp->m_zone_cache, ip->i_ino); |
| if (!mru) |
| return NULL; |
| oz = xfs_zone_cache_item(mru)->oz; |
| if (oz) { |
| /* |
| * GC only steals open zones at mount time, so no GC zones |
| * should end up in the cache. |
| */ |
| ASSERT(!oz->oz_is_gc); |
| ASSERT(atomic_read(&oz->oz_ref) > 0); |
| atomic_inc(&oz->oz_ref); |
| } |
| xfs_mru_cache_done(mp->m_zone_cache); |
| return oz; |
| } |
| |
| /* |
| * Update the last used zone cache for a given inode. |
| * |
| * The caller must have a reference on the open zone. |
| */ |
| static void |
| xfs_zone_cache_create_association( |
| struct xfs_inode *ip, |
| struct xfs_open_zone *oz) |
| { |
| struct xfs_mount *mp = ip->i_mount; |
| struct xfs_zone_cache_item *item = NULL; |
| struct xfs_mru_cache_elem *mru; |
| |
| ASSERT(atomic_read(&oz->oz_ref) > 0); |
| atomic_inc(&oz->oz_ref); |
| |
| mru = xfs_mru_cache_lookup(mp->m_zone_cache, ip->i_ino); |
| if (mru) { |
| /* |
| * If we have an association already, update it to point to the |
| * new zone. |
| */ |
| item = xfs_zone_cache_item(mru); |
| xfs_open_zone_put(item->oz); |
| item->oz = oz; |
| xfs_mru_cache_done(mp->m_zone_cache); |
| return; |
| } |
| |
| item = kmalloc(sizeof(*item), GFP_KERNEL); |
| if (!item) { |
| xfs_open_zone_put(oz); |
| return; |
| } |
| item->oz = oz; |
| xfs_mru_cache_insert(mp->m_zone_cache, ip->i_ino, &item->mru); |
| } |
| |
| void |
| xfs_zone_alloc_and_submit( |
| struct iomap_ioend *ioend, |
| struct xfs_open_zone **oz) |
| { |
| struct xfs_inode *ip = XFS_I(ioend->io_inode); |
| struct xfs_mount *mp = ip->i_mount; |
| enum rw_hint write_hint = xfs_inode_write_hint(ip); |
| bool pack_tight = xfs_zoned_pack_tight(ip); |
| unsigned int alloc_len; |
| struct iomap_ioend *split; |
| bool is_seq; |
| |
| if (xfs_is_shutdown(mp)) |
| goto out_error; |
| |
| /* |
| * If we don't have a cached zone in this write context, see if the |
| * last extent before the one we are writing to points to an active |
| * zone. If so, just continue writing to it. |
| */ |
| if (!*oz && ioend->io_offset) |
| *oz = xfs_last_used_zone(ioend); |
| if (!*oz) |
| *oz = xfs_cached_zone(mp, ip); |
| |
| if (!*oz) { |
| select_zone: |
| *oz = xfs_select_zone(mp, write_hint, pack_tight); |
| if (!*oz) |
| goto out_error; |
| |
| xfs_zone_cache_create_association(ip, *oz); |
| } |
| |
| alloc_len = xfs_zone_alloc_blocks(*oz, XFS_B_TO_FSB(mp, ioend->io_size), |
| &ioend->io_sector, &is_seq); |
| if (!alloc_len) { |
| xfs_open_zone_put(*oz); |
| goto select_zone; |
| } |
| |
| while ((split = iomap_split_ioend(ioend, alloc_len, is_seq))) { |
| if (IS_ERR(split)) |
| goto out_split_error; |
| alloc_len -= split->io_bio.bi_iter.bi_size; |
| xfs_submit_zoned_bio(split, *oz, is_seq); |
| if (!alloc_len) { |
| xfs_open_zone_put(*oz); |
| goto select_zone; |
| } |
| } |
| |
| xfs_submit_zoned_bio(ioend, *oz, is_seq); |
| return; |
| |
| out_split_error: |
| ioend->io_bio.bi_status = errno_to_blk_status(PTR_ERR(split)); |
| out_error: |
| bio_io_error(&ioend->io_bio); |
| } |
| |
| /* |
| * Wake up all threads waiting for a zoned space allocation when the file system |
| * is shut down. |
| */ |
| void |
| xfs_zoned_wake_all( |
| struct xfs_mount *mp) |
| { |
| /* |
| * Don't wake up if there is no m_zone_info. This is complicated by the |
| * fact that unmount can't atomically clear m_zone_info and thus we need |
| * to check SB_ACTIVE for that, but mount temporarily enables SB_ACTIVE |
| * during log recovery so we can't entirely rely on that either. |
| */ |
| if ((mp->m_super->s_flags & SB_ACTIVE) && mp->m_zone_info) |
| wake_up_all(&mp->m_zone_info->zi_zone_wait); |
| } |
| |
| /* |
| * Check if @rgbno in @rgb is a potentially valid block. It might still be |
| * unused, but that information is only found in the rmap. |
| */ |
| bool |
| xfs_zone_rgbno_is_valid( |
| struct xfs_rtgroup *rtg, |
| xfs_rgnumber_t rgbno) |
| { |
| lockdep_assert_held(&rtg_rmap(rtg)->i_lock); |
| |
| if (rtg->rtg_open_zone) |
| return rgbno < rtg->rtg_open_zone->oz_write_pointer; |
| return !xa_get_mark(&rtg_mount(rtg)->m_groups[XG_TYPE_RTG].xa, |
| rtg_rgno(rtg), XFS_RTG_FREE); |
| } |
| |
| static void |
| xfs_free_open_zones( |
| struct xfs_zone_info *zi) |
| { |
| struct xfs_open_zone *oz; |
| |
| spin_lock(&zi->zi_open_zones_lock); |
| while ((oz = list_first_entry_or_null(&zi->zi_open_zones, |
| struct xfs_open_zone, oz_entry))) { |
| list_del(&oz->oz_entry); |
| xfs_open_zone_put(oz); |
| } |
| spin_unlock(&zi->zi_open_zones_lock); |
| } |
| |
| struct xfs_init_zones { |
| struct xfs_mount *mp; |
| uint64_t available; |
| uint64_t reclaimable; |
| }; |
| |
| static int |
| xfs_init_zone( |
| struct xfs_init_zones *iz, |
| struct xfs_rtgroup *rtg, |
| struct blk_zone *zone) |
| { |
| struct xfs_mount *mp = rtg_mount(rtg); |
| struct xfs_zone_info *zi = mp->m_zone_info; |
| uint64_t used = rtg_rmap(rtg)->i_used_blocks; |
| xfs_rgblock_t write_pointer, highest_rgbno; |
| int error; |
| |
| if (zone && !xfs_zone_validate(zone, rtg, &write_pointer)) |
| return -EFSCORRUPTED; |
| |
| /* |
| * For sequential write required zones we retrieved the hardware write |
| * pointer above. |
| * |
| * For conventional zones or conventional devices we don't have that |
| * luxury. Instead query the rmap to find the highest recorded block |
| * and set the write pointer to the block after that. In case of a |
| * power loss this misses blocks where the data I/O has completed but |
| * not recorded in the rmap yet, and it also rewrites blocks if the most |
| * recently written ones got deleted again before unmount, but this is |
| * the best we can do without hardware support. |
| */ |
| if (!zone || zone->cond == BLK_ZONE_COND_NOT_WP) { |
| xfs_rtgroup_lock(rtg, XFS_RTGLOCK_RMAP); |
| highest_rgbno = xfs_rtrmap_highest_rgbno(rtg); |
| if (highest_rgbno == NULLRGBLOCK) |
| write_pointer = 0; |
| else |
| write_pointer = highest_rgbno + 1; |
| xfs_rtgroup_unlock(rtg, XFS_RTGLOCK_RMAP); |
| } |
| |
| /* |
| * If there are no used blocks, but the zone is not in empty state yet |
| * we lost power before the zoned reset. In that case finish the work |
| * here. |
| */ |
| if (write_pointer == rtg_blocks(rtg) && used == 0) { |
| error = xfs_zone_gc_reset_sync(rtg); |
| if (error) |
| return error; |
| write_pointer = 0; |
| } |
| |
| if (write_pointer == 0) { |
| /* zone is empty */ |
| atomic_inc(&zi->zi_nr_free_zones); |
| xfs_group_set_mark(&rtg->rtg_group, XFS_RTG_FREE); |
| iz->available += rtg_blocks(rtg); |
| } else if (write_pointer < rtg_blocks(rtg)) { |
| /* zone is open */ |
| struct xfs_open_zone *oz; |
| |
| atomic_inc(&rtg_group(rtg)->xg_active_ref); |
| oz = xfs_init_open_zone(rtg, write_pointer, WRITE_LIFE_NOT_SET, |
| false); |
| list_add_tail(&oz->oz_entry, &zi->zi_open_zones); |
| zi->zi_nr_open_zones++; |
| |
| iz->available += (rtg_blocks(rtg) - write_pointer); |
| iz->reclaimable += write_pointer - used; |
| } else if (used < rtg_blocks(rtg)) { |
| /* zone fully written, but has freed blocks */ |
| xfs_zone_account_reclaimable(rtg, rtg_blocks(rtg) - used); |
| iz->reclaimable += (rtg_blocks(rtg) - used); |
| } |
| |
| return 0; |
| } |
| |
| static int |
| xfs_get_zone_info_cb( |
| struct blk_zone *zone, |
| unsigned int idx, |
| void *data) |
| { |
| struct xfs_init_zones *iz = data; |
| struct xfs_mount *mp = iz->mp; |
| xfs_fsblock_t zsbno = xfs_daddr_to_rtb(mp, zone->start); |
| xfs_rgnumber_t rgno; |
| struct xfs_rtgroup *rtg; |
| int error; |
| |
| if (xfs_rtb_to_rgbno(mp, zsbno) != 0) { |
| xfs_warn(mp, "mismatched zone start 0x%llx.", zsbno); |
| return -EFSCORRUPTED; |
| } |
| |
| rgno = xfs_rtb_to_rgno(mp, zsbno); |
| rtg = xfs_rtgroup_grab(mp, rgno); |
| if (!rtg) { |
| xfs_warn(mp, "realtime group not found for zone %u.", rgno); |
| return -EFSCORRUPTED; |
| } |
| error = xfs_init_zone(iz, rtg, zone); |
| xfs_rtgroup_rele(rtg); |
| return error; |
| } |
| |
| /* |
| * Calculate the max open zone limit based on the of number of |
| * backing zones available |
| */ |
| static inline uint32_t |
| xfs_max_open_zones( |
| struct xfs_mount *mp) |
| { |
| unsigned int max_open, max_open_data_zones; |
| /* |
| * We need two zones for every open data zone, |
| * one in reserve as we don't reclaim open zones. One data zone |
| * and its spare is included in XFS_MIN_ZONES. |
| */ |
| max_open_data_zones = (mp->m_sb.sb_rgcount - XFS_MIN_ZONES) / 2 + 1; |
| max_open = max_open_data_zones + XFS_OPEN_GC_ZONES; |
| |
| /* |
| * Cap the max open limit to 1/4 of available space |
| */ |
| max_open = min(max_open, mp->m_sb.sb_rgcount / 4); |
| |
| return max(XFS_MIN_OPEN_ZONES, max_open); |
| } |
| |
| /* |
| * Normally we use the open zone limit that the device reports. If there is |
| * none let the user pick one from the command line. |
| * |
| * If the device doesn't report an open zone limit and there is no override, |
| * allow to hold about a quarter of the zones open. In theory we could allow |
| * all to be open, but at that point we run into GC deadlocks because we can't |
| * reclaim open zones. |
| * |
| * When used on conventional SSDs a lower open limit is advisable as we'll |
| * otherwise overwhelm the FTL just as much as a conventional block allocator. |
| * |
| * Note: To debug the open zone management code, force max_open to 1 here. |
| */ |
| static int |
| xfs_calc_open_zones( |
| struct xfs_mount *mp) |
| { |
| struct block_device *bdev = mp->m_rtdev_targp->bt_bdev; |
| unsigned int bdev_open_zones = bdev_max_open_zones(bdev); |
| |
| if (!mp->m_max_open_zones) { |
| if (bdev_open_zones) |
| mp->m_max_open_zones = bdev_open_zones; |
| else |
| mp->m_max_open_zones = xfs_max_open_zones(mp); |
| } |
| |
| if (mp->m_max_open_zones < XFS_MIN_OPEN_ZONES) { |
| xfs_notice(mp, "need at least %u open zones.", |
| XFS_MIN_OPEN_ZONES); |
| return -EIO; |
| } |
| |
| if (bdev_open_zones && bdev_open_zones < mp->m_max_open_zones) { |
| mp->m_max_open_zones = bdev_open_zones; |
| xfs_info(mp, "limiting open zones to %u due to hardware limit.\n", |
| bdev_open_zones); |
| } |
| |
| if (mp->m_max_open_zones > xfs_max_open_zones(mp)) { |
| mp->m_max_open_zones = xfs_max_open_zones(mp); |
| xfs_info(mp, |
| "limiting open zones to %u due to total zone count (%u)", |
| mp->m_max_open_zones, mp->m_sb.sb_rgcount); |
| } |
| |
| return 0; |
| } |
| |
| static unsigned long * |
| xfs_alloc_bucket_bitmap( |
| struct xfs_mount *mp) |
| { |
| return kvmalloc_array(BITS_TO_LONGS(mp->m_sb.sb_rgcount), |
| sizeof(unsigned long), GFP_KERNEL | __GFP_ZERO); |
| } |
| |
| static struct xfs_zone_info * |
| xfs_alloc_zone_info( |
| struct xfs_mount *mp) |
| { |
| struct xfs_zone_info *zi; |
| int i; |
| |
| zi = kzalloc(sizeof(*zi), GFP_KERNEL); |
| if (!zi) |
| return NULL; |
| INIT_LIST_HEAD(&zi->zi_open_zones); |
| INIT_LIST_HEAD(&zi->zi_reclaim_reservations); |
| spin_lock_init(&zi->zi_reset_list_lock); |
| spin_lock_init(&zi->zi_open_zones_lock); |
| spin_lock_init(&zi->zi_reservation_lock); |
| init_waitqueue_head(&zi->zi_zone_wait); |
| spin_lock_init(&zi->zi_used_buckets_lock); |
| for (i = 0; i < XFS_ZONE_USED_BUCKETS; i++) { |
| zi->zi_used_bucket_bitmap[i] = xfs_alloc_bucket_bitmap(mp); |
| if (!zi->zi_used_bucket_bitmap[i]) |
| goto out_free_bitmaps; |
| } |
| return zi; |
| |
| out_free_bitmaps: |
| while (--i > 0) |
| kvfree(zi->zi_used_bucket_bitmap[i]); |
| kfree(zi); |
| return NULL; |
| } |
| |
| static void |
| xfs_free_zone_info( |
| struct xfs_zone_info *zi) |
| { |
| int i; |
| |
| xfs_free_open_zones(zi); |
| for (i = 0; i < XFS_ZONE_USED_BUCKETS; i++) |
| kvfree(zi->zi_used_bucket_bitmap[i]); |
| kfree(zi); |
| } |
| |
| int |
| xfs_mount_zones( |
| struct xfs_mount *mp) |
| { |
| struct xfs_init_zones iz = { |
| .mp = mp, |
| }; |
| struct xfs_buftarg *bt = mp->m_rtdev_targp; |
| int error; |
| |
| if (!bt) { |
| xfs_notice(mp, "RT device missing."); |
| return -EINVAL; |
| } |
| |
| if (!xfs_has_rtgroups(mp) || !xfs_has_rmapbt(mp)) { |
| xfs_notice(mp, "invalid flag combination."); |
| return -EFSCORRUPTED; |
| } |
| if (mp->m_sb.sb_rextsize != 1) { |
| xfs_notice(mp, "zoned file systems do not support rextsize."); |
| return -EFSCORRUPTED; |
| } |
| if (mp->m_sb.sb_rgcount < XFS_MIN_ZONES) { |
| xfs_notice(mp, |
| "zoned file systems need to have at least %u zones.", XFS_MIN_ZONES); |
| return -EFSCORRUPTED; |
| } |
| |
| error = xfs_calc_open_zones(mp); |
| if (error) |
| return error; |
| |
| mp->m_zone_info = xfs_alloc_zone_info(mp); |
| if (!mp->m_zone_info) |
| return -ENOMEM; |
| |
| xfs_info(mp, "%u zones of %u blocks size (%u max open)", |
| mp->m_sb.sb_rgcount, mp->m_groups[XG_TYPE_RTG].blocks, |
| mp->m_max_open_zones); |
| trace_xfs_zones_mount(mp); |
| |
| if (bdev_is_zoned(bt->bt_bdev)) { |
| error = blkdev_report_zones(bt->bt_bdev, |
| XFS_FSB_TO_BB(mp, mp->m_sb.sb_rtstart), |
| mp->m_sb.sb_rgcount, xfs_get_zone_info_cb, &iz); |
| if (error < 0) |
| goto out_free_zone_info; |
| } else { |
| struct xfs_rtgroup *rtg = NULL; |
| |
| while ((rtg = xfs_rtgroup_next(mp, rtg))) { |
| error = xfs_init_zone(&iz, rtg, NULL); |
| if (error) |
| goto out_free_zone_info; |
| } |
| } |
| |
| xfs_set_freecounter(mp, XC_FREE_RTAVAILABLE, iz.available); |
| xfs_set_freecounter(mp, XC_FREE_RTEXTENTS, |
| iz.available + iz.reclaimable); |
| |
| /* |
| * The user may configure GC to free up a percentage of unused blocks. |
| * By default this is 0. GC will always trigger at the minimum level |
| * for keeping max_open_zones available for data placement. |
| */ |
| mp->m_zonegc_low_space = 0; |
| |
| error = xfs_zone_gc_mount(mp); |
| if (error) |
| goto out_free_zone_info; |
| |
| /* |
| * Set up a mru cache to track inode to open zone for data placement |
| * purposes. The magic values for group count and life time is the |
| * same as the defaults for file streams, which seems sane enough. |
| */ |
| xfs_mru_cache_create(&mp->m_zone_cache, mp, |
| 5000, 10, xfs_zone_cache_free_func); |
| return 0; |
| |
| out_free_zone_info: |
| xfs_free_zone_info(mp->m_zone_info); |
| return error; |
| } |
| |
| void |
| xfs_unmount_zones( |
| struct xfs_mount *mp) |
| { |
| xfs_zone_gc_unmount(mp); |
| xfs_free_zone_info(mp->m_zone_info); |
| xfs_mru_cache_destroy(mp->m_zone_cache); |
| } |