| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Copyright 2023 Red Hat |
| */ |
| |
| /* |
| * This file contains the main entry points for normal operations on a vdo as well as functions for |
| * constructing and destroying vdo instances (in memory). |
| */ |
| |
| /** |
| * DOC: |
| * |
| * A read_only_notifier has a single completion which is used to perform read-only notifications, |
| * however, vdo_enter_read_only_mode() may be called from any thread. A pair of fields, protected |
| * by a spinlock, are used to control the read-only mode entry process. The first field holds the |
| * read-only error. The second is the state field, which may hold any of the four special values |
| * enumerated here. |
| * |
| * When vdo_enter_read_only_mode() is called from some vdo thread, if the read_only_error field |
| * already contains an error (i.e. its value is not VDO_SUCCESS), then some other error has already |
| * initiated the read-only process, and nothing more is done. Otherwise, the new error is stored in |
| * the read_only_error field, and the state field is consulted. If the state is MAY_NOTIFY, it is |
| * set to NOTIFYING, and the notification process begins. If the state is MAY_NOT_NOTIFY, then |
| * notifications are currently disallowed, generally due to the vdo being suspended. In this case, |
| * the nothing more will be done until the vdo is resumed, at which point the notification will be |
| * performed. In any other case, the vdo is already read-only, and there is nothing more to do. |
| */ |
| |
| #include "vdo.h" |
| |
| #include <linux/completion.h> |
| #include <linux/device-mapper.h> |
| #include <linux/kernel.h> |
| #include <linux/lz4.h> |
| #include <linux/module.h> |
| #include <linux/mutex.h> |
| #include <linux/spinlock.h> |
| #include <linux/types.h> |
| |
| #include "logger.h" |
| #include "memory-alloc.h" |
| #include "permassert.h" |
| #include "string-utils.h" |
| |
| #include "block-map.h" |
| #include "completion.h" |
| #include "data-vio.h" |
| #include "dedupe.h" |
| #include "encodings.h" |
| #include "funnel-workqueue.h" |
| #include "io-submitter.h" |
| #include "logical-zone.h" |
| #include "packer.h" |
| #include "physical-zone.h" |
| #include "recovery-journal.h" |
| #include "slab-depot.h" |
| #include "statistics.h" |
| #include "status-codes.h" |
| #include "vio.h" |
| |
| #define PARANOID_THREAD_CONSISTENCY_CHECKS 0 |
| |
| struct sync_completion { |
| struct vdo_completion vdo_completion; |
| struct completion completion; |
| }; |
| |
| /* A linked list is adequate for the small number of entries we expect. */ |
| struct device_registry { |
| struct list_head links; |
| /* TODO: Convert to rcu per kernel recommendation. */ |
| rwlock_t lock; |
| }; |
| |
| static struct device_registry registry; |
| |
| /** |
| * vdo_initialize_device_registry_once() - Initialize the necessary structures for the device |
| * registry. |
| */ |
| void vdo_initialize_device_registry_once(void) |
| { |
| INIT_LIST_HEAD(®istry.links); |
| rwlock_init(®istry.lock); |
| } |
| |
| /** vdo_is_equal() - Implements vdo_filter_fn. */ |
| static bool vdo_is_equal(struct vdo *vdo, const void *context) |
| { |
| return (vdo == context); |
| } |
| |
| /** |
| * filter_vdos_locked() - Find a vdo in the registry if it exists there. |
| * @filter: The filter function to apply to devices. |
| * @context: A bit of context to provide the filter. |
| * |
| * Context: Must be called holding the lock. |
| * |
| * Return: the vdo object found, if any. |
| */ |
| static struct vdo * __must_check filter_vdos_locked(vdo_filter_fn filter, |
| const void *context) |
| { |
| struct vdo *vdo; |
| |
| list_for_each_entry(vdo, ®istry.links, registration) { |
| if (filter(vdo, context)) |
| return vdo; |
| } |
| |
| return NULL; |
| } |
| |
| /** |
| * vdo_find_matching() - Find and return the first (if any) vdo matching a given filter function. |
| * @filter: The filter function to apply to vdos. |
| * @context: A bit of context to provide the filter. |
| */ |
| struct vdo *vdo_find_matching(vdo_filter_fn filter, const void *context) |
| { |
| struct vdo *vdo; |
| |
| read_lock(®istry.lock); |
| vdo = filter_vdos_locked(filter, context); |
| read_unlock(®istry.lock); |
| |
| return vdo; |
| } |
| |
| static void start_vdo_request_queue(void *ptr) |
| { |
| struct vdo_thread *thread = vdo_get_work_queue_owner(vdo_get_current_work_queue()); |
| |
| vdo_register_allocating_thread(&thread->allocating_thread, |
| &thread->vdo->allocations_allowed); |
| } |
| |
| static void finish_vdo_request_queue(void *ptr) |
| { |
| vdo_unregister_allocating_thread(); |
| } |
| |
| #ifdef MODULE |
| #define MODULE_NAME THIS_MODULE->name |
| #else |
| #define MODULE_NAME "dm-vdo" |
| #endif /* MODULE */ |
| |
| static const struct vdo_work_queue_type default_queue_type = { |
| .start = start_vdo_request_queue, |
| .finish = finish_vdo_request_queue, |
| .max_priority = VDO_DEFAULT_Q_MAX_PRIORITY, |
| .default_priority = VDO_DEFAULT_Q_COMPLETION_PRIORITY, |
| }; |
| |
| static const struct vdo_work_queue_type bio_ack_q_type = { |
| .start = NULL, |
| .finish = NULL, |
| .max_priority = BIO_ACK_Q_MAX_PRIORITY, |
| .default_priority = BIO_ACK_Q_ACK_PRIORITY, |
| }; |
| |
| static const struct vdo_work_queue_type cpu_q_type = { |
| .start = NULL, |
| .finish = NULL, |
| .max_priority = CPU_Q_MAX_PRIORITY, |
| .default_priority = CPU_Q_MAX_PRIORITY, |
| }; |
| |
| static void uninitialize_thread_config(struct thread_config *config) |
| { |
| vdo_free(vdo_forget(config->logical_threads)); |
| vdo_free(vdo_forget(config->physical_threads)); |
| vdo_free(vdo_forget(config->hash_zone_threads)); |
| vdo_free(vdo_forget(config->bio_threads)); |
| memset(config, 0, sizeof(struct thread_config)); |
| } |
| |
| static void assign_thread_ids(struct thread_config *config, |
| thread_id_t thread_ids[], zone_count_t count) |
| { |
| zone_count_t zone; |
| |
| for (zone = 0; zone < count; zone++) |
| thread_ids[zone] = config->thread_count++; |
| } |
| |
| /** |
| * initialize_thread_config() - Initialize the thread mapping |
| * |
| * If the logical, physical, and hash zone counts are all 0, a single thread will be shared by all |
| * three plus the packer and recovery journal. Otherwise, there must be at least one of each type, |
| * and each will have its own thread, as will the packer and recovery journal. |
| * |
| * Return: VDO_SUCCESS or an error. |
| */ |
| static int __must_check initialize_thread_config(struct thread_count_config counts, |
| struct thread_config *config) |
| { |
| int result; |
| bool single = ((counts.logical_zones + counts.physical_zones + counts.hash_zones) == 0); |
| |
| config->bio_thread_count = counts.bio_threads; |
| if (single) { |
| config->logical_zone_count = 1; |
| config->physical_zone_count = 1; |
| config->hash_zone_count = 1; |
| } else { |
| config->logical_zone_count = counts.logical_zones; |
| config->physical_zone_count = counts.physical_zones; |
| config->hash_zone_count = counts.hash_zones; |
| } |
| |
| result = vdo_allocate(config->logical_zone_count, thread_id_t, |
| "logical thread array", &config->logical_threads); |
| if (result != VDO_SUCCESS) { |
| uninitialize_thread_config(config); |
| return result; |
| } |
| |
| result = vdo_allocate(config->physical_zone_count, thread_id_t, |
| "physical thread array", &config->physical_threads); |
| if (result != VDO_SUCCESS) { |
| uninitialize_thread_config(config); |
| return result; |
| } |
| |
| result = vdo_allocate(config->hash_zone_count, thread_id_t, |
| "hash thread array", &config->hash_zone_threads); |
| if (result != VDO_SUCCESS) { |
| uninitialize_thread_config(config); |
| return result; |
| } |
| |
| result = vdo_allocate(config->bio_thread_count, thread_id_t, |
| "bio thread array", &config->bio_threads); |
| if (result != VDO_SUCCESS) { |
| uninitialize_thread_config(config); |
| return result; |
| } |
| |
| if (single) { |
| config->logical_threads[0] = config->thread_count; |
| config->physical_threads[0] = config->thread_count; |
| config->hash_zone_threads[0] = config->thread_count++; |
| } else { |
| config->admin_thread = config->thread_count; |
| config->journal_thread = config->thread_count++; |
| config->packer_thread = config->thread_count++; |
| assign_thread_ids(config, config->logical_threads, counts.logical_zones); |
| assign_thread_ids(config, config->physical_threads, counts.physical_zones); |
| assign_thread_ids(config, config->hash_zone_threads, counts.hash_zones); |
| } |
| |
| config->dedupe_thread = config->thread_count++; |
| config->bio_ack_thread = |
| ((counts.bio_ack_threads > 0) ? config->thread_count++ : VDO_INVALID_THREAD_ID); |
| config->cpu_thread = config->thread_count++; |
| assign_thread_ids(config, config->bio_threads, counts.bio_threads); |
| return VDO_SUCCESS; |
| } |
| |
| /** |
| * read_geometry_block() - Synchronously read the geometry block from a vdo's underlying block |
| * device. |
| * @vdo: The vdo whose geometry is to be read. |
| * |
| * Return: VDO_SUCCESS or an error code. |
| */ |
| static int __must_check read_geometry_block(struct vdo *vdo) |
| { |
| struct vio *vio; |
| char *block; |
| int result; |
| |
| result = vdo_allocate(VDO_BLOCK_SIZE, u8, __func__, &block); |
| if (result != VDO_SUCCESS) |
| return result; |
| |
| result = create_metadata_vio(vdo, VIO_TYPE_GEOMETRY, VIO_PRIORITY_HIGH, NULL, |
| block, &vio); |
| if (result != VDO_SUCCESS) { |
| vdo_free(block); |
| return result; |
| } |
| |
| /* |
| * This is only safe because, having not already loaded the geometry, the vdo's geometry's |
| * bio_offset field is 0, so the fact that vio_reset_bio() will subtract that offset from |
| * the supplied pbn is not a problem. |
| */ |
| result = vio_reset_bio(vio, block, NULL, REQ_OP_READ, |
| VDO_GEOMETRY_BLOCK_LOCATION); |
| if (result != VDO_SUCCESS) { |
| free_vio(vdo_forget(vio)); |
| vdo_free(block); |
| return result; |
| } |
| |
| bio_set_dev(vio->bio, vdo_get_backing_device(vdo)); |
| submit_bio_wait(vio->bio); |
| result = blk_status_to_errno(vio->bio->bi_status); |
| free_vio(vdo_forget(vio)); |
| if (result != 0) { |
| vdo_log_error_strerror(result, "synchronous read failed"); |
| vdo_free(block); |
| return -EIO; |
| } |
| |
| result = vdo_parse_geometry_block((u8 *) block, &vdo->geometry); |
| vdo_free(block); |
| return result; |
| } |
| |
| static bool get_zone_thread_name(const thread_id_t thread_ids[], zone_count_t count, |
| thread_id_t id, const char *prefix, |
| char *buffer, size_t buffer_length) |
| { |
| if (id >= thread_ids[0]) { |
| thread_id_t index = id - thread_ids[0]; |
| |
| if (index < count) { |
| snprintf(buffer, buffer_length, "%s%d", prefix, index); |
| return true; |
| } |
| } |
| |
| return false; |
| } |
| |
| /** |
| * get_thread_name() - Format the name of the worker thread desired to support a given work queue. |
| * @thread_config: The thread configuration. |
| * @thread_id: The thread id. |
| * @buffer: Where to put the formatted name. |
| * @buffer_length: Size of the output buffer. |
| * |
| * The physical layer may add a prefix identifying the product; the output from this function |
| * should just identify the thread. |
| */ |
| static void get_thread_name(const struct thread_config *thread_config, |
| thread_id_t thread_id, char *buffer, size_t buffer_length) |
| { |
| if (thread_id == thread_config->journal_thread) { |
| if (thread_config->packer_thread == thread_id) { |
| /* |
| * This is the "single thread" config where one thread is used for the |
| * journal, packer, logical, physical, and hash zones. In that case, it is |
| * known as the "request queue." |
| */ |
| snprintf(buffer, buffer_length, "reqQ"); |
| return; |
| } |
| |
| snprintf(buffer, buffer_length, "journalQ"); |
| return; |
| } else if (thread_id == thread_config->admin_thread) { |
| /* Theoretically this could be different from the journal thread. */ |
| snprintf(buffer, buffer_length, "adminQ"); |
| return; |
| } else if (thread_id == thread_config->packer_thread) { |
| snprintf(buffer, buffer_length, "packerQ"); |
| return; |
| } else if (thread_id == thread_config->dedupe_thread) { |
| snprintf(buffer, buffer_length, "dedupeQ"); |
| return; |
| } else if (thread_id == thread_config->bio_ack_thread) { |
| snprintf(buffer, buffer_length, "ackQ"); |
| return; |
| } else if (thread_id == thread_config->cpu_thread) { |
| snprintf(buffer, buffer_length, "cpuQ"); |
| return; |
| } |
| |
| if (get_zone_thread_name(thread_config->logical_threads, |
| thread_config->logical_zone_count, |
| thread_id, "logQ", buffer, buffer_length)) |
| return; |
| |
| if (get_zone_thread_name(thread_config->physical_threads, |
| thread_config->physical_zone_count, |
| thread_id, "physQ", buffer, buffer_length)) |
| return; |
| |
| if (get_zone_thread_name(thread_config->hash_zone_threads, |
| thread_config->hash_zone_count, |
| thread_id, "hashQ", buffer, buffer_length)) |
| return; |
| |
| if (get_zone_thread_name(thread_config->bio_threads, |
| thread_config->bio_thread_count, |
| thread_id, "bioQ", buffer, buffer_length)) |
| return; |
| |
| /* Some sort of misconfiguration? */ |
| snprintf(buffer, buffer_length, "reqQ%d", thread_id); |
| } |
| |
| /** |
| * vdo_make_thread() - Construct a single vdo work_queue and its associated thread (or threads for |
| * round-robin queues). |
| * @vdo: The vdo which owns the thread. |
| * @thread_id: The id of the thread to create (as determined by the thread_config). |
| * @type: The description of the work queue for this thread. |
| * @queue_count: The number of actual threads/queues contained in the "thread". |
| * @contexts: An array of queue_count contexts, one for each individual queue; may be NULL. |
| * |
| * Each "thread" constructed by this method is represented by a unique thread id in the thread |
| * config, and completions can be enqueued to the queue and run on the threads comprising this |
| * entity. |
| * |
| * Return: VDO_SUCCESS or an error. |
| */ |
| int vdo_make_thread(struct vdo *vdo, thread_id_t thread_id, |
| const struct vdo_work_queue_type *type, |
| unsigned int queue_count, void *contexts[]) |
| { |
| struct vdo_thread *thread = &vdo->threads[thread_id]; |
| char queue_name[MAX_VDO_WORK_QUEUE_NAME_LEN]; |
| |
| if (type == NULL) |
| type = &default_queue_type; |
| |
| if (thread->queue != NULL) { |
| return VDO_ASSERT(vdo_work_queue_type_is(thread->queue, type), |
| "already constructed vdo thread %u is of the correct type", |
| thread_id); |
| } |
| |
| thread->vdo = vdo; |
| thread->thread_id = thread_id; |
| get_thread_name(&vdo->thread_config, thread_id, queue_name, sizeof(queue_name)); |
| return vdo_make_work_queue(vdo->thread_name_prefix, queue_name, thread, |
| type, queue_count, contexts, &thread->queue); |
| } |
| |
| /** |
| * register_vdo() - Register a VDO; it must not already be registered. |
| * @vdo: The vdo to register. |
| * |
| * Return: VDO_SUCCESS or an error. |
| */ |
| static int register_vdo(struct vdo *vdo) |
| { |
| int result; |
| |
| write_lock(®istry.lock); |
| result = VDO_ASSERT(filter_vdos_locked(vdo_is_equal, vdo) == NULL, |
| "VDO not already registered"); |
| if (result == VDO_SUCCESS) { |
| INIT_LIST_HEAD(&vdo->registration); |
| list_add_tail(&vdo->registration, ®istry.links); |
| } |
| write_unlock(®istry.lock); |
| |
| return result; |
| } |
| |
| /** |
| * initialize_vdo() - Do the portion of initializing a vdo which will clean up after itself on |
| * error. |
| * @vdo: The vdo being initialized |
| * @config: The configuration of the vdo |
| * @instance: The instance number of the vdo |
| * @reason: The buffer to hold the failure reason on error |
| */ |
| static int initialize_vdo(struct vdo *vdo, struct device_config *config, |
| unsigned int instance, char **reason) |
| { |
| int result; |
| zone_count_t i; |
| |
| vdo->device_config = config; |
| vdo->starting_sector_offset = config->owning_target->begin; |
| vdo->instance = instance; |
| vdo->allocations_allowed = true; |
| vdo_set_admin_state_code(&vdo->admin.state, VDO_ADMIN_STATE_NEW); |
| INIT_LIST_HEAD(&vdo->device_config_list); |
| vdo_initialize_completion(&vdo->admin.completion, vdo, VDO_ADMIN_COMPLETION); |
| init_completion(&vdo->admin.callback_sync); |
| mutex_init(&vdo->stats_mutex); |
| result = read_geometry_block(vdo); |
| if (result != VDO_SUCCESS) { |
| *reason = "Could not load geometry block"; |
| return result; |
| } |
| |
| result = initialize_thread_config(config->thread_counts, &vdo->thread_config); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot create thread configuration"; |
| return result; |
| } |
| |
| vdo_log_info("zones: %d logical, %d physical, %d hash; total threads: %d", |
| config->thread_counts.logical_zones, |
| config->thread_counts.physical_zones, |
| config->thread_counts.hash_zones, vdo->thread_config.thread_count); |
| |
| /* Compression context storage */ |
| result = vdo_allocate(config->thread_counts.cpu_threads, char *, "LZ4 context", |
| &vdo->compression_context); |
| if (result != VDO_SUCCESS) { |
| *reason = "cannot allocate LZ4 context"; |
| return result; |
| } |
| |
| for (i = 0; i < config->thread_counts.cpu_threads; i++) { |
| result = vdo_allocate(LZ4_MEM_COMPRESS, char, "LZ4 context", |
| &vdo->compression_context[i]); |
| if (result != VDO_SUCCESS) { |
| *reason = "cannot allocate LZ4 context"; |
| return result; |
| } |
| } |
| |
| result = register_vdo(vdo); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot add VDO to device registry"; |
| return result; |
| } |
| |
| vdo_set_admin_state_code(&vdo->admin.state, VDO_ADMIN_STATE_INITIALIZED); |
| return result; |
| } |
| |
| /** |
| * vdo_make() - Allocate and initialize a vdo. |
| * @instance: Device instantiation counter. |
| * @config: The device configuration. |
| * @reason: The reason for any failure during this call. |
| * @vdo_ptr: A pointer to hold the created vdo. |
| * |
| * Return: VDO_SUCCESS or an error. |
| */ |
| int vdo_make(unsigned int instance, struct device_config *config, char **reason, |
| struct vdo **vdo_ptr) |
| { |
| int result; |
| struct vdo *vdo; |
| |
| /* Initialize with a generic failure reason to prevent returning garbage. */ |
| *reason = "Unspecified error"; |
| |
| result = vdo_allocate(1, struct vdo, __func__, &vdo); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot allocate VDO"; |
| return result; |
| } |
| |
| result = initialize_vdo(vdo, config, instance, reason); |
| if (result != VDO_SUCCESS) { |
| vdo_destroy(vdo); |
| return result; |
| } |
| |
| /* From here on, the caller will clean up if there is an error. */ |
| *vdo_ptr = vdo; |
| |
| snprintf(vdo->thread_name_prefix, sizeof(vdo->thread_name_prefix), |
| "%s%u", MODULE_NAME, instance); |
| BUG_ON(vdo->thread_name_prefix[0] == '\0'); |
| result = vdo_allocate(vdo->thread_config.thread_count, |
| struct vdo_thread, __func__, &vdo->threads); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot allocate thread structures"; |
| return result; |
| } |
| |
| result = vdo_make_thread(vdo, vdo->thread_config.admin_thread, |
| &default_queue_type, 1, NULL); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot make admin thread"; |
| return result; |
| } |
| |
| result = vdo_make_flusher(vdo); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot make flusher zones"; |
| return result; |
| } |
| |
| result = vdo_make_packer(vdo, DEFAULT_PACKER_BINS, &vdo->packer); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot make packer zones"; |
| return result; |
| } |
| |
| BUG_ON(vdo->device_config->logical_block_size <= 0); |
| BUG_ON(vdo->device_config->owned_device == NULL); |
| result = make_data_vio_pool(vdo, MAXIMUM_VDO_USER_VIOS, |
| MAXIMUM_VDO_USER_VIOS * 3 / 4, |
| &vdo->data_vio_pool); |
| if (result != VDO_SUCCESS) { |
| *reason = "Cannot allocate data_vio pool"; |
| return result; |
| } |
| |
| result = vdo_make_io_submitter(config->thread_counts.bio_threads, |
| config->thread_counts.bio_rotation_interval, |
| get_data_vio_pool_request_limit(vdo->data_vio_pool), |
| vdo, &vdo->io_submitter); |
| if (result != VDO_SUCCESS) { |
| *reason = "bio submission initialization failed"; |
| return result; |
| } |
| |
| if (vdo_uses_bio_ack_queue(vdo)) { |
| result = vdo_make_thread(vdo, vdo->thread_config.bio_ack_thread, |
| &bio_ack_q_type, |
| config->thread_counts.bio_ack_threads, NULL); |
| if (result != VDO_SUCCESS) { |
| *reason = "bio ack queue initialization failed"; |
| return result; |
| } |
| } |
| |
| result = vdo_make_thread(vdo, vdo->thread_config.cpu_thread, &cpu_q_type, |
| config->thread_counts.cpu_threads, |
| (void **) vdo->compression_context); |
| if (result != VDO_SUCCESS) { |
| *reason = "CPU queue initialization failed"; |
| return result; |
| } |
| |
| return VDO_SUCCESS; |
| } |
| |
| static void finish_vdo(struct vdo *vdo) |
| { |
| int i; |
| |
| if (vdo->threads == NULL) |
| return; |
| |
| vdo_cleanup_io_submitter(vdo->io_submitter); |
| vdo_finish_dedupe_index(vdo->hash_zones); |
| |
| for (i = 0; i < vdo->thread_config.thread_count; i++) |
| vdo_finish_work_queue(vdo->threads[i].queue); |
| } |
| |
| /** |
| * free_listeners() - Free the list of read-only listeners associated with a thread. |
| * @thread_data: The thread holding the list to free. |
| */ |
| static void free_listeners(struct vdo_thread *thread) |
| { |
| struct read_only_listener *listener, *next; |
| |
| for (listener = vdo_forget(thread->listeners); listener != NULL; listener = next) { |
| next = vdo_forget(listener->next); |
| vdo_free(listener); |
| } |
| } |
| |
| static void uninitialize_super_block(struct vdo_super_block *super_block) |
| { |
| free_vio_components(&super_block->vio); |
| vdo_free(super_block->buffer); |
| } |
| |
| /** |
| * unregister_vdo() - Remove a vdo from the device registry. |
| * @vdo: The vdo to remove. |
| */ |
| static void unregister_vdo(struct vdo *vdo) |
| { |
| write_lock(®istry.lock); |
| if (filter_vdos_locked(vdo_is_equal, vdo) == vdo) |
| list_del_init(&vdo->registration); |
| |
| write_unlock(®istry.lock); |
| } |
| |
| /** |
| * vdo_destroy() - Destroy a vdo instance. |
| * @vdo: The vdo to destroy (may be NULL). |
| */ |
| void vdo_destroy(struct vdo *vdo) |
| { |
| unsigned int i; |
| |
| if (vdo == NULL) |
| return; |
| |
| /* A running VDO should never be destroyed without suspending first. */ |
| BUG_ON(vdo_get_admin_state(vdo)->normal); |
| |
| vdo->allocations_allowed = true; |
| |
| finish_vdo(vdo); |
| unregister_vdo(vdo); |
| free_data_vio_pool(vdo->data_vio_pool); |
| vdo_free_io_submitter(vdo_forget(vdo->io_submitter)); |
| vdo_free_flusher(vdo_forget(vdo->flusher)); |
| vdo_free_packer(vdo_forget(vdo->packer)); |
| vdo_free_recovery_journal(vdo_forget(vdo->recovery_journal)); |
| vdo_free_slab_depot(vdo_forget(vdo->depot)); |
| vdo_uninitialize_layout(&vdo->layout); |
| vdo_uninitialize_layout(&vdo->next_layout); |
| if (vdo->partition_copier) |
| dm_kcopyd_client_destroy(vdo_forget(vdo->partition_copier)); |
| uninitialize_super_block(&vdo->super_block); |
| vdo_free_block_map(vdo_forget(vdo->block_map)); |
| vdo_free_hash_zones(vdo_forget(vdo->hash_zones)); |
| vdo_free_physical_zones(vdo_forget(vdo->physical_zones)); |
| vdo_free_logical_zones(vdo_forget(vdo->logical_zones)); |
| |
| if (vdo->threads != NULL) { |
| for (i = 0; i < vdo->thread_config.thread_count; i++) { |
| free_listeners(&vdo->threads[i]); |
| vdo_free_work_queue(vdo_forget(vdo->threads[i].queue)); |
| } |
| vdo_free(vdo_forget(vdo->threads)); |
| } |
| |
| uninitialize_thread_config(&vdo->thread_config); |
| |
| if (vdo->compression_context != NULL) { |
| for (i = 0; i < vdo->device_config->thread_counts.cpu_threads; i++) |
| vdo_free(vdo_forget(vdo->compression_context[i])); |
| |
| vdo_free(vdo_forget(vdo->compression_context)); |
| } |
| vdo_free(vdo); |
| } |
| |
| static int initialize_super_block(struct vdo *vdo, struct vdo_super_block *super_block) |
| { |
| int result; |
| |
| result = vdo_allocate(VDO_BLOCK_SIZE, char, "encoded super block", |
| (char **) &vdo->super_block.buffer); |
| if (result != VDO_SUCCESS) |
| return result; |
| |
| return allocate_vio_components(vdo, VIO_TYPE_SUPER_BLOCK, |
| VIO_PRIORITY_METADATA, NULL, 1, |
| (char *) super_block->buffer, |
| &vdo->super_block.vio); |
| } |
| |
| /** |
| * finish_reading_super_block() - Continue after loading the super block. |
| * @completion: The super block vio. |
| * |
| * This callback is registered in vdo_load_super_block(). |
| */ |
| static void finish_reading_super_block(struct vdo_completion *completion) |
| { |
| struct vdo_super_block *super_block = |
| container_of(as_vio(completion), struct vdo_super_block, vio); |
| |
| vdo_continue_completion(vdo_forget(completion->parent), |
| vdo_decode_super_block(super_block->buffer)); |
| } |
| |
| /** |
| * handle_super_block_read_error() - Handle an error reading the super block. |
| * @completion: The super block vio. |
| * |
| * This error handler is registered in vdo_load_super_block(). |
| */ |
| static void handle_super_block_read_error(struct vdo_completion *completion) |
| { |
| vio_record_metadata_io_error(as_vio(completion)); |
| finish_reading_super_block(completion); |
| } |
| |
| static void read_super_block_endio(struct bio *bio) |
| { |
| struct vio *vio = bio->bi_private; |
| struct vdo_completion *parent = vio->completion.parent; |
| |
| continue_vio_after_io(vio, finish_reading_super_block, |
| parent->callback_thread_id); |
| } |
| |
| /** |
| * vdo_load_super_block() - Allocate a super block and read its contents from storage. |
| * @vdo: The vdo containing the super block on disk. |
| * @parent: The completion to notify after loading the super block. |
| */ |
| void vdo_load_super_block(struct vdo *vdo, struct vdo_completion *parent) |
| { |
| int result; |
| |
| result = initialize_super_block(vdo, &vdo->super_block); |
| if (result != VDO_SUCCESS) { |
| vdo_continue_completion(parent, result); |
| return; |
| } |
| |
| vdo->super_block.vio.completion.parent = parent; |
| vdo_submit_metadata_vio(&vdo->super_block.vio, |
| vdo_get_data_region_start(vdo->geometry), |
| read_super_block_endio, |
| handle_super_block_read_error, |
| REQ_OP_READ); |
| } |
| |
| /** |
| * vdo_get_backing_device() - Get the block device object underlying a vdo. |
| * @vdo: The vdo. |
| * |
| * Return: The vdo's current block device. |
| */ |
| struct block_device *vdo_get_backing_device(const struct vdo *vdo) |
| { |
| return vdo->device_config->owned_device->bdev; |
| } |
| |
| /** |
| * vdo_get_device_name() - Get the device name associated with the vdo target. |
| * @target: The target device interface. |
| * |
| * Return: The block device name. |
| */ |
| const char *vdo_get_device_name(const struct dm_target *target) |
| { |
| return dm_device_name(dm_table_get_md(target->table)); |
| } |
| |
| /** |
| * vdo_synchronous_flush() - Issue a flush request and wait for it to complete. |
| * @vdo: The vdo. |
| * |
| * Return: VDO_SUCCESS or an error. |
| */ |
| int vdo_synchronous_flush(struct vdo *vdo) |
| { |
| int result; |
| struct bio bio; |
| |
| bio_init(&bio, vdo_get_backing_device(vdo), NULL, 0, |
| REQ_OP_WRITE | REQ_PREFLUSH); |
| submit_bio_wait(&bio); |
| result = blk_status_to_errno(bio.bi_status); |
| |
| atomic64_inc(&vdo->stats.flush_out); |
| if (result != 0) { |
| vdo_log_error_strerror(result, "synchronous flush failed"); |
| result = -EIO; |
| } |
| |
| bio_uninit(&bio); |
| return result; |
| } |
| |
| /** |
| * vdo_get_state() - Get the current state of the vdo. |
| * @vdo: The vdo. |
| |
| * Context: This method may be called from any thread. |
| * |
| * Return: The current state of the vdo. |
| */ |
| enum vdo_state vdo_get_state(const struct vdo *vdo) |
| { |
| enum vdo_state state = atomic_read(&vdo->state); |
| |
| /* pairs with barriers where state field is changed */ |
| smp_rmb(); |
| return state; |
| } |
| |
| /** |
| * vdo_set_state() - Set the current state of the vdo. |
| * @vdo: The vdo whose state is to be set. |
| * @state: The new state of the vdo. |
| * |
| * Context: This method may be called from any thread. |
| */ |
| void vdo_set_state(struct vdo *vdo, enum vdo_state state) |
| { |
| /* pairs with barrier in vdo_get_state */ |
| smp_wmb(); |
| atomic_set(&vdo->state, state); |
| } |
| |
| /** |
| * vdo_get_admin_state() - Get the admin state of the vdo. |
| * @vdo: The vdo. |
| * |
| * Return: The code for the vdo's current admin state. |
| */ |
| const struct admin_state_code *vdo_get_admin_state(const struct vdo *vdo) |
| { |
| return vdo_get_admin_state_code(&vdo->admin.state); |
| } |
| |
| /** |
| * record_vdo() - Record the state of the VDO for encoding in the super block. |
| */ |
| static void record_vdo(struct vdo *vdo) |
| { |
| /* This is for backwards compatibility. */ |
| vdo->states.unused = vdo->geometry.unused; |
| vdo->states.vdo.state = vdo_get_state(vdo); |
| vdo->states.block_map = vdo_record_block_map(vdo->block_map); |
| vdo->states.recovery_journal = vdo_record_recovery_journal(vdo->recovery_journal); |
| vdo->states.slab_depot = vdo_record_slab_depot(vdo->depot); |
| vdo->states.layout = vdo->layout; |
| } |
| |
| /** |
| * continue_super_block_parent() - Continue the parent of a super block save operation. |
| * @completion: The super block vio. |
| * |
| * This callback is registered in vdo_save_components(). |
| */ |
| static void continue_super_block_parent(struct vdo_completion *completion) |
| { |
| vdo_continue_completion(vdo_forget(completion->parent), completion->result); |
| } |
| |
| /** |
| * handle_save_error() - Log a super block save error. |
| * @completion: The super block vio. |
| * |
| * This error handler is registered in vdo_save_components(). |
| */ |
| static void handle_save_error(struct vdo_completion *completion) |
| { |
| struct vdo_super_block *super_block = |
| container_of(as_vio(completion), struct vdo_super_block, vio); |
| |
| vio_record_metadata_io_error(&super_block->vio); |
| vdo_log_error_strerror(completion->result, "super block save failed"); |
| /* |
| * Mark the super block as unwritable so that we won't attempt to write it again. This |
| * avoids the case where a growth attempt fails writing the super block with the new size, |
| * but the subsequent attempt to write out the read-only state succeeds. In this case, |
| * writes which happened just before the suspend would not be visible if the VDO is |
| * restarted without rebuilding, but, after a read-only rebuild, the effects of those |
| * writes would reappear. |
| */ |
| super_block->unwritable = true; |
| completion->callback(completion); |
| } |
| |
| static void super_block_write_endio(struct bio *bio) |
| { |
| struct vio *vio = bio->bi_private; |
| struct vdo_completion *parent = vio->completion.parent; |
| |
| continue_vio_after_io(vio, continue_super_block_parent, |
| parent->callback_thread_id); |
| } |
| |
| /** |
| * vdo_save_components() - Encode the vdo and save the super block asynchronously. |
| * @vdo: The vdo whose state is being saved. |
| * @parent: The completion to notify when the save is complete. |
| */ |
| void vdo_save_components(struct vdo *vdo, struct vdo_completion *parent) |
| { |
| struct vdo_super_block *super_block = &vdo->super_block; |
| |
| if (super_block->unwritable) { |
| vdo_continue_completion(parent, VDO_READ_ONLY); |
| return; |
| } |
| |
| if (super_block->vio.completion.parent != NULL) { |
| vdo_continue_completion(parent, VDO_COMPONENT_BUSY); |
| return; |
| } |
| |
| record_vdo(vdo); |
| |
| vdo_encode_super_block(super_block->buffer, &vdo->states); |
| super_block->vio.completion.parent = parent; |
| super_block->vio.completion.callback_thread_id = parent->callback_thread_id; |
| vdo_submit_metadata_vio(&super_block->vio, |
| vdo_get_data_region_start(vdo->geometry), |
| super_block_write_endio, handle_save_error, |
| REQ_OP_WRITE | REQ_PREFLUSH | REQ_FUA); |
| } |
| |
| /** |
| * vdo_register_read_only_listener() - Register a listener to be notified when the VDO goes |
| * read-only. |
| * @vdo: The vdo to register with. |
| * @listener: The object to notify. |
| * @notification: The function to call to send the notification. |
| * @thread_id: The id of the thread on which to send the notification. |
| * |
| * Return: VDO_SUCCESS or an error. |
| */ |
| int vdo_register_read_only_listener(struct vdo *vdo, void *listener, |
| vdo_read_only_notification_fn notification, |
| thread_id_t thread_id) |
| { |
| struct vdo_thread *thread = &vdo->threads[thread_id]; |
| struct read_only_listener *read_only_listener; |
| int result; |
| |
| result = VDO_ASSERT(thread_id != vdo->thread_config.dedupe_thread, |
| "read only listener not registered on dedupe thread"); |
| if (result != VDO_SUCCESS) |
| return result; |
| |
| result = vdo_allocate(1, struct read_only_listener, __func__, |
| &read_only_listener); |
| if (result != VDO_SUCCESS) |
| return result; |
| |
| *read_only_listener = (struct read_only_listener) { |
| .listener = listener, |
| .notify = notification, |
| .next = thread->listeners, |
| }; |
| |
| thread->listeners = read_only_listener; |
| return VDO_SUCCESS; |
| } |
| |
| /** |
| * notify_vdo_of_read_only_mode() - Notify a vdo that it is going read-only. |
| * @listener: The vdo. |
| * @parent: The completion to notify in order to acknowledge the notification. |
| * |
| * This will save the read-only state to the super block. |
| * |
| * Implements vdo_read_only_notification_fn. |
| */ |
| static void notify_vdo_of_read_only_mode(void *listener, struct vdo_completion *parent) |
| { |
| struct vdo *vdo = listener; |
| |
| if (vdo_in_read_only_mode(vdo)) |
| vdo_finish_completion(parent); |
| |
| vdo_set_state(vdo, VDO_READ_ONLY_MODE); |
| vdo_save_components(vdo, parent); |
| } |
| |
| /** |
| * vdo_enable_read_only_entry() - Enable a vdo to enter read-only mode on errors. |
| * @vdo: The vdo to enable. |
| * |
| * Return: VDO_SUCCESS or an error. |
| */ |
| int vdo_enable_read_only_entry(struct vdo *vdo) |
| { |
| thread_id_t id; |
| bool is_read_only = vdo_in_read_only_mode(vdo); |
| struct read_only_notifier *notifier = &vdo->read_only_notifier; |
| |
| if (is_read_only) { |
| notifier->read_only_error = VDO_READ_ONLY; |
| notifier->state = NOTIFIED; |
| } else { |
| notifier->state = MAY_NOT_NOTIFY; |
| } |
| |
| spin_lock_init(¬ifier->lock); |
| vdo_initialize_completion(¬ifier->completion, vdo, |
| VDO_READ_ONLY_MODE_COMPLETION); |
| |
| for (id = 0; id < vdo->thread_config.thread_count; id++) |
| vdo->threads[id].is_read_only = is_read_only; |
| |
| return vdo_register_read_only_listener(vdo, vdo, notify_vdo_of_read_only_mode, |
| vdo->thread_config.admin_thread); |
| } |
| |
| /** |
| * vdo_wait_until_not_entering_read_only_mode() - Wait until no read-only notifications are in |
| * progress and prevent any subsequent |
| * notifications. |
| * @parent: The completion to notify when no threads are entering read-only mode. |
| * |
| * Notifications may be re-enabled by calling vdo_allow_read_only_mode_entry(). |
| */ |
| void vdo_wait_until_not_entering_read_only_mode(struct vdo_completion *parent) |
| { |
| struct vdo *vdo = parent->vdo; |
| struct read_only_notifier *notifier = &vdo->read_only_notifier; |
| |
| vdo_assert_on_admin_thread(vdo, __func__); |
| |
| if (notifier->waiter != NULL) { |
| vdo_continue_completion(parent, VDO_COMPONENT_BUSY); |
| return; |
| } |
| |
| spin_lock(¬ifier->lock); |
| if (notifier->state == NOTIFYING) |
| notifier->waiter = parent; |
| else if (notifier->state == MAY_NOTIFY) |
| notifier->state = MAY_NOT_NOTIFY; |
| spin_unlock(¬ifier->lock); |
| |
| if (notifier->waiter == NULL) { |
| /* |
| * A notification was not in progress, and now they are |
| * disallowed. |
| */ |
| vdo_launch_completion(parent); |
| return; |
| } |
| } |
| |
| /** |
| * as_notifier() - Convert a generic vdo_completion to a read_only_notifier. |
| * @completion: The completion to convert. |
| * |
| * Return: The completion as a read_only_notifier. |
| */ |
| static inline struct read_only_notifier *as_notifier(struct vdo_completion *completion) |
| { |
| vdo_assert_completion_type(completion, VDO_READ_ONLY_MODE_COMPLETION); |
| return container_of(completion, struct read_only_notifier, completion); |
| } |
| |
| /** |
| * finish_entering_read_only_mode() - Complete the process of entering read only mode. |
| * @completion: The read-only mode completion. |
| */ |
| static void finish_entering_read_only_mode(struct vdo_completion *completion) |
| { |
| struct read_only_notifier *notifier = as_notifier(completion); |
| |
| vdo_assert_on_admin_thread(completion->vdo, __func__); |
| |
| spin_lock(¬ifier->lock); |
| notifier->state = NOTIFIED; |
| spin_unlock(¬ifier->lock); |
| |
| if (notifier->waiter != NULL) |
| vdo_continue_completion(vdo_forget(notifier->waiter), |
| completion->result); |
| } |
| |
| /** |
| * make_thread_read_only() - Inform each thread that the VDO is in read-only mode. |
| * @completion: The read-only mode completion. |
| */ |
| static void make_thread_read_only(struct vdo_completion *completion) |
| { |
| struct vdo *vdo = completion->vdo; |
| thread_id_t thread_id = completion->callback_thread_id; |
| struct read_only_notifier *notifier = as_notifier(completion); |
| struct read_only_listener *listener = completion->parent; |
| |
| if (listener == NULL) { |
| /* This is the first call on this thread */ |
| struct vdo_thread *thread = &vdo->threads[thread_id]; |
| |
| thread->is_read_only = true; |
| listener = thread->listeners; |
| if (thread_id == 0) |
| vdo_log_error_strerror(READ_ONCE(notifier->read_only_error), |
| "Unrecoverable error, entering read-only mode"); |
| } else { |
| /* We've just finished notifying a listener */ |
| listener = listener->next; |
| } |
| |
| if (listener != NULL) { |
| /* We have a listener to notify */ |
| vdo_prepare_completion(completion, make_thread_read_only, |
| make_thread_read_only, thread_id, |
| listener); |
| listener->notify(listener->listener, completion); |
| return; |
| } |
| |
| /* We're done with this thread */ |
| if (++thread_id == vdo->thread_config.dedupe_thread) { |
| /* |
| * We don't want to notify the dedupe thread since it may be |
| * blocked rebuilding the index. |
| */ |
| thread_id++; |
| } |
| |
| if (thread_id >= vdo->thread_config.thread_count) { |
| /* There are no more threads */ |
| vdo_prepare_completion(completion, finish_entering_read_only_mode, |
| finish_entering_read_only_mode, |
| vdo->thread_config.admin_thread, NULL); |
| } else { |
| vdo_prepare_completion(completion, make_thread_read_only, |
| make_thread_read_only, thread_id, NULL); |
| } |
| |
| vdo_launch_completion(completion); |
| } |
| |
| /** |
| * vdo_allow_read_only_mode_entry() - Allow the notifier to put the VDO into read-only mode, |
| * reversing the effects of |
| * vdo_wait_until_not_entering_read_only_mode(). |
| * @parent: The object to notify once the operation is complete. |
| * |
| * If some thread tried to put the vdo into read-only mode while notifications were disallowed, it |
| * will be done when this method is called. If that happens, the parent will not be notified until |
| * the vdo has actually entered read-only mode and attempted to save the super block. |
| * |
| * Context: This method may only be called from the admin thread. |
| */ |
| void vdo_allow_read_only_mode_entry(struct vdo_completion *parent) |
| { |
| struct vdo *vdo = parent->vdo; |
| struct read_only_notifier *notifier = &vdo->read_only_notifier; |
| |
| vdo_assert_on_admin_thread(vdo, __func__); |
| |
| if (notifier->waiter != NULL) { |
| vdo_continue_completion(parent, VDO_COMPONENT_BUSY); |
| return; |
| } |
| |
| spin_lock(¬ifier->lock); |
| if (notifier->state == MAY_NOT_NOTIFY) { |
| if (notifier->read_only_error == VDO_SUCCESS) { |
| notifier->state = MAY_NOTIFY; |
| } else { |
| notifier->state = NOTIFYING; |
| notifier->waiter = parent; |
| } |
| } |
| spin_unlock(¬ifier->lock); |
| |
| if (notifier->waiter == NULL) { |
| /* We're done */ |
| vdo_launch_completion(parent); |
| return; |
| } |
| |
| /* Do the pending notification. */ |
| make_thread_read_only(¬ifier->completion); |
| } |
| |
| /** |
| * vdo_enter_read_only_mode() - Put a VDO into read-only mode and save the read-only state in the |
| * super block. |
| * @vdo: The vdo. |
| * @error_code: The error which caused the VDO to enter read-only mode. |
| * |
| * This method is a no-op if the VDO is already read-only. |
| */ |
| void vdo_enter_read_only_mode(struct vdo *vdo, int error_code) |
| { |
| bool notify = false; |
| thread_id_t thread_id = vdo_get_callback_thread_id(); |
| struct read_only_notifier *notifier = &vdo->read_only_notifier; |
| struct vdo_thread *thread; |
| |
| if (thread_id != VDO_INVALID_THREAD_ID) { |
| thread = &vdo->threads[thread_id]; |
| if (thread->is_read_only) { |
| /* This thread has already gone read-only. */ |
| return; |
| } |
| |
| /* Record for this thread that the VDO is read-only. */ |
| thread->is_read_only = true; |
| } |
| |
| spin_lock(¬ifier->lock); |
| if (notifier->read_only_error == VDO_SUCCESS) { |
| WRITE_ONCE(notifier->read_only_error, error_code); |
| if (notifier->state == MAY_NOTIFY) { |
| notifier->state = NOTIFYING; |
| notify = true; |
| } |
| } |
| spin_unlock(¬ifier->lock); |
| |
| if (!notify) { |
| /* The notifier is already aware of a read-only error */ |
| return; |
| } |
| |
| /* Initiate a notification starting on the lowest numbered thread. */ |
| vdo_launch_completion_callback(¬ifier->completion, make_thread_read_only, 0); |
| } |
| |
| /** |
| * vdo_is_read_only() - Check whether the VDO is read-only. |
| * @vdo: The vdo. |
| * |
| * Return: true if the vdo is read-only. |
| * |
| * This method may be called from any thread, as opposed to examining the VDO's state field which |
| * is only safe to check from the admin thread. |
| */ |
| bool vdo_is_read_only(struct vdo *vdo) |
| { |
| return vdo->threads[vdo_get_callback_thread_id()].is_read_only; |
| } |
| |
| /** |
| * vdo_in_read_only_mode() - Check whether a vdo is in read-only mode. |
| * @vdo: The vdo to query. |
| * |
| * Return: true if the vdo is in read-only mode. |
| */ |
| bool vdo_in_read_only_mode(const struct vdo *vdo) |
| { |
| return (vdo_get_state(vdo) == VDO_READ_ONLY_MODE); |
| } |
| |
| /** |
| * vdo_in_recovery_mode() - Check whether the vdo is in recovery mode. |
| * @vdo: The vdo to query. |
| * |
| * Return: true if the vdo is in recovery mode. |
| */ |
| bool vdo_in_recovery_mode(const struct vdo *vdo) |
| { |
| return (vdo_get_state(vdo) == VDO_RECOVERING); |
| } |
| |
| /** |
| * vdo_enter_recovery_mode() - Put the vdo into recovery mode. |
| * @vdo: The vdo. |
| */ |
| void vdo_enter_recovery_mode(struct vdo *vdo) |
| { |
| vdo_assert_on_admin_thread(vdo, __func__); |
| |
| if (vdo_in_read_only_mode(vdo)) |
| return; |
| |
| vdo_log_info("Entering recovery mode"); |
| vdo_set_state(vdo, VDO_RECOVERING); |
| } |
| |
| /** |
| * complete_synchronous_action() - Signal the waiting thread that a synchronous action is complete. |
| * @completion: The sync completion. |
| */ |
| static void complete_synchronous_action(struct vdo_completion *completion) |
| { |
| vdo_assert_completion_type(completion, VDO_SYNC_COMPLETION); |
| complete(&(container_of(completion, struct sync_completion, |
| vdo_completion)->completion)); |
| } |
| |
| /** |
| * perform_synchronous_action() - Launch an action on a VDO thread and wait for it to complete. |
| * @vdo: The vdo. |
| * @action: The callback to launch. |
| * @thread_id: The thread on which to run the action. |
| * @parent: The parent of the sync completion (may be NULL). |
| */ |
| static int perform_synchronous_action(struct vdo *vdo, vdo_action_fn action, |
| thread_id_t thread_id, void *parent) |
| { |
| struct sync_completion sync; |
| |
| vdo_initialize_completion(&sync.vdo_completion, vdo, VDO_SYNC_COMPLETION); |
| init_completion(&sync.completion); |
| sync.vdo_completion.parent = parent; |
| vdo_launch_completion_callback(&sync.vdo_completion, action, thread_id); |
| wait_for_completion(&sync.completion); |
| return sync.vdo_completion.result; |
| } |
| |
| /** |
| * set_compression_callback() - Callback to turn compression on or off. |
| * @completion: The completion. |
| */ |
| static void set_compression_callback(struct vdo_completion *completion) |
| { |
| struct vdo *vdo = completion->vdo; |
| bool *enable = completion->parent; |
| bool was_enabled = vdo_get_compressing(vdo); |
| |
| if (*enable != was_enabled) { |
| WRITE_ONCE(vdo->compressing, *enable); |
| if (was_enabled) { |
| /* Signal the packer to flush since compression has been disabled. */ |
| vdo_flush_packer(vdo->packer); |
| } |
| } |
| |
| vdo_log_info("compression is %s", (*enable ? "enabled" : "disabled")); |
| *enable = was_enabled; |
| complete_synchronous_action(completion); |
| } |
| |
| /** |
| * vdo_set_compressing() - Turn compression on or off. |
| * @vdo: The vdo. |
| * @enable: Whether to enable or disable compression. |
| * |
| * Return: Whether compression was previously on or off. |
| */ |
| bool vdo_set_compressing(struct vdo *vdo, bool enable) |
| { |
| perform_synchronous_action(vdo, set_compression_callback, |
| vdo->thread_config.packer_thread, |
| &enable); |
| return enable; |
| } |
| |
| /** |
| * vdo_get_compressing() - Get whether compression is enabled in a vdo. |
| * @vdo: The vdo. |
| * |
| * Return: State of compression. |
| */ |
| bool vdo_get_compressing(struct vdo *vdo) |
| { |
| return READ_ONCE(vdo->compressing); |
| } |
| |
| static size_t get_block_map_cache_size(const struct vdo *vdo) |
| { |
| return ((size_t) vdo->device_config->cache_size) * VDO_BLOCK_SIZE; |
| } |
| |
| static struct error_statistics __must_check get_vdo_error_statistics(const struct vdo *vdo) |
| { |
| /* |
| * The error counts can be incremented from arbitrary threads and so must be incremented |
| * atomically, but they are just statistics with no semantics that could rely on memory |
| * order, so unfenced reads are sufficient. |
| */ |
| const struct atomic_statistics *atoms = &vdo->stats; |
| |
| return (struct error_statistics) { |
| .invalid_advice_pbn_count = atomic64_read(&atoms->invalid_advice_pbn_count), |
| .no_space_error_count = atomic64_read(&atoms->no_space_error_count), |
| .read_only_error_count = atomic64_read(&atoms->read_only_error_count), |
| }; |
| } |
| |
| static void copy_bio_stat(struct bio_stats *b, const struct atomic_bio_stats *a) |
| { |
| b->read = atomic64_read(&a->read); |
| b->write = atomic64_read(&a->write); |
| b->discard = atomic64_read(&a->discard); |
| b->flush = atomic64_read(&a->flush); |
| b->empty_flush = atomic64_read(&a->empty_flush); |
| b->fua = atomic64_read(&a->fua); |
| } |
| |
| static struct bio_stats subtract_bio_stats(struct bio_stats minuend, |
| struct bio_stats subtrahend) |
| { |
| return (struct bio_stats) { |
| .read = minuend.read - subtrahend.read, |
| .write = minuend.write - subtrahend.write, |
| .discard = minuend.discard - subtrahend.discard, |
| .flush = minuend.flush - subtrahend.flush, |
| .empty_flush = minuend.empty_flush - subtrahend.empty_flush, |
| .fua = minuend.fua - subtrahend.fua, |
| }; |
| } |
| |
| /** |
| * vdo_get_physical_blocks_allocated() - Get the number of physical blocks in use by user data. |
| * @vdo: The vdo. |
| * |
| * Return: The number of blocks allocated for user data. |
| */ |
| static block_count_t __must_check vdo_get_physical_blocks_allocated(const struct vdo *vdo) |
| { |
| return (vdo_get_slab_depot_allocated_blocks(vdo->depot) - |
| vdo_get_journal_block_map_data_blocks_used(vdo->recovery_journal)); |
| } |
| |
| /** |
| * vdo_get_physical_blocks_overhead() - Get the number of physical blocks used by vdo metadata. |
| * @vdo: The vdo. |
| * |
| * Return: The number of overhead blocks. |
| */ |
| static block_count_t __must_check vdo_get_physical_blocks_overhead(const struct vdo *vdo) |
| { |
| /* |
| * config.physical_blocks is mutated during resize and is in a packed structure, |
| * but resize runs on admin thread. |
| * TODO: Verify that this is always safe. |
| */ |
| return (vdo->states.vdo.config.physical_blocks - |
| vdo_get_slab_depot_data_blocks(vdo->depot) + |
| vdo_get_journal_block_map_data_blocks_used(vdo->recovery_journal)); |
| } |
| |
| static const char *vdo_describe_state(enum vdo_state state) |
| { |
| /* These strings should all fit in the 15 chars of VDOStatistics.mode. */ |
| switch (state) { |
| case VDO_RECOVERING: |
| return "recovering"; |
| |
| case VDO_READ_ONLY_MODE: |
| return "read-only"; |
| |
| default: |
| return "normal"; |
| } |
| } |
| |
| /** |
| * get_vdo_statistics() - Populate a vdo_statistics structure on the admin thread. |
| * @vdo: The vdo. |
| * @stats: The statistics structure to populate. |
| */ |
| static void get_vdo_statistics(const struct vdo *vdo, struct vdo_statistics *stats) |
| { |
| struct recovery_journal *journal = vdo->recovery_journal; |
| enum vdo_state state = vdo_get_state(vdo); |
| |
| vdo_assert_on_admin_thread(vdo, __func__); |
| |
| /* start with a clean slate */ |
| memset(stats, 0, sizeof(struct vdo_statistics)); |
| |
| /* |
| * These are immutable properties of the vdo object, so it is safe to query them from any |
| * thread. |
| */ |
| stats->version = STATISTICS_VERSION; |
| stats->logical_blocks = vdo->states.vdo.config.logical_blocks; |
| /* |
| * config.physical_blocks is mutated during resize and is in a packed structure, but resize |
| * runs on the admin thread. |
| * TODO: verify that this is always safe |
| */ |
| stats->physical_blocks = vdo->states.vdo.config.physical_blocks; |
| stats->block_size = VDO_BLOCK_SIZE; |
| stats->complete_recoveries = vdo->states.vdo.complete_recoveries; |
| stats->read_only_recoveries = vdo->states.vdo.read_only_recoveries; |
| stats->block_map_cache_size = get_block_map_cache_size(vdo); |
| |
| /* The callees are responsible for thread-safety. */ |
| stats->data_blocks_used = vdo_get_physical_blocks_allocated(vdo); |
| stats->overhead_blocks_used = vdo_get_physical_blocks_overhead(vdo); |
| stats->logical_blocks_used = vdo_get_recovery_journal_logical_blocks_used(journal); |
| vdo_get_slab_depot_statistics(vdo->depot, stats); |
| stats->journal = vdo_get_recovery_journal_statistics(journal); |
| stats->packer = vdo_get_packer_statistics(vdo->packer); |
| stats->block_map = vdo_get_block_map_statistics(vdo->block_map); |
| vdo_get_dedupe_statistics(vdo->hash_zones, stats); |
| stats->errors = get_vdo_error_statistics(vdo); |
| stats->in_recovery_mode = (state == VDO_RECOVERING); |
| snprintf(stats->mode, sizeof(stats->mode), "%s", vdo_describe_state(state)); |
| |
| stats->instance = vdo->instance; |
| stats->current_vios_in_progress = get_data_vio_pool_active_requests(vdo->data_vio_pool); |
| stats->max_vios = get_data_vio_pool_maximum_requests(vdo->data_vio_pool); |
| |
| stats->flush_out = atomic64_read(&vdo->stats.flush_out); |
| stats->logical_block_size = vdo->device_config->logical_block_size; |
| copy_bio_stat(&stats->bios_in, &vdo->stats.bios_in); |
| copy_bio_stat(&stats->bios_in_partial, &vdo->stats.bios_in_partial); |
| copy_bio_stat(&stats->bios_out, &vdo->stats.bios_out); |
| copy_bio_stat(&stats->bios_meta, &vdo->stats.bios_meta); |
| copy_bio_stat(&stats->bios_journal, &vdo->stats.bios_journal); |
| copy_bio_stat(&stats->bios_page_cache, &vdo->stats.bios_page_cache); |
| copy_bio_stat(&stats->bios_out_completed, &vdo->stats.bios_out_completed); |
| copy_bio_stat(&stats->bios_meta_completed, &vdo->stats.bios_meta_completed); |
| copy_bio_stat(&stats->bios_journal_completed, |
| &vdo->stats.bios_journal_completed); |
| copy_bio_stat(&stats->bios_page_cache_completed, |
| &vdo->stats.bios_page_cache_completed); |
| copy_bio_stat(&stats->bios_acknowledged, &vdo->stats.bios_acknowledged); |
| copy_bio_stat(&stats->bios_acknowledged_partial, &vdo->stats.bios_acknowledged_partial); |
| stats->bios_in_progress = |
| subtract_bio_stats(stats->bios_in, stats->bios_acknowledged); |
| vdo_get_memory_stats(&stats->memory_usage.bytes_used, |
| &stats->memory_usage.peak_bytes_used); |
| } |
| |
| /** |
| * vdo_fetch_statistics_callback() - Action to populate a vdo_statistics |
| * structure on the admin thread. |
| * @completion: The completion. |
| * |
| * This callback is registered in vdo_fetch_statistics(). |
| */ |
| static void vdo_fetch_statistics_callback(struct vdo_completion *completion) |
| { |
| get_vdo_statistics(completion->vdo, completion->parent); |
| complete_synchronous_action(completion); |
| } |
| |
| /** |
| * vdo_fetch_statistics() - Fetch statistics on the correct thread. |
| * @vdo: The vdo. |
| * @stats: The vdo statistics are returned here. |
| */ |
| void vdo_fetch_statistics(struct vdo *vdo, struct vdo_statistics *stats) |
| { |
| perform_synchronous_action(vdo, vdo_fetch_statistics_callback, |
| vdo->thread_config.admin_thread, stats); |
| } |
| |
| /** |
| * vdo_get_callback_thread_id() - Get the id of the callback thread on which a completion is |
| * currently running. |
| * |
| * Return: The current thread ID, or -1 if no such thread. |
| */ |
| thread_id_t vdo_get_callback_thread_id(void) |
| { |
| struct vdo_work_queue *queue = vdo_get_current_work_queue(); |
| struct vdo_thread *thread; |
| thread_id_t thread_id; |
| |
| if (queue == NULL) |
| return VDO_INVALID_THREAD_ID; |
| |
| thread = vdo_get_work_queue_owner(queue); |
| thread_id = thread->thread_id; |
| |
| if (PARANOID_THREAD_CONSISTENCY_CHECKS) { |
| BUG_ON(thread_id >= thread->vdo->thread_config.thread_count); |
| BUG_ON(thread != &thread->vdo->threads[thread_id]); |
| } |
| |
| return thread_id; |
| } |
| |
| /** |
| * vdo_dump_status() - Dump status information about a vdo to the log for debugging. |
| * @vdo: The vdo to dump. |
| */ |
| void vdo_dump_status(const struct vdo *vdo) |
| { |
| zone_count_t zone; |
| |
| vdo_dump_flusher(vdo->flusher); |
| vdo_dump_recovery_journal_statistics(vdo->recovery_journal); |
| vdo_dump_packer(vdo->packer); |
| vdo_dump_slab_depot(vdo->depot); |
| |
| for (zone = 0; zone < vdo->thread_config.logical_zone_count; zone++) |
| vdo_dump_logical_zone(&vdo->logical_zones->zones[zone]); |
| |
| for (zone = 0; zone < vdo->thread_config.physical_zone_count; zone++) |
| vdo_dump_physical_zone(&vdo->physical_zones->zones[zone]); |
| |
| vdo_dump_hash_zones(vdo->hash_zones); |
| } |
| |
| /** |
| * vdo_assert_on_admin_thread() - Assert that we are running on the admin thread. |
| * @vdo: The vdo. |
| * @name: The name of the function which should be running on the admin thread (for logging). |
| */ |
| void vdo_assert_on_admin_thread(const struct vdo *vdo, const char *name) |
| { |
| VDO_ASSERT_LOG_ONLY((vdo_get_callback_thread_id() == vdo->thread_config.admin_thread), |
| "%s called on admin thread", name); |
| } |
| |
| /** |
| * vdo_assert_on_logical_zone_thread() - Assert that this function was called on the specified |
| * logical zone thread. |
| * @vdo: The vdo. |
| * @logical_zone: The number of the logical zone. |
| * @name: The name of the calling function. |
| */ |
| void vdo_assert_on_logical_zone_thread(const struct vdo *vdo, zone_count_t logical_zone, |
| const char *name) |
| { |
| VDO_ASSERT_LOG_ONLY((vdo_get_callback_thread_id() == |
| vdo->thread_config.logical_threads[logical_zone]), |
| "%s called on logical thread", name); |
| } |
| |
| /** |
| * vdo_assert_on_physical_zone_thread() - Assert that this function was called on the specified |
| * physical zone thread. |
| * @vdo: The vdo. |
| * @physical_zone: The number of the physical zone. |
| * @name: The name of the calling function. |
| */ |
| void vdo_assert_on_physical_zone_thread(const struct vdo *vdo, |
| zone_count_t physical_zone, const char *name) |
| { |
| VDO_ASSERT_LOG_ONLY((vdo_get_callback_thread_id() == |
| vdo->thread_config.physical_threads[physical_zone]), |
| "%s called on physical thread", name); |
| } |
| |
| /** |
| * vdo_get_physical_zone() - Get the physical zone responsible for a given physical block number. |
| * @vdo: The vdo containing the physical zones. |
| * @pbn: The PBN of the data block. |
| * @zone_ptr: A pointer to return the physical zone. |
| * |
| * Gets the physical zone responsible for a given physical block number of a data block in this vdo |
| * instance, or of the zero block (for which a NULL zone is returned). For any other block number |
| * that is not in the range of valid data block numbers in any slab, an error will be returned. |
| * This function is safe to call on invalid block numbers; it will not put the vdo into read-only |
| * mode. |
| * |
| * Return: VDO_SUCCESS or VDO_OUT_OF_RANGE if the block number is invalid or an error code for any |
| * other failure. |
| */ |
| int vdo_get_physical_zone(const struct vdo *vdo, physical_block_number_t pbn, |
| struct physical_zone **zone_ptr) |
| { |
| struct vdo_slab *slab; |
| int result; |
| |
| if (pbn == VDO_ZERO_BLOCK) { |
| *zone_ptr = NULL; |
| return VDO_SUCCESS; |
| } |
| |
| /* |
| * Used because it does a more restrictive bounds check than vdo_get_slab(), and done first |
| * because it won't trigger read-only mode on an invalid PBN. |
| */ |
| if (!vdo_is_physical_data_block(vdo->depot, pbn)) |
| return VDO_OUT_OF_RANGE; |
| |
| /* With the PBN already checked, we should always succeed in finding a slab. */ |
| slab = vdo_get_slab(vdo->depot, pbn); |
| result = VDO_ASSERT(slab != NULL, "vdo_get_slab must succeed on all valid PBNs"); |
| if (result != VDO_SUCCESS) |
| return result; |
| |
| *zone_ptr = &vdo->physical_zones->zones[slab->allocator->zone_number]; |
| return VDO_SUCCESS; |
| } |