| // SPDX-License-Identifier: GPL-2.0-or-later |
| |
| #include <linux/blkdev.h> |
| #include <linux/dax.h> |
| #include <linux/vmalloc.h> |
| #include <linux/parser.h> |
| |
| #include "cache_dev.h" |
| #include "backing_dev.h" |
| #include "cache.h" |
| #include "dm_pcache.h" |
| |
| static void cache_dev_dax_exit(struct pcache_cache_dev *cache_dev) |
| { |
| if (cache_dev->use_vmap) |
| vunmap(cache_dev->mapping); |
| } |
| |
| static int build_vmap(struct dax_device *dax_dev, long total_pages, void **vaddr) |
| { |
| struct page **pages; |
| long i = 0, chunk; |
| unsigned long pfn; |
| int ret; |
| |
| pages = vmalloc_array(total_pages, sizeof(struct page *)); |
| if (!pages) |
| return -ENOMEM; |
| |
| do { |
| chunk = dax_direct_access(dax_dev, i, total_pages - i, |
| DAX_ACCESS, NULL, &pfn); |
| if (chunk <= 0) { |
| ret = chunk ? chunk : -EINVAL; |
| goto out_free; |
| } |
| |
| if (!pfn_valid(pfn)) { |
| ret = -EOPNOTSUPP; |
| goto out_free; |
| } |
| |
| while (chunk-- && i < total_pages) { |
| pages[i++] = pfn_to_page(pfn); |
| pfn++; |
| if (!(i & 15)) |
| cond_resched(); |
| } |
| } while (i < total_pages); |
| |
| *vaddr = vmap(pages, total_pages, VM_MAP, PAGE_KERNEL); |
| if (!*vaddr) { |
| ret = -ENOMEM; |
| goto out_free; |
| } |
| |
| ret = 0; |
| |
| out_free: |
| vfree(pages); |
| return ret; |
| } |
| |
| static int cache_dev_dax_init(struct pcache_cache_dev *cache_dev) |
| { |
| struct dm_pcache *pcache = CACHE_DEV_TO_PCACHE(cache_dev); |
| struct dax_device *dax_dev; |
| long total_pages, mapped_pages; |
| u64 bdev_size; |
| void *vaddr; |
| int ret; |
| int id; |
| unsigned long pfn; |
| |
| dax_dev = cache_dev->dm_dev->dax_dev; |
| /* total size check */ |
| bdev_size = bdev_nr_bytes(cache_dev->dm_dev->bdev); |
| if (bdev_size < PCACHE_CACHE_DEV_SIZE_MIN) { |
| pcache_dev_err(pcache, "dax device is too small, required at least %llu", |
| PCACHE_CACHE_DEV_SIZE_MIN); |
| ret = -ENOSPC; |
| goto out; |
| } |
| |
| total_pages = bdev_size >> PAGE_SHIFT; |
| /* attempt: direct-map the whole range */ |
| id = dax_read_lock(); |
| mapped_pages = dax_direct_access(dax_dev, 0, total_pages, |
| DAX_ACCESS, &vaddr, &pfn); |
| if (mapped_pages < 0) { |
| pcache_dev_err(pcache, "dax_direct_access failed: %ld\n", mapped_pages); |
| ret = mapped_pages; |
| goto unlock; |
| } |
| |
| if (!pfn_valid(pfn)) { |
| ret = -EOPNOTSUPP; |
| goto unlock; |
| } |
| |
| if (mapped_pages == total_pages) { |
| /* success: contiguous direct mapping */ |
| cache_dev->mapping = vaddr; |
| } else { |
| /* need vmap fallback */ |
| ret = build_vmap(dax_dev, total_pages, &vaddr); |
| if (ret) { |
| pcache_dev_err(pcache, "vmap fallback failed: %d\n", ret); |
| goto unlock; |
| } |
| |
| cache_dev->mapping = vaddr; |
| cache_dev->use_vmap = true; |
| } |
| dax_read_unlock(id); |
| |
| return 0; |
| unlock: |
| dax_read_unlock(id); |
| out: |
| return ret; |
| } |
| |
| void cache_dev_zero_range(struct pcache_cache_dev *cache_dev, void *pos, u32 size) |
| { |
| memset(pos, 0, size); |
| dax_flush(cache_dev->dm_dev->dax_dev, pos, size); |
| } |
| |
| static int sb_read(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb) |
| { |
| struct pcache_sb *sb_addr = CACHE_DEV_SB(cache_dev); |
| |
| if (copy_mc_to_kernel(sb, sb_addr, sizeof(struct pcache_sb))) |
| return -EIO; |
| |
| return 0; |
| } |
| |
| static void sb_write(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb) |
| { |
| struct pcache_sb *sb_addr = CACHE_DEV_SB(cache_dev); |
| |
| memcpy_flushcache(sb_addr, sb, sizeof(struct pcache_sb)); |
| pmem_wmb(); |
| } |
| |
| static int sb_init(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb) |
| { |
| struct dm_pcache *pcache = CACHE_DEV_TO_PCACHE(cache_dev); |
| u64 nr_segs; |
| u64 cache_dev_size; |
| u64 magic; |
| u32 flags = 0; |
| |
| magic = le64_to_cpu(sb->magic); |
| if (magic) |
| return -EEXIST; |
| |
| cache_dev_size = bdev_nr_bytes(file_bdev(cache_dev->dm_dev->bdev_file)); |
| if (cache_dev_size < PCACHE_CACHE_DEV_SIZE_MIN) { |
| pcache_dev_err(pcache, "dax device is too small, required at least %llu", |
| PCACHE_CACHE_DEV_SIZE_MIN); |
| return -ENOSPC; |
| } |
| |
| nr_segs = (cache_dev_size - PCACHE_SEGMENTS_OFF) / ((PCACHE_SEG_SIZE)); |
| |
| #if defined(__BYTE_ORDER) ? (__BIG_ENDIAN == __BYTE_ORDER) : defined(__BIG_ENDIAN) |
| flags |= PCACHE_SB_F_BIGENDIAN; |
| #endif |
| sb->flags = cpu_to_le32(flags); |
| sb->magic = cpu_to_le64(PCACHE_MAGIC); |
| sb->seg_num = cpu_to_le32(nr_segs); |
| sb->crc = cpu_to_le32(crc32c(PCACHE_CRC_SEED, (void *)(sb) + 4, sizeof(struct pcache_sb) - 4)); |
| |
| cache_dev_zero_range(cache_dev, CACHE_DEV_CACHE_INFO(cache_dev), |
| PCACHE_CACHE_INFO_SIZE * PCACHE_META_INDEX_MAX + |
| PCACHE_CACHE_CTRL_SIZE); |
| |
| return 0; |
| } |
| |
| static int sb_validate(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb) |
| { |
| struct dm_pcache *pcache = CACHE_DEV_TO_PCACHE(cache_dev); |
| u32 flags; |
| u32 crc; |
| |
| if (le64_to_cpu(sb->magic) != PCACHE_MAGIC) { |
| pcache_dev_err(pcache, "unexpected magic: %llx\n", |
| le64_to_cpu(sb->magic)); |
| return -EINVAL; |
| } |
| |
| crc = crc32c(PCACHE_CRC_SEED, (void *)(sb) + 4, sizeof(struct pcache_sb) - 4); |
| if (crc != le32_to_cpu(sb->crc)) { |
| pcache_dev_err(pcache, "corrupted sb: %u, expected: %u\n", crc, le32_to_cpu(sb->crc)); |
| return -EINVAL; |
| } |
| |
| flags = le32_to_cpu(sb->flags); |
| #if defined(__BYTE_ORDER) ? (__BIG_ENDIAN == __BYTE_ORDER) : defined(__BIG_ENDIAN) |
| if (!(flags & PCACHE_SB_F_BIGENDIAN)) { |
| pcache_dev_err(pcache, "cache_dev is not big endian\n"); |
| return -EINVAL; |
| } |
| #else |
| if (flags & PCACHE_SB_F_BIGENDIAN) { |
| pcache_dev_err(pcache, "cache_dev is big endian\n"); |
| return -EINVAL; |
| } |
| #endif |
| return 0; |
| } |
| |
| static int cache_dev_init(struct pcache_cache_dev *cache_dev, u32 seg_num) |
| { |
| cache_dev->seg_num = seg_num; |
| cache_dev->seg_bitmap = kvcalloc(BITS_TO_LONGS(cache_dev->seg_num), sizeof(unsigned long), GFP_KERNEL); |
| if (!cache_dev->seg_bitmap) |
| return -ENOMEM; |
| |
| return 0; |
| } |
| |
| static void cache_dev_exit(struct pcache_cache_dev *cache_dev) |
| { |
| kvfree(cache_dev->seg_bitmap); |
| } |
| |
| void cache_dev_stop(struct dm_pcache *pcache) |
| { |
| struct pcache_cache_dev *cache_dev = &pcache->cache_dev; |
| |
| cache_dev_exit(cache_dev); |
| cache_dev_dax_exit(cache_dev); |
| } |
| |
| int cache_dev_start(struct dm_pcache *pcache) |
| { |
| struct pcache_cache_dev *cache_dev = &pcache->cache_dev; |
| struct pcache_sb sb; |
| bool format = false; |
| int ret; |
| |
| mutex_init(&cache_dev->seg_lock); |
| |
| ret = cache_dev_dax_init(cache_dev); |
| if (ret) { |
| pcache_dev_err(pcache, "failed to init cache_dev %s via dax way: %d.", |
| cache_dev->dm_dev->name, ret); |
| goto err; |
| } |
| |
| ret = sb_read(cache_dev, &sb); |
| if (ret) |
| goto dax_release; |
| |
| if (le64_to_cpu(sb.magic) == 0) { |
| format = true; |
| ret = sb_init(cache_dev, &sb); |
| if (ret < 0) |
| goto dax_release; |
| } |
| |
| ret = sb_validate(cache_dev, &sb); |
| if (ret) |
| goto dax_release; |
| |
| cache_dev->sb_flags = le32_to_cpu(sb.flags); |
| ret = cache_dev_init(cache_dev, le32_to_cpu(sb.seg_num)); |
| if (ret) |
| goto dax_release; |
| |
| if (format) |
| sb_write(cache_dev, &sb); |
| |
| return 0; |
| |
| dax_release: |
| cache_dev_dax_exit(cache_dev); |
| err: |
| return ret; |
| } |
| |
| int cache_dev_get_empty_segment_id(struct pcache_cache_dev *cache_dev, u32 *seg_id) |
| { |
| int ret; |
| |
| mutex_lock(&cache_dev->seg_lock); |
| *seg_id = find_next_zero_bit(cache_dev->seg_bitmap, cache_dev->seg_num, 0); |
| if (*seg_id == cache_dev->seg_num) { |
| ret = -ENOSPC; |
| goto unlock; |
| } |
| |
| __set_bit(*seg_id, cache_dev->seg_bitmap); |
| ret = 0; |
| unlock: |
| mutex_unlock(&cache_dev->seg_lock); |
| return ret; |
| } |