| // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB |
| /* Copyright (c) 2020 Mellanox Technologies Ltd. */ |
| |
| #include <linux/vhost_types.h> |
| #include <linux/vdpa.h> |
| #include <linux/gcd.h> |
| #include <linux/string.h> |
| #include <linux/mlx5/qp.h> |
| #include "mlx5_vdpa.h" |
| |
| /* DIV_ROUND_UP where the divider is a power of 2 give by its log base 2 value */ |
| #define MLX5_DIV_ROUND_UP_POW2(_n, _s) \ |
| ({ \ |
| u64 __s = _s; \ |
| u64 _res; \ |
| _res = (((_n) + (1 << (__s)) - 1) >> (__s)); \ |
| _res; \ |
| }) |
| |
| static int get_octo_len(u64 len, int page_shift) |
| { |
| u64 page_size = 1ULL << page_shift; |
| int npages; |
| |
| npages = ALIGN(len, page_size) >> page_shift; |
| return (npages + 1) / 2; |
| } |
| |
| static void mlx5_set_access_mode(void *mkc, int mode) |
| { |
| MLX5_SET(mkc, mkc, access_mode_1_0, mode & 0x3); |
| MLX5_SET(mkc, mkc, access_mode_4_2, mode >> 2); |
| } |
| |
| static void populate_mtts(struct mlx5_vdpa_direct_mr *mr, __be64 *mtt) |
| { |
| struct scatterlist *sg; |
| int nsg = mr->nsg; |
| u64 dma_addr; |
| u64 dma_len; |
| int j = 0; |
| int i; |
| |
| for_each_sg(mr->sg_head.sgl, sg, mr->nent, i) { |
| for (dma_addr = sg_dma_address(sg), dma_len = sg_dma_len(sg); |
| nsg && dma_len; |
| nsg--, dma_addr += BIT(mr->log_size), dma_len -= BIT(mr->log_size)) |
| mtt[j++] = cpu_to_be64(dma_addr); |
| } |
| } |
| |
| struct mlx5_create_mkey_mem { |
| u8 out[MLX5_ST_SZ_BYTES(create_mkey_out)]; |
| u8 in[MLX5_ST_SZ_BYTES(create_mkey_in)]; |
| __be64 mtt[]; |
| }; |
| |
| struct mlx5_destroy_mkey_mem { |
| u8 out[MLX5_ST_SZ_BYTES(destroy_mkey_out)]; |
| u8 in[MLX5_ST_SZ_BYTES(destroy_mkey_in)]; |
| }; |
| |
| static void fill_create_direct_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_direct_mr *mr, |
| struct mlx5_create_mkey_mem *mem) |
| { |
| void *in = &mem->in; |
| void *mkc; |
| |
| MLX5_SET(create_mkey_in, in, uid, mvdev->res.uid); |
| mkc = MLX5_ADDR_OF(create_mkey_in, in, memory_key_mkey_entry); |
| MLX5_SET(mkc, mkc, lw, !!(mr->perm & VHOST_MAP_WO)); |
| MLX5_SET(mkc, mkc, lr, !!(mr->perm & VHOST_MAP_RO)); |
| mlx5_set_access_mode(mkc, MLX5_MKC_ACCESS_MODE_MTT); |
| MLX5_SET(mkc, mkc, qpn, 0xffffff); |
| MLX5_SET(mkc, mkc, pd, mvdev->res.pdn); |
| MLX5_SET64(mkc, mkc, start_addr, mr->offset); |
| MLX5_SET64(mkc, mkc, len, mr->end - mr->start); |
| MLX5_SET(mkc, mkc, log_page_size, mr->log_size); |
| MLX5_SET(mkc, mkc, translations_octword_size, |
| get_octo_len(mr->end - mr->start, mr->log_size)); |
| MLX5_SET(create_mkey_in, in, translations_octword_actual_size, |
| get_octo_len(mr->end - mr->start, mr->log_size)); |
| populate_mtts(mr, MLX5_ADDR_OF(create_mkey_in, in, klm_pas_mtt)); |
| |
| MLX5_SET(create_mkey_in, in, opcode, MLX5_CMD_OP_CREATE_MKEY); |
| MLX5_SET(create_mkey_in, in, uid, mvdev->res.uid); |
| } |
| |
| static void create_direct_mr_end(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_direct_mr *mr, |
| struct mlx5_create_mkey_mem *mem) |
| { |
| u32 mkey_index = MLX5_GET(create_mkey_out, mem->out, mkey_index); |
| |
| mr->mr = mlx5_idx_to_mkey(mkey_index); |
| } |
| |
| static void fill_destroy_direct_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_direct_mr *mr, |
| struct mlx5_destroy_mkey_mem *mem) |
| { |
| void *in = &mem->in; |
| |
| MLX5_SET(destroy_mkey_in, in, uid, mvdev->res.uid); |
| MLX5_SET(destroy_mkey_in, in, opcode, MLX5_CMD_OP_DESTROY_MKEY); |
| MLX5_SET(destroy_mkey_in, in, mkey_index, mlx5_mkey_to_idx(mr->mr)); |
| } |
| |
| static void destroy_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr) |
| { |
| if (!mr->mr) |
| return; |
| |
| mlx5_vdpa_destroy_mkey(mvdev, mr->mr); |
| } |
| |
| static u64 map_start(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr) |
| { |
| return max_t(u64, map->start, mr->start); |
| } |
| |
| static u64 map_end(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr) |
| { |
| return min_t(u64, map->last + 1, mr->end); |
| } |
| |
| static u64 maplen(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr) |
| { |
| return map_end(map, mr) - map_start(map, mr); |
| } |
| |
| #define MLX5_VDPA_INVALID_START_ADDR ((u64)-1) |
| #define MLX5_VDPA_INVALID_LEN ((u64)-1) |
| |
| static u64 indir_start_addr(struct mlx5_vdpa_mr *mkey) |
| { |
| struct mlx5_vdpa_direct_mr *s; |
| |
| s = list_first_entry_or_null(&mkey->head, struct mlx5_vdpa_direct_mr, list); |
| if (!s) |
| return MLX5_VDPA_INVALID_START_ADDR; |
| |
| return s->start; |
| } |
| |
| static u64 indir_len(struct mlx5_vdpa_mr *mkey) |
| { |
| struct mlx5_vdpa_direct_mr *s; |
| struct mlx5_vdpa_direct_mr *e; |
| |
| s = list_first_entry_or_null(&mkey->head, struct mlx5_vdpa_direct_mr, list); |
| if (!s) |
| return MLX5_VDPA_INVALID_LEN; |
| |
| e = list_last_entry(&mkey->head, struct mlx5_vdpa_direct_mr, list); |
| |
| return e->end - s->start; |
| } |
| |
| #define LOG_MAX_KLM_SIZE 30 |
| #define MAX_KLM_SIZE BIT(LOG_MAX_KLM_SIZE) |
| |
| static u32 klm_bcount(u64 size) |
| { |
| return (u32)size; |
| } |
| |
| static void fill_indir(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mkey, void *in) |
| { |
| struct mlx5_vdpa_direct_mr *dmr; |
| struct mlx5_klm *klmarr; |
| struct mlx5_klm *klm; |
| bool first = true; |
| u64 preve; |
| int i; |
| |
| klmarr = MLX5_ADDR_OF(create_mkey_in, in, klm_pas_mtt); |
| i = 0; |
| list_for_each_entry(dmr, &mkey->head, list) { |
| again: |
| klm = &klmarr[i++]; |
| if (first) { |
| preve = dmr->start; |
| first = false; |
| } |
| |
| if (preve == dmr->start) { |
| klm->key = cpu_to_be32(dmr->mr); |
| klm->bcount = cpu_to_be32(klm_bcount(dmr->end - dmr->start)); |
| preve = dmr->end; |
| } else { |
| klm->key = cpu_to_be32(mvdev->res.null_mkey); |
| klm->bcount = cpu_to_be32(klm_bcount(dmr->start - preve)); |
| preve = dmr->start; |
| goto again; |
| } |
| } |
| } |
| |
| static int klm_byte_size(int nklms) |
| { |
| return 16 * ALIGN(nklms, 4); |
| } |
| |
| #define MLX5_VDPA_MTT_ALIGN 16 |
| |
| static int create_direct_keys(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) |
| { |
| struct mlx5_vdpa_async_cmd *cmds; |
| struct mlx5_vdpa_direct_mr *dmr; |
| int err = 0; |
| int i = 0; |
| |
| cmds = kvcalloc(mr->num_directs, sizeof(*cmds), GFP_KERNEL); |
| if (!cmds) |
| return -ENOMEM; |
| |
| list_for_each_entry(dmr, &mr->head, list) { |
| struct mlx5_create_mkey_mem *cmd_mem; |
| int mttlen, mttcount; |
| |
| mttlen = roundup(MLX5_ST_SZ_BYTES(mtt) * dmr->nsg, MLX5_VDPA_MTT_ALIGN); |
| mttcount = mttlen / sizeof(cmd_mem->mtt[0]); |
| cmd_mem = kvcalloc(1, struct_size(cmd_mem, mtt, mttcount), GFP_KERNEL); |
| if (!cmd_mem) { |
| err = -ENOMEM; |
| goto done; |
| } |
| |
| cmds[i].out = cmd_mem->out; |
| cmds[i].outlen = sizeof(cmd_mem->out); |
| cmds[i].in = cmd_mem->in; |
| cmds[i].inlen = struct_size(cmd_mem, mtt, mttcount); |
| |
| fill_create_direct_mr(mvdev, dmr, cmd_mem); |
| |
| i++; |
| } |
| |
| err = mlx5_vdpa_exec_async_cmds(mvdev, cmds, mr->num_directs); |
| if (err) { |
| |
| mlx5_vdpa_err(mvdev, "error issuing MTT mkey creation for direct mrs: %d\n", err); |
| goto done; |
| } |
| |
| i = 0; |
| list_for_each_entry(dmr, &mr->head, list) { |
| struct mlx5_vdpa_async_cmd *cmd = &cmds[i++]; |
| struct mlx5_create_mkey_mem *cmd_mem; |
| |
| cmd_mem = container_of(cmd->out, struct mlx5_create_mkey_mem, out); |
| |
| if (!cmd->err) { |
| create_direct_mr_end(mvdev, dmr, cmd_mem); |
| } else { |
| err = err ? err : cmd->err; |
| mlx5_vdpa_err(mvdev, "error creating MTT mkey [0x%llx, 0x%llx]: %d\n", |
| dmr->start, dmr->end, cmd->err); |
| } |
| } |
| |
| done: |
| for (i = i-1; i >= 0; i--) { |
| struct mlx5_create_mkey_mem *cmd_mem; |
| |
| cmd_mem = container_of(cmds[i].out, struct mlx5_create_mkey_mem, out); |
| kvfree(cmd_mem); |
| } |
| |
| kvfree(cmds); |
| return err; |
| } |
| |
| DEFINE_FREE(free_cmds, struct mlx5_vdpa_async_cmd *, kvfree(_T)) |
| DEFINE_FREE(free_cmd_mem, struct mlx5_destroy_mkey_mem *, kvfree(_T)) |
| |
| static int destroy_direct_keys(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) |
| { |
| struct mlx5_destroy_mkey_mem *cmd_mem __free(free_cmd_mem) = NULL; |
| struct mlx5_vdpa_async_cmd *cmds __free(free_cmds) = NULL; |
| struct mlx5_vdpa_direct_mr *dmr; |
| int err = 0; |
| int i = 0; |
| |
| cmds = kvcalloc(mr->num_directs, sizeof(*cmds), GFP_KERNEL); |
| cmd_mem = kvcalloc(mr->num_directs, sizeof(*cmd_mem), GFP_KERNEL); |
| if (!cmds || !cmd_mem) |
| return -ENOMEM; |
| |
| list_for_each_entry(dmr, &mr->head, list) { |
| cmds[i].out = cmd_mem[i].out; |
| cmds[i].outlen = sizeof(cmd_mem[i].out); |
| cmds[i].in = cmd_mem[i].in; |
| cmds[i].inlen = sizeof(cmd_mem[i].in); |
| fill_destroy_direct_mr(mvdev, dmr, &cmd_mem[i]); |
| i++; |
| } |
| |
| err = mlx5_vdpa_exec_async_cmds(mvdev, cmds, mr->num_directs); |
| if (err) { |
| |
| mlx5_vdpa_err(mvdev, "error issuing MTT mkey deletion for direct mrs: %d\n", err); |
| return err; |
| } |
| |
| i = 0; |
| list_for_each_entry(dmr, &mr->head, list) { |
| struct mlx5_vdpa_async_cmd *cmd = &cmds[i++]; |
| |
| dmr->mr = 0; |
| if (cmd->err) { |
| err = err ? err : cmd->err; |
| mlx5_vdpa_err(mvdev, "error deleting MTT mkey [0x%llx, 0x%llx]: %d\n", |
| dmr->start, dmr->end, cmd->err); |
| } |
| } |
| |
| return err; |
| } |
| |
| static int create_indirect_key(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) |
| { |
| int inlen; |
| void *mkc; |
| void *in; |
| int err; |
| u64 start; |
| u64 len; |
| |
| start = indir_start_addr(mr); |
| len = indir_len(mr); |
| if (start == MLX5_VDPA_INVALID_START_ADDR || len == MLX5_VDPA_INVALID_LEN) |
| return -EINVAL; |
| |
| inlen = MLX5_ST_SZ_BYTES(create_mkey_in) + klm_byte_size(mr->num_klms); |
| in = kzalloc(inlen, GFP_KERNEL); |
| if (!in) |
| return -ENOMEM; |
| |
| MLX5_SET(create_mkey_in, in, uid, mvdev->res.uid); |
| mkc = MLX5_ADDR_OF(create_mkey_in, in, memory_key_mkey_entry); |
| MLX5_SET(mkc, mkc, lw, 1); |
| MLX5_SET(mkc, mkc, lr, 1); |
| mlx5_set_access_mode(mkc, MLX5_MKC_ACCESS_MODE_KLMS); |
| MLX5_SET(mkc, mkc, qpn, 0xffffff); |
| MLX5_SET(mkc, mkc, pd, mvdev->res.pdn); |
| MLX5_SET64(mkc, mkc, start_addr, start); |
| MLX5_SET64(mkc, mkc, len, len); |
| MLX5_SET(mkc, mkc, translations_octword_size, klm_byte_size(mr->num_klms) / 16); |
| MLX5_SET(create_mkey_in, in, translations_octword_actual_size, mr->num_klms); |
| fill_indir(mvdev, mr, in); |
| err = mlx5_vdpa_create_mkey(mvdev, &mr->mkey, in, inlen); |
| kfree(in); |
| return err; |
| } |
| |
| static void destroy_indirect_key(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mkey) |
| { |
| mlx5_vdpa_destroy_mkey(mvdev, mkey->mkey); |
| } |
| |
| static int map_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr, |
| struct vhost_iotlb *iotlb) |
| { |
| struct vhost_iotlb_map *map; |
| unsigned long lgcd = 0; |
| int log_entity_size; |
| unsigned long size; |
| u64 start = 0; |
| int err; |
| struct page *pg; |
| unsigned int nsg; |
| int sglen; |
| u64 pa; |
| u64 paend; |
| struct scatterlist *sg; |
| struct device *dma = mvdev->vdev.dma_dev; |
| |
| for (map = vhost_iotlb_itree_first(iotlb, mr->start, mr->end - 1); |
| map; map = vhost_iotlb_itree_next(map, start, mr->end - 1)) { |
| size = maplen(map, mr); |
| lgcd = gcd(lgcd, size); |
| start += size; |
| } |
| log_entity_size = ilog2(lgcd); |
| |
| sglen = 1 << log_entity_size; |
| nsg = MLX5_DIV_ROUND_UP_POW2(mr->end - mr->start, log_entity_size); |
| |
| err = sg_alloc_table(&mr->sg_head, nsg, GFP_KERNEL); |
| if (err) |
| return err; |
| |
| sg = mr->sg_head.sgl; |
| for (map = vhost_iotlb_itree_first(iotlb, mr->start, mr->end - 1); |
| map; map = vhost_iotlb_itree_next(map, mr->start, mr->end - 1)) { |
| paend = map->addr + maplen(map, mr); |
| for (pa = map->addr; pa < paend; pa += sglen) { |
| pg = pfn_to_page(__phys_to_pfn(pa)); |
| if (!sg) { |
| mlx5_vdpa_warn(mvdev, "sg null. start 0x%llx, end 0x%llx\n", |
| map->start, map->last + 1); |
| err = -ENOMEM; |
| goto err_map; |
| } |
| sg_set_page(sg, pg, sglen, 0); |
| sg = sg_next(sg); |
| if (!sg) |
| goto done; |
| } |
| } |
| done: |
| mr->log_size = log_entity_size; |
| mr->nsg = nsg; |
| mr->nent = dma_map_sg_attrs(dma, mr->sg_head.sgl, mr->nsg, DMA_BIDIRECTIONAL, 0); |
| if (!mr->nent) { |
| err = -ENOMEM; |
| goto err_map; |
| } |
| |
| return 0; |
| |
| err_map: |
| sg_free_table(&mr->sg_head); |
| return err; |
| } |
| |
| static void unmap_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr) |
| { |
| struct device *dma = mvdev->vdev.dma_dev; |
| |
| destroy_direct_mr(mvdev, mr); |
| dma_unmap_sg_attrs(dma, mr->sg_head.sgl, mr->nsg, DMA_BIDIRECTIONAL, 0); |
| sg_free_table(&mr->sg_head); |
| } |
| |
| static int add_direct_chain(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *mr, |
| u64 start, |
| u64 size, |
| u8 perm, |
| struct vhost_iotlb *iotlb) |
| { |
| struct mlx5_vdpa_direct_mr *dmr; |
| struct mlx5_vdpa_direct_mr *n; |
| LIST_HEAD(tmp); |
| u64 st; |
| u64 sz; |
| int err; |
| |
| st = start; |
| while (size) { |
| sz = (u32)min_t(u64, MAX_KLM_SIZE, size); |
| dmr = kzalloc(sizeof(*dmr), GFP_KERNEL); |
| if (!dmr) { |
| err = -ENOMEM; |
| goto err_alloc; |
| } |
| |
| dmr->start = st; |
| dmr->end = st + sz; |
| dmr->perm = perm; |
| err = map_direct_mr(mvdev, dmr, iotlb); |
| if (err) { |
| kfree(dmr); |
| goto err_alloc; |
| } |
| |
| list_add_tail(&dmr->list, &tmp); |
| size -= sz; |
| mr->num_directs++; |
| mr->num_klms++; |
| st += sz; |
| } |
| list_splice_tail(&tmp, &mr->head); |
| return 0; |
| |
| err_alloc: |
| list_for_each_entry_safe(dmr, n, &mr->head, list) { |
| list_del_init(&dmr->list); |
| unmap_direct_mr(mvdev, dmr); |
| kfree(dmr); |
| } |
| return err; |
| } |
| |
| /* The iotlb pointer contains a list of maps. Go over the maps, possibly |
| * merging mergeable maps, and create direct memory keys that provide the |
| * device access to memory. The direct mkeys are then referred to by the |
| * indirect memory key that provides access to the enitre address space given |
| * by iotlb. |
| */ |
| static int create_user_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *mr, |
| struct vhost_iotlb *iotlb) |
| { |
| struct mlx5_vdpa_direct_mr *dmr; |
| struct mlx5_vdpa_direct_mr *n; |
| struct vhost_iotlb_map *map; |
| u32 pperm = U16_MAX; |
| u64 last = U64_MAX; |
| u64 ps = U64_MAX; |
| u64 pe = U64_MAX; |
| u64 start = 0; |
| int err = 0; |
| int nnuls; |
| |
| INIT_LIST_HEAD(&mr->head); |
| for (map = vhost_iotlb_itree_first(iotlb, start, last); map; |
| map = vhost_iotlb_itree_next(map, start, last)) { |
| start = map->start; |
| if (pe == map->start && pperm == map->perm) { |
| pe = map->last + 1; |
| } else { |
| if (ps != U64_MAX) { |
| if (pe < map->start) { |
| /* We have a hole in the map. Check how |
| * many null keys are required to fill it. |
| */ |
| nnuls = MLX5_DIV_ROUND_UP_POW2(map->start - pe, |
| LOG_MAX_KLM_SIZE); |
| mr->num_klms += nnuls; |
| } |
| err = add_direct_chain(mvdev, mr, ps, pe - ps, pperm, iotlb); |
| if (err) |
| goto err_chain; |
| } |
| ps = map->start; |
| pe = map->last + 1; |
| pperm = map->perm; |
| } |
| } |
| err = add_direct_chain(mvdev, mr, ps, pe - ps, pperm, iotlb); |
| if (err) |
| goto err_chain; |
| |
| err = create_direct_keys(mvdev, mr); |
| if (err) |
| goto err_chain; |
| |
| /* Create the memory key that defines the guests's address space. This |
| * memory key refers to the direct keys that contain the MTT |
| * translations |
| */ |
| err = create_indirect_key(mvdev, mr); |
| if (err) |
| goto err_chain; |
| |
| mr->user_mr = true; |
| return 0; |
| |
| err_chain: |
| list_for_each_entry_safe_reverse(dmr, n, &mr->head, list) { |
| list_del_init(&dmr->list); |
| unmap_direct_mr(mvdev, dmr); |
| kfree(dmr); |
| } |
| return err; |
| } |
| |
| static int create_dma_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) |
| { |
| int inlen = MLX5_ST_SZ_BYTES(create_mkey_in); |
| void *mkc; |
| u32 *in; |
| int err; |
| |
| in = kzalloc(inlen, GFP_KERNEL); |
| if (!in) |
| return -ENOMEM; |
| |
| mkc = MLX5_ADDR_OF(create_mkey_in, in, memory_key_mkey_entry); |
| |
| MLX5_SET(mkc, mkc, access_mode_1_0, MLX5_MKC_ACCESS_MODE_PA); |
| MLX5_SET(mkc, mkc, length64, 1); |
| MLX5_SET(mkc, mkc, lw, 1); |
| MLX5_SET(mkc, mkc, lr, 1); |
| MLX5_SET(mkc, mkc, pd, mvdev->res.pdn); |
| MLX5_SET(mkc, mkc, qpn, 0xffffff); |
| |
| err = mlx5_vdpa_create_mkey(mvdev, &mr->mkey, in, inlen); |
| if (!err) |
| mr->user_mr = false; |
| |
| kfree(in); |
| return err; |
| } |
| |
| static void destroy_dma_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) |
| { |
| mlx5_vdpa_destroy_mkey(mvdev, mr->mkey); |
| } |
| |
| static int dup_iotlb(struct vhost_iotlb *dst, struct vhost_iotlb *src) |
| { |
| struct vhost_iotlb_map *map; |
| u64 start = 0, last = ULLONG_MAX; |
| int err; |
| |
| if (dst == src) |
| return -EINVAL; |
| |
| if (!src) { |
| err = vhost_iotlb_add_range(dst, start, last, start, VHOST_ACCESS_RW); |
| return err; |
| } |
| |
| for (map = vhost_iotlb_itree_first(src, start, last); map; |
| map = vhost_iotlb_itree_next(map, start, last)) { |
| err = vhost_iotlb_add_range(dst, map->start, map->last, |
| map->addr, map->perm); |
| if (err) |
| return err; |
| } |
| return 0; |
| } |
| |
| static void prune_iotlb(struct vhost_iotlb *iotlb) |
| { |
| vhost_iotlb_del_range(iotlb, 0, ULLONG_MAX); |
| } |
| |
| static void destroy_user_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) |
| { |
| struct mlx5_vdpa_direct_mr *dmr; |
| struct mlx5_vdpa_direct_mr *n; |
| |
| destroy_indirect_key(mvdev, mr); |
| destroy_direct_keys(mvdev, mr); |
| list_for_each_entry_safe_reverse(dmr, n, &mr->head, list) { |
| list_del_init(&dmr->list); |
| unmap_direct_mr(mvdev, dmr); |
| kfree(dmr); |
| } |
| } |
| |
| static void _mlx5_vdpa_destroy_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) |
| { |
| if (WARN_ON(!mr)) |
| return; |
| |
| if (mr->user_mr) |
| destroy_user_mr(mvdev, mr); |
| else |
| destroy_dma_mr(mvdev, mr); |
| |
| vhost_iotlb_free(mr->iotlb); |
| |
| list_del(&mr->mr_list); |
| |
| kfree(mr); |
| } |
| |
| /* There can be multiple .set_map() operations in quick succession. |
| * This large delay is a simple way to prevent the MR cleanup from blocking |
| * .set_map() MR creation in this scenario. |
| */ |
| #define MLX5_VDPA_MR_GC_TRIGGER_MS 2000 |
| |
| static void mlx5_vdpa_mr_gc_handler(struct work_struct *work) |
| { |
| struct mlx5_vdpa_mr_resources *mres; |
| struct mlx5_vdpa_mr *mr, *tmp; |
| struct mlx5_vdpa_dev *mvdev; |
| |
| mres = container_of(work, struct mlx5_vdpa_mr_resources, gc_dwork_ent.work); |
| |
| if (atomic_read(&mres->shutdown)) { |
| mutex_lock(&mres->lock); |
| } else if (!mutex_trylock(&mres->lock)) { |
| queue_delayed_work(mres->wq_gc, &mres->gc_dwork_ent, |
| msecs_to_jiffies(MLX5_VDPA_MR_GC_TRIGGER_MS)); |
| return; |
| } |
| |
| mvdev = container_of(mres, struct mlx5_vdpa_dev, mres); |
| |
| list_for_each_entry_safe(mr, tmp, &mres->mr_gc_list_head, mr_list) { |
| _mlx5_vdpa_destroy_mr(mvdev, mr); |
| } |
| |
| mutex_unlock(&mres->lock); |
| } |
| |
| static void _mlx5_vdpa_put_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *mr) |
| { |
| struct mlx5_vdpa_mr_resources *mres = &mvdev->mres; |
| |
| if (!mr) |
| return; |
| |
| if (refcount_dec_and_test(&mr->refcount)) { |
| list_move_tail(&mr->mr_list, &mres->mr_gc_list_head); |
| queue_delayed_work(mres->wq_gc, &mres->gc_dwork_ent, |
| msecs_to_jiffies(MLX5_VDPA_MR_GC_TRIGGER_MS)); |
| } |
| } |
| |
| void mlx5_vdpa_put_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *mr) |
| { |
| mutex_lock(&mvdev->mres.lock); |
| _mlx5_vdpa_put_mr(mvdev, mr); |
| mutex_unlock(&mvdev->mres.lock); |
| } |
| |
| static void _mlx5_vdpa_get_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *mr) |
| { |
| if (!mr) |
| return; |
| |
| refcount_inc(&mr->refcount); |
| } |
| |
| void mlx5_vdpa_get_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *mr) |
| { |
| mutex_lock(&mvdev->mres.lock); |
| _mlx5_vdpa_get_mr(mvdev, mr); |
| mutex_unlock(&mvdev->mres.lock); |
| } |
| |
| void mlx5_vdpa_update_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *new_mr, |
| unsigned int asid) |
| { |
| struct mlx5_vdpa_mr *old_mr = mvdev->mres.mr[asid]; |
| |
| mutex_lock(&mvdev->mres.lock); |
| |
| _mlx5_vdpa_put_mr(mvdev, old_mr); |
| mvdev->mres.mr[asid] = new_mr; |
| |
| mutex_unlock(&mvdev->mres.lock); |
| } |
| |
| static void mlx5_vdpa_show_mr_leaks(struct mlx5_vdpa_dev *mvdev) |
| { |
| struct mlx5_vdpa_mr *mr; |
| |
| mutex_lock(&mvdev->mres.lock); |
| |
| list_for_each_entry(mr, &mvdev->mres.mr_list_head, mr_list) { |
| |
| mlx5_vdpa_warn(mvdev, "mkey still alive after resource delete: " |
| "mr: %p, mkey: 0x%x, refcount: %u\n", |
| mr, mr->mkey, refcount_read(&mr->refcount)); |
| } |
| |
| mutex_unlock(&mvdev->mres.lock); |
| |
| } |
| |
| void mlx5_vdpa_clean_mrs(struct mlx5_vdpa_dev *mvdev) |
| { |
| if (!mvdev->res.valid) |
| return; |
| |
| for (int i = 0; i < MLX5_VDPA_NUM_AS; i++) |
| mlx5_vdpa_update_mr(mvdev, NULL, i); |
| |
| prune_iotlb(mvdev->cvq.iotlb); |
| |
| mlx5_vdpa_show_mr_leaks(mvdev); |
| } |
| |
| static int _mlx5_vdpa_create_mr(struct mlx5_vdpa_dev *mvdev, |
| struct mlx5_vdpa_mr *mr, |
| struct vhost_iotlb *iotlb) |
| { |
| int err; |
| |
| if (iotlb) |
| err = create_user_mr(mvdev, mr, iotlb); |
| else |
| err = create_dma_mr(mvdev, mr); |
| |
| if (err) |
| return err; |
| |
| mr->iotlb = vhost_iotlb_alloc(0, 0); |
| if (!mr->iotlb) { |
| err = -ENOMEM; |
| goto err_mr; |
| } |
| |
| err = dup_iotlb(mr->iotlb, iotlb); |
| if (err) |
| goto err_iotlb; |
| |
| list_add_tail(&mr->mr_list, &mvdev->mres.mr_list_head); |
| |
| return 0; |
| |
| err_iotlb: |
| vhost_iotlb_free(mr->iotlb); |
| |
| err_mr: |
| if (iotlb) |
| destroy_user_mr(mvdev, mr); |
| else |
| destroy_dma_mr(mvdev, mr); |
| |
| return err; |
| } |
| |
| struct mlx5_vdpa_mr *mlx5_vdpa_create_mr(struct mlx5_vdpa_dev *mvdev, |
| struct vhost_iotlb *iotlb) |
| { |
| struct mlx5_vdpa_mr *mr; |
| int err; |
| |
| mr = kzalloc(sizeof(*mr), GFP_KERNEL); |
| if (!mr) |
| return ERR_PTR(-ENOMEM); |
| |
| mutex_lock(&mvdev->mres.lock); |
| err = _mlx5_vdpa_create_mr(mvdev, mr, iotlb); |
| mutex_unlock(&mvdev->mres.lock); |
| |
| if (err) |
| goto out_err; |
| |
| refcount_set(&mr->refcount, 1); |
| |
| return mr; |
| |
| out_err: |
| kfree(mr); |
| return ERR_PTR(err); |
| } |
| |
| int mlx5_vdpa_update_cvq_iotlb(struct mlx5_vdpa_dev *mvdev, |
| struct vhost_iotlb *iotlb, |
| unsigned int asid) |
| { |
| int err; |
| |
| if (mvdev->mres.group2asid[MLX5_VDPA_CVQ_GROUP] != asid) |
| return 0; |
| |
| spin_lock(&mvdev->cvq.iommu_lock); |
| |
| prune_iotlb(mvdev->cvq.iotlb); |
| err = dup_iotlb(mvdev->cvq.iotlb, iotlb); |
| |
| spin_unlock(&mvdev->cvq.iommu_lock); |
| |
| return err; |
| } |
| |
| int mlx5_vdpa_create_dma_mr(struct mlx5_vdpa_dev *mvdev) |
| { |
| struct mlx5_vdpa_mr *mr; |
| |
| mr = mlx5_vdpa_create_mr(mvdev, NULL); |
| if (IS_ERR(mr)) |
| return PTR_ERR(mr); |
| |
| mlx5_vdpa_update_mr(mvdev, mr, 0); |
| |
| return mlx5_vdpa_update_cvq_iotlb(mvdev, NULL, 0); |
| } |
| |
| int mlx5_vdpa_reset_mr(struct mlx5_vdpa_dev *mvdev, unsigned int asid) |
| { |
| if (asid >= MLX5_VDPA_NUM_AS) |
| return -EINVAL; |
| |
| mlx5_vdpa_update_mr(mvdev, NULL, asid); |
| |
| if (asid == 0 && MLX5_CAP_GEN(mvdev->mdev, umem_uid_0)) { |
| if (mlx5_vdpa_create_dma_mr(mvdev)) |
| mlx5_vdpa_warn(mvdev, "create DMA MR failed\n"); |
| } else { |
| mlx5_vdpa_update_cvq_iotlb(mvdev, NULL, asid); |
| } |
| |
| return 0; |
| } |
| |
| int mlx5_vdpa_init_mr_resources(struct mlx5_vdpa_dev *mvdev) |
| { |
| struct mlx5_vdpa_mr_resources *mres = &mvdev->mres; |
| |
| mres->wq_gc = create_singlethread_workqueue("mlx5_vdpa_mr_gc"); |
| if (!mres->wq_gc) |
| return -ENOMEM; |
| |
| INIT_DELAYED_WORK(&mres->gc_dwork_ent, mlx5_vdpa_mr_gc_handler); |
| |
| mutex_init(&mres->lock); |
| |
| INIT_LIST_HEAD(&mres->mr_list_head); |
| INIT_LIST_HEAD(&mres->mr_gc_list_head); |
| |
| return 0; |
| } |
| |
| void mlx5_vdpa_destroy_mr_resources(struct mlx5_vdpa_dev *mvdev) |
| { |
| struct mlx5_vdpa_mr_resources *mres = &mvdev->mres; |
| |
| atomic_set(&mres->shutdown, 1); |
| |
| flush_delayed_work(&mres->gc_dwork_ent); |
| destroy_workqueue(mres->wq_gc); |
| mres->wq_gc = NULL; |
| mutex_destroy(&mres->lock); |
| } |