| /* |
| * Copyright 2017 Red Hat Inc. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in |
| * all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR |
| * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
| * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR |
| * OTHER DEALINGS IN THE SOFTWARE. |
| */ |
| #include "uvmm.h" |
| #include "umem.h" |
| #include "ummu.h" |
| |
| #include <core/client.h> |
| #include <core/memory.h> |
| |
| #include <nvif/if000c.h> |
| #include <nvif/unpack.h> |
| |
| static const struct nvkm_object_func nvkm_uvmm; |
| struct nvkm_vmm * |
| nvkm_uvmm_search(struct nvkm_client *client, u64 handle) |
| { |
| struct nvkm_object *object; |
| |
| object = nvkm_object_search(client, handle, &nvkm_uvmm); |
| if (IS_ERR(object)) |
| return (void *)object; |
| |
| return nvkm_uvmm(object)->vmm; |
| } |
| |
| static int |
| nvkm_uvmm_mthd_unmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc) |
| { |
| struct nvkm_client *client = uvmm->object.client; |
| union { |
| struct nvif_vmm_unmap_v0 v0; |
| } *args = argv; |
| struct nvkm_vmm *vmm = uvmm->vmm; |
| struct nvkm_vma *vma; |
| int ret = -ENOSYS; |
| u64 addr; |
| |
| if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { |
| addr = args->v0.addr; |
| } else |
| return ret; |
| |
| mutex_lock(&vmm->mutex); |
| vma = nvkm_vmm_node_search(vmm, addr); |
| if (ret = -ENOENT, !vma || vma->addr != addr) { |
| VMM_DEBUG(vmm, "lookup %016llx: %016llx", |
| addr, vma ? vma->addr : ~0ULL); |
| goto done; |
| } |
| |
| if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) { |
| VMM_DEBUG(vmm, "denied %016llx: %d %d %d", addr, |
| vma->user, !client->super, vma->busy); |
| goto done; |
| } |
| |
| if (ret = -EINVAL, !vma->memory) { |
| VMM_DEBUG(vmm, "unmapped"); |
| goto done; |
| } |
| |
| nvkm_vmm_unmap_locked(vmm, vma); |
| ret = 0; |
| done: |
| mutex_unlock(&vmm->mutex); |
| return ret; |
| } |
| |
| static int |
| nvkm_uvmm_mthd_map(struct nvkm_uvmm *uvmm, void *argv, u32 argc) |
| { |
| struct nvkm_client *client = uvmm->object.client; |
| union { |
| struct nvif_vmm_map_v0 v0; |
| } *args = argv; |
| u64 addr, size, handle, offset; |
| struct nvkm_vmm *vmm = uvmm->vmm; |
| struct nvkm_vma *vma; |
| struct nvkm_memory *memory; |
| int ret = -ENOSYS; |
| |
| if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { |
| addr = args->v0.addr; |
| size = args->v0.size; |
| handle = args->v0.memory; |
| offset = args->v0.offset; |
| } else |
| return ret; |
| |
| memory = nvkm_umem_search(client, handle); |
| if (IS_ERR(memory)) { |
| VMM_DEBUG(vmm, "memory %016llx %ld\n", handle, PTR_ERR(memory)); |
| return PTR_ERR(memory); |
| } |
| |
| mutex_lock(&vmm->mutex); |
| if (ret = -ENOENT, !(vma = nvkm_vmm_node_search(vmm, addr))) { |
| VMM_DEBUG(vmm, "lookup %016llx", addr); |
| goto fail; |
| } |
| |
| if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) { |
| VMM_DEBUG(vmm, "denied %016llx: %d %d %d", addr, |
| vma->user, !client->super, vma->busy); |
| goto fail; |
| } |
| |
| if (ret = -EINVAL, vma->addr != addr || vma->size != size) { |
| if (addr + size > vma->addr + vma->size || vma->memory || |
| (vma->refd == NVKM_VMA_PAGE_NONE && !vma->mapref)) { |
| VMM_DEBUG(vmm, "split %d %d %d " |
| "%016llx %016llx %016llx %016llx", |
| !!vma->memory, vma->refd, vma->mapref, |
| addr, size, vma->addr, (u64)vma->size); |
| goto fail; |
| } |
| |
| if (vma->addr != addr) { |
| const u64 tail = vma->size + vma->addr - addr; |
| if (ret = -ENOMEM, !(vma = nvkm_vma_tail(vma, tail))) |
| goto fail; |
| vma->part = true; |
| nvkm_vmm_node_insert(vmm, vma); |
| } |
| |
| if (vma->size != size) { |
| const u64 tail = vma->size - size; |
| struct nvkm_vma *tmp; |
| if (ret = -ENOMEM, !(tmp = nvkm_vma_tail(vma, tail))) { |
| nvkm_vmm_unmap_region(vmm, vma); |
| goto fail; |
| } |
| tmp->part = true; |
| nvkm_vmm_node_insert(vmm, tmp); |
| } |
| } |
| vma->busy = true; |
| mutex_unlock(&vmm->mutex); |
| |
| ret = nvkm_memory_map(memory, offset, vmm, vma, argv, argc); |
| if (ret == 0) { |
| /* Successful map will clear vma->busy. */ |
| nvkm_memory_unref(&memory); |
| return 0; |
| } |
| |
| mutex_lock(&vmm->mutex); |
| vma->busy = false; |
| nvkm_vmm_unmap_region(vmm, vma); |
| fail: |
| mutex_unlock(&vmm->mutex); |
| nvkm_memory_unref(&memory); |
| return ret; |
| } |
| |
| static int |
| nvkm_uvmm_mthd_put(struct nvkm_uvmm *uvmm, void *argv, u32 argc) |
| { |
| struct nvkm_client *client = uvmm->object.client; |
| union { |
| struct nvif_vmm_put_v0 v0; |
| } *args = argv; |
| struct nvkm_vmm *vmm = uvmm->vmm; |
| struct nvkm_vma *vma; |
| int ret = -ENOSYS; |
| u64 addr; |
| |
| if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { |
| addr = args->v0.addr; |
| } else |
| return ret; |
| |
| mutex_lock(&vmm->mutex); |
| vma = nvkm_vmm_node_search(vmm, args->v0.addr); |
| if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) { |
| VMM_DEBUG(vmm, "lookup %016llx: %016llx %d", addr, |
| vma ? vma->addr : ~0ULL, vma ? vma->part : 0); |
| goto done; |
| } |
| |
| if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) { |
| VMM_DEBUG(vmm, "denied %016llx: %d %d %d", addr, |
| vma->user, !client->super, vma->busy); |
| goto done; |
| } |
| |
| nvkm_vmm_put_locked(vmm, vma); |
| ret = 0; |
| done: |
| mutex_unlock(&vmm->mutex); |
| return ret; |
| } |
| |
| static int |
| nvkm_uvmm_mthd_get(struct nvkm_uvmm *uvmm, void *argv, u32 argc) |
| { |
| struct nvkm_client *client = uvmm->object.client; |
| union { |
| struct nvif_vmm_get_v0 v0; |
| } *args = argv; |
| struct nvkm_vmm *vmm = uvmm->vmm; |
| struct nvkm_vma *vma; |
| int ret = -ENOSYS; |
| bool getref, mapref, sparse; |
| u8 page, align; |
| u64 size; |
| |
| if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { |
| getref = args->v0.type == NVIF_VMM_GET_V0_PTES; |
| mapref = args->v0.type == NVIF_VMM_GET_V0_ADDR; |
| sparse = args->v0.sparse; |
| page = args->v0.page; |
| align = args->v0.align; |
| size = args->v0.size; |
| } else |
| return ret; |
| |
| mutex_lock(&vmm->mutex); |
| ret = nvkm_vmm_get_locked(vmm, getref, mapref, sparse, |
| page, align, size, &vma); |
| mutex_unlock(&vmm->mutex); |
| if (ret) |
| return ret; |
| |
| args->v0.addr = vma->addr; |
| vma->user = !client->super; |
| return ret; |
| } |
| |
| static int |
| nvkm_uvmm_mthd_page(struct nvkm_uvmm *uvmm, void *argv, u32 argc) |
| { |
| union { |
| struct nvif_vmm_page_v0 v0; |
| } *args = argv; |
| const struct nvkm_vmm_page *page; |
| int ret = -ENOSYS; |
| u8 type, index, nr; |
| |
| page = uvmm->vmm->func->page; |
| for (nr = 0; page[nr].shift; nr++); |
| |
| if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { |
| if ((index = args->v0.index) >= nr) |
| return -EINVAL; |
| type = page[index].type; |
| args->v0.shift = page[index].shift; |
| args->v0.sparse = !!(type & NVKM_VMM_PAGE_SPARSE); |
| args->v0.vram = !!(type & NVKM_VMM_PAGE_VRAM); |
| args->v0.host = !!(type & NVKM_VMM_PAGE_HOST); |
| args->v0.comp = !!(type & NVKM_VMM_PAGE_COMP); |
| } else |
| return -ENOSYS; |
| |
| return 0; |
| } |
| |
| static int |
| nvkm_uvmm_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc) |
| { |
| struct nvkm_uvmm *uvmm = nvkm_uvmm(object); |
| switch (mthd) { |
| case NVIF_VMM_V0_PAGE : return nvkm_uvmm_mthd_page (uvmm, argv, argc); |
| case NVIF_VMM_V0_GET : return nvkm_uvmm_mthd_get (uvmm, argv, argc); |
| case NVIF_VMM_V0_PUT : return nvkm_uvmm_mthd_put (uvmm, argv, argc); |
| case NVIF_VMM_V0_MAP : return nvkm_uvmm_mthd_map (uvmm, argv, argc); |
| case NVIF_VMM_V0_UNMAP : return nvkm_uvmm_mthd_unmap (uvmm, argv, argc); |
| default: |
| break; |
| } |
| return -EINVAL; |
| } |
| |
| static void * |
| nvkm_uvmm_dtor(struct nvkm_object *object) |
| { |
| struct nvkm_uvmm *uvmm = nvkm_uvmm(object); |
| nvkm_vmm_unref(&uvmm->vmm); |
| return uvmm; |
| } |
| |
| static const struct nvkm_object_func |
| nvkm_uvmm = { |
| .dtor = nvkm_uvmm_dtor, |
| .mthd = nvkm_uvmm_mthd, |
| }; |
| |
| int |
| nvkm_uvmm_new(const struct nvkm_oclass *oclass, void *argv, u32 argc, |
| struct nvkm_object **pobject) |
| { |
| struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu; |
| const bool more = oclass->base.maxver >= 0; |
| union { |
| struct nvif_vmm_v0 v0; |
| } *args = argv; |
| const struct nvkm_vmm_page *page; |
| struct nvkm_uvmm *uvmm; |
| int ret = -ENOSYS; |
| u64 addr, size; |
| |
| if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, more))) { |
| addr = args->v0.addr; |
| size = args->v0.size; |
| } else |
| return ret; |
| |
| if (!(uvmm = kzalloc(sizeof(*uvmm), GFP_KERNEL))) |
| return -ENOMEM; |
| nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object); |
| *pobject = &uvmm->object; |
| |
| if (!mmu->vmm) { |
| ret = mmu->func->vmm.ctor(mmu, addr, size, argv, argc, |
| NULL, "user", &uvmm->vmm); |
| if (ret) |
| return ret; |
| |
| uvmm->vmm->debug = max(uvmm->vmm->debug, oclass->client->debug); |
| } else { |
| if (size) |
| return -EINVAL; |
| |
| uvmm->vmm = nvkm_vmm_ref(mmu->vmm); |
| } |
| |
| page = uvmm->vmm->func->page; |
| args->v0.page_nr = 0; |
| while (page && (page++)->shift) |
| args->v0.page_nr++; |
| args->v0.addr = uvmm->vmm->start; |
| args->v0.size = uvmm->vmm->limit; |
| return 0; |
| } |