| // SPDX-License-Identifier: GPL-2.0 OR MIT |
| /* |
| * Copyright (c) 2007-2008 Tungsten Graphics, Inc., Cedar Park, TX., USA, |
| * Copyright (c) 2009 VMware, Inc., Palo Alto, CA., USA, |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sub license, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice (including the |
| * next paragraph) shall be included in all copies or substantial portions |
| * of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL |
| * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, |
| * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR |
| * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE |
| * USE OR OTHER DEALINGS IN THE SOFTWARE. |
| */ |
| |
| #include <linux/limits.h> |
| #include <linux/swiotlb.h> |
| |
| #include "nouveau_drv.h" |
| #include "nouveau_gem.h" |
| #include "nouveau_mem.h" |
| #include "nouveau_ttm.h" |
| |
| #include <drm/drm_legacy.h> |
| |
| #include <core/tegra.h> |
| |
| static void |
| nouveau_manager_del(struct ttm_resource_manager *man, struct ttm_resource *reg) |
| { |
| nouveau_mem_del(reg); |
| } |
| |
| static int |
| nouveau_vram_manager_new(struct ttm_resource_manager *man, |
| struct ttm_buffer_object *bo, |
| const struct ttm_place *place, |
| struct ttm_resource *reg) |
| { |
| struct nouveau_bo *nvbo = nouveau_bo(bo); |
| struct nouveau_drm *drm = nouveau_bdev(bo->bdev); |
| int ret; |
| |
| if (drm->client.device.info.ram_size == 0) |
| return -ENOMEM; |
| |
| ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, reg); |
| if (ret) |
| return ret; |
| |
| ret = nouveau_mem_vram(reg, nvbo->contig, nvbo->page); |
| if (ret) { |
| nouveau_mem_del(reg); |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| const struct ttm_resource_manager_func nouveau_vram_manager = { |
| .alloc = nouveau_vram_manager_new, |
| .free = nouveau_manager_del, |
| }; |
| |
| static int |
| nouveau_gart_manager_new(struct ttm_resource_manager *man, |
| struct ttm_buffer_object *bo, |
| const struct ttm_place *place, |
| struct ttm_resource *reg) |
| { |
| struct nouveau_bo *nvbo = nouveau_bo(bo); |
| struct nouveau_drm *drm = nouveau_bdev(bo->bdev); |
| int ret; |
| |
| ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, reg); |
| if (ret) |
| return ret; |
| |
| reg->start = 0; |
| return 0; |
| } |
| |
| const struct ttm_resource_manager_func nouveau_gart_manager = { |
| .alloc = nouveau_gart_manager_new, |
| .free = nouveau_manager_del, |
| }; |
| |
| static int |
| nv04_gart_manager_new(struct ttm_resource_manager *man, |
| struct ttm_buffer_object *bo, |
| const struct ttm_place *place, |
| struct ttm_resource *reg) |
| { |
| struct nouveau_bo *nvbo = nouveau_bo(bo); |
| struct nouveau_drm *drm = nouveau_bdev(bo->bdev); |
| struct nouveau_mem *mem; |
| int ret; |
| |
| ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, reg); |
| mem = nouveau_mem(reg); |
| if (ret) |
| return ret; |
| |
| ret = nvif_vmm_get(&mem->cli->vmm.vmm, PTES, false, 12, 0, |
| (long)reg->num_pages << PAGE_SHIFT, &mem->vma[0]); |
| if (ret) { |
| nouveau_mem_del(reg); |
| return ret; |
| } |
| |
| reg->start = mem->vma[0].addr >> PAGE_SHIFT; |
| return 0; |
| } |
| |
| const struct ttm_resource_manager_func nv04_gart_manager = { |
| .alloc = nv04_gart_manager_new, |
| .free = nouveau_manager_del, |
| }; |
| |
| static vm_fault_t nouveau_ttm_fault(struct vm_fault *vmf) |
| { |
| struct vm_area_struct *vma = vmf->vma; |
| struct ttm_buffer_object *bo = vma->vm_private_data; |
| pgprot_t prot; |
| vm_fault_t ret; |
| |
| ret = ttm_bo_vm_reserve(bo, vmf); |
| if (ret) |
| return ret; |
| |
| ret = nouveau_ttm_fault_reserve_notify(bo); |
| if (ret) |
| goto error_unlock; |
| |
| nouveau_bo_del_io_reserve_lru(bo); |
| prot = vm_get_page_prot(vma->vm_flags); |
| ret = ttm_bo_vm_fault_reserved(vmf, prot, TTM_BO_VM_NUM_PREFAULT, 1); |
| nouveau_bo_add_io_reserve_lru(bo); |
| if (ret == VM_FAULT_RETRY && !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) |
| return ret; |
| |
| error_unlock: |
| dma_resv_unlock(bo->base.resv); |
| return ret; |
| } |
| |
| static struct vm_operations_struct nouveau_ttm_vm_ops = { |
| .fault = nouveau_ttm_fault, |
| .open = ttm_bo_vm_open, |
| .close = ttm_bo_vm_close, |
| .access = ttm_bo_vm_access |
| }; |
| |
| int |
| nouveau_ttm_mmap(struct file *filp, struct vm_area_struct *vma) |
| { |
| struct drm_file *file_priv = filp->private_data; |
| struct nouveau_drm *drm = nouveau_drm(file_priv->minor->dev); |
| int ret; |
| |
| ret = ttm_bo_mmap(filp, vma, &drm->ttm.bdev); |
| if (ret) |
| return ret; |
| |
| vma->vm_ops = &nouveau_ttm_vm_ops; |
| return 0; |
| } |
| |
| static int |
| nouveau_ttm_init_host(struct nouveau_drm *drm, u8 kind) |
| { |
| struct nvif_mmu *mmu = &drm->client.mmu; |
| int typei; |
| |
| typei = nvif_mmu_type(mmu, NVIF_MEM_HOST | NVIF_MEM_MAPPABLE | |
| kind | NVIF_MEM_COHERENT); |
| if (typei < 0) |
| return -ENOSYS; |
| |
| drm->ttm.type_host[!!kind] = typei; |
| |
| typei = nvif_mmu_type(mmu, NVIF_MEM_HOST | NVIF_MEM_MAPPABLE | kind); |
| if (typei < 0) |
| return -ENOSYS; |
| |
| drm->ttm.type_ncoh[!!kind] = typei; |
| return 0; |
| } |
| |
| static int |
| nouveau_ttm_init_vram(struct nouveau_drm *drm) |
| { |
| if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_TESLA) { |
| struct ttm_resource_manager *man = kzalloc(sizeof(*man), GFP_KERNEL); |
| |
| if (!man) |
| return -ENOMEM; |
| |
| man->func = &nouveau_vram_manager; |
| |
| ttm_resource_manager_init(man, |
| drm->gem.vram_available >> PAGE_SHIFT); |
| ttm_set_driver_manager(&drm->ttm.bdev, TTM_PL_VRAM, man); |
| ttm_resource_manager_set_used(man, true); |
| return 0; |
| } else { |
| return ttm_range_man_init(&drm->ttm.bdev, TTM_PL_VRAM, false, |
| drm->gem.vram_available >> PAGE_SHIFT); |
| } |
| } |
| |
| static void |
| nouveau_ttm_fini_vram(struct nouveau_drm *drm) |
| { |
| struct ttm_resource_manager *man = ttm_manager_type(&drm->ttm.bdev, TTM_PL_VRAM); |
| |
| if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_TESLA) { |
| ttm_resource_manager_set_used(man, false); |
| ttm_resource_manager_evict_all(&drm->ttm.bdev, man); |
| ttm_resource_manager_cleanup(man); |
| ttm_set_driver_manager(&drm->ttm.bdev, TTM_PL_VRAM, NULL); |
| kfree(man); |
| } else |
| ttm_range_man_fini(&drm->ttm.bdev, TTM_PL_VRAM); |
| } |
| |
| static int |
| nouveau_ttm_init_gtt(struct nouveau_drm *drm) |
| { |
| struct ttm_resource_manager *man; |
| unsigned long size_pages = drm->gem.gart_available >> PAGE_SHIFT; |
| const struct ttm_resource_manager_func *func = NULL; |
| |
| if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_TESLA) |
| func = &nouveau_gart_manager; |
| else if (!drm->agp.bridge) |
| func = &nv04_gart_manager; |
| else |
| return ttm_range_man_init(&drm->ttm.bdev, TTM_PL_TT, true, |
| size_pages); |
| |
| man = kzalloc(sizeof(*man), GFP_KERNEL); |
| if (!man) |
| return -ENOMEM; |
| |
| man->func = func; |
| man->use_tt = true; |
| ttm_resource_manager_init(man, size_pages); |
| ttm_set_driver_manager(&drm->ttm.bdev, TTM_PL_TT, man); |
| ttm_resource_manager_set_used(man, true); |
| return 0; |
| } |
| |
| static void |
| nouveau_ttm_fini_gtt(struct nouveau_drm *drm) |
| { |
| struct ttm_resource_manager *man = ttm_manager_type(&drm->ttm.bdev, TTM_PL_TT); |
| |
| if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA && |
| drm->agp.bridge) |
| ttm_range_man_fini(&drm->ttm.bdev, TTM_PL_TT); |
| else { |
| ttm_resource_manager_set_used(man, false); |
| ttm_resource_manager_evict_all(&drm->ttm.bdev, man); |
| ttm_resource_manager_cleanup(man); |
| ttm_set_driver_manager(&drm->ttm.bdev, TTM_PL_TT, NULL); |
| kfree(man); |
| } |
| } |
| |
| int |
| nouveau_ttm_init(struct nouveau_drm *drm) |
| { |
| struct nvkm_device *device = nvxx_device(&drm->client.device); |
| struct nvkm_pci *pci = device->pci; |
| struct nvif_mmu *mmu = &drm->client.mmu; |
| struct drm_device *dev = drm->dev; |
| bool need_swiotlb = false; |
| int typei, ret; |
| |
| ret = nouveau_ttm_init_host(drm, 0); |
| if (ret) |
| return ret; |
| |
| if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_TESLA && |
| drm->client.device.info.chipset != 0x50) { |
| ret = nouveau_ttm_init_host(drm, NVIF_MEM_KIND); |
| if (ret) |
| return ret; |
| } |
| |
| if (drm->client.device.info.platform != NV_DEVICE_INFO_V0_SOC && |
| drm->client.device.info.family >= NV_DEVICE_INFO_V0_TESLA) { |
| typei = nvif_mmu_type(mmu, NVIF_MEM_VRAM | NVIF_MEM_MAPPABLE | |
| NVIF_MEM_KIND | |
| NVIF_MEM_COMP | |
| NVIF_MEM_DISP); |
| if (typei < 0) |
| return -ENOSYS; |
| |
| drm->ttm.type_vram = typei; |
| } else { |
| drm->ttm.type_vram = -1; |
| } |
| |
| if (pci && pci->agp.bridge) { |
| drm->agp.bridge = pci->agp.bridge; |
| drm->agp.base = pci->agp.base; |
| drm->agp.size = pci->agp.size; |
| drm->agp.cma = pci->agp.cma; |
| } |
| |
| #if IS_ENABLED(CONFIG_SWIOTLB) && IS_ENABLED(CONFIG_X86) |
| need_swiotlb = is_swiotlb_active(); |
| #endif |
| |
| ret = ttm_bo_device_init(&drm->ttm.bdev, &nouveau_bo_driver, |
| drm->dev->dev, dev->anon_inode->i_mapping, |
| dev->vma_offset_manager, need_swiotlb, |
| drm->client.mmu.dmabits <= 32); |
| if (ret) { |
| NV_ERROR(drm, "error initialising bo driver, %d\n", ret); |
| return ret; |
| } |
| |
| /* VRAM init */ |
| drm->gem.vram_available = drm->client.device.info.ram_user; |
| |
| arch_io_reserve_memtype_wc(device->func->resource_addr(device, 1), |
| device->func->resource_size(device, 1)); |
| |
| ret = nouveau_ttm_init_vram(drm); |
| if (ret) { |
| NV_ERROR(drm, "VRAM mm init failed, %d\n", ret); |
| return ret; |
| } |
| |
| drm->ttm.mtrr = arch_phys_wc_add(device->func->resource_addr(device, 1), |
| device->func->resource_size(device, 1)); |
| |
| /* GART init */ |
| if (!drm->agp.bridge) { |
| drm->gem.gart_available = drm->client.vmm.vmm.limit; |
| } else { |
| drm->gem.gart_available = drm->agp.size; |
| } |
| |
| ret = nouveau_ttm_init_gtt(drm); |
| if (ret) { |
| NV_ERROR(drm, "GART mm init failed, %d\n", ret); |
| return ret; |
| } |
| |
| mutex_init(&drm->ttm.io_reserve_mutex); |
| INIT_LIST_HEAD(&drm->ttm.io_reserve_lru); |
| |
| NV_INFO(drm, "VRAM: %d MiB\n", (u32)(drm->gem.vram_available >> 20)); |
| NV_INFO(drm, "GART: %d MiB\n", (u32)(drm->gem.gart_available >> 20)); |
| return 0; |
| } |
| |
| void |
| nouveau_ttm_fini(struct nouveau_drm *drm) |
| { |
| struct nvkm_device *device = nvxx_device(&drm->client.device); |
| |
| nouveau_ttm_fini_vram(drm); |
| nouveau_ttm_fini_gtt(drm); |
| |
| ttm_bo_device_release(&drm->ttm.bdev); |
| |
| arch_phys_wc_del(drm->ttm.mtrr); |
| drm->ttm.mtrr = 0; |
| arch_io_free_memtype_wc(device->func->resource_addr(device, 1), |
| device->func->resource_size(device, 1)); |
| |
| } |