| // SPDX-License-Identifier: MIT |
| /* |
| * Copyright © 2019 Intel Corporation |
| */ |
| |
| #include "i915_drv.h" |
| #include "intel_memory_region.h" |
| #include "gem/i915_gem_lmem.h" |
| #include "gem/i915_gem_region.h" |
| #include "intel_region_lmem.h" |
| |
| static int init_fake_lmem_bar(struct intel_memory_region *mem) |
| { |
| struct drm_i915_private *i915 = mem->i915; |
| struct i915_ggtt *ggtt = &i915->ggtt; |
| unsigned long n; |
| int ret; |
| |
| /* We want to 1:1 map the mappable aperture to our reserved region */ |
| |
| mem->fake_mappable.start = 0; |
| mem->fake_mappable.size = resource_size(&mem->region); |
| mem->fake_mappable.color = I915_COLOR_UNEVICTABLE; |
| |
| ret = drm_mm_reserve_node(&ggtt->vm.mm, &mem->fake_mappable); |
| if (ret) |
| return ret; |
| |
| mem->remap_addr = dma_map_resource(&i915->drm.pdev->dev, |
| mem->region.start, |
| mem->fake_mappable.size, |
| PCI_DMA_BIDIRECTIONAL, |
| DMA_ATTR_FORCE_CONTIGUOUS); |
| if (dma_mapping_error(&i915->drm.pdev->dev, mem->remap_addr)) { |
| drm_mm_remove_node(&mem->fake_mappable); |
| return -EINVAL; |
| } |
| |
| for (n = 0; n < mem->fake_mappable.size >> PAGE_SHIFT; ++n) { |
| ggtt->vm.insert_page(&ggtt->vm, |
| mem->remap_addr + (n << PAGE_SHIFT), |
| n << PAGE_SHIFT, |
| I915_CACHE_NONE, 0); |
| } |
| |
| mem->region = (struct resource)DEFINE_RES_MEM(mem->remap_addr, |
| mem->fake_mappable.size); |
| |
| return 0; |
| } |
| |
| static void release_fake_lmem_bar(struct intel_memory_region *mem) |
| { |
| if (drm_mm_node_allocated(&mem->fake_mappable)) |
| drm_mm_remove_node(&mem->fake_mappable); |
| |
| dma_unmap_resource(&mem->i915->drm.pdev->dev, |
| mem->remap_addr, |
| mem->fake_mappable.size, |
| PCI_DMA_BIDIRECTIONAL, |
| DMA_ATTR_FORCE_CONTIGUOUS); |
| } |
| |
| static void |
| region_lmem_release(struct intel_memory_region *mem) |
| { |
| release_fake_lmem_bar(mem); |
| io_mapping_fini(&mem->iomap); |
| intel_memory_region_release_buddy(mem); |
| } |
| |
| static int |
| region_lmem_init(struct intel_memory_region *mem) |
| { |
| int ret; |
| |
| if (i915_modparams.fake_lmem_start) { |
| ret = init_fake_lmem_bar(mem); |
| GEM_BUG_ON(ret); |
| } |
| |
| if (!io_mapping_init_wc(&mem->iomap, |
| mem->io_start, |
| resource_size(&mem->region))) |
| return -EIO; |
| |
| ret = intel_memory_region_init_buddy(mem); |
| if (ret) |
| io_mapping_fini(&mem->iomap); |
| |
| return ret; |
| } |
| |
| const struct intel_memory_region_ops intel_region_lmem_ops = { |
| .init = region_lmem_init, |
| .release = region_lmem_release, |
| .create_object = __i915_gem_lmem_object_create, |
| }; |
| |
| struct intel_memory_region * |
| intel_setup_fake_lmem(struct drm_i915_private *i915) |
| { |
| struct pci_dev *pdev = i915->drm.pdev; |
| struct intel_memory_region *mem; |
| resource_size_t mappable_end; |
| resource_size_t io_start; |
| resource_size_t start; |
| |
| GEM_BUG_ON(i915_ggtt_has_aperture(&i915->ggtt)); |
| GEM_BUG_ON(!i915_modparams.fake_lmem_start); |
| |
| /* Your mappable aperture belongs to me now! */ |
| mappable_end = pci_resource_len(pdev, 2); |
| io_start = pci_resource_start(pdev, 2), |
| start = i915_modparams.fake_lmem_start; |
| |
| mem = intel_memory_region_create(i915, |
| start, |
| mappable_end, |
| PAGE_SIZE, |
| io_start, |
| &intel_region_lmem_ops); |
| if (!IS_ERR(mem)) { |
| DRM_INFO("Intel graphics fake LMEM: %pR\n", &mem->region); |
| DRM_INFO("Intel graphics fake LMEM IO start: %llx\n", |
| (u64)mem->io_start); |
| DRM_INFO("Intel graphics fake LMEM size: %llx\n", |
| (u64)resource_size(&mem->region)); |
| } |
| |
| return mem; |
| } |