| // SPDX-License-Identifier: MIT |
| /* |
| * Copyright © 2020 Intel Corporation |
| */ |
| |
| #include <drm/drm_fourcc.h> |
| |
| #include "display/intel_display.h" |
| #include "gem/i915_gem_ioctls.h" |
| #include "gem/i915_gem_lmem.h" |
| #include "gem/i915_gem_region.h" |
| #include "pxp/intel_pxp.h" |
| |
| #include "i915_drv.h" |
| #include "i915_gem_create.h" |
| #include "i915_trace.h" |
| #include "i915_user_extensions.h" |
| |
| static u32 object_max_page_size(struct intel_memory_region **placements, |
| unsigned int n_placements) |
| { |
| u32 max_page_size = 0; |
| int i; |
| |
| for (i = 0; i < n_placements; i++) { |
| struct intel_memory_region *mr = placements[i]; |
| |
| GEM_BUG_ON(!is_power_of_2(mr->min_page_size)); |
| max_page_size = max_t(u32, max_page_size, mr->min_page_size); |
| } |
| |
| GEM_BUG_ON(!max_page_size); |
| return max_page_size; |
| } |
| |
| static int object_set_placements(struct drm_i915_gem_object *obj, |
| struct intel_memory_region **placements, |
| unsigned int n_placements) |
| { |
| struct intel_memory_region **arr; |
| unsigned int i; |
| |
| GEM_BUG_ON(!n_placements); |
| |
| /* |
| * For the common case of one memory region, skip storing an |
| * allocated array and just point at the region directly. |
| */ |
| if (n_placements == 1) { |
| struct intel_memory_region *mr = placements[0]; |
| struct drm_i915_private *i915 = mr->i915; |
| |
| obj->mm.placements = &i915->mm.regions[mr->id]; |
| obj->mm.n_placements = 1; |
| } else { |
| arr = kmalloc_array(n_placements, |
| sizeof(struct intel_memory_region *), |
| GFP_KERNEL); |
| if (!arr) |
| return -ENOMEM; |
| |
| for (i = 0; i < n_placements; i++) |
| arr[i] = placements[i]; |
| |
| obj->mm.placements = arr; |
| obj->mm.n_placements = n_placements; |
| } |
| |
| return 0; |
| } |
| |
| static int i915_gem_publish(struct drm_i915_gem_object *obj, |
| struct drm_file *file, |
| u64 *size_p, |
| u32 *handle_p) |
| { |
| u64 size = obj->base.size; |
| int ret; |
| |
| ret = drm_gem_handle_create(file, &obj->base, handle_p); |
| /* drop reference from allocate - handle holds it now */ |
| i915_gem_object_put(obj); |
| if (ret) |
| return ret; |
| |
| *size_p = size; |
| return 0; |
| } |
| |
| static struct drm_i915_gem_object * |
| __i915_gem_object_create_user_ext(struct drm_i915_private *i915, u64 size, |
| struct intel_memory_region **placements, |
| unsigned int n_placements, |
| unsigned int ext_flags) |
| { |
| struct intel_memory_region *mr = placements[0]; |
| struct drm_i915_gem_object *obj; |
| unsigned int flags; |
| int ret; |
| |
| i915_gem_flush_free_objects(i915); |
| |
| size = round_up(size, object_max_page_size(placements, n_placements)); |
| if (size == 0) |
| return ERR_PTR(-EINVAL); |
| |
| /* For most of the ABI (e.g. mmap) we think in system pages */ |
| GEM_BUG_ON(!IS_ALIGNED(size, PAGE_SIZE)); |
| |
| if (i915_gem_object_size_2big(size)) |
| return ERR_PTR(-E2BIG); |
| |
| obj = i915_gem_object_alloc(); |
| if (!obj) |
| return ERR_PTR(-ENOMEM); |
| |
| ret = object_set_placements(obj, placements, n_placements); |
| if (ret) |
| goto object_free; |
| |
| /* |
| * I915_BO_ALLOC_USER will make sure the object is cleared before |
| * any user access. |
| */ |
| flags = I915_BO_ALLOC_USER; |
| |
| ret = mr->ops->init_object(mr, obj, I915_BO_INVALID_OFFSET, size, 0, flags); |
| if (ret) |
| goto object_free; |
| |
| GEM_BUG_ON(size != obj->base.size); |
| |
| /* Add any flag set by create_ext options */ |
| obj->flags |= ext_flags; |
| |
| trace_i915_gem_object_create(obj); |
| return obj; |
| |
| object_free: |
| if (obj->mm.n_placements > 1) |
| kfree(obj->mm.placements); |
| i915_gem_object_free(obj); |
| return ERR_PTR(ret); |
| } |
| |
| /** |
| * __i915_gem_object_create_user - Creates a new object using the same path as |
| * DRM_I915_GEM_CREATE_EXT |
| * @i915: i915 private |
| * @size: size of the buffer, in bytes |
| * @placements: possible placement regions, in priority order |
| * @n_placements: number of possible placement regions |
| * |
| * This function is exposed primarily for selftests and does very little |
| * error checking. It is assumed that the set of placement regions has |
| * already been verified to be valid. |
| */ |
| struct drm_i915_gem_object * |
| __i915_gem_object_create_user(struct drm_i915_private *i915, u64 size, |
| struct intel_memory_region **placements, |
| unsigned int n_placements) |
| { |
| return __i915_gem_object_create_user_ext(i915, size, placements, |
| n_placements, 0); |
| } |
| |
| int |
| i915_gem_dumb_create(struct drm_file *file, |
| struct drm_device *dev, |
| struct drm_mode_create_dumb *args) |
| { |
| struct drm_i915_gem_object *obj; |
| struct intel_memory_region *mr; |
| enum intel_memory_type mem_type; |
| int cpp = DIV_ROUND_UP(args->bpp, 8); |
| u32 format; |
| |
| switch (cpp) { |
| case 1: |
| format = DRM_FORMAT_C8; |
| break; |
| case 2: |
| format = DRM_FORMAT_RGB565; |
| break; |
| case 4: |
| format = DRM_FORMAT_XRGB8888; |
| break; |
| default: |
| return -EINVAL; |
| } |
| |
| /* have to work out size/pitch and return them */ |
| args->pitch = ALIGN(args->width * cpp, 64); |
| |
| /* align stride to page size so that we can remap */ |
| if (args->pitch > intel_plane_fb_max_stride(to_i915(dev), format, |
| DRM_FORMAT_MOD_LINEAR)) |
| args->pitch = ALIGN(args->pitch, 4096); |
| |
| if (args->pitch < args->width) |
| return -EINVAL; |
| |
| args->size = mul_u32_u32(args->pitch, args->height); |
| |
| mem_type = INTEL_MEMORY_SYSTEM; |
| if (HAS_LMEM(to_i915(dev))) |
| mem_type = INTEL_MEMORY_LOCAL; |
| |
| mr = intel_memory_region_by_type(to_i915(dev), mem_type); |
| |
| obj = __i915_gem_object_create_user(to_i915(dev), args->size, &mr, 1); |
| if (IS_ERR(obj)) |
| return PTR_ERR(obj); |
| |
| return i915_gem_publish(obj, file, &args->size, &args->handle); |
| } |
| |
| /** |
| * i915_gem_create_ioctl - Creates a new mm object and returns a handle to it. |
| * @dev: drm device pointer |
| * @data: ioctl data blob |
| * @file: drm file pointer |
| */ |
| int |
| i915_gem_create_ioctl(struct drm_device *dev, void *data, |
| struct drm_file *file) |
| { |
| struct drm_i915_private *i915 = to_i915(dev); |
| struct drm_i915_gem_create *args = data; |
| struct drm_i915_gem_object *obj; |
| struct intel_memory_region *mr; |
| |
| mr = intel_memory_region_by_type(i915, INTEL_MEMORY_SYSTEM); |
| |
| obj = __i915_gem_object_create_user(i915, args->size, &mr, 1); |
| if (IS_ERR(obj)) |
| return PTR_ERR(obj); |
| |
| return i915_gem_publish(obj, file, &args->size, &args->handle); |
| } |
| |
| struct create_ext { |
| struct drm_i915_private *i915; |
| struct intel_memory_region *placements[INTEL_REGION_UNKNOWN]; |
| unsigned int n_placements; |
| unsigned int placement_mask; |
| unsigned long flags; |
| unsigned int pat_index; |
| }; |
| |
| static void repr_placements(char *buf, size_t size, |
| struct intel_memory_region **placements, |
| int n_placements) |
| { |
| int i; |
| |
| buf[0] = '\0'; |
| |
| for (i = 0; i < n_placements; i++) { |
| struct intel_memory_region *mr = placements[i]; |
| int r; |
| |
| r = snprintf(buf, size, "\n %s -> { class: %d, inst: %d }", |
| mr->name, mr->type, mr->instance); |
| if (r >= size) |
| return; |
| |
| buf += r; |
| size -= r; |
| } |
| } |
| |
| static int set_placements(struct drm_i915_gem_create_ext_memory_regions *args, |
| struct create_ext *ext_data) |
| { |
| struct drm_i915_private *i915 = ext_data->i915; |
| struct drm_i915_gem_memory_class_instance __user *uregions = |
| u64_to_user_ptr(args->regions); |
| struct intel_memory_region *placements[INTEL_REGION_UNKNOWN]; |
| u32 mask; |
| int i, ret = 0; |
| |
| if (args->pad) { |
| drm_dbg(&i915->drm, "pad should be zero\n"); |
| ret = -EINVAL; |
| } |
| |
| if (!args->num_regions) { |
| drm_dbg(&i915->drm, "num_regions is zero\n"); |
| ret = -EINVAL; |
| } |
| |
| BUILD_BUG_ON(ARRAY_SIZE(i915->mm.regions) != ARRAY_SIZE(placements)); |
| BUILD_BUG_ON(ARRAY_SIZE(ext_data->placements) != ARRAY_SIZE(placements)); |
| if (args->num_regions > ARRAY_SIZE(i915->mm.regions)) { |
| drm_dbg(&i915->drm, "num_regions is too large\n"); |
| ret = -EINVAL; |
| } |
| |
| if (ret) |
| return ret; |
| |
| mask = 0; |
| for (i = 0; i < args->num_regions; i++) { |
| struct drm_i915_gem_memory_class_instance region; |
| struct intel_memory_region *mr; |
| |
| if (copy_from_user(®ion, uregions, sizeof(region))) |
| return -EFAULT; |
| |
| mr = intel_memory_region_lookup(i915, |
| region.memory_class, |
| region.memory_instance); |
| if (!mr || mr->private) { |
| drm_dbg(&i915->drm, "Device is missing region { class: %d, inst: %d } at index = %d\n", |
| region.memory_class, region.memory_instance, i); |
| ret = -EINVAL; |
| goto out_dump; |
| } |
| |
| if (mask & BIT(mr->id)) { |
| drm_dbg(&i915->drm, "Found duplicate placement %s -> { class: %d, inst: %d } at index = %d\n", |
| mr->name, region.memory_class, |
| region.memory_instance, i); |
| ret = -EINVAL; |
| goto out_dump; |
| } |
| |
| placements[i] = mr; |
| mask |= BIT(mr->id); |
| |
| ++uregions; |
| } |
| |
| if (ext_data->n_placements) { |
| ret = -EINVAL; |
| goto out_dump; |
| } |
| |
| ext_data->n_placements = args->num_regions; |
| for (i = 0; i < args->num_regions; i++) |
| ext_data->placements[i] = placements[i]; |
| |
| ext_data->placement_mask = mask; |
| return 0; |
| |
| out_dump: |
| if (1) { |
| char buf[256]; |
| |
| if (ext_data->n_placements) { |
| repr_placements(buf, |
| sizeof(buf), |
| ext_data->placements, |
| ext_data->n_placements); |
| drm_dbg(&i915->drm, |
| "Placements were already set in previous EXT. Existing placements: %s\n", |
| buf); |
| } |
| |
| repr_placements(buf, sizeof(buf), placements, i); |
| drm_dbg(&i915->drm, "New placements(so far validated): %s\n", buf); |
| } |
| |
| return ret; |
| } |
| |
| static int ext_set_placements(struct i915_user_extension __user *base, |
| void *data) |
| { |
| struct drm_i915_gem_create_ext_memory_regions ext; |
| |
| if (copy_from_user(&ext, base, sizeof(ext))) |
| return -EFAULT; |
| |
| return set_placements(&ext, data); |
| } |
| |
| static int ext_set_protected(struct i915_user_extension __user *base, void *data) |
| { |
| struct drm_i915_gem_create_ext_protected_content ext; |
| struct create_ext *ext_data = data; |
| |
| if (copy_from_user(&ext, base, sizeof(ext))) |
| return -EFAULT; |
| |
| if (ext.flags) |
| return -EINVAL; |
| |
| if (!intel_pxp_is_enabled(ext_data->i915->pxp)) |
| return -ENODEV; |
| |
| ext_data->flags |= I915_BO_PROTECTED; |
| |
| return 0; |
| } |
| |
| static int ext_set_pat(struct i915_user_extension __user *base, void *data) |
| { |
| struct create_ext *ext_data = data; |
| struct drm_i915_private *i915 = ext_data->i915; |
| struct drm_i915_gem_create_ext_set_pat ext; |
| unsigned int max_pat_index; |
| |
| BUILD_BUG_ON(sizeof(struct drm_i915_gem_create_ext_set_pat) != |
| offsetofend(struct drm_i915_gem_create_ext_set_pat, rsvd)); |
| |
| /* Limiting the extension only to Xe_LPG and beyond */ |
| if (GRAPHICS_VER_FULL(i915) < IP_VER(12, 70)) |
| return -ENODEV; |
| |
| if (copy_from_user(&ext, base, sizeof(ext))) |
| return -EFAULT; |
| |
| max_pat_index = INTEL_INFO(i915)->max_pat_index; |
| |
| if (ext.pat_index > max_pat_index) { |
| drm_dbg(&i915->drm, "PAT index is invalid: %u\n", |
| ext.pat_index); |
| return -EINVAL; |
| } |
| |
| ext_data->pat_index = ext.pat_index; |
| |
| return 0; |
| } |
| |
| static const i915_user_extension_fn create_extensions[] = { |
| [I915_GEM_CREATE_EXT_MEMORY_REGIONS] = ext_set_placements, |
| [I915_GEM_CREATE_EXT_PROTECTED_CONTENT] = ext_set_protected, |
| [I915_GEM_CREATE_EXT_SET_PAT] = ext_set_pat, |
| }; |
| |
| #define PAT_INDEX_NOT_SET 0xffff |
| /** |
| * i915_gem_create_ext_ioctl - Creates a new mm object and returns a handle to it. |
| * @dev: drm device pointer |
| * @data: ioctl data blob |
| * @file: drm file pointer |
| */ |
| int |
| i915_gem_create_ext_ioctl(struct drm_device *dev, void *data, |
| struct drm_file *file) |
| { |
| struct drm_i915_private *i915 = to_i915(dev); |
| struct drm_i915_gem_create_ext *args = data; |
| struct create_ext ext_data = { .i915 = i915 }; |
| struct drm_i915_gem_object *obj; |
| int ret; |
| |
| if (args->flags & ~I915_GEM_CREATE_EXT_FLAG_NEEDS_CPU_ACCESS) |
| return -EINVAL; |
| |
| ext_data.pat_index = PAT_INDEX_NOT_SET; |
| ret = i915_user_extensions(u64_to_user_ptr(args->extensions), |
| create_extensions, |
| ARRAY_SIZE(create_extensions), |
| &ext_data); |
| if (ret) |
| return ret; |
| |
| if (!ext_data.n_placements) { |
| ext_data.placements[0] = |
| intel_memory_region_by_type(i915, INTEL_MEMORY_SYSTEM); |
| ext_data.n_placements = 1; |
| } |
| |
| if (args->flags & I915_GEM_CREATE_EXT_FLAG_NEEDS_CPU_ACCESS) { |
| if (ext_data.n_placements == 1) |
| return -EINVAL; |
| |
| /* |
| * We always need to be able to spill to system memory, if we |
| * can't place in the mappable part of LMEM. |
| */ |
| if (!(ext_data.placement_mask & BIT(INTEL_REGION_SMEM))) |
| return -EINVAL; |
| } else { |
| if (ext_data.n_placements > 1 || |
| ext_data.placements[0]->type != INTEL_MEMORY_SYSTEM) |
| ext_data.flags |= I915_BO_ALLOC_GPU_ONLY; |
| } |
| |
| obj = __i915_gem_object_create_user_ext(i915, args->size, |
| ext_data.placements, |
| ext_data.n_placements, |
| ext_data.flags); |
| if (IS_ERR(obj)) |
| return PTR_ERR(obj); |
| |
| if (ext_data.pat_index != PAT_INDEX_NOT_SET) { |
| i915_gem_object_set_pat_index(obj, ext_data.pat_index); |
| /* Mark pat_index is set by UMD */ |
| obj->pat_set_by_user = true; |
| } |
| |
| return i915_gem_publish(obj, file, &args->size, &args->handle); |
| } |