| /* SPDX-License-Identifier: MIT */ |
| /* Copyright (C) 2023 Collabora ltd. */ |
| #ifndef _PANTHOR_DRM_H_ |
| #define _PANTHOR_DRM_H_ |
| |
| #include "drm.h" |
| |
| #if defined(__cplusplus) |
| extern "C" { |
| #endif |
| |
| /** |
| * DOC: Introduction |
| * |
| * This documentation describes the Panthor IOCTLs. |
| * |
| * Just a few generic rules about the data passed to the Panthor IOCTLs: |
| * |
| * - Structures must be aligned on 64-bit/8-byte. If the object is not |
| * naturally aligned, a padding field must be added. |
| * - Fields must be explicitly aligned to their natural type alignment with |
| * pad[0..N] fields. |
| * - All padding fields will be checked by the driver to make sure they are |
| * zeroed. |
| * - Flags can be added, but not removed/replaced. |
| * - New fields can be added to the main structures (the structures |
| * directly passed to the ioctl). Those fields can be added at the end of |
| * the structure, or replace existing padding fields. Any new field being |
| * added must preserve the behavior that existed before those fields were |
| * added when a value of zero is passed. |
| * - New fields can be added to indirect objects (objects pointed by the |
| * main structure), iff those objects are passed a size to reflect the |
| * size known by the userspace driver (see drm_panthor_obj_array::stride |
| * or drm_panthor_dev_query::size). |
| * - If the kernel driver is too old to know some fields, those will be |
| * ignored if zero, and otherwise rejected (and so will be zero on output). |
| * - If userspace is too old to know some fields, those will be zeroed |
| * (input) before the structure is parsed by the kernel driver. |
| * - Each new flag/field addition must come with a driver version update so |
| * the userspace driver doesn't have to trial and error to know which |
| * flags are supported. |
| * - Structures should not contain unions, as this would defeat the |
| * extensibility of such structures. |
| * - IOCTLs can't be removed or replaced. New IOCTL IDs should be placed |
| * at the end of the drm_panthor_ioctl_id enum. |
| */ |
| |
| /** |
| * DOC: MMIO regions exposed to userspace. |
| * |
| * .. c:macro:: DRM_PANTHOR_USER_MMIO_OFFSET |
| * |
| * File offset for all MMIO regions being exposed to userspace. Don't use |
| * this value directly, use DRM_PANTHOR_USER_<name>_OFFSET values instead. |
| * pgoffset passed to mmap2() is an unsigned long, which forces us to use a |
| * different offset on 32-bit and 64-bit systems. |
| * |
| * .. c:macro:: DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET |
| * |
| * File offset for the LATEST_FLUSH_ID register. The Userspace driver controls |
| * GPU cache flushing through CS instructions, but the flush reduction |
| * mechanism requires a flush_id. This flush_id could be queried with an |
| * ioctl, but Arm provides a well-isolated register page containing only this |
| * read-only register, so let's expose this page through a static mmap offset |
| * and allow direct mapping of this MMIO region so we can avoid the |
| * user <-> kernel round-trip. |
| */ |
| #define DRM_PANTHOR_USER_MMIO_OFFSET_32BIT (1ull << 43) |
| #define DRM_PANTHOR_USER_MMIO_OFFSET_64BIT (1ull << 56) |
| #define DRM_PANTHOR_USER_MMIO_OFFSET (sizeof(unsigned long) < 8 ? \ |
| DRM_PANTHOR_USER_MMIO_OFFSET_32BIT : \ |
| DRM_PANTHOR_USER_MMIO_OFFSET_64BIT) |
| #define DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET (DRM_PANTHOR_USER_MMIO_OFFSET | 0) |
| |
| /** |
| * DOC: IOCTL IDs |
| * |
| * enum drm_panthor_ioctl_id - IOCTL IDs |
| * |
| * Place new ioctls at the end, don't re-order, don't replace or remove entries. |
| * |
| * These IDs are not meant to be used directly. Use the DRM_IOCTL_PANTHOR_xxx |
| * definitions instead. |
| */ |
| enum drm_panthor_ioctl_id { |
| /** @DRM_PANTHOR_DEV_QUERY: Query device information. */ |
| DRM_PANTHOR_DEV_QUERY = 0, |
| |
| /** @DRM_PANTHOR_VM_CREATE: Create a VM. */ |
| DRM_PANTHOR_VM_CREATE, |
| |
| /** @DRM_PANTHOR_VM_DESTROY: Destroy a VM. */ |
| DRM_PANTHOR_VM_DESTROY, |
| |
| /** @DRM_PANTHOR_VM_BIND: Bind/unbind memory to a VM. */ |
| DRM_PANTHOR_VM_BIND, |
| |
| /** @DRM_PANTHOR_VM_GET_STATE: Get VM state. */ |
| DRM_PANTHOR_VM_GET_STATE, |
| |
| /** @DRM_PANTHOR_BO_CREATE: Create a buffer object. */ |
| DRM_PANTHOR_BO_CREATE, |
| |
| /** |
| * @DRM_PANTHOR_BO_MMAP_OFFSET: Get the file offset to pass to |
| * mmap to map a GEM object. |
| */ |
| DRM_PANTHOR_BO_MMAP_OFFSET, |
| |
| /** @DRM_PANTHOR_GROUP_CREATE: Create a scheduling group. */ |
| DRM_PANTHOR_GROUP_CREATE, |
| |
| /** @DRM_PANTHOR_GROUP_DESTROY: Destroy a scheduling group. */ |
| DRM_PANTHOR_GROUP_DESTROY, |
| |
| /** |
| * @DRM_PANTHOR_GROUP_SUBMIT: Submit jobs to queues belonging |
| * to a specific scheduling group. |
| */ |
| DRM_PANTHOR_GROUP_SUBMIT, |
| |
| /** @DRM_PANTHOR_GROUP_GET_STATE: Get the state of a scheduling group. */ |
| DRM_PANTHOR_GROUP_GET_STATE, |
| |
| /** @DRM_PANTHOR_TILER_HEAP_CREATE: Create a tiler heap. */ |
| DRM_PANTHOR_TILER_HEAP_CREATE, |
| |
| /** @DRM_PANTHOR_TILER_HEAP_DESTROY: Destroy a tiler heap. */ |
| DRM_PANTHOR_TILER_HEAP_DESTROY, |
| }; |
| |
| /** |
| * DRM_IOCTL_PANTHOR() - Build a Panthor IOCTL number |
| * @__access: Access type. Must be R, W or RW. |
| * @__id: One of the DRM_PANTHOR_xxx id. |
| * @__type: Suffix of the type being passed to the IOCTL. |
| * |
| * Don't use this macro directly, use the DRM_IOCTL_PANTHOR_xxx |
| * values instead. |
| * |
| * Return: An IOCTL number to be passed to ioctl() from userspace. |
| */ |
| #define DRM_IOCTL_PANTHOR(__access, __id, __type) \ |
| DRM_IO ## __access(DRM_COMMAND_BASE + DRM_PANTHOR_ ## __id, \ |
| struct drm_panthor_ ## __type) |
| |
| #define DRM_IOCTL_PANTHOR_DEV_QUERY \ |
| DRM_IOCTL_PANTHOR(WR, DEV_QUERY, dev_query) |
| #define DRM_IOCTL_PANTHOR_VM_CREATE \ |
| DRM_IOCTL_PANTHOR(WR, VM_CREATE, vm_create) |
| #define DRM_IOCTL_PANTHOR_VM_DESTROY \ |
| DRM_IOCTL_PANTHOR(WR, VM_DESTROY, vm_destroy) |
| #define DRM_IOCTL_PANTHOR_VM_BIND \ |
| DRM_IOCTL_PANTHOR(WR, VM_BIND, vm_bind) |
| #define DRM_IOCTL_PANTHOR_VM_GET_STATE \ |
| DRM_IOCTL_PANTHOR(WR, VM_GET_STATE, vm_get_state) |
| #define DRM_IOCTL_PANTHOR_BO_CREATE \ |
| DRM_IOCTL_PANTHOR(WR, BO_CREATE, bo_create) |
| #define DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET \ |
| DRM_IOCTL_PANTHOR(WR, BO_MMAP_OFFSET, bo_mmap_offset) |
| #define DRM_IOCTL_PANTHOR_GROUP_CREATE \ |
| DRM_IOCTL_PANTHOR(WR, GROUP_CREATE, group_create) |
| #define DRM_IOCTL_PANTHOR_GROUP_DESTROY \ |
| DRM_IOCTL_PANTHOR(WR, GROUP_DESTROY, group_destroy) |
| #define DRM_IOCTL_PANTHOR_GROUP_SUBMIT \ |
| DRM_IOCTL_PANTHOR(WR, GROUP_SUBMIT, group_submit) |
| #define DRM_IOCTL_PANTHOR_GROUP_GET_STATE \ |
| DRM_IOCTL_PANTHOR(WR, GROUP_GET_STATE, group_get_state) |
| #define DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE \ |
| DRM_IOCTL_PANTHOR(WR, TILER_HEAP_CREATE, tiler_heap_create) |
| #define DRM_IOCTL_PANTHOR_TILER_HEAP_DESTROY \ |
| DRM_IOCTL_PANTHOR(WR, TILER_HEAP_DESTROY, tiler_heap_destroy) |
| |
| /** |
| * DOC: IOCTL arguments |
| */ |
| |
| /** |
| * struct drm_panthor_obj_array - Object array. |
| * |
| * This object is used to pass an array of objects whose size is subject to changes in |
| * future versions of the driver. In order to support this mutability, we pass a stride |
| * describing the size of the object as known by userspace. |
| * |
| * You shouldn't fill drm_panthor_obj_array fields directly. You should instead use |
| * the DRM_PANTHOR_OBJ_ARRAY() macro that takes care of initializing the stride to |
| * the object size. |
| */ |
| struct drm_panthor_obj_array { |
| /** @stride: Stride of object struct. Used for versioning. */ |
| __u32 stride; |
| |
| /** @count: Number of objects in the array. */ |
| __u32 count; |
| |
| /** @array: User pointer to an array of objects. */ |
| __u64 array; |
| }; |
| |
| /** |
| * DRM_PANTHOR_OBJ_ARRAY() - Initialize a drm_panthor_obj_array field. |
| * @cnt: Number of elements in the array. |
| * @ptr: Pointer to the array to pass to the kernel. |
| * |
| * Macro initializing a drm_panthor_obj_array based on the object size as known |
| * by userspace. |
| */ |
| #define DRM_PANTHOR_OBJ_ARRAY(cnt, ptr) \ |
| { .stride = sizeof((ptr)[0]), .count = (cnt), .array = (__u64)(uintptr_t)(ptr) } |
| |
| /** |
| * enum drm_panthor_sync_op_flags - Synchronization operation flags. |
| */ |
| enum drm_panthor_sync_op_flags { |
| /** @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK: Synchronization handle type mask. */ |
| DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK = 0xff, |
| |
| /** @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ: Synchronization object type. */ |
| DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ = 0, |
| |
| /** |
| * @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ: Timeline synchronization |
| * object type. |
| */ |
| DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ = 1, |
| |
| /** @DRM_PANTHOR_SYNC_OP_WAIT: Wait operation. */ |
| DRM_PANTHOR_SYNC_OP_WAIT = 0 << 31, |
| |
| /** @DRM_PANTHOR_SYNC_OP_SIGNAL: Signal operation. */ |
| DRM_PANTHOR_SYNC_OP_SIGNAL = (int)(1u << 31), |
| }; |
| |
| /** |
| * struct drm_panthor_sync_op - Synchronization operation. |
| */ |
| struct drm_panthor_sync_op { |
| /** @flags: Synchronization operation flags. Combination of DRM_PANTHOR_SYNC_OP values. */ |
| __u32 flags; |
| |
| /** @handle: Sync handle. */ |
| __u32 handle; |
| |
| /** |
| * @timeline_value: MBZ if |
| * (flags & DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK) != |
| * DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ. |
| */ |
| __u64 timeline_value; |
| }; |
| |
| /** |
| * enum drm_panthor_dev_query_type - Query type |
| * |
| * Place new types at the end, don't re-order, don't remove or replace. |
| */ |
| enum drm_panthor_dev_query_type { |
| /** @DRM_PANTHOR_DEV_QUERY_GPU_INFO: Query GPU information. */ |
| DRM_PANTHOR_DEV_QUERY_GPU_INFO = 0, |
| |
| /** @DRM_PANTHOR_DEV_QUERY_CSIF_INFO: Query command-stream interface information. */ |
| DRM_PANTHOR_DEV_QUERY_CSIF_INFO, |
| }; |
| |
| /** |
| * struct drm_panthor_gpu_info - GPU information |
| * |
| * Structure grouping all queryable information relating to the GPU. |
| */ |
| struct drm_panthor_gpu_info { |
| /** @gpu_id : GPU ID. */ |
| __u32 gpu_id; |
| #define DRM_PANTHOR_ARCH_MAJOR(x) ((x) >> 28) |
| #define DRM_PANTHOR_ARCH_MINOR(x) (((x) >> 24) & 0xf) |
| #define DRM_PANTHOR_ARCH_REV(x) (((x) >> 20) & 0xf) |
| #define DRM_PANTHOR_PRODUCT_MAJOR(x) (((x) >> 16) & 0xf) |
| #define DRM_PANTHOR_VERSION_MAJOR(x) (((x) >> 12) & 0xf) |
| #define DRM_PANTHOR_VERSION_MINOR(x) (((x) >> 4) & 0xff) |
| #define DRM_PANTHOR_VERSION_STATUS(x) ((x) & 0xf) |
| |
| /** @gpu_rev: GPU revision. */ |
| __u32 gpu_rev; |
| |
| /** @csf_id: Command stream frontend ID. */ |
| __u32 csf_id; |
| #define DRM_PANTHOR_CSHW_MAJOR(x) (((x) >> 26) & 0x3f) |
| #define DRM_PANTHOR_CSHW_MINOR(x) (((x) >> 20) & 0x3f) |
| #define DRM_PANTHOR_CSHW_REV(x) (((x) >> 16) & 0xf) |
| #define DRM_PANTHOR_MCU_MAJOR(x) (((x) >> 10) & 0x3f) |
| #define DRM_PANTHOR_MCU_MINOR(x) (((x) >> 4) & 0x3f) |
| #define DRM_PANTHOR_MCU_REV(x) ((x) & 0xf) |
| |
| /** @l2_features: L2-cache features. */ |
| __u32 l2_features; |
| |
| /** @tiler_features: Tiler features. */ |
| __u32 tiler_features; |
| |
| /** @mem_features: Memory features. */ |
| __u32 mem_features; |
| |
| /** @mmu_features: MMU features. */ |
| __u32 mmu_features; |
| #define DRM_PANTHOR_MMU_VA_BITS(x) ((x) & 0xff) |
| |
| /** @thread_features: Thread features. */ |
| __u32 thread_features; |
| |
| /** @max_threads: Maximum number of threads. */ |
| __u32 max_threads; |
| |
| /** @thread_max_workgroup_size: Maximum workgroup size. */ |
| __u32 thread_max_workgroup_size; |
| |
| /** |
| * @thread_max_barrier_size: Maximum number of threads that can wait |
| * simultaneously on a barrier. |
| */ |
| __u32 thread_max_barrier_size; |
| |
| /** @coherency_features: Coherency features. */ |
| __u32 coherency_features; |
| |
| /** @texture_features: Texture features. */ |
| __u32 texture_features[4]; |
| |
| /** @as_present: Bitmask encoding the number of address-space exposed by the MMU. */ |
| __u32 as_present; |
| |
| /** @shader_present: Bitmask encoding the shader cores exposed by the GPU. */ |
| __u64 shader_present; |
| |
| /** @l2_present: Bitmask encoding the L2 caches exposed by the GPU. */ |
| __u64 l2_present; |
| |
| /** @tiler_present: Bitmask encoding the tiler units exposed by the GPU. */ |
| __u64 tiler_present; |
| |
| /** @core_features: Used to discriminate core variants when they exist. */ |
| __u32 core_features; |
| |
| /** @pad: MBZ. */ |
| __u32 pad; |
| }; |
| |
| /** |
| * struct drm_panthor_csif_info - Command stream interface information |
| * |
| * Structure grouping all queryable information relating to the command stream interface. |
| */ |
| struct drm_panthor_csif_info { |
| /** @csg_slot_count: Number of command stream group slots exposed by the firmware. */ |
| __u32 csg_slot_count; |
| |
| /** @cs_slot_count: Number of command stream slots per group. */ |
| __u32 cs_slot_count; |
| |
| /** @cs_reg_count: Number of command stream registers. */ |
| __u32 cs_reg_count; |
| |
| /** @scoreboard_slot_count: Number of scoreboard slots. */ |
| __u32 scoreboard_slot_count; |
| |
| /** |
| * @unpreserved_cs_reg_count: Number of command stream registers reserved by |
| * the kernel driver to call a userspace command stream. |
| * |
| * All registers can be used by a userspace command stream, but the |
| * [cs_slot_count - unpreserved_cs_reg_count .. cs_slot_count] registers are |
| * used by the kernel when DRM_PANTHOR_IOCTL_GROUP_SUBMIT is called. |
| */ |
| __u32 unpreserved_cs_reg_count; |
| |
| /** |
| * @pad: Padding field, set to zero. |
| */ |
| __u32 pad; |
| }; |
| |
| /** |
| * struct drm_panthor_dev_query - Arguments passed to DRM_PANTHOR_IOCTL_DEV_QUERY |
| */ |
| struct drm_panthor_dev_query { |
| /** @type: the query type (see drm_panthor_dev_query_type). */ |
| __u32 type; |
| |
| /** |
| * @size: size of the type being queried. |
| * |
| * If pointer is NULL, size is updated by the driver to provide the |
| * output structure size. If pointer is not NULL, the driver will |
| * only copy min(size, actual_structure_size) bytes to the pointer, |
| * and update the size accordingly. This allows us to extend query |
| * types without breaking userspace. |
| */ |
| __u32 size; |
| |
| /** |
| * @pointer: user pointer to a query type struct. |
| * |
| * Pointer can be NULL, in which case, nothing is copied, but the |
| * actual structure size is returned. If not NULL, it must point to |
| * a location that's large enough to hold size bytes. |
| */ |
| __u64 pointer; |
| }; |
| |
| /** |
| * struct drm_panthor_vm_create - Arguments passed to DRM_PANTHOR_IOCTL_VM_CREATE |
| */ |
| struct drm_panthor_vm_create { |
| /** @flags: VM flags, MBZ. */ |
| __u32 flags; |
| |
| /** @id: Returned VM ID. */ |
| __u32 id; |
| |
| /** |
| * @user_va_range: Size of the VA space reserved for user objects. |
| * |
| * The kernel will pick the remaining space to map kernel-only objects to the |
| * VM (heap chunks, heap context, ring buffers, kernel synchronization objects, |
| * ...). If the space left for kernel objects is too small, kernel object |
| * allocation will fail further down the road. One can use |
| * drm_panthor_gpu_info::mmu_features to extract the total virtual address |
| * range, and chose a user_va_range that leaves some space to the kernel. |
| * |
| * If user_va_range is zero, the kernel will pick a sensible value based on |
| * TASK_SIZE and the virtual range supported by the GPU MMU (the kernel/user |
| * split should leave enough VA space for userspace processes to support SVM, |
| * while still allowing the kernel to map some amount of kernel objects in |
| * the kernel VA range). The value chosen by the driver will be returned in |
| * @user_va_range. |
| * |
| * User VA space always starts at 0x0, kernel VA space is always placed after |
| * the user VA range. |
| */ |
| __u64 user_va_range; |
| }; |
| |
| /** |
| * struct drm_panthor_vm_destroy - Arguments passed to DRM_PANTHOR_IOCTL_VM_DESTROY |
| */ |
| struct drm_panthor_vm_destroy { |
| /** @id: ID of the VM to destroy. */ |
| __u32 id; |
| |
| /** @pad: MBZ. */ |
| __u32 pad; |
| }; |
| |
| /** |
| * enum drm_panthor_vm_bind_op_flags - VM bind operation flags |
| */ |
| enum drm_panthor_vm_bind_op_flags { |
| /** |
| * @DRM_PANTHOR_VM_BIND_OP_MAP_READONLY: Map the memory read-only. |
| * |
| * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP. |
| */ |
| DRM_PANTHOR_VM_BIND_OP_MAP_READONLY = 1 << 0, |
| |
| /** |
| * @DRM_PANTHOR_VM_BIND_OP_MAP_NOEXEC: Map the memory not-executable. |
| * |
| * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP. |
| */ |
| DRM_PANTHOR_VM_BIND_OP_MAP_NOEXEC = 1 << 1, |
| |
| /** |
| * @DRM_PANTHOR_VM_BIND_OP_MAP_UNCACHED: Map the memory uncached. |
| * |
| * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP. |
| */ |
| DRM_PANTHOR_VM_BIND_OP_MAP_UNCACHED = 1 << 2, |
| |
| /** |
| * @DRM_PANTHOR_VM_BIND_OP_TYPE_MASK: Mask used to determine the type of operation. |
| */ |
| DRM_PANTHOR_VM_BIND_OP_TYPE_MASK = (int)(0xfu << 28), |
| |
| /** @DRM_PANTHOR_VM_BIND_OP_TYPE_MAP: Map operation. */ |
| DRM_PANTHOR_VM_BIND_OP_TYPE_MAP = 0 << 28, |
| |
| /** @DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP: Unmap operation. */ |
| DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP = 1 << 28, |
| |
| /** |
| * @DRM_PANTHOR_VM_BIND_OP_TYPE_SYNC_ONLY: No VM operation. |
| * |
| * Just serves as a synchronization point on a VM queue. |
| * |
| * Only valid if %DRM_PANTHOR_VM_BIND_ASYNC is set in drm_panthor_vm_bind::flags, |
| * and drm_panthor_vm_bind_op::syncs contains at least one element. |
| */ |
| DRM_PANTHOR_VM_BIND_OP_TYPE_SYNC_ONLY = 2 << 28, |
| }; |
| |
| /** |
| * struct drm_panthor_vm_bind_op - VM bind operation |
| */ |
| struct drm_panthor_vm_bind_op { |
| /** @flags: Combination of drm_panthor_vm_bind_op_flags flags. */ |
| __u32 flags; |
| |
| /** |
| * @bo_handle: Handle of the buffer object to map. |
| * MBZ for unmap or sync-only operations. |
| */ |
| __u32 bo_handle; |
| |
| /** |
| * @bo_offset: Buffer object offset. |
| * MBZ for unmap or sync-only operations. |
| */ |
| __u64 bo_offset; |
| |
| /** |
| * @va: Virtual address to map/unmap. |
| * MBZ for sync-only operations. |
| */ |
| __u64 va; |
| |
| /** |
| * @size: Size to map/unmap. |
| * MBZ for sync-only operations. |
| */ |
| __u64 size; |
| |
| /** |
| * @syncs: Array of struct drm_panthor_sync_op synchronization |
| * operations. |
| * |
| * This array must be empty if %DRM_PANTHOR_VM_BIND_ASYNC is not set on |
| * the drm_panthor_vm_bind object containing this VM bind operation. |
| * |
| * This array shall not be empty for sync-only operations. |
| */ |
| struct drm_panthor_obj_array syncs; |
| |
| }; |
| |
| /** |
| * enum drm_panthor_vm_bind_flags - VM bind flags |
| */ |
| enum drm_panthor_vm_bind_flags { |
| /** |
| * @DRM_PANTHOR_VM_BIND_ASYNC: VM bind operations are queued to the VM |
| * queue instead of being executed synchronously. |
| */ |
| DRM_PANTHOR_VM_BIND_ASYNC = 1 << 0, |
| }; |
| |
| /** |
| * struct drm_panthor_vm_bind - Arguments passed to DRM_IOCTL_PANTHOR_VM_BIND |
| */ |
| struct drm_panthor_vm_bind { |
| /** @vm_id: VM targeted by the bind request. */ |
| __u32 vm_id; |
| |
| /** @flags: Combination of drm_panthor_vm_bind_flags flags. */ |
| __u32 flags; |
| |
| /** @ops: Array of struct drm_panthor_vm_bind_op bind operations. */ |
| struct drm_panthor_obj_array ops; |
| }; |
| |
| /** |
| * enum drm_panthor_vm_state - VM states. |
| */ |
| enum drm_panthor_vm_state { |
| /** |
| * @DRM_PANTHOR_VM_STATE_USABLE: VM is usable. |
| * |
| * New VM operations will be accepted on this VM. |
| */ |
| DRM_PANTHOR_VM_STATE_USABLE, |
| |
| /** |
| * @DRM_PANTHOR_VM_STATE_UNUSABLE: VM is unusable. |
| * |
| * Something put the VM in an unusable state (like an asynchronous |
| * VM_BIND request failing for any reason). |
| * |
| * Once the VM is in this state, all new MAP operations will be |
| * rejected, and any GPU job targeting this VM will fail. |
| * UNMAP operations are still accepted. |
| * |
| * The only way to recover from an unusable VM is to create a new |
| * VM, and destroy the old one. |
| */ |
| DRM_PANTHOR_VM_STATE_UNUSABLE, |
| }; |
| |
| /** |
| * struct drm_panthor_vm_get_state - Get VM state. |
| */ |
| struct drm_panthor_vm_get_state { |
| /** @vm_id: VM targeted by the get_state request. */ |
| __u32 vm_id; |
| |
| /** |
| * @state: state returned by the driver. |
| * |
| * Must be one of the enum drm_panthor_vm_state values. |
| */ |
| __u32 state; |
| }; |
| |
| /** |
| * enum drm_panthor_bo_flags - Buffer object flags, passed at creation time. |
| */ |
| enum drm_panthor_bo_flags { |
| /** @DRM_PANTHOR_BO_NO_MMAP: The buffer object will never be CPU-mapped in userspace. */ |
| DRM_PANTHOR_BO_NO_MMAP = (1 << 0), |
| }; |
| |
| /** |
| * struct drm_panthor_bo_create - Arguments passed to DRM_IOCTL_PANTHOR_BO_CREATE. |
| */ |
| struct drm_panthor_bo_create { |
| /** |
| * @size: Requested size for the object |
| * |
| * The (page-aligned) allocated size for the object will be returned. |
| */ |
| __u64 size; |
| |
| /** |
| * @flags: Flags. Must be a combination of drm_panthor_bo_flags flags. |
| */ |
| __u32 flags; |
| |
| /** |
| * @exclusive_vm_id: Exclusive VM this buffer object will be mapped to. |
| * |
| * If not zero, the field must refer to a valid VM ID, and implies that: |
| * - the buffer object will only ever be bound to that VM |
| * - cannot be exported as a PRIME fd |
| */ |
| __u32 exclusive_vm_id; |
| |
| /** |
| * @handle: Returned handle for the object. |
| * |
| * Object handles are nonzero. |
| */ |
| __u32 handle; |
| |
| /** @pad: MBZ. */ |
| __u32 pad; |
| }; |
| |
| /** |
| * struct drm_panthor_bo_mmap_offset - Arguments passed to DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET. |
| */ |
| struct drm_panthor_bo_mmap_offset { |
| /** @handle: Handle of the object we want an mmap offset for. */ |
| __u32 handle; |
| |
| /** @pad: MBZ. */ |
| __u32 pad; |
| |
| /** @offset: The fake offset to use for subsequent mmap calls. */ |
| __u64 offset; |
| }; |
| |
| /** |
| * struct drm_panthor_queue_create - Queue creation arguments. |
| */ |
| struct drm_panthor_queue_create { |
| /** |
| * @priority: Defines the priority of queues inside a group. Goes from 0 to 15, |
| * 15 being the highest priority. |
| */ |
| __u8 priority; |
| |
| /** @pad: Padding fields, MBZ. */ |
| __u8 pad[3]; |
| |
| /** @ringbuf_size: Size of the ring buffer to allocate to this queue. */ |
| __u32 ringbuf_size; |
| }; |
| |
| /** |
| * enum drm_panthor_group_priority - Scheduling group priority |
| */ |
| enum drm_panthor_group_priority { |
| /** @PANTHOR_GROUP_PRIORITY_LOW: Low priority group. */ |
| PANTHOR_GROUP_PRIORITY_LOW = 0, |
| |
| /** @PANTHOR_GROUP_PRIORITY_MEDIUM: Medium priority group. */ |
| PANTHOR_GROUP_PRIORITY_MEDIUM, |
| |
| /** |
| * @PANTHOR_GROUP_PRIORITY_HIGH: High priority group. |
| * |
| * Requires CAP_SYS_NICE or DRM_MASTER. |
| */ |
| PANTHOR_GROUP_PRIORITY_HIGH, |
| }; |
| |
| /** |
| * struct drm_panthor_group_create - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_CREATE |
| */ |
| struct drm_panthor_group_create { |
| /** @queues: Array of drm_panthor_queue_create elements. */ |
| struct drm_panthor_obj_array queues; |
| |
| /** |
| * @max_compute_cores: Maximum number of cores that can be used by compute |
| * jobs across CS queues bound to this group. |
| * |
| * Must be less or equal to the number of bits set in @compute_core_mask. |
| */ |
| __u8 max_compute_cores; |
| |
| /** |
| * @max_fragment_cores: Maximum number of cores that can be used by fragment |
| * jobs across CS queues bound to this group. |
| * |
| * Must be less or equal to the number of bits set in @fragment_core_mask. |
| */ |
| __u8 max_fragment_cores; |
| |
| /** |
| * @max_tiler_cores: Maximum number of tilers that can be used by tiler jobs |
| * across CS queues bound to this group. |
| * |
| * Must be less or equal to the number of bits set in @tiler_core_mask. |
| */ |
| __u8 max_tiler_cores; |
| |
| /** @priority: Group priority (see enum drm_panthor_group_priority). */ |
| __u8 priority; |
| |
| /** @pad: Padding field, MBZ. */ |
| __u32 pad; |
| |
| /** |
| * @compute_core_mask: Mask encoding cores that can be used for compute jobs. |
| * |
| * This field must have at least @max_compute_cores bits set. |
| * |
| * The bits set here should also be set in drm_panthor_gpu_info::shader_present. |
| */ |
| __u64 compute_core_mask; |
| |
| /** |
| * @fragment_core_mask: Mask encoding cores that can be used for fragment jobs. |
| * |
| * This field must have at least @max_fragment_cores bits set. |
| * |
| * The bits set here should also be set in drm_panthor_gpu_info::shader_present. |
| */ |
| __u64 fragment_core_mask; |
| |
| /** |
| * @tiler_core_mask: Mask encoding cores that can be used for tiler jobs. |
| * |
| * This field must have at least @max_tiler_cores bits set. |
| * |
| * The bits set here should also be set in drm_panthor_gpu_info::tiler_present. |
| */ |
| __u64 tiler_core_mask; |
| |
| /** |
| * @vm_id: VM ID to bind this group to. |
| * |
| * All submission to queues bound to this group will use this VM. |
| */ |
| __u32 vm_id; |
| |
| /** |
| * @group_handle: Returned group handle. Passed back when submitting jobs or |
| * destroying a group. |
| */ |
| __u32 group_handle; |
| }; |
| |
| /** |
| * struct drm_panthor_group_destroy - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_DESTROY |
| */ |
| struct drm_panthor_group_destroy { |
| /** @group_handle: Group to destroy */ |
| __u32 group_handle; |
| |
| /** @pad: Padding field, MBZ. */ |
| __u32 pad; |
| }; |
| |
| /** |
| * struct drm_panthor_queue_submit - Job submission arguments. |
| * |
| * This is describing the userspace command stream to call from the kernel |
| * command stream ring-buffer. Queue submission is always part of a group |
| * submission, taking one or more jobs to submit to the underlying queues. |
| */ |
| struct drm_panthor_queue_submit { |
| /** @queue_index: Index of the queue inside a group. */ |
| __u32 queue_index; |
| |
| /** |
| * @stream_size: Size of the command stream to execute. |
| * |
| * Must be 64-bit/8-byte aligned (the size of a CS instruction) |
| * |
| * Can be zero if stream_addr is zero too. |
| * |
| * When the stream size is zero, the queue submit serves as a |
| * synchronization point. |
| */ |
| __u32 stream_size; |
| |
| /** |
| * @stream_addr: GPU address of the command stream to execute. |
| * |
| * Must be aligned on 64-byte. |
| * |
| * Can be zero is stream_size is zero too. |
| */ |
| __u64 stream_addr; |
| |
| /** |
| * @latest_flush: FLUSH_ID read at the time the stream was built. |
| * |
| * This allows cache flush elimination for the automatic |
| * flush+invalidate(all) done at submission time, which is needed to |
| * ensure the GPU doesn't get garbage when reading the indirect command |
| * stream buffers. If you want the cache flush to happen |
| * unconditionally, pass a zero here. |
| * |
| * Ignored when stream_size is zero. |
| */ |
| __u32 latest_flush; |
| |
| /** @pad: MBZ. */ |
| __u32 pad; |
| |
| /** @syncs: Array of struct drm_panthor_sync_op sync operations. */ |
| struct drm_panthor_obj_array syncs; |
| }; |
| |
| /** |
| * struct drm_panthor_group_submit - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_SUBMIT |
| */ |
| struct drm_panthor_group_submit { |
| /** @group_handle: Handle of the group to queue jobs to. */ |
| __u32 group_handle; |
| |
| /** @pad: MBZ. */ |
| __u32 pad; |
| |
| /** @queue_submits: Array of drm_panthor_queue_submit objects. */ |
| struct drm_panthor_obj_array queue_submits; |
| }; |
| |
| /** |
| * enum drm_panthor_group_state_flags - Group state flags |
| */ |
| enum drm_panthor_group_state_flags { |
| /** |
| * @DRM_PANTHOR_GROUP_STATE_TIMEDOUT: Group had unfinished jobs. |
| * |
| * When a group ends up with this flag set, no jobs can be submitted to its queues. |
| */ |
| DRM_PANTHOR_GROUP_STATE_TIMEDOUT = 1 << 0, |
| |
| /** |
| * @DRM_PANTHOR_GROUP_STATE_FATAL_FAULT: Group had fatal faults. |
| * |
| * When a group ends up with this flag set, no jobs can be submitted to its queues. |
| */ |
| DRM_PANTHOR_GROUP_STATE_FATAL_FAULT = 1 << 1, |
| }; |
| |
| /** |
| * struct drm_panthor_group_get_state - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_GET_STATE |
| * |
| * Used to query the state of a group and decide whether a new group should be created to |
| * replace it. |
| */ |
| struct drm_panthor_group_get_state { |
| /** @group_handle: Handle of the group to query state on */ |
| __u32 group_handle; |
| |
| /** |
| * @state: Combination of DRM_PANTHOR_GROUP_STATE_* flags encoding the |
| * group state. |
| */ |
| __u32 state; |
| |
| /** @fatal_queues: Bitmask of queues that faced fatal faults. */ |
| __u32 fatal_queues; |
| |
| /** @pad: MBZ */ |
| __u32 pad; |
| }; |
| |
| /** |
| * struct drm_panthor_tiler_heap_create - Arguments passed to DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE |
| */ |
| struct drm_panthor_tiler_heap_create { |
| /** @vm_id: VM ID the tiler heap should be mapped to */ |
| __u32 vm_id; |
| |
| /** @initial_chunk_count: Initial number of chunks to allocate. Must be at least one. */ |
| __u32 initial_chunk_count; |
| |
| /** |
| * @chunk_size: Chunk size. |
| * |
| * Must be page-aligned and lie in the [128k:8M] range. |
| */ |
| __u32 chunk_size; |
| |
| /** |
| * @max_chunks: Maximum number of chunks that can be allocated. |
| * |
| * Must be at least @initial_chunk_count. |
| */ |
| __u32 max_chunks; |
| |
| /** |
| * @target_in_flight: Maximum number of in-flight render passes. |
| * |
| * If the heap has more than tiler jobs in-flight, the FW will wait for render |
| * passes to finish before queuing new tiler jobs. |
| */ |
| __u32 target_in_flight; |
| |
| /** @handle: Returned heap handle. Passed back to DESTROY_TILER_HEAP. */ |
| __u32 handle; |
| |
| /** @tiler_heap_ctx_gpu_va: Returned heap GPU virtual address returned */ |
| __u64 tiler_heap_ctx_gpu_va; |
| |
| /** |
| * @first_heap_chunk_gpu_va: First heap chunk. |
| * |
| * The tiler heap is formed of heap chunks forming a single-link list. This |
| * is the first element in the list. |
| */ |
| __u64 first_heap_chunk_gpu_va; |
| }; |
| |
| /** |
| * struct drm_panthor_tiler_heap_destroy - Arguments passed to DRM_IOCTL_PANTHOR_TILER_HEAP_DESTROY |
| */ |
| struct drm_panthor_tiler_heap_destroy { |
| /** |
| * @handle: Handle of the tiler heap to destroy. |
| * |
| * Must be a valid heap handle returned by DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE. |
| */ |
| __u32 handle; |
| |
| /** @pad: Padding field, MBZ. */ |
| __u32 pad; |
| }; |
| |
| #if defined(__cplusplus) |
| } |
| #endif |
| |
| #endif /* _PANTHOR_DRM_H_ */ |