| // SPDX-License-Identifier: MIT |
| /* |
| * Copyright © 2022 Intel Corporation |
| */ |
| |
| #include "xe_bb.h" |
| |
| #include "instructions/xe_mi_commands.h" |
| #include "regs/xe_gpu_commands.h" |
| #include "xe_device.h" |
| #include "xe_exec_queue_types.h" |
| #include "xe_gt.h" |
| #include "xe_hw_fence.h" |
| #include "xe_sa.h" |
| #include "xe_sched_job.h" |
| #include "xe_vm_types.h" |
| |
| static int bb_prefetch(struct xe_gt *gt) |
| { |
| struct xe_device *xe = gt_to_xe(gt); |
| |
| if (GRAPHICS_VERx100(xe) >= 1250 && !xe_gt_is_media_type(gt)) |
| /* |
| * RCS and CCS require 1K, although other engines would be |
| * okay with 512. |
| */ |
| return SZ_1K; |
| else |
| return SZ_512; |
| } |
| |
| struct xe_bb *xe_bb_new(struct xe_gt *gt, u32 dwords, bool usm) |
| { |
| struct xe_tile *tile = gt_to_tile(gt); |
| struct xe_bb *bb = kmalloc(sizeof(*bb), GFP_KERNEL); |
| int err; |
| |
| if (!bb) |
| return ERR_PTR(-ENOMEM); |
| |
| /* |
| * We need to allocate space for the requested number of dwords, |
| * one additional MI_BATCH_BUFFER_END dword, and additional buffer |
| * space to accomodate the platform-specific hardware prefetch |
| * requirements. |
| */ |
| bb->bo = xe_sa_bo_new(!usm ? tile->mem.kernel_bb_pool : gt->usm.bb_pool, |
| 4 * (dwords + 1) + bb_prefetch(gt)); |
| if (IS_ERR(bb->bo)) { |
| err = PTR_ERR(bb->bo); |
| goto err; |
| } |
| |
| bb->cs = xe_sa_bo_cpu_addr(bb->bo); |
| bb->len = 0; |
| |
| return bb; |
| err: |
| kfree(bb); |
| return ERR_PTR(err); |
| } |
| |
| static struct xe_sched_job * |
| __xe_bb_create_job(struct xe_exec_queue *q, struct xe_bb *bb, u64 *addr) |
| { |
| u32 size = drm_suballoc_size(bb->bo); |
| |
| bb->cs[bb->len++] = MI_BATCH_BUFFER_END; |
| |
| xe_gt_assert(q->gt, bb->len * 4 + bb_prefetch(q->gt) <= size); |
| |
| xe_sa_bo_flush_write(bb->bo); |
| |
| return xe_sched_job_create(q, addr); |
| } |
| |
| struct xe_sched_job *xe_bb_create_migration_job(struct xe_exec_queue *q, |
| struct xe_bb *bb, |
| u64 batch_base_ofs, |
| u32 second_idx) |
| { |
| u64 addr[2] = { |
| batch_base_ofs + drm_suballoc_soffset(bb->bo), |
| batch_base_ofs + drm_suballoc_soffset(bb->bo) + |
| 4 * second_idx, |
| }; |
| |
| xe_gt_assert(q->gt, second_idx <= bb->len); |
| xe_gt_assert(q->gt, xe_sched_job_is_migration(q)); |
| xe_gt_assert(q->gt, q->width == 1); |
| |
| return __xe_bb_create_job(q, bb, addr); |
| } |
| |
| struct xe_sched_job *xe_bb_create_job(struct xe_exec_queue *q, |
| struct xe_bb *bb) |
| { |
| u64 addr = xe_sa_bo_gpu_addr(bb->bo); |
| |
| xe_gt_assert(q->gt, !xe_sched_job_is_migration(q)); |
| xe_gt_assert(q->gt, q->width == 1); |
| return __xe_bb_create_job(q, bb, &addr); |
| } |
| |
| void xe_bb_free(struct xe_bb *bb, struct dma_fence *fence) |
| { |
| if (!bb) |
| return; |
| |
| xe_sa_bo_free(bb->bo, fence); |
| kfree(bb); |
| } |