18c2ecf20Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0 28c2ecf20Sopenharmony_ci 38c2ecf20Sopenharmony_ci/* 48c2ecf20Sopenharmony_ci * Copyright 2016-2019 HabanaLabs, Ltd. 58c2ecf20Sopenharmony_ci * All Rights Reserved. 68c2ecf20Sopenharmony_ci */ 78c2ecf20Sopenharmony_ci 88c2ecf20Sopenharmony_ci#include <uapi/misc/habanalabs.h> 98c2ecf20Sopenharmony_ci#include "habanalabs.h" 108c2ecf20Sopenharmony_ci 118c2ecf20Sopenharmony_ci#include <linux/uaccess.h> 128c2ecf20Sopenharmony_ci#include <linux/slab.h> 138c2ecf20Sopenharmony_ci 148c2ecf20Sopenharmony_ci#define HL_CS_FLAGS_SIG_WAIT (HL_CS_FLAGS_SIGNAL | HL_CS_FLAGS_WAIT) 158c2ecf20Sopenharmony_ci 168c2ecf20Sopenharmony_cistatic void job_wq_completion(struct work_struct *work); 178c2ecf20Sopenharmony_cistatic long _hl_cs_wait_ioctl(struct hl_device *hdev, 188c2ecf20Sopenharmony_ci struct hl_ctx *ctx, u64 timeout_us, u64 seq); 198c2ecf20Sopenharmony_cistatic void cs_do_release(struct kref *ref); 208c2ecf20Sopenharmony_ci 218c2ecf20Sopenharmony_cistatic void hl_sob_reset(struct kref *ref) 228c2ecf20Sopenharmony_ci{ 238c2ecf20Sopenharmony_ci struct hl_hw_sob *hw_sob = container_of(ref, struct hl_hw_sob, 248c2ecf20Sopenharmony_ci kref); 258c2ecf20Sopenharmony_ci struct hl_device *hdev = hw_sob->hdev; 268c2ecf20Sopenharmony_ci 278c2ecf20Sopenharmony_ci hdev->asic_funcs->reset_sob(hdev, hw_sob); 288c2ecf20Sopenharmony_ci} 298c2ecf20Sopenharmony_ci 308c2ecf20Sopenharmony_civoid hl_sob_reset_error(struct kref *ref) 318c2ecf20Sopenharmony_ci{ 328c2ecf20Sopenharmony_ci struct hl_hw_sob *hw_sob = container_of(ref, struct hl_hw_sob, 338c2ecf20Sopenharmony_ci kref); 348c2ecf20Sopenharmony_ci struct hl_device *hdev = hw_sob->hdev; 358c2ecf20Sopenharmony_ci 368c2ecf20Sopenharmony_ci dev_crit(hdev->dev, 378c2ecf20Sopenharmony_ci "SOB release shouldn't be called here, q_idx: %d, sob_id: %d\n", 388c2ecf20Sopenharmony_ci hw_sob->q_idx, hw_sob->sob_id); 398c2ecf20Sopenharmony_ci} 408c2ecf20Sopenharmony_ci 418c2ecf20Sopenharmony_cistatic void hl_fence_release(struct kref *kref) 428c2ecf20Sopenharmony_ci{ 438c2ecf20Sopenharmony_ci struct hl_fence *fence = 448c2ecf20Sopenharmony_ci container_of(kref, struct hl_fence, refcount); 458c2ecf20Sopenharmony_ci struct hl_cs_compl *hl_cs_cmpl = 468c2ecf20Sopenharmony_ci container_of(fence, struct hl_cs_compl, base_fence); 478c2ecf20Sopenharmony_ci struct hl_device *hdev = hl_cs_cmpl->hdev; 488c2ecf20Sopenharmony_ci 498c2ecf20Sopenharmony_ci /* EBUSY means the CS was never submitted and hence we don't have 508c2ecf20Sopenharmony_ci * an attached hw_sob object that we should handle here 518c2ecf20Sopenharmony_ci */ 528c2ecf20Sopenharmony_ci if (fence->error == -EBUSY) 538c2ecf20Sopenharmony_ci goto free; 548c2ecf20Sopenharmony_ci 558c2ecf20Sopenharmony_ci if ((hl_cs_cmpl->type == CS_TYPE_SIGNAL) || 568c2ecf20Sopenharmony_ci (hl_cs_cmpl->type == CS_TYPE_WAIT)) { 578c2ecf20Sopenharmony_ci 588c2ecf20Sopenharmony_ci dev_dbg(hdev->dev, 598c2ecf20Sopenharmony_ci "CS 0x%llx type %d finished, sob_id: %d, sob_val: 0x%x\n", 608c2ecf20Sopenharmony_ci hl_cs_cmpl->cs_seq, 618c2ecf20Sopenharmony_ci hl_cs_cmpl->type, 628c2ecf20Sopenharmony_ci hl_cs_cmpl->hw_sob->sob_id, 638c2ecf20Sopenharmony_ci hl_cs_cmpl->sob_val); 648c2ecf20Sopenharmony_ci 658c2ecf20Sopenharmony_ci /* 668c2ecf20Sopenharmony_ci * A signal CS can get completion while the corresponding wait 678c2ecf20Sopenharmony_ci * for signal CS is on its way to the PQ. The wait for signal CS 688c2ecf20Sopenharmony_ci * will get stuck if the signal CS incremented the SOB to its 698c2ecf20Sopenharmony_ci * max value and there are no pending (submitted) waits on this 708c2ecf20Sopenharmony_ci * SOB. 718c2ecf20Sopenharmony_ci * We do the following to void this situation: 728c2ecf20Sopenharmony_ci * 1. The wait for signal CS must get a ref for the signal CS as 738c2ecf20Sopenharmony_ci * soon as possible in cs_ioctl_signal_wait() and put it 748c2ecf20Sopenharmony_ci * before being submitted to the PQ but after it incremented 758c2ecf20Sopenharmony_ci * the SOB refcnt in init_signal_wait_cs(). 768c2ecf20Sopenharmony_ci * 2. Signal/Wait for signal CS will decrement the SOB refcnt 778c2ecf20Sopenharmony_ci * here. 788c2ecf20Sopenharmony_ci * These two measures guarantee that the wait for signal CS will 798c2ecf20Sopenharmony_ci * reset the SOB upon completion rather than the signal CS and 808c2ecf20Sopenharmony_ci * hence the above scenario is avoided. 818c2ecf20Sopenharmony_ci */ 828c2ecf20Sopenharmony_ci kref_put(&hl_cs_cmpl->hw_sob->kref, hl_sob_reset); 838c2ecf20Sopenharmony_ci } 848c2ecf20Sopenharmony_ci 858c2ecf20Sopenharmony_cifree: 868c2ecf20Sopenharmony_ci kfree(hl_cs_cmpl); 878c2ecf20Sopenharmony_ci} 888c2ecf20Sopenharmony_ci 898c2ecf20Sopenharmony_civoid hl_fence_put(struct hl_fence *fence) 908c2ecf20Sopenharmony_ci{ 918c2ecf20Sopenharmony_ci if (fence) 928c2ecf20Sopenharmony_ci kref_put(&fence->refcount, hl_fence_release); 938c2ecf20Sopenharmony_ci} 948c2ecf20Sopenharmony_ci 958c2ecf20Sopenharmony_civoid hl_fence_get(struct hl_fence *fence) 968c2ecf20Sopenharmony_ci{ 978c2ecf20Sopenharmony_ci if (fence) 988c2ecf20Sopenharmony_ci kref_get(&fence->refcount); 998c2ecf20Sopenharmony_ci} 1008c2ecf20Sopenharmony_ci 1018c2ecf20Sopenharmony_cistatic void hl_fence_init(struct hl_fence *fence) 1028c2ecf20Sopenharmony_ci{ 1038c2ecf20Sopenharmony_ci kref_init(&fence->refcount); 1048c2ecf20Sopenharmony_ci fence->error = 0; 1058c2ecf20Sopenharmony_ci init_completion(&fence->completion); 1068c2ecf20Sopenharmony_ci} 1078c2ecf20Sopenharmony_ci 1088c2ecf20Sopenharmony_cistatic void cs_get(struct hl_cs *cs) 1098c2ecf20Sopenharmony_ci{ 1108c2ecf20Sopenharmony_ci kref_get(&cs->refcount); 1118c2ecf20Sopenharmony_ci} 1128c2ecf20Sopenharmony_ci 1138c2ecf20Sopenharmony_cistatic int cs_get_unless_zero(struct hl_cs *cs) 1148c2ecf20Sopenharmony_ci{ 1158c2ecf20Sopenharmony_ci return kref_get_unless_zero(&cs->refcount); 1168c2ecf20Sopenharmony_ci} 1178c2ecf20Sopenharmony_ci 1188c2ecf20Sopenharmony_cistatic void cs_put(struct hl_cs *cs) 1198c2ecf20Sopenharmony_ci{ 1208c2ecf20Sopenharmony_ci kref_put(&cs->refcount, cs_do_release); 1218c2ecf20Sopenharmony_ci} 1228c2ecf20Sopenharmony_ci 1238c2ecf20Sopenharmony_cistatic bool is_cb_patched(struct hl_device *hdev, struct hl_cs_job *job) 1248c2ecf20Sopenharmony_ci{ 1258c2ecf20Sopenharmony_ci /* 1268c2ecf20Sopenharmony_ci * Patched CB is created for external queues jobs, and for H/W queues 1278c2ecf20Sopenharmony_ci * jobs if the user CB was allocated by driver and MMU is disabled. 1288c2ecf20Sopenharmony_ci */ 1298c2ecf20Sopenharmony_ci return (job->queue_type == QUEUE_TYPE_EXT || 1308c2ecf20Sopenharmony_ci (job->queue_type == QUEUE_TYPE_HW && 1318c2ecf20Sopenharmony_ci job->is_kernel_allocated_cb && 1328c2ecf20Sopenharmony_ci !hdev->mmu_enable)); 1338c2ecf20Sopenharmony_ci} 1348c2ecf20Sopenharmony_ci 1358c2ecf20Sopenharmony_ci/* 1368c2ecf20Sopenharmony_ci * cs_parser - parse the user command submission 1378c2ecf20Sopenharmony_ci * 1388c2ecf20Sopenharmony_ci * @hpriv : pointer to the private data of the fd 1398c2ecf20Sopenharmony_ci * @job : pointer to the job that holds the command submission info 1408c2ecf20Sopenharmony_ci * 1418c2ecf20Sopenharmony_ci * The function parses the command submission of the user. It calls the 1428c2ecf20Sopenharmony_ci * ASIC specific parser, which returns a list of memory blocks to send 1438c2ecf20Sopenharmony_ci * to the device as different command buffers 1448c2ecf20Sopenharmony_ci * 1458c2ecf20Sopenharmony_ci */ 1468c2ecf20Sopenharmony_cistatic int cs_parser(struct hl_fpriv *hpriv, struct hl_cs_job *job) 1478c2ecf20Sopenharmony_ci{ 1488c2ecf20Sopenharmony_ci struct hl_device *hdev = hpriv->hdev; 1498c2ecf20Sopenharmony_ci struct hl_cs_parser parser; 1508c2ecf20Sopenharmony_ci int rc; 1518c2ecf20Sopenharmony_ci 1528c2ecf20Sopenharmony_ci parser.ctx_id = job->cs->ctx->asid; 1538c2ecf20Sopenharmony_ci parser.cs_sequence = job->cs->sequence; 1548c2ecf20Sopenharmony_ci parser.job_id = job->id; 1558c2ecf20Sopenharmony_ci 1568c2ecf20Sopenharmony_ci parser.hw_queue_id = job->hw_queue_id; 1578c2ecf20Sopenharmony_ci parser.job_userptr_list = &job->userptr_list; 1588c2ecf20Sopenharmony_ci parser.patched_cb = NULL; 1598c2ecf20Sopenharmony_ci parser.user_cb = job->user_cb; 1608c2ecf20Sopenharmony_ci parser.user_cb_size = job->user_cb_size; 1618c2ecf20Sopenharmony_ci parser.queue_type = job->queue_type; 1628c2ecf20Sopenharmony_ci parser.is_kernel_allocated_cb = job->is_kernel_allocated_cb; 1638c2ecf20Sopenharmony_ci job->patched_cb = NULL; 1648c2ecf20Sopenharmony_ci 1658c2ecf20Sopenharmony_ci rc = hdev->asic_funcs->cs_parser(hdev, &parser); 1668c2ecf20Sopenharmony_ci 1678c2ecf20Sopenharmony_ci if (is_cb_patched(hdev, job)) { 1688c2ecf20Sopenharmony_ci if (!rc) { 1698c2ecf20Sopenharmony_ci job->patched_cb = parser.patched_cb; 1708c2ecf20Sopenharmony_ci job->job_cb_size = parser.patched_cb_size; 1718c2ecf20Sopenharmony_ci job->contains_dma_pkt = parser.contains_dma_pkt; 1728c2ecf20Sopenharmony_ci 1738c2ecf20Sopenharmony_ci spin_lock(&job->patched_cb->lock); 1748c2ecf20Sopenharmony_ci job->patched_cb->cs_cnt++; 1758c2ecf20Sopenharmony_ci spin_unlock(&job->patched_cb->lock); 1768c2ecf20Sopenharmony_ci } 1778c2ecf20Sopenharmony_ci 1788c2ecf20Sopenharmony_ci /* 1798c2ecf20Sopenharmony_ci * Whether the parsing worked or not, we don't need the 1808c2ecf20Sopenharmony_ci * original CB anymore because it was already parsed and 1818c2ecf20Sopenharmony_ci * won't be accessed again for this CS 1828c2ecf20Sopenharmony_ci */ 1838c2ecf20Sopenharmony_ci spin_lock(&job->user_cb->lock); 1848c2ecf20Sopenharmony_ci job->user_cb->cs_cnt--; 1858c2ecf20Sopenharmony_ci spin_unlock(&job->user_cb->lock); 1868c2ecf20Sopenharmony_ci hl_cb_put(job->user_cb); 1878c2ecf20Sopenharmony_ci job->user_cb = NULL; 1888c2ecf20Sopenharmony_ci } else if (!rc) { 1898c2ecf20Sopenharmony_ci job->job_cb_size = job->user_cb_size; 1908c2ecf20Sopenharmony_ci } 1918c2ecf20Sopenharmony_ci 1928c2ecf20Sopenharmony_ci return rc; 1938c2ecf20Sopenharmony_ci} 1948c2ecf20Sopenharmony_ci 1958c2ecf20Sopenharmony_cistatic void free_job(struct hl_device *hdev, struct hl_cs_job *job) 1968c2ecf20Sopenharmony_ci{ 1978c2ecf20Sopenharmony_ci struct hl_cs *cs = job->cs; 1988c2ecf20Sopenharmony_ci 1998c2ecf20Sopenharmony_ci if (is_cb_patched(hdev, job)) { 2008c2ecf20Sopenharmony_ci hl_userptr_delete_list(hdev, &job->userptr_list); 2018c2ecf20Sopenharmony_ci 2028c2ecf20Sopenharmony_ci /* 2038c2ecf20Sopenharmony_ci * We might arrive here from rollback and patched CB wasn't 2048c2ecf20Sopenharmony_ci * created, so we need to check it's not NULL 2058c2ecf20Sopenharmony_ci */ 2068c2ecf20Sopenharmony_ci if (job->patched_cb) { 2078c2ecf20Sopenharmony_ci spin_lock(&job->patched_cb->lock); 2088c2ecf20Sopenharmony_ci job->patched_cb->cs_cnt--; 2098c2ecf20Sopenharmony_ci spin_unlock(&job->patched_cb->lock); 2108c2ecf20Sopenharmony_ci 2118c2ecf20Sopenharmony_ci hl_cb_put(job->patched_cb); 2128c2ecf20Sopenharmony_ci } 2138c2ecf20Sopenharmony_ci } 2148c2ecf20Sopenharmony_ci 2158c2ecf20Sopenharmony_ci /* For H/W queue jobs, if a user CB was allocated by driver and MMU is 2168c2ecf20Sopenharmony_ci * enabled, the user CB isn't released in cs_parser() and thus should be 2178c2ecf20Sopenharmony_ci * released here. 2188c2ecf20Sopenharmony_ci */ 2198c2ecf20Sopenharmony_ci if (job->queue_type == QUEUE_TYPE_HW && 2208c2ecf20Sopenharmony_ci job->is_kernel_allocated_cb && hdev->mmu_enable) { 2218c2ecf20Sopenharmony_ci spin_lock(&job->user_cb->lock); 2228c2ecf20Sopenharmony_ci job->user_cb->cs_cnt--; 2238c2ecf20Sopenharmony_ci spin_unlock(&job->user_cb->lock); 2248c2ecf20Sopenharmony_ci 2258c2ecf20Sopenharmony_ci hl_cb_put(job->user_cb); 2268c2ecf20Sopenharmony_ci } 2278c2ecf20Sopenharmony_ci 2288c2ecf20Sopenharmony_ci /* 2298c2ecf20Sopenharmony_ci * This is the only place where there can be multiple threads 2308c2ecf20Sopenharmony_ci * modifying the list at the same time 2318c2ecf20Sopenharmony_ci */ 2328c2ecf20Sopenharmony_ci spin_lock(&cs->job_lock); 2338c2ecf20Sopenharmony_ci list_del(&job->cs_node); 2348c2ecf20Sopenharmony_ci spin_unlock(&cs->job_lock); 2358c2ecf20Sopenharmony_ci 2368c2ecf20Sopenharmony_ci hl_debugfs_remove_job(hdev, job); 2378c2ecf20Sopenharmony_ci 2388c2ecf20Sopenharmony_ci if (job->queue_type == QUEUE_TYPE_EXT || 2398c2ecf20Sopenharmony_ci job->queue_type == QUEUE_TYPE_HW) 2408c2ecf20Sopenharmony_ci cs_put(cs); 2418c2ecf20Sopenharmony_ci 2428c2ecf20Sopenharmony_ci kfree(job); 2438c2ecf20Sopenharmony_ci} 2448c2ecf20Sopenharmony_ci 2458c2ecf20Sopenharmony_cistatic void cs_counters_aggregate(struct hl_device *hdev, struct hl_ctx *ctx) 2468c2ecf20Sopenharmony_ci{ 2478c2ecf20Sopenharmony_ci hdev->aggregated_cs_counters.device_in_reset_drop_cnt += 2488c2ecf20Sopenharmony_ci ctx->cs_counters.device_in_reset_drop_cnt; 2498c2ecf20Sopenharmony_ci hdev->aggregated_cs_counters.out_of_mem_drop_cnt += 2508c2ecf20Sopenharmony_ci ctx->cs_counters.out_of_mem_drop_cnt; 2518c2ecf20Sopenharmony_ci hdev->aggregated_cs_counters.parsing_drop_cnt += 2528c2ecf20Sopenharmony_ci ctx->cs_counters.parsing_drop_cnt; 2538c2ecf20Sopenharmony_ci hdev->aggregated_cs_counters.queue_full_drop_cnt += 2548c2ecf20Sopenharmony_ci ctx->cs_counters.queue_full_drop_cnt; 2558c2ecf20Sopenharmony_ci hdev->aggregated_cs_counters.max_cs_in_flight_drop_cnt += 2568c2ecf20Sopenharmony_ci ctx->cs_counters.max_cs_in_flight_drop_cnt; 2578c2ecf20Sopenharmony_ci} 2588c2ecf20Sopenharmony_ci 2598c2ecf20Sopenharmony_cistatic void cs_do_release(struct kref *ref) 2608c2ecf20Sopenharmony_ci{ 2618c2ecf20Sopenharmony_ci struct hl_cs *cs = container_of(ref, struct hl_cs, 2628c2ecf20Sopenharmony_ci refcount); 2638c2ecf20Sopenharmony_ci struct hl_device *hdev = cs->ctx->hdev; 2648c2ecf20Sopenharmony_ci struct hl_cs_job *job, *tmp; 2658c2ecf20Sopenharmony_ci 2668c2ecf20Sopenharmony_ci cs->completed = true; 2678c2ecf20Sopenharmony_ci 2688c2ecf20Sopenharmony_ci /* 2698c2ecf20Sopenharmony_ci * Although if we reached here it means that all external jobs have 2708c2ecf20Sopenharmony_ci * finished, because each one of them took refcnt to CS, we still 2718c2ecf20Sopenharmony_ci * need to go over the internal jobs and free them. Otherwise, we 2728c2ecf20Sopenharmony_ci * will have leaked memory and what's worse, the CS object (and 2738c2ecf20Sopenharmony_ci * potentially the CTX object) could be released, while the JOB 2748c2ecf20Sopenharmony_ci * still holds a pointer to them (but no reference). 2758c2ecf20Sopenharmony_ci */ 2768c2ecf20Sopenharmony_ci list_for_each_entry_safe(job, tmp, &cs->job_list, cs_node) 2778c2ecf20Sopenharmony_ci free_job(hdev, job); 2788c2ecf20Sopenharmony_ci 2798c2ecf20Sopenharmony_ci /* We also need to update CI for internal queues */ 2808c2ecf20Sopenharmony_ci if (cs->submitted) { 2818c2ecf20Sopenharmony_ci hdev->asic_funcs->hw_queues_lock(hdev); 2828c2ecf20Sopenharmony_ci 2838c2ecf20Sopenharmony_ci hdev->cs_active_cnt--; 2848c2ecf20Sopenharmony_ci if (!hdev->cs_active_cnt) { 2858c2ecf20Sopenharmony_ci struct hl_device_idle_busy_ts *ts; 2868c2ecf20Sopenharmony_ci 2878c2ecf20Sopenharmony_ci ts = &hdev->idle_busy_ts_arr[hdev->idle_busy_ts_idx++]; 2888c2ecf20Sopenharmony_ci ts->busy_to_idle_ts = ktime_get(); 2898c2ecf20Sopenharmony_ci 2908c2ecf20Sopenharmony_ci if (hdev->idle_busy_ts_idx == HL_IDLE_BUSY_TS_ARR_SIZE) 2918c2ecf20Sopenharmony_ci hdev->idle_busy_ts_idx = 0; 2928c2ecf20Sopenharmony_ci } else if (hdev->cs_active_cnt < 0) { 2938c2ecf20Sopenharmony_ci dev_crit(hdev->dev, "CS active cnt %d is negative\n", 2948c2ecf20Sopenharmony_ci hdev->cs_active_cnt); 2958c2ecf20Sopenharmony_ci } 2968c2ecf20Sopenharmony_ci 2978c2ecf20Sopenharmony_ci hdev->asic_funcs->hw_queues_unlock(hdev); 2988c2ecf20Sopenharmony_ci 2998c2ecf20Sopenharmony_ci hl_int_hw_queue_update_ci(cs); 3008c2ecf20Sopenharmony_ci 3018c2ecf20Sopenharmony_ci spin_lock(&hdev->hw_queues_mirror_lock); 3028c2ecf20Sopenharmony_ci /* remove CS from hw_queues mirror list */ 3038c2ecf20Sopenharmony_ci list_del_init(&cs->mirror_node); 3048c2ecf20Sopenharmony_ci spin_unlock(&hdev->hw_queues_mirror_lock); 3058c2ecf20Sopenharmony_ci 3068c2ecf20Sopenharmony_ci /* 3078c2ecf20Sopenharmony_ci * Don't cancel TDR in case this CS was timedout because we 3088c2ecf20Sopenharmony_ci * might be running from the TDR context 3098c2ecf20Sopenharmony_ci */ 3108c2ecf20Sopenharmony_ci if ((!cs->timedout) && 3118c2ecf20Sopenharmony_ci (hdev->timeout_jiffies != MAX_SCHEDULE_TIMEOUT)) { 3128c2ecf20Sopenharmony_ci struct hl_cs *next; 3138c2ecf20Sopenharmony_ci 3148c2ecf20Sopenharmony_ci if (cs->tdr_active) 3158c2ecf20Sopenharmony_ci cancel_delayed_work_sync(&cs->work_tdr); 3168c2ecf20Sopenharmony_ci 3178c2ecf20Sopenharmony_ci spin_lock(&hdev->hw_queues_mirror_lock); 3188c2ecf20Sopenharmony_ci 3198c2ecf20Sopenharmony_ci /* queue TDR for next CS */ 3208c2ecf20Sopenharmony_ci next = list_first_entry_or_null( 3218c2ecf20Sopenharmony_ci &hdev->hw_queues_mirror_list, 3228c2ecf20Sopenharmony_ci struct hl_cs, mirror_node); 3238c2ecf20Sopenharmony_ci 3248c2ecf20Sopenharmony_ci if ((next) && (!next->tdr_active)) { 3258c2ecf20Sopenharmony_ci next->tdr_active = true; 3268c2ecf20Sopenharmony_ci schedule_delayed_work(&next->work_tdr, 3278c2ecf20Sopenharmony_ci hdev->timeout_jiffies); 3288c2ecf20Sopenharmony_ci } 3298c2ecf20Sopenharmony_ci 3308c2ecf20Sopenharmony_ci spin_unlock(&hdev->hw_queues_mirror_lock); 3318c2ecf20Sopenharmony_ci } 3328c2ecf20Sopenharmony_ci } else if (cs->type == CS_TYPE_WAIT) { 3338c2ecf20Sopenharmony_ci /* 3348c2ecf20Sopenharmony_ci * In case the wait for signal CS was submitted, the put occurs 3358c2ecf20Sopenharmony_ci * in init_signal_wait_cs() right before hanging on the PQ. 3368c2ecf20Sopenharmony_ci */ 3378c2ecf20Sopenharmony_ci hl_fence_put(cs->signal_fence); 3388c2ecf20Sopenharmony_ci } 3398c2ecf20Sopenharmony_ci 3408c2ecf20Sopenharmony_ci /* 3418c2ecf20Sopenharmony_ci * Must be called before hl_ctx_put because inside we use ctx to get 3428c2ecf20Sopenharmony_ci * the device 3438c2ecf20Sopenharmony_ci */ 3448c2ecf20Sopenharmony_ci hl_debugfs_remove_cs(cs); 3458c2ecf20Sopenharmony_ci 3468c2ecf20Sopenharmony_ci hl_ctx_put(cs->ctx); 3478c2ecf20Sopenharmony_ci 3488c2ecf20Sopenharmony_ci /* We need to mark an error for not submitted because in that case 3498c2ecf20Sopenharmony_ci * the hl fence release flow is different. Mainly, we don't need 3508c2ecf20Sopenharmony_ci * to handle hw_sob for signal/wait 3518c2ecf20Sopenharmony_ci */ 3528c2ecf20Sopenharmony_ci if (cs->timedout) 3538c2ecf20Sopenharmony_ci cs->fence->error = -ETIMEDOUT; 3548c2ecf20Sopenharmony_ci else if (cs->aborted) 3558c2ecf20Sopenharmony_ci cs->fence->error = -EIO; 3568c2ecf20Sopenharmony_ci else if (!cs->submitted) 3578c2ecf20Sopenharmony_ci cs->fence->error = -EBUSY; 3588c2ecf20Sopenharmony_ci 3598c2ecf20Sopenharmony_ci complete_all(&cs->fence->completion); 3608c2ecf20Sopenharmony_ci hl_fence_put(cs->fence); 3618c2ecf20Sopenharmony_ci cs_counters_aggregate(hdev, cs->ctx); 3628c2ecf20Sopenharmony_ci 3638c2ecf20Sopenharmony_ci kfree(cs->jobs_in_queue_cnt); 3648c2ecf20Sopenharmony_ci kfree(cs); 3658c2ecf20Sopenharmony_ci} 3668c2ecf20Sopenharmony_ci 3678c2ecf20Sopenharmony_cistatic void cs_timedout(struct work_struct *work) 3688c2ecf20Sopenharmony_ci{ 3698c2ecf20Sopenharmony_ci struct hl_device *hdev; 3708c2ecf20Sopenharmony_ci int rc; 3718c2ecf20Sopenharmony_ci struct hl_cs *cs = container_of(work, struct hl_cs, 3728c2ecf20Sopenharmony_ci work_tdr.work); 3738c2ecf20Sopenharmony_ci rc = cs_get_unless_zero(cs); 3748c2ecf20Sopenharmony_ci if (!rc) 3758c2ecf20Sopenharmony_ci return; 3768c2ecf20Sopenharmony_ci 3778c2ecf20Sopenharmony_ci if ((!cs->submitted) || (cs->completed)) { 3788c2ecf20Sopenharmony_ci cs_put(cs); 3798c2ecf20Sopenharmony_ci return; 3808c2ecf20Sopenharmony_ci } 3818c2ecf20Sopenharmony_ci 3828c2ecf20Sopenharmony_ci /* Mark the CS is timed out so we won't try to cancel its TDR */ 3838c2ecf20Sopenharmony_ci cs->timedout = true; 3848c2ecf20Sopenharmony_ci 3858c2ecf20Sopenharmony_ci hdev = cs->ctx->hdev; 3868c2ecf20Sopenharmony_ci 3878c2ecf20Sopenharmony_ci dev_err(hdev->dev, 3888c2ecf20Sopenharmony_ci "Command submission %llu has not finished in time!\n", 3898c2ecf20Sopenharmony_ci cs->sequence); 3908c2ecf20Sopenharmony_ci 3918c2ecf20Sopenharmony_ci cs_put(cs); 3928c2ecf20Sopenharmony_ci 3938c2ecf20Sopenharmony_ci if (hdev->reset_on_lockup) 3948c2ecf20Sopenharmony_ci hl_device_reset(hdev, false, false); 3958c2ecf20Sopenharmony_ci} 3968c2ecf20Sopenharmony_ci 3978c2ecf20Sopenharmony_cistatic int allocate_cs(struct hl_device *hdev, struct hl_ctx *ctx, 3988c2ecf20Sopenharmony_ci enum hl_cs_type cs_type, struct hl_cs **cs_new) 3998c2ecf20Sopenharmony_ci{ 4008c2ecf20Sopenharmony_ci struct hl_cs_compl *cs_cmpl; 4018c2ecf20Sopenharmony_ci struct hl_fence *other = NULL; 4028c2ecf20Sopenharmony_ci struct hl_cs *cs; 4038c2ecf20Sopenharmony_ci int rc; 4048c2ecf20Sopenharmony_ci 4058c2ecf20Sopenharmony_ci cs = kzalloc(sizeof(*cs), GFP_ATOMIC); 4068c2ecf20Sopenharmony_ci if (!cs) 4078c2ecf20Sopenharmony_ci return -ENOMEM; 4088c2ecf20Sopenharmony_ci 4098c2ecf20Sopenharmony_ci cs->ctx = ctx; 4108c2ecf20Sopenharmony_ci cs->submitted = false; 4118c2ecf20Sopenharmony_ci cs->completed = false; 4128c2ecf20Sopenharmony_ci cs->type = cs_type; 4138c2ecf20Sopenharmony_ci INIT_LIST_HEAD(&cs->job_list); 4148c2ecf20Sopenharmony_ci INIT_DELAYED_WORK(&cs->work_tdr, cs_timedout); 4158c2ecf20Sopenharmony_ci kref_init(&cs->refcount); 4168c2ecf20Sopenharmony_ci spin_lock_init(&cs->job_lock); 4178c2ecf20Sopenharmony_ci 4188c2ecf20Sopenharmony_ci cs_cmpl = kmalloc(sizeof(*cs_cmpl), GFP_ATOMIC); 4198c2ecf20Sopenharmony_ci if (!cs_cmpl) { 4208c2ecf20Sopenharmony_ci rc = -ENOMEM; 4218c2ecf20Sopenharmony_ci goto free_cs; 4228c2ecf20Sopenharmony_ci } 4238c2ecf20Sopenharmony_ci 4248c2ecf20Sopenharmony_ci cs_cmpl->hdev = hdev; 4258c2ecf20Sopenharmony_ci cs_cmpl->type = cs->type; 4268c2ecf20Sopenharmony_ci spin_lock_init(&cs_cmpl->lock); 4278c2ecf20Sopenharmony_ci cs->fence = &cs_cmpl->base_fence; 4288c2ecf20Sopenharmony_ci 4298c2ecf20Sopenharmony_ci spin_lock(&ctx->cs_lock); 4308c2ecf20Sopenharmony_ci 4318c2ecf20Sopenharmony_ci cs_cmpl->cs_seq = ctx->cs_sequence; 4328c2ecf20Sopenharmony_ci other = ctx->cs_pending[cs_cmpl->cs_seq & 4338c2ecf20Sopenharmony_ci (hdev->asic_prop.max_pending_cs - 1)]; 4348c2ecf20Sopenharmony_ci 4358c2ecf20Sopenharmony_ci if (other && !completion_done(&other->completion)) { 4368c2ecf20Sopenharmony_ci dev_dbg_ratelimited(hdev->dev, 4378c2ecf20Sopenharmony_ci "Rejecting CS because of too many in-flights CS\n"); 4388c2ecf20Sopenharmony_ci ctx->cs_counters.max_cs_in_flight_drop_cnt++; 4398c2ecf20Sopenharmony_ci rc = -EAGAIN; 4408c2ecf20Sopenharmony_ci goto free_fence; 4418c2ecf20Sopenharmony_ci } 4428c2ecf20Sopenharmony_ci 4438c2ecf20Sopenharmony_ci cs->jobs_in_queue_cnt = kcalloc(hdev->asic_prop.max_queues, 4448c2ecf20Sopenharmony_ci sizeof(*cs->jobs_in_queue_cnt), GFP_ATOMIC); 4458c2ecf20Sopenharmony_ci if (!cs->jobs_in_queue_cnt) { 4468c2ecf20Sopenharmony_ci rc = -ENOMEM; 4478c2ecf20Sopenharmony_ci goto free_fence; 4488c2ecf20Sopenharmony_ci } 4498c2ecf20Sopenharmony_ci 4508c2ecf20Sopenharmony_ci /* init hl_fence */ 4518c2ecf20Sopenharmony_ci hl_fence_init(&cs_cmpl->base_fence); 4528c2ecf20Sopenharmony_ci 4538c2ecf20Sopenharmony_ci cs->sequence = cs_cmpl->cs_seq; 4548c2ecf20Sopenharmony_ci 4558c2ecf20Sopenharmony_ci ctx->cs_pending[cs_cmpl->cs_seq & 4568c2ecf20Sopenharmony_ci (hdev->asic_prop.max_pending_cs - 1)] = 4578c2ecf20Sopenharmony_ci &cs_cmpl->base_fence; 4588c2ecf20Sopenharmony_ci ctx->cs_sequence++; 4598c2ecf20Sopenharmony_ci 4608c2ecf20Sopenharmony_ci hl_fence_get(&cs_cmpl->base_fence); 4618c2ecf20Sopenharmony_ci 4628c2ecf20Sopenharmony_ci hl_fence_put(other); 4638c2ecf20Sopenharmony_ci 4648c2ecf20Sopenharmony_ci spin_unlock(&ctx->cs_lock); 4658c2ecf20Sopenharmony_ci 4668c2ecf20Sopenharmony_ci *cs_new = cs; 4678c2ecf20Sopenharmony_ci 4688c2ecf20Sopenharmony_ci return 0; 4698c2ecf20Sopenharmony_ci 4708c2ecf20Sopenharmony_cifree_fence: 4718c2ecf20Sopenharmony_ci spin_unlock(&ctx->cs_lock); 4728c2ecf20Sopenharmony_ci kfree(cs_cmpl); 4738c2ecf20Sopenharmony_cifree_cs: 4748c2ecf20Sopenharmony_ci kfree(cs); 4758c2ecf20Sopenharmony_ci return rc; 4768c2ecf20Sopenharmony_ci} 4778c2ecf20Sopenharmony_ci 4788c2ecf20Sopenharmony_cistatic void cs_rollback(struct hl_device *hdev, struct hl_cs *cs) 4798c2ecf20Sopenharmony_ci{ 4808c2ecf20Sopenharmony_ci struct hl_cs_job *job, *tmp; 4818c2ecf20Sopenharmony_ci 4828c2ecf20Sopenharmony_ci list_for_each_entry_safe(job, tmp, &cs->job_list, cs_node) 4838c2ecf20Sopenharmony_ci free_job(hdev, job); 4848c2ecf20Sopenharmony_ci} 4858c2ecf20Sopenharmony_ci 4868c2ecf20Sopenharmony_civoid hl_cs_rollback_all(struct hl_device *hdev) 4878c2ecf20Sopenharmony_ci{ 4888c2ecf20Sopenharmony_ci int i; 4898c2ecf20Sopenharmony_ci struct hl_cs *cs, *tmp; 4908c2ecf20Sopenharmony_ci 4918c2ecf20Sopenharmony_ci /* flush all completions */ 4928c2ecf20Sopenharmony_ci for (i = 0 ; i < hdev->asic_prop.completion_queues_count ; i++) 4938c2ecf20Sopenharmony_ci flush_workqueue(hdev->cq_wq[i]); 4948c2ecf20Sopenharmony_ci 4958c2ecf20Sopenharmony_ci /* Make sure we don't have leftovers in the H/W queues mirror list */ 4968c2ecf20Sopenharmony_ci list_for_each_entry_safe(cs, tmp, &hdev->hw_queues_mirror_list, 4978c2ecf20Sopenharmony_ci mirror_node) { 4988c2ecf20Sopenharmony_ci cs_get(cs); 4998c2ecf20Sopenharmony_ci cs->aborted = true; 5008c2ecf20Sopenharmony_ci dev_warn_ratelimited(hdev->dev, "Killing CS %d.%llu\n", 5018c2ecf20Sopenharmony_ci cs->ctx->asid, cs->sequence); 5028c2ecf20Sopenharmony_ci cs_rollback(hdev, cs); 5038c2ecf20Sopenharmony_ci cs_put(cs); 5048c2ecf20Sopenharmony_ci } 5058c2ecf20Sopenharmony_ci} 5068c2ecf20Sopenharmony_ci 5078c2ecf20Sopenharmony_cistatic void job_wq_completion(struct work_struct *work) 5088c2ecf20Sopenharmony_ci{ 5098c2ecf20Sopenharmony_ci struct hl_cs_job *job = container_of(work, struct hl_cs_job, 5108c2ecf20Sopenharmony_ci finish_work); 5118c2ecf20Sopenharmony_ci struct hl_cs *cs = job->cs; 5128c2ecf20Sopenharmony_ci struct hl_device *hdev = cs->ctx->hdev; 5138c2ecf20Sopenharmony_ci 5148c2ecf20Sopenharmony_ci /* job is no longer needed */ 5158c2ecf20Sopenharmony_ci free_job(hdev, job); 5168c2ecf20Sopenharmony_ci} 5178c2ecf20Sopenharmony_ci 5188c2ecf20Sopenharmony_cistatic int validate_queue_index(struct hl_device *hdev, 5198c2ecf20Sopenharmony_ci struct hl_cs_chunk *chunk, 5208c2ecf20Sopenharmony_ci enum hl_queue_type *queue_type, 5218c2ecf20Sopenharmony_ci bool *is_kernel_allocated_cb) 5228c2ecf20Sopenharmony_ci{ 5238c2ecf20Sopenharmony_ci struct asic_fixed_properties *asic = &hdev->asic_prop; 5248c2ecf20Sopenharmony_ci struct hw_queue_properties *hw_queue_prop; 5258c2ecf20Sopenharmony_ci 5268c2ecf20Sopenharmony_ci /* This must be checked here to prevent out-of-bounds access to 5278c2ecf20Sopenharmony_ci * hw_queues_props array 5288c2ecf20Sopenharmony_ci */ 5298c2ecf20Sopenharmony_ci if (chunk->queue_index >= asic->max_queues) { 5308c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Queue index %d is invalid\n", 5318c2ecf20Sopenharmony_ci chunk->queue_index); 5328c2ecf20Sopenharmony_ci return -EINVAL; 5338c2ecf20Sopenharmony_ci } 5348c2ecf20Sopenharmony_ci 5358c2ecf20Sopenharmony_ci hw_queue_prop = &asic->hw_queues_props[chunk->queue_index]; 5368c2ecf20Sopenharmony_ci 5378c2ecf20Sopenharmony_ci if (hw_queue_prop->type == QUEUE_TYPE_NA) { 5388c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Queue index %d is invalid\n", 5398c2ecf20Sopenharmony_ci chunk->queue_index); 5408c2ecf20Sopenharmony_ci return -EINVAL; 5418c2ecf20Sopenharmony_ci } 5428c2ecf20Sopenharmony_ci 5438c2ecf20Sopenharmony_ci if (hw_queue_prop->driver_only) { 5448c2ecf20Sopenharmony_ci dev_err(hdev->dev, 5458c2ecf20Sopenharmony_ci "Queue index %d is restricted for the kernel driver\n", 5468c2ecf20Sopenharmony_ci chunk->queue_index); 5478c2ecf20Sopenharmony_ci return -EINVAL; 5488c2ecf20Sopenharmony_ci } 5498c2ecf20Sopenharmony_ci 5508c2ecf20Sopenharmony_ci *queue_type = hw_queue_prop->type; 5518c2ecf20Sopenharmony_ci *is_kernel_allocated_cb = !!hw_queue_prop->requires_kernel_cb; 5528c2ecf20Sopenharmony_ci 5538c2ecf20Sopenharmony_ci return 0; 5548c2ecf20Sopenharmony_ci} 5558c2ecf20Sopenharmony_ci 5568c2ecf20Sopenharmony_cistatic struct hl_cb *get_cb_from_cs_chunk(struct hl_device *hdev, 5578c2ecf20Sopenharmony_ci struct hl_cb_mgr *cb_mgr, 5588c2ecf20Sopenharmony_ci struct hl_cs_chunk *chunk) 5598c2ecf20Sopenharmony_ci{ 5608c2ecf20Sopenharmony_ci struct hl_cb *cb; 5618c2ecf20Sopenharmony_ci u32 cb_handle; 5628c2ecf20Sopenharmony_ci 5638c2ecf20Sopenharmony_ci cb_handle = (u32) (chunk->cb_handle >> PAGE_SHIFT); 5648c2ecf20Sopenharmony_ci 5658c2ecf20Sopenharmony_ci cb = hl_cb_get(hdev, cb_mgr, cb_handle); 5668c2ecf20Sopenharmony_ci if (!cb) { 5678c2ecf20Sopenharmony_ci dev_err(hdev->dev, "CB handle 0x%x invalid\n", cb_handle); 5688c2ecf20Sopenharmony_ci return NULL; 5698c2ecf20Sopenharmony_ci } 5708c2ecf20Sopenharmony_ci 5718c2ecf20Sopenharmony_ci if ((chunk->cb_size < 8) || (chunk->cb_size > cb->size)) { 5728c2ecf20Sopenharmony_ci dev_err(hdev->dev, "CB size %u invalid\n", chunk->cb_size); 5738c2ecf20Sopenharmony_ci goto release_cb; 5748c2ecf20Sopenharmony_ci } 5758c2ecf20Sopenharmony_ci 5768c2ecf20Sopenharmony_ci spin_lock(&cb->lock); 5778c2ecf20Sopenharmony_ci cb->cs_cnt++; 5788c2ecf20Sopenharmony_ci spin_unlock(&cb->lock); 5798c2ecf20Sopenharmony_ci 5808c2ecf20Sopenharmony_ci return cb; 5818c2ecf20Sopenharmony_ci 5828c2ecf20Sopenharmony_cirelease_cb: 5838c2ecf20Sopenharmony_ci hl_cb_put(cb); 5848c2ecf20Sopenharmony_ci return NULL; 5858c2ecf20Sopenharmony_ci} 5868c2ecf20Sopenharmony_ci 5878c2ecf20Sopenharmony_cistruct hl_cs_job *hl_cs_allocate_job(struct hl_device *hdev, 5888c2ecf20Sopenharmony_ci enum hl_queue_type queue_type, bool is_kernel_allocated_cb) 5898c2ecf20Sopenharmony_ci{ 5908c2ecf20Sopenharmony_ci struct hl_cs_job *job; 5918c2ecf20Sopenharmony_ci 5928c2ecf20Sopenharmony_ci job = kzalloc(sizeof(*job), GFP_ATOMIC); 5938c2ecf20Sopenharmony_ci if (!job) 5948c2ecf20Sopenharmony_ci return NULL; 5958c2ecf20Sopenharmony_ci 5968c2ecf20Sopenharmony_ci job->queue_type = queue_type; 5978c2ecf20Sopenharmony_ci job->is_kernel_allocated_cb = is_kernel_allocated_cb; 5988c2ecf20Sopenharmony_ci 5998c2ecf20Sopenharmony_ci if (is_cb_patched(hdev, job)) 6008c2ecf20Sopenharmony_ci INIT_LIST_HEAD(&job->userptr_list); 6018c2ecf20Sopenharmony_ci 6028c2ecf20Sopenharmony_ci if (job->queue_type == QUEUE_TYPE_EXT) 6038c2ecf20Sopenharmony_ci INIT_WORK(&job->finish_work, job_wq_completion); 6048c2ecf20Sopenharmony_ci 6058c2ecf20Sopenharmony_ci return job; 6068c2ecf20Sopenharmony_ci} 6078c2ecf20Sopenharmony_ci 6088c2ecf20Sopenharmony_cistatic int cs_ioctl_default(struct hl_fpriv *hpriv, void __user *chunks, 6098c2ecf20Sopenharmony_ci u32 num_chunks, u64 *cs_seq) 6108c2ecf20Sopenharmony_ci{ 6118c2ecf20Sopenharmony_ci struct hl_device *hdev = hpriv->hdev; 6128c2ecf20Sopenharmony_ci struct hl_cs_chunk *cs_chunk_array; 6138c2ecf20Sopenharmony_ci struct hl_cs_job *job; 6148c2ecf20Sopenharmony_ci struct hl_cs *cs; 6158c2ecf20Sopenharmony_ci struct hl_cb *cb; 6168c2ecf20Sopenharmony_ci bool int_queues_only = true; 6178c2ecf20Sopenharmony_ci u32 size_to_copy; 6188c2ecf20Sopenharmony_ci int rc, i; 6198c2ecf20Sopenharmony_ci 6208c2ecf20Sopenharmony_ci *cs_seq = ULLONG_MAX; 6218c2ecf20Sopenharmony_ci 6228c2ecf20Sopenharmony_ci if (num_chunks > HL_MAX_JOBS_PER_CS) { 6238c2ecf20Sopenharmony_ci dev_err(hdev->dev, 6248c2ecf20Sopenharmony_ci "Number of chunks can NOT be larger than %d\n", 6258c2ecf20Sopenharmony_ci HL_MAX_JOBS_PER_CS); 6268c2ecf20Sopenharmony_ci rc = -EINVAL; 6278c2ecf20Sopenharmony_ci goto out; 6288c2ecf20Sopenharmony_ci } 6298c2ecf20Sopenharmony_ci 6308c2ecf20Sopenharmony_ci cs_chunk_array = kmalloc_array(num_chunks, sizeof(*cs_chunk_array), 6318c2ecf20Sopenharmony_ci GFP_ATOMIC); 6328c2ecf20Sopenharmony_ci if (!cs_chunk_array) { 6338c2ecf20Sopenharmony_ci rc = -ENOMEM; 6348c2ecf20Sopenharmony_ci goto out; 6358c2ecf20Sopenharmony_ci } 6368c2ecf20Sopenharmony_ci 6378c2ecf20Sopenharmony_ci size_to_copy = num_chunks * sizeof(struct hl_cs_chunk); 6388c2ecf20Sopenharmony_ci if (copy_from_user(cs_chunk_array, chunks, size_to_copy)) { 6398c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Failed to copy cs chunk array from user\n"); 6408c2ecf20Sopenharmony_ci rc = -EFAULT; 6418c2ecf20Sopenharmony_ci goto free_cs_chunk_array; 6428c2ecf20Sopenharmony_ci } 6438c2ecf20Sopenharmony_ci 6448c2ecf20Sopenharmony_ci /* increment refcnt for context */ 6458c2ecf20Sopenharmony_ci hl_ctx_get(hdev, hpriv->ctx); 6468c2ecf20Sopenharmony_ci 6478c2ecf20Sopenharmony_ci rc = allocate_cs(hdev, hpriv->ctx, CS_TYPE_DEFAULT, &cs); 6488c2ecf20Sopenharmony_ci if (rc) { 6498c2ecf20Sopenharmony_ci hl_ctx_put(hpriv->ctx); 6508c2ecf20Sopenharmony_ci goto free_cs_chunk_array; 6518c2ecf20Sopenharmony_ci } 6528c2ecf20Sopenharmony_ci 6538c2ecf20Sopenharmony_ci *cs_seq = cs->sequence; 6548c2ecf20Sopenharmony_ci 6558c2ecf20Sopenharmony_ci hl_debugfs_add_cs(cs); 6568c2ecf20Sopenharmony_ci 6578c2ecf20Sopenharmony_ci /* Validate ALL the CS chunks before submitting the CS */ 6588c2ecf20Sopenharmony_ci for (i = 0 ; i < num_chunks ; i++) { 6598c2ecf20Sopenharmony_ci struct hl_cs_chunk *chunk = &cs_chunk_array[i]; 6608c2ecf20Sopenharmony_ci enum hl_queue_type queue_type; 6618c2ecf20Sopenharmony_ci bool is_kernel_allocated_cb; 6628c2ecf20Sopenharmony_ci 6638c2ecf20Sopenharmony_ci rc = validate_queue_index(hdev, chunk, &queue_type, 6648c2ecf20Sopenharmony_ci &is_kernel_allocated_cb); 6658c2ecf20Sopenharmony_ci if (rc) { 6668c2ecf20Sopenharmony_ci hpriv->ctx->cs_counters.parsing_drop_cnt++; 6678c2ecf20Sopenharmony_ci goto free_cs_object; 6688c2ecf20Sopenharmony_ci } 6698c2ecf20Sopenharmony_ci 6708c2ecf20Sopenharmony_ci if (is_kernel_allocated_cb) { 6718c2ecf20Sopenharmony_ci cb = get_cb_from_cs_chunk(hdev, &hpriv->cb_mgr, chunk); 6728c2ecf20Sopenharmony_ci if (!cb) { 6738c2ecf20Sopenharmony_ci hpriv->ctx->cs_counters.parsing_drop_cnt++; 6748c2ecf20Sopenharmony_ci rc = -EINVAL; 6758c2ecf20Sopenharmony_ci goto free_cs_object; 6768c2ecf20Sopenharmony_ci } 6778c2ecf20Sopenharmony_ci } else { 6788c2ecf20Sopenharmony_ci cb = (struct hl_cb *) (uintptr_t) chunk->cb_handle; 6798c2ecf20Sopenharmony_ci } 6808c2ecf20Sopenharmony_ci 6818c2ecf20Sopenharmony_ci if (queue_type == QUEUE_TYPE_EXT || queue_type == QUEUE_TYPE_HW) 6828c2ecf20Sopenharmony_ci int_queues_only = false; 6838c2ecf20Sopenharmony_ci 6848c2ecf20Sopenharmony_ci job = hl_cs_allocate_job(hdev, queue_type, 6858c2ecf20Sopenharmony_ci is_kernel_allocated_cb); 6868c2ecf20Sopenharmony_ci if (!job) { 6878c2ecf20Sopenharmony_ci hpriv->ctx->cs_counters.out_of_mem_drop_cnt++; 6888c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Failed to allocate a new job\n"); 6898c2ecf20Sopenharmony_ci rc = -ENOMEM; 6908c2ecf20Sopenharmony_ci if (is_kernel_allocated_cb) 6918c2ecf20Sopenharmony_ci goto release_cb; 6928c2ecf20Sopenharmony_ci 6938c2ecf20Sopenharmony_ci goto free_cs_object; 6948c2ecf20Sopenharmony_ci } 6958c2ecf20Sopenharmony_ci 6968c2ecf20Sopenharmony_ci job->id = i + 1; 6978c2ecf20Sopenharmony_ci job->cs = cs; 6988c2ecf20Sopenharmony_ci job->user_cb = cb; 6998c2ecf20Sopenharmony_ci job->user_cb_size = chunk->cb_size; 7008c2ecf20Sopenharmony_ci job->hw_queue_id = chunk->queue_index; 7018c2ecf20Sopenharmony_ci 7028c2ecf20Sopenharmony_ci cs->jobs_in_queue_cnt[job->hw_queue_id]++; 7038c2ecf20Sopenharmony_ci 7048c2ecf20Sopenharmony_ci list_add_tail(&job->cs_node, &cs->job_list); 7058c2ecf20Sopenharmony_ci 7068c2ecf20Sopenharmony_ci /* 7078c2ecf20Sopenharmony_ci * Increment CS reference. When CS reference is 0, CS is 7088c2ecf20Sopenharmony_ci * done and can be signaled to user and free all its resources 7098c2ecf20Sopenharmony_ci * Only increment for JOB on external or H/W queues, because 7108c2ecf20Sopenharmony_ci * only for those JOBs we get completion 7118c2ecf20Sopenharmony_ci */ 7128c2ecf20Sopenharmony_ci if (job->queue_type == QUEUE_TYPE_EXT || 7138c2ecf20Sopenharmony_ci job->queue_type == QUEUE_TYPE_HW) 7148c2ecf20Sopenharmony_ci cs_get(cs); 7158c2ecf20Sopenharmony_ci 7168c2ecf20Sopenharmony_ci hl_debugfs_add_job(hdev, job); 7178c2ecf20Sopenharmony_ci 7188c2ecf20Sopenharmony_ci rc = cs_parser(hpriv, job); 7198c2ecf20Sopenharmony_ci if (rc) { 7208c2ecf20Sopenharmony_ci hpriv->ctx->cs_counters.parsing_drop_cnt++; 7218c2ecf20Sopenharmony_ci dev_err(hdev->dev, 7228c2ecf20Sopenharmony_ci "Failed to parse JOB %d.%llu.%d, err %d, rejecting the CS\n", 7238c2ecf20Sopenharmony_ci cs->ctx->asid, cs->sequence, job->id, rc); 7248c2ecf20Sopenharmony_ci goto free_cs_object; 7258c2ecf20Sopenharmony_ci } 7268c2ecf20Sopenharmony_ci } 7278c2ecf20Sopenharmony_ci 7288c2ecf20Sopenharmony_ci if (int_queues_only) { 7298c2ecf20Sopenharmony_ci hpriv->ctx->cs_counters.parsing_drop_cnt++; 7308c2ecf20Sopenharmony_ci dev_err(hdev->dev, 7318c2ecf20Sopenharmony_ci "Reject CS %d.%llu because only internal queues jobs are present\n", 7328c2ecf20Sopenharmony_ci cs->ctx->asid, cs->sequence); 7338c2ecf20Sopenharmony_ci rc = -EINVAL; 7348c2ecf20Sopenharmony_ci goto free_cs_object; 7358c2ecf20Sopenharmony_ci } 7368c2ecf20Sopenharmony_ci 7378c2ecf20Sopenharmony_ci rc = hl_hw_queue_schedule_cs(cs); 7388c2ecf20Sopenharmony_ci if (rc) { 7398c2ecf20Sopenharmony_ci if (rc != -EAGAIN) 7408c2ecf20Sopenharmony_ci dev_err(hdev->dev, 7418c2ecf20Sopenharmony_ci "Failed to submit CS %d.%llu to H/W queues, error %d\n", 7428c2ecf20Sopenharmony_ci cs->ctx->asid, cs->sequence, rc); 7438c2ecf20Sopenharmony_ci goto free_cs_object; 7448c2ecf20Sopenharmony_ci } 7458c2ecf20Sopenharmony_ci 7468c2ecf20Sopenharmony_ci rc = HL_CS_STATUS_SUCCESS; 7478c2ecf20Sopenharmony_ci goto put_cs; 7488c2ecf20Sopenharmony_ci 7498c2ecf20Sopenharmony_cirelease_cb: 7508c2ecf20Sopenharmony_ci spin_lock(&cb->lock); 7518c2ecf20Sopenharmony_ci cb->cs_cnt--; 7528c2ecf20Sopenharmony_ci spin_unlock(&cb->lock); 7538c2ecf20Sopenharmony_ci hl_cb_put(cb); 7548c2ecf20Sopenharmony_cifree_cs_object: 7558c2ecf20Sopenharmony_ci cs_rollback(hdev, cs); 7568c2ecf20Sopenharmony_ci *cs_seq = ULLONG_MAX; 7578c2ecf20Sopenharmony_ci /* The path below is both for good and erroneous exits */ 7588c2ecf20Sopenharmony_ciput_cs: 7598c2ecf20Sopenharmony_ci /* We finished with the CS in this function, so put the ref */ 7608c2ecf20Sopenharmony_ci cs_put(cs); 7618c2ecf20Sopenharmony_cifree_cs_chunk_array: 7628c2ecf20Sopenharmony_ci kfree(cs_chunk_array); 7638c2ecf20Sopenharmony_ciout: 7648c2ecf20Sopenharmony_ci return rc; 7658c2ecf20Sopenharmony_ci} 7668c2ecf20Sopenharmony_ci 7678c2ecf20Sopenharmony_cistatic int cs_ioctl_signal_wait(struct hl_fpriv *hpriv, enum hl_cs_type cs_type, 7688c2ecf20Sopenharmony_ci void __user *chunks, u32 num_chunks, 7698c2ecf20Sopenharmony_ci u64 *cs_seq) 7708c2ecf20Sopenharmony_ci{ 7718c2ecf20Sopenharmony_ci struct hl_device *hdev = hpriv->hdev; 7728c2ecf20Sopenharmony_ci struct hl_ctx *ctx = hpriv->ctx; 7738c2ecf20Sopenharmony_ci struct hl_cs_chunk *cs_chunk_array, *chunk; 7748c2ecf20Sopenharmony_ci struct hw_queue_properties *hw_queue_prop; 7758c2ecf20Sopenharmony_ci struct hl_fence *sig_fence = NULL; 7768c2ecf20Sopenharmony_ci struct hl_cs_job *job; 7778c2ecf20Sopenharmony_ci struct hl_cs *cs; 7788c2ecf20Sopenharmony_ci struct hl_cb *cb; 7798c2ecf20Sopenharmony_ci enum hl_queue_type q_type; 7808c2ecf20Sopenharmony_ci u64 *signal_seq_arr = NULL, signal_seq; 7818c2ecf20Sopenharmony_ci u32 size_to_copy, q_idx, signal_seq_arr_len, cb_size; 7828c2ecf20Sopenharmony_ci int rc; 7838c2ecf20Sopenharmony_ci 7848c2ecf20Sopenharmony_ci *cs_seq = ULLONG_MAX; 7858c2ecf20Sopenharmony_ci 7868c2ecf20Sopenharmony_ci if (num_chunks > HL_MAX_JOBS_PER_CS) { 7878c2ecf20Sopenharmony_ci dev_err(hdev->dev, 7888c2ecf20Sopenharmony_ci "Number of chunks can NOT be larger than %d\n", 7898c2ecf20Sopenharmony_ci HL_MAX_JOBS_PER_CS); 7908c2ecf20Sopenharmony_ci rc = -EINVAL; 7918c2ecf20Sopenharmony_ci goto out; 7928c2ecf20Sopenharmony_ci } 7938c2ecf20Sopenharmony_ci 7948c2ecf20Sopenharmony_ci cs_chunk_array = kmalloc_array(num_chunks, sizeof(*cs_chunk_array), 7958c2ecf20Sopenharmony_ci GFP_ATOMIC); 7968c2ecf20Sopenharmony_ci if (!cs_chunk_array) { 7978c2ecf20Sopenharmony_ci rc = -ENOMEM; 7988c2ecf20Sopenharmony_ci goto out; 7998c2ecf20Sopenharmony_ci } 8008c2ecf20Sopenharmony_ci 8018c2ecf20Sopenharmony_ci size_to_copy = num_chunks * sizeof(struct hl_cs_chunk); 8028c2ecf20Sopenharmony_ci if (copy_from_user(cs_chunk_array, chunks, size_to_copy)) { 8038c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Failed to copy cs chunk array from user\n"); 8048c2ecf20Sopenharmony_ci rc = -EFAULT; 8058c2ecf20Sopenharmony_ci goto free_cs_chunk_array; 8068c2ecf20Sopenharmony_ci } 8078c2ecf20Sopenharmony_ci 8088c2ecf20Sopenharmony_ci /* currently it is guaranteed to have only one chunk */ 8098c2ecf20Sopenharmony_ci chunk = &cs_chunk_array[0]; 8108c2ecf20Sopenharmony_ci 8118c2ecf20Sopenharmony_ci if (chunk->queue_index >= hdev->asic_prop.max_queues) { 8128c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Queue index %d is invalid\n", 8138c2ecf20Sopenharmony_ci chunk->queue_index); 8148c2ecf20Sopenharmony_ci rc = -EINVAL; 8158c2ecf20Sopenharmony_ci goto free_cs_chunk_array; 8168c2ecf20Sopenharmony_ci } 8178c2ecf20Sopenharmony_ci 8188c2ecf20Sopenharmony_ci q_idx = chunk->queue_index; 8198c2ecf20Sopenharmony_ci hw_queue_prop = &hdev->asic_prop.hw_queues_props[q_idx]; 8208c2ecf20Sopenharmony_ci q_type = hw_queue_prop->type; 8218c2ecf20Sopenharmony_ci 8228c2ecf20Sopenharmony_ci if ((q_idx >= hdev->asic_prop.max_queues) || 8238c2ecf20Sopenharmony_ci (!hw_queue_prop->supports_sync_stream)) { 8248c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Queue index %d is invalid\n", q_idx); 8258c2ecf20Sopenharmony_ci rc = -EINVAL; 8268c2ecf20Sopenharmony_ci goto free_cs_chunk_array; 8278c2ecf20Sopenharmony_ci } 8288c2ecf20Sopenharmony_ci 8298c2ecf20Sopenharmony_ci if (cs_type == CS_TYPE_WAIT) { 8308c2ecf20Sopenharmony_ci struct hl_cs_compl *sig_waitcs_cmpl; 8318c2ecf20Sopenharmony_ci 8328c2ecf20Sopenharmony_ci signal_seq_arr_len = chunk->num_signal_seq_arr; 8338c2ecf20Sopenharmony_ci 8348c2ecf20Sopenharmony_ci /* currently only one signal seq is supported */ 8358c2ecf20Sopenharmony_ci if (signal_seq_arr_len != 1) { 8368c2ecf20Sopenharmony_ci dev_err(hdev->dev, 8378c2ecf20Sopenharmony_ci "Wait for signal CS supports only one signal CS seq\n"); 8388c2ecf20Sopenharmony_ci rc = -EINVAL; 8398c2ecf20Sopenharmony_ci goto free_cs_chunk_array; 8408c2ecf20Sopenharmony_ci } 8418c2ecf20Sopenharmony_ci 8428c2ecf20Sopenharmony_ci signal_seq_arr = kmalloc_array(signal_seq_arr_len, 8438c2ecf20Sopenharmony_ci sizeof(*signal_seq_arr), 8448c2ecf20Sopenharmony_ci GFP_ATOMIC); 8458c2ecf20Sopenharmony_ci if (!signal_seq_arr) { 8468c2ecf20Sopenharmony_ci rc = -ENOMEM; 8478c2ecf20Sopenharmony_ci goto free_cs_chunk_array; 8488c2ecf20Sopenharmony_ci } 8498c2ecf20Sopenharmony_ci 8508c2ecf20Sopenharmony_ci size_to_copy = chunk->num_signal_seq_arr * 8518c2ecf20Sopenharmony_ci sizeof(*signal_seq_arr); 8528c2ecf20Sopenharmony_ci if (copy_from_user(signal_seq_arr, 8538c2ecf20Sopenharmony_ci u64_to_user_ptr(chunk->signal_seq_arr), 8548c2ecf20Sopenharmony_ci size_to_copy)) { 8558c2ecf20Sopenharmony_ci dev_err(hdev->dev, 8568c2ecf20Sopenharmony_ci "Failed to copy signal seq array from user\n"); 8578c2ecf20Sopenharmony_ci rc = -EFAULT; 8588c2ecf20Sopenharmony_ci goto free_signal_seq_array; 8598c2ecf20Sopenharmony_ci } 8608c2ecf20Sopenharmony_ci 8618c2ecf20Sopenharmony_ci /* currently it is guaranteed to have only one signal seq */ 8628c2ecf20Sopenharmony_ci signal_seq = signal_seq_arr[0]; 8638c2ecf20Sopenharmony_ci sig_fence = hl_ctx_get_fence(ctx, signal_seq); 8648c2ecf20Sopenharmony_ci if (IS_ERR(sig_fence)) { 8658c2ecf20Sopenharmony_ci dev_err(hdev->dev, 8668c2ecf20Sopenharmony_ci "Failed to get signal CS with seq 0x%llx\n", 8678c2ecf20Sopenharmony_ci signal_seq); 8688c2ecf20Sopenharmony_ci rc = PTR_ERR(sig_fence); 8698c2ecf20Sopenharmony_ci goto free_signal_seq_array; 8708c2ecf20Sopenharmony_ci } 8718c2ecf20Sopenharmony_ci 8728c2ecf20Sopenharmony_ci if (!sig_fence) { 8738c2ecf20Sopenharmony_ci /* signal CS already finished */ 8748c2ecf20Sopenharmony_ci rc = 0; 8758c2ecf20Sopenharmony_ci goto free_signal_seq_array; 8768c2ecf20Sopenharmony_ci } 8778c2ecf20Sopenharmony_ci 8788c2ecf20Sopenharmony_ci sig_waitcs_cmpl = 8798c2ecf20Sopenharmony_ci container_of(sig_fence, struct hl_cs_compl, base_fence); 8808c2ecf20Sopenharmony_ci 8818c2ecf20Sopenharmony_ci if (sig_waitcs_cmpl->type != CS_TYPE_SIGNAL) { 8828c2ecf20Sopenharmony_ci dev_err(hdev->dev, 8838c2ecf20Sopenharmony_ci "CS seq 0x%llx is not of a signal CS\n", 8848c2ecf20Sopenharmony_ci signal_seq); 8858c2ecf20Sopenharmony_ci hl_fence_put(sig_fence); 8868c2ecf20Sopenharmony_ci rc = -EINVAL; 8878c2ecf20Sopenharmony_ci goto free_signal_seq_array; 8888c2ecf20Sopenharmony_ci } 8898c2ecf20Sopenharmony_ci 8908c2ecf20Sopenharmony_ci if (completion_done(&sig_fence->completion)) { 8918c2ecf20Sopenharmony_ci /* signal CS already finished */ 8928c2ecf20Sopenharmony_ci hl_fence_put(sig_fence); 8938c2ecf20Sopenharmony_ci rc = 0; 8948c2ecf20Sopenharmony_ci goto free_signal_seq_array; 8958c2ecf20Sopenharmony_ci } 8968c2ecf20Sopenharmony_ci } 8978c2ecf20Sopenharmony_ci 8988c2ecf20Sopenharmony_ci /* increment refcnt for context */ 8998c2ecf20Sopenharmony_ci hl_ctx_get(hdev, ctx); 9008c2ecf20Sopenharmony_ci 9018c2ecf20Sopenharmony_ci rc = allocate_cs(hdev, ctx, cs_type, &cs); 9028c2ecf20Sopenharmony_ci if (rc) { 9038c2ecf20Sopenharmony_ci if (cs_type == CS_TYPE_WAIT) 9048c2ecf20Sopenharmony_ci hl_fence_put(sig_fence); 9058c2ecf20Sopenharmony_ci hl_ctx_put(ctx); 9068c2ecf20Sopenharmony_ci goto free_signal_seq_array; 9078c2ecf20Sopenharmony_ci } 9088c2ecf20Sopenharmony_ci 9098c2ecf20Sopenharmony_ci /* 9108c2ecf20Sopenharmony_ci * Save the signal CS fence for later initialization right before 9118c2ecf20Sopenharmony_ci * hanging the wait CS on the queue. 9128c2ecf20Sopenharmony_ci */ 9138c2ecf20Sopenharmony_ci if (cs->type == CS_TYPE_WAIT) 9148c2ecf20Sopenharmony_ci cs->signal_fence = sig_fence; 9158c2ecf20Sopenharmony_ci 9168c2ecf20Sopenharmony_ci hl_debugfs_add_cs(cs); 9178c2ecf20Sopenharmony_ci 9188c2ecf20Sopenharmony_ci *cs_seq = cs->sequence; 9198c2ecf20Sopenharmony_ci 9208c2ecf20Sopenharmony_ci job = hl_cs_allocate_job(hdev, q_type, true); 9218c2ecf20Sopenharmony_ci if (!job) { 9228c2ecf20Sopenharmony_ci ctx->cs_counters.out_of_mem_drop_cnt++; 9238c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Failed to allocate a new job\n"); 9248c2ecf20Sopenharmony_ci rc = -ENOMEM; 9258c2ecf20Sopenharmony_ci goto put_cs; 9268c2ecf20Sopenharmony_ci } 9278c2ecf20Sopenharmony_ci 9288c2ecf20Sopenharmony_ci if (cs->type == CS_TYPE_WAIT) 9298c2ecf20Sopenharmony_ci cb_size = hdev->asic_funcs->get_wait_cb_size(hdev); 9308c2ecf20Sopenharmony_ci else 9318c2ecf20Sopenharmony_ci cb_size = hdev->asic_funcs->get_signal_cb_size(hdev); 9328c2ecf20Sopenharmony_ci 9338c2ecf20Sopenharmony_ci cb = hl_cb_kernel_create(hdev, cb_size, 9348c2ecf20Sopenharmony_ci q_type == QUEUE_TYPE_HW && hdev->mmu_enable); 9358c2ecf20Sopenharmony_ci if (!cb) { 9368c2ecf20Sopenharmony_ci ctx->cs_counters.out_of_mem_drop_cnt++; 9378c2ecf20Sopenharmony_ci kfree(job); 9388c2ecf20Sopenharmony_ci rc = -EFAULT; 9398c2ecf20Sopenharmony_ci goto put_cs; 9408c2ecf20Sopenharmony_ci } 9418c2ecf20Sopenharmony_ci 9428c2ecf20Sopenharmony_ci job->id = 0; 9438c2ecf20Sopenharmony_ci job->cs = cs; 9448c2ecf20Sopenharmony_ci job->user_cb = cb; 9458c2ecf20Sopenharmony_ci job->user_cb->cs_cnt++; 9468c2ecf20Sopenharmony_ci job->user_cb_size = cb_size; 9478c2ecf20Sopenharmony_ci job->hw_queue_id = q_idx; 9488c2ecf20Sopenharmony_ci 9498c2ecf20Sopenharmony_ci /* 9508c2ecf20Sopenharmony_ci * No need in parsing, user CB is the patched CB. 9518c2ecf20Sopenharmony_ci * We call hl_cb_destroy() out of two reasons - we don't need the CB in 9528c2ecf20Sopenharmony_ci * the CB idr anymore and to decrement its refcount as it was 9538c2ecf20Sopenharmony_ci * incremented inside hl_cb_kernel_create(). 9548c2ecf20Sopenharmony_ci */ 9558c2ecf20Sopenharmony_ci job->patched_cb = job->user_cb; 9568c2ecf20Sopenharmony_ci job->job_cb_size = job->user_cb_size; 9578c2ecf20Sopenharmony_ci hl_cb_destroy(hdev, &hdev->kernel_cb_mgr, cb->id << PAGE_SHIFT); 9588c2ecf20Sopenharmony_ci 9598c2ecf20Sopenharmony_ci cs->jobs_in_queue_cnt[job->hw_queue_id]++; 9608c2ecf20Sopenharmony_ci 9618c2ecf20Sopenharmony_ci list_add_tail(&job->cs_node, &cs->job_list); 9628c2ecf20Sopenharmony_ci 9638c2ecf20Sopenharmony_ci /* increment refcount as for external queues we get completion */ 9648c2ecf20Sopenharmony_ci cs_get(cs); 9658c2ecf20Sopenharmony_ci 9668c2ecf20Sopenharmony_ci hl_debugfs_add_job(hdev, job); 9678c2ecf20Sopenharmony_ci 9688c2ecf20Sopenharmony_ci rc = hl_hw_queue_schedule_cs(cs); 9698c2ecf20Sopenharmony_ci if (rc) { 9708c2ecf20Sopenharmony_ci if (rc != -EAGAIN) 9718c2ecf20Sopenharmony_ci dev_err(hdev->dev, 9728c2ecf20Sopenharmony_ci "Failed to submit CS %d.%llu to H/W queues, error %d\n", 9738c2ecf20Sopenharmony_ci ctx->asid, cs->sequence, rc); 9748c2ecf20Sopenharmony_ci goto free_cs_object; 9758c2ecf20Sopenharmony_ci } 9768c2ecf20Sopenharmony_ci 9778c2ecf20Sopenharmony_ci rc = HL_CS_STATUS_SUCCESS; 9788c2ecf20Sopenharmony_ci goto put_cs; 9798c2ecf20Sopenharmony_ci 9808c2ecf20Sopenharmony_cifree_cs_object: 9818c2ecf20Sopenharmony_ci cs_rollback(hdev, cs); 9828c2ecf20Sopenharmony_ci *cs_seq = ULLONG_MAX; 9838c2ecf20Sopenharmony_ci /* The path below is both for good and erroneous exits */ 9848c2ecf20Sopenharmony_ciput_cs: 9858c2ecf20Sopenharmony_ci /* We finished with the CS in this function, so put the ref */ 9868c2ecf20Sopenharmony_ci cs_put(cs); 9878c2ecf20Sopenharmony_cifree_signal_seq_array: 9888c2ecf20Sopenharmony_ci if (cs_type == CS_TYPE_WAIT) 9898c2ecf20Sopenharmony_ci kfree(signal_seq_arr); 9908c2ecf20Sopenharmony_cifree_cs_chunk_array: 9918c2ecf20Sopenharmony_ci kfree(cs_chunk_array); 9928c2ecf20Sopenharmony_ciout: 9938c2ecf20Sopenharmony_ci return rc; 9948c2ecf20Sopenharmony_ci} 9958c2ecf20Sopenharmony_ci 9968c2ecf20Sopenharmony_ciint hl_cs_ioctl(struct hl_fpriv *hpriv, void *data) 9978c2ecf20Sopenharmony_ci{ 9988c2ecf20Sopenharmony_ci struct hl_device *hdev = hpriv->hdev; 9998c2ecf20Sopenharmony_ci union hl_cs_args *args = data; 10008c2ecf20Sopenharmony_ci struct hl_ctx *ctx = hpriv->ctx; 10018c2ecf20Sopenharmony_ci void __user *chunks_execute, *chunks_restore; 10028c2ecf20Sopenharmony_ci enum hl_cs_type cs_type; 10038c2ecf20Sopenharmony_ci u32 num_chunks_execute, num_chunks_restore, sig_wait_flags; 10048c2ecf20Sopenharmony_ci u64 cs_seq = ULONG_MAX; 10058c2ecf20Sopenharmony_ci int rc, do_ctx_switch; 10068c2ecf20Sopenharmony_ci bool need_soft_reset = false; 10078c2ecf20Sopenharmony_ci 10088c2ecf20Sopenharmony_ci if (hl_device_disabled_or_in_reset(hdev)) { 10098c2ecf20Sopenharmony_ci dev_warn_ratelimited(hdev->dev, 10108c2ecf20Sopenharmony_ci "Device is %s. Can't submit new CS\n", 10118c2ecf20Sopenharmony_ci atomic_read(&hdev->in_reset) ? "in_reset" : "disabled"); 10128c2ecf20Sopenharmony_ci rc = -EBUSY; 10138c2ecf20Sopenharmony_ci goto out; 10148c2ecf20Sopenharmony_ci } 10158c2ecf20Sopenharmony_ci 10168c2ecf20Sopenharmony_ci sig_wait_flags = args->in.cs_flags & HL_CS_FLAGS_SIG_WAIT; 10178c2ecf20Sopenharmony_ci 10188c2ecf20Sopenharmony_ci if (unlikely(sig_wait_flags == HL_CS_FLAGS_SIG_WAIT)) { 10198c2ecf20Sopenharmony_ci dev_err(hdev->dev, 10208c2ecf20Sopenharmony_ci "Signal and wait CS flags are mutually exclusive, context %d\n", 10218c2ecf20Sopenharmony_ci ctx->asid); 10228c2ecf20Sopenharmony_ci rc = -EINVAL; 10238c2ecf20Sopenharmony_ci goto out; 10248c2ecf20Sopenharmony_ci } 10258c2ecf20Sopenharmony_ci 10268c2ecf20Sopenharmony_ci if (unlikely((sig_wait_flags & HL_CS_FLAGS_SIG_WAIT) && 10278c2ecf20Sopenharmony_ci (!hdev->supports_sync_stream))) { 10288c2ecf20Sopenharmony_ci dev_err(hdev->dev, "Sync stream CS is not supported\n"); 10298c2ecf20Sopenharmony_ci rc = -EINVAL; 10308c2ecf20Sopenharmony_ci goto out; 10318c2ecf20Sopenharmony_ci } 10328c2ecf20Sopenharmony_ci 10338c2ecf20Sopenharmony_ci if (args->in.cs_flags & HL_CS_FLAGS_SIGNAL) 10348c2ecf20Sopenharmony_ci cs_type = CS_TYPE_SIGNAL; 10358c2ecf20Sopenharmony_ci else if (args->in.cs_flags & HL_CS_FLAGS_WAIT) 10368c2ecf20Sopenharmony_ci cs_type = CS_TYPE_WAIT; 10378c2ecf20Sopenharmony_ci else 10388c2ecf20Sopenharmony_ci cs_type = CS_TYPE_DEFAULT; 10398c2ecf20Sopenharmony_ci 10408c2ecf20Sopenharmony_ci chunks_execute = (void __user *) (uintptr_t) args->in.chunks_execute; 10418c2ecf20Sopenharmony_ci num_chunks_execute = args->in.num_chunks_execute; 10428c2ecf20Sopenharmony_ci 10438c2ecf20Sopenharmony_ci if (cs_type == CS_TYPE_DEFAULT) { 10448c2ecf20Sopenharmony_ci if (!num_chunks_execute) { 10458c2ecf20Sopenharmony_ci dev_err(hdev->dev, 10468c2ecf20Sopenharmony_ci "Got execute CS with 0 chunks, context %d\n", 10478c2ecf20Sopenharmony_ci ctx->asid); 10488c2ecf20Sopenharmony_ci rc = -EINVAL; 10498c2ecf20Sopenharmony_ci goto out; 10508c2ecf20Sopenharmony_ci } 10518c2ecf20Sopenharmony_ci } else if (num_chunks_execute != 1) { 10528c2ecf20Sopenharmony_ci dev_err(hdev->dev, 10538c2ecf20Sopenharmony_ci "Sync stream CS mandates one chunk only, context %d\n", 10548c2ecf20Sopenharmony_ci ctx->asid); 10558c2ecf20Sopenharmony_ci rc = -EINVAL; 10568c2ecf20Sopenharmony_ci goto out; 10578c2ecf20Sopenharmony_ci } 10588c2ecf20Sopenharmony_ci 10598c2ecf20Sopenharmony_ci do_ctx_switch = atomic_cmpxchg(&ctx->thread_ctx_switch_token, 1, 0); 10608c2ecf20Sopenharmony_ci 10618c2ecf20Sopenharmony_ci if (do_ctx_switch || (args->in.cs_flags & HL_CS_FLAGS_FORCE_RESTORE)) { 10628c2ecf20Sopenharmony_ci long ret; 10638c2ecf20Sopenharmony_ci 10648c2ecf20Sopenharmony_ci chunks_restore = 10658c2ecf20Sopenharmony_ci (void __user *) (uintptr_t) args->in.chunks_restore; 10668c2ecf20Sopenharmony_ci num_chunks_restore = args->in.num_chunks_restore; 10678c2ecf20Sopenharmony_ci 10688c2ecf20Sopenharmony_ci mutex_lock(&hpriv->restore_phase_mutex); 10698c2ecf20Sopenharmony_ci 10708c2ecf20Sopenharmony_ci if (do_ctx_switch) { 10718c2ecf20Sopenharmony_ci rc = hdev->asic_funcs->context_switch(hdev, ctx->asid); 10728c2ecf20Sopenharmony_ci if (rc) { 10738c2ecf20Sopenharmony_ci dev_err_ratelimited(hdev->dev, 10748c2ecf20Sopenharmony_ci "Failed to switch to context %d, rejecting CS! %d\n", 10758c2ecf20Sopenharmony_ci ctx->asid, rc); 10768c2ecf20Sopenharmony_ci /* 10778c2ecf20Sopenharmony_ci * If we timedout, or if the device is not IDLE 10788c2ecf20Sopenharmony_ci * while we want to do context-switch (-EBUSY), 10798c2ecf20Sopenharmony_ci * we need to soft-reset because QMAN is 10808c2ecf20Sopenharmony_ci * probably stuck. However, we can't call to 10818c2ecf20Sopenharmony_ci * reset here directly because of deadlock, so 10828c2ecf20Sopenharmony_ci * need to do it at the very end of this 10838c2ecf20Sopenharmony_ci * function 10848c2ecf20Sopenharmony_ci */ 10858c2ecf20Sopenharmony_ci if ((rc == -ETIMEDOUT) || (rc == -EBUSY)) 10868c2ecf20Sopenharmony_ci need_soft_reset = true; 10878c2ecf20Sopenharmony_ci mutex_unlock(&hpriv->restore_phase_mutex); 10888c2ecf20Sopenharmony_ci goto out; 10898c2ecf20Sopenharmony_ci } 10908c2ecf20Sopenharmony_ci } 10918c2ecf20Sopenharmony_ci 10928c2ecf20Sopenharmony_ci hdev->asic_funcs->restore_phase_topology(hdev); 10938c2ecf20Sopenharmony_ci 10948c2ecf20Sopenharmony_ci if (!num_chunks_restore) { 10958c2ecf20Sopenharmony_ci dev_dbg(hdev->dev, 10968c2ecf20Sopenharmony_ci "Need to run restore phase but restore CS is empty\n"); 10978c2ecf20Sopenharmony_ci rc = 0; 10988c2ecf20Sopenharmony_ci } else { 10998c2ecf20Sopenharmony_ci rc = cs_ioctl_default(hpriv, chunks_restore, 11008c2ecf20Sopenharmony_ci num_chunks_restore, &cs_seq); 11018c2ecf20Sopenharmony_ci } 11028c2ecf20Sopenharmony_ci 11038c2ecf20Sopenharmony_ci mutex_unlock(&hpriv->restore_phase_mutex); 11048c2ecf20Sopenharmony_ci 11058c2ecf20Sopenharmony_ci if (rc) { 11068c2ecf20Sopenharmony_ci dev_err(hdev->dev, 11078c2ecf20Sopenharmony_ci "Failed to submit restore CS for context %d (%d)\n", 11088c2ecf20Sopenharmony_ci ctx->asid, rc); 11098c2ecf20Sopenharmony_ci goto out; 11108c2ecf20Sopenharmony_ci } 11118c2ecf20Sopenharmony_ci 11128c2ecf20Sopenharmony_ci /* Need to wait for restore completion before execution phase */ 11138c2ecf20Sopenharmony_ci if (num_chunks_restore) { 11148c2ecf20Sopenharmony_ci ret = _hl_cs_wait_ioctl(hdev, ctx, 11158c2ecf20Sopenharmony_ci jiffies_to_usecs(hdev->timeout_jiffies), 11168c2ecf20Sopenharmony_ci cs_seq); 11178c2ecf20Sopenharmony_ci if (ret <= 0) { 11188c2ecf20Sopenharmony_ci dev_err(hdev->dev, 11198c2ecf20Sopenharmony_ci "Restore CS for context %d failed to complete %ld\n", 11208c2ecf20Sopenharmony_ci ctx->asid, ret); 11218c2ecf20Sopenharmony_ci rc = -ENOEXEC; 11228c2ecf20Sopenharmony_ci goto out; 11238c2ecf20Sopenharmony_ci } 11248c2ecf20Sopenharmony_ci } 11258c2ecf20Sopenharmony_ci 11268c2ecf20Sopenharmony_ci ctx->thread_ctx_switch_wait_token = 1; 11278c2ecf20Sopenharmony_ci } else if (!ctx->thread_ctx_switch_wait_token) { 11288c2ecf20Sopenharmony_ci u32 tmp; 11298c2ecf20Sopenharmony_ci 11308c2ecf20Sopenharmony_ci rc = hl_poll_timeout_memory(hdev, 11318c2ecf20Sopenharmony_ci &ctx->thread_ctx_switch_wait_token, tmp, (tmp == 1), 11328c2ecf20Sopenharmony_ci 100, jiffies_to_usecs(hdev->timeout_jiffies), false); 11338c2ecf20Sopenharmony_ci 11348c2ecf20Sopenharmony_ci if (rc == -ETIMEDOUT) { 11358c2ecf20Sopenharmony_ci dev_err(hdev->dev, 11368c2ecf20Sopenharmony_ci "context switch phase timeout (%d)\n", tmp); 11378c2ecf20Sopenharmony_ci goto out; 11388c2ecf20Sopenharmony_ci } 11398c2ecf20Sopenharmony_ci } 11408c2ecf20Sopenharmony_ci 11418c2ecf20Sopenharmony_ci if (cs_type == CS_TYPE_DEFAULT) 11428c2ecf20Sopenharmony_ci rc = cs_ioctl_default(hpriv, chunks_execute, num_chunks_execute, 11438c2ecf20Sopenharmony_ci &cs_seq); 11448c2ecf20Sopenharmony_ci else 11458c2ecf20Sopenharmony_ci rc = cs_ioctl_signal_wait(hpriv, cs_type, chunks_execute, 11468c2ecf20Sopenharmony_ci num_chunks_execute, &cs_seq); 11478c2ecf20Sopenharmony_ci 11488c2ecf20Sopenharmony_ciout: 11498c2ecf20Sopenharmony_ci if (rc != -EAGAIN) { 11508c2ecf20Sopenharmony_ci memset(args, 0, sizeof(*args)); 11518c2ecf20Sopenharmony_ci args->out.status = rc; 11528c2ecf20Sopenharmony_ci args->out.seq = cs_seq; 11538c2ecf20Sopenharmony_ci } 11548c2ecf20Sopenharmony_ci 11558c2ecf20Sopenharmony_ci if (((rc == -ETIMEDOUT) || (rc == -EBUSY)) && (need_soft_reset)) 11568c2ecf20Sopenharmony_ci hl_device_reset(hdev, false, false); 11578c2ecf20Sopenharmony_ci 11588c2ecf20Sopenharmony_ci return rc; 11598c2ecf20Sopenharmony_ci} 11608c2ecf20Sopenharmony_ci 11618c2ecf20Sopenharmony_cistatic long _hl_cs_wait_ioctl(struct hl_device *hdev, 11628c2ecf20Sopenharmony_ci struct hl_ctx *ctx, u64 timeout_us, u64 seq) 11638c2ecf20Sopenharmony_ci{ 11648c2ecf20Sopenharmony_ci struct hl_fence *fence; 11658c2ecf20Sopenharmony_ci unsigned long timeout; 11668c2ecf20Sopenharmony_ci long rc; 11678c2ecf20Sopenharmony_ci 11688c2ecf20Sopenharmony_ci if (timeout_us == MAX_SCHEDULE_TIMEOUT) 11698c2ecf20Sopenharmony_ci timeout = timeout_us; 11708c2ecf20Sopenharmony_ci else 11718c2ecf20Sopenharmony_ci timeout = usecs_to_jiffies(timeout_us); 11728c2ecf20Sopenharmony_ci 11738c2ecf20Sopenharmony_ci hl_ctx_get(hdev, ctx); 11748c2ecf20Sopenharmony_ci 11758c2ecf20Sopenharmony_ci fence = hl_ctx_get_fence(ctx, seq); 11768c2ecf20Sopenharmony_ci if (IS_ERR(fence)) { 11778c2ecf20Sopenharmony_ci rc = PTR_ERR(fence); 11788c2ecf20Sopenharmony_ci if (rc == -EINVAL) 11798c2ecf20Sopenharmony_ci dev_notice_ratelimited(hdev->dev, 11808c2ecf20Sopenharmony_ci "Can't wait on CS %llu because current CS is at seq %llu\n", 11818c2ecf20Sopenharmony_ci seq, ctx->cs_sequence); 11828c2ecf20Sopenharmony_ci } else if (fence) { 11838c2ecf20Sopenharmony_ci if (!timeout_us) 11848c2ecf20Sopenharmony_ci rc = completion_done(&fence->completion); 11858c2ecf20Sopenharmony_ci else 11868c2ecf20Sopenharmony_ci rc = wait_for_completion_interruptible_timeout( 11878c2ecf20Sopenharmony_ci &fence->completion, timeout); 11888c2ecf20Sopenharmony_ci 11898c2ecf20Sopenharmony_ci if (fence->error == -ETIMEDOUT) 11908c2ecf20Sopenharmony_ci rc = -ETIMEDOUT; 11918c2ecf20Sopenharmony_ci else if (fence->error == -EIO) 11928c2ecf20Sopenharmony_ci rc = -EIO; 11938c2ecf20Sopenharmony_ci 11948c2ecf20Sopenharmony_ci hl_fence_put(fence); 11958c2ecf20Sopenharmony_ci } else { 11968c2ecf20Sopenharmony_ci dev_dbg(hdev->dev, 11978c2ecf20Sopenharmony_ci "Can't wait on seq %llu because current CS is at seq %llu (Fence is gone)\n", 11988c2ecf20Sopenharmony_ci seq, ctx->cs_sequence); 11998c2ecf20Sopenharmony_ci rc = 1; 12008c2ecf20Sopenharmony_ci } 12018c2ecf20Sopenharmony_ci 12028c2ecf20Sopenharmony_ci hl_ctx_put(ctx); 12038c2ecf20Sopenharmony_ci 12048c2ecf20Sopenharmony_ci return rc; 12058c2ecf20Sopenharmony_ci} 12068c2ecf20Sopenharmony_ci 12078c2ecf20Sopenharmony_ciint hl_cs_wait_ioctl(struct hl_fpriv *hpriv, void *data) 12088c2ecf20Sopenharmony_ci{ 12098c2ecf20Sopenharmony_ci struct hl_device *hdev = hpriv->hdev; 12108c2ecf20Sopenharmony_ci union hl_wait_cs_args *args = data; 12118c2ecf20Sopenharmony_ci u64 seq = args->in.seq; 12128c2ecf20Sopenharmony_ci long rc; 12138c2ecf20Sopenharmony_ci 12148c2ecf20Sopenharmony_ci rc = _hl_cs_wait_ioctl(hdev, hpriv->ctx, args->in.timeout_us, seq); 12158c2ecf20Sopenharmony_ci 12168c2ecf20Sopenharmony_ci memset(args, 0, sizeof(*args)); 12178c2ecf20Sopenharmony_ci 12188c2ecf20Sopenharmony_ci if (rc < 0) { 12198c2ecf20Sopenharmony_ci if (rc == -ERESTARTSYS) { 12208c2ecf20Sopenharmony_ci dev_err_ratelimited(hdev->dev, 12218c2ecf20Sopenharmony_ci "user process got signal while waiting for CS handle %llu\n", 12228c2ecf20Sopenharmony_ci seq); 12238c2ecf20Sopenharmony_ci args->out.status = HL_WAIT_CS_STATUS_INTERRUPTED; 12248c2ecf20Sopenharmony_ci rc = -EINTR; 12258c2ecf20Sopenharmony_ci } else if (rc == -ETIMEDOUT) { 12268c2ecf20Sopenharmony_ci dev_err_ratelimited(hdev->dev, 12278c2ecf20Sopenharmony_ci "CS %llu has timed-out while user process is waiting for it\n", 12288c2ecf20Sopenharmony_ci seq); 12298c2ecf20Sopenharmony_ci args->out.status = HL_WAIT_CS_STATUS_TIMEDOUT; 12308c2ecf20Sopenharmony_ci } else if (rc == -EIO) { 12318c2ecf20Sopenharmony_ci dev_err_ratelimited(hdev->dev, 12328c2ecf20Sopenharmony_ci "CS %llu has been aborted while user process is waiting for it\n", 12338c2ecf20Sopenharmony_ci seq); 12348c2ecf20Sopenharmony_ci args->out.status = HL_WAIT_CS_STATUS_ABORTED; 12358c2ecf20Sopenharmony_ci } 12368c2ecf20Sopenharmony_ci return rc; 12378c2ecf20Sopenharmony_ci } 12388c2ecf20Sopenharmony_ci 12398c2ecf20Sopenharmony_ci if (rc == 0) 12408c2ecf20Sopenharmony_ci args->out.status = HL_WAIT_CS_STATUS_BUSY; 12418c2ecf20Sopenharmony_ci else 12428c2ecf20Sopenharmony_ci args->out.status = HL_WAIT_CS_STATUS_COMPLETED; 12438c2ecf20Sopenharmony_ci 12448c2ecf20Sopenharmony_ci return 0; 12458c2ecf20Sopenharmony_ci} 1246