162306a36Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0 262306a36Sopenharmony_ci/* Copyright(c) 2019 Intel Corporation. All rights rsvd. */ 362306a36Sopenharmony_ci#include <linux/init.h> 462306a36Sopenharmony_ci#include <linux/kernel.h> 562306a36Sopenharmony_ci#include <linux/module.h> 662306a36Sopenharmony_ci#include <linux/pci.h> 762306a36Sopenharmony_ci#include <linux/io-64-nonatomic-lo-hi.h> 862306a36Sopenharmony_ci#include <linux/dmaengine.h> 962306a36Sopenharmony_ci#include <linux/delay.h> 1062306a36Sopenharmony_ci#include <linux/iommu.h> 1162306a36Sopenharmony_ci#include <linux/sched/mm.h> 1262306a36Sopenharmony_ci#include <uapi/linux/idxd.h> 1362306a36Sopenharmony_ci#include "../dmaengine.h" 1462306a36Sopenharmony_ci#include "idxd.h" 1562306a36Sopenharmony_ci#include "registers.h" 1662306a36Sopenharmony_ci 1762306a36Sopenharmony_cienum irq_work_type { 1862306a36Sopenharmony_ci IRQ_WORK_NORMAL = 0, 1962306a36Sopenharmony_ci IRQ_WORK_PROCESS_FAULT, 2062306a36Sopenharmony_ci}; 2162306a36Sopenharmony_ci 2262306a36Sopenharmony_cistruct idxd_resubmit { 2362306a36Sopenharmony_ci struct work_struct work; 2462306a36Sopenharmony_ci struct idxd_desc *desc; 2562306a36Sopenharmony_ci}; 2662306a36Sopenharmony_ci 2762306a36Sopenharmony_cistruct idxd_int_handle_revoke { 2862306a36Sopenharmony_ci struct work_struct work; 2962306a36Sopenharmony_ci struct idxd_device *idxd; 3062306a36Sopenharmony_ci}; 3162306a36Sopenharmony_ci 3262306a36Sopenharmony_cistatic void idxd_device_reinit(struct work_struct *work) 3362306a36Sopenharmony_ci{ 3462306a36Sopenharmony_ci struct idxd_device *idxd = container_of(work, struct idxd_device, work); 3562306a36Sopenharmony_ci struct device *dev = &idxd->pdev->dev; 3662306a36Sopenharmony_ci int rc, i; 3762306a36Sopenharmony_ci 3862306a36Sopenharmony_ci idxd_device_reset(idxd); 3962306a36Sopenharmony_ci rc = idxd_device_config(idxd); 4062306a36Sopenharmony_ci if (rc < 0) 4162306a36Sopenharmony_ci goto out; 4262306a36Sopenharmony_ci 4362306a36Sopenharmony_ci rc = idxd_device_enable(idxd); 4462306a36Sopenharmony_ci if (rc < 0) 4562306a36Sopenharmony_ci goto out; 4662306a36Sopenharmony_ci 4762306a36Sopenharmony_ci for (i = 0; i < idxd->max_wqs; i++) { 4862306a36Sopenharmony_ci if (test_bit(i, idxd->wq_enable_map)) { 4962306a36Sopenharmony_ci struct idxd_wq *wq = idxd->wqs[i]; 5062306a36Sopenharmony_ci 5162306a36Sopenharmony_ci rc = idxd_wq_enable(wq); 5262306a36Sopenharmony_ci if (rc < 0) { 5362306a36Sopenharmony_ci clear_bit(i, idxd->wq_enable_map); 5462306a36Sopenharmony_ci dev_warn(dev, "Unable to re-enable wq %s\n", 5562306a36Sopenharmony_ci dev_name(wq_confdev(wq))); 5662306a36Sopenharmony_ci } 5762306a36Sopenharmony_ci } 5862306a36Sopenharmony_ci } 5962306a36Sopenharmony_ci 6062306a36Sopenharmony_ci return; 6162306a36Sopenharmony_ci 6262306a36Sopenharmony_ci out: 6362306a36Sopenharmony_ci idxd_device_clear_state(idxd); 6462306a36Sopenharmony_ci} 6562306a36Sopenharmony_ci 6662306a36Sopenharmony_ci/* 6762306a36Sopenharmony_ci * The function sends a drain descriptor for the interrupt handle. The drain ensures 6862306a36Sopenharmony_ci * all descriptors with this interrupt handle is flushed and the interrupt 6962306a36Sopenharmony_ci * will allow the cleanup of the outstanding descriptors. 7062306a36Sopenharmony_ci */ 7162306a36Sopenharmony_cistatic void idxd_int_handle_revoke_drain(struct idxd_irq_entry *ie) 7262306a36Sopenharmony_ci{ 7362306a36Sopenharmony_ci struct idxd_wq *wq = ie_to_wq(ie); 7462306a36Sopenharmony_ci struct idxd_device *idxd = wq->idxd; 7562306a36Sopenharmony_ci struct device *dev = &idxd->pdev->dev; 7662306a36Sopenharmony_ci struct dsa_hw_desc desc = {}; 7762306a36Sopenharmony_ci void __iomem *portal; 7862306a36Sopenharmony_ci int rc; 7962306a36Sopenharmony_ci 8062306a36Sopenharmony_ci /* Issue a simple drain operation with interrupt but no completion record */ 8162306a36Sopenharmony_ci desc.flags = IDXD_OP_FLAG_RCI; 8262306a36Sopenharmony_ci desc.opcode = DSA_OPCODE_DRAIN; 8362306a36Sopenharmony_ci desc.priv = 1; 8462306a36Sopenharmony_ci 8562306a36Sopenharmony_ci if (ie->pasid != IOMMU_PASID_INVALID) 8662306a36Sopenharmony_ci desc.pasid = ie->pasid; 8762306a36Sopenharmony_ci desc.int_handle = ie->int_handle; 8862306a36Sopenharmony_ci portal = idxd_wq_portal_addr(wq); 8962306a36Sopenharmony_ci 9062306a36Sopenharmony_ci /* 9162306a36Sopenharmony_ci * The wmb() makes sure that the descriptor is all there before we 9262306a36Sopenharmony_ci * issue. 9362306a36Sopenharmony_ci */ 9462306a36Sopenharmony_ci wmb(); 9562306a36Sopenharmony_ci if (wq_dedicated(wq)) { 9662306a36Sopenharmony_ci iosubmit_cmds512(portal, &desc, 1); 9762306a36Sopenharmony_ci } else { 9862306a36Sopenharmony_ci rc = idxd_enqcmds(wq, portal, &desc); 9962306a36Sopenharmony_ci /* This should not fail unless hardware failed. */ 10062306a36Sopenharmony_ci if (rc < 0) 10162306a36Sopenharmony_ci dev_warn(dev, "Failed to submit drain desc on wq %d\n", wq->id); 10262306a36Sopenharmony_ci } 10362306a36Sopenharmony_ci} 10462306a36Sopenharmony_ci 10562306a36Sopenharmony_cistatic void idxd_abort_invalid_int_handle_descs(struct idxd_irq_entry *ie) 10662306a36Sopenharmony_ci{ 10762306a36Sopenharmony_ci LIST_HEAD(flist); 10862306a36Sopenharmony_ci struct idxd_desc *d, *t; 10962306a36Sopenharmony_ci struct llist_node *head; 11062306a36Sopenharmony_ci 11162306a36Sopenharmony_ci spin_lock(&ie->list_lock); 11262306a36Sopenharmony_ci head = llist_del_all(&ie->pending_llist); 11362306a36Sopenharmony_ci if (head) { 11462306a36Sopenharmony_ci llist_for_each_entry_safe(d, t, head, llnode) 11562306a36Sopenharmony_ci list_add_tail(&d->list, &ie->work_list); 11662306a36Sopenharmony_ci } 11762306a36Sopenharmony_ci 11862306a36Sopenharmony_ci list_for_each_entry_safe(d, t, &ie->work_list, list) { 11962306a36Sopenharmony_ci if (d->completion->status == DSA_COMP_INT_HANDLE_INVAL) 12062306a36Sopenharmony_ci list_move_tail(&d->list, &flist); 12162306a36Sopenharmony_ci } 12262306a36Sopenharmony_ci spin_unlock(&ie->list_lock); 12362306a36Sopenharmony_ci 12462306a36Sopenharmony_ci list_for_each_entry_safe(d, t, &flist, list) { 12562306a36Sopenharmony_ci list_del(&d->list); 12662306a36Sopenharmony_ci idxd_dma_complete_txd(d, IDXD_COMPLETE_ABORT, true); 12762306a36Sopenharmony_ci } 12862306a36Sopenharmony_ci} 12962306a36Sopenharmony_ci 13062306a36Sopenharmony_cistatic void idxd_int_handle_revoke(struct work_struct *work) 13162306a36Sopenharmony_ci{ 13262306a36Sopenharmony_ci struct idxd_int_handle_revoke *revoke = 13362306a36Sopenharmony_ci container_of(work, struct idxd_int_handle_revoke, work); 13462306a36Sopenharmony_ci struct idxd_device *idxd = revoke->idxd; 13562306a36Sopenharmony_ci struct pci_dev *pdev = idxd->pdev; 13662306a36Sopenharmony_ci struct device *dev = &pdev->dev; 13762306a36Sopenharmony_ci int i, new_handle, rc; 13862306a36Sopenharmony_ci 13962306a36Sopenharmony_ci if (!idxd->request_int_handles) { 14062306a36Sopenharmony_ci kfree(revoke); 14162306a36Sopenharmony_ci dev_warn(dev, "Unexpected int handle refresh interrupt.\n"); 14262306a36Sopenharmony_ci return; 14362306a36Sopenharmony_ci } 14462306a36Sopenharmony_ci 14562306a36Sopenharmony_ci /* 14662306a36Sopenharmony_ci * The loop attempts to acquire new interrupt handle for all interrupt 14762306a36Sopenharmony_ci * vectors that supports a handle. If a new interrupt handle is acquired and the 14862306a36Sopenharmony_ci * wq is kernel type, the driver will kill the percpu_ref to pause all 14962306a36Sopenharmony_ci * ongoing descriptor submissions. The interrupt handle is then changed. 15062306a36Sopenharmony_ci * After change, the percpu_ref is revived and all the pending submissions 15162306a36Sopenharmony_ci * are woken to try again. A drain is sent to for the interrupt handle 15262306a36Sopenharmony_ci * at the end to make sure all invalid int handle descriptors are processed. 15362306a36Sopenharmony_ci */ 15462306a36Sopenharmony_ci for (i = 1; i < idxd->irq_cnt; i++) { 15562306a36Sopenharmony_ci struct idxd_irq_entry *ie = idxd_get_ie(idxd, i); 15662306a36Sopenharmony_ci struct idxd_wq *wq = ie_to_wq(ie); 15762306a36Sopenharmony_ci 15862306a36Sopenharmony_ci if (ie->int_handle == INVALID_INT_HANDLE) 15962306a36Sopenharmony_ci continue; 16062306a36Sopenharmony_ci 16162306a36Sopenharmony_ci rc = idxd_device_request_int_handle(idxd, i, &new_handle, IDXD_IRQ_MSIX); 16262306a36Sopenharmony_ci if (rc < 0) { 16362306a36Sopenharmony_ci dev_warn(dev, "get int handle %d failed: %d\n", i, rc); 16462306a36Sopenharmony_ci /* 16562306a36Sopenharmony_ci * Failed to acquire new interrupt handle. Kill the WQ 16662306a36Sopenharmony_ci * and release all the pending submitters. The submitters will 16762306a36Sopenharmony_ci * get error return code and handle appropriately. 16862306a36Sopenharmony_ci */ 16962306a36Sopenharmony_ci ie->int_handle = INVALID_INT_HANDLE; 17062306a36Sopenharmony_ci idxd_wq_quiesce(wq); 17162306a36Sopenharmony_ci idxd_abort_invalid_int_handle_descs(ie); 17262306a36Sopenharmony_ci continue; 17362306a36Sopenharmony_ci } 17462306a36Sopenharmony_ci 17562306a36Sopenharmony_ci /* No change in interrupt handle, nothing needs to be done */ 17662306a36Sopenharmony_ci if (ie->int_handle == new_handle) 17762306a36Sopenharmony_ci continue; 17862306a36Sopenharmony_ci 17962306a36Sopenharmony_ci if (wq->state != IDXD_WQ_ENABLED || wq->type != IDXD_WQT_KERNEL) { 18062306a36Sopenharmony_ci /* 18162306a36Sopenharmony_ci * All the MSIX interrupts are allocated at once during probe. 18262306a36Sopenharmony_ci * Therefore we need to update all interrupts even if the WQ 18362306a36Sopenharmony_ci * isn't supporting interrupt operations. 18462306a36Sopenharmony_ci */ 18562306a36Sopenharmony_ci ie->int_handle = new_handle; 18662306a36Sopenharmony_ci continue; 18762306a36Sopenharmony_ci } 18862306a36Sopenharmony_ci 18962306a36Sopenharmony_ci mutex_lock(&wq->wq_lock); 19062306a36Sopenharmony_ci reinit_completion(&wq->wq_resurrect); 19162306a36Sopenharmony_ci 19262306a36Sopenharmony_ci /* Kill percpu_ref to pause additional descriptor submissions */ 19362306a36Sopenharmony_ci percpu_ref_kill(&wq->wq_active); 19462306a36Sopenharmony_ci 19562306a36Sopenharmony_ci /* Wait for all submitters quiesce before we change interrupt handle */ 19662306a36Sopenharmony_ci wait_for_completion(&wq->wq_dead); 19762306a36Sopenharmony_ci 19862306a36Sopenharmony_ci ie->int_handle = new_handle; 19962306a36Sopenharmony_ci 20062306a36Sopenharmony_ci /* Revive percpu ref and wake up all the waiting submitters */ 20162306a36Sopenharmony_ci percpu_ref_reinit(&wq->wq_active); 20262306a36Sopenharmony_ci complete_all(&wq->wq_resurrect); 20362306a36Sopenharmony_ci mutex_unlock(&wq->wq_lock); 20462306a36Sopenharmony_ci 20562306a36Sopenharmony_ci /* 20662306a36Sopenharmony_ci * The delay here is to wait for all possible MOVDIR64B that 20762306a36Sopenharmony_ci * are issued before percpu_ref_kill() has happened to have 20862306a36Sopenharmony_ci * reached the PCIe domain before the drain is issued. The driver 20962306a36Sopenharmony_ci * needs to ensure that the drain descriptor issued does not pass 21062306a36Sopenharmony_ci * all the other issued descriptors that contain the invalid 21162306a36Sopenharmony_ci * interrupt handle in order to ensure that the drain descriptor 21262306a36Sopenharmony_ci * interrupt will allow the cleanup of all the descriptors with 21362306a36Sopenharmony_ci * invalid interrupt handle. 21462306a36Sopenharmony_ci */ 21562306a36Sopenharmony_ci if (wq_dedicated(wq)) 21662306a36Sopenharmony_ci udelay(100); 21762306a36Sopenharmony_ci idxd_int_handle_revoke_drain(ie); 21862306a36Sopenharmony_ci } 21962306a36Sopenharmony_ci kfree(revoke); 22062306a36Sopenharmony_ci} 22162306a36Sopenharmony_ci 22262306a36Sopenharmony_cistatic void idxd_evl_fault_work(struct work_struct *work) 22362306a36Sopenharmony_ci{ 22462306a36Sopenharmony_ci struct idxd_evl_fault *fault = container_of(work, struct idxd_evl_fault, work); 22562306a36Sopenharmony_ci struct idxd_wq *wq = fault->wq; 22662306a36Sopenharmony_ci struct idxd_device *idxd = wq->idxd; 22762306a36Sopenharmony_ci struct device *dev = &idxd->pdev->dev; 22862306a36Sopenharmony_ci struct idxd_evl *evl = idxd->evl; 22962306a36Sopenharmony_ci struct __evl_entry *entry_head = fault->entry; 23062306a36Sopenharmony_ci void *cr = (void *)entry_head + idxd->data->evl_cr_off; 23162306a36Sopenharmony_ci int cr_size = idxd->data->compl_size; 23262306a36Sopenharmony_ci u8 *status = (u8 *)cr + idxd->data->cr_status_off; 23362306a36Sopenharmony_ci u8 *result = (u8 *)cr + idxd->data->cr_result_off; 23462306a36Sopenharmony_ci int copied, copy_size; 23562306a36Sopenharmony_ci bool *bf; 23662306a36Sopenharmony_ci 23762306a36Sopenharmony_ci switch (fault->status) { 23862306a36Sopenharmony_ci case DSA_COMP_CRA_XLAT: 23962306a36Sopenharmony_ci if (entry_head->batch && entry_head->first_err_in_batch) 24062306a36Sopenharmony_ci evl->batch_fail[entry_head->batch_id] = false; 24162306a36Sopenharmony_ci 24262306a36Sopenharmony_ci copy_size = cr_size; 24362306a36Sopenharmony_ci idxd_user_counter_increment(wq, entry_head->pasid, COUNTER_FAULTS); 24462306a36Sopenharmony_ci break; 24562306a36Sopenharmony_ci case DSA_COMP_BATCH_EVL_ERR: 24662306a36Sopenharmony_ci bf = &evl->batch_fail[entry_head->batch_id]; 24762306a36Sopenharmony_ci 24862306a36Sopenharmony_ci copy_size = entry_head->rcr || *bf ? cr_size : 0; 24962306a36Sopenharmony_ci if (*bf) { 25062306a36Sopenharmony_ci if (*status == DSA_COMP_SUCCESS) 25162306a36Sopenharmony_ci *status = DSA_COMP_BATCH_FAIL; 25262306a36Sopenharmony_ci *result = 1; 25362306a36Sopenharmony_ci *bf = false; 25462306a36Sopenharmony_ci } 25562306a36Sopenharmony_ci idxd_user_counter_increment(wq, entry_head->pasid, COUNTER_FAULTS); 25662306a36Sopenharmony_ci break; 25762306a36Sopenharmony_ci case DSA_COMP_DRAIN_EVL: 25862306a36Sopenharmony_ci copy_size = cr_size; 25962306a36Sopenharmony_ci break; 26062306a36Sopenharmony_ci default: 26162306a36Sopenharmony_ci copy_size = 0; 26262306a36Sopenharmony_ci dev_dbg_ratelimited(dev, "Unrecognized error code: %#x\n", fault->status); 26362306a36Sopenharmony_ci break; 26462306a36Sopenharmony_ci } 26562306a36Sopenharmony_ci 26662306a36Sopenharmony_ci if (copy_size == 0) 26762306a36Sopenharmony_ci return; 26862306a36Sopenharmony_ci 26962306a36Sopenharmony_ci /* 27062306a36Sopenharmony_ci * Copy completion record to fault_addr in user address space 27162306a36Sopenharmony_ci * that is found by wq and PASID. 27262306a36Sopenharmony_ci */ 27362306a36Sopenharmony_ci copied = idxd_copy_cr(wq, entry_head->pasid, entry_head->fault_addr, 27462306a36Sopenharmony_ci cr, copy_size); 27562306a36Sopenharmony_ci /* 27662306a36Sopenharmony_ci * The task that triggered the page fault is unknown currently 27762306a36Sopenharmony_ci * because multiple threads may share the user address 27862306a36Sopenharmony_ci * space or the task exits already before this fault. 27962306a36Sopenharmony_ci * So if the copy fails, SIGSEGV can not be sent to the task. 28062306a36Sopenharmony_ci * Just print an error for the failure. The user application 28162306a36Sopenharmony_ci * waiting for the completion record will time out on this 28262306a36Sopenharmony_ci * failure. 28362306a36Sopenharmony_ci */ 28462306a36Sopenharmony_ci switch (fault->status) { 28562306a36Sopenharmony_ci case DSA_COMP_CRA_XLAT: 28662306a36Sopenharmony_ci if (copied != copy_size) { 28762306a36Sopenharmony_ci idxd_user_counter_increment(wq, entry_head->pasid, COUNTER_FAULT_FAILS); 28862306a36Sopenharmony_ci dev_dbg_ratelimited(dev, "Failed to write to completion record: (%d:%d)\n", 28962306a36Sopenharmony_ci copy_size, copied); 29062306a36Sopenharmony_ci if (entry_head->batch) 29162306a36Sopenharmony_ci evl->batch_fail[entry_head->batch_id] = true; 29262306a36Sopenharmony_ci } 29362306a36Sopenharmony_ci break; 29462306a36Sopenharmony_ci case DSA_COMP_BATCH_EVL_ERR: 29562306a36Sopenharmony_ci if (copied != copy_size) { 29662306a36Sopenharmony_ci idxd_user_counter_increment(wq, entry_head->pasid, COUNTER_FAULT_FAILS); 29762306a36Sopenharmony_ci dev_dbg_ratelimited(dev, "Failed to write to batch completion record: (%d:%d)\n", 29862306a36Sopenharmony_ci copy_size, copied); 29962306a36Sopenharmony_ci } 30062306a36Sopenharmony_ci break; 30162306a36Sopenharmony_ci case DSA_COMP_DRAIN_EVL: 30262306a36Sopenharmony_ci if (copied != copy_size) 30362306a36Sopenharmony_ci dev_dbg_ratelimited(dev, "Failed to write to drain completion record: (%d:%d)\n", 30462306a36Sopenharmony_ci copy_size, copied); 30562306a36Sopenharmony_ci break; 30662306a36Sopenharmony_ci } 30762306a36Sopenharmony_ci 30862306a36Sopenharmony_ci kmem_cache_free(idxd->evl_cache, fault); 30962306a36Sopenharmony_ci} 31062306a36Sopenharmony_ci 31162306a36Sopenharmony_cistatic void process_evl_entry(struct idxd_device *idxd, 31262306a36Sopenharmony_ci struct __evl_entry *entry_head, unsigned int index) 31362306a36Sopenharmony_ci{ 31462306a36Sopenharmony_ci struct device *dev = &idxd->pdev->dev; 31562306a36Sopenharmony_ci struct idxd_evl *evl = idxd->evl; 31662306a36Sopenharmony_ci u8 status; 31762306a36Sopenharmony_ci 31862306a36Sopenharmony_ci if (test_bit(index, evl->bmap)) { 31962306a36Sopenharmony_ci clear_bit(index, evl->bmap); 32062306a36Sopenharmony_ci } else { 32162306a36Sopenharmony_ci status = DSA_COMP_STATUS(entry_head->error); 32262306a36Sopenharmony_ci 32362306a36Sopenharmony_ci if (status == DSA_COMP_CRA_XLAT || status == DSA_COMP_DRAIN_EVL || 32462306a36Sopenharmony_ci status == DSA_COMP_BATCH_EVL_ERR) { 32562306a36Sopenharmony_ci struct idxd_evl_fault *fault; 32662306a36Sopenharmony_ci int ent_size = evl_ent_size(idxd); 32762306a36Sopenharmony_ci 32862306a36Sopenharmony_ci if (entry_head->rci) 32962306a36Sopenharmony_ci dev_dbg(dev, "Completion Int Req set, ignoring!\n"); 33062306a36Sopenharmony_ci 33162306a36Sopenharmony_ci if (!entry_head->rcr && status == DSA_COMP_DRAIN_EVL) 33262306a36Sopenharmony_ci return; 33362306a36Sopenharmony_ci 33462306a36Sopenharmony_ci fault = kmem_cache_alloc(idxd->evl_cache, GFP_ATOMIC); 33562306a36Sopenharmony_ci if (fault) { 33662306a36Sopenharmony_ci struct idxd_wq *wq = idxd->wqs[entry_head->wq_idx]; 33762306a36Sopenharmony_ci 33862306a36Sopenharmony_ci fault->wq = wq; 33962306a36Sopenharmony_ci fault->status = status; 34062306a36Sopenharmony_ci memcpy(&fault->entry, entry_head, ent_size); 34162306a36Sopenharmony_ci INIT_WORK(&fault->work, idxd_evl_fault_work); 34262306a36Sopenharmony_ci queue_work(wq->wq, &fault->work); 34362306a36Sopenharmony_ci } else { 34462306a36Sopenharmony_ci dev_warn(dev, "Failed to service fault work.\n"); 34562306a36Sopenharmony_ci } 34662306a36Sopenharmony_ci } else { 34762306a36Sopenharmony_ci dev_warn_ratelimited(dev, "Device error %#x operation: %#x fault addr: %#llx\n", 34862306a36Sopenharmony_ci status, entry_head->operation, 34962306a36Sopenharmony_ci entry_head->fault_addr); 35062306a36Sopenharmony_ci } 35162306a36Sopenharmony_ci } 35262306a36Sopenharmony_ci} 35362306a36Sopenharmony_ci 35462306a36Sopenharmony_cistatic void process_evl_entries(struct idxd_device *idxd) 35562306a36Sopenharmony_ci{ 35662306a36Sopenharmony_ci union evl_status_reg evl_status; 35762306a36Sopenharmony_ci unsigned int h, t; 35862306a36Sopenharmony_ci struct idxd_evl *evl = idxd->evl; 35962306a36Sopenharmony_ci struct __evl_entry *entry_head; 36062306a36Sopenharmony_ci unsigned int ent_size = evl_ent_size(idxd); 36162306a36Sopenharmony_ci u32 size; 36262306a36Sopenharmony_ci 36362306a36Sopenharmony_ci evl_status.bits = 0; 36462306a36Sopenharmony_ci evl_status.int_pending = 1; 36562306a36Sopenharmony_ci 36662306a36Sopenharmony_ci spin_lock(&evl->lock); 36762306a36Sopenharmony_ci /* Clear interrupt pending bit */ 36862306a36Sopenharmony_ci iowrite32(evl_status.bits_upper32, 36962306a36Sopenharmony_ci idxd->reg_base + IDXD_EVLSTATUS_OFFSET + sizeof(u32)); 37062306a36Sopenharmony_ci evl_status.bits = ioread64(idxd->reg_base + IDXD_EVLSTATUS_OFFSET); 37162306a36Sopenharmony_ci t = evl_status.tail; 37262306a36Sopenharmony_ci h = evl_status.head; 37362306a36Sopenharmony_ci size = idxd->evl->size; 37462306a36Sopenharmony_ci 37562306a36Sopenharmony_ci while (h != t) { 37662306a36Sopenharmony_ci entry_head = (struct __evl_entry *)(evl->log + (h * ent_size)); 37762306a36Sopenharmony_ci process_evl_entry(idxd, entry_head, h); 37862306a36Sopenharmony_ci h = (h + 1) % size; 37962306a36Sopenharmony_ci } 38062306a36Sopenharmony_ci 38162306a36Sopenharmony_ci evl_status.head = h; 38262306a36Sopenharmony_ci iowrite32(evl_status.bits_lower32, idxd->reg_base + IDXD_EVLSTATUS_OFFSET); 38362306a36Sopenharmony_ci spin_unlock(&evl->lock); 38462306a36Sopenharmony_ci} 38562306a36Sopenharmony_ci 38662306a36Sopenharmony_ciirqreturn_t idxd_misc_thread(int vec, void *data) 38762306a36Sopenharmony_ci{ 38862306a36Sopenharmony_ci struct idxd_irq_entry *irq_entry = data; 38962306a36Sopenharmony_ci struct idxd_device *idxd = ie_to_idxd(irq_entry); 39062306a36Sopenharmony_ci struct device *dev = &idxd->pdev->dev; 39162306a36Sopenharmony_ci union gensts_reg gensts; 39262306a36Sopenharmony_ci u32 val = 0; 39362306a36Sopenharmony_ci int i; 39462306a36Sopenharmony_ci bool err = false; 39562306a36Sopenharmony_ci u32 cause; 39662306a36Sopenharmony_ci 39762306a36Sopenharmony_ci cause = ioread32(idxd->reg_base + IDXD_INTCAUSE_OFFSET); 39862306a36Sopenharmony_ci if (!cause) 39962306a36Sopenharmony_ci return IRQ_NONE; 40062306a36Sopenharmony_ci 40162306a36Sopenharmony_ci iowrite32(cause, idxd->reg_base + IDXD_INTCAUSE_OFFSET); 40262306a36Sopenharmony_ci 40362306a36Sopenharmony_ci if (cause & IDXD_INTC_HALT_STATE) 40462306a36Sopenharmony_ci goto halt; 40562306a36Sopenharmony_ci 40662306a36Sopenharmony_ci if (cause & IDXD_INTC_ERR) { 40762306a36Sopenharmony_ci spin_lock(&idxd->dev_lock); 40862306a36Sopenharmony_ci for (i = 0; i < 4; i++) 40962306a36Sopenharmony_ci idxd->sw_err.bits[i] = ioread64(idxd->reg_base + 41062306a36Sopenharmony_ci IDXD_SWERR_OFFSET + i * sizeof(u64)); 41162306a36Sopenharmony_ci 41262306a36Sopenharmony_ci iowrite64(idxd->sw_err.bits[0] & IDXD_SWERR_ACK, 41362306a36Sopenharmony_ci idxd->reg_base + IDXD_SWERR_OFFSET); 41462306a36Sopenharmony_ci 41562306a36Sopenharmony_ci if (idxd->sw_err.valid && idxd->sw_err.wq_idx_valid) { 41662306a36Sopenharmony_ci int id = idxd->sw_err.wq_idx; 41762306a36Sopenharmony_ci struct idxd_wq *wq = idxd->wqs[id]; 41862306a36Sopenharmony_ci 41962306a36Sopenharmony_ci if (wq->type == IDXD_WQT_USER) 42062306a36Sopenharmony_ci wake_up_interruptible(&wq->err_queue); 42162306a36Sopenharmony_ci } else { 42262306a36Sopenharmony_ci int i; 42362306a36Sopenharmony_ci 42462306a36Sopenharmony_ci for (i = 0; i < idxd->max_wqs; i++) { 42562306a36Sopenharmony_ci struct idxd_wq *wq = idxd->wqs[i]; 42662306a36Sopenharmony_ci 42762306a36Sopenharmony_ci if (wq->type == IDXD_WQT_USER) 42862306a36Sopenharmony_ci wake_up_interruptible(&wq->err_queue); 42962306a36Sopenharmony_ci } 43062306a36Sopenharmony_ci } 43162306a36Sopenharmony_ci 43262306a36Sopenharmony_ci spin_unlock(&idxd->dev_lock); 43362306a36Sopenharmony_ci val |= IDXD_INTC_ERR; 43462306a36Sopenharmony_ci 43562306a36Sopenharmony_ci for (i = 0; i < 4; i++) 43662306a36Sopenharmony_ci dev_warn(dev, "err[%d]: %#16.16llx\n", 43762306a36Sopenharmony_ci i, idxd->sw_err.bits[i]); 43862306a36Sopenharmony_ci err = true; 43962306a36Sopenharmony_ci } 44062306a36Sopenharmony_ci 44162306a36Sopenharmony_ci if (cause & IDXD_INTC_INT_HANDLE_REVOKED) { 44262306a36Sopenharmony_ci struct idxd_int_handle_revoke *revoke; 44362306a36Sopenharmony_ci 44462306a36Sopenharmony_ci val |= IDXD_INTC_INT_HANDLE_REVOKED; 44562306a36Sopenharmony_ci 44662306a36Sopenharmony_ci revoke = kzalloc(sizeof(*revoke), GFP_ATOMIC); 44762306a36Sopenharmony_ci if (revoke) { 44862306a36Sopenharmony_ci revoke->idxd = idxd; 44962306a36Sopenharmony_ci INIT_WORK(&revoke->work, idxd_int_handle_revoke); 45062306a36Sopenharmony_ci queue_work(idxd->wq, &revoke->work); 45162306a36Sopenharmony_ci 45262306a36Sopenharmony_ci } else { 45362306a36Sopenharmony_ci dev_err(dev, "Failed to allocate work for int handle revoke\n"); 45462306a36Sopenharmony_ci idxd_wqs_quiesce(idxd); 45562306a36Sopenharmony_ci } 45662306a36Sopenharmony_ci } 45762306a36Sopenharmony_ci 45862306a36Sopenharmony_ci if (cause & IDXD_INTC_CMD) { 45962306a36Sopenharmony_ci val |= IDXD_INTC_CMD; 46062306a36Sopenharmony_ci complete(idxd->cmd_done); 46162306a36Sopenharmony_ci } 46262306a36Sopenharmony_ci 46362306a36Sopenharmony_ci if (cause & IDXD_INTC_OCCUPY) { 46462306a36Sopenharmony_ci /* Driver does not utilize occupancy interrupt */ 46562306a36Sopenharmony_ci val |= IDXD_INTC_OCCUPY; 46662306a36Sopenharmony_ci } 46762306a36Sopenharmony_ci 46862306a36Sopenharmony_ci if (cause & IDXD_INTC_PERFMON_OVFL) { 46962306a36Sopenharmony_ci val |= IDXD_INTC_PERFMON_OVFL; 47062306a36Sopenharmony_ci perfmon_counter_overflow(idxd); 47162306a36Sopenharmony_ci } 47262306a36Sopenharmony_ci 47362306a36Sopenharmony_ci if (cause & IDXD_INTC_EVL) { 47462306a36Sopenharmony_ci val |= IDXD_INTC_EVL; 47562306a36Sopenharmony_ci process_evl_entries(idxd); 47662306a36Sopenharmony_ci } 47762306a36Sopenharmony_ci 47862306a36Sopenharmony_ci val ^= cause; 47962306a36Sopenharmony_ci if (val) 48062306a36Sopenharmony_ci dev_warn_once(dev, "Unexpected interrupt cause bits set: %#x\n", 48162306a36Sopenharmony_ci val); 48262306a36Sopenharmony_ci 48362306a36Sopenharmony_ci if (!err) 48462306a36Sopenharmony_ci goto out; 48562306a36Sopenharmony_ci 48662306a36Sopenharmony_cihalt: 48762306a36Sopenharmony_ci gensts.bits = ioread32(idxd->reg_base + IDXD_GENSTATS_OFFSET); 48862306a36Sopenharmony_ci if (gensts.state == IDXD_DEVICE_STATE_HALT) { 48962306a36Sopenharmony_ci idxd->state = IDXD_DEV_HALTED; 49062306a36Sopenharmony_ci if (gensts.reset_type == IDXD_DEVICE_RESET_SOFTWARE) { 49162306a36Sopenharmony_ci /* 49262306a36Sopenharmony_ci * If we need a software reset, we will throw the work 49362306a36Sopenharmony_ci * on a system workqueue in order to allow interrupts 49462306a36Sopenharmony_ci * for the device command completions. 49562306a36Sopenharmony_ci */ 49662306a36Sopenharmony_ci INIT_WORK(&idxd->work, idxd_device_reinit); 49762306a36Sopenharmony_ci queue_work(idxd->wq, &idxd->work); 49862306a36Sopenharmony_ci } else { 49962306a36Sopenharmony_ci idxd->state = IDXD_DEV_HALTED; 50062306a36Sopenharmony_ci idxd_wqs_quiesce(idxd); 50162306a36Sopenharmony_ci idxd_wqs_unmap_portal(idxd); 50262306a36Sopenharmony_ci idxd_device_clear_state(idxd); 50362306a36Sopenharmony_ci dev_err(&idxd->pdev->dev, 50462306a36Sopenharmony_ci "idxd halted, need %s.\n", 50562306a36Sopenharmony_ci gensts.reset_type == IDXD_DEVICE_RESET_FLR ? 50662306a36Sopenharmony_ci "FLR" : "system reset"); 50762306a36Sopenharmony_ci } 50862306a36Sopenharmony_ci } 50962306a36Sopenharmony_ci 51062306a36Sopenharmony_ciout: 51162306a36Sopenharmony_ci return IRQ_HANDLED; 51262306a36Sopenharmony_ci} 51362306a36Sopenharmony_ci 51462306a36Sopenharmony_cistatic void idxd_int_handle_resubmit_work(struct work_struct *work) 51562306a36Sopenharmony_ci{ 51662306a36Sopenharmony_ci struct idxd_resubmit *irw = container_of(work, struct idxd_resubmit, work); 51762306a36Sopenharmony_ci struct idxd_desc *desc = irw->desc; 51862306a36Sopenharmony_ci struct idxd_wq *wq = desc->wq; 51962306a36Sopenharmony_ci int rc; 52062306a36Sopenharmony_ci 52162306a36Sopenharmony_ci desc->completion->status = 0; 52262306a36Sopenharmony_ci rc = idxd_submit_desc(wq, desc); 52362306a36Sopenharmony_ci if (rc < 0) { 52462306a36Sopenharmony_ci dev_dbg(&wq->idxd->pdev->dev, "Failed to resubmit desc %d to wq %d.\n", 52562306a36Sopenharmony_ci desc->id, wq->id); 52662306a36Sopenharmony_ci /* 52762306a36Sopenharmony_ci * If the error is not -EAGAIN, it means the submission failed due to wq 52862306a36Sopenharmony_ci * has been killed instead of ENQCMDS failure. Here the driver needs to 52962306a36Sopenharmony_ci * notify the submitter of the failure by reporting abort status. 53062306a36Sopenharmony_ci * 53162306a36Sopenharmony_ci * -EAGAIN comes from ENQCMDS failure. idxd_submit_desc() will handle the 53262306a36Sopenharmony_ci * abort. 53362306a36Sopenharmony_ci */ 53462306a36Sopenharmony_ci if (rc != -EAGAIN) { 53562306a36Sopenharmony_ci desc->completion->status = IDXD_COMP_DESC_ABORT; 53662306a36Sopenharmony_ci idxd_dma_complete_txd(desc, IDXD_COMPLETE_ABORT, false); 53762306a36Sopenharmony_ci } 53862306a36Sopenharmony_ci idxd_free_desc(wq, desc); 53962306a36Sopenharmony_ci } 54062306a36Sopenharmony_ci kfree(irw); 54162306a36Sopenharmony_ci} 54262306a36Sopenharmony_ci 54362306a36Sopenharmony_cibool idxd_queue_int_handle_resubmit(struct idxd_desc *desc) 54462306a36Sopenharmony_ci{ 54562306a36Sopenharmony_ci struct idxd_wq *wq = desc->wq; 54662306a36Sopenharmony_ci struct idxd_device *idxd = wq->idxd; 54762306a36Sopenharmony_ci struct idxd_resubmit *irw; 54862306a36Sopenharmony_ci 54962306a36Sopenharmony_ci irw = kzalloc(sizeof(*irw), GFP_KERNEL); 55062306a36Sopenharmony_ci if (!irw) 55162306a36Sopenharmony_ci return false; 55262306a36Sopenharmony_ci 55362306a36Sopenharmony_ci irw->desc = desc; 55462306a36Sopenharmony_ci INIT_WORK(&irw->work, idxd_int_handle_resubmit_work); 55562306a36Sopenharmony_ci queue_work(idxd->wq, &irw->work); 55662306a36Sopenharmony_ci return true; 55762306a36Sopenharmony_ci} 55862306a36Sopenharmony_ci 55962306a36Sopenharmony_cistatic void irq_process_pending_llist(struct idxd_irq_entry *irq_entry) 56062306a36Sopenharmony_ci{ 56162306a36Sopenharmony_ci struct idxd_desc *desc, *t; 56262306a36Sopenharmony_ci struct llist_node *head; 56362306a36Sopenharmony_ci 56462306a36Sopenharmony_ci head = llist_del_all(&irq_entry->pending_llist); 56562306a36Sopenharmony_ci if (!head) 56662306a36Sopenharmony_ci return; 56762306a36Sopenharmony_ci 56862306a36Sopenharmony_ci llist_for_each_entry_safe(desc, t, head, llnode) { 56962306a36Sopenharmony_ci u8 status = desc->completion->status & DSA_COMP_STATUS_MASK; 57062306a36Sopenharmony_ci 57162306a36Sopenharmony_ci if (status) { 57262306a36Sopenharmony_ci /* 57362306a36Sopenharmony_ci * Check against the original status as ABORT is software defined 57462306a36Sopenharmony_ci * and 0xff, which DSA_COMP_STATUS_MASK can mask out. 57562306a36Sopenharmony_ci */ 57662306a36Sopenharmony_ci if (unlikely(desc->completion->status == IDXD_COMP_DESC_ABORT)) { 57762306a36Sopenharmony_ci idxd_dma_complete_txd(desc, IDXD_COMPLETE_ABORT, true); 57862306a36Sopenharmony_ci continue; 57962306a36Sopenharmony_ci } 58062306a36Sopenharmony_ci 58162306a36Sopenharmony_ci idxd_dma_complete_txd(desc, IDXD_COMPLETE_NORMAL, true); 58262306a36Sopenharmony_ci } else { 58362306a36Sopenharmony_ci spin_lock(&irq_entry->list_lock); 58462306a36Sopenharmony_ci list_add_tail(&desc->list, 58562306a36Sopenharmony_ci &irq_entry->work_list); 58662306a36Sopenharmony_ci spin_unlock(&irq_entry->list_lock); 58762306a36Sopenharmony_ci } 58862306a36Sopenharmony_ci } 58962306a36Sopenharmony_ci} 59062306a36Sopenharmony_ci 59162306a36Sopenharmony_cistatic void irq_process_work_list(struct idxd_irq_entry *irq_entry) 59262306a36Sopenharmony_ci{ 59362306a36Sopenharmony_ci LIST_HEAD(flist); 59462306a36Sopenharmony_ci struct idxd_desc *desc, *n; 59562306a36Sopenharmony_ci 59662306a36Sopenharmony_ci /* 59762306a36Sopenharmony_ci * This lock protects list corruption from access of list outside of the irq handler 59862306a36Sopenharmony_ci * thread. 59962306a36Sopenharmony_ci */ 60062306a36Sopenharmony_ci spin_lock(&irq_entry->list_lock); 60162306a36Sopenharmony_ci if (list_empty(&irq_entry->work_list)) { 60262306a36Sopenharmony_ci spin_unlock(&irq_entry->list_lock); 60362306a36Sopenharmony_ci return; 60462306a36Sopenharmony_ci } 60562306a36Sopenharmony_ci 60662306a36Sopenharmony_ci list_for_each_entry_safe(desc, n, &irq_entry->work_list, list) { 60762306a36Sopenharmony_ci if (desc->completion->status) { 60862306a36Sopenharmony_ci list_move_tail(&desc->list, &flist); 60962306a36Sopenharmony_ci } 61062306a36Sopenharmony_ci } 61162306a36Sopenharmony_ci 61262306a36Sopenharmony_ci spin_unlock(&irq_entry->list_lock); 61362306a36Sopenharmony_ci 61462306a36Sopenharmony_ci list_for_each_entry(desc, &flist, list) { 61562306a36Sopenharmony_ci /* 61662306a36Sopenharmony_ci * Check against the original status as ABORT is software defined 61762306a36Sopenharmony_ci * and 0xff, which DSA_COMP_STATUS_MASK can mask out. 61862306a36Sopenharmony_ci */ 61962306a36Sopenharmony_ci if (unlikely(desc->completion->status == IDXD_COMP_DESC_ABORT)) { 62062306a36Sopenharmony_ci idxd_dma_complete_txd(desc, IDXD_COMPLETE_ABORT, true); 62162306a36Sopenharmony_ci continue; 62262306a36Sopenharmony_ci } 62362306a36Sopenharmony_ci 62462306a36Sopenharmony_ci idxd_dma_complete_txd(desc, IDXD_COMPLETE_NORMAL, true); 62562306a36Sopenharmony_ci } 62662306a36Sopenharmony_ci} 62762306a36Sopenharmony_ci 62862306a36Sopenharmony_ciirqreturn_t idxd_wq_thread(int irq, void *data) 62962306a36Sopenharmony_ci{ 63062306a36Sopenharmony_ci struct idxd_irq_entry *irq_entry = data; 63162306a36Sopenharmony_ci 63262306a36Sopenharmony_ci /* 63362306a36Sopenharmony_ci * There are two lists we are processing. The pending_llist is where 63462306a36Sopenharmony_ci * submmiter adds all the submitted descriptor after sending it to 63562306a36Sopenharmony_ci * the workqueue. It's a lockless singly linked list. The work_list 63662306a36Sopenharmony_ci * is the common linux double linked list. We are in a scenario of 63762306a36Sopenharmony_ci * multiple producers and a single consumer. The producers are all 63862306a36Sopenharmony_ci * the kernel submitters of descriptors, and the consumer is the 63962306a36Sopenharmony_ci * kernel irq handler thread for the msix vector when using threaded 64062306a36Sopenharmony_ci * irq. To work with the restrictions of llist to remain lockless, 64162306a36Sopenharmony_ci * we are doing the following steps: 64262306a36Sopenharmony_ci * 1. Iterate through the work_list and process any completed 64362306a36Sopenharmony_ci * descriptor. Delete the completed entries during iteration. 64462306a36Sopenharmony_ci * 2. llist_del_all() from the pending list. 64562306a36Sopenharmony_ci * 3. Iterate through the llist that was deleted from the pending list 64662306a36Sopenharmony_ci * and process the completed entries. 64762306a36Sopenharmony_ci * 4. If the entry is still waiting on hardware, list_add_tail() to 64862306a36Sopenharmony_ci * the work_list. 64962306a36Sopenharmony_ci */ 65062306a36Sopenharmony_ci irq_process_work_list(irq_entry); 65162306a36Sopenharmony_ci irq_process_pending_llist(irq_entry); 65262306a36Sopenharmony_ci 65362306a36Sopenharmony_ci return IRQ_HANDLED; 65462306a36Sopenharmony_ci} 655