162306a36Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0 or BSD-3-Clause 262306a36Sopenharmony_ci/* 362306a36Sopenharmony_ci * Copyright(c) 2015-2018 Intel Corporation. 462306a36Sopenharmony_ci */ 562306a36Sopenharmony_ci 662306a36Sopenharmony_ci#include <linux/delay.h> 762306a36Sopenharmony_ci#include "hfi.h" 862306a36Sopenharmony_ci#include "qp.h" 962306a36Sopenharmony_ci#include "trace.h" 1062306a36Sopenharmony_ci 1162306a36Sopenharmony_ci#define SC(name) SEND_CTXT_##name 1262306a36Sopenharmony_ci/* 1362306a36Sopenharmony_ci * Send Context functions 1462306a36Sopenharmony_ci */ 1562306a36Sopenharmony_cistatic void sc_wait_for_packet_egress(struct send_context *sc, int pause); 1662306a36Sopenharmony_ci 1762306a36Sopenharmony_ci/* 1862306a36Sopenharmony_ci * Set the CM reset bit and wait for it to clear. Use the provided 1962306a36Sopenharmony_ci * sendctrl register. This routine has no locking. 2062306a36Sopenharmony_ci */ 2162306a36Sopenharmony_civoid __cm_reset(struct hfi1_devdata *dd, u64 sendctrl) 2262306a36Sopenharmony_ci{ 2362306a36Sopenharmony_ci write_csr(dd, SEND_CTRL, sendctrl | SEND_CTRL_CM_RESET_SMASK); 2462306a36Sopenharmony_ci while (1) { 2562306a36Sopenharmony_ci udelay(1); 2662306a36Sopenharmony_ci sendctrl = read_csr(dd, SEND_CTRL); 2762306a36Sopenharmony_ci if ((sendctrl & SEND_CTRL_CM_RESET_SMASK) == 0) 2862306a36Sopenharmony_ci break; 2962306a36Sopenharmony_ci } 3062306a36Sopenharmony_ci} 3162306a36Sopenharmony_ci 3262306a36Sopenharmony_ci/* global control of PIO send */ 3362306a36Sopenharmony_civoid pio_send_control(struct hfi1_devdata *dd, int op) 3462306a36Sopenharmony_ci{ 3562306a36Sopenharmony_ci u64 reg, mask; 3662306a36Sopenharmony_ci unsigned long flags; 3762306a36Sopenharmony_ci int write = 1; /* write sendctrl back */ 3862306a36Sopenharmony_ci int flush = 0; /* re-read sendctrl to make sure it is flushed */ 3962306a36Sopenharmony_ci int i; 4062306a36Sopenharmony_ci 4162306a36Sopenharmony_ci spin_lock_irqsave(&dd->sendctrl_lock, flags); 4262306a36Sopenharmony_ci 4362306a36Sopenharmony_ci reg = read_csr(dd, SEND_CTRL); 4462306a36Sopenharmony_ci switch (op) { 4562306a36Sopenharmony_ci case PSC_GLOBAL_ENABLE: 4662306a36Sopenharmony_ci reg |= SEND_CTRL_SEND_ENABLE_SMASK; 4762306a36Sopenharmony_ci fallthrough; 4862306a36Sopenharmony_ci case PSC_DATA_VL_ENABLE: 4962306a36Sopenharmony_ci mask = 0; 5062306a36Sopenharmony_ci for (i = 0; i < ARRAY_SIZE(dd->vld); i++) 5162306a36Sopenharmony_ci if (!dd->vld[i].mtu) 5262306a36Sopenharmony_ci mask |= BIT_ULL(i); 5362306a36Sopenharmony_ci /* Disallow sending on VLs not enabled */ 5462306a36Sopenharmony_ci mask = (mask & SEND_CTRL_UNSUPPORTED_VL_MASK) << 5562306a36Sopenharmony_ci SEND_CTRL_UNSUPPORTED_VL_SHIFT; 5662306a36Sopenharmony_ci reg = (reg & ~SEND_CTRL_UNSUPPORTED_VL_SMASK) | mask; 5762306a36Sopenharmony_ci break; 5862306a36Sopenharmony_ci case PSC_GLOBAL_DISABLE: 5962306a36Sopenharmony_ci reg &= ~SEND_CTRL_SEND_ENABLE_SMASK; 6062306a36Sopenharmony_ci break; 6162306a36Sopenharmony_ci case PSC_GLOBAL_VLARB_ENABLE: 6262306a36Sopenharmony_ci reg |= SEND_CTRL_VL_ARBITER_ENABLE_SMASK; 6362306a36Sopenharmony_ci break; 6462306a36Sopenharmony_ci case PSC_GLOBAL_VLARB_DISABLE: 6562306a36Sopenharmony_ci reg &= ~SEND_CTRL_VL_ARBITER_ENABLE_SMASK; 6662306a36Sopenharmony_ci break; 6762306a36Sopenharmony_ci case PSC_CM_RESET: 6862306a36Sopenharmony_ci __cm_reset(dd, reg); 6962306a36Sopenharmony_ci write = 0; /* CSR already written (and flushed) */ 7062306a36Sopenharmony_ci break; 7162306a36Sopenharmony_ci case PSC_DATA_VL_DISABLE: 7262306a36Sopenharmony_ci reg |= SEND_CTRL_UNSUPPORTED_VL_SMASK; 7362306a36Sopenharmony_ci flush = 1; 7462306a36Sopenharmony_ci break; 7562306a36Sopenharmony_ci default: 7662306a36Sopenharmony_ci dd_dev_err(dd, "%s: invalid control %d\n", __func__, op); 7762306a36Sopenharmony_ci break; 7862306a36Sopenharmony_ci } 7962306a36Sopenharmony_ci 8062306a36Sopenharmony_ci if (write) { 8162306a36Sopenharmony_ci write_csr(dd, SEND_CTRL, reg); 8262306a36Sopenharmony_ci if (flush) 8362306a36Sopenharmony_ci (void)read_csr(dd, SEND_CTRL); /* flush write */ 8462306a36Sopenharmony_ci } 8562306a36Sopenharmony_ci 8662306a36Sopenharmony_ci spin_unlock_irqrestore(&dd->sendctrl_lock, flags); 8762306a36Sopenharmony_ci} 8862306a36Sopenharmony_ci 8962306a36Sopenharmony_ci/* number of send context memory pools */ 9062306a36Sopenharmony_ci#define NUM_SC_POOLS 2 9162306a36Sopenharmony_ci 9262306a36Sopenharmony_ci/* Send Context Size (SCS) wildcards */ 9362306a36Sopenharmony_ci#define SCS_POOL_0 -1 9462306a36Sopenharmony_ci#define SCS_POOL_1 -2 9562306a36Sopenharmony_ci 9662306a36Sopenharmony_ci/* Send Context Count (SCC) wildcards */ 9762306a36Sopenharmony_ci#define SCC_PER_VL -1 9862306a36Sopenharmony_ci#define SCC_PER_CPU -2 9962306a36Sopenharmony_ci#define SCC_PER_KRCVQ -3 10062306a36Sopenharmony_ci 10162306a36Sopenharmony_ci/* Send Context Size (SCS) constants */ 10262306a36Sopenharmony_ci#define SCS_ACK_CREDITS 32 10362306a36Sopenharmony_ci#define SCS_VL15_CREDITS 102 /* 3 pkts of 2048B data + 128B header */ 10462306a36Sopenharmony_ci 10562306a36Sopenharmony_ci#define PIO_THRESHOLD_CEILING 4096 10662306a36Sopenharmony_ci 10762306a36Sopenharmony_ci#define PIO_WAIT_BATCH_SIZE 5 10862306a36Sopenharmony_ci 10962306a36Sopenharmony_ci/* default send context sizes */ 11062306a36Sopenharmony_cistatic struct sc_config_sizes sc_config_sizes[SC_MAX] = { 11162306a36Sopenharmony_ci [SC_KERNEL] = { .size = SCS_POOL_0, /* even divide, pool 0 */ 11262306a36Sopenharmony_ci .count = SCC_PER_VL }, /* one per NUMA */ 11362306a36Sopenharmony_ci [SC_ACK] = { .size = SCS_ACK_CREDITS, 11462306a36Sopenharmony_ci .count = SCC_PER_KRCVQ }, 11562306a36Sopenharmony_ci [SC_USER] = { .size = SCS_POOL_0, /* even divide, pool 0 */ 11662306a36Sopenharmony_ci .count = SCC_PER_CPU }, /* one per CPU */ 11762306a36Sopenharmony_ci [SC_VL15] = { .size = SCS_VL15_CREDITS, 11862306a36Sopenharmony_ci .count = 1 }, 11962306a36Sopenharmony_ci 12062306a36Sopenharmony_ci}; 12162306a36Sopenharmony_ci 12262306a36Sopenharmony_ci/* send context memory pool configuration */ 12362306a36Sopenharmony_cistruct mem_pool_config { 12462306a36Sopenharmony_ci int centipercent; /* % of memory, in 100ths of 1% */ 12562306a36Sopenharmony_ci int absolute_blocks; /* absolute block count */ 12662306a36Sopenharmony_ci}; 12762306a36Sopenharmony_ci 12862306a36Sopenharmony_ci/* default memory pool configuration: 100% in pool 0 */ 12962306a36Sopenharmony_cistatic struct mem_pool_config sc_mem_pool_config[NUM_SC_POOLS] = { 13062306a36Sopenharmony_ci /* centi%, abs blocks */ 13162306a36Sopenharmony_ci { 10000, -1 }, /* pool 0 */ 13262306a36Sopenharmony_ci { 0, -1 }, /* pool 1 */ 13362306a36Sopenharmony_ci}; 13462306a36Sopenharmony_ci 13562306a36Sopenharmony_ci/* memory pool information, used when calculating final sizes */ 13662306a36Sopenharmony_cistruct mem_pool_info { 13762306a36Sopenharmony_ci int centipercent; /* 13862306a36Sopenharmony_ci * 100th of 1% of memory to use, -1 if blocks 13962306a36Sopenharmony_ci * already set 14062306a36Sopenharmony_ci */ 14162306a36Sopenharmony_ci int count; /* count of contexts in the pool */ 14262306a36Sopenharmony_ci int blocks; /* block size of the pool */ 14362306a36Sopenharmony_ci int size; /* context size, in blocks */ 14462306a36Sopenharmony_ci}; 14562306a36Sopenharmony_ci 14662306a36Sopenharmony_ci/* 14762306a36Sopenharmony_ci * Convert a pool wildcard to a valid pool index. The wildcards 14862306a36Sopenharmony_ci * start at -1 and increase negatively. Map them as: 14962306a36Sopenharmony_ci * -1 => 0 15062306a36Sopenharmony_ci * -2 => 1 15162306a36Sopenharmony_ci * etc. 15262306a36Sopenharmony_ci * 15362306a36Sopenharmony_ci * Return -1 on non-wildcard input, otherwise convert to a pool number. 15462306a36Sopenharmony_ci */ 15562306a36Sopenharmony_cistatic int wildcard_to_pool(int wc) 15662306a36Sopenharmony_ci{ 15762306a36Sopenharmony_ci if (wc >= 0) 15862306a36Sopenharmony_ci return -1; /* non-wildcard */ 15962306a36Sopenharmony_ci return -wc - 1; 16062306a36Sopenharmony_ci} 16162306a36Sopenharmony_ci 16262306a36Sopenharmony_cistatic const char *sc_type_names[SC_MAX] = { 16362306a36Sopenharmony_ci "kernel", 16462306a36Sopenharmony_ci "ack", 16562306a36Sopenharmony_ci "user", 16662306a36Sopenharmony_ci "vl15" 16762306a36Sopenharmony_ci}; 16862306a36Sopenharmony_ci 16962306a36Sopenharmony_cistatic const char *sc_type_name(int index) 17062306a36Sopenharmony_ci{ 17162306a36Sopenharmony_ci if (index < 0 || index >= SC_MAX) 17262306a36Sopenharmony_ci return "unknown"; 17362306a36Sopenharmony_ci return sc_type_names[index]; 17462306a36Sopenharmony_ci} 17562306a36Sopenharmony_ci 17662306a36Sopenharmony_ci/* 17762306a36Sopenharmony_ci * Read the send context memory pool configuration and send context 17862306a36Sopenharmony_ci * size configuration. Replace any wildcards and come up with final 17962306a36Sopenharmony_ci * counts and sizes for the send context types. 18062306a36Sopenharmony_ci */ 18162306a36Sopenharmony_ciint init_sc_pools_and_sizes(struct hfi1_devdata *dd) 18262306a36Sopenharmony_ci{ 18362306a36Sopenharmony_ci struct mem_pool_info mem_pool_info[NUM_SC_POOLS] = { { 0 } }; 18462306a36Sopenharmony_ci int total_blocks = (chip_pio_mem_size(dd) / PIO_BLOCK_SIZE) - 1; 18562306a36Sopenharmony_ci int total_contexts = 0; 18662306a36Sopenharmony_ci int fixed_blocks; 18762306a36Sopenharmony_ci int pool_blocks; 18862306a36Sopenharmony_ci int used_blocks; 18962306a36Sopenharmony_ci int cp_total; /* centipercent total */ 19062306a36Sopenharmony_ci int ab_total; /* absolute block total */ 19162306a36Sopenharmony_ci int extra; 19262306a36Sopenharmony_ci int i; 19362306a36Sopenharmony_ci 19462306a36Sopenharmony_ci /* 19562306a36Sopenharmony_ci * When SDMA is enabled, kernel context pio packet size is capped by 19662306a36Sopenharmony_ci * "piothreshold". Reduce pio buffer allocation for kernel context by 19762306a36Sopenharmony_ci * setting it to a fixed size. The allocation allows 3-deep buffering 19862306a36Sopenharmony_ci * of the largest pio packets plus up to 128 bytes header, sufficient 19962306a36Sopenharmony_ci * to maintain verbs performance. 20062306a36Sopenharmony_ci * 20162306a36Sopenharmony_ci * When SDMA is disabled, keep the default pooling allocation. 20262306a36Sopenharmony_ci */ 20362306a36Sopenharmony_ci if (HFI1_CAP_IS_KSET(SDMA)) { 20462306a36Sopenharmony_ci u16 max_pkt_size = (piothreshold < PIO_THRESHOLD_CEILING) ? 20562306a36Sopenharmony_ci piothreshold : PIO_THRESHOLD_CEILING; 20662306a36Sopenharmony_ci sc_config_sizes[SC_KERNEL].size = 20762306a36Sopenharmony_ci 3 * (max_pkt_size + 128) / PIO_BLOCK_SIZE; 20862306a36Sopenharmony_ci } 20962306a36Sopenharmony_ci 21062306a36Sopenharmony_ci /* 21162306a36Sopenharmony_ci * Step 0: 21262306a36Sopenharmony_ci * - copy the centipercents/absolute sizes from the pool config 21362306a36Sopenharmony_ci * - sanity check these values 21462306a36Sopenharmony_ci * - add up centipercents, then later check for full value 21562306a36Sopenharmony_ci * - add up absolute blocks, then later check for over-commit 21662306a36Sopenharmony_ci */ 21762306a36Sopenharmony_ci cp_total = 0; 21862306a36Sopenharmony_ci ab_total = 0; 21962306a36Sopenharmony_ci for (i = 0; i < NUM_SC_POOLS; i++) { 22062306a36Sopenharmony_ci int cp = sc_mem_pool_config[i].centipercent; 22162306a36Sopenharmony_ci int ab = sc_mem_pool_config[i].absolute_blocks; 22262306a36Sopenharmony_ci 22362306a36Sopenharmony_ci /* 22462306a36Sopenharmony_ci * A negative value is "unused" or "invalid". Both *can* 22562306a36Sopenharmony_ci * be valid, but centipercent wins, so check that first 22662306a36Sopenharmony_ci */ 22762306a36Sopenharmony_ci if (cp >= 0) { /* centipercent valid */ 22862306a36Sopenharmony_ci cp_total += cp; 22962306a36Sopenharmony_ci } else if (ab >= 0) { /* absolute blocks valid */ 23062306a36Sopenharmony_ci ab_total += ab; 23162306a36Sopenharmony_ci } else { /* neither valid */ 23262306a36Sopenharmony_ci dd_dev_err( 23362306a36Sopenharmony_ci dd, 23462306a36Sopenharmony_ci "Send context memory pool %d: both the block count and centipercent are invalid\n", 23562306a36Sopenharmony_ci i); 23662306a36Sopenharmony_ci return -EINVAL; 23762306a36Sopenharmony_ci } 23862306a36Sopenharmony_ci 23962306a36Sopenharmony_ci mem_pool_info[i].centipercent = cp; 24062306a36Sopenharmony_ci mem_pool_info[i].blocks = ab; 24162306a36Sopenharmony_ci } 24262306a36Sopenharmony_ci 24362306a36Sopenharmony_ci /* do not use both % and absolute blocks for different pools */ 24462306a36Sopenharmony_ci if (cp_total != 0 && ab_total != 0) { 24562306a36Sopenharmony_ci dd_dev_err( 24662306a36Sopenharmony_ci dd, 24762306a36Sopenharmony_ci "All send context memory pools must be described as either centipercent or blocks, no mixing between pools\n"); 24862306a36Sopenharmony_ci return -EINVAL; 24962306a36Sopenharmony_ci } 25062306a36Sopenharmony_ci 25162306a36Sopenharmony_ci /* if any percentages are present, they must add up to 100% x 100 */ 25262306a36Sopenharmony_ci if (cp_total != 0 && cp_total != 10000) { 25362306a36Sopenharmony_ci dd_dev_err( 25462306a36Sopenharmony_ci dd, 25562306a36Sopenharmony_ci "Send context memory pool centipercent is %d, expecting 10000\n", 25662306a36Sopenharmony_ci cp_total); 25762306a36Sopenharmony_ci return -EINVAL; 25862306a36Sopenharmony_ci } 25962306a36Sopenharmony_ci 26062306a36Sopenharmony_ci /* the absolute pool total cannot be more than the mem total */ 26162306a36Sopenharmony_ci if (ab_total > total_blocks) { 26262306a36Sopenharmony_ci dd_dev_err( 26362306a36Sopenharmony_ci dd, 26462306a36Sopenharmony_ci "Send context memory pool absolute block count %d is larger than the memory size %d\n", 26562306a36Sopenharmony_ci ab_total, total_blocks); 26662306a36Sopenharmony_ci return -EINVAL; 26762306a36Sopenharmony_ci } 26862306a36Sopenharmony_ci 26962306a36Sopenharmony_ci /* 27062306a36Sopenharmony_ci * Step 2: 27162306a36Sopenharmony_ci * - copy from the context size config 27262306a36Sopenharmony_ci * - replace context type wildcard counts with real values 27362306a36Sopenharmony_ci * - add up non-memory pool block sizes 27462306a36Sopenharmony_ci * - add up memory pool user counts 27562306a36Sopenharmony_ci */ 27662306a36Sopenharmony_ci fixed_blocks = 0; 27762306a36Sopenharmony_ci for (i = 0; i < SC_MAX; i++) { 27862306a36Sopenharmony_ci int count = sc_config_sizes[i].count; 27962306a36Sopenharmony_ci int size = sc_config_sizes[i].size; 28062306a36Sopenharmony_ci int pool; 28162306a36Sopenharmony_ci 28262306a36Sopenharmony_ci /* 28362306a36Sopenharmony_ci * Sanity check count: Either a positive value or 28462306a36Sopenharmony_ci * one of the expected wildcards is valid. The positive 28562306a36Sopenharmony_ci * value is checked later when we compare against total 28662306a36Sopenharmony_ci * memory available. 28762306a36Sopenharmony_ci */ 28862306a36Sopenharmony_ci if (i == SC_ACK) { 28962306a36Sopenharmony_ci count = dd->n_krcv_queues; 29062306a36Sopenharmony_ci } else if (i == SC_KERNEL) { 29162306a36Sopenharmony_ci count = INIT_SC_PER_VL * num_vls; 29262306a36Sopenharmony_ci } else if (count == SCC_PER_CPU) { 29362306a36Sopenharmony_ci count = dd->num_rcv_contexts - dd->n_krcv_queues; 29462306a36Sopenharmony_ci } else if (count < 0) { 29562306a36Sopenharmony_ci dd_dev_err( 29662306a36Sopenharmony_ci dd, 29762306a36Sopenharmony_ci "%s send context invalid count wildcard %d\n", 29862306a36Sopenharmony_ci sc_type_name(i), count); 29962306a36Sopenharmony_ci return -EINVAL; 30062306a36Sopenharmony_ci } 30162306a36Sopenharmony_ci if (total_contexts + count > chip_send_contexts(dd)) 30262306a36Sopenharmony_ci count = chip_send_contexts(dd) - total_contexts; 30362306a36Sopenharmony_ci 30462306a36Sopenharmony_ci total_contexts += count; 30562306a36Sopenharmony_ci 30662306a36Sopenharmony_ci /* 30762306a36Sopenharmony_ci * Sanity check pool: The conversion will return a pool 30862306a36Sopenharmony_ci * number or -1 if a fixed (non-negative) value. The fixed 30962306a36Sopenharmony_ci * value is checked later when we compare against 31062306a36Sopenharmony_ci * total memory available. 31162306a36Sopenharmony_ci */ 31262306a36Sopenharmony_ci pool = wildcard_to_pool(size); 31362306a36Sopenharmony_ci if (pool == -1) { /* non-wildcard */ 31462306a36Sopenharmony_ci fixed_blocks += size * count; 31562306a36Sopenharmony_ci } else if (pool < NUM_SC_POOLS) { /* valid wildcard */ 31662306a36Sopenharmony_ci mem_pool_info[pool].count += count; 31762306a36Sopenharmony_ci } else { /* invalid wildcard */ 31862306a36Sopenharmony_ci dd_dev_err( 31962306a36Sopenharmony_ci dd, 32062306a36Sopenharmony_ci "%s send context invalid pool wildcard %d\n", 32162306a36Sopenharmony_ci sc_type_name(i), size); 32262306a36Sopenharmony_ci return -EINVAL; 32362306a36Sopenharmony_ci } 32462306a36Sopenharmony_ci 32562306a36Sopenharmony_ci dd->sc_sizes[i].count = count; 32662306a36Sopenharmony_ci dd->sc_sizes[i].size = size; 32762306a36Sopenharmony_ci } 32862306a36Sopenharmony_ci if (fixed_blocks > total_blocks) { 32962306a36Sopenharmony_ci dd_dev_err( 33062306a36Sopenharmony_ci dd, 33162306a36Sopenharmony_ci "Send context fixed block count, %u, larger than total block count %u\n", 33262306a36Sopenharmony_ci fixed_blocks, total_blocks); 33362306a36Sopenharmony_ci return -EINVAL; 33462306a36Sopenharmony_ci } 33562306a36Sopenharmony_ci 33662306a36Sopenharmony_ci /* step 3: calculate the blocks in the pools, and pool context sizes */ 33762306a36Sopenharmony_ci pool_blocks = total_blocks - fixed_blocks; 33862306a36Sopenharmony_ci if (ab_total > pool_blocks) { 33962306a36Sopenharmony_ci dd_dev_err( 34062306a36Sopenharmony_ci dd, 34162306a36Sopenharmony_ci "Send context fixed pool sizes, %u, larger than pool block count %u\n", 34262306a36Sopenharmony_ci ab_total, pool_blocks); 34362306a36Sopenharmony_ci return -EINVAL; 34462306a36Sopenharmony_ci } 34562306a36Sopenharmony_ci /* subtract off the fixed pool blocks */ 34662306a36Sopenharmony_ci pool_blocks -= ab_total; 34762306a36Sopenharmony_ci 34862306a36Sopenharmony_ci for (i = 0; i < NUM_SC_POOLS; i++) { 34962306a36Sopenharmony_ci struct mem_pool_info *pi = &mem_pool_info[i]; 35062306a36Sopenharmony_ci 35162306a36Sopenharmony_ci /* % beats absolute blocks */ 35262306a36Sopenharmony_ci if (pi->centipercent >= 0) 35362306a36Sopenharmony_ci pi->blocks = (pool_blocks * pi->centipercent) / 10000; 35462306a36Sopenharmony_ci 35562306a36Sopenharmony_ci if (pi->blocks == 0 && pi->count != 0) { 35662306a36Sopenharmony_ci dd_dev_err( 35762306a36Sopenharmony_ci dd, 35862306a36Sopenharmony_ci "Send context memory pool %d has %u contexts, but no blocks\n", 35962306a36Sopenharmony_ci i, pi->count); 36062306a36Sopenharmony_ci return -EINVAL; 36162306a36Sopenharmony_ci } 36262306a36Sopenharmony_ci if (pi->count == 0) { 36362306a36Sopenharmony_ci /* warn about wasted blocks */ 36462306a36Sopenharmony_ci if (pi->blocks != 0) 36562306a36Sopenharmony_ci dd_dev_err( 36662306a36Sopenharmony_ci dd, 36762306a36Sopenharmony_ci "Send context memory pool %d has %u blocks, but zero contexts\n", 36862306a36Sopenharmony_ci i, pi->blocks); 36962306a36Sopenharmony_ci pi->size = 0; 37062306a36Sopenharmony_ci } else { 37162306a36Sopenharmony_ci pi->size = pi->blocks / pi->count; 37262306a36Sopenharmony_ci } 37362306a36Sopenharmony_ci } 37462306a36Sopenharmony_ci 37562306a36Sopenharmony_ci /* step 4: fill in the context type sizes from the pool sizes */ 37662306a36Sopenharmony_ci used_blocks = 0; 37762306a36Sopenharmony_ci for (i = 0; i < SC_MAX; i++) { 37862306a36Sopenharmony_ci if (dd->sc_sizes[i].size < 0) { 37962306a36Sopenharmony_ci unsigned pool = wildcard_to_pool(dd->sc_sizes[i].size); 38062306a36Sopenharmony_ci 38162306a36Sopenharmony_ci WARN_ON_ONCE(pool >= NUM_SC_POOLS); 38262306a36Sopenharmony_ci dd->sc_sizes[i].size = mem_pool_info[pool].size; 38362306a36Sopenharmony_ci } 38462306a36Sopenharmony_ci /* make sure we are not larger than what is allowed by the HW */ 38562306a36Sopenharmony_ci#define PIO_MAX_BLOCKS 1024 38662306a36Sopenharmony_ci if (dd->sc_sizes[i].size > PIO_MAX_BLOCKS) 38762306a36Sopenharmony_ci dd->sc_sizes[i].size = PIO_MAX_BLOCKS; 38862306a36Sopenharmony_ci 38962306a36Sopenharmony_ci /* calculate our total usage */ 39062306a36Sopenharmony_ci used_blocks += dd->sc_sizes[i].size * dd->sc_sizes[i].count; 39162306a36Sopenharmony_ci } 39262306a36Sopenharmony_ci extra = total_blocks - used_blocks; 39362306a36Sopenharmony_ci if (extra != 0) 39462306a36Sopenharmony_ci dd_dev_info(dd, "unused send context blocks: %d\n", extra); 39562306a36Sopenharmony_ci 39662306a36Sopenharmony_ci return total_contexts; 39762306a36Sopenharmony_ci} 39862306a36Sopenharmony_ci 39962306a36Sopenharmony_ciint init_send_contexts(struct hfi1_devdata *dd) 40062306a36Sopenharmony_ci{ 40162306a36Sopenharmony_ci u16 base; 40262306a36Sopenharmony_ci int ret, i, j, context; 40362306a36Sopenharmony_ci 40462306a36Sopenharmony_ci ret = init_credit_return(dd); 40562306a36Sopenharmony_ci if (ret) 40662306a36Sopenharmony_ci return ret; 40762306a36Sopenharmony_ci 40862306a36Sopenharmony_ci dd->hw_to_sw = kmalloc_array(TXE_NUM_CONTEXTS, sizeof(u8), 40962306a36Sopenharmony_ci GFP_KERNEL); 41062306a36Sopenharmony_ci dd->send_contexts = kcalloc(dd->num_send_contexts, 41162306a36Sopenharmony_ci sizeof(struct send_context_info), 41262306a36Sopenharmony_ci GFP_KERNEL); 41362306a36Sopenharmony_ci if (!dd->send_contexts || !dd->hw_to_sw) { 41462306a36Sopenharmony_ci kfree(dd->hw_to_sw); 41562306a36Sopenharmony_ci kfree(dd->send_contexts); 41662306a36Sopenharmony_ci free_credit_return(dd); 41762306a36Sopenharmony_ci return -ENOMEM; 41862306a36Sopenharmony_ci } 41962306a36Sopenharmony_ci 42062306a36Sopenharmony_ci /* hardware context map starts with invalid send context indices */ 42162306a36Sopenharmony_ci for (i = 0; i < TXE_NUM_CONTEXTS; i++) 42262306a36Sopenharmony_ci dd->hw_to_sw[i] = INVALID_SCI; 42362306a36Sopenharmony_ci 42462306a36Sopenharmony_ci /* 42562306a36Sopenharmony_ci * All send contexts have their credit sizes. Allocate credits 42662306a36Sopenharmony_ci * for each context one after another from the global space. 42762306a36Sopenharmony_ci */ 42862306a36Sopenharmony_ci context = 0; 42962306a36Sopenharmony_ci base = 1; 43062306a36Sopenharmony_ci for (i = 0; i < SC_MAX; i++) { 43162306a36Sopenharmony_ci struct sc_config_sizes *scs = &dd->sc_sizes[i]; 43262306a36Sopenharmony_ci 43362306a36Sopenharmony_ci for (j = 0; j < scs->count; j++) { 43462306a36Sopenharmony_ci struct send_context_info *sci = 43562306a36Sopenharmony_ci &dd->send_contexts[context]; 43662306a36Sopenharmony_ci sci->type = i; 43762306a36Sopenharmony_ci sci->base = base; 43862306a36Sopenharmony_ci sci->credits = scs->size; 43962306a36Sopenharmony_ci 44062306a36Sopenharmony_ci context++; 44162306a36Sopenharmony_ci base += scs->size; 44262306a36Sopenharmony_ci } 44362306a36Sopenharmony_ci } 44462306a36Sopenharmony_ci 44562306a36Sopenharmony_ci return 0; 44662306a36Sopenharmony_ci} 44762306a36Sopenharmony_ci 44862306a36Sopenharmony_ci/* 44962306a36Sopenharmony_ci * Allocate a software index and hardware context of the given type. 45062306a36Sopenharmony_ci * 45162306a36Sopenharmony_ci * Must be called with dd->sc_lock held. 45262306a36Sopenharmony_ci */ 45362306a36Sopenharmony_cistatic int sc_hw_alloc(struct hfi1_devdata *dd, int type, u32 *sw_index, 45462306a36Sopenharmony_ci u32 *hw_context) 45562306a36Sopenharmony_ci{ 45662306a36Sopenharmony_ci struct send_context_info *sci; 45762306a36Sopenharmony_ci u32 index; 45862306a36Sopenharmony_ci u32 context; 45962306a36Sopenharmony_ci 46062306a36Sopenharmony_ci for (index = 0, sci = &dd->send_contexts[0]; 46162306a36Sopenharmony_ci index < dd->num_send_contexts; index++, sci++) { 46262306a36Sopenharmony_ci if (sci->type == type && sci->allocated == 0) { 46362306a36Sopenharmony_ci sci->allocated = 1; 46462306a36Sopenharmony_ci /* use a 1:1 mapping, but make them non-equal */ 46562306a36Sopenharmony_ci context = chip_send_contexts(dd) - index - 1; 46662306a36Sopenharmony_ci dd->hw_to_sw[context] = index; 46762306a36Sopenharmony_ci *sw_index = index; 46862306a36Sopenharmony_ci *hw_context = context; 46962306a36Sopenharmony_ci return 0; /* success */ 47062306a36Sopenharmony_ci } 47162306a36Sopenharmony_ci } 47262306a36Sopenharmony_ci dd_dev_err(dd, "Unable to locate a free type %d send context\n", type); 47362306a36Sopenharmony_ci return -ENOSPC; 47462306a36Sopenharmony_ci} 47562306a36Sopenharmony_ci 47662306a36Sopenharmony_ci/* 47762306a36Sopenharmony_ci * Free the send context given by its software index. 47862306a36Sopenharmony_ci * 47962306a36Sopenharmony_ci * Must be called with dd->sc_lock held. 48062306a36Sopenharmony_ci */ 48162306a36Sopenharmony_cistatic void sc_hw_free(struct hfi1_devdata *dd, u32 sw_index, u32 hw_context) 48262306a36Sopenharmony_ci{ 48362306a36Sopenharmony_ci struct send_context_info *sci; 48462306a36Sopenharmony_ci 48562306a36Sopenharmony_ci sci = &dd->send_contexts[sw_index]; 48662306a36Sopenharmony_ci if (!sci->allocated) { 48762306a36Sopenharmony_ci dd_dev_err(dd, "%s: sw_index %u not allocated? hw_context %u\n", 48862306a36Sopenharmony_ci __func__, sw_index, hw_context); 48962306a36Sopenharmony_ci } 49062306a36Sopenharmony_ci sci->allocated = 0; 49162306a36Sopenharmony_ci dd->hw_to_sw[hw_context] = INVALID_SCI; 49262306a36Sopenharmony_ci} 49362306a36Sopenharmony_ci 49462306a36Sopenharmony_ci/* return the base context of a context in a group */ 49562306a36Sopenharmony_cistatic inline u32 group_context(u32 context, u32 group) 49662306a36Sopenharmony_ci{ 49762306a36Sopenharmony_ci return (context >> group) << group; 49862306a36Sopenharmony_ci} 49962306a36Sopenharmony_ci 50062306a36Sopenharmony_ci/* return the size of a group */ 50162306a36Sopenharmony_cistatic inline u32 group_size(u32 group) 50262306a36Sopenharmony_ci{ 50362306a36Sopenharmony_ci return 1 << group; 50462306a36Sopenharmony_ci} 50562306a36Sopenharmony_ci 50662306a36Sopenharmony_ci/* 50762306a36Sopenharmony_ci * Obtain the credit return addresses, kernel virtual and bus, for the 50862306a36Sopenharmony_ci * given sc. 50962306a36Sopenharmony_ci * 51062306a36Sopenharmony_ci * To understand this routine: 51162306a36Sopenharmony_ci * o va and dma are arrays of struct credit_return. One for each physical 51262306a36Sopenharmony_ci * send context, per NUMA. 51362306a36Sopenharmony_ci * o Each send context always looks in its relative location in a struct 51462306a36Sopenharmony_ci * credit_return for its credit return. 51562306a36Sopenharmony_ci * o Each send context in a group must have its return address CSR programmed 51662306a36Sopenharmony_ci * with the same value. Use the address of the first send context in the 51762306a36Sopenharmony_ci * group. 51862306a36Sopenharmony_ci */ 51962306a36Sopenharmony_cistatic void cr_group_addresses(struct send_context *sc, dma_addr_t *dma) 52062306a36Sopenharmony_ci{ 52162306a36Sopenharmony_ci u32 gc = group_context(sc->hw_context, sc->group); 52262306a36Sopenharmony_ci u32 index = sc->hw_context & 0x7; 52362306a36Sopenharmony_ci 52462306a36Sopenharmony_ci sc->hw_free = &sc->dd->cr_base[sc->node].va[gc].cr[index]; 52562306a36Sopenharmony_ci *dma = (unsigned long) 52662306a36Sopenharmony_ci &((struct credit_return *)sc->dd->cr_base[sc->node].dma)[gc]; 52762306a36Sopenharmony_ci} 52862306a36Sopenharmony_ci 52962306a36Sopenharmony_ci/* 53062306a36Sopenharmony_ci * Work queue function triggered in error interrupt routine for 53162306a36Sopenharmony_ci * kernel contexts. 53262306a36Sopenharmony_ci */ 53362306a36Sopenharmony_cistatic void sc_halted(struct work_struct *work) 53462306a36Sopenharmony_ci{ 53562306a36Sopenharmony_ci struct send_context *sc; 53662306a36Sopenharmony_ci 53762306a36Sopenharmony_ci sc = container_of(work, struct send_context, halt_work); 53862306a36Sopenharmony_ci sc_restart(sc); 53962306a36Sopenharmony_ci} 54062306a36Sopenharmony_ci 54162306a36Sopenharmony_ci/* 54262306a36Sopenharmony_ci * Calculate PIO block threshold for this send context using the given MTU. 54362306a36Sopenharmony_ci * Trigger a return when one MTU plus optional header of credits remain. 54462306a36Sopenharmony_ci * 54562306a36Sopenharmony_ci * Parameter mtu is in bytes. 54662306a36Sopenharmony_ci * Parameter hdrqentsize is in DWORDs. 54762306a36Sopenharmony_ci * 54862306a36Sopenharmony_ci * Return value is what to write into the CSR: trigger return when 54962306a36Sopenharmony_ci * unreturned credits pass this count. 55062306a36Sopenharmony_ci */ 55162306a36Sopenharmony_ciu32 sc_mtu_to_threshold(struct send_context *sc, u32 mtu, u32 hdrqentsize) 55262306a36Sopenharmony_ci{ 55362306a36Sopenharmony_ci u32 release_credits; 55462306a36Sopenharmony_ci u32 threshold; 55562306a36Sopenharmony_ci 55662306a36Sopenharmony_ci /* add in the header size, then divide by the PIO block size */ 55762306a36Sopenharmony_ci mtu += hdrqentsize << 2; 55862306a36Sopenharmony_ci release_credits = DIV_ROUND_UP(mtu, PIO_BLOCK_SIZE); 55962306a36Sopenharmony_ci 56062306a36Sopenharmony_ci /* check against this context's credits */ 56162306a36Sopenharmony_ci if (sc->credits <= release_credits) 56262306a36Sopenharmony_ci threshold = 1; 56362306a36Sopenharmony_ci else 56462306a36Sopenharmony_ci threshold = sc->credits - release_credits; 56562306a36Sopenharmony_ci 56662306a36Sopenharmony_ci return threshold; 56762306a36Sopenharmony_ci} 56862306a36Sopenharmony_ci 56962306a36Sopenharmony_ci/* 57062306a36Sopenharmony_ci * Calculate credit threshold in terms of percent of the allocated credits. 57162306a36Sopenharmony_ci * Trigger when unreturned credits equal or exceed the percentage of the whole. 57262306a36Sopenharmony_ci * 57362306a36Sopenharmony_ci * Return value is what to write into the CSR: trigger return when 57462306a36Sopenharmony_ci * unreturned credits pass this count. 57562306a36Sopenharmony_ci */ 57662306a36Sopenharmony_ciu32 sc_percent_to_threshold(struct send_context *sc, u32 percent) 57762306a36Sopenharmony_ci{ 57862306a36Sopenharmony_ci return (sc->credits * percent) / 100; 57962306a36Sopenharmony_ci} 58062306a36Sopenharmony_ci 58162306a36Sopenharmony_ci/* 58262306a36Sopenharmony_ci * Set the credit return threshold. 58362306a36Sopenharmony_ci */ 58462306a36Sopenharmony_civoid sc_set_cr_threshold(struct send_context *sc, u32 new_threshold) 58562306a36Sopenharmony_ci{ 58662306a36Sopenharmony_ci unsigned long flags; 58762306a36Sopenharmony_ci u32 old_threshold; 58862306a36Sopenharmony_ci int force_return = 0; 58962306a36Sopenharmony_ci 59062306a36Sopenharmony_ci spin_lock_irqsave(&sc->credit_ctrl_lock, flags); 59162306a36Sopenharmony_ci 59262306a36Sopenharmony_ci old_threshold = (sc->credit_ctrl >> 59362306a36Sopenharmony_ci SC(CREDIT_CTRL_THRESHOLD_SHIFT)) 59462306a36Sopenharmony_ci & SC(CREDIT_CTRL_THRESHOLD_MASK); 59562306a36Sopenharmony_ci 59662306a36Sopenharmony_ci if (new_threshold != old_threshold) { 59762306a36Sopenharmony_ci sc->credit_ctrl = 59862306a36Sopenharmony_ci (sc->credit_ctrl 59962306a36Sopenharmony_ci & ~SC(CREDIT_CTRL_THRESHOLD_SMASK)) 60062306a36Sopenharmony_ci | ((new_threshold 60162306a36Sopenharmony_ci & SC(CREDIT_CTRL_THRESHOLD_MASK)) 60262306a36Sopenharmony_ci << SC(CREDIT_CTRL_THRESHOLD_SHIFT)); 60362306a36Sopenharmony_ci write_kctxt_csr(sc->dd, sc->hw_context, 60462306a36Sopenharmony_ci SC(CREDIT_CTRL), sc->credit_ctrl); 60562306a36Sopenharmony_ci 60662306a36Sopenharmony_ci /* force a credit return on change to avoid a possible stall */ 60762306a36Sopenharmony_ci force_return = 1; 60862306a36Sopenharmony_ci } 60962306a36Sopenharmony_ci 61062306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->credit_ctrl_lock, flags); 61162306a36Sopenharmony_ci 61262306a36Sopenharmony_ci if (force_return) 61362306a36Sopenharmony_ci sc_return_credits(sc); 61462306a36Sopenharmony_ci} 61562306a36Sopenharmony_ci 61662306a36Sopenharmony_ci/* 61762306a36Sopenharmony_ci * set_pio_integrity 61862306a36Sopenharmony_ci * 61962306a36Sopenharmony_ci * Set the CHECK_ENABLE register for the send context 'sc'. 62062306a36Sopenharmony_ci */ 62162306a36Sopenharmony_civoid set_pio_integrity(struct send_context *sc) 62262306a36Sopenharmony_ci{ 62362306a36Sopenharmony_ci struct hfi1_devdata *dd = sc->dd; 62462306a36Sopenharmony_ci u32 hw_context = sc->hw_context; 62562306a36Sopenharmony_ci int type = sc->type; 62662306a36Sopenharmony_ci 62762306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, 62862306a36Sopenharmony_ci SC(CHECK_ENABLE), 62962306a36Sopenharmony_ci hfi1_pkt_default_send_ctxt_mask(dd, type)); 63062306a36Sopenharmony_ci} 63162306a36Sopenharmony_ci 63262306a36Sopenharmony_cistatic u32 get_buffers_allocated(struct send_context *sc) 63362306a36Sopenharmony_ci{ 63462306a36Sopenharmony_ci int cpu; 63562306a36Sopenharmony_ci u32 ret = 0; 63662306a36Sopenharmony_ci 63762306a36Sopenharmony_ci for_each_possible_cpu(cpu) 63862306a36Sopenharmony_ci ret += *per_cpu_ptr(sc->buffers_allocated, cpu); 63962306a36Sopenharmony_ci return ret; 64062306a36Sopenharmony_ci} 64162306a36Sopenharmony_ci 64262306a36Sopenharmony_cistatic void reset_buffers_allocated(struct send_context *sc) 64362306a36Sopenharmony_ci{ 64462306a36Sopenharmony_ci int cpu; 64562306a36Sopenharmony_ci 64662306a36Sopenharmony_ci for_each_possible_cpu(cpu) 64762306a36Sopenharmony_ci (*per_cpu_ptr(sc->buffers_allocated, cpu)) = 0; 64862306a36Sopenharmony_ci} 64962306a36Sopenharmony_ci 65062306a36Sopenharmony_ci/* 65162306a36Sopenharmony_ci * Allocate a NUMA relative send context structure of the given type along 65262306a36Sopenharmony_ci * with a HW context. 65362306a36Sopenharmony_ci */ 65462306a36Sopenharmony_cistruct send_context *sc_alloc(struct hfi1_devdata *dd, int type, 65562306a36Sopenharmony_ci uint hdrqentsize, int numa) 65662306a36Sopenharmony_ci{ 65762306a36Sopenharmony_ci struct send_context_info *sci; 65862306a36Sopenharmony_ci struct send_context *sc = NULL; 65962306a36Sopenharmony_ci dma_addr_t dma; 66062306a36Sopenharmony_ci unsigned long flags; 66162306a36Sopenharmony_ci u64 reg; 66262306a36Sopenharmony_ci u32 thresh; 66362306a36Sopenharmony_ci u32 sw_index; 66462306a36Sopenharmony_ci u32 hw_context; 66562306a36Sopenharmony_ci int ret; 66662306a36Sopenharmony_ci u8 opval, opmask; 66762306a36Sopenharmony_ci 66862306a36Sopenharmony_ci /* do not allocate while frozen */ 66962306a36Sopenharmony_ci if (dd->flags & HFI1_FROZEN) 67062306a36Sopenharmony_ci return NULL; 67162306a36Sopenharmony_ci 67262306a36Sopenharmony_ci sc = kzalloc_node(sizeof(*sc), GFP_KERNEL, numa); 67362306a36Sopenharmony_ci if (!sc) 67462306a36Sopenharmony_ci return NULL; 67562306a36Sopenharmony_ci 67662306a36Sopenharmony_ci sc->buffers_allocated = alloc_percpu(u32); 67762306a36Sopenharmony_ci if (!sc->buffers_allocated) { 67862306a36Sopenharmony_ci kfree(sc); 67962306a36Sopenharmony_ci dd_dev_err(dd, 68062306a36Sopenharmony_ci "Cannot allocate buffers_allocated per cpu counters\n" 68162306a36Sopenharmony_ci ); 68262306a36Sopenharmony_ci return NULL; 68362306a36Sopenharmony_ci } 68462306a36Sopenharmony_ci 68562306a36Sopenharmony_ci spin_lock_irqsave(&dd->sc_lock, flags); 68662306a36Sopenharmony_ci ret = sc_hw_alloc(dd, type, &sw_index, &hw_context); 68762306a36Sopenharmony_ci if (ret) { 68862306a36Sopenharmony_ci spin_unlock_irqrestore(&dd->sc_lock, flags); 68962306a36Sopenharmony_ci free_percpu(sc->buffers_allocated); 69062306a36Sopenharmony_ci kfree(sc); 69162306a36Sopenharmony_ci return NULL; 69262306a36Sopenharmony_ci } 69362306a36Sopenharmony_ci 69462306a36Sopenharmony_ci sci = &dd->send_contexts[sw_index]; 69562306a36Sopenharmony_ci sci->sc = sc; 69662306a36Sopenharmony_ci 69762306a36Sopenharmony_ci sc->dd = dd; 69862306a36Sopenharmony_ci sc->node = numa; 69962306a36Sopenharmony_ci sc->type = type; 70062306a36Sopenharmony_ci spin_lock_init(&sc->alloc_lock); 70162306a36Sopenharmony_ci spin_lock_init(&sc->release_lock); 70262306a36Sopenharmony_ci spin_lock_init(&sc->credit_ctrl_lock); 70362306a36Sopenharmony_ci seqlock_init(&sc->waitlock); 70462306a36Sopenharmony_ci INIT_LIST_HEAD(&sc->piowait); 70562306a36Sopenharmony_ci INIT_WORK(&sc->halt_work, sc_halted); 70662306a36Sopenharmony_ci init_waitqueue_head(&sc->halt_wait); 70762306a36Sopenharmony_ci 70862306a36Sopenharmony_ci /* grouping is always single context for now */ 70962306a36Sopenharmony_ci sc->group = 0; 71062306a36Sopenharmony_ci 71162306a36Sopenharmony_ci sc->sw_index = sw_index; 71262306a36Sopenharmony_ci sc->hw_context = hw_context; 71362306a36Sopenharmony_ci cr_group_addresses(sc, &dma); 71462306a36Sopenharmony_ci sc->credits = sci->credits; 71562306a36Sopenharmony_ci sc->size = sc->credits * PIO_BLOCK_SIZE; 71662306a36Sopenharmony_ci 71762306a36Sopenharmony_ci/* PIO Send Memory Address details */ 71862306a36Sopenharmony_ci#define PIO_ADDR_CONTEXT_MASK 0xfful 71962306a36Sopenharmony_ci#define PIO_ADDR_CONTEXT_SHIFT 16 72062306a36Sopenharmony_ci sc->base_addr = dd->piobase + ((hw_context & PIO_ADDR_CONTEXT_MASK) 72162306a36Sopenharmony_ci << PIO_ADDR_CONTEXT_SHIFT); 72262306a36Sopenharmony_ci 72362306a36Sopenharmony_ci /* set base and credits */ 72462306a36Sopenharmony_ci reg = ((sci->credits & SC(CTRL_CTXT_DEPTH_MASK)) 72562306a36Sopenharmony_ci << SC(CTRL_CTXT_DEPTH_SHIFT)) 72662306a36Sopenharmony_ci | ((sci->base & SC(CTRL_CTXT_BASE_MASK)) 72762306a36Sopenharmony_ci << SC(CTRL_CTXT_BASE_SHIFT)); 72862306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CTRL), reg); 72962306a36Sopenharmony_ci 73062306a36Sopenharmony_ci set_pio_integrity(sc); 73162306a36Sopenharmony_ci 73262306a36Sopenharmony_ci /* unmask all errors */ 73362306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(ERR_MASK), (u64)-1); 73462306a36Sopenharmony_ci 73562306a36Sopenharmony_ci /* set the default partition key */ 73662306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CHECK_PARTITION_KEY), 73762306a36Sopenharmony_ci (SC(CHECK_PARTITION_KEY_VALUE_MASK) & 73862306a36Sopenharmony_ci DEFAULT_PKEY) << 73962306a36Sopenharmony_ci SC(CHECK_PARTITION_KEY_VALUE_SHIFT)); 74062306a36Sopenharmony_ci 74162306a36Sopenharmony_ci /* per context type checks */ 74262306a36Sopenharmony_ci if (type == SC_USER) { 74362306a36Sopenharmony_ci opval = USER_OPCODE_CHECK_VAL; 74462306a36Sopenharmony_ci opmask = USER_OPCODE_CHECK_MASK; 74562306a36Sopenharmony_ci } else { 74662306a36Sopenharmony_ci opval = OPCODE_CHECK_VAL_DISABLED; 74762306a36Sopenharmony_ci opmask = OPCODE_CHECK_MASK_DISABLED; 74862306a36Sopenharmony_ci } 74962306a36Sopenharmony_ci 75062306a36Sopenharmony_ci /* set the send context check opcode mask and value */ 75162306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CHECK_OPCODE), 75262306a36Sopenharmony_ci ((u64)opmask << SC(CHECK_OPCODE_MASK_SHIFT)) | 75362306a36Sopenharmony_ci ((u64)opval << SC(CHECK_OPCODE_VALUE_SHIFT))); 75462306a36Sopenharmony_ci 75562306a36Sopenharmony_ci /* set up credit return */ 75662306a36Sopenharmony_ci reg = dma & SC(CREDIT_RETURN_ADDR_ADDRESS_SMASK); 75762306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CREDIT_RETURN_ADDR), reg); 75862306a36Sopenharmony_ci 75962306a36Sopenharmony_ci /* 76062306a36Sopenharmony_ci * Calculate the initial credit return threshold. 76162306a36Sopenharmony_ci * 76262306a36Sopenharmony_ci * For Ack contexts, set a threshold for half the credits. 76362306a36Sopenharmony_ci * For User contexts use the given percentage. This has been 76462306a36Sopenharmony_ci * sanitized on driver start-up. 76562306a36Sopenharmony_ci * For Kernel contexts, use the default MTU plus a header 76662306a36Sopenharmony_ci * or half the credits, whichever is smaller. This should 76762306a36Sopenharmony_ci * work for both the 3-deep buffering allocation and the 76862306a36Sopenharmony_ci * pooling allocation. 76962306a36Sopenharmony_ci */ 77062306a36Sopenharmony_ci if (type == SC_ACK) { 77162306a36Sopenharmony_ci thresh = sc_percent_to_threshold(sc, 50); 77262306a36Sopenharmony_ci } else if (type == SC_USER) { 77362306a36Sopenharmony_ci thresh = sc_percent_to_threshold(sc, 77462306a36Sopenharmony_ci user_credit_return_threshold); 77562306a36Sopenharmony_ci } else { /* kernel */ 77662306a36Sopenharmony_ci thresh = min(sc_percent_to_threshold(sc, 50), 77762306a36Sopenharmony_ci sc_mtu_to_threshold(sc, hfi1_max_mtu, 77862306a36Sopenharmony_ci hdrqentsize)); 77962306a36Sopenharmony_ci } 78062306a36Sopenharmony_ci reg = thresh << SC(CREDIT_CTRL_THRESHOLD_SHIFT); 78162306a36Sopenharmony_ci /* add in early return */ 78262306a36Sopenharmony_ci if (type == SC_USER && HFI1_CAP_IS_USET(EARLY_CREDIT_RETURN)) 78362306a36Sopenharmony_ci reg |= SC(CREDIT_CTRL_EARLY_RETURN_SMASK); 78462306a36Sopenharmony_ci else if (HFI1_CAP_IS_KSET(EARLY_CREDIT_RETURN)) /* kernel, ack */ 78562306a36Sopenharmony_ci reg |= SC(CREDIT_CTRL_EARLY_RETURN_SMASK); 78662306a36Sopenharmony_ci 78762306a36Sopenharmony_ci /* set up write-through credit_ctrl */ 78862306a36Sopenharmony_ci sc->credit_ctrl = reg; 78962306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CREDIT_CTRL), reg); 79062306a36Sopenharmony_ci 79162306a36Sopenharmony_ci /* User send contexts should not allow sending on VL15 */ 79262306a36Sopenharmony_ci if (type == SC_USER) { 79362306a36Sopenharmony_ci reg = 1ULL << 15; 79462306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CHECK_VL), reg); 79562306a36Sopenharmony_ci } 79662306a36Sopenharmony_ci 79762306a36Sopenharmony_ci spin_unlock_irqrestore(&dd->sc_lock, flags); 79862306a36Sopenharmony_ci 79962306a36Sopenharmony_ci /* 80062306a36Sopenharmony_ci * Allocate shadow ring to track outstanding PIO buffers _after_ 80162306a36Sopenharmony_ci * unlocking. We don't know the size until the lock is held and 80262306a36Sopenharmony_ci * we can't allocate while the lock is held. No one is using 80362306a36Sopenharmony_ci * the context yet, so allocate it now. 80462306a36Sopenharmony_ci * 80562306a36Sopenharmony_ci * User contexts do not get a shadow ring. 80662306a36Sopenharmony_ci */ 80762306a36Sopenharmony_ci if (type != SC_USER) { 80862306a36Sopenharmony_ci /* 80962306a36Sopenharmony_ci * Size the shadow ring 1 larger than the number of credits 81062306a36Sopenharmony_ci * so head == tail can mean empty. 81162306a36Sopenharmony_ci */ 81262306a36Sopenharmony_ci sc->sr_size = sci->credits + 1; 81362306a36Sopenharmony_ci sc->sr = kcalloc_node(sc->sr_size, 81462306a36Sopenharmony_ci sizeof(union pio_shadow_ring), 81562306a36Sopenharmony_ci GFP_KERNEL, numa); 81662306a36Sopenharmony_ci if (!sc->sr) { 81762306a36Sopenharmony_ci sc_free(sc); 81862306a36Sopenharmony_ci return NULL; 81962306a36Sopenharmony_ci } 82062306a36Sopenharmony_ci } 82162306a36Sopenharmony_ci 82262306a36Sopenharmony_ci hfi1_cdbg(PIO, 82362306a36Sopenharmony_ci "Send context %u(%u) %s group %u credits %u credit_ctrl 0x%llx threshold %u", 82462306a36Sopenharmony_ci sw_index, 82562306a36Sopenharmony_ci hw_context, 82662306a36Sopenharmony_ci sc_type_name(type), 82762306a36Sopenharmony_ci sc->group, 82862306a36Sopenharmony_ci sc->credits, 82962306a36Sopenharmony_ci sc->credit_ctrl, 83062306a36Sopenharmony_ci thresh); 83162306a36Sopenharmony_ci 83262306a36Sopenharmony_ci return sc; 83362306a36Sopenharmony_ci} 83462306a36Sopenharmony_ci 83562306a36Sopenharmony_ci/* free a per-NUMA send context structure */ 83662306a36Sopenharmony_civoid sc_free(struct send_context *sc) 83762306a36Sopenharmony_ci{ 83862306a36Sopenharmony_ci struct hfi1_devdata *dd; 83962306a36Sopenharmony_ci unsigned long flags; 84062306a36Sopenharmony_ci u32 sw_index; 84162306a36Sopenharmony_ci u32 hw_context; 84262306a36Sopenharmony_ci 84362306a36Sopenharmony_ci if (!sc) 84462306a36Sopenharmony_ci return; 84562306a36Sopenharmony_ci 84662306a36Sopenharmony_ci sc->flags |= SCF_IN_FREE; /* ensure no restarts */ 84762306a36Sopenharmony_ci dd = sc->dd; 84862306a36Sopenharmony_ci if (!list_empty(&sc->piowait)) 84962306a36Sopenharmony_ci dd_dev_err(dd, "piowait list not empty!\n"); 85062306a36Sopenharmony_ci sw_index = sc->sw_index; 85162306a36Sopenharmony_ci hw_context = sc->hw_context; 85262306a36Sopenharmony_ci sc_disable(sc); /* make sure the HW is disabled */ 85362306a36Sopenharmony_ci flush_work(&sc->halt_work); 85462306a36Sopenharmony_ci 85562306a36Sopenharmony_ci spin_lock_irqsave(&dd->sc_lock, flags); 85662306a36Sopenharmony_ci dd->send_contexts[sw_index].sc = NULL; 85762306a36Sopenharmony_ci 85862306a36Sopenharmony_ci /* clear/disable all registers set in sc_alloc */ 85962306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CTRL), 0); 86062306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CHECK_ENABLE), 0); 86162306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(ERR_MASK), 0); 86262306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CHECK_PARTITION_KEY), 0); 86362306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CHECK_OPCODE), 0); 86462306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CREDIT_RETURN_ADDR), 0); 86562306a36Sopenharmony_ci write_kctxt_csr(dd, hw_context, SC(CREDIT_CTRL), 0); 86662306a36Sopenharmony_ci 86762306a36Sopenharmony_ci /* release the index and context for re-use */ 86862306a36Sopenharmony_ci sc_hw_free(dd, sw_index, hw_context); 86962306a36Sopenharmony_ci spin_unlock_irqrestore(&dd->sc_lock, flags); 87062306a36Sopenharmony_ci 87162306a36Sopenharmony_ci kfree(sc->sr); 87262306a36Sopenharmony_ci free_percpu(sc->buffers_allocated); 87362306a36Sopenharmony_ci kfree(sc); 87462306a36Sopenharmony_ci} 87562306a36Sopenharmony_ci 87662306a36Sopenharmony_ci/* disable the context */ 87762306a36Sopenharmony_civoid sc_disable(struct send_context *sc) 87862306a36Sopenharmony_ci{ 87962306a36Sopenharmony_ci u64 reg; 88062306a36Sopenharmony_ci struct pio_buf *pbuf; 88162306a36Sopenharmony_ci LIST_HEAD(wake_list); 88262306a36Sopenharmony_ci 88362306a36Sopenharmony_ci if (!sc) 88462306a36Sopenharmony_ci return; 88562306a36Sopenharmony_ci 88662306a36Sopenharmony_ci /* do all steps, even if already disabled */ 88762306a36Sopenharmony_ci spin_lock_irq(&sc->alloc_lock); 88862306a36Sopenharmony_ci reg = read_kctxt_csr(sc->dd, sc->hw_context, SC(CTRL)); 88962306a36Sopenharmony_ci reg &= ~SC(CTRL_CTXT_ENABLE_SMASK); 89062306a36Sopenharmony_ci sc->flags &= ~SCF_ENABLED; 89162306a36Sopenharmony_ci sc_wait_for_packet_egress(sc, 1); 89262306a36Sopenharmony_ci write_kctxt_csr(sc->dd, sc->hw_context, SC(CTRL), reg); 89362306a36Sopenharmony_ci 89462306a36Sopenharmony_ci /* 89562306a36Sopenharmony_ci * Flush any waiters. Once the context is disabled, 89662306a36Sopenharmony_ci * credit return interrupts are stopped (although there 89762306a36Sopenharmony_ci * could be one in-process when the context is disabled). 89862306a36Sopenharmony_ci * Wait one microsecond for any lingering interrupts, then 89962306a36Sopenharmony_ci * proceed with the flush. 90062306a36Sopenharmony_ci */ 90162306a36Sopenharmony_ci udelay(1); 90262306a36Sopenharmony_ci spin_lock(&sc->release_lock); 90362306a36Sopenharmony_ci if (sc->sr) { /* this context has a shadow ring */ 90462306a36Sopenharmony_ci while (sc->sr_tail != sc->sr_head) { 90562306a36Sopenharmony_ci pbuf = &sc->sr[sc->sr_tail].pbuf; 90662306a36Sopenharmony_ci if (pbuf->cb) 90762306a36Sopenharmony_ci (*pbuf->cb)(pbuf->arg, PRC_SC_DISABLE); 90862306a36Sopenharmony_ci sc->sr_tail++; 90962306a36Sopenharmony_ci if (sc->sr_tail >= sc->sr_size) 91062306a36Sopenharmony_ci sc->sr_tail = 0; 91162306a36Sopenharmony_ci } 91262306a36Sopenharmony_ci } 91362306a36Sopenharmony_ci spin_unlock(&sc->release_lock); 91462306a36Sopenharmony_ci 91562306a36Sopenharmony_ci write_seqlock(&sc->waitlock); 91662306a36Sopenharmony_ci list_splice_init(&sc->piowait, &wake_list); 91762306a36Sopenharmony_ci write_sequnlock(&sc->waitlock); 91862306a36Sopenharmony_ci while (!list_empty(&wake_list)) { 91962306a36Sopenharmony_ci struct iowait *wait; 92062306a36Sopenharmony_ci struct rvt_qp *qp; 92162306a36Sopenharmony_ci struct hfi1_qp_priv *priv; 92262306a36Sopenharmony_ci 92362306a36Sopenharmony_ci wait = list_first_entry(&wake_list, struct iowait, list); 92462306a36Sopenharmony_ci qp = iowait_to_qp(wait); 92562306a36Sopenharmony_ci priv = qp->priv; 92662306a36Sopenharmony_ci list_del_init(&priv->s_iowait.list); 92762306a36Sopenharmony_ci priv->s_iowait.lock = NULL; 92862306a36Sopenharmony_ci hfi1_qp_wakeup(qp, RVT_S_WAIT_PIO | HFI1_S_WAIT_PIO_DRAIN); 92962306a36Sopenharmony_ci } 93062306a36Sopenharmony_ci 93162306a36Sopenharmony_ci spin_unlock_irq(&sc->alloc_lock); 93262306a36Sopenharmony_ci} 93362306a36Sopenharmony_ci 93462306a36Sopenharmony_ci/* return SendEgressCtxtStatus.PacketOccupancy */ 93562306a36Sopenharmony_cistatic u64 packet_occupancy(u64 reg) 93662306a36Sopenharmony_ci{ 93762306a36Sopenharmony_ci return (reg & 93862306a36Sopenharmony_ci SEND_EGRESS_CTXT_STATUS_CTXT_EGRESS_PACKET_OCCUPANCY_SMASK) 93962306a36Sopenharmony_ci >> SEND_EGRESS_CTXT_STATUS_CTXT_EGRESS_PACKET_OCCUPANCY_SHIFT; 94062306a36Sopenharmony_ci} 94162306a36Sopenharmony_ci 94262306a36Sopenharmony_ci/* is egress halted on the context? */ 94362306a36Sopenharmony_cistatic bool egress_halted(u64 reg) 94462306a36Sopenharmony_ci{ 94562306a36Sopenharmony_ci return !!(reg & SEND_EGRESS_CTXT_STATUS_CTXT_EGRESS_HALT_STATUS_SMASK); 94662306a36Sopenharmony_ci} 94762306a36Sopenharmony_ci 94862306a36Sopenharmony_ci/* is the send context halted? */ 94962306a36Sopenharmony_cistatic bool is_sc_halted(struct hfi1_devdata *dd, u32 hw_context) 95062306a36Sopenharmony_ci{ 95162306a36Sopenharmony_ci return !!(read_kctxt_csr(dd, hw_context, SC(STATUS)) & 95262306a36Sopenharmony_ci SC(STATUS_CTXT_HALTED_SMASK)); 95362306a36Sopenharmony_ci} 95462306a36Sopenharmony_ci 95562306a36Sopenharmony_ci/** 95662306a36Sopenharmony_ci * sc_wait_for_packet_egress - wait for packet 95762306a36Sopenharmony_ci * @sc: valid send context 95862306a36Sopenharmony_ci * @pause: wait for credit return 95962306a36Sopenharmony_ci * 96062306a36Sopenharmony_ci * Wait for packet egress, optionally pause for credit return 96162306a36Sopenharmony_ci * 96262306a36Sopenharmony_ci * Egress halt and Context halt are not necessarily the same thing, so 96362306a36Sopenharmony_ci * check for both. 96462306a36Sopenharmony_ci * 96562306a36Sopenharmony_ci * NOTE: The context halt bit may not be set immediately. Because of this, 96662306a36Sopenharmony_ci * it is necessary to check the SW SFC_HALTED bit (set in the IRQ) and the HW 96762306a36Sopenharmony_ci * context bit to determine if the context is halted. 96862306a36Sopenharmony_ci */ 96962306a36Sopenharmony_cistatic void sc_wait_for_packet_egress(struct send_context *sc, int pause) 97062306a36Sopenharmony_ci{ 97162306a36Sopenharmony_ci struct hfi1_devdata *dd = sc->dd; 97262306a36Sopenharmony_ci u64 reg = 0; 97362306a36Sopenharmony_ci u64 reg_prev; 97462306a36Sopenharmony_ci u32 loop = 0; 97562306a36Sopenharmony_ci 97662306a36Sopenharmony_ci while (1) { 97762306a36Sopenharmony_ci reg_prev = reg; 97862306a36Sopenharmony_ci reg = read_csr(dd, sc->hw_context * 8 + 97962306a36Sopenharmony_ci SEND_EGRESS_CTXT_STATUS); 98062306a36Sopenharmony_ci /* done if any halt bits, SW or HW are set */ 98162306a36Sopenharmony_ci if (sc->flags & SCF_HALTED || 98262306a36Sopenharmony_ci is_sc_halted(dd, sc->hw_context) || egress_halted(reg)) 98362306a36Sopenharmony_ci break; 98462306a36Sopenharmony_ci reg = packet_occupancy(reg); 98562306a36Sopenharmony_ci if (reg == 0) 98662306a36Sopenharmony_ci break; 98762306a36Sopenharmony_ci /* counter is reset if occupancy count changes */ 98862306a36Sopenharmony_ci if (reg != reg_prev) 98962306a36Sopenharmony_ci loop = 0; 99062306a36Sopenharmony_ci if (loop > 50000) { 99162306a36Sopenharmony_ci /* timed out - bounce the link */ 99262306a36Sopenharmony_ci dd_dev_err(dd, 99362306a36Sopenharmony_ci "%s: context %u(%u) timeout waiting for packets to egress, remaining count %u, bouncing link\n", 99462306a36Sopenharmony_ci __func__, sc->sw_index, 99562306a36Sopenharmony_ci sc->hw_context, (u32)reg); 99662306a36Sopenharmony_ci queue_work(dd->pport->link_wq, 99762306a36Sopenharmony_ci &dd->pport->link_bounce_work); 99862306a36Sopenharmony_ci break; 99962306a36Sopenharmony_ci } 100062306a36Sopenharmony_ci loop++; 100162306a36Sopenharmony_ci udelay(1); 100262306a36Sopenharmony_ci } 100362306a36Sopenharmony_ci 100462306a36Sopenharmony_ci if (pause) 100562306a36Sopenharmony_ci /* Add additional delay to ensure chip returns all credits */ 100662306a36Sopenharmony_ci pause_for_credit_return(dd); 100762306a36Sopenharmony_ci} 100862306a36Sopenharmony_ci 100962306a36Sopenharmony_civoid sc_wait(struct hfi1_devdata *dd) 101062306a36Sopenharmony_ci{ 101162306a36Sopenharmony_ci int i; 101262306a36Sopenharmony_ci 101362306a36Sopenharmony_ci for (i = 0; i < dd->num_send_contexts; i++) { 101462306a36Sopenharmony_ci struct send_context *sc = dd->send_contexts[i].sc; 101562306a36Sopenharmony_ci 101662306a36Sopenharmony_ci if (!sc) 101762306a36Sopenharmony_ci continue; 101862306a36Sopenharmony_ci sc_wait_for_packet_egress(sc, 0); 101962306a36Sopenharmony_ci } 102062306a36Sopenharmony_ci} 102162306a36Sopenharmony_ci 102262306a36Sopenharmony_ci/* 102362306a36Sopenharmony_ci * Restart a context after it has been halted due to error. 102462306a36Sopenharmony_ci * 102562306a36Sopenharmony_ci * If the first step fails - wait for the halt to be asserted, return early. 102662306a36Sopenharmony_ci * Otherwise complain about timeouts but keep going. 102762306a36Sopenharmony_ci * 102862306a36Sopenharmony_ci * It is expected that allocations (enabled flag bit) have been shut off 102962306a36Sopenharmony_ci * already (only applies to kernel contexts). 103062306a36Sopenharmony_ci */ 103162306a36Sopenharmony_ciint sc_restart(struct send_context *sc) 103262306a36Sopenharmony_ci{ 103362306a36Sopenharmony_ci struct hfi1_devdata *dd = sc->dd; 103462306a36Sopenharmony_ci u64 reg; 103562306a36Sopenharmony_ci u32 loop; 103662306a36Sopenharmony_ci int count; 103762306a36Sopenharmony_ci 103862306a36Sopenharmony_ci /* bounce off if not halted, or being free'd */ 103962306a36Sopenharmony_ci if (!(sc->flags & SCF_HALTED) || (sc->flags & SCF_IN_FREE)) 104062306a36Sopenharmony_ci return -EINVAL; 104162306a36Sopenharmony_ci 104262306a36Sopenharmony_ci dd_dev_info(dd, "restarting send context %u(%u)\n", sc->sw_index, 104362306a36Sopenharmony_ci sc->hw_context); 104462306a36Sopenharmony_ci 104562306a36Sopenharmony_ci /* 104662306a36Sopenharmony_ci * Step 1: Wait for the context to actually halt. 104762306a36Sopenharmony_ci * 104862306a36Sopenharmony_ci * The error interrupt is asynchronous to actually setting halt 104962306a36Sopenharmony_ci * on the context. 105062306a36Sopenharmony_ci */ 105162306a36Sopenharmony_ci loop = 0; 105262306a36Sopenharmony_ci while (1) { 105362306a36Sopenharmony_ci reg = read_kctxt_csr(dd, sc->hw_context, SC(STATUS)); 105462306a36Sopenharmony_ci if (reg & SC(STATUS_CTXT_HALTED_SMASK)) 105562306a36Sopenharmony_ci break; 105662306a36Sopenharmony_ci if (loop > 100) { 105762306a36Sopenharmony_ci dd_dev_err(dd, "%s: context %u(%u) not halting, skipping\n", 105862306a36Sopenharmony_ci __func__, sc->sw_index, sc->hw_context); 105962306a36Sopenharmony_ci return -ETIME; 106062306a36Sopenharmony_ci } 106162306a36Sopenharmony_ci loop++; 106262306a36Sopenharmony_ci udelay(1); 106362306a36Sopenharmony_ci } 106462306a36Sopenharmony_ci 106562306a36Sopenharmony_ci /* 106662306a36Sopenharmony_ci * Step 2: Ensure no users are still trying to write to PIO. 106762306a36Sopenharmony_ci * 106862306a36Sopenharmony_ci * For kernel contexts, we have already turned off buffer allocation. 106962306a36Sopenharmony_ci * Now wait for the buffer count to go to zero. 107062306a36Sopenharmony_ci * 107162306a36Sopenharmony_ci * For user contexts, the user handling code has cut off write access 107262306a36Sopenharmony_ci * to the context's PIO pages before calling this routine and will 107362306a36Sopenharmony_ci * restore write access after this routine returns. 107462306a36Sopenharmony_ci */ 107562306a36Sopenharmony_ci if (sc->type != SC_USER) { 107662306a36Sopenharmony_ci /* kernel context */ 107762306a36Sopenharmony_ci loop = 0; 107862306a36Sopenharmony_ci while (1) { 107962306a36Sopenharmony_ci count = get_buffers_allocated(sc); 108062306a36Sopenharmony_ci if (count == 0) 108162306a36Sopenharmony_ci break; 108262306a36Sopenharmony_ci if (loop > 100) { 108362306a36Sopenharmony_ci dd_dev_err(dd, 108462306a36Sopenharmony_ci "%s: context %u(%u) timeout waiting for PIO buffers to zero, remaining %d\n", 108562306a36Sopenharmony_ci __func__, sc->sw_index, 108662306a36Sopenharmony_ci sc->hw_context, count); 108762306a36Sopenharmony_ci } 108862306a36Sopenharmony_ci loop++; 108962306a36Sopenharmony_ci udelay(1); 109062306a36Sopenharmony_ci } 109162306a36Sopenharmony_ci } 109262306a36Sopenharmony_ci 109362306a36Sopenharmony_ci /* 109462306a36Sopenharmony_ci * Step 3: Wait for all packets to egress. 109562306a36Sopenharmony_ci * This is done while disabling the send context 109662306a36Sopenharmony_ci * 109762306a36Sopenharmony_ci * Step 4: Disable the context 109862306a36Sopenharmony_ci * 109962306a36Sopenharmony_ci * This is a superset of the halt. After the disable, the 110062306a36Sopenharmony_ci * errors can be cleared. 110162306a36Sopenharmony_ci */ 110262306a36Sopenharmony_ci sc_disable(sc); 110362306a36Sopenharmony_ci 110462306a36Sopenharmony_ci /* 110562306a36Sopenharmony_ci * Step 5: Enable the context 110662306a36Sopenharmony_ci * 110762306a36Sopenharmony_ci * This enable will clear the halted flag and per-send context 110862306a36Sopenharmony_ci * error flags. 110962306a36Sopenharmony_ci */ 111062306a36Sopenharmony_ci return sc_enable(sc); 111162306a36Sopenharmony_ci} 111262306a36Sopenharmony_ci 111362306a36Sopenharmony_ci/* 111462306a36Sopenharmony_ci * PIO freeze processing. To be called after the TXE block is fully frozen. 111562306a36Sopenharmony_ci * Go through all frozen send contexts and disable them. The contexts are 111662306a36Sopenharmony_ci * already stopped by the freeze. 111762306a36Sopenharmony_ci */ 111862306a36Sopenharmony_civoid pio_freeze(struct hfi1_devdata *dd) 111962306a36Sopenharmony_ci{ 112062306a36Sopenharmony_ci struct send_context *sc; 112162306a36Sopenharmony_ci int i; 112262306a36Sopenharmony_ci 112362306a36Sopenharmony_ci for (i = 0; i < dd->num_send_contexts; i++) { 112462306a36Sopenharmony_ci sc = dd->send_contexts[i].sc; 112562306a36Sopenharmony_ci /* 112662306a36Sopenharmony_ci * Don't disable unallocated, unfrozen, or user send contexts. 112762306a36Sopenharmony_ci * User send contexts will be disabled when the process 112862306a36Sopenharmony_ci * calls into the driver to reset its context. 112962306a36Sopenharmony_ci */ 113062306a36Sopenharmony_ci if (!sc || !(sc->flags & SCF_FROZEN) || sc->type == SC_USER) 113162306a36Sopenharmony_ci continue; 113262306a36Sopenharmony_ci 113362306a36Sopenharmony_ci /* only need to disable, the context is already stopped */ 113462306a36Sopenharmony_ci sc_disable(sc); 113562306a36Sopenharmony_ci } 113662306a36Sopenharmony_ci} 113762306a36Sopenharmony_ci 113862306a36Sopenharmony_ci/* 113962306a36Sopenharmony_ci * Unfreeze PIO for kernel send contexts. The precondition for calling this 114062306a36Sopenharmony_ci * is that all PIO send contexts have been disabled and the SPC freeze has 114162306a36Sopenharmony_ci * been cleared. Now perform the last step and re-enable each kernel context. 114262306a36Sopenharmony_ci * User (PSM) processing will occur when PSM calls into the kernel to 114362306a36Sopenharmony_ci * acknowledge the freeze. 114462306a36Sopenharmony_ci */ 114562306a36Sopenharmony_civoid pio_kernel_unfreeze(struct hfi1_devdata *dd) 114662306a36Sopenharmony_ci{ 114762306a36Sopenharmony_ci struct send_context *sc; 114862306a36Sopenharmony_ci int i; 114962306a36Sopenharmony_ci 115062306a36Sopenharmony_ci for (i = 0; i < dd->num_send_contexts; i++) { 115162306a36Sopenharmony_ci sc = dd->send_contexts[i].sc; 115262306a36Sopenharmony_ci if (!sc || !(sc->flags & SCF_FROZEN) || sc->type == SC_USER) 115362306a36Sopenharmony_ci continue; 115462306a36Sopenharmony_ci if (sc->flags & SCF_LINK_DOWN) 115562306a36Sopenharmony_ci continue; 115662306a36Sopenharmony_ci 115762306a36Sopenharmony_ci sc_enable(sc); /* will clear the sc frozen flag */ 115862306a36Sopenharmony_ci } 115962306a36Sopenharmony_ci} 116062306a36Sopenharmony_ci 116162306a36Sopenharmony_ci/** 116262306a36Sopenharmony_ci * pio_kernel_linkup() - Re-enable send contexts after linkup event 116362306a36Sopenharmony_ci * @dd: valid devive data 116462306a36Sopenharmony_ci * 116562306a36Sopenharmony_ci * When the link goes down, the freeze path is taken. However, a link down 116662306a36Sopenharmony_ci * event is different from a freeze because if the send context is re-enabled 116762306a36Sopenharmony_ci * whowever is sending data will start sending data again, which will hang 116862306a36Sopenharmony_ci * any QP that is sending data. 116962306a36Sopenharmony_ci * 117062306a36Sopenharmony_ci * The freeze path now looks at the type of event that occurs and takes this 117162306a36Sopenharmony_ci * path for link down event. 117262306a36Sopenharmony_ci */ 117362306a36Sopenharmony_civoid pio_kernel_linkup(struct hfi1_devdata *dd) 117462306a36Sopenharmony_ci{ 117562306a36Sopenharmony_ci struct send_context *sc; 117662306a36Sopenharmony_ci int i; 117762306a36Sopenharmony_ci 117862306a36Sopenharmony_ci for (i = 0; i < dd->num_send_contexts; i++) { 117962306a36Sopenharmony_ci sc = dd->send_contexts[i].sc; 118062306a36Sopenharmony_ci if (!sc || !(sc->flags & SCF_LINK_DOWN) || sc->type == SC_USER) 118162306a36Sopenharmony_ci continue; 118262306a36Sopenharmony_ci 118362306a36Sopenharmony_ci sc_enable(sc); /* will clear the sc link down flag */ 118462306a36Sopenharmony_ci } 118562306a36Sopenharmony_ci} 118662306a36Sopenharmony_ci 118762306a36Sopenharmony_ci/* 118862306a36Sopenharmony_ci * Wait for the SendPioInitCtxt.PioInitInProgress bit to clear. 118962306a36Sopenharmony_ci * Returns: 119062306a36Sopenharmony_ci * -ETIMEDOUT - if we wait too long 119162306a36Sopenharmony_ci * -EIO - if there was an error 119262306a36Sopenharmony_ci */ 119362306a36Sopenharmony_cistatic int pio_init_wait_progress(struct hfi1_devdata *dd) 119462306a36Sopenharmony_ci{ 119562306a36Sopenharmony_ci u64 reg; 119662306a36Sopenharmony_ci int max, count = 0; 119762306a36Sopenharmony_ci 119862306a36Sopenharmony_ci /* max is the longest possible HW init time / delay */ 119962306a36Sopenharmony_ci max = (dd->icode == ICODE_FPGA_EMULATION) ? 120 : 5; 120062306a36Sopenharmony_ci while (1) { 120162306a36Sopenharmony_ci reg = read_csr(dd, SEND_PIO_INIT_CTXT); 120262306a36Sopenharmony_ci if (!(reg & SEND_PIO_INIT_CTXT_PIO_INIT_IN_PROGRESS_SMASK)) 120362306a36Sopenharmony_ci break; 120462306a36Sopenharmony_ci if (count >= max) 120562306a36Sopenharmony_ci return -ETIMEDOUT; 120662306a36Sopenharmony_ci udelay(5); 120762306a36Sopenharmony_ci count++; 120862306a36Sopenharmony_ci } 120962306a36Sopenharmony_ci 121062306a36Sopenharmony_ci return reg & SEND_PIO_INIT_CTXT_PIO_INIT_ERR_SMASK ? -EIO : 0; 121162306a36Sopenharmony_ci} 121262306a36Sopenharmony_ci 121362306a36Sopenharmony_ci/* 121462306a36Sopenharmony_ci * Reset all of the send contexts to their power-on state. Used 121562306a36Sopenharmony_ci * only during manual init - no lock against sc_enable needed. 121662306a36Sopenharmony_ci */ 121762306a36Sopenharmony_civoid pio_reset_all(struct hfi1_devdata *dd) 121862306a36Sopenharmony_ci{ 121962306a36Sopenharmony_ci int ret; 122062306a36Sopenharmony_ci 122162306a36Sopenharmony_ci /* make sure the init engine is not busy */ 122262306a36Sopenharmony_ci ret = pio_init_wait_progress(dd); 122362306a36Sopenharmony_ci /* ignore any timeout */ 122462306a36Sopenharmony_ci if (ret == -EIO) { 122562306a36Sopenharmony_ci /* clear the error */ 122662306a36Sopenharmony_ci write_csr(dd, SEND_PIO_ERR_CLEAR, 122762306a36Sopenharmony_ci SEND_PIO_ERR_CLEAR_PIO_INIT_SM_IN_ERR_SMASK); 122862306a36Sopenharmony_ci } 122962306a36Sopenharmony_ci 123062306a36Sopenharmony_ci /* reset init all */ 123162306a36Sopenharmony_ci write_csr(dd, SEND_PIO_INIT_CTXT, 123262306a36Sopenharmony_ci SEND_PIO_INIT_CTXT_PIO_ALL_CTXT_INIT_SMASK); 123362306a36Sopenharmony_ci udelay(2); 123462306a36Sopenharmony_ci ret = pio_init_wait_progress(dd); 123562306a36Sopenharmony_ci if (ret < 0) { 123662306a36Sopenharmony_ci dd_dev_err(dd, 123762306a36Sopenharmony_ci "PIO send context init %s while initializing all PIO blocks\n", 123862306a36Sopenharmony_ci ret == -ETIMEDOUT ? "is stuck" : "had an error"); 123962306a36Sopenharmony_ci } 124062306a36Sopenharmony_ci} 124162306a36Sopenharmony_ci 124262306a36Sopenharmony_ci/* enable the context */ 124362306a36Sopenharmony_ciint sc_enable(struct send_context *sc) 124462306a36Sopenharmony_ci{ 124562306a36Sopenharmony_ci u64 sc_ctrl, reg, pio; 124662306a36Sopenharmony_ci struct hfi1_devdata *dd; 124762306a36Sopenharmony_ci unsigned long flags; 124862306a36Sopenharmony_ci int ret = 0; 124962306a36Sopenharmony_ci 125062306a36Sopenharmony_ci if (!sc) 125162306a36Sopenharmony_ci return -EINVAL; 125262306a36Sopenharmony_ci dd = sc->dd; 125362306a36Sopenharmony_ci 125462306a36Sopenharmony_ci /* 125562306a36Sopenharmony_ci * Obtain the allocator lock to guard against any allocation 125662306a36Sopenharmony_ci * attempts (which should not happen prior to context being 125762306a36Sopenharmony_ci * enabled). On the release/disable side we don't need to 125862306a36Sopenharmony_ci * worry about locking since the releaser will not do anything 125962306a36Sopenharmony_ci * if the context accounting values have not changed. 126062306a36Sopenharmony_ci */ 126162306a36Sopenharmony_ci spin_lock_irqsave(&sc->alloc_lock, flags); 126262306a36Sopenharmony_ci sc_ctrl = read_kctxt_csr(dd, sc->hw_context, SC(CTRL)); 126362306a36Sopenharmony_ci if ((sc_ctrl & SC(CTRL_CTXT_ENABLE_SMASK))) 126462306a36Sopenharmony_ci goto unlock; /* already enabled */ 126562306a36Sopenharmony_ci 126662306a36Sopenharmony_ci /* IMPORTANT: only clear free and fill if transitioning 0 -> 1 */ 126762306a36Sopenharmony_ci 126862306a36Sopenharmony_ci *sc->hw_free = 0; 126962306a36Sopenharmony_ci sc->free = 0; 127062306a36Sopenharmony_ci sc->alloc_free = 0; 127162306a36Sopenharmony_ci sc->fill = 0; 127262306a36Sopenharmony_ci sc->fill_wrap = 0; 127362306a36Sopenharmony_ci sc->sr_head = 0; 127462306a36Sopenharmony_ci sc->sr_tail = 0; 127562306a36Sopenharmony_ci sc->flags = 0; 127662306a36Sopenharmony_ci /* the alloc lock insures no fast path allocation */ 127762306a36Sopenharmony_ci reset_buffers_allocated(sc); 127862306a36Sopenharmony_ci 127962306a36Sopenharmony_ci /* 128062306a36Sopenharmony_ci * Clear all per-context errors. Some of these will be set when 128162306a36Sopenharmony_ci * we are re-enabling after a context halt. Now that the context 128262306a36Sopenharmony_ci * is disabled, the halt will not clear until after the PIO init 128362306a36Sopenharmony_ci * engine runs below. 128462306a36Sopenharmony_ci */ 128562306a36Sopenharmony_ci reg = read_kctxt_csr(dd, sc->hw_context, SC(ERR_STATUS)); 128662306a36Sopenharmony_ci if (reg) 128762306a36Sopenharmony_ci write_kctxt_csr(dd, sc->hw_context, SC(ERR_CLEAR), reg); 128862306a36Sopenharmony_ci 128962306a36Sopenharmony_ci /* 129062306a36Sopenharmony_ci * The HW PIO initialization engine can handle only one init 129162306a36Sopenharmony_ci * request at a time. Serialize access to each device's engine. 129262306a36Sopenharmony_ci */ 129362306a36Sopenharmony_ci spin_lock(&dd->sc_init_lock); 129462306a36Sopenharmony_ci /* 129562306a36Sopenharmony_ci * Since access to this code block is serialized and 129662306a36Sopenharmony_ci * each access waits for the initialization to complete 129762306a36Sopenharmony_ci * before releasing the lock, the PIO initialization engine 129862306a36Sopenharmony_ci * should not be in use, so we don't have to wait for the 129962306a36Sopenharmony_ci * InProgress bit to go down. 130062306a36Sopenharmony_ci */ 130162306a36Sopenharmony_ci pio = ((sc->hw_context & SEND_PIO_INIT_CTXT_PIO_CTXT_NUM_MASK) << 130262306a36Sopenharmony_ci SEND_PIO_INIT_CTXT_PIO_CTXT_NUM_SHIFT) | 130362306a36Sopenharmony_ci SEND_PIO_INIT_CTXT_PIO_SINGLE_CTXT_INIT_SMASK; 130462306a36Sopenharmony_ci write_csr(dd, SEND_PIO_INIT_CTXT, pio); 130562306a36Sopenharmony_ci /* 130662306a36Sopenharmony_ci * Wait until the engine is done. Give the chip the required time 130762306a36Sopenharmony_ci * so, hopefully, we read the register just once. 130862306a36Sopenharmony_ci */ 130962306a36Sopenharmony_ci udelay(2); 131062306a36Sopenharmony_ci ret = pio_init_wait_progress(dd); 131162306a36Sopenharmony_ci spin_unlock(&dd->sc_init_lock); 131262306a36Sopenharmony_ci if (ret) { 131362306a36Sopenharmony_ci dd_dev_err(dd, 131462306a36Sopenharmony_ci "sctxt%u(%u): Context not enabled due to init failure %d\n", 131562306a36Sopenharmony_ci sc->sw_index, sc->hw_context, ret); 131662306a36Sopenharmony_ci goto unlock; 131762306a36Sopenharmony_ci } 131862306a36Sopenharmony_ci 131962306a36Sopenharmony_ci /* 132062306a36Sopenharmony_ci * All is well. Enable the context. 132162306a36Sopenharmony_ci */ 132262306a36Sopenharmony_ci sc_ctrl |= SC(CTRL_CTXT_ENABLE_SMASK); 132362306a36Sopenharmony_ci write_kctxt_csr(dd, sc->hw_context, SC(CTRL), sc_ctrl); 132462306a36Sopenharmony_ci /* 132562306a36Sopenharmony_ci * Read SendCtxtCtrl to force the write out and prevent a timing 132662306a36Sopenharmony_ci * hazard where a PIO write may reach the context before the enable. 132762306a36Sopenharmony_ci */ 132862306a36Sopenharmony_ci read_kctxt_csr(dd, sc->hw_context, SC(CTRL)); 132962306a36Sopenharmony_ci sc->flags |= SCF_ENABLED; 133062306a36Sopenharmony_ci 133162306a36Sopenharmony_ciunlock: 133262306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->alloc_lock, flags); 133362306a36Sopenharmony_ci 133462306a36Sopenharmony_ci return ret; 133562306a36Sopenharmony_ci} 133662306a36Sopenharmony_ci 133762306a36Sopenharmony_ci/* force a credit return on the context */ 133862306a36Sopenharmony_civoid sc_return_credits(struct send_context *sc) 133962306a36Sopenharmony_ci{ 134062306a36Sopenharmony_ci if (!sc) 134162306a36Sopenharmony_ci return; 134262306a36Sopenharmony_ci 134362306a36Sopenharmony_ci /* a 0->1 transition schedules a credit return */ 134462306a36Sopenharmony_ci write_kctxt_csr(sc->dd, sc->hw_context, SC(CREDIT_FORCE), 134562306a36Sopenharmony_ci SC(CREDIT_FORCE_FORCE_RETURN_SMASK)); 134662306a36Sopenharmony_ci /* 134762306a36Sopenharmony_ci * Ensure that the write is flushed and the credit return is 134862306a36Sopenharmony_ci * scheduled. We care more about the 0 -> 1 transition. 134962306a36Sopenharmony_ci */ 135062306a36Sopenharmony_ci read_kctxt_csr(sc->dd, sc->hw_context, SC(CREDIT_FORCE)); 135162306a36Sopenharmony_ci /* set back to 0 for next time */ 135262306a36Sopenharmony_ci write_kctxt_csr(sc->dd, sc->hw_context, SC(CREDIT_FORCE), 0); 135362306a36Sopenharmony_ci} 135462306a36Sopenharmony_ci 135562306a36Sopenharmony_ci/* allow all in-flight packets to drain on the context */ 135662306a36Sopenharmony_civoid sc_flush(struct send_context *sc) 135762306a36Sopenharmony_ci{ 135862306a36Sopenharmony_ci if (!sc) 135962306a36Sopenharmony_ci return; 136062306a36Sopenharmony_ci 136162306a36Sopenharmony_ci sc_wait_for_packet_egress(sc, 1); 136262306a36Sopenharmony_ci} 136362306a36Sopenharmony_ci 136462306a36Sopenharmony_ci/* drop all packets on the context, no waiting until they are sent */ 136562306a36Sopenharmony_civoid sc_drop(struct send_context *sc) 136662306a36Sopenharmony_ci{ 136762306a36Sopenharmony_ci if (!sc) 136862306a36Sopenharmony_ci return; 136962306a36Sopenharmony_ci 137062306a36Sopenharmony_ci dd_dev_info(sc->dd, "%s: context %u(%u) - not implemented\n", 137162306a36Sopenharmony_ci __func__, sc->sw_index, sc->hw_context); 137262306a36Sopenharmony_ci} 137362306a36Sopenharmony_ci 137462306a36Sopenharmony_ci/* 137562306a36Sopenharmony_ci * Start the software reaction to a context halt or SPC freeze: 137662306a36Sopenharmony_ci * - mark the context as halted or frozen 137762306a36Sopenharmony_ci * - stop buffer allocations 137862306a36Sopenharmony_ci * 137962306a36Sopenharmony_ci * Called from the error interrupt. Other work is deferred until 138062306a36Sopenharmony_ci * out of the interrupt. 138162306a36Sopenharmony_ci */ 138262306a36Sopenharmony_civoid sc_stop(struct send_context *sc, int flag) 138362306a36Sopenharmony_ci{ 138462306a36Sopenharmony_ci unsigned long flags; 138562306a36Sopenharmony_ci 138662306a36Sopenharmony_ci /* stop buffer allocations */ 138762306a36Sopenharmony_ci spin_lock_irqsave(&sc->alloc_lock, flags); 138862306a36Sopenharmony_ci /* mark the context */ 138962306a36Sopenharmony_ci sc->flags |= flag; 139062306a36Sopenharmony_ci sc->flags &= ~SCF_ENABLED; 139162306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->alloc_lock, flags); 139262306a36Sopenharmony_ci wake_up(&sc->halt_wait); 139362306a36Sopenharmony_ci} 139462306a36Sopenharmony_ci 139562306a36Sopenharmony_ci#define BLOCK_DWORDS (PIO_BLOCK_SIZE / sizeof(u32)) 139662306a36Sopenharmony_ci#define dwords_to_blocks(x) DIV_ROUND_UP(x, BLOCK_DWORDS) 139762306a36Sopenharmony_ci 139862306a36Sopenharmony_ci/* 139962306a36Sopenharmony_ci * The send context buffer "allocator". 140062306a36Sopenharmony_ci * 140162306a36Sopenharmony_ci * @sc: the PIO send context we are allocating from 140262306a36Sopenharmony_ci * @len: length of whole packet - including PBC - in dwords 140362306a36Sopenharmony_ci * @cb: optional callback to call when the buffer is finished sending 140462306a36Sopenharmony_ci * @arg: argument for cb 140562306a36Sopenharmony_ci * 140662306a36Sopenharmony_ci * Return a pointer to a PIO buffer, NULL if not enough room, -ECOMM 140762306a36Sopenharmony_ci * when link is down. 140862306a36Sopenharmony_ci */ 140962306a36Sopenharmony_cistruct pio_buf *sc_buffer_alloc(struct send_context *sc, u32 dw_len, 141062306a36Sopenharmony_ci pio_release_cb cb, void *arg) 141162306a36Sopenharmony_ci{ 141262306a36Sopenharmony_ci struct pio_buf *pbuf = NULL; 141362306a36Sopenharmony_ci unsigned long flags; 141462306a36Sopenharmony_ci unsigned long avail; 141562306a36Sopenharmony_ci unsigned long blocks = dwords_to_blocks(dw_len); 141662306a36Sopenharmony_ci u32 fill_wrap; 141762306a36Sopenharmony_ci int trycount = 0; 141862306a36Sopenharmony_ci u32 head, next; 141962306a36Sopenharmony_ci 142062306a36Sopenharmony_ci spin_lock_irqsave(&sc->alloc_lock, flags); 142162306a36Sopenharmony_ci if (!(sc->flags & SCF_ENABLED)) { 142262306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->alloc_lock, flags); 142362306a36Sopenharmony_ci return ERR_PTR(-ECOMM); 142462306a36Sopenharmony_ci } 142562306a36Sopenharmony_ci 142662306a36Sopenharmony_ciretry: 142762306a36Sopenharmony_ci avail = (unsigned long)sc->credits - (sc->fill - sc->alloc_free); 142862306a36Sopenharmony_ci if (blocks > avail) { 142962306a36Sopenharmony_ci /* not enough room */ 143062306a36Sopenharmony_ci if (unlikely(trycount)) { /* already tried to get more room */ 143162306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->alloc_lock, flags); 143262306a36Sopenharmony_ci goto done; 143362306a36Sopenharmony_ci } 143462306a36Sopenharmony_ci /* copy from receiver cache line and recalculate */ 143562306a36Sopenharmony_ci sc->alloc_free = READ_ONCE(sc->free); 143662306a36Sopenharmony_ci avail = 143762306a36Sopenharmony_ci (unsigned long)sc->credits - 143862306a36Sopenharmony_ci (sc->fill - sc->alloc_free); 143962306a36Sopenharmony_ci if (blocks > avail) { 144062306a36Sopenharmony_ci /* still no room, actively update */ 144162306a36Sopenharmony_ci sc_release_update(sc); 144262306a36Sopenharmony_ci sc->alloc_free = READ_ONCE(sc->free); 144362306a36Sopenharmony_ci trycount++; 144462306a36Sopenharmony_ci goto retry; 144562306a36Sopenharmony_ci } 144662306a36Sopenharmony_ci } 144762306a36Sopenharmony_ci 144862306a36Sopenharmony_ci /* there is enough room */ 144962306a36Sopenharmony_ci 145062306a36Sopenharmony_ci preempt_disable(); 145162306a36Sopenharmony_ci this_cpu_inc(*sc->buffers_allocated); 145262306a36Sopenharmony_ci 145362306a36Sopenharmony_ci /* read this once */ 145462306a36Sopenharmony_ci head = sc->sr_head; 145562306a36Sopenharmony_ci 145662306a36Sopenharmony_ci /* "allocate" the buffer */ 145762306a36Sopenharmony_ci sc->fill += blocks; 145862306a36Sopenharmony_ci fill_wrap = sc->fill_wrap; 145962306a36Sopenharmony_ci sc->fill_wrap += blocks; 146062306a36Sopenharmony_ci if (sc->fill_wrap >= sc->credits) 146162306a36Sopenharmony_ci sc->fill_wrap = sc->fill_wrap - sc->credits; 146262306a36Sopenharmony_ci 146362306a36Sopenharmony_ci /* 146462306a36Sopenharmony_ci * Fill the parts that the releaser looks at before moving the head. 146562306a36Sopenharmony_ci * The only necessary piece is the sent_at field. The credits 146662306a36Sopenharmony_ci * we have just allocated cannot have been returned yet, so the 146762306a36Sopenharmony_ci * cb and arg will not be looked at for a "while". Put them 146862306a36Sopenharmony_ci * on this side of the memory barrier anyway. 146962306a36Sopenharmony_ci */ 147062306a36Sopenharmony_ci pbuf = &sc->sr[head].pbuf; 147162306a36Sopenharmony_ci pbuf->sent_at = sc->fill; 147262306a36Sopenharmony_ci pbuf->cb = cb; 147362306a36Sopenharmony_ci pbuf->arg = arg; 147462306a36Sopenharmony_ci pbuf->sc = sc; /* could be filled in at sc->sr init time */ 147562306a36Sopenharmony_ci /* make sure this is in memory before updating the head */ 147662306a36Sopenharmony_ci 147762306a36Sopenharmony_ci /* calculate next head index, do not store */ 147862306a36Sopenharmony_ci next = head + 1; 147962306a36Sopenharmony_ci if (next >= sc->sr_size) 148062306a36Sopenharmony_ci next = 0; 148162306a36Sopenharmony_ci /* 148262306a36Sopenharmony_ci * update the head - must be last! - the releaser can look at fields 148362306a36Sopenharmony_ci * in pbuf once we move the head 148462306a36Sopenharmony_ci */ 148562306a36Sopenharmony_ci smp_wmb(); 148662306a36Sopenharmony_ci sc->sr_head = next; 148762306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->alloc_lock, flags); 148862306a36Sopenharmony_ci 148962306a36Sopenharmony_ci /* finish filling in the buffer outside the lock */ 149062306a36Sopenharmony_ci pbuf->start = sc->base_addr + fill_wrap * PIO_BLOCK_SIZE; 149162306a36Sopenharmony_ci pbuf->end = sc->base_addr + sc->size; 149262306a36Sopenharmony_ci pbuf->qw_written = 0; 149362306a36Sopenharmony_ci pbuf->carry_bytes = 0; 149462306a36Sopenharmony_ci pbuf->carry.val64 = 0; 149562306a36Sopenharmony_cidone: 149662306a36Sopenharmony_ci return pbuf; 149762306a36Sopenharmony_ci} 149862306a36Sopenharmony_ci 149962306a36Sopenharmony_ci/* 150062306a36Sopenharmony_ci * There are at least two entities that can turn on credit return 150162306a36Sopenharmony_ci * interrupts and they can overlap. Avoid problems by implementing 150262306a36Sopenharmony_ci * a count scheme that is enforced by a lock. The lock is needed because 150362306a36Sopenharmony_ci * the count and CSR write must be paired. 150462306a36Sopenharmony_ci */ 150562306a36Sopenharmony_ci 150662306a36Sopenharmony_ci/* 150762306a36Sopenharmony_ci * Start credit return interrupts. This is managed by a count. If already 150862306a36Sopenharmony_ci * on, just increment the count. 150962306a36Sopenharmony_ci */ 151062306a36Sopenharmony_civoid sc_add_credit_return_intr(struct send_context *sc) 151162306a36Sopenharmony_ci{ 151262306a36Sopenharmony_ci unsigned long flags; 151362306a36Sopenharmony_ci 151462306a36Sopenharmony_ci /* lock must surround both the count change and the CSR update */ 151562306a36Sopenharmony_ci spin_lock_irqsave(&sc->credit_ctrl_lock, flags); 151662306a36Sopenharmony_ci if (sc->credit_intr_count == 0) { 151762306a36Sopenharmony_ci sc->credit_ctrl |= SC(CREDIT_CTRL_CREDIT_INTR_SMASK); 151862306a36Sopenharmony_ci write_kctxt_csr(sc->dd, sc->hw_context, 151962306a36Sopenharmony_ci SC(CREDIT_CTRL), sc->credit_ctrl); 152062306a36Sopenharmony_ci } 152162306a36Sopenharmony_ci sc->credit_intr_count++; 152262306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->credit_ctrl_lock, flags); 152362306a36Sopenharmony_ci} 152462306a36Sopenharmony_ci 152562306a36Sopenharmony_ci/* 152662306a36Sopenharmony_ci * Stop credit return interrupts. This is managed by a count. Decrement the 152762306a36Sopenharmony_ci * count, if the last user, then turn the credit interrupts off. 152862306a36Sopenharmony_ci */ 152962306a36Sopenharmony_civoid sc_del_credit_return_intr(struct send_context *sc) 153062306a36Sopenharmony_ci{ 153162306a36Sopenharmony_ci unsigned long flags; 153262306a36Sopenharmony_ci 153362306a36Sopenharmony_ci WARN_ON(sc->credit_intr_count == 0); 153462306a36Sopenharmony_ci 153562306a36Sopenharmony_ci /* lock must surround both the count change and the CSR update */ 153662306a36Sopenharmony_ci spin_lock_irqsave(&sc->credit_ctrl_lock, flags); 153762306a36Sopenharmony_ci sc->credit_intr_count--; 153862306a36Sopenharmony_ci if (sc->credit_intr_count == 0) { 153962306a36Sopenharmony_ci sc->credit_ctrl &= ~SC(CREDIT_CTRL_CREDIT_INTR_SMASK); 154062306a36Sopenharmony_ci write_kctxt_csr(sc->dd, sc->hw_context, 154162306a36Sopenharmony_ci SC(CREDIT_CTRL), sc->credit_ctrl); 154262306a36Sopenharmony_ci } 154362306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->credit_ctrl_lock, flags); 154462306a36Sopenharmony_ci} 154562306a36Sopenharmony_ci 154662306a36Sopenharmony_ci/* 154762306a36Sopenharmony_ci * The caller must be careful when calling this. All needint calls 154862306a36Sopenharmony_ci * must be paired with !needint. 154962306a36Sopenharmony_ci */ 155062306a36Sopenharmony_civoid hfi1_sc_wantpiobuf_intr(struct send_context *sc, u32 needint) 155162306a36Sopenharmony_ci{ 155262306a36Sopenharmony_ci if (needint) 155362306a36Sopenharmony_ci sc_add_credit_return_intr(sc); 155462306a36Sopenharmony_ci else 155562306a36Sopenharmony_ci sc_del_credit_return_intr(sc); 155662306a36Sopenharmony_ci trace_hfi1_wantpiointr(sc, needint, sc->credit_ctrl); 155762306a36Sopenharmony_ci if (needint) 155862306a36Sopenharmony_ci sc_return_credits(sc); 155962306a36Sopenharmony_ci} 156062306a36Sopenharmony_ci 156162306a36Sopenharmony_ci/** 156262306a36Sopenharmony_ci * sc_piobufavail - callback when a PIO buffer is available 156362306a36Sopenharmony_ci * @sc: the send context 156462306a36Sopenharmony_ci * 156562306a36Sopenharmony_ci * This is called from the interrupt handler when a PIO buffer is 156662306a36Sopenharmony_ci * available after hfi1_verbs_send() returned an error that no buffers were 156762306a36Sopenharmony_ci * available. Disable the interrupt if there are no more QPs waiting. 156862306a36Sopenharmony_ci */ 156962306a36Sopenharmony_cistatic void sc_piobufavail(struct send_context *sc) 157062306a36Sopenharmony_ci{ 157162306a36Sopenharmony_ci struct hfi1_devdata *dd = sc->dd; 157262306a36Sopenharmony_ci struct list_head *list; 157362306a36Sopenharmony_ci struct rvt_qp *qps[PIO_WAIT_BATCH_SIZE]; 157462306a36Sopenharmony_ci struct rvt_qp *qp; 157562306a36Sopenharmony_ci struct hfi1_qp_priv *priv; 157662306a36Sopenharmony_ci unsigned long flags; 157762306a36Sopenharmony_ci uint i, n = 0, top_idx = 0; 157862306a36Sopenharmony_ci 157962306a36Sopenharmony_ci if (dd->send_contexts[sc->sw_index].type != SC_KERNEL && 158062306a36Sopenharmony_ci dd->send_contexts[sc->sw_index].type != SC_VL15) 158162306a36Sopenharmony_ci return; 158262306a36Sopenharmony_ci list = &sc->piowait; 158362306a36Sopenharmony_ci /* 158462306a36Sopenharmony_ci * Note: checking that the piowait list is empty and clearing 158562306a36Sopenharmony_ci * the buffer available interrupt needs to be atomic or we 158662306a36Sopenharmony_ci * could end up with QPs on the wait list with the interrupt 158762306a36Sopenharmony_ci * disabled. 158862306a36Sopenharmony_ci */ 158962306a36Sopenharmony_ci write_seqlock_irqsave(&sc->waitlock, flags); 159062306a36Sopenharmony_ci while (!list_empty(list)) { 159162306a36Sopenharmony_ci struct iowait *wait; 159262306a36Sopenharmony_ci 159362306a36Sopenharmony_ci if (n == ARRAY_SIZE(qps)) 159462306a36Sopenharmony_ci break; 159562306a36Sopenharmony_ci wait = list_first_entry(list, struct iowait, list); 159662306a36Sopenharmony_ci iowait_get_priority(wait); 159762306a36Sopenharmony_ci qp = iowait_to_qp(wait); 159862306a36Sopenharmony_ci priv = qp->priv; 159962306a36Sopenharmony_ci list_del_init(&priv->s_iowait.list); 160062306a36Sopenharmony_ci priv->s_iowait.lock = NULL; 160162306a36Sopenharmony_ci if (n) { 160262306a36Sopenharmony_ci priv = qps[top_idx]->priv; 160362306a36Sopenharmony_ci top_idx = iowait_priority_update_top(wait, 160462306a36Sopenharmony_ci &priv->s_iowait, 160562306a36Sopenharmony_ci n, top_idx); 160662306a36Sopenharmony_ci } 160762306a36Sopenharmony_ci 160862306a36Sopenharmony_ci /* refcount held until actual wake up */ 160962306a36Sopenharmony_ci qps[n++] = qp; 161062306a36Sopenharmony_ci } 161162306a36Sopenharmony_ci /* 161262306a36Sopenharmony_ci * If there had been waiters and there are more 161362306a36Sopenharmony_ci * insure that we redo the force to avoid a potential hang. 161462306a36Sopenharmony_ci */ 161562306a36Sopenharmony_ci if (n) { 161662306a36Sopenharmony_ci hfi1_sc_wantpiobuf_intr(sc, 0); 161762306a36Sopenharmony_ci if (!list_empty(list)) 161862306a36Sopenharmony_ci hfi1_sc_wantpiobuf_intr(sc, 1); 161962306a36Sopenharmony_ci } 162062306a36Sopenharmony_ci write_sequnlock_irqrestore(&sc->waitlock, flags); 162162306a36Sopenharmony_ci 162262306a36Sopenharmony_ci /* Wake up the top-priority one first */ 162362306a36Sopenharmony_ci if (n) 162462306a36Sopenharmony_ci hfi1_qp_wakeup(qps[top_idx], 162562306a36Sopenharmony_ci RVT_S_WAIT_PIO | HFI1_S_WAIT_PIO_DRAIN); 162662306a36Sopenharmony_ci for (i = 0; i < n; i++) 162762306a36Sopenharmony_ci if (i != top_idx) 162862306a36Sopenharmony_ci hfi1_qp_wakeup(qps[i], 162962306a36Sopenharmony_ci RVT_S_WAIT_PIO | HFI1_S_WAIT_PIO_DRAIN); 163062306a36Sopenharmony_ci} 163162306a36Sopenharmony_ci 163262306a36Sopenharmony_ci/* translate a send credit update to a bit code of reasons */ 163362306a36Sopenharmony_cistatic inline int fill_code(u64 hw_free) 163462306a36Sopenharmony_ci{ 163562306a36Sopenharmony_ci int code = 0; 163662306a36Sopenharmony_ci 163762306a36Sopenharmony_ci if (hw_free & CR_STATUS_SMASK) 163862306a36Sopenharmony_ci code |= PRC_STATUS_ERR; 163962306a36Sopenharmony_ci if (hw_free & CR_CREDIT_RETURN_DUE_TO_PBC_SMASK) 164062306a36Sopenharmony_ci code |= PRC_PBC; 164162306a36Sopenharmony_ci if (hw_free & CR_CREDIT_RETURN_DUE_TO_THRESHOLD_SMASK) 164262306a36Sopenharmony_ci code |= PRC_THRESHOLD; 164362306a36Sopenharmony_ci if (hw_free & CR_CREDIT_RETURN_DUE_TO_ERR_SMASK) 164462306a36Sopenharmony_ci code |= PRC_FILL_ERR; 164562306a36Sopenharmony_ci if (hw_free & CR_CREDIT_RETURN_DUE_TO_FORCE_SMASK) 164662306a36Sopenharmony_ci code |= PRC_SC_DISABLE; 164762306a36Sopenharmony_ci return code; 164862306a36Sopenharmony_ci} 164962306a36Sopenharmony_ci 165062306a36Sopenharmony_ci/* use the jiffies compare to get the wrap right */ 165162306a36Sopenharmony_ci#define sent_before(a, b) time_before(a, b) /* a < b */ 165262306a36Sopenharmony_ci 165362306a36Sopenharmony_ci/* 165462306a36Sopenharmony_ci * The send context buffer "releaser". 165562306a36Sopenharmony_ci */ 165662306a36Sopenharmony_civoid sc_release_update(struct send_context *sc) 165762306a36Sopenharmony_ci{ 165862306a36Sopenharmony_ci struct pio_buf *pbuf; 165962306a36Sopenharmony_ci u64 hw_free; 166062306a36Sopenharmony_ci u32 head, tail; 166162306a36Sopenharmony_ci unsigned long old_free; 166262306a36Sopenharmony_ci unsigned long free; 166362306a36Sopenharmony_ci unsigned long extra; 166462306a36Sopenharmony_ci unsigned long flags; 166562306a36Sopenharmony_ci int code; 166662306a36Sopenharmony_ci 166762306a36Sopenharmony_ci if (!sc) 166862306a36Sopenharmony_ci return; 166962306a36Sopenharmony_ci 167062306a36Sopenharmony_ci spin_lock_irqsave(&sc->release_lock, flags); 167162306a36Sopenharmony_ci /* update free */ 167262306a36Sopenharmony_ci hw_free = le64_to_cpu(*sc->hw_free); /* volatile read */ 167362306a36Sopenharmony_ci old_free = sc->free; 167462306a36Sopenharmony_ci extra = (((hw_free & CR_COUNTER_SMASK) >> CR_COUNTER_SHIFT) 167562306a36Sopenharmony_ci - (old_free & CR_COUNTER_MASK)) 167662306a36Sopenharmony_ci & CR_COUNTER_MASK; 167762306a36Sopenharmony_ci free = old_free + extra; 167862306a36Sopenharmony_ci trace_hfi1_piofree(sc, extra); 167962306a36Sopenharmony_ci 168062306a36Sopenharmony_ci /* call sent buffer callbacks */ 168162306a36Sopenharmony_ci code = -1; /* code not yet set */ 168262306a36Sopenharmony_ci head = READ_ONCE(sc->sr_head); /* snapshot the head */ 168362306a36Sopenharmony_ci tail = sc->sr_tail; 168462306a36Sopenharmony_ci while (head != tail) { 168562306a36Sopenharmony_ci pbuf = &sc->sr[tail].pbuf; 168662306a36Sopenharmony_ci 168762306a36Sopenharmony_ci if (sent_before(free, pbuf->sent_at)) { 168862306a36Sopenharmony_ci /* not sent yet */ 168962306a36Sopenharmony_ci break; 169062306a36Sopenharmony_ci } 169162306a36Sopenharmony_ci if (pbuf->cb) { 169262306a36Sopenharmony_ci if (code < 0) /* fill in code on first user */ 169362306a36Sopenharmony_ci code = fill_code(hw_free); 169462306a36Sopenharmony_ci (*pbuf->cb)(pbuf->arg, code); 169562306a36Sopenharmony_ci } 169662306a36Sopenharmony_ci 169762306a36Sopenharmony_ci tail++; 169862306a36Sopenharmony_ci if (tail >= sc->sr_size) 169962306a36Sopenharmony_ci tail = 0; 170062306a36Sopenharmony_ci } 170162306a36Sopenharmony_ci sc->sr_tail = tail; 170262306a36Sopenharmony_ci /* make sure tail is updated before free */ 170362306a36Sopenharmony_ci smp_wmb(); 170462306a36Sopenharmony_ci sc->free = free; 170562306a36Sopenharmony_ci spin_unlock_irqrestore(&sc->release_lock, flags); 170662306a36Sopenharmony_ci sc_piobufavail(sc); 170762306a36Sopenharmony_ci} 170862306a36Sopenharmony_ci 170962306a36Sopenharmony_ci/* 171062306a36Sopenharmony_ci * Send context group releaser. Argument is the send context that caused 171162306a36Sopenharmony_ci * the interrupt. Called from the send context interrupt handler. 171262306a36Sopenharmony_ci * 171362306a36Sopenharmony_ci * Call release on all contexts in the group. 171462306a36Sopenharmony_ci * 171562306a36Sopenharmony_ci * This routine takes the sc_lock without an irqsave because it is only 171662306a36Sopenharmony_ci * called from an interrupt handler. Adjust if that changes. 171762306a36Sopenharmony_ci */ 171862306a36Sopenharmony_civoid sc_group_release_update(struct hfi1_devdata *dd, u32 hw_context) 171962306a36Sopenharmony_ci{ 172062306a36Sopenharmony_ci struct send_context *sc; 172162306a36Sopenharmony_ci u32 sw_index; 172262306a36Sopenharmony_ci u32 gc, gc_end; 172362306a36Sopenharmony_ci 172462306a36Sopenharmony_ci spin_lock(&dd->sc_lock); 172562306a36Sopenharmony_ci sw_index = dd->hw_to_sw[hw_context]; 172662306a36Sopenharmony_ci if (unlikely(sw_index >= dd->num_send_contexts)) { 172762306a36Sopenharmony_ci dd_dev_err(dd, "%s: invalid hw (%u) to sw (%u) mapping\n", 172862306a36Sopenharmony_ci __func__, hw_context, sw_index); 172962306a36Sopenharmony_ci goto done; 173062306a36Sopenharmony_ci } 173162306a36Sopenharmony_ci sc = dd->send_contexts[sw_index].sc; 173262306a36Sopenharmony_ci if (unlikely(!sc)) 173362306a36Sopenharmony_ci goto done; 173462306a36Sopenharmony_ci 173562306a36Sopenharmony_ci gc = group_context(hw_context, sc->group); 173662306a36Sopenharmony_ci gc_end = gc + group_size(sc->group); 173762306a36Sopenharmony_ci for (; gc < gc_end; gc++) { 173862306a36Sopenharmony_ci sw_index = dd->hw_to_sw[gc]; 173962306a36Sopenharmony_ci if (unlikely(sw_index >= dd->num_send_contexts)) { 174062306a36Sopenharmony_ci dd_dev_err(dd, 174162306a36Sopenharmony_ci "%s: invalid hw (%u) to sw (%u) mapping\n", 174262306a36Sopenharmony_ci __func__, hw_context, sw_index); 174362306a36Sopenharmony_ci continue; 174462306a36Sopenharmony_ci } 174562306a36Sopenharmony_ci sc_release_update(dd->send_contexts[sw_index].sc); 174662306a36Sopenharmony_ci } 174762306a36Sopenharmony_cidone: 174862306a36Sopenharmony_ci spin_unlock(&dd->sc_lock); 174962306a36Sopenharmony_ci} 175062306a36Sopenharmony_ci 175162306a36Sopenharmony_ci/* 175262306a36Sopenharmony_ci * pio_select_send_context_vl() - select send context 175362306a36Sopenharmony_ci * @dd: devdata 175462306a36Sopenharmony_ci * @selector: a spreading factor 175562306a36Sopenharmony_ci * @vl: this vl 175662306a36Sopenharmony_ci * 175762306a36Sopenharmony_ci * This function returns a send context based on the selector and a vl. 175862306a36Sopenharmony_ci * The mapping fields are protected by RCU 175962306a36Sopenharmony_ci */ 176062306a36Sopenharmony_cistruct send_context *pio_select_send_context_vl(struct hfi1_devdata *dd, 176162306a36Sopenharmony_ci u32 selector, u8 vl) 176262306a36Sopenharmony_ci{ 176362306a36Sopenharmony_ci struct pio_vl_map *m; 176462306a36Sopenharmony_ci struct pio_map_elem *e; 176562306a36Sopenharmony_ci struct send_context *rval; 176662306a36Sopenharmony_ci 176762306a36Sopenharmony_ci /* 176862306a36Sopenharmony_ci * NOTE This should only happen if SC->VL changed after the initial 176962306a36Sopenharmony_ci * checks on the QP/AH 177062306a36Sopenharmony_ci * Default will return VL0's send context below 177162306a36Sopenharmony_ci */ 177262306a36Sopenharmony_ci if (unlikely(vl >= num_vls)) { 177362306a36Sopenharmony_ci rval = NULL; 177462306a36Sopenharmony_ci goto done; 177562306a36Sopenharmony_ci } 177662306a36Sopenharmony_ci 177762306a36Sopenharmony_ci rcu_read_lock(); 177862306a36Sopenharmony_ci m = rcu_dereference(dd->pio_map); 177962306a36Sopenharmony_ci if (unlikely(!m)) { 178062306a36Sopenharmony_ci rcu_read_unlock(); 178162306a36Sopenharmony_ci return dd->vld[0].sc; 178262306a36Sopenharmony_ci } 178362306a36Sopenharmony_ci e = m->map[vl & m->mask]; 178462306a36Sopenharmony_ci rval = e->ksc[selector & e->mask]; 178562306a36Sopenharmony_ci rcu_read_unlock(); 178662306a36Sopenharmony_ci 178762306a36Sopenharmony_cidone: 178862306a36Sopenharmony_ci rval = !rval ? dd->vld[0].sc : rval; 178962306a36Sopenharmony_ci return rval; 179062306a36Sopenharmony_ci} 179162306a36Sopenharmony_ci 179262306a36Sopenharmony_ci/* 179362306a36Sopenharmony_ci * pio_select_send_context_sc() - select send context 179462306a36Sopenharmony_ci * @dd: devdata 179562306a36Sopenharmony_ci * @selector: a spreading factor 179662306a36Sopenharmony_ci * @sc5: the 5 bit sc 179762306a36Sopenharmony_ci * 179862306a36Sopenharmony_ci * This function returns an send context based on the selector and an sc 179962306a36Sopenharmony_ci */ 180062306a36Sopenharmony_cistruct send_context *pio_select_send_context_sc(struct hfi1_devdata *dd, 180162306a36Sopenharmony_ci u32 selector, u8 sc5) 180262306a36Sopenharmony_ci{ 180362306a36Sopenharmony_ci u8 vl = sc_to_vlt(dd, sc5); 180462306a36Sopenharmony_ci 180562306a36Sopenharmony_ci return pio_select_send_context_vl(dd, selector, vl); 180662306a36Sopenharmony_ci} 180762306a36Sopenharmony_ci 180862306a36Sopenharmony_ci/* 180962306a36Sopenharmony_ci * Free the indicated map struct 181062306a36Sopenharmony_ci */ 181162306a36Sopenharmony_cistatic void pio_map_free(struct pio_vl_map *m) 181262306a36Sopenharmony_ci{ 181362306a36Sopenharmony_ci int i; 181462306a36Sopenharmony_ci 181562306a36Sopenharmony_ci for (i = 0; m && i < m->actual_vls; i++) 181662306a36Sopenharmony_ci kfree(m->map[i]); 181762306a36Sopenharmony_ci kfree(m); 181862306a36Sopenharmony_ci} 181962306a36Sopenharmony_ci 182062306a36Sopenharmony_ci/* 182162306a36Sopenharmony_ci * Handle RCU callback 182262306a36Sopenharmony_ci */ 182362306a36Sopenharmony_cistatic void pio_map_rcu_callback(struct rcu_head *list) 182462306a36Sopenharmony_ci{ 182562306a36Sopenharmony_ci struct pio_vl_map *m = container_of(list, struct pio_vl_map, list); 182662306a36Sopenharmony_ci 182762306a36Sopenharmony_ci pio_map_free(m); 182862306a36Sopenharmony_ci} 182962306a36Sopenharmony_ci 183062306a36Sopenharmony_ci/* 183162306a36Sopenharmony_ci * Set credit return threshold for the kernel send context 183262306a36Sopenharmony_ci */ 183362306a36Sopenharmony_cistatic void set_threshold(struct hfi1_devdata *dd, int scontext, int i) 183462306a36Sopenharmony_ci{ 183562306a36Sopenharmony_ci u32 thres; 183662306a36Sopenharmony_ci 183762306a36Sopenharmony_ci thres = min(sc_percent_to_threshold(dd->kernel_send_context[scontext], 183862306a36Sopenharmony_ci 50), 183962306a36Sopenharmony_ci sc_mtu_to_threshold(dd->kernel_send_context[scontext], 184062306a36Sopenharmony_ci dd->vld[i].mtu, 184162306a36Sopenharmony_ci dd->rcd[0]->rcvhdrqentsize)); 184262306a36Sopenharmony_ci sc_set_cr_threshold(dd->kernel_send_context[scontext], thres); 184362306a36Sopenharmony_ci} 184462306a36Sopenharmony_ci 184562306a36Sopenharmony_ci/* 184662306a36Sopenharmony_ci * pio_map_init - called when #vls change 184762306a36Sopenharmony_ci * @dd: hfi1_devdata 184862306a36Sopenharmony_ci * @port: port number 184962306a36Sopenharmony_ci * @num_vls: number of vls 185062306a36Sopenharmony_ci * @vl_scontexts: per vl send context mapping (optional) 185162306a36Sopenharmony_ci * 185262306a36Sopenharmony_ci * This routine changes the mapping based on the number of vls. 185362306a36Sopenharmony_ci * 185462306a36Sopenharmony_ci * vl_scontexts is used to specify a non-uniform vl/send context 185562306a36Sopenharmony_ci * loading. NULL implies auto computing the loading and giving each 185662306a36Sopenharmony_ci * VL an uniform distribution of send contexts per VL. 185762306a36Sopenharmony_ci * 185862306a36Sopenharmony_ci * The auto algorithm computers the sc_per_vl and the number of extra 185962306a36Sopenharmony_ci * send contexts. Any extra send contexts are added from the last VL 186062306a36Sopenharmony_ci * on down 186162306a36Sopenharmony_ci * 186262306a36Sopenharmony_ci * rcu locking is used here to control access to the mapping fields. 186362306a36Sopenharmony_ci * 186462306a36Sopenharmony_ci * If either the num_vls or num_send_contexts are non-power of 2, the 186562306a36Sopenharmony_ci * array sizes in the struct pio_vl_map and the struct pio_map_elem are 186662306a36Sopenharmony_ci * rounded up to the next highest power of 2 and the first entry is 186762306a36Sopenharmony_ci * reused in a round robin fashion. 186862306a36Sopenharmony_ci * 186962306a36Sopenharmony_ci * If an error occurs the map change is not done and the mapping is not 187062306a36Sopenharmony_ci * chaged. 187162306a36Sopenharmony_ci * 187262306a36Sopenharmony_ci */ 187362306a36Sopenharmony_ciint pio_map_init(struct hfi1_devdata *dd, u8 port, u8 num_vls, u8 *vl_scontexts) 187462306a36Sopenharmony_ci{ 187562306a36Sopenharmony_ci int i, j; 187662306a36Sopenharmony_ci int extra, sc_per_vl; 187762306a36Sopenharmony_ci int scontext = 1; 187862306a36Sopenharmony_ci int num_kernel_send_contexts = 0; 187962306a36Sopenharmony_ci u8 lvl_scontexts[OPA_MAX_VLS]; 188062306a36Sopenharmony_ci struct pio_vl_map *oldmap, *newmap; 188162306a36Sopenharmony_ci 188262306a36Sopenharmony_ci if (!vl_scontexts) { 188362306a36Sopenharmony_ci for (i = 0; i < dd->num_send_contexts; i++) 188462306a36Sopenharmony_ci if (dd->send_contexts[i].type == SC_KERNEL) 188562306a36Sopenharmony_ci num_kernel_send_contexts++; 188662306a36Sopenharmony_ci /* truncate divide */ 188762306a36Sopenharmony_ci sc_per_vl = num_kernel_send_contexts / num_vls; 188862306a36Sopenharmony_ci /* extras */ 188962306a36Sopenharmony_ci extra = num_kernel_send_contexts % num_vls; 189062306a36Sopenharmony_ci vl_scontexts = lvl_scontexts; 189162306a36Sopenharmony_ci /* add extras from last vl down */ 189262306a36Sopenharmony_ci for (i = num_vls - 1; i >= 0; i--, extra--) 189362306a36Sopenharmony_ci vl_scontexts[i] = sc_per_vl + (extra > 0 ? 1 : 0); 189462306a36Sopenharmony_ci } 189562306a36Sopenharmony_ci /* build new map */ 189662306a36Sopenharmony_ci newmap = kzalloc(struct_size(newmap, map, roundup_pow_of_two(num_vls)), 189762306a36Sopenharmony_ci GFP_KERNEL); 189862306a36Sopenharmony_ci if (!newmap) 189962306a36Sopenharmony_ci goto bail; 190062306a36Sopenharmony_ci newmap->actual_vls = num_vls; 190162306a36Sopenharmony_ci newmap->vls = roundup_pow_of_two(num_vls); 190262306a36Sopenharmony_ci newmap->mask = (1 << ilog2(newmap->vls)) - 1; 190362306a36Sopenharmony_ci for (i = 0; i < newmap->vls; i++) { 190462306a36Sopenharmony_ci /* save for wrap around */ 190562306a36Sopenharmony_ci int first_scontext = scontext; 190662306a36Sopenharmony_ci 190762306a36Sopenharmony_ci if (i < newmap->actual_vls) { 190862306a36Sopenharmony_ci int sz = roundup_pow_of_two(vl_scontexts[i]); 190962306a36Sopenharmony_ci 191062306a36Sopenharmony_ci /* only allocate once */ 191162306a36Sopenharmony_ci newmap->map[i] = kzalloc(struct_size(newmap->map[i], 191262306a36Sopenharmony_ci ksc, sz), 191362306a36Sopenharmony_ci GFP_KERNEL); 191462306a36Sopenharmony_ci if (!newmap->map[i]) 191562306a36Sopenharmony_ci goto bail; 191662306a36Sopenharmony_ci newmap->map[i]->mask = (1 << ilog2(sz)) - 1; 191762306a36Sopenharmony_ci /* 191862306a36Sopenharmony_ci * assign send contexts and 191962306a36Sopenharmony_ci * adjust credit return threshold 192062306a36Sopenharmony_ci */ 192162306a36Sopenharmony_ci for (j = 0; j < sz; j++) { 192262306a36Sopenharmony_ci if (dd->kernel_send_context[scontext]) { 192362306a36Sopenharmony_ci newmap->map[i]->ksc[j] = 192462306a36Sopenharmony_ci dd->kernel_send_context[scontext]; 192562306a36Sopenharmony_ci set_threshold(dd, scontext, i); 192662306a36Sopenharmony_ci } 192762306a36Sopenharmony_ci if (++scontext >= first_scontext + 192862306a36Sopenharmony_ci vl_scontexts[i]) 192962306a36Sopenharmony_ci /* wrap back to first send context */ 193062306a36Sopenharmony_ci scontext = first_scontext; 193162306a36Sopenharmony_ci } 193262306a36Sopenharmony_ci } else { 193362306a36Sopenharmony_ci /* just re-use entry without allocating */ 193462306a36Sopenharmony_ci newmap->map[i] = newmap->map[i % num_vls]; 193562306a36Sopenharmony_ci } 193662306a36Sopenharmony_ci scontext = first_scontext + vl_scontexts[i]; 193762306a36Sopenharmony_ci } 193862306a36Sopenharmony_ci /* newmap in hand, save old map */ 193962306a36Sopenharmony_ci spin_lock_irq(&dd->pio_map_lock); 194062306a36Sopenharmony_ci oldmap = rcu_dereference_protected(dd->pio_map, 194162306a36Sopenharmony_ci lockdep_is_held(&dd->pio_map_lock)); 194262306a36Sopenharmony_ci 194362306a36Sopenharmony_ci /* publish newmap */ 194462306a36Sopenharmony_ci rcu_assign_pointer(dd->pio_map, newmap); 194562306a36Sopenharmony_ci 194662306a36Sopenharmony_ci spin_unlock_irq(&dd->pio_map_lock); 194762306a36Sopenharmony_ci /* success, free any old map after grace period */ 194862306a36Sopenharmony_ci if (oldmap) 194962306a36Sopenharmony_ci call_rcu(&oldmap->list, pio_map_rcu_callback); 195062306a36Sopenharmony_ci return 0; 195162306a36Sopenharmony_cibail: 195262306a36Sopenharmony_ci /* free any partial allocation */ 195362306a36Sopenharmony_ci pio_map_free(newmap); 195462306a36Sopenharmony_ci return -ENOMEM; 195562306a36Sopenharmony_ci} 195662306a36Sopenharmony_ci 195762306a36Sopenharmony_civoid free_pio_map(struct hfi1_devdata *dd) 195862306a36Sopenharmony_ci{ 195962306a36Sopenharmony_ci /* Free PIO map if allocated */ 196062306a36Sopenharmony_ci if (rcu_access_pointer(dd->pio_map)) { 196162306a36Sopenharmony_ci spin_lock_irq(&dd->pio_map_lock); 196262306a36Sopenharmony_ci pio_map_free(rcu_access_pointer(dd->pio_map)); 196362306a36Sopenharmony_ci RCU_INIT_POINTER(dd->pio_map, NULL); 196462306a36Sopenharmony_ci spin_unlock_irq(&dd->pio_map_lock); 196562306a36Sopenharmony_ci synchronize_rcu(); 196662306a36Sopenharmony_ci } 196762306a36Sopenharmony_ci kfree(dd->kernel_send_context); 196862306a36Sopenharmony_ci dd->kernel_send_context = NULL; 196962306a36Sopenharmony_ci} 197062306a36Sopenharmony_ci 197162306a36Sopenharmony_ciint init_pervl_scs(struct hfi1_devdata *dd) 197262306a36Sopenharmony_ci{ 197362306a36Sopenharmony_ci int i; 197462306a36Sopenharmony_ci u64 mask, all_vl_mask = (u64)0x80ff; /* VLs 0-7, 15 */ 197562306a36Sopenharmony_ci u64 data_vls_mask = (u64)0x00ff; /* VLs 0-7 */ 197662306a36Sopenharmony_ci u32 ctxt; 197762306a36Sopenharmony_ci struct hfi1_pportdata *ppd = dd->pport; 197862306a36Sopenharmony_ci 197962306a36Sopenharmony_ci dd->vld[15].sc = sc_alloc(dd, SC_VL15, 198062306a36Sopenharmony_ci dd->rcd[0]->rcvhdrqentsize, dd->node); 198162306a36Sopenharmony_ci if (!dd->vld[15].sc) 198262306a36Sopenharmony_ci return -ENOMEM; 198362306a36Sopenharmony_ci 198462306a36Sopenharmony_ci hfi1_init_ctxt(dd->vld[15].sc); 198562306a36Sopenharmony_ci dd->vld[15].mtu = enum_to_mtu(OPA_MTU_2048); 198662306a36Sopenharmony_ci 198762306a36Sopenharmony_ci dd->kernel_send_context = kcalloc_node(dd->num_send_contexts, 198862306a36Sopenharmony_ci sizeof(struct send_context *), 198962306a36Sopenharmony_ci GFP_KERNEL, dd->node); 199062306a36Sopenharmony_ci if (!dd->kernel_send_context) 199162306a36Sopenharmony_ci goto freesc15; 199262306a36Sopenharmony_ci 199362306a36Sopenharmony_ci dd->kernel_send_context[0] = dd->vld[15].sc; 199462306a36Sopenharmony_ci 199562306a36Sopenharmony_ci for (i = 0; i < num_vls; i++) { 199662306a36Sopenharmony_ci /* 199762306a36Sopenharmony_ci * Since this function does not deal with a specific 199862306a36Sopenharmony_ci * receive context but we need the RcvHdrQ entry size, 199962306a36Sopenharmony_ci * use the size from rcd[0]. It is guaranteed to be 200062306a36Sopenharmony_ci * valid at this point and will remain the same for all 200162306a36Sopenharmony_ci * receive contexts. 200262306a36Sopenharmony_ci */ 200362306a36Sopenharmony_ci dd->vld[i].sc = sc_alloc(dd, SC_KERNEL, 200462306a36Sopenharmony_ci dd->rcd[0]->rcvhdrqentsize, dd->node); 200562306a36Sopenharmony_ci if (!dd->vld[i].sc) 200662306a36Sopenharmony_ci goto nomem; 200762306a36Sopenharmony_ci dd->kernel_send_context[i + 1] = dd->vld[i].sc; 200862306a36Sopenharmony_ci hfi1_init_ctxt(dd->vld[i].sc); 200962306a36Sopenharmony_ci /* non VL15 start with the max MTU */ 201062306a36Sopenharmony_ci dd->vld[i].mtu = hfi1_max_mtu; 201162306a36Sopenharmony_ci } 201262306a36Sopenharmony_ci for (i = num_vls; i < INIT_SC_PER_VL * num_vls; i++) { 201362306a36Sopenharmony_ci dd->kernel_send_context[i + 1] = 201462306a36Sopenharmony_ci sc_alloc(dd, SC_KERNEL, dd->rcd[0]->rcvhdrqentsize, dd->node); 201562306a36Sopenharmony_ci if (!dd->kernel_send_context[i + 1]) 201662306a36Sopenharmony_ci goto nomem; 201762306a36Sopenharmony_ci hfi1_init_ctxt(dd->kernel_send_context[i + 1]); 201862306a36Sopenharmony_ci } 201962306a36Sopenharmony_ci 202062306a36Sopenharmony_ci sc_enable(dd->vld[15].sc); 202162306a36Sopenharmony_ci ctxt = dd->vld[15].sc->hw_context; 202262306a36Sopenharmony_ci mask = all_vl_mask & ~(1LL << 15); 202362306a36Sopenharmony_ci write_kctxt_csr(dd, ctxt, SC(CHECK_VL), mask); 202462306a36Sopenharmony_ci dd_dev_info(dd, 202562306a36Sopenharmony_ci "Using send context %u(%u) for VL15\n", 202662306a36Sopenharmony_ci dd->vld[15].sc->sw_index, ctxt); 202762306a36Sopenharmony_ci 202862306a36Sopenharmony_ci for (i = 0; i < num_vls; i++) { 202962306a36Sopenharmony_ci sc_enable(dd->vld[i].sc); 203062306a36Sopenharmony_ci ctxt = dd->vld[i].sc->hw_context; 203162306a36Sopenharmony_ci mask = all_vl_mask & ~(data_vls_mask); 203262306a36Sopenharmony_ci write_kctxt_csr(dd, ctxt, SC(CHECK_VL), mask); 203362306a36Sopenharmony_ci } 203462306a36Sopenharmony_ci for (i = num_vls; i < INIT_SC_PER_VL * num_vls; i++) { 203562306a36Sopenharmony_ci sc_enable(dd->kernel_send_context[i + 1]); 203662306a36Sopenharmony_ci ctxt = dd->kernel_send_context[i + 1]->hw_context; 203762306a36Sopenharmony_ci mask = all_vl_mask & ~(data_vls_mask); 203862306a36Sopenharmony_ci write_kctxt_csr(dd, ctxt, SC(CHECK_VL), mask); 203962306a36Sopenharmony_ci } 204062306a36Sopenharmony_ci 204162306a36Sopenharmony_ci if (pio_map_init(dd, ppd->port - 1, num_vls, NULL)) 204262306a36Sopenharmony_ci goto nomem; 204362306a36Sopenharmony_ci return 0; 204462306a36Sopenharmony_ci 204562306a36Sopenharmony_cinomem: 204662306a36Sopenharmony_ci for (i = 0; i < num_vls; i++) { 204762306a36Sopenharmony_ci sc_free(dd->vld[i].sc); 204862306a36Sopenharmony_ci dd->vld[i].sc = NULL; 204962306a36Sopenharmony_ci } 205062306a36Sopenharmony_ci 205162306a36Sopenharmony_ci for (i = num_vls; i < INIT_SC_PER_VL * num_vls; i++) 205262306a36Sopenharmony_ci sc_free(dd->kernel_send_context[i + 1]); 205362306a36Sopenharmony_ci 205462306a36Sopenharmony_ci kfree(dd->kernel_send_context); 205562306a36Sopenharmony_ci dd->kernel_send_context = NULL; 205662306a36Sopenharmony_ci 205762306a36Sopenharmony_cifreesc15: 205862306a36Sopenharmony_ci sc_free(dd->vld[15].sc); 205962306a36Sopenharmony_ci return -ENOMEM; 206062306a36Sopenharmony_ci} 206162306a36Sopenharmony_ci 206262306a36Sopenharmony_ciint init_credit_return(struct hfi1_devdata *dd) 206362306a36Sopenharmony_ci{ 206462306a36Sopenharmony_ci int ret; 206562306a36Sopenharmony_ci int i; 206662306a36Sopenharmony_ci 206762306a36Sopenharmony_ci dd->cr_base = kcalloc( 206862306a36Sopenharmony_ci node_affinity.num_possible_nodes, 206962306a36Sopenharmony_ci sizeof(struct credit_return_base), 207062306a36Sopenharmony_ci GFP_KERNEL); 207162306a36Sopenharmony_ci if (!dd->cr_base) { 207262306a36Sopenharmony_ci ret = -ENOMEM; 207362306a36Sopenharmony_ci goto done; 207462306a36Sopenharmony_ci } 207562306a36Sopenharmony_ci for_each_node_with_cpus(i) { 207662306a36Sopenharmony_ci int bytes = TXE_NUM_CONTEXTS * sizeof(struct credit_return); 207762306a36Sopenharmony_ci 207862306a36Sopenharmony_ci set_dev_node(&dd->pcidev->dev, i); 207962306a36Sopenharmony_ci dd->cr_base[i].va = dma_alloc_coherent(&dd->pcidev->dev, 208062306a36Sopenharmony_ci bytes, 208162306a36Sopenharmony_ci &dd->cr_base[i].dma, 208262306a36Sopenharmony_ci GFP_KERNEL); 208362306a36Sopenharmony_ci if (!dd->cr_base[i].va) { 208462306a36Sopenharmony_ci set_dev_node(&dd->pcidev->dev, dd->node); 208562306a36Sopenharmony_ci dd_dev_err(dd, 208662306a36Sopenharmony_ci "Unable to allocate credit return DMA range for NUMA %d\n", 208762306a36Sopenharmony_ci i); 208862306a36Sopenharmony_ci ret = -ENOMEM; 208962306a36Sopenharmony_ci goto free_cr_base; 209062306a36Sopenharmony_ci } 209162306a36Sopenharmony_ci } 209262306a36Sopenharmony_ci set_dev_node(&dd->pcidev->dev, dd->node); 209362306a36Sopenharmony_ci 209462306a36Sopenharmony_ci ret = 0; 209562306a36Sopenharmony_cidone: 209662306a36Sopenharmony_ci return ret; 209762306a36Sopenharmony_ci 209862306a36Sopenharmony_cifree_cr_base: 209962306a36Sopenharmony_ci free_credit_return(dd); 210062306a36Sopenharmony_ci goto done; 210162306a36Sopenharmony_ci} 210262306a36Sopenharmony_ci 210362306a36Sopenharmony_civoid free_credit_return(struct hfi1_devdata *dd) 210462306a36Sopenharmony_ci{ 210562306a36Sopenharmony_ci int i; 210662306a36Sopenharmony_ci 210762306a36Sopenharmony_ci if (!dd->cr_base) 210862306a36Sopenharmony_ci return; 210962306a36Sopenharmony_ci for (i = 0; i < node_affinity.num_possible_nodes; i++) { 211062306a36Sopenharmony_ci if (dd->cr_base[i].va) { 211162306a36Sopenharmony_ci dma_free_coherent(&dd->pcidev->dev, 211262306a36Sopenharmony_ci TXE_NUM_CONTEXTS * 211362306a36Sopenharmony_ci sizeof(struct credit_return), 211462306a36Sopenharmony_ci dd->cr_base[i].va, 211562306a36Sopenharmony_ci dd->cr_base[i].dma); 211662306a36Sopenharmony_ci } 211762306a36Sopenharmony_ci } 211862306a36Sopenharmony_ci kfree(dd->cr_base); 211962306a36Sopenharmony_ci dd->cr_base = NULL; 212062306a36Sopenharmony_ci} 212162306a36Sopenharmony_ci 212262306a36Sopenharmony_civoid seqfile_dump_sci(struct seq_file *s, u32 i, 212362306a36Sopenharmony_ci struct send_context_info *sci) 212462306a36Sopenharmony_ci{ 212562306a36Sopenharmony_ci struct send_context *sc = sci->sc; 212662306a36Sopenharmony_ci u64 reg; 212762306a36Sopenharmony_ci 212862306a36Sopenharmony_ci seq_printf(s, "SCI %u: type %u base %u credits %u\n", 212962306a36Sopenharmony_ci i, sci->type, sci->base, sci->credits); 213062306a36Sopenharmony_ci seq_printf(s, " flags 0x%x sw_inx %u hw_ctxt %u grp %u\n", 213162306a36Sopenharmony_ci sc->flags, sc->sw_index, sc->hw_context, sc->group); 213262306a36Sopenharmony_ci seq_printf(s, " sr_size %u credits %u sr_head %u sr_tail %u\n", 213362306a36Sopenharmony_ci sc->sr_size, sc->credits, sc->sr_head, sc->sr_tail); 213462306a36Sopenharmony_ci seq_printf(s, " fill %lu free %lu fill_wrap %u alloc_free %lu\n", 213562306a36Sopenharmony_ci sc->fill, sc->free, sc->fill_wrap, sc->alloc_free); 213662306a36Sopenharmony_ci seq_printf(s, " credit_intr_count %u credit_ctrl 0x%llx\n", 213762306a36Sopenharmony_ci sc->credit_intr_count, sc->credit_ctrl); 213862306a36Sopenharmony_ci reg = read_kctxt_csr(sc->dd, sc->hw_context, SC(CREDIT_STATUS)); 213962306a36Sopenharmony_ci seq_printf(s, " *hw_free %llu CurrentFree %llu LastReturned %llu\n", 214062306a36Sopenharmony_ci (le64_to_cpu(*sc->hw_free) & CR_COUNTER_SMASK) >> 214162306a36Sopenharmony_ci CR_COUNTER_SHIFT, 214262306a36Sopenharmony_ci (reg >> SC(CREDIT_STATUS_CURRENT_FREE_COUNTER_SHIFT)) & 214362306a36Sopenharmony_ci SC(CREDIT_STATUS_CURRENT_FREE_COUNTER_MASK), 214462306a36Sopenharmony_ci reg & SC(CREDIT_STATUS_LAST_RETURNED_COUNTER_SMASK)); 214562306a36Sopenharmony_ci} 2146