/third_party/vixl/tools/ |
H A D | make_instruction_doc_aarch64.pl | 41 my %inst = (); # Global hash of instructions. 98 $inst{$p}->{'type'} = $type; 99 $inst{$p}->{'mnemonic'} = $i; 100 $inst{$p}->{'description'} = $d; 101 $inst{$p}->{'initial'} = substr($i, 0, 1); 106 my $links = get_links_list(\%inst); 143 my $inst = $insts->{$i}; 144 $initials{$inst->{type}}->{$inst->{initial}}++; 159 $inst{ [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
H A D | mc.h | 20 int nv04_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 21 int nv11_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 22 int nv17_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 23 int nv44_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 24 int nv50_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 25 int g84_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 26 int g98_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 27 int gt215_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 28 int gf100_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **); 29 int gk104_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struc [all...] |
/third_party/pcre2/pcre2/src/sljit/ |
H A D | sljitNativeX86_64.c | 35 sljit_u8 *inst; in emit_load_imm64() local 37 inst = (sljit_u8*)ensure_buf(compiler, 1 + 2 + sizeof(sljit_sw)); in emit_load_imm64() 38 FAIL_IF(!inst); in emit_load_imm64() 40 *inst++ = REX_W | ((reg_map[reg] <= 7) ? 0 : REX_B); in emit_load_imm64() 41 *inst++ = U8(MOV_r_i32 | (reg_map[reg] & 0x7)); in emit_load_imm64() 42 sljit_unaligned_store_sw(inst, imm); in emit_load_imm64() 48 sljit_u8 *inst; in emit_do_imm32() local 51 inst = (sljit_u8*)ensure_buf(compiler, 1 + length); in emit_do_imm32() 52 FAIL_IF(!inst); in emit_do_imm32() 55 *inst in emit_do_imm32() 67 sljit_u8 *inst; emit_x86_instruction() local 386 sljit_u8 *inst; sljit_emit_enter() local 586 sljit_u8 *inst; emit_stack_frame_release() local 659 sljit_u8 *inst; sljit_emit_return_void() local 868 sljit_u8 *inst; sljit_emit_fast_enter() local 901 sljit_u8 *inst; emit_fast_return() local 945 sljit_u8* inst; sljit_emit_mem() local 1025 sljit_u8* inst; emit_mov_int() local [all...] |
/kernel/linux/linux-6.6/drivers/media/platform/qcom/venus/ |
H A D | pm_helpers.c | 149 static u32 load_per_instance(struct venus_inst *inst) in load_per_instance() argument 153 if (!inst || !(inst->state >= INST_INIT && inst->state < INST_STOP)) in load_per_instance() 156 mbs = (ALIGN(inst->width, 16) / 16) * (ALIGN(inst->height, 16) / 16); in load_per_instance() 158 return mbs * inst->fps; in load_per_instance() 163 struct venus_inst *inst = NULL; in load_per_type() local 166 list_for_each_entry(inst, &core->instances, list) { in load_per_type() 167 if (inst in load_per_type() 176 mbs_to_bw(struct venus_inst *inst, u32 mbs, u32 *avg, u32 *peak) mbs_to_bw() argument 217 struct venus_inst *inst = NULL; load_scale_bw() local 243 load_scale_v1(struct venus_inst *inst) load_scale_v1() argument 525 power_save_mode_enable(struct venus_inst *inst, bool enable) power_save_mode_enable() argument 555 struct venus_inst *inst = NULL; move_core_to_power_save_mode() local 568 min_loaded_core(struct venus_inst *inst, u32 *min_coreid, u32 *min_load, bool low_power) min_loaded_core() argument 621 decide_core(struct venus_inst *inst) decide_core() argument 682 acquire_core(struct venus_inst *inst) acquire_core() argument 709 release_core(struct venus_inst *inst) release_core() argument 742 coreid_power_v4(struct venus_inst *inst, int on) coreid_power_v4() argument 1075 calculate_inst_freq(struct venus_inst *inst, unsigned long filled_len) calculate_inst_freq() argument 1110 load_scale_v4(struct venus_inst *inst) load_scale_v4() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
H A D | nv20.c | 21 u32 inst = nvkm_memory_addr(chan->inst); in nv20_gr_chan_init() local 24 nvkm_wo32(gr->ctxtab, chan->chid * 4, inst >> 4); in nv20_gr_chan_init() 35 u32 inst = nvkm_memory_addr(chan->inst); in nv20_gr_chan_fini() local 42 nvkm_wr32(device, 0x400784, inst >> 4); in nv20_gr_chan_fini() 63 nvkm_memory_unref(&chan->inst); in nv20_gr_chan_dtor() 91 &chan->inst); in nv20_gr_chan_new() 95 nvkm_kmap(chan->inst); in nv20_gr_chan_new() 96 nvkm_wo32(chan->inst, in nv20_gr_chan_new() [all...] |
/third_party/spirv-tools/source/val/ |
H A D | validation_state.cpp | 108 const spv_parsed_instruction_t* inst) { in CountInstructions() 110 if (spv::Op(inst->opcode) == spv::Op::OpFunction) { in CountInstructions() 307 const Instruction* inst) { in diag() 320 if (inst) disassembly = Disassemble(*inst); in diag() 322 return DiagnosticStream({0, 0, inst ? inst->LineNum() : 0}, in diag() 527 const spv_parsed_instruction_t* inst) { in AddOrderedInstruction() 528 ordered_instructions_.emplace_back(inst); in AddOrderedInstruction() 534 void ValidationState_t::RegisterDebugInstruction(const Instruction* inst) { in RegisterDebugInstruction() argument 107 CountInstructions(void* user_data, const spv_parsed_instruction_t* inst) CountInstructions() argument 306 diag(spv_result_t error_code, const Instruction* inst) diag() argument 526 AddOrderedInstruction( const spv_parsed_instruction_t* inst) AddOrderedInstruction() argument 559 RegisterInstruction(Instruction* inst) RegisterInstruction() argument 831 RegisterUniqueTypeDeclaration(const Instruction* inst) RegisterUniqueTypeDeclaration() argument 851 const Instruction* inst = FindDef(id); GetTypeId() local 856 const Instruction* inst = FindDef(id); GetIdOpcode() local 861 const Instruction* inst = FindDef(id); GetComponentType() local 891 const Instruction* inst = FindDef(id); GetDimension() local 921 const Instruction* inst = FindDef(component_type_id); GetBitWidth() local 935 const Instruction* inst = FindDef(id); IsVoidType() local 940 const Instruction* inst = FindDef(id); IsFloatScalarType() local 945 const Instruction* inst = FindDef(id); IsFloatVectorType() local 958 const Instruction* inst = FindDef(id); IsFloatScalarOrVectorType() local 975 const Instruction* inst = FindDef(id); IsIntScalarType() local 980 const Instruction* inst = FindDef(id); IsIntVectorType() local 993 const Instruction* inst = FindDef(id); IsIntScalarOrVectorType() local 1010 const Instruction* inst = FindDef(id); IsUnsignedIntScalarType() local 1015 const Instruction* inst = FindDef(id); IsUnsignedIntVectorType() local 1028 const Instruction* inst = FindDef(id); IsUnsignedIntScalarOrVectorType() local 1045 const Instruction* inst = FindDef(id); IsSignedIntScalarType() local 1050 const Instruction* inst = FindDef(id); IsSignedIntVectorType() local 1063 const Instruction* inst = FindDef(id); IsBoolScalarType() local 1068 const Instruction* inst = FindDef(id); IsBoolVectorType() local 1081 const Instruction* inst = FindDef(id); IsBoolScalarOrVectorType() local 1098 const Instruction* inst = FindDef(id); IsFloatMatrixType() local 1142 const Instruction* inst = FindDef(struct_type_id); GetStructMemberTypes() local 1155 const Instruction* inst = FindDef(id); IsPointerType() local 1164 const Instruction* inst = FindDef(id); GetPointerTypeInfo() local 1174 const Instruction* inst = FindDef(id); IsAccelerationStructureType() local 1179 const Instruction* inst = FindDef(id); IsCooperativeMatrixType() local 1185 const Instruction* inst = FindDef(id); IsCooperativeMatrixNVType() local 1190 const Instruction* inst = FindDef(id); IsCooperativeMatrixKHRType() local 1196 const Instruction* inst = FindDef(id); IsCooperativeMatrixAType() local 1207 const Instruction* inst = FindDef(id); IsCooperativeMatrixBType() local 1217 const Instruction* inst = FindDef(id); IsCooperativeMatrixAccType() local 1251 CooperativeMatrixShapesMatch( const Instruction* inst, uint32_t m1, uint32_t m2) CooperativeMatrixShapesMatch() argument 1324 GetOperandTypeId(const Instruction* inst, size_t operand_index) const GetOperandTypeId() argument 1330 const Instruction* inst = FindDef(id); GetConstantValUint64() local 1354 const Instruction* const inst = FindDef(id); EvalInt32IfConst() local 1445 const auto inst = FindDef(id); EntryPointReferences() local 1573 const auto inst = FindDef(id); ContainsType() local [all...] |
H A D | validate_mode_setting.cpp | 27 spv_result_t ValidateEntryPoint(ValidationState_t& _, const Instruction* inst) { in ValidateEntryPoint() argument 28 const auto entry_point_id = inst->GetOperandAs<uint32_t>(1); in ValidateEntryPoint() 31 return _.diag(SPV_ERROR_INVALID_ID, inst) in ValidateEntryPoint() 38 inst->GetOperandAs<spv::ExecutionModel>(0); in ValidateEntryPoint() 43 return _.diag(SPV_ERROR_INVALID_ID, inst) in ValidateEntryPoint() 52 return _.diag(SPV_ERROR_INVALID_ID, inst) in ValidateEntryPoint() 65 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint() 73 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint() 89 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint() 111 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint() 329 ValidateExecutionMode(ValidationState_t& _, const Instruction* inst) ValidateExecutionMode() argument 600 ValidateMemoryModel(ValidationState_t& _, const Instruction* inst) ValidateMemoryModel() argument 639 ModeSettingPass(ValidationState_t& _, const Instruction* inst) ModeSettingPass() argument [all...] |
H A D | validate_mesh_shading.cpp | 25 spv_result_t MeshShadingPass(ValidationState_t& _, const Instruction* inst) { in MeshShadingPass() argument 26 const spv::Op opcode = inst->opcode(); in MeshShadingPass() 29 _.function(inst->function()->id()) in MeshShadingPass() 42 const uint32_t group_count_x = _.GetOperandTypeId(inst, 0); in MeshShadingPass() 45 return _.diag(SPV_ERROR_INVALID_DATA, inst) in MeshShadingPass() 49 const uint32_t group_count_y = _.GetOperandTypeId(inst, 1); in MeshShadingPass() 52 return _.diag(SPV_ERROR_INVALID_DATA, inst) in MeshShadingPass() 56 const uint32_t group_count_z = _.GetOperandTypeId(inst, 2); in MeshShadingPass() 59 return _.diag(SPV_ERROR_INVALID_DATA, inst) in MeshShadingPass() 63 if (inst in MeshShadingPass() [all...] |
/third_party/mesa3d/src/gallium/auxiliary/tgsi/ |
H A D | tgsi_parse.c | 184 struct tgsi_full_instruction *inst = &ctx->FullToken.FullInstruction; in tgsi_parse_token() local 186 memset(inst, 0, sizeof *inst); in tgsi_parse_token() 187 copy_token(&inst->Instruction, &token); in tgsi_parse_token() 189 if (inst->Instruction.Label) { in tgsi_parse_token() 190 next_token( ctx, &inst->Label); in tgsi_parse_token() 193 if (inst->Instruction.Texture) { in tgsi_parse_token() 194 next_token( ctx, &inst->Texture); in tgsi_parse_token() 195 for (i = 0; i < inst->Texture.NumOffsets; i++) { in tgsi_parse_token() 196 next_token( ctx, &inst in tgsi_parse_token() [all...] |
/kernel/linux/linux-5.10/drivers/mailbox/ |
H A D | ti-msgmgr.c | 206 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_queue_rx_interrupt() local 214 if (WARN_ON(!inst)) { in ti_msgmgr_queue_rx_interrupt() 226 desc = inst->desc; in ti_msgmgr_queue_rx_interrupt() 289 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_queue_peek_data() local 290 const struct ti_msgmgr_desc *desc = inst->desc; in ti_msgmgr_queue_peek_data() 316 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_last_tx_done() local 317 const struct ti_msgmgr_desc *desc = inst->desc; in ti_msgmgr_last_tx_done() 349 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_send_data() local 357 if (WARN_ON(!inst)) { in ti_msgmgr_send_data() 361 desc = inst in ti_msgmgr_send_data() 462 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); ti_msgmgr_queue_startup() local 526 struct ti_msgmgr_inst *inst; ti_msgmgr_of_xlate() local 587 ti_msgmgr_queue_setup(int idx, struct device *dev, struct device_node *np, struct ti_msgmgr_inst *inst, const struct ti_msgmgr_desc *d, const struct ti_msgmgr_valid_queue_desc *qd, struct ti_queue_inst *qinst, struct mbox_chan *chan) ti_msgmgr_queue_setup() argument 719 struct ti_msgmgr_inst *inst; ti_msgmgr_probe() local [all...] |
/third_party/mesa3d/src/intel/compiler/ |
H A D | brw_fs_bank_conflicts.cpp | 514 foreach_block_and_inst(block, fs_inst, inst, v->cfg) { 515 if (is_grf(inst->dst)) 516 p.require_contiguous(reg_of(inst->dst), regs_written(inst)); 518 for (int i = 0; i < inst->sources; i++) { 519 if (is_grf(inst->src[i])) 520 p.require_contiguous(reg_of(inst->src[i]), regs_read(inst, i)); 555 foreach_block_and_inst(block, fs_inst, inst, v->cfg) { in foreach_block_and_inst() 561 if (inst in foreach_block_and_inst() 600 is_conflict_optimized_out(const intel_device_info *devinfo, const fs_inst *inst) is_conflict_optimized_out() argument 945 has_bank_conflict(const struct brw_isa_info *isa, const fs_inst *inst) has_bank_conflict() argument [all...] |
H A D | brw_fs_scoreboard.cpp | 74 inferred_sync_pipe(const struct intel_device_info *devinfo, const fs_inst *inst) in inferred_sync_pipe() argument 79 if (is_send(inst)) in inferred_sync_pipe() 82 for (unsigned i = 0; i < inst->sources; i++) { in inferred_sync_pipe() 83 if (inst->src[i].file != BAD_FILE && in inferred_sync_pipe() 84 !inst->is_control_source(i)) { in inferred_sync_pipe() 85 const brw_reg_type t = inst->src[i].type; in inferred_sync_pipe() 106 inferred_exec_pipe(const struct intel_device_info *devinfo, const fs_inst *inst) in inferred_exec_pipe() argument 108 const brw_reg_type t = get_exec_type(inst); in inferred_exec_pipe() 110 ((inst->opcode == BRW_OPCODE_MUL && in inferred_exec_pipe() 111 MIN2(type_sz(inst in inferred_exec_pipe() 152 ordered_unit(const struct intel_device_info *devinfo, const fs_inst *inst, unsigned p) ordered_unit() argument 390 assert(from < n); if (is[from] != from) assign(is[from], to); is[from] = to; } } unsigned *is; unsigned n; }; struct dependency { dependency() : ordered(TGL_REGDIST_NULL), jp(), unordered(TGL_SBID_NULL), id(0), exec_all(false) {} dependency(tgl_regdist_mode mode, const ordered_address &jp, bool exec_all) : ordered(mode), jp(jp), unordered(TGL_SBID_NULL), id(0), exec_all(exec_all) {} dependency(tgl_sbid_mode mode, unsigned id, bool exec_all) : ordered(TGL_REGDIST_NULL), jp(), unordered(mode), id(id), exec_all(exec_all) {} tgl_regdist_mode ordered; ordered_address jp; tgl_sbid_mode unordered; unsigned id; bool exec_all; static const dependency done; friend bool operator==(const dependency &dep0, const dependency &dep1) { return dep0.ordered == dep1.ordered && dep0.jp == dep1.jp && dep0.unordered == dep1.unordered && dep0.id == dep1.id && dep0.exec_all == dep1.exec_all; } friend bool operator!=(const dependency &dep0, const dependency &dep1) { return !(dep0 == dep1); } }; const dependency dependency::done = dependency(TGL_REGDIST_DST, ordered_address(), false); bool is_valid(const dependency &dep) { return dep.ordered || dep.unordered; } dependency merge(equivalence_relation &eq, const dependency &dep0, const dependency &dep1) { dependency dep; if (dep0.ordered || dep1.ordered) { dep.ordered = dep0.ordered | dep1.ordered; for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) dep.jp.jp[p] = MAX2(dep0.jp.jp[p], dep1.jp.jp[p]); } if (dep0.unordered || dep1.unordered) { dep.unordered = dep0.unordered | dep1.unordered; dep.id = eq.link(dep0.unordered ? dep0.id : dep1.id, dep1.unordered ? dep1.id : dep0.id); } dep.exec_all = dep0.exec_all || dep1.exec_all; return dep; } dependency shadow(const dependency &dep0, const dependency &dep1) { if (dep0.ordered == TGL_REGDIST_SRC && is_valid(dep1) && !(dep1.unordered & TGL_SBID_DST) && !(dep1.ordered & TGL_REGDIST_DST)) { dependency dep = dep1; dep.ordered |= dep0.ordered; for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) dep.jp.jp[p] = MAX2(dep.jp.jp[p], dep0.jp.jp[p]); return dep; } else { return is_valid(dep1) ? dep1 : dep0; } } dependency transport(dependency dep, int delta[IDX(TGL_PIPE_ALL)]) { if (dep.ordered) { for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) { if (dep.jp.jp[p] > INT_MIN) dep.jp.jp[p] += delta[p]; } } return dep; } dependency dependency_for_read(dependency dep) { dep.ordered &= TGL_REGDIST_DST; return dep; } dependency dependency_for_write(const struct intel_device_info *devinfo, const fs_inst *inst, dependency dep) { if (!is_unordered(inst) && is_single_pipe(dep.jp, inferred_exec_pipe(devinfo, inst))) dep.ordered &= TGL_REGDIST_DST; return dep; } class scoreboard { public: dependency get(const fs_reg &r) const { if (const dependency *p = const_cast<scoreboard *>(this)->dep(r)) return *p; else return dependency(); } void set(const fs_reg &r, const dependency &d) { if (dependency *p = dep(r)) *p = d; } friend scoreboard merge(equivalence_relation &eq, const scoreboard &sb0, const scoreboard &sb1) { scoreboard sb; for (unsigned i = 0; i < ARRAY_SIZE(sb.grf_deps); i++) sb.grf_deps[i] = merge(eq, sb0.grf_deps[i], sb1.grf_deps[i]); sb.addr_dep = merge(eq, sb0.addr_dep, sb1.addr_dep); sb.accum_dep = merge(eq, sb0.accum_dep, sb1.accum_dep); return sb; } friend scoreboard shadow(const scoreboard &sb0, const scoreboard &sb1) { scoreboard sb; for (unsigned i = 0; i < ARRAY_SIZE(sb.grf_deps); i++) sb.grf_deps[i] = shadow(sb0.grf_deps[i], sb1.grf_deps[i]); sb.addr_dep = shadow(sb0.addr_dep, sb1.addr_dep); sb.accum_dep = shadow(sb0.accum_dep, sb1.accum_dep); return sb; } friend scoreboard transport(const scoreboard &sb0, int delta[IDX(TGL_PIPE_ALL)]) { scoreboard sb; for (unsigned i = 0; i < ARRAY_SIZE(sb.grf_deps); i++) sb.grf_deps[i] = transport(sb0.grf_deps[i], delta); sb.addr_dep = transport(sb0.addr_dep, delta); sb.accum_dep = transport(sb0.accum_dep, delta); return sb; } friend bool operator==(const scoreboard &sb0, const scoreboard &sb1) { for (unsigned i = 0; i < ARRAY_SIZE(sb0.grf_deps); i++) { if (sb0.grf_deps[i] != sb1.grf_deps[i]) return false; } if (sb0.addr_dep != sb1.addr_dep) return false; if (sb0.accum_dep != sb1.accum_dep) return false; return true; } friend bool operator!=(const scoreboard &sb0, const scoreboard &sb1) { return !(sb0 == sb1); } private: dependency grf_deps[BRW_MAX_GRF]; dependency addr_dep; dependency accum_dep; dependency * dep(const fs_reg &r) { const unsigned reg = (r.file == VGRF ? r.nr + r.offset / REG_SIZE : reg_offset(r) / REG_SIZE); return (r.file == VGRF || r.file == FIXED_GRF ? &grf_deps[reg] : r.file == MRF ? &grf_deps[GFX7_MRF_HACK_START + reg] : r.file == ARF && reg >= BRW_ARF_ADDRESS && reg < BRW_ARF_ACCUMULATOR ? &addr_dep : r.file == ARF && reg >= BRW_ARF_ACCUMULATOR && reg < BRW_ARF_FLAG ? &accum_dep : NULL); } }; struct dependency_list { dependency_list() : deps(NULL), n(0) {} ~dependency_list() { free(deps); } void push_back(const dependency &dep) { deps = (dependency *)realloc(deps, (n + 1) * sizeof(*deps)); deps[n++] = dep; } unsigned size() const { return n; } const dependency & operator[](unsigned i) const { assert(i < n); return deps[i]; } dependency & operator[](unsigned i) { assert(i < n); return deps[i]; } private: dependency_list(const dependency_list &); dependency_list & operator=(const dependency_list &); dependency *deps; unsigned n; }; void add_dependency(const unsigned *ids, dependency_list &deps, dependency dep) { if (is_valid(dep)) { if (dep.unordered) dep.id = ids[dep.id]; for (unsigned i = 0; i < deps.size(); i++) { if (deps[i].exec_all != dep.exec_all && (!deps[i].exec_all || (dep.unordered & TGL_SBID_SET)) && (!dep.exec_all || (deps[i].unordered & TGL_SBID_SET))) continue; if (dep.ordered && deps[i].ordered) { for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) deps[i].jp.jp[p] = MAX2(deps[i].jp.jp[p], dep.jp.jp[p]); deps[i].ordered |= dep.ordered; deps[i].exec_all |= dep.exec_all; dep.ordered = TGL_REGDIST_NULL; } if (dep.unordered && deps[i].unordered && deps[i].id == dep.id) { deps[i].unordered |= dep.unordered; deps[i].exec_all |= dep.exec_all; dep.unordered = TGL_SBID_NULL; } } if (is_valid(dep)) deps.push_back(dep); } } tgl_swsb ordered_dependency_swsb(const dependency_list &deps, const ordered_address &jp, bool exec_all) { tgl_pipe p = TGL_PIPE_NONE; unsigned min_dist = ~0u; for (unsigned i = 0; i < deps.size(); i++) { if (deps[i].ordered && exec_all >= deps[i].exec_all) { for (unsigned q = 0; q < IDX(TGL_PIPE_ALL); q++) { const unsigned dist = jp.jp[q] - int64_t(deps[i].jp.jp[q]); const unsigned max_dist = (q == IDX(TGL_PIPE_LONG) ? 14 : 10); assert(jp.jp[q] > deps[i].jp.jp[q]); if (dist <= max_dist) { p = (p && IDX(p) != q ? TGL_PIPE_ALL : tgl_pipe(TGL_PIPE_FLOAT + q)); min_dist = MIN3(min_dist, dist, 7); } } } } return { p ? min_dist : 0, p }; } bool find_ordered_dependency(const dependency_list &deps, const ordered_address &jp, bool exec_all) { return ordered_dependency_swsb(deps, jp, exec_all).regdist; } tgl_sbid_mode find_unordered_dependency(const dependency_list &deps, tgl_sbid_mode unordered, bool exec_all) { if (unordered) { for (unsigned i = 0; i < deps.size(); i++) { if ((unordered & deps[i].unordered) && exec_all >= deps[i].exec_all) return deps[i].unordered; } } return TGL_SBID_NULL; } tgl_sbid_mode baked_unordered_dependency_mode(const struct intel_device_info *devinfo, const fs_inst *inst, const dependency_list &deps, const ordered_address &jp) { const bool exec_all = inst->force_writemask_all; const bool has_ordered = find_ordered_dependency(deps, jp, exec_all); const tgl_pipe ordered_pipe = ordered_dependency_swsb(deps, jp, exec_all).pipe; if (find_unordered_dependency(deps, TGL_SBID_SET, exec_all)) return find_unordered_dependency(deps, TGL_SBID_SET, exec_all); else if (has_ordered && is_unordered(inst)) return TGL_SBID_NULL; else if (find_unordered_dependency(deps, TGL_SBID_DST, exec_all) && (!has_ordered || ordered_pipe == inferred_sync_pipe(devinfo, inst))) return find_unordered_dependency(deps, TGL_SBID_DST, exec_all); else if (!has_ordered) return find_unordered_dependency(deps, TGL_SBID_SRC, exec_all); else return TGL_SBID_NULL; } bool baked_ordered_dependency_mode(const struct intel_device_info *devinfo, const fs_inst *inst, const dependency_list &deps, const ordered_address &jp) { const bool exec_all = inst->force_writemask_all; const bool has_ordered = find_ordered_dependency(deps, jp, exec_all); const tgl_pipe ordered_pipe = ordered_dependency_swsb(deps, jp, exec_all).pipe; const tgl_sbid_mode unordered_mode = baked_unordered_dependency_mode(devinfo, inst, deps, jp); if (!has_ordered) return false; else if (!unordered_mode) return true; else return ordered_pipe == inferred_sync_pipe(devinfo, inst) && unordered_mode == (is_unordered(inst) ? TGL_SBID_SET : TGL_SBID_DST); } void update_inst_scoreboard(const fs_visitor *shader, const ordered_address *jps, const fs_inst *inst, unsigned ip, scoreboard &sb) { const bool exec_all = inst->force_writemask_all; const struct intel_device_info *devinfo = shader->devinfo; const tgl_pipe p = inferred_exec_pipe(devinfo, inst); const ordered_address jp = p ? ordered_address(p, jps[ip].jp[IDX(p)]) : ordered_address(); const bool is_ordered = ordered_unit(devinfo, inst, IDX(TGL_PIPE_ALL)); for (unsigned i = 0; i < inst->sources; i++) { const dependency rd_dep = (inst->is_payload(i) || inst->is_math()) ? dependency(TGL_SBID_SRC, ip, exec_all) : is_ordered ? dependency(TGL_REGDIST_SRC, jp, exec_all) : dependency::done; for (unsigned j = 0; j < regs_read(inst, i); j++) { const fs_reg r = byte_offset(inst->src[i], REG_SIZE * j); sb.set(r, shadow(sb.get(r), rd_dep)); } } if (inst->reads_accumulator_implicitly()) sb.set(brw_acc_reg(8), dependency(TGL_REGDIST_SRC, jp, exec_all)); if (is_send(inst) && inst->base_mrf != -1) { const dependency rd_dep = dependency(TGL_SBID_SRC, ip, exec_all); for (unsigned j = 0; j < inst->mlen; j++) sb.set(brw_uvec_mrf(8, inst->base_mrf + j, 0), rd_dep); } const dependency wr_dep = is_unordered(inst) ? dependency(TGL_SBID_DST, ip, exec_all) : is_ordered ? dependency(TGL_REGDIST_DST, jp, exec_all) : dependency(); if (inst->writes_accumulator_implicitly(devinfo)) sb.set(brw_acc_reg(8), wr_dep); if (is_valid(wr_dep) && inst->dst.file != BAD_FILE && !inst->dst.is_null()) assign() argument [all...] |
H A D | test_eu_validate.cpp | 1254 } inst[] = { in TEST_P() local 1290 for (unsigned i = 0; i < ARRAY_SIZE(inst); i++) { in TEST_P() 1292 inst[i].src_type == BRW_REGISTER_TYPE_DF) in TEST_P() 1296 (inst[i].src_type == BRW_REGISTER_TYPE_Q || in TEST_P() 1297 inst[i].src_type == BRW_REGISTER_TYPE_UQ)) in TEST_P() 1300 brw_MOV(p, retype(g0, inst[i].dst_type), retype(g0, inst[i].src_type)); in TEST_P() 1301 brw_inst_set_dst_hstride(&devinfo, last_inst, inst[i].dst_stride); in TEST_P() 1302 EXPECT_EQ(inst[i].expected_result, validate(p)); in TEST_P() 1317 } inst[] in TEST_P() local 1441 } inst[] = { TEST_P() local 1503 } inst[] = { TEST_P() local 1559 } inst[] = { TEST_P() local 1630 } inst[] = { TEST_P() local 1711 } inst[] = { TEST_P() local 1776 } inst[] = { TEST_P() local 1854 } inst[] = { TEST_P() local 1911 } inst[] = { TEST_P() local 1964 } inst[] = { TEST_P() local 2014 } inst[] = { TEST_P() local 2157 } inst[] = { TEST_P() local 2346 } inst[] = { TEST_P() local 2480 } inst[] = { TEST_P() local 2634 } inst[] = { TEST_P() local 2716 } inst[] = { TEST_P() local 2846 } inst[] = { TEST_P() local [all...] |
/third_party/vulkan-loader/tests/ |
H A D | loader_regression_tests.cpp | 43 InstWrapper inst{env.vulkan_functions}; in TEST() 44 inst.CheckCreate(); in TEST() 51 VkInstance inst = VK_NULL_HANDLE; in TEST() local 53 ASSERT_EQ(env.vulkan_functions.vkCreateInstance(&info, VK_NULL_HANDLE, &inst), VK_SUCCESS); in TEST() 55 env.vulkan_functions.vkDestroyInstance(inst, nullptr); in TEST() 77 InstWrapper inst{env.vulkan_functions}; in TEST() 78 inst.create_info.add_extension("VK_EXT_validation_features"); // test icd won't report this as supported in TEST() 79 inst.CheckCreate(VK_ERROR_EXTENSION_NOT_PRESENT); in TEST() 82 InstWrapper inst{env.vulkan_functions}; in TEST() 83 inst in TEST() 411 exercise_EnumerateDeviceExtensionProperties(InstWrapper& inst, VkPhysicalDevice physical_device, std::vector<Extension>& exts_to_expect) exercise_EnumerateDeviceExtensionProperties() argument 2729 VkInstance inst = VK_NULL_HANDLE; TEST() local 2739 VkInstance inst = VK_NULL_HANDLE; TEST() local [all...] |
/third_party/skia/third_party/externals/spirv-tools/source/lint/ |
H A D | divergence_analysis.cpp | 27 void DivergenceAnalysis::EnqueueSuccessors(opt::Instruction* inst) { in EnqueueSuccessors() argument 33 if (inst->IsBlockTerminator()) { in EnqueueSuccessors() 34 block_id = context().get_instr_block(inst)->id(); in EnqueueSuccessors() 35 } else if (inst->opcode() == SpvOpLabel) { in EnqueueSuccessors() 36 block_id = inst->result_id(); in EnqueueSuccessors() 41 opt::ForwardDataFlowAnalysis::EnqueueUsers(inst); in EnqueueSuccessors() 56 opt::Instruction* inst) { in Visit() 57 if (inst->opcode() == SpvOpLabel) { in Visit() 58 return VisitBlock(inst->result_id()); in Visit() 60 return VisitInstruction(inst); in Visit() 55 Visit( opt::Instruction* inst) Visit() argument 102 VisitInstruction( opt::Instruction* inst) VisitInstruction() argument 123 ComputeInstructionDivergence(opt::Instruction* inst) ComputeInstructionDivergence() argument [all...] |
/third_party/skia/third_party/externals/spirv-tools/source/opt/ |
H A D | code_sink.cpp | 43 for (auto inst = bb->rbegin(); inst != bb->rend(); ++inst) { in SinkInstructionsInBB() 44 if (SinkInstruction(&*inst)) { in SinkInstructionsInBB() 45 inst = bb->rbegin(); in SinkInstructionsInBB() 52 bool CodeSinkingPass::SinkInstruction(Instruction* inst) { in SinkInstruction() argument 53 if (inst->opcode() != SpvOpLoad && inst->opcode() != SpvOpAccessChain) { in SinkInstruction() 57 if (ReferencesMutableMemory(inst)) { in SinkInstruction() 61 if (BasicBlock* target_bb = FindNewBasicBlockFor(inst)) { in SinkInstruction() 74 FindNewBasicBlockFor(Instruction* inst) FindNewBasicBlockFor() argument 170 ReferencesMutableMemory(Instruction* inst) ReferencesMutableMemory() argument [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/source/lint/ |
H A D | divergence_analysis.cpp | 27 void DivergenceAnalysis::EnqueueSuccessors(opt::Instruction* inst) { in EnqueueSuccessors() argument 33 if (inst->IsBlockTerminator()) { in EnqueueSuccessors() 34 block_id = context().get_instr_block(inst)->id(); in EnqueueSuccessors() 35 } else if (inst->opcode() == SpvOpLabel) { in EnqueueSuccessors() 36 block_id = inst->result_id(); in EnqueueSuccessors() 41 opt::ForwardDataFlowAnalysis::EnqueueUsers(inst); in EnqueueSuccessors() 56 opt::Instruction* inst) { in Visit() 57 if (inst->opcode() == SpvOpLabel) { in Visit() 58 return VisitBlock(inst->result_id()); in Visit() 60 return VisitInstruction(inst); in Visit() 55 Visit( opt::Instruction* inst) Visit() argument 102 VisitInstruction( opt::Instruction* inst) VisitInstruction() argument 123 ComputeInstructionDivergence(opt::Instruction* inst) ComputeInstructionDivergence() argument [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/source/opt/ |
H A D | code_sink.cpp | 43 for (auto inst = bb->rbegin(); inst != bb->rend(); ++inst) { in SinkInstructionsInBB() 44 if (SinkInstruction(&*inst)) { in SinkInstructionsInBB() 45 inst = bb->rbegin(); in SinkInstructionsInBB() 52 bool CodeSinkingPass::SinkInstruction(Instruction* inst) { in SinkInstruction() argument 53 if (inst->opcode() != SpvOpLoad && inst->opcode() != SpvOpAccessChain) { in SinkInstruction() 57 if (ReferencesMutableMemory(inst)) { in SinkInstruction() 61 if (BasicBlock* target_bb = FindNewBasicBlockFor(inst)) { in SinkInstruction() 74 FindNewBasicBlockFor(Instruction* inst) FindNewBasicBlockFor() argument 170 ReferencesMutableMemory(Instruction* inst) ReferencesMutableMemory() argument [all...] |
/third_party/spirv-tools/source/opt/ |
H A D | private_to_local_pass.cpp | 41 for (auto& inst : context()->types_values()) { in Process() 42 if (inst.opcode() != spv::Op::OpVariable) { in Process() 46 if (spv::StorageClass(inst.GetSingleWordInOperand( in Process() 51 Function* target_function = FindLocalFunction(inst); in Process() 53 variables_to_move.push_back({&inst, target_function}); in Process() 88 Function* PrivateToLocalPass::FindLocalFunction(const Instruction& inst) const { in FindLocalFunction() 92 inst.result_id(), in FindLocalFunction() 156 bool PrivateToLocalPass::IsValidUse(const Instruction* inst) const { in IsValidUse() 159 if (inst->GetCommonDebugOpcode() == CommonDebugInfoDebugGlobalVariable) { in IsValidUse() 162 switch (inst in IsValidUse() 180 UpdateUse(Instruction* inst, Instruction* user) UpdateUse() argument 221 UpdateUses(Instruction* inst) UpdateUses() argument [all...] |
H A D | loop_fission.cpp | 77 // Checks if |inst| is safe to move. We can only move instructions which don't 79 bool MovableInstruction(const Instruction& inst) const; 82 // Traverse the def use chain of |inst| and add the users and uses of |inst| 84 void TraverseUseDef(Instruction* inst, std::set<Instruction*>* returned_set, 112 bool LoopFissionImpl::MovableInstruction(const Instruction& inst) const { in MovableInstruction() 113 return inst.opcode() == spv::Op::OpLoad || in MovableInstruction() 114 inst.opcode() == spv::Op::OpStore || in MovableInstruction() 115 inst.opcode() == spv::Op::OpSelectionMerge || in MovableInstruction() 116 inst in MovableInstruction() 119 TraverseUseDef(Instruction* inst, std::set<Instruction*>* returned_set, bool ignore_phi_users, bool report_loads) TraverseUseDef() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
H A D | nv04.c | 445 nv04_gr_set_ctx1(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx1() argument 450 tmp = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx1() 453 nvkm_wr32(device, 0x700000 + inst, tmp); in nv04_gr_set_ctx1() 460 nv04_gr_set_ctx_val(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx_val() argument 465 ctx1 = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx_val() 469 tmp = nvkm_rd32(device, 0x70000c + inst); in nv04_gr_set_ctx_val() 472 nvkm_wr32(device, 0x70000c + inst, tmp); in nv04_gr_set_ctx_val() 504 nv04_gr_set_ctx1(device, inst, 0x01000000, valid << 24); in nv04_gr_set_ctx_val() 508 nv04_gr_mthd_set_operation(struct nvkm_device *device, u32 inst, u32 data) in nv04_gr_mthd_set_operation() argument 516 nv04_gr_set_ctx1(device, inst, in nv04_gr_mthd_set_operation() 523 nv04_gr_mthd_surf3d_clip_h(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_surf3d_clip_h() argument 541 nv04_gr_mthd_surf3d_clip_v(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_surf3d_clip_v() argument 559 nv04_gr_mthd_bind_class(struct nvkm_device *device, u32 inst) nv04_gr_mthd_bind_class() argument 565 nv04_gr_mthd_bind_surf2d(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf2d() argument 581 nv04_gr_mthd_bind_surf2d_swzsurf(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf2d_swzsurf() argument 601 nv01_gr_mthd_bind_patt(struct nvkm_device *device, u32 inst, u32 data) nv01_gr_mthd_bind_patt() argument 615 nv04_gr_mthd_bind_patt(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_patt() argument 629 nv04_gr_mthd_bind_rop(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_rop() argument 643 nv04_gr_mthd_bind_beta1(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_beta1() argument 657 nv04_gr_mthd_bind_beta4(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_beta4() argument 671 nv04_gr_mthd_bind_surf_dst(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_dst() argument 685 nv04_gr_mthd_bind_surf_src(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_src() argument 699 nv04_gr_mthd_bind_surf_color(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_color() argument 713 nv04_gr_mthd_bind_surf_zeta(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_zeta() argument 727 nv01_gr_mthd_bind_clip(struct nvkm_device *device, u32 inst, u32 data) nv01_gr_mthd_bind_clip() argument 741 nv01_gr_mthd_bind_chroma(struct nvkm_device *device, u32 inst, u32 data) nv01_gr_mthd_bind_chroma() argument 758 nv03_gr_mthd_gdi(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_gdi() argument 774 nv04_gr_mthd_gdi(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_gdi() argument 791 nv01_gr_mthd_blit(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv01_gr_mthd_blit() argument 810 nv04_gr_mthd_blit(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_blit() argument 829 nv04_gr_mthd_iifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_iifc() argument 848 nv01_gr_mthd_ifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv01_gr_mthd_ifc() argument 866 nv04_gr_mthd_ifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_ifc() argument 885 nv03_gr_mthd_sifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_sifc() argument 902 nv04_gr_mthd_sifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_sifc() argument 920 nv03_gr_mthd_sifm(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_sifm() argument 936 nv04_gr_mthd_sifm(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_sifm() argument 953 nv04_gr_mthd_surf3d(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_surf3d() argument 966 nv03_gr_mthd_ttri(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_ttri() argument 980 nv01_gr_mthd_prim(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv01_gr_mthd_prim() argument 997 nv04_gr_mthd_prim(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_prim() argument 1015 nv04_gr_mthd(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd() argument 1286 u32 inst = (nvkm_rd32(device, 0x40016c) & 0xffff) << 4; nv04_gr_intr() local 1416 nv04_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr) nv04_gr_new() argument [all...] |
/third_party/spirv-tools/source/fuzz/ |
H A D | transformation_move_instruction_down.cpp | 53 auto* inst = FindInstruction(message_.instruction(), ir_context); in IsApplicable() local 54 if (!inst) { in IsApplicable() 59 if (!IsInstructionSupported(ir_context, *inst)) { in IsApplicable() 63 auto* inst_block = ir_context->get_instr_block(inst); in IsApplicable() 67 auto inst_it = fuzzerutil::GetIteratorForInstruction(inst_block, inst); in IsApplicable() 78 if (!IsSimpleInstruction(ir_context, *inst) && in IsApplicable() 86 !CanSafelySwapInstructions(ir_context, *inst, *successor_it, in IsApplicable() 94 !fuzzerutil::CanInsertOpcodeBeforeInstruction(inst->opcode(), in IsApplicable() 100 if (inst->result_id()) { in IsApplicable() 104 operand.words[0] == inst in IsApplicable() 115 auto* inst = FindInstruction(message_.instruction(), ir_context); Apply() local 134 IsInstructionSupported( opt::IRContext* ir_context, const opt::Instruction& inst) IsInstructionSupported() argument 142 IsSimpleInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsSimpleInstruction() argument 345 IsMemoryReadInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsMemoryReadInstruction() argument 414 GetMemoryReadTarget( opt::IRContext* ir_context, const opt::Instruction& inst) GetMemoryReadTarget() argument 492 IsMemoryWriteInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsMemoryWriteInstruction() argument 533 GetMemoryWriteTarget( opt::IRContext* ir_context, const opt::Instruction& inst) GetMemoryWriteTarget() argument 584 IsMemoryInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsMemoryInstruction() argument 590 IsBarrierInstruction( const opt::Instruction& inst) IsBarrierInstruction() argument [all...] |
/third_party/musl/porting/liteos_a/user/src/thread/ |
H A D | pthread_barrier_wait.c | 63 struct instance *inst; in pthread_barrier_wait() local 74 inst = b->_b_inst; in pthread_barrier_wait() 77 if (!inst) { in pthread_barrier_wait() 80 b->_b_inst = inst = &new_inst; in pthread_barrier_wait() 83 while (spins-- && !inst->finished) in pthread_barrier_wait() 85 a_inc(&inst->finished); in pthread_barrier_wait() 86 while (inst->finished == 1) in pthread_barrier_wait() 87 __syscall(SYS_futex,&inst->finished,FUTEX_WAIT|FUTEX_PRIVATE,1,0xffffffffu) != -ENOSYS in pthread_barrier_wait() 88 || __syscall(SYS_futex,&inst->finished,FUTEX_WAIT,1,0xffffffffu); in pthread_barrier_wait() 93 if (++inst in pthread_barrier_wait() [all...] |
/third_party/skia/third_party/externals/tint/src/reader/spirv/ |
H A D | function.cc | 550 auto* inst = bb.GetMergeInst(); in MergeFor() local 551 return inst == nullptr ? 0 : inst->GetSingleWordInOperand(0); in MergeFor() 558 auto* inst = bb.GetLoopMergeInst(); in ContinueTargetFor() local 559 return inst == nullptr ? 0 : inst->GetSingleWordInOperand(1); in ContinueTargetFor() 727 : inst(def_inst), block_pos(the_block_pos), index(the_index) {} in DefInfo() 1540 if (const auto* inst = block.GetMergeInst()) { in RegisterMerges() 1542 switch (inst->opcode()) { in RegisterMerges() 1565 const uint32_t merge = inst in RegisterMerges() 2520 AddressOfIfNeeded( TypedExpression expr, const spvtools::opt::Instruction* inst) AddressOfIfNeeded() argument 2594 const auto* inst = def_use_mgr_->GetDef(id); MakeExpression() local 3473 EmitConstDefinition( const spvtools::opt::Instruction& inst, TypedExpression expr) EmitConstDefinition() argument 3497 EmitConstDefOrWriteToHoistedVar( const spvtools::opt::Instruction& inst, TypedExpression expr) EmitConstDefOrWriteToHoistedVar() argument 3503 WriteIfHoistedVar(const spvtools::opt::Instruction& inst, TypedExpression expr) WriteIfHoistedVar() argument 3520 EmitStatement(const spvtools::opt::Instruction& inst) EmitStatement() argument 3810 MakeOperand( const spvtools::opt::Instruction& inst, uint32_t operand_index) MakeOperand() argument 3835 MaybeEmitCombinatorialValue( const spvtools::opt::Instruction& inst) MaybeEmitCombinatorialValue() argument 4017 EmitGlslStd450ExtInst( const spvtools::opt::Instruction& inst) EmitGlslStd450ExtInst() argument 4202 MakeFMod( const spvtools::opt::Instruction& inst) MakeFMod() argument 4217 MakeAccessChain( const spvtools::opt::Instruction& inst) MakeAccessChain() argument 4438 MakeCompositeExtract( const spvtools::opt::Instruction& inst) MakeCompositeExtract() argument 4457 MakeCompositeValueDecomposition( const spvtools::opt::Instruction& inst, TypedExpression composite, uint32_t composite_type_id, int index_start) MakeCompositeValueDecomposition() argument 4602 MakeVectorShuffle( const spvtools::opt::Instruction& inst) MakeVectorShuffle() argument 4802 const auto& inst = id_def_info_pair.second->inst; FindValuesNeedingNamedOrHoistedDefinition() local 4977 MakeNumericConversion( const spvtools::opt::Instruction& inst) MakeNumericConversion() argument 5033 EmitFunctionCall(const spvtools::opt::Instruction& inst) EmitFunctionCall() argument 5070 EmitControlBarrier( const spvtools::opt::Instruction& inst) EmitControlBarrier() argument 5115 MakeIntrinsicCall( const spvtools::opt::Instruction& inst) MakeIntrinsicCall() argument 5143 MakeSimpleSelect( const spvtools::opt::Instruction& inst) MakeSimpleSelect() argument 5176 GetImage( const spvtools::opt::Instruction& inst) GetImage() argument 5212 GetImageExpression( const spvtools::opt::Instruction& inst) GetImageExpression() argument 5223 GetSamplerExpression( const spvtools::opt::Instruction& inst) GetSamplerExpression() argument 5239 EmitImageAccess(const spvtools::opt::Instruction& inst) EmitImageAccess() argument 5513 EmitImageQuery(const spvtools::opt::Instruction& inst) EmitImageQuery() argument 5591 MakeCoordinateOperandsForImageAccess( const spvtools::opt::Instruction& inst) MakeCoordinateOperandsForImageAccess() argument 5725 ConvertTexelForStorage( const spvtools::opt::Instruction& inst, TypedExpression texel, const Texture* texture_type) ConvertTexelForStorage() argument 5819 MakeArrayLength( const spvtools::opt::Instruction& inst) MakeArrayLength() argument 5860 MakeOuterProduct( const spvtools::opt::Instruction& inst) MakeOuterProduct() argument 5904 MakeVectorInsertDynamic( const spvtools::opt::Instruction& inst) MakeVectorInsertDynamic() argument 5972 MakeCompositeInsert( const spvtools::opt::Instruction& inst) MakeCompositeInsert() argument [all...] |
/kernel/linux/linux-6.6/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_compression.c | 18 void qat_compression_put_instance(struct qat_compression_instance *inst) in qat_compression_put_instance() argument 20 atomic_dec(&inst->refctr); in qat_compression_put_instance() 21 adf_dev_put(inst->accel_dev); in qat_compression_put_instance() 26 struct qat_compression_instance *inst; in qat_compression_free_instances() local 31 inst = list_entry(list_ptr, in qat_compression_free_instances() 34 for (i = 0; i < atomic_read(&inst->refctr); i++) in qat_compression_free_instances() 35 qat_compression_put_instance(inst); in qat_compression_free_instances() 37 if (inst->dc_tx) in qat_compression_free_instances() 38 adf_remove_ring(inst->dc_tx); in qat_compression_free_instances() 40 if (inst in qat_compression_free_instances() 51 struct qat_compression_instance *inst = NULL; qat_compression_get_instance_node() local 116 struct qat_compression_instance *inst; qat_compression_create_instances() local [all...] |