Home
last modified time | relevance | path

Searched refs:inst (Results 326 - 350 of 3173) sorted by relevance

1...<<11121314151617181920>>...127

/third_party/vixl/tools/
H A Dmake_instruction_doc_aarch64.pl41 my %inst = (); # Global hash of instructions.
98 $inst{$p}->{'type'} = $type;
99 $inst{$p}->{'mnemonic'} = $i;
100 $inst{$p}->{'description'} = $d;
101 $inst{$p}->{'initial'} = substr($i, 0, 1);
106 my $links = get_links_list(\%inst);
143 my $inst = $insts->{$i};
144 $initials{$inst->{type}}->{$inst->{initial}}++;
159 $inst{
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/include/nvkm/subdev/
H A Dmc.h20 int nv04_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
21 int nv11_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
22 int nv17_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
23 int nv44_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
24 int nv50_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
25 int g84_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
26 int g98_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
27 int gt215_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
28 int gf100_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mc **);
29 int gk104_mc_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struc
[all...]
/third_party/pcre2/pcre2/src/sljit/
H A DsljitNativeX86_64.c35 sljit_u8 *inst; in emit_load_imm64() local
37 inst = (sljit_u8*)ensure_buf(compiler, 1 + 2 + sizeof(sljit_sw)); in emit_load_imm64()
38 FAIL_IF(!inst); in emit_load_imm64()
40 *inst++ = REX_W | ((reg_map[reg] <= 7) ? 0 : REX_B); in emit_load_imm64()
41 *inst++ = U8(MOV_r_i32 | (reg_map[reg] & 0x7)); in emit_load_imm64()
42 sljit_unaligned_store_sw(inst, imm); in emit_load_imm64()
48 sljit_u8 *inst; in emit_do_imm32() local
51 inst = (sljit_u8*)ensure_buf(compiler, 1 + length); in emit_do_imm32()
52 FAIL_IF(!inst); in emit_do_imm32()
55 *inst in emit_do_imm32()
67 sljit_u8 *inst; emit_x86_instruction() local
386 sljit_u8 *inst; sljit_emit_enter() local
586 sljit_u8 *inst; emit_stack_frame_release() local
659 sljit_u8 *inst; sljit_emit_return_void() local
868 sljit_u8 *inst; sljit_emit_fast_enter() local
901 sljit_u8 *inst; emit_fast_return() local
945 sljit_u8* inst; sljit_emit_mem() local
1025 sljit_u8* inst; emit_mov_int() local
[all...]
/kernel/linux/linux-6.6/drivers/media/platform/qcom/venus/
H A Dpm_helpers.c149 static u32 load_per_instance(struct venus_inst *inst) in load_per_instance() argument
153 if (!inst || !(inst->state >= INST_INIT && inst->state < INST_STOP)) in load_per_instance()
156 mbs = (ALIGN(inst->width, 16) / 16) * (ALIGN(inst->height, 16) / 16); in load_per_instance()
158 return mbs * inst->fps; in load_per_instance()
163 struct venus_inst *inst = NULL; in load_per_type() local
166 list_for_each_entry(inst, &core->instances, list) { in load_per_type()
167 if (inst in load_per_type()
176 mbs_to_bw(struct venus_inst *inst, u32 mbs, u32 *avg, u32 *peak) mbs_to_bw() argument
217 struct venus_inst *inst = NULL; load_scale_bw() local
243 load_scale_v1(struct venus_inst *inst) load_scale_v1() argument
525 power_save_mode_enable(struct venus_inst *inst, bool enable) power_save_mode_enable() argument
555 struct venus_inst *inst = NULL; move_core_to_power_save_mode() local
568 min_loaded_core(struct venus_inst *inst, u32 *min_coreid, u32 *min_load, bool low_power) min_loaded_core() argument
621 decide_core(struct venus_inst *inst) decide_core() argument
682 acquire_core(struct venus_inst *inst) acquire_core() argument
709 release_core(struct venus_inst *inst) release_core() argument
742 coreid_power_v4(struct venus_inst *inst, int on) coreid_power_v4() argument
1075 calculate_inst_freq(struct venus_inst *inst, unsigned long filled_len) calculate_inst_freq() argument
1110 load_scale_v4(struct venus_inst *inst) load_scale_v4() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dnv20.c21 u32 inst = nvkm_memory_addr(chan->inst); in nv20_gr_chan_init() local
24 nvkm_wo32(gr->ctxtab, chan->chid * 4, inst >> 4); in nv20_gr_chan_init()
35 u32 inst = nvkm_memory_addr(chan->inst); in nv20_gr_chan_fini() local
42 nvkm_wr32(device, 0x400784, inst >> 4); in nv20_gr_chan_fini()
63 nvkm_memory_unref(&chan->inst); in nv20_gr_chan_dtor()
91 &chan->inst); in nv20_gr_chan_new()
95 nvkm_kmap(chan->inst); in nv20_gr_chan_new()
96 nvkm_wo32(chan->inst, in nv20_gr_chan_new()
[all...]
/third_party/spirv-tools/source/val/
H A Dvalidation_state.cpp108 const spv_parsed_instruction_t* inst) { in CountInstructions()
110 if (spv::Op(inst->opcode) == spv::Op::OpFunction) { in CountInstructions()
307 const Instruction* inst) { in diag()
320 if (inst) disassembly = Disassemble(*inst); in diag()
322 return DiagnosticStream({0, 0, inst ? inst->LineNum() : 0}, in diag()
527 const spv_parsed_instruction_t* inst) { in AddOrderedInstruction()
528 ordered_instructions_.emplace_back(inst); in AddOrderedInstruction()
534 void ValidationState_t::RegisterDebugInstruction(const Instruction* inst) { in RegisterDebugInstruction() argument
107 CountInstructions(void* user_data, const spv_parsed_instruction_t* inst) CountInstructions() argument
306 diag(spv_result_t error_code, const Instruction* inst) diag() argument
526 AddOrderedInstruction( const spv_parsed_instruction_t* inst) AddOrderedInstruction() argument
559 RegisterInstruction(Instruction* inst) RegisterInstruction() argument
831 RegisterUniqueTypeDeclaration(const Instruction* inst) RegisterUniqueTypeDeclaration() argument
851 const Instruction* inst = FindDef(id); GetTypeId() local
856 const Instruction* inst = FindDef(id); GetIdOpcode() local
861 const Instruction* inst = FindDef(id); GetComponentType() local
891 const Instruction* inst = FindDef(id); GetDimension() local
921 const Instruction* inst = FindDef(component_type_id); GetBitWidth() local
935 const Instruction* inst = FindDef(id); IsVoidType() local
940 const Instruction* inst = FindDef(id); IsFloatScalarType() local
945 const Instruction* inst = FindDef(id); IsFloatVectorType() local
958 const Instruction* inst = FindDef(id); IsFloatScalarOrVectorType() local
975 const Instruction* inst = FindDef(id); IsIntScalarType() local
980 const Instruction* inst = FindDef(id); IsIntVectorType() local
993 const Instruction* inst = FindDef(id); IsIntScalarOrVectorType() local
1010 const Instruction* inst = FindDef(id); IsUnsignedIntScalarType() local
1015 const Instruction* inst = FindDef(id); IsUnsignedIntVectorType() local
1028 const Instruction* inst = FindDef(id); IsUnsignedIntScalarOrVectorType() local
1045 const Instruction* inst = FindDef(id); IsSignedIntScalarType() local
1050 const Instruction* inst = FindDef(id); IsSignedIntVectorType() local
1063 const Instruction* inst = FindDef(id); IsBoolScalarType() local
1068 const Instruction* inst = FindDef(id); IsBoolVectorType() local
1081 const Instruction* inst = FindDef(id); IsBoolScalarOrVectorType() local
1098 const Instruction* inst = FindDef(id); IsFloatMatrixType() local
1142 const Instruction* inst = FindDef(struct_type_id); GetStructMemberTypes() local
1155 const Instruction* inst = FindDef(id); IsPointerType() local
1164 const Instruction* inst = FindDef(id); GetPointerTypeInfo() local
1174 const Instruction* inst = FindDef(id); IsAccelerationStructureType() local
1179 const Instruction* inst = FindDef(id); IsCooperativeMatrixType() local
1185 const Instruction* inst = FindDef(id); IsCooperativeMatrixNVType() local
1190 const Instruction* inst = FindDef(id); IsCooperativeMatrixKHRType() local
1196 const Instruction* inst = FindDef(id); IsCooperativeMatrixAType() local
1207 const Instruction* inst = FindDef(id); IsCooperativeMatrixBType() local
1217 const Instruction* inst = FindDef(id); IsCooperativeMatrixAccType() local
1251 CooperativeMatrixShapesMatch( const Instruction* inst, uint32_t m1, uint32_t m2) CooperativeMatrixShapesMatch() argument
1324 GetOperandTypeId(const Instruction* inst, size_t operand_index) const GetOperandTypeId() argument
1330 const Instruction* inst = FindDef(id); GetConstantValUint64() local
1354 const Instruction* const inst = FindDef(id); EvalInt32IfConst() local
1445 const auto inst = FindDef(id); EntryPointReferences() local
1573 const auto inst = FindDef(id); ContainsType() local
[all...]
H A Dvalidate_mode_setting.cpp27 spv_result_t ValidateEntryPoint(ValidationState_t& _, const Instruction* inst) { in ValidateEntryPoint() argument
28 const auto entry_point_id = inst->GetOperandAs<uint32_t>(1); in ValidateEntryPoint()
31 return _.diag(SPV_ERROR_INVALID_ID, inst) in ValidateEntryPoint()
38 inst->GetOperandAs<spv::ExecutionModel>(0); in ValidateEntryPoint()
43 return _.diag(SPV_ERROR_INVALID_ID, inst) in ValidateEntryPoint()
52 return _.diag(SPV_ERROR_INVALID_ID, inst) in ValidateEntryPoint()
65 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint()
73 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint()
89 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint()
111 return _.diag(SPV_ERROR_INVALID_DATA, inst) in ValidateEntryPoint()
329 ValidateExecutionMode(ValidationState_t& _, const Instruction* inst) ValidateExecutionMode() argument
600 ValidateMemoryModel(ValidationState_t& _, const Instruction* inst) ValidateMemoryModel() argument
639 ModeSettingPass(ValidationState_t& _, const Instruction* inst) ModeSettingPass() argument
[all...]
H A Dvalidate_mesh_shading.cpp25 spv_result_t MeshShadingPass(ValidationState_t& _, const Instruction* inst) { in MeshShadingPass() argument
26 const spv::Op opcode = inst->opcode(); in MeshShadingPass()
29 _.function(inst->function()->id()) in MeshShadingPass()
42 const uint32_t group_count_x = _.GetOperandTypeId(inst, 0); in MeshShadingPass()
45 return _.diag(SPV_ERROR_INVALID_DATA, inst) in MeshShadingPass()
49 const uint32_t group_count_y = _.GetOperandTypeId(inst, 1); in MeshShadingPass()
52 return _.diag(SPV_ERROR_INVALID_DATA, inst) in MeshShadingPass()
56 const uint32_t group_count_z = _.GetOperandTypeId(inst, 2); in MeshShadingPass()
59 return _.diag(SPV_ERROR_INVALID_DATA, inst) in MeshShadingPass()
63 if (inst in MeshShadingPass()
[all...]
/third_party/mesa3d/src/gallium/auxiliary/tgsi/
H A Dtgsi_parse.c184 struct tgsi_full_instruction *inst = &ctx->FullToken.FullInstruction; in tgsi_parse_token() local
186 memset(inst, 0, sizeof *inst); in tgsi_parse_token()
187 copy_token(&inst->Instruction, &token); in tgsi_parse_token()
189 if (inst->Instruction.Label) { in tgsi_parse_token()
190 next_token( ctx, &inst->Label); in tgsi_parse_token()
193 if (inst->Instruction.Texture) { in tgsi_parse_token()
194 next_token( ctx, &inst->Texture); in tgsi_parse_token()
195 for (i = 0; i < inst->Texture.NumOffsets; i++) { in tgsi_parse_token()
196 next_token( ctx, &inst in tgsi_parse_token()
[all...]
/kernel/linux/linux-5.10/drivers/mailbox/
H A Dti-msgmgr.c206 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_queue_rx_interrupt() local
214 if (WARN_ON(!inst)) { in ti_msgmgr_queue_rx_interrupt()
226 desc = inst->desc; in ti_msgmgr_queue_rx_interrupt()
289 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_queue_peek_data() local
290 const struct ti_msgmgr_desc *desc = inst->desc; in ti_msgmgr_queue_peek_data()
316 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_last_tx_done() local
317 const struct ti_msgmgr_desc *desc = inst->desc; in ti_msgmgr_last_tx_done()
349 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); in ti_msgmgr_send_data() local
357 if (WARN_ON(!inst)) { in ti_msgmgr_send_data()
361 desc = inst in ti_msgmgr_send_data()
462 struct ti_msgmgr_inst *inst = dev_get_drvdata(dev); ti_msgmgr_queue_startup() local
526 struct ti_msgmgr_inst *inst; ti_msgmgr_of_xlate() local
587 ti_msgmgr_queue_setup(int idx, struct device *dev, struct device_node *np, struct ti_msgmgr_inst *inst, const struct ti_msgmgr_desc *d, const struct ti_msgmgr_valid_queue_desc *qd, struct ti_queue_inst *qinst, struct mbox_chan *chan) ti_msgmgr_queue_setup() argument
719 struct ti_msgmgr_inst *inst; ti_msgmgr_probe() local
[all...]
/third_party/mesa3d/src/intel/compiler/
H A Dbrw_fs_bank_conflicts.cpp514 foreach_block_and_inst(block, fs_inst, inst, v->cfg) {
515 if (is_grf(inst->dst))
516 p.require_contiguous(reg_of(inst->dst), regs_written(inst));
518 for (int i = 0; i < inst->sources; i++) {
519 if (is_grf(inst->src[i]))
520 p.require_contiguous(reg_of(inst->src[i]), regs_read(inst, i));
555 foreach_block_and_inst(block, fs_inst, inst, v->cfg) { in foreach_block_and_inst()
561 if (inst in foreach_block_and_inst()
600 is_conflict_optimized_out(const intel_device_info *devinfo, const fs_inst *inst) is_conflict_optimized_out() argument
945 has_bank_conflict(const struct brw_isa_info *isa, const fs_inst *inst) has_bank_conflict() argument
[all...]
H A Dbrw_fs_scoreboard.cpp74 inferred_sync_pipe(const struct intel_device_info *devinfo, const fs_inst *inst) in inferred_sync_pipe() argument
79 if (is_send(inst)) in inferred_sync_pipe()
82 for (unsigned i = 0; i < inst->sources; i++) { in inferred_sync_pipe()
83 if (inst->src[i].file != BAD_FILE && in inferred_sync_pipe()
84 !inst->is_control_source(i)) { in inferred_sync_pipe()
85 const brw_reg_type t = inst->src[i].type; in inferred_sync_pipe()
106 inferred_exec_pipe(const struct intel_device_info *devinfo, const fs_inst *inst) in inferred_exec_pipe() argument
108 const brw_reg_type t = get_exec_type(inst); in inferred_exec_pipe()
110 ((inst->opcode == BRW_OPCODE_MUL && in inferred_exec_pipe()
111 MIN2(type_sz(inst in inferred_exec_pipe()
152 ordered_unit(const struct intel_device_info *devinfo, const fs_inst *inst, unsigned p) ordered_unit() argument
390 assert(from < n); if (is[from] != from) assign(is[from], to); is[from] = to; } } unsigned *is; unsigned n; }; struct dependency { dependency() : ordered(TGL_REGDIST_NULL), jp(), unordered(TGL_SBID_NULL), id(0), exec_all(false) {} dependency(tgl_regdist_mode mode, const ordered_address &jp, bool exec_all) : ordered(mode), jp(jp), unordered(TGL_SBID_NULL), id(0), exec_all(exec_all) {} dependency(tgl_sbid_mode mode, unsigned id, bool exec_all) : ordered(TGL_REGDIST_NULL), jp(), unordered(mode), id(id), exec_all(exec_all) {} tgl_regdist_mode ordered; ordered_address jp; tgl_sbid_mode unordered; unsigned id; bool exec_all; static const dependency done; friend bool operator==(const dependency &dep0, const dependency &dep1) { return dep0.ordered == dep1.ordered && dep0.jp == dep1.jp && dep0.unordered == dep1.unordered && dep0.id == dep1.id && dep0.exec_all == dep1.exec_all; } friend bool operator!=(const dependency &dep0, const dependency &dep1) { return !(dep0 == dep1); } }; const dependency dependency::done = dependency(TGL_REGDIST_DST, ordered_address(), false); bool is_valid(const dependency &dep) { return dep.ordered || dep.unordered; } dependency merge(equivalence_relation &eq, const dependency &dep0, const dependency &dep1) { dependency dep; if (dep0.ordered || dep1.ordered) { dep.ordered = dep0.ordered | dep1.ordered; for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) dep.jp.jp[p] = MAX2(dep0.jp.jp[p], dep1.jp.jp[p]); } if (dep0.unordered || dep1.unordered) { dep.unordered = dep0.unordered | dep1.unordered; dep.id = eq.link(dep0.unordered ? dep0.id : dep1.id, dep1.unordered ? dep1.id : dep0.id); } dep.exec_all = dep0.exec_all || dep1.exec_all; return dep; } dependency shadow(const dependency &dep0, const dependency &dep1) { if (dep0.ordered == TGL_REGDIST_SRC && is_valid(dep1) && !(dep1.unordered & TGL_SBID_DST) && !(dep1.ordered & TGL_REGDIST_DST)) { dependency dep = dep1; dep.ordered |= dep0.ordered; for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) dep.jp.jp[p] = MAX2(dep.jp.jp[p], dep0.jp.jp[p]); return dep; } else { return is_valid(dep1) ? dep1 : dep0; } } dependency transport(dependency dep, int delta[IDX(TGL_PIPE_ALL)]) { if (dep.ordered) { for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) { if (dep.jp.jp[p] > INT_MIN) dep.jp.jp[p] += delta[p]; } } return dep; } dependency dependency_for_read(dependency dep) { dep.ordered &= TGL_REGDIST_DST; return dep; } dependency dependency_for_write(const struct intel_device_info *devinfo, const fs_inst *inst, dependency dep) { if (!is_unordered(inst) && is_single_pipe(dep.jp, inferred_exec_pipe(devinfo, inst))) dep.ordered &= TGL_REGDIST_DST; return dep; } class scoreboard { public: dependency get(const fs_reg &r) const { if (const dependency *p = const_cast<scoreboard *>(this)->dep(r)) return *p; else return dependency(); } void set(const fs_reg &r, const dependency &d) { if (dependency *p = dep(r)) *p = d; } friend scoreboard merge(equivalence_relation &eq, const scoreboard &sb0, const scoreboard &sb1) { scoreboard sb; for (unsigned i = 0; i < ARRAY_SIZE(sb.grf_deps); i++) sb.grf_deps[i] = merge(eq, sb0.grf_deps[i], sb1.grf_deps[i]); sb.addr_dep = merge(eq, sb0.addr_dep, sb1.addr_dep); sb.accum_dep = merge(eq, sb0.accum_dep, sb1.accum_dep); return sb; } friend scoreboard shadow(const scoreboard &sb0, const scoreboard &sb1) { scoreboard sb; for (unsigned i = 0; i < ARRAY_SIZE(sb.grf_deps); i++) sb.grf_deps[i] = shadow(sb0.grf_deps[i], sb1.grf_deps[i]); sb.addr_dep = shadow(sb0.addr_dep, sb1.addr_dep); sb.accum_dep = shadow(sb0.accum_dep, sb1.accum_dep); return sb; } friend scoreboard transport(const scoreboard &sb0, int delta[IDX(TGL_PIPE_ALL)]) { scoreboard sb; for (unsigned i = 0; i < ARRAY_SIZE(sb.grf_deps); i++) sb.grf_deps[i] = transport(sb0.grf_deps[i], delta); sb.addr_dep = transport(sb0.addr_dep, delta); sb.accum_dep = transport(sb0.accum_dep, delta); return sb; } friend bool operator==(const scoreboard &sb0, const scoreboard &sb1) { for (unsigned i = 0; i < ARRAY_SIZE(sb0.grf_deps); i++) { if (sb0.grf_deps[i] != sb1.grf_deps[i]) return false; } if (sb0.addr_dep != sb1.addr_dep) return false; if (sb0.accum_dep != sb1.accum_dep) return false; return true; } friend bool operator!=(const scoreboard &sb0, const scoreboard &sb1) { return !(sb0 == sb1); } private: dependency grf_deps[BRW_MAX_GRF]; dependency addr_dep; dependency accum_dep; dependency * dep(const fs_reg &r) { const unsigned reg = (r.file == VGRF ? r.nr + r.offset / REG_SIZE : reg_offset(r) / REG_SIZE); return (r.file == VGRF || r.file == FIXED_GRF ? &grf_deps[reg] : r.file == MRF ? &grf_deps[GFX7_MRF_HACK_START + reg] : r.file == ARF && reg >= BRW_ARF_ADDRESS && reg < BRW_ARF_ACCUMULATOR ? &addr_dep : r.file == ARF && reg >= BRW_ARF_ACCUMULATOR && reg < BRW_ARF_FLAG ? &accum_dep : NULL); } }; struct dependency_list { dependency_list() : deps(NULL), n(0) {} ~dependency_list() { free(deps); } void push_back(const dependency &dep) { deps = (dependency *)realloc(deps, (n + 1) * sizeof(*deps)); deps[n++] = dep; } unsigned size() const { return n; } const dependency & operator[](unsigned i) const { assert(i < n); return deps[i]; } dependency & operator[](unsigned i) { assert(i < n); return deps[i]; } private: dependency_list(const dependency_list &); dependency_list & operator=(const dependency_list &); dependency *deps; unsigned n; }; void add_dependency(const unsigned *ids, dependency_list &deps, dependency dep) { if (is_valid(dep)) { if (dep.unordered) dep.id = ids[dep.id]; for (unsigned i = 0; i < deps.size(); i++) { if (deps[i].exec_all != dep.exec_all && (!deps[i].exec_all || (dep.unordered & TGL_SBID_SET)) && (!dep.exec_all || (deps[i].unordered & TGL_SBID_SET))) continue; if (dep.ordered && deps[i].ordered) { for (unsigned p = 0; p < IDX(TGL_PIPE_ALL); p++) deps[i].jp.jp[p] = MAX2(deps[i].jp.jp[p], dep.jp.jp[p]); deps[i].ordered |= dep.ordered; deps[i].exec_all |= dep.exec_all; dep.ordered = TGL_REGDIST_NULL; } if (dep.unordered && deps[i].unordered && deps[i].id == dep.id) { deps[i].unordered |= dep.unordered; deps[i].exec_all |= dep.exec_all; dep.unordered = TGL_SBID_NULL; } } if (is_valid(dep)) deps.push_back(dep); } } tgl_swsb ordered_dependency_swsb(const dependency_list &deps, const ordered_address &jp, bool exec_all) { tgl_pipe p = TGL_PIPE_NONE; unsigned min_dist = ~0u; for (unsigned i = 0; i < deps.size(); i++) { if (deps[i].ordered && exec_all >= deps[i].exec_all) { for (unsigned q = 0; q < IDX(TGL_PIPE_ALL); q++) { const unsigned dist = jp.jp[q] - int64_t(deps[i].jp.jp[q]); const unsigned max_dist = (q == IDX(TGL_PIPE_LONG) ? 14 : 10); assert(jp.jp[q] > deps[i].jp.jp[q]); if (dist <= max_dist) { p = (p && IDX(p) != q ? TGL_PIPE_ALL : tgl_pipe(TGL_PIPE_FLOAT + q)); min_dist = MIN3(min_dist, dist, 7); } } } } return { p ? min_dist : 0, p }; } bool find_ordered_dependency(const dependency_list &deps, const ordered_address &jp, bool exec_all) { return ordered_dependency_swsb(deps, jp, exec_all).regdist; } tgl_sbid_mode find_unordered_dependency(const dependency_list &deps, tgl_sbid_mode unordered, bool exec_all) { if (unordered) { for (unsigned i = 0; i < deps.size(); i++) { if ((unordered & deps[i].unordered) && exec_all >= deps[i].exec_all) return deps[i].unordered; } } return TGL_SBID_NULL; } tgl_sbid_mode baked_unordered_dependency_mode(const struct intel_device_info *devinfo, const fs_inst *inst, const dependency_list &deps, const ordered_address &jp) { const bool exec_all = inst->force_writemask_all; const bool has_ordered = find_ordered_dependency(deps, jp, exec_all); const tgl_pipe ordered_pipe = ordered_dependency_swsb(deps, jp, exec_all).pipe; if (find_unordered_dependency(deps, TGL_SBID_SET, exec_all)) return find_unordered_dependency(deps, TGL_SBID_SET, exec_all); else if (has_ordered && is_unordered(inst)) return TGL_SBID_NULL; else if (find_unordered_dependency(deps, TGL_SBID_DST, exec_all) && (!has_ordered || ordered_pipe == inferred_sync_pipe(devinfo, inst))) return find_unordered_dependency(deps, TGL_SBID_DST, exec_all); else if (!has_ordered) return find_unordered_dependency(deps, TGL_SBID_SRC, exec_all); else return TGL_SBID_NULL; } bool baked_ordered_dependency_mode(const struct intel_device_info *devinfo, const fs_inst *inst, const dependency_list &deps, const ordered_address &jp) { const bool exec_all = inst->force_writemask_all; const bool has_ordered = find_ordered_dependency(deps, jp, exec_all); const tgl_pipe ordered_pipe = ordered_dependency_swsb(deps, jp, exec_all).pipe; const tgl_sbid_mode unordered_mode = baked_unordered_dependency_mode(devinfo, inst, deps, jp); if (!has_ordered) return false; else if (!unordered_mode) return true; else return ordered_pipe == inferred_sync_pipe(devinfo, inst) && unordered_mode == (is_unordered(inst) ? TGL_SBID_SET : TGL_SBID_DST); } void update_inst_scoreboard(const fs_visitor *shader, const ordered_address *jps, const fs_inst *inst, unsigned ip, scoreboard &sb) { const bool exec_all = inst->force_writemask_all; const struct intel_device_info *devinfo = shader->devinfo; const tgl_pipe p = inferred_exec_pipe(devinfo, inst); const ordered_address jp = p ? ordered_address(p, jps[ip].jp[IDX(p)]) : ordered_address(); const bool is_ordered = ordered_unit(devinfo, inst, IDX(TGL_PIPE_ALL)); for (unsigned i = 0; i < inst->sources; i++) { const dependency rd_dep = (inst->is_payload(i) || inst->is_math()) ? dependency(TGL_SBID_SRC, ip, exec_all) : is_ordered ? dependency(TGL_REGDIST_SRC, jp, exec_all) : dependency::done; for (unsigned j = 0; j < regs_read(inst, i); j++) { const fs_reg r = byte_offset(inst->src[i], REG_SIZE * j); sb.set(r, shadow(sb.get(r), rd_dep)); } } if (inst->reads_accumulator_implicitly()) sb.set(brw_acc_reg(8), dependency(TGL_REGDIST_SRC, jp, exec_all)); if (is_send(inst) && inst->base_mrf != -1) { const dependency rd_dep = dependency(TGL_SBID_SRC, ip, exec_all); for (unsigned j = 0; j < inst->mlen; j++) sb.set(brw_uvec_mrf(8, inst->base_mrf + j, 0), rd_dep); } const dependency wr_dep = is_unordered(inst) ? dependency(TGL_SBID_DST, ip, exec_all) : is_ordered ? dependency(TGL_REGDIST_DST, jp, exec_all) : dependency(); if (inst->writes_accumulator_implicitly(devinfo)) sb.set(brw_acc_reg(8), wr_dep); if (is_valid(wr_dep) && inst->dst.file != BAD_FILE && !inst->dst.is_null()) assign() argument
[all...]
H A Dtest_eu_validate.cpp1254 } inst[] = { in TEST_P() local
1290 for (unsigned i = 0; i < ARRAY_SIZE(inst); i++) { in TEST_P()
1292 inst[i].src_type == BRW_REGISTER_TYPE_DF) in TEST_P()
1296 (inst[i].src_type == BRW_REGISTER_TYPE_Q || in TEST_P()
1297 inst[i].src_type == BRW_REGISTER_TYPE_UQ)) in TEST_P()
1300 brw_MOV(p, retype(g0, inst[i].dst_type), retype(g0, inst[i].src_type)); in TEST_P()
1301 brw_inst_set_dst_hstride(&devinfo, last_inst, inst[i].dst_stride); in TEST_P()
1302 EXPECT_EQ(inst[i].expected_result, validate(p)); in TEST_P()
1317 } inst[] in TEST_P() local
1441 } inst[] = { TEST_P() local
1503 } inst[] = { TEST_P() local
1559 } inst[] = { TEST_P() local
1630 } inst[] = { TEST_P() local
1711 } inst[] = { TEST_P() local
1776 } inst[] = { TEST_P() local
1854 } inst[] = { TEST_P() local
1911 } inst[] = { TEST_P() local
1964 } inst[] = { TEST_P() local
2014 } inst[] = { TEST_P() local
2157 } inst[] = { TEST_P() local
2346 } inst[] = { TEST_P() local
2480 } inst[] = { TEST_P() local
2634 } inst[] = { TEST_P() local
2716 } inst[] = { TEST_P() local
2846 } inst[] = { TEST_P() local
[all...]
/third_party/vulkan-loader/tests/
H A Dloader_regression_tests.cpp43 InstWrapper inst{env.vulkan_functions}; in TEST()
44 inst.CheckCreate(); in TEST()
51 VkInstance inst = VK_NULL_HANDLE; in TEST() local
53 ASSERT_EQ(env.vulkan_functions.vkCreateInstance(&info, VK_NULL_HANDLE, &inst), VK_SUCCESS); in TEST()
55 env.vulkan_functions.vkDestroyInstance(inst, nullptr); in TEST()
77 InstWrapper inst{env.vulkan_functions}; in TEST()
78 inst.create_info.add_extension("VK_EXT_validation_features"); // test icd won't report this as supported in TEST()
79 inst.CheckCreate(VK_ERROR_EXTENSION_NOT_PRESENT); in TEST()
82 InstWrapper inst{env.vulkan_functions}; in TEST()
83 inst in TEST()
411 exercise_EnumerateDeviceExtensionProperties(InstWrapper& inst, VkPhysicalDevice physical_device, std::vector<Extension>& exts_to_expect) exercise_EnumerateDeviceExtensionProperties() argument
2729 VkInstance inst = VK_NULL_HANDLE; TEST() local
2739 VkInstance inst = VK_NULL_HANDLE; TEST() local
[all...]
/third_party/skia/third_party/externals/spirv-tools/source/lint/
H A Ddivergence_analysis.cpp27 void DivergenceAnalysis::EnqueueSuccessors(opt::Instruction* inst) { in EnqueueSuccessors() argument
33 if (inst->IsBlockTerminator()) { in EnqueueSuccessors()
34 block_id = context().get_instr_block(inst)->id(); in EnqueueSuccessors()
35 } else if (inst->opcode() == SpvOpLabel) { in EnqueueSuccessors()
36 block_id = inst->result_id(); in EnqueueSuccessors()
41 opt::ForwardDataFlowAnalysis::EnqueueUsers(inst); in EnqueueSuccessors()
56 opt::Instruction* inst) { in Visit()
57 if (inst->opcode() == SpvOpLabel) { in Visit()
58 return VisitBlock(inst->result_id()); in Visit()
60 return VisitInstruction(inst); in Visit()
55 Visit( opt::Instruction* inst) Visit() argument
102 VisitInstruction( opt::Instruction* inst) VisitInstruction() argument
123 ComputeInstructionDivergence(opt::Instruction* inst) ComputeInstructionDivergence() argument
[all...]
/third_party/skia/third_party/externals/spirv-tools/source/opt/
H A Dcode_sink.cpp43 for (auto inst = bb->rbegin(); inst != bb->rend(); ++inst) { in SinkInstructionsInBB()
44 if (SinkInstruction(&*inst)) { in SinkInstructionsInBB()
45 inst = bb->rbegin(); in SinkInstructionsInBB()
52 bool CodeSinkingPass::SinkInstruction(Instruction* inst) { in SinkInstruction() argument
53 if (inst->opcode() != SpvOpLoad && inst->opcode() != SpvOpAccessChain) { in SinkInstruction()
57 if (ReferencesMutableMemory(inst)) { in SinkInstruction()
61 if (BasicBlock* target_bb = FindNewBasicBlockFor(inst)) { in SinkInstruction()
74 FindNewBasicBlockFor(Instruction* inst) FindNewBasicBlockFor() argument
170 ReferencesMutableMemory(Instruction* inst) ReferencesMutableMemory() argument
[all...]
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/source/lint/
H A Ddivergence_analysis.cpp27 void DivergenceAnalysis::EnqueueSuccessors(opt::Instruction* inst) { in EnqueueSuccessors() argument
33 if (inst->IsBlockTerminator()) { in EnqueueSuccessors()
34 block_id = context().get_instr_block(inst)->id(); in EnqueueSuccessors()
35 } else if (inst->opcode() == SpvOpLabel) { in EnqueueSuccessors()
36 block_id = inst->result_id(); in EnqueueSuccessors()
41 opt::ForwardDataFlowAnalysis::EnqueueUsers(inst); in EnqueueSuccessors()
56 opt::Instruction* inst) { in Visit()
57 if (inst->opcode() == SpvOpLabel) { in Visit()
58 return VisitBlock(inst->result_id()); in Visit()
60 return VisitInstruction(inst); in Visit()
55 Visit( opt::Instruction* inst) Visit() argument
102 VisitInstruction( opt::Instruction* inst) VisitInstruction() argument
123 ComputeInstructionDivergence(opt::Instruction* inst) ComputeInstructionDivergence() argument
[all...]
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/source/opt/
H A Dcode_sink.cpp43 for (auto inst = bb->rbegin(); inst != bb->rend(); ++inst) { in SinkInstructionsInBB()
44 if (SinkInstruction(&*inst)) { in SinkInstructionsInBB()
45 inst = bb->rbegin(); in SinkInstructionsInBB()
52 bool CodeSinkingPass::SinkInstruction(Instruction* inst) { in SinkInstruction() argument
53 if (inst->opcode() != SpvOpLoad && inst->opcode() != SpvOpAccessChain) { in SinkInstruction()
57 if (ReferencesMutableMemory(inst)) { in SinkInstruction()
61 if (BasicBlock* target_bb = FindNewBasicBlockFor(inst)) { in SinkInstruction()
74 FindNewBasicBlockFor(Instruction* inst) FindNewBasicBlockFor() argument
170 ReferencesMutableMemory(Instruction* inst) ReferencesMutableMemory() argument
[all...]
/third_party/spirv-tools/source/opt/
H A Dprivate_to_local_pass.cpp41 for (auto& inst : context()->types_values()) { in Process()
42 if (inst.opcode() != spv::Op::OpVariable) { in Process()
46 if (spv::StorageClass(inst.GetSingleWordInOperand( in Process()
51 Function* target_function = FindLocalFunction(inst); in Process()
53 variables_to_move.push_back({&inst, target_function}); in Process()
88 Function* PrivateToLocalPass::FindLocalFunction(const Instruction& inst) const { in FindLocalFunction()
92 inst.result_id(), in FindLocalFunction()
156 bool PrivateToLocalPass::IsValidUse(const Instruction* inst) const { in IsValidUse()
159 if (inst->GetCommonDebugOpcode() == CommonDebugInfoDebugGlobalVariable) { in IsValidUse()
162 switch (inst in IsValidUse()
180 UpdateUse(Instruction* inst, Instruction* user) UpdateUse() argument
221 UpdateUses(Instruction* inst) UpdateUses() argument
[all...]
H A Dloop_fission.cpp77 // Checks if |inst| is safe to move. We can only move instructions which don't
79 bool MovableInstruction(const Instruction& inst) const;
82 // Traverse the def use chain of |inst| and add the users and uses of |inst|
84 void TraverseUseDef(Instruction* inst, std::set<Instruction*>* returned_set,
112 bool LoopFissionImpl::MovableInstruction(const Instruction& inst) const { in MovableInstruction()
113 return inst.opcode() == spv::Op::OpLoad || in MovableInstruction()
114 inst.opcode() == spv::Op::OpStore || in MovableInstruction()
115 inst.opcode() == spv::Op::OpSelectionMerge || in MovableInstruction()
116 inst in MovableInstruction()
119 TraverseUseDef(Instruction* inst, std::set<Instruction*>* returned_set, bool ignore_phi_users, bool report_loads) TraverseUseDef() argument
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dnv04.c445 nv04_gr_set_ctx1(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx1() argument
450 tmp = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx1()
453 nvkm_wr32(device, 0x700000 + inst, tmp); in nv04_gr_set_ctx1()
460 nv04_gr_set_ctx_val(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx_val() argument
465 ctx1 = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx_val()
469 tmp = nvkm_rd32(device, 0x70000c + inst); in nv04_gr_set_ctx_val()
472 nvkm_wr32(device, 0x70000c + inst, tmp); in nv04_gr_set_ctx_val()
504 nv04_gr_set_ctx1(device, inst, 0x01000000, valid << 24); in nv04_gr_set_ctx_val()
508 nv04_gr_mthd_set_operation(struct nvkm_device *device, u32 inst, u32 data) in nv04_gr_mthd_set_operation() argument
516 nv04_gr_set_ctx1(device, inst, in nv04_gr_mthd_set_operation()
523 nv04_gr_mthd_surf3d_clip_h(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_surf3d_clip_h() argument
541 nv04_gr_mthd_surf3d_clip_v(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_surf3d_clip_v() argument
559 nv04_gr_mthd_bind_class(struct nvkm_device *device, u32 inst) nv04_gr_mthd_bind_class() argument
565 nv04_gr_mthd_bind_surf2d(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf2d() argument
581 nv04_gr_mthd_bind_surf2d_swzsurf(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf2d_swzsurf() argument
601 nv01_gr_mthd_bind_patt(struct nvkm_device *device, u32 inst, u32 data) nv01_gr_mthd_bind_patt() argument
615 nv04_gr_mthd_bind_patt(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_patt() argument
629 nv04_gr_mthd_bind_rop(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_rop() argument
643 nv04_gr_mthd_bind_beta1(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_beta1() argument
657 nv04_gr_mthd_bind_beta4(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_beta4() argument
671 nv04_gr_mthd_bind_surf_dst(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_dst() argument
685 nv04_gr_mthd_bind_surf_src(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_src() argument
699 nv04_gr_mthd_bind_surf_color(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_color() argument
713 nv04_gr_mthd_bind_surf_zeta(struct nvkm_device *device, u32 inst, u32 data) nv04_gr_mthd_bind_surf_zeta() argument
727 nv01_gr_mthd_bind_clip(struct nvkm_device *device, u32 inst, u32 data) nv01_gr_mthd_bind_clip() argument
741 nv01_gr_mthd_bind_chroma(struct nvkm_device *device, u32 inst, u32 data) nv01_gr_mthd_bind_chroma() argument
758 nv03_gr_mthd_gdi(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_gdi() argument
774 nv04_gr_mthd_gdi(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_gdi() argument
791 nv01_gr_mthd_blit(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv01_gr_mthd_blit() argument
810 nv04_gr_mthd_blit(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_blit() argument
829 nv04_gr_mthd_iifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_iifc() argument
848 nv01_gr_mthd_ifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv01_gr_mthd_ifc() argument
866 nv04_gr_mthd_ifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_ifc() argument
885 nv03_gr_mthd_sifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_sifc() argument
902 nv04_gr_mthd_sifc(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_sifc() argument
920 nv03_gr_mthd_sifm(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_sifm() argument
936 nv04_gr_mthd_sifm(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_sifm() argument
953 nv04_gr_mthd_surf3d(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_surf3d() argument
966 nv03_gr_mthd_ttri(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv03_gr_mthd_ttri() argument
980 nv01_gr_mthd_prim(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv01_gr_mthd_prim() argument
997 nv04_gr_mthd_prim(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd_prim() argument
1015 nv04_gr_mthd(struct nvkm_device *device, u32 inst, u32 mthd, u32 data) nv04_gr_mthd() argument
1286 u32 inst = (nvkm_rd32(device, 0x40016c) & 0xffff) << 4; nv04_gr_intr() local
1416 nv04_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr) nv04_gr_new() argument
[all...]
/third_party/spirv-tools/source/fuzz/
H A Dtransformation_move_instruction_down.cpp53 auto* inst = FindInstruction(message_.instruction(), ir_context); in IsApplicable() local
54 if (!inst) { in IsApplicable()
59 if (!IsInstructionSupported(ir_context, *inst)) { in IsApplicable()
63 auto* inst_block = ir_context->get_instr_block(inst); in IsApplicable()
67 auto inst_it = fuzzerutil::GetIteratorForInstruction(inst_block, inst); in IsApplicable()
78 if (!IsSimpleInstruction(ir_context, *inst) && in IsApplicable()
86 !CanSafelySwapInstructions(ir_context, *inst, *successor_it, in IsApplicable()
94 !fuzzerutil::CanInsertOpcodeBeforeInstruction(inst->opcode(), in IsApplicable()
100 if (inst->result_id()) { in IsApplicable()
104 operand.words[0] == inst in IsApplicable()
115 auto* inst = FindInstruction(message_.instruction(), ir_context); Apply() local
134 IsInstructionSupported( opt::IRContext* ir_context, const opt::Instruction& inst) IsInstructionSupported() argument
142 IsSimpleInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsSimpleInstruction() argument
345 IsMemoryReadInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsMemoryReadInstruction() argument
414 GetMemoryReadTarget( opt::IRContext* ir_context, const opt::Instruction& inst) GetMemoryReadTarget() argument
492 IsMemoryWriteInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsMemoryWriteInstruction() argument
533 GetMemoryWriteTarget( opt::IRContext* ir_context, const opt::Instruction& inst) GetMemoryWriteTarget() argument
584 IsMemoryInstruction( opt::IRContext* ir_context, const opt::Instruction& inst) IsMemoryInstruction() argument
590 IsBarrierInstruction( const opt::Instruction& inst) IsBarrierInstruction() argument
[all...]
/third_party/musl/porting/liteos_a/user/src/thread/
H A Dpthread_barrier_wait.c63 struct instance *inst; in pthread_barrier_wait() local
74 inst = b->_b_inst; in pthread_barrier_wait()
77 if (!inst) { in pthread_barrier_wait()
80 b->_b_inst = inst = &new_inst; in pthread_barrier_wait()
83 while (spins-- && !inst->finished) in pthread_barrier_wait()
85 a_inc(&inst->finished); in pthread_barrier_wait()
86 while (inst->finished == 1) in pthread_barrier_wait()
87 __syscall(SYS_futex,&inst->finished,FUTEX_WAIT|FUTEX_PRIVATE,1,0xffffffffu) != -ENOSYS in pthread_barrier_wait()
88 || __syscall(SYS_futex,&inst->finished,FUTEX_WAIT,1,0xffffffffu); in pthread_barrier_wait()
93 if (++inst in pthread_barrier_wait()
[all...]
/third_party/skia/third_party/externals/tint/src/reader/spirv/
H A Dfunction.cc550 auto* inst = bb.GetMergeInst(); in MergeFor() local
551 return inst == nullptr ? 0 : inst->GetSingleWordInOperand(0); in MergeFor()
558 auto* inst = bb.GetLoopMergeInst(); in ContinueTargetFor() local
559 return inst == nullptr ? 0 : inst->GetSingleWordInOperand(1); in ContinueTargetFor()
727 : inst(def_inst), block_pos(the_block_pos), index(the_index) {} in DefInfo()
1540 if (const auto* inst = block.GetMergeInst()) { in RegisterMerges()
1542 switch (inst->opcode()) { in RegisterMerges()
1565 const uint32_t merge = inst in RegisterMerges()
2520 AddressOfIfNeeded( TypedExpression expr, const spvtools::opt::Instruction* inst) AddressOfIfNeeded() argument
2594 const auto* inst = def_use_mgr_->GetDef(id); MakeExpression() local
3473 EmitConstDefinition( const spvtools::opt::Instruction& inst, TypedExpression expr) EmitConstDefinition() argument
3497 EmitConstDefOrWriteToHoistedVar( const spvtools::opt::Instruction& inst, TypedExpression expr) EmitConstDefOrWriteToHoistedVar() argument
3503 WriteIfHoistedVar(const spvtools::opt::Instruction& inst, TypedExpression expr) WriteIfHoistedVar() argument
3520 EmitStatement(const spvtools::opt::Instruction& inst) EmitStatement() argument
3810 MakeOperand( const spvtools::opt::Instruction& inst, uint32_t operand_index) MakeOperand() argument
3835 MaybeEmitCombinatorialValue( const spvtools::opt::Instruction& inst) MaybeEmitCombinatorialValue() argument
4017 EmitGlslStd450ExtInst( const spvtools::opt::Instruction& inst) EmitGlslStd450ExtInst() argument
4202 MakeFMod( const spvtools::opt::Instruction& inst) MakeFMod() argument
4217 MakeAccessChain( const spvtools::opt::Instruction& inst) MakeAccessChain() argument
4438 MakeCompositeExtract( const spvtools::opt::Instruction& inst) MakeCompositeExtract() argument
4457 MakeCompositeValueDecomposition( const spvtools::opt::Instruction& inst, TypedExpression composite, uint32_t composite_type_id, int index_start) MakeCompositeValueDecomposition() argument
4602 MakeVectorShuffle( const spvtools::opt::Instruction& inst) MakeVectorShuffle() argument
4802 const auto& inst = id_def_info_pair.second->inst; FindValuesNeedingNamedOrHoistedDefinition() local
4977 MakeNumericConversion( const spvtools::opt::Instruction& inst) MakeNumericConversion() argument
5033 EmitFunctionCall(const spvtools::opt::Instruction& inst) EmitFunctionCall() argument
5070 EmitControlBarrier( const spvtools::opt::Instruction& inst) EmitControlBarrier() argument
5115 MakeIntrinsicCall( const spvtools::opt::Instruction& inst) MakeIntrinsicCall() argument
5143 MakeSimpleSelect( const spvtools::opt::Instruction& inst) MakeSimpleSelect() argument
5176 GetImage( const spvtools::opt::Instruction& inst) GetImage() argument
5212 GetImageExpression( const spvtools::opt::Instruction& inst) GetImageExpression() argument
5223 GetSamplerExpression( const spvtools::opt::Instruction& inst) GetSamplerExpression() argument
5239 EmitImageAccess(const spvtools::opt::Instruction& inst) EmitImageAccess() argument
5513 EmitImageQuery(const spvtools::opt::Instruction& inst) EmitImageQuery() argument
5591 MakeCoordinateOperandsForImageAccess( const spvtools::opt::Instruction& inst) MakeCoordinateOperandsForImageAccess() argument
5725 ConvertTexelForStorage( const spvtools::opt::Instruction& inst, TypedExpression texel, const Texture* texture_type) ConvertTexelForStorage() argument
5819 MakeArrayLength( const spvtools::opt::Instruction& inst) MakeArrayLength() argument
5860 MakeOuterProduct( const spvtools::opt::Instruction& inst) MakeOuterProduct() argument
5904 MakeVectorInsertDynamic( const spvtools::opt::Instruction& inst) MakeVectorInsertDynamic() argument
5972 MakeCompositeInsert( const spvtools::opt::Instruction& inst) MakeCompositeInsert() argument
[all...]
/kernel/linux/linux-6.6/drivers/crypto/intel/qat/qat_common/
H A Dqat_compression.c18 void qat_compression_put_instance(struct qat_compression_instance *inst) in qat_compression_put_instance() argument
20 atomic_dec(&inst->refctr); in qat_compression_put_instance()
21 adf_dev_put(inst->accel_dev); in qat_compression_put_instance()
26 struct qat_compression_instance *inst; in qat_compression_free_instances() local
31 inst = list_entry(list_ptr, in qat_compression_free_instances()
34 for (i = 0; i < atomic_read(&inst->refctr); i++) in qat_compression_free_instances()
35 qat_compression_put_instance(inst); in qat_compression_free_instances()
37 if (inst->dc_tx) in qat_compression_free_instances()
38 adf_remove_ring(inst->dc_tx); in qat_compression_free_instances()
40 if (inst in qat_compression_free_instances()
51 struct qat_compression_instance *inst = NULL; qat_compression_get_instance_node() local
116 struct qat_compression_instance *inst; qat_compression_create_instances() local
[all...]

Completed in 21 milliseconds

1...<<11121314151617181920>>...127