/third_party/pcre2/pcre2/src/sljit/ |
H A D | sljitNativePPC_32.c | 330 sljit_ins *inst = (sljit_ins *)addr; in sljit_set_jump_addr() local 333 SLJIT_UPDATE_WX_FLAGS(inst, inst + 2, 0); in sljit_set_jump_addr() 334 SLJIT_ASSERT((inst[0] & 0xfc1f0000) == ADDIS && (inst[1] & 0xfc000000) == ORI); in sljit_set_jump_addr() 335 inst[0] = (inst[0] & 0xffff0000) | ((new_target >> 16) & 0xffff); in sljit_set_jump_addr() 336 inst[1] = (inst[1] & 0xffff0000) | (new_target & 0xffff); in sljit_set_jump_addr() 337 SLJIT_UPDATE_WX_FLAGS(inst, ins in sljit_set_jump_addr() [all...] |
/third_party/spirv-tools/test/opt/ |
H A D | ir_context_test.cpp | 230 for (auto& inst : context->debugs2()) { in TEST_F() 231 EXPECT_EQ(inst.opcode(), spv::Op::OpNop); in TEST_F() 235 for (auto& inst : context->annotations()) { in TEST_F() 236 EXPECT_EQ(inst.opcode(), spv::Op::OpNop); in TEST_F() 277 auto inst = context->annotation_begin(); in TEST_F() local 278 EXPECT_EQ(inst->opcode(), spv::Op::OpDecorate); in TEST_F() 279 EXPECT_EQ(inst->GetSingleWordInOperand(0), 3); in TEST_F() 282 ++inst; in TEST_F() 283 EXPECT_EQ(inst->opcode(), spv::Op::OpDecorationGroup); in TEST_F() 284 EXPECT_EQ(inst in TEST_F() 341 auto inst = context->annotation_begin(); TEST_F() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
H A D | uvd_v4_2.c | 111 r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 124, &adev->uvd.inst->irq); in uvd_v4_2_sw_init() 119 ring = &adev->uvd.inst->ring; in uvd_v4_2_sw_init() 121 r = amdgpu_ring_init(adev, ring, 512, &adev->uvd.inst->irq, 0, in uvd_v4_2_sw_init() 159 struct amdgpu_ring *ring = &adev->uvd.inst->ring; in uvd_v4_2_hw_init() 254 struct amdgpu_ring *ring = &adev->uvd.inst->ring; in uvd_v4_2_start() 545 addr = (adev->uvd.inst->gpu_addr + AMDGPU_UVD_FIRMWARE_OFFSET) >> 3; in uvd_v4_2_mc_resume() 562 addr = (adev->uvd.inst->gpu_addr >> 28) & 0xF; in uvd_v4_2_mc_resume() 566 addr = (adev->uvd.inst->gpu_addr >> 32) & 0xFF; in uvd_v4_2_mc_resume() 673 amdgpu_fence_process(&adev->uvd.inst->ring); in uvd_v4_2_process_interrupt() 762 adev->uvd.inst in uvd_v4_2_set_ring_funcs() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmnv50.c | 155 nvkm_kmap(join->inst); in nv50_vmm_pgd_pde() 156 nvkm_wo64(join->inst, pdeo, data); in nv50_vmm_pgd_pde() 157 nvkm_done(join->inst); in nv50_vmm_pgd_pde() 325 nv50_vmm_part(struct nvkm_vmm *vmm, struct nvkm_memory *inst) in nv50_vmm_part() argument 330 if (join->inst == inst) { in nv50_vmm_part() 339 nv50_vmm_join(struct nvkm_vmm *vmm, struct nvkm_memory *inst) in nv50_vmm_join() argument 349 join->inst = inst; in nv50_vmm_join() 352 nvkm_kmap(join->inst); in nv50_vmm_join() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | chan.c | 286 nvkm_vmm_part(chan->vmm, chan->inst->memory); in nvkm_chan_del() 291 nvkm_gpuobj_del(&chan->inst); in nvkm_chan_del() 308 nvkm_chan_get_inst(struct nvkm_engine *engine, u64 inst, unsigned long *pirqflags) in nvkm_chan_get_inst() argument 318 chan = nvkm_runl_chan_get_inst(runl, inst, pirqflags); in nvkm_chan_get_inst() 358 (!func->inst->vmm != !vmm) || in nvkm_chan_new_() 365 runl->func->runqs, runq, func->inst->vmm, vmm, in nvkm_chan_new_() 410 ret = nvkm_gpuobj_new(device, func->inst->size, 0x1000, func->inst->zero, NULL, in nvkm_chan_new_() 411 &chan->inst); in nvkm_chan_new_() 413 RUNL_DEBUG(runl, "inst in nvkm_chan_new_() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmnv50.c | 155 nvkm_kmap(join->inst); in nv50_vmm_pgd_pde() 156 nvkm_wo64(join->inst, pdeo, data); in nv50_vmm_pgd_pde() 157 nvkm_done(join->inst); in nv50_vmm_pgd_pde() 327 nv50_vmm_part(struct nvkm_vmm *vmm, struct nvkm_memory *inst) in nv50_vmm_part() argument 332 if (join->inst == inst) { in nv50_vmm_part() 341 nv50_vmm_join(struct nvkm_vmm *vmm, struct nvkm_memory *inst) in nv50_vmm_join() argument 351 join->inst = inst; in nv50_vmm_join() 354 nvkm_kmap(join->inst); in nv50_vmm_join() [all...] |
H A D | gm20b.c | 50 gm20b_mmu_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, in gm20b_mmu_new() argument 54 return nvkm_mmu_new_(&gm20b_mmu_fixed, device, type, inst, pmmu); in gm20b_mmu_new() 55 return nvkm_mmu_new_(&gm20b_mmu, device, type, inst, pmmu); in gm20b_mmu_new()
|
H A D | gp100.c | 40 gp100_mmu_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, in gp100_mmu_new() argument 44 return gm200_mmu_new(device, type, inst, pmmu); in gp100_mmu_new() 45 return nvkm_mmu_new_(&gp100_mmu, device, type, inst, pmmu); in gp100_mmu_new()
|
H A D | gp10b.c | 40 gp10b_mmu_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, in gp10b_mmu_new() argument 44 return gm20b_mmu_new(device, type, inst, pmmu); in gp10b_mmu_new() 45 return nvkm_mmu_new_(&gp10b_mmu, device, type, inst, pmmu); in gp10b_mmu_new()
|
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/api/ |
H A D | vktApiPhysicalDeviceFormatPropertiesMaint5Tests.cpp | 147 const InstanceInterface& inst = m_context.getInstanceInterface(); in iterate() local 191 inst.getPhysicalDeviceFormatProperties(dev, format, &props1); in iterate() 197 inst.getPhysicalDeviceFormatProperties2(dev, format, &props2); in iterate() 203 res = inst.getPhysicalDeviceImageFormatProperties(dev, format, imageType, tiling, usage, createFlags, &imageProps1); in iterate() 210 res = inst.getPhysicalDeviceImageFormatProperties2(dev, &imageFormatInfo, &imageProps2); in iterate() 216 inst.getPhysicalDeviceSparseImageFormatProperties(dev, format, imageType, sampling, usage, tiling, &propsCount, nullptr); in iterate() 223 inst.getPhysicalDeviceSparseImageFormatProperties2(dev, &sparseFormatInfo, &propsCount, nullptr); in iterate() 239 const InstanceInterface& inst = m_context.getInstanceInterface(); in iterate() local 279 res = inst.getPhysicalDeviceImageFormatProperties(dev, format, imageType, tiling, usage, createFlags, &imageProps1); in iterate() 286 res = inst in iterate() [all...] |
/kernel/linux/linux-5.10/include/uapi/linux/ |
H A D | kcov.h | 56 static inline __u64 kcov_remote_handle(__u64 subsys, __u64 inst) in kcov_remote_handle() argument 58 if (subsys & ~KCOV_SUBSYSTEM_MASK || inst & ~KCOV_INSTANCE_MASK) in kcov_remote_handle() 60 return subsys | inst; in kcov_remote_handle()
|
/kernel/linux/linux-6.6/include/uapi/linux/ |
H A D | kcov.h | 56 static inline __u64 kcov_remote_handle(__u64 subsys, __u64 inst) in kcov_remote_handle() argument 58 if (subsys & ~KCOV_SUBSYSTEM_MASK || inst & ~KCOV_INSTANCE_MASK) in kcov_remote_handle() 60 return subsys | inst; in kcov_remote_handle()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
H A D | top.h | 14 int inst; member 31 int gk104_top_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_top **); 32 int ga100_top_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_top **);
|
/third_party/skia/third_party/externals/spirv-tools/source/fuzz/ |
H A D | fuzzer_pass_add_no_contraction_decorations.cpp | 34 for (auto& inst : block) { in Apply() 38 inst.opcode())) { in Apply() 45 inst.result_id()); in Apply()
|
H A D | fuzzer_pass_invert_comparison_operators.cpp | 33 GetIRContext()->module()->ForEachInst([this](const opt::Instruction* inst) { in Apply() 35 inst->opcode())) { in Apply() 45 inst->result_id(), GetFuzzerContext()->GetFreshId())); in Apply()
|
/third_party/skia/third_party/externals/spirv-tools/source/opt/ |
H A D | relax_float_ops_pass.h | 41 // Return true if |inst| can have the RelaxedPrecision decoration applied 43 bool IsRelaxable(Instruction* inst); 45 // Return true if |inst| returns scalar, vector or matrix type with base 47 bool IsFloat32(Instruction* inst); 52 // If |inst| is an instruction of float32-based type and is not decorated 54 bool ProcessInst(Instruction* inst);
|
/third_party/skia/third_party/externals/spirv-tools/source/reduce/ |
H A D | change_operand_to_undef_reduction_opportunity.h | 28 // Constructs the opportunity to replace operand |operand_index| of |inst| 31 opt::Instruction* inst, in ChangeOperandToUndefReductionOpportunity() 34 inst_(inst), in ChangeOperandToUndefReductionOpportunity() 36 original_id_(inst->GetOperand(operand_index).words[0]) {} in ChangeOperandToUndefReductionOpportunity() 30 ChangeOperandToUndefReductionOpportunity(opt::IRContext* context, opt::Instruction* inst, uint32_t operand_index) ChangeOperandToUndefReductionOpportunity() argument
|
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/source/fuzz/ |
H A D | fuzzer_pass_add_no_contraction_decorations.cpp | 34 for (auto& inst : block) { in Apply() 38 inst.opcode())) { in Apply() 45 inst.result_id()); in Apply()
|
H A D | fuzzer_pass_invert_comparison_operators.cpp | 33 GetIRContext()->module()->ForEachInst([this](const opt::Instruction* inst) { in Apply() 35 inst->opcode())) { in Apply() 45 inst->result_id(), GetFuzzerContext()->GetFreshId())); in Apply()
|
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/source/opt/ |
H A D | relax_float_ops_pass.h | 41 // Return true if |inst| can have the RelaxedPrecision decoration applied 43 bool IsRelaxable(Instruction* inst); 45 // Return true if |inst| returns scalar, vector or matrix type with base 47 bool IsFloat32(Instruction* inst); 52 // If |inst| is an instruction of float32-based type and is not decorated 54 bool ProcessInst(Instruction* inst);
|
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/source/reduce/ |
H A D | change_operand_to_undef_reduction_opportunity.h | 28 // Constructs the opportunity to replace operand |operand_index| of |inst| 31 opt::Instruction* inst, in ChangeOperandToUndefReductionOpportunity() 34 inst_(inst), in ChangeOperandToUndefReductionOpportunity() 36 original_id_(inst->GetOperand(operand_index).words[0]) {} in ChangeOperandToUndefReductionOpportunity() 30 ChangeOperandToUndefReductionOpportunity(opt::IRContext* context, opt::Instruction* inst, uint32_t operand_index) ChangeOperandToUndefReductionOpportunity() argument
|
/third_party/spirv-tools/source/fuzz/ |
H A D | fuzzer_pass_add_no_contraction_decorations.cpp | 34 for (auto& inst : block) { in Apply() 38 inst.opcode())) { in Apply() 45 inst.result_id()); in Apply()
|
H A D | fuzzer_pass_invert_comparison_operators.cpp | 33 GetIRContext()->module()->ForEachInst([this](const opt::Instruction* inst) { in Apply() 35 inst->opcode())) { in Apply() 45 inst->result_id(), GetFuzzerContext()->GetFreshId())); in Apply()
|
/third_party/spirv-tools/source/reduce/ |
H A D | change_operand_to_undef_reduction_opportunity.h | 28 // Constructs the opportunity to replace operand |operand_index| of |inst| 31 opt::Instruction* inst, in ChangeOperandToUndefReductionOpportunity() 34 inst_(inst), in ChangeOperandToUndefReductionOpportunity() 36 original_id_(inst->GetOperand(operand_index).words[0]) {} in ChangeOperandToUndefReductionOpportunity() 30 ChangeOperandToUndefReductionOpportunity(opt::IRContext* context, opt::Instruction* inst, uint32_t operand_index) ChangeOperandToUndefReductionOpportunity() argument
|
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_mqd_manager_v9.c | 63 struct mqd_update_info *minfo, uint32_t inst) in update_cu_mask() 72 minfo->cu_mask.ptr, minfo->cu_mask.count, se_mask, inst); in update_cu_mask() 96 pr_debug("inst: %u, update cu mask to %#x %#x %#x %#x\n", in update_cu_mask() 97 inst, m->compute_static_thread_mgmt_se0, in update_cu_mask() 555 int xcc_id, err, inst = 0; in hiq_load_mqd_kiq_v9_4_3() local 560 xcc_mqd = mqd + hiq_mqd_size * inst; in hiq_load_mqd_kiq_v9_4_3() 565 pr_debug("Failed to load HIQ MQD for XCC: %d\n", inst); in hiq_load_mqd_kiq_v9_4_3() 568 ++inst; in hiq_load_mqd_kiq_v9_4_3() 579 int xcc_id, err, inst = 0; in destroy_hiq_mqd_v9_4_3() local 585 m = get_mqd(mqd + hiq_mqd_size * inst); in destroy_hiq_mqd_v9_4_3() 62 update_cu_mask(struct mqd_manager *mm, void *mqd, struct mqd_update_info *minfo, uint32_t inst) update_cu_mask() argument 716 int xcc_id, err, inst = 0; destroy_mqd_v9_4_3() local 746 int xcc_id, err, inst = 0; load_mqd_v9_4_3() local [all...] |