/third_party/skia/third_party/externals/dawn/src/dawn_native/d3d12/ |
H A D | CommandBufferD3D12.cpp | 160 count, offsets.data(), 0); in RecordFirstIndexOffset() 254 // data pre-computed in the frontend. Also performs lazy initialization if required. 301 commandList->ResourceBarrier(barriers.size(), barriers.data()); in TransitionAndClearForSyncScope() 377 mDynamicOffsetCounts[index], mDynamicOffsets[index].data()); in Apply() 391 commandList->SetDescriptorHeaps(descriptorHeaps.size(), descriptorHeaps.data()); in SetID3D12DescriptorHeaps() 528 dynamicStorageBufferLengths.data(), firstRegisterOffset); in ApplyBindGroup() 532 dynamicStorageBufferLengths.data(), firstRegisterOffset); in ApplyBindGroup() 591 // mD3D12BufferViews is kept up to date with the most recent data passed in Apply() 607 // data in the middle of the range). 1034 uint8_t* data in RecordCommands() local [all...] |
/third_party/skia/third_party/externals/tint/src/ast/ |
H A D | intrinsic_texture_helper_test.cc | 119 std::ostream& operator<<(std::ostream& out, const TextureOverloadCase& data) { in operator <<() argument 120 out << "TextureOverloadCase " << static_cast<int>(data.overload) << "\n"; in operator <<() 121 out << data.description << "\n"; in operator <<() 122 out << "texture_kind: " << data.texture_kind << "\n"; in operator <<() 124 if (data.texture_kind != TextureKind::kStorage) { in operator <<() 125 out << data.sampler_kind; in operator <<() 130 out << "access: " << data.access << "\n"; in operator <<() 131 out << "image_format: " << data.image_format << "\n"; in operator <<() 132 out << "texture_dimension: " << data.texture_dimension << "\n"; in operator <<() 133 out << "texture_data_type: " << data in operator <<() [all...] |
/third_party/rust/crates/serde/serde_derive_internals/src/ |
H A D | attr.rs | 359 match item.data { in from_ast() 401 match &item.data { in from_ast() 423 match &item.data { in from_ast() 451 match item.data { in from_ast() 467 match &item.data { in from_ast() 489 match &item.data { in from_ast() 702 if let syn::Data::Enum(data) = &item.data { in decide_tag() 703 for variant in &data.variants { in decide_tag() 754 &item.data, in decide_identifier() [all...] |
/third_party/rust/crates/serde/serde_derive/src/internals/ |
H A D | attr.rs | 359 match item.data { in from_ast() 401 match &item.data { in from_ast() 423 match &item.data { in from_ast() 451 match item.data { in from_ast() 467 match &item.data { in from_ast() 489 match &item.data { in from_ast() 702 if let syn::Data::Enum(data) = &item.data { in decide_tag() 703 for variant in &data.variants { in decide_tag() 754 &item.data, in decide_identifier() [all...] |
/third_party/vk-gl-cts/framework/common/ |
H A D | tcuTexture.hpp | 180 //! get the swizzle used to expand texture data with a given channel order to RGBA form 183 //! get the swizzle used to narrow RGBA form data to native texture data with a given channel order 334 // Calculate pitches for pixel data with no padding. 342 * \brief Read-only pixel data access 344 * ConstPixelBufferAccess encapsulates pixel data pointer along with 356 ConstPixelBufferAccess (const TextureFormat& format, int width, int height, int depth, const void* data); 357 ConstPixelBufferAccess (const TextureFormat& format, const IVec3& size, const void* data); 358 ConstPixelBufferAccess (const TextureFormat& format, int width, int height, int depth, int rowPitch, int slicePitch, const void* data); 359 ConstPixelBufferAccess (const TextureFormat& format, const IVec3& size, const IVec3& pitch, const void* data); [all...] |
/third_party/ffmpeg/libavcodec/ |
H A D | dpcm.c | 27 * for more information on the specific data formats, visit: 224 bytestream2_init(&gb, avpkt->data, buf_size); in dpcm_decode_frame() 261 output_samples = (int16_t *)frame->data[0]; in dpcm_decode_frame() 340 uint8_t *output_samples_u8 = frame->data[0], in dpcm_decode_frame()
|
H A D | dvbsubenc.c | 362 uint32_t x= ((uint32_t*)h->rects[clut_id]->data[1])[i]; in dvbsub_encode() 467 h->rects[object_id]->data[0], in dvbsub_encode() 476 h->rects[object_id]->data[0] + h->rects[object_id]->w, in dvbsub_encode()
|
H A D | libjxlenc.c | 355 if (sd && sd->size && JxlEncoderSetICCProfile(ctx->encoder, sd->data, sd->size) != JXL_ENC_SUCCESS) in libjxl_encode_frame() 370 if (JxlEncoderAddImageFrame(ctx->options, &jxl_fmt, frame->data[0], jxl_fmt.align * info.ysize) != JXL_ENC_SUCCESS) { in libjxl_encode_frame() 388 /* all data passed has been encoded */ in libjxl_encode_frame() 416 memcpy(pkt->data, ctx->buffer, bytes_written); in libjxl_encode_frame()
|
H A D | hevc_sei.c | 141 sps = (HEVCSPS*)ps->sps_list[s->active_seq_parameter_set_id]->data; in decode_nal_sei_pic_timing() 195 bytestream2_get_bufferu(gb, buf_ref->data, size); in decode_nal_sei_user_data_unregistered() 196 buf_ref->data[size] = 0; in decode_nal_sei_user_data_unregistered()
|
H A D | fic.c | 189 uint8_t* dst = ctx->frame->data[p] + (y_off >> !!p) * stride; in fic_decode_slice() 246 dstptr[i] = ctx->final_frame->data[i] + in fic_draw_cursor() 274 const uint8_t *src = avpkt->data; in fic_decode_frame() 288 av_log(avctx, AV_LOG_ERROR, "Frame data is too small.\n"); in fic_decode_frame() 314 /* Skip cursor data. */ in fic_decode_frame() 328 "Cursor data too small. Skipping cursor.\n"); in fic_decode_frame() 357 /* First slice offset and remaining data. */ in fic_decode_frame() 362 av_log(avctx, AV_LOG_ERROR, "Not enough frame data to decode.\n"); in fic_decode_frame() 366 /* Allocate slice data. */ in fic_decode_frame() 370 av_log(avctx, AV_LOG_ERROR, "Could not allocate slice data in fic_decode_frame() [all...] |
H A D | ffwavesynth.c | 429 ts = AV_RL64(packet->data); in wavesynth_decode() 432 duration = AV_RL32(packet->data + 8); in wavesynth_decode() 439 pcm = (int16_t *)frame->data[0]; in wavesynth_decode()
|
H A D | aic.c | 334 Y = ctx->frame->data[0] + mb_x * 16 + y_pos * ystride; in aic_decode_slice() 336 C[i] = ctx->frame->data[i + 1] + mb_x * 8 in aic_decode_slice() 388 const uint8_t *buf = avpkt->data; in aic_decode_frame()
|
H A D | rscc.c | 28 * Lossless codec, data stored in tiles, with optional deflate compression. 31 * and it can be deflated or not. Similarly, pixel data comes after the header 162 bytestream2_init(gbc, avpkt->data, avpkt->size); in rscc_decode_frame() 224 * and point it to read the newly uncompressed data */ in rscc_decode_frame() 271 /* Extract how much pixel data the tiles contain */ in rscc_decode_frame() 322 /* Pointer to actual pixels, will be updated when data is consumed */ in rscc_decode_frame() 325 uint8_t *dst = ctx->reference->data[0] + ctx->reference->linesize[0] * in rscc_decode_frame() 351 memcpy(frame->data[1], ctx->palette, AVPALETTE_SIZE); in rscc_decode_frame()
|
H A D | smc.c | 28 * The SMC decoder outputs PAL8 colorspace data. 98 uint8_t * const pixels = s->frame->data[0]; in smc_decode_stream() 118 memcpy(s->frame->data[1], s->pal, AVPALETTE_SIZE); in smc_decode_stream() 434 const uint8_t *buf = avpkt->data; in smc_decode_frame()
|
/third_party/elfutils/src/ |
H A D | elfclassify.c | 364 Elf_Data *data = elf_getdata_rawchunk (elf, dyn_seg.p_offset, in run_classify() local 367 if (data != NULL) in run_classify() 371 GElf_Dyn *dyn = gelf_getdyn (data, dyn_idx, &dyn_storage); in run_classify()
|
/third_party/ffmpeg/libavfilter/ |
H A D | vf_avgblur_vulkan.c | 212 AVVkFrame *in = (AVVkFrame *)in_f->data[0]; in process_frames() 213 AVVkFrame *tmp = (AVVkFrame *)tmp_f->data[0]; in process_frames() 214 AVVkFrame *out = (AVVkFrame *)out_f->data[0]; in process_frames()
|
H A D | vf_deinterlace_vaapi.c | 219 input_surface = (VASurfaceID)(uintptr_t)input_frame->data[3]; in deint_vaapi_filter_frame() 222 ctx->frame_queue[current_frame_index - i - 1]->data[3]; in deint_vaapi_filter_frame() 225 ctx->frame_queue[current_frame_index + i + 1]->data[3]; in deint_vaapi_filter_frame()
|
H A D | vf_convolution_opencl.c | 215 for (p = 0; p < FF_ARRAY_ELEMS(output->data); p++) { in convolution_opencl_filter_frame() 216 src = (cl_mem) input->data[p]; in convolution_opencl_filter_frame() 217 dst = (cl_mem)output->data[p]; in convolution_opencl_filter_frame()
|
H A D | avf_abitscope.c | 167 uint8_t *dst = outpicref->data[0] + (b * h + j) * outpicref->linesize[0] + w * ch * 4; \ 195 uint8_t *dst = outpicref->data[0] + w * ch * 4 + wb * b * 4 + \ 224 memset(outpicref->data[0] + i * outpicref->linesize[0], 0, outlink->w * 4); in filter_frame()
|
H A D | vf_frei0r.c | 378 (const uint32_t *)in->data[0], in filter_frame() 379 (uint32_t *)out->data[0]); in filter_frame() 493 NULL, (uint32_t *)frame->data[0]); in source_request_frame()
|
H A D | vf_zoompan.c | 214 for (k = 0; in->data[k]; k++) in output_single_frame() 215 input[k] = in->data[k] + py[k] * in->linesize[k] + px[k]; in output_single_frame() 228 sws_scale(s->sws, (const uint8_t *const *)&input, in->linesize, 0, h, out->data, out->linesize); in output_single_frame()
|
H A D | vf_transpose.c | 291 dst = out->data[plane] + start * dstlinesize; in filter_slice() 292 src = in->data[plane]; in filter_slice() 301 dst = out->data[plane] + dstlinesize * (outh - start - 1); in filter_slice()
|
H A D | vf_unsharp_opencl.c | 244 for (p = 0; p < FF_ARRAY_ELEMS(output->data); p++) { in unsharp_opencl_filter_frame() 245 src = (cl_mem) input->data[p]; in unsharp_opencl_filter_frame() 246 dst = (cl_mem)output->data[p]; in unsharp_opencl_filter_frame()
|
/third_party/ffmpeg/libavformat/ |
H A D | amvenc.c | 192 memset(amv->apad->data, 0, amv->ablock_align); in amv_init() 193 AV_WL32(amv->apad->data + 4, amv->aframe_size); in amv_init() 315 avio_write(s->pb, pkt->data, pkt->size); in amv_write_packet_internal()
|
H A D | avformat.h | 37 * data in a specified container format. It also has an @ref lavf_io 38 * "I/O module" which supports a number of protocols for accessing the data (e.g. 95 * Demuxers read a media file and split it into chunks of data (@em packets). A 122 * for reading input data instead of lavf internal I/O layer. 142 * cannot know how to interpret raw video data otherwise. If the format turns 160 * Reading data from an opened AVFormatContext is done by repeatedly calling 162 * containing encoded data for one AVStream, identified by 165 * caller wishes to decode the data. 183 * Muxers take encoded data in the form of @ref AVPacket "AVPackets" and write 226 * The data i [all...] |