/third_party/ffmpeg/doc/examples/ |
H A D | decode_audio.c | 34 #include <libavutil/frame.h> 71 static void decode(AVCodecContext *dec_ctx, AVPacket *pkt, AVFrame *frame, in decode() argument 86 ret = avcodec_receive_frame(dec_ctx, frame); in decode() 99 for (i = 0; i < frame->nb_samples; i++) in decode() 101 fwrite(frame->data[ch] + data_size*i, 1, data_size, outfile); in decode() 174 fprintf(stderr, "Could not allocate audio frame\n"); in main()
|
/third_party/ffmpeg/libavcodec/ |
H A D | adxdec.c | 96 static int adx_decode_frame(AVCodecContext *avctx, AVFrame *frame, in adx_decode_frame() argument 160 frame->nb_samples = num_blocks * BLOCK_SAMPLES; in adx_decode_frame() 161 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in adx_decode_frame() 163 samples = (int16_t **)frame->extended_data; in adx_decode_frame() 180 frame->nb_samples = samples_offset; in adx_decode_frame()
|
H A D | roqaudioenc.c | 123 const AVFrame *frame, int *got_packet_ptr) in roq_dpcm_encode_frame() 126 const int16_t *in = frame ? (const int16_t *)frame->data[0] : NULL; in roq_dpcm_encode_frame() 141 context->first_pts = frame->pts; in roq_dpcm_encode_frame() 179 avpkt->pts = context->input_frames <= 7 ? context->first_pts : frame->pts; in roq_dpcm_encode_frame() 122 roq_dpcm_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) roq_dpcm_encode_frame() argument
|
H A D | screenpresso.c | 30 * rebuilt frame (not the reference), and since there is no coordinate system 78 /* Allocate current frame */ in screenpresso_init() 83 /* Allocate maximum size possible, a full RGBA frame */ in screenpresso_init() 105 static int screenpresso_decode_frame(AVCodecContext *avctx, AVFrame *frame, in screenpresso_decode_frame() argument 141 /* Inflate the frame after the 2 byte header */ in screenpresso_decode_frame() 163 /* Otherwise sum the delta on top of the current frame */ in screenpresso_decode_frame() 170 ret = av_frame_ref(frame, ctx->current); in screenpresso_decode_frame() 176 frame->pict_type = AV_PICTURE_TYPE_I; in screenpresso_decode_frame() 177 frame->key_frame = 1; in screenpresso_decode_frame() 179 frame in screenpresso_decode_frame() [all...] |
H A D | rl2.c | 48 uint8_t *back_frame; ///< background frame 53 * Run Length Decode a single 320x200 frame 59 * @param video_base offset of the rle data inside the frame 73 /** copy start of the background frame */ in rl2_rle_decode() 84 /** decode the variable part of the frame */ in rl2_rle_decode() 116 /** copy the rest from the background frame */ in rl2_rle_decode() 166 /** decode background frame if present */ in rl2_decode_init() 181 static int rl2_decode_frame(AVCodecContext *avctx, AVFrame *frame, in rl2_decode_frame() argument 188 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in rl2_decode_frame() 192 rl2_rle_decode(s, buf, buf_size, frame in rl2_decode_frame() [all...] |
H A D | libvpxenc.c | 54 * One encoded frame returned from the library. 59 int64_t pts; /**< time stamp to show frame 61 unsigned long duration; /**< duration to show frame 63 uint32_t flags; /**< flags for this frame */ 97 #define VP8F_AUTO_ALT_REF 0x00000002 ///< Enable automatic alternate reference frame generation 136 * encounter a frame with ROI side data. 484 * 2-layers, 2-frame period. in set_temporal_layer_pattern() 505 * 3-layers structure with one reference frame. in set_temporal_layer_pattern() 508 * 3-layers, 4-frame period. in set_temporal_layer_pattern() 543 * 3-layers, 4-frame perio in set_temporal_layer_pattern() 1600 vpx_encode(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *frame, int *got_packet) vpx_encode() argument [all...] |
H A D | vt_internal.h | 36 CVImageBufferRef frame; member 52 int ff_videotoolbox_alloc_frame(AVCodecContext *avctx, AVFrame *frame); 66 int ff_videotoolbox_common_end_frame(AVCodecContext *avctx, AVFrame *frame);
|
H A D | intrax8.c | 295 w->frame->linesize[chroma > 0], in x8_setup_spatial_predictor() 623 w->frame->linesize[!!chroma]); in x8_decode_intra_mb() 648 w->frame->linesize[!!chroma]); in x8_decode_intra_mb() 652 w->frame->linesize[!!chroma]); in x8_decode_intra_mb() 656 w->frame->linesize[!!chroma], in x8_decode_intra_mb() 665 ptrdiff_t linesize = w->frame->linesize[!!chroma]; in x8_decode_intra_mb() 677 static void x8_init_block_index(IntraX8Context *w, AVFrame *frame) in x8_init_block_index() argument 681 const ptrdiff_t linesize = frame->linesize[0]; in x8_init_block_index() 682 const ptrdiff_t uvlinesize = frame->linesize[1]; in x8_init_block_index() 684 w->dest[0] = frame in x8_init_block_index() [all...] |
H A D | scpr.c | 498 static int decode_frame(AVCodecContext *avctx, AVFrame *frame, in decode_frame() argument 506 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in decode_frame() 521 frame->key_frame = 1; in decode_frame() 528 frame->key_frame = 1; in decode_frame() 532 frame->key_frame = 1; in decode_frame() 543 frame->key_frame = 1; in decode_frame() 562 frame->key_frame = 0; in decode_frame() 587 ret = av_frame_ref(frame, s->current_frame); in decode_frame() 591 uint8_t *dst = frame->data[0]; in decode_frame() 594 ret = av_frame_copy(frame, in decode_frame() [all...] |
H A D | videotoolboxenc.c | 646 av_log(avctx, AV_LOG_ERROR, "Error encoding frame: %d\n", (int)status); in vtenc_output_callback() 1246 av_log(avctx, AV_LOG_ERROR, "Error setting 'max key-frame interval' property: %d\n", status); in vtenc_create_encoder() 1387 av_log(avctx, AV_LOG_ERROR, "Error setting 'allow frame reordering' property: %d\n", status); in vtenc_create_encoder() 2113 const AVFrame *frame, in get_cv_pixel_info() 2123 int av_format = frame->format; in get_cv_pixel_info() 2124 int av_color_range = frame->color_range; in get_cv_pixel_info() 2167 strides[p] = frame->linesize[p]; in get_cv_pixel_info() 2173 frame->data[i] + strides[i] * heights[i] != frame->data[i + 1]) { in get_cv_pixel_info() 2184 //Not used on OSX - frame i 2111 get_cv_pixel_info( AVCodecContext *avctx, const AVFrame *frame, int *color, int *plane_count, size_t *widths, size_t *heights, size_t *strides, size_t *contiguous_buf_size) get_cv_pixel_info() argument 2185 copy_avframe_to_pixel_buffer(AVCodecContext *avctx, const AVFrame *frame, CVPixelBufferRef cv_img, const size_t *plane_strides, const size_t *plane_rows) copy_avframe_to_pixel_buffer() argument 2277 create_cv_pixel_buffer(AVCodecContext *avctx, const AVFrame *frame, CVPixelBufferRef *cv_img) create_cv_pixel_buffer() argument 2374 create_encoder_dict_h264(const AVFrame *frame, CFDictionaryRef* dict_out) create_encoder_dict_h264() argument 2390 vtenc_send_frame(AVCodecContext *avctx, VTEncContext *vtctx, const AVFrame *frame) vtenc_send_frame() argument 2448 vtenc_frame( AVCodecContext *avctx, AVPacket *pkt, const AVFrame *frame, int *got_packet) vtenc_frame() argument [all...] |
H A D | dstdec.c | 243 static int decode_frame(AVCodecContext *avctx, AVFrame *frame, in decode_frame() argument 262 frame->nb_samples = samples_per_frame / 8; in decode_frame() 263 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in decode_frame() 265 dsd = frame->data[0]; in decode_frame() 266 pcm = (float *)frame->data[0]; in decode_frame() 275 memcpy(frame->data[0], avpkt->data + 1, FFMIN(avpkt->size - 1, frame->nb_samples * channels)); in decode_frame() 340 memset(dsd, 0, frame->nb_samples * 4 * channels); in decode_frame() 377 ff_dsd2pcm_translate(&s->dsdctx[i], frame->nb_samples, 0, in decode_frame() 378 frame in decode_frame() [all...] |
/third_party/ffmpeg/libavfilter/ |
H A D | asrc_hilbert.c | 49 { "nb_samples", "set the number of samples per requested frame", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, FLAGS }, 50 { "n", "set the number of samples per requested frame", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, FLAGS }, 131 AVFrame *frame; in activate() local 143 if (!(frame = ff_get_audio_buffer(outlink, nb_samples))) in activate() 146 memcpy(frame->data[0], s->taps + s->pts, nb_samples * sizeof(float)); in activate() 148 frame->pts = s->pts; in activate() 150 return ff_filter_frame(outlink, frame); in activate()
|
H A D | vf_separatefields.c | 54 static void extract_field(AVFrame *frame, int nb_planes, int type) in extract_field() argument 60 frame->data[i] = frame->data[i] + frame->linesize[i]; in extract_field() 61 frame->linesize[i] *= 2; in extract_field()
|
H A D | af_firequalizer.c | 134 { "fixed", "set fixed frame samples", OFFSET(fixed), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS }, 816 static int filter_frame(AVFilterLink *inlink, AVFrame *frame) in filter_frame() argument 825 s->conv_idx + ch, (float *) frame->extended_data[ch], in filter_frame() 826 (float *) frame->extended_data[ch+1], frame->nb_samples); in filter_frame() 832 (float *) frame->extended_data[ch], frame->nb_samples); in filter_frame() 838 (float *) frame->extended_data[ch], frame->nb_samples); in filter_frame() 843 if (frame in filter_frame() 860 AVFrame *frame = ff_get_audio_buffer(outlink, FFMIN(s->remaining, s->frame_nsamples_max)); request_frame() local [all...] |
H A D | vf_chromakey_cuda.c | 69 AVFrame *frame; member 83 s->frame = av_frame_alloc(); in cudachromakey_init() 84 if (!s->frame) in cudachromakey_init() 109 av_frame_free(&s->frame); in cudachromakey_uninit() 134 av_frame_unref(s->frame); in init_hwframe_ctx() 135 ret = av_hwframe_get_buffer(out_ref, s->frame, 0); in init_hwframe_ctx() 394 ret = cudachromakey_process_internal(ctx, s->frame, src); in cudachromakey_process() 398 src = s->frame; in cudachromakey_process() 403 av_frame_move_ref(out, s->frame); in cudachromakey_process() 404 av_frame_move_ref(s->frame, in cudachromakey_process() [all...] |
H A D | vf_transpose_npp.c | 71 AVFrame *frame; member 89 s->stages[i].frame = av_frame_alloc(); in npptranspose_init() 90 if (!s->stages[i].frame) in npptranspose_init() 107 av_frame_free(&s->stages[i].frame); in npptranspose_uninit() 150 av_frame_unref(stage->frame); in init_stage() 151 ret = av_hwframe_get_buffer(out_ref, stage->frame, 0); in init_stage() 155 stage->frame->width = stage->planes_out[0].width; in init_stage() 156 stage->frame->height = stage->planes_out[0].height; in init_stage() 360 ret = npptranspose_process[i](ctx, &s->stages[i], s->stages[i].frame, src); in npptranspose_filter() 364 src = s->stages[i].frame; in npptranspose_filter() [all...] |
/third_party/ffmpeg/libavdevice/ |
H A D | alsa_enc.c | 121 AVFrame **frame, unsigned flags) in audio_write_frame() 131 pkt.data = (*frame)->data[0]; in audio_write_frame() 132 pkt.size = (*frame)->nb_samples * s->frame_size; in audio_write_frame() 133 pkt.dts = (*frame)->pkt_dts; in audio_write_frame() 134 pkt.duration = (*frame)->pkt_duration; in audio_write_frame() 120 audio_write_frame(AVFormatContext *s1, int stream_index, AVFrame **frame, unsigned flags) audio_write_frame() argument
|
/third_party/node/test/fixtures/wpt/encoding/resources/ |
H A D | encode-form-common.js | 72 var frame = document.createElement("iframe"); 73 frame.id = frame.name = "frame-" + i; 74 document.body.appendChild(frame); 80 form.target = frame.id;
|
/third_party/skia/tools/ |
H A D | skottie2movie.cpp | 33 static void produce_frame(SkSurface* surf, skottie::Animation* anim, double frame) { in produce_frame() argument 34 anim->seekFrame(frame); in produce_frame() 130 const double frame = i * fps_scale; in main() local 132 SkDebugf("rendering frame %g\n", frame); in main() 135 produce_frame(surf.get(), animation.get(), frame); in main()
|
/third_party/nghttp2/src/ |
H A D | HttpServer.cc | 1407 int on_header_callback2(nghttp2_session *session, const nghttp2_frame *frame, in on_header_callback2() argument 1417 verbose_on_header_callback(session, frame, namebuf.base, namebuf.len, in on_header_callback2() 1420 if (frame->hd.type != NGHTTP2_HEADERS || in on_header_callback2() 1421 frame->headers.cat != NGHTTP2_HCAT_REQUEST) { in on_header_callback2() 1424 auto stream = hd->get_stream(frame->hd.stream_id); in on_header_callback2() 1484 const nghttp2_frame *frame, void *user_data) { in on_begin_headers_callback() 1487 if (frame->hd.type != NGHTTP2_HEADERS || in on_begin_headers_callback() 1488 frame->headers.cat != NGHTTP2_HCAT_REQUEST) { in on_begin_headers_callback() 1492 auto stream = std::make_unique<Stream>(hd, frame->hd.stream_id); in on_begin_headers_callback() 1496 hd->add_stream(frame in on_begin_headers_callback() 1483 on_begin_headers_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_begin_headers_callback() argument 1503 hd_on_frame_recv_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) hd_on_frame_recv_callback() argument 1580 hd_on_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) hd_on_frame_send_callback() argument 1642 send_data_callback(nghttp2_session *session, nghttp2_frame *frame, const uint8_t *framehd, size_t length, nghttp2_data_source *source, void *user_data) send_data_callback() argument 1694 select_padding_callback(nghttp2_session *session, const nghttp2_frame *frame, size_t max_payload, void *user_data) select_padding_callback() argument [all...] |
H A D | nghttp.cc | 1787 const nghttp2_frame *frame, size_t max_payload, in select_padding_callback() 1789 return std::min(max_payload, frame->hd.length + config.padding); in select_padding_callback() 1855 const nghttp2_frame *frame, void *user_data) { in on_begin_headers_callback() 1857 switch (frame->hd.type) { in on_begin_headers_callback() 1860 nghttp2_session_get_stream_user_data(session, frame->hd.stream_id)); in on_begin_headers_callback() 1865 switch (frame->headers.cat) { in on_begin_headers_callback() 1877 auto stream_id = frame->push_promise.promised_stream_id; in on_begin_headers_callback() 1901 int on_header_callback(nghttp2_session *session, const nghttp2_frame *frame, in on_header_callback() argument 1906 verbose_on_header_callback(session, frame, name, namelen, value, valuelen, in on_header_callback() 1910 switch (frame in on_header_callback() 1786 select_padding_callback(nghttp2_session *session, const nghttp2_frame *frame, size_t max_payload, void *user_data) select_padding_callback() argument 1854 on_begin_headers_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_begin_headers_callback() argument 1970 on_frame_recv_callback2(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_frame_recv_callback2() argument 2097 before_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) before_frame_send_callback() argument 2113 on_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_frame_send_callback() argument 2140 on_frame_not_send_callback(nghttp2_session *session, const nghttp2_frame *frame, int lib_error_code, void *user_data) on_frame_not_send_callback() argument [all...] |
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/video/ |
H A D | vktVideoDecodeTests.cpp | 404 DecodedFrame frame = initDecodeFrame(); in iterateSingleFrame() local 415 framesInQueue = m_decoder->GetVideoFrameBuffer()->DequeueDecodedPicture(&frame); in iterateSingleFrame() 430 framesInQueue = m_decoder->GetVideoFrameBuffer()->DequeueDecodedPicture(&frame); in iterateSingleFrame() 433 if (frame.pictureIndex >= 0) in iterateSingleFrame() 435 const VkExtent2D imageExtent = frame.pDecodedImage->getExtent(); in iterateSingleFrame() 436 const VkImage image = frame.pDecodedImage->getImage(); in iterateSingleFrame() 437 const VkFormat format = frame.pDecodedImage->getFormat(); in iterateSingleFrame() 438 const VkImageLayout layout = frame.decodedImageLayout; in iterateSingleFrame() 444 m_decoder->ReleaseDisplayedFrame(&frame); in iterateSingleFrame() 512 DecodedFrame& frame in iterateDoubleFrame() local 519 DecodedFrame& frame = frames[frameNdx]; iterateDoubleFrame() local 601 DecodedFrame& frame = frames[frameNdx]; iterateMultipleFrame() local 610 DecodedFrame& frame = frames[frameNdx]; iterateMultipleFrame() local [all...] |
/third_party/node/deps/v8/src/execution/ |
H A D | isolate.cc | 592 it.frame()->Iterate(v); in Iterate() 599 it.frame()->Iterate(v); in Iterate() 694 reinterpret_cast<void*>(it.frame()->unchecked_code().ptr()); in StackTraceFailureMessage() 822 // Poison stack frames below the first strict mode frame. 834 // Determines whether the given stack frame should be displayed in a stack 841 // trace. This can be be the first frame (which will be a builtin-exit frame 842 // for the error constructor builtin) or every frame until encountering a 956 // Append async frame corresponding to the {generator_object}. in CaptureAsyncStackTrace() 1104 // find an async frame i in CaptureAsyncStackTrace() 1122 StackFrame* frame = it.frame(); VisitStack() local 1293 JavaScriptFrame* frame = it.frame(); GetAbstractPC() local 1929 StackFrame* frame = iter.frame(); ThrowInternal() local 2163 PredictException(JavaScriptFrame* frame) ThrowInternal() argument 2223 StackFrame* frame = iter.frame(); ThrowInternal() local 2727 StackFrame* frame = it.frame(); ThrowInternal() local 5434 IsBelowFrame(CommonFrame* frame) ThrowInternal() argument [all...] |
/third_party/python/Lib/idlelib/idle_test/ |
H A D | test_help.py | 19 cls.frame = help.HelpFrame(root, helpfile) 23 del cls.frame 29 text = self.frame.text
|
/third_party/python/Lib/ |
H A D | pdb.py | 258 def sigint_handler(self, signum, frame): 263 self.set_trace(frame) 287 # The f_locals dictionary is updated from the actual frame 314 def user_call(self, frame, argument_list): 319 if self.stop_here(frame): 321 self.interaction(frame, None) 323 def user_line(self, frame): 326 if (self.mainpyfile != self.canonic(frame.f_code.co_filename) 327 or frame.f_lineno <= 0): 330 if self.bp_commands(frame) [all...] |