Home
last modified time | relevance | path

Searched refs:frame (Results 476 - 500 of 1514) sorted by relevance

1...<<11121314151617181920>>...61

/third_party/ffmpeg/doc/examples/
H A Ddecode_audio.c34 #include <libavutil/frame.h>
71 static void decode(AVCodecContext *dec_ctx, AVPacket *pkt, AVFrame *frame, in decode() argument
86 ret = avcodec_receive_frame(dec_ctx, frame); in decode()
99 for (i = 0; i < frame->nb_samples; i++) in decode()
101 fwrite(frame->data[ch] + data_size*i, 1, data_size, outfile); in decode()
174 fprintf(stderr, "Could not allocate audio frame\n"); in main()
/third_party/ffmpeg/libavcodec/
H A Dadxdec.c96 static int adx_decode_frame(AVCodecContext *avctx, AVFrame *frame, in adx_decode_frame() argument
160 frame->nb_samples = num_blocks * BLOCK_SAMPLES; in adx_decode_frame()
161 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in adx_decode_frame()
163 samples = (int16_t **)frame->extended_data; in adx_decode_frame()
180 frame->nb_samples = samples_offset; in adx_decode_frame()
H A Droqaudioenc.c123 const AVFrame *frame, int *got_packet_ptr) in roq_dpcm_encode_frame()
126 const int16_t *in = frame ? (const int16_t *)frame->data[0] : NULL; in roq_dpcm_encode_frame()
141 context->first_pts = frame->pts; in roq_dpcm_encode_frame()
179 avpkt->pts = context->input_frames <= 7 ? context->first_pts : frame->pts; in roq_dpcm_encode_frame()
122 roq_dpcm_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) roq_dpcm_encode_frame() argument
H A Dscreenpresso.c30 * rebuilt frame (not the reference), and since there is no coordinate system
78 /* Allocate current frame */ in screenpresso_init()
83 /* Allocate maximum size possible, a full RGBA frame */ in screenpresso_init()
105 static int screenpresso_decode_frame(AVCodecContext *avctx, AVFrame *frame, in screenpresso_decode_frame() argument
141 /* Inflate the frame after the 2 byte header */ in screenpresso_decode_frame()
163 /* Otherwise sum the delta on top of the current frame */ in screenpresso_decode_frame()
170 ret = av_frame_ref(frame, ctx->current); in screenpresso_decode_frame()
176 frame->pict_type = AV_PICTURE_TYPE_I; in screenpresso_decode_frame()
177 frame->key_frame = 1; in screenpresso_decode_frame()
179 frame in screenpresso_decode_frame()
[all...]
H A Drl2.c48 uint8_t *back_frame; ///< background frame
53 * Run Length Decode a single 320x200 frame
59 * @param video_base offset of the rle data inside the frame
73 /** copy start of the background frame */ in rl2_rle_decode()
84 /** decode the variable part of the frame */ in rl2_rle_decode()
116 /** copy the rest from the background frame */ in rl2_rle_decode()
166 /** decode background frame if present */ in rl2_decode_init()
181 static int rl2_decode_frame(AVCodecContext *avctx, AVFrame *frame, in rl2_decode_frame() argument
188 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in rl2_decode_frame()
192 rl2_rle_decode(s, buf, buf_size, frame in rl2_decode_frame()
[all...]
H A Dlibvpxenc.c54 * One encoded frame returned from the library.
59 int64_t pts; /**< time stamp to show frame
61 unsigned long duration; /**< duration to show frame
63 uint32_t flags; /**< flags for this frame */
97 #define VP8F_AUTO_ALT_REF 0x00000002 ///< Enable automatic alternate reference frame generation
136 * encounter a frame with ROI side data.
484 * 2-layers, 2-frame period. in set_temporal_layer_pattern()
505 * 3-layers structure with one reference frame. in set_temporal_layer_pattern()
508 * 3-layers, 4-frame period. in set_temporal_layer_pattern()
543 * 3-layers, 4-frame perio in set_temporal_layer_pattern()
1600 vpx_encode(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *frame, int *got_packet) vpx_encode() argument
[all...]
H A Dvt_internal.h36 CVImageBufferRef frame; member
52 int ff_videotoolbox_alloc_frame(AVCodecContext *avctx, AVFrame *frame);
66 int ff_videotoolbox_common_end_frame(AVCodecContext *avctx, AVFrame *frame);
H A Dintrax8.c295 w->frame->linesize[chroma > 0], in x8_setup_spatial_predictor()
623 w->frame->linesize[!!chroma]); in x8_decode_intra_mb()
648 w->frame->linesize[!!chroma]); in x8_decode_intra_mb()
652 w->frame->linesize[!!chroma]); in x8_decode_intra_mb()
656 w->frame->linesize[!!chroma], in x8_decode_intra_mb()
665 ptrdiff_t linesize = w->frame->linesize[!!chroma]; in x8_decode_intra_mb()
677 static void x8_init_block_index(IntraX8Context *w, AVFrame *frame) in x8_init_block_index() argument
681 const ptrdiff_t linesize = frame->linesize[0]; in x8_init_block_index()
682 const ptrdiff_t uvlinesize = frame->linesize[1]; in x8_init_block_index()
684 w->dest[0] = frame in x8_init_block_index()
[all...]
H A Dscpr.c498 static int decode_frame(AVCodecContext *avctx, AVFrame *frame, in decode_frame() argument
506 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in decode_frame()
521 frame->key_frame = 1; in decode_frame()
528 frame->key_frame = 1; in decode_frame()
532 frame->key_frame = 1; in decode_frame()
543 frame->key_frame = 1; in decode_frame()
562 frame->key_frame = 0; in decode_frame()
587 ret = av_frame_ref(frame, s->current_frame); in decode_frame()
591 uint8_t *dst = frame->data[0]; in decode_frame()
594 ret = av_frame_copy(frame, in decode_frame()
[all...]
H A Dvideotoolboxenc.c646 av_log(avctx, AV_LOG_ERROR, "Error encoding frame: %d\n", (int)status); in vtenc_output_callback()
1246 av_log(avctx, AV_LOG_ERROR, "Error setting 'max key-frame interval' property: %d\n", status); in vtenc_create_encoder()
1387 av_log(avctx, AV_LOG_ERROR, "Error setting 'allow frame reordering' property: %d\n", status); in vtenc_create_encoder()
2113 const AVFrame *frame, in get_cv_pixel_info()
2123 int av_format = frame->format; in get_cv_pixel_info()
2124 int av_color_range = frame->color_range; in get_cv_pixel_info()
2167 strides[p] = frame->linesize[p]; in get_cv_pixel_info()
2173 frame->data[i] + strides[i] * heights[i] != frame->data[i + 1]) { in get_cv_pixel_info()
2184 //Not used on OSX - frame i
2111 get_cv_pixel_info( AVCodecContext *avctx, const AVFrame *frame, int *color, int *plane_count, size_t *widths, size_t *heights, size_t *strides, size_t *contiguous_buf_size) get_cv_pixel_info() argument
2185 copy_avframe_to_pixel_buffer(AVCodecContext *avctx, const AVFrame *frame, CVPixelBufferRef cv_img, const size_t *plane_strides, const size_t *plane_rows) copy_avframe_to_pixel_buffer() argument
2277 create_cv_pixel_buffer(AVCodecContext *avctx, const AVFrame *frame, CVPixelBufferRef *cv_img) create_cv_pixel_buffer() argument
2374 create_encoder_dict_h264(const AVFrame *frame, CFDictionaryRef* dict_out) create_encoder_dict_h264() argument
2390 vtenc_send_frame(AVCodecContext *avctx, VTEncContext *vtctx, const AVFrame *frame) vtenc_send_frame() argument
2448 vtenc_frame( AVCodecContext *avctx, AVPacket *pkt, const AVFrame *frame, int *got_packet) vtenc_frame() argument
[all...]
H A Ddstdec.c243 static int decode_frame(AVCodecContext *avctx, AVFrame *frame, in decode_frame() argument
262 frame->nb_samples = samples_per_frame / 8; in decode_frame()
263 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in decode_frame()
265 dsd = frame->data[0]; in decode_frame()
266 pcm = (float *)frame->data[0]; in decode_frame()
275 memcpy(frame->data[0], avpkt->data + 1, FFMIN(avpkt->size - 1, frame->nb_samples * channels)); in decode_frame()
340 memset(dsd, 0, frame->nb_samples * 4 * channels); in decode_frame()
377 ff_dsd2pcm_translate(&s->dsdctx[i], frame->nb_samples, 0, in decode_frame()
378 frame in decode_frame()
[all...]
/third_party/ffmpeg/libavfilter/
H A Dasrc_hilbert.c49 { "nb_samples", "set the number of samples per requested frame", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, FLAGS },
50 { "n", "set the number of samples per requested frame", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, FLAGS },
131 AVFrame *frame; in activate() local
143 if (!(frame = ff_get_audio_buffer(outlink, nb_samples))) in activate()
146 memcpy(frame->data[0], s->taps + s->pts, nb_samples * sizeof(float)); in activate()
148 frame->pts = s->pts; in activate()
150 return ff_filter_frame(outlink, frame); in activate()
H A Dvf_separatefields.c54 static void extract_field(AVFrame *frame, int nb_planes, int type) in extract_field() argument
60 frame->data[i] = frame->data[i] + frame->linesize[i]; in extract_field()
61 frame->linesize[i] *= 2; in extract_field()
H A Daf_firequalizer.c134 { "fixed", "set fixed frame samples", OFFSET(fixed), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS },
816 static int filter_frame(AVFilterLink *inlink, AVFrame *frame) in filter_frame() argument
825 s->conv_idx + ch, (float *) frame->extended_data[ch], in filter_frame()
826 (float *) frame->extended_data[ch+1], frame->nb_samples); in filter_frame()
832 (float *) frame->extended_data[ch], frame->nb_samples); in filter_frame()
838 (float *) frame->extended_data[ch], frame->nb_samples); in filter_frame()
843 if (frame in filter_frame()
860 AVFrame *frame = ff_get_audio_buffer(outlink, FFMIN(s->remaining, s->frame_nsamples_max)); request_frame() local
[all...]
H A Dvf_chromakey_cuda.c69 AVFrame *frame; member
83 s->frame = av_frame_alloc(); in cudachromakey_init()
84 if (!s->frame) in cudachromakey_init()
109 av_frame_free(&s->frame); in cudachromakey_uninit()
134 av_frame_unref(s->frame); in init_hwframe_ctx()
135 ret = av_hwframe_get_buffer(out_ref, s->frame, 0); in init_hwframe_ctx()
394 ret = cudachromakey_process_internal(ctx, s->frame, src); in cudachromakey_process()
398 src = s->frame; in cudachromakey_process()
403 av_frame_move_ref(out, s->frame); in cudachromakey_process()
404 av_frame_move_ref(s->frame, in cudachromakey_process()
[all...]
H A Dvf_transpose_npp.c71 AVFrame *frame; member
89 s->stages[i].frame = av_frame_alloc(); in npptranspose_init()
90 if (!s->stages[i].frame) in npptranspose_init()
107 av_frame_free(&s->stages[i].frame); in npptranspose_uninit()
150 av_frame_unref(stage->frame); in init_stage()
151 ret = av_hwframe_get_buffer(out_ref, stage->frame, 0); in init_stage()
155 stage->frame->width = stage->planes_out[0].width; in init_stage()
156 stage->frame->height = stage->planes_out[0].height; in init_stage()
360 ret = npptranspose_process[i](ctx, &s->stages[i], s->stages[i].frame, src); in npptranspose_filter()
364 src = s->stages[i].frame; in npptranspose_filter()
[all...]
/third_party/ffmpeg/libavdevice/
H A Dalsa_enc.c121 AVFrame **frame, unsigned flags) in audio_write_frame()
131 pkt.data = (*frame)->data[0]; in audio_write_frame()
132 pkt.size = (*frame)->nb_samples * s->frame_size; in audio_write_frame()
133 pkt.dts = (*frame)->pkt_dts; in audio_write_frame()
134 pkt.duration = (*frame)->pkt_duration; in audio_write_frame()
120 audio_write_frame(AVFormatContext *s1, int stream_index, AVFrame **frame, unsigned flags) audio_write_frame() argument
/third_party/node/test/fixtures/wpt/encoding/resources/
H A Dencode-form-common.js72 var frame = document.createElement("iframe");
73 frame.id = frame.name = "frame-" + i;
74 document.body.appendChild(frame);
80 form.target = frame.id;
/third_party/skia/tools/
H A Dskottie2movie.cpp33 static void produce_frame(SkSurface* surf, skottie::Animation* anim, double frame) { in produce_frame() argument
34 anim->seekFrame(frame); in produce_frame()
130 const double frame = i * fps_scale; in main() local
132 SkDebugf("rendering frame %g\n", frame); in main()
135 produce_frame(surf.get(), animation.get(), frame); in main()
/third_party/nghttp2/src/
H A DHttpServer.cc1407 int on_header_callback2(nghttp2_session *session, const nghttp2_frame *frame, in on_header_callback2() argument
1417 verbose_on_header_callback(session, frame, namebuf.base, namebuf.len, in on_header_callback2()
1420 if (frame->hd.type != NGHTTP2_HEADERS || in on_header_callback2()
1421 frame->headers.cat != NGHTTP2_HCAT_REQUEST) { in on_header_callback2()
1424 auto stream = hd->get_stream(frame->hd.stream_id); in on_header_callback2()
1484 const nghttp2_frame *frame, void *user_data) { in on_begin_headers_callback()
1487 if (frame->hd.type != NGHTTP2_HEADERS || in on_begin_headers_callback()
1488 frame->headers.cat != NGHTTP2_HCAT_REQUEST) { in on_begin_headers_callback()
1492 auto stream = std::make_unique<Stream>(hd, frame->hd.stream_id); in on_begin_headers_callback()
1496 hd->add_stream(frame in on_begin_headers_callback()
1483 on_begin_headers_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_begin_headers_callback() argument
1503 hd_on_frame_recv_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) hd_on_frame_recv_callback() argument
1580 hd_on_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) hd_on_frame_send_callback() argument
1642 send_data_callback(nghttp2_session *session, nghttp2_frame *frame, const uint8_t *framehd, size_t length, nghttp2_data_source *source, void *user_data) send_data_callback() argument
1694 select_padding_callback(nghttp2_session *session, const nghttp2_frame *frame, size_t max_payload, void *user_data) select_padding_callback() argument
[all...]
H A Dnghttp.cc1787 const nghttp2_frame *frame, size_t max_payload, in select_padding_callback()
1789 return std::min(max_payload, frame->hd.length + config.padding); in select_padding_callback()
1855 const nghttp2_frame *frame, void *user_data) { in on_begin_headers_callback()
1857 switch (frame->hd.type) { in on_begin_headers_callback()
1860 nghttp2_session_get_stream_user_data(session, frame->hd.stream_id)); in on_begin_headers_callback()
1865 switch (frame->headers.cat) { in on_begin_headers_callback()
1877 auto stream_id = frame->push_promise.promised_stream_id; in on_begin_headers_callback()
1901 int on_header_callback(nghttp2_session *session, const nghttp2_frame *frame, in on_header_callback() argument
1906 verbose_on_header_callback(session, frame, name, namelen, value, valuelen, in on_header_callback()
1910 switch (frame in on_header_callback()
1786 select_padding_callback(nghttp2_session *session, const nghttp2_frame *frame, size_t max_payload, void *user_data) select_padding_callback() argument
1854 on_begin_headers_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_begin_headers_callback() argument
1970 on_frame_recv_callback2(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_frame_recv_callback2() argument
2097 before_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) before_frame_send_callback() argument
2113 on_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_frame_send_callback() argument
2140 on_frame_not_send_callback(nghttp2_session *session, const nghttp2_frame *frame, int lib_error_code, void *user_data) on_frame_not_send_callback() argument
[all...]
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/video/
H A DvktVideoDecodeTests.cpp404 DecodedFrame frame = initDecodeFrame(); in iterateSingleFrame() local
415 framesInQueue = m_decoder->GetVideoFrameBuffer()->DequeueDecodedPicture(&frame); in iterateSingleFrame()
430 framesInQueue = m_decoder->GetVideoFrameBuffer()->DequeueDecodedPicture(&frame); in iterateSingleFrame()
433 if (frame.pictureIndex >= 0) in iterateSingleFrame()
435 const VkExtent2D imageExtent = frame.pDecodedImage->getExtent(); in iterateSingleFrame()
436 const VkImage image = frame.pDecodedImage->getImage(); in iterateSingleFrame()
437 const VkFormat format = frame.pDecodedImage->getFormat(); in iterateSingleFrame()
438 const VkImageLayout layout = frame.decodedImageLayout; in iterateSingleFrame()
444 m_decoder->ReleaseDisplayedFrame(&frame); in iterateSingleFrame()
512 DecodedFrame& frame in iterateDoubleFrame() local
519 DecodedFrame& frame = frames[frameNdx]; iterateDoubleFrame() local
601 DecodedFrame& frame = frames[frameNdx]; iterateMultipleFrame() local
610 DecodedFrame& frame = frames[frameNdx]; iterateMultipleFrame() local
[all...]
/third_party/node/deps/v8/src/execution/
H A Disolate.cc592 it.frame()->Iterate(v); in Iterate()
599 it.frame()->Iterate(v); in Iterate()
694 reinterpret_cast<void*>(it.frame()->unchecked_code().ptr()); in StackTraceFailureMessage()
822 // Poison stack frames below the first strict mode frame.
834 // Determines whether the given stack frame should be displayed in a stack
841 // trace. This can be be the first frame (which will be a builtin-exit frame
842 // for the error constructor builtin) or every frame until encountering a
956 // Append async frame corresponding to the {generator_object}. in CaptureAsyncStackTrace()
1104 // find an async frame i in CaptureAsyncStackTrace()
1122 StackFrame* frame = it.frame(); VisitStack() local
1293 JavaScriptFrame* frame = it.frame(); GetAbstractPC() local
1929 StackFrame* frame = iter.frame(); ThrowInternal() local
2163 PredictException(JavaScriptFrame* frame) ThrowInternal() argument
2223 StackFrame* frame = iter.frame(); ThrowInternal() local
2727 StackFrame* frame = it.frame(); ThrowInternal() local
5434 IsBelowFrame(CommonFrame* frame) ThrowInternal() argument
[all...]
/third_party/python/Lib/idlelib/idle_test/
H A Dtest_help.py19 cls.frame = help.HelpFrame(root, helpfile)
23 del cls.frame
29 text = self.frame.text
/third_party/python/Lib/
H A Dpdb.py258 def sigint_handler(self, signum, frame):
263 self.set_trace(frame)
287 # The f_locals dictionary is updated from the actual frame
314 def user_call(self, frame, argument_list):
319 if self.stop_here(frame):
321 self.interaction(frame, None)
323 def user_line(self, frame):
326 if (self.mainpyfile != self.canonic(frame.f_code.co_filename)
327 or frame.f_lineno <= 0):
330 if self.bp_commands(frame)
[all...]

Completed in 28 milliseconds

1...<<11121314151617181920>>...61