/third_party/ffmpeg/libavcodec/ |
H A D | av1_frame_split_bsf.c | 24 * just one frame, plus any leading and trailing OBUs that may be present at 27 * Temporal Units already containing only one frame will be passed through 30 * one after the last successfully split frame. 91 AV1RawFrameHeader *frame = NULL; in av1_frame_split_filter() local 101 if (frame) { in av1_frame_split_filter() 103 "previous frame was expected.\n"); in av1_frame_split_filter() 107 frame = &obu->obu.frame.header; in av1_frame_split_filter() 113 // split here unless it's the last frame, in which case in av1_frame_split_filter() 120 if (frame) { in av1_frame_split_filter() [all...] |
H A D | vbndec.c | 57 AVFrame *frame, int *got_frame, in vbn_decode_frame() 149 ret = ff_get_buffer(avctx, frame, 0); in vbn_decode_frame() 153 frame->pict_type = AV_PICTURE_TYPE_I; in vbn_decode_frame() 154 frame->key_frame = 1; in vbn_decode_frame() 157 uint8_t *flipped = frame->data[0] + frame->linesize[0] * (frame->height - 1); in vbn_decode_frame() 158 av_image_copy_plane(flipped, -frame->linesize[0], image_buf ? image_buf : gb->buffer, linesize, linesize, frame->height); in vbn_decode_frame() 163 ctx->dec.frame_data.out = frame in vbn_decode_frame() 56 vbn_decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt) vbn_decode_frame() argument [all...] |
H A D | dxa.c | 211 static int decode_frame(AVCodecContext *avctx, AVFrame *frame, in decode_frame() argument 233 if ((ret = ff_get_buffer(avctx, frame, AV_GET_BUFFER_FLAG_REF)) < 0) in decode_frame() 235 memcpy(frame->data[1], c->pal, AVPALETTE_SIZE); in decode_frame() 236 frame->palette_has_changed = pc; in decode_frame() 238 outptr = frame->data[0]; in decode_frame() 241 stride = frame->linesize[0]; in decode_frame() 264 frame->key_frame = 0; in decode_frame() 265 frame->pict_type = AV_PICTURE_TYPE_P; in decode_frame() 267 memcpy(frame->data[0], c->prev->data[0], frame in decode_frame() [all...] |
H A D | escape124.c | 45 AVFrame *frame; member 66 s->frame = av_frame_alloc(); in escape124_decode_init() 67 if (!s->frame) in escape124_decode_init() 81 av_frame_free(&s->frame); in escape124_decode_close() 194 static int escape124_decode_frame(AVCodecContext *avctx, AVFrame *frame, in escape124_decode_frame() argument 228 // Leave last frame unchanged in escape124_decode_frame() 231 if (!s->frame->data[0]) in escape124_decode_frame() 234 av_log(avctx, AV_LOG_DEBUG, "Skipping frame\n"); in escape124_decode_frame() 237 if ((ret = av_frame_ref(frame, s->frame)) < in escape124_decode_frame() [all...] |
H A D | libopencore-amr.c | 99 static int amr_nb_decode_frame(AVCodecContext *avctx, AVFrame *frame, in amr_nb_decode_frame() argument 113 frame->nb_samples = 160; in amr_nb_decode_frame() 114 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in amr_nb_decode_frame() 121 av_log(avctx, AV_LOG_ERROR, "AMR frame too short (%d, should be %d)\n", in amr_nb_decode_frame() 129 Decoder_Interface_Decode(s->dec_state, buf, (short *)frame->data[0], 0); in amr_nb_decode_frame() 237 const AVFrame *frame, int *got_packet_ptr) in amr_nb_encode_frame() 242 const int16_t *samples = frame ? (const int16_t *)frame->data[0] : NULL; in amr_nb_encode_frame() 252 if (frame) { in amr_nb_encode_frame() 253 if (frame in amr_nb_encode_frame() 236 amr_nb_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) amr_nb_encode_frame() argument 332 amr_wb_decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame_ptr, AVPacket *avpkt) amr_wb_decode_frame() argument [all...] |
H A D | pafvideo.c | 58 uint8_t *frame[4]; member 74 av_freep(&c->frame[i]); in paf_video_close() 107 c->frame[i] = av_mallocz(c->frame_size); in paf_video_init() 108 if (!c->frame[i]) in paf_video_init() 158 *p = c->frame[page] + x * 2 + y * 2 * c->width; in set_src_position() 159 *pend = c->frame[page] + c->frame_size; in set_src_position() 184 dst = c->frame[page] + x + y * c->width; in decode_0() 185 dend = c->frame[page] + c->frame_size; in decode_0() 203 dst = c->frame[c->current_frame]; in decode_0() 204 dend = c->frame[ in decode_0() 276 int i, frame, ret; paf_video_decode() local [all...] |
H A D | ulti.c | 41 AVFrame *frame; member 59 s->frame = av_frame_alloc(); in ulti_decode_init() 60 if (!s->frame) in ulti_decode_init() 70 av_frame_free(&s->frame); in ulti_decode_end() 96 two 4-bit chroma samples) into standard YUV and put it into frame */ 97 static void ulti_convert_yuv(AVFrame *frame, int x, int y, in ulti_convert_yuv() argument 103 y_plane = frame->data[0] + x + y * frame->linesize[0]; in ulti_convert_yuv() 104 cr_plane = frame->data[1] + (x / 4) + (y / 4) * frame in ulti_convert_yuv() 121 ulti_pattern(AVFrame *frame, int x, int y, int f0, int f1, int Y0, int Y1, int chroma) ulti_pattern() argument 144 ulti_grad(AVFrame *frame, int x, int y, uint8_t *Y, int chroma, int angle) ulti_grad() argument [all...] |
/third_party/ffmpeg/libavfilter/ |
H A D | vf_pad.c | 231 AVFrame *frame; in get_video_buffer() local 237 frame = ff_get_video_buffer(inlink->dst->outputs[0], in get_video_buffer() 241 if (!frame) in get_video_buffer() 244 frame->width = w; in get_video_buffer() 245 frame->height = h; in get_video_buffer() 247 for (plane = 0; plane < 4 && frame->data[plane] && frame->linesize[plane]; plane++) { in get_video_buffer() 250 frame->data[plane] += (s->x >> hsub) * s->draw.pixelstep[plane] + in get_video_buffer() 251 (s->y >> vsub) * frame->linesize[plane]; in get_video_buffer() 254 return frame; in get_video_buffer() 258 buffer_needs_copy(PadContext *s, AVFrame *frame, AVBufferRef *buf) buffer_needs_copy() argument 310 frame_needs_copy(PadContext *s, AVFrame *frame) frame_needs_copy() argument [all...] |
H A D | lavfutils.c | 43 AVFrame *frame = NULL; in ff_load_image() local 87 if (!(frame = av_frame_alloc()) ) { in ff_load_image() 88 av_log(log_ctx, AV_LOG_ERROR, "Failed to alloc frame\n"); in ff_load_image() 95 av_log(log_ctx, AV_LOG_ERROR, "Failed to read frame from file\n"); in ff_load_image() 106 ret = avcodec_receive_frame(codec_ctx, frame); in ff_load_image() 112 *w = frame->width; in ff_load_image() 113 *h = frame->height; in ff_load_image() 114 *pix_fmt = frame->format; in ff_load_image() 120 av_image_copy(data, linesize, (const uint8_t **)frame->data, frame in ff_load_image() [all...] |
H A D | vf_exposure.c | 44 AVFrame *frame = arg; in exposure_slice() local 45 const int width = frame->width; in exposure_slice() 46 const int height = frame->height; in exposure_slice() 53 const int linesize = frame->linesize[p] / 4; in exposure_slice() 54 float *ptr = (float *)frame->data[p] + slice_start * linesize; in exposure_slice() 66 static int filter_frame(AVFilterLink *inlink, AVFrame *frame) in filter_frame() argument 72 ff_filter_execute(ctx, s->do_slice, frame, NULL, in filter_frame() 73 FFMIN(frame->height, ff_filter_get_nb_threads(ctx))); in filter_frame() 75 return ff_filter_frame(ctx->outputs[0], frame); in filter_frame()
|
H A D | vf_colorkey.c | 71 AVFrame *frame = arg; \ 72 const int slice_start = (frame->height * jobnr) / nb_jobs; \ 73 const int slice_end = (frame->height * (jobnr + 1)) / nb_jobs; \ 83 type *dst = (type *)(frame->data[0] + y * frame->linesize[0]);\ 85 for (int x = 0; x < frame->width; x++) { \ 106 AVFrame *frame = arg; \ 107 const int slice_start = (frame->height * jobnr) / nb_jobs; \ 108 const int slice_end = (frame->height * (jobnr + 1)) / nb_jobs; \ 120 type *dst = (type *)(frame 150 filter_frame(AVFilterLink *link, AVFrame *frame) filter_frame() argument [all...] |
H A D | vf_deinterlace_qsv.c | 94 av_frame_free(&cur->frame); in qsvdeint_uninit() 362 av_frame_free(&cur->frame); in clear_unused_frames() 371 QSVFrame *frame, **last; in get_free_frame() local 375 frame = s->work_frames; in get_free_frame() 377 while (frame) { in get_free_frame() 378 if (!frame->queued) { in get_free_frame() 379 *f = frame; in get_free_frame() 383 last = &frame->next; in get_free_frame() 384 frame = frame in get_free_frame() 396 submit_frame(AVFilterContext *ctx, AVFrame *frame, mfxFrameSurface1 **surface) submit_frame() argument [all...] |
H A D | af_aecho.c | 244 static int filter_frame(AVFilterLink *inlink, AVFrame *frame) in filter_frame() argument 250 if (av_frame_is_writable(frame)) { in filter_frame() 251 out_frame = frame; in filter_frame() 253 out_frame = ff_get_audio_buffer(ctx->outputs[0], frame->nb_samples); in filter_frame() 255 av_frame_free(&frame); in filter_frame() 258 av_frame_copy_props(out_frame, frame); in filter_frame() 261 s->echo_samples(s, s->delayptrs, frame->extended_data, out_frame->extended_data, in filter_frame() 262 frame->nb_samples, inlink->ch_layout.nb_channels); in filter_frame() 264 s->next_pts = frame->pts + av_rescale_q(frame in filter_frame() 277 AVFrame *frame = ff_get_audio_buffer(outlink, nb_samples); request_frame() local [all...] |
H A D | vf_despill.c | 46 AVFrame *frame = arg; in do_despill_slice() local 48 const int slice_start = (frame->height * jobnr) / nb_jobs; in do_despill_slice() 49 const int slice_end = (frame->height * (jobnr + 1)) / nb_jobs; in do_despill_slice() 60 uint8_t *dst = frame->data[0] + y * frame->linesize[0]; in do_despill_slice() 62 for (x = 0; x < frame->width; x++) { in do_despill_slice() 92 static int filter_frame(AVFilterLink *link, AVFrame *frame) in filter_frame() argument 97 if (ret = ff_filter_execute(ctx, do_despill_slice, frame, NULL, in filter_frame() 98 FFMIN(frame->height, ff_filter_get_nb_threads(ctx)))) in filter_frame() 101 return ff_filter_frame(ctx->outputs[0], frame); in filter_frame() [all...] |
H A D | af_bs2b.c | 120 static int filter_frame(AVFilterLink *inlink, AVFrame *frame) in filter_frame() argument 128 if (av_frame_is_writable(frame)) { in filter_frame() 129 out_frame = frame; in filter_frame() 131 out_frame = ff_get_audio_buffer(outlink, frame->nb_samples); in filter_frame() 133 av_frame_free(&frame); in filter_frame() 136 av_frame_copy(out_frame, frame); in filter_frame() 137 ret = av_frame_copy_props(out_frame, frame); in filter_frame() 140 av_frame_free(&frame); in filter_frame() 147 if (frame != out_frame) in filter_frame() 148 av_frame_free(&frame); in filter_frame() [all...] |
/third_party/python/Lib/ |
H A D | warnings.py | 74 for frame in tb: 76 % (frame.filename, frame.lineno)) 80 line = linecache.getline(frame.filename, frame.lineno) 272 def _is_internal_frame(frame): 273 """Signal whether the frame is an internal CPython implementation detail.""" 274 filename = frame.f_code.co_filename 278 def _next_external_frame(frame): 279 """Find the next frame tha [all...] |
/third_party/skia/experimental/ffmpeg/ |
H A D | SkVideoDecoder.cpp | 199 double SkVideoDecoder::computeTimeStamp(const AVFrame* frame) const { in computeTimeStamp() 201 return 1.0 * frame->pts * base.num / base.den; in computeTimeStamp() 204 sk_sp<SkImage> SkVideoDecoder::convertFrame(const AVFrame* frame) { in convertFrame() argument 205 auto yuv_space = get_yuvspace(frame->colorspace); in convertFrame() 208 fCSCache.update(frame->color_primaries, frame->color_trc); in convertFrame() 210 // Are these always true? If so, we don't need to check our "cache" on each frame... in convertFrame() 211 SkASSERT(fDecoderCtx->colorspace == frame->colorspace); in convertFrame() 212 SkASSERT(fDecoderCtx->color_primaries == frame->color_primaries); in convertFrame() 213 SkASSERT(fDecoderCtx->color_trc == frame in convertFrame() [all...] |
/third_party/node/deps/v8/third_party/wasm-api/example/ |
H A D | trap.c | 23 void print_frame(wasm_frame_t* frame) { in print_frame() argument 25 wasm_frame_instance(frame), in print_frame() 26 wasm_frame_module_offset(frame), in print_frame() 27 wasm_frame_func_index(frame), in print_frame() 28 wasm_frame_func_offset(frame) in print_frame() 121 own wasm_frame_t* frame = wasm_trap_origin(trap); in main() local 122 if (frame) { in main() 123 print_frame(frame); in main() 124 wasm_frame_delete(frame); in main()
|
/third_party/ltp/testcases/network/can/filter-tests/ |
H A D | can_rcv_own_msgs.c | 23 struct can_frame frame; in test_sockets() local 26 frame.can_id = can_id; in test_sockets() 27 frame.can_dlc = 0; in test_sockets() 28 SAFE_WRITE(SAFE_WRITE_ALL, s, &frame, sizeof(frame)); in test_sockets() 45 SAFE_READ(1, s, &frame, sizeof(struct can_frame)); in test_sockets() 47 if (frame.can_id != can_id) in test_sockets() 56 SAFE_READ(1, t, &frame, sizeof(struct can_frame)); in test_sockets() 58 if (frame.can_id != can_id) in test_sockets() 115 tst_res(TINFO, "Starting PF_CAN frame flo in run() [all...] |
/foundation/CastEngine/castengine_wifi_display/tests/unittest/impl/screen_capture/ |
H A D | wfd_screen_capture_test.cpp | 53 Frame::Ptr frame = FrameImpl::Create(); in HWTEST_F() local 54 consumer_->HandleSpsFrame(consumer_->dispatcher_, frame); in HWTEST_F() 62 Frame::Ptr frame = FrameImpl::Create(); in HWTEST_F() local 63 consumer_->HandleSpsFrame(consumer_->dispatcher_, frame); in HWTEST_F() 71 Frame::Ptr frame = FrameImpl::Create(); in HWTEST_F() local 72 consumer_->HandlePpsFrame(consumer_->dispatcher_, frame); in HWTEST_F() 80 Frame::Ptr frame = FrameImpl::Create(); in HWTEST_F() local 81 consumer_->HandlePpsFrame(consumer_->dispatcher_, frame); in HWTEST_F() 95 Frame::Ptr frame = FrameImpl::Create(); in HWTEST_F() local 96 consumer_->OnFrame(frame, FRAME_TYP in HWTEST_F() 103 Frame::Ptr frame = FrameImpl::Create(); HWTEST_F() local 111 Frame::Ptr frame = FrameImpl::Create(); HWTEST_F() local [all...] |
/third_party/nghttp2/src/ |
H A D | h2load_http2_session.cc | 45 int on_header_callback(nghttp2_session *session, const nghttp2_frame *frame, in on_header_callback() argument 50 if (frame->hd.type != NGHTTP2_HEADERS || in on_header_callback() 51 frame->headers.cat != NGHTTP2_HCAT_RESPONSE) { in on_header_callback() 54 client->on_header(frame->hd.stream_id, name, namelen, value, valuelen); in on_header_callback() 58 std::cout << "[stream_id=" << frame->hd.stream_id << "] "; in on_header_callback() 70 int on_frame_recv_callback(nghttp2_session *session, const nghttp2_frame *frame, in on_frame_recv_callback() argument 73 if (frame->hd.type != NGHTTP2_HEADERS || in on_frame_recv_callback() 74 frame->headers.cat != NGHTTP2_HCAT_RESPONSE) { in on_frame_recv_callback() 78 frame->hd.length - frame in on_frame_recv_callback() 109 before_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) before_frame_send_callback() argument [all...] |
/third_party/python/Lib/idlelib/idle_test/ |
H A D | test_textview.py | 94 frame = tv.ScrollableTextFrame(self.root, wrap=wrap, **kwargs) 96 frame.update_idletasks() 97 frame.destroy() 99 return frame 102 frame = self.make_frame() 103 frame.text.insert('1.0', 'test text') 104 self.assertEqual(frame.text.get('1.0', '1.end'), 'test text') 112 frame = self.make_frame(wrap=NONE) 113 self.assertEqual(frame.text.cget('wrap'), NONE) 114 self.assertIsNotNone(frame [all...] |
/third_party/node/deps/v8/src/objects/ |
H A D | call-site-info.cc | 606 void AppendFileLocation(Isolate* isolate, Handle<CallSiteInfo> frame, in AppendFileLocation() argument 608 Handle<Object> script_name_or_source_url(frame->GetScriptNameOrSourceURL(), in AppendFileLocation() 610 if (!script_name_or_source_url->IsString() && frame->IsEval()) { in AppendFileLocation() 612 Handle<String>::cast(CallSiteInfo::GetEvalOrigin(frame))); in AppendFileLocation() 626 int line_number = CallSiteInfo::GetLineNumber(frame); in AppendFileLocation() 631 int column_number = CallSiteInfo::GetColumnNumber(frame); in AppendFileLocation() 670 void AppendMethodCall(Isolate* isolate, Handle<CallSiteInfo> frame, in AppendMethodCall() argument 672 Handle<Object> type_name = CallSiteInfo::GetTypeName(frame); in AppendMethodCall() 673 Handle<Object> method_name = CallSiteInfo::GetMethodName(frame); in AppendMethodCall() 674 Handle<Object> function_name = CallSiteInfo::GetFunctionName(frame); in AppendMethodCall() 717 SerializeJSStackFrame(Isolate* isolate, Handle<CallSiteInfo> frame, IncrementalStringBuilder* builder) SerializeJSStackFrame() argument 753 SerializeWasmStackFrame(Isolate* isolate, Handle<CallSiteInfo> frame, IncrementalStringBuilder* builder) SerializeWasmStackFrame() argument 795 SerializeCallSiteInfo(Isolate* isolate, Handle<CallSiteInfo> frame, IncrementalStringBuilder* builder) SerializeCallSiteInfo() argument 806 SerializeCallSiteInfo(Isolate* isolate, Handle<CallSiteInfo> frame) SerializeCallSiteInfo() argument [all...] |
/foundation/multimedia/av_codec/services/media_engine/modules/sink/ |
H A D | calc_max_amplitude.h | 23 float CalculateMaxAmplitudeForPCM8Bit(int8_t *frame, uint64_t nSamples);
24 float CalculateMaxAmplitudeForPCM16Bit(int16_t *frame, uint64_t nSamples);
25 float CalculateMaxAmplitudeForPCM24Bit(char *frame, uint64_t nSamples);
26 float CalculateMaxAmplitudeForPCM32Bit(int32_t *frame, uint64_t nSamples);
27 float UpdateMaxAmplitude(char *frame, uint64_t replyBytes, int32_t adapterFormat);
|
/third_party/ffmpeg/doc/examples/ |
H A D | qsvdec.c | 63 AVFrame *frame, AVFrame *sw_frame, in decode_packet() 77 ret = avcodec_receive_frame(decoder_ctx, frame); in decode_packet() 85 /* A real program would do something useful with the decoded frame here. in decode_packet() 88 ret = av_hwframe_transfer_data(sw_frame, frame, 0); in decode_packet() 100 av_frame_unref(frame); in decode_packet() 117 AVFrame *frame = NULL, *sw_frame = NULL; in main() local 201 frame = av_frame_alloc(); in main() 204 if (!frame || !sw_frame || !pkt) { in main() 216 ret = decode_packet(decoder_ctx, frame, sw_frame, pkt, output_ctx); in main() 222 ret = decode_packet(decoder_ctx, frame, sw_fram in main() 62 decode_packet(AVCodecContext *decoder_ctx, AVFrame *frame, AVFrame *sw_frame, AVPacket *pkt, AVIOContext *output_ctx) decode_packet() argument [all...] |