/third_party/ffmpeg/libavcodec/ |
H A D | sgienc.c | 94 const AVFrame *frame, int *got_packet) in encode_frame() 97 const AVFrame * const p = frame; in encode_frame() 93 encode_frame(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *frame, int *got_packet) encode_frame() argument
|
H A D | qsvenc_hevc.c | 209 const AVFrame *frame, int *got_packet) in qsv_enc_frame() 213 return ff_qsv_encode(avctx, &q->qsv, pkt, frame, got_packet); in qsv_enc_frame() 239 { "begin_only", "Output an IDR-frame only at the beginning of the stream", 0, AV_OPT_TYPE_CONST, { .i64 = -1 }, 0, 0, VE, "idr_interval" }, 259 { "gpb", "1: GPB (generalized P/B frame); 0: regular P frame", OFFSET(qsv.gpb), AV_OPT_TYPE_BOOL, { .i64 = 1 }, 0, 1, VE}, 208 qsv_enc_frame(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *frame, int *got_packet) qsv_enc_frame() argument
|
H A D | wcmv.c | 44 static int decode_frame(AVCodecContext *avctx, AVFrame *frame, in decode_frame() argument 203 if ((ret = av_frame_ref(frame, s->prev_frame)) < 0) in decode_frame()
|
H A D | atrac9dec.c | 646 ATRAC9BlockData *b, AVFrame *frame, in atrac9_decode_block() 779 float *dst = (float *)(frame->extended_data[dst_idx] + offset); in atrac9_decode_block() 790 static int atrac9_decode_frame(AVCodecContext *avctx, AVFrame *frame, in atrac9_decode_frame() argument 798 frame->nb_samples = (1 << s->frame_log2) * frames; in atrac9_decode_frame() 799 ret = ff_get_buffer(avctx, frame, 0); in atrac9_decode_frame() 807 ret = atrac9_decode_block(s, &gb, &s->block[j], frame, i, j); in atrac9_decode_frame() 950 /* Average frame size in bytes */ in atrac9_decode_init() 645 atrac9_decode_block(ATRAC9Context *s, GetBitContext *gb, ATRAC9BlockData *b, AVFrame *frame, int frame_idx, int block_idx) atrac9_decode_block() argument
|
H A D | dolby_e.c | 1058 static int filter_frame(DBEDecodeContext *s, AVFrame *frame) in filter_frame() argument 1073 frame->nb_samples = FRAME_SAMPLES; in filter_frame() 1074 if ((ret = ff_get_buffer(s->avctx, frame, 0)) < 0) in filter_frame() 1078 float *output = (float *)frame->extended_data[reorder[ch]]; in filter_frame() 1087 static int dolby_e_decode_frame(AVCodecContext *avctx, AVFrame *frame, in dolby_e_decode_frame() argument 1138 if ((ret = filter_frame(s1, frame)) < 0) in dolby_e_decode_frame()
|
H A D | imc.c | 884 /* Check the frame header */ in imc_decode_block() 887 av_log(avctx, AV_LOG_ERROR, "frame header check failed!\n"); in imc_decode_block() 1021 static int imc_decode_frame(AVCodecContext *avctx, AVFrame *frame, in imc_decode_frame() argument 1035 av_log(avctx, AV_LOG_ERROR, "frame too small!\n"); in imc_decode_frame() 1040 frame->nb_samples = COEFFS; in imc_decode_frame() 1041 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in imc_decode_frame() 1045 q->out_samples = (float *)frame->extended_data[i]; in imc_decode_frame() 1058 q->butterflies_float((float *)frame->extended_data[0], in imc_decode_frame() 1059 (float *)frame->extended_data[1], COEFFS); in imc_decode_frame()
|
H A D | sonic.c | 698 av_log(avctx, AV_LOG_INFO, "Sonic: ver: %d.%d ls: %d dr: %d taps: %d block: %d frame: %d downsamp: %d\n", in sonic_encode_init() 721 const AVFrame *frame, int *got_packet_ptr) in sonic_encode_frame() 727 const short *samples = (const int16_t*)frame->data[0]; in sonic_encode_frame() 932 "number of taps times channels (%d * %d) larger than frame size %d\n", in sonic_decode_init() 937 av_log(avctx, AV_LOG_INFO, "Sonic: ver: %d.%d ls: %d dr: %d taps: %d block: %d frame: %d downsamp: %d\n", in sonic_decode_init() 983 static int sonic_decode_frame(AVCodecContext *avctx, AVFrame *frame, in sonic_decode_frame() argument 996 frame->nb_samples = s->frame_size / avctx->ch_layout.nb_channels; in sonic_decode_frame() 997 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in sonic_decode_frame() 999 samples = (int16_t *)frame->data[0]; in sonic_decode_frame() 720 sonic_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) sonic_encode_frame() argument
|
H A D | vc2enc.c | 898 static int encode_frame(VC2EncContext *s, AVPacket *avpkt, const AVFrame *frame, in encode_frame() argument 909 s->transform_args[i].idata = frame->data[i]; in encode_frame() 910 s->transform_args[i].istride = frame->linesize[i]; in encode_frame() 952 const AVFrame *frame, int *got_packet) in vc2_encode_frame() 989 ret = encode_frame(s, avpkt, frame, aux_data, header_size, s->interlaced); in vc2_encode_frame() 993 ret = encode_frame(s, avpkt, frame, aux_data, header_size, 2); in vc2_encode_frame() 951 vc2_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet) vc2_encode_frame() argument
|
/third_party/ffmpeg/libavfilter/ |
H A D | vf_dnn_classify.c | 64 static int dnn_classify_post_proc(AVFrame *frame, DNNData *output, uint32_t bbox_index, AVFilterContext *filter_ctx) in dnn_classify_post_proc() argument 79 sd = av_frame_get_side_data(frame, AV_FRAME_DATA_DETECTION_BBOXES); in dnn_classify_post_proc() 275 // if frame got, schedule to next filter in dnn_classify_activate()
|
H A D | fflcms2.c | 173 int ff_icc_profile_attach(FFIccContext *s, cmsHPROFILE profile, AVFrame *frame) in ff_icc_profile_attach() argument 190 if (!av_frame_new_side_data_from_buf(frame, AV_FRAME_DATA_ICC_PROFILE, buf)) { in ff_icc_profile_attach()
|
H A D | vf_kerndeint.c | 40 int frame; ///< frame count, starting from 0 member 127 int x, y, plane, val, hi, lo, g, h, n = kerndeint->frame++; in filter_frame()
|
H A D | af_loudnorm.c | 651 AVFrame *frame; in request_frame() local 656 frame = ff_get_audio_buffer(outlink, nb_samples); in request_frame() 657 if (!frame) in request_frame() 659 frame->nb_samples = nb_samples; in request_frame() 662 src = (double *)frame->data[0]; in request_frame() 679 ret = filter_frame(inlink, frame); in request_frame()
|
/third_party/ffmpeg/libavutil/ |
H A D | Makefile | 35 frame.h \ 135 frame.o \
|
/third_party/backends/backend/ |
H A D | pieusb_scancmd.h | 242 SANE_Int index; /* scan frame index (0-7) */ 333 void sanei_pieusb_cmd_get_scan_frame(SANE_Int device_number, SANE_Int index, struct Pieusb_Scan_Frame* frame, struct Pieusb_Command_Status *status); 334 void sanei_pieusb_cmd_set_scan_frame(SANE_Int device_number, SANE_Int index, struct Pieusb_Scan_Frame* frame, struct Pieusb_Command_Status *status);
|
/third_party/libunwind/libunwind/include/tdep-x86_64/ |
H A D | libunwind_i.h | 45 UNW_X86_64_FRAME_ALIGNED = -3, /* frame stack pointer aligned */ 47 UNW_X86_64_FRAME_SIGRETURN = -1, /* special sigreturn frame */ 57 int64_t last_frame : 1; /* non-zero if last frame in chain */ 86 X86_64_SCF_NONE, /* no signal frame encountered */ 88 X86_64_SCF_FREEBSD_SIGFRAME, /* FreeBSD signal frame */ 90 X86_64_SCF_SOLARIS_SIGFRAME, /* illumos/Solaris signal frame */ 244 # define tdep_reuse_frame(c,frame) do {} while(0) 290 int frame);
|
/third_party/node/tools/inspector_protocol/jinja2/ |
H A D | debug.py | 37 """Proxies a traceback frame.""" 66 def make_frame_proxy(frame): 67 proxy = TracebackFrameProxy(frame) 118 """Exception info tuple with a proxy around the frame objects.""" 125 # the frame will be an actual traceback (or transparent proxy) if 175 # save a reference to the next frame if we override the current 250 # frame for the ProcessedTraceback 294 # return without this frame
|
/third_party/nghttp2/src/ |
H A D | shrpx_http.cc | 171 const nghttp2_frame *frame, size_t max_payload, in select_padding_callback() 173 return std::min(max_payload, frame->hd.length + get_config()->padding); in select_padding_callback() 170 select_padding_callback(nghttp2_session *session, const nghttp2_frame *frame, size_t max_payload, void *user_data) select_padding_callback() argument
|
/third_party/libinput/tools/ |
H A D | libinput-replay.py | 171 def collect_events(frame): 174 for (sec, usec, evtype, evcode, value) in frame:
|
/third_party/skia/third_party/externals/dawn/examples/ |
H A D | CubeReflection.cpp | 253 void frame() { in frame() function 306 frame(); in main()
|
/third_party/python/Lib/unittest/test/ |
H A D | test_break.py | 162 def new_handler(frame, signum): 163 handler(frame, signum)
|
/third_party/skia/experimental/ffmpeg/ |
H A D | SkVideoEncoder.cpp | 264 /* make sure the frame data is writable */ in addFrame() 279 bool SkVideoEncoder::sendFrame(AVFrame* frame) { in sendFrame() argument 280 if (check_err(avcodec_send_frame(fEncoderCtx, frame))) { in sendFrame()
|
/third_party/mesa3d/src/gallium/drivers/r600/ |
H A D | radeon_uvd.c | 247 // aligned size of a single frame in calc_dpb_size() 541 /* extract the frame number from a referenced video buffer */ 546 uintptr_t frame; in get_ref_pic_idx() local 552 /* get the frame number from the associated data */ in get_ref_pic_idx() 553 frame = (uintptr_t)vl_video_buffer_get_associated_data(ref, &dec->base); in get_ref_pic_idx() 555 /* limit the frame number to a valid range */ in get_ref_pic_idx() 556 return MAX2(MIN2(frame, max), min); in get_ref_pic_idx() 827 * start decoding of a new frame 834 uintptr_t frame; in ruvd_begin_frame() local 838 frame in ruvd_begin_frame() [all...] |
/third_party/mesa3d/src/vulkan/wsi/ |
H A D | wsi_common_wayland.c | 930 struct wl_callback * frame; member 1037 chain->frame = NULL; in frame_handle_done() 1085 chain->frame = wl_surface_frame(chain->surface); in wsi_wl_swapchain_queue_present() 1086 wl_callback_add_listener(chain->frame, &frame_listener, chain); in wsi_wl_swapchain_queue_present() 1235 if (chain->frame) in wsi_wl_swapchain_chain_free() 1236 wl_callback_destroy(chain->frame); in wsi_wl_swapchain_chain_free()
|
/third_party/libdrm/tests/amdgpu/ |
H A D | vce_tests.c | 36 #include "frame.h" 515 memcpy(enc.vbuf.ptr, (frame + i * enc.width), enc.width); in amdgpu_cs_vce_encode() 519 memcpy(enc.vbuf.ptr, ((frame + enc.height * enc.width) + i * enc.width), enc.width); in amdgpu_cs_vce_encode() 697 memcpy(enc.vbuf.ptr, (frame + i * enc.width), enc.width); in amdgpu_cs_vce_encode_mv() 701 memcpy(enc.vbuf.ptr, ((frame + enc.height * enc.width) + i * enc.width), enc.width); in amdgpu_cs_vce_encode_mv() 713 memcpy(enc.mvrefbuf.ptr, (frame + (enc.height - i -1) * enc.width), enc.width); in amdgpu_cs_vce_encode_mv() 718 ((frame + enc.height * enc.width) + (enc.height / 2 - i -1) * enc.width), enc.width); in amdgpu_cs_vce_encode_mv()
|
/third_party/skia/third_party/externals/angle2/src/tests/ |
H A D | capture_replay_tests.py | 12 # 1. Build all tests in angle_end2end with frame capture enabled 13 # 2. Run each test with frame capture 539 frame = 1 542 frame) 543 replay_file = "{}/{}_ContextReplayed{}.json".format(replay_build_dir, test_name, frame) 552 if frame > self.CAPTURE_FRAME_END: 554 frame = frame + 1
|