/third_party/mesa3d/src/gallium/drivers/nouveau/nv50/ |
H A D | nv84_video_bsp.c | 116 struct nv84_video_buffer *frame = (struct nv84_video_buffer *)desc->ref[i]; in nv84_decoder_bsp() local 117 if (!frame) break; in nv84_decoder_bsp() 118 /* The frame index is relative to the last IDR frame. So once the frame in nv84_decoder_bsp() 122 if (desc->frame_num >= frame->frame_num_max) { in nv84_decoder_bsp() 123 frame->frame_num_max = desc->frame_num; in nv84_decoder_bsp() 125 frame->frame_num -= frame->frame_num_max + 1; in nv84_decoder_bsp() 126 frame in nv84_decoder_bsp() [all...] |
/third_party/lwip/src/netif/ppp/ |
H A D | demand.c | 61 char *frame; variable 92 frame = malloc(framemax); in demand_conf() 93 if (frame == NULL) in demand_conf() 94 novm("demand frame"); in demand_conf() 221 * Calls loop_frame when a complete frame has been accumulated. 246 if (loop_frame((unsigned char *)frame, framelen)) 268 frame[framelen++] = c; 275 * loop_frame - given a frame obtained from the loopback, 277 * to transmit this frame later, put it on the pending queue. 285 loop_frame(frame, le [all...] |
/third_party/node/deps/v8/src/debug/ |
H A D | debug.h | 67 JavaScriptFrame* frame); 70 JavaScriptFrame* frame, 89 JavaScriptFrame* frame) const; 232 // The break target may not be the top-most frame, since we may be 234 void Break(JavaScriptFrame* frame, Handle<JSFunction> break_target); 322 // Check whether this frame is just about to return. 323 bool IsBreakAtReturn(JavaScriptFrame* frame); 359 bool PerformSideEffectCheckAtBytecode(InterpretedFrame* frame); 457 bool IsFrameBlackboxed(JavaScriptFrame* frame); 459 void ActivateStepOut(StackFrame* frame); [all...] |
/third_party/rust/crates/proc-macro2/src/ |
H A D | parse.rs | 205 let frame = (open_delimiter, trees); 207 let frame = (lo, frame); 208 stack.push(frame); 216 let frame = match stack.pop() { 217 Some(frame) => frame, 221 let (lo, frame) = frame; 222 let (open_delimiter, outer) = frame; [all...] |
/third_party/wpa_supplicant/wpa_supplicant-2.9/src/drivers/ |
H A D | driver_nl80211_event.c | 148 const u8 *frame, size_t len) in mlme_event_auth() 165 mgmt = (const struct ieee80211_mgmt *) frame; in mlme_event_auth() 168 "frame"); in mlme_event_auth() 210 const u8 *frame, size_t len, struct nlattr *wmm, in mlme_event_assoc() 230 mgmt = (const struct ieee80211_mgmt *) frame; in mlme_event_assoc() 233 "frame"); in mlme_event_assoc() 258 event.assoc_info.resp_frame = frame; in mlme_event_assoc() 635 const u8 *frame, size_t len) in mlme_event_mgmt() 645 mgmt = (const struct ieee80211_mgmt *) frame; in mlme_event_mgmt() 647 wpa_printf(MSG_DEBUG, "nl80211: Too short management frame"); in mlme_event_mgmt() 147 mlme_event_auth(struct wpa_driver_nl80211_data *drv, const u8 *frame, size_t len) mlme_event_auth() argument 209 mlme_event_assoc(struct wpa_driver_nl80211_data *drv, const u8 *frame, size_t len, struct nlattr *wmm, struct nlattr *req_ie) mlme_event_assoc() argument 633 mlme_event_mgmt(struct i802_bss *bss, struct nlattr *freq, struct nlattr *sig, const u8 *frame, size_t len) mlme_event_mgmt() argument 677 mlme_event_mgmt_tx_status(struct wpa_driver_nl80211_data *drv, struct nlattr *cookie, const u8 *frame, size_t len, struct nlattr *ack) mlme_event_mgmt_tx_status() argument 716 mlme_event_deauth_disassoc(struct wpa_driver_nl80211_data *drv, enum wpa_event_type type, const u8 *frame, size_t len) mlme_event_deauth_disassoc() argument 842 mlme_event_unprot_disconnect(struct wpa_driver_nl80211_data *drv, enum wpa_event_type type, const u8 *frame, size_t len) mlme_event_unprot_disconnect() argument 879 mlme_event(struct i802_bss *bss, enum nl80211_commands cmd, struct nlattr *frame, struct nlattr *addr, struct nlattr *timed_out, struct nlattr *freq, struct nlattr *ack, struct nlattr *cookie, struct nlattr *sig, struct nlattr *wmm, struct nlattr *req_ie) mlme_event() argument [all...] |
/third_party/ffmpeg/libavcodec/ |
H A D | audiotoolboxenc.c | 458 AVFrame *frame; in ffat_encode_callback() local 471 frame = ff_bufqueue_get(&at->frame_queue); in ffat_encode_callback() 475 data->mBuffers[0].mDataByteSize = frame->nb_samples * in ffat_encode_callback() 478 data->mBuffers[0].mData = frame->data[0]; in ffat_encode_callback() 479 if (*nb_packets > frame->nb_samples) in ffat_encode_callback() 480 *nb_packets = frame->nb_samples; in ffat_encode_callback() 483 ret = av_frame_ref(at->encoding_frame, frame); in ffat_encode_callback() 489 ff_bufqueue_add(avctx, &at->used_frame_queue, frame); in ffat_encode_callback() 495 const AVFrame *frame, int *got_packet_ptr) in ffat_encode() 511 if (frame) { in ffat_encode() 494 ffat_encode(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) ffat_encode() argument [all...] |
H A D | libopenjpegenc.c | 360 av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); in libopenjpeg_copy_packed8() 401 av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); in libopenjpeg_copy_packed12() 442 av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); in libopenjpeg_copy_packed16() 484 av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); in libopenjpeg_copy_unpacked8() 526 av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); in libopenjpeg_copy_unpacked16() 556 const AVFrame *frame, int *got_packet) in libopenjpeg_encode_frame() 565 const uint8_t *data[4] = { frame->data[0], frame->data[1], in libopenjpeg_encode_frame() 566 frame->data[2], frame in libopenjpeg_encode_frame() 555 libopenjpeg_encode_frame(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *frame, int *got_packet) libopenjpeg_encode_frame() argument [all...] |
H A D | g722dec.c | 87 static int g722_decode_frame(AVCodecContext *avctx, AVFrame *frame, in g722_decode_frame() argument 98 frame->nb_samples = avpkt->size * 2; in g722_decode_frame() 99 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in g722_decode_frame() 101 out_buf = (int16_t *)frame->data[0]; in g722_decode_frame()
|
H A D | libvo-amrwbenc.c | 116 const AVFrame *frame, int *got_packet_ptr) in amr_wb_encode_frame() 119 const int16_t *samples = (const int16_t *)frame->data[0]; in amr_wb_encode_frame() 131 av_log(avctx, AV_LOG_ERROR, "Error encoding frame\n"); in amr_wb_encode_frame() 135 if (frame->pts != AV_NOPTS_VALUE) in amr_wb_encode_frame() 136 avpkt->pts = frame->pts - ff_samples_to_time_base(avctx, avctx->initial_padding); in amr_wb_encode_frame() 115 amr_wb_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) amr_wb_encode_frame() argument
|
H A D | cngenc.c | 64 const AVFrame *frame, int *got_packet_ptr) in cng_encode_frame() 70 int16_t *samples = (int16_t*) frame->data[0]; in cng_encode_frame() 77 for (i = 0; i < frame->nb_samples; i++) { in cng_encode_frame() 81 energy /= frame->nb_samples; in cng_encode_frame() 63 cng_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) cng_encode_frame() argument
|
H A D | gsmdec.c | 68 static int gsm_decode_frame(AVCodecContext *avctx, AVFrame *frame, in gsm_decode_frame() argument 83 frame->nb_samples = avctx->frame_size; in gsm_decode_frame() 84 if ((res = ff_get_buffer(avctx, frame, 0)) < 0) in gsm_decode_frame() 86 samples = (int16_t *)frame->data[0]; in gsm_decode_frame()
|
H A D | libgsmdec.c | 85 static int libgsm_decode_frame(AVCodecContext *avctx, AVFrame *frame, in libgsm_decode_frame() argument 100 frame->nb_samples = avctx->frame_size; in libgsm_decode_frame() 101 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in libgsm_decode_frame() 103 samples = (int16_t *)frame->data[0]; in libgsm_decode_frame()
|
H A D | qsv_internal.h | 44 #include "libavutil/frame.h" 78 AVFrame *frame; member 112 * from the frame allocator. */ 149 int ff_qsv_find_surface_idx(QSVFramesContext *ctx, QSVFrame *frame); 151 void ff_qsv_frame_add_ext_param(AVCodecContext *avctx, QSVFrame *frame, 154 int ff_qsv_map_frame_to_surface(const AVFrame *frame, mfxFrameSurface1 *surface);
|
H A D | vima.c | 120 static int decode_frame(AVCodecContext *avctx, AVFrame *frame, in decode_frame() argument 158 frame->nb_samples = samples; in decode_frame() 159 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in decode_frame() 163 uint16_t *dest = (uint16_t *)frame->data[0] + chan; in decode_frame()
|
H A D | ra144dec.c | 63 static int ra144_decode_frame(AVCodecContext * avctx, AVFrame *frame, in ra144_decode_frame() argument 71 unsigned int lpc_refl[LPC_ORDER]; // LPC reflection coefficients of the frame in ra144_decode_frame() 88 frame->nb_samples = NBLOCKS * BLOCKSIZE; in ra144_decode_frame() 89 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in ra144_decode_frame() 91 samples = (int16_t *)frame->data[0]; in ra144_decode_frame()
|
/third_party/skia/third_party/externals/libwebp/src/mux/ |
H A D | muxread.c | 342 // single image and not forced canvas size => use dimension of first frame in MuxGetCanvasInfo() 456 WebPMuxFrameInfo* const frame) { in MuxGetFrameInternal() 461 // Get frame chunk. in MuxGetFrameInternal() 465 frame->x_offset = 2 * GetLE24(frame_data->bytes + 0); in MuxGetFrameInternal() 466 frame->y_offset = 2 * GetLE24(frame_data->bytes + 3); in MuxGetFrameInternal() 469 frame->duration = GetLE24(frame_data->bytes + 12); in MuxGetFrameInternal() 470 frame->dispose_method = in MuxGetFrameInternal() 472 frame->blend_method = (bits & 2) ? WEBP_MUX_NO_BLEND : WEBP_MUX_BLEND; in MuxGetFrameInternal() 474 frame->id = ChunkGetIdFromTag(wpi->header_->tag_); in MuxGetFrameInternal() 475 return SynthesizeBitstream(wpi, &frame in MuxGetFrameInternal() 455 MuxGetFrameInternal(const WebPMuxImage* const wpi, WebPMuxFrameInfo* const frame) MuxGetFrameInternal() argument 478 WebPMuxGetFrame( const WebPMux* mux, uint32_t nth, WebPMuxFrameInfo* frame) WebPMuxGetFrame() argument [all...] |
/third_party/ffmpeg/tests/fate/ |
H A D | mov.mak | 62 # Makes sure that we handle edit lists ending on a B-frame correctly. 63 # The last frame in decoding order which is B-frame should be output, but the last but-one P-frame shouldn't be 67 # Makes sure that we handle timestamps of packets in case of multiple edit lists with one of them ending on a B-frame correctly. 70 # Makes sure that if edit list ends on a B-frame but before the I-frame, then we output the B-frame but discard the I-frame. 79 # Makes sure that 1st key-frame i [all...] |
/third_party/node/deps/v8/src/runtime/ |
H A D | runtime-wasm.cc | 11 #include "src/execution/frame-constants.h" 42 // We skip at least one frame. 46 DCHECK_EQ(type, frame_iterator_.frame()->type()); 50 // Type check the frame where the iterator stopped now. 51 DCHECK_NOT_NULL(frame()); 54 FrameType* frame() { return FrameType::cast(frame_iterator_.frame()); } in frame() function in v8::internal::__anon15054::FrameFinder 65 .frame() 219 DCHECK_EQ(*instance, frame_finder.frame()->wasm_instance()); in RUNTIME_FUNCTION() 316 int func_index = frame_finder.frame() in RUNTIME_FUNCTION() 561 WasmFrame* frame = frame_finder.frame(); RUNTIME_FUNCTION() local [all...] |
/third_party/vk-gl-cts/modules/egl/ |
H A D | teglGetFrameTimestampsTests.cpp | 600 // Do extra rendering to allow frame pacing to stabilize. in executeForConfig() 601 // The frame timestamp validation below assumes there is no frame janking, in executeForConfig() 604 // to jank. This will cause frame timestamps read back not matching with in executeForConfig() 607 // and verifying frame timestamps. in executeForConfig() 628 FrameTimes& frame = frameTimes[i]; in executeForConfig() local 643 // Get the frame id. in executeForConfig() 646 frame.frameId = nextFrameId; in executeForConfig() 652 frame.compositeDeadline = compositorTimingValues[0]; in executeForConfig() 653 frame in executeForConfig() [all...] |
/third_party/ffmpeg/libavfilter/ |
H A D | f_realtime.c | 37 static int filter_frame(AVFilterLink *inlink, AVFrame *frame) in filter_frame() argument 42 if (frame->pts != AV_NOPTS_VALUE) { in filter_frame() 43 int64_t pts = av_rescale_q(frame->pts, inlink->time_base, AV_TIME_BASE_Q) / s->speed; in filter_frame() 65 return ff_filter_frame(inlink->dst->outputs[0], frame); in filter_frame()
|
/third_party/mesa3d/include/drm-uapi/ |
H A D | lima_drm.h | 72 /* frame used to setup GP for each task */ 74 __u32 frame[LIMA_GP_FRAME_REG_NUM]; member 80 /* frame used to setup mali400 GPU PP for each task */ 82 __u32 frame[LIMA_PP_FRAME_REG_NUM]; member 89 /* frame used to setup mali450 GPU PP for each task */ 91 __u32 frame[LIMA_PP_FRAME_REG_NUM]; member 120 __u32 frame_size; /* in, size of frame field */ 122 __u64 frame; /* in, GP/PP frame */ member
|
/third_party/jinja2/ |
H A D | nativetypes.py | 63 self, node: nodes.Expr, frame: Frame, finalize: CodeGenerator._FinalizeInfo 65 const = node.as_const(frame.eval_ctx) 76 self, node: nodes.Expr, frame: Frame, finalize: CodeGenerator._FinalizeInfo 82 self, node: nodes.Expr, frame: Frame, finalize: CodeGenerator._FinalizeInfo
|
/third_party/mesa3d/src/gallium/drivers/r600/sfn/ |
H A D | sfn_conditionaljumptracker.cpp | 115 auto& frame = *impl->m_jump_stack.top(); in pop() 116 if (frame.type != type) in pop() 119 frame.fixup_pop(final); in pop() 120 if (frame.type == jt_loop) in pop()
|
/third_party/node/test/sequential/ |
H A D | test-diagnostic-dir-heap-prof.js | 37 return { frame: first, roots: profile.head.children }; 41 const { frame, roots } = findFirstFrame(file, func); 42 if (!frame) { 47 assert.notStrictEqual(frame, undefined);
|
/third_party/python/Lib/idlelib/ |
H A D | search.py | 145 frame = Frame(top) 146 frame.pack() 147 text = Text(frame, inactiveselectbackground='gray') 156 button = Button(frame, text="Search (selection ignored)", command=show_find)
|