/third_party/ffmpeg/libavcodec/ |
H A D | opus_silk.c | 58 SilkFrame frame[2]; member 305 static inline void silk_decode_lpc(SilkContext *s, SilkFrame *frame, in silk_decode_lpc() argument 378 if (offset != 4 && frame->coded) { in silk_decode_lpc() 383 nlsf_leadin[i] = frame->nlsf[i] + in silk_decode_lpc() 384 ((nlsf[i] - frame->nlsf[i]) * offset >> 2); in silk_decode_lpc() 387 memcpy(lpc_leadin, frame->lpc, 16 * sizeof(float)); in silk_decode_lpc() 398 memcpy(frame->nlsf, nlsf, order * sizeof(nlsf[0])); in silk_decode_lpc() 399 memcpy(frame->lpc, lpc, order * sizeof(lpc[0])); in silk_decode_lpc() 513 /* per frame */ in silk_decode_frame() 528 SilkFrame * const frame in silk_decode_frame() local 768 silk_flush_frame(SilkFrame *frame) silk_flush_frame() argument [all...] |
H A D | eatqi.c | 97 static inline void tqi_idct_put(AVCodecContext *avctx, AVFrame *frame, in tqi_idct_put() argument 101 ptrdiff_t linesize = frame->linesize[0]; in tqi_idct_put() 102 uint8_t *dest_y = frame->data[0] + t->mb_y * 16 * linesize + t->mb_x * 16; in tqi_idct_put() 103 uint8_t *dest_cb = frame->data[1] + t->mb_y * 8 * frame->linesize[1] + t->mb_x * 8; in tqi_idct_put() 104 uint8_t *dest_cr = frame->data[2] + t->mb_y * 8 * frame->linesize[2] + t->mb_x * 8; in tqi_idct_put() 112 ff_ea_idct_put_c(dest_cb, frame->linesize[1], block[4]); in tqi_idct_put() 113 ff_ea_idct_put_c(dest_cr, frame->linesize[2], block[5]); in tqi_idct_put() 127 static int tqi_decode_frame(AVCodecContext *avctx, AVFrame *frame, in tqi_decode_frame() argument [all...] |
H A D | mediacodecdec_common.c | 277 AVFrame *frame) in mediacodec_wrap_hw_buffer() 283 frame->buf[0] = NULL; in mediacodec_wrap_hw_buffer() 284 frame->width = avctx->width; in mediacodec_wrap_hw_buffer() 285 frame->height = avctx->height; in mediacodec_wrap_hw_buffer() 286 frame->format = avctx->pix_fmt; in mediacodec_wrap_hw_buffer() 287 frame->sample_aspect_ratio = avctx->sample_aspect_ratio; in mediacodec_wrap_hw_buffer() 290 frame->pts = av_rescale_q(info->presentationTimeUs, in mediacodec_wrap_hw_buffer() 294 frame->pts = info->presentationTimeUs; in mediacodec_wrap_hw_buffer() 296 frame->pkt_dts = AV_NOPTS_VALUE; in mediacodec_wrap_hw_buffer() 297 frame in mediacodec_wrap_hw_buffer() 273 mediacodec_wrap_hw_buffer(AVCodecContext *avctx, MediaCodecDecContext *s, ssize_t index, FFAMediaCodecBufferInfo *info, AVFrame *frame) mediacodec_wrap_hw_buffer() argument 350 mediacodec_wrap_sw_buffer(AVCodecContext *avctx, MediaCodecDecContext *s, uint8_t *data, size_t size, ssize_t index, FFAMediaCodecBufferInfo *info, AVFrame *frame) mediacodec_wrap_sw_buffer() argument 750 ff_mediacodec_dec_receive(AVCodecContext *avctx, MediaCodecDecContext *s, AVFrame *frame, bool wait) ff_mediacodec_dec_receive() argument [all...] |
H A D | eamad.c | 47 #define MADk_TAG MKTAG('M', 'A', 'D', 'k') /* MAD I-frame */ 48 #define MADm_TAG MKTAG('M', 'A', 'D', 'm') /* MAD P-frame */ 49 #define MADe_TAG MKTAG('M', 'A', 'D', 'e') /* MAD lqp-frame */ 95 static inline void comp_block(MadContext *t, AVFrame *frame, in comp_block() argument 103 comp(frame->data[0] + (mb_y*16 + ((j&2)<<2))*frame->linesize[0] + mb_x*16 + ((j&1)<<3), in comp_block() 104 frame->linesize[0], in comp_block() 112 comp(frame->data[index] + (mb_y*8)*frame->linesize[index] + mb_x * 8, in comp_block() 113 frame in comp_block() 119 idct_put(MadContext *t, AVFrame *frame, int16_t *block, int mb_x, int mb_y, int j) idct_put() argument 211 decode_mb(MadContext *s, AVFrame *frame, int inter) decode_mb() argument 250 decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt) decode_frame() argument [all...] |
H A D | dxtory.c | 417 typedef int (*decode_slice_func)(GetBitContext *gb, AVFrame *frame, 487 static int dx2_decode_slice_5x5(GetBitContext *gb, AVFrame *frame, in dx2_decode_slice_5x5() argument 493 int width = frame->width; in dx2_decode_slice_5x5() 494 int stride = frame->linesize[0]; in dx2_decode_slice_5x5() 495 uint8_t *dst = frame->data[0] + stride * line; in dx2_decode_slice_5x5() 527 static int dx2_decode_slice_555(GetBitContext *gb, AVFrame *frame, in dx2_decode_slice_555() argument 530 return dx2_decode_slice_5x5(gb, frame, line, left, lru, 0); in dx2_decode_slice_555() 533 static int dx2_decode_slice_565(GetBitContext *gb, AVFrame *frame, in dx2_decode_slice_565() argument 536 return dx2_decode_slice_5x5(gb, frame, line, left, lru, 1); in dx2_decode_slice_565() 556 static int dx2_decode_slice_rgb(GetBitContext *gb, AVFrame *frame, in dx2_decode_slice_rgb() argument 595 dx2_decode_slice_410(GetBitContext *gb, AVFrame *frame, int line, int left, uint8_t lru[3][8]) dx2_decode_slice_410() argument 674 dx2_decode_slice_420(GetBitContext *gb, AVFrame *frame, int line, int left, uint8_t lru[3][8]) dx2_decode_slice_420() argument 746 dx2_decode_slice_444(GetBitContext *gb, AVFrame *frame, int line, int left, uint8_t lru[3][8]) dx2_decode_slice_444() argument [all...] |
H A D | videotoolbox.c | 93 static int videotoolbox_postproc_frame(void *avctx, AVFrame *frame) in videotoolbox_postproc_frame() argument 96 VTHWFrame *ref = (VTHWFrame *)frame->buf[0]->data; in videotoolbox_postproc_frame() 99 av_log(avctx, AV_LOG_ERROR, "No frame decoded?\n"); in videotoolbox_postproc_frame() 100 av_frame_unref(frame); in videotoolbox_postproc_frame() 104 frame->crop_right = 0; in videotoolbox_postproc_frame() 105 frame->crop_left = 0; in videotoolbox_postproc_frame() 106 frame->crop_top = 0; in videotoolbox_postproc_frame() 107 frame->crop_bottom = 0; in videotoolbox_postproc_frame() 109 if ((ret = av_vt_pixbuf_set_attachments(avctx, ref->pixbuf, frame)) < 0) in videotoolbox_postproc_frame() 112 frame in videotoolbox_postproc_frame() 124 ff_videotoolbox_alloc_frame(AVCodecContext *avctx, AVFrame *frame) ff_videotoolbox_alloc_frame() argument 515 videotoolbox_buffer_create(AVCodecContext *avctx, AVFrame *frame) videotoolbox_buffer_create() argument 1003 ff_videotoolbox_common_end_frame(AVCodecContext *avctx, AVFrame *frame) ff_videotoolbox_common_end_frame() argument 1040 AVFrame *frame = h->cur_pic_ptr->f; videotoolbox_h264_end_frame() local 1073 AVFrame *frame = h->ref->frame; videotoolbox_hevc_end_frame() local 1105 AVFrame *frame = s->current_picture_ptr->f; videotoolbox_mpeg_end_frame() local 1129 AVFrame *frame = ctx->frame; videotoolbox_prores_end_frame() local [all...] |
H A D | mv30.c | 404 static int decode_intra(AVCodecContext *avctx, GetBitContext *gb, AVFrame *frame) in decode_intra() argument 418 linesize[0] = frame->linesize[0]; in decode_intra() 419 linesize[1] = frame->linesize[0]; in decode_intra() 420 linesize[2] = frame->linesize[0]; in decode_intra() 421 linesize[3] = frame->linesize[0]; in decode_intra() 422 linesize[4] = frame->linesize[1]; in decode_intra() 423 linesize[5] = frame->linesize[2]; in decode_intra() 440 dst[0] = frame->data[0] + linesize[0] * y + x; in decode_intra() 441 dst[1] = frame->data[0] + linesize[0] * y + x + 8; in decode_intra() 442 dst[2] = frame in decode_intra() 463 decode_inter(AVCodecContext *avctx, GetBitContext *gb, AVFrame *frame, AVFrame *prev) decode_inter() argument 604 decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt) decode_frame() argument [all...] |
H A D | librsvgdec.c | 36 static int librsvg_decode_frame(AVCodecContext *avctx, AVFrame *frame, in librsvg_decode_frame() argument 73 if ((ret = ff_get_buffer(avctx, frame, 0))) in librsvg_decode_frame() 75 frame->pict_type = AV_PICTURE_TYPE_I; in librsvg_decode_frame() 76 frame->key_frame = 1; in librsvg_decode_frame() 78 image = cairo_image_surface_create_for_data(frame->data[0], CAIRO_FORMAT_ARGB32, in librsvg_decode_frame() 79 frame->width, frame->height, in librsvg_decode_frame() 80 frame->linesize[0]); in librsvg_decode_frame()
|
H A D | libsvtav1.c | 28 #include "libavutil/frame.h" 57 AVFrame *frame; member 318 static int read_in_data(EbSvtAv1EncConfiguration *param, const AVFrame *frame, in read_in_data() argument 328 linesizes[i] = frame->linesize[i]; in read_in_data() 330 ret = av_image_fill_plane_sizes(sizes, frame->format, frame->height, in read_in_data() 342 in_data->luma = frame->data[0]; in read_in_data() 343 in_data->cb = frame->data[1]; in read_in_data() 344 in_data->cr = frame->data[2]; in read_in_data() 346 in_data->y_stride = AV_CEIL_RSHIFT(frame in read_in_data() 415 eb_send_frame(AVCodecContext *avctx, const AVFrame *frame) eb_send_frame() argument 487 AVFrame *frame = svt_enc->frame; eb_receive_packet() local [all...] |
/third_party/ffmpeg/libavfilter/ |
H A D | vf_phase.c | 68 AVFrame *frame; /* previous frame */ member 163 av_frame_free(&s->frame); in filter_frame() 164 /* we keep a reference to the previous frame so the filter can start in filter_frame() 165 * being useful as soon as it's not disabled, avoiding the 1-frame in filter_frame() 167 s->frame = av_frame_clone(in); in filter_frame() 178 if (!s->frame) { in filter_frame() 179 s->frame = in; in filter_frame() 182 mode = s->analyze_plane(ctx, s->mode, s->frame, in); in filter_frame() 186 const uint8_t *buf = s->frame in filter_frame() [all...] |
H A D | src_movie.c | 26 * @todo use direct rendering (no allocation of a new frame) 413 AVFrame *frame, enum AVMediaType frame_type, in describe_frame_to_str() 420 av_ts2str(frame->pts), av_ts2timestr(frame->pts, &link->time_base), in describe_frame_to_str() 421 frame->width, frame->height, in describe_frame_to_str() 422 frame->sample_aspect_ratio.num, in describe_frame_to_str() 423 frame->sample_aspect_ratio.den); in describe_frame_to_str() 428 av_ts2str(frame->pts), av_ts2timestr(frame in describe_frame_to_str() 412 describe_frame_to_str(char *dst, size_t dst_size, AVFrame *frame, enum AVMediaType frame_type, AVFilterLink *link) describe_frame_to_str() argument 504 AVFrame *frame; movie_push_frame() local [all...] |
H A D | vf_swapuv.c | 43 static void do_swap(AVFrame *frame) in do_swap() argument 45 FFSWAP(uint8_t*, frame->data[1], frame->data[2]); in do_swap() 46 FFSWAP(int, frame->linesize[1], frame->linesize[2]); in do_swap() 47 FFSWAP(AVBufferRef*, frame->buf[1], frame->buf[2]); in do_swap()
|
H A D | vf_siti.c | 145 static int is_full_range(AVFrame* frame) in is_full_range() argument 148 if (frame->color_range == AVCOL_RANGE_UNSPECIFIED || frame->color_range == AVCOL_RANGE_NB) in is_full_range() 149 return frame->format == AV_PIX_FMT_YUVJ420P || frame->format == AV_PIX_FMT_YUVJ422P; in is_full_range() 150 return frame->color_range == AVCOL_RANGE_JPEG; in is_full_range() 153 // Check frame's color range and convert to full range if needed 214 // Calculate pixel difference between current and previous frame, and update previous 226 // Previous frame is already converted to full range in calculate_motion() 281 static int filter_frame(AVFilterLink *inlink, AVFrame *frame) in filter_frame() argument [all...] |
H A D | vf_fillborders.c | 55 void (*fillborders)(struct FillBordersContext *s, AVFrame *frame); 80 static void smear_borders8(FillBordersContext *s, AVFrame *frame) in smear_borders8() argument 85 uint8_t *ptr = frame->data[p]; in smear_borders8() 86 int linesize = frame->linesize[p]; in smear_borders8() 110 static void smear_borders16(FillBordersContext *s, AVFrame *frame) in smear_borders16() argument 115 uint16_t *ptr = (uint16_t *)frame->data[p]; in smear_borders16() 116 int linesize = frame->linesize[p] / 2; in smear_borders16() 142 static void mirror_borders8(FillBordersContext *s, AVFrame *frame) in mirror_borders8() argument 147 uint8_t *ptr = frame->data[p]; in mirror_borders8() 148 int linesize = frame in mirror_borders8() 175 mirror_borders16(FillBordersContext *s, AVFrame *frame) mirror_borders16() argument 208 fixed_borders8(FillBordersContext *s, AVFrame *frame) fixed_borders8() argument 233 fixed_borders16(FillBordersContext *s, AVFrame *frame) fixed_borders16() argument 266 reflect_borders8(FillBordersContext *s, AVFrame *frame) reflect_borders8() argument 299 reflect_borders16(FillBordersContext *s, AVFrame *frame) reflect_borders16() argument 332 wrap_borders8(FillBordersContext *s, AVFrame *frame) wrap_borders8() argument 365 wrap_borders16(FillBordersContext *s, AVFrame *frame) wrap_borders16() argument 408 fade_borders8(FillBordersContext *s, AVFrame *frame) fade_borders8() argument 449 fade_borders16(FillBordersContext *s, AVFrame *frame) fade_borders16() argument 491 margins_borders8(FillBordersContext *s, AVFrame *frame) margins_borders8() argument 536 margins_borders16(FillBordersContext *s, AVFrame *frame) margins_borders16() argument 584 filter_frame(AVFilterLink *inlink, AVFrame *frame) filter_frame() argument [all...] |
H A D | f_cue.c | 47 AVFrame *frame = ff_inlink_peek_frame(inlink, 0); in activate() local 48 int64_t pts = av_rescale_q(frame->pts, inlink->time_base, AV_TIME_BASE_Q); in activate() 56 int ret = ff_inlink_consume_frame(inlink, &frame); in activate() 59 return ff_filter_frame(outlink, frame); in activate() 65 frame = ff_inlink_peek_frame(inlink, ff_inlink_queued_frames(inlink) - 1); in activate() 66 pts = av_rescale_q(frame->pts, inlink->time_base, AV_TIME_BASE_Q); in activate() 77 int ret = ff_inlink_consume_frame(inlink, &frame); in activate() 80 return ff_filter_frame(outlink, frame); in activate()
|
H A D | vf_iccdetect.c | 68 static int iccdetect_filter_frame(AVFilterLink *inlink, AVFrame *frame) in iccdetect_filter_frame() argument 77 sd = av_frame_get_side_data(frame, AV_FRAME_DATA_ICC_PROFILE); in iccdetect_filter_frame() 79 return ff_filter_frame(inlink->dst->outputs[0], frame); in iccdetect_filter_frame() 106 if (s->force || frame->color_primaries == AVCOL_PRI_UNSPECIFIED) in iccdetect_filter_frame() 107 frame->color_primaries = s->profile_prim; in iccdetect_filter_frame() 111 if (s->force || frame->color_trc == AVCOL_TRC_UNSPECIFIED) in iccdetect_filter_frame() 112 frame->color_trc = s->profile_trc; in iccdetect_filter_frame() 115 return ff_filter_frame(inlink->dst->outputs[0], frame); in iccdetect_filter_frame()
|
/foundation/graphic/graphic_2d/rosen/samples/2d_graphics/drawing_engine/drawing_surface/ |
H A D | surface_ohos_gl.cpp | 61 bool SurfaceOhosGl::NativeFlushFrame(std::unique_ptr<SurfaceFrame>& frame) in NativeFlushFrame() argument 63 return FlushFrame(frame); in NativeFlushFrame() 66 bool SurfaceOhosGl::FlushFrame(std::unique_ptr<SurfaceFrame>& frame) in FlushFrame() argument 78 SkCanvas* SurfaceOhosGl::GetSkCanvas(std::unique_ptr<SurfaceFrame>& frame) in GetSkCanvas() argument 84 return drawingProxy_->AcquireSkCanvas(frame); in GetSkCanvas() 87 Drawing::Canvas* SurfaceOhosGl::GetCanvas(std::unique_ptr<SurfaceFrame>& frame) in GetCanvas() argument 93 return drawingProxy_->AcquireDrCanvas(frame); in GetCanvas()
|
/third_party/ffmpeg/libavformat/ |
H A D | sierravmd.c | 210 av_log(s, AV_LOG_ERROR, "Failed to read frame record\n"); in vmd_read_header() 218 av_log(s, AV_LOG_ERROR, "Invalid frame size\n"); in vmd_read_header() 270 vmd_frame *frame; in vmd_read_packet() local 275 frame = &vmd->frame_table[vmd->current_frame]; in vmd_read_packet() 277 avio_seek(pb, frame->frame_offset, SEEK_SET); in vmd_read_packet() 279 if(ffio_limit(pb, frame->frame_size) != frame->frame_size) in vmd_read_packet() 281 ret = av_new_packet(pkt, frame->frame_size + BYTES_PER_FRAME_RECORD); in vmd_read_packet() 285 memcpy(pkt->data, frame->frame_record, BYTES_PER_FRAME_RECORD); in vmd_read_packet() 286 if(vmd->is_indeo3 && frame in vmd_read_packet() [all...] |
H A D | aqtitledec.c | 42 int frame; in aqt_probe() local 45 if (sscanf(ptr, "-->> %d", &frame) == 1) in aqt_probe() 55 int64_t pos = 0, frame = AV_NOPTS_VALUE; in aqt_read_header() local 73 if (sscanf(line, "-->> %"SCNd64, &frame) == 1) { in aqt_read_header() 77 if (frame >= sub->pts && (uint64_t)frame - sub->pts < INT64_MAX) in aqt_read_header() 78 sub->duration = frame - sub->pts; in aqt_read_header() 91 sub->pts = frame; in aqt_read_header() 127 { "subfps", "set the movie frame rate", OFFSET(frame_rate), AV_OPT_TYPE_RATIONAL, {.dbl=25}, 0, INT_MAX, SD },
|
/device/soc/rockchip/rk3568/hardware/mpp/mpp/legacy/inc/ |
H A D | mpp_frame_impl.h | 43 * 0 - frame 52 * current decoded frame whether to display 59 * send decoded frame belong which view 75 * info_change - set when buffer resized or frame infomation changed 114 * frame buffer compression (FBC) information 129 * pointer for multiple frame output at one time 138 MPP_RET mpp_frame_set_next(MppFrame frame, MppFrame next); 139 MPP_RET mpp_frame_copy(MppFrame frame, MppFrame next); 141 RK_U32 mpp_frame_get_fbc_offset(MppFrame frame); 142 RK_U32 mpp_frame_get_fbc_stride(MppFrame frame); [all...] |
/third_party/skia/third_party/externals/oboe/src/flowgraph/resampler/ |
H A D | PolyphaseResamplerStereo.cpp | 28 void PolyphaseResamplerStereo::writeFrame(const float *frame) { in writeFrame() argument 29 // Move cursor before write so that cursor points to last written frame in read. in writeFrame() 36 const float left = frame[0]; in writeFrame() 37 const float right = frame[1]; in writeFrame() 45 void PolyphaseResamplerStereo::readFrame(float *frame) { in readFrame() argument 76 frame[0] = left; in readFrame() 77 frame[1] = right; in readFrame()
|
/third_party/nghttp2/src/ |
H A D | shrpx_http2_session.cc | 901 int on_header_callback2(nghttp2_session *session, const nghttp2_frame *frame, in on_header_callback2() argument 906 nghttp2_session_get_stream_user_data(session, frame->hd.stream_id)); in on_header_callback2() 918 switch (frame->hd.type) { in on_header_callback2() 920 auto trailer = frame->headers.cat == NGHTTP2_HCAT_HEADERS && in on_header_callback2() 962 auto promised_stream_id = frame->push_promise.promised_stream_id; in on_header_callback2() 1011 const nghttp2_frame *frame, nghttp2_rcbuf *name, in on_invalid_header_callback2() 1016 nghttp2_session_get_stream_user_data(session, frame->hd.stream_id)); in on_invalid_header_callback2() 1023 if (frame->hd.type == NGHTTP2_PUSH_PROMISE) { in on_invalid_header_callback2() 1024 stream_id = frame->push_promise.promised_stream_id; in on_invalid_header_callback2() 1026 stream_id = frame in on_invalid_header_callback2() 1010 on_invalid_header_callback2(nghttp2_session *session, const nghttp2_frame *frame, nghttp2_rcbuf *name, nghttp2_rcbuf *value, uint8_t flags, void *user_data) on_invalid_header_callback2() argument 1047 on_begin_headers_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_begin_headers_callback() argument 1095 on_response_headers(Http2Session *http2session, Downstream *downstream, nghttp2_session *session, const nghttp2_frame *frame) on_response_headers() argument 1222 on_frame_recv_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_frame_recv_callback() argument 1468 on_frame_send_callback(nghttp2_session *session, const nghttp2_frame *frame, void *user_data) on_frame_send_callback() argument 1514 on_frame_not_send_callback(nghttp2_session *session, const nghttp2_frame *frame, int lib_error_code, void *user_data) on_frame_not_send_callback() argument 1566 send_data_callback(nghttp2_session *session, nghttp2_frame *frame, const uint8_t *framehd, size_t length, nghttp2_data_source *source, void *user_data) send_data_callback() argument 2411 on_settings_received(const nghttp2_frame *frame) on_settings_received() argument [all...] |
/third_party/nghttp2/tests/ |
H A D | failmalloc_test.c | 277 nghttp2_frame frame; in run_nghttp2_session_recv() local 324 nghttp2_frame_settings_init(&frame.settings, NGHTTP2_FLAG_NONE, in run_nghttp2_session_recv() 327 nghttp2_frame_pack_settings(&bufs, &frame.settings); in run_nghttp2_session_recv() 328 nghttp2_frame_settings_free(&frame.settings, nghttp2_mem_fm()); in run_nghttp2_session_recv() 342 nghttp2_frame_headers_init(&frame.headers, NGHTTP2_FLAG_END_STREAM, 1, in run_nghttp2_session_recv() 344 nghttp2_frame_pack_headers(&bufs, &frame.headers, &deflater); in run_nghttp2_session_recv() 345 nghttp2_frame_headers_free(&frame.headers, nghttp2_mem_fm()); in run_nghttp2_session_recv() 357 nghttp2_frame_ping_init(&frame.ping, NGHTTP2_FLAG_NONE, NULL); in run_nghttp2_session_recv() 358 nghttp2_frame_pack_ping(&bufs, &frame.ping); in run_nghttp2_session_recv() 359 nghttp2_frame_ping_free(&frame in run_nghttp2_session_recv() 397 nghttp2_frame frame, oframe; run_nghttp2_frame_pack_headers() local 447 nghttp2_frame frame, oframe; run_nghttp2_frame_pack_settings() local [all...] |
/third_party/libunwind/libunwind/src/arm/ |
H A D | Gstep.c | 94 /* Check if this is a signal frame. */ in unw_step() 131 /* Fall back on APCS frame parsing. in unw_step() 141 Debug (13, "%s%s%s%s(ret=%d), trying frame-chain\n", in unw_step() 148 /* DWARF unwinding failed, try to follow APCS/optimized APCS frame chain */ in unw_step() 151 unw_word_t frame; in unw_step() local 155 if (dwarf_get(&c->dwarf, c->dwarf.loc[UNW_ARM_R11], &frame) < 0) in unw_step() 162 if (frame) in unw_step() 164 if (dwarf_get(&c->dwarf, DWARF_LOC(frame, 0), &instr) < 0) in unw_step() 175 /* Standard APCS frame. */ in unw_step() 176 ip_loc = DWARF_LOC(frame in unw_step() [all...] |
/third_party/node/deps/v8/third_party/wasm-api/example/ |
H A D | start.cc | 10 void print_frame(const wasm::Frame* frame) { in print_frame() argument 11 std::cout << "> " << frame->instance(); in print_frame() 12 std::cout << " @ 0x" << std::hex << frame->module_offset(); in print_frame() 13 std::cout << " = " << frame->func_index(); in print_frame() 14 std::cout << ".0x" << std::hex << frame->func_offset() << std::endl; in print_frame() 61 auto frame = trap->origin(); in run() local 62 if (frame) { in run() 63 print_frame(frame.get()); in run()
|