/third_party/ffmpeg/libavcodec/ |
H A D | fastaudio.c | 108 static int fastaudio_decode(AVCodecContext *avctx, AVFrame *frame, in fastaudio_decode() argument 117 frame->nb_samples = subframes * 256; in fastaudio_decode() 118 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in fastaudio_decode() 172 memcpy(frame->extended_data[channel] + 1024 * subframe, result, 256 * sizeof(float)); in fastaudio_decode()
|
H A D | mace.c | 236 static int mace_decode_frame(AVCodecContext *avctx, AVFrame *frame, in mace_decode_frame() argument 255 frame->nb_samples = 3 * (buf_size << (1 - is_mace3)) / channels; in mace_decode_frame() 256 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in mace_decode_frame() 258 samples = (int16_t **)frame->extended_data; in mace_decode_frame()
|
H A D | sunrastenc.c | 174 const AVFrame *frame, int *got_packet_ptr) in sunrast_encode_frame() 184 sunrast_image_write_image(avctx, frame->data[0], in sunrast_encode_frame() 185 (const uint32_t *)frame->data[1], in sunrast_encode_frame() 186 frame->linesize[0]); in sunrast_encode_frame() 173 sunrast_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) sunrast_encode_frame() argument
|
H A D | ra288.c | 190 static int ra288_decode_frame(AVCodecContext * avctx, AVFrame *frame, in ra288_decode_frame() argument 212 frame->nb_samples = RA288_BLOCK_SIZE * RA288_BLOCKS_PER_FRAME; in ra288_decode_frame() 213 if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) in ra288_decode_frame() 215 out = (float *)frame->data[0]; in ra288_decode_frame()
|
H A D | libtheoraenc.c | 277 const AVFrame *frame, int *got_packet) in encode_frame() 285 if (!frame) { in encode_frame() 297 t_yuv_buffer[i].stride = frame->linesize[i]; in encode_frame() 298 t_yuv_buffer[i].data = frame->data[i]; in encode_frame() 311 message = "differing frame sizes"; in encode_frame() 349 pkt->pts = pkt->dts = frame->pts; in encode_frame() 276 encode_frame(AVCodecContext* avc_context, AVPacket *pkt, const AVFrame *frame, int *got_packet) encode_frame() argument
|
H A D | aptxenc.c | 213 const AVFrame *frame, int *got_packet_ptr) in aptx_encode_frame() 218 if ((ret = ff_af_queue_add(&s->afq, frame)) < 0) in aptx_encode_frame() 221 output_size = s->block_size * frame->nb_samples/4; in aptx_encode_frame() 230 samples[channel][sample] = (int32_t)AV_RN32A(&frame->data[channel][4*(ipos+sample)]) >> 8; in aptx_encode_frame() 235 ff_af_queue_remove(&s->afq, frame->nb_samples, &avpkt->pts, &avpkt->duration); in aptx_encode_frame() 212 aptx_encode_frame(AVCodecContext *avctx, AVPacket *avpkt, const AVFrame *frame, int *got_packet_ptr) aptx_encode_frame() argument
|
/third_party/ffmpeg/libavformat/ |
H A D | thp.c | 36 unsigned frame; member 171 /* Terminate when last frame is reached. */ in thp_read_packet() 172 if (thp->frame >= thp->framecnt) in thp_read_packet() 177 /* Locate the next frame and read out its size. */ in thp_read_packet() 182 size = avio_rb32(pb); /* Total size of this frame. */ in thp_read_packet() 189 thp->frame++; in thp_read_packet() 212 thp->frame++; in thp_read_packet()
|
/third_party/ffmpeg/libavfilter/ |
H A D | vf_vidstabdetect.c | 57 " reference frame (frame # is the value)", OFFSETC(virtualTripod), AV_OPT_TYPE_INT, {.i64 = 0}, 0, INT_MAX, FLAGS}, 150 VSFrame frame; in filter_frame() local 157 frame.data[plane] = in->data[plane]; in filter_frame() 158 frame.linesize[plane] = in->linesize[plane]; in filter_frame() 160 if (vsMotionDetection(md, &localmotions, &frame) != VS_OK) { in filter_frame()
|
H A D | asrc_sine.c | 79 OPT_STR("samples_per_frame", samples_per_frame, "1024", 0, 0, "set the number of samples per frame",), 209 AVFrame *frame; in activate() local 235 if (!(frame = ff_get_audio_buffer(outlink, nb_samples))) in activate() 237 samples = (int16_t *)frame->data[0]; in activate() 250 frame->pts = sine->pts; in activate() 252 return ff_filter_frame(outlink, frame); in activate()
|
H A D | asrc_afirsrc.c | 75 { "nb_samples", "set the number of samples per requested frame", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, FLAGS }, 76 { "n", "set the number of samples per requested frame", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, FLAGS }, 261 AVFrame *frame; in activate() local 273 if (!(frame = ff_get_audio_buffer(outlink, nb_samples))) in activate() 276 memcpy(frame->data[0], s->taps + s->pts, nb_samples * sizeof(float)); in activate() 278 frame->pts = s->pts; in activate() 280 return ff_filter_frame(outlink, frame); in activate()
|
/third_party/libunwind/libunwind/doc/ |
H A D | unw_set_reg.tex | 20 \Var{reg} in the stack frame identified by cursor \Var{cp} to the 26 type of frame that \Var{cp} is referring to. For ordinary stack 28 (``callee-saved'') registers and frame-related registers (such as the 54 that is either invalid or not accessible in the current frame.
|
H A D | unw_get_fpreg.tex | 20 register \Var{reg} in the stack frame identified by cursor \Var{cp} 26 type of frame that \Var{cp} is referring to. For ordinary stack 28 (``callee-saved'') registers and frame-related registers (such as the 53 that is either invalid or not accessible in the current frame.
|
H A D | unw_get_reg.tex | 20 \Var{reg} in the stack frame identified by cursor \Var{cp} and stores 26 type of frame that \Var{cp} is referring to. For ordinary stack 28 (``callee-saved'') registers and frame-related registers (such as the 54 that is either invalid or not accessible in the current frame.
|
H A D | unw_set_fpreg.tex | 20 \Var{reg} in the stack frame identified by cursor \Var{cp} to the 26 type of frame that \Var{cp} is referring to. For ordinary stack 28 (``callee-saved'') registers and frame-related registers (such as the 53 that is either invalid or not accessible in the current frame.
|
/third_party/python/Tools/demo/ |
H A D | redemo.py | 82 frame = Frame(self.master) 83 frame.pack(fill=X) 84 self.frames.append(frame) 87 box = Checkbutton(frame,
|
/third_party/python/Lib/idlelib/ |
H A D | grep.py | 203 frame = Frame(top) 204 frame.pack() 205 text = Text(frame, height=5) 213 button = Button(frame, text="Show GrepDialog", command=show_grep_dialog)
|
/third_party/vk-gl-cts/external/vulkan-docs/src/scripts/spec_tools/ |
H A D | shared.py | 215 and a stack frame where the message was created. 219 replacement=None, see_also=None, fix=None, frame=None): 251 if frame: 253 frameinfo = getframeinfo(frame) 257 del frame
|
/third_party/skia/third_party/externals/oboe/src/common/ |
H A D | AudioStream.cpp | 190 FrameTimestamp frame; in getTimestamp() local 191 Result result = getTimestamp(clockId, &frame.position, &frame.timestamp); in getTimestamp() 193 return ResultWithValue<FrameTimestamp>(frame); in getTimestamp()
|
/third_party/python/Lib/test/ |
H A D | test_threadsignals.py | 29 def handle_signals(sig,frame): 79 def alarm_interrupt(self, sig, frame): 153 def my_handler(signal, frame): 204 def my_handler(signum, frame):
|
/drivers/peripheral/audio/test/hal_mock/ |
H A D | audio_render_mock.c | 33 static int32_t RenderFrame(struct IAudioRenderVdi* render, const int8_t *frame, uint32_t frameLen, uint64_t *replyBytes)
in RenderFrame() argument 35 if (render == NULL || frame == NULL || replyBytes == NULL) {
in RenderFrame() 39 (void *)frame;
in RenderFrame() 47 if (render == NULL || frame == NULL || time == NULL) {
in GetPosition() 327 if (render == NULL || frame == NULL) {
in GetMmapPosition()
|
/drivers/peripheral/distributed_audio/hdi_service/audio/v1_0/src/ |
H A D | audio_capture_interface_impl.cpp | 69 int32_t AudioCaptureInterfaceImpl::CaptureFrame(std::vector<int8_t> &frame, uint64_t &replyBytes) in CaptureFrame() argument 71 DHLOGD("Capture frame[sampleRate:%{public}d, channelCount: %{public}d, format: %{public}d, frameSize: %{public}d].", in CaptureFrame() 93 frame.clear(); in CaptureFrame() 94 frame.resize(devAttrs_.frameSize, 0); in CaptureFrame() 95 if (!muteState_.load() && memcpy_s(frame.data(), frame.size(), audioData.data.data(), audioData.data.size()) != in CaptureFrame() 97 DHLOGE("Copy capture frame failed"); in CaptureFrame() 102 DHLOGD("Capture audio frame success."); in CaptureFrame() 105 DHLOGD("This time capture frame spend: %" PRId64" us, The interval of this time and the last time: %" PRId64 in CaptureFrame() 364 DHLOGI("Get frame buffe in GetFrameBufferSize() [all...] |
/third_party/ffmpeg/libavdevice/ |
H A D | libdc1394.c | 40 dc1394video_frame_t *frame; member 82 { "video_size", "A string describing frame size, such as 640x480 or hd720.", OFFSET(video_size), AV_OPT_TYPE_STRING, {.str = "qvga"}, 0, 0, DEC }, 258 /* discard stale frame */ in dc1394_read_packet() 260 if (dc1394_capture_enqueue(dc1394->camera, dc1394->frame) != DC1394_SUCCESS) in dc1394_read_packet() 261 av_log(c, AV_LOG_ERROR, "failed to release %d frame\n", dc1394->current_frame); in dc1394_read_packet() 264 res = dc1394_capture_dequeue(dc1394->camera, DC1394_CAPTURE_POLICY_WAIT, &dc1394->frame); in dc1394_read_packet() 266 pkt->data = (uint8_t *)dc1394->frame->image; in dc1394_read_packet() 267 pkt->size = dc1394->frame->image_bytes; in dc1394_read_packet()
|
/third_party/skia/gm/ |
H A D | animated_gif.cpp | 67 SkDebugf("Could not getPixels for frame %i: %s", frameIndex, FLAGS_animatedGif[0]); in drawFrame() 150 // I'm assuming this gets called *after* onOnceBeforeDraw, so our first frame should 163 // Retrieve the next frame. 169 // Note that we loop here. This is not safe if we need to draw the intermediate frame 249 // We'll draw one of each frame, so make it big enough to hold them all in init() 280 for (int frame = 0; duration < fPlayer->duration(); frame++) { 282 const int xTranslate = (frame % factor) * dimensions.width(); 283 const int yTranslate = (frame / factor) * dimensions.height(); 289 duration += fFrameInfos[frame] [all...] |
/third_party/skia/modules/particles/src/ |
H A D | SkParticleDrawable.cpp | 135 int frame = static_cast<int>(spriteFrames[i] * frameCount + 0.5f); variable 136 frame = SkTPin(frame, 0, frameCount - 1); 137 int row = frame / cols; 138 int col = frame % cols;
|
/third_party/skia/experimental/docs/ |
H A D | animationCommon.js | 18 function addActions(frame, timeline) { 19 var keyframe = keyframes[frame]; 236 function keyframeInit(frame) { 239 addActions(frame, animationState.timeline); 250 keyframes[frame][0]);
|