/external/v8/src/inspector/ |
H A D | v8-stack-trace-impl.cc | 24 V8StackTraceImpl::Frame toFrame(v8::Local<v8::StackFrame> frame, argument 27 String16 scriptId = String16::fromInteger(frame->GetScriptId()); 29 v8::Local<v8::String> sourceNameValue(frame->GetScriptNameOrSourceURL()); 34 v8::Local<v8::String> functionNameValue(frame->GetFunctionName()); 38 int sourceLineNumber = frame->GetLineNumber() - 1; 39 int sourceColumn = frame->GetColumn() - 1; 290 const Frame& frame = m_frames[i]; local 291 stackTrace.append("\n at " + (frame.functionName().length() 292 ? frame.functionName() 295 stackTrace.append(frame [all...] |
/external/v8/src/profiler/ |
H A D | sampling-heap-profiler.cc | 162 JavaScriptFrame* frame = it.frame(); local 163 SharedFunctionInfo* shared = frame->function()->shared();
|
/external/v8/src/runtime/ |
H A D | runtime-compiler.cc | 152 JavaScriptFrame* frame = it->frame(); local 153 if (code_->contains(frame->pc())) has_code_activations_ = true; 183 JavaScriptFrame* top_frame = top_it.frame(); 199 JavaScriptFrame* top_frame = top_it.frame(); 208 // At this point {it} is at the topmost frame of all the frames materialized 209 // by the deoptimizer. Note that this frame does not necessarily represent 249 JavaScriptFrame* frame = it.frame(); local 250 if (frame 258 DetermineEntryAndDisarmOSRForBaseline(JavaScriptFrame* frame) argument 282 DetermineEntryAndDisarmOSRForInterpreter(JavaScriptFrame* frame) argument 318 JavaScriptFrame* frame = it.frame(); local [all...] |
/external/valgrind/coregrind/m_sigframe/ |
H A D | sigframe-amd64-linux.c | 73 /*--- Signal frame layouts ---*/ 81 // frame is. Unfortunately, things like gcc's exception unwinding 82 // make assumptions about the locations of various parts of the frame, 85 /* Valgrind-specific parts of the signal frame */ 375 /* Build the Valgrind-specific part of a signal frame. */ 377 static void build_vg_sigframe(struct vg_sigframe *frame, argument 383 frame->sigNo_private = sigNo; 384 frame->magicPI = 0x31415927; 385 frame->vex_shadow1 = tst->arch.vex_shadow1; 386 frame 404 struct rt_sigframe *frame; local 464 struct rt_sigframe *frame; local 508 restore_vg_sigframe( ThreadState *tst, struct vg_sigframe *frame, Int *sigNo ) argument 565 restore_rt_sigframe( ThreadState *tst, struct rt_sigframe *frame, Int *sigNo ) argument [all...] |
H A D | sigframe-arm-linux.c | 56 shadows in the frame, and restoring it afterwards from there, 124 struct sigframe *frame, 134 struct vg_sig_private *priv = &frame->vp; 136 VG_TRACK( pre_mem_write, Vg_CoreSignal, tst->tid, "signal handler frame", 137 (Addr)frame, offsetof(struct sigframe, vp)); 147 synth_ucontext(tst->tid, siginfo, trapno, err, mask, &frame->uc); 150 (Addr)frame, offsetof(struct sigframe, vp)); 259 struct rt_sigframe *frame = (struct rt_sigframe *)sp; local 260 frame_size = sizeof(*frame); 261 priv = &frame 123 build_sigframe(ThreadState *tst, struct sigframe *frame, const vki_siginfo_t *siginfo, const struct vki_ucontext *siguc, void *handler, UInt flags, const vki_sigset_t *mask, void *restorer) argument 265 struct sigframe *frame = (struct sigframe *)sp; local [all...] |
H A D | sigframe-arm64-linux.c | 51 shadows in the frame, and restoring it afterwards from there, 117 struct sigframe *frame, 127 struct vg_sig_private *priv = &frame->vp; 129 VG_TRACK( pre_mem_write, Vg_CoreSignal, tst->tid, "signal handler frame", 130 (Addr)frame, offsetof(struct sigframe, vp)); 140 synth_ucontext(tst->tid, siginfo, trapno, err, mask, &frame->uc); 143 (Addr)frame, offsetof(struct sigframe, vp)); 235 struct rt_sigframe *frame = (struct rt_sigframe *)sp; local 236 frame_size = sizeof(*frame); 237 //mc = &frame 116 build_sigframe(ThreadState *tst, struct sigframe *frame, const vki_siginfo_t *siginfo, const struct vki_ucontext *siguc, void *handler, UInt flags, const vki_sigset_t *mask, void *restorer) argument [all...] |
H A D | sigframe-s390x-linux.c | 72 /*--- Signal frame layouts ---*/ 78 // Linux has 2 signal frame structures: one for normal signal 84 // frame is. Unfortunately, things like gcc's exception unwinding 85 // make assumptions about the locations of various parts of the frame, 88 /* Valgrind-specific parts of the signal frame */ 262 /* Build the Valgrind-specific part of a signal frame. */ 264 static void build_vg_sigframe(struct vg_sigframe *frame, argument 269 frame->sigNo_private = sigNo; 270 frame->magicPI = 0x31415927; 271 frame 290 struct sigframe *frame; local 349 struct rt_sigframe *frame; local 447 restore_vg_sigframe( ThreadState *tst, struct vg_sigframe *frame, Int *sigNo ) argument 472 restore_sigframe( ThreadState *tst, struct sigframe *frame, Int *sigNo ) argument 482 restore_rt_sigframe( ThreadState *tst, struct rt_sigframe *frame, Int *sigNo ) argument [all...] |
H A D | sigframe-x86-linux.c | 74 /*--- Signal frame layouts ---*/ 80 // Linux has 2 signal frame structures: one for normal signal 86 // frame is. Unfortunately, things like gcc's exception unwinding 87 // make assumptions about the locations of various parts of the frame, 90 /* Valgrind-specific parts of the signal frame */ 395 /* Build the Valgrind-specific part of a signal frame. */ 397 static void build_vg_sigframe(struct vg_sigframe *frame, argument 402 frame->sigNo_private = sigNo; 403 frame->magicPI = 0x31415927; 404 frame 423 struct sigframe *frame; local 481 struct rt_sigframe *frame; local 581 restore_vg_sigframe( ThreadState *tst, struct vg_sigframe *frame, Int *sigNo ) argument 632 restore_sigframe( ThreadState *tst, struct sigframe *frame, Int *sigNo ) argument 642 restore_rt_sigframe( ThreadState *tst, struct rt_sigframe *frame, Int *sigNo ) argument [all...] |
/external/webrtc/talk/media/base/ |
H A D | fakevideorenderer.h | 59 virtual bool RenderFrame(const VideoFrame* frame) { argument 63 black_frame_ = CheckFrameColorYuv(6, 48, 128, 128, 128, 128, frame); 64 // Treat unexpected frame size as error. 65 if (!frame || 66 frame->GetWidth() != static_cast<size_t>(width_) || 67 frame->GetHeight() != static_cast<size_t>(height_)) { 68 if (!frame) { 69 LOG(LS_WARNING) << "RenderFrame expected non-null frame."; 71 LOG(LS_WARNING) << "RenderFrame expected frame of size " << width_ 72 << "x" << height_ << " but received frame o 109 CheckFrameColorYuv(uint8_t y_min, uint8_t y_max, uint8_t u_min, uint8_t u_max, uint8_t v_min, uint8_t v_max, const cricket::VideoFrame* frame) argument [all...] |
/external/webrtc/talk/media/devices/ |
H A D | filevideocapturer.cc | 57 bool VideoRecorder::RecordFrame(const CapturedFrame& frame) { argument 64 if (!frame.GetDataSize(&size)) { 65 LOG(LS_ERROR) << "Unable to calculate the data size of the frame"; 70 // Convert the frame header to bytebuffer. 72 buffer.WriteUInt32(frame.width); 73 buffer.WriteUInt32(frame.height); 74 buffer.WriteUInt32(frame.fourcc); 75 buffer.WriteUInt32(frame.pixel_width); 76 buffer.WriteUInt32(frame.pixel_height); 80 buffer.WriteUInt64(frame 193 CapturedFrame frame; local 279 ReadFrameHeader( CapturedFrame* frame) argument [all...] |
H A D | gdivideorenderer.cc | 56 // Called when a new frame is available. Upon this call, we send 60 bool RenderFrame(const VideoFrame* frame); 100 void OnRenderFrame(const VideoFrame* frame); 154 const VideoFrame* frame = video_frame->GetCopyWithRotationApplied(); local 156 if (!SetSize(static_cast<int>(frame->GetWidth()), 157 static_cast<int>(frame->GetHeight()))) { 161 SendMessage(handle(), kRenderFrameMsg, reinterpret_cast<WPARAM>(frame), 0); 244 void GdiVideoRenderer::VideoWindow::OnRenderFrame(const VideoFrame* frame) { argument 245 if (!frame) { 248 // Convert frame t 271 RenderFrame(const VideoFrame* frame) argument [all...] |
/external/webrtc/talk/media/webrtc/ |
H A D | webrtcvideocapturer.cc | 402 const webrtc::VideoFrame& frame) { 409 // Log the size and pixel aspect ratio of the first captured frame. 411 LOG(LS_INFO) << "Captured frame size " 412 << frame.width() << "x" << frame.height() 416 // Signal down stream components on captured frame. 420 webrtc::CalcBufferSize(webrtc::kI420, frame.width(), frame.height()); 423 // take over ownership of the buffer held by |frame| if that's possible. 424 webrtc::ExtractBuffer(frame, lengt 401 SignalFrameCapturedOnStartThread( const webrtc::VideoFrame& frame) argument [all...] |
H A D | webrtcvideoframe.cc | 77 bool WebRtcVideoFrame::Init(const CapturedFrame* frame, int dw, int dh, argument 79 return Reset(frame->fourcc, frame->width, frame->height, dw, dh, 80 static_cast<uint8_t*>(frame->data), frame->data_size, 81 frame->pixel_width, frame->pixel_height, frame->time_stamp, 82 frame 252 WebRtcVideoFrame* frame = new WebRtcVideoFrame(); local [all...] |
/external/webrtc/webrtc/common_video/libyuv/ |
H A D | libyuv_unittest.cc | 40 int PrintFrame(const VideoFrame* frame, const char* str) { argument 41 if (frame == NULL) 43 printf("%s %dx%d \n", str, frame->width(), frame->height()); 48 int width = (plane_num ? (frame->width() + 1) / 2 : frame->width()); 49 int height = (plane_num ? (frame->height() + 1) / 2 : frame->height()); 50 ret += PrintBuffer(frame->buffer(plane_type), width, height, 51 frame 59 CreateImage(VideoFrame* frame, int plane_offset[kNumOfPlanes]) argument [all...] |
H A D | webrtc_libyuv.cc | 105 int PrintVideoFrame(const VideoFrame& frame, FILE* file) { argument 108 if (frame.IsZeroSize()) 111 int width = (planeNum ? (frame.width() + 1) / 2 : frame.width()); 112 int height = (planeNum ? (frame.height() + 1) / 2 : frame.height()); 114 const uint8_t* plane_buffer = frame.buffer(plane_type); 120 plane_buffer += frame.stride(plane_type); 294 // Compute PSNR for an I420 frame (all planes) 318 // In case of 0 mse in one frame, 12 [all...] |
/external/webrtc/webrtc/common_video/ |
H A D | video_frame.cc | 71 // Creating empty frame - reset all values. 219 VideoFrame frame; local 220 frame.ShallowCopy(*this); 221 frame.set_video_frame_buffer(video_frame_buffer_->NativeToI420Buffer()); 222 return frame; 225 bool VideoFrame::EqualsFrame(const VideoFrame& frame) const { 226 if (width() != frame.width() || height() != frame.height() || 227 stride(kYPlane) != frame.stride(kYPlane) || 228 stride(kUPlane) != frame [all...] |
/external/webrtc/webrtc/modules/audio_coding/acm2/ |
H A D | acm_receiver_unittest_oldapi.cc | 99 AudioFrame frame; local 101 frame.sample_rate_hz_ = codec.plfreq; 102 frame.samples_per_channel_ = codec.plfreq / 100; // 10 ms. 103 frame.num_channels_ = codec.channels; 104 memset(frame.data_, 0, frame.samples_per_channel_ * frame.num_channels_ * 109 frame.timestamp_ = timestamp_; 110 timestamp_ += frame.samples_per_channel_; 111 ASSERT_GE(acm_->Add10MsData(frame), 279 AudioFrame frame; local 306 AudioFrame frame; local [all...] |
/external/webrtc/webrtc/modules/audio_coding/test/ |
H A D | target_delay_unittest.cc | 136 // payload-len = frame-samples * 2 bytes/sample. 151 AudioFrame frame; local 152 for (int k = 0; k < kNum10msPerFrame; ++k) { // Pull one frame. 153 ASSERT_EQ(0, acm_->PlayoutData10Ms(-1, &frame)); 155 ASSERT_TRUE(kSampleRateHz == frame.sample_rate_hz_); 156 ASSERT_EQ(1u, frame.num_channels_); 157 ASSERT_TRUE(kSampleRateHz / 100 == frame.samples_per_channel_);
|
/external/webrtc/webrtc/modules/audio_processing/ |
H A D | audio_buffer.cc | 370 void AudioBuffer::DeinterleaveFrom(AudioFrame* frame) { argument 371 assert(frame->num_channels_ == num_input_channels_); 372 assert(frame->samples_per_channel_ == input_num_frames_); 379 activity_ = frame->vad_activity_; 389 DownmixInterleavedToMono(frame->data_, input_num_frames_, 393 Deinterleave(frame->data_, 410 void AudioBuffer::InterleaveTo(AudioFrame* frame, bool data_changed) { argument 411 frame->vad_activity_ = activity_; 416 assert(frame->num_channels_ == num_channels_ || num_channels_ == 1); 417 assert(frame [all...] |
/external/webrtc/webrtc/modules/audio_processing/vad/ |
H A D | vad_audio_proc.cc | 78 int VadAudioProc::ExtractFeatures(const int16_t* frame, argument 89 if (high_pass_filter_->Filter(frame, kNumSubframeSamples, 137 // each 10ms sub-frame. This is equivalent to computing LPC coefficients for the
|
/external/webrtc/webrtc/modules/desktop_capture/ |
H A D | desktop_and_cursor_composer_unittest.cc | 45 uint32_t GetFramePixel(const DesktopFrame& frame, const DesktopVector& pos) { argument 46 return *reinterpret_cast<uint32_t*>(frame.data() + pos.y() * frame.stride() + 62 DesktopFrame* frame = local 64 uint32_t* data = reinterpret_cast<uint32_t*>(frame->data()); 70 return frame; 139 void VerifyFrame(const DesktopFrame& frame, argument 153 GetFramePixel(frame, p)); 155 EXPECT_EQ(GetFakeFramePixelValue(p), GetFramePixel(frame, p)); 173 void OnCaptureCompleted(DesktopFrame* frame) overrid [all...] |
H A D | screen_capturer_x11.cc | 35 // A class to perform video frame capturing for Linux. 252 // If the current frame is from an older generation then allocate a new one. 256 rtc::scoped_ptr<DesktopFrame> frame( 258 queue_.ReplaceCurrentFrame(frame.release()); 262 DesktopFrame* frame = queue_.current_frame(); local 265 (differ_->width() != frame->size().width()) || 266 (differ_->height() != frame->size().height()) || 267 (differ_->bytes_per_row() != frame->stride()))) { 268 differ_.reset(new Differ(frame->size().width(), frame 310 DesktopFrame* frame = queue_.current_frame()->Share(); local [all...] |
/external/webrtc/webrtc/modules/desktop_capture/win/ |
H A D | screen_capturer_win_magnifier.cc | 152 // No previous frame is available, or the screen is resized. Invalidate the 159 // Emit the current frame. 160 DesktopFrame* frame = queue_.current_frame()->Share(); local 161 frame->set_dpi(DesktopVector(GetDeviceCaps(desktop_dc_, LOGPIXELSX), 163 frame->mutable_updated_region()->Clear(); 164 helper_.TakeInvalidRegion(frame->mutable_updated_region()); 165 frame->set_capture_time_ms( 167 callback_->OnCaptureCompleted(frame); 218 // frame before set_window_source_func_ returns. 410 // Copy the data into the frame [all...] |
/external/webrtc/webrtc/modules/desktop_capture/x11/ |
H A D | x_server_pixel_buffer.cc | 237 DesktopFrame* frame) { 262 FastBlit(data, rect, frame); 264 SlowBlit(data, rect, frame); 270 DesktopFrame* frame) { 275 uint8_t* dst_pos = frame->data() + frame->stride() * dst_y; 283 dst_pos += frame->stride(); 289 DesktopFrame* frame) { 304 uint8_t* dst_pos = frame->data() + frame 236 CaptureRect(const DesktopRect& rect, DesktopFrame* frame) argument 268 FastBlit(uint8_t* image, const DesktopRect& rect, DesktopFrame* frame) argument 287 SlowBlit(uint8_t* image, const DesktopRect& rect, DesktopFrame* frame) argument [all...] |
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/test/ |
H A D | vp8_impl_unittest.cc | 32 Vp8UnitTestEncodeCompleteCallback(EncodedImage* frame, argument 35 : encoded_frame_(frame), encode_complete_(false) {} 79 explicit Vp8UnitTestDecodeCompleteCallback(VideoFrame* frame) argument 80 : decoded_frame_(frame), decode_complete(false) {} 81 int32_t Decoded(VideoFrame& frame) override; 82 int32_t Decoded(VideoFrame& frame, int64_t decode_time_ms) override { 120 // Processing only one frame. 125 // Set input frame. 232 // First frame should be a key frame [all...] |