/external/python/cpython2/Demo/tkinter/ttk/ |
H A D | roundframe.py | 98 frame2 = ttk.Frame(style="RoundedFrame", padding=10) variable 99 frame2.pack(fill='both', expand=1) 106 text = Tkinter.Text(frame2, borderwidth=0, bg="white", highlightthickness=0) 108 text.bind("<FocusIn>", lambda evt: frame2.state(["focus"])) 109 text.bind("<FocusOut>", lambda evt: frame2.state(["!focus"]))
|
/external/valgrind/memcheck/tests/ |
H A D | xml1.c | 37 int frame2 ( void ) function 44 return frame2() + 1;
|
/external/webrtc/talk/media/webrtc/ |
H A D | webrtcvideoframe_unittest.cc | 318 cricket::VideoFrame* frame2 = frame1.Copy(); local 319 EXPECT_EQ(frame1.GetNativeHandle(), frame2->GetNativeHandle()); 320 EXPECT_EQ(frame1.GetWidth(), frame2->GetWidth()); 321 EXPECT_EQ(frame1.GetHeight(), frame2->GetHeight()); 322 EXPECT_EQ(frame1.GetTimeStamp(), frame2->GetTimeStamp()); 323 delete frame2;
|
/external/webrtc/webrtc/modules/utility/source/ |
H A D | audio_frame_operations_unittest.cc | 43 void VerifyFramesAreEqual(const AudioFrame& frame1, const AudioFrame& frame2) { argument 44 EXPECT_EQ(frame1.num_channels_, frame2.num_channels_); 46 frame2.samples_per_channel_); 50 EXPECT_EQ(frame1.data_[i], frame2.data_[i]);
|
/external/webrtc/webrtc/common_video/ |
H A D | i420_video_frame_unittest.cc | 37 VideoFrame frame2; local 38 frame2.CopyFrame(frame); 143 VideoFrame frame2; local 144 frame2.ShallowCopy(frame1); 148 const VideoFrame* const_frame2_ptr = &frame2; 157 EXPECT_EQ(frame2.timestamp(), frame1.timestamp()); 158 EXPECT_EQ(frame2.ntp_time_ms(), frame1.ntp_time_ms()); 159 EXPECT_EQ(frame2.render_time_ms(), frame1.render_time_ms()); 160 EXPECT_EQ(frame2.rotation(), frame1.rotation()); 162 frame2 189 VideoFrame frame1, frame2; local [all...] |
/external/google-breakpad/src/processor/ |
H A D | stackwalker_amd64_unittest.cc | 296 StackFrameAMD64 *frame2 = static_cast<StackFrameAMD64 *>(frames->at(2)); local 297 EXPECT_EQ(StackFrame::FRAME_TRUST_SCAN, frame2->trust); 300 frame2->context_validity); 301 EXPECT_EQ(return_address2, frame2->context.rip); 302 EXPECT_EQ(frame2_sp.Value(), frame2->context.rsp);
|
H A D | stackwalker_arm64_unittest.cc | 253 StackFrameARM64 *frame2 = static_cast<StackFrameARM64 *>(frames->at(2)); local 254 EXPECT_EQ(StackFrame::FRAME_TRUST_SCAN, frame2->trust); 257 frame2->context_validity); 258 EXPECT_EQ(return_address2, frame2->context.iregs[MD_CONTEXT_ARM64_REG_PC]); 259 EXPECT_EQ(frame2_sp.Value(), frame2->context.iregs[MD_CONTEXT_ARM64_REG_SP]); 524 StackFrameARM64 *frame2 = static_cast<StackFrameARM64 *>(frames->at(2)); local 525 EXPECT_EQ(StackFrame::FRAME_TRUST_FP, frame2->trust); 530 frame2->context_validity); 531 EXPECT_EQ(return_address2, frame2->context.iregs[MD_CONTEXT_ARM64_REG_PC]); 532 EXPECT_EQ(0U, frame2 [all...] |
H A D | stackwalker_arm_unittest.cc | 275 StackFrameARM *frame2 = static_cast<StackFrameARM *>(frames->at(2)); local 276 EXPECT_EQ(StackFrame::FRAME_TRUST_SCAN, frame2->trust); 279 frame2->context_validity); 280 EXPECT_EQ(return_address2, frame2->context.iregs[MD_CONTEXT_ARM_REG_PC]); 281 EXPECT_EQ(frame2_sp.Value(), frame2->context.iregs[MD_CONTEXT_ARM_REG_SP]); 864 StackFrameARM *frame2 = static_cast<StackFrameARM *>(frames->at(2)); local 865 EXPECT_EQ(StackFrame::FRAME_TRUST_FP, frame2->trust); 870 frame2->context_validity); 871 EXPECT_EQ(return_address2, frame2->context.iregs[MD_CONTEXT_ARM_REG_PC]); 872 EXPECT_EQ(0U, frame2 963 StackFrameARM *frame2 = static_cast<StackFrameARM *>(frames->at(2)); local [all...] |
H A D | stackwalker_mips_unittest.cc | 289 StackFrameMIPS* frame2 = static_cast<StackFrameMIPS*>(frames->at(2)); local 290 EXPECT_EQ(StackFrame::FRAME_TRUST_SCAN, frame2->trust); 295 frame2->context_validity); 296 EXPECT_EQ(return_address2 - 2 * sizeof(return_address2), frame2->context.epc); 297 EXPECT_EQ(frame2_sp.Value(), frame2->context.iregs[MD_CONTEXT_MIPS_REG_SP]);
|
H A D | stackwalker_x86_unittest.cc | 739 StackFrameX86 *frame2 = static_cast<StackFrameX86 *>(frames->at(2)); local 740 EXPECT_EQ(StackFrame::FRAME_TRUST_CFI, frame2->trust); 745 frame2->context_validity); 746 EXPECT_EQ(0x2a179e38U, frame2->instruction + 1); 747 EXPECT_EQ(0x2a179e38U, frame2->context.eip); 748 EXPECT_EQ(frame2_esp.Value(), frame2->context.esp); 749 EXPECT_EQ(frame2_ebp.Value(), frame2->context.ebp); 750 EXPECT_EQ(0x2558c7f3U, frame2->context.ebx); 751 EXPECT_EQ(NULL, frame2->module); 752 EXPECT_EQ(NULL, frame2 1443 StackFrameX86 *frame2 = static_cast<StackFrameX86 *>(frames->at(2)); local 1881 StackFrameX86 *frame2 = static_cast<StackFrameX86 *>(frames->at(2)); local [all...] |
/external/skia/tests/ |
H A D | PathOpsConicIntersectionTest.cpp | 195 const SkDConic frame2[] = { variable 231 frame0, frame1, frame2, frame3, frame4, frame5, frame6 235 (int) SK_ARRAY_COUNT(frame2), (int) SK_ARRAY_COUNT(frame3),
|
/external/webrtc/webrtc/modules/video_processing/test/ |
H A D | video_processing_unittest.cc | 54 const webrtc::VideoFrame& frame2); 381 const webrtc::VideoFrame& frame2) { 385 int allocated_size2 = frame2.allocated_size(plane_type); 389 const uint8_t* plane_buffer2 = frame2.buffer(plane_type); 380 CompareFrames(const webrtc::VideoFrame& frame1, const webrtc::VideoFrame& frame2) argument
|
/external/webrtc/webrtc/video/ |
H A D | video_capture_input_unittest.cc | 42 bool EqualFrames(const VideoFrame& frame1, const VideoFrame& frame2); 43 bool EqualTextureFrames(const VideoFrame& frame1, const VideoFrame& frame2); 44 bool EqualBufferFrames(const VideoFrame& frame1, const VideoFrame& frame2); 250 bool EqualFrames(const VideoFrame& frame1, const VideoFrame& frame2) { argument 251 if (frame1.native_handle() != NULL || frame2.native_handle() != NULL) 252 return EqualTextureFrames(frame1, frame2); 253 return EqualBufferFrames(frame1, frame2); 256 bool EqualTextureFrames(const VideoFrame& frame1, const VideoFrame& frame2) { argument 257 return ((frame1.native_handle() == frame2.native_handle()) && 258 (frame1.width() == frame2 263 EqualBufferFrames(const VideoFrame& frame1, const VideoFrame& frame2) argument [all...] |
H A D | video_send_stream_tests.cc | 47 void ExpectEqualFrames(const VideoFrame& frame1, const VideoFrame& frame2); 49 const VideoFrame& frame2); 51 const VideoFrame& frame2); 1191 void ExpectEqualFrames(const VideoFrame& frame1, const VideoFrame& frame2) { argument 1192 if (frame1.native_handle() != nullptr || frame2.native_handle() != nullptr) 1193 ExpectEqualTextureFrames(frame1, frame2); 1195 ExpectEqualBufferFrames(frame1, frame2); 1199 const VideoFrame& frame2) { 1200 EXPECT_EQ(frame1.native_handle(), frame2.native_handle()); 1201 EXPECT_EQ(frame1.width(), frame2 1198 ExpectEqualTextureFrames(const VideoFrame& frame1, const VideoFrame& frame2) argument 1206 ExpectEqualBufferFrames(const VideoFrame& frame1, const VideoFrame& frame2) argument [all...] |
/external/libvpx/libvpx/vp8/encoder/mips/msa/ |
H A D | temporal_filter_msa.c | 129 v16i8 frame2 = { 0 }; local 161 INSERT_D2_SB(f2, f3, frame2); 164 ILVRL_B2_UB(frame1, frame2, frame_l, frame_h); 190 UNPCK_UB_SH(frame2, frame2_0_h, frame2_1_h); 244 uint8_t *frame2, uint32_t block_size, 248 temporal_filter_apply_8size_msa(frame1, stride, frame2, strength, 251 temporal_filter_apply_16size_msa(frame1, stride, frame2, strength, 262 int pixel_value = *frame2++; 243 vp8_temporal_filter_apply_msa(uint8_t *frame1, uint32_t stride, uint8_t *frame2, uint32_t block_size, int32_t strength, int32_t filter_weight, uint32_t *accumulator, uint16_t *count) argument
|
/external/libvpx/libvpx/vp8/encoder/ |
H A D | temporal_filter.c | 70 unsigned char *frame2, unsigned int block_size, 82 int pixel_value = *frame2++; 69 vp8_temporal_filter_apply_c(unsigned char *frame1, unsigned int stride, unsigned char *frame2, unsigned int block_size, int strength, int filter_weight, unsigned int *accumulator, unsigned short *count) argument
|
/external/webrtc/webrtc/modules/video_capture/test/ |
H A D | video_capture_unittest.cc | 66 const webrtc::VideoFrame& frame2) { 68 (frame1.stride(webrtc::kYPlane) == frame2.stride(webrtc::kYPlane)) && 69 (frame1.stride(webrtc::kUPlane) == frame2.stride(webrtc::kUPlane)) && 70 (frame1.stride(webrtc::kVPlane) == frame2.stride(webrtc::kVPlane)) && 71 (frame1.width() == frame2.width()) && 72 (frame1.height() == frame2.height()); 79 int allocated_size2 = frame2.allocated_size(plane_type); 83 const uint8_t* plane_buffer2 = frame2.buffer(plane_type); 65 CompareFrames(const webrtc::VideoFrame& frame1, const webrtc::VideoFrame& frame2) argument
|
/external/deqp/modules/egl/ |
H A D | teglGetFrameTimestampsTests.cpp | 257 void verifyNeighboringFrames (const FrameTimes& frame1, const FrameTimes& frame2, tcu::ResultCollector& result, bool verifyReadsDone) argument 260 check_lt(result, frame1.swapBufferBeginNs, frame2.swapBufferBeginNs, "Swap begin times not monotonic."); 261 check_lt(result, frame1.latch, frame2.latch, "Latch times not monotonic."); 262 check_lt(result, frame1.lastCompositionStart, frame2.latch, "Old buffer composited after new buffer latched."); 263 check_lt(result, frame1.lastCompositionStart, frame2.firstCompositionStart, "Composition times overlap."); 264 check_lt(result, frame1.dequeueReady, frame2.dequeueReady, "Dequeue ready times not monotonic."); 267 check_lt(result, frame1.renderingComplete, frame2.renderingComplete, "Rendering complete times not monotonic."); 269 if (timestampValid(frame1.firstCompositionGpuFinished) && timestampValid(frame2.firstCompositionGpuFinished)) 270 check_lt(result, frame1.firstCompositionGpuFinished, frame2.firstCompositionGpuFinished, "Composition GPU work complete times not monotonic."); 272 if (timestampValid(frame1.displayPresent) && timestampValid(frame2 [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
H A D | vp9_temporal_filter.c | 98 const uint8_t *frame2, 116 int pixel_value = *frame2; 130 frame2[idy * (int)block_width + idx]; 145 ++frame2; 171 const uint16_t *frame2 = CONVERT_TO_SHORTPTR(frame2_8); local 179 int pixel_value = *frame2; 191 frame2[idy * (int)block_width + idx]; 205 ++frame2; 97 vp9_temporal_filter_apply_c(const uint8_t *frame1, unsigned int stride, const uint8_t *frame2, unsigned int block_width, unsigned int block_height, int strength, int filter_weight, uint32_t *accumulator, uint16_t *count) argument
|
/external/valgrind/exp-sgcheck/ |
H A D | sg_main.c | 2440 StackFrame *frame, *frame2; local 2452 frame2 = frame->inner; 2453 if (frame2) tl_assert(1 + frame->depth == frame2->depth); 2455 frame = frame2;
|
/external/webrtc/talk/media/base/ |
H A D | videoframe_unittest.h | 495 const cricket::VideoFrame& frame2, 498 frame2.GetWidth(), frame2.GetHeight(), 499 frame2.GetPixelWidth(), frame2.GetPixelHeight(), 500 frame2.GetTimeStamp(), 501 frame2.GetYPlane(), frame2.GetYPitch(), 502 frame2.GetUPlane(), frame2 494 IsEqual(const cricket::VideoFrame& frame1, const cricket::VideoFrame& frame2, int max_error) argument 507 IsEqualWithCrop(const cricket::VideoFrame& frame1, const cricket::VideoFrame& frame2, int hcrop, int vcrop, int max_error) argument 573 T frame1, frame2; local 594 T frame1, frame2; local 611 T frame1, frame2; local 629 T frame1, frame2; local 643 T frame1, frame2; local 686 T frame1, frame2; local 700 T frame1, frame2; local 714 T frame1, frame2; local 728 T frame1, frame2; local 742 T frame1, frame2; local 756 T frame1, frame2; local 769 T frame1, frame2; local 785 T frame1, frame2; local 801 T frame1, frame2; local 903 T frame2; local 913 T frame2; local 924 T frame2; local 935 T frame2; local 945 T frame2; local 956 T frame2; local 1058 T frame1, frame2; local 1068 T frame1, frame2; local 1082 T frame1, frame2; local 1096 T frame1, frame2; local 1106 T frame1, frame2, frame3; local 1119 T frame1, frame2; local 1128 T frame1, frame2; local 1137 T frame1, frame2; local 1146 T frame1, frame2; local 1156 T frame1, frame2; local 1289 T frame1, frame2, frame3, frame4; local 1308 T frame1, frame2, frame3, frame4; local 1327 T frame1, frame2; local 1337 T frame1, frame2; local 1361 T frame1, frame2; local 1379 T frame1, frame2; local 1412 T frame1, frame2; local 1475 T frame1, frame2; local 1797 T frame1, frame2; local [all...] |
/external/webrtc/webrtc/modules/audio_processing/test/ |
H A D | audio_processing_unittest.cc | 151 bool FrameDataAreEqual(const AudioFrame& frame1, const AudioFrame& frame2) { argument 152 if (frame1.samples_per_channel_ != frame2.samples_per_channel_) { 155 if (frame1.num_channels_ != frame2.num_channels_) { 158 if (memcmp(frame1.data_, frame2.data_,
|
/external/owasp/sanitizer/tools/findbugs/lib/ |
H A D | findbugs.jar | META-INF/ META-INF/MANIFEST.MF default.xsl edu/ edu/umd/ edu/umd/cs/ edu/ ... |
/external/icu/tools/srcgen/currysrc/libs/ |
H A D | org.eclipse.jdt.core_3.11.0.v20150602-1242.jar | META-INF/MANIFEST.MF META-INF/ECLIPSE_.SF META-INF/ECLIPSE_.RSA META ... |
/external/dagger2/lib/ |
H A D | google-java-format-0.1-20151017.042846-2.jar | META-INF/ META-INF/MANIFEST.MF com/ com/google/ com/google/googlejavaformat/ com/google/googlejavaformat/CloseOp ... |