/external/chromium_org/third_party/libjingle/source/talk/app/webrtc/ |
H A D | mediastreamsignaling_unittest.cc | 65 // id "audio_1" and a video track with id "video_1; 77 "m=video 1 RTP/AVPF 120\r\n" 78 "a=mid:video\r\n" 85 // MediaStreams have one audio track and one video track. 100 "m=video 1 RTP/AVPF 120\r\n" 101 "a=mid:video\r\n" 117 "m=video 1 RTP/AVPF 120\r\n" 118 "a=mid:video\r\n" 131 "m=video 1 RTP/AVPF 120\r\n" 132 "a=mid:video\ 200 webrtc::VideoTrackInterface* video = video_tracks[j]; local [all...] |
H A D | peerconnection_unittest.cc | 141 virtual void Negotiate(bool audio, bool video) = 0; 148 void AddMediaStream(bool audio, bool video) { argument 169 if (video && can_receive_video()) { 532 // This is consistent with the video pipeline that us a a separate thread for 580 virtual void Negotiate(bool audio, bool video) { argument 587 if (offer->description()->GetContentByName("video")) { 588 offer->description()->GetContentByName("video")->rejected = !video; 635 void SetReceiveAudioVideo(bool audio, bool video) { argument 637 session_description_constraints_.SetMandatoryReceiveVideo(video); [all...] |
H A D | webrtcsession.cc | 835 // support video. 854 LOG(LS_WARNING) << "SetVideoPlayout: No video channel exists."; 869 LOG(LS_WARNING) << "SetVideoSend: No video channel exists."; 1136 // Enabling voice and video channel. 1297 const cricket::ContentInfo* video = cricket::GetFirstVideoContent(desc); local 1298 if (video && !video->rejected && !video_channel_) { 1299 if (!CreateVideoChannel(video)) { 1300 LOG(LS_ERROR) << "Failed to create video channel.";
|
H A D | webrtcsession_unittest.cc | 108 static const char kMediaContentName1[] = "video"; 457 const TransportInfo* video = sdp->GetTransportInfoByName("video"); local 458 ASSERT_TRUE(video != NULL); 459 ASSERT_EQ(expected, video->description.identity_fingerprint.get() != NULL); 461 ASSERT_EQ(std::string(talk_base::DIGEST_SHA_256), video->description. 1474 // "video". Goal is to modify these content names and verify transport channel 1481 const std::string kVideoMid = "a=mid:video"; 1489 // Replacing |video| with |video_content_name|. 1581 // Test that an offer contains audio and video medi [all...] |
/external/chromium_org/third_party/libjingle/source/talk/session/media/ |
H A D | call.cc | 271 // If desired, create video channel and start a media monitor. 350 // Destroy video channel 487 << " because there is no video channel."; 500 LOG(LS_WARNING) << "Could not start video capture."; 523 VideoContentDescription* video = CreateVideoStreamUpdate(stream); local 525 // TODO(pthatcher): Wait until view request before sending video. 526 video_channel->SetLocalContent(video, CA_UPDATE); 527 SendVideoStreamUpdate(session, video); 540 << " because there is no video channel."; 547 VideoContentDescription* video local 592 VideoContentDescription* video = new VideoContentDescription(); local 599 SendVideoStreamUpdate( Session* session, VideoContentDescription* video) argument 881 UpdateVideoChannelRemoteContent( Session* session, const VideoContentDescription* video) argument [all...] |
H A D | mediasessionclient.cc | 81 // Register ourselves as the handler of audio and video sessions. 405 VideoContentDescription* video = new VideoContentDescription(); local 413 video->AddCodec(codec); 417 ParseGingleSsrc(content_elem, QN_GINGLE_VIDEO_SRCID, video); 418 ParseBandwidth(content_elem, video); 421 video, error)) { 425 *content = video; 605 talk_base::scoped_ptr<VideoContentDescription> video( 618 video->AddCodec(codec); 622 if (!ParseJingleStreamsOrLegacySsrc(content_elem, video 850 CreateGingleVideoContentElem( const VideoContentDescription* video, bool crypto_required) argument 1001 CreateJingleVideoContentElem( const VideoContentDescription* video, bool crypto_required) argument 1125 const VideoContentDescription* video = local [all...] |
H A D | channel_unittest.cc | 501 // kPcmuCodec is used as audio codec and kH264Codec is used as video codec. 661 // video options to the media channel. 1967 cricket::VideoContentDescription* video) { 1968 video->AddCodec(video_codec); 1969 video->set_rtcp_mux((flags & RTCP_MUX) != 0); 1971 video->AddCrypto(cricket::CryptoParams( 1980 cricket::VideoContentDescription* video) { 1981 *video = source; 1993 uint32 ssrc, int flags, cricket::VideoContentDescription* video) { 1994 video 1963 CreateContent( int flags, const cricket::AudioCodec& audio_codec, const cricket::VideoCodec& video_codec, cricket::VideoContentDescription* video) argument 1978 CopyContent( const cricket::VideoContentDescription& source, cricket::VideoContentDescription* video) argument 1992 AddLegacyStreamInContent( uint32 ssrc, int flags, cricket::VideoContentDescription* video) argument [all...] |
H A D | mediasession.cc | 169 // For video support only 80-bit SHA1 HMAC. For audio 32-bit HMAC is 373 // Helper class used for finding duplicate RTP payload types among audio, video 388 // audio and video extensions. 1069 // Handle m=video. 1071 scoped_ptr<VideoContentDescription> video(new VideoContentDescription()); 1083 video.get())) { 1087 video->set_bandwidth(options.video_bandwidth); 1088 SetMediaProtocol(secure_transport, video.get()); 1089 offer->AddContent(CN_VIDEO, NS_JINGLE_RTP, video.release()); 1233 // Handle m=video 1400 const VideoContentDescription* video = local 1439 const VideoContentDescription* video = local [all...] |
H A D | mediasessionclient_unittest.cc | 403 " <content name='test video'> " \ 404 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 431 " <description xmlns='http://www.google.com/session/video'> " \ 434 " <payload-type xmlns='http://www.google.com/session/video' " \ 452 " <content name='test video'> " \ 453 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 476 " <content name='test video'> " \ 477 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 507 " <content name='test video'> " \ 508 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 1724 const cricket::VideoContentDescription* video = local 1733 const cricket::VideoContentDescription* video = local 1793 const cricket::VideoContentDescription* video = local [all...] |
H A D | channel.cc | 1053 // Sets the maximum video bandwidth for automatic bandwidth adjustment. 1657 // Set remote video codecs (what the other side wants to receive). 1989 LOG(LS_ERROR) << "Failed to SetRender on video channel"; 1997 LOG(LS_ERROR) << "Failed to SetSend on video channel"; 2001 LOG(LS_INFO) << "Changing video state, recv=" << recv << " send=" << send; 2033 LOG(LS_INFO) << "Setting local video description"; 2035 const VideoContentDescription* video = local 2037 ASSERT(video != NULL); 2038 if (!video) return false; 2041 // Set local video codec 2071 const VideoContentDescription* video = local [all...] |
/external/qemu/distrib/sdl-1.2.15/src/video/ |
H A D | SDL_video.c | 24 /* The high-level video driver subsystem */ 34 /* Available video drivers */ 149 * Initialize the video and event subsystems -- determine native pixel format 153 SDL_VideoDevice *video; local 174 /* Select the proper video driver */ 176 video = NULL; 186 video = bootstrap[i]->create(index); 194 video = bootstrap[i]->create(index); 195 if ( video != NULL ) { 201 if ( video 333 SDL_VideoDevice *video = current_video; local 585 SDL_VideoDevice *video, *this; local 1040 SDL_VideoDevice *video = current_video; local 1107 SDL_VideoDevice *video = current_video; local 1186 SDL_VideoDevice *video = current_video; local 1304 SDL_VideoDevice *video = current_video; local 1349 SDL_VideoDevice *video = current_video; local 1411 SDL_VideoDevice *video = current_video; local 1430 SDL_VideoDevice *video = current_video; local 1451 SDL_VideoDevice *video = current_video; local 1518 SDL_VideoDevice* video = current_video; local 1533 SDL_VideoDevice *video = current_video; local 1701 SDL_VideoDevice *video = current_video; local 1730 SDL_VideoDevice *video = current_video; local 1804 SDL_VideoDevice *video = current_video; local 1837 SDL_VideoDevice *video = current_video; local 1876 SDL_VideoDevice *video = current_video; local 1924 SDL_VideoDevice *video = current_video; local 1940 SDL_VideoDevice *video = current_video; local 1973 SDL_VideoDevice* video = current_video; local 1984 SDL_VideoDevice* video = current_video; local 1998 SDL_VideoDevice* video = current_video; local 2009 SDL_VideoDevice* video = current_video; local 2020 SDL_VideoDevice* video = current_video; local 2033 SDL_VideoDevice *video = current_video; local [all...] |
/external/qemu/distrib/sdl-1.2.15/src/video/gapi/ |
H A D | SDL_gapivideo.c | 24 /* Pocket PC GAPI SDL video driver implementation; 41 // wince goes to portrait mode, but does not update video memory 224 /* Set up the new video mode rectangle */ 379 GAPIVID_DRIVER_NAME, "WinCE GAPI video driver", 621 SDL_Surface *video; local 634 video = current; 667 video = SDL_CreateRGBSurface(SDL_SWSURFACE, 669 if ( video == NULL ) { 677 video->flags = SDL_FULLSCREEN; /* Clear flags, GAPI supports fullscreen only */ 690 SDL_SetError("Couldn't get address of video memor [all...] |
/external/qemu/distrib/sdl-1.2.15/src/video/windib/ |
H A D | SDL_dibvideo.c | 273 /* Set up the new video mode rectangle */ 382 /* Query for the list of available video modes */ 598 SDL_Surface *video; local 623 * and uploaded textures. So if we're requesting the same video mode 647 video = current; 677 video = SDL_CreateRGBSurface(SDL_SWSURFACE, 679 if ( video == NULL ) { 685 /* Fill in part of the video surface */ 686 video->flags = 0; /* Clear flags */ 687 video [all...] |
/external/qemu/distrib/sdl-1.2.15/src/video/windx5/ |
H A D | SDL_dx5video.c | 58 /* DirectX function pointers for video and events */ 762 error = "Out of video memory"; 896 /* Find out how much video memory is available */ 995 /* Fill in the video hardware capabilities */ 1036 SDL_Surface *video; local 1054 fprintf(stderr, "Setting %dx%dx%d video mode\n", width, height, bpp); 1085 video = current; 1115 video = SDL_CreateRGBSurface(SDL_SWSURFACE, 0, 0, bpp, 1117 if ( video == NULL ) { 1123 /* Fill in part of the video surfac [all...] |
/external/qemu-pc-bios/bochs/ |
H A D | bochs.h | 436 bx_bool video; member in struct:__anon25134
|
/external/chromium/chrome/browser/resources/file_manager/js/ |
H A D | file_manager.js | 131 'video': /\.(mov|mp4|m4v|mpe?g4?|ogm|ogv|ogx|webm)$/i 138 'video': 'images/filetype_large_video.png'
|
/external/chromium_org/media/filters/ |
H A D | chunk_demuxer_unittest.cc | 274 type = "video/webm"; 469 // The resulting video stream returns data from each file for the following 565 // The first video frame must be a keyframe. 890 // Test no streams, audio-only, video-only, and audio & video scenarios. 891 // Audio and video streams can be encrypted or not encrypted. 958 EXPECT_EQ(AddId("video", false, true), ChunkDemuxer::kOk); 1015 // Note: We skip trying to read a video buffer here because computing 1210 // Request a read on the audio and video streams. 1216 DemuxerStream* video local 2138 DemuxerStream* video = demuxer_->GetStream(DemuxerStream::VIDEO); local 2229 DemuxerStream* video = demuxer_->GetStream(DemuxerStream::VIDEO); local [all...] |
/external/chromium_org/third_party/WebKit/Source/core/rendering/ |
H A D | RenderLayerCompositor.cpp | 788 // own layers to draw on top of the accelerated video. 1651 return "video"; 1793 RenderVideo* video = toRenderVideo(renderer); local 1794 return video->shouldDisplayVideo() && canAccelerateVideoRendering(video);
|
/external/kernel-headers/original/linux/ |
H A D | videodev.h | 200 int video; /* Video minor */ member in struct:video_unit 256 #define VIDIOCGWIN _IOR('v',9, struct video_window) /* Get the video overlay window */ 257 #define VIDIOCSWIN _IOW('v',10, struct video_window) /* Set the video overlay window - passes clip list for hardware smarts , chromakey etc */ 271 #define VIDIOCSPLAYMODE _IOW('v',24, struct video_play_mode) /* Set output video mode/feature */
|
/external/chromium_org/third_party/WebKit/Source/core/html/canvas/ |
H A D | CanvasRenderingContext2D.cpp | 1181 static IntSize sizeFor(HTMLVideoElement* video) argument 1183 if (MediaPlayer* player = video->player()) 1479 void CanvasRenderingContext2D::drawImage(HTMLVideoElement* video, float x, float y, ExceptionState& es) argument 1481 if (!video) { 1485 IntSize size = sizeFor(video); 1486 drawImage(video, x, y, size.width(), size.height(), es); 1489 void CanvasRenderingContext2D::drawImage(HTMLVideoElement* video, argument 1492 if (!video) { 1496 IntSize size = sizeFor(video); 1497 drawImage(video, FloatRec 1500 drawImage(HTMLVideoElement* video, float sx, float sy, float sw, float sh, float dx, float dy, float dw, float dh, ExceptionState& es) argument 1507 drawImage(HTMLVideoElement* video, const FloatRect& srcRect, const FloatRect& dstRect, ExceptionState& es) argument [all...] |
H A D | WebGLRenderingContext.cpp | 3409 PassRefPtr<Image> WebGLRenderingContext::videoFrameToImage(HTMLVideoElement* video, BackingStoreCopy backingStoreCopy) argument 3411 IntSize size(video->videoWidth(), video->videoHeight()); 3419 video->paintCurrentFrameInContext(buf->context(), destRect); 3424 GC3Denum format, GC3Denum type, HTMLVideoElement* video, ExceptionState& es) 3426 if (isContextLost() || !validateHTMLVideoElement("texImage2D", video, es) 3427 || !validateTexFunc("texImage2D", NotTexSubImage2D, SourceHTMLVideoElement, target, level, internalformat, video->videoWidth(), video->videoHeight(), 0, format, type, 0, 0)) 3434 if (video->copyVideoTextureToPlatformTexture(m_context.get(), texture->object(), level, type, internalformat, m_unpackPremultiplyAlpha, m_unpackFlipY)) { 3435 texture->setLevelInfo(target, level, internalformat, video 3423 texImage2D(GC3Denum target, GC3Dint level, GC3Denum internalformat, GC3Denum format, GC3Denum type, HTMLVideoElement* video, ExceptionState& es) argument 3624 texSubImage2D(GC3Denum target, GC3Dint level, GC3Dint xoffset, GC3Dint yoffset, GC3Denum format, GC3Denum type, HTMLVideoElement* video, ExceptionState& es) argument 5095 validateHTMLVideoElement(const char* functionName, HTMLVideoElement* video, ExceptionState& es) argument [all...] |
/external/libvpx/libvpx/nestegg/src/ |
H A D | nestegg.c | 202 struct video { struct 230 struct video video; member in struct:track_entry 377 E_FIELD(ID_PIXEL_WIDTH, TYPE_UINT, struct video, pixel_width), 378 E_FIELD(ID_PIXEL_HEIGHT, TYPE_UINT, struct video, pixel_height), 379 E_FIELD(ID_PIXEL_CROP_BOTTOM, TYPE_UINT, struct video, pixel_crop_bottom), 380 E_FIELD(ID_PIXEL_CROP_TOP, TYPE_UINT, struct video, pixel_crop_top), 381 E_FIELD(ID_PIXEL_CROP_LEFT, TYPE_UINT, struct video, pixel_crop_left), 382 E_FIELD(ID_PIXEL_CROP_RIGHT, TYPE_UINT, struct video, pixel_crop_right), 383 E_FIELD(ID_DISPLAY_WIDTH, TYPE_UINT, struct video, display_widt [all...] |