/external/qemu/distrib/sdl-1.2.15/src/video/ |
H A D | SDL_blit.c | 276 SDL_VideoDevice *video = current_video; local 278 video->CheckHWBlit(this, surface, surface->map->dst); 287 SDL_VideoDevice *video = current_video; local 289 video->CheckHWBlit(this, surface, surface->map->dst);
|
/external/chromium_org/third_party/libjingle/source/talk/app/webrtc/ |
H A D | jsepsessiondescription_unittest.cc | 56 // This creates a session description with both audio and video media contents. 57 // In SDP this is described by two m lines, one audio and one video. 65 scoped_ptr<cricket::VideoContentDescription> video( 72 video->AddCodec(cricket::VideoCodec(120, "VP8", 640, 480, 30, 0)); 74 video.release()); 165 JsepIceCandidate jsep_candidate("video", 0, candidate_);
|
H A D | peerconnection_unittest.cc | 150 virtual void Negotiate(bool audio, bool video) = 0; 157 void AddMediaStream(bool audio, bool video) { argument 178 if (video && can_receive_video()) { 585 // Store references to the video capturers we've created, so that we can stop 608 virtual void Negotiate(bool audio, bool video) { argument 615 if (offer->description()->GetContentByName("video")) { 616 offer->description()->GetContentByName("video")->rejected = !video; 663 void SetReceiveAudioVideo(bool audio, bool video) { argument 665 SetReceiveVideo(video); 676 SetReceiveVideo(bool video) argument [all...] |
H A D | peerconnectionendtoend_unittest.cc | 97 InjectAfter("a=mid:video\r\n", kVideoSdes, sdp); 115 InjectAfter("a=mid:video\r\n", ufrag_pwd, sdp); 162 bool video, FakeConstraints video_constraints) { 164 video, video_constraints); 166 video, video_constraints); 161 GetAndAddUserMedia(bool audio, FakeConstraints audio_constraints, bool video, FakeConstraints video_constraints) argument
|
/external/chromium_org/third_party/libjingle/source/talk/session/media/ |
H A D | call.cc | 281 // If desired, create video channel and start a media monitor. 360 // Destroy video channel 495 << " because there is no video channel."; 515 LOG(LS_WARNING) << "Could not start video capture."; 538 VideoContentDescription* video = CreateVideoStreamUpdate(stream); local 540 // TODO(pthatcher): Wait until view request before sending video. 541 video_channel->SetLocalContent(video, CA_UPDATE, NULL); 542 SendVideoStreamUpdate(session, video); 555 << " because there is no video channel."; 562 VideoContentDescription* video local 607 VideoContentDescription* video = new VideoContentDescription(); local 614 SendVideoStreamUpdate( Session* session, VideoContentDescription* video) argument 899 UpdateVideoChannelRemoteContent( Session* session, const VideoContentDescription* video) argument [all...] |
H A D | mediamessages.h | 50 // A collection of audio and video and data streams. Most of the 67 const std::vector<StreamParams>& video() const { return video_; } function in struct:cricket::MediaStreams
|
H A D | mediasessionclient_unittest.cc | 310 " <content name='test video'> " \ 311 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 338 " <description xmlns='http://www.google.com/session/video'> " \ 341 " <payload-type xmlns='http://www.google.com/session/video' " \ 359 " <content name='test video'> " \ 360 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 383 " <content name='test video'> " \ 384 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 414 " <content name='test video'> " \ 415 " <description xmlns='urn:xmpp:jingle:apps:rtp:1' media='video'> " \ 1630 const cricket::VideoContentDescription* video = local 1639 const cricket::VideoContentDescription* video = local 1699 const cricket::VideoContentDescription* video = local [all...] |
H A D | mediasession.cc | 179 // For video support only 80-bit SHA1 HMAC. For audio 32-bit HMAC is 396 // Helper class used for finding duplicate RTP payload types among audio, video 411 // audio and video extensions. 1066 type_str = "video"; 1326 const VideoContentDescription* video = local 1328 if (video) { 1329 *video_codecs = video->codecs(); 1367 const VideoContentDescription* video = local 1369 if (video) { 1370 *video_extensions = video [all...] |
H A D | channel.cc | 1508 // Set remote video codecs (what the other side wants to receive). 1765 LOG(LS_ERROR) << "Failed to SetRender on video channel"; 1773 LOG(LS_ERROR) << "Failed to SetSend on video channel"; 1777 LOG(LS_INFO) << "Changing video state, recv=" << recv << " send=" << send; 1810 LOG(LS_INFO) << "Setting local video description"; 1812 const VideoContentDescription* video = local 1814 ASSERT(video != NULL); 1815 if (!video) { 1816 SafeSetError("Can't find video content in local description.", error_desc); 1821 // Set local video codec 1859 const VideoContentDescription* video = local [all...] |
/external/chromium_org/ui/file_manager/video_player/js/ |
H A D | video_player.js | 9 * @param {Element} videoContainer Container for the video element. 10 * @param {Element} controlsContainer Container for video controls. 99 // The window is hidden if the video has not loaded yet. 123 document.querySelector('#video-player').setAttribute('disabled', 'true'); 125 // Detach the video element, since it may be unreliable and reset stored 130 // Avoid reusing a video element. 178 * Initializes the video player window. This method must be called after DOM 224 document.querySelector('#video-player'), 225 document.querySelector('#video-container'), 244 var videoPlayerElement = document.querySelector('#video [all...] |
/external/chromium_org/chrome/browser/resources/chromeos/ |
H A D | user_images_grid.js | 204 * Whether camera is actually streaming video. May be |false| even when 225 {video: true}, 420 this.cameraVideo_ = value.querySelector('video'); 496 * Capture a single still frame from a <video> element, placing it at the 498 * @param {HTMLVideoElement} video Video element to capture from. 503 captureFrame_: function(video, ctx, destSize) { 504 var width = video.videoWidth; 505 var height = video.videoHeight; 522 ctx.drawImage(video, src.x, src.y, src.width, src.height,
|
/external/chromium_org/content/browser/media/ |
H A D | webrtc_internals.h | 72 // is requested, |video| is true if the video stream is requested, 74 // is the constraints for the video. 79 bool video, 166 // "video" -- the serialized video constraints if video is requested.
|
H A D | webrtc_internals.cc | 188 bool video, 199 if (video) 200 dict->SetString("video", video_constraints); 184 OnGetUserMedia(int rid, base::ProcessId pid, const std::string& origin, bool audio, bool video, const std::string& audio_constraints, const std::string& video_constraints) argument
|
H A D | webrtc_internals_unittest.cc | 79 const std::string& video) { 87 VerifyString(dict, "video", video); 74 VerifyGetUserMediaData(base::Value* actual_data, int rid, int pid, const std::string& origin, const std::string& audio, const std::string& video) argument
|
/external/libvpx/libvpx/third_party/nestegg/src/ |
H A D | nestegg.c | 208 struct video { struct 239 struct video video; member in struct:track_entry 390 E_FIELD(ID_STEREO_MODE, TYPE_UINT, struct video, stereo_mode), 391 E_FIELD(ID_PIXEL_WIDTH, TYPE_UINT, struct video, pixel_width), 392 E_FIELD(ID_PIXEL_HEIGHT, TYPE_UINT, struct video, pixel_height), 393 E_FIELD(ID_PIXEL_CROP_BOTTOM, TYPE_UINT, struct video, pixel_crop_bottom), 394 E_FIELD(ID_PIXEL_CROP_TOP, TYPE_UINT, struct video, pixel_crop_top), 395 E_FIELD(ID_PIXEL_CROP_LEFT, TYPE_UINT, struct video, pixel_crop_left), 396 E_FIELD(ID_PIXEL_CROP_RIGHT, TYPE_UINT, struct video, pixel_crop_righ [all...] |
/external/chromium_org/content/shell/renderer/test_runner/ |
H A D | mock_web_user_media_client.cc | 169 WebVector<WebMediaStreamTrack> video_tracks(request.video() ? one : zero); 179 if (request.video()) { 183 "Mock video device");
|
/external/chromium_org/third_party/libjingle/source/talk/examples/objc/AppRTCDemo/ |
H A D | APPRTCConnectionManager.m | 144 // TODO(tkchin): local video capture for OSX. See 187 RTCPair* video = 189 NSArray* mandatory = @[ audio, video ]; 268 @"Expected audio or video track"); 272 @"Expected at most 1 video stream"); 383 RTCPair* video = [[RTCPair alloc] initWithKey:@"OfferToReceiveVideo" 385 NSArray* mandatory = @[ audio, video ];
|
/external/chromium_org/third_party/webrtc/modules/media_file/source/ |
H A D | media_file_impl.h | 116 // provide a non-NULL codecInst. Only video will be read if videoOnly is 139 // provide a non-NULL codecInst. Only video will be read if videoOnly is 158 // written to audioBuffer. If video is true the data written is a video 161 bool video); 164 // to file. The frame is an audio frame if video is true otherwise it is an 168 const bool video); 177 // specifies the encoding of the video data. maxSizeBytes specifies the 180 // only contain video frames. 198 // specifies the encoding of the video dat [all...] |
/external/chromium_org/tools/telemetry/telemetry/core/ |
H A D | tab_unittest.py | 12 from telemetry.core import video namespace 44 return video.Video(tempfile.NamedTemporaryFile())
|
/external/chromium_org/remoting/webapp/ |
H A D | main.css | 757 /* video-container needs relative position so that mediasource-video-output can 759 #video-container { 763 /* mediasource-video-output is hidden by default. */ 764 #video-container video { 768 /* Use absolute positioning for mediasource-video-output so that it's rendered 770 #video-container.mediasource-rendering video { 780 * top of the <video> elemen [all...] |
/external/chromium_org/media/base/android/ |
H A D | media_source_player.h | 35 // MediaCodec to decode audio and video streams in two separate threads. 110 // Functions check whether audio/video is present. 114 // Functions that check whether audio/video stream has reached end of output 150 // video playback. 155 // |start_time_ticks_| and |start_presentation_timestamp_| so that video can 190 // If the player is previously waiting for audio or video decoder job, retry 191 // creating the decoders identified by |audio| and |video|. 192 void RetryDecoderCreation(bool audio, bool video); 212 // Then we use this information to estimate when the next audio/video frame 214 // TODO(qinmin): Need to fix the problem if audio/video lagge [all...] |
/external/chromium_org/media/formats/mp4/ |
H A D | mp4_stream_parser.cc | 26 static const char kMp4InitDataType[] = "video/mp4"; 187 // TODO(strobe): Only the first audio and video track present in a file are 282 MEDIA_LOG(log_cb_) << "Unsupported video format 0x" 298 // SPS/PPS are embedded in the video stream 444 bool video = has_video_ && video_track_id_ == runs_->track_id(); local 447 if (!audio && !video) { 481 if (video) { 510 (video && is_video_track_encrypted_)) {
|
/external/kernel-headers/original/uapi/asm-x86/asm/ |
H A D | bootparam.h | 37 #include <video/edid.h>
|
/external/chromium_org/tools/telemetry/telemetry/core/platform/ |
H A D | android_platform_backend.py | 12 from telemetry.core import video namespace 253 """Starts the video capture at specified bitrate.""" 256 raise ValueError('Android video capture cannot capture at %dmbps. ' 270 assert self.is_video_capture_running, 'Must start video capture first' 276 return video.Video(video_file_obj)
|
/external/chromium_org/third_party/WebKit/Source/core/html/canvas/ |
H A D | CanvasRenderingContext2D.idl | 119 [RaisesException] void drawImage(HTMLVideoElement video, unrestricted float x, unrestricted float y); 120 [RaisesException] void drawImage(HTMLVideoElement video, unrestricted float x, unrestricted float y, unrestricted float width, unrestricted float height); 121 [RaisesException] void drawImage(HTMLVideoElement video, unrestricted float sx, unrestricted float sy, unrestricted float sw, unrestricted float sh, unrestricted float dx, unrestricted float dy, unrestricted float dw, unrestricted float dh);
|