/external/webrtc/webrtc/video/ |
H A D | video_decoder.cc | 86 const EncodedImage& input_image, 93 if (!fallback_decoder_ || input_image._frameType == kVideoFrameKey) { 94 int32_t ret = decoder_->Decode(input_image, missing_frames, fragmentation, 112 return fallback_decoder_->Decode(input_image, missing_frames, fragmentation, 85 Decode( const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t render_time_ms) argument
|
/external/webrtc/webrtc/modules/video_coding/codecs/h264/ |
H A D | h264_video_toolbox_decoder.cc | 122 const EncodedImage& input_image, 127 RTC_DCHECK(input_image._buffer); 130 if (!H264AnnexBBufferToCMSampleBuffer(input_image._buffer, 131 input_image._length, video_format_, 147 new internal::FrameDecodeParams(callback_, input_image._timeStamp)); 121 Decode( const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t render_time_ms) argument
|
H A D | h264_video_toolbox_encoder.cc | 236 const VideoFrame& input_image, 239 if (input_image.IsZeroSize()) { 263 if (!internal::CopyVideoFrameToPixelBuffer(input_image, pixel_buffer)) { 281 CMTimeMake(input_image.render_time_ms(), 1000); 291 input_image.render_time_ms(), input_image.timestamp())); 235 Encode( const VideoFrame& input_image, const CodecSpecificInfo* codec_specific_info, const std::vector<FrameType>* frame_types) argument
|
/external/webrtc/webrtc/test/ |
H A D | fake_encoder.cc | 48 int32_t FakeEncoder::Encode(const VideoFrame& input_image, argument 102 encoded._timeStamp = input_image.timestamp(); 103 encoded.capture_time_ms_ = input_image.render_time_ms(); 200 int32_t DelayedEncoder::Encode(const VideoFrame& input_image, argument 204 return FakeEncoder::Encode(input_image, codec_specific_info, frame_types);
|
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/ |
H A D | simulcast_encoder_adapter.cc | 232 const VideoFrame& input_image, 261 int src_width = input_image.width(); 262 int src_height = input_image.height(); 284 input_image.IsZeroSize()) { 285 streaminfos_[stream_idx].encoder->Encode(input_image, codec_specific_info, 294 input_image.buffer(kYPlane), input_image.stride(kYPlane), 295 input_image.buffer(kUPlane), input_image.stride(kUPlane), 296 input_image 231 Encode( const VideoFrame& input_image, const CodecSpecificInfo* codec_specific_info, const std::vector<FrameType>* frame_types) argument [all...] |
H A D | vp8_impl.cc | 724 const VideoFrame& input_image = local 727 if (quality_scaler_enabled_ && (input_image.width() != codec_.width || 728 input_image.height() != codec_.height)) { 729 int ret = UpdateCodecFrameSize(input_image); 734 // Since we are extracting raw pointers from |input_image| to 736 // |input_image| might be scaled from |frame|. In that case, the resolution of 738 RTC_DCHECK_EQ(input_image.width(), static_cast<int>(raw_images_[0].d_w)); 739 RTC_DCHECK_EQ(input_image.height(), static_cast<int>(raw_images_[0].d_h)); 744 const_cast<uint8_t*>(input_image.buffer(kYPlane)); 746 const_cast<uint8_t*>(input_image 899 UpdateCodecFrameSize(const VideoFrame& input_image) argument 951 GetEncodedPartitions(const VideoFrame& input_image, bool only_predicting_from_key_frame) argument 1122 Decode(const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t ) argument 1309 DecodePartitions( const EncodedImage& input_image, const RTPFragmentationHeader* fragmentation) argument [all...] |
/external/webrtc/webrtc/modules/video_coding/codecs/vp9/ |
H A D | vp9_impl.cc | 475 int VP9EncoderImpl::Encode(const VideoFrame& input_image, argument 481 if (input_image.IsZeroSize()) { 492 RTC_DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w)); 493 RTC_DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h)); 496 // This was necessary since you need some information from input_image. 499 input_image_ = &input_image; 503 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane)); 504 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane)); 505 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane)); 506 raw_->stride[VPX_PLANE_Y] = input_image 884 Decode(const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t ) argument [all...] |
/external/ImageMagick/MagickCore/ |
H A D | opencl.c | 1539 % const size_t *local_work_size,const Image *input_image, 1561 % o input_image: the input image of the operation. 1589 const size_t *lsize,const Image *input_image,const Image *output_image, 1606 assert(input_image != (const Image *) NULL); 1607 input_info=(CacheInfo *) input_image->cache; 1586 EnqueueOpenCLKernel(cl_command_queue queue, cl_kernel kernel,cl_uint work_dim,const size_t *offset,const size_t *gsize, const size_t *lsize,const Image *input_image,const Image *output_image, ExceptionInfo *exception) argument
|