Searched defs:input_image (Results 1 - 8 of 8) sorted by relevance

/external/webrtc/webrtc/video/
H A Dvideo_decoder.cc86 const EncodedImage& input_image,
93 if (!fallback_decoder_ || input_image._frameType == kVideoFrameKey) {
94 int32_t ret = decoder_->Decode(input_image, missing_frames, fragmentation,
112 return fallback_decoder_->Decode(input_image, missing_frames, fragmentation,
85 Decode( const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t render_time_ms) argument
/external/webrtc/webrtc/modules/video_coding/codecs/h264/
H A Dh264_video_toolbox_decoder.cc122 const EncodedImage& input_image,
127 RTC_DCHECK(input_image._buffer);
130 if (!H264AnnexBBufferToCMSampleBuffer(input_image._buffer,
131 input_image._length, video_format_,
147 new internal::FrameDecodeParams(callback_, input_image._timeStamp));
121 Decode( const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t render_time_ms) argument
H A Dh264_video_toolbox_encoder.cc236 const VideoFrame& input_image,
239 if (input_image.IsZeroSize()) {
263 if (!internal::CopyVideoFrameToPixelBuffer(input_image, pixel_buffer)) {
281 CMTimeMake(input_image.render_time_ms(), 1000);
291 input_image.render_time_ms(), input_image.timestamp()));
235 Encode( const VideoFrame& input_image, const CodecSpecificInfo* codec_specific_info, const std::vector<FrameType>* frame_types) argument
/external/webrtc/webrtc/test/
H A Dfake_encoder.cc48 int32_t FakeEncoder::Encode(const VideoFrame& input_image, argument
102 encoded._timeStamp = input_image.timestamp();
103 encoded.capture_time_ms_ = input_image.render_time_ms();
200 int32_t DelayedEncoder::Encode(const VideoFrame& input_image, argument
204 return FakeEncoder::Encode(input_image, codec_specific_info, frame_types);
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/
H A Dsimulcast_encoder_adapter.cc232 const VideoFrame& input_image,
261 int src_width = input_image.width();
262 int src_height = input_image.height();
284 input_image.IsZeroSize()) {
285 streaminfos_[stream_idx].encoder->Encode(input_image, codec_specific_info,
294 input_image.buffer(kYPlane), input_image.stride(kYPlane),
295 input_image.buffer(kUPlane), input_image.stride(kUPlane),
296 input_image
231 Encode( const VideoFrame& input_image, const CodecSpecificInfo* codec_specific_info, const std::vector<FrameType>* frame_types) argument
[all...]
H A Dvp8_impl.cc724 const VideoFrame& input_image = local
727 if (quality_scaler_enabled_ && (input_image.width() != codec_.width ||
728 input_image.height() != codec_.height)) {
729 int ret = UpdateCodecFrameSize(input_image);
734 // Since we are extracting raw pointers from |input_image| to
736 // |input_image| might be scaled from |frame|. In that case, the resolution of
738 RTC_DCHECK_EQ(input_image.width(), static_cast<int>(raw_images_[0].d_w));
739 RTC_DCHECK_EQ(input_image.height(), static_cast<int>(raw_images_[0].d_h));
744 const_cast<uint8_t*>(input_image.buffer(kYPlane));
746 const_cast<uint8_t*>(input_image
899 UpdateCodecFrameSize(const VideoFrame& input_image) argument
951 GetEncodedPartitions(const VideoFrame& input_image, bool only_predicting_from_key_frame) argument
1122 Decode(const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t ) argument
1309 DecodePartitions( const EncodedImage& input_image, const RTPFragmentationHeader* fragmentation) argument
[all...]
/external/webrtc/webrtc/modules/video_coding/codecs/vp9/
H A Dvp9_impl.cc475 int VP9EncoderImpl::Encode(const VideoFrame& input_image, argument
481 if (input_image.IsZeroSize()) {
492 RTC_DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w));
493 RTC_DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h));
496 // This was necessary since you need some information from input_image.
499 input_image_ = &input_image;
503 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane));
504 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane));
505 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane));
506 raw_->stride[VPX_PLANE_Y] = input_image
884 Decode(const EncodedImage& input_image, bool missing_frames, const RTPFragmentationHeader* fragmentation, const CodecSpecificInfo* codec_specific_info, int64_t ) argument
[all...]
/external/ImageMagick/MagickCore/
H A Dopencl.c1539 % const size_t *local_work_size,const Image *input_image,
1561 % o input_image: the input image of the operation.
1589 const size_t *lsize,const Image *input_image,const Image *output_image,
1606 assert(input_image != (const Image *) NULL);
1607 input_info=(CacheInfo *) input_image->cache;
1586 EnqueueOpenCLKernel(cl_command_queue queue, cl_kernel kernel,cl_uint work_dim,const size_t *offset,const size_t *gsize, const size_t *lsize,const Image *input_image,const Image *output_image, ExceptionInfo *exception) argument

Completed in 369 milliseconds