/external/webrtc/webrtc/common_video/libyuv/ |
H A D | scaler.cc | 50 int Scaler::Scale(const VideoFrame& src_frame, VideoFrame* dst_frame) { argument 51 assert(dst_frame); 58 dst_frame->set_video_frame_buffer( 89 dst_frame->buffer(kYPlane), 90 dst_frame->stride(kYPlane), 91 dst_frame->buffer(kUPlane), 92 dst_frame->stride(kUPlane), 93 dst_frame->buffer(kVPlane), 94 dst_frame->stride(kVPlane),
|
H A D | webrtc_libyuv.cc | 157 uint8_t* dst_frame, 165 dst_frame, width, 169 int ConvertRGB24ToARGB(const uint8_t* src_frame, uint8_t* dst_frame, argument 174 dst_frame, dst_stride, 239 VideoFrame* dst_frame) { 240 int dst_width = dst_frame->width(); 241 int dst_height = dst_frame->height(); 245 dst_width = dst_frame->height(); 246 dst_height = dst_frame->width(); 249 dst_frame 156 ConvertNV12ToRGB565(const uint8_t* src_frame, uint8_t* dst_frame, int width, int height) argument 231 ConvertToI420(VideoType src_video_type, const uint8_t* src_frame, int crop_x, int crop_y, int src_width, int src_height, size_t sample_size, VideoRotation rotation, VideoFrame* dst_frame) argument 262 ConvertFromI420(const VideoFrame& src_frame, VideoType dst_video_type, int dst_sample_size, uint8_t* dst_frame) argument 278 ConvertFromYV12(const VideoFrame& src_frame, VideoType dst_video_type, int dst_sample_size, uint8_t* dst_frame) argument [all...] |
/external/webrtc/webrtc/voice_engine/ |
H A D | utility.cc | 26 AudioFrame* dst_frame) { 29 resampler, dst_frame); 30 dst_frame->timestamp_ = src_frame.timestamp_; 31 dst_frame->elapsed_time_ms_ = src_frame.elapsed_time_ms_; 32 dst_frame->ntp_time_ms_ = src_frame.ntp_time_ms_; 40 AudioFrame* dst_frame) { 46 if (num_channels == 2 && dst_frame->num_channels_ == 1) { 53 if (resampler->InitializeIfNeeded(sample_rate_hz, dst_frame->sample_rate_hz_, 56 << sample_rate_hz << ", dst_frame->sample_rate_hz_ = " 57 << dst_frame 24 RemixAndResample(const AudioFrame& src_frame, PushResampler<int16_t>* resampler, AudioFrame* dst_frame) argument 35 RemixAndResample(const int16_t* src_data, size_t samples_per_channel, size_t num_channels, int sample_rate_hz, PushResampler<int16_t>* resampler, AudioFrame* dst_frame) argument [all...] |
/external/adhd/cras/src/tests/ |
H A D | mix_unittest.cc | 845 int32_t dst_frame = 0; local 847 memcpy((uint8_t *)&dst_frame + 1, mix_buffer_ + 3*i, 3); 849 tmp = (int64_t)dst_frame + (int64_t)src_frame * scaler; 851 tmp = (int64_t)dst_frame + (int64_t)src_frame; 856 dst_frame = (int32_t)tmp; 857 memcpy(compare_buffer_ + 3*i, (uint8_t *)&dst_frame + 1, 3);
|
/external/adhd/cras/src/server/ |
H A D | cras_mix_ops.c | 594 int32_t dst_frame; local 599 convert_single_s243le_to_s32le(&dst_frame, dst); 601 sum = (int64_t)dst_frame + (int64_t)src_frame; 606 dst_frame = (int32_t)sum; 607 convert_single_s32le_to_s243le(dst, &dst_frame); 619 int32_t dst_frame; local 627 convert_single_s243le_to_s32le(&dst_frame, dst); 629 sum = (int64_t)dst_frame + (int64_t)(src_frame * vol); 634 dst_frame = (int32_t)sum; 635 convert_single_s32le_to_s243le(dst, &dst_frame); 743 int32_t dst_frame; local [all...] |
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/ |
H A D | simulcast_encoder_adapter.cc | 288 VideoFrame dst_frame; local 291 dst_frame.CreateEmptyFrame(dst_width, dst_height, dst_width, 297 src_height, dst_frame.buffer(kYPlane), dst_frame.stride(kYPlane), 298 dst_frame.buffer(kUPlane), dst_frame.stride(kUPlane), 299 dst_frame.buffer(kVPlane), dst_frame.stride(kVPlane), dst_width, 301 dst_frame.set_timestamp(input_image.timestamp()); 302 dst_frame [all...] |
/external/tensorflow/tensorflow/core/graph/ |
H A D | graph_partition.cc | 643 const Node* dst_frame = InputFrame(dst, cf_info); local 645 const string& dst_frame_name = cf_info[dst_frame->id()].frame_name; 723 const Node* dst_frame = InputFrame(dst, cf_info); local 725 const string& dst_frame_name = cf_info[dst_frame->id()].frame_name;
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
H A D | row_common.cc | 2251 uint8* dst_frame, int width) { 2254 dst_frame[0] = src_y[0]; 2255 dst_frame[1] = src_u[0]; 2256 dst_frame[2] = src_y[1]; 2257 dst_frame[3] = src_v[0]; 2258 dst_frame += 4; 2264 dst_frame[0] = src_y[0]; 2265 dst_frame[1] = src_u[0]; 2266 dst_frame[2] = 0; 2267 dst_frame[ 2248 I422ToYUY2Row_C(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument 2271 I422ToUYVYRow_C(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument [all...] |
H A D | row_gcc.cc | 5207 uint8* dst_frame, int width) { 5229 "+r"(dst_frame), // %3 5242 uint8* dst_frame, int width) { 5264 "+r"(dst_frame), // %3 5204 I422ToYUY2Row_SSE2(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument 5239 I422ToUYVYRow_SSE2(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument
|
/external/libyuv/files/source/ |
H A D | row_common.cc | 2243 uint8* dst_frame, 2247 dst_frame[0] = src_y[0]; 2248 dst_frame[1] = src_u[0]; 2249 dst_frame[2] = src_y[1]; 2250 dst_frame[3] = src_v[0]; 2251 dst_frame += 4; 2257 dst_frame[0] = src_y[0]; 2258 dst_frame[1] = src_u[0]; 2259 dst_frame[2] = 0; 2260 dst_frame[ 2240 I422ToYUY2Row_C(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument 2264 I422ToUYVYRow_C(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument [all...] |
H A D | row_gcc.cc | 5287 uint8* dst_frame, 5310 "+r"(dst_frame), // %3 5323 uint8* dst_frame, 5346 "+r"(dst_frame), // %3 5284 I422ToYUY2Row_SSE2(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument 5320 I422ToUYVYRow_SSE2(const uint8* src_y, const uint8* src_u, const uint8* src_v, uint8* dst_frame, int width) argument
|