1// Copyright 2013 The Chromium Authors. All rights reserved.
2// Use of this source code is governed by a BSD-style license that can be
3// found in the LICENSE file.
4
5#include "remoting/codec/video_encoder_vpx.h"
6
7#include "base/bind.h"
8#include "base/command_line.h"
9#include "base/logging.h"
10#include "base/sys_info.h"
11#include "remoting/base/util.h"
12#include "remoting/proto/video.pb.h"
13#include "third_party/libyuv/include/libyuv/convert_from_argb.h"
14#include "third_party/webrtc/modules/desktop_capture/desktop_frame.h"
15#include "third_party/webrtc/modules/desktop_capture/desktop_geometry.h"
16#include "third_party/webrtc/modules/desktop_capture/desktop_region.h"
17
18extern "C" {
19#define VPX_CODEC_DISABLE_COMPAT 1
20#include "third_party/libvpx/source/libvpx/vpx/vpx_encoder.h"
21#include "third_party/libvpx/source/libvpx/vpx/vp8cx.h"
22}
23
24namespace remoting {
25
26namespace {
27
28// Name of command-line flag to enable VP9 to use I444 by default.
29const char kEnableI444SwitchName[] = "enable-i444";
30
31// Number of bytes in an RGBx pixel.
32const int kBytesPerRgbPixel = 4;
33
34// Defines the dimension of a macro block. This is used to compute the active
35// map for the encoder.
36const int kMacroBlockSize = 16;
37
38// Magic encoder profile numbers for I420 and I444 input formats.
39const int kVp9I420ProfileNumber = 0;
40const int kVp9I444ProfileNumber = 1;
41
42void SetCommonCodecParameters(const webrtc::DesktopSize& size,
43                              vpx_codec_enc_cfg_t* config) {
44  // Use millisecond granularity time base.
45  config->g_timebase.num = 1;
46  config->g_timebase.den = 1000;
47
48  // Adjust default target bit-rate to account for actual desktop size.
49  config->rc_target_bitrate = size.width() * size.height() *
50      config->rc_target_bitrate / config->g_w / config->g_h;
51
52  config->g_w = size.width();
53  config->g_h = size.height();
54  config->g_pass = VPX_RC_ONE_PASS;
55
56  // Start emitting packets immediately.
57  config->g_lag_in_frames = 0;
58
59  // Using 2 threads gives a great boost in performance for most systems with
60  // adequate processing power. NB: Going to multiple threads on low end
61  // windows systems can really hurt performance.
62  // http://crbug.com/99179
63  config->g_threads = (base::SysInfo::NumberOfProcessors() > 2) ? 2 : 1;
64}
65
66ScopedVpxCodec CreateVP8Codec(const webrtc::DesktopSize& size) {
67  ScopedVpxCodec codec(new vpx_codec_ctx_t);
68
69  // Configure the encoder.
70  vpx_codec_enc_cfg_t config;
71  const vpx_codec_iface_t* algo = vpx_codec_vp8_cx();
72  CHECK(algo);
73  vpx_codec_err_t ret = vpx_codec_enc_config_default(algo, &config, 0);
74  if (ret != VPX_CODEC_OK)
75    return ScopedVpxCodec();
76
77  SetCommonCodecParameters(size, &config);
78
79  // Value of 2 means using the real time profile. This is basically a
80  // redundant option since we explicitly select real time mode when doing
81  // encoding.
82  config.g_profile = 2;
83
84  // Clamping the quantizer constrains the worst-case quality and CPU usage.
85  config.rc_min_quantizer = 20;
86  config.rc_max_quantizer = 30;
87
88  if (vpx_codec_enc_init(codec.get(), algo, &config, 0))
89    return ScopedVpxCodec();
90
91  // Value of 16 will have the smallest CPU load. This turns off subpixel
92  // motion search.
93  if (vpx_codec_control(codec.get(), VP8E_SET_CPUUSED, 16))
94    return ScopedVpxCodec();
95
96  // Use the lowest level of noise sensitivity so as to spend less time
97  // on motion estimation and inter-prediction mode.
98  if (vpx_codec_control(codec.get(), VP8E_SET_NOISE_SENSITIVITY, 0))
99    return ScopedVpxCodec();
100
101  return codec.Pass();
102}
103
104ScopedVpxCodec CreateVP9Codec(const webrtc::DesktopSize& size,
105                              bool lossless_color,
106                              bool lossless_encode) {
107  ScopedVpxCodec codec(new vpx_codec_ctx_t);
108
109  // Configure the encoder.
110  vpx_codec_enc_cfg_t config;
111  const vpx_codec_iface_t* algo = vpx_codec_vp9_cx();
112  CHECK(algo);
113  vpx_codec_err_t ret = vpx_codec_enc_config_default(algo, &config, 0);
114  if (ret != VPX_CODEC_OK)
115    return ScopedVpxCodec();
116
117  SetCommonCodecParameters(size, &config);
118
119  // Configure VP9 for I420 or I444 source frames.
120  config.g_profile =
121      lossless_color ? kVp9I444ProfileNumber : kVp9I420ProfileNumber;
122
123  if (lossless_encode) {
124    // Disable quantization entirely, putting the encoder in "lossless" mode.
125    config.rc_min_quantizer = 0;
126    config.rc_max_quantizer = 0;
127  } else {
128    // Lossy encode using the same settings as for VP8.
129    config.rc_min_quantizer = 20;
130    config.rc_max_quantizer = 30;
131  }
132
133  if (vpx_codec_enc_init(codec.get(), algo, &config, 0))
134    return ScopedVpxCodec();
135
136  // Request the lowest-CPU usage that VP9 supports, which depends on whether
137  // we are encoding lossy or lossless.
138  // Note that this is configured via the same parameter as for VP8.
139  int cpu_used = lossless_encode ? 5 : 7;
140  if (vpx_codec_control(codec.get(), VP8E_SET_CPUUSED, cpu_used))
141    return ScopedVpxCodec();
142
143  // Use the lowest level of noise sensitivity so as to spend less time
144  // on motion estimation and inter-prediction mode.
145  if (vpx_codec_control(codec.get(), VP9E_SET_NOISE_SENSITIVITY, 0))
146    return ScopedVpxCodec();
147
148  return codec.Pass();
149}
150
151void CreateImage(bool use_i444,
152                 const webrtc::DesktopSize& size,
153                 scoped_ptr<vpx_image_t>* out_image,
154                 scoped_ptr<uint8[]>* out_image_buffer) {
155  DCHECK(!size.is_empty());
156
157  scoped_ptr<vpx_image_t> image(new vpx_image_t());
158  memset(image.get(), 0, sizeof(vpx_image_t));
159
160  // libvpx seems to require both to be assigned.
161  image->d_w = size.width();
162  image->w = size.width();
163  image->d_h = size.height();
164  image->h = size.height();
165
166  // libvpx should derive chroma shifts from|fmt| but currently has a bug:
167  // https://code.google.com/p/webm/issues/detail?id=627
168  if (use_i444) {
169    image->fmt = VPX_IMG_FMT_I444;
170    image->x_chroma_shift = 0;
171    image->y_chroma_shift = 0;
172  } else { // I420
173    image->fmt = VPX_IMG_FMT_YV12;
174    image->x_chroma_shift = 1;
175    image->y_chroma_shift = 1;
176  }
177
178  // libyuv's fast-path requires 16-byte aligned pointers and strides, so pad
179  // the Y, U and V planes' strides to multiples of 16 bytes.
180  const int y_stride = ((image->w - 1) & ~15) + 16;
181  const int uv_unaligned_stride = y_stride >> image->x_chroma_shift;
182  const int uv_stride = ((uv_unaligned_stride - 1) & ~15) + 16;
183
184  // libvpx accesses the source image in macro blocks, and will over-read
185  // if the image is not padded out to the next macroblock: crbug.com/119633.
186  // Pad the Y, U and V planes' height out to compensate.
187  // Assuming macroblocks are 16x16, aligning the planes' strides above also
188  // macroblock aligned them.
189  DCHECK_EQ(16, kMacroBlockSize);
190  const int y_rows = ((image->h - 1) & ~(kMacroBlockSize-1)) + kMacroBlockSize;
191  const int uv_rows = y_rows >> image->y_chroma_shift;
192
193  // Allocate a YUV buffer large enough for the aligned data & padding.
194  const int buffer_size = y_stride * y_rows + 2*uv_stride * uv_rows;
195  scoped_ptr<uint8[]> image_buffer(new uint8[buffer_size]);
196
197  // Reset image value to 128 so we just need to fill in the y plane.
198  memset(image_buffer.get(), 128, buffer_size);
199
200  // Fill in the information for |image_|.
201  unsigned char* uchar_buffer =
202      reinterpret_cast<unsigned char*>(image_buffer.get());
203  image->planes[0] = uchar_buffer;
204  image->planes[1] = image->planes[0] + y_stride * y_rows;
205  image->planes[2] = image->planes[1] + uv_stride * uv_rows;
206  image->stride[0] = y_stride;
207  image->stride[1] = uv_stride;
208  image->stride[2] = uv_stride;
209
210  *out_image = image.Pass();
211  *out_image_buffer = image_buffer.Pass();
212}
213
214} // namespace
215
216// static
217scoped_ptr<VideoEncoderVpx> VideoEncoderVpx::CreateForVP8() {
218  return scoped_ptr<VideoEncoderVpx>(new VideoEncoderVpx(false));
219}
220
221// static
222scoped_ptr<VideoEncoderVpx> VideoEncoderVpx::CreateForVP9() {
223  return scoped_ptr<VideoEncoderVpx>(new VideoEncoderVpx(true));
224}
225
226VideoEncoderVpx::~VideoEncoderVpx() {}
227
228void VideoEncoderVpx::SetLosslessEncode(bool want_lossless) {
229  if (use_vp9_ && (want_lossless != lossless_encode_)) {
230    lossless_encode_ = want_lossless;
231    codec_.reset(); // Force encoder re-initialization.
232  }
233}
234
235void VideoEncoderVpx::SetLosslessColor(bool want_lossless) {
236  if (use_vp9_ && (want_lossless != lossless_color_)) {
237    lossless_color_ = want_lossless;
238    codec_.reset(); // Force encoder re-initialization.
239  }
240}
241
242scoped_ptr<VideoPacket> VideoEncoderVpx::Encode(
243    const webrtc::DesktopFrame& frame) {
244  DCHECK_LE(32, frame.size().width());
245  DCHECK_LE(32, frame.size().height());
246
247  base::TimeTicks encode_start_time = base::TimeTicks::Now();
248
249  if (!codec_ ||
250      !frame.size().equals(webrtc::DesktopSize(image_->w, image_->h))) {
251    bool ret = Initialize(frame.size());
252    // TODO(hclam): Handle error better.
253    CHECK(ret) << "Initialization of encoder failed";
254
255    // Set now as the base for timestamp calculation.
256    timestamp_base_ = encode_start_time;
257  }
258
259  // Convert the updated capture data ready for encode.
260  webrtc::DesktopRegion updated_region;
261  PrepareImage(frame, &updated_region);
262
263  // Update active map based on updated region.
264  PrepareActiveMap(updated_region);
265
266  // Apply active map to the encoder.
267  vpx_active_map_t act_map;
268  act_map.rows = active_map_height_;
269  act_map.cols = active_map_width_;
270  act_map.active_map = active_map_.get();
271  if (vpx_codec_control(codec_.get(), VP8E_SET_ACTIVEMAP, &act_map)) {
272    LOG(ERROR) << "Unable to apply active map";
273  }
274
275  // Do the actual encoding.
276  int timestamp = (encode_start_time - timestamp_base_).InMilliseconds();
277  vpx_codec_err_t ret = vpx_codec_encode(
278      codec_.get(), image_.get(), timestamp, 1, 0, VPX_DL_REALTIME);
279  DCHECK_EQ(ret, VPX_CODEC_OK)
280      << "Encoding error: " << vpx_codec_err_to_string(ret) << "\n"
281      << "Details: " << vpx_codec_error(codec_.get()) << "\n"
282      << vpx_codec_error_detail(codec_.get());
283
284  // Read the encoded data.
285  vpx_codec_iter_t iter = NULL;
286  bool got_data = false;
287
288  // TODO(hclam): Make sure we get exactly one frame from the packet.
289  // TODO(hclam): We should provide the output buffer to avoid one copy.
290  scoped_ptr<VideoPacket> packet(
291      helper_.CreateVideoPacketWithUpdatedRegion(frame, updated_region));
292  packet->mutable_format()->set_encoding(VideoPacketFormat::ENCODING_VP8);
293
294  while (!got_data) {
295    const vpx_codec_cx_pkt_t* vpx_packet =
296        vpx_codec_get_cx_data(codec_.get(), &iter);
297    if (!vpx_packet)
298      continue;
299
300    switch (vpx_packet->kind) {
301      case VPX_CODEC_CX_FRAME_PKT:
302        got_data = true;
303        packet->set_data(vpx_packet->data.frame.buf, vpx_packet->data.frame.sz);
304        break;
305      default:
306        break;
307    }
308  }
309
310  // Note the time taken to encode the pixel data.
311  packet->set_encode_time_ms(
312      (base::TimeTicks::Now() - encode_start_time).InMillisecondsRoundedUp());
313
314  return packet.Pass();
315}
316
317VideoEncoderVpx::VideoEncoderVpx(bool use_vp9)
318    : use_vp9_(use_vp9),
319      lossless_encode_(false),
320      lossless_color_(false),
321      active_map_width_(0),
322      active_map_height_(0) {
323  if (use_vp9_) {
324    // Use I444 colour space, by default, if specified on the command-line.
325    if (CommandLine::ForCurrentProcess()->HasSwitch(kEnableI444SwitchName)) {
326      SetLosslessColor(true);
327    }
328  }
329}
330
331bool VideoEncoderVpx::Initialize(const webrtc::DesktopSize& size) {
332  DCHECK(use_vp9_ || !lossless_color_);
333  DCHECK(use_vp9_ || !lossless_encode_);
334
335  codec_.reset();
336
337  // (Re)Create the VPX image structure and pixel buffer.
338  CreateImage(lossless_color_, size, &image_, &image_buffer_);
339
340  // Initialize active map.
341  active_map_width_ = (image_->w + kMacroBlockSize - 1) / kMacroBlockSize;
342  active_map_height_ = (image_->h + kMacroBlockSize - 1) / kMacroBlockSize;
343  active_map_.reset(new uint8[active_map_width_ * active_map_height_]);
344
345  // (Re)Initialize the codec.
346  if (use_vp9_) {
347    codec_ = CreateVP9Codec(size, lossless_color_, lossless_encode_);
348  } else {
349    codec_ = CreateVP8Codec(size);
350  }
351
352  return codec_;
353}
354
355void VideoEncoderVpx::PrepareImage(const webrtc::DesktopFrame& frame,
356                                   webrtc::DesktopRegion* updated_region) {
357  if (frame.updated_region().is_empty()) {
358    updated_region->Clear();
359    return;
360  }
361
362  // Align the region to macroblocks, to avoid encoding artefacts.
363  // This also ensures that all rectangles have even-aligned top-left, which
364  // is required for ConvertRGBToYUVWithRect() to work.
365  std::vector<webrtc::DesktopRect> aligned_rects;
366  for (webrtc::DesktopRegion::Iterator r(frame.updated_region());
367       !r.IsAtEnd(); r.Advance()) {
368    const webrtc::DesktopRect& rect = r.rect();
369    aligned_rects.push_back(AlignRect(webrtc::DesktopRect::MakeLTRB(
370        rect.left(), rect.top(), rect.right(), rect.bottom())));
371  }
372  DCHECK(!aligned_rects.empty());
373  updated_region->Clear();
374  updated_region->AddRects(&aligned_rects[0], aligned_rects.size());
375
376  // Clip back to the screen dimensions, in case they're not macroblock aligned.
377  // The conversion routines don't require even width & height, so this is safe
378  // even if the source dimensions are not even.
379  updated_region->IntersectWith(
380      webrtc::DesktopRect::MakeWH(image_->w, image_->h));
381
382  // Convert the updated region to YUV ready for encoding.
383  const uint8* rgb_data = frame.data();
384  const int rgb_stride = frame.stride();
385  const int y_stride = image_->stride[0];
386  DCHECK_EQ(image_->stride[1], image_->stride[2]);
387  const int uv_stride = image_->stride[1];
388  uint8* y_data = image_->planes[0];
389  uint8* u_data = image_->planes[1];
390  uint8* v_data = image_->planes[2];
391
392  switch (image_->fmt) {
393    case VPX_IMG_FMT_I444:
394      for (webrtc::DesktopRegion::Iterator r(*updated_region); !r.IsAtEnd();
395           r.Advance()) {
396        const webrtc::DesktopRect& rect = r.rect();
397        int rgb_offset = rgb_stride * rect.top() +
398                         rect.left() * kBytesPerRgbPixel;
399        int yuv_offset = uv_stride * rect.top() + rect.left();
400        libyuv::ARGBToI444(rgb_data + rgb_offset, rgb_stride,
401                           y_data + yuv_offset, y_stride,
402                           u_data + yuv_offset, uv_stride,
403                           v_data + yuv_offset, uv_stride,
404                           rect.width(), rect.height());
405      }
406      break;
407    case VPX_IMG_FMT_YV12:
408      for (webrtc::DesktopRegion::Iterator r(*updated_region); !r.IsAtEnd();
409           r.Advance()) {
410        const webrtc::DesktopRect& rect = r.rect();
411        int rgb_offset = rgb_stride * rect.top() +
412                         rect.left() * kBytesPerRgbPixel;
413        int y_offset = y_stride * rect.top() + rect.left();
414        int uv_offset = uv_stride * rect.top() / 2 + rect.left() / 2;
415        libyuv::ARGBToI420(rgb_data + rgb_offset, rgb_stride,
416                           y_data + y_offset, y_stride,
417                           u_data + uv_offset, uv_stride,
418                           v_data + uv_offset, uv_stride,
419                           rect.width(), rect.height());
420      }
421      break;
422    default:
423      NOTREACHED();
424      break;
425  }
426}
427
428void VideoEncoderVpx::PrepareActiveMap(
429    const webrtc::DesktopRegion& updated_region) {
430  // Clear active map first.
431  memset(active_map_.get(), 0, active_map_width_ * active_map_height_);
432
433  // Mark updated areas active.
434  for (webrtc::DesktopRegion::Iterator r(updated_region); !r.IsAtEnd();
435       r.Advance()) {
436    const webrtc::DesktopRect& rect = r.rect();
437    int left = rect.left() / kMacroBlockSize;
438    int right = (rect.right() - 1) / kMacroBlockSize;
439    int top = rect.top() / kMacroBlockSize;
440    int bottom = (rect.bottom() - 1) / kMacroBlockSize;
441    DCHECK_LT(right, active_map_width_);
442    DCHECK_LT(bottom, active_map_height_);
443
444    uint8* map = active_map_.get() + top * active_map_width_;
445    for (int y = top; y <= bottom; ++y) {
446      for (int x = left; x <= right; ++x)
447        map[x] = 1;
448      map += active_map_width_;
449    }
450  }
451}
452
453}  // namespace remoting
454