SoftAVCEncoder.cpp revision 84333e0475bc911adc16417f4ca327c975cf6c36
1/*
2 * Copyright (C) 2012 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 *      http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17//#define LOG_NDEBUG 0
18#define LOG_TAG "SoftAVCEncoder"
19#include <utils/Log.h>
20
21#include "avcenc_api.h"
22#include "avcenc_int.h"
23#include "OMX_Video.h"
24
25#include <HardwareAPI.h>
26#include <MetadataBufferType.h>
27#include <media/stagefright/foundation/ADebug.h>
28#include <media/stagefright/MediaDefs.h>
29#include <media/stagefright/MediaErrors.h>
30#include <media/stagefright/MetaData.h>
31#include <media/stagefright/Utils.h>
32#include <ui/Rect.h>
33#include <ui/GraphicBufferMapper.h>
34
35#include "SoftAVCEncoder.h"
36
37#if LOG_NDEBUG
38#define UNUSED_UNLESS_VERBOSE(x) (void)(x)
39#else
40#define UNUSED_UNLESS_VERBOSE(x)
41#endif
42
43namespace android {
44
45template<class T>
46static void InitOMXParams(T *params) {
47    params->nSize = sizeof(T);
48    params->nVersion.s.nVersionMajor = 1;
49    params->nVersion.s.nVersionMinor = 0;
50    params->nVersion.s.nRevision = 0;
51    params->nVersion.s.nStep = 0;
52}
53
54typedef struct LevelConversion {
55    OMX_U32 omxLevel;
56    AVCLevel avcLevel;
57} LevelConcersion;
58
59static LevelConversion ConversionTable[] = {
60    { OMX_VIDEO_AVCLevel1,  AVC_LEVEL1_B },
61    { OMX_VIDEO_AVCLevel1b, AVC_LEVEL1   },
62    { OMX_VIDEO_AVCLevel11, AVC_LEVEL1_1 },
63    { OMX_VIDEO_AVCLevel12, AVC_LEVEL1_2 },
64    { OMX_VIDEO_AVCLevel13, AVC_LEVEL1_3 },
65    { OMX_VIDEO_AVCLevel2,  AVC_LEVEL2 },
66#if 0
67    // encoding speed is very poor if video
68    // resolution is higher than CIF
69    { OMX_VIDEO_AVCLevel21, AVC_LEVEL2_1 },
70    { OMX_VIDEO_AVCLevel22, AVC_LEVEL2_2 },
71    { OMX_VIDEO_AVCLevel3,  AVC_LEVEL3   },
72    { OMX_VIDEO_AVCLevel31, AVC_LEVEL3_1 },
73    { OMX_VIDEO_AVCLevel32, AVC_LEVEL3_2 },
74    { OMX_VIDEO_AVCLevel4,  AVC_LEVEL4   },
75    { OMX_VIDEO_AVCLevel41, AVC_LEVEL4_1 },
76    { OMX_VIDEO_AVCLevel42, AVC_LEVEL4_2 },
77    { OMX_VIDEO_AVCLevel5,  AVC_LEVEL5   },
78    { OMX_VIDEO_AVCLevel51, AVC_LEVEL5_1 },
79#endif
80};
81
82static status_t ConvertOmxAvcLevelToAvcSpecLevel(
83        OMX_U32 omxLevel, AVCLevel *avcLevel) {
84    for (size_t i = 0, n = sizeof(ConversionTable)/sizeof(ConversionTable[0]);
85        i < n; ++i) {
86        if (omxLevel == ConversionTable[i].omxLevel) {
87            *avcLevel = ConversionTable[i].avcLevel;
88            return OK;
89        }
90    }
91
92    ALOGE("ConvertOmxAvcLevelToAvcSpecLevel: %d level not supported",
93            (int32_t)omxLevel);
94
95    return BAD_VALUE;
96}
97
98static status_t ConvertAvcSpecLevelToOmxAvcLevel(
99    AVCLevel avcLevel, OMX_U32 *omxLevel) {
100    for (size_t i = 0, n = sizeof(ConversionTable)/sizeof(ConversionTable[0]);
101        i < n; ++i) {
102        if (avcLevel == ConversionTable[i].avcLevel) {
103            *omxLevel = ConversionTable[i].omxLevel;
104            return OK;
105        }
106    }
107
108    ALOGE("ConvertAvcSpecLevelToOmxAvcLevel: %d level not supported",
109            (int32_t) avcLevel);
110
111    return BAD_VALUE;
112}
113
114inline static void ConvertYUV420SemiPlanarToYUV420Planar(
115        uint8_t *inyuv, uint8_t* outyuv,
116        int32_t width, int32_t height) {
117
118    int32_t outYsize = width * height;
119    uint32_t *outy =  (uint32_t *) outyuv;
120    uint16_t *outcb = (uint16_t *) (outyuv + outYsize);
121    uint16_t *outcr = (uint16_t *) (outyuv + outYsize + (outYsize >> 2));
122
123    /* Y copying */
124    memcpy(outy, inyuv, outYsize);
125
126    /* U & V copying */
127    uint32_t *inyuv_4 = (uint32_t *) (inyuv + outYsize);
128    for (int32_t i = height >> 1; i > 0; --i) {
129        for (int32_t j = width >> 2; j > 0; --j) {
130            uint32_t temp = *inyuv_4++;
131            uint32_t tempU = temp & 0xFF;
132            tempU = tempU | ((temp >> 8) & 0xFF00);
133
134            uint32_t tempV = (temp >> 8) & 0xFF;
135            tempV = tempV | ((temp >> 16) & 0xFF00);
136
137            // Flip U and V
138            *outcb++ = tempV;
139            *outcr++ = tempU;
140        }
141    }
142}
143
144static void* MallocWrapper(
145        void * /* userData */, int32_t size, int32_t /* attrs */) {
146    void *ptr = malloc(size);
147    if (ptr)
148        memset(ptr, 0, size);
149    return ptr;
150}
151
152static void FreeWrapper(void * /* userData */, void* ptr) {
153    free(ptr);
154}
155
156static int32_t DpbAllocWrapper(void *userData,
157        unsigned int sizeInMbs, unsigned int numBuffers) {
158    SoftAVCEncoder *encoder = static_cast<SoftAVCEncoder *>(userData);
159    CHECK(encoder != NULL);
160    return encoder->allocOutputBuffers(sizeInMbs, numBuffers);
161}
162
163static int32_t BindFrameWrapper(
164        void *userData, int32_t index, uint8_t **yuv) {
165    SoftAVCEncoder *encoder = static_cast<SoftAVCEncoder *>(userData);
166    CHECK(encoder != NULL);
167    return encoder->bindOutputBuffer(index, yuv);
168}
169
170static void UnbindFrameWrapper(void *userData, int32_t index) {
171    SoftAVCEncoder *encoder = static_cast<SoftAVCEncoder *>(userData);
172    CHECK(encoder != NULL);
173    return encoder->unbindOutputBuffer(index);
174}
175
176SoftAVCEncoder::SoftAVCEncoder(
177            const char *name,
178            const OMX_CALLBACKTYPE *callbacks,
179            OMX_PTR appData,
180            OMX_COMPONENTTYPE **component)
181    : SimpleSoftOMXComponent(name, callbacks, appData, component),
182      mVideoWidth(176),
183      mVideoHeight(144),
184      mVideoFrameRate(30),
185      mVideoBitRate(192000),
186      mVideoColorFormat(OMX_COLOR_FormatYUV420Planar),
187      mStoreMetaDataInBuffers(false),
188      mIDRFrameRefreshIntervalInSec(1),
189      mAVCEncProfile(AVC_BASELINE),
190      mAVCEncLevel(AVC_LEVEL2),
191      mNumInputFrames(-1),
192      mPrevTimestampUs(-1),
193      mStarted(false),
194      mSawInputEOS(false),
195      mSignalledError(false),
196      mHandle(new tagAVCHandle),
197      mEncParams(new tagAVCEncParam),
198      mInputFrameData(NULL),
199      mSliceGroup(NULL) {
200
201    initPorts();
202    ALOGI("Construct SoftAVCEncoder");
203}
204
205SoftAVCEncoder::~SoftAVCEncoder() {
206    ALOGV("Destruct SoftAVCEncoder");
207    releaseEncoder();
208    List<BufferInfo *> &outQueue = getPortQueue(1);
209    List<BufferInfo *> &inQueue = getPortQueue(0);
210    CHECK(outQueue.empty());
211    CHECK(inQueue.empty());
212}
213
214OMX_ERRORTYPE SoftAVCEncoder::initEncParams() {
215    CHECK(mHandle != NULL);
216    memset(mHandle, 0, sizeof(tagAVCHandle));
217    mHandle->AVCObject = NULL;
218    mHandle->userData = this;
219    mHandle->CBAVC_DPBAlloc = DpbAllocWrapper;
220    mHandle->CBAVC_FrameBind = BindFrameWrapper;
221    mHandle->CBAVC_FrameUnbind = UnbindFrameWrapper;
222    mHandle->CBAVC_Malloc = MallocWrapper;
223    mHandle->CBAVC_Free = FreeWrapper;
224
225    CHECK(mEncParams != NULL);
226    memset(mEncParams, 0, sizeof(*mEncParams));
227    mEncParams->rate_control = AVC_ON;
228    mEncParams->initQP = 0;
229    mEncParams->init_CBP_removal_delay = 1600;
230
231    mEncParams->intramb_refresh = 0;
232    mEncParams->auto_scd = AVC_ON;
233    mEncParams->out_of_band_param_set = AVC_ON;
234    mEncParams->poc_type = 2;
235    mEncParams->log2_max_poc_lsb_minus_4 = 12;
236    mEncParams->delta_poc_zero_flag = 0;
237    mEncParams->offset_poc_non_ref = 0;
238    mEncParams->offset_top_bottom = 0;
239    mEncParams->num_ref_in_cycle = 0;
240    mEncParams->offset_poc_ref = NULL;
241
242    mEncParams->num_ref_frame = 1;
243    mEncParams->num_slice_group = 1;
244    mEncParams->fmo_type = 0;
245
246    mEncParams->db_filter = AVC_ON;
247    mEncParams->disable_db_idc = 0;
248
249    mEncParams->alpha_offset = 0;
250    mEncParams->beta_offset = 0;
251    mEncParams->constrained_intra_pred = AVC_OFF;
252
253    mEncParams->data_par = AVC_OFF;
254    mEncParams->fullsearch = AVC_OFF;
255    mEncParams->search_range = 16;
256    mEncParams->sub_pel = AVC_OFF;
257    mEncParams->submb_pred = AVC_OFF;
258    mEncParams->rdopt_mode = AVC_OFF;
259    mEncParams->bidir_pred = AVC_OFF;
260
261    mEncParams->use_overrun_buffer = AVC_OFF;
262
263    if (mVideoColorFormat == OMX_COLOR_FormatYUV420SemiPlanar) {
264        // Color conversion is needed.
265        CHECK(mInputFrameData == NULL);
266        mInputFrameData =
267            (uint8_t *) malloc((mVideoWidth * mVideoHeight * 3 ) >> 1);
268        CHECK(mInputFrameData != NULL);
269    }
270
271    // PV's AVC encoder requires the video dimension of multiple
272    if (mVideoWidth % 16 != 0 || mVideoHeight % 16 != 0) {
273        ALOGE("Video frame size %dx%d must be a multiple of 16",
274            mVideoWidth, mVideoHeight);
275        return OMX_ErrorBadParameter;
276    }
277
278    mEncParams->width = mVideoWidth;
279    mEncParams->height = mVideoHeight;
280    mEncParams->bitrate = mVideoBitRate;
281    mEncParams->frame_rate = 1000 * mVideoFrameRate;  // In frames/ms!
282    mEncParams->CPB_size = (uint32_t) (mVideoBitRate >> 1);
283
284    int32_t nMacroBlocks = ((((mVideoWidth + 15) >> 4) << 4) *
285            (((mVideoHeight + 15) >> 4) << 4)) >> 8;
286    CHECK(mSliceGroup == NULL);
287    mSliceGroup = (uint32_t *) malloc(sizeof(uint32_t) * nMacroBlocks);
288    CHECK(mSliceGroup != NULL);
289    for (int ii = 0, idx = 0; ii < nMacroBlocks; ++ii) {
290        mSliceGroup[ii] = idx++;
291        if (idx >= mEncParams->num_slice_group) {
292            idx = 0;
293        }
294    }
295    mEncParams->slice_group = mSliceGroup;
296
297    // Set IDR frame refresh interval
298    if (mIDRFrameRefreshIntervalInSec < 0) {
299        mEncParams->idr_period = -1;
300    } else if (mIDRFrameRefreshIntervalInSec == 0) {
301        mEncParams->idr_period = 1;  // All I frames
302    } else {
303        mEncParams->idr_period =
304            (mIDRFrameRefreshIntervalInSec * mVideoFrameRate);
305    }
306
307    // Set profile and level
308    mEncParams->profile = mAVCEncProfile;
309    mEncParams->level = mAVCEncLevel;
310
311    return OMX_ErrorNone;
312}
313
314OMX_ERRORTYPE SoftAVCEncoder::initEncoder() {
315    CHECK(!mStarted);
316
317    OMX_ERRORTYPE errType = OMX_ErrorNone;
318    if (OMX_ErrorNone != (errType = initEncParams())) {
319        ALOGE("Failed to initialized encoder params");
320        mSignalledError = true;
321        notify(OMX_EventError, OMX_ErrorUndefined, 0, 0);
322        return errType;
323    }
324
325    AVCEnc_Status err;
326    err = PVAVCEncInitialize(mHandle, mEncParams, NULL, NULL);
327    if (err != AVCENC_SUCCESS) {
328        ALOGE("Failed to initialize the encoder: %d", err);
329        mSignalledError = true;
330        notify(OMX_EventError, OMX_ErrorUndefined, 0, 0);
331        return OMX_ErrorUndefined;
332    }
333
334    mNumInputFrames = -2;  // 1st two buffers contain SPS and PPS
335    mSpsPpsHeaderReceived = false;
336    mReadyForNextFrame = true;
337    mIsIDRFrame = false;
338    mStarted = true;
339
340    return OMX_ErrorNone;
341}
342
343OMX_ERRORTYPE SoftAVCEncoder::releaseEncoder() {
344    if (!mStarted) {
345        return OMX_ErrorNone;
346    }
347
348    PVAVCCleanUpEncoder(mHandle);
349    releaseOutputBuffers();
350
351    delete mInputFrameData;
352    mInputFrameData = NULL;
353
354    delete mSliceGroup;
355    mSliceGroup = NULL;
356
357    delete mEncParams;
358    mEncParams = NULL;
359
360    delete mHandle;
361    mHandle = NULL;
362
363    mStarted = false;
364
365    return OMX_ErrorNone;
366}
367
368void SoftAVCEncoder::releaseOutputBuffers() {
369    for (size_t i = 0; i < mOutputBuffers.size(); ++i) {
370        MediaBuffer *buffer = mOutputBuffers.editItemAt(i);
371        buffer->setObserver(NULL);
372        buffer->release();
373    }
374    mOutputBuffers.clear();
375}
376
377void SoftAVCEncoder::initPorts() {
378    OMX_PARAM_PORTDEFINITIONTYPE def;
379    InitOMXParams(&def);
380
381    const size_t kInputBufferSize = (mVideoWidth * mVideoHeight * 3) >> 1;
382
383    // 31584 is PV's magic number.  Not sure why.
384    const size_t kOutputBufferSize =
385            (kInputBufferSize > 31584) ? kInputBufferSize: 31584;
386
387    def.nPortIndex = 0;
388    def.eDir = OMX_DirInput;
389    def.nBufferCountMin = kNumBuffers;
390    def.nBufferCountActual = def.nBufferCountMin;
391    def.nBufferSize = kInputBufferSize;
392    def.bEnabled = OMX_TRUE;
393    def.bPopulated = OMX_FALSE;
394    def.eDomain = OMX_PortDomainVideo;
395    def.bBuffersContiguous = OMX_FALSE;
396    def.nBufferAlignment = 1;
397
398    def.format.video.cMIMEType = const_cast<char *>("video/raw");
399    def.format.video.eCompressionFormat = OMX_VIDEO_CodingUnused;
400    def.format.video.eColorFormat = OMX_COLOR_FormatYUV420Planar;
401    def.format.video.xFramerate = (mVideoFrameRate << 16);  // Q16 format
402    def.format.video.nBitrate = mVideoBitRate;
403    def.format.video.nFrameWidth = mVideoWidth;
404    def.format.video.nFrameHeight = mVideoHeight;
405    def.format.video.nStride = mVideoWidth;
406    def.format.video.nSliceHeight = mVideoHeight;
407
408    addPort(def);
409
410    def.nPortIndex = 1;
411    def.eDir = OMX_DirOutput;
412    def.nBufferCountMin = kNumBuffers;
413    def.nBufferCountActual = def.nBufferCountMin;
414    def.nBufferSize = kOutputBufferSize;
415    def.bEnabled = OMX_TRUE;
416    def.bPopulated = OMX_FALSE;
417    def.eDomain = OMX_PortDomainVideo;
418    def.bBuffersContiguous = OMX_FALSE;
419    def.nBufferAlignment = 2;
420
421    def.format.video.cMIMEType = const_cast<char *>("video/avc");
422    def.format.video.eCompressionFormat = OMX_VIDEO_CodingAVC;
423    def.format.video.eColorFormat = OMX_COLOR_FormatUnused;
424    def.format.video.xFramerate = (0 << 16);  // Q16 format
425    def.format.video.nBitrate = mVideoBitRate;
426    def.format.video.nFrameWidth = mVideoWidth;
427    def.format.video.nFrameHeight = mVideoHeight;
428    def.format.video.nStride = mVideoWidth;
429    def.format.video.nSliceHeight = mVideoHeight;
430
431    addPort(def);
432}
433
434OMX_ERRORTYPE SoftAVCEncoder::internalGetParameter(
435        OMX_INDEXTYPE index, OMX_PTR params) {
436    switch (index) {
437        case OMX_IndexParamVideoErrorCorrection:
438        {
439            return OMX_ErrorNotImplemented;
440        }
441
442        case OMX_IndexParamVideoBitrate:
443        {
444            OMX_VIDEO_PARAM_BITRATETYPE *bitRate =
445                (OMX_VIDEO_PARAM_BITRATETYPE *) params;
446
447            if (bitRate->nPortIndex != 1) {
448                return OMX_ErrorUndefined;
449            }
450
451            bitRate->eControlRate = OMX_Video_ControlRateVariable;
452            bitRate->nTargetBitrate = mVideoBitRate;
453            return OMX_ErrorNone;
454        }
455
456        case OMX_IndexParamVideoPortFormat:
457        {
458            OMX_VIDEO_PARAM_PORTFORMATTYPE *formatParams =
459                (OMX_VIDEO_PARAM_PORTFORMATTYPE *)params;
460
461            if (formatParams->nPortIndex > 1) {
462                return OMX_ErrorUndefined;
463            }
464
465            if (formatParams->nIndex > 2) {
466                return OMX_ErrorNoMore;
467            }
468
469            if (formatParams->nPortIndex == 0) {
470                formatParams->eCompressionFormat = OMX_VIDEO_CodingUnused;
471                if (formatParams->nIndex == 0) {
472                    formatParams->eColorFormat = OMX_COLOR_FormatYUV420Planar;
473                } else if (formatParams->nIndex == 1) {
474                    formatParams->eColorFormat = OMX_COLOR_FormatYUV420SemiPlanar;
475                } else {
476                    formatParams->eColorFormat = OMX_COLOR_FormatAndroidOpaque;
477                }
478            } else {
479                formatParams->eCompressionFormat = OMX_VIDEO_CodingAVC;
480                formatParams->eColorFormat = OMX_COLOR_FormatUnused;
481            }
482
483            return OMX_ErrorNone;
484        }
485
486        case OMX_IndexParamVideoAvc:
487        {
488            OMX_VIDEO_PARAM_AVCTYPE *avcParams =
489                (OMX_VIDEO_PARAM_AVCTYPE *)params;
490
491            if (avcParams->nPortIndex != 1) {
492                return OMX_ErrorUndefined;
493            }
494
495            avcParams->eProfile = OMX_VIDEO_AVCProfileBaseline;
496            OMX_U32 omxLevel = AVC_LEVEL2;
497            if (OMX_ErrorNone !=
498                ConvertAvcSpecLevelToOmxAvcLevel(mAVCEncLevel, &omxLevel)) {
499                return OMX_ErrorUndefined;
500            }
501
502            avcParams->eLevel = (OMX_VIDEO_AVCLEVELTYPE) omxLevel;
503            avcParams->nRefFrames = 1;
504            avcParams->nBFrames = 0;
505            avcParams->bUseHadamard = OMX_TRUE;
506            avcParams->nAllowedPictureTypes =
507                    (OMX_VIDEO_PictureTypeI | OMX_VIDEO_PictureTypeP);
508            avcParams->nRefIdx10ActiveMinus1 = 0;
509            avcParams->nRefIdx11ActiveMinus1 = 0;
510            avcParams->bWeightedPPrediction = OMX_FALSE;
511            avcParams->bEntropyCodingCABAC = OMX_FALSE;
512            avcParams->bconstIpred = OMX_FALSE;
513            avcParams->bDirect8x8Inference = OMX_FALSE;
514            avcParams->bDirectSpatialTemporal = OMX_FALSE;
515            avcParams->nCabacInitIdc = 0;
516            return OMX_ErrorNone;
517        }
518
519        case OMX_IndexParamVideoProfileLevelQuerySupported:
520        {
521            OMX_VIDEO_PARAM_PROFILELEVELTYPE *profileLevel =
522                (OMX_VIDEO_PARAM_PROFILELEVELTYPE *)params;
523
524            if (profileLevel->nPortIndex != 1) {
525                return OMX_ErrorUndefined;
526            }
527
528            const size_t size =
529                    sizeof(ConversionTable) / sizeof(ConversionTable[0]);
530
531            if (profileLevel->nProfileIndex >= size) {
532                return OMX_ErrorNoMore;
533            }
534
535            profileLevel->eProfile = OMX_VIDEO_AVCProfileBaseline;
536            profileLevel->eLevel = ConversionTable[profileLevel->nProfileIndex].omxLevel;
537
538            return OMX_ErrorNone;
539        }
540
541        default:
542            return SimpleSoftOMXComponent::internalGetParameter(index, params);
543    }
544}
545
546OMX_ERRORTYPE SoftAVCEncoder::internalSetParameter(
547        OMX_INDEXTYPE index, const OMX_PTR params) {
548    int32_t indexFull = index;
549
550    switch (indexFull) {
551        case OMX_IndexParamVideoErrorCorrection:
552        {
553            return OMX_ErrorNotImplemented;
554        }
555
556        case OMX_IndexParamVideoBitrate:
557        {
558            OMX_VIDEO_PARAM_BITRATETYPE *bitRate =
559                (OMX_VIDEO_PARAM_BITRATETYPE *) params;
560
561            if (bitRate->nPortIndex != 1 ||
562                bitRate->eControlRate != OMX_Video_ControlRateVariable) {
563                return OMX_ErrorUndefined;
564            }
565
566            mVideoBitRate = bitRate->nTargetBitrate;
567            return OMX_ErrorNone;
568        }
569
570        case OMX_IndexParamPortDefinition:
571        {
572            OMX_PARAM_PORTDEFINITIONTYPE *def =
573                (OMX_PARAM_PORTDEFINITIONTYPE *)params;
574            if (def->nPortIndex > 1) {
575                return OMX_ErrorUndefined;
576            }
577
578            if (def->nPortIndex == 0) {
579                if (def->format.video.eCompressionFormat != OMX_VIDEO_CodingUnused ||
580                    (def->format.video.eColorFormat != OMX_COLOR_FormatYUV420Planar &&
581                     def->format.video.eColorFormat != OMX_COLOR_FormatYUV420SemiPlanar &&
582                     def->format.video.eColorFormat != OMX_COLOR_FormatAndroidOpaque)) {
583                    return OMX_ErrorUndefined;
584                }
585            } else {
586                if (def->format.video.eCompressionFormat != OMX_VIDEO_CodingAVC ||
587                    (def->format.video.eColorFormat != OMX_COLOR_FormatUnused)) {
588                    return OMX_ErrorUndefined;
589                }
590            }
591
592            OMX_ERRORTYPE err = SimpleSoftOMXComponent::internalSetParameter(index, params);
593            if (OMX_ErrorNone != err) {
594                return err;
595            }
596
597            if (def->nPortIndex == 0) {
598                mVideoWidth = def->format.video.nFrameWidth;
599                mVideoHeight = def->format.video.nFrameHeight;
600                mVideoFrameRate = def->format.video.xFramerate >> 16;
601                mVideoColorFormat = def->format.video.eColorFormat;
602
603                OMX_PARAM_PORTDEFINITIONTYPE *portDef =
604                    &editPortInfo(0)->mDef;
605                portDef->format.video.nFrameWidth = mVideoWidth;
606                portDef->format.video.nFrameHeight = mVideoHeight;
607                portDef->format.video.xFramerate = def->format.video.xFramerate;
608                portDef->format.video.eColorFormat =
609                    (OMX_COLOR_FORMATTYPE) mVideoColorFormat;
610                portDef = &editPortInfo(1)->mDef;
611                portDef->format.video.nFrameWidth = mVideoWidth;
612                portDef->format.video.nFrameHeight = mVideoHeight;
613            } else {
614                mVideoBitRate = def->format.video.nBitrate;
615            }
616
617            return OMX_ErrorNone;
618        }
619
620        case OMX_IndexParamStandardComponentRole:
621        {
622            const OMX_PARAM_COMPONENTROLETYPE *roleParams =
623                (const OMX_PARAM_COMPONENTROLETYPE *)params;
624
625            if (strncmp((const char *)roleParams->cRole,
626                        "video_encoder.avc",
627                        OMX_MAX_STRINGNAME_SIZE - 1)) {
628                return OMX_ErrorUndefined;
629            }
630
631            return OMX_ErrorNone;
632        }
633
634        case OMX_IndexParamVideoPortFormat:
635        {
636            const OMX_VIDEO_PARAM_PORTFORMATTYPE *formatParams =
637                (const OMX_VIDEO_PARAM_PORTFORMATTYPE *)params;
638
639            if (formatParams->nPortIndex > 1) {
640                return OMX_ErrorUndefined;
641            }
642
643            if (formatParams->nIndex > 2) {
644                return OMX_ErrorNoMore;
645            }
646
647            if (formatParams->nPortIndex == 0) {
648                if (formatParams->eCompressionFormat != OMX_VIDEO_CodingUnused ||
649                    ((formatParams->nIndex == 0 &&
650                      formatParams->eColorFormat != OMX_COLOR_FormatYUV420Planar) ||
651                    (formatParams->nIndex == 1 &&
652                     formatParams->eColorFormat != OMX_COLOR_FormatYUV420SemiPlanar) ||
653                    (formatParams->nIndex == 2 &&
654                     formatParams->eColorFormat != OMX_COLOR_FormatAndroidOpaque) )) {
655                    return OMX_ErrorUndefined;
656                }
657                mVideoColorFormat = formatParams->eColorFormat;
658            } else {
659                if (formatParams->eCompressionFormat != OMX_VIDEO_CodingAVC ||
660                    formatParams->eColorFormat != OMX_COLOR_FormatUnused) {
661                    return OMX_ErrorUndefined;
662                }
663            }
664
665            return OMX_ErrorNone;
666        }
667
668        case OMX_IndexParamVideoAvc:
669        {
670            OMX_VIDEO_PARAM_AVCTYPE *avcType =
671                (OMX_VIDEO_PARAM_AVCTYPE *)params;
672
673            if (avcType->nPortIndex != 1) {
674                return OMX_ErrorUndefined;
675            }
676
677            // PV's AVC encoder only supports baseline profile
678            if (avcType->eProfile != OMX_VIDEO_AVCProfileBaseline ||
679                avcType->nRefFrames != 1 ||
680                avcType->nBFrames != 0 ||
681                avcType->bUseHadamard != OMX_TRUE ||
682                (avcType->nAllowedPictureTypes & OMX_VIDEO_PictureTypeB) != 0 ||
683                avcType->nRefIdx10ActiveMinus1 != 0 ||
684                avcType->nRefIdx11ActiveMinus1 != 0 ||
685                avcType->bWeightedPPrediction != OMX_FALSE ||
686                avcType->bEntropyCodingCABAC != OMX_FALSE ||
687                avcType->bconstIpred != OMX_FALSE ||
688                avcType->bDirect8x8Inference != OMX_FALSE ||
689                avcType->bDirectSpatialTemporal != OMX_FALSE ||
690                avcType->nCabacInitIdc != 0) {
691                return OMX_ErrorUndefined;
692            }
693
694            if (OK != ConvertOmxAvcLevelToAvcSpecLevel(avcType->eLevel, &mAVCEncLevel)) {
695                return OMX_ErrorUndefined;
696            }
697
698            return OMX_ErrorNone;
699        }
700
701        case kStoreMetaDataExtensionIndex:
702        {
703            StoreMetaDataInBuffersParams *storeParams =
704                    (StoreMetaDataInBuffersParams*)params;
705            if (storeParams->nPortIndex != 0) {
706                ALOGE("%s: StoreMetadataInBuffersParams.nPortIndex not zero!",
707                        __FUNCTION__);
708                return OMX_ErrorUndefined;
709            }
710
711            mStoreMetaDataInBuffers = storeParams->bStoreMetaData;
712            ALOGV("StoreMetaDataInBuffers set to: %s",
713                    mStoreMetaDataInBuffers ? " true" : "false");
714
715            if (mStoreMetaDataInBuffers) {
716                mVideoColorFormat == OMX_COLOR_FormatYUV420SemiPlanar;
717                if (mInputFrameData == NULL) {
718                    mInputFrameData =
719                            (uint8_t *) malloc((mVideoWidth * mVideoHeight * 3 ) >> 1);
720                }
721            }
722
723            return OMX_ErrorNone;
724        }
725
726        default:
727            return SimpleSoftOMXComponent::internalSetParameter(index, params);
728    }
729}
730
731void SoftAVCEncoder::onQueueFilled(OMX_U32 /* portIndex */) {
732    if (mSignalledError || mSawInputEOS) {
733        return;
734    }
735
736    if (!mStarted) {
737        if (OMX_ErrorNone != initEncoder()) {
738            return;
739        }
740    }
741
742    List<BufferInfo *> &inQueue = getPortQueue(0);
743    List<BufferInfo *> &outQueue = getPortQueue(1);
744
745    while (!mSawInputEOS && !inQueue.empty() && !outQueue.empty()) {
746        BufferInfo *inInfo = *inQueue.begin();
747        OMX_BUFFERHEADERTYPE *inHeader = inInfo->mHeader;
748        BufferInfo *outInfo = *outQueue.begin();
749        OMX_BUFFERHEADERTYPE *outHeader = outInfo->mHeader;
750
751        outHeader->nTimeStamp = 0;
752        outHeader->nFlags = 0;
753        outHeader->nOffset = 0;
754        outHeader->nFilledLen = 0;
755        outHeader->nOffset = 0;
756
757        uint8_t *outPtr = (uint8_t *) outHeader->pBuffer;
758        uint32_t dataLength = outHeader->nAllocLen;
759
760        if (!mSpsPpsHeaderReceived && mNumInputFrames < 0) {
761            // 4 bytes are reserved for holding the start code 0x00000001
762            // of the sequence parameter set at the beginning.
763            outPtr += 4;
764            dataLength -= 4;
765        }
766
767        int32_t type;
768        AVCEnc_Status encoderStatus = AVCENC_SUCCESS;
769
770        // Combine SPS and PPS and place them in the very first output buffer
771        // SPS and PPS are separated by start code 0x00000001
772        // Assume that we have exactly one SPS and exactly one PPS.
773        while (!mSpsPpsHeaderReceived && mNumInputFrames <= 0) {
774            encoderStatus = PVAVCEncodeNAL(mHandle, outPtr, &dataLength, &type);
775            if (encoderStatus == AVCENC_WRONG_STATE) {
776                mSpsPpsHeaderReceived = true;
777                CHECK_EQ(0, mNumInputFrames);  // 1st video frame is 0
778                outHeader->nFlags = OMX_BUFFERFLAG_CODECCONFIG;
779                outQueue.erase(outQueue.begin());
780                outInfo->mOwnedByUs = false;
781                notifyFillBufferDone(outHeader);
782                return;
783            } else {
784                switch (type) {
785                    case AVC_NALTYPE_SPS:
786                        ++mNumInputFrames;
787                        memcpy((uint8_t *)outHeader->pBuffer, "\x00\x00\x00\x01", 4);
788                        outHeader->nFilledLen = 4 + dataLength;
789                        outPtr += (dataLength + 4);  // 4 bytes for next start code
790                        dataLength = outHeader->nAllocLen - outHeader->nFilledLen;
791                        break;
792                    default:
793                        CHECK_EQ(AVC_NALTYPE_PPS, type);
794                        ++mNumInputFrames;
795                        memcpy((uint8_t *) outHeader->pBuffer + outHeader->nFilledLen,
796                                "\x00\x00\x00\x01", 4);
797                        outHeader->nFilledLen += (dataLength + 4);
798                        outPtr += (dataLength + 4);
799                        break;
800                }
801            }
802        }
803
804        buffer_handle_t srcBuffer; // for MetaDataMode only
805
806        // Get next input video frame
807        if (mReadyForNextFrame) {
808            // Save the input buffer info so that it can be
809            // passed to an output buffer
810            InputBufferInfo info;
811            info.mTimeUs = inHeader->nTimeStamp;
812            info.mFlags = inHeader->nFlags;
813            mInputBufferInfoVec.push(info);
814            mPrevTimestampUs = inHeader->nTimeStamp;
815
816            if (inHeader->nFlags & OMX_BUFFERFLAG_EOS) {
817                mSawInputEOS = true;
818            }
819
820            if (inHeader->nFilledLen > 0) {
821                AVCFrameIO videoInput;
822                memset(&videoInput, 0, sizeof(videoInput));
823                videoInput.height = ((mVideoHeight  + 15) >> 4) << 4;
824                videoInput.pitch = ((mVideoWidth + 15) >> 4) << 4;
825                videoInput.coding_timestamp = (inHeader->nTimeStamp + 500) / 1000;  // in ms
826                uint8_t *inputData = NULL;
827                if (mStoreMetaDataInBuffers) {
828                    if (inHeader->nFilledLen != 8) {
829                        ALOGE("MetaData buffer is wrong size! "
830                                "(got %lu bytes, expected 8)", inHeader->nFilledLen);
831                        mSignalledError = true;
832                        notify(OMX_EventError, OMX_ErrorUndefined, 0, 0);
833                        return;
834                    }
835                    inputData =
836                            extractGrallocData(inHeader->pBuffer + inHeader->nOffset,
837                                    &srcBuffer);
838                    if (inputData == NULL) {
839                        ALOGE("Unable to extract gralloc buffer in metadata mode");
840                        mSignalledError = true;
841                        notify(OMX_EventError, OMX_ErrorUndefined, 0, 0);
842                        return;
843                    }
844                    // TODO: Verify/convert pixel format enum
845                } else {
846                    inputData = (uint8_t *)inHeader->pBuffer + inHeader->nOffset;
847                }
848
849                if (mVideoColorFormat != OMX_COLOR_FormatYUV420Planar) {
850                    ConvertYUV420SemiPlanarToYUV420Planar(
851                        inputData, mInputFrameData, mVideoWidth, mVideoHeight);
852                    inputData = mInputFrameData;
853                }
854                CHECK(inputData != NULL);
855                videoInput.YCbCr[0] = inputData;
856                videoInput.YCbCr[1] = videoInput.YCbCr[0] + videoInput.height * videoInput.pitch;
857                videoInput.YCbCr[2] = videoInput.YCbCr[1] +
858                    ((videoInput.height * videoInput.pitch) >> 2);
859                videoInput.disp_order = mNumInputFrames;
860
861                encoderStatus = PVAVCEncSetInput(mHandle, &videoInput);
862                if (encoderStatus == AVCENC_SUCCESS || encoderStatus == AVCENC_NEW_IDR) {
863                    mReadyForNextFrame = false;
864                    ++mNumInputFrames;
865                    if (encoderStatus == AVCENC_NEW_IDR) {
866                        mIsIDRFrame = 1;
867                    }
868                } else {
869                    if (encoderStatus < AVCENC_SUCCESS) {
870                        ALOGE("encoderStatus = %d at line %d", encoderStatus, __LINE__);
871                        mSignalledError = true;
872                        releaseGrallocData(srcBuffer);
873                        notify(OMX_EventError, OMX_ErrorUndefined, 0, 0);
874                        return;
875                    } else {
876                        ALOGV("encoderStatus = %d at line %d", encoderStatus, __LINE__);
877                        inQueue.erase(inQueue.begin());
878                        inInfo->mOwnedByUs = false;
879                        releaseGrallocData(srcBuffer);
880                        notifyEmptyBufferDone(inHeader);
881                        return;
882                    }
883                }
884            }
885        }
886
887        // Encode an input video frame
888        CHECK(encoderStatus == AVCENC_SUCCESS || encoderStatus == AVCENC_NEW_IDR);
889        dataLength = outHeader->nAllocLen;  // Reset the output buffer length
890        if (inHeader->nFilledLen > 0) {
891            if (outHeader->nAllocLen >= 4) {
892                memcpy(outPtr, "\x00\x00\x00\x01", 4);
893                outPtr += 4;
894                dataLength -= 4;
895            }
896            encoderStatus = PVAVCEncodeNAL(mHandle, outPtr, &dataLength, &type);
897            dataLength = outPtr + dataLength - outHeader->pBuffer;
898            if (encoderStatus == AVCENC_SUCCESS) {
899                CHECK(NULL == PVAVCEncGetOverrunBuffer(mHandle));
900            } else if (encoderStatus == AVCENC_PICTURE_READY) {
901                CHECK(NULL == PVAVCEncGetOverrunBuffer(mHandle));
902                if (mIsIDRFrame) {
903                    outHeader->nFlags |= OMX_BUFFERFLAG_SYNCFRAME;
904                    mIsIDRFrame = false;
905                }
906                mReadyForNextFrame = true;
907                AVCFrameIO recon;
908                if (PVAVCEncGetRecon(mHandle, &recon) == AVCENC_SUCCESS) {
909                    PVAVCEncReleaseRecon(mHandle, &recon);
910                }
911            } else {
912                dataLength = 0;
913                mReadyForNextFrame = true;
914            }
915
916            if (encoderStatus < AVCENC_SUCCESS) {
917                ALOGE("encoderStatus = %d at line %d", encoderStatus, __LINE__);
918                mSignalledError = true;
919                releaseGrallocData(srcBuffer);
920                notify(OMX_EventError, OMX_ErrorUndefined, 0, 0);
921                return;
922            }
923        } else {
924            dataLength = 0;
925        }
926
927        inQueue.erase(inQueue.begin());
928        inInfo->mOwnedByUs = false;
929        releaseGrallocData(srcBuffer);
930        notifyEmptyBufferDone(inHeader);
931
932        outQueue.erase(outQueue.begin());
933        CHECK(!mInputBufferInfoVec.empty());
934        InputBufferInfo *inputBufInfo = mInputBufferInfoVec.begin();
935        outHeader->nTimeStamp = inputBufInfo->mTimeUs;
936        outHeader->nFlags |= (inputBufInfo->mFlags | OMX_BUFFERFLAG_ENDOFFRAME);
937        if (mSawInputEOS) {
938            outHeader->nFlags |= OMX_BUFFERFLAG_EOS;
939        }
940        outHeader->nFilledLen = dataLength;
941        outInfo->mOwnedByUs = false;
942        notifyFillBufferDone(outHeader);
943        mInputBufferInfoVec.erase(mInputBufferInfoVec.begin());
944    }
945}
946
947int32_t SoftAVCEncoder::allocOutputBuffers(
948        unsigned int sizeInMbs, unsigned int numBuffers) {
949    CHECK(mOutputBuffers.isEmpty());
950    size_t frameSize = (sizeInMbs << 7) * 3;
951    for (unsigned int i = 0; i <  numBuffers; ++i) {
952        MediaBuffer *buffer = new MediaBuffer(frameSize);
953        buffer->setObserver(this);
954        mOutputBuffers.push(buffer);
955    }
956
957    return 1;
958}
959
960void SoftAVCEncoder::unbindOutputBuffer(int32_t index) {
961    CHECK(index >= 0);
962}
963
964int32_t SoftAVCEncoder::bindOutputBuffer(int32_t index, uint8_t **yuv) {
965    CHECK(index >= 0);
966    CHECK(index < (int32_t) mOutputBuffers.size());
967    *yuv = (uint8_t *) mOutputBuffers[index]->data();
968
969    return 1;
970}
971
972void SoftAVCEncoder::signalBufferReturned(MediaBuffer *buffer) {
973    UNUSED_UNLESS_VERBOSE(buffer);
974    ALOGV("signalBufferReturned: %p", buffer);
975}
976
977OMX_ERRORTYPE SoftAVCEncoder::getExtensionIndex(
978        const char *name, OMX_INDEXTYPE *index) {
979    if (!strcmp(name, "OMX.google.android.index.storeMetaDataInBuffers")) {
980        *(int32_t*)index = kStoreMetaDataExtensionIndex;
981        return OMX_ErrorNone;
982    }
983    return OMX_ErrorUndefined;
984}
985
986uint8_t *SoftAVCEncoder::extractGrallocData(void *data, buffer_handle_t *buffer) {
987    OMX_U32 type = *(OMX_U32*)data;
988    status_t res;
989    if (type != kMetadataBufferTypeGrallocSource) {
990        ALOGE("Data passed in with metadata mode does not have type "
991                "kMetadataBufferTypeGrallocSource (%d), has type %ld instead",
992                kMetadataBufferTypeGrallocSource, type);
993        return NULL;
994    }
995    buffer_handle_t imgBuffer = *(buffer_handle_t*)((uint8_t*)data + 4);
996
997    const Rect rect(mVideoWidth, mVideoHeight);
998    uint8_t *img;
999    res = GraphicBufferMapper::get().lock(imgBuffer,
1000            GRALLOC_USAGE_HW_VIDEO_ENCODER,
1001            rect, (void**)&img);
1002    if (res != OK) {
1003        ALOGE("%s: Unable to lock image buffer %p for access", __FUNCTION__,
1004                imgBuffer);
1005        return NULL;
1006    }
1007
1008    *buffer = imgBuffer;
1009    return img;
1010}
1011
1012void SoftAVCEncoder::releaseGrallocData(buffer_handle_t buffer) {
1013    if (mStoreMetaDataInBuffers) {
1014        GraphicBufferMapper::get().unlock(buffer);
1015    }
1016}
1017
1018}  // namespace android
1019
1020android::SoftOMXComponent *createSoftOMXComponent(
1021        const char *name, const OMX_CALLBACKTYPE *callbacks,
1022        OMX_PTR appData, OMX_COMPONENTTYPE **component) {
1023    return new android::SoftAVCEncoder(name, callbacks, appData, component);
1024}
1025