/frameworks/av/media/libnbaio/ |
H A D | MonoPipe.cpp | 205 status_t MonoPipe::getNextWriteTimestamp(int64_t *timestamp) argument 209 ALOG_ASSERT(NULL != timestamp); 214 observeFrontAndNRPTS(&front, timestamp); 216 if (AudioBufferProvider::kInvalidPTS != *timestamp) { 217 // If we have a valid read-pointer and next read timestamp pair, then 219 // frames are in the buffer, and offset the timestamp by that amt. Then 221 // the next read timestamp plus the current amount of data in the 224 *timestamp = offsetTimestampByAudioFrames(*timestamp, pendingFrames);
|
/frameworks/av/media/libstagefright/ |
H A D | CameraSource.cpp | 46 nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr); 81 nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) { 85 source->dataCallbackTimestamp(timestamp/1000, msgType, dataPtr); 804 ALOGV("dataCallbackTimestamp: timestamp %lld us", timestampUs); 819 // May need to skip frame or modify timestamp. Currently implemented 861 nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) { 862 mSource->dataCallbackTimestamp(timestamp / 1000, msgType, dataPtr); 80 postDataTimestamp( nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) argument 860 dataCallbackTimestamp( nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) argument
|
H A D | OMXClient.cpp | 103 OMX_U32 flags, OMX_TICKS timestamp); 304 OMX_U32 flags, OMX_TICKS timestamp) { 306 node, buffer, range_offset, range_length, flags, timestamp); 300 emptyBuffer( node_id node, buffer_id buffer, OMX_U32 range_offset, OMX_U32 range_length, OMX_U32 flags, OMX_TICKS timestamp) argument
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/ |
H A D | SoftMPEG4.cpp | 419 uint32_t timestamp = local 429 mHandle, &bitstream, ×tamp, &tmp, 444 outHeader->nTimeStamp = timestamp * 1000;
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/include/ |
H A D | mp4dec_api.h | 109 uint32 timestamp; member in struct:tagVopHeaderInfo 155 Bool PVDecSetReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp); 156 Bool PVDecSetEnhReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp); 157 OSCL_IMPORT_REF Bool PVDecodeVideoFrame(VideoDecControls *decCtrl, uint8 *bitstream[], uint32 *timestamp, int32 *buffer_size, uint use_ext_timestamp[], uint8* currYUV); 158 Bool PVDecodeVopHeader(VideoDecControls *decCtrl, uint8 *buffer[], uint32 timestamp[], int32 buffer_size[], VopHeaderInfo *header_info, uint use_ext_timestamp[], uint8 *currYUV);
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
H A D | pvdec_api.cpp | 1002 uint32 timestamp[], int32 buffer_size[], uint use_ext_timestamp[], uint8 *currYUV) 1007 status = (PV_STATUS)PVDecodeVopHeader(decCtrl, buffer, timestamp, buffer_size, &header_info, use_ext_timestamp, currYUV); 1030 uint32 timestamp[], int32 buffer_size[], VopHeaderInfo *header_info, uint use_ext_timestamp [], uint8 *currYUV) 1054 timestamp[idx] = -1; 1058 /* did the application provide a timestamp for this vop? */ 1059 if (timestamp[idx] < 0) 1063 /* decode the timestamp in the bitstream */ 1082 timestamp[idx] = CalcVopDisplayTime(video->vol[idx], vopHeader[idx], video->shortVideoHeader); 1089 timestamp[idx] = vopHeader[idx]->timeStamp; 1094 if (timestamp[id 1001 PVDecodeVideoFrame(VideoDecControls *decCtrl, uint8 *buffer[], uint32 timestamp[], int32 buffer_size[], uint use_ext_timestamp[], uint8 *currYUV) argument 1029 PVDecodeVopHeader(VideoDecControls *decCtrl, uint8 *buffer[], uint32 timestamp[], int32 buffer_size[], VopHeaderInfo *header_info, uint use_ext_timestamp [], uint8 *currYUV) argument 1598 PVDecSetReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp) argument 1639 PVDecSetEnhReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp) argument [all...] |
/frameworks/av/media/libstagefright/codecs/m4v_h263/enc/include/ |
H A D | mp4enc_api.h | 123 ULong timestamp; /* modulo timestamp in millisecond*/ member in struct:tagvideoEncFrameIO 331 * @param nextModTime is the timestamp encoder expects from the next input 354 * @brief This function sets the input YUV frame and timestamp to be encoded by the slice-based encoding function PVEncodeSlice(). 356 * The encoder library processes the timestamp and determine if this frame is to be encoded or not. If the current frame 361 * @param nextModTime is the timestamp encoder expects from the next input if this input is rejected and nLayer is set to -1. 373 * At the end-of-frame, the next video input address and the next video modulo timestamp will be set. 380 * @param nextModTime is the timestamp encoder expects from the next input
|