/frameworks/av/camera/ |
H A D | Camera.cpp | 316 void Camera::dataCallbackTimestamp(nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) argument 327 proxylistener->dataCallbackTimestamp(timestamp, msgType, dataPtr); 338 listener->postDataTimestamp(timestamp, msgType, dataPtr);
|
H A D | ICameraClient.cpp | 70 void dataCallbackTimestamp(nsecs_t timestamp, int32_t msgType, const sp<IMemory>& imageData) argument 75 data.writeInt64(timestamp); 118 nsecs_t timestamp = data.readInt64(); local 121 dataCallbackTimestamp(timestamp, msgType, imageData);
|
H A D | ICameraRecordingProxyListener.cpp | 38 void dataCallbackTimestamp(nsecs_t timestamp, int32_t msgType, const sp<IMemory>& imageData) argument 43 data.writeInt64(timestamp); 61 nsecs_t timestamp = data.readInt64(); local 64 dataCallbackTimestamp(timestamp, msgType, imageData);
|
/frameworks/av/camera/camera2/ |
H A D | ICameraDeviceCallbacks.cpp | 71 void onCaptureStarted(const CaptureResultExtras& result, int64_t timestamp) argument 78 data.writeInt64(timestamp); 139 int64_t timestamp = data.readInt64(); local 140 onCaptureStarted(result, timestamp);
|
/frameworks/av/include/media/ |
H A D | ExtendedAudioBufferProvider.h | 32 // Invoked by buffer consumer when a new timestamp is available. 33 // Default implementation ignores the timestamp. 34 virtual void onTimestamp(const AudioTimestamp& timestamp) { } argument
|
H A D | IOMX.h | 139 OMX_U32 flags, OMX_TICKS timestamp) = 0; 190 OMX_TICKS timestamp; member in struct:android::omx_message::__anon42::__anon45
|
/frameworks/av/include/media/nbaio/ |
H A D | NBAIO.h | 228 // Returns NO_ERROR if a timestamp is available. The timestamp includes the total number 230 // as of this presentation count. The timestamp parameter is undefined if error is returned. 231 virtual status_t getTimestamp(AudioTimestamp& timestamp) { return INVALID_OPERATION; } argument 320 // Invoked asynchronously by corresponding sink when a new timestamp is available. 321 // Default implementation ignores the timestamp. 322 virtual void onTimestamp(const AudioTimestamp& timestamp) { } argument
|
/frameworks/av/media/libmedia/ |
H A D | AudioTrack.cpp | 607 // the same configuration. A timestamp query for a paused track 1932 status_t AudioTrack::getTimestamp(AudioTimestamp& timestamp) argument 1964 status_t status = mAudioTrack->getTimestamp(timestamp); 1972 timestamp.mPosition = mPausedPosition; 1973 clock_gettime(CLOCK_MONOTONIC, ×tamp.mTime); 1986 const int64_t timestampTimeUs = convertTimespecToUs(timestamp.mTime); 1988 return WOULD_BLOCK; // stale timestamp time, occurs before start. 1991 const int64_t deltaPositionByUs = timestamp.mPosition * 1000000LL / mSampleRate; 2000 timestamp.mPosition); 2004 mStartUs = 0; // no need to check again, start timestamp ha [all...] |
H A D | IAudioTrack.cpp | 178 virtual status_t getTimestamp(AudioTimestamp& timestamp) { argument 185 timestamp.mPosition = reply.readInt32(); 186 timestamp.mTime.tv_sec = reply.readInt32(); 187 timestamp.mTime.tv_nsec = reply.readInt32(); 275 AudioTimestamp timestamp; local 276 status_t status = getTimestamp(timestamp); 279 reply->writeInt32(timestamp.mPosition); 280 reply->writeInt32(timestamp.mTime.tv_sec); 281 reply->writeInt32(timestamp.mTime.tv_nsec);
|
H A D | IOMX.cpp | 462 OMX_U32 flags, OMX_TICKS timestamp) { 470 data.writeInt64(timestamp); 921 OMX_TICKS timestamp = data.readInt64(); local 926 flags, timestamp)); 458 emptyBuffer( node_id node, buffer_id buffer, OMX_U32 range_offset, OMX_U32 range_length, OMX_U32 flags, OMX_TICKS timestamp) argument
|
/frameworks/av/media/libnbaio/ |
H A D | AudioStreamOutSink.cpp | 69 status_t AudioStreamOutSink::getNextWriteTimestamp(int64_t *timestamp) { argument 70 ALOG_ASSERT(timestamp != NULL); 78 return mStream->get_next_write_timestamp(mStream, timestamp); 81 status_t AudioStreamOutSink::getTimestamp(AudioTimestamp& timestamp) argument 88 int ok = mStream->get_presentation_position(mStream, &position64, ×tamp.mTime); 92 timestamp.mPosition = position64;
|
H A D | MonoPipe.cpp | 226 status_t MonoPipe::getNextWriteTimestamp(int64_t *timestamp) argument 230 ALOG_ASSERT(NULL != timestamp); 235 observeFrontAndNRPTS(&front, timestamp); 237 if (AudioBufferProvider::kInvalidPTS != *timestamp) { 238 // If we have a valid read-pointer and next read timestamp pair, then 240 // frames are in the buffer, and offset the timestamp by that amt. Then 242 // the next read timestamp plus the current amount of data in the 245 *timestamp = offsetTimestampByAudioFrames(*timestamp, pendingFrames); 334 status_t MonoPipe::getTimestamp(AudioTimestamp& timestamp) argument [all...] |
H A D | MonoPipeReader.cpp | 89 void MonoPipeReader::onTimestamp(const AudioTimestamp& timestamp) argument 91 mPipe->mTimestampMutator.push(timestamp);
|
H A D | NBLog.cpp | 343 String8 timestamp, body; local 347 // TODO timestamp empty here, only other choice to wait for the first timestamp event in the 348 // log to push it out. Consider keeping the timestamp/body between calls to readAt(). 349 dumpLine(timestamp, body); 357 timestamp.appendFormat("[%*s]", (int) width + 4, ""); 402 dumpLine(timestamp, body); 405 timestamp.clear(); 407 timestamp.appendFormat("[%d.%03d to .%.03d by .%.03d to .%.03d]", 415 timestamp 438 dumpLine(const String8& timestamp, String8& body) argument [all...] |
H A D | SourceAudioBufferProvider.cpp | 110 void SourceAudioBufferProvider::onTimestamp(const AudioTimestamp& timestamp) argument 112 mSource->onTimestamp(timestamp);
|
/frameworks/av/media/libstagefright/ |
H A D | CameraSource.cpp | 54 nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr); 92 nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) { 96 source->dataCallbackTimestamp(timestamp/1000, msgType, dataPtr); 855 ALOGV("dataCallbackTimestamp: timestamp %" PRId64 " us", timestampUs); 870 // May need to skip frame or modify timestamp. Currently implemented 912 nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) { 913 mSource->dataCallbackTimestamp(timestamp / 1000, msgType, dataPtr); 91 postDataTimestamp( nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) argument 911 dataCallbackTimestamp( nsecs_t timestamp, int32_t msgType, const sp<IMemory>& dataPtr) argument
|
H A D | OMXClient.cpp | 126 OMX_U32 flags, OMX_TICKS timestamp); 374 OMX_U32 flags, OMX_TICKS timestamp) { 376 node, buffer, range_offset, range_length, flags, timestamp); 370 emptyBuffer( node_id node, buffer_id buffer, OMX_U32 range_offset, OMX_U32 range_length, OMX_U32 flags, OMX_TICKS timestamp) argument
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/ |
H A D | SoftMPEG4.cpp | 214 // so use fake timestamp instead 215 uint32_t timestamp = 0xFFFFFFFF; local 218 timestamp = mPvTime; 229 mHandle, &bitstream, ×tamp, &tmp, 246 outHeader->nTimeStamp = mPvToOmxTimeMap.valueFor(timestamp); 247 mPvToOmxTimeMap.removeItem(timestamp);
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/include/ |
H A D | mp4dec_api.h | 109 uint32 timestamp; member in struct:tagVopHeaderInfo 155 Bool PVDecSetReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp); 156 Bool PVDecSetEnhReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp); 157 OSCL_IMPORT_REF Bool PVDecodeVideoFrame(VideoDecControls *decCtrl, uint8 *bitstream[], uint32 *timestamp, int32 *buffer_size, uint use_ext_timestamp[], uint8* currYUV); 158 Bool PVDecodeVopHeader(VideoDecControls *decCtrl, uint8 *buffer[], uint32 timestamp[], int32 buffer_size[], VopHeaderInfo *header_info, uint use_ext_timestamp[], uint8 *currYUV);
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
H A D | pvdec_api.cpp | 1007 uint32 timestamp[], int32 buffer_size[], uint use_ext_timestamp[], uint8 *currYUV) 1012 status = (PV_STATUS)PVDecodeVopHeader(decCtrl, buffer, timestamp, buffer_size, &header_info, use_ext_timestamp, currYUV); 1035 uint32 timestamp[], int32 buffer_size[], VopHeaderInfo *header_info, uint use_ext_timestamp [], uint8 *currYUV) 1059 timestamp[idx] = -1; 1063 /* did the application provide a timestamp for this vop? */ 1064 if (timestamp[idx] < 0) 1068 /* decode the timestamp in the bitstream */ 1087 timestamp[idx] = CalcVopDisplayTime(video->vol[idx], vopHeader[idx], video->shortVideoHeader); 1094 timestamp[idx] = vopHeader[idx]->timeStamp; 1099 if (timestamp[id 1006 PVDecodeVideoFrame(VideoDecControls *decCtrl, uint8 *buffer[], uint32 timestamp[], int32 buffer_size[], uint use_ext_timestamp[], uint8 *currYUV) argument 1034 PVDecodeVopHeader(VideoDecControls *decCtrl, uint8 *buffer[], uint32 timestamp[], int32 buffer_size[], VopHeaderInfo *header_info, uint use_ext_timestamp [], uint8 *currYUV) argument 1603 PVDecSetReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp) argument 1644 PVDecSetEnhReference(VideoDecControls *decCtrl, uint8 *refYUV, uint32 timestamp) argument [all...] |
/frameworks/av/media/libstagefright/codecs/m4v_h263/enc/include/ |
H A D | mp4enc_api.h | 123 ULong timestamp; /* modulo timestamp in millisecond*/ member in struct:tagvideoEncFrameIO 331 * @param nextModTime is the timestamp encoder expects from the next input 354 * @brief This function sets the input YUV frame and timestamp to be encoded by the slice-based encoding function PVEncodeSlice(). 356 * The encoder library processes the timestamp and determine if this frame is to be encoded or not. If the current frame 361 * @param nextModTime is the timestamp encoder expects from the next input if this input is rejected and nLayer is set to -1. 373 * At the end-of-frame, the next video input address and the next video modulo timestamp will be set. 380 * @param nextModTime is the timestamp encoder expects from the next input
|
/frameworks/av/media/libstagefright/omx/ |
H A D | GraphicBufferSource.cpp | 335 ALOGV("OUT timestamp: %lld -> %lld", 342 ALOGW("giving up limiting timestamp gap (pts = %lld)", 611 /* Cap timestamp gap between adjacent frames to specified max 622 // Drop the frame if it's going backward in time. Bad timestamp 634 ALOGV("IN timestamp: %lld -> %lld", 717 uint64_t timestamp = 0; // does this matter? local 721 timestamp);
|
H A D | OMX.cpp | 411 OMX_U32 flags, OMX_TICKS timestamp) { 413 buffer, range_offset, range_length, flags, timestamp); 481 msg.u.extended_buffer_data.timestamp = pBuffer->nTimeStamp; 407 emptyBuffer( node_id node, buffer_id buffer, OMX_U32 range_offset, OMX_U32 range_length, OMX_U32 flags, OMX_TICKS timestamp) argument
|
H A D | OMXNodeInstance.cpp | 979 OMX_U32 flags, OMX_TICKS timestamp) { 990 return emptyBuffer_l(header, flags, timestamp, (intptr_t)buffer); 1018 OMX_BUFFERHEADERTYPE *header, OMX_U32 flags, OMX_TICKS timestamp, intptr_t debugAddr) { 1020 header->nTimeStamp = timestamp; 1053 OMX_U32 flags, OMX_TICKS timestamp) { 1059 return emptyBuffer_l(header, flags, timestamp, (intptr_t)header->pBuffer); 1184 // fix up the buffer info (especially timestamp) if needed 1188 newMsg.u.extended_buffer_data.timestamp = buffer->nTimeStamp; 976 emptyBuffer( OMX::buffer_id buffer, OMX_U32 rangeOffset, OMX_U32 rangeLength, OMX_U32 flags, OMX_TICKS timestamp) argument 1017 emptyBuffer_l( OMX_BUFFERHEADERTYPE *header, OMX_U32 flags, OMX_TICKS timestamp, intptr_t debugAddr) argument 1050 emptyDirectBuffer( OMX_BUFFERHEADERTYPE *header, OMX_U32 rangeOffset, OMX_U32 rangeLength, OMX_U32 flags, OMX_TICKS timestamp) argument
|
/frameworks/av/services/audioflinger/ |
H A D | FastMixer.h | 83 AudioTimestamp timestamp; member in class:android::FastMixer
|