/frameworks/av/media/libmediaplayerservice/nuplayer/ |
H A D | NuPlayerSource.h | 37 virtual sp<AMessage> getFormat(bool audio); 40 bool audio, sp<ABuffer> *accessUnit) = 0; 57 virtual sp<MetaData> getFormatMeta(bool audio) { return NULL; } argument
|
H A D | StreamingSource.cpp | 118 sp<MetaData> NuPlayer::StreamingSource::getFormatMeta(bool audio) { argument 120 audio ? ATSParser::AUDIO : ATSParser::VIDEO; 133 bool audio, sp<ABuffer> *accessUnit) { 135 audio ? ATSParser::AUDIO : ATSParser::VIDEO; 132 dequeueAccessUnit( bool audio, sp<ABuffer> *accessUnit) argument
|
H A D | HTTPLiveSource.cpp | 84 sp<MetaData> NuPlayer::HTTPLiveSource::getFormatMeta(bool audio) { argument 86 audio ? ATSParser::AUDIO : ATSParser::VIDEO; 149 bool audio, sp<ABuffer> *accessUnit) { 151 audio ? ATSParser::AUDIO : ATSParser::VIDEO; 148 dequeueAccessUnit( bool audio, sp<ABuffer> *accessUnit) argument
|
H A D | GenericSource.cpp | 75 if (!strncasecmp(mime, "audio/", 6)) { 114 readBuffer(true /* audio */); 123 readBuffer(false /* audio */); 131 sp<MetaData> NuPlayer::GenericSource::getFormatMeta(bool audio) { argument 132 sp<MediaSource> source = audio ? mAudioTrack.mSource : mVideoTrack.mSource; 142 bool audio, sp<ABuffer> *accessUnit) { 143 Track *track = audio ? &mAudioTrack : &mVideoTrack; 156 readBuffer(audio, -1ll); 169 readBuffer(false /* audio */, seekTimeUs, &actualTimeUs); 175 readBuffer(true /* audio */, seekTimeU 141 dequeueAccessUnit( bool audio, sp<ABuffer> *accessUnit) argument 181 readBuffer( bool audio, int64_t seekTimeUs, int64_t *actualTimeUs) argument [all...] |
H A D | NuPlayerRenderer.cpp | 59 bool audio, 63 msg->setInt32("audio", static_cast<int32_t>(audio)); 69 void NuPlayer::Renderer::queueEOS(bool audio, status_t finalResult) { argument 73 msg->setInt32("audio", static_cast<int32_t>(audio)); 78 void NuPlayer::Renderer::flush(bool audio) { argument 81 if (audio) { 91 msg->setInt32("audio", static_cast<int32_t>(audio)); 58 queueBuffer( bool audio, const sp<ABuffer> &buffer, const sp<AMessage> ¬ifyConsumed) argument 405 notifyEOS(bool audio, status_t finalResult) argument 414 int32_t audio; local 500 int32_t audio; local 524 int32_t audio; local 571 notifyFlushComplete(bool audio) argument 578 dropBufferWhileFlushing( bool audio, const sp<AMessage> &msg) argument [all...] |
H A D | RTSPSource.cpp | 99 sp<MetaData> NuPlayer::RTSPSource::getFormatMeta(bool audio) { argument 100 sp<AnotherPacketSource> source = getSource(audio); 121 ALOGV("audio track doesn't have enough data yet. (%.2f secs buffered)", 139 bool audio, sp<ABuffer> *accessUnit) { 148 sp<AnotherPacketSource> source = getSource(audio); 162 sp<AnotherPacketSource> NuPlayer::RTSPSource::getSource(bool audio) { argument 165 audio ? ATSParser::AUDIO : ATSParser::VIDEO); 170 return audio ? mAudioTrack : mVideoTrack; 301 sp<AnotherPacketSource> source = getSource(false /* audio */); 306 source = getSource(true /* audio */); 138 dequeueAccessUnit( bool audio, sp<ABuffer> *accessUnit) argument [all...] |
H A D | NuPlayer.cpp | 288 // We're not currently decoding anything (no audio or 311 bool audio = msg->what() == kWhatAudioNotify; local 321 audio, codecRequest); 333 ALOGV("got %s decoder EOS", audio ? "audio" : "video"); 336 audio ? "audio" : "video", 340 mRenderer->queueEOS(audio, err); 344 if (audio) { 354 ALOGV("decoder %s flush completed", audio 474 int32_t audio; local 518 int32_t audio; local 693 instantiateDecoder(bool audio, sp<Decoder> *decoder) argument 731 feedDecoderInputData(bool audio, const sp<AMessage> &msg) argument 848 renderBuffer(bool audio, const sp<AMessage> &msg) argument 908 flushDecoder(bool audio, bool needShutdown) argument 949 getFormat(bool audio) argument [all...] |
/frameworks/base/media/tests/MediaFrameworkTest/src/com/android/mediaframeworktest/functional/audio/ |
H A D | SimTonesTest.java | 17 package com.android.mediaframeworktest.functional.audio;
|
H A D | MediaBassBoostTest.java | 17 package com.android.mediaframeworktest.functional.audio;
|
H A D | MediaPresetReverbTest.java | 17 package com.android.mediaframeworktest.functional.audio; 210 // creating a volume controller on output mix ensures that ro.audio.silent mutes 211 // audio after the effects and not before 281 // creating a volume controller on output mix ensures that ro.audio.silent mutes 282 // audio after the effects and not before
|
H A D | MediaVirtualizerTest.java | 17 package com.android.mediaframeworktest.functional.audio;
|
H A D | MediaAudioManagerTest.java | 17 package com.android.mediaframeworktest.functional.audio; 148 // the audio focus request is async, so wait a bit to verify it had the expected effect 168 //Test case 1: test audio focus listener loses audio focus: 179 //Test case 2: test audio focus listener loses audio focus: 191 //Test case 3: test audio focus listener loses audio focus: 203 //Test case 4: test audio focus registering and use over 3000 iterations 213 assertTrue("audio focu [all...] |
H A D | MediaEnvReverbTest.java | 17 package com.android.mediaframeworktest.functional.audio; 365 // creating a volume controller on output mix ensures that ro.audio.silent mutes 366 // audio after the effects and not before 438 // creating a volume controller on output mix ensures that ro.audio.silent mutes 439 // audio after the effects and not before
|
H A D | MediaEqualizerTest.java | 17 package com.android.mediaframeworktest.functional.audio;
|
H A D | MediaVisualizerTest.java | 17 package com.android.mediaframeworktest.functional.audio; 242 // creating a volume controller on output mix ensures that ro.audio.silent mutes 243 // audio after the effects and not before 321 // creating a volume controller on output mix ensures that ro.audio.silent mutes 322 // audio after the effects and not before 454 // creating a volume controller on output mix ensures that ro.audio.silent mutes 455 // audio after the effects and not before
|
H A D | MediaAudioEffectTest.java | 17 package com.android.mediaframeworktest.functional.audio; 276 //Test case 1.4: test contructor on mediaPlayer audio session
|
H A D | MediaAudioTrackTest.java | 17 package com.android.mediaframeworktest.functional.audio; 752 // count (given the audio track properties), and add 77.
|
/frameworks/av/media/libmediaplayerservice/nuplayer/mp4/ |
H A D | MP4Source.cpp | 127 sp<AMessage> MP4Source::getFormat(bool audio) { argument 128 return mParser->getFormat(audio); 132 bool audio, sp<ABuffer> *accessUnit) { 133 return mParser->dequeueAccessUnit(audio, accessUnit); 131 dequeueAccessUnit( bool audio, sp<ABuffer> *accessUnit) argument
|
/frameworks/av/media/libstagefright/ |
H A D | FragmentedMP4Extractor.cpp | 52 FragmentedMPEG4Source(bool audio, 104 bool hasVideo = mParser->getFormat(false /* audio */, true /* synchronous */) != NULL; 105 bool hasAudio = mParser->getFormat(true /* audio */, true /* synchronous */) != NULL; 113 mFileMetaData->setCString(kKeyMIMEType, "audio/mp4"); 115 ALOGE("no audio and no video, no idea what file type this is"); 117 // tracks are numbered such that video track is first, audio track is second 189 bool audio, 196 mIsAudioTrack(audio), 188 FragmentedMPEG4Source( bool audio, const sp<MetaData> &format, const sp<FragmentedMP4Parser> &parser, const sp<FragmentedMP4Extractor> &extractor) argument
|
/frameworks/base/core/java/android/speech/srec/ |
H A D | Recognizer.java | 55 * // create and start audio input 56 * InputStream audio = new MicrophoneInputStream(11025, 11025*5); 87 * // put more audio in the Recognizer 88 * recognizer.putAudio(audio); 100 * // stop the audio device 101 * audio.close(); 159 * A separate config file is needed for each audio sample rate. 265 * Process some audio and return the current status. 288 * Put audio samples into the <code>Recognizer</code>. 289 * @param buf holds the audio sample 304 putAudio(InputStream audio) argument [all...] |
/frameworks/native/include/media/openmax/ |
H A D | OMX_Component.h | 96 OMX_AUDIO_PORTDEFINITIONTYPE audio; member in union:OMX_PARAM_PORTDEFINITIONTYPE::__anon1371
|
/frameworks/av/media/libstagefright/mp4/ |
H A D | FragmentedMP4Parser.cpp | 253 sp<AMessage> FragmentedMP4Parser::getFormat(bool audio, bool synchronous) { argument 258 msg->setInt32("audio", audio); 289 msg->setInt32("audio", wantAudio); 342 status_t FragmentedMP4Parser::dequeueAccessUnit(bool audio, sp<ABuffer> *accessUnit, argument 347 msg->setInt32("audio", audio); 480 CHECK(msg->findInt32("audio", &wantAudio)); 518 CHECK(msg->findInt32("audio", &wantAudio)); 549 CHECK(msg->findInt32("audio", [all...] |