/frameworks/base/core/java/android/bluetooth/ |
H A D | BluetoothAudioConfig.java | 35 public BluetoothAudioConfig(int sampleRate, int channelConfig, int audioFormat) { argument 38 mAudioFormat = audioFormat; 72 int audioFormat = in.readInt(); 73 return new BluetoothAudioConfig(sampleRate, channelConfig, audioFormat);
|
/frameworks/base/core/java/android/speech/tts/ |
H A D | FileSynthesisCallback.java | 102 public int start(int sampleRateInHz, int audioFormat, int channelCount) { argument 104 Log.d(TAG, "FileSynthesisRequest.start(" + sampleRateInHz + "," + audioFormat 107 if (audioFormat != AudioFormat.ENCODING_PCM_8BIT && 108 audioFormat != AudioFormat.ENCODING_PCM_16BIT && 109 audioFormat != AudioFormat.ENCODING_PCM_FLOAT) { 110 Log.e(TAG, "Audio format encoding " + audioFormat + " not supported. Please use one " + 114 mDispatcher.dispatchOnBeginSynthesis(sampleRateInHz, audioFormat, channelCount); 132 mAudioFormat = audioFormat; 203 int audioFormat = 0; 228 audioFormat 284 makeWavHeader(int sampleRateInHz, int audioFormat, int channelCount, int dataLength) argument [all...] |
H A D | ITextToSpeechCallback.aidl | 65 * @param audioFormat The audio format of the generated audio in the {@link #onAudioAvailable} 71 void onBeginSynthesis(String utteranceId, int sampleRateInHz, int audioFormat, int channelCount);
|
H A D | PlaybackSynthesisCallback.java | 125 public int start(int sampleRateInHz, int audioFormat, int channelCount) { argument 126 if (DBG) Log.d(TAG, "start(" + sampleRateInHz + "," + audioFormat + "," + channelCount 128 if (audioFormat != AudioFormat.ENCODING_PCM_8BIT && 129 audioFormat != AudioFormat.ENCODING_PCM_16BIT && 130 audioFormat != AudioFormat.ENCODING_PCM_FLOAT) { 131 Log.w(TAG, "Audio format encoding " + audioFormat + " not supported. Please use one " + 135 mDispatcher.dispatchOnBeginSynthesis(sampleRateInHz, audioFormat, channelCount); 158 mAudioParams, sampleRateInHz, audioFormat, channelCount,
|
H A D | SynthesisCallback.java | 65 * @param audioFormat Audio format of the generated audio. Must be one of {@link 74 @SupportedAudioFormat int audioFormat, 72 start( int sampleRateInHz, @SupportedAudioFormat int audioFormat, @IntRange(from = 1, to = 2) int channelCount) argument
|
H A D | BlockingAudioTrack.java | 79 int audioFormat, int channelCount) { 82 mAudioFormat = audioFormat; 217 AudioFormat audioFormat = (new AudioFormat.Builder()) 222 audioFormat, bufferSizeInBytes, AudioTrack.MODE_STREAM, 78 BlockingAudioTrack(AudioOutputParams audioParams, int sampleRate, int audioFormat, int channelCount) argument
|
H A D | UtteranceProgressListener.java | 94 * @param audioFormat Audio format of the generated audio. Should be one of 99 public void onBeginSynthesis(String utteranceId, int sampleRateInHz, int audioFormat, int channelCount) { argument
|
H A D | SynthesisPlaybackQueueItem.java | 80 int audioFormat, int channelCount, UtteranceProgressDispatcher dispatcher, 90 mAudioTrack = new BlockingAudioTrack(audioParams, sampleRate, audioFormat, channelCount); 79 SynthesisPlaybackQueueItem(AudioOutputParams audioParams, int sampleRate, int audioFormat, int channelCount, UtteranceProgressDispatcher dispatcher, Object callerIdentity, AbstractEventLogger logger) argument
|
H A D | TextToSpeechService.java | 686 void dispatchOnBeginSynthesis(int sampleRateInHz, int audioFormat, int channelCount); argument 895 public void dispatchOnBeginSynthesis(int sampleRateInHz, int audioFormat, int channelCount) { argument 898 mCallbacks.dispatchOnBeginSynthesis(getCallerIdentity(), utteranceId, sampleRateInHz, audioFormat, channelCount); 1576 public void dispatchOnBeginSynthesis(Object callerIdentity, String utteranceId, int sampleRateInHz, int audioFormat, int channelCount) { argument 1580 cb.onBeginSynthesis(utteranceId, sampleRateInHz, audioFormat, channelCount);
|
/frameworks/base/media/java/android/media/ |
H A D | AudioFormat.java | 499 public static int getBytesPerSample(int audioFormat) argument 501 switch (audioFormat) { 512 throw new IllegalArgumentException("Bad audio format " + audioFormat); 517 public static boolean isValidEncoding(int audioFormat) argument 519 switch (audioFormat) { 543 public static boolean isPublicEncoding(int audioFormat) argument 545 switch (audioFormat) { 570 public static boolean isEncodingLinearPcm(int audioFormat) argument 572 switch (audioFormat) { 594 throw new IllegalArgumentException("Bad audio format " + audioFormat); 599 isEncodingLinearFrames(int audioFormat) argument 1110 toDisplayName(@urroundSoundEncoding int audioFormat) argument [all...] |
H A D | AudioTrack.java | 427 * @param audioFormat the format in which the audio data is represented. 454 public AudioTrack(int streamType, int sampleRateInHz, int channelConfig, int audioFormat, argument 457 this(streamType, sampleRateInHz, channelConfig, audioFormat, 483 * @param audioFormat the format in which the audio data is represented. 512 public AudioTrack(int streamType, int sampleRateInHz, int channelConfig, int audioFormat, argument 521 .setEncoding(audioFormat) 1063 int audioFormat, int mode) { 1077 if (audioFormat == AudioFormat.ENCODING_IEC61937 1130 if (audioFormat == AudioFormat.ENCODING_DEFAULT) { 1131 audioFormat 1062 audioParamCheck(int sampleRateInHz, int channelConfig, int channelIndexMask, int audioFormat, int mode) argument 1576 getMinBufferSize(int sampleRateInHz, int channelConfig, int audioFormat) argument 3184 native_setup(Object audiotrack_this, Object attributes, int[] sampleRate, int channelMask, int channelIndexMask, int audioFormat, int buffSizeInBytes, int mode, int[] sessionId, long nativeAudioTrack, boolean offload) argument 3257 native_get_min_buff_size( int sampleRateInHz, int channelConfig, int audioFormat) argument [all...] |
H A D | AudioRecord.java | 271 * @param audioFormat the format in which the audio data is to be returned. 281 public AudioRecord(int audioSource, int sampleRateInHz, int channelConfig, int audioFormat, argument 290 .setEncoding(audioFormat) 677 private void audioParamCheck(int audioSource, int sampleRateInHz, int audioFormat) argument 702 switch (audioFormat) { 709 mAudioFormat = audioFormat; 712 throw new IllegalArgumentException("Unsupported sample encoding " + audioFormat 924 * @param audioFormat the format in which the audio data is represented. 933 static public int getMinBufferSize(int sampleRateInHz, int channelConfig, int audioFormat) { argument 952 int size = native_get_min_buff_size(sampleRateInHz, channelCount, audioFormat); 1749 native_setup(Object audiorecord_this, Object attributes, int[] sampleRate, int channelMask, int channelIndexMask, int audioFormat, int buffSizeInBytes, int[] sessionId, String opPackageName, long nativeRecordInJavaObj) argument 1787 native_get_min_buff_size( int sampleRateInHz, int channelCount, int audioFormat) argument [all...] |
/frameworks/av/media/libaaudio/src/core/ |
H A D | AAudioStreamParameters.h | 63 void setFormat(aaudio_format_t audioFormat) { argument 64 mAudioFormat = audioFormat;
|
/frameworks/base/core/jni/ |
H A D | android_media_AudioFormat.h | 49 static inline audio_format_t audioFormatToNative(int audioFormat) argument 51 switch (audioFormat) {
|
H A D | android_media_AudioRecord.cpp | 193 jint audioFormat, jint buffSizeInBytes, jintArray jSession, jstring opPackageName, 197 //ALOGV("sampleRate=%d, audioFormat=%d, channel mask=%x, buffSizeInBytes=%d " 199 // sampleRateInHertz, audioFormat, channelMask, buffSizeInBytes, nativeRecordInJavaObj); 257 audio_format_t format = audioFormatToNative(audioFormat); 259 ALOGE("Error creating AudioRecord: unsupported audio format %d.", audioFormat); 662 jint sampleRateInHertz, jint channelCount, jint audioFormat) { 665 sampleRateInHertz, channelCount, audioFormat); 668 audio_format_t format = audioFormatToNative(audioFormat); 191 android_media_AudioRecord_setup(JNIEnv *env, jobject thiz, jobject weak_this, jobject jaa, jintArray jSampleRate, jint channelMask, jint channelIndexMask, jint audioFormat, jint buffSizeInBytes, jintArray jSession, jstring opPackageName, jlong nativeRecordInJavaObj) argument 661 android_media_AudioRecord_get_min_buff_size(JNIEnv *env, jobject thiz, jint sampleRateInHertz, jint channelCount, jint audioFormat) argument
|
H A D | android_media_AudioTrack.cpp | 215 jint audioFormat, jint buffSizeInBytes, jint memoryMode, jintArray jSession, 218 ALOGV("sampleRates=%p, channel mask=%x, index mask=%x, audioFormat(Java)=%d, buffSize=%d" 220 jSampleRate, channelPositionMask, channelIndexMask, audioFormat, buffSizeInBytes, 277 audio_format_t format = audioFormatToNative(audioFormat); 279 ALOGE("Error creating AudioTrack: unsupported audio format %d.", audioFormat); 631 static jint writeToTrack(const sp<AudioTrack>& track, jint audioFormat, const T *data, argument 1112 jint sampleRateInHertz, jint channelCount, jint audioFormat) { 1122 const audio_format_t format = audioFormatToNative(audioFormat); 213 android_media_AudioTrack_setup(JNIEnv *env, jobject thiz, jobject weak_this, jobject jaa, jintArray jSampleRate, jint channelPositionMask, jint channelIndexMask, jint audioFormat, jint buffSizeInBytes, jint memoryMode, jintArray jSession, jlong nativeAudioTrack, jboolean offload) argument 1111 android_media_AudioTrack_get_min_buff_size(JNIEnv *env, jobject thiz, jint sampleRateInHertz, jint channelCount, jint audioFormat) argument
|
/frameworks/av/media/libstagefright/ |
H A D | AudioPlayer.cpp | 140 audio_format_t audioFormat = AUDIO_FORMAT_PCM_16_BIT; local 143 sAudioFormatToPcmEncoding.map(pcmEncoding, &audioFormat); 147 if (mapMimeToAudioFormat(audioFormat, mime) != OK) { 149 audioFormat = AUDIO_FORMAT_INVALID; 151 ALOGV("Mime type \"%s\" mapped to audio_format 0x%x", mime, audioFormat); 155 if ((audioFormat == AUDIO_FORMAT_AAC) && format->findInt32(kKeyAACAOT, &aacaot)) { 157 mapAACProfileToAudioFormat(audioFormat,(OMX_AUDIO_AACPROFILETYPE) aacaot); 184 offloadInfo.format = audioFormat; 192 mSampleRate, numChannels, channelMask, audioFormat,
|
/frameworks/base/media/java/android/media/soundtrigger/ |
H A D | SoundTriggerDetector.java | 123 AudioFormat audioFormat, int captureSession, byte[] data) { 127 mAudioFormat = audioFormat; 122 EventPayload(boolean triggerAvailable, boolean captureAvailable, AudioFormat audioFormat, int captureSession, byte[] data) argument
|
/frameworks/base/tests/Camera2Tests/SmartCamera/SimpleCamera/src/androidx/media/filterfw/decoder/ |
H A D | MediaDecoder.java | 264 MediaFormat audioFormat = mMediaExtractor.getTrackFormat(mAudioTrackIndex); 265 mAudioTrackDecoder = new AudioTrackDecoder(mAudioTrackIndex, audioFormat, this);
|
/frameworks/base/media/java/android/media/projection/ |
H A D | MediaProjection.java | 148 int audioFormat, int bufferSizeInBytes) { 146 createAudioRecord( int sampleRateInHz, int channelConfig, int audioFormat, int bufferSizeInBytes) argument
|
/frameworks/base/core/java/android/service/voice/ |
H A D | AlwaysOnHotwordDetector.java | 216 AudioFormat audioFormat, int captureSession, byte[] data) { 220 mAudioFormat = audioFormat; 215 EventPayload(boolean triggerAvailable, boolean captureAvailable, AudioFormat audioFormat, int captureSession, byte[] data) argument
|
/frameworks/av/include/media/ |
H A D | IAudioPolicyService.h | 174 virtual status_t setSurroundFormatEnabled(audio_format_t audioFormat, bool enabled) = 0;
|
/frameworks/av/media/libaudioclient/include/media/ |
H A D | IAudioPolicyService.h | 174 virtual status_t setSurroundFormatEnabled(audio_format_t audioFormat, bool enabled) = 0;
|
/frameworks/av/media/libmediaplayer2/nuplayer2/ |
H A D | NuPlayer2.h | 285 void determineAudioModeChange(const sp<AMessage> &audioFormat);
|
/frameworks/av/media/libmediaplayerservice/nuplayer/ |
H A D | NuPlayer.h | 286 void determineAudioModeChange(const sp<AMessage> &audioFormat);
|