/frameworks/av/services/audioflinger/ |
H A D | AudioPolicyService.cpp | 219 audio_io_handle_t AudioPolicyService::getOutput(audio_stream_type_t stream, argument 230 return mpAudioPolicy->get_output(mpAudioPolicy, stream, samplingRate, format, channelMask, flags); 234 audio_stream_type_t stream, 242 return mpAudioPolicy->start_output(mpAudioPolicy, output, stream, session); 246 audio_stream_type_t stream, 254 return mpAudioPolicy->stop_output(mpAudioPolicy, output, stream, session); 359 status_t AudioPolicyService::initStreamVolume(audio_stream_type_t stream, argument 369 if (uint32_t(stream) >= AUDIO_STREAM_CNT) { 373 mpAudioPolicy->init_stream_volume(mpAudioPolicy, stream, indexMin, indexMax); 377 status_t AudioPolicyService::setStreamVolumeIndex(audio_stream_type_t stream, argument 233 startOutput(audio_io_handle_t output, audio_stream_type_t stream, int session) argument 245 stopOutput(audio_io_handle_t output, audio_stream_type_t stream, int session) argument 401 getStreamVolumeIndex(audio_stream_type_t stream, int *index, audio_devices_t device) argument 422 getStrategyForStream(audio_stream_type_t stream) argument 432 getDevicesForStream(audio_stream_type_t stream) argument 477 isStreamActive(audio_stream_type_t stream, uint32_t inPastMs) const argument 782 startToneCommand(ToneGenerator::tone_type type, audio_stream_type_t stream) argument 808 volumeCommand(audio_stream_type_t stream, float volume, audio_io_handle_t output, int delayMs) argument 1013 setStreamVolume(audio_stream_type_t stream, float volume, audio_io_handle_t output, int delayMs) argument 1022 startTone(audio_policy_tone_t tone, audio_stream_type_t stream) argument 1474 aps_set_stream_output(void *service, audio_stream_type_t stream, audio_io_handle_t output) argument 1510 aps_set_stream_volume(void *service, audio_stream_type_t stream, float volume, audio_io_handle_t output, int delay_ms) argument 1520 aps_start_tone(void *service, audio_policy_tone_t tone, audio_stream_type_t stream) argument [all...] |
/frameworks/av/include/media/ |
H A D | IAudioFlinger.h | 106 /* set/get stream type state. This will probably be used by 109 virtual status_t setStreamVolume(audio_stream_type_t stream, float value, 111 virtual status_t setStreamMute(audio_stream_type_t stream, bool muted) = 0; 113 virtual float streamVolume(audio_stream_type_t stream, 115 virtual bool streamMute(audio_stream_type_t stream) const = 0; 155 virtual status_t setStreamOutput(audio_stream_type_t stream, audio_io_handle_t output) = 0; 193 // that looks on primary device for a stream with fast flag, primary flag, or first one.
|
/frameworks/base/media/java/android/media/ |
H A D | AudioSystem.java | 35 /* The audio stream for phone calls */ 37 /* The audio stream for system sounds */ 39 /* The audio stream for the phone ring and message alerts */ 41 /* The audio stream for music playback */ 43 /* The audio stream for alarms */ 45 /* The audio stream for notifications */ 47 /* @hide The audio stream for phone calls when connected on bluetooth */ 49 /* @hide The audio stream for enforced system sounds in certain countries (e.g camera in Japan) */ 51 /* @hide The audio stream for DTMF tones */ 53 /* @hide The audio stream fo 111 isStreamActive(int stream, int inPastMs) argument 384 initStreamVolume(int stream, int indexMin, int indexMax) argument 385 setStreamVolumeIndex(int stream, int index, int device) argument 386 getStreamVolumeIndex(int stream, int device) argument 391 getDevicesForStream(int stream) argument [all...] |
/frameworks/rs/ |
H A D | rsAllocation.cpp | 297 void Allocation::packVec3Allocation(Context *rsc, OStream *stream) const { 306 stream->addByteArray(dst, getPackedSize()); 312 void Allocation::serialize(Context *rsc, OStream *stream) const { 314 stream->addU32((uint32_t)getClassId()); 317 stream->addString(&name); 321 mHal.state.type->serialize(rsc, stream); 327 stream->addU32(packedSize); 330 stream->addByteArray(rsc->mHal.funcs.allocation.lock1D(rsc, this), dataSize); 334 packVec3Allocation(rsc, stream); 338 Allocation *Allocation::createFromStream(Context *rsc, IStream *stream) { argument [all...] |
H A D | rsProgramRaster.cpp | 56 void ProgramRaster::serialize(Context *rsc, OStream *stream) const { 59 ProgramRaster *ProgramRaster::createFromStream(Context *rsc, IStream *stream) { argument
|
H A D | rsScript.h | 37 virtual void serialize(Context *rsc, OStream *stream) const; 51 virtual void serialize(Context *rsc, OStream *stream) const;
|
H A D | rsFileA3D.h | 74 virtual void serialize(Context *rsc, OStream *stream) const {
|
H A D | rsPath.cpp | 62 void Path::serialize(Context *rsc, OStream *stream) const {
|
H A D | rsScriptIntrinsic.cpp | 69 void ScriptIntrinsic::serialize(Context *rsc, OStream *stream) const {
|
/frameworks/support/renderscript/v8/rs_support/ |
H A D | rsAllocation.cpp | 278 void Allocation::packVec3Allocation(Context *rsc, OStream *stream) const { 287 stream->addByteArray(dst, getPackedSize()); 293 void Allocation::serialize(Context *rsc, OStream *stream) const { 295 stream->addU32((uint32_t)getClassId()); 298 stream->addString(&name); 302 mHal.state.type->serialize(rsc, stream); 308 stream->addU32(packedSize); 311 stream->addByteArray(rsc->mHal.funcs.allocation.lock1D(rsc, this), dataSize); 315 packVec3Allocation(rsc, stream); 319 Allocation *Allocation::createFromStream(Context *rsc, IStream *stream) { argument [all...] |
H A D | rsScriptIntrinsic.cpp | 69 void ScriptIntrinsic::serialize(Context *rsc, OStream *stream) const {
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
H A D | conceal.cpp | 34 BitstreamDecVideo *stream = video->bitstream; local 38 movePointerTo(stream, startFirstPartition); 48 PV_VlcDecMCBPC_com_intra(stream);
|
/frameworks/base/cmds/screencap/ |
H A D | screencap.cpp | 177 SkDynamicMemoryWStream stream; local 178 SkImageEncoder::EncodeStream(&stream, b, 180 SkData* streamData = stream.copyToData();
|
/frameworks/base/core/java/android/os/ |
H A D | FileUtils.java | 65 * Perform an fsync on the given FileOutputStream. The stream at this 68 public static boolean sync(FileOutputStream stream) { argument 70 if (stream != null) { 71 stream.getFD().sync(); 97 * Copy data from a source stream to destFile. 148 // input stream, bytes read not equal to buffer size is not necessarily the correct
|
/frameworks/base/core/jni/android/graphics/ |
H A D | Typeface.cpp | 133 SkStream* stream = new AssetStream(asset, true); local 134 SkTypeface* face = SkTypeface::CreateFromStream(stream); 135 // SkTypeFace::CreateFromStream calls ref() on the stream, so we 137 stream->unref();
|
H A D | YuvToJpegEncoder.h | 23 /** Encode YUV data to jpeg, which is output to a stream. 25 * @param stream The jpeg output stream. 31 * @return true if successfully compressed the stream. 33 bool encode(SkWStream* stream, void* inYuv, int width,
|
/frameworks/base/voip/java/android/net/sip/ |
H A D | SipAudioCall.java | 924 * represents the RTP stream that carries the audio data to and from the 929 * @return the {@link AudioStream} object or null if the RTP stream has not 948 * @return the {@link AudioGroup} object or null if the RTP stream has not 1011 AudioStream stream = mAudioStream; 1032 stream.associate(InetAddress.getByName(address), 1035 stream.setDtmfType(-1); 1036 stream.setCodec(codec); 1042 stream.setDtmfType(type); 1048 stream.setMode(RtpStream.MODE_NORMAL); 1050 stream [all...] |
/frameworks/base/media/tests/MediaFrameworkTest/src/com/android/mediaframeworktest/unit/ |
H A D | MediaMetadataRetrieverTest.java | 92 java.io.OutputStream stream = new FileOutputStream(MediaNames.THUMBNAIL_METADATA_TEST_FILES[i] + ".jpg"); 93 bitmap.compress(Bitmap.CompressFormat.JPEG, 75, stream); 94 stream.close(); 150 java.io.OutputStream stream = new FileOutputStream("/sdcard/thumbnailout.jpg"); 151 bitmap.compress(Bitmap.CompressFormat.JPEG, 75, stream); 152 stream.close();
|
/frameworks/native/opengl/libs/GLES_trace/src/ |
H A D | gltrace_context.h | 60 BufferedOutputStream *mBufferedOutputStream; /* stream where trace info is sent */ 69 GLTraceContext(int id, int version, GLTraceState *state, BufferedOutputStream *stream); 103 GLTraceState(TCPStream *stream);
|
H A D | gltrace_transport.cpp | 131 ALOGE("Error receiving data from stream: %d", errno); 141 BufferedOutputStream::BufferedOutputStream(TCPStream *stream, size_t bufferSize) { argument 142 mStream = stream;
|
/frameworks/base/media/tests/omxjpegdecoder/ |
H A D | omx_jpeg_decoder.cpp | 99 bool OmxJpegImageDecoder::onDecode(SkStream* stream, argument 101 sp<MediaSource> source = prepareMediaSource(stream); 124 JPEGSource* OmxJpegImageDecoder::prepareMediaSource(SkStream* stream) { argument 126 sp<DataSource> dataSource = new StreamSource(stream);
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/source/ |
H A D | DecTestBench.c | 44 /* Global variables for stream handling */ 136 DEBUG(("\t-U NAL unit stream mode\n")); 181 /* check size of the input file -> length of the stream in bytes */ 186 /* allocate memory for stream buffer. if unsuccessful -> exit */ 194 /* read input stream from file to buffer and close input file */ 233 * -> stream information is available for query now */ 271 * "consumed" is computed as difference between the new stream 272 * pointer and old stream pointer */ 286 * "consumed" is computed as difference between the new stream 287 * pointer and old stream pointe 492 u8 *stream; local [all...] |
/frameworks/base/core/java/android/content/ |
H A D | ClipData.java | 103 * content URI can make available as a stream and retrieve the stream of data. 106 * itself uses this to try to retrieve a URI clip as a stream of text: 134 * text) into a stream of plain text data. 137 * stream} 297 * as a text stream from its content provider. If this succeeds, copy 322 // First see if the URI can be opened as a plain text stream 325 FileInputStream stream = null; 327 // Ask for a stream of the desired type. 330 stream [all...] |
/frameworks/av/media/libmedia/ |
H A D | IAudioFlinger.cpp | 261 virtual status_t setStreamVolume(audio_stream_type_t stream, float value, argument 266 data.writeInt32((int32_t) stream); 273 virtual status_t setStreamMute(audio_stream_type_t stream, bool muted) argument 277 data.writeInt32((int32_t) stream); 283 virtual float streamVolume(audio_stream_type_t stream, audio_io_handle_t output) const argument 287 data.writeInt32((int32_t) stream); 293 virtual bool streamMute(audio_stream_type_t stream) const 297 data.writeInt32((int32_t) stream); 485 virtual status_t setStreamOutput(audio_stream_type_t stream, audio_io_handle_t output) argument 489 data.writeInt32((int32_t) stream); 809 int stream = data.readInt32(); local 817 int stream = data.readInt32(); local 823 int stream = data.readInt32(); local 830 int stream = data.readInt32(); local 954 uint32_t stream = data.readInt32(); local [all...] |
/frameworks/base/graphics/java/android/graphics/ |
H A D | BitmapFactory.java | 55 * resource or as a stream). The {@link android.graphics.Bitmap.Config 299 InputStream stream = null; 301 stream = new FileInputStream(pathName); 302 bm = decodeStream(stream, null, opts); 307 Log.e("BitmapFactory", "Unable to decode stream: " + e); 309 if (stream != null) { 311 stream.close(); 450 * Decode an input stream into a bitmap. If the input stream is null, or 452 * The stream' [all...] |