/frameworks/ex/framesequence/jni/ |
H A D | FrameSequence_webp.cpp | 87 FrameSequence_webp::FrameSequence_webp(Stream* stream) { argument 88 if (stream->getRawBuffer() != NULL) { 89 mData.size = stream->getRawBufferSize(); 90 mData.bytes = stream->getRawBufferAddr(); 91 mRawByteBuffer = stream->getRawBuffer(); 95 if (stream->read(riff_header, RIFF_HEADER_SIZE) != RIFF_HEADER_SIZE) { 106 if (stream->read(remaining_bytes, remaining_size) != remaining_size) { 381 static FrameSequence* createFramesequence(Stream* stream) { argument 382 return new FrameSequence_webp(stream);
|
H A D | FrameSequence_gif.cpp | 32 Stream* stream = (Stream*) fileType->UserData; local 33 return (int) stream->read(out, size); 56 FrameSequence_gif::FrameSequence_gif(Stream* stream) : argument 58 mGif = DGifOpen(stream, streamReader, NULL); 359 static FrameSequence* createFramesequence(Stream* stream) { argument 360 return new FrameSequence_gif(stream);
|
H A D | FrameSequence_gif.h | 29 FrameSequence_gif(Stream* stream);
|
H A D | FrameSequence_webp.h | 31 FrameSequence_webp(Stream* stream);
|
/frameworks/av/include/media/ |
H A D | IAudioFlinger.h | 120 /* set/get stream type state. This will probably be used by 123 virtual status_t setStreamVolume(audio_stream_type_t stream, float value, 125 virtual status_t setStreamMute(audio_stream_type_t stream, bool muted) = 0; 127 virtual float streamVolume(audio_stream_type_t stream, 129 virtual bool streamMute(audio_stream_type_t stream) const = 0; 175 virtual status_t invalidateStream(audio_stream_type_t stream) = 0; 215 // that looks on primary device for a stream with fast flag, primary flag, or first one.
|
/frameworks/base/graphics/java/android/graphics/ |
H A D | BitmapFactory.java | 80 * resource or as a stream) must be in jpeg or png format. Only equal 388 InputStream stream = null; 390 stream = new FileInputStream(pathName); 391 bm = decodeStream(stream, null, opts); 396 Log.e("BitmapFactory", "Unable to decode stream: " + e); 398 if (stream != null) { 400 stream.close(); 574 * Decode an input stream into a bitmap. If the input stream is null, or 576 * The stream' [all...] |
/frameworks/av/services/audioflinger/ |
H A D | AudioFlinger.h | 151 virtual status_t setStreamVolume(audio_stream_type_t stream, float value, 153 virtual status_t setStreamMute(audio_stream_type_t stream, bool muted); 155 virtual float streamVolume(audio_stream_type_t stream, 157 virtual bool streamMute(audio_stream_type_t stream) const; 199 virtual status_t invalidateStream(audio_stream_type_t stream); 562 bool streamMute_l(audio_stream_type_t stream) const 563 { return mStreamTypes[stream].mute; } 564 // no range check, doesn't check per-thread stream volume, AudioFlinger::mLock held 565 float streamVolume_l(audio_stream_type_t stream) const 566 { return mStreamTypes[stream] 622 audio_stream_in_t* const stream; member in struct:android::AudioFlinger::AudioStreamIn [all...] |
H A D | AudioStreamOut.h | 33 * Managed access to a HAL output stream. 41 audio_stream_out_t *stream; member in class:android::AudioStreamOut
|
/frameworks/base/tools/layoutlib/bridge/src/com/android/layoutlib/bridge/bars/ |
H A D | StatusBar.java | 100 InputStream stream = iconLoader.getIcon(); 102 if (stream != null) { 105 ParserFactory.create(stream, null), (BridgeContext) mContext, true);
|
/frameworks/opt/net/voip/src/java/android/net/sip/ |
H A D | SipAudioCall.java | 931 * represents the RTP stream that carries the audio data to and from the 936 * @return the {@link AudioStream} object or null if the RTP stream has not 955 * @return the {@link AudioGroup} object or null if the RTP stream has not 1019 AudioStream stream = mAudioStream; 1040 stream.associate(InetAddress.getByName(address), 1043 stream.setDtmfType(-1); 1044 stream.setCodec(codec); 1050 stream.setDtmfType(type); 1056 stream.setMode(RtpStream.MODE_NORMAL); 1058 stream [all...] |
/frameworks/av/services/audiopolicy/service/ |
H A D | AudioPolicyClientImplLegacy.cpp | 245 int aps_invalidate_stream(void *service __unused, audio_stream_type_t stream) argument 252 return af->invalidateStream(stream); 282 int aps_set_stream_volume(void *service, audio_stream_type_t stream, argument 288 return audioPolicyService->setStreamVolume(stream, volume, output, 293 audio_stream_type_t stream) 297 return audioPolicyService->startTone(tone, stream); 292 aps_start_tone(void *service, audio_policy_tone_t tone, audio_stream_type_t stream) argument
|
H A D | AudioPolicyEffects.cpp | 222 audio_stream_type_t stream, 228 // create audio processors according to stream 231 if (stream >= AUDIO_STREAM_PUBLIC_CNT) { 232 stream = AUDIO_STREAM_MUSIC; 234 ssize_t index = mOutputStreams.indexOfKey(stream); 236 ALOGV("addOutputSessionEffects(): no output processing needed for this stream"); 266 ALOGV("addOutputSessionEffects(): added Fx %s on session: %d for stream: %d", 267 effect->mName, audioSession, (int32_t)stream); 277 audio_stream_type_t stream, 282 (void) stream; // argumen 221 addOutputSessionEffects(audio_io_handle_t output, audio_stream_type_t stream, int audioSession) argument 276 releaseOutputSessionEffects(audio_io_handle_t output, audio_stream_type_t stream, int audioSession) argument 592 audio_stream_type_t stream = streamNameToEnum(node->name); local [all...] |
/frameworks/base/core/java/android/widget/ |
H A D | TextClock.java | 583 protected void encodeProperties(@NonNull ViewHierarchyEncoder stream) { argument 584 super.encodeProperties(stream); 587 stream.addProperty("format12Hour", s == null ? null : s.toString()); 590 stream.addProperty("format24Hour", s == null ? null : s.toString()); 591 stream.addProperty("format", mFormat == null ? null : mFormat.toString()); 592 stream.addProperty("hasSeconds", mHasSeconds);
|
/frameworks/base/media/tests/MediaFrameworkTest/src/com/android/mediaframeworktest/unit/ |
H A D | MediaMetadataRetrieverTest.java | 92 java.io.OutputStream stream = new FileOutputStream(MediaNames.THUMBNAIL_METADATA_TEST_FILES[i] + ".jpg"); 93 bitmap.compress(Bitmap.CompressFormat.JPEG, 75, stream); 94 stream.close(); 150 java.io.OutputStream stream = new FileOutputStream("/sdcard/thumbnailout.jpg"); 151 bitmap.compress(Bitmap.CompressFormat.JPEG, 75, stream); 152 stream.close();
|
/frameworks/base/tools/layoutlib/bridge/tests/src/com/android/layoutlib/bridge/intensive/setup/ |
H A D | LayoutPullParser.java | 66 private void init(InputStream stream) throws XmlPullParserException { argument 68 setInput(stream, null);
|
/frameworks/compile/slang/ |
H A D | llvm-rs-as.cpp | 138 raw_string_ostream stream(Err); 139 if (verifyModule(*M.get(), &stream)) {
|
/frameworks/native/opengl/libs/GLES_trace/src/ |
H A D | gltrace_context.h | 63 BufferedOutputStream *mBufferedOutputStream; /* stream where trace info is sent */ 75 GLTraceContext(int id, int version, GLTraceState *state, BufferedOutputStream *stream); 111 GLTraceState(TCPStream *stream);
|
H A D | gltrace_transport.cpp | 131 ALOGE("Error receiving data from stream: %d", errno); 141 BufferedOutputStream::BufferedOutputStream(TCPStream *stream, size_t bufferSize) { argument 142 mStream = stream;
|
/frameworks/rs/ |
H A D | rsProgramRaster.cpp | 56 void ProgramRaster::serialize(Context *rsc, OStream *stream) const { 59 ProgramRaster *ProgramRaster::createFromStream(Context *rsc, IStream *stream) { argument
|
H A D | rsClosure.h | 42 virtual void serialize(Context *rsc, OStream *stream) const {}
|
H A D | rsFileA3D.h | 75 virtual void serialize(Context *rsc, OStream *stream) const {
|
/frameworks/base/core/jni/android/graphics/ |
H A D | YuvToJpegEncoder.h | 23 /** Encode YUV data to jpeg, which is output to a stream. 25 * @param stream The jpeg output stream. 31 * @return true if successfully compressed the stream. 33 bool encode(SkWStream* stream, void* inYuv, int width,
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/source/ |
H A D | DecTestBench.c | 44 /* Global variables for stream handling */ 136 DEBUG(("\t-U NAL unit stream mode\n")); 181 /* check size of the input file -> length of the stream in bytes */ 186 /* allocate memory for stream buffer. if unsuccessful -> exit */ 194 /* read input stream from file to buffer and close input file */ 233 * -> stream information is available for query now */ 271 * "consumed" is computed as difference between the new stream 272 * pointer and old stream pointer */ 286 * "consumed" is computed as difference between the new stream 287 * pointer and old stream pointe 492 u8 *stream; local [all...] |
/frameworks/base/core/java/android/content/ |
H A D | ClipData.java | 105 * content URI can make available as a stream and retrieve the stream of data. 108 * itself uses this to try to retrieve a URI clip as a stream of text: 136 * text) into a stream of plain text data. 139 * stream} 299 * as a text stream from its content provider. If this succeeds, copy 324 // First see if the URI can be opened as a plain text stream 327 FileInputStream stream = null; 329 // Ask for a stream of the desired type. 332 stream [all...] |
/frameworks/base/tools/velocityplot/ |
H A D | velocityplot.py | 42 # Non-blocking stream wrapper. 44 def __init__(self, stream): 45 fcntl.fcntl(stream, fcntl.F_SETFL, os.O_NONBLOCK) 46 self.stream = stream 61 chunk = os.read(self.stream.fileno(), 4096)
|