/frameworks/rs/ |
H A D | rsAllocation.cpp | 320 void Allocation::packVec3Allocation(Context *rsc, OStream *stream) const { 329 stream->addByteArray(dst, getPackedSize()); 335 void Allocation::serialize(Context *rsc, OStream *stream) const { 337 stream->addU32((uint32_t)getClassId()); 338 stream->addString(getName()); 342 mHal.state.type->serialize(rsc, stream); 348 stream->addU32(packedSize); 351 stream->addByteArray(rsc->mHal.funcs.allocation.lock1D(rsc, this), dataSize); 355 packVec3Allocation(rsc, stream); 359 Allocation *Allocation::createFromStream(Context *rsc, IStream *stream) { argument [all...] |
H A D | rsProgramRaster.cpp | 56 void ProgramRaster::serialize(Context *rsc, OStream *stream) const { 59 ProgramRaster *ProgramRaster::createFromStream(Context *rsc, IStream *stream) { argument
|
H A D | rsScript.h | 40 virtual void serialize(Context *rsc, OStream *stream) const; 54 virtual void serialize(Context *rsc, OStream *stream) const;
|
H A D | rsFileA3D.h | 75 virtual void serialize(Context *rsc, OStream *stream) const {
|
H A D | rsPath.cpp | 62 void Path::serialize(Context *rsc, OStream *stream) const {
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
H A D | conceal.cpp | 34 BitstreamDecVideo *stream = video->bitstream; local 38 movePointerTo(stream, startFirstPartition); 48 PV_VlcDecMCBPC_com_intra(stream);
|
H A D | mp4lib_int.h | 276 PV_STATUS(*vlcDecCoeffIntra)(BitstreamDecVideo *stream, Tcoef *pTcoef/*, int intra_luma*/); 277 PV_STATUS(*vlcDecCoeffInter)(BitstreamDecVideo *stream, Tcoef *pTcoef);
|
/frameworks/av/media/libstagefright/httplive/ |
H A D | LiveSession.h | 57 status_t dequeueAccessUnit(StreamType stream, sp<ABuffer> *accessUnit); 59 status_t getStreamFormat(StreamType stream, sp<AMessage> *format); 246 void swapPacketSource(StreamType stream);
|
H A D | LiveSession.cpp | 98 void LiveSession::swapPacketSource(StreamType stream) { argument 99 sp<AnotherPacketSource> &aps = mPacketSources.editValueFor(stream); 100 sp<AnotherPacketSource> &aps2 = mPacketSources2.editValueFor(stream); 108 StreamType stream, sp<ABuffer> *accessUnit) { 109 if (!(mStreamMask & stream)) { 115 sp<AnotherPacketSource> packetSource = mPacketSources.valueFor(stream); 125 switch (stream) { 163 swapPacketSource(stream); 165 msg->setInt32("stream", stream); 107 dequeueAccessUnit( StreamType stream, sp<ABuffer> *accessUnit) argument 191 getStreamFormat(StreamType stream, sp<AMessage> *format) argument 1260 int32_t stream; local [all...] |
/frameworks/base/cmds/screencap/ |
H A D | screencap.cpp | 180 SkDynamicMemoryWStream stream; local 181 SkImageEncoder::EncodeStream(&stream, b, 183 SkData* streamData = stream.copyToData();
|
/frameworks/base/core/jni/android/graphics/ |
H A D | Typeface.cpp | 95 SkStream* stream = new AssetStreamAdaptor(asset, local 98 SkTypeface* face = SkTypeface::CreateFromStream(stream); 99 // SkTypeFace::CreateFromStream calls ref() on the stream, so we 101 stream->unref();
|
H A D | YuvToJpegEncoder.h | 23 /** Encode YUV data to jpeg, which is output to a stream. 25 * @param stream The jpeg output stream. 31 * @return true if successfully compressed the stream. 33 bool encode(SkWStream* stream, void* inYuv, int width,
|
/frameworks/opt/net/voip/src/java/android/net/sip/ |
H A D | SipAudioCall.java | 931 * represents the RTP stream that carries the audio data to and from the 936 * @return the {@link AudioStream} object or null if the RTP stream has not 955 * @return the {@link AudioGroup} object or null if the RTP stream has not 1019 AudioStream stream = mAudioStream; 1040 stream.associate(InetAddress.getByName(address), 1043 stream.setDtmfType(-1); 1044 stream.setCodec(codec); 1050 stream.setDtmfType(type); 1056 stream.setMode(RtpStream.MODE_NORMAL); 1058 stream [all...] |
/frameworks/base/media/tests/MediaFrameworkTest/src/com/android/mediaframeworktest/unit/ |
H A D | MediaMetadataRetrieverTest.java | 92 java.io.OutputStream stream = new FileOutputStream(MediaNames.THUMBNAIL_METADATA_TEST_FILES[i] + ".jpg"); 93 bitmap.compress(Bitmap.CompressFormat.JPEG, 75, stream); 94 stream.close(); 150 java.io.OutputStream stream = new FileOutputStream("/sdcard/thumbnailout.jpg"); 151 bitmap.compress(Bitmap.CompressFormat.JPEG, 75, stream); 152 stream.close();
|
/frameworks/av/services/camera/libcameraservice/device3/ |
H A D | Camera3Stream.h | 36 * A class for managing a single stream of input or output data from the camera 39 * The stream has an internal state machine to track whether it's 44 * STATE_ERROR: A serious error has occurred, stream is unusable. Outstanding 47 * STATE_CONSTRUCTED: The stream is ready for configuration, but buffers cannot 55 * STATE_IN_RE_CONFIG: Configuration has started, and the stream has been 60 * HAL. The stream's getBuffer/returnBuffer work. The priv pointer may still be 76 * When startConfiguration() is called again, after making sure stream is 83 * When disconnect() is called after making sure stream is idle with 94 static Camera3Stream* cast(camera3_stream *stream); 95 static const Camera3Stream* cast(const camera3_stream *stream); [all...] |
H A D | Camera3Stream.cpp | 37 Camera3Stream* Camera3Stream::cast(camera3_stream *stream) { argument 38 return static_cast<Camera3Stream*>(stream); 41 const Camera3Stream* Camera3Stream::cast(const camera3_stream *stream) { argument 42 return static_cast<const Camera3Stream*>(stream); 104 ALOGE("%s: Cannot configure an input stream twice", 108 ALOGE("%s: Cannot configure stream; has outstanding buffers", 178 // Check if the stream configuration is unchanged, and skip reallocation if 190 ALOGE("%s: Unable to configure stream %d queue: %s (%d)", 198 ALOGE("%s: Unable to register stream buffers with HAL: %s (%d)", 325 bufferSet.stream [all...] |
/frameworks/native/opengl/libs/GLES_trace/src/ |
H A D | gltrace_context.h | 60 BufferedOutputStream *mBufferedOutputStream; /* stream where trace info is sent */ 69 GLTraceContext(int id, int version, GLTraceState *state, BufferedOutputStream *stream); 103 GLTraceState(TCPStream *stream);
|
H A D | gltrace_transport.cpp | 131 ALOGE("Error receiving data from stream: %d", errno); 141 BufferedOutputStream::BufferedOutputStream(TCPStream *stream, size_t bufferSize) { argument 142 mStream = stream;
|
/frameworks/base/media/tests/omxjpegdecoder/ |
H A D | omx_jpeg_decoder.cpp | 100 bool OmxJpegImageDecoder::onDecode(SkStream* stream, argument 102 sp<MediaSource> source = prepareMediaSource(stream); 125 JPEGSource* OmxJpegImageDecoder::prepareMediaSource(SkStream* stream) { argument 127 sp<DataSource> dataSource = new StreamSource(stream);
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/source/ |
H A D | DecTestBench.c | 44 /* Global variables for stream handling */ 136 DEBUG(("\t-U NAL unit stream mode\n")); 181 /* check size of the input file -> length of the stream in bytes */ 186 /* allocate memory for stream buffer. if unsuccessful -> exit */ 194 /* read input stream from file to buffer and close input file */ 233 * -> stream information is available for query now */ 271 * "consumed" is computed as difference between the new stream 272 * pointer and old stream pointer */ 286 * "consumed" is computed as difference between the new stream 287 * pointer and old stream pointe 492 u8 *stream; local [all...] |
/frameworks/base/core/java/android/content/ |
H A D | ClipData.java | 104 * content URI can make available as a stream and retrieve the stream of data. 107 * itself uses this to try to retrieve a URI clip as a stream of text: 135 * text) into a stream of plain text data. 138 * stream} 298 * as a text stream from its content provider. If this succeeds, copy 323 // First see if the URI can be opened as a plain text stream 326 FileInputStream stream = null; 328 // Ask for a stream of the desired type. 331 stream [all...] |
/frameworks/av/media/libmedia/ |
H A D | IAudioFlinger.cpp | 277 virtual status_t setStreamVolume(audio_stream_type_t stream, float value, argument 282 data.writeInt32((int32_t) stream); 289 virtual status_t setStreamMute(audio_stream_type_t stream, bool muted) argument 293 data.writeInt32((int32_t) stream); 299 virtual float streamVolume(audio_stream_type_t stream, audio_io_handle_t output) const argument 303 data.writeInt32((int32_t) stream); 309 virtual bool streamMute(audio_stream_type_t stream) const 313 data.writeInt32((int32_t) stream); 509 virtual status_t setStreamOutput(audio_stream_type_t stream, audio_io_handle_t output) argument 513 data.writeInt32((int32_t) stream); 850 int stream = data.readInt32(); local 858 int stream = data.readInt32(); local 864 int stream = data.readInt32(); local 871 int stream = data.readInt32(); local 1002 uint32_t stream = data.readInt32(); local [all...] |
/frameworks/base/tools/velocityplot/ |
H A D | velocityplot.py | 42 # Non-blocking stream wrapper. 44 def __init__(self, stream): 45 fcntl.fcntl(stream, fcntl.F_SETFL, os.O_NONBLOCK) 46 self.stream = stream 61 chunk = os.read(self.stream.fileno(), 4096)
|
/frameworks/base/core/jni/android/graphics/pdf/ |
H A D | PdfDocument.cpp | 82 void write(SkWStream* stream) { argument 83 SkDocument* document = SkDocument::CreatePDF(stream);
|
/frameworks/base/graphics/java/android/renderscript/ |
H A D | FileA3D.java | 172 FileA3D(int id, RenderScript rs, InputStream stream) { argument 174 mInputStream = stream; 303 throw new RSRuntimeException("Unsupported asset stream");
|