/frameworks/base/tests/TileBenchmark/src/com/test/tilebenchmark/ |
H A D | RunData.java | 23 public TileData[][] frames; field in class:RunData 26 public RunData(int frames) { argument 27 this.frames = new TileData[frames][];
|
H A D | PlaybackView.java | 150 if (mProfData == null || mProfData.frames.length == 0) { 159 for (TileData t : mProfData.frames[frame]) { 160 if (t == mProfData.frames[frame][0]){ 185 for (TileData t : mProfData.frames[frame]) { 186 if (t.level == INVAL_FLAG && t != mProfData.frames[frame][0]) { 212 TileDrawable viewShape = new TileDrawable(mProfData.frames[frame][0],
|
H A D | ProfiledWebView.java | 211 for (int frame = 0; frame < data.frames.length; frame++) { 212 data.frames[frame] = new TileData[ 214 for (int tile = 0; tile < data.frames[frame].length; tile++) { 226 data.frames[frame][tile] = data.new TileData(left, top, right, bottom,
|
H A D | PlaybackGraphs.java | 216 for (int frameIndex = 0; frameIndex < data.frames.length; frameIndex++) { 217 TileData frame[] = data.frames[frameIndex]; 240 double metricValues[] = new double[data.frames.length]; 244 if (data.frames.length == 0) { 249 // calculate metric based on list of frames
|
H A D | PlaybackActivity.java | 113 mFrameMax = data.frames.length - 1;
|
H A D | ProfileActivity.java | 255 Log.d("ProfileActivity", "stored " + data.frames.length + " frames in file");
|
/frameworks/av/media/libstagefright/rtsp/ |
H A D | ARTPAssembler.h | 53 const List<sp<ABuffer> > &frames); 56 const List<sp<ABuffer> > &frames);
|
H A D | ARTPAssembler.cpp | 76 const List<sp<ABuffer> > &frames) { 78 for (List<sp<ABuffer> >::const_iterator it = frames.begin(); 79 it != frames.end(); ++it) { 86 for (List<sp<ABuffer> >::const_iterator it = frames.begin(); 87 it != frames.end(); ++it) { 116 CopyTimes(accessUnit, *frames.begin()); 72 MakeADTSCompoundFromAACFrames( unsigned profile, unsigned samplingFreqIndex, unsigned channelConfig, const List<sp<ABuffer> > &frames) argument
|
/frameworks/wilhelm/tools/permute/ |
H A D | permute.c | 75 /** Split the specified range of frames, using the allowed budget of segments. 202 switch (sfinfo_in.frames) { 205 fprintf(stderr, "%s: unsupported frames %d\n", path_in, (int) sfinfo_in.frames); 212 double durationSeconds = (double) sfinfo_in.frames / (double) sfinfo_in.samplerate; 224 used = split(&s, 0, sfinfo_in.frames, s.mSegmentMax); 241 void *ptr = malloc(sfinfo_in.frames * frameSizeRead); 244 count = sf_readf_short(sf_in, ptr, sfinfo_in.frames); 245 if (count != sfinfo_in.frames) { 246 fprintf(stderr, "%s: expected to read %d frames bu [all...] |
/frameworks/native/opengl/tests/hwc/ |
H A D | hwcStress.cpp | 199 static vector <vector <sp<GraphicBuffer> > > frames; variable 396 // Regenerate a new set of test frames when this pass is 413 list = hwcTestCreateLayerList(testRandMod(frames.size()) + 1); 419 // Prandomly select a subset of frames to be used by this pass. 421 selectedFrames = vectorRandSelect(frames, list->numHwLayers); 545 * named frames. The graphic buffers are contained within a vector of 562 frames.clear(); 563 frames.resize(rows); 566 // All frames within a row have to have the same format and 591 frames[ro [all...] |
/frameworks/av/media/libstagefright/ |
H A D | XINGSeeker.cpp | 139 int32_t frames = U32_AT(buffer); local 144 if (frames) { 145 seeker->mDurationUs = (int64_t)frames * samples_per_frame * 1000000LL / sampling_rate;
|
/frameworks/av/media/libnbaio/ |
H A D | LibsndfileSource.cpp | 29 mEstimatedFramesUntilEOF(sfinfo.frames), 55 // Detect EOF by zero frames read, not by mFramesUntilEOF as it could be inaccurate 62 // We didn't read any frames during the current loop cycle, so disable
|
/frameworks/base/cmds/bootanimation/ |
H A D | BootAnimation.h | 72 SortedVector<Frame> frames; member in struct:android::BootAnimation::Animation::Part
|
H A D | BootAnimation.cpp | 461 part.frames.add(frame); 499 const size_t fcount = part.frames.size(); 508 const Animation::Frame& frame(part.frames[j]); 568 const Animation::Frame& frame(part.frames[j]);
|
/frameworks/av/media/libmedia/ |
H A D | AudioTrack.cpp | 1197 uint32_t frames; local 1248 frames = 0; 1250 frames = mRemainingFrames; 1263 audioBuffer.frameCount = frames; 1310 frames -= audioBuffer.frameCount; 1314 while (frames); 1316 if (frames == 0) { 1319 mRemainingFrames = frames; 1370 // the actual amount of audio frames played (e.g SoundPool) receives them. 1372 uint32_t frames local 1648 uint32_t frames = UINT_MAX; local [all...] |
H A D | AudioRecord.cpp | 659 uint32_t frames = mRemainingFrames; local 697 audioBuffer.frameCount = frames; 728 frames -= audioBuffer.frameCount; 732 } while (frames); 745 if (frames == 0) { 748 mRemainingFrames = frames;
|
/frameworks/av/include/media/ |
H A D | MediaProfiles.h | 54 * Set default maximum prefetch YUV frames to 6, which means video editor can 55 * queue up to 6 YUV frames in the video encoder source. 57 * engine when the encoder consumes YUV frames at a lower speed 134 * enc.vid.fps.min - min frame rate in frames per second 135 * enc.vid.fps.max - max frame rate in frames per second 147 * maxPrefetchYUVFrames - max prefetch YUV frames in video editor engine. This value is used 368 int outFrameWidth, int outFrameHeight, int frames) 373 mMaxPrefetchYUVFrames(frames) {} 367 VideoEditorCap(int inFrameWidth, int inFrameHeight, int outFrameWidth, int outFrameHeight, int frames) argument
|
/frameworks/wilhelm/tests/sandbox/ |
H A D | playbq.c | 172 fprintf(stderr, " -f# frames per buffer (default 512)\n"); 378 sf_count_t frames = framesPerBuffer; local 380 count = sf_readf_short(sndfile, buffer, frames);
|
/frameworks/wilhelm/src/desktop/ |
H A D | SndFile.c | 261 thiz->mPlay.mDuration = (SLmillisecond) (((long long) thiz->mSndFile.mSfInfo.frames *
|
/frameworks/base/services/java/com/android/server/ |
H A D | TextServicesManagerService.java | 972 final StackTraceElement[] frames = e.getStackTrace(); 974 for (int j = 1; j < frames.length; ++j) { 975 sb.append(frames[j].toString() + "\n");
|
H A D | InputMethodManagerService.java | 4058 final StackTraceElement[] frames = e.getStackTrace(); 4060 for (int j = 1; j < frames.length; ++j) { 4061 sb.append(frames[j].toString() + "\n");
|
/frameworks/av/services/audioflinger/ |
H A D | AudioFlinger.cpp | 137 // minimum normal mix buffer size, expressed in milliseconds rather than frames 2059 ALOGW("HAL output buffer size is %u frames but AudioMixer requires multiples of 16 frames", 2068 // round up minimum and round down maximum to nearest 16 frames to satisfy AudioMixer 2098 // round up to nearest 16 frames to satisfy AudioMixer 2100 ALOGI("HAL output buffer size %u frames, normal mix buffer size %u frames", mFrameCount, mNormalFrameCount); 2131 // return an estimation of rendered frames when the output is suspended 2132 int32_t frames = mBytesWritten - latency_l(); local 2133 if (frames < 5542 write(int16_t* data, uint32_t frames) argument [all...] |
H A D | AudioFlinger.h | 429 // Parameter 'frames' is the requested length, also in frame units. 432 // function with invalid 'offset' or 'frames' is undefined. 433 void* getBuffer(uint32_t offset, uint32_t frames) const; 879 size_t mPresentationCompleteFrames; // number of frames written to the audio HAL 1005 bool write(int16_t* data, uint32_t frames); 1571 // number of captured frames to drop after the start sync event has been received. 1572 // when < 0, maximum frames to drop before starting capture even if sync event is
|