/frameworks/base/media/mca/filterfw/java/android/filterfw/core/ |
H A D | OutputPort.java | 90 public void pushFrame(Frame frame) { argument 93 "Attempting to push frame on unconnected port: " + this + "!"); 95 mTargetPort.pushFrame(frame); 99 public void setFrame(Frame frame) { argument 103 "Attempting to set frame on unconnected port: " + this + "!"); 105 mTargetPort.setFrame(frame); 110 throw new RuntimeException("Cannot pull frame on " + this + "!");
|
H A D | VertexFrame.java | 37 throw new IllegalArgumentException("Initializing vertex frame with zero size!"); 40 throw new RuntimeException("Could not allocate vertex frame!"); 65 throw new RuntimeException("Could not set int values for vertex frame!"); 78 throw new RuntimeException("Could not set int values for vertex frame!"); 92 throw new RuntimeException("Data size in setData does not match vertex frame size!"); 94 throw new RuntimeException("Could not set vertex frame data!"); 105 throw new RuntimeException("Unsupported: Cannot set vertex frame bitmap value!"); 114 public void setDataFromFrame(Frame frame) { argument 116 super.setDataFromFrame(frame);
|
H A D | FilterPort.java | 98 public abstract void pushFrame(Frame frame); argument 100 public abstract void setFrame(Frame frame); argument 118 protected void checkFrameType(Frame frame, boolean forceCheck) { argument 121 && !frame.getFormat().isCompatibleWith(mPortFormat)) { 123 + "Expected " + mPortFormat + " but got " + frame.getFormat()); 127 protected void checkFrameManager(Frame frame, FilterContext context) { argument 128 if (frame.getFrameManager() != null 129 && frame.getFrameManager() != context.getFrameManager()) { 130 throw new RuntimeException("Frame " + frame + " is managed by foreign FrameManager! ");
|
H A D | NativeFrame.java | 60 * Returns the native frame's Object value. 62 * If the frame's base-type is not TYPE_OBJECT, this returns a data buffer containing the native 64 * If the frame is based on an object type, this type is expected to be a subclass of 65 * NativeBuffer. The NativeBuffer returned is only valid for as long as the frame is alive. If 70 // If this is not a structured frame, return our data 78 throw new RuntimeException("Attempting to get object data from frame that does " + 99 throw new RuntimeException("Could not get the native structured data for frame!"); 102 // Attach this frame to it 116 throw new RuntimeException("Could not set int values for native frame!"); 133 throw new RuntimeException("Could not set int values for native frame!"); 202 setDataFromFrame(Frame frame) argument 260 nativeCopyFromNative(NativeFrame frame) argument 262 nativeCopyFromGL(GLFrame frame) argument [all...] |
/frameworks/base/media/mca/filterpacks/java/android/filterpacks/base/ |
H A D | CallbackFilter.java | 52 public CallbackRunnable(FilterContext.OnFrameReceivedListener listener, Filter filter, Frame frame, Object userData) { argument 55 mFrame = frame; 71 addInputPort("frame"); 81 // Get frame and forward to listener 82 final Frame input = pullInput("frame"); 94 throw new RuntimeException("CallbackFilter received frame, but no listener set!");
|
/frameworks/base/media/mca/filterpacks/java/android/filterpacks/imageproc/ |
H A D | ImageSlicer.java | 73 private void calcOutputFormatForInput(Frame frame) { argument 76 mInputWidth = frame.getFormat().getWidth(); 77 mInputHeight = frame.getFormat().getHeight(); 90 // Get input frame 100 // Create output frame 129 // Retain the original frame so it can be used next time. 137 // Release pushed frame
|
/frameworks/av/media/libmediaplayerservice/ |
H A D | MetadataRetrieverClient.cpp | 206 VideoFrame *frame = mRetriever->getFrameAtTime(timeUs, option); local 207 if (frame == NULL) { 208 ALOGE("failed to capture a video frame"); 211 size_t size = sizeof(VideoFrame) + frame->mSize; 215 delete frame; 221 delete frame; 225 frameCopy->mWidth = frame->mWidth; 226 frameCopy->mHeight = frame->mHeight; 227 frameCopy->mDisplayWidth = frame->mDisplayWidth; 228 frameCopy->mDisplayHeight = frame [all...] |
/frameworks/av/services/camera/libcameraservice/api1/client2/ |
H A D | FrameProcessor.cpp | 65 bool FrameProcessor::processSingleFrame(CaptureResult &frame, argument 75 isPartialResult = frame.mResultExtras.partialResultCount < mNumPartialResults; 78 if (!isPartialResult && processFaceDetect(frame.mMetadata, client) != OK) { 83 process3aState(frame, client); 86 return FrameProcessorBase::processSingleFrame(frame, device); 89 status_t FrameProcessor::processFaceDetect(const CameraMetadata &frame, 100 entry = frame.find(ANDROID_STATISTICS_FACE_DETECT_MODE); 117 entry = frame.find(ANDROID_STATISTICS_FACE_RECTANGLES); 119 // No faces this frame 134 entry = frame 234 process3aState(const CaptureResult &frame, const sp<Camera2Client> &client) argument [all...] |
H A D | ZslProcessor.h | 122 CameraMetadata frame; member in struct:android::camera2::ZslProcessor::ZslPair
|
/frameworks/base/core/java/android/view/ |
H A D | DisplayEventReceiver.java | 49 * Surface Flinger is processing a frame. 131 * The recipient should render a frame and then call {@link #scheduleVsync} 138 * @param frame The frame number. Increases by one for each vertical sync interval. 140 public void onVsync(long timestampNanos, int builtInDisplayId, int frame) { argument 157 * display frame begins. 170 private void dispatchVsync(long timestampNanos, int builtInDisplayId, int frame) { argument 171 onVsync(timestampNanos, builtInDisplayId, frame);
|
/frameworks/base/tests/Camera2Tests/SmartCamera/SimpleCamera/src/androidx/media/filterfw/ |
H A D | FrameQueue.java | 85 public void pushFrame(Frame frame); argument 117 // Note that we cannot use a cached frame here, as we do not know where that cached 124 public void pushFrame(Frame frame) { argument 125 mFrame = frame.retain(); 162 public void pushFrame(Frame frame) { argument 163 mQueueImpl.pushFrame(frame);
|
H A D | OutputPort.java | 24 * Input and output ports share their Frame slot, meaning that when a frame is waiting on an output 27 * Only one frame can be pushed onto an output port at a time. In other words, a Frame must first 28 * be consumed by the target filter before a new frame can be pushed on the output port. If the 52 * Returns true, if there is no frame waiting on this port. 60 * Returns a frame for writing. 62 * Call this method to fetch a new frame to write into. When you have finished writing the 63 * frame data, you can push it into the output queue using {@link #pushFrame(Frame)}. Note, 64 * that the Frame returned is owned by the queue. If you wish to hold on to the frame, you 71 Frame frame = getQueue().fetchAvailableFrame(dimensions); 72 if (frame ! 91 pushFrame(Frame frame) argument [all...] |
/frameworks/base/tests/Camera2Tests/SmartCamera/SimpleCamera/tests/src/androidx/media/filterfw/ |
H A D | MffFilterTestCase.java | 91 for (Frame frame : mOutputFrames.values()) { 92 frame.release(); 104 protected void injectInputFrame(String portName, Frame frame) { argument 106 filter.injectFrame(frame); 110 * Returns the frame pushed out by the filter under test. Should only be called after 164 public void onFramePushed(String filterName, Frame frame) { argument 165 mOutputFrames.put(filterName, frame); 169 "A frame has been pushed twice to the same output port.");
|
/frameworks/av/media/libstagefright/codecs/amrnb/dec/src/ |
H A D | dtx_dec.cpp | 204 // initialize decoder log frame energy 312 /* initialize decoder log frame energy */ 580 Set_zero(lsfState->past_r_q, M); // reset for next speech frame 1073 Set_zero(lsfState->past_r_q, M); /* reset for next speech frame */ 1396 frame = 1430 Word16 frame[]) 1445 // compute log energy based on frame energy 1449 L_frame_en = L_mac(L_frame_en, frame[i], frame[i]); 1495 Word16 frame[], 1493 dtx_dec_activity_update(dtx_decState *st, Word16 lsf[], Word16 frame[], Flag *pOverflow) argument [all...] |
/frameworks/base/core/java/com/android/internal/view/ |
H A D | BaseIWindow.java | 41 public void resized(Rect frame, Rect overscanInsets, Rect contentInsets, Rect visibleInsets, argument
|
/frameworks/base/graphics/java/android/graphics/drawable/ |
H A D | AnimationDrawable.java | 34 * An object used to create frame-by-frame animations, defined by a series of 37 * The simplest way to create a frame-by-frame animation is to define the 44 * {@code <item>} tags. Each item defines a frame of the animation. See 90 /** The current frame, ranging from 0 to {@link #mAnimationState#getChildCount() - 1} */ 110 * first frame. If <code>restart</code> is false, the drawable will resume from the most recent 111 * frame. If the drawable has already reached the last frame, it will then loop back to the 112 * first frame, unles 250 addFrame(@onNull Drawable frame, int duration) argument 270 setFrame(int frame, boolean unschedule, boolean animate) argument [all...] |
/frameworks/base/libs/hwui/ |
H A D | DamageAccumulator.cpp | 40 // When this frame is pop'd, this rect is mapped through the above transform 41 // and applied to the previous (aka parent) frame 103 LOG_ALWAYS_FATAL_IF(mHead->prev == mHead, "Cannot pop the root frame!"); 135 void DamageAccumulator::applyMatrix4Transform(DirtyStack* frame) { argument 136 mapRect(frame->matrix4, frame->pendingDirty, &mHead->pendingDirty); 157 static DirtyStack* findParentRenderNode(DirtyStack* frame) { argument 158 while (frame->prev != frame) { 159 frame 167 findProjectionReceiver(DirtyStack* frame) argument 180 applyTransforms(DirtyStack* frame, DirtyStack* end) argument 192 applyRenderNodeTransform(DirtyStack* frame) argument [all...] |
H A D | JankTracker.cpp | 72 // If a frame is <= to this it goes in bucket 0 74 // If a frame is > this, start counting in increments of 2ms 76 // If a frame is > this, start counting in increments of 4ms 80 // consider any unknown delay of frame start as part of the test infrastructure 81 // and filter it out of the frame profile data 84 // The interval of the slow frame histogram 86 // The start point of the slow frame bucket in ms 89 // This will be called every frame, performance sensitive 253 void JankTracker::addFrame(const FrameInfo& frame) { argument 256 int64_t totalDuration = frame [all...] |
/frameworks/base/libs/hwui/pipeline/skia/ |
H A D | SkiaOpenGLPipeline.cpp | 59 bool SkiaOpenGLPipeline::draw(const Frame& frame, const SkRect& screenDirty, argument 68 mEglManager.damageFrame(frame, dirty); 72 renderTargetDesc.fWidth = frame.width(); 73 renderTargetDesc.fHeight = frame.height(); 107 bool SkiaOpenGLPipeline::swapBuffers(const Frame& frame, bool drew, argument 112 // Even if we decided to cancel the frame, from the perspective of jank 113 // metrics the frame was swapped at this point 118 if (*requireSwap && (CC_UNLIKELY(!mEglManager.swapBuffers(frame, screenDirty)))) {
|
H A D | SkiaVulkanPipeline.cpp | 61 Frame frame(backBuffer->width(), backBuffer->height(), mVkManager.getAge(mVkSurface)); 62 return frame; 65 bool SkiaVulkanPipeline::draw(const Frame& frame, const SkRect& screenDirty, argument 99 bool SkiaVulkanPipeline::swapBuffers(const Frame& frame, bool drew, argument 104 // Even if we decided to cancel the frame, from the perspective of jank 105 // metrics the frame was swapped at this point
|
/frameworks/base/libs/hwui/renderthread/ |
H A D | EglManager.cpp | 306 Frame frame; local 307 frame.mSurface = surface; 308 eglQuerySurface(mEglDisplay, surface, EGL_WIDTH, &frame.mWidth); 309 eglQuerySurface(mEglDisplay, surface, EGL_HEIGHT, &frame.mHeight); 310 frame.mBufferAge = queryBufferAge(surface); 312 return frame; 315 void EglManager::damageFrame(const Frame& frame, const SkRect& dirty) { argument 319 frame.map(dirty, rects); 320 if (!eglSetDamageRegionKHR(mEglDisplay, frame.mSurface, rects, 1)) { 322 (void*)frame 332 swapBuffers(const Frame& frame, const SkRect& screenDirty) argument [all...] |
/frameworks/base/media/mca/filterfw/jni/ |
H A D | jni_gl_frame.cpp | 51 std::unique_ptr<GLFrame> frame(new GLFrame(gl_env_ptr)); 52 if (frame->Init(width, height)) { 53 return ToJBool(WrapOwnedObjectInJava(std::move(frame), env, thiz, true)); 66 std::unique_ptr<GLFrame> frame(new GLFrame(gl_env_ptr)); 67 if (frame->InitWithTexture(tex_id, width, height)) { 68 return ToJBool(WrapOwnedObjectInJava(std::move(frame), env, thiz, true)); 81 std::unique_ptr<GLFrame> frame(new GLFrame(gl_env_ptr)); 82 if (frame->InitWithFbo(fbo_id, width, height)) { 83 return ToJBool(WrapOwnedObjectInJava(std::move(frame), env, thiz, true)); 93 std::unique_ptr<GLFrame> frame(ne 109 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 122 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 136 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 151 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 165 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 187 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 212 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 228 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 246 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 251 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 256 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 261 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 269 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 274 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 278 Java_android_filterfw_core_GLFrame_nativeCopyFromNative(JNIEnv* env, jobject thiz, jobject frame) argument 289 Java_android_filterfw_core_GLFrame_nativeCopyFromGL(JNIEnv* env, jobject thiz, jobject frame) argument 301 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 306 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local 311 GLFrame* frame = ConvertFromJava<GLFrame>(env, thiz); local [all...] |
H A D | jni_native_frame.cpp | 38 std::unique_ptr<NativeFrame> frame(new NativeFrame(size)); 39 return ToJBool(WrapOwnedObjectInJava(std::move(frame), env, thiz, true)); 59 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 60 if (frame && data) { 63 const bool success = frame->WriteData(reinterpret_cast<const uint8_t*>(bytes + offset), 76 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 77 if (frame) { 78 const uint8_t* data = frame->Data(); 79 if (!data || size > frame->Size()) 91 NativeFrame* frame local 102 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 120 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 136 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 154 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 172 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 222 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 270 NativeFrame* frame = ConvertFromJava<NativeFrame>(env, thiz); local 274 Java_android_filterfw_core_NativeFrame_nativeCopyFromNative(JNIEnv* env, jobject thiz, jobject frame) argument 285 Java_android_filterfw_core_NativeFrame_nativeCopyFromGL(JNIEnv* env, jobject thiz, jobject frame) argument [all...] |
/frameworks/base/media/mca/filterfw/native/core/ |
H A D | gl_env.cpp | 400 void GLEnv::AttachVertexFrame(int key, VertexFrame* frame) { argument 404 attached_vframes_[key] = frame;
|
H A D | gl_frame.cpp | 202 bool GLFrame::CopyPixelsFrom(const GLFrame* frame) { argument 203 if (frame == this) { 205 } else if (frame && frame->width_ == width_ && frame->height_ == height_) { 207 sources.push_back(frame); 366 // render the current texture frame into it. As this is quite inefficient, 370 ALOGW("Warning: Reading pixel data from unfiltered GL frame. This is highly " 374 // Create source frame set (unfortunately this requires an ugly const-cast, 375 // as we need to wrap ourselves in a frame [all...] |