diff options
Diffstat (limited to 'media/HdrViewfinder/README.md')
-rw-r--r-- | media/HdrViewfinder/README.md | 65 |
1 files changed, 37 insertions, 28 deletions
diff --git a/media/HdrViewfinder/README.md b/media/HdrViewfinder/README.md index 46adfbbc..73433ca0 100644 --- a/media/HdrViewfinder/README.md +++ b/media/HdrViewfinder/README.md @@ -1,54 +1,63 @@ + Android HdrViewfinder Sample -============================== +=================================== -This demo shows how to use Camera2 API and RenderScript to implement an HDR viewfinder. +This demo implements a real-time high-dynamic-range camera viewfinder, by alternating +the sensor's exposure time between two exposure values on even and odd frames, and then +compositing together the latest two frames whenever a new frame is captured. Introduction ------------ -This demo implements a real-time high-dynamic-range camera viewfinder, by alternating the sensor's -exposure time between two exposure values on even and odd frames, and then compositing together the -latest two frames whenever a new frame is captured. +A small demo of advanced camera functionality with the Android camera2 API. + +This demo implements a real-time high-dynamic-range camera viewfinder, +by alternating the sensor's exposure time between two exposure values on even and odd +frames, and then compositing together the latest two frames whenever a new frame is +captured. -The demo has three modes: Regular auto-exposure viewfinder, split-screen manual exposure, and the -fused HDR viewfinder. The latter two use manual exposure controlled by the user, by swiping up/down -on the right and left halves of the viewfinder. The left half controls the exposure time of even -frames, and the right half controls the exposure time of odd frames. +The demo has three modes: Regular auto-exposure viewfinder, split-screen manual exposure, +and the fused HDR viewfinder. The latter two use manual exposure controlled by the user, +by swiping up/down on the right and left halves of the viewfinder. The left half controls +the exposure time of even frames, and the right half controls the exposure time of odd frames. -In split-screen mode, the even frames are shown on the left and the odd frames on the right, so the -user can see two different exposures of the scene simultaneously. In fused HDR mode, the even/odd -frames are merged together into a single image. By selecting different exposure values for the -even/odd frames, the fused image has a higher dynamic range than the regular viewfinder. +In split-screen mode, the even frames are shown on the left and the odd frames on the right, +so the user can see two different exposures of the scene simultaneously. In fused HDR mode, +the even/odd frames are merged together into a single image. By selecting different exposure +values for the even/odd frames, the fused image has a higher dynamic range than the regular +viewfinder. The HDR fusion and the split-screen viewfinder processing is done with RenderScript; as is the -necessary YUV->RGB conversion. The camera subsystem outputs YUV images naturally, while the GPU and -display subsystems generally only accept RGB data. Therefore, after the images are -fused/composited, a standard YUV->RGB color transform is applied before the the data is written to -the output Allocation. The HDR fusion algorithm is very simple, and tends to result in +necessary YUV->RGB conversion. The camera subsystem outputs YUV images naturally, while the GPU +and display subsystems generally only accept RGB data. Therefore, after the images are +fused/composited, a standard YUV->RGB color transform is applied before the the data is written +to the output Allocation. The HDR fusion algorithm is very simple, and tends to result in lower-contrast scenes, but has very few artifacts and can run very fast. -Data is passed between the subsystems (camera, RenderScript, and display) using the Android {@link -android.view.Surface} class, which allows for zero-copy transport of large buffers between processes -and subsystems. +Data is passed between the subsystems (camera, RenderScript, and display) using the +Android [android.view.Surface][1] class, which allows for zero-copy transport of large +buffers between processes and subsystems. + +[1]: http://developer.android.com/reference/android/view/Surface.html Pre-requisites -------------- -- Android SDK v21 -- Android Build Tools v21 +- Android SDK v23 +- Android Build Tools v22.0.1 - Android Support Repository +Screenshots +------------- + +<img src="screenshots/image1.png" height="400" alt="Screenshot"/> + Getting Started --------------- This sample uses the Gradle build system. To build this project, use the "gradlew build" command or use "Import Project" in Android Studio. -Screenshots ------------ - -![Split mode](screenshots/image1.png) - Support ------- @@ -73,7 +82,7 @@ file to you under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at - http://www.apache.org/licenses/LICENSE-2.0 +http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT |