blob: 9e43512b31b22b4e8bcfbf568c054eaccab0eb71 [file] [log] [blame]
Clay Murphyb6e5f5b2013-10-21 17:01:06 -07001page.title=Metadata and Controls
2@jd:body
3
4<!--
5 Copyright 2013 The Android Open Source Project
6
7 Licensed under the Apache License, Version 2.0 (the "License");
8 you may not use this file except in compliance with the License.
9 You may obtain a copy of the License at
10
11 http://www.apache.org/licenses/LICENSE-2.0
12
13 Unless required by applicable law or agreed to in writing, software
14 distributed under the License is distributed on an "AS IS" BASIS,
15 WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
16 See the License for the specific language governing permissions and
17 limitations under the License.
18-->
19<div id="qv-wrapper">
20 <div id="qv">
21 <h2>In this document</h2>
22 <ol id="auto-toc">
23 </ol>
24 </div>
25</div>
26
27<h2 id="metadata">Metadata support</h2>
28<p> To support the saving of raw image files by the Android framework, substantial
29 metadata is required about the sensor's characteristics. This includes
30 information such as color spaces and lens shading functions.</p>
31<p>Most of this information is a static property of the camera subsystem and can
32 therefore be queried before configuring any output pipelines or submitting any
33 requests. The new camera APIs greatly expand the information provided by the
34 getCameraInfo() method to provide this information to the application.</p>
35<p>In addition, manual control of the camera subsystem requires feedback from the
36 assorted devices about their current state, and the actual parameters used in
37 capturing a given frame. The actual values of the controls (exposure time, frame
38 duration, and sensitivity) as actually used by the hardware must be included in
39 the output metadata. This is essential so that applications know when either
40 clamping or rounding took place, and so that the application can compensate for
41 the real settings used for image capture.</p>
42<p>For example, if an application sets frame duration to 0 in a request, the HAL
43 must clamp the frame duration to the real minimum frame duration for that
44 request, and report that clamped minimum duration in the output result metadata.</p>
45<p>So if an application needs to implement a custom 3A routine (for example, to
46 properly meter for an HDR burst), it needs to know the settings used to capture
47 the latest set of results it has received in order to update the settings for
48 the next request. Therefore, the new camera API adds a substantial amount of
49 dynamic metadata to each captured frame. This includes the requested and actual
50 parameters used for the capture, as well as additional per-frame metadata such
51 as timestamps and statistics generator output.</p>
52<h2 id="per-setting">Per-setting control</h2>
53<p> For most settings, the expectation is that they can be changed every frame,
54 without introducing significant stutter or delay to the output frame stream.
55 Ideally, the output frame rate should solely be controlled by the capture
56 request's frame duration field, and be independent of any changes to processing
57 blocks' configuration. In reality, some specific controls are known to be slow
58 to change; these include the output resolution and output format of the camera
59 pipeline, as well as controls that affect physical devices, such as lens focus
60 distance. The exact requirements for each control set are detailed later.</p>
61<h2 id="raw-sensor">Raw sensor data support</h2>
62<p>In addition to the pixel formats supported by
63 the old API, the new API adds a requirement for support for raw sensor data
64 (Bayer RAW), both for advanced camera applications as well as to support raw
65 image files.</p>