blob: a23546105c255fb5031943ccc7f9b787836d0083 [file] [log] [blame]
/*
**
** Copyright 2008, The Android Open Source Project
** Copyright 2012, Samsung Electronics Co. LTD
**
** Licensed under the Apache License, Version 2.0 (the "License");
** you may not use this file except in compliance with the License.
** You may obtain a copy of the License at
**
** http://www.apache.org/licenses/LICENSE-2.0
**
** Unless required by applicable law or agreed to in writing, software
** distributed under the License is distributed on an "AS IS" BASIS,
** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
** See the License for the specific language governing permissions and
** limitations under the License.
*/
/*!
* \file MetadataConverter.cpp
* \brief source file for Metadata converter ( for camera hal2 implementation )
* \author Sungjoong Kang(sj3.kang@samsung.com)
* \date 2012/05/31
*
* <b>Revision History: </b>
* - 2012/05/31 : Sungjoong Kang(sj3.kang@samsung.com) \n
* Initial Release
*/
//#define LOG_NDEBUG 0
#define LOG_TAG "MetadataConverter"
#include <utils/Log.h>
#include "MetadataConverter.h"
#include "exynos_format.h"
namespace android {
MetadataConverter::MetadataConverter()
{
return;
}
MetadataConverter::~MetadataConverter()
{
ALOGV("DEBUG(%s)destroy!!:", __FUNCTION__);
return;
}
status_t MetadataConverter::CheckEntryTypeMismatch(camera_metadata_entry_t * entry,
uint8_t type)
{
if (!(entry->type==type))
{
ALOGV("DEBUG(%s):Metadata Missmatch tag(%s) type (%d) count(%d)",
__FUNCTION__, get_camera_metadata_tag_name(entry->tag), entry->type, entry->count);
return BAD_VALUE;
}
return NO_ERROR;
}
status_t MetadataConverter::CheckEntryTypeMismatch(camera_metadata_entry_t * entry,
uint8_t type, size_t count)
{
if (!((entry->type==type)&&(entry->count==count)))
{
ALOGV("DEBUG(%s):Metadata Missmatch tag(%s) type (%d) count(%d)",
__FUNCTION__, get_camera_metadata_tag_name(entry->tag), entry->type, entry->count);
return BAD_VALUE;
}
return NO_ERROR;
}
status_t MetadataConverter::ToInternalShot(camera_metadata_t * request, struct camera2_shot_ext * dst_ext)
{
uint32_t num_entry = 0;
uint32_t index = 0;
uint32_t i = 0;
uint32_t cnt = 0;
camera_metadata_entry_t curr_entry;
struct camera2_shot * dst = NULL;
if (request == NULL || dst_ext == NULL)
return BAD_VALUE;
memset((void*)dst_ext, 0, sizeof(struct camera2_shot_ext));
dst = &dst_ext->shot;
dst->magicNumber = 0x23456789;
dst->ctl.aa.aeTargetFpsRange[0] = 15;
dst->ctl.aa.aeTargetFpsRange[1] = 30;
dst->ctl.aa.aeExpCompensation = 5;
num_entry = (uint32_t)get_camera_metadata_entry_count(request);
for (index = 0 ; index < num_entry ; index++) {
if (get_camera_metadata_entry(request, index, &curr_entry)==0) {
switch (curr_entry.tag) {
case ANDROID_LENS_FOCUS_DISTANCE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_FLOAT, 1))
break;
dst->ctl.lens.focusDistance = curr_entry.data.f[0];
break;
case ANDROID_LENS_APERTURE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_FLOAT, 1))
break;
dst->ctl.lens.aperture = curr_entry.data.f[0];
break;
case ANDROID_LENS_FOCAL_LENGTH:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_FLOAT, 1))
break;
dst->ctl.lens.focalLength = curr_entry.data.f[0];
break;
case ANDROID_LENS_FILTER_DENSITY:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_FLOAT, 1))
break;
dst->ctl.lens.filterDensity = curr_entry.data.f[0];
break;
case ANDROID_LENS_OPTICAL_STABILIZATION_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.lens.opticalStabilizationMode =
(enum optical_stabilization_mode)curr_entry.data.u8[0];
break;
case ANDROID_SENSOR_TIMESTAMP:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT64, 1))
break;
dst->dm.sensor.timeStamp = curr_entry.data.i64[0];
break;
case ANDROID_SENSOR_SENSITIVITY:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 1))
break;
dst->dm.aa.isoValue = curr_entry.data.i32[0];
break;
case ANDROID_SENSOR_EXPOSURE_TIME:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT64, 1))
break;
dst->dm.sensor.exposureTime = curr_entry.data.i64[0];
break;
case ANDROID_FLASH_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.flash.flashMode = (enum flash_mode)(curr_entry.data.u8[0] + 1);
break;
case ANDROID_FLASH_FIRING_POWER:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.flash.firingPower = curr_entry.data.u8[0];
break;
case ANDROID_FLASH_FIRING_TIME:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT64, 1))
break;
dst->ctl.flash.firingTime = curr_entry.data.i64[0];
break;
case ANDROID_SCALER_CROP_REGION:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 4))
break;
for (i=0 ; i<3; i++)
dst->ctl.scaler.cropRegion[i] = ALIGN(curr_entry.data.i32[i], 2);
break;
case ANDROID_JPEG_QUALITY:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 1))
break;
dst->ctl.jpeg.quality= curr_entry.data.i32[0];
break;
case ANDROID_JPEG_THUMBNAIL_SIZE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 2))
break;
for (i=0 ; i<curr_entry.count ; i++)
dst->ctl.jpeg.thumbnailSize[i] = curr_entry.data.i32[i];
break;
case ANDROID_JPEG_THUMBNAIL_QUALITY:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 1))
break;
dst->ctl.jpeg.thumbnailQuality= curr_entry.data.i32[0];
break;
case ANDROID_JPEG_GPS_COORDINATES:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_DOUBLE, 3))
break;
for (i=0 ; i<curr_entry.count ; i++)
dst->ctl.jpeg.gpsCoordinates[i] = curr_entry.data.d[i];
break;
case ANDROID_JPEG_GPS_PROCESSING_METHOD:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE))
break;
if (curr_entry.count > 32)
cnt = 32;
else
cnt = curr_entry.count;
for (i = 0 ; i < cnt ; i++)
dst_ext->gpsProcessingMethod[i] = curr_entry.data.u8[i];
break;
case ANDROID_JPEG_GPS_TIMESTAMP:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT64, 1))
break;
dst->ctl.jpeg.gpsTimestamp = curr_entry.data.i64[0];
break;
case ANDROID_JPEG_ORIENTATION:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 1))
break;
dst->ctl.jpeg.orientation = curr_entry.data.i32[0];
break;
case ANDROID_STATISTICS_FACE_DETECT_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.stats.faceDetectMode = (enum facedetect_mode)(curr_entry.data.u8[0] + 1);
break;
case ANDROID_CONTROL_CAPTURE_INTENT:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.aa.captureIntent = (enum aa_capture_intent)curr_entry.data.u8[0];
break;
case ANDROID_CONTROL_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.aa.mode = (enum aa_mode)(curr_entry.data.u8[0] + 1);
break;
case ANDROID_CONTROL_VIDEO_STABILIZATION_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.aa.videoStabilizationMode = curr_entry.data.u8[0];
break;
case ANDROID_CONTROL_AE_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.aa.aeMode = (enum aa_aemode)(curr_entry.data.u8[0] + 2);
// skip locked mode
if (dst->ctl.aa.aeMode == AA_AEMODE_LOCKED)
dst->ctl.aa.aeMode = AA_AEMODE_OFF;
ALOGV("DEBUG(%s): ANDROID_CONTROL_AE_MODE (%d)", __FUNCTION__, dst->ctl.aa.aeMode);
break;
case ANDROID_CONTROL_AE_LOCK:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst_ext->ae_lock = (enum ae_lockmode)(curr_entry.data.u8[0]);
break;
case ANDROID_CONTROL_AE_EXPOSURE_COMPENSATION:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 1))
break;
dst->ctl.aa.aeExpCompensation = curr_entry.data.i32[0] + 5;
break;
case ANDROID_CONTROL_AWB_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.aa.awbMode = (enum aa_awbmode)(curr_entry.data.u8[0] + 2);
// skip locked mode
if (dst->ctl.aa.awbMode == AA_AWBMODE_LOCKED)
dst->ctl.aa.awbMode = AA_AWBMODE_OFF;
dst_ext->awb_mode_dm = (enum aa_awbmode)(curr_entry.data.u8[0] + 2);
break;
case ANDROID_CONTROL_AWB_LOCK:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst_ext->awb_lock = (enum awb_lockmode)(curr_entry.data.u8[0]);
break;
case ANDROID_CONTROL_AF_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.aa.afMode = (enum aa_afmode)(curr_entry.data.u8[0] + 1);
if (dst->ctl.aa.afMode == AA_AFMODE_OFF)
dst->ctl.aa.afMode = AA_AFMODE_MANUAL;
break;
case ANDROID_CONTROL_AF_REGIONS:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 5))
break;
for (i=0 ; i<curr_entry.count ; i++)
dst->ctl.aa.afRegions[i] = curr_entry.data.i32[i];
break;
case ANDROID_CONTROL_AE_REGIONS:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 5))
break;
for (i=0 ; i<curr_entry.count ; i++)
dst->ctl.aa.aeRegions[i] = curr_entry.data.i32[i];
break;
case ANDROID_REQUEST_ID:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 1))
break;
dst->ctl.request.id = curr_entry.data.i32[0];
ALOGV("DEBUG(%s): ANDROID_REQUEST_ID (%d)", __FUNCTION__, dst->ctl.request.id);
break;
case ANDROID_REQUEST_METADATA_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.request.metadataMode = (enum metadata_mode)curr_entry.data.u8[0];
ALOGV("DEBUG(%s): ANDROID_REQUEST_METADATA_MODE (%d)", __FUNCTION__, (int)( dst->ctl.request.metadataMode));
break;
case ANDROID_REQUEST_OUTPUT_STREAMS:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32))
break;
dst->ctl.request.outputStreams[0] = 0;
for (i=0 ; i<curr_entry.count ; i++) {
ALOGV("DEBUG(%s): OUTPUT_STREAM[%d] = %d ", __FUNCTION__, i, curr_entry.data.i32[i]);
dst->ctl.request.outputStreams[0] |= (1 << curr_entry.data.i32[i]);
}
break;
case ANDROID_REQUEST_INPUT_STREAMS:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32))
break;
for (i=0 ; i<curr_entry.count ; i++) {
dst_ext->reprocessInput = curr_entry.data.i32[0];
ALOGV("DEBUG(%s): ANDROID_REQUEST_INPUT_STREAMS[%d] = %d ", __FUNCTION__, i, dst_ext->reprocessInput);
}
break;
case ANDROID_REQUEST_TYPE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst_ext->isReprocessing = curr_entry.data.u8[0];
ALOGV("DEBUG(%s): ANDROID_REQUEST_TYPE (%d)", __FUNCTION__, dst_ext->isReprocessing);
break;
case ANDROID_REQUEST_FRAME_COUNT:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 1))
break;
dst->ctl.request.frameCount = curr_entry.data.i32[0];
ALOGV("DEBUG(%s): ANDROID_REQUEST_FRAME_COUNT (%d)", __FUNCTION__, dst->ctl.request.frameCount);
break;
case ANDROID_CONTROL_SCENE_MODE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_BYTE, 1))
break;
dst->ctl.aa.sceneMode = (enum aa_scene_mode)(curr_entry.data.u8[0] + 1);
break;
case ANDROID_CONTROL_AE_TARGET_FPS_RANGE:
if (NO_ERROR != CheckEntryTypeMismatch(&curr_entry, TYPE_INT32, 2))
break;
for (i=0 ; i<curr_entry.count ; i++)
dst->ctl.aa.aeTargetFpsRange[i] = curr_entry.data.i32[i];
break;
default:
ALOGV("DEBUG(%s):Bad Metadata tag (%d)", __FUNCTION__, curr_entry.tag);
break;
}
}
}
if (dst->ctl.aa.mode != AA_CONTROL_USE_SCENE_MODE)
dst->ctl.aa.sceneMode = AA_SCENE_MODE_UNSUPPORTED;
ApplySceneModeParameters(request, dst_ext);
return NO_ERROR;
}
status_t MetadataConverter::ApplySceneModeParameters(camera_metadata_t * request, struct camera2_shot_ext * dst_ext)
{
camera_metadata_entry_t curr_entry;
struct camera2_shot * dst = NULL;
ALOGV("DEBUG(%s):", __FUNCTION__);
dst = &(dst_ext->shot);
switch (dst->ctl.aa.sceneMode) {
case AA_SCENE_MODE_ACTION:
dst->ctl.aa.mode = AA_CONTROL_USE_SCENE_MODE;
if (dst->ctl.aa.aeMode != AA_AEMODE_LOCKED)
dst->ctl.aa.aeMode = AA_AEMODE_ON;
dst->ctl.aa.awbMode = AA_AWBMODE_WB_AUTO;
dst_ext->awb_mode_dm = AA_AWBMODE_WB_AUTO;
dst->ctl.aa.isoMode = AA_ISOMODE_AUTO;
dst->ctl.aa.isoValue = 0;
dst->ctl.aa.aeTargetFpsRange[0] = 30;
dst->ctl.aa.aeTargetFpsRange[1] = 30;
dst->ctl.noise.mode = PROCESSING_MODE_FAST;
dst->ctl.noise.strength = 0;
dst->ctl.edge.mode = PROCESSING_MODE_FAST;
dst->ctl.edge.strength = 0;
dst->ctl.color.saturation = 3; // means '0'
break;
case AA_SCENE_MODE_PARTY:
dst->ctl.aa.mode = AA_CONTROL_USE_SCENE_MODE;
if (dst->ctl.aa.aeMode != AA_AEMODE_LOCKED)
dst->ctl.aa.aeMode = AA_AEMODE_ON;
dst->ctl.aa.awbMode = AA_AWBMODE_WB_AUTO;
dst_ext->awb_mode_dm = AA_AWBMODE_WB_AUTO;
dst->ctl.aa.isoMode = AA_ISOMODE_MANUAL;
dst->ctl.aa.isoValue = 200;
dst->ctl.aa.aeTargetFpsRange[0] = 15;
dst->ctl.aa.aeTargetFpsRange[1] = 30;
dst->ctl.noise.mode = PROCESSING_MODE_FAST;
dst->ctl.noise.strength = 0;
dst->ctl.edge.mode = PROCESSING_MODE_FAST;
dst->ctl.edge.strength = 0;
dst->ctl.color.saturation = 4; // means '+1'
break;
case AA_SCENE_MODE_SUNSET:
dst->ctl.aa.mode = AA_CONTROL_USE_SCENE_MODE;
if (dst->ctl.aa.aeMode != AA_AEMODE_LOCKED)
dst->ctl.aa.aeMode = AA_AEMODE_ON;
dst->ctl.aa.awbMode = AA_AWBMODE_WB_DAYLIGHT;
dst_ext->awb_mode_dm = AA_AWBMODE_WB_DAYLIGHT;
dst->ctl.aa.isoMode = AA_ISOMODE_AUTO;
dst->ctl.aa.isoValue = 0;
dst->ctl.aa.aeTargetFpsRange[0] = 15;
dst->ctl.aa.aeTargetFpsRange[1] = 30;
dst->ctl.noise.mode = PROCESSING_MODE_FAST;
dst->ctl.noise.strength = 0;
dst->ctl.edge.mode = PROCESSING_MODE_FAST;
dst->ctl.edge.strength = 0;
dst->ctl.color.saturation = 3; // means '0'
break;
case AA_SCENE_MODE_NIGHT:
dst->ctl.aa.mode = AA_CONTROL_USE_SCENE_MODE;
dst->ctl.aa.aeMode = AA_AEMODE_ON; // AE_LOCK is prohibited
dst->ctl.aa.awbMode = AA_AWBMODE_WB_AUTO;
dst_ext->awb_mode_dm = AA_AWBMODE_WB_AUTO;
dst->ctl.aa.isoMode = AA_ISOMODE_AUTO;
dst->ctl.aa.isoValue = 0;
dst->ctl.aa.aeTargetFpsRange[0] = 8;
dst->ctl.aa.aeTargetFpsRange[1] = 30;
dst->ctl.noise.mode = PROCESSING_MODE_FAST;
dst->ctl.noise.strength = 0;
dst->ctl.edge.mode = PROCESSING_MODE_FAST;
dst->ctl.edge.strength = 0;
dst->ctl.color.saturation = 3; // means '0'
break;
case AA_SCENE_MODE_FACE_PRIORITY:
dst->ctl.aa.mode = AA_CONTROL_AUTO;
if ((dst->ctl.aa.aeMode != AA_AEMODE_LOCKED) && (dst->ctl.aa.aeMode < AA_AEMODE_ON_AUTO_FLASH))
dst->ctl.aa.aeMode = AA_AEMODE_ON;
dst->ctl.aa.sceneMode = AA_SCENE_MODE_FACE_PRIORITY;
dst->ctl.aa.isoMode = AA_ISOMODE_AUTO;
dst->ctl.aa.isoValue = 0;
dst->ctl.noise.mode = PROCESSING_MODE_OFF;
dst->ctl.noise.strength = 0;
dst->ctl.edge.mode = PROCESSING_MODE_OFF;
dst->ctl.edge.strength = 0;
dst->ctl.color.saturation = 3; // means '0'
break;
default:
dst->ctl.aa.mode = AA_CONTROL_AUTO;
if ((dst->ctl.aa.aeMode != AA_AEMODE_LOCKED) && (dst->ctl.aa.aeMode < AA_AEMODE_ON_AUTO_FLASH))
dst->ctl.aa.aeMode = AA_AEMODE_ON;
dst->ctl.aa.sceneMode = AA_SCENE_MODE_UNSUPPORTED;
dst->ctl.aa.isoMode = AA_ISOMODE_AUTO;
dst->ctl.aa.isoValue = 0;
dst->ctl.noise.mode = PROCESSING_MODE_OFF;
dst->ctl.noise.strength = 0;
dst->ctl.edge.mode = PROCESSING_MODE_OFF;
dst->ctl.edge.strength = 0;
dst->ctl.color.saturation = 3; // means '0'
break;
}
return NO_ERROR;
}
status_t MetadataConverter::ToDynamicMetadata(struct camera2_shot_ext * metadata_ext, camera_metadata_t * dst)
{
status_t res;
struct camera2_shot * metadata = &metadata_ext->shot;
uint8_t byteData;
uint32_t intData;
if (0 != add_camera_metadata_entry(dst, ANDROID_REQUEST_ID,
&(metadata->ctl.request.id), 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_REQUEST_METADATA_MODE,
&(metadata->ctl.request.metadataMode), 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_REQUEST_FRAME_COUNT,
&(metadata->ctl.request.frameCount), 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_SENSOR_TIMESTAMP,
&metadata->dm.sensor.timeStamp, 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_SENSOR_EXPOSURE_TIME,
&metadata->dm.sensor.exposureTime, 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_LENS_APERTURE,
&metadata->dm.lens.aperture, 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_AE_TARGET_FPS_RANGE,
&(metadata->ctl.aa.aeTargetFpsRange), 2))
return NO_MEMORY;
ALOGV("(%s): ID(%d) METAMODE(%d) FrameCnt(%d) Timestamp(%lld) exposure(%lld) aper(%f)", __FUNCTION__,
metadata->ctl.request.id, metadata->ctl.request.metadataMode, metadata->ctl.request.frameCount,
metadata->dm.sensor.timeStamp, metadata->dm.sensor.exposureTime, metadata->dm.lens.aperture);
byteData = metadata_ext->awb_mode_dm- 2;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_AWB_MODE,
&byteData, 1))
return NO_MEMORY;
byteData = metadata->dm.aa.aeMode - 1;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_AE_MODE,
&byteData, 1))
return NO_MEMORY;
byteData = metadata->ctl.aa.afMode - 1;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_AF_MODE,
&byteData, 1))
return NO_MEMORY;
byteData = metadata->ctl.aa.sceneMode - 1;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_SCENE_MODE,
&byteData, 1))
return NO_MEMORY;
intData = metadata->ctl.aa.aeExpCompensation - 5;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_AE_EXPOSURE_COMPENSATION,
&intData, 1))
return NO_MEMORY;
byteData = metadata->dm.stats.faceDetectMode - 1;
if (0 != add_camera_metadata_entry(dst, ANDROID_STATISTICS_FACE_DETECT_MODE,
&byteData, 1))
return NO_MEMORY;
int maxFacecount = CAMERA2_MAX_FACES;
if (0 != add_camera_metadata_entry(dst, ANDROID_STATISTICS_INFO_MAX_FACE_COUNT,
&maxFacecount, 1))
return NO_MEMORY;
int tempFaceCount = 0;
for (int i = 0; i < CAMERA2_MAX_FACES; i++) {
if (metadata->dm.stats.faceIds[i] > 0) {
mataFaceIds[tempFaceCount] = metadata->dm.stats.faceIds[i];
// clipping fd score because the max face score of android is 100
if (metadata->dm.stats.faceScores[i] > 100)
metaFaceScores[tempFaceCount] = 100;
else
metaFaceScores[tempFaceCount] = metadata->dm.stats.faceScores[i];
memcpy(&mataFaceLandmarks[tempFaceCount][0], &metadata->dm.stats.faceLandmarks[i][0], 6*sizeof(uint32_t));
memcpy(&metaFaceRectangles[tempFaceCount][0], &metadata->dm.stats.faceRectangles[i][0], 4*sizeof(uint32_t));
tempFaceCount++;
}
}
if (tempFaceCount > 0) {
if (0 != add_camera_metadata_entry(dst, ANDROID_STATISTICS_FACE_RECTANGLES,
&metaFaceRectangles, 4 * tempFaceCount))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_STATISTICS_FACE_LANDMARKS,
&mataFaceLandmarks, 6 * tempFaceCount))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_STATISTICS_FACE_IDS,
&mataFaceIds, tempFaceCount))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_STATISTICS_FACE_SCORES,
&metaFaceScores, tempFaceCount))
return NO_MEMORY;
}
if (0 != add_camera_metadata_entry(dst, ANDROID_SENSOR_SENSITIVITY,
&metadata->dm.aa.isoValue, 1))
return NO_MEMORY;
// Need a four-entry crop region
uint32_t cropRegion[4] = {
metadata->ctl.scaler.cropRegion[0],
metadata->ctl.scaler.cropRegion[1],
metadata->ctl.scaler.cropRegion[2],
0
};
if (0 != add_camera_metadata_entry(dst, ANDROID_SCALER_CROP_REGION,
cropRegion, 4))
return NO_MEMORY;
byteData = metadata->dm.aa.aeState - 1;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_AE_STATE,
&byteData, 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_CONTROL_AWB_STATE,
&(metadata->dm.aa.awbState), 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_JPEG_ORIENTATION,
&metadata->ctl.jpeg.orientation, 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_JPEG_QUALITY,
&metadata->ctl.jpeg.quality, 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_JPEG_THUMBNAIL_SIZE,
&metadata->ctl.jpeg.thumbnailSize, 2))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_JPEG_THUMBNAIL_QUALITY,
&metadata->ctl.jpeg.thumbnailQuality, 1))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_JPEG_GPS_COORDINATES,
&(metadata->ctl.jpeg.gpsCoordinates), 3))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_JPEG_GPS_PROCESSING_METHOD,
&(metadata_ext->gpsProcessingMethod), 32))
return NO_MEMORY;
if (0 != add_camera_metadata_entry(dst, ANDROID_JPEG_GPS_TIMESTAMP,
&(metadata->ctl.jpeg.gpsTimestamp), 1))
return NO_MEMORY;
ALOGV("(%s): AWB(%d) AE(%d) SCENE(%d) AEComp(%d) AF(%d)", __FUNCTION__,
metadata_ext->awb_mode_dm- 2, metadata->dm.aa.aeMode - 1, metadata->ctl.aa.sceneMode - 1,
metadata->ctl.aa.aeExpCompensation, metadata->ctl.aa.afMode - 1);
if (metadata->ctl.request.metadataMode == METADATA_MODE_NONE) {
ALOGV("DEBUG(%s): METADATA_MODE_NONE", __FUNCTION__);
return NO_ERROR;
}
return NO_ERROR;
}
}; // namespace android