| /* |
| Copyright (c) 2017-2020, The Linux Foundation. All rights reserved. |
| |
| Redistribution and use in source and binary forms, with or without |
| modification, are permitted provided that the following conditions are |
| met: |
| * Redistributions of source code must retain the above copyright |
| notice, this list of conditions and the following disclaimer. |
| * Redistributions in binary form must reproduce the above |
| copyright notice, this list of conditions and the following |
| disclaimer in the documentation and/or other materials provided |
| with the distribution. |
| * Neither the name of The Linux Foundation nor the names of its |
| contributors may be used to endorse or promote products derived |
| from this software without specific prior written permission. |
| |
| THIS SOFTWARE IS PROVIDED "AS IS" AND ANY EXPRESS OR IMPLIED |
| WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF |
| MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT |
| ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS |
| BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR |
| BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, |
| WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE |
| OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN |
| IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include <fcntl.h> |
| #include <utils/debug.h> |
| #include <utils/sys.h> |
| #include <vector> |
| #include <string> |
| #include <cstring> |
| #include <algorithm> |
| |
| #include "hw_peripheral_drm.h" |
| |
| #define __CLASS__ "HWPeripheralDRM" |
| |
| using sde_drm::DRMDisplayType; |
| using sde_drm::DRMOps; |
| using sde_drm::DRMPowerMode; |
| using sde_drm::DppsFeaturePayload; |
| using sde_drm::DRMDppsFeatureInfo; |
| using sde_drm::DRMSecureMode; |
| using sde_drm::DRMCWbCaptureMode; |
| |
| namespace sdm { |
| |
| HWPeripheralDRM::HWPeripheralDRM(int32_t display_id, BufferAllocator *buffer_allocator, |
| HWInfoInterface *hw_info_intf) |
| : HWDeviceDRM(buffer_allocator, hw_info_intf) { |
| disp_type_ = DRMDisplayType::PERIPHERAL; |
| device_name_ = "Peripheral"; |
| display_id_ = display_id; |
| } |
| |
| DisplayError HWPeripheralDRM::Init() { |
| DisplayError ret = HWDeviceDRM::Init(); |
| if (ret != kErrorNone) { |
| DLOGE("Init failed for %s", device_name_); |
| return ret; |
| } |
| |
| scalar_data_.resize(hw_resource_.hw_dest_scalar_info.count); |
| dest_scalar_cache_.resize(hw_resource_.hw_dest_scalar_info.count); |
| PopulateBitClkRates(); |
| |
| topology_control_ = UINT32(sde_drm::DRMTopologyControl::DSPP); |
| if (hw_panel_info_.is_primary_panel) { |
| topology_control_ |= UINT32(sde_drm::DRMTopologyControl::DEST_SCALER); |
| } |
| |
| return kErrorNone; |
| } |
| |
| void HWPeripheralDRM::PopulateBitClkRates() { |
| if (!hw_panel_info_.dyn_bitclk_support) { |
| return; |
| } |
| |
| // Group all bit_clk_rates corresponding to DRM_PREFERRED mode. |
| uint32_t width = connector_info_.modes[current_mode_index_].mode.hdisplay; |
| uint32_t height = connector_info_.modes[current_mode_index_].mode.vdisplay; |
| |
| for (auto &mode_info : connector_info_.modes) { |
| auto &mode = mode_info.mode; |
| if (mode.hdisplay == width && mode.vdisplay == height) { |
| if (std::find(bitclk_rates_.begin(), bitclk_rates_.end(), mode_info.bit_clk_rate) == |
| bitclk_rates_.end()) { |
| bitclk_rates_.push_back(mode_info.bit_clk_rate); |
| DLOGI("Possible bit_clk_rates %d", mode_info.bit_clk_rate); |
| } |
| } |
| } |
| |
| hw_panel_info_.bitclk_rates = bitclk_rates_; |
| DLOGI("bit_clk_rates Size %d", bitclk_rates_.size()); |
| } |
| |
| DisplayError HWPeripheralDRM::SetDynamicDSIClock(uint64_t bit_clk_rate) { |
| if (last_power_mode_ == DRMPowerMode::DOZE_SUSPEND || last_power_mode_ == DRMPowerMode::OFF) { |
| return kErrorNotSupported; |
| } |
| |
| bit_clk_rate_ = bit_clk_rate; |
| update_mode_ = true; |
| |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::GetDynamicDSIClock(uint64_t *bit_clk_rate) { |
| // Update bit_rate corresponding to current refresh rate. |
| *bit_clk_rate = (uint32_t)connector_info_.modes[current_mode_index_].bit_clk_rate; |
| |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::SetDisplayMode(const HWDisplayMode hw_display_mode) { |
| DisplayError error = HWDeviceDRM::SetDisplayMode(hw_display_mode); |
| if (error != kErrorNone) { |
| return error; |
| } |
| |
| // update bit clk rates. |
| hw_panel_info_.bitclk_rates = bitclk_rates_; |
| |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::Validate(HWLayers *hw_layers) { |
| HWLayersInfo &hw_layer_info = hw_layers->info; |
| SetDestScalarData(hw_layer_info); |
| SetupConcurrentWriteback(hw_layer_info, true, nullptr); |
| SetIdlePCState(); |
| |
| return HWDeviceDRM::Validate(hw_layers); |
| } |
| |
| DisplayError HWPeripheralDRM::Commit(HWLayers *hw_layers) { |
| HWLayersInfo &hw_layer_info = hw_layers->info; |
| SetDestScalarData(hw_layer_info); |
| |
| int64_t cwb_fence_fd = -1; |
| bool has_fence = SetupConcurrentWriteback(hw_layer_info, false, &cwb_fence_fd); |
| |
| SetIdlePCState(); |
| |
| DisplayError error = HWDeviceDRM::Commit(hw_layers); |
| if (error != kErrorNone) { |
| return error; |
| } |
| |
| if (has_fence) { |
| hw_layer_info.stack->output_buffer->release_fence = Fence::Create(INT(cwb_fence_fd), |
| "release_cwb"); |
| } |
| |
| CacheDestScalarData(); |
| if (cwb_config_.enabled && (error == kErrorNone)) { |
| PostCommitConcurrentWriteback(hw_layer_info.stack->output_buffer); |
| } |
| |
| // Initialize to default after successful commit |
| synchronous_commit_ = false; |
| idle_pc_state_ = sde_drm::DRMIdlePCState::NONE; |
| |
| return error; |
| } |
| |
| void HWPeripheralDRM::ResetDestScalarCache() { |
| for (uint32_t j = 0; j < scalar_data_.size(); j++) { |
| dest_scalar_cache_[j] = {}; |
| } |
| } |
| |
| void HWPeripheralDRM::SetDestScalarData(const HWLayersInfo &hw_layer_info) { |
| if (!hw_scale_ || !hw_resource_.hw_dest_scalar_info.count) { |
| return; |
| } |
| |
| for (uint32_t i = 0; i < hw_resource_.hw_dest_scalar_info.count; i++) { |
| auto it = hw_layer_info.dest_scale_info_map.find(i); |
| |
| if (it == hw_layer_info.dest_scale_info_map.end()) { |
| continue; |
| } |
| |
| HWDestScaleInfo *dest_scale_info = it->second; |
| SDEScaler *scale = &scalar_data_[i]; |
| hw_scale_->SetScaler(dest_scale_info->scale_data, scale); |
| |
| sde_drm_dest_scaler_cfg *dest_scalar_data = &sde_dest_scalar_data_.ds_cfg[i]; |
| dest_scalar_data->flags = 0; |
| if (scale->scaler_v2.enable) { |
| dest_scalar_data->flags |= SDE_DRM_DESTSCALER_ENABLE; |
| } |
| if (scale->scaler_v2.de.enable) { |
| dest_scalar_data->flags |= SDE_DRM_DESTSCALER_ENHANCER_UPDATE; |
| } |
| if (dest_scale_info->scale_update) { |
| dest_scalar_data->flags |= SDE_DRM_DESTSCALER_SCALE_UPDATE; |
| } |
| if (hw_panel_info_.partial_update) { |
| dest_scalar_data->flags |= SDE_DRM_DESTSCALER_PU_ENABLE; |
| } |
| dest_scalar_data->index = i; |
| dest_scalar_data->lm_width = dest_scale_info->mixer_width; |
| dest_scalar_data->lm_height = dest_scale_info->mixer_height; |
| dest_scalar_data->scaler_cfg = reinterpret_cast<uint64_t>(&scale->scaler_v2); |
| |
| if (std::memcmp(&dest_scalar_cache_[i].scalar_data, scale, sizeof(SDEScaler)) || |
| dest_scalar_cache_[i].flags != dest_scalar_data->flags) { |
| needs_ds_update_ = true; |
| } |
| } |
| |
| if (needs_ds_update_) { |
| sde_dest_scalar_data_.num_dest_scaler = UINT32(hw_layer_info.dest_scale_info_map.size()); |
| drm_atomic_intf_->Perform(DRMOps::CRTC_SET_DEST_SCALER_CONFIG, token_.crtc_id, |
| reinterpret_cast<uint64_t>(&sde_dest_scalar_data_)); |
| } |
| } |
| |
| void HWPeripheralDRM::CacheDestScalarData() { |
| if (needs_ds_update_) { |
| // Cache the destination scalar data during commit |
| for (uint32_t i = 0; i < sde_dest_scalar_data_.num_dest_scaler; i++) { |
| dest_scalar_cache_[i].flags = sde_dest_scalar_data_.ds_cfg[i].flags; |
| dest_scalar_cache_[i].scalar_data = scalar_data_[i]; |
| } |
| needs_ds_update_ = false; |
| } |
| } |
| |
| DisplayError HWPeripheralDRM::Flush(HWLayers *hw_layers) { |
| DisplayError err = HWDeviceDRM::Flush(hw_layers); |
| if (err != kErrorNone) { |
| return err; |
| } |
| |
| ResetDestScalarCache(); |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::SetDppsFeature(void *payload, size_t size) { |
| uint32_t obj_id = 0, object_type = 0, feature_id = 0; |
| uint64_t value = 0; |
| |
| if (size != sizeof(DppsFeaturePayload)) { |
| DLOGE("invalid payload size %d, expected %d", size, sizeof(DppsFeaturePayload)); |
| return kErrorParameters; |
| } |
| |
| DppsFeaturePayload *feature_payload = reinterpret_cast<DppsFeaturePayload *>(payload); |
| object_type = feature_payload->object_type; |
| feature_id = feature_payload->feature_id; |
| value = feature_payload->value; |
| |
| if (feature_id == sde_drm::kFeatureAd4Roi) { |
| if (feature_payload->value) { |
| DisplayDppsAd4RoiCfg *params = reinterpret_cast<DisplayDppsAd4RoiCfg *> |
| (feature_payload->value); |
| if (!params) { |
| DLOGE("invalid playload value %d", feature_payload->value); |
| return kErrorNotSupported; |
| } |
| |
| ad4_roi_cfg_.h_x = params->h_start; |
| ad4_roi_cfg_.h_y = params->h_end; |
| ad4_roi_cfg_.v_x = params->v_start; |
| ad4_roi_cfg_.v_y = params->v_end; |
| ad4_roi_cfg_.factor_in = params->factor_in; |
| ad4_roi_cfg_.factor_out = params->factor_out; |
| |
| value = (uint64_t)&ad4_roi_cfg_; |
| } |
| } |
| |
| if (object_type == DRM_MODE_OBJECT_CRTC) { |
| obj_id = token_.crtc_id; |
| } else if (object_type == DRM_MODE_OBJECT_CONNECTOR) { |
| obj_id = token_.conn_id; |
| } else { |
| DLOGE("invalid object type 0x%x", object_type); |
| return kErrorUndefined; |
| } |
| |
| drm_atomic_intf_->Perform(DRMOps::DPPS_CACHE_FEATURE, obj_id, feature_id, value); |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::GetDppsFeatureInfo(void *payload, size_t size) { |
| if (size != sizeof(DRMDppsFeatureInfo)) { |
| DLOGE("invalid payload size %d, expected %d", size, sizeof(DRMDppsFeatureInfo)); |
| return kErrorParameters; |
| } |
| DRMDppsFeatureInfo *feature_info = reinterpret_cast<DRMDppsFeatureInfo *>(payload); |
| feature_info->obj_id = token_.crtc_id; |
| drm_mgr_intf_->GetDppsFeatureInfo(feature_info); |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::HandleSecureEvent(SecureEvent secure_event, HWLayers *hw_layers) { |
| switch (secure_event) { |
| case kSecureDisplayStart: { |
| secure_display_active_ = true; |
| if (hw_panel_info_.mode != kModeCommand) { |
| DisplayError err = Flush(hw_layers); |
| if (err != kErrorNone) { |
| return err; |
| } |
| } |
| } |
| break; |
| |
| case kSecureDisplayEnd: { |
| if (hw_panel_info_.mode != kModeCommand) { |
| DisplayError err = Flush(hw_layers); |
| if (err != kErrorNone) { |
| return err; |
| } |
| } |
| secure_display_active_ = false; |
| synchronous_commit_ = true; |
| } |
| break; |
| |
| default: |
| DLOGE("Invalid secure event %d", secure_event); |
| return kErrorNotSupported; |
| } |
| |
| return kErrorNone; |
| } |
| |
| bool HWPeripheralDRM::SetupConcurrentWriteback(const HWLayersInfo &hw_layer_info, bool validate, |
| int64_t *release_fence_fd) { |
| bool enable = hw_resource_.has_concurrent_writeback && hw_layer_info.stack->output_buffer; |
| if (!(enable || cwb_config_.enabled)) { |
| return false; |
| } |
| |
| bool setup_modes = enable && !cwb_config_.enabled && validate; |
| if (setup_modes && (SetupConcurrentWritebackModes() == kErrorNone)) { |
| cwb_config_.enabled = true; |
| } |
| |
| if (cwb_config_.enabled) { |
| if (enable) { |
| // Set DRM properties for Concurrent Writeback. |
| ConfigureConcurrentWriteback(hw_layer_info.stack); |
| |
| if (!validate && release_fence_fd) { |
| // Set GET_RETIRE_FENCE property to get Concurrent Writeback fence. |
| drm_atomic_intf_->Perform(DRMOps::CONNECTOR_GET_RETIRE_FENCE, |
| cwb_config_.token.conn_id, release_fence_fd); |
| return true; |
| } |
| } else { |
| // Tear down the Concurrent Writeback topology. |
| drm_atomic_intf_->Perform(DRMOps::CONNECTOR_SET_CRTC, cwb_config_.token.conn_id, 0); |
| } |
| } |
| |
| return false; |
| } |
| |
| DisplayError HWPeripheralDRM::TeardownConcurrentWriteback(void) { |
| if (cwb_config_.enabled) { |
| drm_mgr_intf_->UnregisterDisplay(&(cwb_config_.token)); |
| cwb_config_.enabled = false; |
| registry_.Clear(); |
| } |
| |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::SetupConcurrentWritebackModes() { |
| // To setup Concurrent Writeback topology, get the Connector ID of Virtual display |
| if (drm_mgr_intf_->RegisterDisplay(DRMDisplayType::VIRTUAL, &cwb_config_.token)) { |
| DLOGE("RegisterDisplay failed for Concurrent Writeback"); |
| return kErrorResources; |
| } |
| |
| // Set the modes based on Primary display. |
| std::vector<drmModeModeInfo> modes; |
| for (auto &item : connector_info_.modes) { |
| modes.push_back(item.mode); |
| } |
| |
| // Inform the mode list to driver. |
| struct sde_drm_wb_cfg cwb_cfg = {}; |
| cwb_cfg.connector_id = cwb_config_.token.conn_id; |
| cwb_cfg.flags = SDE_DRM_WB_CFG_FLAGS_CONNECTED; |
| cwb_cfg.count_modes = UINT32(modes.size()); |
| cwb_cfg.modes = (uint64_t)modes.data(); |
| |
| int ret = -EINVAL; |
| #ifdef DRM_IOCTL_SDE_WB_CONFIG |
| ret = drmIoctl(dev_fd_, DRM_IOCTL_SDE_WB_CONFIG, &cwb_cfg); |
| #endif |
| if (ret) { |
| drm_mgr_intf_->UnregisterDisplay(&(cwb_config_.token)); |
| DLOGE("Dump CWBConfig: mode_count %d flags %x", cwb_cfg.count_modes, cwb_cfg.flags); |
| DumpConnectorModeInfo(); |
| return kErrorHardware; |
| } |
| |
| return kErrorNone; |
| } |
| |
| void HWPeripheralDRM::ConfigureConcurrentWriteback(LayerStack *layer_stack) { |
| LayerBuffer *output_buffer = layer_stack->output_buffer; |
| registry_.MapOutputBufferToFbId(output_buffer); |
| |
| // Set the topology for Concurrent Writeback: [CRTC_PRIMARY_DISPLAY - CONNECTOR_VIRTUAL_DISPLAY]. |
| drm_atomic_intf_->Perform(DRMOps::CONNECTOR_SET_CRTC, cwb_config_.token.conn_id, token_.crtc_id); |
| |
| // Set CRTC Capture Mode |
| DRMCWbCaptureMode capture_mode = layer_stack->flags.post_processed_output ? |
| DRMCWbCaptureMode::DSPP_OUT : DRMCWbCaptureMode::MIXER_OUT; |
| drm_atomic_intf_->Perform(DRMOps::CRTC_SET_CAPTURE_MODE, token_.crtc_id, capture_mode); |
| |
| // Set Connector Output FB |
| uint32_t fb_id = registry_.GetOutputFbId(output_buffer->handle_id); |
| drm_atomic_intf_->Perform(DRMOps::CONNECTOR_SET_OUTPUT_FB_ID, cwb_config_.token.conn_id, fb_id); |
| |
| // Set Connector Secure Mode |
| bool secure = output_buffer->flags.secure; |
| DRMSecureMode mode = secure ? DRMSecureMode::SECURE : DRMSecureMode::NON_SECURE; |
| drm_atomic_intf_->Perform(DRMOps::CONNECTOR_SET_FB_SECURE_MODE, cwb_config_.token.conn_id, mode); |
| |
| // Set Connector Output Rect |
| sde_drm::DRMRect dst = {}; |
| dst.left = 0; |
| dst.top = 0; |
| dst.right = display_attributes_[current_mode_index_].x_pixels; |
| dst.bottom = display_attributes_[current_mode_index_].y_pixels; |
| drm_atomic_intf_->Perform(DRMOps::CONNECTOR_SET_OUTPUT_RECT, cwb_config_.token.conn_id, dst); |
| } |
| |
| void HWPeripheralDRM::PostCommitConcurrentWriteback(LayerBuffer *output_buffer) { |
| bool enabled = hw_resource_.has_concurrent_writeback && output_buffer; |
| |
| if (!enabled) { |
| TeardownConcurrentWriteback(); |
| } |
| } |
| |
| DisplayError HWPeripheralDRM::ControlIdlePowerCollapse(bool enable, bool synchronous) { |
| if (enable == idle_pc_enabled_) { |
| return kErrorNone; |
| } |
| idle_pc_state_ = enable ? sde_drm::DRMIdlePCState::ENABLE : sde_drm::DRMIdlePCState::DISABLE; |
| // As idle PC is disabled after subsequent commit, Make sure to have synchrounous commit and |
| // ensure TA accesses the display_cc registers after idle PC is disabled. |
| synchronous_commit_ = !enable ? synchronous : false; |
| idle_pc_enabled_ = enable; |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::PowerOn(const HWQosData &qos_data, |
| shared_ptr<Fence> *release_fence) { |
| DTRACE_SCOPED(); |
| if (!drm_atomic_intf_) { |
| DLOGE("DRM Atomic Interface is null!"); |
| return kErrorUndefined; |
| } |
| |
| if (first_cycle_) { |
| return kErrorDeferred; |
| } |
| |
| if (!idle_pc_enabled_) { |
| drm_atomic_intf_->Perform(sde_drm::DRMOps::CRTC_SET_IDLE_PC_STATE, token_.crtc_id, |
| sde_drm::DRMIdlePCState::ENABLE); |
| } |
| |
| if (sde_dest_scalar_data_.num_dest_scaler) { |
| drm_atomic_intf_->Perform(DRMOps::CRTC_SET_DEST_SCALER_CONFIG, token_.crtc_id, |
| reinterpret_cast<uint64_t>(&sde_dest_scalar_data_)); |
| needs_ds_update_ = true; |
| } |
| |
| DisplayError err = HWDeviceDRM::PowerOn(qos_data, release_fence); |
| if (err != kErrorNone) { |
| return err; |
| } |
| idle_pc_state_ = sde_drm::DRMIdlePCState::NONE; |
| idle_pc_enabled_ = true; |
| |
| CacheDestScalarData(); |
| |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::SetDisplayAttributes(uint32_t index) { |
| HWDeviceDRM::SetDisplayAttributes(index); |
| // update bit clk rates. |
| hw_panel_info_.bitclk_rates = bitclk_rates_; |
| |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::SetDisplayDppsAdROI(void *payload) { |
| DisplayError err = kErrorNone; |
| struct sde_drm::DppsFeaturePayload feature_payload = {}; |
| |
| if (!payload) { |
| DLOGE("Invalid payload parameter"); |
| return kErrorParameters; |
| } |
| |
| feature_payload.object_type = DRM_MODE_OBJECT_CRTC; |
| feature_payload.feature_id = sde_drm::kFeatureAd4Roi; |
| feature_payload.value = (uint64_t)(payload); |
| |
| err = SetDppsFeature(&feature_payload, sizeof(feature_payload)); |
| if (err != kErrorNone) { |
| DLOGE("Faid to SetDppsFeature feature_id = %d, err = %d", |
| sde_drm::kFeatureAd4Roi, err); |
| } |
| |
| return err; |
| } |
| |
| DisplayError HWPeripheralDRM::SetFrameTrigger(FrameTriggerMode mode) { |
| sde_drm::DRMFrameTriggerMode drm_mode = sde_drm::DRMFrameTriggerMode::FRAME_DONE_WAIT_DEFAULT; |
| switch (mode) { |
| case kFrameTriggerDefault: |
| drm_mode = sde_drm::DRMFrameTriggerMode::FRAME_DONE_WAIT_DEFAULT; |
| break; |
| case kFrameTriggerSerialize: |
| drm_mode = sde_drm::DRMFrameTriggerMode::FRAME_DONE_WAIT_SERIALIZE; |
| break; |
| case kFrameTriggerPostedStart: |
| drm_mode = sde_drm::DRMFrameTriggerMode::FRAME_DONE_WAIT_POSTED_START; |
| break; |
| default: |
| DLOGE("Invalid frame trigger mode %d", (int32_t)mode); |
| return kErrorParameters; |
| } |
| |
| int ret = drm_atomic_intf_->Perform(DRMOps::CONNECTOR_SET_FRAME_TRIGGER, |
| token_.conn_id, drm_mode); |
| if (ret) { |
| DLOGE("Failed to perform CONNECTOR_SET_FRAME_TRIGGER, drm_mode %d, ret %d", drm_mode, ret); |
| return kErrorUndefined; |
| } |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::SetPanelBrightness(int level) { |
| if (pending_doze_) { |
| DLOGI("Doze state pending!! Skip for now"); |
| return kErrorDeferred; |
| } |
| |
| char buffer[kMaxSysfsCommandLength] = {0}; |
| |
| if (brightness_base_path_.empty()) { |
| return kErrorHardware; |
| } |
| |
| std::string brightness_node(brightness_base_path_ + "brightness"); |
| int fd = Sys::open_(brightness_node.c_str(), O_RDWR); |
| if (fd < 0) { |
| DLOGE("Failed to open node = %s, error = %s ", brightness_node.c_str(), |
| strerror(errno)); |
| return kErrorFileDescriptor; |
| } |
| |
| int32_t bytes = snprintf(buffer, kMaxSysfsCommandLength, "%d\n", level); |
| ssize_t ret = Sys::pwrite_(fd, buffer, static_cast<size_t>(bytes), 0); |
| if (ret <= 0) { |
| DLOGE("Failed to write to node = %s, error = %s ", brightness_node.c_str(), |
| strerror(errno)); |
| Sys::close_(fd); |
| return kErrorHardware; |
| } |
| |
| Sys::close_(fd); |
| |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::GetPanelBrightness(int *level) { |
| char value[kMaxStringLength] = {0}; |
| |
| if (!level) { |
| DLOGE("Invalid input, null pointer."); |
| return kErrorParameters; |
| } |
| |
| if (brightness_base_path_.empty()) { |
| return kErrorHardware; |
| } |
| |
| std::string brightness_node(brightness_base_path_ + "brightness"); |
| int fd = Sys::open_(brightness_node.c_str(), O_RDWR); |
| if (fd < 0) { |
| DLOGE("Failed to open brightness node = %s, error = %s", brightness_node.c_str(), |
| strerror(errno)); |
| return kErrorFileDescriptor; |
| } |
| |
| if (Sys::pread_(fd, value, sizeof(value), 0) > 0) { |
| *level = atoi(value); |
| } else { |
| DLOGE("Failed to read panel brightness"); |
| Sys::close_(fd); |
| return kErrorHardware; |
| } |
| |
| Sys::close_(fd); |
| |
| return kErrorNone; |
| } |
| |
| void HWPeripheralDRM::GetHWPanelMaxBrightness() { |
| char value[kMaxStringLength] = {0}; |
| hw_panel_info_.panel_max_brightness = 255.0f; |
| |
| // Panel nodes, driver connector creation, and DSI probing all occur in sync, for each DSI. This |
| // means that the connector_type_id - 1 will reflect the same # as the panel # for panel node. |
| char s[kMaxStringLength] = {}; |
| snprintf(s, sizeof(s), "/sys/class/backlight/panel%d-backlight/", |
| static_cast<int>(connector_info_.type_id - 1)); |
| brightness_base_path_.assign(s); |
| |
| std::string brightness_node(brightness_base_path_ + "max_brightness"); |
| int fd = Sys::open_(brightness_node.c_str(), O_RDONLY); |
| if (fd < 0) { |
| DLOGE("Failed to open max brightness node = %s, error = %s", brightness_node.c_str(), |
| strerror(errno)); |
| return; |
| } |
| |
| if (Sys::pread_(fd, value, sizeof(value), 0) > 0) { |
| hw_panel_info_.panel_max_brightness = static_cast<float>(atof(value)); |
| DLOGI_IF(kTagDriverConfig, "Max brightness = %f", hw_panel_info_.panel_max_brightness); |
| } else { |
| DLOGE("Failed to read max brightness. error = %s", strerror(errno)); |
| } |
| |
| Sys::close_(fd); |
| return; |
| } |
| |
| DisplayError HWPeripheralDRM::SetBLScale(uint32_t level) { |
| int ret = drm_atomic_intf_->Perform(DRMOps::DPPS_CACHE_FEATURE, |
| token_.conn_id, sde_drm::kFeatureSvBlScale, level); |
| if (ret) { |
| DLOGE("Failed to set backlight scale level %d, ret %d", level, ret); |
| return kErrorUndefined; |
| } |
| return kErrorNone; |
| } |
| |
| DisplayError HWPeripheralDRM::GetPanelBrightnessBasePath(std::string *base_path) { |
| if (!base_path) { |
| DLOGE("Invalid base_path is null pointer"); |
| return kErrorParameters; |
| } |
| |
| if (brightness_base_path_.empty()) { |
| DLOGE("brightness_base_path_ is empty"); |
| return kErrorHardware; |
| } |
| |
| *base_path = brightness_base_path_; |
| return kErrorNone; |
| } |
| |
| } // namespace sdm |