| /* |
| * Copyright (c) 2013-2017 The Linux Foundation. All rights reserved. |
| * |
| * Previously licensed under the ISC license by Qualcomm Atheros, Inc. |
| * |
| * |
| * Permission to use, copy, modify, and/or distribute this software for |
| * any purpose with or without fee is hereby granted, provided that the |
| * above copyright notice and this permission notice appear in all |
| * copies. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL |
| * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED |
| * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE |
| * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL |
| * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR |
| * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER |
| * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR |
| * PERFORMANCE OF THIS SOFTWARE. |
| */ |
| |
| /* |
| * This file was originally distributed by Qualcomm Atheros, Inc. |
| * under proprietary terms before Copyright ownership was assigned |
| * to the Linux Foundation. |
| */ |
| |
| /** |
| * DOC: wma_main.c |
| * |
| * This file contains wma initialization and FW exchange |
| * related functions. |
| */ |
| |
| /* Header files */ |
| |
| #include "wma.h" |
| #include "wma_api.h" |
| #include "cds_api.h" |
| #include "wmi_unified_api.h" |
| #include "wlan_qct_sys.h" |
| #include "wni_api.h" |
| #include "ani_global.h" |
| #include "wmi_unified.h" |
| #include "service_ready_event_handler.h" |
| #include "wni_cfg.h" |
| #include "cfg_api.h" |
| #if defined(CONFIG_HL_SUPPORT) |
| #include "wlan_tgt_def_config_hl.h" |
| #else |
| #include "wlan_tgt_def_config.h" |
| #endif |
| #include "qdf_nbuf.h" |
| #include "qdf_types.h" |
| #include "qdf_mem.h" |
| #include "wma_types.h" |
| #include "lim_api.h" |
| #include "lim_session_utils.h" |
| |
| #include "cds_utils.h" |
| |
| #if !defined(REMOVE_PKT_LOG) |
| #include "pktlog_ac.h" |
| #endif /* REMOVE_PKT_LOG */ |
| |
| #include "dbglog_host.h" |
| #include "wmi_version_whitelist.h" |
| #include "csr_api.h" |
| #include "ol_fw.h" |
| |
| #include "wma_internal.h" |
| |
| #include "wma_ocb.h" |
| #include "wlan_policy_mgr_api.h" |
| #include "cdp_txrx_cfg.h" |
| #include "cdp_txrx_flow_ctrl_legacy.h" |
| #include "cdp_txrx_flow_ctrl_v2.h" |
| #include "cdp_txrx_ipa.h" |
| #include "cdp_txrx_misc.h" |
| #include "wma_fips_api.h" |
| #include "wma_nan_datapath.h" |
| #include "wlan_lmac_if_def.h" |
| #include "wlan_lmac_if_api.h" |
| #include "target_if.h" |
| #include "wlan_global_lmac_if_api.h" |
| #include "target_if_pmo.h" |
| #include "wma_he.h" |
| #include "wlan_pmo_obj_mgmt_api.h" |
| |
| #include "wlan_reg_tgt_api.h" |
| #include "wlan_reg_services_api.h" |
| #include <cdp_txrx_handle.h> |
| #include <wlan_pmo_ucfg_api.h> |
| #include "wifi_pos_api.h" |
| #include "hif_main.h" |
| |
| #define WMA_LOG_COMPLETION_TIMER 3000 /* 3 seconds */ |
| #define WMI_TLV_HEADROOM 128 |
| uint8_t *mac_trace_get_wma_msg_string(uint16_t wmaMsg); |
| static uint32_t g_fw_wlan_feat_caps; |
| /** |
| * wma_get_fw_wlan_feat_caps() - get fw feature capablity |
| * @feature: feature enum value |
| * |
| * Return: true/false |
| */ |
| bool wma_get_fw_wlan_feat_caps(enum cap_bitmap feature) |
| { |
| return (g_fw_wlan_feat_caps & (1 << feature)) ? true : false; |
| } |
| |
| /** |
| * wma_set_fw_wlan_feat_caps() - set fw feature capablity |
| * @feature: feature enum value |
| * |
| * Return: None |
| */ |
| void wma_set_fw_wlan_feat_caps(enum cap_bitmap feature) |
| { |
| g_fw_wlan_feat_caps |= (1 << feature); |
| } |
| |
| /** |
| * wma_service_ready_ext_evt_timeout() - Service ready extended event timeout |
| * @data: Timeout handler data |
| * |
| * This function is called when the FW fails to send WMI_SERVICE_READY_EXT_EVENT |
| * message |
| * |
| * Return: None |
| */ |
| static void wma_service_ready_ext_evt_timeout(void *data) |
| { |
| tp_wma_handle wma_handle; |
| |
| WMA_LOGA("%s: Timeout waiting for WMI_SERVICE_READY_EXT_EVENT", |
| __func__); |
| |
| wma_handle = (tp_wma_handle) data; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| goto end; |
| } |
| |
| end: |
| /* Assert here. Panic is being called in insmod thread */ |
| QDF_ASSERT(0); |
| } |
| |
| /** |
| * wma_get_ini_handle() - API to get WMA ini info handle |
| * @wma: WMA Handle |
| * |
| * Returns the pointer to WMA ini structure. |
| * Return: struct wma_ini_config |
| */ |
| struct wma_ini_config *wma_get_ini_handle(tp_wma_handle wma) |
| { |
| if (!wma) { |
| WMA_LOGE("%s: Invalid WMA context\n", __func__); |
| return NULL; |
| } |
| |
| return &wma->ini_config; |
| } |
| |
| #define MAX_SUPPORTED_PEERS_REV1_1 14 |
| #define MAX_SUPPORTED_PEERS_REV1_3 32 |
| #define MIN_NO_OF_PEERS 1 |
| |
| /** |
| * wma_get_number_of_peers_supported - API to query for number of peers |
| * supported |
| * @wma: WMA Handle |
| * |
| * Return: Max Number of Peers Supported |
| */ |
| static uint8_t wma_get_number_of_peers_supported(tp_wma_handle wma) |
| { |
| struct hif_target_info *tgt_info; |
| struct wma_ini_config *cfg = wma_get_ini_handle(wma); |
| uint8_t max_no_of_peers = cfg ? cfg->max_no_of_peers : MIN_NO_OF_PEERS; |
| struct hif_opaque_softc *scn = cds_get_context(QDF_MODULE_ID_HIF); |
| |
| if (!scn) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return 0; |
| } |
| |
| tgt_info = hif_get_target_info_handle(scn); |
| |
| switch (tgt_info->target_version) { |
| case AR6320_REV1_1_VERSION: |
| if (max_no_of_peers > MAX_SUPPORTED_PEERS_REV1_1) |
| max_no_of_peers = MAX_SUPPORTED_PEERS_REV1_1; |
| break; |
| default: |
| if (max_no_of_peers > MAX_SUPPORTED_PEERS_REV1_3) |
| max_no_of_peers = MAX_SUPPORTED_PEERS_REV1_3; |
| break; |
| } |
| |
| return max_no_of_peers; |
| } |
| |
| /** |
| * wma_get_number_of_tids_supported - API to query for number of tids supported |
| * @no_of_peers_supported: Number of peer supported |
| * |
| * Return: Max number of tids supported |
| */ |
| #if defined(CONFIG_HL_SUPPORT) |
| static uint32_t wma_get_number_of_tids_supported(uint8_t no_of_peers_supported) |
| { |
| return 4 * no_of_peers_supported; |
| } |
| #else |
| static uint32_t wma_get_number_of_tids_supported(uint8_t no_of_peers_supported) |
| { |
| return 2 * (no_of_peers_supported + CFG_TGT_NUM_VDEV + 2); |
| } |
| #endif |
| |
| /** |
| * wma_set_default_tgt_config() - set default tgt config |
| * @wma_handle: wma handle |
| * |
| * Return: none |
| */ |
| static void wma_set_default_tgt_config(tp_wma_handle wma_handle) |
| { |
| uint8_t no_of_peers_supported; |
| target_resource_config tgt_cfg; |
| |
| qdf_mem_zero(&tgt_cfg, sizeof(tgt_cfg)); |
| tgt_cfg.num_vdevs = CFG_TGT_NUM_VDEV; |
| tgt_cfg.num_peers = CFG_TGT_NUM_PEERS + CFG_TGT_NUM_VDEV + 2; |
| tgt_cfg.num_offload_peers = CFG_TGT_NUM_OFFLOAD_PEERS; |
| tgt_cfg.num_offload_reorder_buffs = CFG_TGT_NUM_OFFLOAD_REORDER_BUFFS; |
| tgt_cfg.num_peer_keys = CFG_TGT_NUM_PEER_KEYS; |
| tgt_cfg.num_tids = CFG_TGT_NUM_TIDS; |
| tgt_cfg.ast_skid_limit = CFG_TGT_AST_SKID_LIMIT; |
| tgt_cfg.tx_chain_mask = CFG_TGT_DEFAULT_TX_CHAIN_MASK; |
| tgt_cfg.rx_chain_mask = CFG_TGT_DEFAULT_RX_CHAIN_MASK; |
| tgt_cfg.rx_timeout_pri[0] = CFG_TGT_RX_TIMEOUT_LO_PRI; |
| tgt_cfg.rx_timeout_pri[1] = CFG_TGT_RX_TIMEOUT_LO_PRI; |
| tgt_cfg.rx_timeout_pri[2] = CFG_TGT_RX_TIMEOUT_LO_PRI; |
| tgt_cfg.rx_timeout_pri[3] = CFG_TGT_RX_TIMEOUT_HI_PRI; |
| tgt_cfg.rx_decap_mode = CFG_TGT_RX_DECAP_MODE; |
| tgt_cfg.scan_max_pending_req = CFG_TGT_DEFAULT_SCAN_MAX_REQS; |
| tgt_cfg.bmiss_offload_max_vdev = CFG_TGT_DEFAULT_BMISS_OFFLOAD_MAX_VDEV; |
| tgt_cfg.roam_offload_max_vdev = CFG_TGT_DEFAULT_ROAM_OFFLOAD_MAX_VDEV; |
| tgt_cfg.roam_offload_max_ap_profiles = |
| CFG_TGT_DEFAULT_ROAM_OFFLOAD_MAX_PROFILES; |
| tgt_cfg.num_mcast_groups = CFG_TGT_DEFAULT_NUM_MCAST_GROUPS; |
| tgt_cfg.num_mcast_table_elems = CFG_TGT_DEFAULT_NUM_MCAST_TABLE_ELEMS; |
| tgt_cfg.mcast2ucast_mode = CFG_TGT_DEFAULT_MCAST2UCAST_MODE; |
| tgt_cfg.tx_dbg_log_size = CFG_TGT_DEFAULT_TX_DBG_LOG_SIZE; |
| tgt_cfg.num_wds_entries = CFG_TGT_WDS_ENTRIES; |
| tgt_cfg.dma_burst_size = CFG_TGT_DEFAULT_DMA_BURST_SIZE; |
| tgt_cfg.mac_aggr_delim = CFG_TGT_DEFAULT_MAC_AGGR_DELIM; |
| tgt_cfg.rx_skip_defrag_timeout_dup_detection_check = |
| CFG_TGT_DEFAULT_RX_SKIP_DEFRAG_TIMEOUT_DUP_DETECTION_CHECK, |
| tgt_cfg.vow_config = CFG_TGT_DEFAULT_VOW_CONFIG; |
| tgt_cfg.gtk_offload_max_vdev = CFG_TGT_DEFAULT_GTK_OFFLOAD_MAX_VDEV; |
| tgt_cfg.num_msdu_desc = CFG_TGT_NUM_MSDU_DESC; |
| tgt_cfg.max_frag_entries = CFG_TGT_MAX_FRAG_TABLE_ENTRIES; |
| tgt_cfg.num_tdls_vdevs = CFG_TGT_NUM_TDLS_VDEVS; |
| tgt_cfg.num_tdls_conn_table_entries = |
| CFG_TGT_NUM_TDLS_CONN_TABLE_ENTRIES; |
| tgt_cfg.beacon_tx_offload_max_vdev = |
| CFG_TGT_DEFAULT_BEACON_TX_OFFLOAD_MAX_VDEV; |
| tgt_cfg.num_multicast_filter_entries = |
| CFG_TGT_MAX_MULTICAST_FILTER_ENTRIES; |
| tgt_cfg.num_wow_filters = 0; |
| tgt_cfg.num_keep_alive_pattern = 0; |
| tgt_cfg.keep_alive_pattern_size = 0; |
| tgt_cfg.max_tdls_concurrent_sleep_sta = |
| CFG_TGT_NUM_TDLS_CONC_SLEEP_STAS; |
| tgt_cfg.max_tdls_concurrent_buffer_sta = |
| CFG_TGT_NUM_TDLS_CONC_BUFFER_STAS; |
| tgt_cfg.wmi_send_separate = 0; |
| tgt_cfg.num_ocb_vdevs = CFG_TGT_NUM_OCB_VDEVS; |
| tgt_cfg.num_ocb_channels = CFG_TGT_NUM_OCB_CHANNELS; |
| tgt_cfg.num_ocb_schedules = CFG_TGT_NUM_OCB_SCHEDULES; |
| |
| no_of_peers_supported = wma_get_number_of_peers_supported(wma_handle); |
| tgt_cfg.num_peers = no_of_peers_supported + CFG_TGT_NUM_VDEV + 2; |
| tgt_cfg.num_tids = wma_get_number_of_tids_supported( |
| no_of_peers_supported); |
| tgt_cfg.scan_max_pending_req = wma_handle->max_scan; |
| |
| tgt_cfg.mgmt_comp_evt_bundle_support = true; |
| tgt_cfg.tx_msdu_new_partition_id_support = true; |
| |
| /* reduce the peer/vdev if CFG_TGT_NUM_MSDU_DESC exceeds 1000 */ |
| #ifdef PERE_IP_HDR_ALIGNMENT_WAR |
| if (scn->host_80211_enable) { |
| /* |
| * To make the IP header begins at dword aligned address, |
| * we make the decapsulation mode as Native Wifi. |
| */ |
| tgt_cfg.rx_decap_mode = CFG_TGT_RX_DECAP_MODE_NWIFI; |
| } |
| #endif /* PERE_IP_HDR_ALIGNMENT_WAR */ |
| if (QDF_GLOBAL_MONITOR_MODE == cds_get_conparam()) |
| tgt_cfg.rx_decap_mode = CFG_TGT_RX_DECAP_MODE_RAW; |
| |
| wma_handle->wlan_resource_config = tgt_cfg; |
| } |
| |
| /** |
| * wma_cli_get_command() - WMA "get" command processor |
| * @vdev_id: virtual device for the command |
| * @param_id: parameter id |
| * @vpdev: parameter category |
| * |
| * Return: parameter value on success, -EINVAL on failure |
| */ |
| int wma_cli_get_command(int vdev_id, int param_id, int vpdev) |
| { |
| int ret = 0; |
| tp_wma_handle wma; |
| struct wma_txrx_node *intr = NULL; |
| |
| wma = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| if (NULL == wma) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return -EINVAL; |
| } |
| |
| intr = wma->interfaces; |
| |
| if (VDEV_CMD == vpdev) { |
| switch (param_id) { |
| case WMI_VDEV_PARAM_NSS: |
| ret = intr[vdev_id].config.nss; |
| break; |
| #ifdef QCA_SUPPORT_GTX |
| case WMI_VDEV_PARAM_GTX_HT_MCS: |
| ret = intr[vdev_id].config.gtx_info.gtxRTMask[0]; |
| break; |
| case WMI_VDEV_PARAM_GTX_VHT_MCS: |
| ret = intr[vdev_id].config.gtx_info.gtxRTMask[1]; |
| break; |
| case WMI_VDEV_PARAM_GTX_USR_CFG: |
| ret = intr[vdev_id].config.gtx_info.gtxUsrcfg; |
| break; |
| case WMI_VDEV_PARAM_GTX_THRE: |
| ret = intr[vdev_id].config.gtx_info.gtxPERThreshold; |
| break; |
| case WMI_VDEV_PARAM_GTX_MARGIN: |
| ret = intr[vdev_id].config.gtx_info.gtxPERMargin; |
| break; |
| case WMI_VDEV_PARAM_GTX_STEP: |
| ret = intr[vdev_id].config.gtx_info.gtxTPCstep; |
| break; |
| case WMI_VDEV_PARAM_GTX_MINTPC: |
| ret = intr[vdev_id].config.gtx_info.gtxTPCMin; |
| break; |
| case WMI_VDEV_PARAM_GTX_BW_MASK: |
| ret = intr[vdev_id].config.gtx_info.gtxBWMask; |
| break; |
| #endif /* QCA_SUPPORT_GTX */ |
| case WMI_VDEV_PARAM_LDPC: |
| ret = intr[vdev_id].config.ldpc; |
| break; |
| case WMI_VDEV_PARAM_TX_STBC: |
| ret = intr[vdev_id].config.tx_stbc; |
| break; |
| case WMI_VDEV_PARAM_RX_STBC: |
| ret = intr[vdev_id].config.rx_stbc; |
| break; |
| case WMI_VDEV_PARAM_SGI: |
| ret = intr[vdev_id].config.shortgi; |
| break; |
| case WMI_VDEV_PARAM_ENABLE_RTSCTS: |
| ret = intr[vdev_id].config.rtscts_en; |
| break; |
| case WMI_VDEV_PARAM_CHWIDTH: |
| ret = intr[vdev_id].config.chwidth; |
| break; |
| case WMI_VDEV_PARAM_FIXED_RATE: |
| ret = intr[vdev_id].config.tx_rate; |
| break; |
| case WMI_VDEV_PARAM_HE_DCM: |
| case WMI_VDEV_PARAM_HE_RANGE_EXT: |
| ret = wma_get_he_vdev_param(&intr[vdev_id], param_id); |
| break; |
| default: |
| WMA_LOGE("Invalid cli_get vdev command/Not yet implemented 0x%x", |
| param_id); |
| return -EINVAL; |
| } |
| } else if (PDEV_CMD == vpdev) { |
| switch (param_id) { |
| case WMI_PDEV_PARAM_ANI_ENABLE: |
| ret = wma->pdevconfig.ani_enable; |
| break; |
| case WMI_PDEV_PARAM_ANI_POLL_PERIOD: |
| ret = wma->pdevconfig.ani_poll_len; |
| break; |
| case WMI_PDEV_PARAM_ANI_LISTEN_PERIOD: |
| ret = wma->pdevconfig.ani_listen_len; |
| break; |
| case WMI_PDEV_PARAM_ANI_OFDM_LEVEL: |
| ret = wma->pdevconfig.ani_ofdm_level; |
| break; |
| case WMI_PDEV_PARAM_ANI_CCK_LEVEL: |
| ret = wma->pdevconfig.ani_cck_level; |
| break; |
| case WMI_PDEV_PARAM_DYNAMIC_BW: |
| ret = wma->pdevconfig.cwmenable; |
| break; |
| case WMI_PDEV_PARAM_CTS_CBW: |
| ret = wma->pdevconfig.cts_cbw; |
| break; |
| case WMI_PDEV_PARAM_TX_CHAIN_MASK: |
| ret = wma->pdevconfig.txchainmask; |
| break; |
| case WMI_PDEV_PARAM_RX_CHAIN_MASK: |
| ret = wma->pdevconfig.rxchainmask; |
| break; |
| case WMI_PDEV_PARAM_TXPOWER_LIMIT2G: |
| ret = wma->pdevconfig.txpow2g; |
| break; |
| case WMI_PDEV_PARAM_TXPOWER_LIMIT5G: |
| ret = wma->pdevconfig.txpow5g; |
| break; |
| case WMI_PDEV_PARAM_BURST_ENABLE: |
| ret = wma->pdevconfig.burst_enable; |
| break; |
| case WMI_PDEV_PARAM_BURST_DUR: |
| ret = wma->pdevconfig.burst_dur; |
| break; |
| default: |
| WMA_LOGE("Invalid cli_get pdev command/Not yet implemented 0x%x", |
| param_id); |
| return -EINVAL; |
| } |
| } else if (GEN_CMD == vpdev) { |
| switch (param_id) { |
| case GEN_VDEV_PARAM_AMPDU: |
| ret = intr[vdev_id].config.ampdu; |
| break; |
| case GEN_VDEV_PARAM_AMSDU: |
| ret = intr[vdev_id].config.amsdu; |
| break; |
| case GEN_VDEV_ROAM_SYNCH_DELAY: |
| ret = intr[vdev_id].roam_synch_delay; |
| break; |
| default: |
| WMA_LOGE("Invalid generic vdev command/Not yet implemented 0x%x", |
| param_id); |
| return -EINVAL; |
| } |
| } else if (PPS_CMD == vpdev) { |
| switch (param_id) { |
| case WMI_VDEV_PPS_PAID_MATCH: |
| ret = intr[vdev_id].config.pps_params.paid_match_enable; |
| break; |
| case WMI_VDEV_PPS_GID_MATCH: |
| ret = intr[vdev_id].config.pps_params.gid_match_enable; |
| break; |
| case WMI_VDEV_PPS_EARLY_TIM_CLEAR: |
| ret = intr[vdev_id].config.pps_params.tim_clear; |
| break; |
| case WMI_VDEV_PPS_EARLY_DTIM_CLEAR: |
| ret = intr[vdev_id].config.pps_params.dtim_clear; |
| break; |
| case WMI_VDEV_PPS_EOF_PAD_DELIM: |
| ret = intr[vdev_id].config.pps_params.eof_delim; |
| break; |
| case WMI_VDEV_PPS_MACADDR_MISMATCH: |
| ret = intr[vdev_id].config.pps_params.mac_match; |
| break; |
| case WMI_VDEV_PPS_DELIM_CRC_FAIL: |
| ret = intr[vdev_id].config.pps_params.delim_fail; |
| break; |
| case WMI_VDEV_PPS_GID_NSTS_ZERO: |
| ret = intr[vdev_id].config.pps_params.nsts_zero; |
| break; |
| case WMI_VDEV_PPS_RSSI_CHECK: |
| ret = intr[vdev_id].config.pps_params.rssi_chk; |
| break; |
| default: |
| WMA_LOGE("Invalid pps vdev command/Not yet implemented 0x%x", |
| param_id); |
| return -EINVAL; |
| } |
| } else if (QPOWER_CMD == vpdev) { |
| switch (param_id) { |
| case WMI_STA_PS_PARAM_QPOWER_PSPOLL_COUNT: |
| ret = intr[vdev_id].config.qpower_params. |
| max_ps_poll_cnt; |
| break; |
| case WMI_STA_PS_PARAM_QPOWER_MAX_TX_BEFORE_WAKE: |
| ret = intr[vdev_id].config.qpower_params. |
| max_tx_before_wake; |
| break; |
| case WMI_STA_PS_PARAM_QPOWER_SPEC_PSPOLL_WAKE_INTERVAL: |
| ret = intr[vdev_id].config.qpower_params. |
| spec_ps_poll_wake_interval; |
| break; |
| case WMI_STA_PS_PARAM_QPOWER_SPEC_MAX_SPEC_NODATA_PSPOLL: |
| ret = intr[vdev_id].config.qpower_params. |
| max_spec_nodata_ps_poll; |
| break; |
| default: |
| WMA_LOGE("Invalid generic vdev command/Not yet implemented 0x%x", |
| param_id); |
| return -EINVAL; |
| } |
| } else if (GTX_CMD == vpdev) { |
| switch (param_id) { |
| case WMI_VDEV_PARAM_GTX_HT_MCS: |
| ret = intr[vdev_id].config.gtx_info.gtxRTMask[0]; |
| break; |
| case WMI_VDEV_PARAM_GTX_VHT_MCS: |
| ret = intr[vdev_id].config.gtx_info.gtxRTMask[1]; |
| break; |
| case WMI_VDEV_PARAM_GTX_USR_CFG: |
| ret = intr[vdev_id].config.gtx_info.gtxUsrcfg; |
| break; |
| case WMI_VDEV_PARAM_GTX_THRE: |
| ret = intr[vdev_id].config.gtx_info.gtxPERThreshold; |
| break; |
| case WMI_VDEV_PARAM_GTX_MARGIN: |
| ret = intr[vdev_id].config.gtx_info.gtxPERMargin; |
| break; |
| case WMI_VDEV_PARAM_GTX_STEP: |
| ret = intr[vdev_id].config.gtx_info.gtxTPCstep; |
| break; |
| case WMI_VDEV_PARAM_GTX_MINTPC: |
| ret = intr[vdev_id].config.gtx_info.gtxTPCMin; |
| break; |
| case WMI_VDEV_PARAM_GTX_BW_MASK: |
| ret = intr[vdev_id].config.gtx_info.gtxBWMask; |
| break; |
| default: |
| WMA_LOGE("Invalid generic vdev command/Not yet implemented 0x%x", |
| param_id); |
| return -EINVAL; |
| } |
| } |
| return ret; |
| } |
| |
| /** |
| * wma_cli_set2_command() - WMA "set 2 params" command processor |
| * @vdev_id: virtual device for the command |
| * @param_id: parameter id |
| * @sval1: first parameter value |
| * @sval2: second parameter value |
| * @vpdev: parameter category |
| * |
| * Command handler for set operations which require 2 parameters |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| int wma_cli_set2_command(int vdev_id, int param_id, int sval1, |
| int sval2, int vpdev) |
| { |
| struct scheduler_msg msg = { 0 }; |
| wma_cli_set_cmd_t *iwcmd; |
| |
| iwcmd = qdf_mem_malloc(sizeof(*iwcmd)); |
| if (!iwcmd) { |
| WMA_LOGE("%s: Failed alloc memory for iwcmd", __func__); |
| return -ENOMEM; |
| } |
| |
| qdf_mem_zero(iwcmd, sizeof(*iwcmd)); |
| iwcmd->param_value = sval1; |
| iwcmd->param_sec_value = sval2; |
| iwcmd->param_vdev_id = vdev_id; |
| iwcmd->param_id = param_id; |
| iwcmd->param_vp_dev = vpdev; |
| msg.type = WMA_CLI_SET_CMD; |
| msg.reserved = 0; |
| msg.bodyptr = iwcmd; |
| |
| if (QDF_STATUS_SUCCESS != |
| scheduler_post_msg(QDF_MODULE_ID_WMA, &msg)) { |
| WMA_LOGE("%s: Failed to post WMA_CLI_SET_CMD msg", |
| __func__); |
| qdf_mem_free(iwcmd); |
| return -EIO; |
| } |
| return 0; |
| } |
| |
| /** |
| * wma_cli_set_command() - WMA "set" command processor |
| * @vdev_id: virtual device for the command |
| * @param_id: parameter id |
| * @sval: parameter value |
| * @vpdev: parameter category |
| * |
| * Command handler for set operations |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| int wma_cli_set_command(int vdev_id, int param_id, int sval, int vpdev) |
| { |
| return wma_cli_set2_command(vdev_id, param_id, sval, 0, vpdev); |
| |
| } |
| |
| QDF_STATUS wma_form_unit_test_cmd_and_send(uint32_t vdev_id, |
| uint32_t module_id, uint32_t arg_count, uint32_t *arg) |
| { |
| struct wmi_unit_test_cmd *unit_test_args; |
| tp_wma_handle wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| uint32_t i; |
| QDF_STATUS status; |
| |
| WMA_LOGD(FL("enter")); |
| if (arg_count >= WMA_MAX_NUM_ARGS) { |
| WMA_LOGE(FL("arg_count is crossed the boundary")); |
| return QDF_STATUS_E_FAILURE; |
| } |
| if (!wma_handle || !wma_handle->wmi_handle) { |
| WMA_LOGE(FL("Invalid WMA/WMI handle")); |
| return QDF_STATUS_E_FAILURE; |
| } |
| unit_test_args = qdf_mem_malloc(sizeof(*unit_test_args)); |
| if (NULL == unit_test_args) { |
| WMA_LOGE(FL("qdf_mem_malloc failed for unit_test_args")); |
| return QDF_STATUS_E_NOMEM; |
| } |
| unit_test_args->vdev_id = vdev_id; |
| unit_test_args->module_id = module_id; |
| unit_test_args->num_args = arg_count; |
| for (i = 0; i < arg_count; i++) |
| unit_test_args->args[i] = arg[i]; |
| |
| status = wmi_unified_unit_test_cmd(wma_handle->wmi_handle, |
| unit_test_args); |
| qdf_mem_free(unit_test_args); |
| WMA_LOGD(FL("exit")); |
| |
| return status; |
| } |
| |
| /** |
| * wma_ipa_get_stat() - get IPA data path stats from FW |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| #ifdef IPA_OFFLOAD |
| static int wma_ipa_get_stat(void) |
| { |
| struct cdp_pdev *pdev; |
| |
| pdev = cds_get_context(QDF_MODULE_ID_TXRX); |
| if (!pdev) { |
| WMA_LOGE("pdev NULL for uc stat"); |
| return -EINVAL; |
| } |
| cdp_ipa_get_stat(cds_get_context(QDF_MODULE_ID_SOC), pdev); |
| |
| return 0; |
| } |
| #else |
| static int wma_ipa_get_stat(void) |
| { |
| return 0; |
| } |
| #endif |
| |
| /** |
| * wma_ipa_uc_get_share_stats() - get Tx/Rx byte stats from FW |
| * @privcmd: private command |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| #if defined(IPA_OFFLOAD) && defined(FEATURE_METERING) |
| static int wma_ipa_uc_get_share_stats(wma_cli_set_cmd_t *privcmd) |
| { |
| void *soc = cds_get_context(QDF_MODULE_ID_SOC); |
| struct cdp_pdev *pdev; |
| uint8_t reset_stats = privcmd->param_value; |
| |
| WMA_LOGD("%s: reset_stats=%d", |
| "WMA_VDEV_TXRX_GET_IPA_UC_SHARING_STATS_CMDID", |
| reset_stats); |
| pdev = cds_get_context(QDF_MODULE_ID_TXRX); |
| if (!pdev) { |
| WMA_LOGE("pdev NULL for uc get share stats"); |
| return -EINVAL; |
| } |
| cdp_ipa_uc_get_share_stats(soc, pdev, reset_stats); |
| |
| return 0; |
| } |
| #else |
| static int wma_ipa_uc_get_share_stats(wma_cli_set_cmd_t *privcmd) |
| { |
| return 0; |
| } |
| #endif |
| |
| /** |
| * wma_ipa_uc_set_quota() - set quota limit to FW |
| * @privcmd: private command |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| #if defined(IPA_OFFLOAD) && defined(FEATURE_METERING) |
| static int wma_ipa_uc_set_quota(wma_cli_set_cmd_t *privcmd) |
| { |
| void *soc = cds_get_context(QDF_MODULE_ID_SOC); |
| struct cdp_pdev *pdev; |
| uint64_t quota_bytes = privcmd->param_sec_value; |
| |
| quota_bytes <<= 32; |
| quota_bytes |= privcmd->param_value; |
| |
| WMA_LOGD("%s: quota_bytes=%llu", |
| "WMA_VDEV_TXRX_SET_IPA_UC_QUOTA_CMDID", |
| quota_bytes); |
| pdev = cds_get_context(QDF_MODULE_ID_TXRX); |
| if (!pdev) { |
| WMA_LOGE("pdev NULL for uc set quota"); |
| return -EINVAL; |
| } |
| cdp_ipa_uc_set_quota(soc, pdev, quota_bytes); |
| |
| return 0; |
| } |
| #else |
| static int wma_ipa_uc_set_quota(wma_cli_set_cmd_t *privcmd) |
| { |
| return 0; |
| } |
| #endif |
| |
| /** |
| * wma_set_priv_cfg() - set private config parameters |
| * @wma_handle: wma handle |
| * @privcmd: private command |
| * |
| * Return: 0 for success or error code |
| */ |
| static int32_t wma_set_priv_cfg(tp_wma_handle wma_handle, |
| wma_cli_set_cmd_t *privcmd) |
| { |
| int32_t ret = 0; |
| |
| switch (privcmd->param_id) { |
| case WMA_VDEV_TXRX_FWSTATS_ENABLE_CMDID: |
| ret = wma_set_txrx_fw_stats_level(wma_handle, |
| privcmd->param_vdev_id, |
| privcmd->param_value); |
| break; |
| case WMA_VDEV_TXRX_FWSTATS_RESET_CMDID: |
| ret = wma_txrx_fw_stats_reset(wma_handle, |
| privcmd->param_vdev_id, |
| privcmd->param_value); |
| break; |
| case WMI_STA_SMPS_FORCE_MODE_CMDID: |
| ret = wma_set_mimops(wma_handle, |
| privcmd->param_vdev_id, |
| privcmd->param_value); |
| break; |
| case WMI_STA_SMPS_PARAM_CMDID: |
| wma_set_smps_params(wma_handle, privcmd->param_vdev_id, |
| privcmd->param_value); |
| break; |
| case WMA_VDEV_MCC_SET_TIME_LATENCY: |
| { |
| /* Extract first MCC adapter/vdev channel number and latency */ |
| uint8_t mcc_channel = privcmd->param_value & 0x000000FF; |
| uint8_t mcc_channel_latency = |
| (privcmd->param_value & 0x0000FF00) >> 8; |
| int ret = -1; |
| |
| WMA_LOGD("%s: Parsed input: Channel #1:%d, latency:%dms", |
| __func__, mcc_channel, mcc_channel_latency); |
| ret = wma_set_mcc_channel_time_latency(wma_handle, |
| mcc_channel, |
| mcc_channel_latency); |
| } |
| break; |
| case WMA_VDEV_MCC_SET_TIME_QUOTA: |
| { |
| /* Extract the MCC 2 adapters/vdevs channel numbers and time |
| * quota value for the first adapter only (which is specified |
| * in iwpriv command. |
| */ |
| uint8_t adapter_2_chan_number = |
| privcmd->param_value & 0x000000FF; |
| uint8_t adapter_1_chan_number = |
| (privcmd->param_value & 0x0000FF00) >> 8; |
| uint8_t adapter_1_quota = |
| (privcmd->param_value & 0x00FF0000) >> 16; |
| int ret = -1; |
| |
| WMA_LOGD("%s: Parsed input: Channel #1:%d, Channel #2:%d, quota 1:%dms", |
| __func__, adapter_1_chan_number, |
| adapter_2_chan_number, adapter_1_quota); |
| |
| ret = wma_set_mcc_channel_time_quota(wma_handle, |
| adapter_1_chan_number, |
| adapter_1_quota, |
| adapter_2_chan_number); |
| } |
| break; |
| case WMA_VDEV_IBSS_SET_ATIM_WINDOW_SIZE: |
| { |
| wma_handle->wma_ibss_power_save_params.atimWindowLength = |
| privcmd->param_value; |
| WMA_LOGD("%s: IBSS power save ATIM Window = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| atimWindowLength); |
| } |
| break; |
| case WMA_VDEV_IBSS_SET_POWER_SAVE_ALLOWED: |
| { |
| wma_handle->wma_ibss_power_save_params.isPowerSaveAllowed = |
| privcmd->param_value; |
| WMA_LOGD("%s: IBSS is Power Save Allowed = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| isPowerSaveAllowed); |
| } |
| break; |
| case WMA_VDEV_IBSS_SET_POWER_COLLAPSE_ALLOWED: |
| { |
| wma_handle->wma_ibss_power_save_params. isPowerCollapseAllowed = |
| privcmd->param_value; |
| WMA_LOGD("%s: IBSS is Power Collapse Allowed = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| isPowerCollapseAllowed); |
| } |
| break; |
| case WMA_VDEV_IBSS_SET_AWAKE_ON_TX_RX: |
| { |
| wma_handle->wma_ibss_power_save_params.isAwakeonTxRxEnabled = |
| privcmd->param_value; |
| WMA_LOGD("%s: IBSS Power Save Awake on Tx/Rx Enabled = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| isAwakeonTxRxEnabled); |
| } |
| break; |
| case WMA_VDEV_IBSS_SET_INACTIVITY_TIME: |
| { |
| wma_handle->wma_ibss_power_save_params.inactivityCount = |
| privcmd->param_value; |
| WMA_LOGD("%s: IBSS Power Save Data Inactivity Count = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| inactivityCount); |
| } |
| break; |
| case WMA_VDEV_IBSS_SET_TXSP_END_INACTIVITY_TIME: |
| { |
| wma_handle->wma_ibss_power_save_params.txSPEndInactivityTime = |
| privcmd->param_value; |
| WMA_LOGD("%s: IBSS Power Save Transmit EOSP inactivity time out = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| txSPEndInactivityTime); |
| } |
| break; |
| case WMA_VDEV_IBSS_PS_SET_WARMUP_TIME_SECS: |
| { |
| wma_handle->wma_ibss_power_save_params.ibssPsWarmupTime = |
| privcmd->param_value; |
| WMA_LOGD("%s: IBSS Power Save Warm Up Time in Seconds = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| ibssPsWarmupTime); |
| } |
| break; |
| case WMA_VDEV_IBSS_PS_SET_1RX_CHAIN_IN_ATIM_WINDOW: |
| { |
| wma_handle->wma_ibss_power_save_params.ibssPs1RxChainInAtimEnable |
| = privcmd->param_value; |
| WMA_LOGD("%s: IBSS Power Save single RX Chain Enable In ATIM = %d", |
| __func__, wma_handle->wma_ibss_power_save_params. |
| ibssPs1RxChainInAtimEnable); |
| } |
| break; |
| |
| case WMA_VDEV_TXRX_GET_IPA_UC_FW_STATS_CMDID: |
| { |
| wma_ipa_get_stat(); |
| } |
| break; |
| |
| case WMA_VDEV_TXRX_GET_IPA_UC_SHARING_STATS_CMDID: |
| { |
| wma_ipa_uc_get_share_stats(privcmd); |
| } |
| break; |
| |
| case WMA_VDEV_TXRX_SET_IPA_UC_QUOTA_CMDID: |
| { |
| wma_ipa_uc_set_quota(privcmd); |
| |
| } |
| break; |
| |
| default: |
| WMA_LOGE("Invalid wma config command id:%d", privcmd->param_id); |
| ret = -EINVAL; |
| } |
| return ret; |
| } |
| |
| /** |
| * wma_set_dtim_period() - set dtim period to FW |
| * @wma: wma handle |
| * @dtim_params: dtim params |
| * |
| * Return: none |
| */ |
| static void wma_set_dtim_period(tp_wma_handle wma, |
| struct set_dtim_params *dtim_params) |
| { |
| QDF_STATUS ret; |
| uint8_t vdev_id = dtim_params->session_id; |
| struct wma_txrx_node *iface = |
| &wma->interfaces[vdev_id]; |
| |
| WMA_LOGD("%s: set dtim_period %d", __func__, |
| dtim_params->dtim_period); |
| iface->dtimPeriod = dtim_params->dtim_period; |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| vdev_id, |
| WMI_VDEV_PARAM_LISTEN_INTERVAL, |
| dtim_params->dtim_period); |
| if (QDF_IS_STATUS_ERROR(ret)) |
| WMA_LOGW("Failed to set listen interval"); |
| |
| } |
| /** |
| * wma_set_modulated_dtim() - function to configure modulated dtim |
| * @wma: wma handle |
| * @privcmd: structure containing parameters |
| * |
| * This function configures the modulated dtim in firmware |
| * |
| * Return: none |
| */ |
| static void wma_set_modulated_dtim(tp_wma_handle wma, |
| wma_cli_set_cmd_t *privcmd) |
| { |
| uint8_t vdev_id = privcmd->param_vdev_id; |
| struct wma_txrx_node *iface = |
| &wma->interfaces[vdev_id]; |
| bool prev_dtim_enabled; |
| uint32_t listen_interval; |
| QDF_STATUS ret; |
| |
| iface->alt_modulated_dtim = privcmd->param_value; |
| |
| prev_dtim_enabled = iface->alt_modulated_dtim_enabled; |
| |
| if (1 != privcmd->param_value) |
| iface->alt_modulated_dtim_enabled = true; |
| else |
| iface->alt_modulated_dtim_enabled = false; |
| |
| if ((true == iface->alt_modulated_dtim_enabled) || |
| (true == prev_dtim_enabled)) { |
| |
| listen_interval = iface->alt_modulated_dtim |
| * iface->dtimPeriod; |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| WMI_VDEV_PARAM_LISTEN_INTERVAL, |
| listen_interval); |
| if (QDF_IS_STATUS_ERROR(ret)) |
| /* Even if it fails, continue */ |
| WMA_LOGW("Failed to set listen interval %d", |
| listen_interval); |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| WMI_VDEV_PARAM_DTIM_POLICY, |
| NORMAL_DTIM); |
| if (QDF_IS_STATUS_ERROR(ret)) |
| WMA_LOGE("Failed to Set to Normal DTIM policy"); |
| } |
| } |
| |
| /** |
| * wma_override_listen_interval() - function to override static/ini based LI |
| * @wma: wma handle |
| * @privcmd: structure containing parameters |
| * |
| * This function override static/ini based LI in firmware |
| * |
| * Return: none |
| */ |
| static void wma_override_listen_interval(tp_wma_handle wma, |
| wma_cli_set_cmd_t *privcmd) |
| { |
| uint8_t vdev_id = privcmd->param_vdev_id; |
| struct wma_txrx_node *iface = |
| &wma->interfaces[vdev_id]; |
| u32 old_override_li, new_override_li, listen_interval; |
| struct sAniSirGlobal *mac; |
| QDF_STATUS ret; |
| |
| mac = cds_get_context(QDF_MODULE_ID_PE); |
| if (!mac) { |
| WMA_LOGE(FL("Failed to get mac context")); |
| return; |
| } |
| |
| old_override_li = iface->override_li; |
| new_override_li = privcmd->param_value; |
| iface->override_li = new_override_li; |
| |
| if (new_override_li && |
| (new_override_li != old_override_li)) { |
| listen_interval = new_override_li; |
| } else if (!new_override_li && |
| (new_override_li != old_override_li)) { |
| /* Configure default LI as we do on resume */ |
| if ((wlan_cfg_get_int(mac, WNI_CFG_LISTEN_INTERVAL, |
| &listen_interval) != eSIR_SUCCESS)) { |
| QDF_TRACE(QDF_MODULE_ID_WMA, QDF_TRACE_LEVEL_ERROR, |
| "Failed to get value for listen interval"); |
| listen_interval = POWERSAVE_DEFAULT_LISTEN_INTERVAL; |
| } |
| } else { |
| return; |
| } |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_LISTEN_INTERVAL, |
| listen_interval); |
| if (QDF_IS_STATUS_ERROR(ret)) { |
| /* Even it fails continue Fw will take default LI */ |
| WMA_LOGE("Failed to Set Listen Interval vdevId %d", |
| vdev_id); |
| } |
| WMA_LOGD("%s: Set Listen Interval vdevId %d Listen Intv %d", |
| __func__, vdev_id, listen_interval); |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| WMI_VDEV_PARAM_DTIM_POLICY, |
| NORMAL_DTIM); |
| if (QDF_IS_STATUS_ERROR(ret)) |
| WMA_LOGE("Failed to Set to Normal DTIM policy"); |
| |
| } |
| |
| |
| /** |
| * wma_process_cli_set_cmd() - set parameters to fw |
| * @wma: wma handle |
| * @privcmd: command |
| * |
| * Return: none |
| */ |
| static void wma_process_cli_set_cmd(tp_wma_handle wma, |
| wma_cli_set_cmd_t *privcmd) |
| { |
| int vid = privcmd->param_vdev_id, pps_val = 0; |
| QDF_STATUS ret; |
| struct wma_txrx_node *intr = wma->interfaces; |
| tpAniSirGlobal pMac = cds_get_context(QDF_MODULE_ID_PE); |
| struct qpower_params *qparams = &intr[vid].config.qpower_params; |
| struct pdev_params pdev_param; |
| void *soc = cds_get_context(QDF_MODULE_ID_SOC); |
| |
| WMA_LOGD("wmihandle %pK", wma->wmi_handle); |
| |
| if (NULL == pMac) { |
| WMA_LOGE("%s: Failed to get pMac", __func__); |
| return; |
| } |
| |
| if (privcmd->param_id >= WMI_CMDID_MAX) { |
| /* |
| * This configuration setting is not done using any wmi |
| * command, call appropriate handler. |
| */ |
| if (wma_set_priv_cfg(wma, privcmd)) |
| WMA_LOGE("Failed to set wma priv congiuration"); |
| return; |
| } |
| |
| switch (privcmd->param_vp_dev) { |
| case VDEV_CMD: |
| if (!wma->interfaces[privcmd->param_vdev_id].is_vdev_valid) { |
| WMA_LOGE("%s Vdev id is not valid", __func__); |
| return; |
| } |
| |
| WMA_LOGD("vdev id %d pid %d pval %d", privcmd->param_vdev_id, |
| privcmd->param_id, privcmd->param_value); |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| privcmd->param_id, |
| privcmd->param_value); |
| if (QDF_IS_STATUS_ERROR(ret)) { |
| WMA_LOGE("wma_vdev_set_param failed ret %d", |
| ret); |
| return; |
| } |
| break; |
| case PDEV_CMD: |
| WMA_LOGD("pdev pid %d pval %d", privcmd->param_id, |
| privcmd->param_value); |
| if ((privcmd->param_id == WMI_PDEV_PARAM_RX_CHAIN_MASK) || |
| (privcmd->param_id == WMI_PDEV_PARAM_TX_CHAIN_MASK)) { |
| if (QDF_STATUS_SUCCESS != |
| wma_check_txrx_chainmask( |
| wma->num_rf_chains, |
| privcmd->param_value)) { |
| WMA_LOGD("Chainmask value is invalid"); |
| return; |
| } |
| } |
| pdev_param.param_id = privcmd->param_id; |
| pdev_param.param_value = privcmd->param_value; |
| ret = wmi_unified_pdev_param_send(wma->wmi_handle, |
| &pdev_param, |
| WMA_WILDCARD_PDEV_ID); |
| if (QDF_IS_STATUS_ERROR(ret)) { |
| WMA_LOGE("wma_vdev_set_param failed ret %d", |
| ret); |
| return; |
| } |
| break; |
| case GEN_CMD: |
| { |
| struct cdp_vdev *vdev = NULL; |
| struct wma_txrx_node *intr = wma->interfaces; |
| |
| vdev = wma_find_vdev_by_id(wma, privcmd->param_vdev_id); |
| if (!vdev) { |
| WMA_LOGE("%s:Invalid vdev handle", __func__); |
| return; |
| } |
| |
| WMA_LOGD("gen pid %d pval %d", privcmd->param_id, |
| privcmd->param_value); |
| |
| switch (privcmd->param_id) { |
| case GEN_VDEV_PARAM_AMPDU: |
| if (soc) { |
| ret = cdp_aggr_cfg(soc, vdev, |
| privcmd->param_value, 0); |
| if (ret) |
| WMA_LOGE("cdp_aggr_cfg set ampdu failed ret %d", |
| ret); |
| else |
| intr[privcmd->param_vdev_id].config. |
| ampdu = privcmd->param_value; |
| } else { |
| WMA_LOGE("%s:SOC context is NULL", __func__); |
| return; |
| } |
| break; |
| case GEN_VDEV_PARAM_AMSDU: |
| /* |
| * Firmware currently does not support set operation |
| * for AMSDU. It may cause crash if the configuration |
| * is sent to firmware. |
| * Firmware enhancement will advertise a service bit |
| * to enable AMSDU configuration through WMI. Then |
| * add the WMI command to configure AMSDU parameter. |
| * For the older chipset that does not advertise the |
| * service bit, enable the following legacy code: |
| * ol_txrx_aggr_cfg(vdev, 0, privcmd->param_value); |
| * intr[privcmd->param_vdev_id].config.amsdu = |
| * privcmd->param_value; |
| */ |
| WMA_LOGE("SET GEN_VDEV_PARAM_AMSDU command is currently not supported"); |
| break; |
| case GEN_PARAM_CRASH_INJECT: |
| if (QDF_GLOBAL_FTM_MODE == cds_get_conparam()) |
| WMA_LOGE("Crash inject not allowed in FTM mode"); |
| else |
| ret = wma_crash_inject(wma, |
| privcmd->param_value, |
| privcmd->param_sec_value); |
| break; |
| case GEN_PARAM_CAPTURE_TSF: |
| ret = wma_capture_tsf(wma, privcmd->param_value); |
| break; |
| case GEN_PARAM_RESET_TSF_GPIO: |
| ret = wma_reset_tsf_gpio(wma, privcmd->param_value); |
| break; |
| case GEN_PARAM_MODULATED_DTIM: |
| wma_set_modulated_dtim(wma, privcmd); |
| break; |
| case GEN_PARAM_LISTEN_INTERVAL: |
| wma_override_listen_interval(wma, privcmd); |
| break; |
| default: |
| WMA_LOGE("Invalid param id 0x%x", |
| privcmd->param_id); |
| break; |
| } |
| break; |
| } |
| case DBG_CMD: |
| WMA_LOGD("dbg pid %d pval %d", privcmd->param_id, |
| privcmd->param_value); |
| switch (privcmd->param_id) { |
| case WMI_DBGLOG_LOG_LEVEL: |
| ret = dbglog_set_log_lvl(wma->wmi_handle, |
| privcmd->param_value); |
| if (ret) |
| WMA_LOGE("dbglog_set_log_lvl failed ret %d", |
| ret); |
| break; |
| case WMI_DBGLOG_VAP_ENABLE: |
| ret = dbglog_vap_log_enable(wma->wmi_handle, |
| privcmd->param_value, true); |
| if (ret) |
| WMA_LOGE("dbglog_vap_log_enable failed ret %d", |
| ret); |
| break; |
| case WMI_DBGLOG_VAP_DISABLE: |
| ret = dbglog_vap_log_enable(wma->wmi_handle, |
| privcmd->param_value, false); |
| if (ret) |
| WMA_LOGE("dbglog_vap_log_enable failed ret %d", |
| ret); |
| break; |
| case WMI_DBGLOG_MODULE_ENABLE: |
| ret = dbglog_module_log_enable(wma->wmi_handle, |
| privcmd->param_value, true); |
| if (ret) |
| WMA_LOGE("dbglog_module_log_enable failed ret %d", |
| ret); |
| break; |
| case WMI_DBGLOG_MODULE_DISABLE: |
| ret = dbglog_module_log_enable(wma->wmi_handle, |
| privcmd->param_value, false); |
| if (ret) |
| WMA_LOGE("dbglog_module_log_enable failed ret %d", |
| ret); |
| break; |
| case WMI_DBGLOG_MOD_LOG_LEVEL: |
| ret = dbglog_set_mod_log_lvl(wma->wmi_handle, |
| privcmd->param_value); |
| if (ret) |
| WMA_LOGE("dbglog_module_log_enable failed ret %d", |
| ret); |
| break; |
| case WMI_DBGLOG_TYPE: |
| ret = dbglog_parser_type_init(wma->wmi_handle, |
| privcmd->param_value); |
| if (ret) |
| WMA_LOGE("dbglog_parser_type_init failed ret %d", |
| ret); |
| break; |
| case WMI_DBGLOG_REPORT_ENABLE: |
| ret = dbglog_report_enable(wma->wmi_handle, |
| privcmd->param_value); |
| if (ret) |
| WMA_LOGE("dbglog_report_enable failed ret %d", |
| ret); |
| break; |
| case WMI_WLAN_PROFILE_TRIGGER_CMDID: |
| ret = wma_unified_fw_profiling_cmd(wma->wmi_handle, |
| WMI_WLAN_PROFILE_TRIGGER_CMDID, |
| privcmd->param_value, 0); |
| if (ret) |
| WMA_LOGE("Profile cmd failed for %d ret %d", |
| WMI_WLAN_PROFILE_TRIGGER_CMDID, ret); |
| break; |
| case WMI_WLAN_PROFILE_ENABLE_PROFILE_ID_CMDID: |
| ret = wma_unified_fw_profiling_cmd(wma->wmi_handle, |
| WMI_WLAN_PROFILE_ENABLE_PROFILE_ID_CMDID, |
| privcmd->param_value, |
| privcmd->param_sec_value); |
| if (ret) |
| WMA_LOGE("Profile cmd failed for %d ret %d", |
| WMI_WLAN_PROFILE_ENABLE_PROFILE_ID_CMDID, |
| ret); |
| break; |
| case WMI_WLAN_PROFILE_SET_HIST_INTVL_CMDID: |
| ret = wma_unified_fw_profiling_cmd(wma->wmi_handle, |
| WMI_WLAN_PROFILE_SET_HIST_INTVL_CMDID, |
| privcmd->param_value, |
| privcmd->param_sec_value); |
| if (ret) |
| WMA_LOGE("Profile cmd failed for %d ret %d", |
| WMI_WLAN_PROFILE_SET_HIST_INTVL_CMDID, |
| ret); |
| break; |
| case WMI_WLAN_PROFILE_LIST_PROFILE_ID_CMDID: |
| ret = wma_unified_fw_profiling_cmd(wma->wmi_handle, |
| WMI_WLAN_PROFILE_LIST_PROFILE_ID_CMDID, |
| 0, 0); |
| if (ret) |
| WMA_LOGE("Profile cmd failed for %d ret %d", |
| WMI_WLAN_PROFILE_LIST_PROFILE_ID_CMDID, |
| ret); |
| break; |
| case WMI_WLAN_PROFILE_GET_PROFILE_DATA_CMDID: |
| ret = wma_unified_fw_profiling_cmd(wma->wmi_handle, |
| WMI_WLAN_PROFILE_GET_PROFILE_DATA_CMDID, |
| 0, 0); |
| if (ret) |
| WMA_LOGE("Profile cmd failed for %d ret %d", |
| WMI_WLAN_PROFILE_GET_PROFILE_DATA_CMDID, |
| ret); |
| break; |
| case WMI_PDEV_GREEN_AP_PS_ENABLE_CMDID: |
| /* Set the Green AP */ |
| ret = wmi_unified_green_ap_ps_send |
| (wma->wmi_handle, privcmd->param_value, |
| WMA_WILDCARD_PDEV_ID); |
| if (ret) { |
| WMA_LOGE("Set GreenAP Failed val %d", |
| privcmd->param_value); |
| } |
| break; |
| |
| default: |
| WMA_LOGE("Invalid param id 0x%x", privcmd->param_id); |
| break; |
| } |
| break; |
| case PPS_CMD: |
| WMA_LOGD("dbg pid %d pval %d", privcmd->param_id, |
| privcmd->param_value); |
| switch (privcmd->param_id) { |
| |
| case WMI_VDEV_PPS_PAID_MATCH: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_PAID_MATCH & 0xffff); |
| intr[vid].config.pps_params.paid_match_enable = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_GID_MATCH: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_GID_MATCH & 0xffff); |
| intr[vid].config.pps_params.gid_match_enable = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_EARLY_TIM_CLEAR: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_EARLY_TIM_CLEAR & 0xffff); |
| intr[vid].config.pps_params.tim_clear = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_EARLY_DTIM_CLEAR: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_EARLY_DTIM_CLEAR & 0xffff); |
| intr[vid].config.pps_params.dtim_clear = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_EOF_PAD_DELIM: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_EOF_PAD_DELIM & 0xffff); |
| intr[vid].config.pps_params.eof_delim = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_MACADDR_MISMATCH: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_MACADDR_MISMATCH & 0xffff); |
| intr[vid].config.pps_params.mac_match = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_DELIM_CRC_FAIL: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_DELIM_CRC_FAIL & 0xffff); |
| intr[vid].config.pps_params.delim_fail = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_GID_NSTS_ZERO: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_GID_NSTS_ZERO & 0xffff); |
| intr[vid].config.pps_params.nsts_zero = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_RSSI_CHECK: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_RSSI_CHECK & 0xffff); |
| intr[vid].config.pps_params.rssi_chk = |
| privcmd->param_value; |
| break; |
| case WMI_VDEV_PPS_5G_EBT: |
| pps_val = ((privcmd->param_value << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_5G_EBT & 0xffff); |
| intr[vid].config.pps_params.ebt_5g = |
| privcmd->param_value; |
| break; |
| default: |
| WMA_LOGE("Invalid param id 0x%x", privcmd->param_id); |
| break; |
| } |
| break; |
| |
| case QPOWER_CMD: |
| WMA_LOGD("QPOWER CLI CMD pid %d pval %d", privcmd->param_id, |
| privcmd->param_value); |
| switch (privcmd->param_id) { |
| case WMI_STA_PS_PARAM_QPOWER_PSPOLL_COUNT: |
| WMA_LOGD("QPOWER CLI CMD:Ps Poll Cnt val %d", |
| privcmd->param_value); |
| /* Set the QPower Ps Poll Count */ |
| ret = wma_unified_set_sta_ps_param(wma->wmi_handle, |
| vid, WMI_STA_PS_PARAM_QPOWER_PSPOLL_COUNT, |
| privcmd->param_value); |
| if (ret) { |
| WMA_LOGE("Set Q-PsPollCnt Failed vdevId %d val %d", |
| vid, privcmd->param_value); |
| } else { |
| qparams->max_ps_poll_cnt = privcmd->param_value; |
| } |
| break; |
| case WMI_STA_PS_PARAM_QPOWER_MAX_TX_BEFORE_WAKE: |
| WMA_LOGD("QPOWER CLI CMD:Max Tx Before wake val %d", |
| privcmd->param_value); |
| /* Set the QPower Max Tx Before Wake */ |
| ret = wma_unified_set_sta_ps_param(wma->wmi_handle, |
| vid, WMI_STA_PS_PARAM_QPOWER_MAX_TX_BEFORE_WAKE, |
| privcmd->param_value); |
| if (ret) { |
| WMA_LOGE("Set Q-MaxTxBefWake Failed vId %d val %d", |
| vid, privcmd->param_value); |
| } else { |
| qparams->max_tx_before_wake = |
| privcmd->param_value; |
| } |
| break; |
| case WMI_STA_PS_PARAM_QPOWER_SPEC_PSPOLL_WAKE_INTERVAL: |
| WMA_LOGD("QPOWER CLI CMD:Ps Poll Wake Inv val %d", |
| privcmd->param_value); |
| /* Set the QPower Spec Ps Poll Wake Inv */ |
| ret = wma_unified_set_sta_ps_param(wma->wmi_handle, vid, |
| WMI_STA_PS_PARAM_QPOWER_SPEC_PSPOLL_WAKE_INTERVAL, |
| privcmd->param_value); |
| if (ret) { |
| WMA_LOGE("Set Q-PsPoll WakeIntv Failed vId %d val %d", |
| vid, privcmd->param_value); |
| } else { |
| qparams->spec_ps_poll_wake_interval = |
| privcmd->param_value; |
| } |
| break; |
| case WMI_STA_PS_PARAM_QPOWER_SPEC_MAX_SPEC_NODATA_PSPOLL: |
| WMA_LOGD("QPOWER CLI CMD:Spec NoData Ps Poll val %d", |
| privcmd->param_value); |
| /* Set the QPower Spec NoData PsPoll */ |
| ret = wma_unified_set_sta_ps_param(wma->wmi_handle, vid, |
| WMI_STA_PS_PARAM_QPOWER_SPEC_MAX_SPEC_NODATA_PSPOLL, |
| privcmd->param_value); |
| if (ret) { |
| WMA_LOGE("Set Q-SpecNoDataPsPoll Failed vId %d val %d", |
| vid, privcmd->param_value); |
| } else { |
| qparams->max_spec_nodata_ps_poll = |
| privcmd->param_value; |
| } |
| break; |
| |
| default: |
| WMA_LOGE("Invalid param id 0x%x", privcmd->param_id); |
| break; |
| } |
| break; |
| case GTX_CMD: |
| WMA_LOGD("vdev id %d pid %d pval %d", privcmd->param_vdev_id, |
| privcmd->param_id, privcmd->param_value); |
| switch (privcmd->param_id) { |
| case WMI_VDEV_PARAM_GTX_HT_MCS: |
| intr[vid].config.gtx_info.gtxRTMask[0] = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| break; |
| case WMI_VDEV_PARAM_GTX_VHT_MCS: |
| intr[vid].config.gtx_info.gtxRTMask[1] = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| break; |
| |
| case WMI_VDEV_PARAM_GTX_USR_CFG: |
| intr[vid].config.gtx_info.gtxUsrcfg = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| break; |
| |
| case WMI_VDEV_PARAM_GTX_THRE: |
| intr[vid].config.gtx_info.gtxPERThreshold = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| break; |
| |
| case WMI_VDEV_PARAM_GTX_MARGIN: |
| intr[vid].config.gtx_info.gtxPERMargin = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| break; |
| |
| case WMI_VDEV_PARAM_GTX_STEP: |
| intr[vid].config.gtx_info.gtxTPCstep = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| break; |
| |
| case WMI_VDEV_PARAM_GTX_MINTPC: |
| intr[vid].config.gtx_info.gtxTPCMin = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| break; |
| |
| case WMI_VDEV_PARAM_GTX_BW_MASK: |
| intr[vid].config.gtx_info.gtxBWMask = |
| privcmd->param_value; |
| ret = wmi_unified_vdev_set_gtx_cfg_send(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| &intr[vid].config.gtx_info); |
| if (ret) { |
| WMA_LOGE("wma_vdev_set_param failed ret %d", |
| ret); |
| return; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| |
| default: |
| WMA_LOGE("Invalid vpdev command id"); |
| } |
| if (1 == privcmd->param_vp_dev) { |
| switch (privcmd->param_id) { |
| case WMI_VDEV_PARAM_NSS: |
| intr[vid].config.nss = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_LDPC: |
| intr[vid].config.ldpc = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_TX_STBC: |
| intr[vid].config.tx_stbc = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_RX_STBC: |
| intr[vid].config.rx_stbc = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_SGI: |
| intr[vid].config.shortgi = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_ENABLE_RTSCTS: |
| intr[vid].config.rtscts_en = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_CHWIDTH: |
| intr[vid].config.chwidth = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_FIXED_RATE: |
| intr[vid].config.tx_rate = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_EARLY_RX_ADJUST_ENABLE: |
| intr[vid].config.erx_adjust = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_EARLY_RX_TGT_BMISS_NUM: |
| intr[vid].config.erx_bmiss_num = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_EARLY_RX_BMISS_SAMPLE_CYCLE: |
| intr[vid].config.erx_bmiss_cycle = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_EARLY_RX_SLOP_STEP: |
| intr[vid].config.erx_slop_step = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_EARLY_RX_INIT_SLOP: |
| intr[vid].config.erx_init_slop = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_EARLY_RX_ADJUST_PAUSE: |
| intr[vid].config.erx_adj_pause = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_EARLY_RX_DRIFT_SAMPLE: |
| intr[vid].config.erx_dri_sample = privcmd->param_value; |
| break; |
| case WMI_VDEV_PARAM_HE_DCM: |
| case WMI_VDEV_PARAM_HE_RANGE_EXT: |
| wma_set_he_vdev_param(&intr[vid], privcmd->param_id, |
| privcmd->param_value); |
| break; |
| default: |
| WMA_LOGE("Invalid wma_cli_set vdev command/Not yet implemented 0x%x", |
| privcmd->param_id); |
| break; |
| } |
| } else if (2 == privcmd->param_vp_dev) { |
| switch (privcmd->param_id) { |
| case WMI_PDEV_PARAM_ANI_ENABLE: |
| wma->pdevconfig.ani_enable = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_ANI_POLL_PERIOD: |
| wma->pdevconfig.ani_poll_len = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_ANI_LISTEN_PERIOD: |
| wma->pdevconfig.ani_listen_len = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_ANI_OFDM_LEVEL: |
| wma->pdevconfig.ani_ofdm_level = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_ANI_CCK_LEVEL: |
| wma->pdevconfig.ani_cck_level = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_DYNAMIC_BW: |
| wma->pdevconfig.cwmenable = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_CTS_CBW: |
| wma->pdevconfig.cts_cbw = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_TX_CHAIN_MASK: |
| wma->pdevconfig.txchainmask = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_RX_CHAIN_MASK: |
| wma->pdevconfig.rxchainmask = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_BURST_ENABLE: |
| wma->pdevconfig.burst_enable = privcmd->param_value; |
| if ((wma->pdevconfig.burst_enable == 1) && |
| (wma->pdevconfig.burst_dur == 0)) |
| wma->pdevconfig.burst_dur = |
| WMA_DEFAULT_SIFS_BURST_DURATION; |
| else if (wma->pdevconfig.burst_enable == 0) |
| wma->pdevconfig.burst_dur = 0; |
| break; |
| case WMI_PDEV_PARAM_BURST_DUR: |
| wma->pdevconfig.burst_dur = privcmd->param_value; |
| break; |
| case WMI_PDEV_PARAM_TXPOWER_LIMIT2G: |
| wma->pdevconfig.txpow2g = privcmd->param_value; |
| if ((pMac->roam.configParam.bandCapability == |
| eCSR_BAND_ALL) || |
| (pMac->roam.configParam.bandCapability == |
| eCSR_BAND_24)) { |
| if (cfg_set_int(pMac, |
| WNI_CFG_CURRENT_TX_POWER_LEVEL, |
| privcmd->param_value) != |
| eSIR_SUCCESS) |
| WMA_LOGE("could not set WNI_CFG_CURRENT_TX_POWER_LEVEL"); |
| |
| } else { |
| WMA_LOGE("Current band is not 2G"); |
| } |
| break; |
| case WMI_PDEV_PARAM_TXPOWER_LIMIT5G: |
| wma->pdevconfig.txpow5g = privcmd->param_value; |
| if ((pMac->roam.configParam.bandCapability == |
| eCSR_BAND_ALL) || |
| (pMac->roam.configParam.bandCapability == |
| eCSR_BAND_5G)) { |
| if (cfg_set_int(pMac, |
| WNI_CFG_CURRENT_TX_POWER_LEVEL, |
| privcmd->param_value) != |
| eSIR_SUCCESS) |
| WMA_LOGE("could not set WNI_CFG_CURRENT_TX_POWER_LEVEL"); |
| |
| } else { |
| WMA_LOGE("Current band is not 5G"); |
| } |
| break; |
| default: |
| WMA_LOGD("Invalid wma_cli_set pdev command/Not yet implemented 0x%x", |
| privcmd->param_id); |
| break; |
| } |
| } else if (5 == privcmd->param_vp_dev) { |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| privcmd->param_vdev_id, |
| WMI_VDEV_PARAM_PACKET_POWERSAVE, |
| pps_val); |
| if (ret) |
| WMA_LOGE("Failed to send wmi packet power save cmd"); |
| else |
| WMA_LOGD("Sent packet power save cmd %d value %x to target", |
| privcmd->param_id, pps_val); |
| } |
| } |
| |
| uint32_t wma_critical_events_in_flight(void) |
| { |
| t_wma_handle *wma; |
| |
| wma = cds_get_context(QDF_MODULE_ID_WMA); |
| if (!wma) |
| return 0; |
| |
| return qdf_atomic_read(&wma->critical_events_in_flight); |
| } |
| |
| static bool wma_event_is_critical(uint32_t event_id) |
| { |
| switch (event_id) { |
| case WMI_ROAM_SYNCH_EVENTID: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| /** |
| * wma_process_fw_event() - process any fw event |
| * @wma: wma handle |
| * @buf: fw event buffer |
| * |
| * This function process any fw event to serialize it through mc thread. |
| * |
| * Return: none |
| */ |
| static int wma_process_fw_event(tp_wma_handle wma, |
| wma_process_fw_event_params *buf) |
| { |
| struct wmi_unified *wmi_handle = (struct wmi_unified *)buf->wmi_handle; |
| uint32_t event_id = WMI_GET_FIELD(qdf_nbuf_data(buf->evt_buf), |
| WMI_CMD_HDR, COMMANDID); |
| |
| wmi_process_fw_event(wmi_handle, buf->evt_buf); |
| |
| if (wma_event_is_critical(event_id)) |
| qdf_atomic_dec(&wma->critical_events_in_flight); |
| |
| return 0; |
| } |
| |
| /** |
| * wmi_process_fw_event_tasklet_ctx() - process in tasklet context |
| * @ctx: handle to wmi |
| * @ev: wmi event buffer |
| * |
| * Event process by below function will be in tasket context, |
| * need to use this method only for time sensitive functions. |
| * |
| * Return: none |
| */ |
| static int wma_process_fw_event_tasklet_ctx(void *ctx, void *ev) |
| { |
| wmi_process_fw_event(ctx, ev); |
| |
| return 0; |
| } |
| |
| /** |
| * wma_process_hal_pwr_dbg_cmd() - send hal pwr dbg cmd to fw. |
| * @handle: wma handle |
| * @sir_pwr_dbg_params: unit test command |
| * |
| * This function send unit test command to fw. |
| * |
| * Return: QDF_STATUS_SUCCESS on success, QDF_STATUS_E_** on error |
| */ |
| QDF_STATUS wma_process_hal_pwr_dbg_cmd(WMA_HANDLE handle, |
| struct sir_mac_pwr_dbg_cmd * |
| sir_pwr_dbg_params) |
| { |
| tp_wma_handle wma_handle = (tp_wma_handle)handle; |
| int i; |
| struct wmi_power_dbg_params wmi_pwr_dbg_params; |
| QDF_STATUS status; |
| |
| if (!sir_pwr_dbg_params) { |
| WMA_LOGE("%s: sir_pwr_dbg_params is null", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| wmi_pwr_dbg_params.module_id = sir_pwr_dbg_params->module_id; |
| wmi_pwr_dbg_params.pdev_id = sir_pwr_dbg_params->pdev_id; |
| wmi_pwr_dbg_params.num_args = sir_pwr_dbg_params->num_args; |
| |
| for (i = 0; i < wmi_pwr_dbg_params.num_args; i++) |
| wmi_pwr_dbg_params.args[i] = sir_pwr_dbg_params->args[i]; |
| |
| status = wmi_unified_send_power_dbg_cmd(wma_handle->wmi_handle, |
| &wmi_pwr_dbg_params); |
| |
| return status; |
| } |
| |
| static void wma_discard_fw_event(struct scheduler_msg *msg) |
| { |
| switch (msg->type) { |
| case WMA_PROCESS_FW_EVENT: |
| qdf_nbuf_free(((wma_process_fw_event_params *)msg->bodyptr) |
| ->evt_buf); |
| break; |
| case WMA_SET_LINK_STATE: |
| qdf_mem_free(((tpLinkStateParams) msg->bodyptr)->callbackArg); |
| break; |
| } |
| if (msg->bodyptr) |
| qdf_mem_free(msg->bodyptr); |
| msg->bodyptr = NULL; |
| msg->bodyval = 0; |
| msg->type = 0; |
| } |
| |
| /** |
| * wma_process_fw_event_handler() - common event handler to serialize |
| * event processing through mc_thread |
| * @ctx: wmi context |
| * @ev: event buffer |
| * @rx_ctx: rx execution context |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| static int wma_process_fw_event_mc_thread_ctx(void *ctx, void *ev) |
| { |
| wma_process_fw_event_params *params_buf; |
| struct scheduler_msg cds_msg = { 0 }; |
| tp_wma_handle wma; |
| uint32_t event_id; |
| |
| params_buf = qdf_mem_malloc(sizeof(wma_process_fw_event_params)); |
| if (!params_buf) { |
| WMA_LOGE("%s: Failed alloc memory for params_buf", __func__); |
| qdf_nbuf_free(ev); |
| return -ENOMEM; |
| } |
| |
| params_buf->wmi_handle = (struct wmi_unified *)ctx; |
| params_buf->evt_buf = (wmi_buf_t *)ev; |
| |
| wma = cds_get_context(QDF_MODULE_ID_WMA); |
| event_id = WMI_GET_FIELD(qdf_nbuf_data(params_buf->evt_buf), |
| WMI_CMD_HDR, COMMANDID); |
| if (wma && wma_event_is_critical(event_id)) |
| qdf_atomic_inc(&wma->critical_events_in_flight); |
| |
| cds_msg.type = WMA_PROCESS_FW_EVENT; |
| cds_msg.bodyptr = params_buf; |
| cds_msg.bodyval = 0; |
| cds_msg.flush_callback = wma_discard_fw_event; |
| |
| if (QDF_STATUS_SUCCESS != |
| scheduler_post_msg(QDF_MODULE_ID_WMA, &cds_msg)) { |
| WMA_LOGE("%s: Failed to post WMA_PROCESS_FW_EVENT msg", |
| __func__); |
| qdf_nbuf_free(ev); |
| qdf_mem_free(params_buf); |
| return -EFAULT; |
| } |
| return 0; |
| |
| } |
| |
| /** |
| * wma_process_fw_event_handler() - common event handler to serialize |
| * event processing through mc_thread |
| * @ctx: wmi context |
| * @ev: event buffer |
| * @rx_ctx: rx execution context |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| int wma_process_fw_event_handler(void *ctx, void *ev, uint8_t rx_ctx) |
| { |
| int err = 0; |
| |
| if (rx_ctx == WMA_RX_SERIALIZER_CTX) { |
| err = wma_process_fw_event_mc_thread_ctx(ctx, ev); |
| } else if (rx_ctx == WMA_RX_TASKLET_CTX) { |
| wma_process_fw_event_tasklet_ctx(ctx, ev); |
| } else { |
| WMA_LOGE("%s: invalid wmi event execution context", __func__); |
| qdf_nbuf_free(ev); |
| } |
| |
| return err; |
| } |
| |
| #ifdef WLAN_FEATURE_NAN |
| /** |
| * wma_set_nan_enable() - set nan enable flag in WMA handle |
| * @wma_handle: Pointer to wma handle |
| * @cds_cfg: Pointer to CDS Configuration |
| * |
| * Return: none |
| */ |
| static void wma_set_nan_enable(tp_wma_handle wma_handle, |
| struct cds_config_info *cds_cfg) |
| { |
| wma_handle->is_nan_enabled = cds_cfg->is_nan_enabled; |
| } |
| #else |
| static void wma_set_nan_enable(tp_wma_handle wma_handle, |
| struct cds_config_info *cds_cfg) |
| { |
| } |
| #endif |
| |
| /** |
| * wma_init_max_no_of_peers - API to initialize wma configuration params |
| * @wma_handle: WMA Handle |
| * @max_peers: Max Peers supported |
| * |
| * Return: void |
| */ |
| static void wma_init_max_no_of_peers(tp_wma_handle wma_handle, |
| uint16_t max_peers) |
| { |
| struct wma_ini_config *cfg = wma_get_ini_handle(wma_handle); |
| |
| if (cfg == NULL) { |
| WMA_LOGE("%s: NULL WMA ini handle", __func__); |
| return; |
| } |
| |
| cfg->max_no_of_peers = max_peers; |
| } |
| |
| /** |
| * wma_cleanup_vdev_resp_queue() - cleanup vdev response queue |
| * @wma: wma handle |
| * |
| * Return: none |
| */ |
| static void wma_cleanup_vdev_resp_queue(tp_wma_handle wma) |
| { |
| struct wma_target_req *req_msg = NULL; |
| qdf_list_node_t *node1 = NULL; |
| |
| qdf_spin_lock_bh(&wma->vdev_respq_lock); |
| if (!qdf_list_size(&wma->vdev_resp_queue)) { |
| qdf_spin_unlock_bh(&wma->vdev_respq_lock); |
| WMA_LOGD(FL("request queue maybe empty")); |
| return; |
| } |
| |
| WMA_LOGD(FL("Cleaning up vdev resp queue")); |
| |
| /* peek front, and then cleanup it in wma_vdev_resp_timer */ |
| while (qdf_list_peek_front(&wma->vdev_resp_queue, &node1) == |
| QDF_STATUS_SUCCESS) { |
| req_msg = qdf_container_of(node1, struct wma_target_req, node); |
| qdf_spin_unlock_bh(&wma->vdev_respq_lock); |
| qdf_mc_timer_stop(&req_msg->event_timeout); |
| wma_vdev_resp_timer(req_msg); |
| qdf_spin_lock_bh(&wma->vdev_respq_lock); |
| } |
| qdf_spin_unlock_bh(&wma->vdev_respq_lock); |
| } |
| |
| /** |
| * wma_cleanup_hold_req() - cleanup hold request queue |
| * @wma: wma handle |
| * |
| * Return: none |
| */ |
| static void wma_cleanup_hold_req(tp_wma_handle wma) |
| { |
| struct wma_target_req *req_msg = NULL; |
| qdf_list_node_t *node1 = NULL; |
| |
| qdf_spin_lock_bh(&wma->wma_hold_req_q_lock); |
| if (!qdf_list_size(&wma->wma_hold_req_queue)) { |
| qdf_spin_unlock_bh(&wma->wma_hold_req_q_lock); |
| WMA_LOGD(FL("request queue is empty")); |
| return; |
| } |
| |
| /* peek front, and then cleanup it in wma_hold_req_timer */ |
| while (QDF_STATUS_SUCCESS == |
| qdf_list_peek_front(&wma->wma_hold_req_queue, &node1)) { |
| req_msg = qdf_container_of(node1, struct wma_target_req, node); |
| qdf_spin_unlock_bh(&wma->wma_hold_req_q_lock); |
| /* Cleanup timeout handler */ |
| qdf_mc_timer_stop(&req_msg->event_timeout); |
| wma_hold_req_timer(req_msg); |
| qdf_spin_lock_bh(&wma->wma_hold_req_q_lock); |
| } |
| qdf_spin_unlock_bh(&wma->wma_hold_req_q_lock); |
| } |
| |
| /** |
| * wma_shutdown_notifier_cb - Shutdown notifer call back |
| * @priv : WMA handle |
| * |
| * During recovery, WMA may wait for resume to complete if the crash happens |
| * while in suspend. This may cause delays in completing the recovery. This call |
| * back would be called during recovery and the event is completed so that if |
| * the resume is waiting on FW to respond then it can get out of the wait so |
| * that recovery thread can start bringing down all the modules. |
| * |
| * Return: None |
| */ |
| static void wma_shutdown_notifier_cb(void *priv) |
| { |
| tp_wma_handle wma_handle = priv; |
| |
| qdf_event_set(&wma_handle->wma_resume_event); |
| wma_cleanup_vdev_resp_queue(wma_handle); |
| wma_cleanup_hold_req(wma_handle); |
| pmo_ucfg_psoc_wakeup_host_event_received(wma_handle->psoc); |
| } |
| |
| struct wma_version_info g_wmi_version_info; |
| |
| /** |
| * wma_state_info_dump() - prints state information of wma layer |
| * @buf: buffer pointer |
| * @size: size of buffer to be filled |
| * |
| * This function is used to dump state information of wma layer |
| * |
| * Return: None |
| */ |
| static void wma_state_info_dump(char **buf_ptr, uint16_t *size) |
| { |
| t_wma_handle *wma; |
| struct sir_vdev_wow_stats *stats; |
| uint16_t len = 0; |
| char *buf = *buf_ptr; |
| struct wma_txrx_node *iface; |
| uint8_t vdev_id; |
| |
| wma = cds_get_context(QDF_MODULE_ID_WMA); |
| if (!wma) { |
| WMA_LOGE("%s: WMA context is invald!", __func__); |
| return; |
| } |
| |
| WMA_LOGE("%s: size of buffer: %d", __func__, *size); |
| |
| for (vdev_id = 0; vdev_id < wma->max_bssid; vdev_id++) { |
| iface = &wma->interfaces[vdev_id]; |
| if (!iface->handle) |
| continue; |
| |
| stats = &iface->wow_stats; |
| len += qdf_scnprintf(buf + len, *size - len, |
| "\n" |
| "vdev_id %d\n" |
| "WoW Stats\n" |
| "\tpno_match %u\n" |
| "\tpno_complete %u\n" |
| "\tgscan %u\n" |
| "\tlow_rssi %u\n" |
| "\trssi_breach %u\n" |
| "\tucast %u\n" |
| "\tbcast %u\n" |
| "\ticmpv4 %u\n" |
| "\ticmpv6 %u\n" |
| "\tipv4_mcast %u\n" |
| "\tipv6_mcast %u\n" |
| "\tipv6_mcast_ra %u\n" |
| "\tipv6_mcast_ns %u\n" |
| "\tipv6_mcast_na %u\n" |
| "\toem_response %u\n" |
| "conn_state %d\n" |
| "dtimPeriod %d\n" |
| "chanmode %d\n" |
| "vht_capable %d\n" |
| "ht_capable %d\n" |
| "chan_width %d\n" |
| "vdev_active %d\n" |
| "vdev_up %d\n" |
| "aid %d\n" |
| "rate_flags %d\n" |
| "nss %d\n" |
| "tx_power %d\n" |
| "max_tx_power %d\n" |
| "nwType %d\n" |
| "tx_streams %d\n" |
| "rx_streams %d\n" |
| "chain_mask %d\n" |
| "nss_2g %d\n" |
| "nss_5g %d", |
| vdev_id, |
| stats->pno_match, |
| stats->pno_complete, |
| stats->gscan, |
| stats->low_rssi, |
| stats->rssi_breach, |
| stats->ucast, |
| stats->bcast, |
| stats->icmpv4, |
| stats->icmpv6, |
| stats->ipv4_mcast, |
| stats->ipv6_mcast, |
| stats->ipv6_mcast_ra, |
| stats->ipv6_mcast_ns, |
| stats->ipv6_mcast_na, |
| stats->oem_response, |
| iface->conn_state, |
| iface->dtimPeriod, |
| iface->chanmode, |
| iface->vht_capable, |
| iface->ht_capable, |
| iface->chan_width, |
| iface->vdev_active, |
| wma_is_vdev_up(vdev_id), |
| iface->aid, |
| iface->rate_flags, |
| iface->nss, |
| iface->tx_power, |
| iface->max_tx_power, |
| iface->nwType, |
| iface->tx_streams, |
| iface->rx_streams, |
| iface->chain_mask, |
| iface->nss_2g, |
| iface->nss_5g); |
| } |
| |
| *size -= len; |
| *buf_ptr += len; |
| } |
| |
| /** |
| * wma_register_debug_callback() - registration function for wma layer |
| * to print wma state information |
| */ |
| static void wma_register_debug_callback(void) |
| { |
| qdf_register_debug_callback(QDF_MODULE_ID_WMA, &wma_state_info_dump); |
| } |
| |
| /** |
| * wma_register_tx_ops_handler() - register tx_ops of southbound |
| * @tx_ops: tx_ops pointer in southbound |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| static QDF_STATUS |
| wma_register_tx_ops_handler(struct wlan_lmac_if_tx_ops *tx_ops) |
| { |
| /* |
| * Assign tx_ops, it's up to UMAC modules to declare and define these |
| * functions which are used to send wmi command to target. |
| */ |
| |
| if (!tx_ops) { |
| WMA_LOGE("%s: pointer to lmac if tx ops is NULL", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* mgmt_txrx component's tx ops */ |
| tx_ops->mgmt_txrx_tx_ops.mgmt_tx_send = wma_mgmt_unified_cmd_send; |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_target_if_open() - Attach UMAC modules' interface with wmi layer |
| * @wma_handle: wma handle |
| * |
| * Separate module defines below functions: |
| * 1. tgt_wmi_<module>_<action> api sends wmi command, assigned to south bound |
| * tx_ops function pointers; |
| * 2. module's south dispatcher handles information from lower layer, assigned |
| * to south bound rx_ops function pointers; |
| * 3. wmi event handler deals with wmi event, extracts umac needed information, |
| * and call rx_ops(module's dispatcher). It executes in tasklet context and |
| * is up to dispatcher to decide the context to reside in tasklet or in |
| * thread context. |
| * |
| * Return: None |
| */ |
| static void wma_target_if_open(tp_wma_handle wma_handle) |
| { |
| struct wlan_objmgr_psoc *psoc = wma_handle->psoc; |
| |
| if (!psoc) |
| return; |
| |
| wlan_global_lmac_if_set_txops_registration_cb(WLAN_DEV_OL, |
| target_if_register_tx_ops); |
| wlan_lmac_if_set_umac_txops_registration_cb( |
| wma_register_tx_ops_handler); |
| wlan_global_lmac_if_open(psoc); |
| |
| } |
| |
| /** |
| * wma_target_if_close() - Detach UMAC modules' interface with wmi layer |
| * @wma_handle: wma handle |
| * |
| * Return: None |
| */ |
| static void wma_target_if_close(tp_wma_handle wma_handle) |
| { |
| struct wlan_objmgr_psoc *psoc = wma_handle->psoc; |
| |
| if (!psoc) |
| return; |
| |
| wlan_global_lmac_if_close(psoc); |
| } |
| |
| /** |
| * wma_get_psoc_from_scn_handle() - API to get psoc from scn handle |
| * @scn_handle: opaque wma handle |
| * |
| * API to get psoc from scn handle |
| * |
| * Return: None |
| */ |
| static struct wlan_objmgr_psoc *wma_get_psoc_from_scn_handle(void *scn_handle) |
| { |
| tp_wma_handle wma_handle; |
| |
| if (!scn_handle) { |
| WMA_LOGE("invalid scn handle"); |
| return NULL; |
| } |
| wma_handle = (tp_wma_handle)scn_handle; |
| |
| return wma_handle->psoc; |
| } |
| /** |
| * wma_legacy_service_ready_event_handler() - legacy (ext)service ready handler |
| * @event_id: event_id |
| * @handle: wma handle |
| * @event_data: event data |
| * @length: event length |
| * |
| * Return: 0 for success, negative error code for failure |
| */ |
| static int wma_legacy_service_ready_event_handler(uint32_t event_id, |
| void *handle, |
| uint8_t *event_data, |
| uint32_t length) |
| { |
| if (wmi_service_ready_event_id == event_id) |
| return wma_rx_service_ready_event(handle, event_data, length); |
| else if (wmi_service_ready_ext_event_id == event_id) |
| return wma_rx_service_ready_ext_event(handle, event_data, |
| length); |
| else |
| QDF_BUG(0); |
| |
| return 0; |
| } |
| |
| /** |
| * wma_flush_complete_evt_handler() - FW log flush complete event handler |
| * @handle: WMI handle |
| * @event: Event recevied from FW |
| * @len: Length of the event |
| * |
| */ |
| static int wma_flush_complete_evt_handler(void *handle, |
| u_int8_t *event, |
| u_int32_t len) |
| { |
| QDF_STATUS status; |
| tp_wma_handle wma = (tp_wma_handle) handle; |
| |
| WMI_DEBUG_MESG_FLUSH_COMPLETE_EVENTID_param_tlvs *param_buf; |
| wmi_debug_mesg_flush_complete_fixed_param *wmi_event; |
| wmi_debug_mesg_fw_data_stall_param *data_stall_event; |
| void *soc = cds_get_context(QDF_MODULE_ID_SOC); |
| uint8_t *buf_ptr; |
| uint32_t reason_code; |
| |
| param_buf = (WMI_DEBUG_MESG_FLUSH_COMPLETE_EVENTID_param_tlvs *) event; |
| if (!param_buf) { |
| WMA_LOGE("Invalid log flush complete event buffer"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| wmi_event = param_buf->fixed_param; |
| reason_code = wmi_event->reserved0; |
| |
| buf_ptr = (uint8_t *)wmi_event; |
| buf_ptr = buf_ptr + sizeof(wmi_debug_mesg_flush_complete_fixed_param) + |
| WMI_TLV_HDR_SIZE; |
| data_stall_event = (wmi_debug_mesg_fw_data_stall_param *) buf_ptr; |
| |
| if (((data_stall_event->tlv_header & 0xFFFF0000) >> 16 == |
| WMITLV_TAG_STRUC_wmi_debug_mesg_fw_data_stall_param)) { |
| /** |
| * Log data stall info received from FW: |
| * |
| * Possible data stall recovery types: |
| * WLAN_DBG_DATA_STALL_RECOVERY_CONNECT_DISCONNECT |
| * WLAN_DBG_DATA_STALL_RECOVERY_CONNECT_MAC_PHY_RESET |
| * WLAN_DBG_DATA_STALL_RECOVERY_CONNECT_PDR |
| * |
| * Possible data stall event types: |
| * WLAN_DBG_DATA_STALL_VDEV_PAUSE |
| * WLAN_DBG_DATA_STALL_HWSCHED_CMD_FILTER |
| * WLAN_DBG_DATA_STALL_HWSCHED_CMD_FLUSH |
| * WLAN_DBG_DATA_STALL_RX_REFILL_FAILED |
| * WLAN_DBG_DATA_STALL_RX_FCS_LEN_ERROR |
| * |
| * reason_code1: |
| * The information stored in reason_code1 varies based on the |
| * data stall type values: |
| * |
| * data_stall_type | reason_code1 |
| * ----------------------------------------------------- |
| * HWSCHED_CMD_FLUSH | flush req reason (0-40) |
| * RX_REFILL_FAILED | ring_id (0-7) |
| * RX_FCS_LEN_ERROR | exact error type |
| * |
| * reasone_code2: |
| * on which tid/hwq stall happened |
| * |
| */ |
| QDF_TRACE(QDF_MODULE_ID_WMA, QDF_TRACE_LEVEL_DEBUG, |
| "Data Stall event:"); |
| QDF_TRACE(QDF_MODULE_ID_WMA, QDF_TRACE_LEVEL_DEBUG, |
| "data_stall_type: %x vdev_id_bitmap: %x reason_code1: %x reason_code2: %x recovery_type: %x ", |
| data_stall_event->data_stall_type, |
| data_stall_event->vdev_id_bitmap, |
| data_stall_event->reason_code1, |
| data_stall_event->reason_code2, |
| data_stall_event->recovery_type); |
| |
| cdp_post_data_stall_event(soc, |
| DATA_STALL_LOG_INDICATOR_FIRMWARE, |
| data_stall_event->data_stall_type, |
| 0XFF, |
| data_stall_event->vdev_id_bitmap, |
| data_stall_event->recovery_type); |
| } |
| |
| /* |
| * reason_code = 0; Flush event in response to flush command |
| * reason_code = other value; Asynchronous flush event for fatal events |
| */ |
| if (!reason_code && (cds_is_log_report_in_progress() == false)) { |
| WMA_LOGD("Received WMI flush event without sending CMD"); |
| return -EINVAL; |
| } else if (!reason_code && cds_is_log_report_in_progress() == true) { |
| /* Flush event in response to flush command */ |
| WMA_LOGD("Received WMI flush event in response to flush CMD"); |
| status = qdf_mc_timer_stop(&wma->log_completion_timer); |
| if (status != QDF_STATUS_SUCCESS) |
| WMA_LOGE("Failed to stop the log completion timeout"); |
| cds_logging_set_fw_flush_complete(); |
| return QDF_STATUS_SUCCESS; |
| } else if (reason_code && cds_is_log_report_in_progress() == false) { |
| /* Asynchronous flush event for fatal events */ |
| status = cds_set_log_completion(WLAN_LOG_TYPE_FATAL, |
| WLAN_LOG_INDICATOR_FIRMWARE, |
| reason_code, false); |
| if (QDF_STATUS_SUCCESS != status) { |
| WMA_LOGE("%s: Failed to set log trigger params", |
| __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| cds_logging_set_fw_flush_complete(); |
| return status; |
| } else { |
| /* Asynchronous flush event for fatal event, |
| * but, report in progress already |
| */ |
| WMA_LOGD("%s: Bug report already in progress - dropping! type:%d, indicator=%d reason_code=%d", |
| __func__, WLAN_LOG_TYPE_FATAL, |
| WLAN_LOG_INDICATOR_FIRMWARE, reason_code); |
| return QDF_STATUS_E_FAILURE; |
| } |
| /* Asynchronous flush event for fatal event, |
| * but, report in progress already |
| */ |
| WMA_LOGW("%s: Bug report already in progress - dropping! type:%d, indicator=%d reason_code=%d", |
| __func__, WLAN_LOG_TYPE_FATAL, |
| WLAN_LOG_INDICATOR_FIRMWARE, reason_code); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| /** |
| * wma_unified_phyerr_rx_event_handler() - phyerr event handler |
| * @handle: wma handle |
| * @data: data buffer |
| * @datalen: buffer length |
| * |
| * WMI Handler for WMI_PHYERR_EVENTID event from firmware. |
| * This handler is currently handling DFS and spectral scan |
| * phy errors. |
| * |
| * Return: 0 for success, other value for failure |
| */ |
| static int wma_unified_phyerr_rx_event_handler(void *handle, |
| uint8_t *data, uint32_t datalen) { |
| /* phyerr handling is moved to cmn project |
| * As WIN still uses handler registration in non-cmn code. |
| * need complete testing of non offloaded DFS code before we enable |
| * it in cmn code. |
| **/ |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| void wma_vdev_init(struct wma_txrx_node *vdev) |
| { |
| qdf_wake_lock_create(&vdev->vdev_start_wakelock, "vdev_start"); |
| qdf_wake_lock_create(&vdev->vdev_stop_wakelock, "vdev_stop"); |
| qdf_wake_lock_create(&vdev->vdev_set_key_wakelock, "vdev_set_key"); |
| } |
| |
| void wma_vdev_deinit(struct wma_txrx_node *vdev) |
| { |
| qdf_wake_lock_destroy(&vdev->vdev_start_wakelock); |
| qdf_wake_lock_destroy(&vdev->vdev_stop_wakelock); |
| qdf_wake_lock_destroy(&vdev->vdev_set_key_wakelock); |
| } |
| |
| /** |
| * wma_rx_service_available_event() - service available event handler |
| * @handle: pointer to wma handle |
| * @cmd_param_info: Pointer to service available event TLVs |
| * @length: length of the event buffer received |
| * |
| * Return: zero on success, error code on failure |
| */ |
| static int wma_rx_service_available_event(void *handle, uint8_t *cmd_param_info, |
| uint32_t length) |
| { |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| WMI_SERVICE_AVAILABLE_EVENTID_param_tlvs *param_buf; |
| wmi_service_available_event_fixed_param *ev; |
| |
| param_buf = (WMI_SERVICE_AVAILABLE_EVENTID_param_tlvs *) cmd_param_info; |
| if (!(handle && param_buf)) { |
| WMA_LOGE("%s: Invalid arguments", __func__); |
| return -EINVAL; |
| } |
| |
| ev = param_buf->fixed_param; |
| if (!ev) { |
| WMA_LOGE("%s: Invalid buffer", __func__); |
| return -EINVAL; |
| } |
| |
| WMA_LOGD("WMA <-- WMI_SERVICE_AVAILABLE_EVENTID"); |
| |
| wma_handle->wmi_service_ext_offset = ev->wmi_service_segment_offset; |
| qdf_mem_copy(wma_handle->wmi_service_ext_bitmap, |
| &ev->wmi_service_segment_bitmap[0], |
| WMI_SERVICE_EXT_BM_SIZE32); |
| |
| return 0; |
| } |
| |
| /** |
| * wma_open() - Allocate wma context and initialize it. |
| * @psoc: Psoc pointer |
| * @wma_tgt_cfg_cb: tgt config callback fun |
| * @cds_cfg: mac parameters |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| QDF_STATUS wma_open(struct wlan_objmgr_psoc *psoc, |
| wma_tgt_cfg_cb tgt_cfg_cb, |
| struct cds_config_info *cds_cfg) |
| { |
| tp_wma_handle wma_handle; |
| HTC_HANDLE htc_handle; |
| qdf_device_t qdf_dev; |
| void *wmi_handle; |
| QDF_STATUS qdf_status; |
| struct wmi_rx_ops ops; |
| struct policy_mgr_wma_cbacks wma_cbacks; |
| struct target_psoc_info *tgt_psoc_info; |
| bool use_cookie = false; |
| int i; |
| void *cds_context; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| cds_context = cds_get_global_context(); |
| if (!cds_context) { |
| WMA_LOGE("%s: Invalid CDS context", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| g_wmi_version_info.major = __WMI_VER_MAJOR_; |
| g_wmi_version_info.minor = __WMI_VER_MINOR_; |
| g_wmi_version_info.revision = __WMI_REVISION_; |
| |
| qdf_dev = cds_get_context(QDF_MODULE_ID_QDF_DEVICE); |
| htc_handle = cds_get_context(QDF_MODULE_ID_HTC); |
| |
| if (!htc_handle) { |
| WMA_LOGE("%s: Invalid HTC handle", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* Alloc memory for WMA Context */ |
| qdf_status = cds_alloc_context(QDF_MODULE_ID_WMA, |
| (void **)&wma_handle, |
| sizeof(*wma_handle)); |
| |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: Memory allocation failed for wma_handle", |
| __func__); |
| return qdf_status; |
| } |
| |
| qdf_mem_zero(wma_handle, sizeof(t_wma_handle)); |
| |
| tgt_psoc_info = qdf_mem_malloc(sizeof(*tgt_psoc_info)); |
| if (!tgt_psoc_info) { |
| WMA_LOGE("%s: failed to allocate mem for tgt info", __func__); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| if (cds_get_conparam() != QDF_GLOBAL_FTM_MODE) { |
| #ifdef FEATURE_WLAN_EXTSCAN |
| qdf_wake_lock_create(&wma_handle->extscan_wake_lock, |
| "wlan_extscan_wl"); |
| #endif /* FEATURE_WLAN_EXTSCAN */ |
| qdf_wake_lock_create(&wma_handle->wow_wake_lock, |
| "wlan_wow_wl"); |
| qdf_wake_lock_create(&wma_handle->wow_auth_req_wl, |
| "wlan_auth_req_wl"); |
| qdf_wake_lock_create(&wma_handle->wow_assoc_req_wl, |
| "wlan_assoc_req_wl"); |
| qdf_wake_lock_create(&wma_handle->wow_deauth_rec_wl, |
| "wlan_deauth_rec_wl"); |
| qdf_wake_lock_create(&wma_handle->wow_disassoc_rec_wl, |
| "wlan_disassoc_rec_wl"); |
| qdf_wake_lock_create(&wma_handle->wow_ap_assoc_lost_wl, |
| "wlan_ap_assoc_lost_wl"); |
| qdf_wake_lock_create(&wma_handle->wow_auto_shutdown_wl, |
| "wlan_auto_shutdown_wl"); |
| qdf_wake_lock_create(&wma_handle->roam_ho_wl, |
| "wlan_roam_ho_wl"); |
| } |
| |
| qdf_status = wlan_objmgr_psoc_try_get_ref(psoc, WLAN_LEGACY_WMA_ID); |
| if (QDF_IS_STATUS_ERROR(qdf_status)) { |
| WMA_LOGE("%s: PSOC get_ref fails", __func__); |
| qdf_mem_free(tgt_psoc_info); |
| return qdf_status; |
| } |
| wma_handle->psoc = psoc; |
| |
| /* Open target_if layer and register wma callback */ |
| wma_target_if_open(wma_handle); |
| target_if_open(wma_get_psoc_from_scn_handle); |
| |
| /* Attach mc_thread context processing function */ |
| ops.wma_process_fw_event_handler_cbk = wma_process_fw_event_handler; |
| /* attach the wmi */ |
| wmi_handle = wmi_unified_attach(wma_handle, NULL, |
| WMI_TLV_TARGET, use_cookie, &ops, psoc); |
| if (!wmi_handle) { |
| WMA_LOGE("%s: failed to attach WMI", __func__); |
| qdf_status = QDF_STATUS_E_NOMEM; |
| goto err_wma_handle; |
| } |
| |
| target_if_register_legacy_service_ready_cb( |
| wma_legacy_service_ready_event_handler); |
| |
| WMA_LOGA("WMA --> wmi_unified_attach - success"); |
| |
| /* store the wmi handle in tgt_if_handle */ |
| tgt_psoc_info->wmi_handle = wmi_handle; |
| |
| wlan_psoc_obj_lock(psoc); |
| wlan_psoc_set_tgt_if_handle(psoc, tgt_psoc_info); |
| wlan_psoc_obj_unlock(psoc); |
| |
| wmi_unified_register_event_handler(wmi_handle, |
| WMI_SERVICE_AVAILABLE_EVENTID, |
| wma_rx_service_available_event, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wmi_handle, |
| WMI_SERVICE_READY_EVENTID, |
| init_deinit_service_ready_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wmi_handle, |
| WMI_SERVICE_READY_EXT_EVENTID, |
| init_deinit_service_ext_ready_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wmi_handle, |
| WMI_READY_EVENTID, |
| wma_rx_ready_event, |
| WMA_RX_SERIALIZER_CTX); |
| /* Save the WMI & HTC handle */ |
| wma_handle->wmi_handle = wmi_handle; |
| wma_handle->htc_handle = htc_handle; |
| wma_handle->cds_context = cds_context; |
| wma_handle->qdf_dev = qdf_dev; |
| wma_handle->max_scan = cds_cfg->max_scan; |
| |
| /* Initialize max_no_of_peers for wma_get_number_of_peers_supported() */ |
| wma_init_max_no_of_peers(wma_handle, cds_cfg->max_station); |
| /* Cap maxStation based on the target version */ |
| cds_cfg->max_station = wma_get_number_of_peers_supported(wma_handle); |
| /* Reinitialize max_no_of_peers based on the capped maxStation value */ |
| wma_init_max_no_of_peers(wma_handle, cds_cfg->max_station); |
| |
| /* initialize default target config */ |
| wma_set_default_tgt_config(wma_handle); |
| |
| wma_handle->tx_chain_mask_cck = cds_cfg->tx_chain_mask_cck; |
| wma_handle->self_gen_frm_pwr = cds_cfg->self_gen_frm_pwr; |
| |
| #if defined(QCA_WIFI_FTM) |
| if (cds_get_conparam() == QDF_GLOBAL_FTM_MODE) |
| wma_utf_attach(wma_handle); |
| #endif /* QCA_WIFI_FTM */ |
| wma_init_max_no_of_peers(wma_handle, cds_cfg->max_station); |
| cds_cfg->max_station = wma_get_number_of_peers_supported(wma_handle); |
| |
| cds_cfg->max_bssid = WMA_MAX_SUPPORTED_BSS; |
| |
| wma_handle->wlan_resource_config.num_wow_filters = |
| cds_cfg->max_wow_filters; |
| wma_handle->wlan_resource_config.num_keep_alive_pattern = |
| WMA_MAXNUM_PERIODIC_TX_PTRNS; |
| |
| /* The current firmware implementation requires the number of |
| * offload peers should be (number of vdevs + 1). |
| */ |
| wma_handle->wlan_resource_config.num_offload_peers = |
| cds_cfg->ap_maxoffload_peers + 1; |
| |
| wma_handle->wlan_resource_config.num_offload_reorder_buffs = |
| cds_cfg->ap_maxoffload_reorderbuffs + 1; |
| |
| wma_handle->ol_ini_info = cds_cfg->ol_ini_info; |
| wma_handle->max_station = cds_cfg->max_station; |
| wma_handle->max_bssid = cds_cfg->max_bssid; |
| wma_handle->driver_type = cds_cfg->driver_type; |
| wma_handle->ssdp = cds_cfg->ssdp; |
| wma_handle->enable_mc_list = cds_cfg->enable_mc_list; |
| wma_handle->bpf_packet_filter_enable = |
| cds_cfg->bpf_packet_filter_enable; |
| wma_handle->active_uc_bpf_mode = cds_cfg->active_uc_bpf_mode; |
| wma_handle->active_mc_bc_bpf_mode = cds_cfg->active_mc_bc_bpf_mode; |
| wma_handle->link_stats_results = NULL; |
| #ifdef FEATURE_WLAN_RA_FILTERING |
| wma_handle->IsRArateLimitEnabled = cds_cfg->is_ra_ratelimit_enabled; |
| wma_handle->RArateLimitInterval = cds_cfg->ra_ratelimit_interval; |
| #endif /* FEATURE_WLAN_RA_FILTERING */ |
| #ifdef WLAN_FEATURE_LPSS |
| wma_handle->is_lpass_enabled = cds_cfg->is_lpass_enabled; |
| #endif |
| wma_set_nan_enable(wma_handle, cds_cfg); |
| wma_handle->interfaces = qdf_mem_malloc(sizeof(struct wma_txrx_node) * |
| wma_handle->max_bssid); |
| if (!wma_handle->interfaces) { |
| WMA_LOGE("%s: failed to allocate interface table", __func__); |
| qdf_status = QDF_STATUS_E_NOMEM; |
| goto err_scn_context; |
| } |
| |
| for (i = 0; i < wma_handle->max_bssid; ++i) |
| wma_vdev_init(&wma_handle->interfaces[i]); |
| |
| /* Register the debug print event handler */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_DEBUG_PRINT_EVENTID, |
| wma_unified_debug_print_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| /* Register profiling event Handler */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_WLAN_PROFILE_DATA_EVENTID, |
| wma_profile_data_report_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| wma_handle->tgt_cfg_update_cb = tgt_cfg_cb; |
| wma_handle->old_hw_mode_index = WMA_DEFAULT_HW_MODE_INDEX; |
| wma_handle->new_hw_mode_index = WMA_DEFAULT_HW_MODE_INDEX; |
| wma_handle->saved_chan.num_channels = 0; |
| wma_handle->fw_timeout_crash = cds_cfg->fw_timeout_crash; |
| |
| qdf_status = qdf_event_create(&wma_handle->wma_ready_event); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: wma_ready_event initialization failed", __func__); |
| goto err_event_init; |
| } |
| |
| qdf_status = qdf_mc_timer_init(&wma_handle->service_ready_ext_timer, |
| QDF_TIMER_TYPE_SW, |
| wma_service_ready_ext_evt_timeout, |
| wma_handle); |
| if (!QDF_IS_STATUS_SUCCESS(qdf_status)) { |
| WMA_LOGE("Failed to initialize service ready ext timeout"); |
| goto err_event_init; |
| } |
| |
| qdf_status = qdf_event_create(&wma_handle->target_suspend); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: target suspend event initialization failed", |
| __func__); |
| goto err_event_init; |
| } |
| |
| /* Init Tx Frame Complete event */ |
| qdf_status = qdf_event_create(&wma_handle->tx_frm_download_comp_event); |
| if (!QDF_IS_STATUS_SUCCESS(qdf_status)) { |
| WMA_LOGE("%s: failed to init tx_frm_download_comp_event", |
| __func__); |
| goto err_event_init; |
| } |
| |
| /* Init tx queue empty check event */ |
| qdf_status = qdf_event_create(&wma_handle->tx_queue_empty_event); |
| if (!QDF_IS_STATUS_SUCCESS(qdf_status)) { |
| WMA_LOGE("%s: failed to init tx_queue_empty_event", __func__); |
| goto err_event_init; |
| } |
| |
| qdf_status = qdf_event_create(&wma_handle->wma_resume_event); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: wma_resume_event initialization failed", |
| __func__); |
| goto err_event_init; |
| } |
| |
| qdf_status = cds_shutdown_notifier_register(wma_shutdown_notifier_cb, |
| wma_handle); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: Shutdown notifier register failed: %d", |
| __func__, qdf_status); |
| goto err_event_init; |
| } |
| |
| qdf_status = qdf_event_create(&wma_handle->runtime_suspend); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: runtime_suspend event initialization failed", |
| __func__); |
| goto err_event_init; |
| } |
| |
| qdf_status = qdf_event_create(&wma_handle->recovery_event); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: recovery event initialization failed", __func__); |
| goto err_event_init; |
| } |
| |
| qdf_list_create(&wma_handle->vdev_resp_queue, |
| MAX_ENTRY_VDEV_RESP_QUEUE); |
| qdf_spinlock_create(&wma_handle->vdev_respq_lock); |
| qdf_list_create(&wma_handle->wma_hold_req_queue, |
| MAX_ENTRY_HOLD_REQ_QUEUE); |
| qdf_spinlock_create(&wma_handle->wma_hold_req_q_lock); |
| qdf_atomic_init(&wma_handle->is_wow_bus_suspended); |
| qdf_atomic_init(&wma_handle->scan_id_counter); |
| |
| /* Register vdev start response event handler */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_VDEV_START_RESP_EVENTID, |
| wma_vdev_start_resp_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| /* Register vdev stop response event handler */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_VDEV_STOPPED_EVENTID, |
| wma_vdev_stop_resp_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| /* register for STA kickout function */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PEER_STA_KICKOUT_EVENTID, |
| wma_peer_sta_kickout_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| /* register for stats response event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_UPDATE_STATS_EVENTID, |
| wma_stats_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| /* register for stats response event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_VDEV_GET_ARP_STAT_EVENTID, |
| wma_get_arp_stats_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| /* register for peer info response event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PEER_STATS_INFO_EVENTID, |
| wma_peer_info_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| #ifdef WLAN_POWER_DEBUGFS |
| /* register for Chip Power stats event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PDEV_CHIP_POWER_STATS_EVENTID, |
| wma_unified_power_debug_stats_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| #endif |
| |
| /* register for linkspeed response event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PEER_ESTIMATED_LINKSPEED_EVENTID, |
| wma_link_speed_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| #ifdef FEATURE_OEM_DATA_SUPPORT |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_OEM_RESPONSE_EVENTID, |
| wma_oem_data_response_handler, |
| WMA_RX_SERIALIZER_CTX); |
| #endif /* FEATURE_OEM_DATA_SUPPORT */ |
| |
| /* Register peer change event handler */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PEER_STATE_EVENTID, |
| wma_peer_state_change_event_handler, |
| WMA_RX_WORK_CTX); |
| |
| /* Register beacon tx complete event id. The event is required |
| * for sending channel switch announcement frames |
| */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_OFFLOAD_BCN_TX_STATUS_EVENTID, |
| wma_unified_bcntx_status_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_UPDATE_VDEV_RATE_STATS_EVENTID, |
| wma_link_status_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| #ifdef WLAN_FEATURE_LINK_LAYER_STATS |
| /* Register event handler for processing Link Layer Stats |
| * response from the FW |
| */ |
| wma_register_ll_stats_event_handler(wma_handle); |
| |
| #endif /* WLAN_FEATURE_LINK_LAYER_STATS */ |
| |
| /* |
| * Register event handler to receive firmware mem dump |
| * copy complete indication |
| */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_UPDATE_FW_MEM_DUMP_EVENTID, |
| wma_fw_mem_dump_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| wmi_set_tgt_assert(wma_handle->wmi_handle, |
| cds_cfg->force_target_assert_enabled); |
| /* Firmware debug log */ |
| qdf_status = dbglog_init(wma_handle->wmi_handle); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: Firmware Dbglog initialization failed", __func__); |
| goto err_dbglog_init; |
| } |
| |
| /* |
| * Update Powersave mode |
| * 1 - Legacy Powersave + Deepsleep Disabled |
| * 2 - QPower + Deepsleep Disabled |
| * 3 - Legacy Powersave + Deepsleep Enabled |
| * 4 - QPower + Deepsleep Enabled |
| */ |
| wma_handle->powersave_mode = cds_cfg->powersave_offload_enabled; |
| wma_handle->staMaxLIModDtim = cds_cfg->sta_maxlimod_dtim; |
| wma_handle->staModDtim = cds_cfg->sta_mod_dtim; |
| wma_handle->staDynamicDtim = cds_cfg->sta_dynamic_dtim; |
| |
| /* |
| * Value of cds_cfg->wow_enable can be, |
| * 0 - Disable both magic pattern match and pattern byte match. |
| * 1 - Enable magic pattern match on all interfaces. |
| * 2 - Enable pattern byte match on all interfaces. |
| * 3 - Enable both magic patter and pattern byte match on |
| * all interfaces. |
| */ |
| wma_handle->wow.magic_ptrn_enable = |
| (cds_cfg->wow_enable & 0x01) ? true : false; |
| wma_handle->ptrn_match_enable_all_vdev = |
| (cds_cfg->wow_enable & 0x02) ? true : false; |
| |
| /* register for install key completion event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_VDEV_INSTALL_KEY_COMPLETE_EVENTID, |
| wma_vdev_install_key_complete_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| #ifdef WLAN_FEATURE_NAN |
| /* register for nan response event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_NAN_EVENTID, |
| wma_nan_rsp_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| #endif /* WLAN_FEATURE_NAN */ |
| |
| #ifdef WLAN_FEATURE_STATS_EXT |
| /* register for extended stats event */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_STATS_EXT_EVENTID, |
| wma_stats_ext_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| #endif /* WLAN_FEATURE_STATS_EXT */ |
| #ifdef FEATURE_WLAN_EXTSCAN |
| wma_register_extscan_event_handler(wma_handle); |
| #endif /* WLAN_FEATURE_STATS_EXT */ |
| |
| WMA_LOGD("%s: Exit", __func__); |
| |
| #ifdef WLAN_FEATURE_ROAM_OFFLOAD |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_ROAM_SYNCH_EVENTID, |
| wma_roam_synch_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_ROAM_SYNCH_FRAME_EVENTID, |
| wma_roam_synch_frame_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| #endif /* WLAN_FEATURE_ROAM_OFFLOAD */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_RSSI_BREACH_EVENTID, |
| wma_rssi_breached_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| qdf_wake_lock_create(&wma_handle->wmi_cmd_rsp_wake_lock, |
| "wlan_fw_rsp_wakelock"); |
| qdf_runtime_lock_init(&wma_handle->wmi_cmd_rsp_runtime_lock); |
| |
| /* Register peer assoc conf event handler */ |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PEER_ASSOC_CONF_EVENTID, |
| wma_peer_assoc_conf_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_VDEV_DELETE_RESP_EVENTID, |
| wma_vdev_delete_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PEER_DELETE_RESP_EVENTID, |
| wma_peer_delete_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_BPF_CAPABILIY_INFO_EVENTID, |
| wma_get_bpf_caps_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_VDEV_ENCRYPT_DECRYPT_DATA_RESP_EVENTID, |
| wma_encrypt_decrypt_msg_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_CHAN_INFO_EVENTID, |
| wma_chan_info_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_DEBUG_MESG_FLUSH_COMPLETE_EVENTID, |
| wma_flush_complete_evt_handler, |
| WMA_RX_WORK_CTX); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_REPORT_RX_AGGR_FAILURE_EVENTID, |
| wma_rx_aggr_failure_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| |
| wma_handle->ito_repeat_count = cds_cfg->ito_repeat_count; |
| wma_handle->bandcapability = cds_cfg->bandcapability; |
| |
| /* Register PWR_SAVE_FAIL event only in case of recovery(1) */ |
| if (cds_cfg->auto_power_save_fail_mode == |
| PMO_FW_TO_SEND_WOW_IND_ON_PWR_FAILURE) { |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PDEV_CHIP_POWER_SAVE_FAILURE_DETECTED_EVENTID, |
| wma_chip_power_save_failure_detected_handler, |
| WMA_RX_WORK_CTX); |
| } |
| |
| wma_ndp_register_all_event_handlers(wma_handle); |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PEER_ANTDIV_INFO_EVENTID, |
| wma_peer_ant_info_evt_handler, |
| WMA_RX_WORK_CTX); |
| |
| |
| wma_register_debug_callback(); |
| /* Register callback with PMO so PMO can update the vdev pause bitmap*/ |
| pmo_register_pause_bitmap_notifier(wma_handle->psoc, |
| wma_vdev_update_pause_bitmap); |
| pmo_register_get_pause_bitmap(wma_handle->psoc, |
| wma_vdev_get_pause_bitmap); |
| pmo_register_is_device_in_low_pwr_mode(wma_handle->psoc, |
| wma_vdev_is_device_in_low_pwr_mode); |
| wma_cbacks.wma_get_connection_info = wma_get_connection_info; |
| wma_cbacks.wma_is_service_enabled = wma_is_service_enabled; |
| qdf_status = policy_mgr_register_wma_cb(wma_handle->psoc, &wma_cbacks); |
| if (!QDF_IS_STATUS_SUCCESS(qdf_status)) { |
| WMA_LOGE("Failed to register wma cb with Policy Manager"); |
| } |
| |
| wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PHYERR_EVENTID, |
| wma_unified_phyerr_rx_event_handler, |
| WMA_RX_WORK_CTX); |
| |
| return QDF_STATUS_SUCCESS; |
| |
| err_dbglog_init: |
| qdf_wake_lock_destroy(&wma_handle->wmi_cmd_rsp_wake_lock); |
| qdf_runtime_lock_deinit(&wma_handle->wmi_cmd_rsp_runtime_lock); |
| qdf_spinlock_destroy(&wma_handle->vdev_respq_lock); |
| qdf_spinlock_destroy(&wma_handle->wma_hold_req_q_lock); |
| err_event_init: |
| wmi_unified_unregister_event_handler(wma_handle->wmi_handle, |
| WMI_DEBUG_PRINT_EVENTID); |
| |
| for (i = 0; i < wma_handle->max_bssid; ++i) |
| wma_vdev_deinit(&wma_handle->interfaces[i]); |
| |
| qdf_mem_free(wma_handle->interfaces); |
| |
| err_scn_context: |
| #if defined(QCA_WIFI_FTM) |
| wma_utf_detach(wma_handle); |
| #endif /* QCA_WIFI_FTM */ |
| qdf_mem_free(((p_cds_contextType) cds_context)->cfg_ctx); |
| OS_FREE(wmi_handle); |
| |
| err_wma_handle: |
| qdf_mem_free(tgt_psoc_info); |
| if (cds_get_conparam() != QDF_GLOBAL_FTM_MODE) { |
| #ifdef FEATURE_WLAN_EXTSCAN |
| qdf_wake_lock_destroy(&wma_handle->extscan_wake_lock); |
| #endif /* FEATURE_WLAN_EXTSCAN */ |
| qdf_wake_lock_destroy(&wma_handle->wow_wake_lock); |
| qdf_wake_lock_destroy(&wma_handle->wow_auth_req_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_assoc_req_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_deauth_rec_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_disassoc_rec_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_ap_assoc_lost_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_auto_shutdown_wl); |
| qdf_wake_lock_destroy(&wma_handle->roam_ho_wl); |
| } |
| |
| cds_free_context(QDF_MODULE_ID_WMA, wma_handle); |
| |
| WMA_LOGD("%s: Exit", __func__); |
| |
| return qdf_status; |
| } |
| |
| /** |
| * wma_pre_start() - wma pre start |
| * |
| * Return: 0 on success, errno on failure |
| */ |
| QDF_STATUS wma_pre_start(void) |
| { |
| QDF_STATUS qdf_status = QDF_STATUS_SUCCESS; |
| A_STATUS status = A_OK; |
| tp_wma_handle wma_handle; |
| struct scheduler_msg wma_msg = { 0 }; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| /* Validate the wma_handle */ |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: invalid argument", __func__); |
| qdf_status = QDF_STATUS_E_INVAL; |
| goto end; |
| } |
| /* Open endpoint for ctrl path - WMI <--> HTC */ |
| status = wmi_unified_connect_htc_service(wma_handle->wmi_handle, |
| wma_handle->htc_handle); |
| if (A_OK != status) { |
| WMA_LOGE("%s: wmi_unified_connect_htc_service", __func__); |
| |
| if (!cds_is_fw_down()) |
| QDF_BUG(0); |
| |
| qdf_status = QDF_STATUS_E_FAULT; |
| goto end; |
| } |
| |
| WMA_LOGD("WMA --> wmi_unified_connect_htc_service - success"); |
| |
| /* Trigger the CFG DOWNLOAD */ |
| wma_msg.type = WNI_CFG_DNLD_REQ; |
| wma_msg.bodyptr = NULL; |
| wma_msg.bodyval = 0; |
| |
| qdf_status = scheduler_post_msg(QDF_MODULE_ID_WMA, &wma_msg); |
| if (QDF_STATUS_SUCCESS != qdf_status) { |
| WMA_LOGE("%s: Failed to post WNI_CFG_DNLD_REQ msg", __func__); |
| QDF_ASSERT(0); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| } |
| end: |
| WMA_LOGD("%s: Exit", __func__); |
| return qdf_status; |
| } |
| |
| void wma_send_msg_by_priority(tp_wma_handle wma_handle, uint16_t msg_type, |
| void *body_ptr, uint32_t body_val, bool is_high_priority) |
| { |
| struct scheduler_msg msg = {0}; |
| QDF_STATUS status; |
| |
| msg.type = msg_type; |
| msg.bodyval = body_val; |
| msg.bodyptr = body_ptr; |
| |
| status = scheduler_post_msg_by_priority(QDF_MODULE_ID_PE, |
| &msg, is_high_priority); |
| if (!QDF_IS_STATUS_SUCCESS(status)) { |
| WMA_LOGE("Failed to post msg %d to PE", msg_type); |
| if (body_ptr) |
| qdf_mem_free(body_ptr); |
| } |
| } |
| |
| |
| void wma_send_msg(tp_wma_handle wma_handle, uint16_t msg_type, |
| void *body_ptr, uint32_t body_val) |
| { |
| wma_send_msg_by_priority(wma_handle, msg_type, |
| body_ptr, body_val, false); |
| } |
| |
| void wma_send_msg_high_priority(tp_wma_handle wma_handle, uint16_t msg_type, |
| void *body_ptr, uint32_t body_val) |
| { |
| wma_send_msg_by_priority(wma_handle, msg_type, |
| body_ptr, body_val, true); |
| } |
| |
| /** |
| * wma_set_base_macaddr_indicate() - set base mac address in fw |
| * @wma_handle: wma handle |
| * @customAddr: base mac address |
| * |
| * Return: 0 for success or error code |
| */ |
| static int wma_set_base_macaddr_indicate(tp_wma_handle wma_handle, |
| tSirMacAddr *customAddr) |
| { |
| int err; |
| |
| err = wmi_unified_set_base_macaddr_indicate_cmd(wma_handle->wmi_handle, |
| (uint8_t *)customAddr); |
| if (err) |
| return -EIO; |
| WMA_LOGD("Base MAC Addr: " MAC_ADDRESS_STR, |
| MAC_ADDR_ARRAY((*customAddr))); |
| |
| return 0; |
| } |
| |
| /** |
| * wma_log_supported_evt_handler() - Enable/Disable FW diag/log events |
| * @handle: WMA handle |
| * @event: Event received from FW |
| * @len: Length of the event |
| * |
| * Enables the low frequency events and disables the high frequency |
| * events. Bit 17 indicates if the event if low/high frequency. |
| * 1 - high frequency, 0 - low frequency |
| * |
| * Return: 0 on successfully enabling/disabling the events |
| */ |
| static int wma_log_supported_evt_handler(void *handle, |
| uint8_t *event, |
| uint32_t len) |
| { |
| tp_wma_handle wma = (tp_wma_handle) handle; |
| |
| if (wmi_unified_log_supported_evt_cmd(wma->wmi_handle, |
| event, len)) |
| return -EINVAL; |
| |
| return 0; |
| } |
| |
| /** |
| * wma_pdev_set_hw_mode_resp_evt_handler() - Set HW mode resp evt handler |
| * @handle: WMI handle |
| * @event: Event recevied from FW |
| * @len: Length of the event |
| * |
| * Event handler for WMI_PDEV_SET_HW_MODE_RESP_EVENTID that is sent to host |
| * driver in response to a WMI_PDEV_SET_HW_MODE_CMDID being sent to WLAN |
| * firmware |
| * |
| * Return: QDF_STATUS |
| */ |
| static int wma_pdev_set_hw_mode_resp_evt_handler(void *handle, |
| uint8_t *event, |
| uint32_t len) |
| { |
| WMI_PDEV_SET_HW_MODE_RESP_EVENTID_param_tlvs *param_buf; |
| wmi_pdev_set_hw_mode_response_event_fixed_param *wmi_event; |
| wmi_pdev_set_hw_mode_response_vdev_mac_entry *vdev_mac_entry; |
| uint32_t i; |
| struct sir_set_hw_mode_resp *hw_mode_resp; |
| tp_wma_handle wma = (tp_wma_handle) handle; |
| |
| if (!wma) { |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| /* Since WMA handle itself is NULL, we cannot send fail |
| * response back to LIM here |
| */ |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| wma_release_wakelock(&wma->wmi_cmd_rsp_wake_lock); |
| wma_remove_req(wma, 0, WMA_PDEV_SET_HW_MODE_RESP); |
| |
| hw_mode_resp = qdf_mem_malloc(sizeof(*hw_mode_resp)); |
| if (!hw_mode_resp) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| /* Since this memory allocation itself failed, we cannot |
| * send fail response back to LIM here |
| */ |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| param_buf = (WMI_PDEV_SET_HW_MODE_RESP_EVENTID_param_tlvs *) event; |
| if (!param_buf) { |
| WMA_LOGE("Invalid WMI_PDEV_SET_HW_MODE_RESP_EVENTID event"); |
| /* Need to send response back to upper layer to free |
| * active command list |
| */ |
| goto fail; |
| } |
| if (param_buf->fixed_param->num_vdev_mac_entries >= |
| MAX_VDEV_SUPPORTED) { |
| WMA_LOGE("num_vdev_mac_entries crossed max value"); |
| goto fail; |
| } |
| |
| wmi_event = param_buf->fixed_param; |
| hw_mode_resp->status = wmi_event->status; |
| hw_mode_resp->cfgd_hw_mode_index = wmi_event->cfgd_hw_mode_index; |
| hw_mode_resp->num_vdev_mac_entries = wmi_event->num_vdev_mac_entries; |
| |
| WMA_LOGE("%s: status:%d cfgd_hw_mode_index:%d num_vdev_mac_entries:%d", |
| __func__, wmi_event->status, |
| wmi_event->cfgd_hw_mode_index, |
| wmi_event->num_vdev_mac_entries); |
| vdev_mac_entry = |
| param_buf->wmi_pdev_set_hw_mode_response_vdev_mac_mapping; |
| |
| /* Store the vdev-mac map in WMA and prepare to send to PE */ |
| for (i = 0; i < wmi_event->num_vdev_mac_entries; i++) { |
| uint32_t vdev_id, mac_id, pdev_id; |
| |
| vdev_id = vdev_mac_entry[i].vdev_id; |
| pdev_id = vdev_mac_entry[i].pdev_id; |
| if (pdev_id == WMI_PDEV_ID_SOC) { |
| WMA_LOGE("%s: soc level id received for mac id)", |
| __func__); |
| QDF_BUG(0); |
| goto fail; |
| } |
| mac_id = WMA_PDEV_TO_MAC_MAP(vdev_mac_entry[i].pdev_id); |
| |
| WMA_LOGE("%s: vdev_id:%d mac_id:%d", |
| __func__, vdev_id, mac_id); |
| |
| hw_mode_resp->vdev_mac_map[i].vdev_id = vdev_id; |
| hw_mode_resp->vdev_mac_map[i].mac_id = mac_id; |
| wma_update_intf_hw_mode_params(vdev_id, mac_id, |
| wmi_event->cfgd_hw_mode_index); |
| } |
| |
| if (hw_mode_resp->status == SET_HW_MODE_STATUS_OK) { |
| if (WMA_DEFAULT_HW_MODE_INDEX == wma->new_hw_mode_index) { |
| wma->new_hw_mode_index = wmi_event->cfgd_hw_mode_index; |
| } else { |
| wma->old_hw_mode_index = wma->new_hw_mode_index; |
| wma->new_hw_mode_index = wmi_event->cfgd_hw_mode_index; |
| } |
| policy_mgr_update_hw_mode_index(wma->psoc, |
| wmi_event->cfgd_hw_mode_index); |
| } |
| |
| WMA_LOGE("%s: Updated: old_hw_mode_index:%d new_hw_mode_index:%d", |
| __func__, wma->old_hw_mode_index, wma->new_hw_mode_index); |
| |
| wma_send_msg(wma, SIR_HAL_PDEV_SET_HW_MODE_RESP, |
| (void *) hw_mode_resp, 0); |
| |
| return QDF_STATUS_SUCCESS; |
| |
| fail: |
| WMA_LOGE("%s: Sending fail response to LIM", __func__); |
| hw_mode_resp->status = SET_HW_MODE_STATUS_ECANCELED; |
| hw_mode_resp->cfgd_hw_mode_index = 0; |
| hw_mode_resp->num_vdev_mac_entries = 0; |
| wma_send_msg(wma, SIR_HAL_PDEV_SET_HW_MODE_RESP, |
| (void *) hw_mode_resp, 0); |
| |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| /** |
| * wma_process_pdev_hw_mode_trans_ind() - Process HW mode transition info |
| * |
| * @handle: WMA handle |
| * @fixed_param: Event fixed parameters |
| * @vdev_mac_entry - vdev mac entry |
| * @hw_mode_trans_ind - Buffer to store parsed information |
| * |
| * Parses fixed_param, vdev_mac_entry and fills in the information into |
| * hw_mode_trans_ind and wma |
| * |
| * Return: None |
| */ |
| void wma_process_pdev_hw_mode_trans_ind(void *handle, |
| wmi_pdev_hw_mode_transition_event_fixed_param *fixed_param, |
| wmi_pdev_set_hw_mode_response_vdev_mac_entry *vdev_mac_entry, |
| struct sir_hw_mode_trans_ind *hw_mode_trans_ind) |
| { |
| uint32_t i; |
| tp_wma_handle wma = (tp_wma_handle) handle; |
| |
| hw_mode_trans_ind->old_hw_mode_index = fixed_param->old_hw_mode_index; |
| hw_mode_trans_ind->new_hw_mode_index = fixed_param->new_hw_mode_index; |
| hw_mode_trans_ind->num_vdev_mac_entries = |
| fixed_param->num_vdev_mac_entries; |
| WMA_LOGE("%s: old_hw_mode_index:%d new_hw_mode_index:%d entries=%d", |
| __func__, fixed_param->old_hw_mode_index, |
| fixed_param->new_hw_mode_index, |
| fixed_param->num_vdev_mac_entries); |
| |
| /* Store the vdev-mac map in WMA and send to policy manager */ |
| for (i = 0; i < fixed_param->num_vdev_mac_entries; i++) { |
| uint32_t vdev_id, mac_id, pdev_id; |
| |
| vdev_id = vdev_mac_entry[i].vdev_id; |
| pdev_id = vdev_mac_entry[i].pdev_id; |
| |
| if (pdev_id == WMI_PDEV_ID_SOC) { |
| WMA_LOGE("%s: soc level id received for mac id)", |
| __func__); |
| QDF_BUG(0); |
| return; |
| } |
| |
| mac_id = WMA_PDEV_TO_MAC_MAP(vdev_mac_entry[i].pdev_id); |
| |
| WMA_LOGE("%s: vdev_id:%d mac_id:%d", |
| __func__, vdev_id, mac_id); |
| |
| hw_mode_trans_ind->vdev_mac_map[i].vdev_id = vdev_id; |
| hw_mode_trans_ind->vdev_mac_map[i].mac_id = mac_id; |
| wma_update_intf_hw_mode_params(vdev_id, mac_id, |
| fixed_param->new_hw_mode_index); |
| } |
| wma->old_hw_mode_index = fixed_param->old_hw_mode_index; |
| wma->new_hw_mode_index = fixed_param->new_hw_mode_index; |
| policy_mgr_update_new_hw_mode_index(wma->psoc, |
| fixed_param->new_hw_mode_index); |
| policy_mgr_update_old_hw_mode_index(wma->psoc, |
| fixed_param->old_hw_mode_index); |
| |
| WMA_LOGE("%s: Updated: old_hw_mode_index:%d new_hw_mode_index:%d", |
| __func__, wma->old_hw_mode_index, wma->new_hw_mode_index); |
| } |
| |
| /** |
| * wma_pdev_hw_mode_transition_evt_handler() - HW mode transition evt handler |
| * @handle: WMI handle |
| * @event: Event recevied from FW |
| * @len: Length of the event |
| * |
| * Event handler for WMI_PDEV_HW_MODE_TRANSITION_EVENTID that indicates an |
| * asynchronous hardware mode transition. This event notifies the host driver |
| * that firmware independently changed the hardware mode for some reason, such |
| * as Coex, LFR 3.0, etc |
| * |
| * Return: Success on receiving valid params from FW |
| */ |
| static int wma_pdev_hw_mode_transition_evt_handler(void *handle, |
| uint8_t *event, |
| uint32_t len) |
| { |
| WMI_PDEV_HW_MODE_TRANSITION_EVENTID_param_tlvs *param_buf; |
| wmi_pdev_hw_mode_transition_event_fixed_param *wmi_event; |
| wmi_pdev_set_hw_mode_response_vdev_mac_entry *vdev_mac_entry; |
| struct sir_hw_mode_trans_ind *hw_mode_trans_ind; |
| tp_wma_handle wma = (tp_wma_handle) handle; |
| |
| if (!wma) { |
| /* This is an async event. So, not sending any event to LIM */ |
| WMA_LOGE("Invalid WMA handle"); |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| param_buf = (WMI_PDEV_HW_MODE_TRANSITION_EVENTID_param_tlvs *) event; |
| if (!param_buf) { |
| /* This is an async event. So, not sending any event to LIM */ |
| WMA_LOGE("Invalid WMI_PDEV_HW_MODE_TRANSITION_EVENTID event"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| if (param_buf->fixed_param->num_vdev_mac_entries > MAX_VDEV_SUPPORTED) { |
| WMA_LOGE("num_vdev_mac_entries: %d crossed max value: %d", |
| param_buf->fixed_param->num_vdev_mac_entries, |
| MAX_VDEV_SUPPORTED); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| hw_mode_trans_ind = qdf_mem_malloc(sizeof(*hw_mode_trans_ind)); |
| if (!hw_mode_trans_ind) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| wmi_event = param_buf->fixed_param; |
| vdev_mac_entry = |
| param_buf->wmi_pdev_set_hw_mode_response_vdev_mac_mapping; |
| wma_process_pdev_hw_mode_trans_ind(wma, wmi_event, vdev_mac_entry, |
| hw_mode_trans_ind); |
| /* Pass the message to PE */ |
| wma_send_msg(wma, SIR_HAL_PDEV_HW_MODE_TRANS_IND, |
| (void *) hw_mode_trans_ind, 0); |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_pdev_set_dual_mode_config_resp_evt_handler() - Dual mode evt handler |
| * @handle: WMI handle |
| * @event: Event received from FW |
| * @len: Length of the event |
| * |
| * Notifies the host driver of the completion or failure of a |
| * WMI_PDEV_SET_MAC_CONFIG_CMDID command. This event would be returned to |
| * the host driver once the firmware has completed a reconfiguration of the Scan |
| * and FW mode configuration. This changes could include entering or leaving a |
| * dual mac configuration for either scan and/or more permanent firmware mode. |
| * |
| * Return: Success on receiving valid params from FW |
| */ |
| static int wma_pdev_set_dual_mode_config_resp_evt_handler(void *handle, |
| uint8_t *event, |
| uint32_t len) |
| { |
| WMI_PDEV_SET_MAC_CONFIG_RESP_EVENTID_param_tlvs *param_buf; |
| wmi_pdev_set_mac_config_response_event_fixed_param *wmi_event; |
| tp_wma_handle wma = (tp_wma_handle) handle; |
| struct sir_dual_mac_config_resp *dual_mac_cfg_resp; |
| |
| if (!wma) { |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| /* Since the WMA handle is NULL, we cannot send resp to LIM. |
| * So, returning from here. |
| */ |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| wma_release_wakelock(&wma->wmi_cmd_rsp_wake_lock); |
| dual_mac_cfg_resp = qdf_mem_malloc(sizeof(*dual_mac_cfg_resp)); |
| if (!dual_mac_cfg_resp) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| /* Since the mem alloc failed, we cannot send resp to LIM. |
| * So, returning from here. |
| */ |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| param_buf = (WMI_PDEV_SET_MAC_CONFIG_RESP_EVENTID_param_tlvs *) |
| event; |
| if (!param_buf) { |
| WMA_LOGE("%s: Invalid event", __func__); |
| goto fail; |
| } |
| |
| wmi_event = param_buf->fixed_param; |
| WMA_LOGD("%s: status:%d", __func__, wmi_event->status); |
| dual_mac_cfg_resp->status = wmi_event->status; |
| |
| if (SET_HW_MODE_STATUS_OK == dual_mac_cfg_resp->status) { |
| policy_mgr_update_dbs_scan_config(wma->psoc); |
| policy_mgr_update_dbs_fw_config(wma->psoc); |
| } |
| |
| /* Pass the message to PE */ |
| wma_send_msg(wma, SIR_HAL_PDEV_MAC_CFG_RESP, |
| (void *) dual_mac_cfg_resp, 0); |
| |
| return QDF_STATUS_SUCCESS; |
| |
| fail: |
| WMA_LOGE("%s: Sending fail response to LIM", __func__); |
| dual_mac_cfg_resp->status = SET_HW_MODE_STATUS_ECANCELED; |
| wma_send_msg(wma, SIR_HAL_PDEV_MAC_CFG_RESP, |
| (void *) dual_mac_cfg_resp, 0); |
| |
| return QDF_STATUS_E_FAILURE; |
| |
| } |
| |
| /** |
| * wma_start() - wma start function. |
| * Intialize event handlers and timers. |
| * |
| * Return: 0 on success, QDF Error on failure |
| */ |
| QDF_STATUS wma_start(void) |
| { |
| QDF_STATUS qdf_status = QDF_STATUS_SUCCESS; |
| tp_wma_handle wma_handle; |
| int status; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| /* validate the wma_handle */ |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid handle", __func__); |
| qdf_status = QDF_STATUS_E_INVAL; |
| goto end; |
| } |
| |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_ROAM_EVENTID, |
| wma_roam_event_callback, |
| WMA_RX_SERIALIZER_CTX); |
| if (0 != status) { |
| WMA_LOGE("%s: Failed to register Roam callback", __func__); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_WOW_WAKEUP_HOST_EVENTID, |
| wma_wow_wakeup_host_event, |
| WMA_RX_TASKLET_CTX); |
| if (status) { |
| WMA_LOGE("%s: Failed to register wow wakeup host event handler", |
| __func__); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PDEV_RESUME_EVENTID, |
| wma_pdev_resume_event_handler, |
| WMA_RX_TASKLET_CTX); |
| if (status) { |
| WMA_LOGE("%s: Failed to register PDEV resume event handler", |
| __func__); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| #if defined(QCA_LL_LEGACY_TX_FLOW_CONTROL) || \ |
| defined(QCA_LL_TX_FLOW_CONTROL_V2) || defined(CONFIG_HL_SUPPORT) |
| WMA_LOGD("MCC TX Pause Event Handler register"); |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_TX_PAUSE_EVENTID, |
| wma_mcc_vdev_tx_pause_evt_handler, |
| WMA_RX_TASKLET_CTX); |
| #endif /* QCA_LL_LEGACY_TX_FLOW_CONTROL */ |
| |
| #ifdef FEATURE_WLAN_AUTO_SHUTDOWN |
| WMA_LOGD("Registering auto shutdown handler"); |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_HOST_AUTO_SHUTDOWN_EVENTID, |
| wma_auto_shutdown_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register WMI Auto shutdown event handler"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| #endif /* FEATURE_WLAN_AUTO_SHUTDOWN */ |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_THERMAL_MGMT_EVENTID, |
| wma_thermal_mgmt_evt_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register thermal mitigation event cb"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| status = wma_ocb_register_event_handlers(wma_handle); |
| if (status) { |
| WMA_LOGE("Failed to register ocb event handlers"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| qdf_status = QDF_STATUS_SUCCESS; |
| |
| #ifdef QCA_WIFI_FTM |
| /* |
| * Tx mgmt attach requires TXRX context which is not created |
| * in FTM mode. So skip the TX mgmt attach. |
| */ |
| if (cds_get_conparam() == QDF_GLOBAL_FTM_MODE) |
| goto end; |
| #endif /* QCA_WIFI_FTM */ |
| |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_RMC)) { |
| |
| WMA_LOGD("FW supports cesium network, registering event handlers"); |
| |
| status = wmi_unified_register_event_handler( |
| wma_handle->wmi_handle, |
| WMI_PEER_INFO_EVENTID, |
| wma_ibss_peer_info_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register ibss peer info event cb"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| status = wmi_unified_register_event_handler( |
| wma_handle->wmi_handle, |
| WMI_PEER_TX_FAIL_CNT_THR_EVENTID, |
| wma_fast_tx_fail_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register peer fast tx failure event cb"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| } else { |
| WMA_LOGE("Target does not support cesium network"); |
| } |
| |
| qdf_status = wma_tx_attach(wma_handle); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: Failed to register tx management", __func__); |
| goto end; |
| } |
| |
| /* Initialize log completion timeout */ |
| qdf_status = qdf_mc_timer_init(&wma_handle->log_completion_timer, |
| QDF_TIMER_TYPE_SW, |
| wma_log_completion_timeout, |
| wma_handle); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("Failed to initialize log completion timeout"); |
| goto end; |
| } |
| |
| status = wma_fips_register_event_handlers(wma_handle); |
| if (!QDF_IS_STATUS_SUCCESS(status)) { |
| WMA_LOGE("Failed to register FIPS event handler"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| /* Initialize the get temperature event handler */ |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PDEV_TEMPERATURE_EVENTID, |
| wma_pdev_temperature_evt_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("Failed to register get_temperature event cb"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_VDEV_TSF_REPORT_EVENTID, |
| wma_vdev_tsf_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (0 != status) { |
| WMA_LOGE("%s: Failed to register tsf callback", __func__); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| /* Initialize the wma_pdev_set_hw_mode_resp_evt_handler event handler */ |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PDEV_SET_HW_MODE_RESP_EVENTID, |
| wma_pdev_set_hw_mode_resp_evt_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("Failed to register set hw mode resp event cb"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| /* Initialize the WMI_SOC_HW_MODE_TRANSITION_EVENTID event handler */ |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PDEV_HW_MODE_TRANSITION_EVENTID, |
| wma_pdev_hw_mode_transition_evt_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("Failed to register hw mode transition event cb"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| /* Initialize the set dual mac configuration event handler */ |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_PDEV_SET_MAC_CONFIG_RESP_EVENTID, |
| wma_pdev_set_dual_mode_config_resp_evt_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("Failed to register hw mode transition event cb"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_WLAN_COEX_BT_ACTIVITY_EVENTID, |
| wma_wlan_bt_activity_evt_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (!QDF_IS_STATUS_SUCCESS(status)) { |
| WMA_LOGE("Failed to register coex bt activity event handler"); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| |
| end: |
| WMA_LOGD("%s: Exit", __func__); |
| return qdf_status; |
| } |
| |
| /** |
| * wma_stop() - wma stop function. |
| * cleanup timers and suspend target. |
| * @reason: reason for wma_stop. |
| * |
| * Return: 0 on success, QDF Error on failure |
| */ |
| QDF_STATUS wma_stop(uint8_t reason) |
| { |
| tp_wma_handle wma_handle; |
| QDF_STATUS qdf_status = QDF_STATUS_SUCCESS; |
| int i; |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| WMA_LOGD("%s: Enter", __func__); |
| /* validate the wma_handle */ |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid handle", __func__); |
| qdf_status = QDF_STATUS_E_INVAL; |
| goto end; |
| } |
| #ifdef QCA_WIFI_FTM |
| /* |
| * Tx mgmt detach requires TXRX context which is not created |
| * in FTM mode. So skip the TX mgmt detach. |
| */ |
| if (cds_get_conparam() == QDF_GLOBAL_FTM_MODE) { |
| qdf_status = QDF_STATUS_SUCCESS; |
| goto end; |
| } |
| #endif /* QCA_WIFI_FTM */ |
| |
| if (wma_handle->ack_work_ctx) { |
| cds_flush_work(&wma_handle->ack_work_ctx->ack_cmp_work); |
| qdf_mem_free(wma_handle->ack_work_ctx); |
| wma_handle->ack_work_ctx = NULL; |
| } |
| |
| /* Destroy the timer for log completion */ |
| qdf_status = qdf_mc_timer_destroy(&wma_handle->log_completion_timer); |
| if (qdf_status != QDF_STATUS_SUCCESS) |
| WMA_LOGE("Failed to destroy the log completion timer"); |
| /* clean up ll-queue for all vdev */ |
| for (i = 0; i < wma_handle->max_bssid; i++) { |
| if (wma_handle->interfaces[i].handle && |
| wma_is_vdev_up(i)) { |
| cdp_fc_vdev_flush( |
| cds_get_context(QDF_MODULE_ID_SOC), |
| wma_handle-> |
| interfaces[i].handle); |
| } |
| } |
| qdf_status = wma_tx_detach(wma_handle); |
| if (qdf_status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("%s: Failed to deregister tx management", __func__); |
| goto end; |
| } |
| |
| end: |
| WMA_LOGD("%s: Exit", __func__); |
| return qdf_status; |
| } |
| |
| /** |
| * wma_wmi_service_close() - close wma wmi service interface. |
| * |
| * Return: 0 on success, QDF Error on failure |
| */ |
| QDF_STATUS wma_wmi_service_close(void) |
| { |
| void *cds_ctx; |
| tp_wma_handle wma_handle; |
| struct beacon_info *bcn; |
| int i; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| cds_ctx = cds_get_global_context(); |
| if (!cds_ctx) { |
| WMA_LOGE("%s: Invalid CDS context", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| /* validate the wma_handle */ |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* validate the wmi handle */ |
| if (NULL == wma_handle->wmi_handle) { |
| WMA_LOGE("%s: Invalid wmi handle", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* dettach the wmi serice */ |
| WMA_LOGD("calling wmi_unified_detach"); |
| wmi_unified_detach(wma_handle->wmi_handle); |
| wma_handle->wmi_handle = NULL; |
| |
| for (i = 0; i < wma_handle->max_bssid; i++) { |
| bcn = wma_handle->interfaces[i].beacon; |
| |
| if (bcn) { |
| if (bcn->dma_mapped) |
| qdf_nbuf_unmap_single(wma_handle->qdf_dev, |
| bcn->buf, QDF_DMA_TO_DEVICE); |
| qdf_nbuf_free(bcn->buf); |
| qdf_mem_free(bcn); |
| wma_handle->interfaces[i].beacon = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].handle) { |
| qdf_mem_free(wma_handle->interfaces[i].handle); |
| wma_handle->interfaces[i].handle = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].addBssStaContext) { |
| qdf_mem_free(wma_handle-> |
| interfaces[i].addBssStaContext); |
| wma_handle->interfaces[i].addBssStaContext = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].del_staself_req) { |
| qdf_mem_free(wma_handle->interfaces[i].del_staself_req); |
| wma_handle->interfaces[i].del_staself_req = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].stats_rsp) { |
| qdf_mem_free(wma_handle->interfaces[i].stats_rsp); |
| wma_handle->interfaces[i].stats_rsp = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].psnr_req) { |
| qdf_mem_free(wma_handle-> |
| interfaces[i].psnr_req); |
| wma_handle->interfaces[i].psnr_req = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].rcpi_req) { |
| qdf_mem_free(wma_handle-> |
| interfaces[i].rcpi_req); |
| wma_handle->interfaces[i].rcpi_req = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].roam_synch_frame_ind. |
| bcn_probe_rsp) { |
| qdf_mem_free(wma_handle->interfaces[i]. |
| roam_synch_frame_ind.bcn_probe_rsp); |
| wma_handle->interfaces[i].roam_synch_frame_ind. |
| bcn_probe_rsp = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].roam_synch_frame_ind. |
| reassoc_req) { |
| qdf_mem_free(wma_handle->interfaces[i]. |
| roam_synch_frame_ind.reassoc_req); |
| wma_handle->interfaces[i].roam_synch_frame_ind. |
| reassoc_req = NULL; |
| } |
| |
| if (wma_handle->interfaces[i].roam_synch_frame_ind. |
| reassoc_rsp) { |
| qdf_mem_free(wma_handle->interfaces[i]. |
| roam_synch_frame_ind.reassoc_rsp); |
| wma_handle->interfaces[i].roam_synch_frame_ind. |
| reassoc_rsp = NULL; |
| } |
| |
| wma_vdev_deinit(&wma_handle->interfaces[i]); |
| } |
| |
| qdf_mem_free(wma_handle->interfaces); |
| |
| /* free the wma_handle */ |
| cds_free_context(QDF_MODULE_ID_WMA, wma_handle); |
| |
| qdf_mem_free(((p_cds_contextType) cds_ctx)->cfg_ctx); |
| WMA_LOGD("%s: Exit", __func__); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_wmi_work_close() - close the work queue items associated with WMI |
| * |
| * This function closes work queue items associated with WMI, but not fully |
| * closes WMI service. |
| * |
| * Return: QDF_STATUS_SUCCESS if work close is successful. Otherwise |
| * proper error codes. |
| */ |
| QDF_STATUS wma_wmi_work_close(void) |
| { |
| tp_wma_handle wma_handle; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| /* validate the wma_handle */ |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* validate the wmi handle */ |
| if (NULL == wma_handle->wmi_handle) { |
| WMA_LOGE("%s: Invalid wmi handle", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* remove the wmi work */ |
| WMA_LOGD("calling wmi_unified_remove_work"); |
| wmi_unified_remove_work(wma_handle->wmi_handle); |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_cleanup_dbs_phy_caps() - release memory allocated for holding ext cap |
| * @wma_handle: pointer to wma handle |
| * |
| * This function releases all the memory created for holding extended |
| * capabilities per hardware mode and per PHY |
| * |
| * Return: void |
| */ |
| static void wma_cleanup_dbs_phy_caps(t_wma_handle *wma_handle) |
| { |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return; |
| } |
| |
| if (wma_handle->phy_caps.hw_mode_to_mac_cap_map) { |
| qdf_mem_free(wma_handle->phy_caps.hw_mode_to_mac_cap_map); |
| wma_handle->phy_caps.hw_mode_to_mac_cap_map = NULL; |
| WMA_LOGD("%s: hw_mode_to_mac_cap_map freed", __func__); |
| } |
| |
| if (wma_handle->phy_caps.each_hw_mode_cap) { |
| qdf_mem_free(wma_handle->phy_caps.each_hw_mode_cap); |
| wma_handle->phy_caps.each_hw_mode_cap = NULL; |
| WMA_LOGD("%s: each_hw_mode_cap freed", __func__); |
| } |
| |
| if (wma_handle->phy_caps.each_phy_cap_per_hwmode) { |
| qdf_mem_free(wma_handle->phy_caps.each_phy_cap_per_hwmode); |
| wma_handle->phy_caps.each_phy_cap_per_hwmode = NULL; |
| WMA_LOGD("%s: each_phy_cap_per_hwmode freed", __func__); |
| } |
| |
| if (wma_handle->phy_caps.each_phy_hal_reg_cap) { |
| qdf_mem_free(wma_handle->phy_caps.each_phy_hal_reg_cap); |
| wma_handle->phy_caps.each_phy_hal_reg_cap = NULL; |
| WMA_LOGD("%s: each_phy_hal_reg_cap freed", __func__); |
| } |
| } |
| |
| /** |
| * wma_close() - wma close function. |
| * cleanup resources attached with wma. |
| * |
| * Return: 0 on success, QDF Error on failure |
| */ |
| QDF_STATUS wma_close(void) |
| { |
| tp_wma_handle wma_handle; |
| uint32_t idx; |
| QDF_STATUS qdf_status = QDF_STATUS_SUCCESS; |
| struct target_psoc_info *tgt_psoc_info; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| /* validate the wma_handle */ |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* validate the wmi handle */ |
| if (NULL == wma_handle->wmi_handle) { |
| WMA_LOGE("%s: Invalid wmi handle", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| /* Free DBS list */ |
| if (wma_handle->hw_mode.hw_mode_list) { |
| qdf_mem_free(wma_handle->hw_mode.hw_mode_list); |
| wma_handle->hw_mode.hw_mode_list = NULL; |
| WMA_LOGD("%s: DBS list is freed", __func__); |
| } |
| wma_cleanup_dbs_phy_caps(wma_handle); |
| if (cds_get_conparam() != QDF_GLOBAL_FTM_MODE) { |
| #ifdef FEATURE_WLAN_EXTSCAN |
| qdf_wake_lock_destroy(&wma_handle->extscan_wake_lock); |
| #endif /* FEATURE_WLAN_EXTSCAN */ |
| qdf_wake_lock_destroy(&wma_handle->wow_wake_lock); |
| qdf_wake_lock_destroy(&wma_handle->wow_auth_req_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_assoc_req_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_deauth_rec_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_disassoc_rec_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_ap_assoc_lost_wl); |
| qdf_wake_lock_destroy(&wma_handle->wow_auto_shutdown_wl); |
| qdf_wake_lock_destroy(&wma_handle->roam_ho_wl); |
| } |
| |
| /* unregister Firmware debug log */ |
| qdf_status = dbglog_deinit(wma_handle->wmi_handle); |
| if (qdf_status != QDF_STATUS_SUCCESS) |
| WMA_LOGE("%s: dbglog_deinit failed", __func__); |
| |
| /* close the qdf events */ |
| qdf_event_destroy(&wma_handle->wma_ready_event); |
| qdf_status = qdf_mc_timer_destroy(&wma_handle->service_ready_ext_timer); |
| if (!QDF_IS_STATUS_SUCCESS(qdf_status)) |
| WMA_LOGE("%s: Failed to destroy service ready ext event timer", |
| __func__); |
| |
| qdf_event_destroy(&wma_handle->target_suspend); |
| qdf_event_destroy(&wma_handle->wma_resume_event); |
| qdf_event_destroy(&wma_handle->runtime_suspend); |
| qdf_event_destroy(&wma_handle->recovery_event); |
| qdf_event_destroy(&wma_handle->tx_frm_download_comp_event); |
| qdf_event_destroy(&wma_handle->tx_queue_empty_event); |
| wma_cleanup_vdev_resp_queue(wma_handle); |
| wma_cleanup_hold_req(wma_handle); |
| qdf_wake_lock_destroy(&wma_handle->wmi_cmd_rsp_wake_lock); |
| qdf_runtime_lock_deinit(&wma_handle->wmi_cmd_rsp_runtime_lock); |
| qdf_spinlock_destroy(&wma_handle->vdev_respq_lock); |
| qdf_spinlock_destroy(&wma_handle->wma_hold_req_q_lock); |
| for (idx = 0; idx < wma_handle->num_mem_chunks; ++idx) { |
| qdf_mem_free_consistent(wma_handle->qdf_dev, |
| wma_handle->qdf_dev->dev, |
| wma_handle->mem_chunks[idx].len, |
| wma_handle->mem_chunks[idx].vaddr, |
| wma_handle->mem_chunks[idx].paddr, |
| qdf_get_dma_mem_context( |
| (&(wma_handle->mem_chunks[idx])), |
| memctx)); |
| } |
| |
| #if defined(QCA_WIFI_FTM) |
| /* Detach UTF and unregister the handler */ |
| if (cds_get_conparam() == QDF_GLOBAL_FTM_MODE) |
| wma_utf_detach(wma_handle); |
| #endif /* QCA_WIFI_FTM */ |
| |
| if (NULL != wma_handle->pGetRssiReq) { |
| qdf_mem_free(wma_handle->pGetRssiReq); |
| wma_handle->pGetRssiReq = NULL; |
| } |
| |
| wma_unified_radio_tx_mem_free(wma_handle); |
| |
| wma_ndp_unregister_all_event_handlers(wma_handle); |
| if (wma_handle->pdev) { |
| wlan_objmgr_pdev_release_ref(wma_handle->pdev, |
| WLAN_LEGACY_WMA_ID); |
| wma_handle->pdev = NULL; |
| } |
| |
| tgt_psoc_info = wlan_psoc_get_tgt_if_handle(wma_handle->psoc); |
| if (tgt_psoc_info) { |
| qdf_mem_free(tgt_psoc_info); |
| wlan_psoc_set_tgt_if_handle(wma_handle->psoc, NULL); |
| } |
| |
| pmo_unregister_pause_bitmap_notifier(wma_handle->psoc, |
| wma_vdev_update_pause_bitmap); |
| pmo_unregister_get_pause_bitmap(wma_handle->psoc, |
| wma_vdev_get_pause_bitmap); |
| pmo_unregister_is_device_in_low_pwr_mode(wma_handle->psoc, |
| wma_vdev_is_device_in_low_pwr_mode); |
| |
| wlan_objmgr_psoc_release_ref(wma_handle->psoc, WLAN_LEGACY_WMA_ID); |
| wma_handle->psoc = NULL; |
| target_if_close(); |
| wma_target_if_close(wma_handle); |
| |
| WMA_LOGD("%s: Exit", __func__); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_update_fw_config() - update fw configuration |
| * @wma_handle: wma handle |
| * @tgt_cap: target capabality |
| * |
| * Return: none |
| */ |
| static void wma_update_fw_config(tp_wma_handle wma_handle, |
| struct wma_target_cap *tgt_cap) |
| { |
| /* |
| * tgt_cap contains default target resource configuration |
| * which can be modified here, if required |
| */ |
| /* Override the no. of max fragments as per platform configuration */ |
| tgt_cap->wlan_resource_config.max_frag_entries = |
| QDF_MIN(QCA_OL_11AC_TX_MAX_FRAGS, |
| wma_handle->max_frag_entry); |
| wma_handle->max_frag_entry = |
| tgt_cap->wlan_resource_config.max_frag_entries; |
| |
| /* Update no. of maxWoWFilters depending on BPF service */ |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_BPF_OFFLOAD)) |
| tgt_cap->wlan_resource_config.num_wow_filters = |
| WMA_STA_WOW_DEFAULT_PTRN_MAX; |
| } |
| |
| /** |
| * wma_alloc_host_mem_chunk() - allocate host memory |
| * @wma_handle: wma handle |
| * @req_id: request id |
| * @idx: index |
| * @num_units: number of units |
| * @unit_len: unit length |
| * |
| * allocate a chunk of memory at the index indicated and |
| * if allocation fail allocate smallest size possiblr and |
| * return number of units allocated. |
| * |
| * Return: number of units or 0 for error. |
| */ |
| static uint32_t wma_alloc_host_mem_chunk(tp_wma_handle wma_handle, |
| uint32_t req_id, uint32_t idx, |
| uint32_t num_units, uint32_t unit_len) |
| { |
| qdf_dma_addr_t paddr; |
| |
| if (!num_units || !unit_len) |
| return 0; |
| wma_handle->mem_chunks[idx].vaddr = NULL; |
| /** reduce the requested allocation by half until allocation succeeds */ |
| while (wma_handle->mem_chunks[idx].vaddr == NULL && num_units) { |
| wma_handle->mem_chunks[idx].vaddr = |
| qdf_mem_alloc_consistent(wma_handle->qdf_dev, |
| wma_handle->qdf_dev->dev, |
| num_units * unit_len, |
| &paddr); |
| if (wma_handle->mem_chunks[idx].vaddr == NULL) { |
| num_units = (num_units >> 1);/* reduce length by half */ |
| } else { |
| wma_handle->mem_chunks[idx].paddr = paddr; |
| wma_handle->mem_chunks[idx].len = num_units * unit_len; |
| wma_handle->mem_chunks[idx].req_id = req_id; |
| } |
| } |
| return num_units; |
| } |
| |
| #define HOST_MEM_SIZE_UNIT 4 |
| /** |
| * wma_alloc_host_mem() - allocate amount of memory requested by FW. |
| * @wma_handle: wma handle |
| * @req_id: request id |
| * @num_units: number of units |
| * @unit_len: unit length |
| * |
| * Return: none |
| */ |
| static void wma_alloc_host_mem(tp_wma_handle wma_handle, uint32_t req_id, |
| uint32_t num_units, uint32_t unit_len) |
| { |
| uint32_t remaining_units, allocated_units, idx; |
| |
| /* adjust the length to nearest multiple of unit size */ |
| unit_len = (unit_len + (HOST_MEM_SIZE_UNIT - 1)) & |
| (~(HOST_MEM_SIZE_UNIT - 1)); |
| idx = wma_handle->num_mem_chunks; |
| remaining_units = num_units; |
| while (remaining_units) { |
| allocated_units = wma_alloc_host_mem_chunk(wma_handle, req_id, |
| idx, remaining_units, |
| unit_len); |
| if (allocated_units == 0) { |
| WMA_LOGE("FAILED TO ALLOCATED memory unit len %d units requested %d units allocated %d ", |
| unit_len, num_units, |
| (num_units - remaining_units)); |
| wma_handle->num_mem_chunks = idx; |
| break; |
| } |
| remaining_units -= allocated_units; |
| ++idx; |
| if (idx == MAX_MEM_CHUNKS) { |
| WMA_LOGE("RWACHED MAX CHUNK LIMIT for memory units %d unit len %d requested by FW, only allocated %d ", |
| num_units, unit_len, |
| (num_units - remaining_units)); |
| wma_handle->num_mem_chunks = idx; |
| break; |
| } |
| } |
| wma_handle->num_mem_chunks = idx; |
| } |
| |
| /** |
| * wma_set_tx_partition_base() - set TX MSDU ID partition base for IPA |
| * @value: TX MSDU ID partition base |
| * |
| * Return: none |
| */ |
| #ifdef IPA_OFFLOAD |
| static void wma_set_tx_partition_base(uint32_t value) |
| { |
| cdp_ipa_set_uc_tx_partition_base( |
| cds_get_context(QDF_MODULE_ID_SOC), |
| (struct cdp_cfg *)cds_get_context(QDF_MODULE_ID_CFG), |
| value); |
| WMA_LOGD("%s: TX_MSDU_ID_PARTITION=%d", __func__, |
| value); |
| } |
| #else |
| static void wma_set_tx_partition_base(uint32_t value) |
| { |
| } |
| #endif |
| |
| /** |
| * wma_update_target_services() - update target services from wma handle |
| * @wh: wma handle |
| * @cfg: target services |
| * |
| * Return: none |
| */ |
| static inline void wma_update_target_services(tp_wma_handle wh, |
| struct wma_tgt_services *cfg) |
| { |
| /* STA power save */ |
| cfg->sta_power_save = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_STA_PWRSAVE); |
| |
| /* Enable UAPSD */ |
| cfg->uapsd = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_AP_UAPSD); |
| |
| /* Update AP DFS service */ |
| cfg->ap_dfs = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_AP_DFS); |
| |
| /* Enable 11AC */ |
| cfg->en_11ac = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_11AC); |
| if (cfg->en_11ac) |
| g_fw_wlan_feat_caps |= (1 << DOT11AC); |
| |
| /* Proactive ARP response */ |
| g_fw_wlan_feat_caps |= (1 << WLAN_PERIODIC_TX_PTRN); |
| |
| /* Enable WOW */ |
| g_fw_wlan_feat_caps |= (1 << WOW); |
| |
| /* ARP offload */ |
| cfg->arp_offload = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_ARPNS_OFFLOAD); |
| |
| /* Adaptive early-rx */ |
| cfg->early_rx = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_EARLY_RX); |
| #ifdef FEATURE_WLAN_SCAN_PNO |
| /* PNO offload */ |
| if (WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, WMI_SERVICE_NLO)) |
| cfg->pno_offload = true; |
| #endif /* FEATURE_WLAN_SCAN_PNO */ |
| |
| #ifdef FEATURE_WLAN_EXTSCAN |
| if (WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, WMI_SERVICE_EXTSCAN)) |
| g_fw_wlan_feat_caps |= (1 << EXTENDED_SCAN); |
| #endif /* FEATURE_WLAN_EXTSCAN */ |
| cfg->lte_coex_ant_share = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_LTE_ANT_SHARE_SUPPORT); |
| #ifdef FEATURE_WLAN_TDLS |
| /* Enable TDLS */ |
| if (WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, WMI_SERVICE_TDLS)) { |
| cfg->en_tdls = 1; |
| g_fw_wlan_feat_caps |= (1 << TDLS); |
| } |
| /* Enable advanced TDLS features */ |
| if (WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_TDLS_OFFCHAN)) { |
| cfg->en_tdls_offchan = 1; |
| g_fw_wlan_feat_caps |= (1 << TDLS_OFF_CHANNEL); |
| } |
| |
| cfg->en_tdls_uapsd_buf_sta = |
| WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_TDLS_UAPSD_BUFFER_STA); |
| cfg->en_tdls_uapsd_sleep_sta = |
| WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_TDLS_UAPSD_SLEEP_STA); |
| #endif /* FEATURE_WLAN_TDLS */ |
| if (WMI_SERVICE_IS_ENABLED |
| (wh->wmi_service_bitmap, WMI_SERVICE_BEACON_OFFLOAD)) |
| cfg->beacon_offload = true; |
| if (WMI_SERVICE_IS_ENABLED |
| (wh->wmi_service_bitmap, WMI_SERVICE_STA_PMF_OFFLOAD)) |
| cfg->pmf_offload = true; |
| #ifdef WLAN_FEATURE_ROAM_OFFLOAD |
| /* Enable Roam Offload */ |
| cfg->en_roam_offload = WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_ROAM_HO_OFFLOAD); |
| #endif /* WLAN_FEATURE_ROAM_OFFLOAD */ |
| #ifdef WLAN_FEATURE_NAN |
| if (WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, WMI_SERVICE_NAN)) |
| g_fw_wlan_feat_caps |= (1 << NAN); |
| #endif /* WLAN_FEATURE_NAN */ |
| |
| if (WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, WMI_SERVICE_RTT)) |
| g_fw_wlan_feat_caps |= (1 << RTT); |
| |
| if (WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_TX_MSDU_ID_NEW_PARTITION_SUPPORT)) { |
| wma_set_tx_partition_base(HTT_TX_IPA_NEW_MSDU_ID_SPACE_BEGIN); |
| } else { |
| wma_set_tx_partition_base(HTT_TX_IPA_MSDU_ID_SPACE_BEGIN); |
| } |
| |
| wma_he_update_tgt_services(wh, cfg); |
| |
| cfg->get_peer_info_enabled = |
| WMI_SERVICE_IS_ENABLED(wh->wmi_service_bitmap, |
| WMI_SERVICE_PEER_STATS_INFO); |
| if (WMI_SERVICE_EXT_IS_ENABLED(wh->wmi_service_bitmap, |
| wh->wmi_service_ext_bitmap, |
| WMI_SERVICE_FILS_SUPPORT)) |
| cfg->is_fils_roaming_supported = true; |
| |
| if (WMI_SERVICE_EXT_IS_ENABLED(wh->wmi_service_bitmap, |
| wh->wmi_service_ext_bitmap, |
| WMI_SERVICE_MAWC_SUPPORT)) |
| cfg->is_fw_mawc_capable = true; |
| |
| } |
| |
| /** |
| * wma_update_target_ht_cap() - update ht capabality from wma handle |
| * @wh: wma handle |
| * @cfg: ht capabality |
| * |
| * Return: none |
| */ |
| static inline void wma_update_target_ht_cap(tp_wma_handle wh, |
| struct wma_tgt_ht_cap *cfg) |
| { |
| /* RX STBC */ |
| cfg->ht_rx_stbc = !!(wh->ht_cap_info & WMI_HT_CAP_RX_STBC); |
| |
| /* TX STBC */ |
| cfg->ht_tx_stbc = !!(wh->ht_cap_info & WMI_HT_CAP_TX_STBC); |
| |
| /* MPDU density */ |
| cfg->mpdu_density = wh->ht_cap_info & WMI_HT_CAP_MPDU_DENSITY; |
| |
| /* HT RX LDPC */ |
| cfg->ht_rx_ldpc = !!(wh->ht_cap_info & WMI_HT_CAP_LDPC); |
| |
| /* HT SGI */ |
| cfg->ht_sgi_20 = !!(wh->ht_cap_info & WMI_HT_CAP_HT20_SGI); |
| |
| cfg->ht_sgi_40 = !!(wh->ht_cap_info & WMI_HT_CAP_HT40_SGI); |
| |
| /* RF chains */ |
| cfg->num_rf_chains = wh->num_rf_chains; |
| |
| WMA_LOGD("%s: ht_cap_info - %x ht_rx_stbc - %d, ht_tx_stbc - %d\n" |
| "mpdu_density - %d ht_rx_ldpc - %d ht_sgi_20 - %d\n" |
| "ht_sgi_40 - %d num_rf_chains - %d", __func__, wh->ht_cap_info, |
| cfg->ht_rx_stbc, cfg->ht_tx_stbc, cfg->mpdu_density, |
| cfg->ht_rx_ldpc, cfg->ht_sgi_20, cfg->ht_sgi_40, |
| cfg->num_rf_chains); |
| |
| } |
| |
| /** |
| * wma_update_target_vht_cap() - update vht capabality from wma handle |
| * @wh: wma handle |
| * @cfg: vht capabality |
| * |
| * Return: none |
| */ |
| static inline void wma_update_target_vht_cap(tp_wma_handle wh, |
| struct wma_tgt_vht_cap *cfg) |
| { |
| |
| if (wh->vht_cap_info & WMI_VHT_CAP_MAX_MPDU_LEN_11454) |
| cfg->vht_max_mpdu = WMI_VHT_CAP_MAX_MPDU_LEN_11454; |
| else if (wh->vht_cap_info & WMI_VHT_CAP_MAX_MPDU_LEN_7935) |
| cfg->vht_max_mpdu = WMI_VHT_CAP_MAX_MPDU_LEN_7935; |
| else |
| cfg->vht_max_mpdu = 0; |
| |
| |
| if (wh->vht_cap_info & WMI_VHT_CAP_CH_WIDTH_80P80_160MHZ) { |
| cfg->supp_chan_width = 1 << eHT_CHANNEL_WIDTH_80P80MHZ; |
| cfg->supp_chan_width |= 1 << eHT_CHANNEL_WIDTH_160MHZ; |
| } else if (wh->vht_cap_info & WMI_VHT_CAP_CH_WIDTH_160MHZ) |
| cfg->supp_chan_width = 1 << eHT_CHANNEL_WIDTH_160MHZ; |
| else |
| cfg->supp_chan_width = 1 << eHT_CHANNEL_WIDTH_80MHZ; |
| |
| cfg->vht_rx_ldpc = wh->vht_cap_info & WMI_VHT_CAP_RX_LDPC; |
| |
| cfg->vht_short_gi_80 = wh->vht_cap_info & WMI_VHT_CAP_SGI_80MHZ; |
| cfg->vht_short_gi_160 = wh->vht_cap_info & WMI_VHT_CAP_SGI_160MHZ; |
| |
| cfg->vht_tx_stbc = wh->vht_cap_info & WMI_VHT_CAP_TX_STBC; |
| |
| cfg->vht_rx_stbc = (wh->vht_cap_info & WMI_VHT_CAP_RX_STBC_1SS) | |
| (wh->vht_cap_info & WMI_VHT_CAP_RX_STBC_2SS) | |
| (wh->vht_cap_info & WMI_VHT_CAP_RX_STBC_3SS); |
| |
| cfg->vht_max_ampdu_len_exp = (wh->vht_cap_info & |
| WMI_VHT_CAP_MAX_AMPDU_LEN_EXP) |
| >> WMI_VHT_CAP_MAX_AMPDU_LEN_EXP_SHIFT; |
| |
| cfg->vht_su_bformer = wh->vht_cap_info & WMI_VHT_CAP_SU_BFORMER; |
| |
| cfg->vht_su_bformee = wh->vht_cap_info & WMI_VHT_CAP_SU_BFORMEE; |
| |
| cfg->vht_mu_bformer = wh->vht_cap_info & WMI_VHT_CAP_MU_BFORMER; |
| |
| cfg->vht_mu_bformee = wh->vht_cap_info & WMI_VHT_CAP_MU_BFORMEE; |
| |
| cfg->vht_txop_ps = wh->vht_cap_info & WMI_VHT_CAP_TXOP_PS; |
| |
| WMA_LOGD("%s: max_mpdu %d supp_chan_width %x rx_ldpc %x\n" |
| "short_gi_80 %x tx_stbc %x rx_stbc %x txop_ps %x\n" |
| "su_bformee %x mu_bformee %x max_ampdu_len_exp %d", __func__, |
| cfg->vht_max_mpdu, cfg->supp_chan_width, cfg->vht_rx_ldpc, |
| cfg->vht_short_gi_80, cfg->vht_tx_stbc, cfg->vht_rx_stbc, |
| cfg->vht_txop_ps, cfg->vht_su_bformee, cfg->vht_mu_bformee, |
| cfg->vht_max_ampdu_len_exp); |
| } |
| |
| /** |
| * wma_update_supported_bands() - update supported bands from service ready ext |
| * @wma_handle: pointer to wma handle |
| * @supported_bands: Supported band given by FW through service ready ext params |
| * @new_supported_bands: New supported band which needs to be updated by |
| * this API which WMA layer understands |
| * |
| * This API will convert FW given supported band to enum which WMA layer |
| * understands |
| * |
| * Return: QDF_STATUS |
| */ |
| static QDF_STATUS wma_update_supported_bands(t_wma_handle *wma_handle, |
| WLAN_BAND_CAPABILITY supported_bands, |
| WMI_PHY_CAPABILITY *new_supported_bands) |
| { |
| QDF_STATUS status = QDF_STATUS_SUCCESS; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: NULL wma_handle", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| if (!new_supported_bands) { |
| WMA_LOGE("%s: NULL new supported band variable", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| switch (supported_bands) { |
| case WLAN_2G_CAPABILITY: |
| *new_supported_bands |= WMI_11G_CAPABILITY; |
| break; |
| case WLAN_5G_CAPABILITY: |
| *new_supported_bands |= WMI_11A_CAPABILITY; |
| break; |
| default: |
| WMA_LOGE("%s: wrong supported band", __func__); |
| status = QDF_STATUS_E_FAILURE; |
| break; |
| } |
| return status; |
| } |
| |
| /** |
| * wma_derive_ext_ht_cap() - Derive HT caps based on given value |
| * @wma_handle: pointer to wma_handle |
| * @ht_cap: given pointer to HT caps which needs to be updated |
| * @tx_chain: given tx chainmask value |
| * @rx_chain: given rx chainmask value |
| * @value: new HT cap info provided in form of bitmask |
| * |
| * This function takes the value provided in form of bitmask and decodes |
| * it. After decoding, what ever value it gets, it takes the union(max) or |
| * intersection(min) with previously derived values. |
| * |
| * Return: none |
| * |
| */ |
| static void wma_derive_ext_ht_cap(tp_wma_handle wma_handle, |
| struct wma_tgt_ht_cap *ht_cap, uint32_t value, |
| uint32_t tx_chain, uint32_t rx_chain) |
| { |
| struct wma_tgt_ht_cap tmp = {0}; |
| |
| if (NULL == wma_handle || NULL == ht_cap) |
| return; |
| |
| if (!qdf_mem_cmp(ht_cap, &tmp, sizeof(struct wma_tgt_ht_cap))) { |
| ht_cap->ht_rx_stbc = (!!(value & WMI_HT_CAP_RX_STBC)); |
| ht_cap->ht_tx_stbc = (!!(value & WMI_HT_CAP_TX_STBC)); |
| ht_cap->mpdu_density = (!!(value & WMI_HT_CAP_MPDU_DENSITY)); |
| ht_cap->ht_rx_ldpc = (!!(value & WMI_HT_CAP_RX_LDPC)); |
| ht_cap->ht_sgi_20 = (!!(value & WMI_HT_CAP_HT20_SGI)); |
| ht_cap->ht_sgi_40 = (!!(value & WMI_HT_CAP_HT40_SGI)); |
| ht_cap->num_rf_chains = |
| QDF_MAX(wma_get_num_of_setbits_from_bitmask(tx_chain), |
| wma_get_num_of_setbits_from_bitmask(rx_chain)); |
| } else { |
| ht_cap->ht_rx_stbc = QDF_MIN(ht_cap->ht_rx_stbc, |
| (!!(value & WMI_HT_CAP_RX_STBC))); |
| ht_cap->ht_tx_stbc = QDF_MAX(ht_cap->ht_tx_stbc, |
| (!!(value & WMI_HT_CAP_TX_STBC))); |
| ht_cap->mpdu_density = QDF_MIN(ht_cap->mpdu_density, |
| (!!(value & WMI_HT_CAP_MPDU_DENSITY))); |
| ht_cap->ht_rx_ldpc = QDF_MIN(ht_cap->ht_rx_ldpc, |
| (!!(value & WMI_HT_CAP_RX_LDPC))); |
| ht_cap->ht_sgi_20 = QDF_MIN(ht_cap->ht_sgi_20, |
| (!!(value & WMI_HT_CAP_HT20_SGI))); |
| ht_cap->ht_sgi_40 = QDF_MIN(ht_cap->ht_sgi_40, |
| (!!(value & WMI_HT_CAP_HT40_SGI))); |
| ht_cap->num_rf_chains = |
| QDF_MAX(ht_cap->num_rf_chains, |
| QDF_MAX(wma_get_num_of_setbits_from_bitmask( |
| tx_chain), |
| wma_get_num_of_setbits_from_bitmask( |
| rx_chain))); |
| } |
| } |
| |
| /** |
| * wma_update_target_ext_ht_cap() - Update HT caps with given extended cap |
| * @wma_handle: pointer to wma_handle |
| * @ht_cap: HT cap structure to be filled |
| * |
| * This function loop through each hardware mode and for each hardware mode |
| * again it loop through each MAC/PHY and pull the caps 2G and 5G specific |
| * HT caps and derives the final cap. |
| * |
| * Return: none |
| * |
| */ |
| static void wma_update_target_ext_ht_cap(tp_wma_handle wma_handle, |
| struct wma_tgt_ht_cap *ht_cap) |
| { |
| int i, j = 0, max_mac; |
| uint32_t ht_2g, ht_5g; |
| struct wma_tgt_ht_cap tmp_ht_cap = {0}, tmp_cap = {0}; |
| struct extended_caps *phy_caps; |
| WMI_MAC_PHY_CAPABILITIES *mac_cap; |
| |
| /* |
| * for legacy device extended cap might not even come, so in that case |
| * don't overwrite legacy values |
| */ |
| if (!wma_handle || |
| (0 == wma_handle->phy_caps.num_hw_modes.num_hw_modes)) { |
| WMA_LOGD("%s: No extended HT cap for current SOC", __func__); |
| return; |
| } |
| |
| phy_caps = &wma_handle->phy_caps; |
| for (i = 0; i < phy_caps->num_hw_modes.num_hw_modes; i++) { |
| if (phy_caps->each_hw_mode_cap[i].phy_id_map == PHY1_PHY2) |
| max_mac = j + 2; |
| else |
| max_mac = j + 1; |
| for ( ; j < max_mac; j++) { |
| mac_cap = &phy_caps->each_phy_cap_per_hwmode[j]; |
| ht_2g = mac_cap->ht_cap_info_2G; |
| ht_5g = mac_cap->ht_cap_info_5G; |
| if (ht_2g) |
| wma_derive_ext_ht_cap(wma_handle, &tmp_ht_cap, |
| ht_2g, mac_cap->tx_chain_mask_2G, |
| mac_cap->rx_chain_mask_2G); |
| if (ht_5g) |
| wma_derive_ext_ht_cap(wma_handle, &tmp_ht_cap, |
| ht_5g, mac_cap->tx_chain_mask_5G, |
| mac_cap->rx_chain_mask_5G); |
| } |
| } |
| |
| if (qdf_mem_cmp(&tmp_cap, &tmp_ht_cap, |
| sizeof(struct wma_tgt_ht_cap))) { |
| qdf_mem_copy(ht_cap, &tmp_ht_cap, |
| sizeof(struct wma_tgt_ht_cap)); |
| } |
| |
| WMA_LOGD("%s: [ext ht cap] ht_rx_stbc - %d, ht_tx_stbc - %d\n" |
| "mpdu_density - %d ht_rx_ldpc - %d ht_sgi_20 - %d\n" |
| "ht_sgi_40 - %d num_rf_chains - %d", __func__, |
| ht_cap->ht_rx_stbc, ht_cap->ht_tx_stbc, |
| ht_cap->mpdu_density, ht_cap->ht_rx_ldpc, |
| ht_cap->ht_sgi_20, ht_cap->ht_sgi_40, |
| ht_cap->num_rf_chains); |
| } |
| |
| /** |
| * wma_derive_ext_vht_cap() - Derive VHT caps based on given value |
| * @wma_handle: pointer to wma_handle |
| * @vht_cap: pointer to given VHT caps to be filled |
| * @value: new VHT cap info provided in form of bitmask |
| * |
| * This function takes the value provided in form of bitmask and decodes |
| * it. After decoding, what ever value it gets, it takes the union(max) or |
| * intersection(min) with previously derived values. |
| * |
| * Return: none |
| * |
| */ |
| static void wma_derive_ext_vht_cap(t_wma_handle *wma_handle, |
| struct wma_tgt_vht_cap *vht_cap, uint32_t value) |
| { |
| struct wma_tgt_vht_cap tmp_cap = {0}; |
| uint32_t tmp = 0; |
| |
| if (NULL == wma_handle || NULL == vht_cap) |
| return; |
| |
| if (!qdf_mem_cmp(vht_cap, &tmp_cap, |
| sizeof(struct wma_tgt_vht_cap))) { |
| if (value & WMI_VHT_CAP_MAX_MPDU_LEN_11454) |
| vht_cap->vht_max_mpdu = WMI_VHT_CAP_MAX_MPDU_LEN_11454; |
| else if (value & WMI_VHT_CAP_MAX_MPDU_LEN_7935) |
| vht_cap->vht_max_mpdu = WMI_VHT_CAP_MAX_MPDU_LEN_7935; |
| else |
| vht_cap->vht_max_mpdu = 0; |
| |
| if (value & WMI_VHT_CAP_CH_WIDTH_80P80_160MHZ) { |
| vht_cap->supp_chan_width = |
| 1 << eHT_CHANNEL_WIDTH_80P80MHZ; |
| vht_cap->supp_chan_width |= |
| 1 << eHT_CHANNEL_WIDTH_160MHZ; |
| } else if (value & WMI_VHT_CAP_CH_WIDTH_160MHZ) { |
| vht_cap->supp_chan_width = |
| 1 << eHT_CHANNEL_WIDTH_160MHZ; |
| } else { |
| vht_cap->supp_chan_width = 1 << eHT_CHANNEL_WIDTH_80MHZ; |
| } |
| vht_cap->vht_rx_ldpc = value & WMI_VHT_CAP_RX_LDPC; |
| vht_cap->vht_short_gi_80 = value & WMI_VHT_CAP_SGI_80MHZ; |
| vht_cap->vht_short_gi_160 = value & WMI_VHT_CAP_SGI_160MHZ; |
| vht_cap->vht_tx_stbc = value & WMI_VHT_CAP_TX_STBC; |
| vht_cap->vht_rx_stbc = |
| (value & WMI_VHT_CAP_RX_STBC_1SS) | |
| (value & WMI_VHT_CAP_RX_STBC_2SS) | |
| (value & WMI_VHT_CAP_RX_STBC_3SS); |
| vht_cap->vht_max_ampdu_len_exp = |
| (value & WMI_VHT_CAP_MAX_AMPDU_LEN_EXP) >> |
| WMI_VHT_CAP_MAX_AMPDU_LEN_EXP_SHIFT; |
| vht_cap->vht_su_bformer = value & WMI_VHT_CAP_SU_BFORMER; |
| vht_cap->vht_su_bformee = value & WMI_VHT_CAP_SU_BFORMEE; |
| vht_cap->vht_mu_bformer = value & WMI_VHT_CAP_MU_BFORMER; |
| vht_cap->vht_mu_bformee = value & WMI_VHT_CAP_MU_BFORMEE; |
| vht_cap->vht_txop_ps = value & WMI_VHT_CAP_TXOP_PS; |
| } else { |
| if (value & WMI_VHT_CAP_MAX_MPDU_LEN_11454) |
| tmp = WMI_VHT_CAP_MAX_MPDU_LEN_11454; |
| else if (value & WMI_VHT_CAP_MAX_MPDU_LEN_7935) |
| tmp = WMI_VHT_CAP_MAX_MPDU_LEN_7935; |
| else |
| tmp = 0; |
| vht_cap->vht_max_mpdu = QDF_MIN(vht_cap->vht_max_mpdu, tmp); |
| |
| if ((value & WMI_VHT_CAP_CH_WIDTH_80P80_160MHZ)) { |
| tmp = (1 << eHT_CHANNEL_WIDTH_80P80MHZ) | |
| (1 << eHT_CHANNEL_WIDTH_160MHZ); |
| } else if (value & WMI_VHT_CAP_CH_WIDTH_160MHZ) { |
| tmp = 1 << eHT_CHANNEL_WIDTH_160MHZ; |
| } else { |
| tmp = 1 << eHT_CHANNEL_WIDTH_80MHZ; |
| } |
| vht_cap->supp_chan_width = |
| QDF_MAX(vht_cap->supp_chan_width, tmp); |
| vht_cap->vht_rx_ldpc = QDF_MIN(vht_cap->vht_rx_ldpc, |
| value & WMI_VHT_CAP_RX_LDPC); |
| vht_cap->vht_short_gi_80 = QDF_MAX(vht_cap->vht_short_gi_80, |
| value & WMI_VHT_CAP_SGI_80MHZ); |
| vht_cap->vht_short_gi_160 = QDF_MAX(vht_cap->vht_short_gi_160, |
| value & WMI_VHT_CAP_SGI_160MHZ); |
| vht_cap->vht_tx_stbc = QDF_MAX(vht_cap->vht_tx_stbc, |
| value & WMI_VHT_CAP_TX_STBC); |
| vht_cap->vht_rx_stbc = QDF_MIN(vht_cap->vht_rx_stbc, |
| (value & WMI_VHT_CAP_RX_STBC_1SS) | |
| (value & WMI_VHT_CAP_RX_STBC_2SS) | |
| (value & WMI_VHT_CAP_RX_STBC_3SS)); |
| vht_cap->vht_max_ampdu_len_exp = |
| QDF_MIN(vht_cap->vht_max_ampdu_len_exp, |
| (value & WMI_VHT_CAP_MAX_AMPDU_LEN_EXP) >> |
| WMI_VHT_CAP_MAX_AMPDU_LEN_EXP_SHIFT); |
| vht_cap->vht_su_bformer = QDF_MAX(vht_cap->vht_su_bformer, |
| value & WMI_VHT_CAP_SU_BFORMER); |
| vht_cap->vht_su_bformee = QDF_MAX(vht_cap->vht_su_bformee, |
| value & WMI_VHT_CAP_SU_BFORMEE); |
| vht_cap->vht_mu_bformer = QDF_MAX(vht_cap->vht_mu_bformer, |
| value & WMI_VHT_CAP_MU_BFORMER); |
| vht_cap->vht_mu_bformee = QDF_MAX(vht_cap->vht_mu_bformee, |
| value & WMI_VHT_CAP_MU_BFORMEE); |
| vht_cap->vht_txop_ps = QDF_MIN(vht_cap->vht_txop_ps, |
| value & WMI_VHT_CAP_TXOP_PS); |
| } |
| } |
| |
| /** |
| * wma_update_target_ext_vht_cap() - Update VHT caps with given extended cap |
| * @wma_handle: pointer to wma_handle |
| * @vht_cap: VHT cap structure to be filled |
| * |
| * This function loop through each hardware mode and for each hardware mode |
| * again it loop through each MAC/PHY and pull the caps 2G and 5G specific |
| * VHT caps and derives the final cap. |
| * |
| * Return: none |
| * |
| */ |
| static void wma_update_target_ext_vht_cap(t_wma_handle *wma_handle, |
| struct wma_tgt_vht_cap *vht_cap) |
| { |
| int i, j = 0, max_mac; |
| uint32_t vht_cap_info_2g, vht_cap_info_5g; |
| struct wma_tgt_vht_cap tmp_vht_cap = {0}, tmp_cap = {0}; |
| struct extended_caps *phy_caps; |
| WMI_MAC_PHY_CAPABILITIES *mac_cap; |
| |
| /* |
| * for legacy device extended cap might not even come, so in that case |
| * don't overwrite legacy values |
| */ |
| if (!wma_handle || |
| (0 == wma_handle->phy_caps.num_hw_modes.num_hw_modes)) { |
| WMA_LOGD("%s: No extended VHT cap for current SOC", __func__); |
| return; |
| } |
| |
| phy_caps = &wma_handle->phy_caps; |
| for (i = 0; i < phy_caps->num_hw_modes.num_hw_modes; i++) { |
| if (phy_caps->each_hw_mode_cap[i].phy_id_map == PHY1_PHY2) |
| max_mac = j + 2; |
| else |
| max_mac = j + 1; |
| for ( ; j < max_mac; j++) { |
| mac_cap = &phy_caps->each_phy_cap_per_hwmode[j]; |
| vht_cap_info_2g = mac_cap->vht_cap_info_2G; |
| vht_cap_info_5g = mac_cap->vht_cap_info_5G; |
| if (vht_cap_info_2g) |
| wma_derive_ext_vht_cap(wma_handle, &tmp_vht_cap, |
| vht_cap_info_2g); |
| if (vht_cap_info_5g) |
| wma_derive_ext_vht_cap(wma_handle, &tmp_vht_cap, |
| vht_cap_info_5g); |
| } |
| } |
| |
| if (qdf_mem_cmp(&tmp_cap, &tmp_vht_cap, |
| sizeof(struct wma_tgt_vht_cap))) { |
| qdf_mem_copy(vht_cap, &tmp_vht_cap, |
| sizeof(struct wma_tgt_vht_cap)); |
| } |
| |
| WMA_LOGD("%s: [ext vhtcap] max_mpdu %d supp_chan_width %x rx_ldpc %x\n" |
| "short_gi_80 %x tx_stbc %x rx_stbc %x txop_ps %x\n" |
| "su_bformee %x mu_bformee %x max_ampdu_len_exp %d", __func__, |
| vht_cap->vht_max_mpdu, vht_cap->supp_chan_width, |
| vht_cap->vht_rx_ldpc, vht_cap->vht_short_gi_80, |
| vht_cap->vht_tx_stbc, vht_cap->vht_rx_stbc, |
| vht_cap->vht_txop_ps, vht_cap->vht_su_bformee, |
| vht_cap->vht_mu_bformee, vht_cap->vht_max_ampdu_len_exp); |
| } |
| |
| /** |
| * wma_update_ra_rate_limit() - update wma config |
| * @wma_handle: wma handle |
| * @cfg: target config |
| * |
| * Return: none |
| */ |
| #ifdef FEATURE_WLAN_RA_FILTERING |
| static void wma_update_ra_rate_limit(tp_wma_handle wma_handle, |
| struct wma_tgt_cfg *cfg) |
| { |
| cfg->is_ra_rate_limit_enabled = wma_handle->IsRArateLimitEnabled; |
| } |
| #else |
| static void wma_update_ra_rate_limit(tp_wma_handle wma_handle, |
| struct wma_tgt_cfg *cfg) |
| { |
| } |
| #endif |
| |
| /** |
| * wma_update_hdd_band_cap() - update band cap which hdd understands |
| * @supported_band: supported band which has been given by FW |
| * @tgt_cfg: target configuration to be updated |
| * |
| * Convert WMA given supported band to enum which HDD understands |
| * |
| * Return: None |
| */ |
| static void wma_update_hdd_band_cap(WMI_PHY_CAPABILITY supported_band, |
| struct wma_tgt_cfg *tgt_cfg) |
| { |
| switch (supported_band) { |
| case WMI_11G_CAPABILITY: |
| case WMI_11NG_CAPABILITY: |
| tgt_cfg->band_cap = eCSR_BAND_24; |
| break; |
| case WMI_11A_CAPABILITY: |
| case WMI_11NA_CAPABILITY: |
| case WMI_11AC_CAPABILITY: |
| tgt_cfg->band_cap = eCSR_BAND_5G; |
| break; |
| case WMI_11AG_CAPABILITY: |
| case WMI_11NAG_CAPABILITY: |
| default: |
| tgt_cfg->band_cap = eCSR_BAND_ALL; |
| } |
| } |
| |
| /** |
| * wma_update_hdd_cfg() - update HDD config |
| * @wma_handle: wma handle |
| * |
| * Return: none |
| */ |
| static void wma_update_hdd_cfg(tp_wma_handle wma_handle) |
| { |
| struct wma_tgt_cfg tgt_cfg; |
| void *hdd_ctx = cds_get_context(QDF_MODULE_ID_HDD); |
| |
| qdf_mem_zero(&tgt_cfg, sizeof(struct wma_tgt_cfg)); |
| |
| tgt_cfg.sub_20_support = wma_handle->sub_20_support; |
| tgt_cfg.reg_domain = wma_handle->reg_cap.eeprom_rd; |
| tgt_cfg.eeprom_rd_ext = wma_handle->reg_cap.eeprom_rd_ext; |
| |
| tgt_cfg.max_intf_count = wma_handle->wlan_resource_config.num_vdevs; |
| |
| qdf_mem_copy(tgt_cfg.hw_macaddr.bytes, wma_handle->hwaddr, |
| ATH_MAC_LEN); |
| |
| wma_update_target_services(wma_handle, &tgt_cfg.services); |
| wma_update_target_ht_cap(wma_handle, &tgt_cfg.ht_cap); |
| wma_update_target_vht_cap(wma_handle, &tgt_cfg.vht_cap); |
| /* |
| * This will overwrite the structure filled by wma_update_target_ht_cap |
| * and wma_update_target_vht_cap APIs. |
| */ |
| wma_update_target_ext_ht_cap(wma_handle, &tgt_cfg.ht_cap); |
| wma_update_target_ext_vht_cap(wma_handle, &tgt_cfg.vht_cap); |
| |
| wma_update_target_ext_he_cap(wma_handle, &tgt_cfg); |
| |
| tgt_cfg.target_fw_version = wma_handle->target_fw_version; |
| tgt_cfg.target_fw_vers_ext = wma_handle->target_fw_vers_ext; |
| #ifdef WLAN_FEATURE_LPSS |
| tgt_cfg.lpss_support = wma_handle->lpss_support; |
| #endif /* WLAN_FEATURE_LPSS */ |
| tgt_cfg.ap_arpns_support = wma_handle->ap_arpns_support; |
| tgt_cfg.bpf_enabled = wma_handle->bpf_enabled; |
| tgt_cfg.dfs_cac_offload = wma_handle->dfs_cac_offload; |
| tgt_cfg.rcpi_enabled = wma_handle->rcpi_enabled; |
| wma_update_ra_rate_limit(wma_handle, &tgt_cfg); |
| wma_update_hdd_band_cap(wma_handle->phy_capability, &tgt_cfg); |
| tgt_cfg.fine_time_measurement_cap = |
| wma_handle->fine_time_measurement_cap; |
| tgt_cfg.wmi_max_len = wmi_get_max_msg_len(wma_handle->wmi_handle) |
| - WMI_TLV_HEADROOM; |
| wma_setup_egap_support(&tgt_cfg, wma_handle); |
| tgt_cfg.tx_bfee_8ss_enabled = wma_handle->tx_bfee_8ss_enabled; |
| tgt_cfg.fw_mem_dump_enabled = wma_handle->fw_mem_dump_enabled; |
| wma_update_hdd_cfg_ndp(wma_handle, &tgt_cfg); |
| wma_handle->tgt_cfg_update_cb(hdd_ctx, &tgt_cfg); |
| } |
| |
| /** |
| * wma_setup_wmi_init_msg() - fill wmi init message buffer |
| * @wma_handle: wma handle |
| * @ev: ready event fixed params |
| * @param_buf: redy event TLVs |
| * @len: buffer length |
| * |
| * Return: wmi buffer or NULL for error |
| */ |
| static int wma_setup_wmi_init_msg(tp_wma_handle wma_handle, |
| wmi_service_ready_event_fixed_param *ev, |
| WMI_SERVICE_READY_EVENTID_param_tlvs *param_buf) |
| { |
| wlan_host_mem_req *ev_mem_reqs; |
| wmi_abi_version my_vers; |
| wmi_abi_version host_abi_vers; |
| int num_whitelist; |
| uint16_t idx; |
| uint32_t num_units; |
| |
| ev_mem_reqs = param_buf->mem_reqs; |
| |
| /* allocate memory requested by FW */ |
| if (ev->num_mem_reqs > WMI_MAX_MEM_REQS) { |
| QDF_ASSERT(0); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| for (idx = 0; idx < ev->num_mem_reqs; ++idx) { |
| num_units = ev_mem_reqs[idx].num_units; |
| if (ev_mem_reqs[idx].num_unit_info & NUM_UNITS_IS_NUM_PEERS) { |
| /* |
| * number of units to allocate is number |
| * of peers, 1 extra for self peer on |
| * target. this needs to be fied, host |
| * and target can get out of sync |
| */ |
| num_units = wma_handle->wlan_resource_config.num_peers + |
| 1; |
| } |
| WMA_LOGD("idx %d req %d num_units %d num_unit_info %d unit size %d actual units %d ", |
| idx, ev_mem_reqs[idx].req_id, |
| ev_mem_reqs[idx].num_units, |
| ev_mem_reqs[idx].num_unit_info, |
| ev_mem_reqs[idx].unit_size, num_units); |
| wma_alloc_host_mem(wma_handle, ev_mem_reqs[idx].req_id, |
| num_units, ev_mem_reqs[idx].unit_size); |
| } |
| |
| qdf_mem_copy(&wma_handle->target_abi_vers, |
| ¶m_buf->fixed_param->fw_abi_vers, |
| sizeof(wmi_abi_version)); |
| num_whitelist = sizeof(version_whitelist) / |
| sizeof(wmi_whitelist_version_info); |
| my_vers.abi_version_0 = WMI_ABI_VERSION_0; |
| my_vers.abi_version_1 = WMI_ABI_VERSION_1; |
| my_vers.abi_version_ns_0 = WMI_ABI_VERSION_NS_0; |
| my_vers.abi_version_ns_1 = WMI_ABI_VERSION_NS_1; |
| my_vers.abi_version_ns_2 = WMI_ABI_VERSION_NS_2; |
| my_vers.abi_version_ns_3 = WMI_ABI_VERSION_NS_3; |
| |
| wmi_cmp_and_set_abi_version(num_whitelist, version_whitelist, |
| &my_vers, |
| ¶m_buf->fixed_param->fw_abi_vers, |
| &host_abi_vers); |
| |
| qdf_mem_copy(&wma_handle->final_abi_vers, &host_abi_vers, |
| sizeof(wmi_abi_version)); |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_dump_dbs_hw_mode() - Print the DBS HW modes |
| * @wma_handle: WMA handle |
| * |
| * Prints the DBS HW modes sent by the FW as part |
| * of WMI ready event |
| * |
| * Return: None |
| */ |
| static void wma_dump_dbs_hw_mode(tp_wma_handle wma_handle) |
| { |
| uint32_t i, param; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| return; |
| } |
| |
| for (i = 0; i < wma_handle->num_dbs_hw_modes; i++) { |
| param = wma_handle->hw_mode.hw_mode_list[i]; |
| WMA_LOGD("%s:[%d]-MAC0: tx_ss:%d rx_ss:%d bw_idx:%d", |
| __func__, i, |
| WMA_HW_MODE_MAC0_TX_STREAMS_GET(param), |
| WMA_HW_MODE_MAC0_RX_STREAMS_GET(param), |
| WMA_HW_MODE_MAC0_BANDWIDTH_GET(param)); |
| WMA_LOGD("%s:[%d]-MAC1: tx_ss:%d rx_ss:%d bw_idx:%d", |
| __func__, i, |
| WMA_HW_MODE_MAC1_TX_STREAMS_GET(param), |
| WMA_HW_MODE_MAC1_RX_STREAMS_GET(param), |
| WMA_HW_MODE_MAC1_BANDWIDTH_GET(param)); |
| WMA_LOGD("%s:[%d] DBS:%d SBS:%d", __func__, i, |
| WMA_HW_MODE_DBS_MODE_GET(param), |
| WMA_HW_MODE_SBS_MODE_GET(param)); |
| } |
| policy_mgr_dump_dbs_hw_mode(wma_handle->psoc); |
| } |
| |
| /** |
| * wma_init_scan_fw_mode_config() - Initialize scan/fw mode config |
| * @wma_handle: WMA handle |
| * @scan_config: Scam mode configuration |
| * @fw_config: FW mode configuration |
| * |
| * Enables all the valid bits of concurrent_scan_config_bits and |
| * fw_mode_config_bits. |
| * |
| * Return: None |
| */ |
| static void wma_init_scan_fw_mode_config(tp_wma_handle wma_handle, |
| uint32_t scan_config, |
| uint32_t fw_config) |
| { |
| WMA_LOGD("%s: Enter", __func__); |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| return; |
| } |
| |
| policy_mgr_init_dbs_config(wma_handle->psoc, scan_config, fw_config); |
| } |
| |
| /** |
| * wma_update_ra_limit() - update ra limit based on bpf filter |
| * enabled or not |
| * @handle: wma handle |
| * |
| * Return: none |
| */ |
| #ifdef FEATURE_WLAN_RA_FILTERING |
| static void wma_update_ra_limit(tp_wma_handle wma_handle) |
| { |
| if (wma_handle->bpf_enabled) |
| wma_handle->IsRArateLimitEnabled = false; |
| } |
| #else |
| static void wma_update_ra__limit(tp_wma_handle handle) |
| { |
| } |
| #endif |
| |
| /** |
| * wma_rx_service_ready_event() - event handler to process |
| * wmi rx sevice ready event. |
| * @handle: wma handle |
| * @cmd_param_info: command params info |
| * |
| * Return: none |
| */ |
| int wma_rx_service_ready_event(void *handle, uint8_t *cmd_param_info, |
| uint32_t length) |
| { |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| struct wma_target_cap target_cap; |
| WMI_SERVICE_READY_EVENTID_param_tlvs *param_buf; |
| wmi_service_ready_event_fixed_param *ev; |
| int status; |
| uint32_t *ev_wlan_dbs_hw_mode_list; |
| QDF_STATUS ret; |
| void *soc = cds_get_context(QDF_MODULE_ID_SOC); |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| param_buf = (WMI_SERVICE_READY_EVENTID_param_tlvs *) cmd_param_info; |
| if (!(handle && param_buf)) { |
| WMA_LOGE("%s: Invalid arguments", __func__); |
| return -EINVAL; |
| } |
| |
| ev = param_buf->fixed_param; |
| if (!ev) { |
| WMA_LOGE("%s: Invalid buffer", __func__); |
| return -EINVAL; |
| } |
| |
| WMA_LOGD("WMA <-- WMI_SERVICE_READY_EVENTID"); |
| |
| if (ev->num_dbs_hw_modes > param_buf->num_wlan_dbs_hw_mode_list) { |
| WMA_LOGE("FW dbs_hw_mode entry %d more than value %d in TLV hdr", |
| ev->num_dbs_hw_modes, |
| param_buf->num_wlan_dbs_hw_mode_list); |
| return -EINVAL; |
| } |
| |
| wma_handle->num_dbs_hw_modes = ev->num_dbs_hw_modes; |
| ev_wlan_dbs_hw_mode_list = param_buf->wlan_dbs_hw_mode_list; |
| wma_handle->hw_mode.hw_mode_list = |
| qdf_mem_malloc(sizeof(*wma_handle->hw_mode.hw_mode_list) * |
| wma_handle->num_dbs_hw_modes); |
| if (!wma_handle->hw_mode.hw_mode_list) { |
| WMA_LOGE("%s: Memory allocation failed for DBS", __func__); |
| /* Continuing with the rest of the processing */ |
| } |
| qdf_mem_copy(wma_handle->hw_mode.hw_mode_list, |
| ev_wlan_dbs_hw_mode_list, |
| (sizeof(*wma_handle->hw_mode.hw_mode_list) * |
| wma_handle->num_dbs_hw_modes)); |
| policy_mgr_init_dbs_hw_mode(wma_handle->psoc, |
| ev->num_dbs_hw_modes, ev_wlan_dbs_hw_mode_list); |
| wma_dump_dbs_hw_mode(wma_handle); |
| |
| /* Initializes the fw_mode and scan_config to zero. |
| * If ext service ready event is present it will set |
| * the actual values of these two params. |
| * This is to ensure that no garbage values would be |
| * present in the absence of ext service ready event. |
| */ |
| wma_init_scan_fw_mode_config(wma_handle, 0, 0); |
| |
| wma_handle->phy_capability = ev->phy_capability; |
| wma_handle->max_frag_entry = ev->max_frag_entry; |
| wma_handle->num_rf_chains = ev->num_rf_chains; |
| qdf_mem_copy(&wma_handle->reg_cap, param_buf->hal_reg_capabilities, |
| sizeof(HAL_REG_CAPABILITIES)); |
| wma_handle->ht_cap_info = ev->ht_cap_info; |
| wma_handle->vht_cap_info = ev->vht_cap_info; |
| wma_handle->vht_supp_mcs = ev->vht_supp_mcs; |
| wma_handle->num_rf_chains = ev->num_rf_chains; |
| |
| wma_handle->target_fw_version = ev->fw_build_vers; |
| wma_handle->new_hw_mode_index = ev->default_dbs_hw_mode_index; |
| policy_mgr_update_new_hw_mode_index(wma_handle->psoc, |
| ev->default_dbs_hw_mode_index); |
| wma_handle->fine_time_measurement_cap = ev->wmi_fw_sub_feat_caps; |
| |
| WMA_LOGD("%s: Firmware default hw mode index : %d", |
| __func__, ev->default_dbs_hw_mode_index); |
| WMA_LOGI("%s: Firmware build version : %08x", |
| __func__, ev->fw_build_vers); |
| WMA_LOGD("FW fine time meas cap: 0x%x", ev->wmi_fw_sub_feat_caps); |
| |
| if (ev->hw_bd_id) { |
| wma_handle->hw_bd_id = ev->hw_bd_id; |
| qdf_mem_copy(wma_handle->hw_bd_info, |
| ev->hw_bd_info, sizeof(ev->hw_bd_info)); |
| |
| WMA_LOGI("%s: Board version: %x.%x", |
| __func__, |
| wma_handle->hw_bd_info[0], wma_handle->hw_bd_info[1]); |
| } else { |
| wma_handle->hw_bd_id = 0; |
| qdf_mem_zero(wma_handle->hw_bd_info, |
| sizeof(wma_handle->hw_bd_info)); |
| WMA_LOGW("%s: Board version is unknown!", __func__); |
| } |
| |
| /* TODO: Recheck below line to dump service ready event */ |
| /* dbg_print_wmi_service_11ac(ev); */ |
| |
| /* wmi service is ready */ |
| qdf_mem_copy(wma_handle->wmi_service_bitmap, |
| param_buf->wmi_service_bitmap, |
| sizeof(wma_handle->wmi_service_bitmap)); |
| |
| cdp_cfg_tx_set_is_mgmt_over_wmi_enabled(soc, |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_MGMT_TX_WMI)); |
| cdp_set_desc_global_pool_size(soc, ev->num_msdu_desc); |
| /* SWBA event handler for beacon transmission */ |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_HOST_SWBA_EVENTID, |
| wma_beacon_swba_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register swba beacon event cb"); |
| return -EINVAL; |
| } |
| #ifdef WLAN_FEATURE_LPSS |
| wma_handle->lpss_support = |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_LPASS); |
| #endif /* WLAN_FEATURE_LPSS */ |
| |
| /* |
| * This Service bit is added to check for ARP/NS Offload |
| * support for LL/HL targets |
| */ |
| wma_handle->ap_arpns_support = |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_AP_ARPNS_OFFLOAD); |
| |
| wma_handle->bpf_enabled = (wma_handle->bpf_packet_filter_enable && |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_BPF_OFFLOAD)); |
| wma_update_ra_limit(wma_handle); |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_CSA_OFFLOAD)) { |
| WMA_LOGD("%s: FW support CSA offload capability", __func__); |
| status = wmi_unified_register_event_handler( |
| wma_handle->wmi_handle, |
| WMI_CSA_HANDLING_EVENTID, |
| wma_csa_offload_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register CSA offload event cb"); |
| return -EINVAL; |
| } |
| } |
| |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_MGMT_TX_WMI)) { |
| WMA_LOGD("Firmware supports management TX over WMI,use WMI interface instead of HTT for management Tx"); |
| /* |
| * Register Tx completion event handler for MGMT Tx over WMI |
| * case |
| */ |
| status = wmi_unified_register_event_handler( |
| wma_handle->wmi_handle, |
| WMI_MGMT_TX_COMPLETION_EVENTID, |
| wma_mgmt_tx_completion_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register MGMT over WMI completion handler"); |
| return -EINVAL; |
| } |
| |
| status = wmi_unified_register_event_handler( |
| wma_handle->wmi_handle, |
| WMI_MGMT_TX_BUNDLE_COMPLETION_EVENTID, |
| wma_mgmt_tx_bundle_completion_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register MGMT over WMI completion handler"); |
| return -EINVAL; |
| } |
| |
| } else { |
| WMA_LOGE("FW doesnot support WMI_SERVICE_MGMT_TX_WMI, Use HTT interface for Management Tx"); |
| } |
| #ifdef WLAN_FEATURE_GTK_OFFLOAD |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_GTK_OFFLOAD)) { |
| status = |
| wmi_unified_register_event_handler( |
| wma_handle->wmi_handle, |
| WMI_GTK_OFFLOAD_STATUS_EVENTID, |
| target_if_pmo_gtk_offload_status_event, |
| WMA_RX_WORK_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register GTK offload event cb"); |
| return -EINVAL; |
| } |
| } |
| #endif /* WLAN_FEATURE_GTK_OFFLOAD */ |
| |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_FW_MEM_DUMP_SUPPORT)) |
| wma_handle->fw_mem_dump_enabled = true; |
| else |
| wma_handle->fw_mem_dump_enabled = false; |
| |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_TBTTOFFSET_UPDATE_EVENTID, |
| wma_tbttoffset_update_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register WMI_TBTTOFFSET_UPDATE_EVENTID callback"); |
| return -EINVAL; |
| } |
| |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_RCPI_SUPPORT)) { |
| /* register for rcpi response event */ |
| status = wmi_unified_register_event_handler( |
| wma_handle->wmi_handle, |
| WMI_UPDATE_RCPI_EVENTID, |
| wma_rcpi_event_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status) { |
| WMA_LOGE("Failed to register RCPI event handler"); |
| return -EINVAL; |
| } |
| wma_handle->rcpi_enabled = true; |
| } |
| |
| /* mac_id is replaced with pdev_id in converged firmware to have |
| * multi-radio support. In order to maintain backward compatibility |
| * with old fw, host needs to check WMI_SERVICE_DEPRECATED_REPLACE |
| * in service bitmap from FW and host needs to set use_pdev_id in |
| * wmi_resource_config to true. If WMI_SERVICE_DEPRECATED_REPLACE |
| * service is not set, then host shall not expect MAC ID from FW in |
| * VDEV START RESPONSE event and host shall use PDEV ID. |
| */ |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_DEPRECATED_REPLACE)) |
| wma_handle->wlan_resource_config.use_pdev_id = true; |
| else |
| wma_handle->wlan_resource_config.use_pdev_id = false; |
| |
| wma_handle->wlan_resource_config.max_num_dbs_scan_duty_cycle = |
| CDS_DBS_SCAN_CLIENTS_MAX; |
| |
| /* register the Enhanced Green AP event handler */ |
| wma_register_egap_event_handle(wma_handle); |
| |
| /* Initialize the log supported event handler */ |
| status = wmi_unified_register_event_handler(wma_handle->wmi_handle, |
| WMI_DIAG_EVENT_LOG_SUPPORTED_EVENTID, |
| wma_log_supported_evt_handler, |
| WMA_RX_SERIALIZER_CTX); |
| if (status != QDF_STATUS_SUCCESS) { |
| WMA_LOGE("Failed to register log supported event cb"); |
| return -EINVAL; |
| } |
| |
| cdp_mark_first_wakeup_packet(soc, |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_MARK_FIRST_WAKEUP_PACKET)); |
| wma_handle->dfs_cac_offload = |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_DFS_PHYERR_OFFLOAD); |
| wma_handle->nan_datapath_enabled = |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_NAN_DATA); |
| qdf_mem_copy(target_cap.wmi_service_bitmap, |
| param_buf->wmi_service_bitmap, |
| sizeof(wma_handle->wmi_service_bitmap)); |
| target_cap.wlan_resource_config = wma_handle->wlan_resource_config; |
| wma_update_fw_config(wma_handle, &target_cap); |
| qdf_mem_copy(wma_handle->wmi_service_bitmap, |
| target_cap.wmi_service_bitmap, |
| sizeof(wma_handle->wmi_service_bitmap)); |
| wma_handle->wlan_resource_config = target_cap.wlan_resource_config; |
| |
| status = wmi_unified_save_fw_version_cmd(wma_handle->wmi_handle, |
| param_buf); |
| if (status != EOK) { |
| WMA_LOGE("Failed to send WMI_INIT_CMDID command"); |
| return -EINVAL; |
| } |
| |
| status = wma_setup_wmi_init_msg(wma_handle, ev, param_buf); |
| if (status != EOK) { |
| WMA_LOGE("Failed to setup for wma init command"); |
| return -EINVAL; |
| } |
| |
| /* A host, which supports WMI_SERVICE_READY_EXT_EVENTID, would need to |
| * check the WMI_SERVICE_READY message for an "extension" flag, and if |
| * this flag is set, then hold off on sending the WMI_INIT message until |
| * WMI_SERVICE_READY_EXT_EVENTID is received. |
| */ |
| if (!WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_EXT_MSG)) { |
| /* No service extended message support. |
| * Send INIT command immediately |
| */ |
| struct wmi_init_cmd_param init_param = {0}; |
| init_param.res_cfg = &wma_handle->wlan_resource_config; |
| init_param.num_mem_chunks = wma_handle->num_mem_chunks; |
| init_param.mem_chunks = wma_handle->mem_chunks; |
| init_param.hw_mode_id = WMI_HOST_HW_MODE_MAX; |
| WMA_LOGA("WMA --> WMI_INIT_CMDID"); |
| status = wmi_unified_init_cmd_send(wma_handle->wmi_handle, |
| &init_param); |
| if (status != EOK) { |
| WMA_LOGE("Failed to send WMI_INIT_CMDID command"); |
| return -EINVAL; |
| } |
| } else { |
| ret = qdf_mc_timer_start(&wma_handle->service_ready_ext_timer, |
| WMA_SERVICE_READY_EXT_TIMEOUT); |
| if (!QDF_IS_STATUS_SUCCESS(ret)) |
| WMA_LOGE("Failed to start the service ready ext timer"); |
| |
| WMA_LOGD("%s: WMA waiting for WMI_SERVICE_READY_EXT_EVENTID", |
| __func__); |
| } |
| |
| if (WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_8SS_TX_BFEE)) |
| wma_handle->tx_bfee_8ss_enabled = true; |
| else |
| wma_handle->tx_bfee_8ss_enabled = false; |
| |
| return 0; |
| } |
| |
| /** |
| * wma_get_phyid_for_given_band() - to get phyid for band |
| * |
| * @wma_handle: Pointer to wma handle |
| * @map: Pointer to map which is derived from hw mode & has mapping between |
| * hw mode and available PHYs for that hw mode. |
| * @band: enum value of for 2G or 5G band |
| * @phyid: Pointer to phyid which needs to be filled |
| * |
| * This API looks in to the map to find out which particular phy supports |
| * provided band and return the idx (also called phyid) of that phy. Caller |
| * use this phyid to fetch various caps of that phy |
| * |
| * Return: QDF_STATUS |
| */ |
| static QDF_STATUS wma_get_phyid_for_given_band( |
| tp_wma_handle wma_handle, |
| struct hw_mode_idx_to_mac_cap_idx *map, |
| enum cds_band_type band, uint8_t *phyid) |
| { |
| uint8_t idx, i; |
| WMI_MAC_PHY_CAPABILITIES *cap; |
| |
| if (!wma_handle) { |
| WMA_LOGE("Invalid wma handle"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| if (!map) { |
| WMA_LOGE("Invalid given map"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| idx = map->mac_cap_idx; |
| *phyid = idx; |
| |
| for (i = 0; i < map->num_of_macs; i++) { |
| cap = &wma_handle->phy_caps.each_phy_cap_per_hwmode[idx + i]; |
| if ((band == CDS_BAND_2GHZ) && |
| (WLAN_2G_CAPABILITY == cap->supported_bands)) { |
| *phyid = idx + i; |
| WMA_LOGD("Select 2G capable phyid[%d]", *phyid); |
| return QDF_STATUS_SUCCESS; |
| } else if ((band == CDS_BAND_5GHZ) && |
| (WLAN_5G_CAPABILITY == cap->supported_bands)) { |
| *phyid = idx + i; |
| WMA_LOGD("Select 5G capable phyid[%d]", *phyid); |
| return QDF_STATUS_SUCCESS; |
| } |
| } |
| WMA_LOGD("Using default single hw mode phyid[%d]", *phyid); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_get_caps_for_phyidx_hwmode() - to fetch caps for given hw mode and band |
| * @caps_per_phy: Pointer to capabilities structure which needs to be filled |
| * @hw_mode: Provided hardware mode |
| * @band: Provide band i.e. 2G or 5G |
| * |
| * This API finds cap which suitable for provided hw mode and band. If user |
| * is provides some invalid hw mode then it will automatically falls back to |
| * default hw mode |
| * |
| * Return: QDF_STATUS |
| */ |
| QDF_STATUS wma_get_caps_for_phyidx_hwmode(struct wma_caps_per_phy *caps_per_phy, |
| enum hw_mode_dbs_capab hw_mode, enum cds_band_type band) |
| { |
| t_wma_handle *wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| struct hw_mode_idx_to_mac_cap_idx *map; |
| WMI_MAC_PHY_CAPABILITIES *phy_cap; |
| uint8_t phyid, our_hw_mode = hw_mode; |
| |
| if (!wma_handle) { |
| WMA_LOGE("Invalid wma handle"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| if (0 == wma_handle->phy_caps.num_hw_modes.num_hw_modes) { |
| WMA_LOGD("Invalid number of hw modes, use legacy HT/VHT caps"); |
| caps_per_phy->ht_2g = wma_handle->ht_cap_info; |
| caps_per_phy->ht_5g = wma_handle->ht_cap_info; |
| caps_per_phy->vht_2g = wma_handle->vht_cap_info; |
| caps_per_phy->vht_5g = wma_handle->vht_cap_info; |
| /* legacy platform doesn't support HE IE */ |
| caps_per_phy->he_2g = 0; |
| caps_per_phy->he_5g = 0; |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| if (!policy_mgr_is_dbs_enable(wma_handle->psoc)) |
| our_hw_mode = HW_MODE_DBS_NONE; |
| |
| if (!caps_per_phy) { |
| WMA_LOGE("Invalid caps pointer"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| map = &wma_handle->phy_caps.hw_mode_to_mac_cap_map[our_hw_mode]; |
| |
| if (QDF_STATUS_SUCCESS != |
| wma_get_phyid_for_given_band(wma_handle, map, band, &phyid)) { |
| WMA_LOGE("Invalid phyid"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| phy_cap = &wma_handle->phy_caps.each_phy_cap_per_hwmode[phyid]; |
| |
| caps_per_phy->ht_2g = phy_cap->ht_cap_info_2G; |
| caps_per_phy->ht_5g = phy_cap->ht_cap_info_5G; |
| caps_per_phy->vht_2g = phy_cap->vht_cap_info_2G; |
| caps_per_phy->vht_5g = phy_cap->vht_cap_info_5G; |
| caps_per_phy->he_2g = phy_cap->he_cap_info_2G; |
| caps_per_phy->he_5g = phy_cap->he_cap_info_5G; |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_is_rx_ldpc_supported_for_channel() - to find out if ldpc is supported |
| * |
| * @channel: Channel number for which it needs to check if rx ldpc is enabled |
| * |
| * This API takes channel number as argument and takes default hw mode as DBS |
| * to check if rx LDPC support is enabled for that channel or no |
| */ |
| bool wma_is_rx_ldpc_supported_for_channel(uint32_t channel) |
| { |
| t_wma_handle *wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| struct wma_caps_per_phy caps_per_phy = {0}; |
| enum cds_band_type band; |
| bool status; |
| |
| if (!WLAN_REG_IS_24GHZ_CH(channel)) |
| band = CDS_BAND_5GHZ; |
| else |
| band = CDS_BAND_2GHZ; |
| |
| if (QDF_STATUS_SUCCESS != wma_get_caps_for_phyidx_hwmode( |
| &caps_per_phy, |
| HW_MODE_DBS, band)) { |
| return false; |
| } |
| |
| /* |
| * Legacy platforms like Rome set WMI_HT_CAP_LDPC to specify RX LDPC |
| * capability. But new platforms like Helium set WMI_HT_CAP_RX_LDPC |
| * instead. |
| */ |
| if (wma_handle->phy_caps.num_hw_modes.num_hw_modes == 0) { |
| status = (!!(caps_per_phy.ht_2g & WMI_HT_CAP_LDPC)); |
| } else { |
| if (WLAN_REG_IS_24GHZ_CH(channel)) |
| status = (!!(caps_per_phy.ht_2g & WMI_HT_CAP_RX_LDPC)); |
| else |
| status = (!!(caps_per_phy.ht_5g & WMI_HT_CAP_RX_LDPC)); |
| } |
| |
| return status; |
| } |
| |
| /** |
| * wma_print_mac_phy_capabilities() - Prints MAC PHY capabilities |
| * @cap: pointer to WMI_MAC_PHY_CAPABILITIES |
| * @index: MAC_PHY index |
| * |
| * Return: none |
| */ |
| static void wma_print_mac_phy_capabilities(WMI_MAC_PHY_CAPABILITIES *cap, |
| int index) |
| { |
| uint32_t mac_2G, mac_5G; |
| uint32_t phy_2G[WMI_MAX_HECAP_PHY_SIZE]; |
| uint32_t phy_5G[WMI_MAX_HECAP_PHY_SIZE]; |
| wmi_ppe_threshold ppet_2G, ppet_5G; |
| |
| WMA_LOGI("\t: index [%d]", index); |
| WMA_LOGI("\t: cap for hw_mode_id[%d]", cap->hw_mode_id); |
| WMA_LOGI("\t: pdev_id[%d]", cap->pdev_id); |
| WMA_LOGI("\t: phy_id[%d]", cap->phy_id); |
| WMA_LOGI("\t: supports_11b[%d]", |
| WMI_SUPPORT_11B_GET(cap->supported_flags)); |
| WMA_LOGI("\t: supports_11g[%d]", |
| WMI_SUPPORT_11G_GET(cap->supported_flags)); |
| WMA_LOGI("\t: supports_11a[%d]", |
| WMI_SUPPORT_11A_GET(cap->supported_flags)); |
| WMA_LOGI("\t: supports_11n[%d]", |
| WMI_SUPPORT_11N_GET(cap->supported_flags)); |
| WMA_LOGI("\t: supports_11ac[%d]", |
| WMI_SUPPORT_11AC_GET(cap->supported_flags)); |
| WMA_LOGI("\t: supports_11ax[%d]", |
| WMI_SUPPORT_11AX_GET(cap->supported_flags)); |
| WMA_LOGI("\t: supported_flags[%d]", cap->supported_flags); |
| WMA_LOGI("\t: supported_bands[%d]", cap->supported_bands); |
| WMA_LOGI("\t: ampdu_density[%d]", cap->ampdu_density); |
| WMA_LOGI("\t: max_bw_supported_2G[%d]", cap->max_bw_supported_2G); |
| WMA_LOGI("\t: ht_cap_info_2G[%d]", cap->ht_cap_info_2G); |
| WMA_LOGI("\t: vht_cap_info_2G[%d]", cap->vht_cap_info_2G); |
| WMA_LOGI("\t: vht_supp_mcs_2G[%d]", cap->vht_supp_mcs_2G); |
| WMA_LOGI("\t: tx_chain_mask_2G[%d]", cap->tx_chain_mask_2G); |
| WMA_LOGI("\t: rx_chain_mask_2G[%d]", cap->rx_chain_mask_2G); |
| WMA_LOGI("\t: max_bw_supported_5G[%d]", cap->max_bw_supported_5G); |
| WMA_LOGI("\t: ht_cap_info_5G[%d]", cap->ht_cap_info_5G); |
| WMA_LOGI("\t: vht_cap_info_5G[%d]", cap->vht_cap_info_5G); |
| WMA_LOGI("\t: vht_supp_mcs_5G[%d]", cap->vht_supp_mcs_5G); |
| WMA_LOGI("\t: tx_chain_mask_5G[%d]", cap->tx_chain_mask_5G); |
| WMA_LOGI("\t: rx_chain_mask_5G[%d]", cap->rx_chain_mask_5G); |
| WMA_LOGI("\t: he_cap_info_2G[%08x]", cap->he_cap_info_2G); |
| WMA_LOGI("\t: he_supp_mcs_2G[%08x]", cap->he_supp_mcs_2G); |
| WMA_LOGI("\t: he_cap_info_5G[%08x]", cap->he_cap_info_5G); |
| WMA_LOGI("\t: he_supp_mcs_5G[%08x]", cap->he_supp_mcs_5G); |
| mac_2G = cap->he_cap_info_2G; |
| mac_5G = cap->he_cap_info_5G; |
| qdf_mem_copy(phy_2G, cap->he_cap_phy_info_2G, |
| WMI_MAX_HECAP_PHY_SIZE * 4); |
| qdf_mem_copy(phy_5G, cap->he_cap_phy_info_5G, |
| WMI_MAX_HECAP_PHY_SIZE * 4); |
| ppet_2G = cap->he_ppet2G; |
| ppet_5G = cap->he_ppet5G; |
| |
| wma_print_he_mac_cap(mac_2G); |
| wma_print_he_phy_cap(phy_2G); |
| wma_print_he_ppet(&ppet_2G); |
| wma_print_he_mac_cap(mac_5G); |
| wma_print_he_phy_cap(phy_5G); |
| wma_print_he_ppet(&ppet_5G); |
| } |
| |
| /** |
| * wma_print_populate_soc_caps() - Prints all the caps populated per hw mode |
| * @wma_handle: pointer to wma_handle |
| * |
| * This function prints all the caps populater per hw mode and per PHY |
| * |
| * Return: none |
| */ |
| static void wma_print_populate_soc_caps(t_wma_handle *wma_handle) |
| { |
| int i, j = 0, max_mac; |
| WMI_MAC_PHY_CAPABILITIES *tmp; |
| |
| /* print number of hw modes */ |
| WMA_LOGD("%s: num of hw modes [%d]", __func__, |
| wma_handle->phy_caps.num_hw_modes.num_hw_modes); |
| WMA_LOGD("%s: <====== HW mode cap printing starts ======>", __func__); |
| /* print cap of each hw mode */ |
| for (i = 0; i < wma_handle->phy_caps.num_hw_modes.num_hw_modes; i++) { |
| WMA_LOGD("====>: hw mode id[%d], phy_id map[%d]", |
| wma_handle->phy_caps.each_hw_mode_cap[i].hw_mode_id, |
| wma_handle->phy_caps.each_hw_mode_cap[i].phy_id_map); |
| if (wma_handle->phy_caps.each_hw_mode_cap[i].phy_id_map == |
| PHY1_PHY2) |
| max_mac = j + 2; |
| else |
| max_mac = j + 1; |
| |
| for ( ; j < max_mac; j++) { |
| tmp = &wma_handle->phy_caps.each_phy_cap_per_hwmode[j]; |
| wma_print_mac_phy_capabilities(tmp, j); |
| } |
| } |
| WMA_LOGI("%s: <====== HW mode cap printing ends ======>\n", __func__); |
| } |
| |
| /** |
| * wma_map_wmi_channel_width_to_hw_mode_bw() - returns bandwidth |
| * in terms of hw_mode_bandwidth |
| * @width: bandwidth in terms of wmi_channel_width |
| * |
| * This function returns the bandwidth in terms of hw_mode_bandwidth. |
| * |
| * Return: BW in terms of hw_mode_bandwidth. |
| */ |
| static enum hw_mode_bandwidth wma_map_wmi_channel_width_to_hw_mode_bw( |
| wmi_channel_width width) |
| { |
| switch (width) { |
| case WMI_CHAN_WIDTH_20: |
| return HW_MODE_20_MHZ; |
| case WMI_CHAN_WIDTH_40: |
| return HW_MODE_40_MHZ; |
| case WMI_CHAN_WIDTH_80: |
| return HW_MODE_80_MHZ; |
| case WMI_CHAN_WIDTH_160: |
| return HW_MODE_160_MHZ; |
| case WMI_CHAN_WIDTH_80P80: |
| return HW_MODE_80_PLUS_80_MHZ; |
| case WMI_CHAN_WIDTH_5: |
| return HW_MODE_5_MHZ; |
| case WMI_CHAN_WIDTH_10: |
| return HW_MODE_10_MHZ; |
| default: |
| return HW_MODE_BW_NONE; |
| } |
| |
| return HW_MODE_BW_NONE; |
| } |
| |
| /** |
| * wma_get_hw_mode_params() - get TX-RX stream and bandwidth |
| * supported from the capabilities. |
| * @caps: PHY capability |
| * @info: param to store TX-RX stream and BW information |
| * |
| * This function will calculate TX-RX stream and bandwidth supported |
| * as per the PHY capability, and assign to mac_ss_bw_info. |
| * |
| * Return: none |
| */ |
| static void wma_get_hw_mode_params(WMI_MAC_PHY_CAPABILITIES *caps, |
| struct mac_ss_bw_info *info) |
| { |
| if (!caps) { |
| WMA_LOGE("%s: Invalid capabilities", __func__); |
| return; |
| } |
| |
| info->mac_tx_stream = wma_get_num_of_setbits_from_bitmask( |
| QDF_MAX(caps->tx_chain_mask_2G, |
| caps->tx_chain_mask_5G)); |
| info->mac_rx_stream = wma_get_num_of_setbits_from_bitmask( |
| QDF_MAX(caps->rx_chain_mask_2G, |
| caps->rx_chain_mask_5G)); |
| info->mac_bw = wma_map_wmi_channel_width_to_hw_mode_bw( |
| QDF_MAX(caps->max_bw_supported_2G, |
| caps->max_bw_supported_5G)); |
| } |
| |
| /** |
| * wma_set_hw_mode_params() - sets TX-RX stream, bandwidth and |
| * DBS in hw_mode_list |
| * @wma_handle: pointer to wma global structure |
| * @mac0_ss_bw_info: TX-RX streams, BW for MAC0 |
| * @mac1_ss_bw_info: TX-RX streams, BW for MAC1 |
| * @pos: refers to hw_mode_index |
| * @dbs_mode: dbs_mode for the dbs_hw_mode |
| * @sbs_mode: sbs_mode for the sbs_hw_mode |
| * |
| * This function sets TX-RX stream, bandwidth and DBS mode in |
| * hw_mode_list. |
| * |
| * Return: none |
| */ |
| static void wma_set_hw_mode_params(t_wma_handle *wma_handle, |
| struct mac_ss_bw_info mac0_ss_bw_info, |
| struct mac_ss_bw_info mac1_ss_bw_info, |
| uint32_t pos, uint32_t dbs_mode, |
| uint32_t sbs_mode) |
| { |
| WMA_HW_MODE_MAC0_TX_STREAMS_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| mac0_ss_bw_info.mac_tx_stream); |
| WMA_HW_MODE_MAC0_RX_STREAMS_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| mac0_ss_bw_info.mac_rx_stream); |
| WMA_HW_MODE_MAC0_BANDWIDTH_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| mac0_ss_bw_info.mac_bw); |
| WMA_HW_MODE_MAC1_TX_STREAMS_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| mac1_ss_bw_info.mac_tx_stream); |
| WMA_HW_MODE_MAC1_RX_STREAMS_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| mac1_ss_bw_info.mac_rx_stream); |
| WMA_HW_MODE_MAC1_BANDWIDTH_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| mac1_ss_bw_info.mac_bw); |
| WMA_HW_MODE_DBS_MODE_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| dbs_mode); |
| WMA_HW_MODE_AGILE_DFS_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| HW_MODE_AGILE_DFS_NONE); |
| WMA_HW_MODE_SBS_MODE_SET( |
| wma_handle->hw_mode.hw_mode_list[pos], |
| sbs_mode); |
| } |
| |
| /** |
| * wma_update_hw_mode_list() - updates hw_mode_list |
| * @wma_handle: pointer to wma global structure |
| * |
| * This function updates hw_mode_list with tx_streams, rx_streams, |
| * bandwidth, dbs and agile dfs for each hw_mode. |
| * |
| * Returns: 0 for success else failure. |
| */ |
| static QDF_STATUS wma_update_hw_mode_list(t_wma_handle *wma_handle) |
| { |
| struct extended_caps *phy_caps; |
| WMI_MAC_PHY_CAPABILITIES *tmp; |
| uint32_t i, hw_config_type, j = 0; |
| uint32_t dbs_mode, sbs_mode; |
| struct mac_ss_bw_info mac0_ss_bw_info = {0}; |
| struct mac_ss_bw_info mac1_ss_bw_info = {0}; |
| WMI_PHY_CAPABILITY new_supported_band = 0; |
| bool supported_band_update_failure = false; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| phy_caps = &wma_handle->phy_caps; |
| if (!phy_caps) { |
| WMA_LOGE("%s: Invalid phy capabilities", __func__); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| if (!phy_caps->num_hw_modes.num_hw_modes) { |
| WMA_LOGE("%s: Number of HW modes: %d", |
| __func__, phy_caps->num_hw_modes.num_hw_modes); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /* |
| * This list was updated as part of service ready event. Re-populate |
| * HW mode list from the device capabilities. |
| */ |
| if (wma_handle->hw_mode.hw_mode_list) { |
| qdf_mem_free(wma_handle->hw_mode.hw_mode_list); |
| wma_handle->hw_mode.hw_mode_list = NULL; |
| WMA_LOGD("%s: DBS list is freed", __func__); |
| } |
| |
| wma_handle->num_dbs_hw_modes = phy_caps->num_hw_modes.num_hw_modes; |
| wma_handle->hw_mode.hw_mode_list = |
| qdf_mem_malloc(sizeof(*wma_handle->hw_mode.hw_mode_list) * |
| wma_handle->num_dbs_hw_modes); |
| if (!wma_handle->hw_mode.hw_mode_list) { |
| WMA_LOGE("%s: Memory allocation failed for DBS", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| WMA_LOGD("%s: Updated HW mode list: Num modes:%d", |
| __func__, wma_handle->num_dbs_hw_modes); |
| |
| for (i = 0; i < wma_handle->num_dbs_hw_modes; i++) { |
| /* Update for MAC0 */ |
| tmp = &phy_caps->each_phy_cap_per_hwmode[j++]; |
| wma_get_hw_mode_params(tmp, &mac0_ss_bw_info); |
| hw_config_type = |
| phy_caps->each_hw_mode_cap[i].hw_mode_config_type; |
| dbs_mode = HW_MODE_DBS_NONE; |
| sbs_mode = HW_MODE_SBS_NONE; |
| mac1_ss_bw_info.mac_tx_stream = 0; |
| mac1_ss_bw_info.mac_rx_stream = 0; |
| mac1_ss_bw_info.mac_bw = 0; |
| if (QDF_STATUS_SUCCESS != wma_update_supported_bands(wma_handle, |
| tmp->supported_bands, |
| &new_supported_band)) |
| supported_band_update_failure = true; |
| |
| /* SBS and DBS have dual MAC. Upto 2 MACs are considered. */ |
| if ((hw_config_type == WMI_HW_MODE_DBS) || |
| (hw_config_type == WMI_HW_MODE_SBS_PASSIVE) || |
| (hw_config_type == WMI_HW_MODE_SBS)) { |
| /* Update for MAC1 */ |
| tmp = &phy_caps->each_phy_cap_per_hwmode[j++]; |
| wma_get_hw_mode_params(tmp, &mac1_ss_bw_info); |
| if (hw_config_type == WMI_HW_MODE_DBS) |
| dbs_mode = HW_MODE_DBS; |
| if ((hw_config_type == WMI_HW_MODE_SBS_PASSIVE) || |
| (hw_config_type == WMI_HW_MODE_SBS)) |
| sbs_mode = HW_MODE_SBS; |
| if (QDF_STATUS_SUCCESS != |
| wma_update_supported_bands(wma_handle, |
| tmp->supported_bands, |
| &new_supported_band)) |
| supported_band_update_failure = true; |
| } |
| |
| /* Updating HW mode list */ |
| wma_set_hw_mode_params(wma_handle, mac0_ss_bw_info, |
| mac1_ss_bw_info, i, dbs_mode, |
| sbs_mode); |
| } |
| |
| /* overwrite phy_capability which we got from service ready event */ |
| if (!supported_band_update_failure) { |
| WMA_LOGD("%s: updating supported band from old[%d] to new[%d]", |
| __func__, wma_handle->phy_capability, |
| new_supported_band); |
| wma_handle->phy_capability = new_supported_band; |
| } |
| |
| if (QDF_STATUS_SUCCESS != |
| policy_mgr_update_hw_mode_list(wma_handle->psoc, |
| phy_caps)) |
| WMA_LOGE("%s: failed to update policy manager", __func__); |
| wma_dump_dbs_hw_mode(wma_handle); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| static void wma_init_wifi_pos_dma_rings(t_wma_handle *wma_handle, |
| uint8_t num_mac, void *buf) |
| { |
| struct hif_softc *hif_ctx = cds_get_context(QDF_MODULE_ID_HIF); |
| void *hal_soc = hif_get_hal_handle(hif_ctx); |
| |
| wifi_pos_init_cir_cfr_rings(wma_handle->psoc, hal_soc, num_mac, buf); |
| } |
| |
| /** |
| * wma_populate_soc_caps() - populate entire SOC's capabilities |
| * @wma_handle: pointer to wma global structure |
| * @param_buf: pointer to param of service ready extension event from fw |
| * |
| * This API populates all capabilities of entire SOC. For example, |
| * how many number of hw modes are supported by this SOC, what are the |
| * capabilities of each phy per hw mode, what are HAL reg capabilities per |
| * phy. |
| * |
| * Return: none |
| */ |
| static void wma_populate_soc_caps(t_wma_handle *wma_handle, |
| WMI_SERVICE_READY_EXT_EVENTID_param_tlvs *param_buf) |
| { |
| int i, num_of_mac_caps = 0, tmp = 0; |
| struct extended_caps *phy_caps; |
| struct hw_mode_idx_to_mac_cap_idx *map; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| return; |
| } |
| |
| if (!param_buf) { |
| WMA_LOGE("%s: Invalid event", __func__); |
| return; |
| } |
| phy_caps = &wma_handle->phy_caps; |
| |
| /* |
| * first thing to do is to get how many number of hw modes are |
| * supported and populate in wma_handle global structure |
| */ |
| if (NULL == param_buf->soc_hw_mode_caps) { |
| WMA_LOGE("%s: Invalid number of hw modes", __func__); |
| return; |
| } |
| |
| if (param_buf->soc_hw_mode_caps->num_hw_modes > |
| MAX_NUM_HW_MODE) { |
| WMA_LOGE("Invalid num_hw_modes %u received from firmware", |
| param_buf->soc_hw_mode_caps->num_hw_modes); |
| return; |
| } |
| |
| qdf_mem_copy(&phy_caps->num_hw_modes, |
| param_buf->soc_hw_mode_caps, |
| sizeof(WMI_SOC_MAC_PHY_HW_MODE_CAPS)); |
| if (0 == phy_caps->num_hw_modes.num_hw_modes) { |
| WMA_LOGE("%s: Number of hw modes is zero", __func__); |
| return; |
| } |
| WMA_LOGD("%s: Given number of hw modes[%d]", |
| __func__, phy_caps->num_hw_modes.num_hw_modes); |
| |
| /* |
| * next thing is to allocate the memory to map hw mode to phy/mac caps |
| */ |
| phy_caps->hw_mode_to_mac_cap_map = |
| qdf_mem_malloc(phy_caps->num_hw_modes.num_hw_modes * |
| sizeof(struct hw_mode_idx_to_mac_cap_idx)); |
| if (!phy_caps->hw_mode_to_mac_cap_map) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| return; |
| } |
| |
| /* |
| * next thing is to allocate the memory for per hw caps |
| */ |
| phy_caps->each_hw_mode_cap = |
| qdf_mem_malloc(phy_caps->num_hw_modes.num_hw_modes * |
| sizeof(WMI_HW_MODE_CAPABILITIES)); |
| if (!phy_caps->each_hw_mode_cap) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| wma_cleanup_dbs_phy_caps(wma_handle); |
| return; |
| } |
| qdf_mem_copy(phy_caps->each_hw_mode_cap, |
| param_buf->hw_mode_caps, |
| phy_caps->num_hw_modes.num_hw_modes * |
| sizeof(WMI_HW_MODE_CAPABILITIES)); |
| /* |
| * next thing is to count the number of mac cap to populate per |
| * hw mode and generate map, so that our search can be done |
| * efficiently which is O(1) |
| */ |
| for (i = 0; i < phy_caps->num_hw_modes.num_hw_modes; i++) { |
| map = &phy_caps->hw_mode_to_mac_cap_map[i]; |
| if (phy_caps->each_hw_mode_cap[i].phy_id_map == PHY1_PHY2) { |
| tmp = num_of_mac_caps; |
| num_of_mac_caps = num_of_mac_caps + 2; |
| map->num_of_macs = 2; |
| } else { |
| tmp = num_of_mac_caps; |
| num_of_mac_caps = num_of_mac_caps + 1; |
| map->num_of_macs = 1; |
| } |
| map->mac_cap_idx = tmp; |
| map->hw_mode_id = phy_caps->each_hw_mode_cap[i].hw_mode_id; |
| } |
| |
| /* |
| * next thing is to populate each phy caps per hw mode |
| */ |
| phy_caps->each_phy_cap_per_hwmode = |
| qdf_mem_malloc(num_of_mac_caps * |
| sizeof(WMI_MAC_PHY_CAPABILITIES)); |
| if (!phy_caps->each_phy_cap_per_hwmode) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| wma_cleanup_dbs_phy_caps(wma_handle); |
| return; |
| } |
| qdf_mem_copy(phy_caps->each_phy_cap_per_hwmode, |
| param_buf->mac_phy_caps, |
| num_of_mac_caps * sizeof(WMI_MAC_PHY_CAPABILITIES)); |
| |
| /* |
| * next thing is to populate reg caps per phy |
| */ |
| |
| if (param_buf->soc_hal_reg_caps->num_phy > |
| MAX_NUM_PHY) { |
| WMA_LOGE("Invalid num_phy %u received from firmware", |
| param_buf->soc_hal_reg_caps->num_phy); |
| return; |
| } |
| |
| qdf_mem_copy(&phy_caps->num_phy_for_hal_reg_cap, |
| param_buf->soc_hal_reg_caps, |
| sizeof(WMI_SOC_HAL_REG_CAPABILITIES)); |
| if (phy_caps->num_phy_for_hal_reg_cap.num_phy == 0) { |
| WMA_LOGE("%s: incorrect number of phys", __func__); |
| wma_cleanup_dbs_phy_caps(wma_handle); |
| return; |
| } |
| phy_caps->each_phy_hal_reg_cap = |
| qdf_mem_malloc(phy_caps->num_phy_for_hal_reg_cap.num_phy * |
| sizeof(WMI_HAL_REG_CAPABILITIES_EXT)); |
| if (!phy_caps->each_phy_hal_reg_cap) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| wma_cleanup_dbs_phy_caps(wma_handle); |
| return; |
| } |
| qdf_mem_copy(phy_caps->each_phy_hal_reg_cap, |
| param_buf->hal_reg_caps, |
| phy_caps->num_phy_for_hal_reg_cap.num_phy * |
| sizeof(WMI_HAL_REG_CAPABILITIES_EXT)); |
| |
| wma_init_wifi_pos_dma_rings(wma_handle, |
| param_buf->num_oem_dma_ring_caps, |
| param_buf->oem_dma_ring_caps); |
| |
| wma_print_populate_soc_caps(wma_handle); |
| } |
| |
| /** |
| * wma_rx_service_ready_ext_event() - evt handler for sevice ready ext event. |
| * @handle: wma handle |
| * @event: params of the service ready extended event |
| * @length: param length |
| * |
| * Return: none |
| */ |
| int wma_rx_service_ready_ext_event(void *handle, uint8_t *event, |
| uint32_t length) |
| { |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| WMI_SERVICE_READY_EXT_EVENTID_param_tlvs *param_buf; |
| wmi_service_ready_ext_event_fixed_param *ev; |
| int status; |
| QDF_STATUS ret; |
| struct wmi_init_cmd_param init_param = {0}; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| return -EINVAL; |
| } |
| |
| param_buf = (WMI_SERVICE_READY_EXT_EVENTID_param_tlvs *) event; |
| if (!param_buf) { |
| WMA_LOGE("%s: Invalid event", __func__); |
| return -EINVAL; |
| } |
| |
| ev = param_buf->fixed_param; |
| if (!ev) { |
| WMA_LOGE("%s: Invalid buffer", __func__); |
| return -EINVAL; |
| } |
| |
| WMA_LOGD("WMA <-- WMI_SERVICE_READY_EXT_EVENTID"); |
| |
| WMA_LOGD("%s: Defaults: scan config:%x FW mode config:%x", |
| __func__, ev->default_conc_scan_config_bits, |
| ev->default_fw_config_bits); |
| |
| ret = qdf_mc_timer_stop(&wma_handle->service_ready_ext_timer); |
| if (!QDF_IS_STATUS_SUCCESS(ret)) { |
| WMA_LOGE("Failed to stop the service ready ext timer"); |
| return -EINVAL; |
| } |
| wma_populate_soc_caps(wma_handle, param_buf); |
| |
| ret = wma_update_hw_mode_list(wma_handle); |
| if (QDF_IS_STATUS_ERROR(ret)) { |
| WMA_LOGE("Failed to update hw mode list"); |
| return -EINVAL; |
| } |
| |
| WMA_LOGD("WMA --> WMI_INIT_CMDID"); |
| |
| init_param.res_cfg = &wma_handle->wlan_resource_config; |
| init_param.num_mem_chunks = wma_handle->num_mem_chunks; |
| init_param.mem_chunks = wma_handle->mem_chunks; |
| init_param.hw_mode_id = WMI_HOST_HW_MODE_MAX; |
| status = wmi_unified_init_cmd_send(wma_handle->wmi_handle, &init_param); |
| if (status != EOK) |
| /* In success case, WMI layer will free after getting copy |
| * engine TX complete interrupt |
| */ |
| WMA_LOGE("Failed to send WMI_INIT_CMDID command"); |
| |
| wma_init_scan_fw_mode_config(wma_handle, |
| ev->default_conc_scan_config_bits, |
| ev->default_fw_config_bits); |
| wma_handle->target_fw_vers_ext = ev->fw_build_vers_ext; |
| return 0; |
| } |
| |
| /** |
| * wma_rx_ready_event() - event handler to process |
| * wmi rx ready event. |
| * @handle: wma handle |
| * @cmd_param_info: command params info |
| * @length: param length |
| * |
| * Return: none |
| */ |
| int wma_rx_ready_event(void *handle, uint8_t *cmd_param_info, |
| uint32_t length) |
| { |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| WMI_READY_EVENTID_param_tlvs *param_buf = NULL; |
| wmi_ready_event_fixed_param *ev = NULL; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| param_buf = (WMI_READY_EVENTID_param_tlvs *) cmd_param_info; |
| if (!(wma_handle && param_buf)) { |
| WMA_LOGE("%s: Invalid arguments", __func__); |
| QDF_ASSERT(0); |
| return -EINVAL; |
| } |
| |
| WMA_LOGD("WMA <-- WMI_READY_EVENTID"); |
| |
| ev = param_buf->fixed_param; |
| /* Indicate to the waiting thread that the ready |
| * event was received |
| */ |
| wma_handle->sub_20_support = |
| WMI_SERVICE_IS_ENABLED(wma_handle->wmi_service_bitmap, |
| WMI_SERVICE_HALF_RATE_QUARTER_RATE_SUPPORT); |
| wma_handle->wmi_ready = true; |
| wma_handle->wlan_init_status = ev->status; |
| |
| /* |
| * We need to check the WMI versions and make sure both |
| * host and fw are compatible. |
| */ |
| if (!wmi_versions_are_compatible(&wma_handle->final_abi_vers, |
| &ev->fw_abi_vers)) { |
| /* |
| * Error: Our host version and the given firmware version |
| * are incompatible. |
| */ |
| WMA_LOGE("%s: Error: Incompatible WMI version.", __func__); |
| WMA_LOGE("%s: Host: %d,%d,0x%x 0x%x 0x%x 0x%x, FW: %d,%d,0x%x 0x%x 0x%x 0x%x", |
| __func__, |
| WMI_VER_GET_MAJOR(wma_handle->final_abi_vers. |
| abi_version_0), |
| WMI_VER_GET_MINOR(wma_handle->final_abi_vers. |
| abi_version_0), |
| wma_handle->final_abi_vers.abi_version_ns_0, |
| wma_handle->final_abi_vers.abi_version_ns_1, |
| wma_handle->final_abi_vers.abi_version_ns_2, |
| wma_handle->final_abi_vers.abi_version_ns_3, |
| WMI_VER_GET_MAJOR(ev->fw_abi_vers.abi_version_0), |
| WMI_VER_GET_MINOR(ev->fw_abi_vers.abi_version_0), |
| ev->fw_abi_vers.abi_version_ns_0, |
| ev->fw_abi_vers.abi_version_ns_1, |
| ev->fw_abi_vers.abi_version_ns_2, |
| ev->fw_abi_vers.abi_version_ns_3); |
| if (wma_handle->wlan_init_status == WLAN_INIT_STATUS_SUCCESS) { |
| /* Failed this connection to FW */ |
| wma_handle->wlan_init_status = |
| WLAN_INIT_STATUS_GEN_FAILED; |
| } |
| } |
| qdf_mem_copy(&wma_handle->final_abi_vers, &ev->fw_abi_vers, |
| sizeof(wmi_abi_version)); |
| qdf_mem_copy(&wma_handle->target_abi_vers, &ev->fw_abi_vers, |
| sizeof(wmi_abi_version)); |
| |
| /* copy the mac addr */ |
| WMI_MAC_ADDR_TO_CHAR_ARRAY(&ev->mac_addr, wma_handle->myaddr); |
| WMI_MAC_ADDR_TO_CHAR_ARRAY(&ev->mac_addr, wma_handle->hwaddr); |
| |
| wma_update_hdd_cfg(wma_handle); |
| |
| qdf_event_set(&wma_handle->wma_ready_event); |
| |
| WMA_LOGD("Exit"); |
| |
| return 0; |
| } |
| |
| /** |
| * wma_setneedshutdown() - setting wma needshutdown flag |
| * |
| * Return: none |
| */ |
| void wma_setneedshutdown(void) |
| { |
| tp_wma_handle wma_handle; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid arguments", __func__); |
| QDF_ASSERT(0); |
| return; |
| } |
| |
| wma_handle->needShutdown = true; |
| WMA_LOGD("%s: Exit", __func__); |
| } |
| |
| /** |
| * wma_needshutdown() - Is wma needs shutdown? |
| * |
| * Return: returns true/false |
| */ |
| bool wma_needshutdown(void) |
| { |
| tp_wma_handle wma_handle; |
| |
| WMA_LOGD("%s: Enter", __func__); |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: Invalid arguments", __func__); |
| QDF_ASSERT(0); |
| return false; |
| } |
| |
| WMA_LOGD("%s: Exit", __func__); |
| return wma_handle->needShutdown; |
| } |
| |
| /** |
| * wma_wait_for_ready_event() - wait for wma ready event |
| * @handle: wma handle |
| * |
| * Return: 0 for success or QDF error |
| */ |
| QDF_STATUS wma_wait_for_ready_event(WMA_HANDLE handle) |
| { |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| QDF_STATUS qdf_status; |
| |
| /* wait until WMI_READY_EVENTID received from FW */ |
| qdf_status = qdf_wait_for_event_completion( |
| &wma_handle->wma_ready_event, |
| WMA_READY_EVENTID_TIMEOUT); |
| |
| if (QDF_STATUS_SUCCESS != qdf_status) { |
| WMA_LOGE("%s: Timeout waiting for ready event from FW", |
| __func__); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| } |
| return qdf_status; |
| } |
| |
| /** |
| * wma_set_ppsconfig() - set pps config in fw |
| * @vdev_id: vdev id |
| * @pps_param: pps params |
| * @val : param value |
| * |
| * Return: 0 for success or QDF error |
| */ |
| QDF_STATUS wma_set_ppsconfig(uint8_t vdev_id, uint16_t pps_param, |
| int val) |
| { |
| tp_wma_handle wma = cds_get_context(QDF_MODULE_ID_WMA); |
| int ret = -EIO; |
| uint32_t pps_val; |
| |
| if (NULL == wma) { |
| WMA_LOGE("%s: Failed to get wma", __func__); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| switch (pps_param) { |
| case WMA_VHT_PPS_PAID_MATCH: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_PAID_MATCH & 0xffff); |
| goto pkt_pwr_save_config; |
| case WMA_VHT_PPS_GID_MATCH: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_GID_MATCH & 0xffff); |
| goto pkt_pwr_save_config; |
| case WMA_VHT_PPS_DELIM_CRC_FAIL: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_DELIM_CRC_FAIL & 0xffff); |
| goto pkt_pwr_save_config; |
| |
| /* Enable the code below as and when the functionality |
| * is supported/added in host. |
| */ |
| #ifdef NOT_YET |
| case WMA_VHT_PPS_EARLY_TIM_CLEAR: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_EARLY_TIM_CLEAR & 0xffff); |
| goto pkt_pwr_save_config; |
| case WMA_VHT_PPS_EARLY_DTIM_CLEAR: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_EARLY_DTIM_CLEAR & 0xffff); |
| goto pkt_pwr_save_config; |
| case WMA_VHT_PPS_EOF_PAD_DELIM: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_EOF_PAD_DELIM & 0xffff); |
| goto pkt_pwr_save_config; |
| case WMA_VHT_PPS_MACADDR_MISMATCH: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_MACADDR_MISMATCH & 0xffff); |
| goto pkt_pwr_save_config; |
| case WMA_VHT_PPS_GID_NSTS_ZERO: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_GID_NSTS_ZERO & 0xffff); |
| goto pkt_pwr_save_config; |
| case WMA_VHT_PPS_RSSI_CHECK: |
| pps_val = ((val << 31) & 0xffff0000) | |
| (PKT_PWR_SAVE_RSSI_CHECK & 0xffff); |
| goto pkt_pwr_save_config; |
| #endif /* NOT_YET */ |
| pkt_pwr_save_config: |
| WMA_LOGD("vdev_id:%d val:0x%x pps_val:0x%x", vdev_id, |
| val, pps_val); |
| ret = wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_PACKET_POWERSAVE, |
| pps_val); |
| break; |
| default: |
| WMA_LOGE("%s:INVALID PPS CONFIG", __func__); |
| } |
| |
| return (ret) ? QDF_STATUS_E_FAILURE : QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_process_set_mas() - Function to enable/disable MAS |
| * @wma: Pointer to WMA handle |
| * @mas_val: 1-Enable MAS, 0-Disable MAS |
| * |
| * This function enables/disables the MAS value |
| * |
| * Return: QDF_SUCCESS for success otherwise failure |
| */ |
| static QDF_STATUS wma_process_set_mas(tp_wma_handle wma, |
| uint32_t *mas_val) |
| { |
| uint32_t val; |
| |
| if (NULL == wma || NULL == mas_val) { |
| WMA_LOGE("%s: Invalid input to enable/disable MAS", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| val = (*mas_val); |
| |
| if (QDF_STATUS_SUCCESS != |
| wma_set_enable_disable_mcc_adaptive_scheduler(val)) { |
| WMA_LOGE("%s: Unable to enable/disable MAS", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| WMA_LOGE("%s: Value is %d", __func__, val); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_process_set_miracast() - Function to set miracast value in WMA |
| * @wma: Pointer to WMA handle |
| * @miracast_val: 0-Disabled,1-Source,2-Sink |
| * |
| * This function stores the miracast value in WMA |
| * |
| * Return: QDF_SUCCESS for success otherwise failure |
| * |
| */ |
| static QDF_STATUS wma_process_set_miracast(tp_wma_handle wma, |
| uint32_t *miracast_val) |
| { |
| if (NULL == wma || NULL == miracast_val) { |
| WMA_LOGE("%s: Invalid input to store miracast value", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| wma->miracast_value = *miracast_val; |
| WMA_LOGE("%s: Miracast value is %d", __func__, wma->miracast_value); |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_config_stats_factor() - Function to configure stats avg. factor |
| * @wma: pointer to WMA handle |
| * @avg_factor: stats. avg. factor passed down by userspace |
| * |
| * This function configures the avg. stats value in firmware |
| * |
| * Return: QDF_STATUS_SUCCESS for success otherwise failure |
| * |
| */ |
| static QDF_STATUS wma_config_stats_factor(tp_wma_handle wma, |
| struct sir_stats_avg_factor *avg_factor) |
| { |
| QDF_STATUS ret; |
| |
| if (NULL == wma || NULL == avg_factor) { |
| WMA_LOGE("%s: Invalid input of stats avg factor", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| avg_factor->vdev_id, |
| WMI_VDEV_PARAM_STATS_AVG_FACTOR, |
| avg_factor->stats_avg_factor); |
| if (QDF_IS_STATUS_ERROR(ret)) { |
| WMA_LOGE(" failed to set avg_factor for vdev_id %d", |
| avg_factor->vdev_id); |
| } |
| |
| WMA_LOGD("%s: Set stats_avg_factor %d for vdev_id %d", __func__, |
| avg_factor->stats_avg_factor, avg_factor->vdev_id); |
| |
| return ret; |
| } |
| |
| /** |
| * wma_config_guard_time() - Function to set guard time in firmware |
| * @wma: pointer to WMA handle |
| * @guard_time: guard time passed down by userspace |
| * |
| * This function configures the guard time in firmware |
| * |
| * Return: QDF_STATUS_SUCCESS for success otherwise failure |
| * |
| */ |
| static QDF_STATUS wma_config_guard_time(tp_wma_handle wma, |
| struct sir_guard_time_request *guard_time) |
| { |
| QDF_STATUS ret; |
| |
| if (NULL == wma || NULL == guard_time) { |
| WMA_LOGE("%s: Invalid input of guard time", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, |
| guard_time->vdev_id, |
| WMI_VDEV_PARAM_RX_LEAK_WINDOW, |
| guard_time->guard_time); |
| if (QDF_IS_STATUS_ERROR(ret)) { |
| WMA_LOGE(" failed to set guard time for vdev_id %d", |
| guard_time->vdev_id); |
| } |
| |
| WMA_LOGD("Set guard time %d for vdev_id %d", |
| guard_time->guard_time, guard_time->vdev_id); |
| |
| return ret; |
| } |
| |
| /** |
| * wma_enable_specific_fw_logs() - Start/Stop logging of diag event/log id |
| * @wma_handle: WMA handle |
| * @start_log: Start logging related parameters |
| * |
| * Send the command to the FW based on which specific logging of diag |
| * event/log id can be started/stopped |
| * |
| * Return: None |
| */ |
| static void wma_enable_specific_fw_logs(tp_wma_handle wma_handle, |
| struct sir_wifi_start_log *start_log) |
| { |
| |
| if (!start_log) { |
| WMA_LOGE("%s: start_log pointer is NULL", __func__); |
| return; |
| } |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return; |
| } |
| |
| if (!((start_log->ring_id == RING_ID_CONNECTIVITY) || |
| (start_log->ring_id == RING_ID_FIRMWARE_DEBUG))) { |
| WMA_LOGD("%s: Not connectivity or fw debug ring: %d", |
| __func__, start_log->ring_id); |
| return; |
| } |
| |
| wmi_unified_enable_specific_fw_logs_cmd(wma_handle->wmi_handle, |
| (struct wmi_wifi_start_log *)start_log); |
| } |
| |
| #define MEGABYTE (1024 * 1024) |
| /** |
| * wma_set_wifi_start_packet_stats() - Start/stop packet stats |
| * @wma_handle: WMA handle |
| * @start_log: Struture containing the start wifi logger params |
| * |
| * This function is used to send the WMA commands to start/stop logging |
| * of per packet statistics |
| * |
| * Return: None |
| * |
| */ |
| #ifdef REMOVE_PKT_LOG |
| static void wma_set_wifi_start_packet_stats(void *wma_handle, |
| struct sir_wifi_start_log *start_log) |
| { |
| } |
| #else |
| static void wma_set_wifi_start_packet_stats(void *wma_handle, |
| struct sir_wifi_start_log *start_log) |
| { |
| struct hif_opaque_softc *scn; |
| uint32_t log_state; |
| |
| if (!start_log) { |
| WMA_LOGE("%s: start_log pointer is NULL", __func__); |
| return; |
| } |
| if (!wma_handle) { |
| WMA_LOGE("%s: Invalid wma handle", __func__); |
| return; |
| } |
| |
| /* No need to register for ring IDs other than packet stats */ |
| if (start_log->ring_id != RING_ID_PER_PACKET_STATS) { |
| WMA_LOGD("%s: Ring id is not for per packet stats: %d", |
| __func__, start_log->ring_id); |
| return; |
| } |
| |
| scn = cds_get_context(QDF_MODULE_ID_HIF); |
| if (scn == NULL) { |
| WMA_LOGE("%s: Invalid HIF handle", __func__); |
| return; |
| } |
| |
| log_state = ATH_PKTLOG_ANI | ATH_PKTLOG_RCUPDATE | ATH_PKTLOG_RCFIND | |
| ATH_PKTLOG_RX | ATH_PKTLOG_TX | |
| ATH_PKTLOG_TEXT | ATH_PKTLOG_SW_EVENT; |
| |
| if (start_log->size != 0) { |
| pktlog_setsize(scn, start_log->size * MEGABYTE); |
| return; |
| } else if (start_log->is_pktlog_buff_clear == true) { |
| pktlog_clearbuff(scn, start_log->is_pktlog_buff_clear); |
| return; |
| } |
| |
| if (start_log->verbose_level == WLAN_LOG_LEVEL_ACTIVE) { |
| pktlog_enable(scn, log_state, start_log->ini_triggered, |
| start_log->user_triggered, |
| start_log->is_iwpriv_command); |
| WMA_LOGD("%s: Enabling per packet stats", __func__); |
| } else { |
| pktlog_enable(scn, 0, start_log->ini_triggered, |
| start_log->user_triggered, |
| start_log->is_iwpriv_command); |
| WMA_LOGD("%s: Disabling per packet stats", __func__); |
| } |
| } |
| #endif |
| |
| /** |
| * wma_send_flush_logs_to_fw() - Send log flush command to FW |
| * @wma_handle: WMI handle |
| * |
| * This function is used to send the flush command to the FW, |
| * that will flush the fw logs that are residue in the FW |
| * |
| * Return: None |
| */ |
| void wma_send_flush_logs_to_fw(tp_wma_handle wma_handle) |
| { |
| QDF_STATUS status; |
| int ret; |
| |
| ret = wmi_unified_flush_logs_to_fw_cmd(wma_handle->wmi_handle); |
| if (ret != EOK) |
| return; |
| |
| status = qdf_mc_timer_start(&wma_handle->log_completion_timer, |
| WMA_LOG_COMPLETION_TIMER); |
| if (status != QDF_STATUS_SUCCESS) |
| WMA_LOGE("Failed to start the log completion timer"); |
| } |
| |
| /** |
| * wma_update_wep_default_key - To update default key id |
| * @wma: pointer to wma handler |
| * @update_def_key: pointer to wep_update_default_key_idx |
| * |
| * This function makes a copy of default key index to txrx node |
| * |
| * Return: Success |
| */ |
| static QDF_STATUS wma_update_wep_default_key(tp_wma_handle wma, |
| struct wep_update_default_key_idx *update_def_key) |
| { |
| struct wma_txrx_node *iface = |
| &wma->interfaces[update_def_key->session_id]; |
| iface->wep_default_key_idx = update_def_key->default_idx; |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_update_tx_fail_cnt_th() - Set threshold for TX pkt fail |
| * @wma_handle: WMA handle |
| * @tx_fail_cnt_th: sme_tx_fail_cnt_threshold parameter |
| * |
| * This function is used to set Tx pkt fail count threshold, |
| * FW will do disconnect with station once this threshold is reached. |
| * |
| * Return: VOS_STATUS_SUCCESS on success, error number otherwise |
| */ |
| static QDF_STATUS wma_update_tx_fail_cnt_th(tp_wma_handle wma, |
| struct sme_tx_fail_cnt_threshold *tx_fail_cnt_th) |
| { |
| u_int8_t vdev_id; |
| u_int32_t tx_fail_disconn_th; |
| int ret = -EIO; |
| |
| if (!wma || !wma->wmi_handle) { |
| WMA_LOGE(FL("WMA is closed, can not issue Tx pkt fail count threshold")); |
| return QDF_STATUS_E_INVAL; |
| } |
| vdev_id = tx_fail_cnt_th->session_id; |
| tx_fail_disconn_th = tx_fail_cnt_th->tx_fail_cnt_threshold; |
| WMA_LOGD("Set TX pkt fail count threshold vdevId %d count %d", |
| vdev_id, tx_fail_disconn_th); |
| |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_DISCONNECT_TH, |
| tx_fail_disconn_th); |
| |
| if (ret) { |
| WMA_LOGE(FL("Failed to send TX pkt fail count threshold command")); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_update_short_retry_limit() - Set retry limit for short frames |
| * @wma_handle: WMA handle |
| * @short_retry_limit_th: retry limir count for Short frames. |
| * |
| * This function is used to configure the transmission retry limit at which |
| * short frames needs to be retry. |
| * |
| * Return: VOS_STATUS_SUCCESS on success, error number otherwise |
| */ |
| static QDF_STATUS wma_update_short_retry_limit(tp_wma_handle wma, |
| struct sme_short_retry_limit *short_retry_limit_th) |
| { |
| uint8_t vdev_id; |
| uint32_t short_retry_limit; |
| int ret; |
| |
| if (!wma || !wma->wmi_handle) { |
| WMA_LOGE("WMA is closed, can not issue short retry limit threshold"); |
| return QDF_STATUS_E_INVAL; |
| } |
| vdev_id = short_retry_limit_th->session_id; |
| short_retry_limit = short_retry_limit_th->short_retry_limit; |
| WMA_LOGD("Set short retry limit threshold vdevId %d count %d", |
| vdev_id, short_retry_limit); |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_NON_AGG_SW_RETRY_TH, |
| short_retry_limit); |
| |
| if (ret) { |
| WMA_LOGE("Failed to send short limit threshold command"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_update_long_retry_limit() - Set retry limit for long frames |
| * @wma_handle: WMA handle |
| * @long_retry_limit_th: retry limir count for long frames |
| * |
| * This function is used to configure the transmission retry limit at which |
| * long frames needs to be retry |
| * |
| * Return: VOS_STATUS_SUCCESS on success, error number otherwise |
| */ |
| static QDF_STATUS wma_update_long_retry_limit(tp_wma_handle wma, |
| struct sme_long_retry_limit *long_retry_limit_th) |
| { |
| uint8_t vdev_id; |
| uint32_t long_retry_limit; |
| int ret; |
| |
| if (!wma || !wma->wmi_handle) { |
| WMA_LOGE("WMA is closed, can not issue long retry limit threshold"); |
| return QDF_STATUS_E_INVAL; |
| } |
| vdev_id = long_retry_limit_th->session_id; |
| long_retry_limit = long_retry_limit_th->long_retry_limit; |
| WMA_LOGD("Set TX pkt fail count threshold vdevId %d count %d", |
| vdev_id, long_retry_limit); |
| |
| ret = wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_AGG_SW_RETRY_TH, |
| long_retry_limit); |
| |
| if (ret) { |
| WMA_LOGE("Failed to send long limit threshold command"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /* |
| * wma_update_sta_inactivity_timeout() - Set sta_inactivity_timeout to fw |
| * @wma_handle: WMA handle |
| * @sta_inactivity_timer: sme_sta_inactivity_timeout |
| * |
| * This function is used to set sta_inactivity_timeout. |
| * If a station does not send anything in sta_inactivity_timeout seconds, an |
| * empty data frame is sent to it in order to verify whether it is |
| * still in range. If this frame is not ACKed, the station will be |
| * disassociated and then deauthenticated. |
| * |
| * Return: None |
| */ |
| void wma_update_sta_inactivity_timeout(tp_wma_handle wma, |
| struct sme_sta_inactivity_timeout *sta_inactivity_timer) |
| { |
| uint8_t vdev_id; |
| uint32_t max_unresponsive_time; |
| uint32_t min_inactive_time, max_inactive_time; |
| |
| if (!wma || !wma->wmi_handle) { |
| WMA_LOGE("WMA is closed, can not issue sta_inactivity_timeout"); |
| return; |
| } |
| vdev_id = sta_inactivity_timer->session_id; |
| max_unresponsive_time = sta_inactivity_timer->sta_inactivity_timeout; |
| max_inactive_time = max_unresponsive_time * TWO_THIRD; |
| min_inactive_time = max_unresponsive_time - max_inactive_time; |
| |
| if (wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_AP_KEEPALIVE_MIN_IDLE_INACTIVE_TIME_SECS, |
| min_inactive_time)) |
| WMA_LOGE("Failed to Set AP MIN IDLE INACTIVE TIME"); |
| |
| if (wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_AP_KEEPALIVE_MAX_IDLE_INACTIVE_TIME_SECS, |
| max_inactive_time)) |
| WMA_LOGE("Failed to Set AP MAX IDLE INACTIVE TIME"); |
| |
| if (wma_vdev_set_param(wma->wmi_handle, vdev_id, |
| WMI_VDEV_PARAM_AP_KEEPALIVE_MAX_UNRESPONSIVE_TIME_SECS, |
| max_unresponsive_time)) |
| WMA_LOGE("Failed to Set MAX UNRESPONSIVE TIME"); |
| |
| WMA_LOGD("%s:vdev_id:%d min_inactive_time: %u max_inactive_time: %u max_unresponsive_time: %u", |
| __func__, vdev_id, |
| min_inactive_time, max_inactive_time, |
| max_unresponsive_time); |
| } |
| |
| #ifdef WLAN_FEATURE_WOW_PULSE |
| |
| |
| #define WMI_WOW_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM \ |
| WMI_WOW_HOSTWAKEUP_GPIO_PIN_PATTERN_CONFIG_CMD_fixed_param |
| |
| |
| #define WMITLV_TAG_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM \ |
| WMITLV_TAG_STRUC_wmi_wow_hostwakeup_gpio_pin_pattern_config_cmd_fixed_param |
| |
| /** |
| * wma_send_wow_pulse_cmd() - send wmi cmd of wow pulse cmd |
| * information to fw. |
| * @wma_handle: wma handler |
| * @udp_response: wow_pulse_mode pointer |
| * |
| * Return: Return QDF_STATUS |
| */ |
| static QDF_STATUS wma_send_wow_pulse_cmd(tp_wma_handle wma_handle, |
| struct wow_pulse_mode *wow_pulse_cmd) |
| { |
| QDF_STATUS status = QDF_STATUS_SUCCESS; |
| wmi_buf_t buf; |
| WMI_WOW_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM *cmd; |
| u_int16_t len; |
| |
| len = sizeof(*cmd); |
| buf = wmi_buf_alloc(wma_handle->wmi_handle, len); |
| if (!buf) { |
| WMA_LOGE("wmi_buf_alloc failed"); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| cmd = (WMI_WOW_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM *)wmi_buf_data(buf); |
| qdf_mem_zero(cmd, len); |
| |
| WMITLV_SET_HDR(&cmd->tlv_header, |
| WMITLV_TAG_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM, |
| WMITLV_GET_STRUCT_TLVLEN( |
| WMI_WOW_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM)); |
| |
| cmd->enable = wow_pulse_cmd->wow_pulse_enable; |
| cmd->pin = wow_pulse_cmd->wow_pulse_pin; |
| cmd->interval_low = wow_pulse_cmd->wow_pulse_interval_low; |
| cmd->interval_high = wow_pulse_cmd->wow_pulse_interval_high; |
| cmd->repeat_cnt = WMI_WOW_PULSE_REPEAT_CNT; |
| |
| if (wmi_unified_cmd_send(wma_handle->wmi_handle, buf, len, |
| WMI_WOW_HOSTWAKEUP_GPIO_PIN_PATTERN_CONFIG_CMDID)) { |
| WMA_LOGE("Failed to send send wow pulse"); |
| wmi_buf_free(buf); |
| status = QDF_STATUS_E_FAILURE; |
| } |
| |
| WMA_LOGD("%s: Exit", __func__); |
| return status; |
| } |
| |
| #undef WMI_WOW_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM |
| #undef WMITLV_TAG_HOSTWAKEUP_GPIO_CMD_FIXED_PARAM |
| #undef WMI_WOW_PULSE_REPEAT_CNT |
| |
| #else |
| static inline QDF_STATUS wma_send_wow_pulse_cmd(tp_wma_handle wma_handle, |
| struct wow_pulse_mode *wow_pulse_cmd) |
| { |
| return QDF_STATUS_E_FAILURE; |
| } |
| #endif |
| |
| |
| /** |
| * wma_process_power_debug_stats_req() - Process the Chip Power stats collect |
| * request and pass the Power stats request to Fw |
| * @wma_handle: WMA handle |
| * |
| * Return: QDF_STATUS |
| */ |
| #ifdef WLAN_POWER_DEBUGFS |
| static QDF_STATUS wma_process_power_debug_stats_req(tp_wma_handle wma_handle) |
| { |
| wmi_pdev_get_chip_power_stats_cmd_fixed_param *cmd; |
| int32_t len; |
| wmi_buf_t buf; |
| uint8_t *buf_ptr; |
| int ret; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: input pointer is NULL", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| len = sizeof(*cmd); |
| buf = wmi_buf_alloc(wma_handle->wmi_handle, len); |
| if (!buf) { |
| WMA_LOGE("%s: Failed allocate wmi buffer", __func__); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| buf_ptr = (u_int8_t *) wmi_buf_data(buf); |
| cmd = (wmi_pdev_get_chip_power_stats_cmd_fixed_param *) buf_ptr; |
| |
| WMITLV_SET_HDR(&cmd->tlv_header, |
| WMITLV_TAG_STRUC_wmi_get_chip_power_stats_cmd_fixed_param, |
| WMITLV_GET_STRUCT_TLVLEN( |
| wmi_pdev_get_chip_power_stats_cmd_fixed_param)); |
| cmd->pdev_id = 0; |
| |
| WMA_LOGD("POWER_DEBUG_STATS - Get Request Params; Pdev id - %d", |
| cmd->pdev_id); |
| ret = wmi_unified_cmd_send(wma_handle->wmi_handle, buf, len, |
| WMI_PDEV_GET_CHIP_POWER_STATS_CMDID); |
| if (ret) { |
| WMA_LOGE("%s: Failed to send power debug stats request", |
| __func__); |
| wmi_buf_free(buf); |
| return QDF_STATUS_E_FAILURE; |
| } |
| return QDF_STATUS_SUCCESS; |
| } |
| #else |
| static QDF_STATUS wma_process_power_debug_stats_req(tp_wma_handle wma_handle) |
| { |
| return QDF_STATUS_SUCCESS; |
| } |
| #endif |
| |
| /** |
| * wma_set_arp_req_stats() - process set arp stats request command to fw |
| * @wma_handle: WMA handle |
| * @req_buf: set srp stats request buffer |
| * |
| * Return: None |
| */ |
| static void wma_set_arp_req_stats(WMA_HANDLE handle, |
| struct set_arp_stats_params *req_buf) |
| { |
| int status; |
| struct set_arp_stats *arp_stats; |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| |
| if (!wma_handle || !wma_handle->wmi_handle) { |
| WMA_LOGE("%s: WMA is closed, cannot send per roam config", |
| __func__); |
| return; |
| } |
| |
| arp_stats = (struct set_arp_stats *)req_buf; |
| status = wmi_unified_set_arp_stats_req(wma_handle->wmi_handle, |
| arp_stats); |
| if (status != EOK) |
| WMA_LOGE("%s: failed to set arp stats to FW", |
| __func__); |
| } |
| |
| /** |
| * wma_get_arp_req_stats() - process get arp stats request command to fw |
| * @wma_handle: WMA handle |
| * @req_buf: get srp stats request buffer |
| * |
| * Return: None |
| */ |
| static void wma_get_arp_req_stats(WMA_HANDLE handle, |
| struct get_arp_stats_params *req_buf) |
| { |
| int status; |
| struct get_arp_stats *arp_stats; |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| |
| if (!wma_handle || !wma_handle->wmi_handle) { |
| WMA_LOGE("%s: WMA is closed, cannot send per roam config", |
| __func__); |
| return; |
| } |
| |
| arp_stats = (struct get_arp_stats *)req_buf; |
| status = wmi_unified_get_arp_stats_req(wma_handle->wmi_handle, |
| arp_stats); |
| if (status != EOK) |
| WMA_LOGE("%s: failed to send get arp stats to FW", |
| __func__); |
| } |
| |
| /** |
| * wma_set_del_pmkid_cache() - API to set/delete PMKID cache entry in fw |
| * @handle: WMA handle |
| * @pmk_cache: PMK cache entry |
| * |
| * Return: None |
| */ |
| static void wma_set_del_pmkid_cache(WMA_HANDLE handle, |
| struct wmi_unified_pmk_cache *pmk_cache) |
| { |
| int status; |
| tp_wma_handle wma_handle = (tp_wma_handle) handle; |
| |
| if (!wma_handle || !wma_handle->wmi_handle) { |
| WMA_LOGE("WMA is closed, cannot send set del pmkid"); |
| return; |
| } |
| |
| status = wmi_unified_set_del_pmkid_cache(wma_handle->wmi_handle, |
| pmk_cache); |
| if (status != EOK) |
| WMA_LOGE("failed to send set/del pmkid cmd to fw"); |
| } |
| |
| QDF_STATUS wma_set_rx_reorder_timeout_val(tp_wma_handle wma_handle, |
| struct sir_set_rx_reorder_timeout_val *reorder_timeout) |
| { |
| wmi_pdev_set_reorder_timeout_val_cmd_fixed_param *cmd; |
| uint32_t len; |
| wmi_buf_t buf; |
| int ret; |
| |
| if (!reorder_timeout) { |
| WMA_LOGE(FL("invalid pointer")); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| if (!wma_handle) { |
| WMA_LOGE(FL("WMA context is invald!")); |
| return QDF_STATUS_E_INVAL; |
| } |
| len = sizeof(*cmd); |
| buf = wmi_buf_alloc(wma_handle->wmi_handle, len); |
| |
| if (!buf) { |
| WMA_LOGE(FL("Failed allocate wmi buffer")); |
| return QDF_STATUS_E_NOMEM; |
| } |
| cmd = (wmi_pdev_set_reorder_timeout_val_cmd_fixed_param *) |
| wmi_buf_data(buf); |
| |
| WMITLV_SET_HDR(&cmd->tlv_header, |
| WMITLV_TAG_STRUC_wmi_pdev_set_reorder_timeout_val_cmd_fixed_param, |
| WMITLV_GET_STRUCT_TLVLEN(wmi_pdev_set_reorder_timeout_val_cmd_fixed_param)); |
| |
| memcpy(cmd->rx_timeout_pri, reorder_timeout->rx_timeout_pri, |
| sizeof(reorder_timeout->rx_timeout_pri)); |
| |
| WMA_LOGD("rx aggr record timeout: VO: %d, VI: %d, BE: %d, BK: %d", |
| cmd->rx_timeout_pri[0], cmd->rx_timeout_pri[1], |
| cmd->rx_timeout_pri[2], cmd->rx_timeout_pri[3]); |
| |
| ret = wmi_unified_cmd_send(wma_handle->wmi_handle, buf, len, |
| WMI_PDEV_SET_REORDER_TIMEOUT_VAL_CMDID); |
| if (ret) { |
| WMA_LOGE(FL("Failed to send aggregation timeout")); |
| wmi_buf_free(buf); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| QDF_STATUS wma_set_rx_blocksize(tp_wma_handle wma_handle, |
| struct sir_peer_set_rx_blocksize *peer_rx_blocksize) |
| { |
| wmi_peer_set_rx_blocksize_cmd_fixed_param *cmd; |
| int32_t len; |
| wmi_buf_t buf; |
| u_int8_t *buf_ptr; |
| int ret; |
| |
| if (!peer_rx_blocksize) { |
| WMA_LOGE(FL("invalid pointer")); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| if (!wma_handle) { |
| WMA_LOGE(FL(" WMA context is invald!")); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| len = sizeof(*cmd); |
| buf = wmi_buf_alloc(wma_handle->wmi_handle, len); |
| |
| if (!buf) { |
| WMA_LOGE(FL("Failed allocate wmi buffer")); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| buf_ptr = (u_int8_t *) wmi_buf_data(buf); |
| cmd = (wmi_peer_set_rx_blocksize_cmd_fixed_param *) buf_ptr; |
| |
| WMITLV_SET_HDR(&cmd->tlv_header, |
| WMITLV_TAG_STRUC_wmi_peer_set_rx_blocksize_cmd_fixed_param, |
| WMITLV_GET_STRUCT_TLVLEN(wmi_peer_set_rx_blocksize_cmd_fixed_param)); |
| |
| cmd->vdev_id = peer_rx_blocksize->vdev_id; |
| cmd->rx_block_ack_win_limit = |
| peer_rx_blocksize->rx_block_ack_win_limit; |
| WMI_CHAR_ARRAY_TO_MAC_ADDR(peer_rx_blocksize->peer_macaddr.bytes, |
| &cmd->peer_macaddr); |
| |
| WMA_LOGD("rx aggr blocksize: %d", cmd->rx_block_ack_win_limit); |
| |
| ret = wmi_unified_cmd_send(wma_handle->wmi_handle, buf, len, |
| WMI_PEER_SET_RX_BLOCKSIZE_CMDID); |
| if (ret) { |
| WMA_LOGE(FL("Failed to send aggregation size command")); |
| wmi_buf_free(buf); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| QDF_STATUS wma_get_chain_rssi(tp_wma_handle wma_handle, |
| struct get_chain_rssi_req_params *req_params) |
| { |
| wmi_peer_antdiv_info_req_cmd_fixed_param *cmd; |
| wmi_buf_t wmi_buf; |
| uint32_t len = sizeof(wmi_peer_antdiv_info_req_cmd_fixed_param); |
| u_int8_t *buf_ptr; |
| |
| if (!wma_handle) { |
| WMA_LOGE(FL("WMA is closed, can not issue cmd")); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| wmi_buf = wmi_buf_alloc(wma_handle->wmi_handle, len); |
| if (!wmi_buf) { |
| WMA_LOGE(FL("wmi_buf_alloc failed")); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| buf_ptr = (u_int8_t *)wmi_buf_data(wmi_buf); |
| |
| cmd = (wmi_peer_antdiv_info_req_cmd_fixed_param *)buf_ptr; |
| WMITLV_SET_HDR(&cmd->tlv_header, |
| WMITLV_TAG_STRUC_wmi_peer_antdiv_info_req_cmd_fixed_param, |
| WMITLV_GET_STRUCT_TLVLEN( |
| wmi_peer_antdiv_info_req_cmd_fixed_param)); |
| cmd->vdev_id = req_params->session_id; |
| WMI_CHAR_ARRAY_TO_MAC_ADDR(req_params->peer_macaddr.bytes, |
| &cmd->peer_mac_address); |
| |
| if (wmi_unified_cmd_send(wma_handle->wmi_handle, wmi_buf, len, |
| WMI_PEER_ANTDIV_INFO_REQ_CMDID)) { |
| WMA_LOGE(FL("failed to send get chain rssi command")); |
| wmi_buf_free(wmi_buf); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| #if defined(WLAN_FEATURE_FILS_SK) |
| /** |
| * wma_roam_scan_send_hlp() - API to send HLP IE info to fw |
| * @wma_handle: WMA handle |
| * @req: HLP params |
| * |
| * Return: QDF_STATUS |
| */ |
| static QDF_STATUS wma_roam_scan_send_hlp(tp_wma_handle wma_handle, |
| struct hlp_params *req) |
| { |
| struct hlp_params *params; |
| QDF_STATUS status; |
| |
| params = qdf_mem_malloc(sizeof(*params)); |
| if (!params) { |
| WMA_LOGE("%s : Memory allocation failed", __func__); |
| return QDF_STATUS_E_NOMEM; |
| } |
| |
| params->vdev_id = req->vdev_id; |
| params->hlp_ie_len = req->hlp_ie_len; |
| qdf_mem_copy(params->hlp_ie, req->hlp_ie, req->hlp_ie_len); |
| status = wmi_unified_roam_send_hlp_cmd(wma_handle->wmi_handle, params); |
| |
| WMA_LOGD("Send HLP status %d vdev id %d", status, params->vdev_id); |
| qdf_trace_hex_dump(QDF_MODULE_ID_WMI, QDF_TRACE_LEVEL_DEBUG, |
| params->hlp_ie, 10); |
| |
| qdf_mem_free(params); |
| return status; |
| } |
| #else |
| static QDF_STATUS wma_roam_scan_send_hlp(tp_wma_handle wma_handle, |
| struct hlp_params *req) |
| { |
| return QDF_STATUS_SUCCESS; |
| } |
| #endif |
| |
| /** |
| * wma_process_set_limit_off_chan() - set limit off chanel parameters |
| * @wma_handle: pointer to wma handle |
| * @param: pointer to sir_limit_off_chan |
| * |
| * Return: QDF_STATUS_SUCCESS for success or error code. |
| */ |
| static QDF_STATUS wma_process_limit_off_chan(tp_wma_handle wma_handle, |
| struct sir_limit_off_chan *param) |
| { |
| int32_t err; |
| struct wmi_limit_off_chan_param limit_off_chan_param; |
| |
| if (param->vdev_id >= wma_handle->max_bssid) { |
| WMA_LOGE(FL("Invalid vdev_id: %d"), param->vdev_id); |
| return QDF_STATUS_E_INVAL; |
| } |
| if (!wma_is_vdev_up(param->vdev_id)) { |
| WMA_LOGE("vdev %d is not up skipping limit_off_chan_param", |
| param->vdev_id); |
| return QDF_STATUS_E_INVAL; |
| } |
| |
| limit_off_chan_param.vdev_id = param->vdev_id; |
| limit_off_chan_param.status = param->is_tos_active; |
| limit_off_chan_param.max_offchan_time = param->max_off_chan_time; |
| limit_off_chan_param.rest_time = param->rest_time; |
| limit_off_chan_param.skip_dfs_chans = param->skip_dfs_chans; |
| |
| err = wmi_unified_send_limit_off_chan_cmd(wma_handle->wmi_handle, |
| &limit_off_chan_param); |
| if (err) { |
| WMA_LOGE("\n failed to set limit off chan cmd"); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_mc_process_msg() - process wma messages and call appropriate function. |
| * @msg: message |
| * |
| * Return: QDF_SUCCESS for success otherwise failure |
| */ |
| static QDF_STATUS wma_mc_process_msg(struct scheduler_msg *msg) |
| { |
| QDF_STATUS qdf_status = QDF_STATUS_SUCCESS; |
| tp_wma_handle wma_handle; |
| struct cdp_vdev *txrx_vdev_handle = NULL; |
| |
| extern uint8_t *mac_trace_get_wma_msg_string(uint16_t wmaMsg); |
| void *soc = cds_get_context(QDF_MODULE_ID_SOC); |
| |
| if (NULL == msg) { |
| WMA_LOGE("msg is NULL"); |
| QDF_ASSERT(0); |
| qdf_status = QDF_STATUS_E_INVAL; |
| goto end; |
| } |
| |
| WMA_LOGD("msg->type = %x %s", msg->type, |
| mac_trace_get_wma_msg_string(msg->type)); |
| |
| wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| if (NULL == wma_handle) { |
| WMA_LOGE("%s: wma_handle is NULL", __func__); |
| QDF_ASSERT(0); |
| qdf_mem_free(msg->bodyptr); |
| qdf_status = QDF_STATUS_E_INVAL; |
| goto end; |
| } |
| |
| switch (msg->type) { |
| |
| /* Message posted by wmi for all control path related |
| * FW events to serialize through mc_thread. |
| */ |
| case WMA_PROCESS_FW_EVENT: |
| wma_process_fw_event(wma_handle, |
| (wma_process_fw_event_params *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| |
| #ifdef FEATURE_WLAN_ESE |
| case WMA_TSM_STATS_REQ: |
| WMA_LOGD("McThread: WMA_TSM_STATS_REQ"); |
| wma_process_tsm_stats_req(wma_handle, (void *)msg->bodyptr); |
| break; |
| #endif /* FEATURE_WLAN_ESE */ |
| case WNI_CFG_DNLD_REQ: |
| WMA_LOGD("McThread: WNI_CFG_DNLD_REQ"); |
| qdf_status = wma_wni_cfg_dnld(wma_handle); |
| if (QDF_IS_STATUS_SUCCESS(qdf_status)) |
| cds_wma_complete_cback(); |
| else |
| WMA_LOGD("config download failure"); |
| break; |
| case WMA_ADD_STA_SELF_REQ: |
| txrx_vdev_handle = |
| wma_vdev_attach(wma_handle, |
| (struct add_sta_self_params *) msg-> |
| bodyptr, 1); |
| if (!txrx_vdev_handle) { |
| WMA_LOGE("Failed to attach vdev"); |
| } else { |
| /* Register with TxRx Module for Data Ack Complete Cb */ |
| if (soc) { |
| cdp_data_tx_cb_set(soc, txrx_vdev_handle, |
| wma_data_tx_ack_comp_hdlr, |
| wma_handle); |
| } else { |
| WMA_LOGE("%s: SOC context is NULL", __func__); |
| qdf_status = QDF_STATUS_E_FAILURE; |
| goto end; |
| } |
| } |
| break; |
| case WMA_DEL_STA_SELF_REQ: |
| wma_vdev_detach(wma_handle, |
| (struct del_sta_self_params *) msg->bodyptr, 1); |
| break; |
| case WMA_START_SCAN_OFFLOAD_REQ: |
| wma_start_scan(wma_handle, msg->bodyptr, msg->type); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_STOP_SCAN_OFFLOAD_REQ: |
| wma_stop_scan(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_UPDATE_CHAN_LIST_REQ: |
| wma_update_channel_list(wma_handle, |
| (tSirUpdateChanList *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_LINK_STATE: |
| wma_set_linkstate(wma_handle, (tpLinkStateParams) msg->bodyptr); |
| break; |
| case WMA_CHNL_SWITCH_REQ: |
| wma_set_channel(wma_handle, |
| (tpSwitchChannelParams) msg->bodyptr); |
| break; |
| case WMA_ADD_BSS_REQ: |
| wma_add_bss(wma_handle, (tpAddBssParams) msg->bodyptr); |
| break; |
| case WMA_ADD_STA_REQ: |
| wma_add_sta(wma_handle, (tpAddStaParams) msg->bodyptr); |
| break; |
| case WMA_SET_BSSKEY_REQ: |
| wma_set_bsskey(wma_handle, (tpSetBssKeyParams) msg->bodyptr); |
| break; |
| case WMA_SET_STAKEY_REQ: |
| wma_set_stakey(wma_handle, (tpSetStaKeyParams) msg->bodyptr); |
| break; |
| case WMA_DELETE_STA_REQ: |
| wma_delete_sta(wma_handle, (tpDeleteStaParams) msg->bodyptr); |
| break; |
| case WMA_DELETE_BSS_HO_FAIL_REQ: |
| wma_delete_bss_ho_fail(wma_handle, |
| (tpDeleteBssParams) msg->bodyptr); |
| break; |
| case WMA_DELETE_BSS_REQ: |
| wma_delete_bss(wma_handle, (tpDeleteBssParams) msg->bodyptr); |
| break; |
| case WMA_UPDATE_EDCA_PROFILE_IND: |
| wma_process_update_edca_param_req(wma_handle, |
| (tEdcaParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SEND_BEACON_REQ: |
| wma_send_beacon(wma_handle, (tpSendbeaconParams) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SEND_PROBE_RSP_TMPL: |
| wma_send_probe_rsp_tmpl(wma_handle, |
| (tpSendProbeRespParams) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_CLI_SET_CMD: |
| wma_process_cli_set_cmd(wma_handle, |
| (wma_cli_set_cmd_t *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_PDEV_IE_REQ: |
| wma_process_set_pdev_ie_req(wma_handle, |
| (struct set_ie_param *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #if !defined(REMOVE_PKT_LOG) |
| case WMA_PKTLOG_ENABLE_REQ: |
| wma_pktlog_wmi_send_cmd(wma_handle, |
| (struct ath_pktlog_wmi_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* REMOVE_PKT_LOG */ |
| #if defined(QCA_WIFI_FTM) |
| case WMA_FTM_CMD_REQ: |
| wma_process_ftm_command(wma_handle, |
| (struct ar6k_testmode_cmd_data *)msg->bodyptr); |
| break; |
| #endif /* QCA_WIFI_FTM */ |
| case WMA_ENTER_PS_REQ: |
| wma_enable_sta_ps_mode(wma_handle, |
| (tpEnablePsParams) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXIT_PS_REQ: |
| wma_disable_sta_ps_mode(wma_handle, |
| (tpDisablePsParams) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_ENABLE_UAPSD_REQ: |
| wma_enable_uapsd_mode(wma_handle, |
| (tpEnableUapsdParams) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_DISABLE_UAPSD_REQ: |
| wma_disable_uapsd_mode(wma_handle, |
| (tpDisableUapsdParams) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_DTIM_PERIOD: |
| wma_set_dtim_period(wma_handle, |
| (struct set_dtim_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_TX_POWER_REQ: |
| wma_set_tx_power(wma_handle, (tpMaxTxPowerParams) msg->bodyptr); |
| break; |
| case WMA_SET_MAX_TX_POWER_REQ: |
| wma_set_max_tx_power(wma_handle, |
| (tpMaxTxPowerParams) msg->bodyptr); |
| break; |
| case WMA_SET_KEEP_ALIVE: |
| wma_set_keepalive_req(wma_handle, |
| (tSirKeepAliveReq *) msg->bodyptr); |
| break; |
| #ifdef FEATURE_WLAN_ESE |
| case WMA_SET_PLM_REQ: |
| wma_config_plm(wma_handle, (tpSirPlmReq) msg->bodyptr); |
| break; |
| #endif |
| case WMA_GET_STATISTICS_REQ: |
| wma_get_stats_req(wma_handle, |
| (tAniGetPEStatsReq *) msg->bodyptr); |
| break; |
| |
| case WMA_CONFIG_PARAM_UPDATE_REQ: |
| wma_update_cfg_params(wma_handle, msg); |
| break; |
| |
| case WMA_UPDATE_OP_MODE: |
| wma_process_update_opmode(wma_handle, |
| (tUpdateVHTOpMode *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_UPDATE_RX_NSS: |
| wma_process_update_rx_nss(wma_handle, |
| (tUpdateRxNss *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_UPDATE_MEMBERSHIP: |
| wma_process_update_membership(wma_handle, |
| (tUpdateMembership *) msg->bodyptr); |
| break; |
| case WMA_UPDATE_USERPOS: |
| wma_process_update_userpos(wma_handle, |
| (tUpdateUserPos *) msg->bodyptr); |
| break; |
| case WMA_UPDATE_BEACON_IND: |
| wma_process_update_beacon_params(wma_handle, |
| (tUpdateBeaconParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| |
| case WMA_ADD_TS_REQ: |
| wma_add_ts_req(wma_handle, (tAddTsParams *) msg->bodyptr); |
| break; |
| |
| case WMA_DEL_TS_REQ: |
| wma_del_ts_req(wma_handle, (tDelTsParams *) msg->bodyptr); |
| break; |
| |
| case WMA_AGGR_QOS_REQ: |
| wma_aggr_qos_req(wma_handle, (tAggrAddTsParams *) msg->bodyptr); |
| break; |
| |
| case WMA_WOWL_ENTER_REQ: |
| wma_wow_enter(wma_handle, |
| (tpSirHalWowlEnterParams) msg->bodyptr); |
| break; |
| case WMA_WOWL_EXIT_REQ: |
| wma_wow_exit(wma_handle, (tpSirHalWowlExitParams) msg->bodyptr); |
| break; |
| |
| case WMA_8023_MULTICAST_LIST_REQ: |
| wma_process_mcbc_set_filter_req(wma_handle, |
| (tpSirRcvFltMcAddrList) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_ROAM_SCAN_OFFLOAD_REQ: |
| /* |
| * Main entry point or roaming directives from CSR. |
| */ |
| wma_process_roaming_config(wma_handle, |
| (tSirRoamOffloadScanReq *) msg->bodyptr); |
| break; |
| |
| case WMA_RATE_UPDATE_IND: |
| wma_process_rate_update_indicate(wma_handle, |
| (tSirRateUpdateInd *) msg->bodyptr); |
| break; |
| |
| #ifdef FEATURE_WLAN_TDLS |
| case WMA_UPDATE_TDLS_PEER_STATE: |
| wma_update_tdls_peer_state(wma_handle, |
| (tTdlsPeerStateParams *) msg->bodyptr); |
| break; |
| case WMA_TDLS_SET_OFFCHAN_MODE: |
| wma_set_tdls_offchan_mode(wma_handle, |
| (tdls_chan_switch_params *)msg->bodyptr); |
| break; |
| #endif /* FEATURE_WLAN_TDLS */ |
| case WMA_ADD_PERIODIC_TX_PTRN_IND: |
| wma_process_add_periodic_tx_ptrn_ind(wma_handle, |
| (tSirAddPeriodicTxPtrn *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_DEL_PERIODIC_TX_PTRN_IND: |
| wma_process_del_periodic_tx_ptrn_ind(wma_handle, |
| (tSirDelPeriodicTxPtrn *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_TX_POWER_LIMIT: |
| wma_process_tx_power_limits(wma_handle, |
| (tSirTxPowerLimit *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #ifdef FEATURE_WLAN_CH_AVOID |
| case WMA_CH_AVOID_UPDATE_REQ: |
| wma_process_ch_avoid_update_req(wma_handle, |
| (tSirChAvoidUpdateReq *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* FEATURE_WLAN_CH_AVOID */ |
| #ifdef FEATURE_WLAN_AUTO_SHUTDOWN |
| case WMA_SET_AUTO_SHUTDOWN_TIMER_REQ: |
| wma_set_auto_shutdown_timer_req(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* FEATURE_WLAN_AUTO_SHUTDOWN */ |
| case WMA_DHCP_START_IND: |
| case WMA_DHCP_STOP_IND: |
| wma_process_dhcp_ind(wma_handle, (tAniDHCPInd *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| |
| case WMA_IBSS_CESIUM_ENABLE_IND: |
| wma_process_cesium_enable_ind(wma_handle); |
| break; |
| case WMA_GET_IBSS_PEER_INFO_REQ: |
| wma_process_get_peer_info_req(wma_handle, |
| (tSirIbssGetPeerInfoReqParams *) |
| msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_TX_FAIL_MONITOR_IND: |
| wma_process_tx_fail_monitor_ind(wma_handle, |
| (tAniTXFailMonitorInd *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| |
| case WMA_RMC_ENABLE_IND: |
| wma_process_rmc_enable_ind(wma_handle); |
| break; |
| case WMA_RMC_DISABLE_IND: |
| wma_process_rmc_disable_ind(wma_handle); |
| break; |
| case WMA_RMC_ACTION_PERIOD_IND: |
| wma_process_rmc_action_period_ind(wma_handle); |
| break; |
| case WMA_INIT_THERMAL_INFO_CMD: |
| wma_process_init_thermal_info(wma_handle, |
| (t_thermal_mgmt *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| |
| case WMA_SET_THERMAL_LEVEL: |
| wma_process_set_thermal_level(wma_handle, msg->bodyval); |
| break; |
| #ifdef CONFIG_HL_SUPPORT |
| case WMA_INIT_BAD_PEER_TX_CTL_INFO_CMD: |
| wma_process_init_bad_peer_tx_ctl_info( |
| wma_handle, |
| (struct t_bad_peer_txtcl_config *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif |
| case WMA_SET_MIMOPS_REQ: |
| wma_process_set_mimops_req(wma_handle, |
| (tSetMIMOPS *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_SAP_INTRABSS_DIS: |
| wma_set_vdev_intrabss_fwd(wma_handle, |
| (tDisableIntraBssFwd *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_GET_PEER_INFO: |
| wma_get_peer_info(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_GET_PEER_INFO_EXT: |
| wma_get_peer_info_ext(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_MODEM_POWER_STATE_IND: |
| wma_notify_modem_power_state(wma_handle, |
| (tSirModemPowerStateInd *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #ifdef WLAN_FEATURE_STATS_EXT |
| case WMA_STATS_EXT_REQUEST: |
| wma_stats_ext_req(wma_handle, |
| (tpStatsExtRequest) (msg->bodyptr)); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* WLAN_FEATURE_STATS_EXT */ |
| case WMA_HIDDEN_SSID_VDEV_RESTART: |
| wma_hidden_ssid_vdev_restart(wma_handle, |
| (tHalHiddenSsidVdevRestart *) msg->bodyptr); |
| break; |
| #ifdef WLAN_FEATURE_EXTWOW_SUPPORT |
| case WMA_WLAN_EXT_WOW: |
| wma_enable_ext_wow(wma_handle, |
| (tSirExtWoWParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_WLAN_SET_APP_TYPE1_PARAMS: |
| wma_set_app_type1_params_in_fw(wma_handle, |
| (tSirAppType1Params *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_WLAN_SET_APP_TYPE2_PARAMS: |
| wma_set_app_type2_params_in_fw(wma_handle, |
| (tSirAppType2Params *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* WLAN_FEATURE_EXTWOW_SUPPORT */ |
| #ifdef FEATURE_WLAN_EXTSCAN |
| case WMA_EXTSCAN_START_REQ: |
| wma_start_extscan(wma_handle, |
| (tSirWifiScanCmdReqParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXTSCAN_STOP_REQ: |
| wma_stop_extscan(wma_handle, |
| (tSirExtScanStopReqParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXTSCAN_SET_BSSID_HOTLIST_REQ: |
| wma_extscan_start_hotlist_monitor(wma_handle, |
| (tSirExtScanSetBssidHotListReqParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXTSCAN_RESET_BSSID_HOTLIST_REQ: |
| wma_extscan_stop_hotlist_monitor(wma_handle, |
| (tSirExtScanResetBssidHotlistReqParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXTSCAN_SET_SIGNF_CHANGE_REQ: |
| wma_extscan_start_change_monitor(wma_handle, |
| (tSirExtScanSetSigChangeReqParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXTSCAN_RESET_SIGNF_CHANGE_REQ: |
| wma_extscan_stop_change_monitor(wma_handle, |
| (tSirExtScanResetSignificantChangeReqParams *) |
| msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXTSCAN_GET_CACHED_RESULTS_REQ: |
| wma_extscan_get_cached_results(wma_handle, |
| (tSirExtScanGetCachedResultsReqParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_EXTSCAN_GET_CAPABILITIES_REQ: |
| wma_extscan_get_capabilities(wma_handle, |
| (tSirGetExtScanCapabilitiesReqParams *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_EPNO_LIST_REQ: |
| wma_set_epno_network_list(wma_handle, |
| (struct wifi_epno_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_PER_ROAM_CONFIG_CMD: |
| wma_update_per_roam_config(wma_handle, |
| (struct wmi_per_roam_config_req *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_PASSPOINT_LIST_REQ: |
| /* Issue reset passpoint network list first and clear |
| * the entries |
| */ |
| wma_reset_passpoint_network_list(wma_handle, |
| (struct wifi_passpoint_req *)msg->bodyptr); |
| |
| wma_set_passpoint_network_list(wma_handle, |
| (struct wifi_passpoint_req *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_RESET_PASSPOINT_LIST_REQ: |
| wma_reset_passpoint_network_list(wma_handle, |
| (struct wifi_passpoint_req *)msg->bodyptr); |
| break; |
| #endif /* FEATURE_WLAN_EXTSCAN */ |
| case WMA_SET_SCAN_MAC_OUI_REQ: |
| wma_scan_probe_setoui(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #ifdef WLAN_FEATURE_LINK_LAYER_STATS |
| case WMA_LINK_LAYER_STATS_CLEAR_REQ: |
| wma_process_ll_stats_clear_req(wma_handle, |
| (tpSirLLStatsClearReq) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_LINK_LAYER_STATS_SET_REQ: |
| wma_process_ll_stats_set_req(wma_handle, |
| (tpSirLLStatsSetReq) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_LINK_LAYER_STATS_GET_REQ: |
| wma_process_ll_stats_get_req(wma_handle, |
| (tpSirLLStatsGetReq) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WDA_LINK_LAYER_STATS_SET_THRESHOLD: |
| wma_config_stats_ext_threshold(wma_handle, |
| (struct sir_ll_ext_stats_threshold *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* WLAN_FEATURE_LINK_LAYER_STATS */ |
| #ifdef WLAN_FEATURE_ROAM_OFFLOAD |
| case WMA_ROAM_OFFLOAD_SYNCH_FAIL: |
| wma_process_roam_synch_fail(wma_handle, |
| (struct roam_offload_synch_fail *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_ROAM_INVOKE: |
| wma_process_roam_invoke(wma_handle, |
| (struct wma_roam_invoke_cmd *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* WLAN_FEATURE_ROAM_OFFLOAD */ |
| #ifdef WLAN_FEATURE_NAN |
| case WMA_NAN_REQUEST: |
| wma_nan_req(wma_handle, (tNanRequest *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* WLAN_FEATURE_NAN */ |
| case SIR_HAL_SET_BASE_MACADDR_IND: |
| wma_set_base_macaddr_indicate(wma_handle, |
| (tSirMacAddr *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_LINK_STATUS_GET_REQ: |
| wma_process_link_status_req(wma_handle, |
| (tAniGetLinkStatus *) msg->bodyptr); |
| break; |
| case WMA_GET_TEMPERATURE_REQ: |
| wma_get_temperature(wma_handle); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_TSF_GPIO_PIN: |
| wma_set_tsf_gpio_pin(wma_handle, msg->bodyval); |
| break; |
| |
| #ifdef DHCP_SERVER_OFFLOAD |
| case WMA_SET_DHCP_SERVER_OFFLOAD_CMD: |
| wma_process_dhcpserver_offload(wma_handle, |
| (tSirDhcpSrvOffloadInfo *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* DHCP_SERVER_OFFLOAD */ |
| #ifdef WLAN_FEATURE_GPIO_LED_FLASHING |
| case WMA_LED_FLASHING_REQ: |
| wma_set_led_flashing(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| #endif /* WLAN_FEATURE_GPIO_LED_FLASHING */ |
| case SIR_HAL_SET_MAS: |
| wma_process_set_mas(wma_handle, |
| (uint32_t *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_SET_MIRACAST: |
| wma_process_set_miracast(wma_handle, |
| (uint32_t *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_CONFIG_STATS_FACTOR: |
| wma_config_stats_factor(wma_handle, |
| (struct sir_stats_avg_factor *) |
| msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_CONFIG_GUARD_TIME: |
| wma_config_guard_time(wma_handle, |
| (struct sir_guard_time_request *) |
| msg->bodyptr); |
| case WMA_IPA_OFFLOAD_ENABLE_DISABLE: |
| wma_ipa_offload_enable_disable(wma_handle, |
| (struct sir_ipa_offload_enable_disable *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_START_STOP_LOGGING: |
| wma_set_wifi_start_packet_stats(wma_handle, |
| (struct sir_wifi_start_log *)msg->bodyptr); |
| wma_enable_specific_fw_logs(wma_handle, |
| (struct sir_wifi_start_log *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_FLUSH_LOG_TO_FW: |
| wma_send_flush_logs_to_fw(wma_handle); |
| /* Body ptr is NULL here */ |
| break; |
| case WMA_SET_RSSI_MONITOR_REQ: |
| wma_set_rssi_monitoring(wma_handle, |
| (struct rssi_monitor_req *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_FW_MEM_DUMP_REQ: |
| wma_process_fw_mem_dump_req(wma_handle, |
| (struct fw_dump_req *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_PDEV_SET_PCL_TO_FW: |
| wma_send_pdev_set_pcl_cmd(wma_handle, |
| (struct wmi_pcl_chan_weights *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_PDEV_SET_HW_MODE: |
| wma_send_pdev_set_hw_mode_cmd(wma_handle, |
| (struct policy_mgr_hw_mode *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_OCB_SET_CONFIG_CMD: |
| wma_ocb_set_config_req(wma_handle, |
| (struct sir_ocb_config *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_OCB_SET_UTC_TIME_CMD: |
| wma_ocb_set_utc_time(wma_handle, |
| (struct sir_ocb_utc *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_OCB_START_TIMING_ADVERT_CMD: |
| wma_ocb_start_timing_advert(wma_handle, |
| (struct sir_ocb_timing_advert *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_OCB_STOP_TIMING_ADVERT_CMD: |
| wma_ocb_stop_timing_advert(wma_handle, |
| (struct sir_ocb_timing_advert *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_DCC_CLEAR_STATS_CMD: |
| wma_dcc_clear_stats(wma_handle, |
| (struct sir_dcc_clear_stats *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_OCB_GET_TSF_TIMER_CMD: |
| wma_ocb_get_tsf_timer(wma_handle, |
| (struct sir_ocb_get_tsf_timer *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_WISA_PARAMS: |
| wma_set_wisa_params(wma_handle, |
| (struct sir_wisa_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_DCC_GET_STATS_CMD: |
| wma_dcc_get_stats(wma_handle, |
| (struct sir_dcc_get_stats *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_DCC_UPDATE_NDL_CMD: |
| wma_dcc_update_ndl(wma_handle, |
| (struct sir_dcc_update_ndl *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_PDEV_DUAL_MAC_CFG_REQ: |
| wma_send_pdev_set_dual_mac_config(wma_handle, |
| (struct sir_dual_mac_config *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_IE_INFO: |
| wma_process_set_ie_info(wma_handle, |
| (struct vdev_ie_info *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_SOC_ANTENNA_MODE_REQ: |
| wma_send_pdev_set_antenna_mode(wma_handle, |
| (struct sir_antenna_mode_param *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_LRO_CONFIG_CMD: |
| wma_lro_config_cmd(wma_handle, |
| (struct cdp_lro_hash_config *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_GW_PARAM_UPDATE_REQ: |
| wma_set_gateway_params(wma_handle, |
| (struct gateway_param_update_req *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_EGAP_CONF_PARAMS: |
| wma_send_egap_conf_params(wma_handle, |
| (struct egap_conf_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_ADAPT_DWELLTIME_CONF_PARAMS: |
| wma_send_adapt_dwelltime_params(wma_handle, |
| (struct adaptive_dwelltime_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_HT40_OBSS_SCAN_IND: |
| wma_send_ht40_obss_scanind(wma_handle, |
| (struct obss_ht40_scanind *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_ADD_BCN_FILTER_CMDID: |
| wma_add_beacon_filter(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_REMOVE_BCN_FILTER_CMDID: |
| wma_remove_beacon_filter(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WDA_BPF_GET_CAPABILITIES_REQ: |
| wma_get_bpf_capabilities(wma_handle); |
| break; |
| case WDA_BPF_SET_INSTRUCTIONS_REQ: |
| wma_set_bpf_instructions(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_NDP_INITIATOR_REQ: |
| wma_handle_ndp_initiator_req(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| |
| case SIR_HAL_NDP_RESPONDER_REQ: |
| wma_handle_ndp_responder_req(wma_handle, msg->bodyptr); |
| break; |
| |
| case SIR_HAL_NDP_END_REQ: |
| wma_handle_ndp_end_req(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_POWER_DBG_CMD: |
| wma_process_hal_pwr_dbg_cmd(wma_handle, |
| msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_UPDATE_WEP_DEFAULT_KEY: |
| wma_update_wep_default_key(wma_handle, |
| (struct wep_update_default_key_idx *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SEND_FREQ_RANGE_CONTROL_IND: |
| wma_enable_disable_caevent_ind(wma_handle, msg->bodyval); |
| break; |
| case WMA_ENCRYPT_DECRYPT_MSG: |
| wma_encrypt_decrypt_msg(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_UPDATE_TX_FAIL_CNT_TH: |
| wma_update_tx_fail_cnt_th(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_LONG_RETRY_LIMIT_CNT: |
| wma_update_long_retry_limit(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_SHORT_RETRY_LIMIT_CNT: |
| wma_update_short_retry_limit(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_POWER_DEBUG_STATS_REQ: |
| wma_process_power_debug_stats_req(wma_handle); |
| break; |
| case WMA_GET_RCPI_REQ: |
| wma_get_rcpi_req(wma_handle, |
| (struct sme_rcpi_req *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_WOW_PULSE_CMD: |
| wma_send_wow_pulse_cmd(wma_handle, |
| (struct wow_pulse_mode *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_DBS_SCAN_SEL_CONF_PARAMS: |
| wma_send_dbs_scan_selection_params(wma_handle, |
| (struct wmi_dbs_scan_sel_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_ARP_STATS_REQ: |
| wma_set_arp_req_stats(wma_handle, |
| (struct set_arp_stats_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_GET_ARP_STATS_REQ: |
| wma_get_arp_req_stats(wma_handle, |
| (struct get_arp_stats_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_SET_DEL_PMKID_CACHE: |
| wma_set_del_pmkid_cache(wma_handle, |
| (struct wmi_unified_pmk_cache *) msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case SIR_HAL_HLP_IE_INFO: |
| wma_roam_scan_send_hlp(wma_handle, |
| (struct hlp_params *)msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| case WMA_SET_LIMIT_OFF_CHAN: |
| wma_process_limit_off_chan(wma_handle, msg->bodyptr); |
| qdf_mem_free(msg->bodyptr); |
| break; |
| default: |
| WMA_LOGE("Unhandled WMA message of type %d", msg->type); |
| if (msg->bodyptr) |
| qdf_mem_free(msg->bodyptr); |
| } |
| end: |
| return qdf_status; |
| } |
| |
| QDF_STATUS wma_mc_process_handler(struct scheduler_msg *msg) |
| { |
| return wma_mc_process_msg(msg); |
| } |
| |
| /** |
| * wma_log_completion_timeout() - Log completion timeout |
| * @data: Timeout handler data |
| * |
| * This function is called when log completion timer expires |
| * |
| * Return: None |
| */ |
| void wma_log_completion_timeout(void *data) |
| { |
| tp_wma_handle wma_handle; |
| |
| WMA_LOGE("%s: Timeout occured for log completion command", __func__); |
| |
| wma_handle = (tp_wma_handle) data; |
| if (!wma_handle) |
| WMA_LOGE("%s: Invalid WMA handle", __func__); |
| |
| /* Though we did not receive any event from FW, |
| * we can flush whatever logs we have with us |
| */ |
| cds_logging_set_fw_flush_complete(); |
| } |
| |
| /** |
| * wma_map_pcl_weights() - Map PCL weights |
| * @pcl_weight: Internal PCL weights |
| * |
| * Maps the internal weights of PCL to the weights needed by FW |
| * |
| * Return: Mapped channel weight of type wmi_pcl_chan_weight |
| */ |
| static wmi_pcl_chan_weight wma_map_pcl_weights(uint32_t pcl_weight) |
| { |
| switch (pcl_weight) { |
| case WEIGHT_OF_GROUP1_PCL_CHANNELS: |
| return WMI_PCL_WEIGHT_VERY_HIGH; |
| case WEIGHT_OF_GROUP2_PCL_CHANNELS: |
| return WMI_PCL_WEIGHT_HIGH; |
| case WEIGHT_OF_GROUP3_PCL_CHANNELS: |
| return WMI_PCL_WEIGHT_MEDIUM; |
| case WEIGHT_OF_NON_PCL_CHANNELS: |
| return WMI_PCL_WEIGHT_LOW; |
| default: |
| return WMI_PCL_WEIGHT_DISALLOW; |
| } |
| } |
| |
| /** |
| * wma_send_pdev_set_pcl_cmd() - Send WMI_SOC_SET_PCL_CMDID to FW |
| * @wma_handle: WMA handle |
| * @msg: PCL structure containing the PCL and the number of channels |
| * |
| * WMI_PDEV_SET_PCL_CMDID provides a Preferred Channel List (PCL) to the WLAN |
| * firmware. The DBS Manager is the consumer of this information in the WLAN |
| * firmware. The channel list will be used when a Virtual DEVice (VDEV) needs |
| * to migrate to a new channel without host driver involvement. An example of |
| * this behavior is Legacy Fast Roaming (LFR 3.0). Generally, the host will |
| * manage the channel selection without firmware involvement. |
| * |
| * WMI_PDEV_SET_PCL_CMDID will carry only the weight list and not the actual |
| * channel list. The weights corresponds to the channels sent in |
| * WMI_SCAN_CHAN_LIST_CMDID. The channels from PCL would be having a higher |
| * weightage compared to the non PCL channels. |
| * |
| * Return: Success if the cmd is sent successfully to the firmware |
| */ |
| QDF_STATUS wma_send_pdev_set_pcl_cmd(tp_wma_handle wma_handle, |
| struct wmi_pcl_chan_weights *msg) |
| { |
| uint32_t i; |
| QDF_STATUS status; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: WMA handle is NULL. Cannot issue command", |
| __func__); |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| for (i = 0; i < wma_handle->saved_chan.num_channels; i++) { |
| msg->saved_chan_list[i] = |
| wma_handle->saved_chan.channel_list[i]; |
| } |
| |
| msg->saved_num_chan = wma_handle->saved_chan.num_channels; |
| status = policy_mgr_get_valid_chan_weights(wma_handle->psoc, |
| (struct policy_mgr_pcl_chan_weights *)msg); |
| |
| for (i = 0; i < msg->saved_num_chan; i++) { |
| msg->weighed_valid_list[i] = |
| wma_map_pcl_weights(msg->weighed_valid_list[i]); |
| /* Dont allow roaming on 2G when 5G_ONLY configured */ |
| if ((wma_handle->bandcapability == eCSR_BAND_5G) && |
| (msg->saved_chan_list[i] <= MAX_24GHZ_CHANNEL)) { |
| msg->weighed_valid_list[i] = |
| WEIGHT_OF_DISALLOWED_CHANNELS; |
| } |
| WMA_LOGD("%s: chan:%d weight[%d]=%d", __func__, |
| msg->saved_chan_list[i], i, |
| msg->weighed_valid_list[i]); |
| } |
| |
| if (!QDF_IS_STATUS_SUCCESS(status)) { |
| WMA_LOGE("%s: Error in creating weighed pcl", __func__); |
| return status; |
| } |
| |
| if (wmi_unified_pdev_set_pcl_cmd(wma_handle->wmi_handle, msg)) |
| return QDF_STATUS_E_FAILURE; |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_send_pdev_set_hw_mode_cmd() - Send WMI_PDEV_SET_HW_MODE_CMDID to FW |
| * @wma_handle: WMA handle |
| * @msg: Structure containing the following parameters |
| * |
| * - hw_mode_index: The HW_Mode field is a enumerated type that is selected |
| * from the HW_Mode table, which is returned in the WMI_SERVICE_READY_EVENTID. |
| * |
| * Provides notification to the WLAN firmware that host driver is requesting a |
| * HardWare (HW) Mode change. This command is needed to support iHelium in the |
| * configurations that include the Dual Band Simultaneous (DBS) feature. |
| * |
| * Return: Success if the cmd is sent successfully to the firmware |
| */ |
| QDF_STATUS wma_send_pdev_set_hw_mode_cmd(tp_wma_handle wma_handle, |
| struct policy_mgr_hw_mode *msg) |
| { |
| struct sir_set_hw_mode_resp *param; |
| struct wma_target_req *timeout_msg; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: WMA handle is NULL. Cannot issue command", |
| __func__); |
| /* Handle is NULL. Will not be able to send failure |
| * response as well |
| */ |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| if (!msg) { |
| WMA_LOGE("%s: Set HW mode param is NULL", __func__); |
| /* Lets try to free the active command list */ |
| goto fail; |
| } |
| |
| wma_acquire_wakelock(&wma_handle->wmi_cmd_rsp_wake_lock, |
| WMA_VDEV_HW_MODE_REQUEST_TIMEOUT); |
| if (wmi_unified_soc_set_hw_mode_cmd(wma_handle->wmi_handle, |
| msg->hw_mode_index)) { |
| wma_release_wakelock(&wma_handle->wmi_cmd_rsp_wake_lock); |
| goto fail; |
| } |
| timeout_msg = wma_fill_hold_req(wma_handle, 0, |
| SIR_HAL_PDEV_SET_HW_MODE, |
| WMA_PDEV_SET_HW_MODE_RESP, NULL, |
| WMA_VDEV_HW_MODE_REQUEST_TIMEOUT - 1); |
| if (!timeout_msg) { |
| WMA_LOGE("Failed to allocate request for SIR_HAL_PDEV_SET_HW_MODE"); |
| wma_remove_req(wma_handle, 0, WMA_PDEV_SET_HW_MODE_RESP); |
| } |
| |
| return QDF_STATUS_SUCCESS; |
| fail: |
| param = qdf_mem_malloc(sizeof(*param)); |
| if (!param) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| param->status = SET_HW_MODE_STATUS_ECANCELED; |
| param->cfgd_hw_mode_index = 0; |
| param->num_vdev_mac_entries = 0; |
| WMA_LOGE("%s: Sending HW mode fail response to LIM", __func__); |
| wma_send_msg(wma_handle, SIR_HAL_PDEV_SET_HW_MODE_RESP, |
| (void *) param, 0); |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_send_pdev_set_dual_mac_config() - Set dual mac config to FW |
| * @wma_handle: WMA handle |
| * @msg: Dual MAC config parameters |
| * |
| * Configures WLAN firmware with the dual MAC features |
| * |
| * Return: QDF_STATUS. 0 on success. |
| */ |
| QDF_STATUS wma_send_pdev_set_dual_mac_config(tp_wma_handle wma_handle, |
| struct sir_dual_mac_config *msg) |
| { |
| QDF_STATUS status; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: WMA handle is NULL. Cannot issue command", |
| __func__); |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| if (!msg) { |
| WMA_LOGE("%s: Set dual mode config is NULL", __func__); |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| /* |
| * aquire the wake lock here and release it in response handler function |
| * In error condition, release the wake lock right away |
| */ |
| wma_acquire_wakelock(&wma_handle->wmi_cmd_rsp_wake_lock, |
| WMA_VDEV_PLCY_MGR_CMD_TIMEOUT); |
| status = wmi_unified_pdev_set_dual_mac_config_cmd( |
| wma_handle->wmi_handle, |
| (struct wmi_dual_mac_config *)msg); |
| if (QDF_IS_STATUS_ERROR(status)) { |
| WMA_LOGE("%s: Failed to send WMI_PDEV_SET_DUAL_MAC_CONFIG_CMDID: %d", |
| __func__, status); |
| wma_release_wakelock(&wma_handle->wmi_cmd_rsp_wake_lock); |
| return status; |
| } |
| policy_mgr_update_dbs_req_config(wma_handle->psoc, |
| msg->scan_config, msg->fw_mode_config); |
| |
| return QDF_STATUS_SUCCESS; |
| } |
| |
| /** |
| * wma_send_pdev_set_antenna_mode() - Set antenna mode to FW |
| * @wma_handle: WMA handle |
| * @msg: Antenna mode parameters |
| * |
| * Send WMI_PDEV_SET_ANTENNA_MODE_CMDID to FW requesting to |
| * modify the number of TX/RX chains from host |
| * |
| * Return: QDF_STATUS. 0 on success. |
| */ |
| QDF_STATUS wma_send_pdev_set_antenna_mode(tp_wma_handle wma_handle, |
| struct sir_antenna_mode_param *msg) |
| { |
| wmi_pdev_set_antenna_mode_cmd_fixed_param *cmd; |
| wmi_buf_t buf; |
| uint32_t len; |
| QDF_STATUS status = QDF_STATUS_SUCCESS; |
| struct sir_antenna_mode_resp *param; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: WMA handle is NULL. Cannot issue command", |
| __func__); |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| if (!msg) { |
| WMA_LOGE("%s: Set antenna mode param is NULL", __func__); |
| return QDF_STATUS_E_NULL_VALUE; |
| } |
| |
| len = sizeof(*cmd); |
| |
| buf = wmi_buf_alloc(wma_handle->wmi_handle, len); |
| if (!buf) { |
| WMA_LOGE("%s: wmi_buf_alloc failed", __func__); |
| status = QDF_STATUS_E_NOMEM; |
| goto resp; |
| } |
| |
| cmd = (wmi_pdev_set_antenna_mode_cmd_fixed_param *) wmi_buf_data(buf); |
| WMITLV_SET_HDR(&cmd->tlv_header, |
| WMITLV_TAG_STRUC_wmi_pdev_set_antenna_mode_cmd_fixed_param, |
| WMITLV_GET_STRUCT_TLVLEN( |
| wmi_pdev_set_antenna_mode_cmd_fixed_param)); |
| |
| cmd->pdev_id = WMI_PDEV_ID_SOC; |
| /* Bits 0-15 is num of RX chains 16-31 is num of TX chains */ |
| cmd->num_txrx_chains = msg->num_rx_chains; |
| cmd->num_txrx_chains |= (msg->num_tx_chains << 16); |
| |
| WMA_LOGD("%s: Num of chains TX: %d RX: %d txrx_chains: 0x%x", |
| __func__, msg->num_tx_chains, |
| msg->num_rx_chains, cmd->num_txrx_chains); |
| |
| if (wmi_unified_cmd_send(wma_handle->wmi_handle, buf, len, |
| WMI_PDEV_SET_ANTENNA_MODE_CMDID)) { |
| WMA_LOGE("%s: Failed to send WMI_PDEV_SET_ANTENNA_MODE_CMDID", |
| __func__); |
| wmi_buf_free(buf); |
| status = QDF_STATUS_E_FAILURE; |
| goto resp; |
| } |
| status = QDF_STATUS_SUCCESS; |
| |
| resp: |
| param = qdf_mem_malloc(sizeof(*param)); |
| if (!param) { |
| WMA_LOGE("%s: Memory allocation failed", __func__); |
| return QDF_STATUS_E_NOMEM; |
| } |
| param->status = (status) ? |
| SET_ANTENNA_MODE_STATUS_ECANCELED : |
| SET_ANTENNA_MODE_STATUS_OK; |
| WMA_LOGE("%s: Send antenna mode resp to LIM status: %d", |
| __func__, param->status); |
| wma_send_msg(wma_handle, SIR_HAL_SOC_ANTENNA_MODE_RESP, |
| (void *) param, 0); |
| return status; |
| } |
| |
| /** |
| * wma_crash_inject() - sends command to FW to simulate crash |
| * @wma_handle: pointer of WMA context |
| * @type: subtype of the command |
| * @delay_time_ms: time in milliseconds for FW to delay the crash |
| * |
| * This function will send a command to FW in order to simulate different |
| * kinds of FW crashes. |
| * |
| * Return: QDF_STATUS_SUCCESS for success or error code |
| */ |
| QDF_STATUS wma_crash_inject(WMA_HANDLE wma_handle, uint32_t type, |
| uint32_t delay_time_ms) |
| { |
| struct crash_inject param; |
| tp_wma_handle wma = (tp_wma_handle)wma_handle; |
| |
| param.type = type; |
| param.delay_time_ms = delay_time_ms; |
| return wmi_crash_inject(wma->wmi_handle, ¶m); |
| } |
| |
| #if defined(FEATURE_LRO) |
| /** |
| * wma_lro_init() - sends LRO configuration to FW |
| * @lro_config: pointer to the config parameters |
| * |
| * This function ends LRO configuration to FW. |
| * |
| * Return: 0 for success or reasons for failure |
| */ |
| int wma_lro_init(struct cdp_lro_hash_config *lro_config) |
| { |
| struct scheduler_msg msg = {0}; |
| struct cdp_lro_hash_config *iwcmd; |
| |
| iwcmd = qdf_mem_malloc(sizeof(*iwcmd)); |
| if (!iwcmd) { |
| WMA_LOGE("memory allocation for WMA_LRO_CONFIG_CMD failed!"); |
| return -ENOMEM; |
| } |
| |
| *iwcmd = *lro_config; |
| |
| msg.type = WMA_LRO_CONFIG_CMD; |
| msg.reserved = 0; |
| msg.bodyptr = iwcmd; |
| |
| if (QDF_STATUS_SUCCESS != |
| scheduler_post_msg(QDF_MODULE_ID_WMA, &msg)) { |
| WMA_LOGE("Failed to post WMA_LRO_CONFIG_CMD msg!"); |
| qdf_mem_free(iwcmd); |
| return -EAGAIN; |
| } |
| |
| WMA_LOGD("sending the LRO configuration to the fw"); |
| return 0; |
| } |
| #endif |
| |
| void wma_peer_set_default_routing(void *scn_handle, uint8_t *peer_macaddr, |
| uint8_t vdev_id, bool hash_based, uint8_t ring_num) |
| { |
| tp_wma_handle wma = cds_get_context(QDF_MODULE_ID_WMA); |
| struct peer_set_params param; |
| |
| if (!wma) { |
| WMA_LOGE("%s:wma_handle is NULL", __func__); |
| return; |
| } |
| |
| /* TODO: Need bit definitions for ring number and hash based routing |
| * fields in common wmi header file |
| */ |
| param.param_id = WMI_HOST_PEER_SET_DEFAULT_ROUTING; |
| param.vdev_id = vdev_id; |
| param.param_value = ((hash_based) ? 1 : 0) | (ring_num << 1); |
| WMA_LOGD("%s: param_value 0x%x", __func__, param.param_value); |
| wmi_set_peer_param_send(wma->wmi_handle, peer_macaddr, ¶m); |
| } |
| |
| int wma_peer_rx_reorder_queue_setup(void *scn_handle, |
| uint8_t vdev_id, uint8_t *peer_macaddr, qdf_dma_addr_t hw_qdesc, |
| int tid, uint16_t queue_no) |
| { |
| tp_wma_handle wma = cds_get_context(QDF_MODULE_ID_WMA); |
| struct rx_reorder_queue_setup_params param; |
| |
| if (!wma) { |
| WMA_LOGE("%s:wma_handle is NULL", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| param.tid = tid; |
| param.vdev_id = vdev_id; |
| param.peer_macaddr = peer_macaddr; |
| param.hw_qdesc_paddr_lo = hw_qdesc & 0xffffffff; |
| param.hw_qdesc_paddr_hi = (uint64_t)hw_qdesc >> 32; |
| param.queue_no = queue_no; |
| |
| return wmi_unified_peer_rx_reorder_queue_setup_send(wma->wmi_handle, |
| ¶m); |
| } |
| |
| int wma_peer_rx_reorder_queue_remove(void *scn_handle, |
| uint8_t vdev_id, uint8_t *peer_macaddr, uint32_t peer_tid_bitmap) |
| { |
| tp_wma_handle wma = cds_get_context(QDF_MODULE_ID_WMA); |
| struct rx_reorder_queue_remove_params param; |
| |
| if (!wma) { |
| WMA_LOGE("%s:wma_handle is NULL", __func__); |
| return QDF_STATUS_E_FAILURE; |
| } |
| |
| param.vdev_id = vdev_id; |
| param.peer_macaddr = peer_macaddr; |
| param.peer_tid_bitmap = peer_tid_bitmap; |
| |
| return wmi_unified_peer_rx_reorder_queue_remove_send(wma->wmi_handle, |
| ¶m); |
| } |
| |
| QDF_STATUS wma_configure_smps_params(uint32_t vdev_id, uint32_t param_id, |
| uint32_t param_val) |
| { |
| tp_wma_handle wma = cds_get_context(QDF_MODULE_ID_WMA); |
| int smps_cmd_value; |
| int status = QDF_STATUS_E_INVAL; |
| |
| if (!wma) { |
| WMA_LOGE("%s: Failed to get wma", __func__); |
| return status; |
| } |
| |
| smps_cmd_value = param_id << WMI_SMPS_PARAM_VALUE_S; |
| smps_cmd_value = smps_cmd_value | param_val; |
| |
| status = wma_set_smps_params(wma, vdev_id, smps_cmd_value); |
| if (status) |
| WMA_LOGE("Failed to set SMPS Param"); |
| |
| return status; |
| } |
| |
| |
| void wma_ipa_uc_stat_request(wma_cli_set_cmd_t *privcmd) |
| { |
| tp_wma_handle wma = cds_get_context(QDF_MODULE_ID_WMA); |
| |
| if (!wma) { |
| WMA_LOGE("%s: Failed to get wma", __func__); |
| return; |
| } |
| |
| if (wma_set_priv_cfg(wma, privcmd)) |
| WMA_LOGE("Failed to set wma priv congiuration"); |
| } |
| |
| /** |
| * wma_config_bmiss_bcnt_params() - set bmiss config parameters |
| * @vdev_id: virtual device for the command |
| * @first_cnt: bmiss first value |
| * @final_cnt: bmiss final value |
| * |
| * Return: QDF_STATUS_SUCCESS or non-zero on failure |
| */ |
| QDF_STATUS wma_config_bmiss_bcnt_params(uint32_t vdev_id, uint32_t first_cnt, |
| uint32_t final_cnt) |
| { |
| tp_wma_handle wma_handle = cds_get_context(QDF_MODULE_ID_WMA); |
| int status = QDF_STATUS_E_INVAL; |
| |
| if (!wma_handle) { |
| WMA_LOGE("%s: Failed to get wma", __func__); |
| return status; |
| } |
| |
| status = wma_roam_scan_bmiss_cnt(wma_handle, first_cnt, final_cnt, |
| vdev_id); |
| |
| if (status) |
| WMA_LOGE("Failed to set Bmiss Param"); |
| |
| return status; |
| } |
| |