| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Copyright (c) 2017-2019, The Linux Foundation. All rights reserved. |
| */ |
| |
| #include <linux/debugfs.h> |
| #include "ipa_pm.h" |
| #include "ipa_i.h" |
| |
| |
| #define IPA_PM_DRV_NAME "ipa_pm" |
| |
| #define IPA_PM_DBG(fmt, args...) \ |
| do { \ |
| pr_debug(IPA_PM_DRV_NAME " %s:%d " fmt, \ |
| __func__, __LINE__, ## args); \ |
| IPA_IPC_LOGGING(ipa_get_ipc_logbuf(), \ |
| IPA_PM_DRV_NAME " %s:%d " fmt, ## args); \ |
| IPA_IPC_LOGGING(ipa_get_ipc_logbuf_low(), \ |
| IPA_PM_DRV_NAME " %s:%d " fmt, ## args); \ |
| } while (0) |
| #define IPA_PM_DBG_LOW(fmt, args...) \ |
| do { \ |
| pr_debug(IPA_PM_DRV_NAME " %s:%d " fmt, \ |
| __func__, __LINE__, ## args); \ |
| IPA_IPC_LOGGING(ipa_get_ipc_logbuf_low(), \ |
| IPA_PM_DRV_NAME " %s:%d " fmt, ## args); \ |
| } while (0) |
| #define IPA_PM_ERR(fmt, args...) \ |
| do { \ |
| pr_err(IPA_PM_DRV_NAME " %s:%d " fmt, \ |
| __func__, __LINE__, ## args); \ |
| IPA_IPC_LOGGING(ipa_get_ipc_logbuf(), \ |
| IPA_PM_DRV_NAME " %s:%d " fmt, ## args); \ |
| IPA_IPC_LOGGING(ipa_get_ipc_logbuf_low(), \ |
| IPA_PM_DRV_NAME " %s:%d " fmt, ## args); \ |
| } while (0) |
| #define IPA_PM_DBG_STATE(hdl, name, state) \ |
| IPA_PM_DBG_LOW("Client[%d] %s: %s\n", hdl, name, \ |
| client_state_to_str[state]) |
| |
| |
| #if IPA_PM_MAX_CLIENTS > 32 |
| #error max client greater than 32 all bitmask types should be changed |
| #endif |
| |
| /* |
| * struct ipa_pm_exception_list - holds information about an exception |
| * @pending: number of clients in exception that have not yet been adctivated |
| * @bitmask: bitmask of the clients in the exception based on handle |
| * @threshold: the threshold values for the exception |
| */ |
| struct ipa_pm_exception_list { |
| char clients[IPA_PM_MAX_EX_CL]; |
| int pending; |
| u32 bitmask; |
| int threshold[IPA_PM_THRESHOLD_MAX]; |
| }; |
| |
| /* |
| * struct clk_scaling_db - holds information about threshholds and exceptions |
| * @lock: lock the bitmasks and thresholds |
| * @exception_list: pointer to the list of exceptions |
| * @work: work for clock scaling algorithm |
| * @active_client_bitmask: the bits represent handles in the clients array that |
| * contain non-null client |
| * @threshold_size: size of the throughput threshold |
| * @exception_size: size of the exception list |
| * @cur_vote: idx of the threshold |
| * @default_threshold: the thresholds used if no exception passes |
| * @current_threshold: the current threshold of the clock plan |
| */ |
| struct clk_scaling_db { |
| spinlock_t lock; |
| struct ipa_pm_exception_list exception_list[IPA_PM_EXCEPTION_MAX]; |
| struct work_struct work; |
| u32 active_client_bitmask; |
| int threshold_size; |
| int exception_size; |
| int cur_vote; |
| int default_threshold[IPA_PM_THRESHOLD_MAX]; |
| int *current_threshold; |
| }; |
| |
| /* |
| * ipa_pm state names |
| * |
| * Timer free states: |
| * @IPA_PM_DEACTIVATED: client starting state when registered |
| * @IPA_PM_DEACTIVATE_IN_PROGRESS: deactivate was called in progress of a client |
| * activating |
| * @IPA_PM_ACTIVATE_IN_PROGRESS: client is being activated by work_queue |
| * @IPA_PM_ACTIVATED: client is activated without any timers |
| * |
| * Timer set states: |
| * @IPA_PM_ACTIVATED_PENDING_DEACTIVATION: moves to deactivate once timer pass |
| * @IPA_PM_ACTIVATED_TIMER_SET: client was activated while timer was set, so |
| * when the timer pass, client will still be activated |
| *@IPA_PM_ACTIVATED_PENDING_RESCHEDULE: state signifying extended timer when |
| * a client is deferred_deactivated when a time ris still active |
| */ |
| enum ipa_pm_state { |
| IPA_PM_DEACTIVATED, |
| IPA_PM_DEACTIVATE_IN_PROGRESS, |
| IPA_PM_ACTIVATE_IN_PROGRESS, |
| IPA_PM_ACTIVATED, |
| IPA_PM_ACTIVATED_PENDING_DEACTIVATION, |
| IPA_PM_ACTIVATED_TIMER_SET, |
| IPA_PM_ACTIVATED_PENDING_RESCHEDULE, |
| IPA_PM_STATE_MAX |
| }; |
| |
| #define IPA_PM_STATE_ACTIVE(state) \ |
| (state == IPA_PM_ACTIVATED ||\ |
| state == IPA_PM_ACTIVATED_PENDING_DEACTIVATION ||\ |
| state == IPA_PM_ACTIVATED_TIMER_SET ||\ |
| state == IPA_PM_ACTIVATED_PENDING_RESCHEDULE) |
| |
| #define IPA_PM_STATE_IN_PROGRESS(state) \ |
| (state == IPA_PM_ACTIVATE_IN_PROGRESS \ |
| || state == IPA_PM_DEACTIVATE_IN_PROGRESS) |
| |
| /* |
| * struct ipa_pm_client - holds information about a specific IPA client |
| * @name: string name of the client |
| * @callback: pointer to the client's callback function |
| * @callback_params: pointer to the client's callback parameters |
| * @state: Activation state of the client |
| * @skip_clk_vote: 0 if client votes for clock when activated, 1 if no vote |
| * @group: the ipa_pm_group the client belongs to |
| * @hdl: handle of the client |
| * @throughput: the throughput of the client for clock scaling |
| * @state_lock: spinlock to lock the pm_states |
| * @activate_work: work for activate (blocking case) |
| * @deactivate work: delayed work for deferred_deactivate function |
| * @complete: generic wait-for-completion handler |
| * @wlock: wake source to prevent AP suspend |
| */ |
| struct ipa_pm_client { |
| char name[IPA_PM_MAX_EX_CL]; |
| void (*callback)(void *user_data, enum ipa_pm_cb_event); |
| void *callback_params; |
| enum ipa_pm_state state; |
| bool skip_clk_vote; |
| int group; |
| int hdl; |
| int throughput; |
| spinlock_t state_lock; |
| struct work_struct activate_work; |
| struct delayed_work deactivate_work; |
| struct completion complete; |
| struct wakeup_source wlock; |
| }; |
| |
| /* |
| * struct ipa_pm_ctx - global ctx that will hold the client arrays and tput info |
| * @clients: array to the clients with the handle as its index |
| * @clients_by_pipe: array to the clients with endpoint as the index |
| * @wq: work queue for deferred deactivate, activate, and clk_scaling work |
| 8 @clk_scaling: pointer to clock scaling database |
| * @client_mutex: global mutex to lock the client arrays |
| * @aggragated_tput: aggragated tput value of all valid activated clients |
| * @group_tput: combined throughput for the groups |
| */ |
| struct ipa_pm_ctx { |
| struct ipa_pm_client *clients[IPA_PM_MAX_CLIENTS]; |
| struct ipa_pm_client *clients_by_pipe[IPA3_MAX_NUM_PIPES]; |
| struct workqueue_struct *wq; |
| struct clk_scaling_db clk_scaling; |
| struct mutex client_mutex; |
| int aggregated_tput; |
| int group_tput[IPA_PM_GROUP_MAX]; |
| }; |
| |
| static struct ipa_pm_ctx *ipa_pm_ctx; |
| |
| static const char *client_state_to_str[IPA_PM_STATE_MAX] = { |
| __stringify(IPA_PM_DEACTIVATED), |
| __stringify(IPA_PM_DEACTIVATE_IN_PROGRESS), |
| __stringify(IPA_PM_ACTIVATE_IN_PROGRESS), |
| __stringify(IPA_PM_ACTIVATED), |
| __stringify(IPA_PM_ACTIVATED_PENDING_DEACTIVATION), |
| __stringify(IPA_PM_ACTIVATED_TIMER_SET), |
| __stringify(IPA_PM_ACTIVATED_PENDING_RESCHEDULE), |
| }; |
| |
| static const char *ipa_pm_group_to_str[IPA_PM_GROUP_MAX] = { |
| __stringify(IPA_PM_GROUP_DEFAULT), |
| __stringify(IPA_PM_GROUP_APPS), |
| __stringify(IPA_PM_GROUP_MODEM), |
| }; |
| |
| /** |
| * pop_max_from_array() -pop the max and move the last element to where the |
| * max was popped |
| * @arr: array to be searched for max |
| * @n: size of the array |
| * |
| * Returns: max value of the array |
| */ |
| static int pop_max_from_array(int *arr, int *n) |
| { |
| int i; |
| int max, max_idx; |
| |
| max_idx = *n - 1; |
| max = 0; |
| |
| if (*n == 0) |
| return 0; |
| |
| for (i = 0; i < *n; i++) { |
| if (arr[i] > max) { |
| max = arr[i]; |
| max_idx = i; |
| } |
| } |
| (*n)--; |
| arr[max_idx] = arr[*n]; |
| |
| return max; |
| } |
| |
| /** |
| * calculate_throughput() - calculate the aggregated throughput |
| * based on active clients |
| * |
| * Returns: aggregated tput value |
| */ |
| static int calculate_throughput(void) |
| { |
| int client_tput[IPA_PM_MAX_CLIENTS] = { 0 }; |
| bool group_voted[IPA_PM_GROUP_MAX] = { false }; |
| int i, n; |
| int max, second_max, aggregated_tput; |
| struct ipa_pm_client *client; |
| |
| /* Create a basic array to hold throughputs*/ |
| for (i = 1, n = 0; i < IPA_PM_MAX_CLIENTS; i++) { |
| client = ipa_pm_ctx->clients[i]; |
| if (client != NULL && IPA_PM_STATE_ACTIVE(client->state)) { |
| /* default case */ |
| if (client->group == IPA_PM_GROUP_DEFAULT) { |
| client_tput[n++] = client->throughput; |
| } else if (!group_voted[client->group]) { |
| client_tput[n++] = ipa_pm_ctx->group_tput |
| [client->group]; |
| group_voted[client->group] = true; |
| } |
| } |
| } |
| /*the array will only use n+1 spots. n will be the last index used*/ |
| |
| aggregated_tput = 0; |
| |
| /** |
| * throughput algorithm: |
| * 1) pop the max and second_max |
| * 2) add the 2nd max to aggregated tput |
| * 3) insert the value of max - 2nd max |
| * 4) repeat until array is of size 1 |
| */ |
| while (n > 1) { |
| max = pop_max_from_array(client_tput, &n); |
| second_max = pop_max_from_array(client_tput, &n); |
| client_tput[n++] = max - second_max; |
| aggregated_tput += second_max; |
| } |
| |
| IPA_PM_DBG_LOW("Aggregated throughput: %d\n", aggregated_tput); |
| |
| return aggregated_tput; |
| } |
| |
| /** |
| * deactivate_client() - turn off the bit in the active client bitmask based on |
| * the handle passed in |
| * @hdl: The index of the client to be deactivated |
| */ |
| static void deactivate_client(u32 hdl) |
| { |
| unsigned long flags; |
| |
| spin_lock_irqsave(&ipa_pm_ctx->clk_scaling.lock, flags); |
| ipa_pm_ctx->clk_scaling.active_client_bitmask &= ~(1 << hdl); |
| spin_unlock_irqrestore(&ipa_pm_ctx->clk_scaling.lock, flags); |
| IPA_PM_DBG_LOW("active bitmask: %x\n", |
| ipa_pm_ctx->clk_scaling.active_client_bitmask); |
| } |
| |
| /** |
| * activate_client() - turn on the bit in the active client bitmask based on |
| * the handle passed in |
| * @hdl: The index of the client to be activated |
| */ |
| static void activate_client(u32 hdl) |
| { |
| unsigned long flags; |
| |
| spin_lock_irqsave(&ipa_pm_ctx->clk_scaling.lock, flags); |
| ipa_pm_ctx->clk_scaling.active_client_bitmask |= (1 << hdl); |
| spin_unlock_irqrestore(&ipa_pm_ctx->clk_scaling.lock, flags); |
| IPA_PM_DBG_LOW("active bitmask: %x\n", |
| ipa_pm_ctx->clk_scaling.active_client_bitmask); |
| } |
| |
| /** |
| * deactivate_client() - get threshold |
| * |
| * Returns: threshold of the exception that passes or default if none pass |
| */ |
| static void set_current_threshold(void) |
| { |
| int i; |
| struct clk_scaling_db *clk; |
| struct ipa_pm_exception_list *exception; |
| unsigned long flags; |
| |
| clk = &ipa_pm_ctx->clk_scaling; |
| |
| spin_lock_irqsave(&ipa_pm_ctx->clk_scaling.lock, flags); |
| for (i = 0; i < clk->exception_size; i++) { |
| exception = &clk->exception_list[i]; |
| if (exception->pending == 0 && (exception->bitmask |
| & ~clk->active_client_bitmask) == 0) { |
| spin_unlock_irqrestore(&ipa_pm_ctx->clk_scaling.lock, |
| flags); |
| clk->current_threshold = exception->threshold; |
| IPA_PM_DBG("Exception %d set\n", i); |
| return; |
| } |
| } |
| clk->current_threshold = clk->default_threshold; |
| spin_unlock_irqrestore(&ipa_pm_ctx->clk_scaling.lock, flags); |
| } |
| |
| /** |
| * do_clk_scaling() - set the clock based on the activated clients |
| * |
| * Returns: 0 if success, negative otherwise |
| */ |
| static int do_clk_scaling(void) |
| { |
| int i, tput; |
| int new_th_idx = 1; |
| struct clk_scaling_db *clk_scaling; |
| |
| if (atomic_read(&ipa3_ctx->ipa_clk_vote) == 0) { |
| IPA_PM_DBG("IPA clock is gated\n"); |
| return 0; |
| } |
| |
| clk_scaling = &ipa_pm_ctx->clk_scaling; |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| IPA_PM_DBG_LOW("clock scaling started\n"); |
| tput = calculate_throughput(); |
| ipa_pm_ctx->aggregated_tput = tput; |
| set_current_threshold(); |
| |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| for (i = 0; i < clk_scaling->threshold_size; i++) { |
| if (tput >= clk_scaling->current_threshold[i]) |
| new_th_idx++; |
| } |
| |
| IPA_PM_DBG_LOW("old idx was at %d\n", ipa_pm_ctx->clk_scaling.cur_vote); |
| |
| |
| if (ipa_pm_ctx->clk_scaling.cur_vote != new_th_idx) { |
| ipa_pm_ctx->clk_scaling.cur_vote = new_th_idx; |
| ipa3_set_clock_plan_from_pm(ipa_pm_ctx->clk_scaling.cur_vote); |
| } |
| |
| IPA_PM_DBG_LOW("new idx is at %d\n", ipa_pm_ctx->clk_scaling.cur_vote); |
| |
| return 0; |
| } |
| |
| /** |
| * clock_scaling_func() - set the clock on a work queue |
| */ |
| static void clock_scaling_func(struct work_struct *work) |
| { |
| do_clk_scaling(); |
| } |
| |
| /** |
| * activate_work_func - activate a client and vote for clock on a work queue |
| */ |
| static void activate_work_func(struct work_struct *work) |
| { |
| struct ipa_pm_client *client; |
| bool dec_clk = false; |
| unsigned long flags; |
| |
| client = container_of(work, struct ipa_pm_client, activate_work); |
| if (!client->skip_clk_vote) { |
| IPA_ACTIVE_CLIENTS_INC_SPECIAL(client->name); |
| if (client->group == IPA_PM_GROUP_APPS) |
| __pm_stay_awake(&client->wlock); |
| } |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| if (client->state == IPA_PM_ACTIVATE_IN_PROGRESS) { |
| client->state = IPA_PM_ACTIVATED; |
| } else if (client->state == IPA_PM_DEACTIVATE_IN_PROGRESS) { |
| client->state = IPA_PM_DEACTIVATED; |
| dec_clk = true; |
| } else { |
| IPA_PM_ERR("unexpected state %d\n", client->state); |
| WARN_ON(1); |
| } |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| |
| complete_all(&client->complete); |
| |
| if (dec_clk) { |
| if (!client->skip_clk_vote) { |
| IPA_ACTIVE_CLIENTS_DEC_SPECIAL(client->name); |
| if (client->group == IPA_PM_GROUP_APPS) |
| __pm_relax(&client->wlock); |
| } |
| |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| return; |
| } |
| |
| activate_client(client->hdl); |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| if (client->callback) { |
| client->callback(client->callback_params, |
| IPA_PM_CLIENT_ACTIVATED); |
| } else { |
| IPA_PM_ERR("client has no callback"); |
| WARN_ON(1); |
| } |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| do_clk_scaling(); |
| } |
| |
| /** |
| * delayed_deferred_deactivate_work_func - deferred deactivate on a work queue |
| */ |
| static void delayed_deferred_deactivate_work_func(struct work_struct *work) |
| { |
| struct delayed_work *dwork; |
| struct ipa_pm_client *client; |
| unsigned long flags; |
| unsigned long delay; |
| |
| dwork = container_of(work, struct delayed_work, work); |
| client = container_of(dwork, struct ipa_pm_client, deactivate_work); |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| switch (client->state) { |
| case IPA_PM_ACTIVATED_TIMER_SET: |
| client->state = IPA_PM_ACTIVATED; |
| goto bail; |
| case IPA_PM_ACTIVATED_PENDING_RESCHEDULE: |
| delay = IPA_PM_DEFERRED_TIMEOUT; |
| if (ipa3_ctx->ipa3_hw_mode == IPA_HW_MODE_VIRTUAL || |
| ipa3_ctx->ipa3_hw_mode == IPA_HW_MODE_EMULATION) |
| delay *= 5; |
| |
| queue_delayed_work(ipa_pm_ctx->wq, &client->deactivate_work, |
| msecs_to_jiffies(delay)); |
| client->state = IPA_PM_ACTIVATED_PENDING_DEACTIVATION; |
| goto bail; |
| case IPA_PM_ACTIVATED_PENDING_DEACTIVATION: |
| client->state = IPA_PM_DEACTIVATED; |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| if (!client->skip_clk_vote) { |
| IPA_ACTIVE_CLIENTS_DEC_SPECIAL(client->name); |
| if (client->group == IPA_PM_GROUP_APPS) |
| __pm_relax(&client->wlock); |
| } |
| |
| deactivate_client(client->hdl); |
| do_clk_scaling(); |
| return; |
| default: |
| IPA_PM_ERR("unexpected state %d\n", client->state); |
| WARN_ON(1); |
| goto bail; |
| } |
| |
| bail: |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| } |
| |
| static int find_next_open_array_element(const char *name) |
| { |
| int i, n; |
| |
| n = -ENOBUFS; |
| |
| /* 0 is not a valid handle */ |
| for (i = IPA_PM_MAX_CLIENTS - 1; i >= 1; i--) { |
| if (ipa_pm_ctx->clients[i] == NULL) { |
| n = i; |
| continue; |
| } |
| |
| if (strlen(name) == strlen(ipa_pm_ctx->clients[i]->name)) |
| if (!strcmp(name, ipa_pm_ctx->clients[i]->name)) |
| return -EEXIST; |
| } |
| return n; |
| } |
| |
| /** |
| * add_client_to_exception_list() - add client to the exception list and |
| * update pending if necessary |
| * @hdl: index of the IPA client |
| * |
| * Returns: 0 if success, negative otherwise |
| */ |
| static int add_client_to_exception_list(u32 hdl) |
| { |
| int i, len = 0; |
| struct ipa_pm_exception_list *exception; |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| len = strlen(ipa_pm_ctx->clients[hdl]->name); |
| for (i = 0; i < ipa_pm_ctx->clk_scaling.exception_size; i++) { |
| exception = &ipa_pm_ctx->clk_scaling.exception_list[i]; |
| if (strnstr(exception->clients, ipa_pm_ctx->clients[hdl]->name, |
| len) && (strlen(exception->clients) |
| == len)) { |
| exception->pending--; |
| IPA_PM_DBG("Pending: %d\n", |
| exception->pending); |
| |
| if (exception->pending < 0) { |
| WARN_ON(1); |
| exception->pending = 0; |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| return -EPERM; |
| } |
| exception->bitmask |= (1 << hdl); |
| } |
| } |
| IPA_PM_DBG("%s added to exception list\n", |
| ipa_pm_ctx->clients[hdl]->name); |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| return 0; |
| } |
| |
| /** |
| * remove_client_to_exception_list() - remove client from the exception list and |
| * update pending if necessary |
| * @hdl: index of the IPA client |
| * |
| * Returns: 0 if success, negative otherwise |
| */ |
| static int remove_client_from_exception_list(u32 hdl) |
| { |
| int i; |
| struct ipa_pm_exception_list *exception; |
| |
| for (i = 0; i < ipa_pm_ctx->clk_scaling.exception_size; i++) { |
| exception = &ipa_pm_ctx->clk_scaling.exception_list[i]; |
| if (exception->bitmask & (1 << hdl)) { |
| exception->pending++; |
| IPA_PM_DBG("Pending: %d\n", |
| exception->pending); |
| exception->bitmask &= ~(1 << hdl); |
| } |
| } |
| IPA_PM_DBG("Client %d removed from exception list\n", hdl); |
| |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_init() - initialize IPA PM Components |
| * @ipa_pm_init_params: parameters needed to fill exceptions and thresholds |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_init(struct ipa_pm_init_params *params) |
| { |
| int i, j; |
| struct clk_scaling_db *clk_scaling; |
| |
| if (params == NULL) { |
| IPA_PM_ERR("Invalid Params\n"); |
| return -EINVAL; |
| } |
| |
| if (params->threshold_size <= 0 |
| || params->threshold_size > IPA_PM_THRESHOLD_MAX) { |
| IPA_PM_ERR("Invalid threshold size\n"); |
| return -EINVAL; |
| } |
| |
| if (params->exception_size < 0 |
| || params->exception_size > IPA_PM_EXCEPTION_MAX) { |
| IPA_PM_ERR("Invalid exception size\n"); |
| return -EINVAL; |
| } |
| |
| IPA_PM_DBG("IPA PM initialization started\n"); |
| |
| if (ipa_pm_ctx != NULL) { |
| IPA_PM_ERR("Already initialized\n"); |
| return -EPERM; |
| } |
| |
| |
| ipa_pm_ctx = kzalloc(sizeof(*ipa_pm_ctx), GFP_KERNEL); |
| if (!ipa_pm_ctx) { |
| IPA_PM_ERR(":kzalloc err.\n"); |
| return -ENOMEM; |
| } |
| |
| ipa_pm_ctx->wq = create_singlethread_workqueue("ipa_pm_activate"); |
| if (!ipa_pm_ctx->wq) { |
| IPA_PM_ERR("create workqueue failed\n"); |
| kfree(ipa_pm_ctx); |
| return -ENOMEM; |
| } |
| |
| mutex_init(&ipa_pm_ctx->client_mutex); |
| |
| /* Populate and init locks in clk_scaling_db */ |
| clk_scaling = &ipa_pm_ctx->clk_scaling; |
| spin_lock_init(&clk_scaling->lock); |
| clk_scaling->threshold_size = params->threshold_size; |
| clk_scaling->exception_size = params->exception_size; |
| INIT_WORK(&clk_scaling->work, clock_scaling_func); |
| |
| for (i = 0; i < params->threshold_size; i++) |
| clk_scaling->default_threshold[i] = |
| params->default_threshold[i]; |
| |
| /* Populate exception list*/ |
| for (i = 0; i < params->exception_size; i++) { |
| strlcpy(clk_scaling->exception_list[i].clients, |
| params->exceptions[i].usecase, IPA_PM_MAX_EX_CL); |
| IPA_PM_DBG("Usecase: %s\n", params->exceptions[i].usecase); |
| |
| /* Parse the commas to count the size of the clients */ |
| for (j = 0; j < IPA_PM_MAX_EX_CL && |
| clk_scaling->exception_list[i].clients[j]; j++) { |
| if (clk_scaling->exception_list[i].clients[j] == ',') |
| clk_scaling->exception_list[i].pending++; |
| } |
| |
| /* for the first client */ |
| clk_scaling->exception_list[i].pending++; |
| IPA_PM_DBG("Pending: %d\n", |
| clk_scaling->exception_list[i].pending); |
| |
| /* populate the threshold */ |
| for (j = 0; j < params->threshold_size; j++) { |
| clk_scaling->exception_list[i].threshold[j] |
| = params->exceptions[i].threshold[j]; |
| } |
| |
| } |
| IPA_PM_DBG("initialization success"); |
| |
| return 0; |
| } |
| |
| int ipa_pm_destroy(void) |
| { |
| IPA_PM_DBG("IPA PM destroy started\n"); |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("Already destroyed\n"); |
| return -EPERM; |
| } |
| |
| destroy_workqueue(ipa_pm_ctx->wq); |
| |
| kfree(ipa_pm_ctx); |
| ipa_pm_ctx = NULL; |
| |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_register() - register an IPA PM client with the PM |
| * @register_params: params for a client like throughput, callback, etc. |
| * @hdl: int pointer that will be used as an index to access the client |
| * |
| * Returns: 0 on success, negative on failure |
| * |
| * Side effects: *hdl is replaced with the client index or -EEXIST if |
| * client is already registered |
| */ |
| int ipa_pm_register(struct ipa_pm_register_params *params, u32 *hdl) |
| { |
| struct ipa_pm_client *client; |
| struct wakeup_source *wlock; |
| int elem; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (params == NULL || hdl == NULL || params->name == NULL) { |
| IPA_PM_ERR("Invalid Params\n"); |
| return -EINVAL; |
| } |
| |
| IPA_PM_DBG("IPA PM registering client\n"); |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| |
| elem = find_next_open_array_element(params->name); |
| *hdl = elem; |
| if (elem < 0 || elem > IPA_PM_MAX_CLIENTS) { |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| IPA_PM_ERR("client already registered or full array elem=%d\n", |
| elem); |
| return elem; |
| } |
| |
| ipa_pm_ctx->clients[*hdl] = kzalloc(sizeof |
| (struct ipa_pm_client), GFP_KERNEL); |
| if (!ipa_pm_ctx->clients[*hdl]) { |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| IPA_PM_ERR(":kzalloc err.\n"); |
| return -ENOMEM; |
| } |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| client = ipa_pm_ctx->clients[*hdl]; |
| |
| spin_lock_init(&client->state_lock); |
| |
| INIT_DELAYED_WORK(&client->deactivate_work, |
| delayed_deferred_deactivate_work_func); |
| |
| INIT_WORK(&client->activate_work, activate_work_func); |
| |
| /* populate fields */ |
| strlcpy(client->name, params->name, IPA_PM_MAX_EX_CL); |
| client->callback = params->callback; |
| client->callback_params = params->user_data; |
| client->group = params->group; |
| client->hdl = *hdl; |
| client->skip_clk_vote = params->skip_clk_vote; |
| wlock = &client->wlock; |
| wakeup_source_init(wlock, client->name); |
| |
| init_completion(&client->complete); |
| |
| /* add client to exception list */ |
| if (add_client_to_exception_list(*hdl)) { |
| ipa_pm_deregister(*hdl); |
| IPA_PM_ERR("Fail to add client to exception_list\n"); |
| return -EPERM; |
| } |
| |
| IPA_PM_DBG("IPA PM client registered with handle %d\n", *hdl); |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_deregister() - deregister IPA client from the PM |
| * @hdl: index of the client in the array |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_deregister(u32 hdl) |
| { |
| struct ipa_pm_client *client; |
| int i; |
| unsigned long flags; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (hdl >= IPA_PM_MAX_CLIENTS) { |
| IPA_PM_ERR("Invalid Param\n"); |
| return -EINVAL; |
| } |
| |
| if (ipa_pm_ctx->clients[hdl] == NULL) { |
| IPA_PM_ERR("Client is Null\n"); |
| return -EINVAL; |
| } |
| |
| IPA_PM_DBG("IPA PM deregistering client\n"); |
| |
| client = ipa_pm_ctx->clients[hdl]; |
| spin_lock_irqsave(&client->state_lock, flags); |
| if (IPA_PM_STATE_IN_PROGRESS(client->state)) { |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| wait_for_completion(&client->complete); |
| spin_lock_irqsave(&client->state_lock, flags); |
| } |
| |
| if (IPA_PM_STATE_ACTIVE(client->state)) { |
| IPA_PM_DBG("Activated clients cannot be deregistered"); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| return -EPERM; |
| } |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| |
| /* nullify pointers in pipe array */ |
| for (i = 0; i < IPA3_MAX_NUM_PIPES; i++) { |
| if (ipa_pm_ctx->clients_by_pipe[i] == ipa_pm_ctx->clients[hdl]) |
| ipa_pm_ctx->clients_by_pipe[i] = NULL; |
| } |
| wakeup_source_trash(&client->wlock); |
| kfree(client); |
| ipa_pm_ctx->clients[hdl] = NULL; |
| |
| remove_client_from_exception_list(hdl); |
| IPA_PM_DBG("IPA PM client %d deregistered\n", hdl); |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_associate_ipa_cons_to_client() - add mapping to pipe with ipa cllent |
| * @hdl: index of the client to be mapped |
| * @consumer: the pipe/consumer name to be pipped to the client |
| * |
| * Returns: 0 on success, negative on failure |
| * |
| * Side effects: multiple pipes are allowed to be mapped to a single client |
| */ |
| int ipa_pm_associate_ipa_cons_to_client(u32 hdl, enum ipa_client_type consumer) |
| { |
| int idx; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (hdl >= IPA_PM_MAX_CLIENTS || consumer < 0 || |
| consumer >= IPA_CLIENT_MAX) { |
| IPA_PM_ERR("invalid params\n"); |
| return -EINVAL; |
| } |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| if (ipa_pm_ctx->clients[hdl] == NULL) { |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| IPA_PM_ERR("Client is NULL\n"); |
| return -EPERM; |
| } |
| |
| idx = ipa_get_ep_mapping(consumer); |
| |
| if (idx < 0) { |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| IPA_PM_DBG("Pipe is not used\n"); |
| return 0; |
| } |
| |
| IPA_PM_DBG("Mapping pipe %d to client %d\n", idx, hdl); |
| |
| if (ipa_pm_ctx->clients_by_pipe[idx] != NULL) { |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| IPA_PM_ERR("Pipe is already mapped\n"); |
| return -EPERM; |
| } |
| ipa_pm_ctx->clients_by_pipe[idx] = ipa_pm_ctx->clients[hdl]; |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| IPA_PM_DBG("Pipe %d is mapped to client %d\n", idx, hdl); |
| |
| return 0; |
| } |
| |
| static int ipa_pm_activate_helper(struct ipa_pm_client *client, bool sync) |
| { |
| struct ipa_active_client_logging_info log_info; |
| int result = 0; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| |
| if (IPA_PM_STATE_IN_PROGRESS(client->state)) { |
| if (sync) { |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| wait_for_completion(&client->complete); |
| spin_lock_irqsave(&client->state_lock, flags); |
| } else { |
| client->state = IPA_PM_ACTIVATE_IN_PROGRESS; |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| return -EINPROGRESS; |
| } |
| } |
| |
| switch (client->state) { |
| case IPA_PM_ACTIVATED_PENDING_RESCHEDULE: |
| case IPA_PM_ACTIVATED_PENDING_DEACTIVATION: |
| client->state = IPA_PM_ACTIVATED_TIMER_SET; |
| case IPA_PM_ACTIVATED: |
| case IPA_PM_ACTIVATED_TIMER_SET: |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| return 0; |
| case IPA_PM_DEACTIVATED: |
| break; |
| default: |
| IPA_PM_ERR("Invalid State\n"); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| return -EPERM; |
| } |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| |
| IPA_ACTIVE_CLIENTS_PREP_SPECIAL(log_info, client->name); |
| if (!client->skip_clk_vote) { |
| if (sync) { |
| client->state = IPA_PM_ACTIVATE_IN_PROGRESS; |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| IPA_ACTIVE_CLIENTS_INC_SPECIAL(client->name); |
| spin_lock_irqsave(&client->state_lock, flags); |
| } else |
| result = ipa3_inc_client_enable_clks_no_block |
| (&log_info); |
| } |
| |
| /* we got the clocks */ |
| if (result == 0) { |
| client->state = IPA_PM_ACTIVATED; |
| if (client->group == IPA_PM_GROUP_APPS) |
| __pm_stay_awake(&client->wlock); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| activate_client(client->hdl); |
| if (sync) |
| do_clk_scaling(); |
| else |
| queue_work(ipa_pm_ctx->wq, |
| &ipa_pm_ctx->clk_scaling.work); |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| return 0; |
| } |
| |
| client->state = IPA_PM_ACTIVATE_IN_PROGRESS; |
| reinit_completion(&client->complete); |
| queue_work(ipa_pm_ctx->wq, &client->activate_work); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| return -EINPROGRESS; |
| } |
| |
| /** |
| * ipa_pm_activate(): activate ipa client to vote for clock(). Can be called |
| * from atomic context and returns -EINPROGRESS if cannot be done synchronously |
| * @hdl: index of the client in the array |
| * |
| * Returns: 0 on success, -EINPROGRESS if operation cannot be done synchronously |
| * and other negatives on failure |
| */ |
| int ipa_pm_activate(u32 hdl) |
| { |
| if (unlikely(ipa_pm_ctx == NULL)) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (unlikely(hdl >= IPA_PM_MAX_CLIENTS || |
| ipa_pm_ctx->clients[hdl] == NULL)) { |
| IPA_PM_ERR("Invalid Param\n"); |
| return -EINVAL; |
| } |
| |
| return ipa_pm_activate_helper(ipa_pm_ctx->clients[hdl], false); |
| } |
| |
| /** |
| * ipa_pm_activate(): activate ipa client to vote for clock synchronously. |
| * Cannot be called from an atomic contex. |
| * @hdl: index of the client in the array |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_activate_sync(u32 hdl) |
| { |
| if (unlikely(ipa_pm_ctx == NULL)) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (unlikely(hdl >= IPA_PM_MAX_CLIENTS || |
| ipa_pm_ctx->clients[hdl] == NULL)) { |
| IPA_PM_ERR("Invalid Param\n"); |
| return -EINVAL; |
| } |
| |
| return ipa_pm_activate_helper(ipa_pm_ctx->clients[hdl], true); |
| } |
| |
| /** |
| * ipa_pm_deferred_deactivate(): schedule a timer to deactivate client and |
| * devote clock. Can be called from atomic context (asynchronously) |
| * @hdl: index of the client in the array |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_deferred_deactivate(u32 hdl) |
| { |
| struct ipa_pm_client *client; |
| unsigned long flags; |
| unsigned long delay; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (hdl >= IPA_PM_MAX_CLIENTS || ipa_pm_ctx->clients[hdl] == NULL) { |
| IPA_PM_ERR("Invalid Param\n"); |
| return -EINVAL; |
| } |
| |
| client = ipa_pm_ctx->clients[hdl]; |
| IPA_PM_DBG_STATE(hdl, client->name, client->state); |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| switch (client->state) { |
| case IPA_PM_ACTIVATE_IN_PROGRESS: |
| client->state = IPA_PM_DEACTIVATE_IN_PROGRESS; |
| case IPA_PM_DEACTIVATED: |
| IPA_PM_DBG_STATE(hdl, client->name, client->state); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| return 0; |
| case IPA_PM_ACTIVATED: |
| delay = IPA_PM_DEFERRED_TIMEOUT; |
| if (ipa3_ctx->ipa3_hw_mode == IPA_HW_MODE_VIRTUAL || |
| ipa3_ctx->ipa3_hw_mode == IPA_HW_MODE_EMULATION) |
| delay *= 5; |
| |
| client->state = IPA_PM_ACTIVATED_PENDING_DEACTIVATION; |
| queue_delayed_work(ipa_pm_ctx->wq, &client->deactivate_work, |
| msecs_to_jiffies(delay)); |
| break; |
| case IPA_PM_ACTIVATED_TIMER_SET: |
| case IPA_PM_ACTIVATED_PENDING_DEACTIVATION: |
| client->state = IPA_PM_ACTIVATED_PENDING_RESCHEDULE; |
| case IPA_PM_DEACTIVATE_IN_PROGRESS: |
| case IPA_PM_ACTIVATED_PENDING_RESCHEDULE: |
| break; |
| case IPA_PM_STATE_MAX: |
| default: |
| IPA_PM_ERR("Bad State"); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| return -EINVAL; |
| } |
| IPA_PM_DBG_STATE(hdl, client->name, client->state); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_deactivate_all_deferred(): Cancel the deferred deactivation timer and |
| * immediately devotes for IPA clocks |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_deactivate_all_deferred(void) |
| { |
| int i; |
| bool run_algorithm = false; |
| struct ipa_pm_client *client; |
| unsigned long flags; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| for (i = 1; i < IPA_PM_MAX_CLIENTS; i++) { |
| client = ipa_pm_ctx->clients[i]; |
| |
| if (client == NULL) |
| continue; |
| |
| cancel_delayed_work_sync(&client->deactivate_work); |
| |
| if (IPA_PM_STATE_IN_PROGRESS(client->state)) { |
| wait_for_completion(&client->complete); |
| continue; |
| } |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| IPA_PM_DBG_STATE(client->hdl, client->name, client->state); |
| |
| if (client->state == IPA_PM_ACTIVATED_TIMER_SET) { |
| client->state = IPA_PM_ACTIVATED; |
| IPA_PM_DBG_STATE(client->hdl, client->name, |
| client->state); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| } else if (client->state == |
| IPA_PM_ACTIVATED_PENDING_DEACTIVATION || |
| client->state == |
| IPA_PM_ACTIVATED_PENDING_RESCHEDULE) { |
| run_algorithm = true; |
| client->state = IPA_PM_DEACTIVATED; |
| IPA_PM_DBG_STATE(client->hdl, client->name, |
| client->state); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| if (!client->skip_clk_vote) { |
| IPA_ACTIVE_CLIENTS_DEC_SPECIAL(client->name); |
| if (client->group == IPA_PM_GROUP_APPS) |
| __pm_relax(&client->wlock); |
| } |
| deactivate_client(client->hdl); |
| } else /* if activated or deactivated, we do nothing */ |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| } |
| |
| if (run_algorithm) |
| do_clk_scaling(); |
| |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_deactivate_sync(): deactivate ipa client and devote clock. Cannot be |
| * called from atomic context. |
| * @hdl: index of the client in the array |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_deactivate_sync(u32 hdl) |
| { |
| struct ipa_pm_client *client; |
| unsigned long flags; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (hdl >= IPA_PM_MAX_CLIENTS || ipa_pm_ctx->clients[hdl] == NULL) { |
| IPA_PM_ERR("Invalid Param\n"); |
| return -EINVAL; |
| } |
| client = ipa_pm_ctx->clients[hdl]; |
| |
| cancel_delayed_work_sync(&client->deactivate_work); |
| |
| if (IPA_PM_STATE_IN_PROGRESS(client->state)) |
| wait_for_completion(&client->complete); |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| IPA_PM_DBG_STATE(hdl, client->name, client->state); |
| |
| if (client->state == IPA_PM_DEACTIVATED) { |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| return 0; |
| } |
| |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| |
| /* else case (Deactivates all Activated cases)*/ |
| if (!client->skip_clk_vote) { |
| IPA_ACTIVE_CLIENTS_DEC_SPECIAL(client->name); |
| if (client->group == IPA_PM_GROUP_APPS) |
| __pm_relax(&client->wlock); |
| } |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| client->state = IPA_PM_DEACTIVATED; |
| IPA_PM_DBG_STATE(hdl, client->name, client->state); |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| deactivate_client(hdl); |
| do_clk_scaling(); |
| |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_handle_suspend(): calls the callbacks of suspended clients to wake up |
| * @pipe_bitmask: the bits represent the indexes of the clients to be woken up |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_handle_suspend(u32 pipe_bitmask) |
| { |
| int i; |
| struct ipa_pm_client *client; |
| bool client_notified[IPA_PM_MAX_CLIENTS] = { false }; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| IPA_PM_DBG_LOW("bitmask: %d", pipe_bitmask); |
| |
| if (pipe_bitmask == 0) |
| return 0; |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| for (i = 0; i < IPA3_MAX_NUM_PIPES; i++) { |
| if (pipe_bitmask & (1 << i)) { |
| client = ipa_pm_ctx->clients_by_pipe[i]; |
| if (client && !client_notified[client->hdl]) { |
| if (client->callback) { |
| client->callback(client->callback_params |
| , IPA_PM_REQUEST_WAKEUP); |
| client_notified[client->hdl] = true; |
| } else { |
| IPA_PM_ERR("client has no callback"); |
| WARN_ON(1); |
| } |
| } |
| } |
| } |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| return 0; |
| } |
| |
| /** |
| * ipa_pm_set_throughput(): Adds/changes the throughput requirement to IPA PM |
| * to be used for clock scaling |
| * @hdl: index of the client in the array |
| * @throughput: the new throughput value to be set for that client |
| * |
| * Returns: 0 on success, negative on failure |
| */ |
| int ipa_pm_set_throughput(u32 hdl, int throughput) |
| { |
| struct ipa_pm_client *client; |
| unsigned long flags; |
| |
| if (ipa_pm_ctx == NULL) { |
| IPA_PM_ERR("PM_ctx is null\n"); |
| return -EINVAL; |
| } |
| |
| if (hdl >= IPA_PM_MAX_CLIENTS || ipa_pm_ctx->clients[hdl] == NULL |
| || throughput < 0) { |
| IPA_PM_ERR("Invalid Params\n"); |
| return -EINVAL; |
| } |
| client = ipa_pm_ctx->clients[hdl]; |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| if (client->group == IPA_PM_GROUP_DEFAULT) |
| IPA_PM_DBG_LOW("Old throughput: %d\n", client->throughput); |
| else |
| IPA_PM_DBG_LOW("old Group %d throughput: %d\n", |
| client->group, ipa_pm_ctx->group_tput[client->group]); |
| |
| if (client->group == IPA_PM_GROUP_DEFAULT) |
| client->throughput = throughput; |
| else |
| ipa_pm_ctx->group_tput[client->group] = throughput; |
| |
| if (client->group == IPA_PM_GROUP_DEFAULT) |
| IPA_PM_DBG_LOW("New throughput: %d\n", client->throughput); |
| else |
| IPA_PM_DBG_LOW("New Group %d throughput: %d\n", |
| client->group, ipa_pm_ctx->group_tput[client->group]); |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| if (IPA_PM_STATE_ACTIVE(client->state) || (client->group != |
| IPA_PM_GROUP_DEFAULT)) { |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| do_clk_scaling(); |
| return 0; |
| } |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| |
| return 0; |
| } |
| |
| void ipa_pm_set_clock_index(int index) |
| { |
| if (ipa_pm_ctx && index >= 0) |
| ipa_pm_ctx->clk_scaling.cur_vote = index; |
| |
| IPA_PM_DBG("Setting pm clock vote to %d\n", index); |
| } |
| |
| /** |
| * ipa_pm_stat() - print PM stat |
| * @buf: [in] The user buff used to print |
| * @size: [in] The size of buf |
| * Returns: number of bytes used on success, negative on failure |
| * |
| * This function is called by ipa_debugfs in order to receive |
| * a picture of the clients in the PM and the throughput, threshold and cur vote |
| */ |
| int ipa_pm_stat(char *buf, int size) |
| { |
| struct ipa_pm_client *client; |
| struct clk_scaling_db *clk = &ipa_pm_ctx->clk_scaling; |
| int i, j, tput, cnt = 0, result = 0; |
| unsigned long flags; |
| |
| if (!buf || size < 0) |
| return -EINVAL; |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| |
| result = scnprintf(buf + cnt, size - cnt, "\n\nCurrent threshold: ["); |
| cnt += result; |
| |
| for (i = 0; i < clk->threshold_size; i++) { |
| result = scnprintf(buf + cnt, size - cnt, |
| "%d, ", clk->current_threshold[i]); |
| cnt += result; |
| } |
| |
| result = scnprintf(buf + cnt, size - cnt, "\b\b]\n"); |
| cnt += result; |
| |
| result = scnprintf(buf + cnt, size - cnt, |
| "Aggregated tput: %d, Cur vote: %d", |
| ipa_pm_ctx->aggregated_tput, clk->cur_vote); |
| cnt += result; |
| |
| result = scnprintf(buf + cnt, size - cnt, "\n\nRegistered Clients:\n"); |
| cnt += result; |
| |
| |
| for (i = 1; i < IPA_PM_MAX_CLIENTS; i++) { |
| client = ipa_pm_ctx->clients[i]; |
| |
| if (client == NULL) |
| continue; |
| |
| spin_lock_irqsave(&client->state_lock, flags); |
| if (client->group == IPA_PM_GROUP_DEFAULT) |
| tput = client->throughput; |
| else |
| tput = ipa_pm_ctx->group_tput[client->group]; |
| |
| result = scnprintf(buf + cnt, size - cnt, |
| "Client[%d]: %s State:%s\nGroup: %s Throughput: %d Pipes: ", |
| i, client->name, client_state_to_str[client->state], |
| ipa_pm_group_to_str[client->group], tput); |
| cnt += result; |
| |
| for (j = 0; j < IPA3_MAX_NUM_PIPES; j++) { |
| if (ipa_pm_ctx->clients_by_pipe[j] == client) { |
| result = scnprintf(buf + cnt, size - cnt, |
| "%d, ", j); |
| cnt += result; |
| } |
| } |
| |
| result = scnprintf(buf + cnt, size - cnt, "\b\b\n\n"); |
| cnt += result; |
| spin_unlock_irqrestore(&client->state_lock, flags); |
| } |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| return cnt; |
| } |
| |
| /** |
| * ipa_pm_exceptions_stat() - print PM exceptions stat |
| * @buf: [in] The user buff used to print |
| * @size: [in] The size of buf |
| * Returns: number of bytes used on success, negative on failure |
| * |
| * This function is called by ipa_debugfs in order to receive |
| * a full picture of the exceptions in the PM |
| */ |
| int ipa_pm_exceptions_stat(char *buf, int size) |
| { |
| int i, j, cnt = 0, result = 0; |
| struct ipa_pm_exception_list *exception; |
| |
| if (!buf || size < 0) |
| return -EINVAL; |
| |
| result = scnprintf(buf + cnt, size - cnt, "\n"); |
| cnt += result; |
| |
| mutex_lock(&ipa_pm_ctx->client_mutex); |
| for (i = 0; i < ipa_pm_ctx->clk_scaling.exception_size; i++) { |
| exception = &ipa_pm_ctx->clk_scaling.exception_list[i]; |
| if (exception == NULL) { |
| result = scnprintf(buf + cnt, size - cnt, |
| "Exception %d is NULL\n\n", i); |
| cnt += result; |
| continue; |
| } |
| |
| result = scnprintf(buf + cnt, size - cnt, |
| "Exception %d: %s\nPending: %d Bitmask: %d Threshold: [" |
| , i, exception->clients, exception->pending, |
| exception->bitmask); |
| cnt += result; |
| for (j = 0; j < ipa_pm_ctx->clk_scaling.threshold_size; j++) { |
| result = scnprintf(buf + cnt, size - cnt, |
| "%d, ", exception->threshold[j]); |
| cnt += result; |
| } |
| result = scnprintf(buf + cnt, size - cnt, "\b\b]\n\n"); |
| cnt += result; |
| } |
| mutex_unlock(&ipa_pm_ctx->client_mutex); |
| |
| return cnt; |
| } |