qcacmn: Fix compilation issues HTC/HIF
Resolve compilation issues for references to cdf after rebase
Change-Id: I332566deb690fa7de68d89bf51b42f94f7a270a6
CRs-Fixed: 981187
diff --git a/hif/src/ce/ce_service.c b/hif/src/ce/ce_service.c
index 9f90d5b..3c34575 100644
--- a/hif/src/ce/ce_service.c
+++ b/hif/src/ce/ce_service.c
@@ -35,7 +35,7 @@
#include "ce_main.h"
#include "ce_internal.h"
#include "ce_reg.h"
-#include "cdf_lock.h"
+#include "qdf_lock.h"
#include "regtable.h"
#include "epping_main.h"
#include "hif_main.h"
@@ -97,7 +97,7 @@
/* max history to record per copy engine */
#define HIF_CE_HISTORY_MAX 512
-cdf_atomic_t hif_ce_desc_history_index[CE_COUNT_MAX];
+qdf_atomic_t hif_ce_desc_history_index[CE_COUNT_MAX];
struct hif_ce_desc_event hif_ce_desc_history[CE_COUNT_MAX][HIF_CE_HISTORY_MAX];
@@ -113,11 +113,11 @@
* trying to access the array, full locking of the recording process would
* be needed to have sane logging.
*/
-static int get_next_record_index(cdf_atomic_t *table_index, int array_size)
+static int get_next_record_index(qdf_atomic_t *table_index, int array_size)
{
- int record_index = cdf_atomic_inc_return(table_index);
+ int record_index = qdf_atomic_inc_return(table_index);
if (record_index == array_size)
- cdf_atomic_sub(array_size, table_index);
+ qdf_atomic_sub(array_size, table_index);
while (record_index >= array_size)
record_index -= array_size;
@@ -149,8 +149,8 @@
if (cbk && cbk->get_monotonic_boottime)
event->time = cbk->get_monotonic_boottime();
else
- event->time = ((uint64_t)cdf_system_ticks_to_msecs(
- cdf_system_ticks()) * 1000);
+ event->time = ((uint64_t)qdf_system_ticks_to_msecs(
+ qdf_system_ticks()) * 1000);
if (descriptor != NULL)
event->descriptor = *descriptor;
@@ -169,7 +169,7 @@
*/
void ce_init_ce_desc_event_log(int ce_id, int size)
{
- cdf_atomic_init(&hif_ce_desc_history_index[ce_id]);
+ qdf_atomic_init(&hif_ce_desc_history_index[ce_id]);
}
#else
void hif_record_ce_desc_event(struct hif_softc *scn,
@@ -232,7 +232,7 @@
ce_completed_send_next_nolock(struct CE_state *CE_state,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *sw_idx, unsigned int *hw_idx,
@@ -275,7 +275,7 @@
int
ce_send_nolock(struct CE_handle *copyeng,
void *per_transfer_context,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
uint32_t nbytes,
uint32_t transfer_id,
uint32_t flags,
@@ -295,7 +295,7 @@
if (unlikely(CE_RING_DELTA(nentries_mask,
write_index, sw_index - 1) <= 0)) {
OL_ATH_CE_PKT_ERROR_COUNT_INCR(scn, CE_RING_DELTA_FAIL);
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
A_TARGET_ACCESS_END_RET(scn);
return status;
}
@@ -357,7 +357,7 @@
src_ring->write_index);
src_ring->write_index = write_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
}
A_TARGET_ACCESS_END_RET(scn);
@@ -367,7 +367,7 @@
int
ce_send(struct CE_handle *copyeng,
void *per_transfer_context,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
uint32_t nbytes,
uint32_t transfer_id,
uint32_t flags,
@@ -376,10 +376,10 @@
struct CE_state *CE_state = (struct CE_state *)copyeng;
int status;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
status = ce_send_nolock(copyeng, per_transfer_context, buffer, nbytes,
transfer_id, flags, user_flag);
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -397,7 +397,7 @@
int
ce_sendlist_buf_add(struct ce_sendlist *sendlist,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
uint32_t nbytes,
uint32_t flags,
uint32_t user_flags)
@@ -407,8 +407,8 @@
struct ce_sendlist_item *item;
if (num_items >= CE_SENDLIST_ITEMS_MAX) {
- CDF_ASSERT(num_items < CE_SENDLIST_ITEMS_MAX);
- return CDF_STATUS_E_RESOURCES;
+ QDF_ASSERT(num_items < CE_SENDLIST_ITEMS_MAX);
+ return QDF_STATUS_E_RESOURCES;
}
item = &sl->item[num_items];
@@ -418,7 +418,7 @@
item->flags = flags;
item->user_flags = user_flags;
sl->num_items = num_items + 1;
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
int
@@ -435,9 +435,9 @@
unsigned int sw_index;
unsigned int write_index;
- CDF_ASSERT((num_items > 0) && (num_items < src_ring->nentries));
+ QDF_ASSERT((num_items > 0) && (num_items < src_ring->nentries));
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
@@ -450,28 +450,28 @@
for (i = 0; i < num_items - 1; i++) {
item = &sl->item[i];
/* TBDXXX: Support extensible sendlist_types? */
- CDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
+ QDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
status = ce_send_nolock(copyeng, CE_SENDLIST_ITEM_CTXT,
- (cdf_dma_addr_t) item->data,
+ (qdf_dma_addr_t) item->data,
item->u.nbytes, transfer_id,
item->flags | CE_SEND_FLAG_GATHER,
item->user_flags);
- CDF_ASSERT(status == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(status == QDF_STATUS_SUCCESS);
}
/* provide valid context pointer for final item */
item = &sl->item[i];
/* TBDXXX: Support extensible sendlist_types? */
- CDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
+ QDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
status = ce_send_nolock(copyeng, per_transfer_context,
- (cdf_dma_addr_t) item->data,
+ (qdf_dma_addr_t) item->data,
item->u.nbytes,
transfer_id, item->flags,
item->user_flags);
- CDF_ASSERT(status == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(status == QDF_STATUS_SUCCESS);
NBUF_UPDATE_TX_PKT_COUNT((cdf_nbuf_t)per_transfer_context,
NBUF_TX_PKT_CE);
- DPTRACE(cdf_dp_trace((cdf_nbuf_t)per_transfer_context,
- CDF_DP_TRACE_CE_PACKET_PTR_RECORD,
+ DPTRACE(qdf_dp_trace((cdf_nbuf_t)per_transfer_context,
+ QDF_DP_TRACE_CE_PACKET_PTR_RECORD,
(uint8_t *)(((cdf_nbuf_t)per_transfer_context)->data),
sizeof(((cdf_nbuf_t)per_transfer_context)->data)));
} else {
@@ -482,7 +482,7 @@
* the entire request at once, punt it back to the caller.
*/
}
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -543,7 +543,7 @@
uint64_t dma_addr;
uint32_t user_flags = 0;
- cdf_spin_lock_bh(&ce_state->ce_index_lock);
+ qdf_spin_lock_bh(&ce_state->ce_index_lock);
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
@@ -606,7 +606,7 @@
/*
* Clear packet offset for all but the first CE desc.
*/
- user_flags &= ~CDF_CE_TX_PKT_OFFSET_BIT_M;
+ user_flags &= ~QDF_CE_TX_PKT_OFFSET_BIT_M;
ce_buffer_addr_hi_set(shadow_src_desc, dma_addr, user_flags);
shadow_src_desc->meta_data = transfer_id;
@@ -641,7 +641,7 @@
}
}
- cdf_spin_unlock_bh(&ce_state->ce_index_lock);
+ qdf_spin_unlock_bh(&ce_state->ce_index_lock);
/*
* If all packets in the array are transmitted,
@@ -655,7 +655,7 @@
int
ce_recv_buf_enqueue(struct CE_handle *copyeng,
- void *per_recv_context, cdf_dma_addr_t buffer)
+ void *per_recv_context, qdf_dma_addr_t buffer)
{
int status;
struct CE_state *CE_state = (struct CE_state *)copyeng;
@@ -668,13 +668,13 @@
uint64_t dma_addr = buffer;
struct hif_softc *scn = CE_state->scn;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
write_index = dest_ring->write_index;
sw_index = dest_ring->sw_index;
A_TARGET_ACCESS_BEGIN_RET_EXT(scn, val);
if (val == -1) {
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return val;
}
@@ -704,17 +704,17 @@
write_index = CE_RING_IDX_INCR(nentries_mask, write_index);
CE_DEST_RING_WRITE_IDX_SET(scn, ctrl_addr, write_index);
dest_ring->write_index = write_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
} else {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
}
A_TARGET_ACCESS_END_RET_EXT(scn, val);
if (val == -1) {
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return val;
}
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -755,10 +755,10 @@
unsigned int sw_index;
unsigned int write_index;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return CE_RING_DELTA(nentries_mask, write_index, sw_index - 1);
}
@@ -771,10 +771,10 @@
unsigned int sw_index;
unsigned int write_index;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
sw_index = dest_ring->sw_index;
write_index = dest_ring->write_index;
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return CE_RING_DELTA(nentries_mask, write_index, sw_index - 1);
}
@@ -804,9 +804,9 @@
struct CE_state *CE_state = (struct CE_state *)copyeng;
unsigned int nentries;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries = ce_send_entries_done_nolock(CE_state->scn, CE_state);
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return nentries;
}
@@ -836,9 +836,9 @@
struct CE_state *CE_state = (struct CE_state *)copyeng;
unsigned int nentries;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries = ce_recv_entries_done_nolock(CE_state->scn, CE_state);
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return nentries;
}
@@ -857,7 +857,7 @@
ce_completed_recv_next_nolock(struct CE_state *CE_state,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *flagsp)
@@ -886,7 +886,7 @@
* corresponding descriptor has completed. We treat this
* as a descriptor that is not yet done.
*/
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
@@ -916,7 +916,7 @@
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
dest_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
done:
return status;
@@ -926,45 +926,45 @@
ce_completed_recv_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp, unsigned int *flagsp)
{
struct CE_state *CE_state = (struct CE_state *)copyeng;
int status;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
status =
ce_completed_recv_next_nolock(CE_state, per_CE_contextp,
per_transfer_contextp, bufferp,
nbytesp, transfer_idp, flagsp);
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
/* NB: Modeled after ce_completed_recv_next_nolock */
-CDF_STATUS
+QDF_STATUS
ce_revoke_recv_next(struct CE_handle *copyeng,
void **per_CE_contextp,
- void **per_transfer_contextp, cdf_dma_addr_t *bufferp)
+ void **per_transfer_contextp, qdf_dma_addr_t *bufferp)
{
struct CE_state *CE_state;
struct CE_ring_state *dest_ring;
unsigned int nentries_mask;
unsigned int sw_index;
unsigned int write_index;
- CDF_STATUS status;
+ QDF_STATUS status;
struct hif_softc *scn;
CE_state = (struct CE_state *)copyeng;
dest_ring = CE_state->dest_ring;
if (!dest_ring) {
- return CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_E_FAILURE;
}
scn = CE_state->scn;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries_mask = dest_ring->nentries_mask;
sw_index = dest_ring->sw_index;
write_index = dest_ring->write_index;
@@ -992,11 +992,11 @@
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
dest_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
} else {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
}
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return status;
}
@@ -1009,14 +1009,14 @@
ce_completed_send_next_nolock(struct CE_state *CE_state,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *sw_idx,
unsigned int *hw_idx,
uint32_t *toeplitz_hash_result)
{
- int status = CDF_STATUS_E_FAILURE;
+ int status = QDF_STATUS_E_FAILURE;
struct CE_ring_state *src_ring = CE_state->src_ring;
uint32_t ctrl_addr = CE_state->ctrl_addr;
unsigned int nentries_mask = src_ring->nentries_mask;
@@ -1085,18 +1085,18 @@
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
src_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
}
return status;
}
/* NB: Modeled after ce_completed_send_next */
-CDF_STATUS
+QDF_STATUS
ce_cancel_send_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
uint32_t *toeplitz_hash_result)
@@ -1106,17 +1106,17 @@
unsigned int nentries_mask;
unsigned int sw_index;
unsigned int write_index;
- CDF_STATUS status;
+ QDF_STATUS status;
struct hif_softc *scn;
CE_state = (struct CE_state *)copyeng;
src_ring = CE_state->src_ring;
if (!src_ring) {
- return CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_E_FAILURE;
}
scn = CE_state->scn;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries_mask = src_ring->nentries_mask;
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
@@ -1151,11 +1151,11 @@
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
src_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
} else {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
}
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return status;
}
@@ -1167,7 +1167,7 @@
ce_completed_send_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *sw_idx,
@@ -1177,13 +1177,13 @@
struct CE_state *CE_state = (struct CE_state *)copyeng;
int status;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
status =
ce_completed_send_next_nolock(CE_state, per_CE_contextp,
per_transfer_contextp, bufferp,
nbytesp, transfer_idp, sw_idx,
hw_idx, toeplitz_hash_result);
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -1202,7 +1202,7 @@
{
void *CE_context;
void *transfer_context;
- cdf_dma_addr_t buf;
+ qdf_dma_addr_t buf;
unsigned int nbytes;
unsigned int id;
unsigned int sw_idx, hw_idx;
@@ -1226,7 +1226,7 @@
* addressed by change spin_lock to spin_lock_bh also.
*/
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
if (CE_state->send_cb) {
{
@@ -1238,9 +1238,9 @@
&transfer_context, &buf,
&nbytes, &id, &sw_idx, &hw_idx,
&toeplitz_hash_result) ==
- CDF_STATUS_SUCCESS) {
+ QDF_STATUS_SUCCESS) {
if (ce_id != CE_HTT_H2T_MSG) {
- cdf_spin_unlock_bh(
+ qdf_spin_unlock_bh(
&CE_state->ce_index_lock);
CE_state->send_cb(
(struct CE_handle *)
@@ -1248,24 +1248,24 @@
transfer_context, buf,
nbytes, id, sw_idx, hw_idx,
toeplitz_hash_result);
- cdf_spin_lock_bh(
+ qdf_spin_lock_bh(
&CE_state->ce_index_lock);
} else {
struct HIF_CE_pipe_info *pipe_info =
(struct HIF_CE_pipe_info *)
CE_context;
- cdf_spin_lock_bh(&pipe_info->
+ qdf_spin_lock_bh(&pipe_info->
completion_freeq_lock);
pipe_info->num_sends_allowed++;
- cdf_spin_unlock_bh(&pipe_info->
+ qdf_spin_unlock_bh(&pipe_info->
completion_freeq_lock);
}
}
}
}
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
hif_record_ce_desc_event(scn, ce_id, HIF_CE_REAP_EXIT,
NULL, NULL, 0);
@@ -1299,7 +1299,7 @@
uint32_t ctrl_addr = CE_state->ctrl_addr;
void *CE_context;
void *transfer_context;
- cdf_dma_addr_t buf;
+ qdf_dma_addr_t buf;
unsigned int nbytes;
unsigned int id;
unsigned int flags;
@@ -1315,7 +1315,7 @@
return 0; /* no work done */
}
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
/* Clear force_break flag and re-initialize receive_count to 0 */
@@ -1331,8 +1331,8 @@
while (ce_completed_recv_next_nolock
(CE_state, &CE_context, &transfer_context,
&buf, &nbytes, &id, &flags) ==
- CDF_STATUS_SUCCESS) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ QDF_STATUS_SUCCESS) {
+ qdf_spin_unlock(&CE_state->ce_index_lock);
CE_state->recv_cb((struct CE_handle *)CE_state,
CE_context, transfer_context, buf,
nbytes, id, flags);
@@ -1354,8 +1354,8 @@
/* Break the receive processes by
* force if force_break set up
*/
- if (cdf_unlikely(CE_state->force_break)) {
- cdf_atomic_set(&CE_state->rx_pending, 1);
+ if (qdf_unlikely(CE_state->force_break)) {
+ qdf_atomic_set(&CE_state->rx_pending, 1);
CE_ENGINE_INT_STATUS_CLEAR(scn, ctrl_addr,
HOST_IS_COPY_COMPLETE_MASK);
if (Q_TARGET_ACCESS_END(scn) < 0)
@@ -1363,7 +1363,7 @@
CE_state->receive_count);
return CE_state->receive_count;
}
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
}
}
@@ -1383,24 +1383,24 @@
(CE_state, &CE_context,
&transfer_context, &buf, &nbytes,
&id, &sw_idx, &hw_idx,
- &toeplitz_hash_result) == CDF_STATUS_SUCCESS) {
+ &toeplitz_hash_result) == QDF_STATUS_SUCCESS) {
if (CE_id != CE_HTT_H2T_MSG ||
WLAN_IS_EPPING_ENABLED(mode)) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
CE_state->send_cb((struct CE_handle *)CE_state,
CE_context, transfer_context,
buf, nbytes, id, sw_idx,
hw_idx, toeplitz_hash_result);
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
} else {
struct HIF_CE_pipe_info *pipe_info =
(struct HIF_CE_pipe_info *)CE_context;
- cdf_spin_lock(&pipe_info->
+ qdf_spin_lock(&pipe_info->
completion_freeq_lock);
pipe_info->num_sends_allowed++;
- cdf_spin_unlock(&pipe_info->
+ qdf_spin_unlock(&pipe_info->
completion_freeq_lock);
}
}
@@ -1409,13 +1409,13 @@
(CE_state, &CE_context,
&transfer_context, &buf, &nbytes,
&id, &sw_idx, &hw_idx,
- &toeplitz_hash_result) == CDF_STATUS_SUCCESS) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ &toeplitz_hash_result) == QDF_STATUS_SUCCESS) {
+ qdf_spin_unlock(&CE_state->ce_index_lock);
CE_state->send_cb((struct CE_handle *)CE_state,
CE_context, transfer_context, buf,
nbytes, id, sw_idx, hw_idx,
toeplitz_hash_result);
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
}
#endif /*ATH_11AC_TXCOMPACT */
}
@@ -1425,7 +1425,7 @@
CE_int_status = CE_ENGINE_INT_STATUS_GET(scn, ctrl_addr);
if (CE_int_status & CE_WATERMARK_MASK) {
if (CE_state->watermark_cb) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
/* Convert HW IS bits to software flags */
flags =
(CE_int_status & CE_WATERMARK_MASK) >>
@@ -1434,7 +1434,7 @@
CE_state->
watermark_cb((struct CE_handle *)CE_state,
CE_state->wm_context, flags);
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
}
}
}
@@ -1494,8 +1494,8 @@
}
}
- cdf_spin_unlock(&CE_state->ce_index_lock);
- cdf_atomic_set(&CE_state->rx_pending, 0);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_atomic_set(&CE_state->rx_pending, 0);
if (Q_TARGET_ACCESS_END(scn) < 0)
HIF_ERROR("<--[premature rc=%d]\n", CE_state->receive_count);
@@ -1514,11 +1514,11 @@
uint32_t intr_summary;
A_TARGET_ACCESS_BEGIN(scn);
- if (!cdf_atomic_read(&scn->tasklet_from_intr)) {
+ if (!qdf_atomic_read(&scn->tasklet_from_intr)) {
for (CE_id = 0; CE_id < scn->ce_count; CE_id++) {
struct CE_state *CE_state = scn->ce_id_to_state[CE_id];
- if (cdf_atomic_read(&CE_state->rx_pending)) {
- cdf_atomic_set(&CE_state->rx_pending, 0);
+ if (qdf_atomic_read(&CE_state->rx_pending)) {
+ qdf_atomic_set(&CE_state->rx_pending, 0);
ce_per_engine_service(scn, CE_id);
}
}
@@ -1718,11 +1718,11 @@
struct hif_softc *sc = (struct hif_softc *)(hif_sc);
struct CE_state *ce_state = sc->ce_id_to_state[CE_HTT_H2T_MSG];
- cdf_assert_always(ce_state);
+ qdf_assert_always(ce_state);
ce_state->download_len = pkt_download_len;
- cdf_print("%s CE %d Pkt download length %d", __func__,
+ qdf_print("%s CE %d Pkt download length %d", __func__,
ce_state->id, ce_state->download_len);
}
#else
@@ -1737,7 +1737,7 @@
for (CE_id = 0; CE_id < scn->ce_count; CE_id++) {
struct CE_state *CE_state = scn->ce_id_to_state[CE_id];
- if (cdf_atomic_read(&CE_state->rx_pending))
+ if (qdf_atomic_read(&CE_state->rx_pending))
return true;
}
@@ -1754,7 +1754,7 @@
bool ce_check_rx_pending(struct hif_softc *scn, int ce_id)
{
struct CE_state *CE_state = scn->ce_id_to_state[ce_id];
- if (cdf_atomic_read(&CE_state->rx_pending))
+ if (qdf_atomic_read(&CE_state->rx_pending))
return true;
else
return false;
@@ -1813,14 +1813,14 @@
* Return: None
*/
void ce_ipa_get_resource(struct CE_handle *ce,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr)
+ qdf_dma_addr_t *ce_reg_paddr)
{
struct CE_state *CE_state = (struct CE_state *)ce;
uint32_t ring_loop;
struct CE_src_desc *ce_desc;
- cdf_dma_addr_t phy_mem_base;
+ qdf_dma_addr_t phy_mem_base;
struct hif_softc *scn = CE_state->scn;
if (CE_RUNNING != CE_state->state) {