summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorChouhan, Anurag <achouhan@codeaurora.org>2016-03-03 19:05:05 +0530
committerGerrit - the friendly Code Review server <code-review@localhost>2016-03-16 12:18:31 -0700
commitfc06aa94302bf143e998d4e89abd0132d8e6d423 (patch)
tree67c8609aa46f10c32d5e6be91cfa6d995ce54c17
parent5776318d1934b14c3b4eb4eb98c74830d64d00a7 (diff)
qcacmn: Fix compilation issues HTC/HIF
Resolve compilation issues for references to cdf after rebase Change-Id: I332566deb690fa7de68d89bf51b42f94f7a270a6 CRs-Fixed: 981187
-rw-r--r--hif/inc/hif.h44
-rw-r--r--hif/inc/hif_napi.h2
-rw-r--r--hif/src/ath_procfs.c12
-rw-r--r--hif/src/ce/ce_api.h30
-rw-r--r--hif/src/ce/ce_bmi.c74
-rw-r--r--hif/src/ce/ce_bmi.h8
-rw-r--r--hif/src/ce/ce_diag.c129
-rw-r--r--hif/src/ce/ce_internal.h12
-rw-r--r--hif/src/ce/ce_main.c330
-rw-r--r--hif/src/ce/ce_main.h14
-rw-r--r--hif/src/ce/ce_service.c226
-rw-r--r--hif/src/ce/ce_tasklet.c32
-rw-r--r--hif/src/ce/ce_tasklet.h4
-rw-r--r--hif/src/hif_debug.h18
-rw-r--r--hif/src/hif_main.c77
-rw-r--r--hif/src/hif_main.h20
-rw-r--r--hif/src/hif_napi.c4
-rw-r--r--hif/src/icnss_stub/icnss_stub.c4
-rw-r--r--hif/src/mp_dev.c52
-rw-r--r--hif/src/pcie/hif_io32_pci.h8
-rw-r--r--hif/src/pcie/if_pci.c172
-rw-r--r--hif/src/pcie/if_pci.h6
-rw-r--r--hif/src/snoc/hif_io32_snoc.h2
-rw-r--r--hif/src/snoc/if_snoc.c16
-rw-r--r--htc/htc.c68
-rw-r--r--htc/htc_api.h247
-rw-r--r--htc/htc_debug.h10
-rw-r--r--htc/htc_internal.h72
-rw-r--r--htc/htc_packet.h78
-rw-r--r--htc/htc_recv.c36
-rw-r--r--htc/htc_send.c92
-rw-r--r--htc/htc_services.c10
32 files changed, 965 insertions, 944 deletions
diff --git a/hif/inc/hif.h b/hif/inc/hif.h
index 169ce54251cf..f9e96dbeb70a 100644
--- a/hif/inc/hif.h
+++ b/hif/inc/hif.h
@@ -36,8 +36,8 @@ extern "C" {
#include "athdefs.h"
#include "a_types.h"
#include "osapi_linux.h"
-#include "cdf_status.h"
-#include "cdf_nbuf.h"
+#include <qdf_status.h>
+#include "qdf_nbuf.h"
#include "ol_if_athvar.h"
#include <linux/platform_device.h>
#ifdef HIF_PCI
@@ -259,8 +259,8 @@ void hif_detach_htc(struct hif_opaque_softc *scn);
/*
* API to handle HIF-specific BMI message exchanges, this API is synchronous
* and only allowed to be called from a context that can block (sleep) */
-CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *scn,
- cdf_dma_addr_t cmd, cdf_dma_addr_t rsp,
+QDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *scn,
+ qdf_dma_addr_t cmd, qdf_dma_addr_t rsp,
uint8_t *pSendMessage, uint32_t Length,
uint8_t *pResponseMessage,
uint32_t *pResponseLength, uint32_t TimeoutMS);
@@ -275,9 +275,9 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *scn,
*
* hif_diag_read_mem reads an arbitrary length of arbitrarily aligned memory.
*/
-CDF_STATUS hif_diag_read_access(struct hif_opaque_softc *scn, uint32_t address,
+QDF_STATUS hif_diag_read_access(struct hif_opaque_softc *scn, uint32_t address,
uint32_t *data);
-CDF_STATUS hif_diag_read_mem(struct hif_opaque_softc *scn, uint32_t address,
+QDF_STATUS hif_diag_read_mem(struct hif_opaque_softc *scn, uint32_t address,
uint8_t *data, int nbytes);
void hif_dump_target_memory(struct hif_opaque_softc *scn, void *ramdump_base,
uint32_t address, uint32_t size);
@@ -292,9 +292,9 @@ void hif_dump_target_memory(struct hif_opaque_softc *scn, void *ramdump_base,
*
* hif_diag_write_mem writes an arbitrary length of arbitrarily aligned memory.
*/
-CDF_STATUS hif_diag_write_access(struct hif_opaque_softc *scn, uint32_t address,
+QDF_STATUS hif_diag_write_access(struct hif_opaque_softc *scn, uint32_t address,
uint32_t data);
-CDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *scn, uint32_t address,
+QDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *scn, uint32_t address,
uint8_t *data, int nbytes);
/*
@@ -379,9 +379,9 @@ void *hif_get_ce_handle(struct hif_opaque_softc *hif_ctx, int);
#ifdef IPA_OFFLOAD
void hif_ipa_get_ce_resource(struct hif_opaque_softc *scn,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr);
+ qdf_dma_addr_t *ce_reg_paddr);
#else
/**
* hif_ipa_get_ce_resource() - get uc resource on hif
@@ -397,9 +397,9 @@ void hif_ipa_get_ce_resource(struct hif_opaque_softc *scn,
* Return: None
*/
static inline void hif_ipa_get_ce_resource(struct hif_opaque_softc *scn,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr)
+ qdf_dma_addr_t *ce_reg_paddr)
{
return;
}
@@ -411,13 +411,13 @@ static inline void hif_ipa_get_ce_resource(struct hif_opaque_softc *scn,
struct hif_msg_callbacks {
void *Context;
/**< context meaningful to HTC */
- CDF_STATUS (*txCompletionHandler)(void *Context, cdf_nbuf_t wbuf,
+ QDF_STATUS (*txCompletionHandler)(void *Context, qdf_nbuf_t wbuf,
uint32_t transferID,
uint32_t toeplitz_hash_result);
- CDF_STATUS (*rxCompletionHandler)(void *Context, cdf_nbuf_t wbuf,
+ QDF_STATUS (*rxCompletionHandler)(void *Context, qdf_nbuf_t wbuf,
uint8_t pipeID);
void (*txResourceAvailHandler)(void *context, uint8_t pipe);
- void (*fwEventHandler)(void *context, CDF_STATUS status);
+ void (*fwEventHandler)(void *context, QDF_STATUS status);
};
#define HIF_DATA_ATTR_SET_TX_CLASSIFY(attr, v) \
@@ -443,13 +443,13 @@ typedef struct device hif_bus_id;
void hif_post_init(struct hif_opaque_softc *scn, void *hHTC,
struct hif_msg_callbacks *callbacks);
-CDF_STATUS hif_start(struct hif_opaque_softc *scn);
+QDF_STATUS hif_start(struct hif_opaque_softc *scn);
void hif_stop(struct hif_opaque_softc *scn);
void hif_flush_surprise_remove(struct hif_opaque_softc *scn);
void hif_dump(struct hif_opaque_softc *scn, uint8_t CmdId, bool start);
-CDF_STATUS hif_send_head(struct hif_opaque_softc *scn, uint8_t PipeID,
+QDF_STATUS hif_send_head(struct hif_opaque_softc *scn, uint8_t PipeID,
uint32_t transferID, uint32_t nbytes,
- cdf_nbuf_t wbuf, uint32_t data_attr);
+ qdf_nbuf_t wbuf, uint32_t data_attr);
void hif_send_complete_check(struct hif_opaque_softc *scn, uint8_t PipeID,
int force);
void hif_cancel_deferred_target_sleep(struct hif_opaque_softc *scn);
@@ -473,13 +473,13 @@ void hif_reset_soc(struct hif_opaque_softc *scn);
void hif_disable_aspm(struct hif_opaque_softc *);
void hif_save_htc_htt_config_endpoint(struct hif_opaque_softc *hif_ctx,
int htc_endpoint);
-struct hif_opaque_softc *hif_open(cdf_device_t cdf_ctx, uint32_t mode,
- enum ath_hal_bus_type bus_type,
+struct hif_opaque_softc *hif_open(qdf_device_t qdf_ctx, uint32_t mode,
+ enum qdf_bus_type bus_type,
struct hif_callbacks *cbk);
void hif_close(struct hif_opaque_softc *hif_ctx);
-CDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev,
+QDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev,
void *bdev, const hif_bus_id *bid,
- enum ath_hal_bus_type bus_type,
+ enum qdf_bus_type bus_type,
enum hif_enable_type type);
void hif_disable(struct hif_opaque_softc *hif_ctx, enum hif_disable_type type);
void hif_enable_power_gating(struct hif_opaque_softc *hif_ctx);
diff --git a/hif/inc/hif_napi.h b/hif/inc/hif_napi.h
index 40927fcd03b9..f1f07b2258ed 100644
--- a/hif/inc/hif_napi.h
+++ b/hif/inc/hif_napi.h
@@ -99,7 +99,7 @@ int hif_napi_poll(struct hif_opaque_softc *hif_ctx,
#ifdef FEATURE_NAPI_DEBUG
#define NAPI_DEBUG(fmt, ...) \
- cdf_print("wlan: NAPI: %s:%d "fmt, __func__, __LINE__, ##__VA_ARGS__);
+ qdf_print("wlan: NAPI: %s:%d "fmt, __func__, __LINE__, ##__VA_ARGS__);
#else
#define NAPI_DEBUG(fmt, ...) /* NO-OP */
#endif /* FEATURE NAPI_DEBUG */
diff --git a/hif/src/ath_procfs.c b/hif/src/ath_procfs.c
index 3837af6c8479..af31864c3a0f 100644
--- a/hif/src/ath_procfs.c
+++ b/hif/src/ath_procfs.c
@@ -66,7 +66,7 @@ static ssize_t ath_procfs_diag_read(struct file *file, char __user *buf,
int rv;
uint8_t *read_buffer = NULL;
- read_buffer = cdf_mem_malloc(count);
+ read_buffer = qdf_mem_malloc(count);
if (NULL == read_buffer) {
HIF_ERROR("%s: cdf_mem_alloc failed", __func__);
return -ENOMEM;
@@ -86,12 +86,12 @@ static ssize_t ath_procfs_diag_read(struct file *file, char __user *buf,
}
if (copy_to_user(buf, read_buffer, count)) {
- cdf_mem_free(read_buffer);
+ qdf_mem_free(read_buffer);
HIF_ERROR("%s: copy_to_user error in /proc/%s",
__func__, PROCFS_NAME);
return -EFAULT;
} else
- cdf_mem_free(read_buffer);
+ qdf_mem_free(read_buffer);
if (rv == 0) {
return count;
@@ -108,13 +108,13 @@ static ssize_t ath_procfs_diag_write(struct file *file,
int rv;
uint8_t *write_buffer = NULL;
- write_buffer = cdf_mem_malloc(count);
+ write_buffer = qdf_mem_malloc(count);
if (NULL == write_buffer) {
HIF_ERROR("%s: cdf_mem_alloc failed", __func__);
return -ENOMEM;
}
if (copy_from_user(write_buffer, buf, count)) {
- cdf_mem_free(write_buffer);
+ qdf_mem_free(write_buffer);
HIF_ERROR("%s: copy_to_user error in /proc/%s",
__func__, PROCFS_NAME);
return -EFAULT;
@@ -134,7 +134,7 @@ static ssize_t ath_procfs_diag_write(struct file *file,
(uint8_t *)write_buffer, count);
}
- cdf_mem_free(write_buffer);
+ qdf_mem_free(write_buffer);
if (rv == 0) {
return count;
} else {
diff --git a/hif/src/ce/ce_api.h b/hif/src/ce/ce_api.h
index 501828359cad..2dd2de0cc604 100644
--- a/hif/src/ce/ce_api.h
+++ b/hif/src/ce/ce_api.h
@@ -73,7 +73,7 @@ struct CE_handle;
typedef void (*ce_send_cb)(struct CE_handle *copyeng,
void *per_ce_send_context,
void *per_transfer_send_context,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
unsigned int nbytes,
unsigned int transfer_id,
unsigned int sw_index,
@@ -88,7 +88,7 @@ typedef void (*ce_send_cb)(struct CE_handle *copyeng,
typedef void (*CE_recv_cb)(struct CE_handle *copyeng,
void *per_CE_recv_context,
void *per_transfer_recv_context,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
unsigned int nbytes,
unsigned int transfer_id,
unsigned int flags);
@@ -142,7 +142,7 @@ struct CE_attr;
*/
int ce_send(struct CE_handle *copyeng,
void *per_transfer_send_context,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
unsigned int nbytes,
unsigned int transfer_id,
unsigned int flags,
@@ -177,7 +177,7 @@ void ce_sendlist_init(struct ce_sendlist *sendlist);
/* Append a simple buffer (address/length) to a sendlist. */
int ce_sendlist_buf_add(struct ce_sendlist *sendlist,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
unsigned int nbytes,
uint32_t flags, /* OR-ed with internal flags */
uint32_t user_flags);
@@ -211,7 +211,7 @@ int ce_sendlist_send(struct CE_handle *copyeng,
*/
int ce_recv_buf_enqueue(struct CE_handle *copyeng,
void *per_transfer_recv_context,
- cdf_dma_addr_t buffer);
+ qdf_dma_addr_t buffer);
/*
* Register a Receive Callback function.
@@ -314,7 +314,7 @@ void ce_enable_msi(struct hif_softc *scn,
int ce_completed_recv_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *flagsp);
@@ -331,7 +331,7 @@ int ce_completed_recv_next(struct CE_handle *copyeng,
int ce_completed_send_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *sw_idx,
@@ -350,22 +350,22 @@ struct CE_handle *ce_init(struct hif_softc *scn,
* receive buffers. Target DMA must be stopped before using
* this API.
*/
-CDF_STATUS
+QDF_STATUS
ce_revoke_recv_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp);
+ qdf_dma_addr_t *bufferp);
/*
* Support clean shutdown by allowing the caller to cancel
* pending sends. Target DMA must be stopped before using
* this API.
*/
-CDF_STATUS
+QDF_STATUS
ce_cancel_send_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
uint32_t *toeplitz_hash_result);
@@ -433,9 +433,9 @@ struct ce_sendlist {
#ifdef IPA_OFFLOAD
void ce_ipa_get_resource(struct CE_handle *ce,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr);
+ qdf_dma_addr_t *ce_reg_paddr);
#else
/**
* ce_ipa_get_resource() - get uc resource on copyengine
@@ -453,9 +453,9 @@ void ce_ipa_get_resource(struct CE_handle *ce,
* Return: None
*/
static inline void ce_ipa_get_resource(struct CE_handle *ce,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr)
+ qdf_dma_addr_t *ce_reg_paddr)
{
return;
}
diff --git a/hif/src/ce/ce_bmi.c b/hif/src/ce/ce_bmi.c
index a1146877df39..fb3d43b0eee9 100644
--- a/hif/src/ce/ce_bmi.c
+++ b/hif/src/ce/ce_bmi.c
@@ -30,9 +30,10 @@
#include "athdefs.h"
#include "osapi_linux.h"
#include "targcfg.h"
-#include "cdf_lock.h"
-#include "cdf_status.h"
-#include <cdf_atomic.h> /* cdf_atomic_read */
+#include "qdf_lock.h"
+#include "qdf_status.h"
+#include "qdf_status.h"
+#include <qdf_atomic.h> /* qdf_atomic_read */
#include <targaddrs.h>
#include <bmi_msg.h>
#include "hif_io32.h"
@@ -42,7 +43,7 @@
#include <a_debug.h>
#include "hif_main.h"
#include "ce_api.h"
-#include "cdf_trace.h"
+#include "qdf_trace.h"
#ifdef CONFIG_CNSS
#include <net/cnss.h>
#else
@@ -63,12 +64,12 @@ enum {
struct BMI_transaction {
struct HIF_CE_state *hif_state;
- cdf_semaphore_t bmi_transaction_sem;
+ qdf_semaphore_t bmi_transaction_sem;
uint8_t *bmi_request_host; /* Req BMI msg in Host addr space */
- cdf_dma_addr_t bmi_request_CE; /* Req BMI msg in CE addr space */
+ qdf_dma_addr_t bmi_request_CE; /* Req BMI msg in CE addr space */
uint32_t bmi_request_length; /* Length of BMI request */
uint8_t *bmi_response_host; /* Rsp BMI msg in Host addr space */
- cdf_dma_addr_t bmi_response_CE; /* Rsp BMI msg in CE addr space */
+ qdf_dma_addr_t bmi_response_CE; /* Rsp BMI msg in CE addr space */
unsigned int bmi_response_length; /* Length of received response */
unsigned int bmi_timeout_ms;
uint32_t bmi_transaction_flags; /* flags for the transcation */
@@ -80,22 +81,20 @@ struct BMI_transaction {
* straight buffer, not an sk_buff.
*/
void hif_bmi_send_done(struct CE_handle *copyeng, void *ce_context,
- void *transfer_context, cdf_dma_addr_t data,
+ void *transfer_context, qdf_dma_addr_t data,
unsigned int nbytes,
unsigned int transfer_id, unsigned int sw_index,
unsigned int hw_index, uint32_t toeplitz_hash_result)
{
struct BMI_transaction *transaction =
(struct BMI_transaction *)transfer_context;
- struct hif_softc *scn = HIF_GET_SOFTC(transaction->hif_state);
#ifdef BMI_RSP_POLLING
/*
* Fix EV118783, Release a semaphore after sending
* no matter whether a response is been expecting now.
*/
- cdf_semaphore_release(scn->cdf_dev,
- &transaction->bmi_transaction_sem);
+ qdf_semaphore_release(&transaction->bmi_transaction_sem);
#else
/*
* If a response is anticipated, we'll complete the
@@ -109,36 +108,33 @@ void hif_bmi_send_done(struct CE_handle *copyeng, void *ce_context,
* never assume resp comes later then this */
if (!transaction->bmi_response_CE ||
(transaction->bmi_transaction_flags & BMI_RESP_RECV_DONE)) {
- cdf_semaphore_release(scn->cdf_dev,
- &transaction->bmi_transaction_sem);
+ qdf_semaphore_release(&transaction->bmi_transaction_sem);
}
#endif
}
#ifndef BMI_RSP_POLLING
void hif_bmi_recv_data(struct CE_handle *copyeng, void *ce_context,
- void *transfer_context, cdf_dma_addr_t data,
+ void *transfer_context, qdf_dma_addr_t data,
unsigned int nbytes,
unsigned int transfer_id, unsigned int flags)
{
struct BMI_transaction *transaction =
(struct BMI_transaction *)transfer_context;
- struct hif_softc *scn = HIF_GET_SOFTC(transaction->hif_state);
transaction->bmi_response_length = nbytes;
transaction->bmi_transaction_flags |= BMI_RESP_RECV_DONE;
/* when both send/recv are done, the sem can be released */
if (transaction->bmi_transaction_flags & BMI_REQ_SEND_DONE) {
- cdf_semaphore_release(scn->cdf_dev,
- &transaction->bmi_transaction_sem);
+ qdf_semaphore_release(&transaction->bmi_transaction_sem);
}
}
#endif
-CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
- cdf_dma_addr_t bmi_cmd_da,
- cdf_dma_addr_t bmi_rsp_da,
+QDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
+ qdf_dma_addr_t bmi_cmd_da,
+ qdf_dma_addr_t bmi_rsp_da,
uint8_t *bmi_request,
uint32_t request_length,
uint8_t *bmi_response,
@@ -150,9 +146,9 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
struct HIF_CE_pipe_info *send_pipe_info =
&(hif_state->pipe_info[BMI_CE_NUM_TO_TARG]);
struct CE_handle *ce_send_hdl = send_pipe_info->ce_hdl;
- cdf_dma_addr_t CE_request, CE_response = 0;
+ qdf_dma_addr_t CE_request, CE_response = 0;
struct BMI_transaction *transaction = NULL;
- int status = CDF_STATUS_SUCCESS;
+ int status = QDF_STATUS_SUCCESS;
struct HIF_CE_pipe_info *recv_pipe_info =
&(hif_state->pipe_info[BMI_CE_NUM_TO_HOST]);
struct CE_handle *ce_recv = recv_pipe_info->ce_hdl;
@@ -160,16 +156,16 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
unsigned int transaction_id = 0xffff;
unsigned int user_flags = 0;
#ifdef BMI_RSP_POLLING
- cdf_dma_addr_t buf;
+ qdf_dma_addr_t buf;
unsigned int completed_nbytes, id, flags;
int i;
#endif
transaction =
- (struct BMI_transaction *)cdf_mem_malloc(sizeof(*transaction));
+ (struct BMI_transaction *)qdf_mem_malloc(sizeof(*transaction));
if (unlikely(!transaction)) {
HIF_ERROR("%s: no memory", __func__);
- return CDF_STATUS_E_NOMEM;
+ return QDF_STATUS_E_NOMEM;
}
transaction_id = (mux_id & MUX_ID_MASK) |
(transaction_id & TRANSACTION_ID_MASK);
@@ -179,8 +175,8 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
A_TARGET_ACCESS_LIKELY(scn);
/* Initialize bmi_transaction_sem to block */
- cdf_semaphore_init(&transaction->bmi_transaction_sem);
- cdf_semaphore_acquire(scn->cdf_dev, &transaction->bmi_transaction_sem);
+ qdf_semaphore_init(&transaction->bmi_transaction_sem);
+ qdf_semaphore_acquire(&transaction->bmi_transaction_sem);
transaction->hif_state = hif_state;
transaction->bmi_request_host = bmi_request;
@@ -207,7 +203,7 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
transaction->bmi_response_CE = CE_response;
/* dma_cache_sync(dev, bmi_response,
BMI_DATASZ_MAX, DMA_FROM_DEVICE); */
- cdf_os_mem_dma_sync_single_for_device(scn->cdf_dev,
+ qdf_mem_dma_sync_single_for_device(scn->qdf_dev,
CE_response,
BMI_DATASZ_MAX,
DMA_FROM_DEVICE);
@@ -220,14 +216,14 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
}
/* dma_cache_sync(dev, bmi_request, request_length, DMA_TO_DEVICE); */
- cdf_os_mem_dma_sync_single_for_device(scn->cdf_dev, CE_request,
+ qdf_mem_dma_sync_single_for_device(scn->qdf_dev, CE_request,
request_length, DMA_TO_DEVICE);
status =
ce_send(ce_send_hdl, transaction,
CE_request, request_length,
transaction_id, 0, user_flags);
- ASSERT(status == CDF_STATUS_SUCCESS);
+ ASSERT(status == QDF_STATUS_SUCCESS);
/* NB: see hif_bmi_send_done */
/* TBDXXX: handle timeout */
@@ -235,8 +231,8 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
/* Wait for BMI request/response transaction to complete */
/* Always just wait for BMI request here if
* BMI_RSP_POLLING is defined */
- while (cdf_semaphore_acquire
- (scn->cdf_dev, &transaction->bmi_transaction_sem)) {
+ while (qdf_semaphore_acquire
+ (&transaction->bmi_transaction_sem)) {
/*need some break out condition(time out?) */
}
@@ -250,20 +246,20 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
while (ce_completed_recv_next(
ce_recv, NULL, NULL, &buf,
&completed_nbytes, &id,
- &flags) != CDF_STATUS_SUCCESS) {
+ &flags) != QDF_STATUS_SUCCESS) {
if (i++ > BMI_RSP_TO_MILLISEC) {
HIF_ERROR("%s:error, can't get bmi response\n",
__func__);
- status = CDF_STATUS_E_BUSY;
+ status = QDF_STATUS_E_BUSY;
break;
}
OS_DELAY(1000);
}
- if ((status == CDF_STATUS_SUCCESS) && bmi_response_lengthp)
+ if ((status == QDF_STATUS_SUCCESS) && bmi_response_lengthp)
*bmi_response_lengthp = completed_nbytes;
#else
- if ((status == CDF_STATUS_SUCCESS) && bmi_response_lengthp) {
+ if ((status == QDF_STATUS_SUCCESS) && bmi_response_lengthp) {
*bmi_response_lengthp =
transaction->bmi_response_length;
}
@@ -277,8 +273,8 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
transaction->bmi_request_CE,
request_length, BUS_DMA_TODEVICE); */
- if (status != CDF_STATUS_SUCCESS) {
- cdf_dma_addr_t unused_buffer;
+ if (status != QDF_STATUS_SUCCESS) {
+ qdf_dma_addr_t unused_buffer;
unsigned int unused_nbytes;
unsigned int unused_id;
unsigned int toeplitz_hash_result;
@@ -290,6 +286,6 @@ CDF_STATUS hif_exchange_bmi_msg(struct hif_opaque_softc *hif_ctx,
}
A_TARGET_ACCESS_UNLIKELY(scn);
- cdf_mem_free(transaction);
+ qdf_mem_free(transaction);
return status;
}
diff --git a/hif/src/ce/ce_bmi.h b/hif/src/ce/ce_bmi.h
index c09b211b6e3e..1fb7c83a97df 100644
--- a/hif/src/ce/ce_bmi.h
+++ b/hif/src/ce/ce_bmi.h
@@ -28,17 +28,17 @@
#ifndef __CE_BMI_H__
#define __CE_BMI_H__
-#include <cdf_atomic.h> /* cdf_atomic_read */
-#include "cdf_lock.h"
+#include <qdf_atomic.h> /* qdf_atomic_read */
+#include "qdf_lock.h"
#include "ce_api.h"
#include "cepci.h"
void hif_bmi_recv_data(struct CE_handle *copyeng, void *ce_context,
- void *transfer_context, cdf_dma_addr_t data,
+ void *transfer_context, qdf_dma_addr_t data,
unsigned int nbytes,
unsigned int transfer_id, unsigned int flags);
void hif_bmi_send_done(struct CE_handle *copyeng, void *ce_context,
- void *transfer_context, cdf_dma_addr_t data,
+ void *transfer_context, qdf_dma_addr_t data,
unsigned int nbytes,
unsigned int transfer_id, unsigned int sw_index,
unsigned int hw_index, uint32_t toeplitz_hash_result);
diff --git a/hif/src/ce/ce_diag.c b/hif/src/ce/ce_diag.c
index 67d85545ff06..05ee4134e2d9 100644
--- a/hif/src/ce/ce_diag.c
+++ b/hif/src/ce/ce_diag.c
@@ -30,9 +30,10 @@
#include "athdefs.h"
#include "osapi_linux.h"
#include "targcfg.h"
-#include "cdf_lock.h"
-#include "cdf_status.h"
-#include <cdf_atomic.h> /* cdf_atomic_read */
+#include "qdf_lock.h"
+#include "qdf_status.h"
+#include "qdf_status.h"
+#include <qdf_atomic.h> /* qdf_atomic_read */
#include <targaddrs.h>
#include <bmi_msg.h>
#include "hif_io32.h"
@@ -42,7 +43,7 @@
#include <a_debug.h>
#include "hif_main.h"
#include "ce_api.h"
-#include "cdf_trace.h"
+#include "qdf_trace.h"
#ifdef CONFIG_CNSS
#include <net/cnss.h>
#endif
@@ -62,7 +63,7 @@ hif_dump_target_memory(struct hif_opaque_softc *hif_ctx, void *ramdump_base,
A_TARGET_ACCESS_BEGIN(scn);
while (j < size) {
val = hif_read32_mb(scn->mem + loc + j);
- cdf_mem_copy(temp, &val, 4);
+ qdf_mem_copy(temp, &val, 4);
j += 4;
temp += 4;
}
@@ -94,25 +95,25 @@ hif_dump_target_memory(struct hif_opaque_softc *hif_ctx, void *ramdump_base,
* at any moment.
*/
-CDF_STATUS
+QDF_STATUS
hif_diag_read_mem(struct hif_opaque_softc *hif_ctx, uint32_t address,
uint8_t *data, int nbytes)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(scn);
- CDF_STATUS status = CDF_STATUS_SUCCESS;
- cdf_dma_addr_t buf;
+ QDF_STATUS status = QDF_STATUS_SUCCESS;
+ qdf_dma_addr_t buf;
unsigned int completed_nbytes, orig_nbytes, remaining_bytes;
unsigned int id;
unsigned int flags;
struct CE_handle *ce_diag;
- cdf_dma_addr_t CE_data; /* Host buffer address in CE space */
- cdf_dma_addr_t CE_data_base = 0;
+ qdf_dma_addr_t CE_data; /* Host buffer address in CE space */
+ qdf_dma_addr_t CE_data_base = 0;
void *data_buf = NULL;
int i;
unsigned int mux_id = 0;
unsigned int transaction_id = 0xffff;
- cdf_dma_addr_t ce_phy_addr = address;
+ qdf_dma_addr_t ce_phy_addr = address;
unsigned int toeplitz_hash_result;
unsigned int user_flags = 0;
@@ -129,10 +130,10 @@ hif_diag_read_mem(struct hif_opaque_softc *hif_ctx, uint32_t address,
if (address < DRAM_BASE_ADDRESS) {
if ((address & 0x3) || ((uintptr_t) data & 0x3))
- return CDF_STATUS_E_INVAL;
+ return QDF_STATUS_E_INVAL;
while ((nbytes >= 4) &&
- (CDF_STATUS_SUCCESS == (status =
+ (QDF_STATUS_SUCCESS == (status =
hif_diag_read_access(hif_ctx, address,
(uint32_t *)data)))) {
@@ -155,14 +156,14 @@ hif_diag_read_mem(struct hif_opaque_softc *hif_ctx, uint32_t address,
* 2) Buffer in DMA-able space
*/
orig_nbytes = nbytes;
- data_buf = cdf_os_mem_alloc_consistent(scn->cdf_dev,
- orig_nbytes, &CE_data_base, 0);
+ data_buf = qdf_mem_alloc_consistent(scn->qdf_dev, scn->qdf_dev->dev,
+ orig_nbytes, &CE_data_base);
if (!data_buf) {
- status = CDF_STATUS_E_NOMEM;
+ status = QDF_STATUS_E_NOMEM;
goto done;
}
- cdf_mem_set(data_buf, orig_nbytes, 0);
- cdf_os_mem_dma_sync_single_for_device(scn->cdf_dev, CE_data_base,
+ qdf_mem_set(data_buf, orig_nbytes, 0);
+ qdf_mem_dma_sync_single_for_device(scn->qdf_dev, CE_data_base,
orig_nbytes, DMA_FROM_DEVICE);
remaining_bytes = orig_nbytes;
@@ -171,7 +172,7 @@ hif_diag_read_mem(struct hif_opaque_softc *hif_ctx, uint32_t address,
nbytes = min(remaining_bytes, DIAG_TRANSFER_LIMIT);
{
status = ce_recv_buf_enqueue(ce_diag, NULL, CE_data);
- if (status != CDF_STATUS_SUCCESS)
+ if (status != QDF_STATUS_SUCCESS)
goto done;
}
@@ -195,26 +196,26 @@ hif_diag_read_mem(struct hif_opaque_softc *hif_ctx, uint32_t address,
status =
ce_send(ce_diag, NULL, ce_phy_addr, nbytes,
transaction_id, 0, user_flags);
- if (status != CDF_STATUS_SUCCESS)
+ if (status != QDF_STATUS_SUCCESS)
goto done;
}
i = 0;
while (ce_completed_send_next(ce_diag, NULL, NULL, &buf,
&completed_nbytes, &id, NULL, NULL,
- &toeplitz_hash_result) != CDF_STATUS_SUCCESS) {
- cdf_mdelay(1);
+ &toeplitz_hash_result) != QDF_STATUS_SUCCESS) {
+ qdf_mdelay(1);
if (i++ > DIAG_ACCESS_CE_TIMEOUT_MS) {
- status = CDF_STATUS_E_BUSY;
+ status = QDF_STATUS_E_BUSY;
goto done;
}
}
if (nbytes != completed_nbytes) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
if (buf != ce_phy_addr) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
@@ -222,19 +223,19 @@ hif_diag_read_mem(struct hif_opaque_softc *hif_ctx, uint32_t address,
while (ce_completed_recv_next
(ce_diag, NULL, NULL, &buf,
&completed_nbytes, &id,
- &flags) != CDF_STATUS_SUCCESS) {
- cdf_mdelay(1);
+ &flags) != QDF_STATUS_SUCCESS) {
+ qdf_mdelay(1);
if (i++ > DIAG_ACCESS_CE_TIMEOUT_MS) {
- status = CDF_STATUS_E_BUSY;
+ status = QDF_STATUS_E_BUSY;
goto done;
}
}
if (nbytes != completed_nbytes) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
if (buf != CE_data) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
@@ -246,20 +247,20 @@ hif_diag_read_mem(struct hif_opaque_softc *hif_ctx, uint32_t address,
done:
A_TARGET_ACCESS_UNLIKELY(scn);
- if (status == CDF_STATUS_SUCCESS)
- cdf_mem_copy(data, data_buf, orig_nbytes);
+ if (status == QDF_STATUS_SUCCESS)
+ qdf_mem_copy(data, data_buf, orig_nbytes);
else
HIF_ERROR("%s failure (0x%x)", __func__, address);
if (data_buf)
- cdf_os_mem_free_consistent(scn->cdf_dev, orig_nbytes,
- data_buf, CE_data_base, 0);
+ qdf_mem_free_consistent(scn->qdf_dev, scn->qdf_dev->dev,
+ orig_nbytes, data_buf, CE_data_base, 0);
return status;
}
/* Read 4-byte aligned data from Target memory or register */
-CDF_STATUS hif_diag_read_access(struct hif_opaque_softc *hif_ctx,
+QDF_STATUS hif_diag_read_access(struct hif_opaque_softc *hif_ctx,
uint32_t address, uint32_t *data)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
@@ -273,28 +274,28 @@ CDF_STATUS hif_diag_read_access(struct hif_opaque_softc *hif_ctx,
*data = A_TARGET_READ(scn, address);
A_TARGET_ACCESS_END_RET(scn);
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
}
-CDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *hif_ctx,
+QDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *hif_ctx,
uint32_t address, uint8_t *data, int nbytes)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(hif_ctx);
- CDF_STATUS status = CDF_STATUS_SUCCESS;
- cdf_dma_addr_t buf;
+ QDF_STATUS status = QDF_STATUS_SUCCESS;
+ qdf_dma_addr_t buf;
unsigned int completed_nbytes, orig_nbytes, remaining_bytes;
unsigned int id;
unsigned int flags;
struct CE_handle *ce_diag;
void *data_buf = NULL;
- cdf_dma_addr_t CE_data; /* Host buffer address in CE space */
- cdf_dma_addr_t CE_data_base = 0;
+ qdf_dma_addr_t CE_data; /* Host buffer address in CE space */
+ qdf_dma_addr_t CE_data_base = 0;
int i;
unsigned int mux_id = 0;
unsigned int transaction_id = 0xffff;
- cdf_dma_addr_t ce_phy_addr = address;
+ qdf_dma_addr_t ce_phy_addr = address;
unsigned int toeplitz_hash_result;
unsigned int user_flags = 0;
@@ -314,16 +315,16 @@ CDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *hif_ctx,
* 2) Buffer in DMA-able space
*/
orig_nbytes = nbytes;
- data_buf = cdf_os_mem_alloc_consistent(scn->cdf_dev,
- orig_nbytes, &CE_data_base, 0);
+ data_buf = qdf_mem_alloc_consistent(scn->qdf_dev, scn->qdf_dev->dev,
+ orig_nbytes, &CE_data_base);
if (!data_buf) {
status = A_NO_MEMORY;
goto done;
}
/* Copy caller's data to allocated DMA buf */
- cdf_mem_copy(data_buf, data, orig_nbytes);
- cdf_os_mem_dma_sync_single_for_device(scn->cdf_dev, CE_data_base,
+ qdf_mem_copy(data_buf, data, orig_nbytes);
+ qdf_mem_dma_sync_single_for_device(scn->qdf_dev, CE_data_base,
orig_nbytes, DMA_TO_DEVICE);
/*
@@ -348,7 +349,7 @@ CDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *hif_ctx,
{ /* Set up to receive directly into Target(!) address */
status = ce_recv_buf_enqueue(ce_diag,
NULL, ce_phy_addr);
- if (status != CDF_STATUS_SUCCESS)
+ if (status != QDF_STATUS_SUCCESS)
goto done;
}
@@ -359,9 +360,9 @@ CDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *hif_ctx,
*/
status =
ce_send(ce_diag, NULL,
- (cdf_dma_addr_t) CE_data, nbytes,
+ (qdf_dma_addr_t) CE_data, nbytes,
transaction_id, 0, user_flags);
- if (status != CDF_STATUS_SUCCESS)
+ if (status != QDF_STATUS_SUCCESS)
goto done;
}
@@ -369,21 +370,21 @@ CDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *hif_ctx,
while (ce_completed_send_next(ce_diag, NULL, NULL, &buf,
&completed_nbytes, &id,
NULL, NULL, &toeplitz_hash_result) !=
- CDF_STATUS_SUCCESS) {
- cdf_mdelay(1);
+ QDF_STATUS_SUCCESS) {
+ qdf_mdelay(1);
if (i++ > DIAG_ACCESS_CE_TIMEOUT_MS) {
- status = CDF_STATUS_E_BUSY;
+ status = QDF_STATUS_E_BUSY;
goto done;
}
}
if (nbytes != completed_nbytes) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
if (buf != CE_data) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
@@ -391,21 +392,21 @@ CDF_STATUS hif_diag_write_mem(struct hif_opaque_softc *hif_ctx,
while (ce_completed_recv_next
(ce_diag, NULL, NULL, &buf,
&completed_nbytes, &id,
- &flags) != CDF_STATUS_SUCCESS) {
- cdf_mdelay(1);
+ &flags) != QDF_STATUS_SUCCESS) {
+ qdf_mdelay(1);
if (i++ > DIAG_ACCESS_CE_TIMEOUT_MS) {
- status = CDF_STATUS_E_BUSY;
+ status = QDF_STATUS_E_BUSY;
goto done;
}
}
if (nbytes != completed_nbytes) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
if (buf != ce_phy_addr) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
@@ -418,11 +419,11 @@ done:
A_TARGET_ACCESS_UNLIKELY(scn);
if (data_buf) {
- cdf_os_mem_free_consistent(scn->cdf_dev, orig_nbytes,
- data_buf, CE_data_base, 0);
+ qdf_mem_free_consistent(scn->qdf_dev, scn->qdf_dev->dev,
+ orig_nbytes, data_buf, CE_data_base, 0);
}
- if (status != CDF_STATUS_SUCCESS) {
+ if (status != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s failure (0x%llu)", __func__,
(uint64_t)ce_phy_addr);
}
@@ -431,7 +432,7 @@ done:
}
/* Write 4B data to Target memory or register */
-CDF_STATUS hif_diag_write_access(struct hif_opaque_softc *hif_ctx,
+QDF_STATUS hif_diag_write_access(struct hif_opaque_softc *hif_ctx,
uint32_t address, uint32_t data)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
@@ -448,6 +449,6 @@ CDF_STATUS hif_diag_write_access(struct hif_opaque_softc *hif_ctx,
A_TARGET_WRITE(scn, address, data);
A_TARGET_ACCESS_END_RET(scn);
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
}
diff --git a/hif/src/ce/ce_internal.h b/hif/src/ce/ce_internal.h
index dabb7d37fb39..60f483a04dbd 100644
--- a/hif/src/ce/ce_internal.h
+++ b/hif/src/ce/ce_internal.h
@@ -76,7 +76,7 @@ struct CE_ring_state {
/* Start of DMA-coherent area reserved for descriptors */
void *base_addr_owner_space_unaligned; /* Host address space */
- cdf_dma_addr_t base_addr_CE_space_unaligned; /* CE address space */
+ qdf_dma_addr_t base_addr_CE_space_unaligned; /* CE address space */
/*
* Actual start of descriptors.
@@ -84,7 +84,7 @@ struct CE_ring_state {
* Points into reserved DMA-coherent area, above.
*/
void *base_addr_owner_space; /* Host address space */
- cdf_dma_addr_t base_addr_CE_space; /* CE address space */
+ qdf_dma_addr_t base_addr_CE_space; /* CE address space */
/*
* Start of shadow copy of descriptors, within regular memory.
* Aligned to descriptor-size boundary.
@@ -130,7 +130,7 @@ struct CE_state {
struct CE_ring_state *dest_ring;
atomic_t rx_pending;
- cdf_spinlock_t ce_index_lock;
+ qdf_spinlock_t ce_index_lock;
bool force_break; /* Flag to indicate whether to
* break out the DPC context */
@@ -139,7 +139,7 @@ struct CE_state {
* DPC routine */
/* epping */
bool timer_inited;
- cdf_softirq_timer_t poll_timer;
+ qdf_timer_t poll_timer;
void (*lro_flush_cb)(void *);
void *lro_data;
};
@@ -149,11 +149,11 @@ struct CE_state {
#ifdef QCA_WIFI_3_0
#define HIF_CE_DESC_ADDR_TO_DMA(desc) \
- (cdf_dma_addr_t)(((uint64_t)(desc)->buffer_addr + \
+ (qdf_dma_addr_t)(((uint64_t)(desc)->buffer_addr + \
((uint64_t)((desc)->buffer_addr_hi & 0x1F) << 32)))
#else
#define HIF_CE_DESC_ADDR_TO_DMA(desc) \
- (cdf_dma_addr_t)((desc)->buffer_addr)
+ (qdf_dma_addr_t)((desc)->buffer_addr)
#endif
#ifdef QCA_WIFI_3_0
diff --git a/hif/src/ce/ce_main.c b/hif/src/ce/ce_main.c
index 58f87d593881..3320d8c47565 100644
--- a/hif/src/ce/ce_main.c
+++ b/hif/src/ce/ce_main.c
@@ -29,9 +29,10 @@
#include "athdefs.h"
#include "osapi_linux.h"
#include "targcfg.h"
-#include "cdf_lock.h"
-#include "cdf_status.h"
-#include <cdf_atomic.h> /* cdf_atomic_read */
+#include "qdf_lock.h"
+#include "qdf_status.h"
+#include "qdf_status.h"
+#include <qdf_atomic.h> /* qdf_atomic_read */
#include <targaddrs.h>
#include <bmi_msg.h>
#include "hif_io32.h"
@@ -44,7 +45,7 @@
#include "ce_bmi.h"
#endif
#include "ce_api.h"
-#include "cdf_trace.h"
+#include "qdf_trace.h"
#ifdef CONFIG_CNSS
#include <net/cnss.h>
#endif
@@ -82,7 +83,7 @@ static void ce_poll_timeout(void *arg)
struct CE_state *CE_state = (struct CE_state *)arg;
if (CE_state->timer_inited) {
ce_per_engine_service(CE_state->scn, CE_state->id);
- cdf_softirq_timer_mod(&CE_state->poll_timer, CE_POLL_TIMEOUT);
+ qdf_timer_mod(&CE_state->poll_timer, CE_POLL_TIMEOUT);
}
}
@@ -101,7 +102,7 @@ static unsigned int roundup_pwr2(unsigned int n)
test_pwr2 = test_pwr2 << 1;
}
- CDF_ASSERT(0); /* n too large */
+ QDF_ASSERT(0); /* n too large */
return 0;
}
@@ -121,25 +122,25 @@ struct CE_handle *ce_init(struct hif_softc *scn,
struct CE_state *CE_state;
uint32_t ctrl_addr;
unsigned int nentries;
- cdf_dma_addr_t base_addr;
+ qdf_dma_addr_t base_addr;
bool malloc_CE_state = false;
bool malloc_src_ring = false;
- CDF_ASSERT(CE_id < scn->ce_count);
+ QDF_ASSERT(CE_id < scn->ce_count);
ctrl_addr = CE_BASE_ADDRESS(CE_id);
CE_state = scn->ce_id_to_state[CE_id];
if (!CE_state) {
CE_state =
- (struct CE_state *)cdf_mem_malloc(sizeof(*CE_state));
+ (struct CE_state *)qdf_mem_malloc(sizeof(*CE_state));
if (!CE_state) {
HIF_ERROR("%s: CE_state has no mem", __func__);
return NULL;
}
malloc_CE_state = true;
- cdf_mem_zero(CE_state, sizeof(*CE_state));
+ qdf_mem_zero(CE_state, sizeof(*CE_state));
scn->ce_id_to_state[CE_id] = CE_state;
- cdf_spinlock_init(&CE_state->ce_index_lock);
+ qdf_spinlock_create(&CE_state->ce_index_lock);
CE_state->id = CE_id;
CE_state->ctrl_addr = ctrl_addr;
@@ -148,7 +149,7 @@ struct CE_handle *ce_init(struct hif_softc *scn,
}
CE_state->scn = scn;
- cdf_atomic_init(&CE_state->rx_pending);
+ qdf_atomic_init(&CE_state->rx_pending);
if (attr == NULL) {
/* Already initialized; caller wants the handle */
return (struct CE_handle *)CE_state;
@@ -160,7 +161,7 @@ struct CE_handle *ce_init(struct hif_softc *scn,
#endif
if (CE_state->src_sz_max)
- CDF_ASSERT(CE_state->src_sz_max == attr->src_sz_max);
+ QDF_ASSERT(CE_state->src_sz_max == attr->src_sz_max);
else
CE_state->src_sz_max = attr->src_sz_max;
@@ -176,11 +177,11 @@ struct CE_handle *ce_init(struct hif_softc *scn,
uint64_t dma_addr;
nentries = roundup_pwr2(nentries);
if (CE_state->src_ring) {
- CDF_ASSERT(CE_state->src_ring->nentries == nentries);
+ QDF_ASSERT(CE_state->src_ring->nentries == nentries);
} else {
CE_nbytes = sizeof(struct CE_ring_state)
+ (nentries * sizeof(void *));
- ptr = cdf_mem_malloc(CE_nbytes);
+ ptr = qdf_mem_malloc(CE_nbytes);
if (!ptr) {
/* cannot allocate src ring. If the
* CE_state is allocated locally free
@@ -190,7 +191,7 @@ struct CE_handle *ce_init(struct hif_softc *scn,
if (malloc_CE_state) {
/* allocated CE_state locally */
scn->ce_id_to_state[CE_id] = NULL;
- cdf_mem_free(CE_state);
+ qdf_mem_free(CE_state);
malloc_CE_state = false;
}
return NULL;
@@ -201,7 +202,7 @@ struct CE_handle *ce_init(struct hif_softc *scn,
*/
malloc_src_ring = true;
}
- cdf_mem_zero(ptr, CE_nbytes);
+ qdf_mem_zero(ptr, CE_nbytes);
src_ring = CE_state->src_ring =
(struct CE_ring_state *)ptr;
@@ -223,11 +224,12 @@ struct CE_handle *ce_init(struct hif_softc *scn,
* coherent DMA are unsupported
*/
src_ring->base_addr_owner_space_unaligned =
- cdf_os_mem_alloc_consistent(scn->cdf_dev,
+ qdf_mem_alloc_consistent(scn->qdf_dev,
+ scn->qdf_dev->dev,
(nentries *
sizeof(struct CE_src_desc) +
CE_DESC_RING_ALIGN),
- &base_addr, 0);
+ &base_addr);
if (src_ring->base_addr_owner_space_unaligned
== NULL) {
HIF_ERROR("%s: src ring has no DMA mem",
@@ -262,7 +264,7 @@ struct CE_handle *ce_init(struct hif_softc *scn,
* regular mem to use for faster access.
*/
src_ring->shadow_base_unaligned =
- cdf_mem_malloc(nentries *
+ qdf_mem_malloc(nentries *
sizeof(struct CE_src_desc) +
CE_DESC_RING_ALIGN);
if (src_ring->shadow_base_unaligned == NULL) {
@@ -312,11 +314,11 @@ struct CE_handle *ce_init(struct hif_softc *scn,
nentries = roundup_pwr2(nentries);
if (CE_state->dest_ring) {
- CDF_ASSERT(CE_state->dest_ring->nentries == nentries);
+ QDF_ASSERT(CE_state->dest_ring->nentries == nentries);
} else {
CE_nbytes = sizeof(struct CE_ring_state)
+ (nentries * sizeof(void *));
- ptr = cdf_mem_malloc(CE_nbytes);
+ ptr = qdf_mem_malloc(CE_nbytes);
if (!ptr) {
/* cannot allocate dst ring. If the CE_state
* or src ring is allocated locally free
@@ -325,19 +327,19 @@ struct CE_handle *ce_init(struct hif_softc *scn,
HIF_ERROR("%s: dest ring has no mem",
__func__);
if (malloc_src_ring) {
- cdf_mem_free(CE_state->src_ring);
+ qdf_mem_free(CE_state->src_ring);
CE_state->src_ring = NULL;
malloc_src_ring = false;
}
if (malloc_CE_state) {
/* allocated CE_state locally */
scn->ce_id_to_state[CE_id] = NULL;
- cdf_mem_free(CE_state);
+ qdf_mem_free(CE_state);
malloc_CE_state = false;
}
return NULL;
}
- cdf_mem_zero(ptr, CE_nbytes);
+ qdf_mem_zero(ptr, CE_nbytes);
dest_ring = CE_state->dest_ring =
(struct CE_ring_state *)ptr;
@@ -357,11 +359,12 @@ struct CE_handle *ce_init(struct hif_softc *scn,
/* Legacy platforms that do not support cache
* coherent DMA are unsupported */
dest_ring->base_addr_owner_space_unaligned =
- cdf_os_mem_alloc_consistent(scn->cdf_dev,
+ qdf_mem_alloc_consistent(scn->qdf_dev,
+ scn->qdf_dev->dev,
(nentries *
sizeof(struct CE_dest_desc) +
CE_DESC_RING_ALIGN),
- &base_addr, 0);
+ &base_addr);
if (dest_ring->base_addr_owner_space_unaligned
== NULL) {
HIF_ERROR("%s: dest ring has no DMA mem",
@@ -374,7 +377,7 @@ struct CE_handle *ce_init(struct hif_softc *scn,
* prevent garbage data crashing system
* when download firmware
*/
- cdf_mem_zero(dest_ring->base_addr_owner_space_unaligned,
+ qdf_mem_zero(dest_ring->base_addr_owner_space_unaligned,
nentries * sizeof(struct CE_dest_desc) +
CE_DESC_RING_ALIGN);
@@ -429,13 +432,13 @@ struct CE_handle *ce_init(struct hif_softc *scn,
/* epping */
/* poll timer */
if ((CE_state->attr_flags & CE_ATTR_ENABLE_POLL)) {
- cdf_softirq_timer_init(scn->cdf_dev,
+ qdf_timer_init(scn->qdf_dev,
&CE_state->poll_timer,
ce_poll_timeout,
CE_state,
- CDF_TIMER_TYPE_SW);
+ QDF_TIMER_TYPE_SW);
CE_state->timer_inited = true;
- cdf_softirq_timer_mod(&CE_state->poll_timer,
+ qdf_timer_mod(&CE_state->poll_timer,
CE_POLL_TIMEOUT);
}
}
@@ -482,7 +485,7 @@ ce_h2t_tx_ce_cleanup(struct CE_handle *ce_hdl)
write_index = src_ring->sw_index;
/* At this point Tx CE should be clean */
- cdf_assert_always(sw_index == write_index);
+ qdf_assert_always(sw_index == write_index);
}
}
#else
@@ -504,9 +507,10 @@ void ce_fini(struct CE_handle *copyeng)
ce_h2t_tx_ce_cleanup(copyeng);
if (CE_state->src_ring->shadow_base_unaligned)
- cdf_mem_free(CE_state->src_ring->shadow_base_unaligned);
+ qdf_mem_free(CE_state->src_ring->shadow_base_unaligned);
if (CE_state->src_ring->base_addr_owner_space_unaligned)
- cdf_os_mem_free_consistent(scn->cdf_dev,
+ qdf_mem_free_consistent(scn->qdf_dev,
+ scn->qdf_dev->dev,
(CE_state->src_ring->nentries *
sizeof(struct CE_src_desc) +
CE_DESC_RING_ALIGN),
@@ -514,11 +518,12 @@ void ce_fini(struct CE_handle *copyeng)
base_addr_owner_space_unaligned,
CE_state->src_ring->
base_addr_CE_space, 0);
- cdf_mem_free(CE_state->src_ring);
+ qdf_mem_free(CE_state->src_ring);
}
if (CE_state->dest_ring) {
if (CE_state->dest_ring->base_addr_owner_space_unaligned)
- cdf_os_mem_free_consistent(scn->cdf_dev,
+ qdf_mem_free_consistent(scn->qdf_dev,
+ scn->qdf_dev->dev,
(CE_state->dest_ring->nentries *
sizeof(struct CE_dest_desc) +
CE_DESC_RING_ALIGN),
@@ -526,29 +531,29 @@ void ce_fini(struct CE_handle *copyeng)
base_addr_owner_space_unaligned,
CE_state->dest_ring->
base_addr_CE_space, 0);
- cdf_mem_free(CE_state->dest_ring);
+ qdf_mem_free(CE_state->dest_ring);
/* epping */
if (CE_state->timer_inited) {
CE_state->timer_inited = false;
- cdf_softirq_timer_free(&CE_state->poll_timer);
+ qdf_timer_free(&CE_state->poll_timer);
}
}
- cdf_mem_free(CE_state);
+ qdf_mem_free(CE_state);
}
void hif_detach_htc(struct hif_opaque_softc *hif_ctx)
{
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(hif_ctx);
- cdf_mem_zero(&hif_state->msg_callbacks_pending,
+ qdf_mem_zero(&hif_state->msg_callbacks_pending,
sizeof(hif_state->msg_callbacks_pending));
- cdf_mem_zero(&hif_state->msg_callbacks_current,
+ qdf_mem_zero(&hif_state->msg_callbacks_current,
sizeof(hif_state->msg_callbacks_current));
}
/* Send the first nbytes bytes of the buffer */
-CDF_STATUS
+QDF_STATUS
hif_send_head(struct hif_opaque_softc *hif_ctx,
uint8_t pipe, unsigned int transfer_id, unsigned int nbytes,
cdf_nbuf_t nbuf, unsigned int data_attr)
@@ -562,7 +567,7 @@ hif_send_head(struct hif_opaque_softc *hif_ctx,
int status, i = 0;
unsigned int mux_id = 0;
- CDF_ASSERT(nbytes <= cdf_nbuf_len(nbuf));
+ QDF_ASSERT(nbytes <= cdf_nbuf_len(nbuf));
transfer_id =
(mux_id & MUX_ID_MASK) |
@@ -578,7 +583,7 @@ hif_send_head(struct hif_opaque_softc *hif_ctx,
*/
ce_sendlist_init(&sendlist);
do {
- cdf_dma_addr_t frag_paddr;
+ qdf_dma_addr_t frag_paddr;
int frag_bytes;
frag_paddr = cdf_nbuf_get_frag_paddr(nbuf, nfrags);
@@ -587,7 +592,7 @@ hif_send_head(struct hif_opaque_softc *hif_ctx,
* Clear the packet offset for all but the first CE desc.
*/
if (i++ > 0)
- data_attr &= ~CDF_CE_TX_PKT_OFFSET_BIT_M;
+ data_attr &= ~QDF_CE_TX_PKT_OFFSET_BIT_M;
status = ce_sendlist_buf_add(&sendlist, frag_paddr,
frag_bytes >
@@ -597,7 +602,7 @@ hif_send_head(struct hif_opaque_softc *hif_ctx,
nfrags) ? 0 :
CE_SEND_FLAG_SWAP_DISABLE,
data_attr);
- if (status != CDF_STATUS_SUCCESS) {
+ if (status != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: error, frag_num %d larger than limit",
__func__, nfrags);
return status;
@@ -607,26 +612,26 @@ hif_send_head(struct hif_opaque_softc *hif_ctx,
} while (bytes > 0);
/* Make sure we have resources to handle this request */
- cdf_spin_lock_bh(&pipe_info->completion_freeq_lock);
+ qdf_spin_lock_bh(&pipe_info->completion_freeq_lock);
if (pipe_info->num_sends_allowed < nfrags) {
- cdf_spin_unlock_bh(&pipe_info->completion_freeq_lock);
+ qdf_spin_unlock_bh(&pipe_info->completion_freeq_lock);
ce_pkt_error_count_incr(hif_state, HIF_PIPE_NO_RESOURCE);
- return CDF_STATUS_E_RESOURCES;
+ return QDF_STATUS_E_RESOURCES;
}
pipe_info->num_sends_allowed -= nfrags;
- cdf_spin_unlock_bh(&pipe_info->completion_freeq_lock);
+ qdf_spin_unlock_bh(&pipe_info->completion_freeq_lock);
- if (cdf_unlikely(ce_hdl == NULL)) {
+ if (qdf_unlikely(ce_hdl == NULL)) {
HIF_ERROR("%s: error CE handle is null", __func__);
return A_ERROR;
}
NBUF_UPDATE_TX_PKT_COUNT(nbuf, NBUF_TX_PKT_HIF);
- DPTRACE(cdf_dp_trace(nbuf, CDF_DP_TRACE_HIF_PACKET_PTR_RECORD,
+ DPTRACE(qdf_dp_trace(nbuf, QDF_DP_TRACE_HIF_PACKET_PTR_RECORD,
(uint8_t *)(cdf_nbuf_data(nbuf)),
sizeof(cdf_nbuf_data(nbuf))));
status = ce_sendlist_send(ce_hdl, nbuf, &sendlist, transfer_id);
- CDF_ASSERT(status == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(status == QDF_STATUS_SUCCESS);
return status;
}
@@ -668,16 +673,16 @@ hif_get_free_queue_number(struct hif_opaque_softc *hif_ctx, uint8_t pipe)
struct HIF_CE_pipe_info *pipe_info = &(hif_state->pipe_info[pipe]);
uint16_t rv;
- cdf_spin_lock_bh(&pipe_info->completion_freeq_lock);
+ qdf_spin_lock_bh(&pipe_info->completion_freeq_lock);
rv = pipe_info->num_sends_allowed;
- cdf_spin_unlock_bh(&pipe_info->completion_freeq_lock);
+ qdf_spin_unlock_bh(&pipe_info->completion_freeq_lock);
return rv;
}
/* Called by lower (CE) layer when a send to Target completes. */
void
hif_pci_ce_send_done(struct CE_handle *copyeng, void *ce_context,
- void *transfer_context, cdf_dma_addr_t CE_data,
+ void *transfer_context, qdf_dma_addr_t CE_data,
unsigned int nbytes, unsigned int transfer_id,
unsigned int sw_index, unsigned int hw_index,
unsigned int toeplitz_hash_result)
@@ -706,14 +711,14 @@ hif_pci_ce_send_done(struct CE_handle *copyeng, void *ce_context,
toeplitz_hash_result);
}
- cdf_spin_lock(&pipe_info->completion_freeq_lock);
+ qdf_spin_lock(&pipe_info->completion_freeq_lock);
pipe_info->num_sends_allowed++;
- cdf_spin_unlock(&pipe_info->completion_freeq_lock);
+ qdf_spin_unlock(&pipe_info->completion_freeq_lock);
} while (ce_completed_send_next(copyeng,
&ce_context, &transfer_context,
&CE_data, &nbytes, &transfer_id,
&sw_idx, &hw_idx,
- &toeplitz_hash_result) == CDF_STATUS_SUCCESS);
+ &toeplitz_hash_result) == QDF_STATUS_SUCCESS);
}
/**
@@ -746,7 +751,7 @@ static inline void hif_ce_do_recv(struct hif_msg_callbacks *msg_callbacks,
/* Called by lower (CE) layer when data is received from the Target. */
void
hif_pci_ce_recv_data(struct CE_handle *copyeng, void *ce_context,
- void *transfer_context, cdf_dma_addr_t CE_data,
+ void *transfer_context, qdf_dma_addr_t CE_data,
unsigned int nbytes, unsigned int transfer_id,
unsigned int flags)
{
@@ -762,9 +767,9 @@ hif_pci_ce_recv_data(struct CE_handle *copyeng, void *ce_context,
do {
hif_pm_runtime_mark_last_busy(hif_sc->dev);
- cdf_nbuf_unmap_single(scn->cdf_dev,
+ cdf_nbuf_unmap_single(scn->qdf_dev,
(cdf_nbuf_t) transfer_context,
- CDF_DMA_FROM_DEVICE);
+ QDF_DMA_FROM_DEVICE);
atomic_inc(&pipe_info->recv_bufs_needed);
hif_post_recv_buffers_for_pipe(pipe_info);
@@ -778,13 +783,13 @@ hif_pci_ce_recv_data(struct CE_handle *copyeng, void *ce_context,
* MAX_NUM_OF_RECEIVES */
ce_state->receive_count++;
count = ce_state->receive_count;
- if (cdf_unlikely(hif_max_num_receives_reached(scn, count))) {
+ if (qdf_unlikely(hif_max_num_receives_reached(scn, count))) {
ce_state->force_break = 1;
break;
}
} while (ce_completed_recv_next(copyeng, &ce_context, &transfer_context,
&CE_data, &nbytes, &transfer_id,
- &flags) == CDF_STATUS_SUCCESS);
+ &flags) == QDF_STATUS_SUCCESS);
}
@@ -800,7 +805,7 @@ hif_post_init(struct hif_opaque_softc *hif_ctx, void *unused,
spin_lock_init(&pcie_access_log_lock);
#endif
/* Save callbacks for later installation */
- cdf_mem_copy(&hif_state->msg_callbacks_pending, callbacks,
+ qdf_mem_copy(&hif_state->msg_callbacks_pending, callbacks,
sizeof(hif_state->msg_callbacks_pending));
}
@@ -854,7 +859,7 @@ int hif_completion_thread_startup(struct HIF_CE_state *hif_state)
}
if (attr.src_nentries)
- cdf_spinlock_init(&pipe_info->completion_freeq_lock);
+ qdf_spinlock_create(&pipe_info->completion_freeq_lock);
}
A_TARGET_ACCESS_UNLIKELY(scn);
@@ -873,7 +878,7 @@ static void hif_msg_callbacks_install(struct hif_softc *scn)
{
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(scn);
- cdf_mem_copy(&hif_state->msg_callbacks_current,
+ qdf_mem_copy(&hif_state->msg_callbacks_current,
&hif_state->msg_callbacks_pending,
sizeof(hif_state->msg_callbacks_pending));
}
@@ -925,9 +930,9 @@ void hif_dump_pipe_debug_count(struct hif_softc *scn)
static int hif_post_recv_buffers_for_pipe(struct HIF_CE_pipe_info *pipe_info)
{
struct CE_handle *ce_hdl;
- cdf_size_t buf_sz;
+ qdf_size_t buf_sz;
struct hif_softc *scn = HIF_GET_SOFTC(pipe_info->HIF_CE_state);
- CDF_STATUS ret;
+ QDF_STATUS ret;
uint32_t bufs_posted = 0;
buf_sz = pipe_info->buf_sz;
@@ -938,20 +943,20 @@ static int hif_post_recv_buffers_for_pipe(struct HIF_CE_pipe_info *pipe_info)
ce_hdl = pipe_info->ce_hdl;
- cdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
while (atomic_read(&pipe_info->recv_bufs_needed) > 0) {
- cdf_dma_addr_t CE_data; /* CE space buffer address */
+ qdf_dma_addr_t CE_data; /* CE space buffer address */
cdf_nbuf_t nbuf;
int status;
atomic_dec(&pipe_info->recv_bufs_needed);
- cdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
- nbuf = cdf_nbuf_alloc(scn->cdf_dev, buf_sz, 0, 4, false);
+ nbuf = cdf_nbuf_alloc(scn->qdf_dev, buf_sz, 0, 4, false);
if (!nbuf) {
- cdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
pipe_info->nbuf_alloc_err_count++;
- cdf_spin_unlock_bh(
+ qdf_spin_unlock_bh(
&pipe_info->recv_bufs_needed_lock);
HIF_ERROR(
"%s buf alloc error [%d] needed %d, nbuf_alloc_err_count = %u",
@@ -968,13 +973,13 @@ static int hif_post_recv_buffers_for_pipe(struct HIF_CE_pipe_info *pipe_info)
* DMA_FROM_DEVICE);
*/
ret =
- cdf_nbuf_map_single(scn->cdf_dev, nbuf,
- CDF_DMA_FROM_DEVICE);
+ cdf_nbuf_map_single(scn->qdf_dev, nbuf,
+ QDF_DMA_FROM_DEVICE);
- if (unlikely(ret != CDF_STATUS_SUCCESS)) {
- cdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
+ if (unlikely(ret != QDF_STATUS_SUCCESS)) {
+ qdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
pipe_info->nbuf_dma_err_count++;
- cdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
HIF_ERROR(
"%s buf alloc error [%d] needed %d, nbuf_dma_err_count = %u",
__func__, pipe_info->pipe_num,
@@ -987,14 +992,14 @@ static int hif_post_recv_buffers_for_pipe(struct HIF_CE_pipe_info *pipe_info)
CE_data = cdf_nbuf_get_frag_paddr(nbuf, 0);
- cdf_os_mem_dma_sync_single_for_device(scn->cdf_dev, CE_data,
+ qdf_mem_dma_sync_single_for_device(scn->qdf_dev, CE_data,
buf_sz, DMA_FROM_DEVICE);
status = ce_recv_buf_enqueue(ce_hdl, (void *)nbuf, CE_data);
- CDF_ASSERT(status == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(status == QDF_STATUS_SUCCESS);
if (status != EOK) {
- cdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
pipe_info->nbuf_ce_enqueue_err_count++;
- cdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
HIF_ERROR(
"%s buf alloc error [%d] needed %d, nbuf_alloc_err_count = %u",
__func__, pipe_info->pipe_num,
@@ -1005,7 +1010,7 @@ static int hif_post_recv_buffers_for_pipe(struct HIF_CE_pipe_info *pipe_info)
return 1;
}
- cdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_lock_bh(&pipe_info->recv_bufs_needed_lock);
bufs_posted++;
}
pipe_info->nbuf_alloc_err_count =
@@ -1018,7 +1023,7 @@ static int hif_post_recv_buffers_for_pipe(struct HIF_CE_pipe_info *pipe_info)
(pipe_info->nbuf_ce_enqueue_err_count > bufs_posted) ?
pipe_info->nbuf_ce_enqueue_err_count - bufs_posted : 0;
- cdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
+ qdf_spin_unlock_bh(&pipe_info->recv_bufs_needed_lock);
return 0;
}
@@ -1051,7 +1056,7 @@ done:
return rv;
}
-CDF_STATUS hif_start(struct hif_opaque_softc *hif_ctx)
+QDF_STATUS hif_start(struct hif_opaque_softc *hif_ctx)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(scn);
@@ -1059,14 +1064,14 @@ CDF_STATUS hif_start(struct hif_opaque_softc *hif_ctx)
hif_msg_callbacks_install(scn);
if (hif_completion_thread_startup(hif_state))
- return CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_E_FAILURE;
/* Post buffers once to start things off. */
(void)hif_post_recv_buffers(scn);
hif_state->started = true;
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
#ifdef WLAN_FEATURE_FASTPATH
@@ -1125,7 +1130,7 @@ void hif_recv_buffer_cleanup_on_pipe(struct HIF_CE_pipe_info *pipe_info)
uint32_t buf_sz;
struct HIF_CE_state *hif_state;
cdf_nbuf_t netbuf;
- cdf_dma_addr_t CE_data;
+ qdf_dma_addr_t CE_data;
void *per_CE_context;
buf_sz = pipe_info->buf_sz;
@@ -1142,14 +1147,14 @@ void hif_recv_buffer_cleanup_on_pipe(struct HIF_CE_pipe_info *pipe_info)
scn = HIF_GET_SOFTC(hif_state);
ce_hdl = pipe_info->ce_hdl;
- if (scn->cdf_dev == NULL) {
+ if (scn->qdf_dev == NULL) {
return;
}
while (ce_revoke_recv_next
(ce_hdl, &per_CE_context, (void **)&netbuf,
- &CE_data) == CDF_STATUS_SUCCESS) {
- cdf_nbuf_unmap_single(scn->cdf_dev, netbuf,
- CDF_DMA_FROM_DEVICE);
+ &CE_data) == QDF_STATUS_SUCCESS) {
+ cdf_nbuf_unmap_single(scn->qdf_dev, netbuf,
+ QDF_DMA_FROM_DEVICE);
cdf_nbuf_free(netbuf);
}
}
@@ -1161,7 +1166,7 @@ void hif_send_buffer_cleanup_on_pipe(struct HIF_CE_pipe_info *pipe_info)
struct hif_softc *scn;
cdf_nbuf_t netbuf;
void *per_CE_context;
- cdf_dma_addr_t CE_data;
+ qdf_dma_addr_t CE_data;
unsigned int nbytes;
unsigned int id;
uint32_t buf_sz;
@@ -1185,7 +1190,7 @@ void hif_send_buffer_cleanup_on_pipe(struct HIF_CE_pipe_info *pipe_info)
while (ce_cancel_send_next
(ce_hdl, &per_CE_context,
(void **)&netbuf, &CE_data, &nbytes,
- &id, &toeplitz_hash_result) == CDF_STATUS_SUCCESS) {
+ &id, &toeplitz_hash_result) == QDF_STATUS_SUCCESS) {
if (netbuf != CE_SENDLIST_ITEM_CTXT) {
/*
* Packets enqueued by htt_h2t_ver_req_msg() and
@@ -1271,8 +1276,8 @@ void hif_stop(struct hif_opaque_softc *hif_ctx)
}
if (hif_state->sleep_timer_init) {
- cdf_softirq_timer_cancel(&hif_state->sleep_timer);
- cdf_softirq_timer_free(&hif_state->sleep_timer);
+ qdf_timer_stop(&hif_state->sleep_timer);
+ qdf_timer_free(&hif_state->sleep_timer);
hif_state->sleep_timer_init = false;
}
@@ -1497,21 +1502,21 @@ static struct service_to_pipe target_service_to_ce_map_wlan_epping[] = {
*/
void hif_wake_target_cpu(struct hif_softc *scn)
{
- CDF_STATUS rv;
+ QDF_STATUS rv;
uint32_t core_ctrl;
struct hif_opaque_softc *hif_hdl = GET_HIF_OPAQUE_HDL(scn);
rv = hif_diag_read_access(hif_hdl,
SOC_CORE_BASE_ADDRESS | CORE_CTRL_ADDRESS,
&core_ctrl);
- CDF_ASSERT(rv == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(rv == QDF_STATUS_SUCCESS);
/* A_INUM_FIRMWARE interrupt to Target CPU */
core_ctrl |= CORE_CTRL_CPU_INTR_MASK;
rv = hif_diag_write_access(hif_hdl,
SOC_CORE_BASE_ADDRESS | CORE_CTRL_ADDRESS,
core_ctrl);
- CDF_ASSERT(rv == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(rv == QDF_STATUS_SUCCESS);
}
#endif
@@ -1527,26 +1532,26 @@ static void hif_sleep_entry(void *arg)
if (hif_is_driver_unloading(scn))
return;
- cdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
+ qdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
if (hif_state->verified_awake == false) {
- idle_ms = cdf_system_ticks_to_msecs(cdf_system_ticks()
+ idle_ms = qdf_system_ticks_to_msecs(qdf_system_ticks()
- hif_state->sleep_ticks);
if (idle_ms >= HIF_MIN_SLEEP_INACTIVITY_TIME_MS) {
- if (!cdf_atomic_read(&scn->link_suspended)) {
+ if (!qdf_atomic_read(&scn->link_suspended)) {
soc_wake_reset(scn);
hif_state->fake_sleep = false;
}
} else {
- cdf_softirq_timer_cancel(&hif_state->sleep_timer);
- cdf_softirq_timer_start(&hif_state->sleep_timer,
+ qdf_timer_stop(&hif_state->sleep_timer);
+ qdf_timer_start(&hif_state->sleep_timer,
HIF_SLEEP_INACTIVITY_TIMER_PERIOD_MS);
}
} else {
- cdf_softirq_timer_cancel(&hif_state->sleep_timer);
- cdf_softirq_timer_start(&hif_state->sleep_timer,
+ qdf_timer_stop(&hif_state->sleep_timer);
+ qdf_timer_start(&hif_state->sleep_timer,
HIF_SLEEP_INACTIVITY_TIMER_PERIOD_MS);
}
- cdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
+ qdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
}
#define HIF_HIA_MAX_POLL_LOOP 1000000
#define HIF_HIA_POLLING_DELAY_MS 10
@@ -1559,7 +1564,7 @@ int hif_set_hia(struct hif_softc *scn)
#else
int hif_set_hia(struct hif_softc *scn)
{
- CDF_STATUS rv;
+ QDF_STATUS rv;
uint32_t interconnect_targ_addr = 0;
uint32_t pcie_state_targ_addr = 0;
uint32_t pipe_cfg_targ_addr = 0;
@@ -1584,7 +1589,7 @@ int hif_set_hia(struct hif_softc *scn)
HIF_TRACE("%s: E", __func__);
if (ADRASTEA_BU)
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
#ifdef QCA_WIFI_3_0
i = 0;
@@ -1592,7 +1597,7 @@ int hif_set_hia(struct hif_softc *scn)
host_interest_area = hif_read32_mb(scn->mem +
A_SOC_CORE_SCRATCH_0_ADDRESS);
if ((host_interest_area & 0x01) == 0) {
- cdf_mdelay(HIF_HIA_POLLING_DELAY_MS);
+ qdf_mdelay(HIF_HIA_POLLING_DELAY_MS);
host_interest_area = 0;
i++;
if (i > HIF_HIA_MAX_POLL_LOOP && (i % 1000 == 0)) {
@@ -1633,13 +1638,13 @@ int hif_set_hia(struct hif_softc *scn)
/* Supply Target-side CE configuration */
rv = hif_diag_read_access(hif_hdl, interconnect_targ_addr,
&pcie_state_targ_addr);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: interconnect_targ_addr = 0x%0x, ret = %d",
__func__, interconnect_targ_addr, rv);
goto done;
}
if (pcie_state_targ_addr == 0) {
- rv = CDF_STATUS_E_FAILURE;
+ rv = QDF_STATUS_E_FAILURE;
HIF_ERROR("%s: pcie state addr is 0", __func__);
goto done;
}
@@ -1649,13 +1654,13 @@ int hif_set_hia(struct hif_softc *scn)
rv = hif_diag_read_access(hif_hdl,
pipe_cfg_addr,
&pipe_cfg_targ_addr);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: pipe_cfg_addr = 0x%0x, ret = %d",
__func__, pipe_cfg_addr, rv);
goto done;
}
if (pipe_cfg_targ_addr == 0) {
- rv = CDF_STATUS_E_FAILURE;
+ rv = QDF_STATUS_E_FAILURE;
HIF_ERROR("%s: pipe cfg addr is 0", __func__);
goto done;
}
@@ -1664,7 +1669,7 @@ int hif_set_hia(struct hif_softc *scn)
(uint8_t *) target_ce_config,
target_ce_config_sz);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: write pipe cfg (%d)", __func__, rv);
goto done;
}
@@ -1674,12 +1679,12 @@ int hif_set_hia(struct hif_softc *scn)
offsetof(struct pcie_state_s,
svc_to_pipe_map),
&svc_to_pipe_map);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: get svc/pipe map (%d)", __func__, rv);
goto done;
}
if (svc_to_pipe_map == 0) {
- rv = CDF_STATUS_E_FAILURE;
+ rv = QDF_STATUS_E_FAILURE;
HIF_ERROR("%s: svc_to_pipe map is 0", __func__);
goto done;
}
@@ -1688,7 +1693,7 @@ int hif_set_hia(struct hif_softc *scn)
svc_to_pipe_map,
(uint8_t *) target_service_to_ce_map,
target_service_to_ce_map_sz);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: write svc/pipe map (%d)", __func__, rv);
goto done;
}
@@ -1698,7 +1703,7 @@ int hif_set_hia(struct hif_softc *scn)
offsetof(struct pcie_state_s,
config_flags),
&pcie_config_flags);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: get pcie config_flags (%d)", __func__, rv);
goto done;
}
@@ -1717,7 +1722,7 @@ int hif_set_hia(struct hif_softc *scn)
config_flags),
(uint8_t *) &pcie_config_flags,
sizeof(pcie_config_flags));
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: write pcie config_flags (%d)", __func__, rv);
goto done;
}
@@ -1731,7 +1736,7 @@ int hif_set_hia(struct hif_softc *scn)
rv = hif_diag_read_access(hif_hdl, ealloc_targ_addr,
&ealloc_value);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: get early alloc val (%d)", __func__, rv);
goto done;
}
@@ -1744,7 +1749,7 @@ int hif_set_hia(struct hif_softc *scn)
rv = hif_diag_read_access(hif_hdl,
CHIP_ID_ADDRESS |
RTC_SOC_BASE_ADDRESS, &chip_id);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: get chip id val (%d)", __func__, rv);
goto done;
}
@@ -1780,7 +1785,7 @@ int hif_set_hia(struct hif_softc *scn)
rv = hif_diag_write_access(hif_hdl,
ealloc_targ_addr,
ealloc_value);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: set early alloc val (%d)", __func__, rv);
goto done;
}
@@ -1794,7 +1799,7 @@ int hif_set_hia(struct hif_softc *scn)
rv = hif_diag_read_access(hif_hdl, flag2_targ_addr,
&flag2_value);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: get option val (%d)", __func__, rv);
goto done;
}
@@ -1802,7 +1807,7 @@ int hif_set_hia(struct hif_softc *scn)
flag2_value |= HI_OPTION_EARLY_CFG_DONE;
rv = hif_diag_write_access(hif_hdl, flag2_targ_addr,
flag2_value);
- if (rv != CDF_STATUS_SUCCESS) {
+ if (rv != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: set option val (%d)", __func__, rv);
goto done;
}
@@ -1840,7 +1845,7 @@ static int hif_wlan_enable(struct hif_softc *scn)
cfg.shadow_reg_cfg =
(struct icnss_shadow_reg_cfg *) target_shadow_reg_cfg;
- if (CDF_GLOBAL_FTM_MODE == con_mode)
+ if (QDF_GLOBAL_FTM_MODE == con_mode)
mode = ICNSS_FTM;
else if (WLAN_IS_EPPING_ENABLED(con_mode))
mode = ICNSS_EPPING;
@@ -1863,7 +1868,7 @@ int hif_config_ce(struct hif_softc *scn)
#ifdef ADRASTEA_SHADOW_REGISTERS
int i;
#endif
- CDF_STATUS rv = CDF_STATUS_SUCCESS;
+ QDF_STATUS rv = QDF_STATUS_SUCCESS;
int ret;
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(scn);
struct hif_opaque_softc *hif_hdl = GET_HIF_OPAQUE_HDL(scn);
@@ -1888,31 +1893,31 @@ int hif_config_ce(struct hif_softc *scn)
if (ret) {
HIF_ERROR("%s: hif_wlan_enable error = %d", __func__, ret);
- return CDF_STATUS_NOT_INITIALIZED;
+ return QDF_STATUS_NOT_INITIALIZED;
}
scn->notice_send = true;
- cdf_mem_zero(&soc_info, sizeof(soc_info));
+ qdf_mem_zero(&soc_info, sizeof(soc_info));
ret = icnss_get_soc_info(scn, &soc_info);
if (ret < 0) {
HIF_ERROR("%s: icnss_get_soc_info error = %d", __func__, ret);
- return CDF_STATUS_NOT_INITIALIZED;
+ return QDF_STATUS_NOT_INITIALIZED;
}
scn->mem = soc_info.v_addr;
scn->mem_pa = soc_info.p_addr;
tgt_info->soc_version = soc_info.version;
- cdf_spinlock_init(&hif_state->keep_awake_lock);
+ qdf_spinlock_create(&hif_state->keep_awake_lock);
hif_state->keep_awake_count = 0;
hif_state->fake_sleep = false;
hif_state->sleep_ticks = 0;
- cdf_softirq_timer_init(NULL, &hif_state->sleep_timer,
+ qdf_timer_init(NULL, &hif_state->sleep_timer,
hif_sleep_entry, (void *)hif_state,
- CDF_TIMER_TYPE_WAKE_APPS);
+ QDF_TIMER_TYPE_WAKE_APPS);
hif_state->sleep_timer_init = true;
hif_state->fw_indicator_address = FW_INDICATOR_ADDRESS;
#ifdef HIF_PCI
@@ -1936,9 +1941,9 @@ int hif_config_ce(struct hif_softc *scn)
pipe_info->HIF_CE_state = hif_state;
attr = &host_ce_config[pipe_num];
pipe_info->ce_hdl = ce_init(scn, pipe_num, attr);
- CDF_ASSERT(pipe_info->ce_hdl != NULL);
+ QDF_ASSERT(pipe_info->ce_hdl != NULL);
if (pipe_info->ce_hdl == NULL) {
- rv = CDF_STATUS_E_FAILURE;
+ rv = QDF_STATUS_E_FAILURE;
A_TARGET_ACCESS_UNLIKELY(scn);
goto err;
}
@@ -1951,8 +1956,8 @@ int hif_config_ce(struct hif_softc *scn)
continue;
}
- pipe_info->buf_sz = (cdf_size_t) (attr->src_sz_max);
- cdf_spinlock_init(&pipe_info->recv_bufs_needed_lock);
+ pipe_info->buf_sz = (qdf_size_t) (attr->src_sz_max);
+ qdf_spinlock_create(&pipe_info->recv_bufs_needed_lock);
if (attr->dest_nentries > 0) {
atomic_set(&pipe_info->recv_bufs_needed,
init_buffer_count(attr->dest_nentries - 1));
@@ -1992,7 +1997,7 @@ int hif_config_ce(struct hif_softc *scn)
A_TARGET_ACCESS_UNLIKELY(scn);
- if (rv != CDF_STATUS_SUCCESS)
+ if (rv != QDF_STATUS_SUCCESS)
goto err;
else
init_tasklet_workers(hif_hdl);
@@ -2009,7 +2014,7 @@ int hif_config_ce(struct hif_softc *scn)
#endif
- return rv != CDF_STATUS_SUCCESS;
+ return rv != QDF_STATUS_SUCCESS;
err:
/* Failure, so clean up */
@@ -2024,15 +2029,15 @@ err:
}
}
if (hif_state->sleep_timer_init) {
- cdf_softirq_timer_cancel(&hif_state->sleep_timer);
- cdf_softirq_timer_free(&hif_state->sleep_timer);
+ qdf_timer_stop(&hif_state->sleep_timer);
+ qdf_timer_free(&hif_state->sleep_timer);
hif_state->sleep_timer_init = false;
}
athdiag_procfs_remove();
scn->athdiag_procfs_inited = false;
HIF_TRACE("%s: X, ret = %d\n", __func__, rv);
- return CDF_STATUS_SUCCESS != CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_SUCCESS != QDF_STATUS_E_FAILURE;
}
@@ -2055,9 +2060,9 @@ err:
* Return: None
*/
void hif_ipa_get_ce_resource(struct hif_opaque_softc *hif_ctx,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr)
+ qdf_dma_addr_t *ce_reg_paddr)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(scn);
@@ -2164,7 +2169,7 @@ u32 shadow_sr_wr_ind_addr(struct hif_softc *scn, u32 ctrl_addr)
break;
default:
HIF_ERROR("invalid CE ctrl_addr\n");
- CDF_ASSERT(0);
+ QDF_ASSERT(0);
}
return addr;
@@ -2190,7 +2195,7 @@ u32 shadow_dst_wr_ind_addr(struct hif_softc *scn, u32 ctrl_addr)
break;
default:
HIF_ERROR("invalid CE ctrl_addr\n");
- CDF_ASSERT(0);
+ QDF_ASSERT(0);
}
return addr;
@@ -2217,9 +2222,9 @@ void ce_lro_flush_cb_register(struct hif_opaque_softc *hif_hdl,
int ul_polled, dl_polled;
struct hif_softc *scn = HIF_GET_SOFTC(hif_hdl);
- CDF_ASSERT(scn != NULL);
+ QDF_ASSERT(scn != NULL);
- if (CDF_STATUS_SUCCESS !=
+ if (QDF_STATUS_SUCCESS !=
hif_map_service_to_pipe(hif_hdl, HTT_DATA_MSG_SVC,
&ul, &dl, &ul_polled, &dl_polled)) {
printk("%s cannot map service to pipe\n", __FUNCTION__);
@@ -2247,9 +2252,9 @@ void ce_lro_flush_cb_deregister(struct hif_opaque_softc *hif_hdl)
int ul_polled, dl_polled;
struct hif_softc *scn = HIF_GET_SOFTC(hif_hdl);
- CDF_ASSERT(scn != NULL);
+ QDF_ASSERT(scn != NULL);
- if (CDF_STATUS_SUCCESS !=
+ if (QDF_STATUS_SUCCESS !=
hif_map_service_to_pipe(hif_hdl, HTT_DATA_MSG_SVC,
&ul, &dl, &ul_polled, &dl_polled)) {
printk("%s cannot map service to pipe\n", __FUNCTION__);
@@ -2284,7 +2289,7 @@ int hif_map_service_to_pipe(struct hif_opaque_softc *hif_hdl, uint16_t svc_id,
uint8_t *ul_pipe, uint8_t *dl_pipe, int *ul_is_polled,
int *dl_is_polled)
{
- int status = CDF_STATUS_SUCCESS;
+ int status = QDF_STATUS_SUCCESS;
unsigned int i;
struct service_to_pipe element;
struct service_to_pipe *tgt_svc_map_to_use;
@@ -2336,7 +2341,7 @@ inline uint32_t DEBUG_CE_SRC_RING_READ_IDX_GET(struct hif_softc *scn,
HIF_ERROR("error: read from ddr = %d actual read from register = %d, CE_MISC_INT_STATUS_GET = 0x%x\n",
srri_from_ddr, read_from_hw,
CE_MISC_INT_STATUS_GET(scn, CE_ctrl_addr));
- CDF_ASSERT(0);
+ QDF_ASSERT(0);
}
return srri_from_ddr;
}
@@ -2355,7 +2360,7 @@ inline uint32_t DEBUG_CE_DEST_RING_READ_IDX_GET(struct hif_softc *scn,
HIF_ERROR("error: read from ddr = %d actual read from register = %d, CE_MISC_INT_STATUS_GET = 0x%x\n",
drri_from_ddr, read_from_hw,
CE_MISC_INT_STATUS_GET(scn, CE_ctrl_addr));
- CDF_ASSERT(0);
+ QDF_ASSERT(0);
}
return drri_from_ddr;
}
@@ -2426,11 +2431,12 @@ inline unsigned int hif_get_dst_ring_read_index(struct hif_softc *scn,
static inline void hif_config_rri_on_ddr(struct hif_softc *scn)
{
unsigned int i;
- cdf_dma_addr_t paddr_rri_on_ddr;
+ qdf_dma_addr_t paddr_rri_on_ddr;
uint32_t high_paddr, low_paddr;
scn->vaddr_rri_on_ddr =
- (uint32_t *)cdf_os_mem_alloc_consistent(scn->cdf_dev,
- (CE_COUNT*sizeof(uint32_t)), &paddr_rri_on_ddr, 0);
+ (uint32_t *)qdf_mem_alloc_consistent(scn->qdf_dev,
+ scn->qdf_dev->dev, (CE_COUNT*sizeof(uint32_t)),
+ &paddr_rri_on_ddr);
low_paddr = BITS0_TO_31(paddr_rri_on_ddr);
high_paddr = BITS32_TO_35(paddr_rri_on_ddr);
@@ -2443,7 +2449,7 @@ static inline void hif_config_rri_on_ddr(struct hif_softc *scn)
for (i = 0; i < CE_COUNT; i++)
CE_IDX_UPD_EN_SET(scn, CE_BASE_ADDRESS(i));
- cdf_mem_zero(scn->vaddr_rri_on_ddr, CE_COUNT*sizeof(uint32_t));
+ qdf_mem_zero(scn->vaddr_rri_on_ddr, CE_COUNT*sizeof(uint32_t));
return;
}
@@ -2480,19 +2486,19 @@ int hif_dump_ce_registers(struct hif_softc *scn)
uint32_t ce_reg_values[CE_COUNT_MAX][CE_USEFUL_SIZE >> 2];
uint32_t ce_reg_word_size = CE_USEFUL_SIZE >> 2;
uint16_t i;
- CDF_STATUS status;
+ QDF_STATUS status;
for (i = 0; i < CE_COUNT_MAX; i++, ce_reg_address += CE_OFFSET) {
status = hif_diag_read_mem(hif_hdl, ce_reg_address,
(uint8_t *) &ce_reg_values[i][0],
ce_reg_word_size * sizeof(uint32_t));
- if (status != CDF_STATUS_SUCCESS) {
+ if (status != QDF_STATUS_SUCCESS) {
HIF_ERROR("Dumping CE register failed!");
return -EACCES;
}
HIF_ERROR("CE%d Registers:", i);
- cdf_trace_hex_dump(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_DEBUG,
+ qdf_trace_hex_dump(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_DEBUG,
(uint8_t *) &ce_reg_values[i][0],
ce_reg_word_size * sizeof(uint32_t));
}
diff --git a/hif/src/ce/ce_main.h b/hif/src/ce/ce_main.h
index a51e82c55712..c0af85045dbe 100644
--- a/hif/src/ce/ce_main.h
+++ b/hif/src/ce/ce_main.h
@@ -28,8 +28,8 @@
#ifndef __CE_H__
#define __CE_H__
-#include "cdf_atomic.h"
-#include "cdf_lock.h"
+#include "qdf_atomic.h"
+#include "qdf_lock.h"
#include "hif_main.h"
#define CE_HTT_T2H_MSG 1
@@ -78,10 +78,10 @@ struct HIF_CE_pipe_info {
/* Instantaneous number of receive buffers that should be posted */
atomic_t recv_bufs_needed;
- cdf_size_t buf_sz;
- cdf_spinlock_t recv_bufs_needed_lock;
+ qdf_size_t buf_sz;
+ qdf_spinlock_t recv_bufs_needed_lock;
- cdf_spinlock_t completion_freeq_lock;
+ qdf_spinlock_t completion_freeq_lock;
/* Limit the number of outstanding send requests. */
int num_sends_allowed;
@@ -111,11 +111,11 @@ struct HIF_CE_state {
struct hif_softc ol_sc;
bool started;
struct ce_tasklet_entry tasklets[CE_COUNT_MAX];
- cdf_spinlock_t keep_awake_lock;
+ qdf_spinlock_t keep_awake_lock;
unsigned int keep_awake_count;
bool verified_awake;
bool fake_sleep;
- cdf_softirq_timer_t sleep_timer;
+ qdf_timer_t sleep_timer;
bool sleep_timer_init;
unsigned long sleep_ticks;
diff --git a/hif/src/ce/ce_service.c b/hif/src/ce/ce_service.c
index 9f90d5b8235e..3c3457591b39 100644
--- a/hif/src/ce/ce_service.c
+++ b/hif/src/ce/ce_service.c
@@ -35,7 +35,7 @@
#include "ce_main.h"
#include "ce_internal.h"
#include "ce_reg.h"
-#include "cdf_lock.h"
+#include "qdf_lock.h"
#include "regtable.h"
#include "epping_main.h"
#include "hif_main.h"
@@ -97,7 +97,7 @@ struct hif_ce_desc_event {
/* max history to record per copy engine */
#define HIF_CE_HISTORY_MAX 512
-cdf_atomic_t hif_ce_desc_history_index[CE_COUNT_MAX];
+qdf_atomic_t hif_ce_desc_history_index[CE_COUNT_MAX];
struct hif_ce_desc_event hif_ce_desc_history[CE_COUNT_MAX][HIF_CE_HISTORY_MAX];
@@ -113,11 +113,11 @@ struct hif_ce_desc_event hif_ce_desc_history[CE_COUNT_MAX][HIF_CE_HISTORY_MAX];
* trying to access the array, full locking of the recording process would
* be needed to have sane logging.
*/
-static int get_next_record_index(cdf_atomic_t *table_index, int array_size)
+static int get_next_record_index(qdf_atomic_t *table_index, int array_size)
{
- int record_index = cdf_atomic_inc_return(table_index);
+ int record_index = qdf_atomic_inc_return(table_index);
if (record_index == array_size)
- cdf_atomic_sub(array_size, table_index);
+ qdf_atomic_sub(array_size, table_index);
while (record_index >= array_size)
record_index -= array_size;
@@ -149,8 +149,8 @@ void hif_record_ce_desc_event(struct hif_softc *scn, int ce_id,
if (cbk && cbk->get_monotonic_boottime)
event->time = cbk->get_monotonic_boottime();
else
- event->time = ((uint64_t)cdf_system_ticks_to_msecs(
- cdf_system_ticks()) * 1000);
+ event->time = ((uint64_t)qdf_system_ticks_to_msecs(
+ qdf_system_ticks()) * 1000);
if (descriptor != NULL)
event->descriptor = *descriptor;
@@ -169,7 +169,7 @@ void hif_record_ce_desc_event(struct hif_softc *scn, int ce_id,
*/
void ce_init_ce_desc_event_log(int ce_id, int size)
{
- cdf_atomic_init(&hif_ce_desc_history_index[ce_id]);
+ qdf_atomic_init(&hif_ce_desc_history_index[ce_id]);
}
#else
void hif_record_ce_desc_event(struct hif_softc *scn,
@@ -232,7 +232,7 @@ int
ce_completed_send_next_nolock(struct CE_state *CE_state,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *sw_idx, unsigned int *hw_idx,
@@ -275,7 +275,7 @@ void war_ce_src_ring_write_idx_set(struct hif_softc *scn,
int
ce_send_nolock(struct CE_handle *copyeng,
void *per_transfer_context,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
uint32_t nbytes,
uint32_t transfer_id,
uint32_t flags,
@@ -295,7 +295,7 @@ ce_send_nolock(struct CE_handle *copyeng,
if (unlikely(CE_RING_DELTA(nentries_mask,
write_index, sw_index - 1) <= 0)) {
OL_ATH_CE_PKT_ERROR_COUNT_INCR(scn, CE_RING_DELTA_FAIL);
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
A_TARGET_ACCESS_END_RET(scn);
return status;
}
@@ -357,7 +357,7 @@ ce_send_nolock(struct CE_handle *copyeng,
src_ring->write_index);
src_ring->write_index = write_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
}
A_TARGET_ACCESS_END_RET(scn);
@@ -367,7 +367,7 @@ ce_send_nolock(struct CE_handle *copyeng,
int
ce_send(struct CE_handle *copyeng,
void *per_transfer_context,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
uint32_t nbytes,
uint32_t transfer_id,
uint32_t flags,
@@ -376,10 +376,10 @@ ce_send(struct CE_handle *copyeng,
struct CE_state *CE_state = (struct CE_state *)copyeng;
int status;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
status = ce_send_nolock(copyeng, per_transfer_context, buffer, nbytes,
transfer_id, flags, user_flag);
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -397,7 +397,7 @@ void ce_sendlist_init(struct ce_sendlist *sendlist)
int
ce_sendlist_buf_add(struct ce_sendlist *sendlist,
- cdf_dma_addr_t buffer,
+ qdf_dma_addr_t buffer,
uint32_t nbytes,
uint32_t flags,
uint32_t user_flags)
@@ -407,8 +407,8 @@ ce_sendlist_buf_add(struct ce_sendlist *sendlist,
struct ce_sendlist_item *item;
if (num_items >= CE_SENDLIST_ITEMS_MAX) {
- CDF_ASSERT(num_items < CE_SENDLIST_ITEMS_MAX);
- return CDF_STATUS_E_RESOURCES;
+ QDF_ASSERT(num_items < CE_SENDLIST_ITEMS_MAX);
+ return QDF_STATUS_E_RESOURCES;
}
item = &sl->item[num_items];
@@ -418,7 +418,7 @@ ce_sendlist_buf_add(struct ce_sendlist *sendlist,
item->flags = flags;
item->user_flags = user_flags;
sl->num_items = num_items + 1;
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
int
@@ -435,9 +435,9 @@ ce_sendlist_send(struct CE_handle *copyeng,
unsigned int sw_index;
unsigned int write_index;
- CDF_ASSERT((num_items > 0) && (num_items < src_ring->nentries));
+ QDF_ASSERT((num_items > 0) && (num_items < src_ring->nentries));
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
@@ -450,28 +450,28 @@ ce_sendlist_send(struct CE_handle *copyeng,
for (i = 0; i < num_items - 1; i++) {
item = &sl->item[i];
/* TBDXXX: Support extensible sendlist_types? */
- CDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
+ QDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
status = ce_send_nolock(copyeng, CE_SENDLIST_ITEM_CTXT,
- (cdf_dma_addr_t) item->data,
+ (qdf_dma_addr_t) item->data,
item->u.nbytes, transfer_id,
item->flags | CE_SEND_FLAG_GATHER,
item->user_flags);
- CDF_ASSERT(status == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(status == QDF_STATUS_SUCCESS);
}
/* provide valid context pointer for final item */
item = &sl->item[i];
/* TBDXXX: Support extensible sendlist_types? */
- CDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
+ QDF_ASSERT(item->send_type == CE_SIMPLE_BUFFER_TYPE);
status = ce_send_nolock(copyeng, per_transfer_context,
- (cdf_dma_addr_t) item->data,
+ (qdf_dma_addr_t) item->data,
item->u.nbytes,
transfer_id, item->flags,
item->user_flags);
- CDF_ASSERT(status == CDF_STATUS_SUCCESS);
+ QDF_ASSERT(status == QDF_STATUS_SUCCESS);
NBUF_UPDATE_TX_PKT_COUNT((cdf_nbuf_t)per_transfer_context,
NBUF_TX_PKT_CE);
- DPTRACE(cdf_dp_trace((cdf_nbuf_t)per_transfer_context,
- CDF_DP_TRACE_CE_PACKET_PTR_RECORD,
+ DPTRACE(qdf_dp_trace((cdf_nbuf_t)per_transfer_context,
+ QDF_DP_TRACE_CE_PACKET_PTR_RECORD,
(uint8_t *)(((cdf_nbuf_t)per_transfer_context)->data),
sizeof(((cdf_nbuf_t)per_transfer_context)->data)));
} else {
@@ -482,7 +482,7 @@ ce_sendlist_send(struct CE_handle *copyeng,
* the entire request at once, punt it back to the caller.
*/
}
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -543,7 +543,7 @@ int ce_send_fast(struct CE_handle *copyeng, cdf_nbuf_t *msdus,
uint64_t dma_addr;
uint32_t user_flags = 0;
- cdf_spin_lock_bh(&ce_state->ce_index_lock);
+ qdf_spin_lock_bh(&ce_state->ce_index_lock);
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
@@ -606,7 +606,7 @@ int ce_send_fast(struct CE_handle *copyeng, cdf_nbuf_t *msdus,
/*
* Clear packet offset for all but the first CE desc.
*/
- user_flags &= ~CDF_CE_TX_PKT_OFFSET_BIT_M;
+ user_flags &= ~QDF_CE_TX_PKT_OFFSET_BIT_M;
ce_buffer_addr_hi_set(shadow_src_desc, dma_addr, user_flags);
shadow_src_desc->meta_data = transfer_id;
@@ -641,7 +641,7 @@ int ce_send_fast(struct CE_handle *copyeng, cdf_nbuf_t *msdus,
}
}
- cdf_spin_unlock_bh(&ce_state->ce_index_lock);
+ qdf_spin_unlock_bh(&ce_state->ce_index_lock);
/*
* If all packets in the array are transmitted,
@@ -655,7 +655,7 @@ int ce_send_fast(struct CE_handle *copyeng, cdf_nbuf_t *msdus,
int
ce_recv_buf_enqueue(struct CE_handle *copyeng,
- void *per_recv_context, cdf_dma_addr_t buffer)
+ void *per_recv_context, qdf_dma_addr_t buffer)
{
int status;
struct CE_state *CE_state = (struct CE_state *)copyeng;
@@ -668,13 +668,13 @@ ce_recv_buf_enqueue(struct CE_handle *copyeng,
uint64_t dma_addr = buffer;
struct hif_softc *scn = CE_state->scn;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
write_index = dest_ring->write_index;
sw_index = dest_ring->sw_index;
A_TARGET_ACCESS_BEGIN_RET_EXT(scn, val);
if (val == -1) {
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return val;
}
@@ -704,17 +704,17 @@ ce_recv_buf_enqueue(struct CE_handle *copyeng,
write_index = CE_RING_IDX_INCR(nentries_mask, write_index);
CE_DEST_RING_WRITE_IDX_SET(scn, ctrl_addr, write_index);
dest_ring->write_index = write_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
} else {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
}
A_TARGET_ACCESS_END_RET_EXT(scn, val);
if (val == -1) {
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return val;
}
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -755,10 +755,10 @@ unsigned int ce_send_entries_avail(struct CE_handle *copyeng)
unsigned int sw_index;
unsigned int write_index;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return CE_RING_DELTA(nentries_mask, write_index, sw_index - 1);
}
@@ -771,10 +771,10 @@ unsigned int ce_recv_entries_avail(struct CE_handle *copyeng)
unsigned int sw_index;
unsigned int write_index;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
sw_index = dest_ring->sw_index;
write_index = dest_ring->write_index;
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return CE_RING_DELTA(nentries_mask, write_index, sw_index - 1);
}
@@ -804,9 +804,9 @@ unsigned int ce_send_entries_done(struct CE_handle *copyeng)
struct CE_state *CE_state = (struct CE_state *)copyeng;
unsigned int nentries;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries = ce_send_entries_done_nolock(CE_state->scn, CE_state);
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return nentries;
}
@@ -836,9 +836,9 @@ unsigned int ce_recv_entries_done(struct CE_handle *copyeng)
struct CE_state *CE_state = (struct CE_state *)copyeng;
unsigned int nentries;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries = ce_recv_entries_done_nolock(CE_state->scn, CE_state);
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return nentries;
}
@@ -857,7 +857,7 @@ int
ce_completed_recv_next_nolock(struct CE_state *CE_state,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *flagsp)
@@ -886,7 +886,7 @@ ce_completed_recv_next_nolock(struct CE_state *CE_state,
* corresponding descriptor has completed. We treat this
* as a descriptor that is not yet done.
*/
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
goto done;
}
@@ -916,7 +916,7 @@ ce_completed_recv_next_nolock(struct CE_state *CE_state,
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
dest_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
done:
return status;
@@ -926,45 +926,45 @@ int
ce_completed_recv_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp, unsigned int *flagsp)
{
struct CE_state *CE_state = (struct CE_state *)copyeng;
int status;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
status =
ce_completed_recv_next_nolock(CE_state, per_CE_contextp,
per_transfer_contextp, bufferp,
nbytesp, transfer_idp, flagsp);
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
/* NB: Modeled after ce_completed_recv_next_nolock */
-CDF_STATUS
+QDF_STATUS
ce_revoke_recv_next(struct CE_handle *copyeng,
void **per_CE_contextp,
- void **per_transfer_contextp, cdf_dma_addr_t *bufferp)
+ void **per_transfer_contextp, qdf_dma_addr_t *bufferp)
{
struct CE_state *CE_state;
struct CE_ring_state *dest_ring;
unsigned int nentries_mask;
unsigned int sw_index;
unsigned int write_index;
- CDF_STATUS status;
+ QDF_STATUS status;
struct hif_softc *scn;
CE_state = (struct CE_state *)copyeng;
dest_ring = CE_state->dest_ring;
if (!dest_ring) {
- return CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_E_FAILURE;
}
scn = CE_state->scn;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries_mask = dest_ring->nentries_mask;
sw_index = dest_ring->sw_index;
write_index = dest_ring->write_index;
@@ -992,11 +992,11 @@ ce_revoke_recv_next(struct CE_handle *copyeng,
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
dest_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
} else {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
}
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return status;
}
@@ -1009,14 +1009,14 @@ int
ce_completed_send_next_nolock(struct CE_state *CE_state,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *sw_idx,
unsigned int *hw_idx,
uint32_t *toeplitz_hash_result)
{
- int status = CDF_STATUS_E_FAILURE;
+ int status = QDF_STATUS_E_FAILURE;
struct CE_ring_state *src_ring = CE_state->src_ring;
uint32_t ctrl_addr = CE_state->ctrl_addr;
unsigned int nentries_mask = src_ring->nentries_mask;
@@ -1085,18 +1085,18 @@ ce_completed_send_next_nolock(struct CE_state *CE_state,
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
src_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
}
return status;
}
/* NB: Modeled after ce_completed_send_next */
-CDF_STATUS
+QDF_STATUS
ce_cancel_send_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
uint32_t *toeplitz_hash_result)
@@ -1106,17 +1106,17 @@ ce_cancel_send_next(struct CE_handle *copyeng,
unsigned int nentries_mask;
unsigned int sw_index;
unsigned int write_index;
- CDF_STATUS status;
+ QDF_STATUS status;
struct hif_softc *scn;
CE_state = (struct CE_state *)copyeng;
src_ring = CE_state->src_ring;
if (!src_ring) {
- return CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_E_FAILURE;
}
scn = CE_state->scn;
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
nentries_mask = src_ring->nentries_mask;
sw_index = src_ring->sw_index;
write_index = src_ring->write_index;
@@ -1151,11 +1151,11 @@ ce_cancel_send_next(struct CE_handle *copyeng,
/* Update sw_index */
sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
src_ring->sw_index = sw_index;
- status = CDF_STATUS_SUCCESS;
+ status = QDF_STATUS_SUCCESS;
} else {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
}
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
return status;
}
@@ -1167,7 +1167,7 @@ int
ce_completed_send_next(struct CE_handle *copyeng,
void **per_CE_contextp,
void **per_transfer_contextp,
- cdf_dma_addr_t *bufferp,
+ qdf_dma_addr_t *bufferp,
unsigned int *nbytesp,
unsigned int *transfer_idp,
unsigned int *sw_idx,
@@ -1177,13 +1177,13 @@ ce_completed_send_next(struct CE_handle *copyeng,
struct CE_state *CE_state = (struct CE_state *)copyeng;
int status;
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
status =
ce_completed_send_next_nolock(CE_state, per_CE_contextp,
per_transfer_contextp, bufferp,
nbytesp, transfer_idp, sw_idx,
hw_idx, toeplitz_hash_result);
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
return status;
}
@@ -1202,7 +1202,7 @@ void ce_per_engine_servicereap(struct hif_softc *scn, unsigned int ce_id)
{
void *CE_context;
void *transfer_context;
- cdf_dma_addr_t buf;
+ qdf_dma_addr_t buf;
unsigned int nbytes;
unsigned int id;
unsigned int sw_idx, hw_idx;
@@ -1226,7 +1226,7 @@ void ce_per_engine_servicereap(struct hif_softc *scn, unsigned int ce_id)
* addressed by change spin_lock to spin_lock_bh also.
*/
- cdf_spin_lock_bh(&CE_state->ce_index_lock);
+ qdf_spin_lock_bh(&CE_state->ce_index_lock);
if (CE_state->send_cb) {
{
@@ -1238,9 +1238,9 @@ void ce_per_engine_servicereap(struct hif_softc *scn, unsigned int ce_id)
&transfer_context, &buf,
&nbytes, &id, &sw_idx, &hw_idx,
&toeplitz_hash_result) ==
- CDF_STATUS_SUCCESS) {
+ QDF_STATUS_SUCCESS) {
if (ce_id != CE_HTT_H2T_MSG) {
- cdf_spin_unlock_bh(
+ qdf_spin_unlock_bh(
&CE_state->ce_index_lock);
CE_state->send_cb(
(struct CE_handle *)
@@ -1248,24 +1248,24 @@ void ce_per_engine_servicereap(struct hif_softc *scn, unsigned int ce_id)
transfer_context, buf,
nbytes, id, sw_idx, hw_idx,
toeplitz_hash_result);
- cdf_spin_lock_bh(
+ qdf_spin_lock_bh(
&CE_state->ce_index_lock);
} else {
struct HIF_CE_pipe_info *pipe_info =
(struct HIF_CE_pipe_info *)
CE_context;
- cdf_spin_lock_bh(&pipe_info->
+ qdf_spin_lock_bh(&pipe_info->
completion_freeq_lock);
pipe_info->num_sends_allowed++;
- cdf_spin_unlock_bh(&pipe_info->
+ qdf_spin_unlock_bh(&pipe_info->
completion_freeq_lock);
}
}
}
}
- cdf_spin_unlock_bh(&CE_state->ce_index_lock);
+ qdf_spin_unlock_bh(&CE_state->ce_index_lock);
hif_record_ce_desc_event(scn, ce_id, HIF_CE_REAP_EXIT,
NULL, NULL, 0);
@@ -1299,7 +1299,7 @@ int ce_per_engine_service(struct hif_softc *scn, unsigned int CE_id)
uint32_t ctrl_addr = CE_state->ctrl_addr;
void *CE_context;
void *transfer_context;
- cdf_dma_addr_t buf;
+ qdf_dma_addr_t buf;
unsigned int nbytes;
unsigned int id;
unsigned int flags;
@@ -1315,7 +1315,7 @@ int ce_per_engine_service(struct hif_softc *scn, unsigned int CE_id)
return 0; /* no work done */
}
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
/* Clear force_break flag and re-initialize receive_count to 0 */
@@ -1331,8 +1331,8 @@ more_completions:
while (ce_completed_recv_next_nolock
(CE_state, &CE_context, &transfer_context,
&buf, &nbytes, &id, &flags) ==
- CDF_STATUS_SUCCESS) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ QDF_STATUS_SUCCESS) {
+ qdf_spin_unlock(&CE_state->ce_index_lock);
CE_state->recv_cb((struct CE_handle *)CE_state,
CE_context, transfer_context, buf,
nbytes, id, flags);
@@ -1354,8 +1354,8 @@ more_completions:
/* Break the receive processes by
* force if force_break set up
*/
- if (cdf_unlikely(CE_state->force_break)) {
- cdf_atomic_set(&CE_state->rx_pending, 1);
+ if (qdf_unlikely(CE_state->force_break)) {
+ qdf_atomic_set(&CE_state->rx_pending, 1);
CE_ENGINE_INT_STATUS_CLEAR(scn, ctrl_addr,
HOST_IS_COPY_COMPLETE_MASK);
if (Q_TARGET_ACCESS_END(scn) < 0)
@@ -1363,7 +1363,7 @@ more_completions:
CE_state->receive_count);
return CE_state->receive_count;
}
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
}
}
@@ -1383,24 +1383,24 @@ more_completions:
(CE_state, &CE_context,
&transfer_context, &buf, &nbytes,
&id, &sw_idx, &hw_idx,
- &toeplitz_hash_result) == CDF_STATUS_SUCCESS) {
+ &toeplitz_hash_result) == QDF_STATUS_SUCCESS) {
if (CE_id != CE_HTT_H2T_MSG ||
WLAN_IS_EPPING_ENABLED(mode)) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
CE_state->send_cb((struct CE_handle *)CE_state,
CE_context, transfer_context,
buf, nbytes, id, sw_idx,
hw_idx, toeplitz_hash_result);
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
} else {
struct HIF_CE_pipe_info *pipe_info =
(struct HIF_CE_pipe_info *)CE_context;
- cdf_spin_lock(&pipe_info->
+ qdf_spin_lock(&pipe_info->
completion_freeq_lock);
pipe_info->num_sends_allowed++;
- cdf_spin_unlock(&pipe_info->
+ qdf_spin_unlock(&pipe_info->
completion_freeq_lock);
}
}
@@ -1409,13 +1409,13 @@ more_completions:
(CE_state, &CE_context,
&transfer_context, &buf, &nbytes,
&id, &sw_idx, &hw_idx,
- &toeplitz_hash_result) == CDF_STATUS_SUCCESS) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ &toeplitz_hash_result) == QDF_STATUS_SUCCESS) {
+ qdf_spin_unlock(&CE_state->ce_index_lock);
CE_state->send_cb((struct CE_handle *)CE_state,
CE_context, transfer_context, buf,
nbytes, id, sw_idx, hw_idx,
toeplitz_hash_result);
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
}
#endif /*ATH_11AC_TXCOMPACT */
}
@@ -1425,7 +1425,7 @@ more_watermarks:
CE_int_status = CE_ENGINE_INT_STATUS_GET(scn, ctrl_addr);
if (CE_int_status & CE_WATERMARK_MASK) {
if (CE_state->watermark_cb) {
- cdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
/* Convert HW IS bits to software flags */
flags =
(CE_int_status & CE_WATERMARK_MASK) >>
@@ -1434,7 +1434,7 @@ more_watermarks:
CE_state->
watermark_cb((struct CE_handle *)CE_state,
CE_state->wm_context, flags);
- cdf_spin_lock(&CE_state->ce_index_lock);
+ qdf_spin_lock(&CE_state->ce_index_lock);
}
}
}
@@ -1494,8 +1494,8 @@ more_watermarks:
}
}
- cdf_spin_unlock(&CE_state->ce_index_lock);
- cdf_atomic_set(&CE_state->rx_pending, 0);
+ qdf_spin_unlock(&CE_state->ce_index_lock);
+ qdf_atomic_set(&CE_state->rx_pending, 0);
if (Q_TARGET_ACCESS_END(scn) < 0)
HIF_ERROR("<--[premature rc=%d]\n", CE_state->receive_count);
@@ -1514,11 +1514,11 @@ void ce_per_engine_service_any(int irq, struct hif_softc *scn)
uint32_t intr_summary;
A_TARGET_ACCESS_BEGIN(scn);
- if (!cdf_atomic_read(&scn->tasklet_from_intr)) {
+ if (!qdf_atomic_read(&scn->tasklet_from_intr)) {
for (CE_id = 0; CE_id < scn->ce_count; CE_id++) {
struct CE_state *CE_state = scn->ce_id_to_state[CE_id];
- if (cdf_atomic_read(&CE_state->rx_pending)) {
- cdf_atomic_set(&CE_state->rx_pending, 0);
+ if (qdf_atomic_read(&CE_state->rx_pending)) {
+ qdf_atomic_set(&CE_state->rx_pending, 0);
ce_per_engine_service(scn, CE_id);
}
}
@@ -1718,11 +1718,11 @@ void ce_pkt_dl_len_set(void *hif_sc, u_int32_t pkt_download_len)
struct hif_softc *sc = (struct hif_softc *)(hif_sc);
struct CE_state *ce_state = sc->ce_id_to_state[CE_HTT_H2T_MSG];
- cdf_assert_always(ce_state);
+ qdf_assert_always(ce_state);
ce_state->download_len = pkt_download_len;
- cdf_print("%s CE %d Pkt download length %d", __func__,
+ qdf_print("%s CE %d Pkt download length %d", __func__,
ce_state->id, ce_state->download_len);
}
#else
@@ -1737,7 +1737,7 @@ bool ce_get_rx_pending(struct hif_softc *scn)
for (CE_id = 0; CE_id < scn->ce_count; CE_id++) {
struct CE_state *CE_state = scn->ce_id_to_state[CE_id];
- if (cdf_atomic_read(&CE_state->rx_pending))
+ if (qdf_atomic_read(&CE_state->rx_pending))
return true;
}
@@ -1754,7 +1754,7 @@ bool ce_get_rx_pending(struct hif_softc *scn)
bool ce_check_rx_pending(struct hif_softc *scn, int ce_id)
{
struct CE_state *CE_state = scn->ce_id_to_state[ce_id];
- if (cdf_atomic_read(&CE_state->rx_pending))
+ if (qdf_atomic_read(&CE_state->rx_pending))
return true;
else
return false;
@@ -1813,14 +1813,14 @@ void ce_enable_msi(struct hif_softc *scn, unsigned int CE_id,
* Return: None
*/
void ce_ipa_get_resource(struct CE_handle *ce,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr)
+ qdf_dma_addr_t *ce_reg_paddr)
{
struct CE_state *CE_state = (struct CE_state *)ce;
uint32_t ring_loop;
struct CE_src_desc *ce_desc;
- cdf_dma_addr_t phy_mem_base;
+ qdf_dma_addr_t phy_mem_base;
struct hif_softc *scn = CE_state->scn;
if (CE_RUNNING != CE_state->state) {
diff --git a/hif/src/ce/ce_tasklet.c b/hif/src/ce/ce_tasklet.c
index 460d0bb57ac7..7ba68ba2b46e 100644
--- a/hif/src/ce/ce_tasklet.c
+++ b/hif/src/ce/ce_tasklet.c
@@ -32,9 +32,9 @@
#include <linux/if_arp.h>
#include "a_types.h"
#include "athdefs.h"
-#include "cdf_lock.h"
-#include "cdf_types.h"
-#include "cdf_status.h"
+#include "qdf_lock.h"
+#include "qdf_types.h"
+#include "qdf_status.h"
#include "regtable.h"
#include "hif.h"
#include "hif_io32.h"
@@ -199,10 +199,10 @@ static void ce_tasklet(unsigned long data)
hif_record_ce_desc_event(scn, tasklet_entry->ce_id,
HIF_CE_TASKLET_ENTRY, NULL, NULL, 0);
- if (cdf_atomic_read(&scn->link_suspended)) {
+ if (qdf_atomic_read(&scn->link_suspended)) {
HIF_ERROR("%s: ce %d tasklet fired after link suspend.",
__func__, tasklet_entry->ce_id);
- CDF_BUG(0);
+ QDF_BUG(0);
}
ce_per_engine_service(scn, tasklet_entry->ce_id);
@@ -229,7 +229,7 @@ static void ce_tasklet(unsigned long data)
hif_record_ce_desc_event(scn, tasklet_entry->ce_id, HIF_CE_TASKLET_EXIT,
NULL, NULL, 0);
- cdf_atomic_dec(&scn->active_tasklet_cnt);
+ qdf_atomic_dec(&scn->active_tasklet_cnt);
}
/**
@@ -270,7 +270,7 @@ void ce_tasklet_kill(struct hif_softc *scn)
tasklet_kill(&hif_ce_state->tasklets[i].intr_tq);
hif_ce_state->tasklets[i].inited = false;
}
- cdf_atomic_set(&scn->active_tasklet_cnt, 0);
+ qdf_atomic_set(&scn->active_tasklet_cnt, 0);
}
/**
* ce_irq_handler() - ce_irq_handler
@@ -303,7 +303,7 @@ static irqreturn_t ce_irq_handler(int irq, void *context)
#endif
ce_irq_disable(scn, ce_id);
ce_irq_status(scn, ce_id, &host_status);
- cdf_atomic_inc(&scn->active_tasklet_cnt);
+ qdf_atomic_inc(&scn->active_tasklet_cnt);
hif_record_ce_desc_event(scn, ce_id, HIF_IRQ_EVENT, NULL, NULL, 0);
if (hif_napi_enabled(hif_hdl, ce_id))
hif_napi_schedule(hif_hdl, ce_id);
@@ -340,16 +340,16 @@ const char *ce_name[ICNSS_MAX_IRQ_REGISTRATIONS] = {
* Unregisters copy engine irqs matching mask. If a 1 is set at bit x,
* unregister for copy engine x.
*
- * Return: CDF_STATUS
+ * Return: QDF_STATUS
*/
-CDF_STATUS ce_unregister_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
+QDF_STATUS ce_unregister_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
{
int id;
int ret;
if (hif_ce_state == NULL) {
HIF_WARN("%s: hif_ce_state = NULL", __func__);
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
for (id = 0; id < CE_COUNT_MAX; id++) {
if ((mask & (1 << id)) && hif_ce_state->tasklets[id].inited) {
@@ -361,7 +361,7 @@ CDF_STATUS ce_unregister_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
__func__, id, ret);
}
}
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
/**
* ce_register_irq() - ce_register_irq
@@ -371,9 +371,9 @@ CDF_STATUS ce_unregister_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
* Registers copy engine irqs matching mask. If a 1 is set at bit x,
* Register for copy engine x.
*
- * Return: CDF_STATUS
+ * Return: QDF_STATUS
*/
-CDF_STATUS ce_register_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
+QDF_STATUS ce_register_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
{
int id;
int ret;
@@ -390,7 +390,7 @@ CDF_STATUS ce_register_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
"%s: cannot register CE %d irq handler, ret = %d",
__func__, id, ret);
ce_unregister_irq(hif_ce_state, done_mask);
- return CDF_STATUS_E_FAULT;
+ return QDF_STATUS_E_FAULT;
} else {
done_mask |= 1 << id;
}
@@ -402,5 +402,5 @@ CDF_STATUS ce_register_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask)
ce_enable_irq_in_group_reg(HIF_GET_SOFTC(hif_ce_state), done_mask);
#endif
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
diff --git a/hif/src/ce/ce_tasklet.h b/hif/src/ce/ce_tasklet.h
index 5187f68a0cf6..d20402e348eb 100644
--- a/hif/src/ce/ce_tasklet.h
+++ b/hif/src/ce/ce_tasklet.h
@@ -31,6 +31,6 @@
void init_tasklet_workers(struct hif_opaque_softc *scn);
void ce_tasklet_init(struct HIF_CE_state *hif_ce_state, uint32_t mask);
void ce_tasklet_kill(struct hif_softc *scn);
-CDF_STATUS ce_register_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask);
-CDF_STATUS ce_unregister_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask);
+QDF_STATUS ce_register_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask);
+QDF_STATUS ce_unregister_irq(struct HIF_CE_state *hif_ce_state, uint32_t mask);
#endif /* __CE_TASKLET_H__ */
diff --git a/hif/src/hif_debug.h b/hif/src/hif_debug.h
index 4ab7d1f8e05f..0b0a7b6d62b5 100644
--- a/hif/src/hif_debug.h
+++ b/hif/src/hif_debug.h
@@ -21,22 +21,22 @@
#ifndef __HIF_DEBUG_H__
#define __HIF_DEBUG_H__
-#include "cdf_trace.h"
+#include "qdf_trace.h"
#define HIF_ERROR(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_ERROR, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_ERROR, ## args)
#define HIF_WARN(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_WARN, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_WARN, ## args)
#define HIF_INFO(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_INFO, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_INFO, ## args)
#define HIF_INFO_HI(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_INFO_HIGH, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_INFO_HIGH, ## args)
#define HIF_INFO_MED(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_INFO_MED, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_INFO_MED, ## args)
#define HIF_INFO_LO(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_INFO_LOW, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_INFO_LOW, ## args)
#define HIF_TRACE(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_ERROR, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_ERROR, ## args)
#define HIF_DBG(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_DEBUG, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_DEBUG, ## args)
#endif /* __HIF_DEBUG_H__ */
diff --git a/hif/src/hif_main.c b/hif/src/hif_main.c
index 0c69dbfe73af..db8b6ad36ac7 100644
--- a/hif/src/hif_main.c
+++ b/hif/src/hif_main.c
@@ -30,9 +30,10 @@
#include "athdefs.h"
#include "osapi_linux.h"
#include "targcfg.h"
-#include "cdf_lock.h"
-#include "cdf_status.h"
-#include <cdf_atomic.h> /* cdf_atomic_read */
+#include "qdf_lock.h"
+#include "qdf_status.h"
+#include "qdf_status.h"
+#include <qdf_atomic.h> /* qdf_atomic_read */
#include <targaddrs.h>
#include <bmi_msg.h>
#include "hif_io32.h"
@@ -45,8 +46,8 @@
#include "hif_hw_version.h"
#include "ce_api.h"
#include "ce_tasklet.h"
-#include "cdf_trace.h"
-#include "cdf_status.h"
+#include "qdf_trace.h"
+#include "qdf_status.h"
#ifdef CONFIG_CNSS
#include <net/cnss.h>
#endif
@@ -193,7 +194,7 @@ static inline void hif_fw_event_handler(struct HIF_CE_state *hif_state)
return;
msg_callbacks->fwEventHandler(msg_callbacks->Context,
- CDF_STATUS_E_FAILURE);
+ QDF_STATUS_E_FAILURE);
}
/**
@@ -279,7 +280,7 @@ void *hif_get_targetdef(struct hif_opaque_softc *hif_ctx)
void hif_vote_link_down(struct hif_opaque_softc *hif_ctx)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
- CDF_BUG(scn);
+ QDF_BUG(scn);
scn->linkstate_vote--;
if (scn->linkstate_vote == 0)
@@ -300,7 +301,7 @@ void hif_vote_link_down(struct hif_opaque_softc *hif_ctx)
void hif_vote_link_up(struct hif_opaque_softc *hif_ctx)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
- CDF_BUG(scn);
+ QDF_BUG(scn);
scn->linkstate_vote++;
if (scn->linkstate_vote == 1)
@@ -322,7 +323,7 @@ void hif_vote_link_up(struct hif_opaque_softc *hif_ctx)
bool hif_can_suspend_link(struct hif_opaque_softc *hif_ctx)
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
- CDF_BUG(scn);
+ QDF_BUG(scn);
return scn->linkstate_vote == 0;
}
@@ -381,14 +382,14 @@ bool hif_max_num_receives_reached(struct hif_softc *scn, unsigned int count)
/**
* init_buffer_count() - initial buffer count
- * @maxSize: cdf_size_t
+ * @maxSize: qdf_size_t
*
* routine to modify the initial buffer count to be allocated on an os
* platform basis. Platform owner will need to modify this as needed
*
- * Return: cdf_size_t
+ * Return: qdf_size_t
*/
-cdf_size_t init_buffer_count(cdf_size_t maxSize)
+qdf_size_t init_buffer_count(qdf_size_t maxSize)
{
return maxSize;
}
@@ -453,7 +454,7 @@ void hif_get_hw_info(struct hif_opaque_softc *scn, u32 *version, u32 *revision,
/**
* hif_open(): hif_open
- * @cdf_ctx: CDF Context
+ * @qdf_ctx: QDF Context
* @mode: Driver Mode
* @bus_type: Bus Type
* @cbk: CDS Callbacks
@@ -462,35 +463,35 @@ void hif_get_hw_info(struct hif_opaque_softc *scn, u32 *version, u32 *revision,
*
* Return: HIF Opaque Pointer
*/
-struct hif_opaque_softc *hif_open(cdf_device_t cdf_ctx, uint32_t mode,
- enum ath_hal_bus_type bus_type,
+struct hif_opaque_softc *hif_open(qdf_device_t qdf_ctx, uint32_t mode,
+ enum qdf_bus_type bus_type,
struct hif_callbacks *cbk)
{
struct hif_softc *scn;
- CDF_STATUS status = CDF_STATUS_SUCCESS;
+ QDF_STATUS status = QDF_STATUS_SUCCESS;
int bus_context_size = hif_bus_get_context_size();
- scn = (struct hif_softc *)cdf_mem_malloc(bus_context_size);
+ scn = (struct hif_softc *)qdf_mem_malloc(bus_context_size);
if (!scn) {
HIF_ERROR("%s: cannot alloc memory for HIF context of size:%d",
__func__, bus_context_size);
return GET_HIF_OPAQUE_HDL(scn);
}
- cdf_mem_zero(scn, bus_context_size);
+ qdf_mem_zero(scn, bus_context_size);
- scn->cdf_dev = cdf_ctx;
+ scn->qdf_dev = qdf_ctx;
scn->hif_con_param = mode;
- cdf_atomic_init(&scn->active_tasklet_cnt);
- cdf_atomic_init(&scn->link_suspended);
- cdf_atomic_init(&scn->tasklet_from_intr);
- cdf_mem_copy(&scn->callbacks, cbk, sizeof(struct hif_callbacks));
+ qdf_atomic_init(&scn->active_tasklet_cnt);
+ qdf_atomic_init(&scn->link_suspended);
+ qdf_atomic_init(&scn->tasklet_from_intr);
+ qdf_mem_copy(&scn->callbacks, cbk, sizeof(struct hif_callbacks));
status = hif_bus_open(scn, bus_type);
- if (status != CDF_STATUS_SUCCESS) {
+ if (status != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: hif_bus_open error = %d, bus_type = %d",
__func__, status, bus_type);
- cdf_mem_free(scn);
+ qdf_mem_free(scn);
scn = NULL;
}
@@ -518,7 +519,7 @@ void hif_close(struct hif_opaque_softc *hif_ctx)
}
hif_bus_close(scn);
- cdf_mem_free(scn);
+ qdf_mem_free(scn);
}
/**
@@ -530,23 +531,23 @@ void hif_close(struct hif_opaque_softc *hif_ctx)
* @bus_type: bus type
* @type: enable type
*
- * Return: CDF_STATUS
+ * Return: QDF_STATUS
*/
-CDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev,
+QDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev,
void *bdev, const hif_bus_id *bid,
- enum ath_hal_bus_type bus_type,
+ enum qdf_bus_type bus_type,
enum hif_enable_type type)
{
- CDF_STATUS status;
+ QDF_STATUS status;
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
if (scn == NULL) {
HIF_ERROR("%s: hif_ctx = NULL", __func__);
- return CDF_STATUS_E_NULL_VALUE;
+ return QDF_STATUS_E_NULL_VALUE;
}
status = hif_enable_bus(scn, dev, bdev, bid, type);
- if (status != CDF_STATUS_SUCCESS) {
+ if (status != QDF_STATUS_SUCCESS) {
HIF_ERROR("%s: hif_enable_bus error = %d",
__func__, status);
return status;
@@ -558,7 +559,7 @@ CDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev,
if (hif_config_ce(scn)) {
HIF_ERROR("%s: Target probe failed.", __func__);
hif_disable_bus(scn);
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
return status;
}
/*
@@ -574,7 +575,7 @@ CDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev,
if (status < 0) {
HIF_ERROR("%s: ERROR - configure_IRQ_and_CE failed, status = %d",
__func__, status);
- return CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_E_FAILURE;
}
#endif
@@ -582,7 +583,7 @@ CDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev,
HIF_TRACE("%s: X OK", __func__);
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
/**
@@ -599,7 +600,7 @@ void hif_wlan_disable(struct hif_softc *scn)
enum icnss_driver_mode mode;
uint32_t con_mode = hif_get_conparam(scn);
- if (CDF_GLOBAL_FTM_MODE == con_mode)
+ if (QDF_GLOBAL_FTM_MODE == con_mode)
mode = ICNSS_FTM;
else if (WLAN_IS_EPPING_ENABLED(con_mode))
mode = ICNSS_EPPING;
@@ -725,7 +726,7 @@ int hif_check_fw_reg(struct hif_opaque_softc *scn)
*
* Return: n/a
*/
-void hif_read_phy_mem_base(struct hif_softc *scn, cdf_dma_addr_t *phy_mem_base)
+void hif_read_phy_mem_base(struct hif_softc *scn, qdf_dma_addr_t *phy_mem_base)
{
*phy_mem_base = scn->mem_pa;
}
@@ -901,7 +902,7 @@ void hif_init_ini_config(struct hif_opaque_softc *hif_ctx,
{
struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx);
- cdf_mem_copy(&scn->hif_config, cfg, sizeof(struct hif_config_info));
+ qdf_mem_copy(&scn->hif_config, cfg, sizeof(struct hif_config_info));
}
/**
diff --git a/hif/src/hif_main.h b/hif/src/hif_main.h
index 54384ab85f5b..be51e3533e4c 100644
--- a/hif/src/hif_main.h
+++ b/hif/src/hif_main.h
@@ -42,8 +42,8 @@
#ifndef __HIF_MAIN_H__
#define __HIF_MAIN_H__
-#include <cdf_atomic.h> /* cdf_atomic_read */
-#include "cdf_lock.h"
+#include <qdf_atomic.h> /* qdf_atomic_read */
+#include "qdf_lock.h"
#include "cepci.h"
#include "hif.h"
@@ -113,9 +113,9 @@ struct hif_softc {
struct hif_config_info hif_config;
struct hif_target_info target_info;
void __iomem *mem;
- enum ath_hal_bus_type bus_type;
+ enum qdf_bus_type bus_type;
void *ce_id_to_state[CE_COUNT_MAX];
- cdf_device_t cdf_dev;
+ qdf_device_t qdf_dev;
bool hif_init_done;
bool request_irq_done;
/* Packet statistics */
@@ -139,7 +139,7 @@ struct hif_softc {
int fastpath_mode_on;
atomic_t tasklet_from_intr;
int htc_endpoint;
- cdf_dma_addr_t mem_pa;
+ qdf_dma_addr_t mem_pa;
bool athdiag_procfs_inited;
#ifdef FEATURE_NAPI
struct qca_napi_data napi_data;
@@ -159,7 +159,7 @@ void athdiag_procfs_remove(void);
/* routine to modify the initial buffer count to be allocated on an os
* platform basis. Platform owner will need to modify this as needed
*/
-cdf_size_t init_buffer_count(cdf_size_t maxSize);
+qdf_size_t init_buffer_count(qdf_size_t maxSize);
irqreturn_t hif_fw_interrupt_handler(int irq, void *arg);
int hif_get_target_type(struct hif_softc *ol_sc, struct device *dev,
@@ -172,14 +172,14 @@ int hif_get_device_type(uint32_t device_id,
bool hif_targ_is_awake(struct hif_softc *scn, void *__iomem *mem);
void hif_nointrs(struct hif_softc *scn);
void hif_bus_close(struct hif_softc *ol_sc);
-CDF_STATUS hif_bus_open(struct hif_softc *ol_sc,
- enum ath_hal_bus_type bus_type);
-CDF_STATUS hif_enable_bus(struct hif_softc *ol_sc, struct device *dev,
+QDF_STATUS hif_bus_open(struct hif_softc *ol_sc,
+ enum qdf_bus_type bus_type);
+QDF_STATUS hif_enable_bus(struct hif_softc *ol_sc, struct device *dev,
void *bdev, const hif_bus_id *bid, enum hif_enable_type type);
void hif_disable_bus(struct hif_softc *scn);
void hif_bus_prevent_linkdown(struct hif_softc *scn, bool flag);
int hif_bus_get_context_size(void);
-void hif_read_phy_mem_base(struct hif_softc *scn, cdf_dma_addr_t *bar_value);
+void hif_read_phy_mem_base(struct hif_softc *scn, qdf_dma_addr_t *bar_value);
uint32_t hif_get_conparam(struct hif_softc *scn);
struct hif_callbacks *hif_get_callbacks_handle(struct hif_softc *scn);
bool hif_is_driver_unloading(struct hif_softc *scn);
diff --git a/hif/src/hif_napi.c b/hif/src/hif_napi.c
index 64a05e61d87a..58635cba9a87 100644
--- a/hif/src/hif_napi.c
+++ b/hif/src/hif_napi.c
@@ -424,7 +424,7 @@ int hif_napi_poll(struct hif_opaque_softc *hif_ctx, struct napi_struct *napi,
napi_info->stats[cpu].napi_polls++;
if (unlikely(NULL == hif))
- CDF_ASSERT(hif != NULL); /* emit a warning if hif NULL */
+ QDF_ASSERT(hif != NULL); /* emit a warning if hif NULL */
else {
rc = ce_per_engine_service(hif, NAPI_ID2PIPE(napi_info->id));
HIF_INFO_HI("%s: ce_per_engine_service processed %d msgs",
@@ -461,7 +461,7 @@ int hif_napi_poll(struct hif_opaque_softc *hif_ctx, struct napi_struct *napi,
hif_napi_enable_irq(hif_ctx, napi_info->id);
/* support suspend/resume */
- cdf_atomic_dec(&(hif->active_tasklet_cnt));
+ qdf_atomic_dec(&(hif->active_tasklet_cnt));
}
NAPI_DEBUG("%s:%d: napi_complete + enabling the interrupts",
diff --git a/hif/src/icnss_stub/icnss_stub.c b/hif/src/icnss_stub/icnss_stub.c
index 85c2320f5ea1..90540c5df13f 100644
--- a/hif/src/icnss_stub/icnss_stub.c
+++ b/hif/src/icnss_stub/icnss_stub.c
@@ -32,7 +32,7 @@
#include <hif.h>
#include "regtable.h"
#include "hif_debug.h"
-#include "cdf_status.h"
+#include "qdf_status.h"
#include "qwlan_version.h"
#include <net/cnss.h>
@@ -294,7 +294,7 @@ void icnss_dispatch_ce_irq(struct hif_softc *scn)
if (intr_summary == 0) {
if ((scn->target_status != OL_TRGET_STATUS_RESET) &&
- (!cdf_atomic_read(&scn->link_suspended))) {
+ (!qdf_atomic_read(&scn->link_suspended))) {
hif_write32_mb(scn->mem +
(SOC_CORE_BASE_ADDRESS |
diff --git a/hif/src/mp_dev.c b/hif/src/mp_dev.c
index f1611dbda910..b3b0f3124084 100644
--- a/hif/src/mp_dev.c
+++ b/hif/src/mp_dev.c
@@ -77,7 +77,7 @@ struct priv_ctrl_ctx {
static struct priv_ctrl_ctx g_priv_dump_ctx;
-static INLINE void set_target_reg_bits(void __iomem *mem, uint32_t reg,
+static inline void set_target_reg_bits(void __iomem *mem, uint32_t reg,
uint32_t bitmask, uint32_t val)
{
uint32_t value = hif_read32_mb(mem + (reg));
@@ -90,7 +90,7 @@ static INLINE void set_target_reg_bits(void __iomem *mem, uint32_t reg,
hif_write32_mb(mem + (reg), value);
}
-static INLINE uint32_t get_target_reg_bits(void __iomem *mem,
+static inline uint32_t get_target_reg_bits(void __iomem *mem,
uint32_t reg, uint32_t bitmask)
{
uint32_t value = hif_read32_mb(mem + (reg));
@@ -177,9 +177,9 @@ void priv_dump_chaninfo(struct hif_softc *scn)
hif_read32_mb(scn->mem +
BB_chn_tables_intf_data) &
0x0000ffff;
- cdf_print("0x%x\t", val);
+ qdf_print("0x%x\t", val);
if (i % 4 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
} else {
len = (bw == 2) ? 59 : 60;
@@ -187,10 +187,10 @@ void priv_dump_chaninfo(struct hif_softc *scn)
tmp =
hif_read32_mb(scn->mem +
BB_chn_tables_intf_data);
- cdf_print("0x%x\t", ((tmp >> 16) & 0x0000ffff));
- cdf_print("0x%x\t", (tmp & 0x0000ffff));
+ qdf_print("0x%x\t", ((tmp >> 16) & 0x0000ffff));
+ qdf_print("0x%x\t", (tmp & 0x0000ffff));
if (i % 2 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
if (bw > 2) {
/* bw == 3 for vht80 */
@@ -202,11 +202,11 @@ void priv_dump_chaninfo(struct hif_softc *scn)
tmp =
hif_read32_mb(scn->mem +
BB_chn_tables_intf_data);
- cdf_print("0x%x\t",
+ qdf_print("0x%x\t",
((tmp >> 16) & 0x0000ffff));
- cdf_print("0x%x\t", (tmp & 0x0000ffff));
+ qdf_print("0x%x\t", (tmp & 0x0000ffff));
if (i % 2 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
}
}
@@ -219,9 +219,9 @@ void priv_dump_chaninfo(struct hif_softc *scn)
hif_read32_mb(scn->mem +
BB_chn1_tables_intf_data) &
0x0000ffff;
- cdf_print("0x%x\t", val);
+ qdf_print("0x%x\t", val);
if (i % 4 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
} else {
len = (bw == 2) ? 59 : 60;
@@ -229,10 +229,10 @@ void priv_dump_chaninfo(struct hif_softc *scn)
tmp =
hif_read32_mb(scn->mem +
BB_chn1_tables_intf_data);
- cdf_print("0x%x\n", (tmp >> 16) & 0x0000ffff);
- cdf_print("0x%x\n", tmp & 0x0000ffff);
+ qdf_print("0x%x\n", (tmp >> 16) & 0x0000ffff);
+ qdf_print("0x%x\n", tmp & 0x0000ffff);
if (i % 2 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
if (bw > 2) {
/* bw == 3 for vht80 */
@@ -244,11 +244,11 @@ void priv_dump_chaninfo(struct hif_softc *scn)
tmp =
hif_read32_mb(scn->mem +
BB_chn1_tables_intf_data);
- cdf_print("0x%x\t",
+ qdf_print("0x%x\t",
((tmp >> 16) & 0x0000ffff));
- cdf_print("0x%x\t", (tmp & 0x0000ffff));
+ qdf_print("0x%x\t", (tmp & 0x0000ffff));
if (i % 2 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
}
}
@@ -283,9 +283,9 @@ void priv_dump_agc(struct hif_softc *scn)
BB_chaninfo_tab_b0 + i * 4);
val = hif_read32_mb(scn->mem +
PHY_BB_CHN_TABLES_INTF_DATA);
- cdf_print("0x%x\t", val);
+ qdf_print("0x%x\t", val);
if (i % 4 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
}
if (chain1) {
@@ -295,9 +295,9 @@ void priv_dump_agc(struct hif_softc *scn)
BB_chaninfo_tab_b0 + i * 4);
val = hif_read32_mb(scn->mem +
PHY_BB_CHN1_TABLES_INTF_DATA);
- cdf_print("0x%x\t", val);
+ qdf_print("0x%x\t", val);
if (i % 4 == 0)
- cdf_print("\n");
+ qdf_print("\n");
}
}
HIF_TRACE("%s: AGC history buffer dump X", __func__);
@@ -316,12 +316,12 @@ void priv_dump_bbwatchdog(struct hif_softc *scn)
HIF_TRACE("%s: BB watchdog dump E", __func__);
val = hif_read32_mb(scn->mem + BB_watchdog_status);
- cdf_print("0x%x\t", val);
+ qdf_print("0x%x\t", val);
val = hif_read32_mb(scn->mem + BB_watchdog_ctrl_1);
- cdf_print("0x%x\t", val);
+ qdf_print("0x%x\t", val);
val = hif_read32_mb(scn->mem + BB_watchdog_ctrl_2);
- cdf_print("0x%x\t", val);
+ qdf_print("0x%x\t", val);
val = hif_read32_mb(scn->mem + BB_watchdog_status_B);
- cdf_print("0x%x", val);
+ qdf_print("0x%x", val);
HIF_TRACE("%s: BB watchdog dump X", __func__);
}
diff --git a/hif/src/pcie/hif_io32_pci.h b/hif/src/pcie/hif_io32_pci.h
index 269d8541f8a3..10fc06fadf04 100644
--- a/hif/src/pcie/hif_io32_pci.h
+++ b/hif/src/pcie/hif_io32_pci.h
@@ -33,7 +33,7 @@
#include "hif_main.h"
#include "regtable.h"
#include "ce_reg.h"
-#include "cdf_atomic.h"
+#include "qdf_atomic.h"
#include "if_pci.h"
/*
* For maximum performance and no power management, set this to 1.
@@ -258,13 +258,13 @@ static inline void ce_irq_enable(struct hif_softc *scn, int ce_id)
uint32_t tmp = 1 << ce_id;
struct hif_pci_softc *sc = HIF_GET_PCI_SOFTC(scn);
- cdf_spin_lock_irqsave(&sc->irq_lock);
+ qdf_spin_lock_irqsave(&sc->irq_lock);
scn->ce_irq_summary &= ~tmp;
if (scn->ce_irq_summary == 0) {
/* Enable Legacy PCI line interrupts */
if (LEGACY_INTERRUPTS(sc) &&
(scn->target_status != OL_TRGET_STATUS_RESET) &&
- (!cdf_atomic_read(&scn->link_suspended))) {
+ (!qdf_atomic_read(&scn->link_suspended))) {
hif_write32_mb(scn->mem +
(SOC_CORE_BASE_ADDRESS |
@@ -278,7 +278,7 @@ static inline void ce_irq_enable(struct hif_softc *scn, int ce_id)
}
if (scn->hif_init_done == true)
A_TARGET_ACCESS_END(scn);
- cdf_spin_unlock_irqrestore(&sc->irq_lock);
+ qdf_spin_unlock_irqrestore(&sc->irq_lock);
/* check for missed firmware crash */
hif_fw_interrupt_handler(0, scn);
diff --git a/hif/src/pcie/if_pci.c b/hif/src/pcie/if_pci.c
index 55d56a37acdc..687bb1187076 100644
--- a/hif/src/pcie/if_pci.c
+++ b/hif/src/pcie/if_pci.c
@@ -46,9 +46,9 @@
#include <linux/debugfs.h>
#include <linux/seq_file.h>
#include <osapi_linux.h>
-#include "cdf_status.h"
+#include "qdf_status.h"
#include "wma_api.h"
-#include "cdf_atomic.h"
+#include "qdf_atomic.h"
#include "wlan_hdd_power.h"
#include "wlan_hdd_main.h"
#ifdef CONFIG_CNSS
@@ -239,7 +239,7 @@ static irqreturn_t hif_pci_interrupt_handler(int irq, void *arg)
__func__,
hif_read32_mb(sc->mem + 0x80018),
hif_read32_mb(sc->mem + 0x8001c));
- CDF_BUG(0);
+ QDF_BUG(0);
}
PCI_CLR_CAUSE0_REGISTER(sc);
@@ -260,9 +260,9 @@ static irqreturn_t hif_pci_interrupt_handler(int irq, void *arg)
if (ssr_irq) {
sc->irq_event = irq;
- cdf_atomic_set(&scn->tasklet_from_intr, 1);
+ qdf_atomic_set(&scn->tasklet_from_intr, 1);
- cdf_atomic_inc(&scn->active_tasklet_cnt);
+ qdf_atomic_inc(&scn->active_tasklet_cnt);
tasklet_schedule(&sc->intr_tq);
} else {
icnss_dispatch_ce_irq(scn);
@@ -302,13 +302,13 @@ void hif_pci_cancel_deferred_target_sleep(struct hif_softc *scn)
struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(scn);
A_target_id_t pci_addr = scn->mem;
- cdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
+ qdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
/*
* If the deferred sleep timer is running cancel it
* and put the soc into sleep.
*/
if (hif_state->fake_sleep == true) {
- cdf_softirq_timer_cancel(&hif_state->sleep_timer);
+ qdf_timer_stop(&hif_state->sleep_timer);
if (hif_state->verified_awake == false) {
hif_write32_mb(pci_addr + PCIE_LOCAL_BASE_ADDRESS +
PCIE_SOC_WAKE_ADDRESS,
@@ -316,7 +316,7 @@ void hif_pci_cancel_deferred_target_sleep(struct hif_softc *scn)
}
hif_state->fake_sleep = false;
}
- cdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
+ qdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
}
#else
inline void hif_pci_cancel_deferred_target_sleep(struct hif_softc *scn)
@@ -367,7 +367,7 @@ static void hif_pci_device_reset(struct hif_pci_softc *sc)
if (hif_targ_is_awake(scn, mem))
break;
- cdf_mdelay(1);
+ qdf_mdelay(1);
}
/* Put Target, including PCIe, into RESET. */
@@ -379,7 +379,7 @@ static void hif_pci_device_reset(struct hif_pci_softc *sc)
RTC_STATE_COLD_RESET_MASK)
break;
- cdf_mdelay(1);
+ qdf_mdelay(1);
}
/* Pull Target, including PCIe, out of RESET. */
@@ -391,7 +391,7 @@ static void hif_pci_device_reset(struct hif_pci_softc *sc)
RTC_STATE_COLD_RESET_MASK))
break;
- cdf_mdelay(1);
+ qdf_mdelay(1);
}
A_PCIE_LOCAL_REG_WRITE(mem, PCIE_SOC_WAKE_ADDRESS, PCIE_SOC_WAKE_RESET);
@@ -432,7 +432,7 @@ void hif_pci_device_warm_reset(struct hif_pci_softc *sc)
for (i = 0; i < ATH_PCI_RESET_WAIT_MAX; i++) {
if (hif_targ_is_awake(scn, mem))
break;
- cdf_mdelay(1);
+ qdf_mdelay(1);
}
/*
@@ -457,7 +457,7 @@ void hif_pci_device_warm_reset(struct hif_pci_softc *sc)
hif_write32_mb((mem + (SOC_CORE_BASE_ADDRESS + PCIE_INTR_CLR_ADDRESS)),
HOST_GROUP0_MASK);
- cdf_mdelay(100);
+ qdf_mdelay(100);
/* Clear FW_INDICATOR_ADDRESS */
if (HAS_FW_INDICATOR) {
@@ -490,7 +490,7 @@ void hif_pci_device_warm_reset(struct hif_pci_softc *sc)
hif_read32_mb(mem +
(RTC_SOC_BASE_ADDRESS |
SOC_RESET_CONTROL_ADDRESS));
- cdf_mdelay(10);
+ qdf_mdelay(10);
/* CE unreset */
val &= ~SOC_RESET_CONTROL_CE_RST_MASK;
@@ -500,7 +500,7 @@ void hif_pci_device_warm_reset(struct hif_pci_softc *sc)
hif_read32_mb(mem +
(RTC_SOC_BASE_ADDRESS |
SOC_RESET_CONTROL_ADDRESS));
- cdf_mdelay(10);
+ qdf_mdelay(10);
/* Read Target CPU Intr Cause */
val = hif_read32_mb(mem + (SOC_CORE_BASE_ADDRESS | CPU_INTR_ADDRESS));
@@ -522,7 +522,7 @@ void hif_pci_device_warm_reset(struct hif_pci_softc *sc)
HIF_INFO_MED("%s: RESET_CONTROL after cpu warm reset 0x%x",
__func__, val);
- cdf_mdelay(100);
+ qdf_mdelay(100);
HIF_INFO_MED("%s: Target Warm reset complete", __func__);
}
@@ -593,7 +593,7 @@ int hif_check_soc_status(struct hif_opaque_softc *hif_ctx)
hif_write32_mb(sc->mem + PCIE_LOCAL_BASE_ADDRESS +
PCIE_SOC_WAKE_ADDRESS, PCIE_SOC_WAKE_V_MASK);
- cdf_mdelay(100);
+ qdf_mdelay(100);
timeout_count += 100;
}
@@ -816,8 +816,8 @@ static void reschedule_tasklet_work_handler(void *arg)
*/
static void hif_init_reschedule_tasklet_work(struct hif_pci_softc *sc)
{
- cdf_create_work(&sc->reschedule_tasklet_work,
- reschedule_tasklet_work_handler, sc);
+ qdf_create_work(0, &sc->reschedule_tasklet_work,
+ reschedule_tasklet_work_handler, NULL);
}
#else
static void hif_init_reschedule_tasklet_work(struct hif_pci_softc *sc) { }
@@ -831,7 +831,7 @@ static void wlan_tasklet(unsigned long data)
if (scn->hif_init_done == false)
goto end;
- if (cdf_atomic_read(&scn->link_suspended))
+ if (qdf_atomic_read(&scn->link_suspended))
goto end;
if (!ADRASTEA_BU) {
@@ -841,8 +841,8 @@ static void wlan_tasklet(unsigned long data)
}
end:
- cdf_atomic_set(&scn->tasklet_from_intr, 0);
- cdf_atomic_dec(&scn->active_tasklet_cnt);
+ qdf_atomic_set(&scn->tasklet_from_intr, 0);
+ qdf_atomic_dec(&scn->active_tasklet_cnt);
}
#ifdef FEATURE_RUNTIME_PM
@@ -1049,7 +1049,7 @@ static void hif_pm_runtime_start(struct hif_pci_softc *sc)
return;
}
- if (mode == CDF_FTM_MODE || WLAN_IS_EPPING_ENABLED(mode)) {
+ if (mode == QDF_FTM_MODE || WLAN_IS_EPPING_ENABLED(mode)) {
HIF_INFO("%s: RUNTIME PM is disabled for FTM/EPPING mode\n",
__func__);
return;
@@ -1062,7 +1062,7 @@ static void hif_pm_runtime_start(struct hif_pci_softc *sc)
ol_sc->runtime_pm_delay);
cnss_runtime_init(sc->dev, ol_sc->runtime_pm_delay);
- cdf_atomic_set(&sc->pm_state, HIF_PM_RUNTIME_STATE_ON);
+ qdf_atomic_set(&sc->pm_state, HIF_PM_RUNTIME_STATE_ON);
hif_runtime_pm_debugfs_create(sc);
}
@@ -1081,13 +1081,13 @@ static void hif_pm_runtime_stop(struct hif_pci_softc *sc)
if (!ol_sc->enable_runtime_pm)
return;
- if (mode == CDF_FTM_MODE || WLAN_IS_EPPING_ENABLED(mode))
+ if (mode == QDF_FTM_MODE || WLAN_IS_EPPING_ENABLED(mode))
return;
cnss_runtime_exit(sc->dev);
cnss_pm_runtime_request(sc->dev, CNSS_PM_RUNTIME_RESUME);
- cdf_atomic_set(&sc->pm_state, HIF_PM_RUNTIME_STATE_NONE);
+ qdf_atomic_set(&sc->pm_state, HIF_PM_RUNTIME_STATE_NONE);
hif_runtime_pm_debugfs_remove(sc);
del_timer_sync(&sc->runtime_timer);
@@ -1104,10 +1104,10 @@ static void hif_pm_runtime_open(struct hif_pci_softc *sc)
{
spin_lock_init(&sc->runtime_lock);
- cdf_atomic_init(&sc->pm_state);
+ qdf_atomic_init(&sc->pm_state);
sc->prevent_linkdown_lock =
hif_runtime_lock_init("linkdown suspend disabled");
- cdf_atomic_set(&sc->pm_state, HIF_PM_RUNTIME_STATE_NONE);
+ qdf_atomic_set(&sc->pm_state, HIF_PM_RUNTIME_STATE_NONE);
INIT_LIST_HEAD(&sc->prevent_suspend_list);
}
@@ -1119,7 +1119,7 @@ static void hif_pm_runtime_open(struct hif_pci_softc *sc)
*/
static void hif_pm_runtime_close(struct hif_pci_softc *sc)
{
- if (cdf_atomic_read(&sc->pm_state) == HIF_PM_RUNTIME_STATE_NONE)
+ if (qdf_atomic_read(&sc->pm_state) == HIF_PM_RUNTIME_STATE_NONE)
return;
else
hif_pm_runtime_stop(sc);
@@ -1193,16 +1193,16 @@ int hif_bus_get_context_size(void)
*
* Return: n/a
*/
-CDF_STATUS hif_bus_open(struct hif_softc *ol_sc, enum ath_hal_bus_type bus_type)
+QDF_STATUS hif_bus_open(struct hif_softc *ol_sc, enum qdf_bus_type bus_type)
{
struct hif_pci_softc *sc = HIF_GET_PCI_SOFTC(ol_sc);
ol_sc->bus_type = bus_type;
hif_pm_runtime_open(sc);
- cdf_spinlock_init(&sc->irq_lock);
+ qdf_spinlock_create(&sc->irq_lock);
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
/**
@@ -1228,8 +1228,8 @@ int hif_enable_pci(struct hif_pci_softc *sc,
uint16_t device_id;
struct hif_softc *ol_sc = HIF_GET_SOFTC(sc);
- pci_read_config_word(pdev,PCI_DEVICE_ID,&device_id);
- if(device_id != id->device) {
+ pci_read_config_word(pdev, PCI_DEVICE_ID, &device_id);
+ if (device_id != id->device) {
HIF_ERROR(
"%s: dev id mismatch, config id = 0x%x, probing id = 0x%x",
__func__, device_id, id->device);
@@ -1361,7 +1361,7 @@ int hif_pci_probe_tgt_wakeup(struct hif_pci_softc *sc)
ret = -EAGAIN;
goto end;
}
- cdf_mdelay(1);
+ qdf_mdelay(1);
targ_awake_limit--;
}
@@ -1374,7 +1374,7 @@ int hif_pci_probe_tgt_wakeup(struct hif_pci_softc *sc)
PCIE_LOCAL_BASE_ADDRESS +
PCIE_SOC_RDY_STATUS_ADDRESS) \
& PCIE_SOC_RDY_STATUS_BAR_MASK)) {
- cdf_mdelay(10);
+ qdf_mdelay(10);
}
if (wait_limit < 0) {
/* AR6320v1 doesn't support checking of BAR0 configuration,
@@ -1411,10 +1411,10 @@ static void wlan_tasklet_msi(unsigned long data)
if (scn->hif_init_done == false)
goto irq_handled;
- if (cdf_atomic_read(&scn->link_suspended))
+ if (qdf_atomic_read(&scn->link_suspended))
goto irq_handled;
- cdf_atomic_inc(&scn->active_tasklet_cnt);
+ qdf_atomic_inc(&scn->active_tasklet_cnt);
if (entry->id == HIF_MAX_TASKLET_NUM) {
/* the last tasklet is for fw IRQ */
@@ -1430,7 +1430,7 @@ static void wlan_tasklet_msi(unsigned long data)
return;
irq_handled:
- cdf_atomic_dec(&scn->active_tasklet_cnt);
+ qdf_atomic_dec(&scn->active_tasklet_cnt);
}
@@ -1464,17 +1464,17 @@ int hif_configure_msi(struct hif_pci_softc *sc)
int i;
sc->num_msi_intrs = HIF_MAX_TASKLET_NUM;
- sc->tasklet_entries[HIF_MAX_TASKLET_NUM -1].hif_handler =
+ sc->tasklet_entries[HIF_MAX_TASKLET_NUM-1].hif_handler =
(void *)sc;
- sc->tasklet_entries[HIF_MAX_TASKLET_NUM -1].id =
+ sc->tasklet_entries[HIF_MAX_TASKLET_NUM-1].id =
HIF_MAX_TASKLET_NUM;
tasklet_init(&sc->intr_tq, wlan_tasklet_msi,
(unsigned long)&sc->tasklet_entries[
- HIF_MAX_TASKLET_NUM -1]);
+ HIF_MAX_TASKLET_NUM-1]);
ret = request_irq(sc->pdev->irq + MSI_ASSIGN_FW,
hif_pci_msi_fw_handler,
IRQF_SHARED, "wlan_pci", sc);
- if(ret) {
+ if (ret) {
HIF_ERROR("%s: request_irq failed", __func__);
goto err_intr;
}
@@ -1487,7 +1487,7 @@ int hif_configure_msi(struct hif_pci_softc *sc)
i + MSI_ASSIGN_CE_INITIAL),
ce_per_engine_handler, IRQF_SHARED,
"wlan_pci", sc);
- if(ret) {
+ if (ret) {
HIF_ERROR("%s: request_irq failed", __func__);
goto err_intr;
}
@@ -1495,7 +1495,8 @@ int hif_configure_msi(struct hif_pci_softc *sc)
} else if (rv > 0) {
HIF_TRACE("%s: use single msi", __func__);
- if ((ret = pci_enable_msi(sc->pdev)) < 0) {
+ ret = pci_enable_msi(sc->pdev);
+ if (ret < 0) {
HIF_ERROR("%s: single MSI allocation failed",
__func__);
/* Try for legacy PCI line interrupts */
@@ -1507,7 +1508,7 @@ int hif_configure_msi(struct hif_pci_softc *sc)
ret = request_irq(sc->pdev->irq,
hif_pci_interrupt_handler,
IRQF_SHARED, "wlan_pci", sc);
- if(ret) {
+ if (ret) {
HIF_ERROR("%s: request_irq failed", __func__);
goto err_intr;
}
@@ -1517,7 +1518,8 @@ int hif_configure_msi(struct hif_pci_softc *sc)
ret = -EIO;
HIF_ERROR("%s: do not support MSI, rv = %d", __func__, rv);
}
- if ((ret = pci_enable_msi(sc->pdev)) < 0) {
+ ret = pci_enable_msi(sc->pdev);
+ if (ret < 0) {
HIF_ERROR("%s: single MSI interrupt allocation failed",
__func__);
/* Try for legacy PCI line interrupts */
@@ -1528,7 +1530,7 @@ int hif_configure_msi(struct hif_pci_softc *sc)
ret = request_irq(sc->pdev->irq,
hif_pci_interrupt_handler, IRQF_SHARED,
"wlan_pci", sc);
- if(ret) {
+ if (ret) {
HIF_ERROR("%s: request_irq failed", __func__);
goto err_intr;
}
@@ -1564,7 +1566,7 @@ static int hif_pci_configure_legacy_irq(struct hif_pci_softc *sc)
ret = request_irq(sc->pdev->irq,
hif_pci_interrupt_handler, IRQF_SHARED,
"wlan_pci", sc);
- if(ret) {
+ if (ret) {
HIF_ERROR("%s: request_irq failed, ret = %d", __func__, ret);
goto end;
}
@@ -1576,7 +1578,7 @@ static int hif_pci_configure_legacy_irq(struct hif_pci_softc *sc)
PCIE_SOC_WAKE_ADDRESS,
PCIE_SOC_WAKE_RESET);
end:
- CDF_TRACE(CDF_MODULE_ID_HIF, CDF_TRACE_LEVEL_ERROR,
+ QDF_TRACE(QDF_MODULE_ID_HIF, QDF_TRACE_LEVEL_ERROR,
"%s: X, ret = %d", __func__, ret);
return ret;
}
@@ -1724,7 +1726,7 @@ static inline int hif_drain_tasklets(struct hif_softc *scn)
{
uint32_t ce_drain_wait_cnt = 0;
- while (cdf_atomic_read(&scn->active_tasklet_cnt)) {
+ while (qdf_atomic_read(&scn->active_tasklet_cnt)) {
if (++ce_drain_wait_cnt > HIF_CE_DRAIN_WAIT_CNT) {
HIF_ERROR("%s: CE still not done with access",
__func__);
@@ -1827,7 +1829,7 @@ static int hif_bus_suspend_link_down(struct hif_softc *scn)
/* Stop the HIF Sleep Timer */
hif_cancel_deferred_target_sleep(hif_hdl);
- cdf_atomic_set(&scn->link_suspended, 1);
+ qdf_atomic_set(&scn->link_suspended, 1);
return 0;
}
@@ -1854,7 +1856,7 @@ static int hif_bus_resume_link_down(struct hif_softc *scn)
return -EFAULT;
}
- cdf_atomic_set(&scn->link_suspended, 0);
+ qdf_atomic_set(&scn->link_suspended, 0);
enable_irq(pdev->irq);
@@ -1914,7 +1916,7 @@ static void __hif_runtime_pm_set_state(struct hif_softc *scn,
}
sc = scn->hif_sc;
- cdf_atomic_set(&sc->pm_state, state);
+ qdf_atomic_set(&sc->pm_state, state);
}
@@ -2117,12 +2119,12 @@ static void hif_fastpath_resume(struct hif_opaque_softc *hif_ctx)
if (scn->fastpath_mode_on) {
if (Q_TARGET_ACCESS_BEGIN(scn)) {
ce_state = scn->ce_id_to_state[CE_HTT_H2T_MSG];
- cdf_spin_lock_bh(&ce_state->ce_index_lock);
+ qdf_spin_lock_bh(&ce_state->ce_index_lock);
/*war_ce_src_ring_write_idx_set */
CE_SRC_RING_WRITE_IDX_SET(scn, ce_state->ctrl_addr,
ce_state->src_ring->write_index);
- cdf_spin_unlock_bh(&ce_state->ce_index_lock);
+ qdf_spin_unlock_bh(&ce_state->ce_index_lock);
Q_TARGET_ACCESS_END(scn);
}
}
@@ -2151,7 +2153,7 @@ static void hif_free_msi_ctx(struct hif_softc *scn)
{
struct hif_pci_softc *sc = scn->hif_sc;
struct hif_msi_info *info = &sc->msi_info;
- struct device *dev = scn->cdf_dev->dev;
+ struct device *dev = scn->qdf_dev->dev;
OS_FREE_CONSISTENT(dev, 4, info->magic, info->magic_dma,
OS_GET_DMA_MEM_CONTEXT(scn, dmacontext));
@@ -2174,7 +2176,7 @@ void hif_disable_isr(struct hif_opaque_softc *ol_sc)
/* Cancel the pending tasklet */
ce_tasklet_kill(scn);
tasklet_kill(&sc->intr_tq);
- cdf_atomic_set(&scn->active_tasklet_cnt, 0);
+ qdf_atomic_set(&scn->active_tasklet_cnt, 0);
}
/* Function to reset SoC */
@@ -2299,7 +2301,7 @@ static int hif_log_soc_wakeup_timeout(struct hif_pci_softc *sc)
hif_msm_pcie_debug_info(sc);
if (!cfg->enable_self_recovery)
- CDF_BUG(0);
+ QDF_BUG(0);
scn->recovery = true;
@@ -2360,10 +2362,10 @@ hif_target_sleep_state_adjust(struct hif_softc *scn,
if (scn->recovery)
return -EACCES;
- if (cdf_atomic_read(&scn->link_suspended)) {
+ if (qdf_atomic_read(&scn->link_suspended)) {
HIF_ERROR("%s:invalid access, PCIe link is down", __func__);
debug = true;
- CDF_ASSERT(0);
+ QDF_ASSERT(0);
return -EACCES;
}
@@ -2371,29 +2373,29 @@ hif_target_sleep_state_adjust(struct hif_softc *scn,
wait_for_it = true;
HIF_ERROR("%s: doing debug for invalid access, PCIe link is suspended",
__func__);
- CDF_ASSERT(0);
+ QDF_ASSERT(0);
}
if (sleep_ok) {
- cdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
+ qdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
hif_state->keep_awake_count--;
if (hif_state->keep_awake_count == 0) {
/* Allow sleep */
hif_state->verified_awake = false;
- hif_state->sleep_ticks = cdf_system_ticks();
+ hif_state->sleep_ticks = qdf_system_ticks();
}
if (hif_state->fake_sleep == false) {
/* Set the Fake Sleep */
hif_state->fake_sleep = true;
/* Start the Sleep Timer */
- cdf_softirq_timer_cancel(&hif_state->sleep_timer);
- cdf_softirq_timer_start(&hif_state->sleep_timer,
+ qdf_timer_stop(&hif_state->sleep_timer);
+ qdf_timer_start(&hif_state->sleep_timer,
HIF_SLEEP_INACTIVITY_TIMER_PERIOD_MS);
}
- cdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
+ qdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
} else {
- cdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
+ qdf_spin_lock_irqsave(&hif_state->keep_awake_lock);
if (hif_state->fake_sleep) {
hif_state->verified_awake = true;
@@ -2407,7 +2409,7 @@ hif_target_sleep_state_adjust(struct hif_softc *scn,
}
}
hif_state->keep_awake_count++;
- cdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
+ qdf_spin_unlock_irqrestore(&hif_state->keep_awake_lock);
if (wait_for_it && !hif_state->verified_awake) {
#define PCIE_SLEEP_ADJUST_TIMEOUT 8000 /* 8Ms */
@@ -2673,7 +2675,7 @@ void hif_target_sync(struct hif_softc *scn)
PCIE_INTR_ENABLE_ADDRESS),
PCIE_INTR_FIRMWARE_MASK);
- cdf_mdelay(10);
+ qdf_mdelay(10);
}
if (wait_limit < 0)
HIF_TRACE("%s: FW signal timed out",
@@ -2696,9 +2698,9 @@ void hif_target_sync(struct hif_softc *scn)
* @bdev: bus dev pointer
* bid: bus id pointer
* type: enum hif_enable_type such as HIF_ENABLE_TYPE_PROBE
- * Return: CDF_STATUS
+ * Return: QDF_STATUS
*/
-CDF_STATUS hif_enable_bus(struct hif_softc *ol_sc,
+QDF_STATUS hif_enable_bus(struct hif_softc *ol_sc,
struct device *dev, void *bdev,
const hif_bus_id *bid,
enum hif_enable_type type)
@@ -2716,7 +2718,7 @@ CDF_STATUS hif_enable_bus(struct hif_softc *ol_sc,
if (!ol_sc) {
HIF_ERROR("%s: hif_ctx is NULL", __func__);
- return CDF_STATUS_E_NOMEM;
+ return QDF_STATUS_E_NOMEM;
}
HIF_TRACE("%s: con_mode = 0x%x, device_id = 0x%x",
@@ -2787,7 +2789,7 @@ err_tgtstate:
hif_disable_pci(sc);
sc->pci_enabled = false;
HIF_ERROR("%s: error, hif_disable_pci done", __func__);
- return CDF_STATUS_E_ABORTED;
+ return QDF_STATUS_E_ABORTED;
err_enable_pci:
if (probe_again && (probe_again <= ATH_PCI_PROBE_RETRY_MAX)) {
@@ -2796,7 +2798,7 @@ err_enable_pci:
HIF_INFO("%s: pci reprobe", __func__);
/* 10, 40, 90, 100, 100, ... */
delay_time = max(100, 10 * (probe_again * probe_again));
- cdf_mdelay(delay_time);
+ qdf_mdelay(delay_time);
goto again;
}
return ret;
@@ -2869,7 +2871,7 @@ int hif_pm_runtime_get(struct hif_opaque_softc *hif_ctx)
return -EFAULT;
}
- pm_state = cdf_atomic_read(&sc->pm_state);
+ pm_state = qdf_atomic_read(&sc->pm_state);
if (pm_state == HIF_PM_RUNTIME_STATE_ON ||
pm_state == HIF_PM_RUNTIME_STATE_NONE) {
@@ -2888,7 +2890,7 @@ int hif_pm_runtime_get(struct hif_opaque_softc *hif_ctx)
if (ret && ret != -EINPROGRESS) {
sc->pm_stats.runtime_get_err++;
HIF_ERROR("%s: Runtime Get PM Error in pm_state:%d ret: %d",
- __func__, cdf_atomic_read(&sc->pm_state), ret);
+ __func__, qdf_atomic_read(&sc->pm_state), ret);
}
return ret;
@@ -2910,7 +2912,7 @@ int hif_pm_runtime_get(struct hif_opaque_softc *hif_ctx)
* This api will return a failure if runtime pm is stopped
* This api will return failure if it would decrement the usage count below 0.
*
- * return: CDF_STATUS_SUCCESS if the put is performed
+ * return: QDF_STATUS_SUCCESS if the put is performed
*/
int hif_pm_runtime_put(struct hif_opaque_softc *hif_ctx)
{
@@ -2928,7 +2930,7 @@ int hif_pm_runtime_put(struct hif_opaque_softc *hif_ctx)
usage_count = atomic_read(&sc->dev->power.usage_count);
if (usage_count == 1) {
- pm_state = cdf_atomic_read(&sc->pm_state);
+ pm_state = qdf_atomic_read(&sc->pm_state);
if (pm_state == HIF_PM_RUNTIME_STATE_NONE)
error = "Ignoring unexpected put when runtime pm is disabled";
@@ -3002,7 +3004,7 @@ static int __hif_pm_runtime_prevent_suspend(struct hif_pci_softc
hif_sc->pm_stats.prevent_suspend++;
HIF_ERROR("%s: in pm_state:%d ret: %d", __func__,
- cdf_atomic_read(&hif_sc->pm_state), ret);
+ qdf_atomic_read(&hif_sc->pm_state), ret);
return ret;
}
@@ -3029,7 +3031,7 @@ static int __hif_pm_runtime_allow_suspend(struct hif_pci_softc *hif_sc,
* context->active should be active for allow suspend to happen
* Handling this case here to prevent any failures.
*/
- if ((cdf_atomic_read(&hif_sc->pm_state) == HIF_PM_RUNTIME_STATE_NONE
+ if ((qdf_atomic_read(&hif_sc->pm_state) == HIF_PM_RUNTIME_STATE_NONE
&& usage_count == 1) || usage_count == 0) {
hif_pci_runtime_pm_warn(hif_sc,
"Allow without a prevent suspend");
@@ -3047,7 +3049,7 @@ static int __hif_pm_runtime_allow_suspend(struct hif_pci_softc *hif_sc,
ret = hif_pm_runtime_put_auto(hif_sc->dev);
HIF_ERROR("%s: in pm_state:%d ret: %d", __func__,
- cdf_atomic_read(&hif_sc->pm_state), ret);
+ qdf_atomic_read(&hif_sc->pm_state), ret);
hif_sc->pm_stats.allow_suspend++;
return ret;
@@ -3232,7 +3234,7 @@ int hif_pm_runtime_prevent_suspend_timeout(struct hif_opaque_softc *ol_sc,
spin_unlock_irqrestore(&hif_sc->runtime_lock, flags);
HIF_ERROR("%s: pm_state: %d delay: %dms ret: %d\n", __func__,
- cdf_atomic_read(&hif_sc->pm_state), delay, ret);
+ qdf_atomic_read(&hif_sc->pm_state), delay, ret);
return ret;
}
@@ -3250,7 +3252,7 @@ struct hif_pm_runtime_lock *hif_runtime_lock_init(const char *name)
{
struct hif_pm_runtime_lock *context;
- context = cdf_mem_malloc(sizeof(*context));
+ context = qdf_mem_malloc(sizeof(*context));
if (!context) {
HIF_ERROR("%s: No memory for Runtime PM wakelock context\n",
__func__);
@@ -3294,7 +3296,7 @@ void hif_runtime_lock_deinit(struct hif_opaque_softc *hif_ctx,
__hif_pm_runtime_allow_suspend(sc, context);
spin_unlock_irqrestore(&sc->runtime_lock, flags);
- cdf_mem_free(context);
+ qdf_mem_free(context);
}
#endif /* FEATURE_RUNTIME_PM */
diff --git a/hif/src/pcie/if_pci.h b/hif/src/pcie/if_pci.h
index b9ec34a1f94c..5f573b28476c 100644
--- a/hif/src/pcie/if_pci.h
+++ b/hif/src/pcie/if_pci.h
@@ -126,11 +126,11 @@ struct hif_pci_softc {
int irq_event;
int cacheline_sz;
u16 devid;
- cdf_dma_addr_t soc_pcie_bar0;
+ qdf_dma_addr_t soc_pcie_bar0;
struct hif_tasklet_entry tasklet_entries[HIF_MAX_TASKLET_NUM];
bool pci_enabled;
- cdf_spinlock_t irq_lock;
- cdf_work_t reschedule_tasklet_work;
+ qdf_spinlock_t irq_lock;
+ qdf_work_t reschedule_tasklet_work;
uint32_t lcr_val;
#ifdef FEATURE_RUNTIME_PM
atomic_t pm_state;
diff --git a/hif/src/snoc/hif_io32_snoc.h b/hif/src/snoc/hif_io32_snoc.h
index 6b8107c24d9d..bc968dda4e5f 100644
--- a/hif/src/snoc/hif_io32_snoc.h
+++ b/hif/src/snoc/hif_io32_snoc.h
@@ -41,7 +41,7 @@
#include "hif.h"
#include "regtable.h"
#include "ce_reg.h"
-#include "cdf_atomic.h"
+#include "qdf_atomic.h"
#include <soc/qcom/icnss.h>
#include "hif_main.h"
#include "hif_debug.h"
diff --git a/hif/src/snoc/if_snoc.c b/hif/src/snoc/if_snoc.c
index e5c04cd7cad0..fdb4463d2f70 100644
--- a/hif/src/snoc/if_snoc.c
+++ b/hif/src/snoc/if_snoc.c
@@ -97,7 +97,7 @@ void hif_disable_isr(struct hif_opaque_softc *hif_ctx)
hif_nointrs(scn);
ce_tasklet_kill(scn);
- cdf_atomic_set(&scn->active_tasklet_cnt, 0);
+ qdf_atomic_set(&scn->active_tasklet_cnt, 0);
}
/**
@@ -204,9 +204,9 @@ int hif_bus_get_context_size(void)
*
* Return: n/a
*/
-CDF_STATUS hif_bus_open(struct hif_softc *scn, enum ath_hal_bus_type bus_type)
+QDF_STATUS hif_bus_open(struct hif_softc *scn, enum qdf_bus_type bus_type)
{
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
/**
@@ -245,9 +245,9 @@ int hif_get_target_type(struct hif_softc *ol_sc, struct device *dev,
* @bid: bus id
* @type: bus type
*
- * Return: CDF_STATUS
+ * Return: QDF_STATUS
*/
-CDF_STATUS hif_enable_bus(struct hif_softc *ol_sc,
+QDF_STATUS hif_enable_bus(struct hif_softc *ol_sc,
struct device *dev, void *bdev,
const hif_bus_id *bid,
enum hif_enable_type type)
@@ -264,14 +264,14 @@ CDF_STATUS hif_enable_bus(struct hif_softc *ol_sc,
if (!ol_sc) {
HIF_ERROR("%s: hif_ctx is NULL", __func__);
- return CDF_STATUS_E_NOMEM;
+ return QDF_STATUS_E_NOMEM;
}
ret = hif_get_target_type(ol_sc, dev, bdev, bid,
&hif_type, &target_type);
if (ret < 0) {
HIF_ERROR("%s: invalid device id/revision_id", __func__);
- return CDF_STATUS_E_FAILURE;
+ return QDF_STATUS_E_FAILURE;
}
hif_register_tbl_attach(ol_sc, hif_type);
@@ -280,7 +280,7 @@ CDF_STATUS hif_enable_bus(struct hif_softc *ol_sc,
HIF_TRACE("%s: X - hif_type = 0x%x, target_type = 0x%x",
__func__, hif_type, target_type);
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
/**
diff --git a/htc/htc.c b/htc/htc.c
index cd7dbf3b9d00..46b0aa420f4c 100644
--- a/htc/htc.c
+++ b/htc/htc.c
@@ -29,7 +29,7 @@
#include "htc_debug.h"
#include "htc_internal.h"
#include <cdf_nbuf.h> /* cdf_nbuf_t */
-#include <cdf_types.h> /* cdf_print */
+#include <qdf_types.h> /* qdf_print */
#include <hif.h>
#include "epping_main.h"
#include "hif_io32.h"
@@ -69,16 +69,16 @@ static void destroy_htc_tx_ctrl_packet(HTC_PACKET *pPacket)
cdf_nbuf_free(netbuf);
}
- cdf_mem_free(pPacket);
+ qdf_mem_free(pPacket);
}
-static HTC_PACKET *build_htc_tx_ctrl_packet(cdf_device_t osdev)
+static HTC_PACKET *build_htc_tx_ctrl_packet(qdf_device_t osdev)
{
HTC_PACKET *pPacket = NULL;
cdf_nbuf_t netbuf;
do {
- pPacket = (HTC_PACKET *) cdf_mem_malloc(sizeof(HTC_PACKET));
+ pPacket = (HTC_PACKET *) qdf_mem_malloc(sizeof(HTC_PACKET));
if (NULL == pPacket) {
break;
}
@@ -86,9 +86,9 @@ static HTC_PACKET *build_htc_tx_ctrl_packet(cdf_device_t osdev)
netbuf =
cdf_nbuf_alloc(osdev, HTC_CONTROL_BUFFER_SIZE, 20, 4, true);
if (NULL == netbuf) {
- cdf_mem_free(pPacket);
+ qdf_mem_free(pPacket);
pPacket = NULL;
- cdf_print("%s: nbuf alloc failed\n", __func__);
+ qdf_print("%s: nbuf alloc failed\n", __func__);
break;
}
AR_DEBUG_PRINTF(ATH_DEBUG_TRC,
@@ -158,13 +158,13 @@ static void htc_cleanup(HTC_TARGET *target)
if (NULL == pPacket) {
break;
}
- cdf_mem_free(pPacket);
+ qdf_mem_free(pPacket);
}
pPacket = target->pBundleFreeList;
while (pPacket) {
HTC_PACKET *pPacketTmp = (HTC_PACKET *) pPacket->ListLink.pNext;
- cdf_mem_free(pPacket);
+ qdf_mem_free(pPacket);
pPacket = pPacketTmp;
}
#ifdef TODO_FIXME
@@ -178,21 +178,21 @@ static void htc_cleanup(HTC_TARGET *target)
cdf_nbuf_free(netbuf);
}
- cdf_mem_free(pPacket);
+ qdf_mem_free(pPacket);
}
#endif
- cdf_spinlock_destroy(&target->HTCLock);
- cdf_spinlock_destroy(&target->HTCRxLock);
- cdf_spinlock_destroy(&target->HTCTxLock);
- cdf_spinlock_destroy(&target->HTCCreditLock);
+ qdf_spinlock_destroy(&target->HTCLock);
+ qdf_spinlock_destroy(&target->HTCRxLock);
+ qdf_spinlock_destroy(&target->HTCTxLock);
+ qdf_spinlock_destroy(&target->HTCCreditLock);
/* free our instance */
- cdf_mem_free(target);
+ qdf_mem_free(target);
}
/* registered target arrival callback from the HIF layer */
-HTC_HANDLE htc_create(void *ol_sc, HTC_INIT_INFO *pInfo, cdf_device_t osdev)
+HTC_HANDLE htc_create(void *ol_sc, HTC_INIT_INFO *pInfo, qdf_device_t osdev)
{
struct hif_msg_callbacks htcCallbacks;
HTC_ENDPOINT *pEndpoint = NULL;
@@ -207,7 +207,7 @@ HTC_HANDLE htc_create(void *ol_sc, HTC_INIT_INFO *pInfo, cdf_device_t osdev)
A_REGISTER_MODULE_DEBUG_INFO(htc);
- target = (HTC_TARGET *) cdf_mem_malloc(sizeof(HTC_TARGET));
+ target = (HTC_TARGET *) qdf_mem_malloc(sizeof(HTC_TARGET));
if (target == NULL) {
HTC_ERROR("%s: Unable to allocate memory", __func__);
return NULL;
@@ -216,10 +216,10 @@ HTC_HANDLE htc_create(void *ol_sc, HTC_INIT_INFO *pInfo, cdf_device_t osdev)
A_MEMZERO(target, sizeof(HTC_TARGET));
htc_runtime_pm_init(target);
- cdf_spinlock_init(&target->HTCLock);
- cdf_spinlock_init(&target->HTCRxLock);
- cdf_spinlock_init(&target->HTCTxLock);
- cdf_spinlock_init(&target->HTCCreditLock);
+ qdf_spinlock_create(&target->HTCLock);
+ qdf_spinlock_create(&target->HTCRxLock);
+ qdf_spinlock_create(&target->HTCTxLock);
+ qdf_spinlock_create(&target->HTCCreditLock);
do {
A_MEMCPY(&target->HTCInitInfo, pInfo, sizeof(HTC_INIT_INFO));
@@ -232,7 +232,7 @@ HTC_HANDLE htc_create(void *ol_sc, HTC_INIT_INFO *pInfo, cdf_device_t osdev)
for (i = 0; i < HTC_PACKET_CONTAINER_ALLOCATION; i++) {
HTC_PACKET *pPacket =
- (HTC_PACKET *) cdf_mem_malloc(sizeof(HTC_PACKET));
+ (HTC_PACKET *) qdf_mem_malloc(sizeof(HTC_PACKET));
if (pPacket != NULL) {
A_MEMZERO(pPacket, sizeof(HTC_PACKET));
free_htc_packet_container(target, pPacket);
@@ -250,7 +250,7 @@ HTC_HANDLE htc_create(void *ol_sc, HTC_INIT_INFO *pInfo, cdf_device_t osdev)
#endif
/* setup HIF layer callbacks */
- cdf_mem_zero(&htcCallbacks, sizeof(struct hif_msg_callbacks));
+ qdf_mem_zero(&htcCallbacks, sizeof(struct hif_msg_callbacks));
htcCallbacks.Context = target;
htcCallbacks.rxCompletionHandler = htc_rx_completion_handler;
htcCallbacks.txCompletionHandler = htc_tx_completion_handler;
@@ -348,7 +348,7 @@ A_STATUS htc_setup_target_buffer_assignments(HTC_TARGET *target)
* There is no WMI mesage exchanges between host and target
* in endpoint ping case.
* In host side, the endpoint ping driver is a Ethernet driver
- * and it directly sits on HTC. Only HIF, HTC, CDF, ADF are
+ * and it directly sits on HTC. Only HIF, HTC, QDF, ADF are
* used by the endpoint ping driver. There is no wifi stack
* at all in host side also. For tx perf use case,
* the user space mboxping app sends the raw packets to endpoint
@@ -529,7 +529,7 @@ static void reset_endpoint_states(HTC_TARGET *target)
pEndpoint->target = target;
/* pEndpoint->TxCreditFlowEnabled = (A_BOOL)htc_credit_flow; */
pEndpoint->TxCreditFlowEnabled = (A_BOOL) 1;
- cdf_atomic_init(&pEndpoint->TxProcessCount);
+ qdf_atomic_init(&pEndpoint->TxProcessCount);
}
}
@@ -551,7 +551,7 @@ A_STATUS htc_start(HTC_HANDLE HTCHandle)
pSendPacket = htc_alloc_control_tx_packet(target);
if (NULL == pSendPacket) {
AR_DEBUG_ASSERT(false);
- cdf_print("%s: allocControlTxPacket failed\n",
+ qdf_print("%s: allocControlTxPacket failed\n",
__func__);
status = A_NO_MEMORY;
break;
@@ -658,8 +658,8 @@ void htc_stop(HTC_HANDLE HTCHandle)
htc_flush_rx_hold_queue(target, pEndpoint);
htc_flush_endpoint_tx(target, pEndpoint, HTC_TX_PACKET_TAG_ALL);
if (pEndpoint->ul_is_polled) {
- cdf_softirq_timer_cancel(&pEndpoint->ul_poll_timer);
- cdf_softirq_timer_free(&pEndpoint->ul_poll_timer);
+ qdf_timer_stop(&pEndpoint->ul_poll_timer);
+ qdf_timer_free(&pEndpoint->ul_poll_timer);
}
}
@@ -693,7 +693,7 @@ void htc_stop(HTC_HANDLE HTCHandle)
*/
void htc_runtime_pm_init(HTC_TARGET *target)
{
- cdf_create_work(&target->queue_kicker, htc_kick_queues, target);
+ qdf_create_work(0, &target->queue_kicker, htc_kick_queues, target);
}
/**
@@ -706,11 +706,11 @@ void htc_runtime_pm_init(HTC_TARGET *target)
*/
int htc_runtime_suspend(void)
{
- ol_txrx_pdev_handle txrx_pdev = cds_get_context(CDF_MODULE_ID_TXRX);
+ ol_txrx_pdev_handle txrx_pdev = cds_get_context(QDF_MODULE_ID_TXRX);
if (txrx_pdev == NULL) {
HTC_ERROR("%s: txrx context null", __func__);
- return CDF_STATUS_E_FAULT;
+ return QDF_STATUS_E_FAULT;
}
if (ol_txrx_get_tx_pending(txrx_pdev))
@@ -729,13 +729,13 @@ int htc_runtime_suspend(void)
*/
int htc_runtime_resume(void)
{
- HTC_HANDLE htc_ctx = cds_get_context(CDF_MODULE_ID_HTC);
+ HTC_HANDLE htc_ctx = cds_get_context(QDF_MODULE_ID_HTC);
HTC_TARGET *target = GET_HTC_TARGET_FROM_HANDLE(htc_ctx);
if (target == NULL)
return 0;
- cdf_schedule_work(&target->queue_kicker);
+ qdf_sched_work(0, &target->queue_kicker);
return 0;
}
@@ -866,9 +866,9 @@ void htc_cancel_deferred_target_sleep(void *context)
* Return: None
*/
void htc_ipa_get_ce_resource(HTC_HANDLE htc_handle,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr)
+ qdf_dma_addr_t *ce_reg_paddr)
{
HTC_TARGET *target = GET_HTC_TARGET_FROM_HANDLE(htc_handle);
diff --git a/htc/htc_api.h b/htc/htc_api.h
index f26f3d63368f..446aec328be9 100644
--- a/htc/htc_api.h
+++ b/htc/htc_api.h
@@ -33,7 +33,7 @@
#include "htc_packet.h"
#include <htc.h>
#include <htc_services.h>
-#include <cdf_types.h> /* cdf_device_t */
+#include <qdf_types.h> /* qdf_device_t */
#ifdef __cplusplus
extern "C" {
@@ -51,11 +51,11 @@ typedef void *HTC_HANDLE;
typedef A_UINT16 HTC_SERVICE_ID;
-typedef void (*HTC_TARGET_FAILURE)(void *Instance, CDF_STATUS Status);
+typedef void (*HTC_TARGET_FAILURE)(void *Instance, QDF_STATUS Status);
typedef struct _HTC_INIT_INFO {
void *pContext; /* context for target notifications */
- void (*TargetFailure)(void *Instance, CDF_STATUS Status);
+ void (*TargetFailure)(void *Instance, QDF_STATUS Status);
void (*TargetSendSuspendComplete)(void *ctx);
} HTC_INIT_INFO;
@@ -151,18 +151,18 @@ typedef struct _HTC_EP_CALLBACKS {
HTC_EP_RECV_ALLOC EpRecvAlloc; /* OPTIONAL recv allocation callback */
HTC_EP_RECV_ALLOC EpRecvAllocThresh; /* OPTIONAL recv allocation callback based on a threshold */
HTC_EP_SEND_PKT_COMP_MULTIPLE EpTxCompleteMultiple; /* OPTIONAL completion handler for multiple complete
- indications (EpTxComplete must be NULL) */
+ indications (EpTxComplete must be NULL) */
HTC_EP_RECV_PKT_MULTIPLE EpRecvPktMultiple; /* OPTIONAL completion handler for multiple
- recv packet indications (EpRecv must be NULL) */
+ recv packet indications (EpRecv must be NULL) */
HTC_EP_RESUME_TX_QUEUE ep_resume_tx_queue;
int RecvAllocThreshold; /* if EpRecvAllocThresh is non-NULL, HTC will compare the
- threshold value to the current recv packet length and invoke
- the EpRecvAllocThresh callback to acquire a packet buffer */
+ threshold value to the current recv packet length and invoke
+ the EpRecvAllocThresh callback to acquire a packet buffer */
int RecvRefillWaterMark; /* if a EpRecvRefill handler is provided, this value
- can be used to set a trigger refill callback
- when the recv queue drops below this value
- if set to 0, the refill is only called when packets
- are empty */
+ can be used to set a trigger refill callback
+ when the recv queue drops below this value
+ if set to 0, the refill is only called when packets
+ are empty */
} HTC_EP_CALLBACKS;
/* service connection information */
@@ -196,32 +196,32 @@ typedef struct _HTC_ENDPOINT_CREDIT_DIST {
HTC_SERVICE_ID service_id; /* Service ID (set by HTC) */
HTC_ENDPOINT_ID Endpoint; /* endpoint for this distribution struct (set by HTC) */
A_UINT32 DistFlags; /* distribution flags, distribution function can
- set default activity using SET_EP_ACTIVE() macro */
+ set default activity using SET_EP_ACTIVE() macro */
int TxCreditsNorm; /* credits for normal operation, anything above this
- indicates the endpoint is over-subscribed, this field
- is only relevant to the credit distribution function */
+ indicates the endpoint is over-subscribed, this field
+ is only relevant to the credit distribution function */
int TxCreditsMin; /* floor for credit distribution, this field is
- only relevant to the credit distribution function */
+ only relevant to the credit distribution function */
int TxCreditsAssigned; /* number of credits assigned to this EP, this field
- is only relevant to the credit dist function */
+ is only relevant to the credit dist function */
int TxCredits; /* current credits available, this field is used by
- HTC to determine whether a message can be sent or
- must be queued */
+ HTC to determine whether a message can be sent or
+ must be queued */
int TxCreditsToDist; /* pending credits to distribute on this endpoint, this
- is set by HTC when credit reports arrive.
- The credit distribution functions sets this to zero
- when it distributes the credits */
+ is set by HTC when credit reports arrive.
+ The credit distribution functions sets this to zero
+ when it distributes the credits */
int TxCreditsSeek; /* this is the number of credits that the current pending TX
- packet needs to transmit. This is set by HTC when
- and endpoint needs credits in order to transmit */
+ packet needs to transmit. This is set by HTC when
+ and endpoint needs credits in order to transmit */
int TxCreditSize; /* size in bytes of each credit (set by HTC) */
int TxCreditsPerMaxMsg; /* credits required for a maximum sized messages (set by HTC) */
void *pHTCReserved; /* reserved for HTC use */
int TxQueueDepth; /* current depth of TX queue , i.e. messages waiting for credits
- This field is valid only when HTC_CREDIT_DIST_ACTIVITY_CHANGE
- or HTC_CREDIT_DIST_SEND_COMPLETE is indicated on an endpoint
- that has non-zero credits to recover
- */
+ This field is valid only when HTC_CREDIT_DIST_ACTIVITY_CHANGE
+ or HTC_CREDIT_DIST_SEND_COMPLETE is indicated on an endpoint
+ that has non-zero credits to recover
+ */
} HTC_ENDPOINT_CREDIT_DIST;
#define HTC_EP_ACTIVE ((A_UINT32) (1u << 31))
@@ -235,11 +235,11 @@ typedef struct _HTC_ENDPOINT_CREDIT_DIST {
* there are mandatory and optional codes that must be handled */
typedef enum _HTC_CREDIT_DIST_REASON {
HTC_CREDIT_DIST_SEND_COMPLETE = 0, /* credits available as a result of completed
- send operations (MANDATORY) resulting in credit reports */
+ send operations (MANDATORY) resulting in credit reports */
HTC_CREDIT_DIST_ACTIVITY_CHANGE = 1, /* a change in endpoint activity occured (OPTIONAL) */
HTC_CREDIT_DIST_SEEK_CREDITS, /* an endpoint needs to "seek" credits (OPTIONAL) */
HTC_DUMP_CREDIT_STATE /* for debugging, dump any state information that is kept by
- the distribution function */
+ the distribution function */
} HTC_CREDIT_DIST_REASON;
typedef void (*HTC_CREDIT_DIST_CALLBACK)(void *Context,
@@ -263,7 +263,7 @@ typedef enum _HTC_ENDPOINT_STAT_ACTION {
typedef struct _HTC_ENDPOINT_STATS {
A_UINT32 TxPosted; /* number of TX packets posted to the endpoint */
A_UINT32 TxCreditLowIndications; /* number of times the host set the credit-low flag in a send message on
- this endpoint */
+ this endpoint */
A_UINT32 TxIssued; /* running count of total TX packets issued */
A_UINT32 TxPacketsBundled; /* running count of TX packets that were issued in bundles */
A_UINT32 TxBundles; /* running count of TX bundles that were issued */
@@ -279,7 +279,7 @@ typedef struct _HTC_ENDPOINT_STATS {
A_UINT32 TxCreditsReturned; /* count of credits returned */
A_UINT32 RxReceived; /* count of RX packets received */
A_UINT32 RxLookAheads; /* count of lookahead records
- found in messages received on this endpoint */
+ found in messages received on this endpoint */
A_UINT32 RxPacketsBundled; /* count of recv packets received in a bundle */
A_UINT32 RxBundleLookAheads; /* count of number of bundled lookaheads */
A_UINT32 RxBundleIndFromHdr; /* count of the number of bundle indications from the HTC header */
@@ -292,7 +292,7 @@ typedef struct _HTC_ENDPOINT_STATS {
@desc: Create an instance of HTC over the underlying HIF device
@function name: htc_create
@input: HifDevice - hif device handle,
- pInfo - initialization information
+ pInfo - initialization information
@output:
@return: HTC_HANDLE on success, NULL on failure
@notes:
@@ -300,7 +300,7 @@ typedef struct _HTC_ENDPOINT_STATS {
@see also: htc_destroy
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
HTC_HANDLE htc_create(void *HifDevice,
- HTC_INIT_INFO *pInfo, cdf_device_t osdev);
+ HTC_INIT_INFO *pInfo, qdf_device_t osdev);
/*+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
@desc: Get the underlying HIF device handle
@function name: htc_get_hif_device
@@ -316,18 +316,19 @@ void *htc_get_hif_device(HTC_HANDLE HTCHandle);
@desc: Set credit distribution parameters
@function name: htc_set_credit_distribution
@input: HTCHandle - HTC handle
- pCreditDistCont - caller supplied context to pass into distribution functions
- CreditDistFunc - Distribution function callback
- CreditDistInit - Credit Distribution initialization callback
- ServicePriorityOrder - Array containing list of service IDs, lowest index is highest
- priority
- ListLength - number of elements in ServicePriorityOrder
+ pCreditDistCont - caller supplied context to pass into distribution functions
+ CreditDistFunc - Distribution function callback
+ CreditDistInit - Credit Distribution initialization callback
+ ServicePriorityOrder - Array containing list of service IDs, lowest index
+ is highestpriority
+ ListLength - number of elements in ServicePriorityOrder
@output:
@return:
- @notes: The user can set a custom credit distribution function to handle special requirements
- for each endpoint. A default credit distribution routine can be used by setting
- CreditInitFunc to NULL. The default credit distribution is only provided for simple
- "fair" credit distribution without regard to any prioritization.
+ @notes : The user can set a custom credit distribution function to handle
+ special requirementsfor each endpoint. A default credit distribution
+ routine can be used by setting CreditInitFunc to NULL. The default
+ credit distribution is only provided for simple "fair" credit distribution
+ without regard to any prioritization.
@example:
@see also:
@@ -345,8 +346,8 @@ void htc_set_credit_distribution(HTC_HANDLE HTCHandle,
@output:
@return:
@notes: This API blocks until the target responds with an HTC ready message.
- The caller should not connect services until the target has indicated it is
- ready.
+ The caller should not connect services until the target has indicated it is
+ ready.
@example:
@see also: htc_connect_service
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -357,11 +358,12 @@ A_STATUS htc_wait_target(HTC_HANDLE HTCHandle);
@input: HTCHandle - HTC handle
@output:
@return:
- @notes: This API indicates to the target that the service connection phase is complete
- and the target can freely start all connected services. This API should only be
- called AFTER all service connections have been made. TCStart will issue a
- SETUP_COMPLETE message to the target to indicate that all service connections
- have been made and the target can start communicating over the endpoints.
+ @notes: This API indicates to the target that the service connection phase
+ is completeand the target can freely start all connected services. This
+ API should only be called AFTER all service connections have been made.
+ TCStart will issue a SETUP_COMPLETE message to the target to indicate that
+ all service connections have been made and the target can start
+ communicating over the endpoints.
@example:
@see also: htc_connect_service
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -370,12 +372,12 @@ A_STATUS htc_start(HTC_HANDLE HTCHandle);
@desc: Add receive packet to HTC
@function name: htc_add_receive_pkt
@input: HTCHandle - HTC handle
- pPacket - HTC receive packet to add
+ pPacket - HTC receive packet to add
@output:
@return: A_OK on success
- @notes: user must supply HTC packets for capturing incomming HTC frames. The caller
- must initialize each HTC packet using the SET_HTC_PACKET_INFO_RX_REFILL()
- macro.
+ @notes: user must supply HTC packets for capturing incomming HTC frames.
+ The caller must initialize each HTC packet using the
+ SET_HTC_PACKET_INFO_RX_REFILL() macro.
@example:
@see also:
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -384,11 +386,11 @@ A_STATUS htc_add_receive_pkt(HTC_HANDLE HTCHandle, HTC_PACKET *pPacket);
@desc: Connect to an HTC service
@function name: htc_connect_service
@input: HTCHandle - HTC handle
- pReq - connection details
+ pReq - connection details
@output: pResp - connection response
@return:
- @notes: Service connections must be performed before htc_start. User provides callback handlers
- for various endpoint events.
+ @notes: Service connections must be performed before htc_start.
+ User provides callback handlersfor various endpoint events.
@example:
@see also: htc_start
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -399,8 +401,8 @@ A_STATUS htc_connect_service(HTC_HANDLE HTCHandle,
@desc: HTC register log dump
@function name: htc_dump
@input: HTCHandle - HTC handle
- CmdId - Log command
- start - start/print logs
+ CmdId - Log command
+ start - start/print logs
@output:
@return:
@notes: Register logs will be started/printed.
@@ -414,12 +416,12 @@ void htc_dump(HTC_HANDLE HTCHandle, uint8_t CmdId, bool start);
@desc: Send an HTC packet
@function name: htc_send_pkt
@input: HTCHandle - HTC handle
- pPacket - packet to send
+ pPacket - packet to send
@output:
@return: A_OK
@notes: Caller must initialize packet using SET_HTC_PACKET_INFO_TX() macro.
- This interface is fully asynchronous. On error, HTC SendPkt will
- call the registered Endpoint callback to cleanup the packet.
+ This interface is fully asynchronous. On error, HTC SendPkt will
+ call the registered Endpoint callback to cleanup the packet.
@example:
@see also: htc_flush_endpoint
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -428,15 +430,15 @@ A_STATUS htc_send_pkt(HTC_HANDLE HTCHandle, HTC_PACKET *pPacket);
@desc: Send an HTC packet containing a tx descriptor and data
@function name: htc_send_data_pkt
@input: HTCHandle - HTC handle
- pPacket - packet to send
+ pPacket - packet to send
@output:
@return: A_OK
@notes: Caller must initialize packet using SET_HTC_PACKET_INFO_TX() macro.
- Caller must provide headroom in an initial fragment added to the
- network buffer to store a HTC_FRAME_HDR.
- This interface is fully asynchronous. On error, htc_send_data_pkt will
- call the registered Endpoint EpDataTxComplete callback to cleanup
- the packet.
+ Caller must provide headroom in an initial fragment added to the
+ network buffer to store a HTC_FRAME_HDR.
+ This interface is fully asynchronous. On error, htc_send_data_pkt will
+ call the registered Endpoint EpDataTxComplete callback to cleanup
+ the packet.
@example:
@see also: htc_send_pkt
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -454,7 +456,7 @@ A_STATUS htc_send_data_pkt(HTC_HANDLE HTCHandle, HTC_PACKET *pPacket,
@output:
@return:
@notes: All receive and pending TX packets will
- be flushed.
+ be flushed.
@example:
@see also:
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -466,7 +468,7 @@ void htc_flush_surprise_remove(HTC_HANDLE HTCHandle);
@output:
@return:
@notes: HTC communications is halted. All receive and pending TX packets will
- be flushed.
+ be flushed.
@example:
@see also:
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -486,12 +488,12 @@ void htc_destroy(HTC_HANDLE HTCHandle);
@desc: Flush pending TX packets
@function name: htc_flush_endpoint
@input: HTCHandle - HTC handle
- Endpoint - Endpoint to flush
- Tag - flush tag
+ Endpoint - Endpoint to flush
+ Tag - flush tag
@output:
@return:
@notes: The Tag parameter is used to selectively flush packets with matching tags.
- The value of 0 forces all packets to be flush regardless of tag.
+ The value of 0 forces all packets to be flush regardless of tag.
@example:
@see also: htc_send_pkt
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -512,12 +514,12 @@ void htc_dump_credit_states(HTC_HANDLE HTCHandle);
@desc: Indicate a traffic activity change on an endpoint
@function name: htc_indicate_activity_change
@input: HTCHandle - HTC handle
- Endpoint - endpoint in which activity has changed
- Active - true if active, false if it has become inactive
+ Endpoint - endpoint in which activity has changed
+ Active - true if active, false if it has become inactive
@output:
@return:
@notes: This triggers the registered credit distribution function to
- re-adjust credits for active/inactive endpoints.
+ re-adjust credits for active/inactive endpoints.
@example:
@see also:
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -528,24 +530,26 @@ void htc_indicate_activity_change(HTC_HANDLE HTCHandle,
@desc: Get endpoint statistics
@function name: htc_get_endpoint_statistics
@input: HTCHandle - HTC handle
- Endpoint - Endpoint identifier
- Action - action to take with statistics
+ Endpoint - Endpoint identifier
+ Action - action to take with statistics
@output:
- pStats - statistics that were sampled (can be NULL if Action is HTC_EP_STAT_CLEAR)
+ pStats - statistics that were sampled (can be NULL if Action is HTC_EP_STAT_CLEAR)
@return: true if statistics profiling is enabled, otherwise false.
- @notes: Statistics is a compile-time option and this function may return false
- if HTC is not compiled with profiling.
+ @notes : Statistics is a compile-time option and this function may return
+ false if HTC is not compiled with profiling.
+
+ The caller can specify the statistic "action" to take when sampling
+ the statistics. This includes :
- The caller can specify the statistic "action" to take when sampling
- the statistics. This includes:
+ HTC_EP_STAT_SAMPLE : The pStats structure is filled with the current
+ values.
+ HTC_EP_STAT_SAMPLE_AND_CLEAR : The structure is filled and the current
+ statisticsare cleared.
- HTC_EP_STAT_SAMPLE: The pStats structure is filled with the current values.
- HTC_EP_STAT_SAMPLE_AND_CLEAR: The structure is filled and the current statistics
- are cleared.
- HTC_EP_STAT_CLEA : the statistics are cleared, the called can pass a NULL value for
- pStats
+ HTC_EP_STAT_CLEA : the statistics are cleared, the called can pass a NULL
+ value forpStats
@example:
@see also:
@@ -562,12 +566,13 @@ A_BOOL htc_get_endpoint_statistics(HTC_HANDLE HTCHandle,
@output:
@return:
@notes:
- HTC will block the receiver if the EpRecvAlloc callback fails to provide a packet.
- The caller can use this API to indicate to HTC when resources (buffers) are available
- such that the receiver can be unblocked and HTC may re-attempt fetching the pending message.
+ HTC will block the receiver if the EpRecvAlloc callback fails to provide a
+ packet. The caller can use this API to indicate to HTC when resources
+ (buffers) are available such that the receiver can be unblocked and HTC
+ may re-attempt fetching the pending message.
- This API is not required if the user uses the EpRecvRefill callback or uses the HTCAddReceivePacket()
- API to recycle or provide receive packets to HTC.
+ This API is not required if the user uses the EpRecvRefill callback or uses
+ the HTCAddReceivePacket()API to recycle or provide receive packets to HTC.
@example:
@see also:
@@ -578,17 +583,17 @@ void htc_unblock_recv(HTC_HANDLE HTCHandle);
@desc: send a series of HTC packets
@function name: htc_send_pkts_multiple
@input: HTCHandle - HTC handle
- pPktQueue - local queue holding packets to send
+ pPktQueue - local queue holding packets to send
@output:
@return: A_OK
@notes: Caller must initialize each packet using SET_HTC_PACKET_INFO_TX() macro.
- The queue must only contain packets directed at the same endpoint.
- Caller supplies a pointer to an HTC_PACKET_QUEUE structure holding the TX packets in FIFO order.
- This API will remove the packets from the pkt queue and place them into the HTC Tx Queue
- and bundle messages where possible.
- The caller may allocate the pkt queue on the stack to hold the packets.
- This interface is fully asynchronous. On error, htc_send_pkts will
- call the registered Endpoint callback to cleanup the packet.
+ The queue must only contain packets directed at the same endpoint.
+ Caller supplies a pointer to an HTC_PACKET_QUEUE structure holding the TX packets in FIFO order.
+ This API will remove the packets from the pkt queue and place them into the HTC Tx Queue
+ and bundle messages where possible.
+ The caller may allocate the pkt queue on the stack to hold the packets.
+ This interface is fully asynchronous. On error, htc_send_pkts will
+ call the registered Endpoint callback to cleanup the packet.
@example:
@see also: htc_flush_endpoint
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -599,16 +604,16 @@ A_STATUS htc_send_pkts_multiple(HTC_HANDLE HTCHandle,
@desc: Add multiple receive packets to HTC
@function name: htc_add_receive_pkt_multiple
@input: HTCHandle - HTC handle
- pPktQueue - HTC receive packet queue holding packets to add
+ pPktQueue - HTC receive packet queue holding packets to add
@output:
@return: A_OK on success
@notes: user must supply HTC packets for capturing incomming HTC frames. The caller
- must initialize each HTC packet using the SET_HTC_PACKET_INFO_RX_REFILL()
- macro. The queue must only contain recv packets for the same endpoint.
- Caller supplies a pointer to an HTC_PACKET_QUEUE structure holding the recv packet.
- This API will remove the packets from the pkt queue and place them into internal
- recv packet list.
- The caller may allocate the pkt queue on the stack to hold the packets.
+ must initialize each HTC packet using the SET_HTC_PACKET_INFO_RX_REFILL()
+ macro. The queue must only contain recv packets for the same endpoint.
+ Caller supplies a pointer to an HTC_PACKET_QUEUE structure holding the recv packet.
+ This API will remove the packets from the pkt queue and place them into internal
+ recv packet list.
+ The caller may allocate the pkt queue on the stack to hold the packets.
@example:
@see also:
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++*/
@@ -619,7 +624,7 @@ A_STATUS htc_add_receive_pkt_multiple(HTC_HANDLE HTCHandle,
@desc: Check if an endpoint is marked active
@function name: htc_is_endpoint_active
@input: HTCHandle - HTC handle
- Endpoint - endpoint to check for active state
+ Endpoint - endpoint to check for active state
@output:
@return: returns true if Endpoint is Active
@notes:
@@ -633,7 +638,7 @@ A_BOOL htc_is_endpoint_active(HTC_HANDLE HTCHandle,
@desc: Get the number of recv buffers currently queued into an HTC endpoint
@function name: htc_get_num_recv_buffers
@input: HTCHandle - HTC handle
- Endpoint - endpoint to check
+ Endpoint - endpoint to check
@output:
@return: returns number of buffers in queue
@notes:
@@ -647,7 +652,7 @@ int htc_get_num_recv_buffers(HTC_HANDLE HTCHandle,
@desc: Set the target failure handling callback in HTC layer
@function name: htc_set_target_failure_callback
@input: HTCHandle - HTC handle
- Callback - target failure handling callback
+ Callback - target failure handling callback
@output:
@return:
@notes:
@@ -668,14 +673,16 @@ A_STATUS HTCWaitForPendingRecv(HTC_HANDLE HTCHandle,
struct ol_ath_htc_stats *ieee80211_ioctl_get_htc_stats(HTC_HANDLE
HTCHandle);
-#ifdef HIF_USB
-#define HTCReturnReceivePkt(target,p,osbuf) \
- A_NETBUF_FREE(osbuf); \
- if(p->Status == A_CLONE) { \
- cdf_mem_free(p); \
- }
+#ifdef HIF_USB
+#define HTCReturnReceivePkt(target, p, osbuf) \
+ do { \
+ A_NETBUF_FREE(osbuf); \
+ if (p->Status == A_CLONE) { \
+ qdf_mem_free(p); \
+ } \
+ } while (0)
#else
-#define HTCReturnReceivePkt(target,p,osbuf) htc_add_receive_pkt(target,p)
+#define HTCReturnReceivePkt(target, p, osbuf) htc_add_receive_pkt(target, p)
#endif
#ifdef WLAN_FEATURE_FASTPATH
@@ -708,9 +715,9 @@ void htc_vote_link_down(HTC_HANDLE HTCHandle);
void htc_vote_link_up(HTC_HANDLE HTCHandle);
#ifdef IPA_OFFLOAD
void htc_ipa_get_ce_resource(HTC_HANDLE htc_handle,
- cdf_dma_addr_t *ce_sr_base_paddr,
+ qdf_dma_addr_t *ce_sr_base_paddr,
uint32_t *ce_sr_ring_size,
- cdf_dma_addr_t *ce_reg_paddr);
+ qdf_dma_addr_t *ce_reg_paddr);
#else
#define htc_ipa_get_ce_resource(htc_handle, \
ce_sr_base_paddr, \
diff --git a/htc/htc_debug.h b/htc/htc_debug.h
index fe0c8496f813..c3c0a7ae8a4b 100644
--- a/htc/htc_debug.h
+++ b/htc/htc_debug.h
@@ -30,7 +30,7 @@
#define ATH_MODULE_NAME htc
#include "a_debug.h"
-#include "cdf_trace.h"
+#include "qdf_trace.h"
/* ------- Debug related stuff ------- */
@@ -40,11 +40,11 @@
#define ATH_DEBUG_DUMP ATH_DEBUG_MAKE_MODULE_MASK(3)
#define ATH_DEBUG_SETUP ATH_DEBUG_MAKE_MODULE_MASK(4)
#define HTC_ERROR(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HTC, CDF_TRACE_LEVEL_ERROR, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HTC, QDF_TRACE_LEVEL_ERROR, ## args)
#define HTC_WARN(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HTC, CDF_TRACE_LEVEL_WARN, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HTC, QDF_TRACE_LEVEL_WARN, ## args)
#define HTC_INFO(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HTC, CDF_TRACE_LEVEL_INFO, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HTC, QDF_TRACE_LEVEL_INFO, ## args)
#define HTC_TRACE(args ...) \
- CDF_TRACE(CDF_MODULE_ID_HTC, CDF_TRACE_LEVEL_DEBUG, ## args)
+ QDF_TRACE(QDF_MODULE_ID_HTC, QDF_TRACE_LEVEL_DEBUG, ## args)
#endif /*HTC_DEBUG_H_ */
diff --git a/htc/htc_internal.h b/htc/htc_internal.h
index 9a69036a28e8..638d03b27eab 100644
--- a/htc/htc_internal.h
+++ b/htc/htc_internal.h
@@ -36,10 +36,10 @@ extern "C" {
#include "a_types.h"
#include "osapi_linux.h"
#include <cdf_nbuf.h>
-#include <cdf_types.h>
-#include <cdf_lock.h>
-#include <cdf_softirq_timer.h>
-#include <cdf_atomic.h>
+#include <qdf_types.h>
+#include <qdf_lock.h>
+#include <qdf_timer.h>
+#include <qdf_atomic.h>
#include "hif.h"
#include <htc.h>
#include "htc_api.h"
@@ -105,23 +105,23 @@ typedef struct _HTC_ENDPOINT {
HTC_EP_CALLBACKS EpCallBacks; /* callbacks associated with this endpoint */
HTC_PACKET_QUEUE TxQueue; /* HTC frame buffer TX queue */
int MaxTxQueueDepth; /* max depth of the TX queue before we need to
- call driver's full handler */
+ call driver's full handler */
int MaxMsgLength; /* max length of endpoint message */
uint8_t UL_PipeID;
uint8_t DL_PipeID;
int ul_is_polled; /* Need to call HIF to get tx completion callbacks? */
- cdf_softirq_timer_t ul_poll_timer;
+ qdf_timer_t ul_poll_timer;
int ul_poll_timer_active;
int ul_outstanding_cnt;
int dl_is_polled; /* Need to call HIF to fetch rx? (Not currently supported.) */
#if 0 /* not currently supported */
- cdf_softirq_timer_t dl_poll_timer;
+ qdf_timer_t dl_poll_timer;
#endif
HTC_PACKET_QUEUE TxLookupQueue; /* lookup queue to match netbufs to htc packets */
HTC_PACKET_QUEUE RxBufferHoldQueue; /* temporary hold queue for back compatibility */
A_UINT8 SeqNo; /* TX seq no (helpful) for debugging */
- cdf_atomic_t TxProcessCount; /* serialization */
+ qdf_atomic_t TxProcessCount; /* serialization */
struct _HTC_TARGET *target;
int TxCredits; /* TX credits available on this endpoint */
int TxCreditSize; /* size in bytes of each credit (set by HTC) */
@@ -155,10 +155,10 @@ enum ol_ath_htc_pkt_ecodes {
typedef struct _HTC_TARGET {
struct hif_opaque_softc *hif_dev;
HTC_ENDPOINT endpoint[ENDPOINT_MAX];
- cdf_spinlock_t HTCLock;
- cdf_spinlock_t HTCRxLock;
- cdf_spinlock_t HTCTxLock;
- cdf_spinlock_t HTCCreditLock;
+ qdf_spinlock_t HTCLock;
+ qdf_spinlock_t HTCRxLock;
+ qdf_spinlock_t HTCTxLock;
+ qdf_spinlock_t HTCCreditLock;
A_UINT32 HTCStateFlags;
void *host_handle;
HTC_INIT_INFO HTCInitInfo;
@@ -166,7 +166,7 @@ typedef struct _HTC_TARGET {
HTC_PACKET_QUEUE ControlBufferTXFreeList;
A_UINT8 CtrlResponseBuffer[HTC_MAX_CONTROL_MESSAGE_LENGTH];
int CtrlResponseLength;
- cdf_event_t ctrl_response_valid;
+ qdf_event_t ctrl_response_valid;
A_BOOL CtrlResponseProcessing;
int TotalTransmitCredits;
HTC_SERVICE_TX_CREDIT_ALLOCATION
@@ -178,13 +178,13 @@ typedef struct _HTC_TARGET {
A_UINT32 CurRxSgTotalLen; /* current total length */
A_UINT32 ExpRxSgTotalLen; /* expected total length */
#endif
- cdf_device_t osdev;
+ qdf_device_t osdev;
struct ol_ath_htc_stats htc_pkt_stats;
HTC_PACKET *pBundleFreeList;
A_UINT32 ce_send_cnt;
A_UINT32 TX_comp_cnt;
A_UINT8 MaxMsgsPerHTCBundle;
- cdf_work_t queue_kicker;
+ qdf_work_t queue_kicker;
} HTC_TARGET;
#define HTC_ENABLE_BUNDLE(target) (target->MaxMsgsPerHTCBundle > 1)
@@ -197,14 +197,14 @@ typedef struct _HTC_TARGET {
#define HTC_STATE_STOPPING (1 << 0)
#define HTC_STOPPING(t) ((t)->HTCStateFlags & HTC_STATE_STOPPING)
-#define LOCK_HTC(t) cdf_spin_lock_bh(&(t)->HTCLock);
-#define UNLOCK_HTC(t) cdf_spin_unlock_bh(&(t)->HTCLock);
-#define LOCK_HTC_RX(t) cdf_spin_lock_bh(&(t)->HTCRxLock);
-#define UNLOCK_HTC_RX(t) cdf_spin_unlock_bh(&(t)->HTCRxLock);
-#define LOCK_HTC_TX(t) cdf_spin_lock_bh(&(t)->HTCTxLock);
-#define UNLOCK_HTC_TX(t) cdf_spin_unlock_bh(&(t)->HTCTxLock);
-#define LOCK_HTC_CREDIT(t) cdf_spin_lock_bh(&(t)->HTCCreditLock);
-#define UNLOCK_HTC_CREDIT(t) cdf_spin_unlock_bh(&(t)->HTCCreditLock);
+#define LOCK_HTC(t) qdf_spin_lock_bh(&(t)->HTCLock);
+#define UNLOCK_HTC(t) qdf_spin_unlock_bh(&(t)->HTCLock);
+#define LOCK_HTC_RX(t) qdf_spin_lock_bh(&(t)->HTCRxLock);
+#define UNLOCK_HTC_RX(t) qdf_spin_unlock_bh(&(t)->HTCRxLock);
+#define LOCK_HTC_TX(t) qdf_spin_lock_bh(&(t)->HTCTxLock);
+#define UNLOCK_HTC_TX(t) qdf_spin_unlock_bh(&(t)->HTCTxLock);
+#define LOCK_HTC_CREDIT(t) qdf_spin_lock_bh(&(t)->HTCCreditLock);
+#define UNLOCK_HTC_CREDIT(t) qdf_spin_unlock_bh(&(t)->HTCCreditLock);
#define GET_HTC_TARGET_FROM_HANDLE(hnd) ((HTC_TARGET *)(hnd))
@@ -213,17 +213,23 @@ typedef struct _HTC_TARGET {
#define HTC_POLL_CLEANUP_PERIOD_MS 10 /* milliseconds */
/* Macro to Increment the HTC_PACKET_ERRORS for Tx.*/
-#define OL_ATH_HTC_PKT_ERROR_COUNT_INCR(_target,_ecode) \
+#define OL_ATH_HTC_PKT_ERROR_COUNT_INCR(_target, _ecode) \
do { \
- if(_ecode==GET_HTC_PKT_Q_FAIL) (_target->htc_pkt_stats.htc_get_pkt_q_fail_count)+=1; \
- if(_ecode==HTC_PKT_Q_EMPTY) (_target->htc_pkt_stats.htc_pkt_q_empty_count)+=1; \
- if(_ecode==HTC_SEND_Q_EMPTY) (_target->htc_pkt_stats.htc_send_q_empty_count)+=1; \
- } while(0);
+ if (_ecode == GET_HTC_PKT_Q_FAIL) \
+ (_target->htc_pkt_stats.htc_get_pkt_q_fail_count) += 1 \
+ ; \
+ if (_ecode == HTC_PKT_Q_EMPTY) \
+ (_target->htc_pkt_stats.htc_pkt_q_empty_count) += 1 \
+ ; \
+ if (_ecode == HTC_SEND_Q_EMPTY) \
+ (_target->htc_pkt_stats.htc_send_q_empty_count) += 1 \
+ ; \
+ } while (0);
/* internal HTC functions */
-CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
+QDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
uint8_t pipeID);
-CDF_STATUS htc_tx_completion_handler(void *Context, cdf_nbuf_t netbuf,
+QDF_STATUS htc_tx_completion_handler(void *Context, cdf_nbuf_t netbuf,
unsigned int transferID, uint32_t toeplitz_hash_result);
HTC_PACKET *allocate_htc_bundle_packet(HTC_TARGET *target);
@@ -244,7 +250,7 @@ void htc_control_rx_complete(void *Context, HTC_PACKET *pPacket);
void htc_process_credit_rpt(HTC_TARGET *target,
HTC_CREDIT_REPORT *pRpt,
int NumEntries, HTC_ENDPOINT_ID FromEndpoint);
-void htc_fw_event_handler(void *context, CDF_STATUS status);
+void htc_fw_event_handler(void *context, QDF_STATUS status);
void htc_send_complete_check_cleanup(void *context);
void htc_runtime_pm_init(HTC_TARGET *target);
void htc_kick_queues(void *context);
@@ -256,7 +262,7 @@ static inline void htc_send_complete_poll_timer_stop(HTC_ENDPOINT *
pEndpoint) {
LOCK_HTC_TX(pEndpoint->target);
if (pEndpoint->ul_poll_timer_active) {
- /* cdf_softirq_timer_cancel(&pEndpoint->ul_poll_timer); */
+ /* qdf_timer_stop(&pEndpoint->ul_poll_timer); */
pEndpoint->ul_poll_timer_active = 0;
}
UNLOCK_HTC_TX(pEndpoint->target);
@@ -268,7 +274,7 @@ static inline void htc_send_complete_poll_timer_start(HTC_ENDPOINT *
if (pEndpoint->ul_outstanding_cnt
&& !pEndpoint->ul_poll_timer_active) {
/*
- cdf_softirq_timer_start(
+ qdf_timer_start(
&pEndpoint->ul_poll_timer, HTC_POLL_CLEANUP_PERIOD_MS);
*/
pEndpoint->ul_poll_timer_active = 1;
diff --git a/htc/htc_packet.h b/htc/htc_packet.h
index ca3698f019b7..9e8993e017a3 100644
--- a/htc/htc_packet.h
+++ b/htc/htc_packet.h
@@ -95,11 +95,11 @@ typedef struct _HTC_PACKET {
void *pPktContext; /* caller's per packet specific context */
A_UINT8 *pBufferStart; /* the true buffer start , the caller can
- store the real buffer start here. In
- receive callbacks, the HTC layer sets pBuffer
- to the start of the payload past the header. This
- field allows the caller to reset pBuffer when it
- recycles receive packets back to HTC */
+ store the real buffer start here. In
+ receive callbacks, the HTC layer sets pBuffer
+ to the start of the payload past the header. This
+ field allows the caller to reset pBuffer when it
+ recycles receive packets back to HTC */
/*
* Pointer to the start of the buffer. In the transmit
* direction this points to the start of the payload. In the
@@ -122,40 +122,40 @@ typedef struct _HTC_PACKET {
HTC_PACKET_COMPLETION Completion; /* completion */
void *pContext; /* HTC private completion context */
void *pNetBufContext; /* optimization for network-oriented data, the HTC packet
- can pass the network buffer corresponding to the HTC packet
- lower layers may optimized the transfer knowing this is
- a network buffer */
+ can pass the network buffer corresponding to the HTC packet
+ lower layers may optimized the transfer knowing this is
+ a network buffer */
} HTC_PACKET;
-#define COMPLETE_HTC_PACKET(p,status) \
+#define COMPLETE_HTC_PACKET(p, status) \
{ \
(p)->Status = (status); \
- (p)->Completion((p)->pContext,(p)); \
+ (p)->Completion((p)->pContext, (p)); \
}
-#define INIT_HTC_PACKET_INFO(p,b,len) \
+#define INIT_HTC_PACKET_INFO(p, b, len) \
{ \
(p)->pBufferStart = (b); \
(p)->BufferLength = (len); \
}
/* macro to set an initial RX packet for refilling HTC */
-#define SET_HTC_PACKET_INFO_RX_REFILL(p,c,b,len,ep) \
- { \
+#define SET_HTC_PACKET_INFO_RX_REFILL(p, c, b, len, ep) \
+ do { \
(p)->pPktContext = (c); \
(p)->pBuffer = (b); \
(p)->pBufferStart = (b); \
(p)->BufferLength = (len); \
(p)->Endpoint = (ep); \
- }
+ } while (0)
/* fast macro to recycle an RX packet that will be re-queued to HTC */
#define HTC_PACKET_RESET_RX(p) \
{ (p)->pBuffer = (p)->pBufferStart; (p)->ActualLength = 0; }
/* macro to set packet parameters for TX */
-#define SET_HTC_PACKET_INFO_TX(p,c,b,len,ep,tag) \
- { \
+#define SET_HTC_PACKET_INFO_TX(p, c, b, len, ep, tag) \
+ do { \
(p)->pPktContext = (c); \
(p)->pBuffer = (b); \
(p)->ActualLength = (len); \
@@ -163,10 +163,12 @@ typedef struct _HTC_PACKET {
(p)->PktInfo.AsTx.Tag = (tag); \
(p)->PktInfo.AsTx.Flags = 0; \
(p)->PktInfo.AsTx.SendFlags = 0; \
- }
+ } while (0)
-#define SET_HTC_PACKET_NET_BUF_CONTEXT(p,nb) \
- (p)->pNetBufContext = (nb)
+#define SET_HTC_PACKET_NET_BUF_CONTEXT(p, nb) \
+ do { \
+ (p)->pNetBufContext = (nb); \
+ } while (0)
#define GET_HTC_PACKET_NET_BUF_CONTEXT(p) (p)->pNetBufContext
@@ -179,25 +181,25 @@ typedef struct _HTC_PACKET_QUEUE {
/* initialize queue */
#define INIT_HTC_PACKET_QUEUE(pQ) \
{ \
- DL_LIST_INIT(& (pQ)->QueueHead); \
+ DL_LIST_INIT(&(pQ)->QueueHead); \
(pQ)->Depth = 0; \
}
/* enqueue HTC packet to the tail of the queue */
-#define HTC_PACKET_ENQUEUE(pQ,p) \
- { dl_list_insert_tail(& (pQ)->QueueHead,& (p)->ListLink); \
- (pQ)->Depth ++; \
+#define HTC_PACKET_ENQUEUE(pQ, p) \
+ { dl_list_insert_tail(&(pQ)->QueueHead, &(p)->ListLink); \
+ (pQ)->Depth++; \
}
/* enqueue HTC packet to the tail of the queue */
-#define HTC_PACKET_ENQUEUE_TO_HEAD(pQ,p) \
- { dl_list_insert_head(& (pQ)->QueueHead,& (p)->ListLink); \
- (pQ)->Depth ++; \
+#define HTC_PACKET_ENQUEUE_TO_HEAD(pQ, p) \
+ { dl_list_insert_head(&(pQ)->QueueHead, &(p)->ListLink); \
+ (pQ)->Depth++; \
}
/* test if a queue is empty */
#define HTC_QUEUE_EMPTY(pQ) ((pQ)->Depth == 0)
/* get packet at head without removing it */
-static INLINE HTC_PACKET *htc_get_pkt_at_head(HTC_PACKET_QUEUE *queue)
+static inline HTC_PACKET *htc_get_pkt_at_head(HTC_PACKET_QUEUE *queue)
{
if (queue->Depth == 0) {
return NULL;
@@ -208,14 +210,14 @@ static INLINE HTC_PACKET *htc_get_pkt_at_head(HTC_PACKET_QUEUE *queue)
}
/* remove a packet from a queue, where-ever it is in the queue */
-#define HTC_PACKET_REMOVE(pQ,p) \
+#define HTC_PACKET_REMOVE(pQ, p) \
{ \
- dl_list_remove(& (p)->ListLink); \
- (pQ)->Depth --; \
+ dl_list_remove(&(p)->ListLink); \
+ (pQ)->Depth--; \
}
/* dequeue an HTC packet from the head of the queue */
-static INLINE HTC_PACKET *htc_packet_dequeue(HTC_PACKET_QUEUE *queue)
+static inline HTC_PACKET *htc_packet_dequeue(HTC_PACKET_QUEUE *queue)
{
DL_LIST *pItem = dl_list_remove_item_from_head(&queue->QueueHead);
if (pItem != NULL) {
@@ -226,7 +228,7 @@ static INLINE HTC_PACKET *htc_packet_dequeue(HTC_PACKET_QUEUE *queue)
}
/* dequeue an HTC packet from the tail of the queue */
-static INLINE HTC_PACKET *htc_packet_dequeue_tail(HTC_PACKET_QUEUE *queue)
+static inline HTC_PACKET *htc_packet_dequeue_tail(HTC_PACKET_QUEUE *queue)
{
DL_LIST *pItem = dl_list_remove_item_from_tail(&queue->QueueHead);
if (pItem != NULL) {
@@ -242,9 +244,9 @@ static INLINE HTC_PACKET *htc_packet_dequeue_tail(HTC_PACKET_QUEUE *queue)
#define HTC_GET_TAG_FROM_PKT(p) (p)->PktInfo.AsTx.Tag
/* transfer the packets from one queue to the tail of another queue */
-#define HTC_PACKET_QUEUE_TRANSFER_TO_TAIL(pQDest,pQSrc) \
+#define HTC_PACKET_QUEUE_TRANSFER_TO_TAIL(pQDest, pQSrc) \
{ \
- dl_list_transfer_items_to_tail(&(pQDest)->QueueHead,&(pQSrc)->QueueHead); \
+ dl_list_transfer_items_to_tail(&(pQDest)->QueueHead, &(pQSrc)->QueueHead); \
(pQDest)->Depth += (pQSrc)->Depth; \
(pQSrc)->Depth = 0; \
}
@@ -257,20 +259,20 @@ static INLINE HTC_PACKET *htc_packet_dequeue_tail(HTC_PACKET_QUEUE *queue)
*/
#define HTC_PACKET_QUEUE_TRANSFER_TO_HEAD(pQDest, pQSrc) \
{ \
- dl_list_transfer_items_to_head(&(pQDest)->QueueHead,&(pQSrc)->QueueHead); \
+ dl_list_transfer_items_to_head(&(pQDest)->QueueHead, &(pQSrc)->QueueHead); \
(pQDest)->Depth += (pQSrc)->Depth; \
(pQSrc)->Depth = 0; \
}
/* fast version to init and add a single packet to a queue */
-#define INIT_HTC_PACKET_QUEUE_AND_ADD(pQ,pP) \
+#define INIT_HTC_PACKET_QUEUE_AND_ADD(pQ, pP) \
{ \
- DL_LIST_INIT_AND_ADD(&(pQ)->QueueHead,&(pP)->ListLink) \
+ DL_LIST_INIT_AND_ADD(&(pQ)->QueueHead, &(pP)->ListLink) \
(pQ)->Depth = 1; \
}
#define HTC_PACKET_QUEUE_ITERATE_ALLOW_REMOVE(pQ, pPTemp) \
- ITERATE_OVER_LIST_ALLOW_REMOVE(&(pQ)->QueueHead,(pPTemp), HTC_PACKET, ListLink)
+ ITERATE_OVER_LIST_ALLOW_REMOVE(&(pQ)->QueueHead, (pPTemp), HTC_PACKET, ListLink)
#define HTC_PACKET_QUEUE_ITERATE_IS_VALID(pQ) ITERATE_IS_VALID(&(pQ)->QueueHead)
#define HTC_PACKET_QUEUE_ITERATE_RESET(pQ) ITERATE_RESET(&(pQ)->QueueHead)
diff --git a/htc/htc_recv.c b/htc/htc_recv.c
index 79bf6c6847d0..cdd4e18e2c6f 100644
--- a/htc/htc_recv.c
+++ b/htc/htc_recv.c
@@ -240,7 +240,7 @@ cdf_nbuf_t rx_sg_to_single_netbuf(HTC_TARGET *target)
skb = cdf_nbuf_queue_remove(rx_sg_queue);
do {
cdf_nbuf_peek_header(skb, &anbdata, &anblen);
- cdf_mem_copy(anbdata_new, anbdata, cdf_nbuf_len(skb));
+ qdf_mem_copy(anbdata_new, anbdata, cdf_nbuf_len(skb));
cdf_nbuf_put_tail(new_skb, cdf_nbuf_len(skb));
anbdata_new += cdf_nbuf_len(skb);
cdf_nbuf_free(skb);
@@ -261,10 +261,10 @@ _failed:
}
#endif
-CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
+QDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
uint8_t pipeID)
{
- CDF_STATUS status = CDF_STATUS_SUCCESS;
+ QDF_STATUS status = QDF_STATUS_SUCCESS;
HTC_FRAME_HDR *HtcHdr;
HTC_TARGET *target = (HTC_TARGET *) Context;
uint8_t *netdata;
@@ -310,8 +310,8 @@ CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
htc_ep_id));
debug_dump_bytes((A_UINT8 *) HtcHdr,
sizeof(HTC_FRAME_HDR), "BAD HTC Header");
- status = CDF_STATUS_E_FAILURE;
- CDF_BUG(0);
+ status = QDF_STATUS_E_FAILURE;
+ QDF_BUG(0);
break;
}
@@ -347,8 +347,8 @@ CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
debug_dump_bytes((A_UINT8 *) HtcHdr,
sizeof(HTC_FRAME_HDR),
"BAD RX packet length");
- status = CDF_STATUS_E_FAILURE;
- CDF_BUG(0);
+ status = QDF_STATUS_E_FAILURE;
+ QDF_BUG(0);
break;
#endif
}
@@ -374,7 +374,7 @@ CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
AR_DEBUG_PRINTF(ATH_DEBUG_ERR,
("htc_rx_completion_handler, invalid header (payloadlength should be :%d, CB[0] is:%d)\n",
payloadLen, temp));
- status = CDF_STATUS_E_INVAL;
+ status = QDF_STATUS_E_INVAL;
break;
}
@@ -386,7 +386,7 @@ CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
payloadLen - temp),
temp, htc_ep_id);
if (A_FAILED(temp_status)) {
- status = CDF_STATUS_E_FAILURE;
+ status = QDF_STATUS_E_FAILURE;
break;
}
@@ -420,8 +420,8 @@ CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
* on the endpoint 0 */
AR_DEBUG_PRINTF(ATH_DEBUG_ERR,
("HTC Rx Ctrl still processing\n"));
- status = CDF_STATUS_E_FAILURE;
- CDF_BUG(false);
+ status = QDF_STATUS_E_FAILURE;
+ QDF_BUG(false);
break;
}
@@ -436,7 +436,7 @@ CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
target->CtrlResponseProcessing = true;
UNLOCK_HTC_RX(target);
- cdf_event_set(&target->ctrl_response_valid);
+ qdf_event_set(&target->ctrl_response_valid);
break;
case HTC_MSG_SEND_SUSPEND_COMPLETE:
wow_nack = 0;
@@ -476,10 +476,10 @@ CDF_STATUS htc_rx_completion_handler(void *Context, cdf_nbuf_t netbuf,
* TODO_FIXME */
pPacket = allocate_htc_packet_container(target);
if (NULL == pPacket) {
- status = CDF_STATUS_E_RESOURCES;
+ status = QDF_STATUS_E_RESOURCES;
break;
}
- pPacket->Status = CDF_STATUS_SUCCESS;
+ pPacket->Status = QDF_STATUS_SUCCESS;
pPacket->Endpoint = htc_ep_id;
pPacket->pPktContext = netbuf;
pPacket->pBuffer = cdf_nbuf_data(netbuf) + HTC_HDR_LENGTH;
@@ -601,7 +601,7 @@ void htc_flush_rx_hold_queue(HTC_TARGET *target, HTC_ENDPOINT *pEndpoint)
void htc_recv_init(HTC_TARGET *target)
{
/* Initialize ctrl_response_valid to block */
- cdf_event_init(&target->ctrl_response_valid);
+ qdf_event_create(&target->ctrl_response_valid);
}
/* polling routine to wait for a control packet to be received */
@@ -612,9 +612,9 @@ A_STATUS htc_wait_recv_ctrl_message(HTC_TARGET *target)
AR_DEBUG_PRINTF(ATH_DEBUG_TRC, ("+HTCWaitCtrlMessageRecv\n"));
/* Wait for BMI request/response transaction to complete */
- if (cdf_wait_single_event(&target->ctrl_response_valid,
- cdf_system_msecs_to_ticks(HTC_CONTROL_RX_TIMEOUT))) {
- CDF_BUG(0);
+ if (qdf_wait_single_event(&target->ctrl_response_valid,
+ qdf_system_msecs_to_ticks(HTC_CONTROL_RX_TIMEOUT))) {
+ QDF_BUG(0);
return A_ERROR;
}
diff --git a/htc/htc_send.c b/htc/htc_send.c
index 26821fd542a4..b2a1b701e435 100644
--- a/htc/htc_send.c
+++ b/htc/htc_send.c
@@ -28,7 +28,7 @@
#include "htc_debug.h"
#include "htc_internal.h"
#include <cdf_nbuf.h> /* cdf_nbuf_t */
-#include <cdf_memory.h> /* cdf_mem_malloc */
+#include <qdf_mem.h> /* qdf_mem_malloc */
#include "epping_main.h"
/* #define USB_HIF_SINGLE_PIPE_DATA_SCHED */
@@ -80,7 +80,7 @@ void htc_credit_record(htc_credit_exchange_type type, uint32_t tx_credit,
htc_credit_history_buffer[g_htc_credit_history_idx].type = type;
htc_credit_history_buffer[g_htc_credit_history_idx].time =
- cdf_get_log_timestamp();
+ qdf_get_log_timestamp();
htc_credit_history_buffer[g_htc_credit_history_idx].tx_credit =
tx_credit;
htc_credit_history_buffer[g_htc_credit_history_idx].htc_tx_queue_depth =
@@ -130,11 +130,11 @@ void htc_get_control_endpoint_tx_host_credits(HTC_HANDLE HTCHandle, int *credits
UNLOCK_HTC_TX(target);
}
-static INLINE void restore_tx_packet(HTC_TARGET *target, HTC_PACKET *pPacket)
+static inline void restore_tx_packet(HTC_TARGET *target, HTC_PACKET *pPacket)
{
if (pPacket->PktInfo.AsTx.Flags & HTC_TX_PACKET_FLAG_FIXUP_NETBUF) {
cdf_nbuf_t netbuf = GET_HTC_PACKET_NET_BUF_CONTEXT(pPacket);
- cdf_nbuf_unmap(target->osdev, netbuf, CDF_DMA_TO_DEVICE);
+ cdf_nbuf_unmap(target->osdev, netbuf, QDF_DMA_TO_DEVICE);
cdf_nbuf_pull_head(netbuf, sizeof(HTC_FRAME_HDR));
pPacket->PktInfo.AsTx.Flags &= ~HTC_TX_PACKET_FLAG_FIXUP_NETBUF;
}
@@ -216,17 +216,17 @@ HTC_PACKET *allocate_htc_bundle_packet(HTC_TARGET *target)
if (!netbuf) {
return NULL;
}
- pPacket = cdf_mem_malloc(sizeof(HTC_PACKET));
+ pPacket = qdf_mem_malloc(sizeof(HTC_PACKET));
AR_DEBUG_ASSERT(pPacket);
if (!pPacket) {
cdf_nbuf_free(netbuf);
return NULL;
}
- pQueueSave = cdf_mem_malloc(sizeof(HTC_PACKET_QUEUE));
+ pQueueSave = qdf_mem_malloc(sizeof(HTC_PACKET_QUEUE));
AR_DEBUG_ASSERT(pQueueSave);
if (!pQueueSave) {
cdf_nbuf_free(netbuf);
- cdf_mem_free(pPacket);
+ qdf_mem_free(pPacket);
return NULL;
}
INIT_HTC_PACKET_QUEUE(pQueueSave);
@@ -305,7 +305,7 @@ static A_STATUS htc_send_bundled_netbuf(HTC_TARGET *target,
unsigned char *pBundleBuffer,
HTC_PACKET *pPacketTx)
{
- cdf_size_t data_len;
+ qdf_size_t data_len;
A_STATUS status;
cdf_nbuf_t bundleBuf;
uint32_t data_attr = 0;
@@ -322,7 +322,7 @@ static A_STATUS htc_send_bundled_netbuf(HTC_TARGET *target,
HTC_PACKET_ENQUEUE(&pEndpoint->TxLookupQueue, pPacketTx);
UNLOCK_HTC_TX(target);
#if DEBUG_BUNDLE
- cdf_print(" Send bundle EP%d buffer size:0x%x, total:0x%x, count:%d.\n",
+ qdf_print(" Send bundle EP%d buffer size:0x%x, total:0x%x, count:%d.\n",
pEndpoint->Id,
pEndpoint->TxCreditSize,
data_len, data_len / pEndpoint->TxCreditSize);
@@ -331,7 +331,7 @@ static A_STATUS htc_send_bundled_netbuf(HTC_TARGET *target,
pEndpoint->UL_PipeID,
pEndpoint->Id, data_len, bundleBuf, data_attr);
if (status != A_OK) {
- cdf_print("%s:hif_send_head failed(len=%d).\n", __FUNCTION__,
+ qdf_print("%s:hif_send_head failed(len=%d).\n", __FUNCTION__,
data_len);
}
return status;
@@ -522,7 +522,7 @@ static A_STATUS htc_issue_packets(HTC_TARGET *target,
Flags & HTC_TX_PACKET_FLAG_FIXUP_NETBUF) {
cdf_nbuf_map(target->osdev,
GET_HTC_PACKET_NET_BUF_CONTEXT
- (pPacket), CDF_DMA_TO_DEVICE);
+ (pPacket), QDF_DMA_TO_DEVICE);
}
}
LOCK_HTC_TX(target);
@@ -538,7 +538,7 @@ static A_STATUS htc_issue_packets(HTC_TARGET *target,
HTC_HDR_LENGTH + pPacket->ActualLength,
netbuf, data_attr);
#if DEBUG_BUNDLE
- cdf_print(" Send single EP%d buffer size:0x%x, total:0x%x.\n",
+ qdf_print(" Send single EP%d buffer size:0x%x, total:0x%x.\n",
pEndpoint->Id,
pEndpoint->TxCreditSize,
HTC_HDR_LENGTH + pPacket->ActualLength);
@@ -546,7 +546,7 @@ static A_STATUS htc_issue_packets(HTC_TARGET *target,
target->ce_send_cnt++;
- if (cdf_unlikely(A_FAILED(status))) {
+ if (qdf_unlikely(A_FAILED(status))) {
if (status != A_NO_RESOURCE) {
/* TODO : if more than 1 endpoint maps to the same PipeID it is possible
* to run out of resources in the HIF layer. Don't emit the error */
@@ -585,7 +585,7 @@ static A_STATUS htc_issue_packets(HTC_TARGET *target,
if (pPacket->PktInfo.AsTx.Tag == HTC_TX_PACKET_TAG_RUNTIME_PUT)
hif_pm_runtime_put(target->hif_dev);
}
- if (cdf_unlikely(A_FAILED(status))) {
+ if (qdf_unlikely(A_FAILED(status))) {
#if defined(HIF_USB)
if (pEndpoint->Id >= ENDPOINT_2 && pEndpoint->Id <= ENDPOINT_5)
target->avail_tx_credits +=
@@ -688,7 +688,7 @@ void get_htc_send_packets_credit_based(HTC_TARGET *target,
while (true) {
if (do_pm_get && hif_pm_runtime_get(target->hif_dev)) {
/* bus suspended, runtime resume issued */
- CDF_ASSERT(HTC_PACKET_QUEUE_DEPTH(pQueue) == 0);
+ QDF_ASSERT(HTC_PACKET_QUEUE_DEPTH(pQueue) == 0);
break;
}
@@ -840,7 +840,7 @@ void get_htc_send_packets(HTC_TARGET *target,
if (do_pm_get && hif_pm_runtime_get(target->hif_dev)) {
/* bus suspended, runtime resume issued */
- CDF_ASSERT(HTC_PACKET_QUEUE_DEPTH(pQueue) == 0);
+ QDF_ASSERT(HTC_PACKET_QUEUE_DEPTH(pQueue) == 0);
break;
}
@@ -878,7 +878,7 @@ void get_htc_send_packets(HTC_TARGET *target,
num_frags =
(pPacket->PktInfo.AsTx.
Flags & HTC_TX_PACKET_FLAG_FIXUP_NETBUF) ? 1
- /* WMI messages are in a single-fragment network buffer */ :
+ /* WMI messages are in a single-fragment network buffer */ :
cdf_nbuf_get_num_frags(GET_HTC_PACKET_NET_BUF_CONTEXT
(pPacket));
Resources -= num_frags;
@@ -1043,11 +1043,11 @@ static HTC_SEND_QUEUE_RESULT htc_try_send(HTC_TARGET *target,
}
/* increment tx processing count on entry */
- cdf_atomic_inc(&pEndpoint->TxProcessCount);
- if (cdf_atomic_read(&pEndpoint->TxProcessCount) > 1) {
+ qdf_atomic_inc(&pEndpoint->TxProcessCount);
+ if (qdf_atomic_read(&pEndpoint->TxProcessCount) > 1) {
/* another thread or task is draining the TX queues on this endpoint
* that thread will reset the tx processing count when the queue is drained */
- cdf_atomic_dec(&pEndpoint->TxProcessCount);
+ qdf_atomic_dec(&pEndpoint->TxProcessCount);
UNLOCK_HTC_TX(target);
AR_DEBUG_PRINTF(ATH_DEBUG_SEND, ("-htc_try_send (busy) \n"));
return HTC_SEND_QUEUE_OK;
@@ -1125,7 +1125,7 @@ static HTC_SEND_QUEUE_RESULT htc_try_send(HTC_TARGET *target,
UNLOCK_HTC_TX(target);
/* done with this endpoint, we can clear the count */
- cdf_atomic_init(&pEndpoint->TxProcessCount);
+ qdf_atomic_init(&pEndpoint->TxProcessCount);
AR_DEBUG_PRINTF(ATH_DEBUG_SEND, ("-htc_try_send: \n"));
@@ -1157,15 +1157,15 @@ static A_UINT16 htc_send_pkts_sched_check(HTC_HANDLE HTCHandle, HTC_ENDPOINT_ID
switch (id) {
case ENDPOINT_2: /* BE */
- return (acQueueStatus[0] && acQueueStatus[2]
- && acQueueStatus[3]);
+ return acQueueStatus[0] && acQueueStatus[2]
+ && acQueueStatus[3];
case ENDPOINT_3: /* BK */
- return (acQueueStatus[0] && acQueueStatus[1] && acQueueStatus[2]
- && acQueueStatus[3]);
+ return acQueueStatus[0] && acQueueStatus[1] && acQueueStatus[2]
+ && acQueueStatus[3];
case ENDPOINT_4: /* VI */
- return (acQueueStatus[2] && acQueueStatus[3]);
+ return acQueueStatus[2] && acQueueStatus[3];
case ENDPOINT_5: /* VO */
- return (acQueueStatus[3]);
+ return acQueueStatus[3];
default:
return 0;
}
@@ -1291,7 +1291,7 @@ A_STATUS htc_send_pkts_multiple(HTC_HANDLE HTCHandle, HTC_PACKET_QUEUE *pPktQueu
*/
cdf_nbuf_map(target->osdev,
GET_HTC_PACKET_NET_BUF_CONTEXT(pPacket),
- CDF_DMA_TO_DEVICE);
+ QDF_DMA_TO_DEVICE);
pPacket->PktInfo.AsTx.Flags |= HTC_TX_PACKET_FLAG_FIXUP_NETBUF;
}
@@ -1410,7 +1410,7 @@ A_STATUS htc_send_data_pkt(HTC_HANDLE HTCHandle, cdf_nbuf_t netbuf, int Epid,
pEndpoint->SeqNo++;
NBUF_UPDATE_TX_PKT_COUNT(netbuf, NBUF_TX_PKT_HTC);
- DPTRACE(cdf_dp_trace(netbuf, CDF_DP_TRACE_HTC_PACKET_PTR_RECORD,
+ DPTRACE(qdf_dp_trace(netbuf, QDF_DP_TRACE_HTC_PACKET_PTR_RECORD,
(uint8_t *)(cdf_nbuf_data(netbuf)),
sizeof(cdf_nbuf_data(netbuf))));
status = hif_send_head(target->hif_dev,
@@ -1492,14 +1492,14 @@ A_STATUS htc_send_data_pkt(HTC_HANDLE HTCHandle, HTC_PACKET *pPacket,
}
/* increment tx processing count on entry */
- cdf_atomic_inc(&pEndpoint->TxProcessCount);
- if (cdf_atomic_read(&pEndpoint->TxProcessCount) > 1) {
+ qdf_atomic_inc(&pEndpoint->TxProcessCount);
+ if (qdf_atomic_read(&pEndpoint->TxProcessCount) > 1) {
/*
* Another thread or task is draining the TX queues on this endpoint.
* That thread will reset the tx processing count when the queue is
* drained.
*/
- cdf_atomic_dec(&pEndpoint->TxProcessCount);
+ qdf_atomic_dec(&pEndpoint->TxProcessCount);
UNLOCK_HTC_TX(target);
return A_OK;
}
@@ -1545,7 +1545,7 @@ A_STATUS htc_send_data_pkt(HTC_HANDLE HTCHandle, HTC_PACKET *pPacket,
UNLOCK_HTC_TX(target);
}
NBUF_UPDATE_TX_PKT_COUNT(netbuf, NBUF_TX_PKT_HTC);
- DPTRACE(cdf_dp_trace(netbuf, CDF_DP_TRACE_HTC_PACKET_PTR_RECORD,
+ DPTRACE(qdf_dp_trace(netbuf, QDF_DP_TRACE_HTC_PACKET_PTR_RECORD,
(uint8_t *)(cdf_nbuf_data(netbuf)),
sizeof(cdf_nbuf_data(netbuf))));
@@ -1579,13 +1579,13 @@ A_STATUS htc_send_data_pkt(HTC_HANDLE HTCHandle, HTC_PACKET *pPacket,
HTC_HDR_LENGTH + pPacket->ActualLength,
netbuf, data_attr);
#if DEBUG_BUNDLE
- cdf_print(" Send single EP%d buffer size:0x%x, total:0x%x.\n",
+ qdf_print(" Send single EP%d buffer size:0x%x, total:0x%x.\n",
pEndpoint->Id,
pEndpoint->TxCreditSize,
HTC_HDR_LENGTH + pPacket->ActualLength);
#endif
- if (cdf_unlikely(A_FAILED(status))) {
+ if (qdf_unlikely(A_FAILED(status))) {
LOCK_HTC_TX(target);
pEndpoint->ul_outstanding_cnt--;
/* remove this packet from the tx completion queue */
@@ -1612,7 +1612,7 @@ A_STATUS htc_send_data_pkt(HTC_HANDLE HTCHandle, HTC_PACKET *pPacket,
}
}
/* done with this endpoint, we can clear the count */
- cdf_atomic_init(&pEndpoint->TxProcessCount);
+ qdf_atomic_init(&pEndpoint->TxProcessCount);
if (pEndpoint->ul_is_polled) {
/*
@@ -1651,7 +1651,7 @@ static HTC_PACKET *htc_lookup_tx_packet(HTC_TARGET *target,
/* Dequeue first packet directly because of in-order completion */
pPacket = htc_packet_dequeue(&pEndpoint->TxLookupQueue);
- if (cdf_unlikely(!pPacket)) {
+ if (qdf_unlikely(!pPacket)) {
UNLOCK_HTC_TX(target);
return NULL;
}
@@ -1697,7 +1697,7 @@ static HTC_PACKET *htc_lookup_tx_packet(HTC_TARGET *target,
return pFoundPacket;
}
-CDF_STATUS htc_tx_completion_handler(void *Context,
+QDF_STATUS htc_tx_completion_handler(void *Context,
cdf_nbuf_t netbuf, unsigned int EpID,
uint32_t toeplitz_hash_result)
{
@@ -1705,8 +1705,8 @@ CDF_STATUS htc_tx_completion_handler(void *Context,
HTC_ENDPOINT *pEndpoint;
HTC_PACKET *pPacket;
#ifdef USB_HIF_SINGLE_PIPE_DATA_SCHED
- HTC_ENDPOINT_ID eid[DATA_EP_SIZE] =
- { ENDPOINT_5, ENDPOINT_4, ENDPOINT_2, ENDPOINT_3 };
+ HTC_ENDPOINT_ID eid[DATA_EP_SIZE] = { ENDPOINT_5, ENDPOINT_4,
+ ENDPOINT_2, ENDPOINT_3 };
int epidIdx;
A_UINT16 resourcesThresh[DATA_EP_SIZE]; /* urb resources */
A_UINT16 resources;
@@ -1739,11 +1739,11 @@ CDF_STATUS htc_tx_completion_handler(void *Context,
}
HTC_PACKET_QUEUE_ITERATE_END;
free_htc_bundle_packet(target, pPacket);
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
/* will be giving this buffer back to upper layers */
netbuf = NULL;
- pPacket->Status = CDF_STATUS_SUCCESS;
+ pPacket->Status = QDF_STATUS_SUCCESS;
send_packet_completion(target, pPacket);
} while (false);
@@ -1755,7 +1755,7 @@ CDF_STATUS htc_tx_completion_handler(void *Context,
htc_try_send(target, pEndpoint, NULL);
}
- return CDF_STATUS_SUCCESS;
+ return QDF_STATUS_SUCCESS;
}
/* callback when TX resources become available */
@@ -1928,8 +1928,8 @@ void htc_process_credit_rpt(HTC_TARGET *target, HTC_CREDIT_REPORT *pRpt,
#endif
#if defined(HIF_USB)
if (pEndpoint->Id >= ENDPOINT_2 && pEndpoint->Id <= ENDPOINT_5) {
- HTC_ENDPOINT_ID eid[DATA_EP_SIZE] =
- { ENDPOINT_5, ENDPOINT_4, ENDPOINT_2, ENDPOINT_3 };
+ HTC_ENDPOINT_ID eid[DATA_EP_SIZE] = { ENDPOINT_5,
+ ENDPOINT_4, ENDPOINT_2, ENDPOINT_3 };
int epid_idx;
target->avail_tx_credits += rpt_credits;
@@ -1998,5 +1998,5 @@ struct ol_ath_htc_stats *ieee80211_ioctl_get_htc_stats(HTC_HANDLE HTCHandle)
{
HTC_TARGET *target = GET_HTC_TARGET_FROM_HANDLE(HTCHandle);
- return (&(target->htc_pkt_stats));
+ return &(target->htc_pkt_stats);
}
diff --git a/htc/htc_services.c b/htc/htc_services.c
index d2cb0dd27851..99fc20e25473 100644
--- a/htc/htc_services.c
+++ b/htc/htc_services.c
@@ -310,7 +310,7 @@ A_STATUS htc_connect_service(HTC_HANDLE HTCHandle,
pEndpoint->TxCreditsPerMaxMsg++;
}
#if DEBUG_CREDIT
- cdf_print(" Endpoint%d initial credit:%d, size:%d.\n",
+ qdf_print(" Endpoint%d initial credit:%d, size:%d.\n",
pEndpoint->Id, pEndpoint->TxCredits,
pEndpoint->TxCreditSize);
#endif
@@ -328,14 +328,14 @@ A_STATUS htc_connect_service(HTC_HANDLE HTCHandle,
break;
}
- cdf_assert(!pEndpoint->dl_is_polled); /* not currently supported */
+ qdf_assert(!pEndpoint->dl_is_polled); /* not currently supported */
if (pEndpoint->ul_is_polled) {
- cdf_softirq_timer_init(target->osdev,
+ qdf_timer_init(target->osdev,
&pEndpoint->ul_poll_timer,
htc_send_complete_check_cleanup,
pEndpoint,
- CDF_TIMER_TYPE_SW);
+ QDF_TIMER_TYPE_SW);
}
AR_DEBUG_PRINTF(ATH_DEBUG_SETUP,
@@ -372,7 +372,7 @@ void htc_set_credit_distribution(HTC_HANDLE HTCHandle,
}
-void htc_fw_event_handler(void *context, CDF_STATUS status)
+void htc_fw_event_handler(void *context, QDF_STATUS status)
{
HTC_TARGET *target = (HTC_TARGET *) context;
HTC_INIT_INFO *initInfo = &target->HTCInitInfo;