1/*
2 * Copyright (c) 2018 Redpine Signals Inc.
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#include "rsi_main.h"
18#include "rsi_coex.h"
19#include "rsi_mgmt.h"
20#include "rsi_hal.h"
21
22static enum rsi_coex_queues rsi_coex_determine_coex_q
23			(struct rsi_coex_ctrl_block *coex_cb)
24{
25	enum rsi_coex_queues q_num = RSI_COEX_Q_INVALID;
26
27	if (skb_queue_len(&coex_cb->coex_tx_qs[RSI_COEX_Q_COMMON]) > 0)
28		q_num = RSI_COEX_Q_COMMON;
29	if (skb_queue_len(&coex_cb->coex_tx_qs[RSI_COEX_Q_BT]) > 0)
30		q_num = RSI_COEX_Q_BT;
31	if (skb_queue_len(&coex_cb->coex_tx_qs[RSI_COEX_Q_WLAN]) > 0)
32		q_num = RSI_COEX_Q_WLAN;
33
34	return q_num;
35}
36
37static void rsi_coex_sched_tx_pkts(struct rsi_coex_ctrl_block *coex_cb)
38{
39	enum rsi_coex_queues coex_q = RSI_COEX_Q_INVALID;
40	struct sk_buff *skb;
41
42	do {
43		coex_q = rsi_coex_determine_coex_q(coex_cb);
44		rsi_dbg(INFO_ZONE, "queue = %d\n", coex_q);
45
46		if (coex_q == RSI_COEX_Q_BT) {
47			skb = skb_dequeue(&coex_cb->coex_tx_qs[RSI_COEX_Q_BT]);
48			rsi_send_bt_pkt(coex_cb->priv, skb);
49		}
50	} while (coex_q != RSI_COEX_Q_INVALID);
51}
52
53static void rsi_coex_scheduler_thread(struct rsi_common *common)
54{
55	struct rsi_coex_ctrl_block *coex_cb = common->coex_cb;
56	u32 timeout = EVENT_WAIT_FOREVER;
57
58	do {
59		rsi_wait_event(&coex_cb->coex_tx_thread.event, timeout);
60		rsi_reset_event(&coex_cb->coex_tx_thread.event);
61
62		rsi_coex_sched_tx_pkts(coex_cb);
63	} while (atomic_read(&coex_cb->coex_tx_thread.thread_done) == 0);
64
65	kthread_complete_and_exit(&coex_cb->coex_tx_thread.completion, 0);
66}
67
68int rsi_coex_recv_pkt(struct rsi_common *common, u8 *msg)
69{
70	u8 msg_type = msg[RSI_RX_DESC_MSG_TYPE_OFFSET];
71
72	switch (msg_type) {
73	case COMMON_CARD_READY_IND:
74		rsi_dbg(INFO_ZONE, "common card ready received\n");
75		common->hibernate_resume = false;
76		rsi_handle_card_ready(common, msg);
77		break;
78	case SLEEP_NOTIFY_IND:
79		rsi_dbg(INFO_ZONE, "sleep notify received\n");
80		rsi_mgmt_pkt_recv(common, msg);
81		break;
82	}
83
84	return 0;
85}
86
87static inline int rsi_map_coex_q(u8 hal_queue)
88{
89	switch (hal_queue) {
90	case RSI_COEX_Q:
91		return RSI_COEX_Q_COMMON;
92	case RSI_WLAN_Q:
93		return RSI_COEX_Q_WLAN;
94	case RSI_BT_Q:
95		return RSI_COEX_Q_BT;
96	}
97	return RSI_COEX_Q_INVALID;
98}
99
100int rsi_coex_send_pkt(void *priv, struct sk_buff *skb, u8 hal_queue)
101{
102	struct rsi_common *common = priv;
103	struct rsi_coex_ctrl_block *coex_cb = common->coex_cb;
104	struct skb_info *tx_params = NULL;
105	enum rsi_coex_queues coex_q;
106	int status;
107
108	coex_q = rsi_map_coex_q(hal_queue);
109	if (coex_q == RSI_COEX_Q_INVALID) {
110		rsi_dbg(ERR_ZONE, "Invalid coex queue\n");
111		return -EINVAL;
112	}
113	if (coex_q != RSI_COEX_Q_COMMON &&
114	    coex_q != RSI_COEX_Q_WLAN) {
115		skb_queue_tail(&coex_cb->coex_tx_qs[coex_q], skb);
116		rsi_set_event(&coex_cb->coex_tx_thread.event);
117		return 0;
118	}
119	if (common->iface_down) {
120		tx_params =
121			(struct skb_info *)&IEEE80211_SKB_CB(skb)->driver_data;
122
123		if (!(tx_params->flags & INTERNAL_MGMT_PKT)) {
124			rsi_indicate_tx_status(common->priv, skb, -EINVAL);
125			return 0;
126		}
127	}
128
129	/* Send packet to hal */
130	if (skb->priority == MGMT_SOFT_Q)
131		status = rsi_send_mgmt_pkt(common, skb);
132	else
133		status = rsi_send_data_pkt(common, skb);
134
135	return status;
136}
137
138int rsi_coex_attach(struct rsi_common *common)
139{
140	struct rsi_coex_ctrl_block *coex_cb;
141	int cnt;
142
143	coex_cb = kzalloc(sizeof(*coex_cb), GFP_KERNEL);
144	if (!coex_cb)
145		return -ENOMEM;
146
147	common->coex_cb = (void *)coex_cb;
148	coex_cb->priv = common;
149
150	/* Initialize co-ex queues */
151	for (cnt = 0; cnt < NUM_COEX_TX_QUEUES; cnt++)
152		skb_queue_head_init(&coex_cb->coex_tx_qs[cnt]);
153	rsi_init_event(&coex_cb->coex_tx_thread.event);
154
155	/* Initialize co-ex thread */
156	if (rsi_create_kthread(common,
157			       &coex_cb->coex_tx_thread,
158			       rsi_coex_scheduler_thread,
159			       "Coex-Tx-Thread")) {
160		rsi_dbg(ERR_ZONE, "%s: Unable to init tx thrd\n", __func__);
161		kfree(coex_cb);
162		return -EINVAL;
163	}
164	return 0;
165}
166
167void rsi_coex_detach(struct rsi_common *common)
168{
169	struct rsi_coex_ctrl_block *coex_cb = common->coex_cb;
170	int cnt;
171
172	rsi_kill_thread(&coex_cb->coex_tx_thread);
173
174	for (cnt = 0; cnt < NUM_COEX_TX_QUEUES; cnt++)
175		skb_queue_purge(&coex_cb->coex_tx_qs[cnt]);
176
177	kfree(coex_cb);
178}
179