1// SPDX-License-Identifier: ISC
2/*
3 * Copyright (c) 2005-2011 Atheros Communications Inc.
4 * Copyright (c) 2011-2017 Qualcomm Atheros, Inc.
5 * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
6 */
7
8#include "core.h"
9#include "hif.h"
10#include "debug.h"
11
12/********/
13/* Send */
14/********/
15
16static void ath10k_htc_control_tx_complete(struct ath10k *ar,
17					   struct sk_buff *skb)
18{
19	kfree_skb(skb);
20}
21
22static struct sk_buff *ath10k_htc_build_tx_ctrl_skb(void *ar)
23{
24	struct sk_buff *skb;
25	struct ath10k_skb_cb *skb_cb;
26
27	skb = dev_alloc_skb(ATH10K_HTC_CONTROL_BUFFER_SIZE);
28	if (!skb)
29		return NULL;
30
31	skb_reserve(skb, 20); /* FIXME: why 20 bytes? */
32	WARN_ONCE((unsigned long)skb->data & 3, "unaligned skb");
33
34	skb_cb = ATH10K_SKB_CB(skb);
35	memset(skb_cb, 0, sizeof(*skb_cb));
36
37	ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: skb %pK\n", __func__, skb);
38	return skb;
39}
40
41static inline void ath10k_htc_restore_tx_skb(struct ath10k_htc *htc,
42					     struct sk_buff *skb)
43{
44	struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
45
46	if (htc->ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
47		dma_unmap_single(htc->ar->dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
48	skb_pull(skb, sizeof(struct ath10k_htc_hdr));
49}
50
51void ath10k_htc_notify_tx_completion(struct ath10k_htc_ep *ep,
52				     struct sk_buff *skb)
53{
54	struct ath10k *ar = ep->htc->ar;
55	struct ath10k_htc_hdr *hdr;
56
57	ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: ep %d skb %pK\n", __func__,
58		   ep->eid, skb);
59
60	/* A corner case where the copy completion is reaching to host but still
61	 * copy engine is processing it due to which host unmaps corresponding
62	 * memory and causes SMMU fault, hence as workaround adding delay
63	 * the unmapping memory to avoid SMMU faults.
64	 */
65	if (ar->hw_params.delay_unmap_buffer &&
66	    ep->ul_pipe_id == 3)
67		mdelay(2);
68
69	hdr = (struct ath10k_htc_hdr *)skb->data;
70	ath10k_htc_restore_tx_skb(ep->htc, skb);
71
72	if (!ep->ep_ops.ep_tx_complete) {
73		ath10k_warn(ar, "no tx handler for eid %d\n", ep->eid);
74		dev_kfree_skb_any(skb);
75		return;
76	}
77
78	if (hdr->flags & ATH10K_HTC_FLAG_SEND_BUNDLE) {
79		dev_kfree_skb_any(skb);
80		return;
81	}
82
83	ep->ep_ops.ep_tx_complete(ep->htc->ar, skb);
84}
85EXPORT_SYMBOL(ath10k_htc_notify_tx_completion);
86
87static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
88				      struct sk_buff *skb)
89{
90	struct ath10k_htc_hdr *hdr;
91
92	hdr = (struct ath10k_htc_hdr *)skb->data;
93	memset(hdr, 0, sizeof(struct ath10k_htc_hdr));
94
95	hdr->eid = ep->eid;
96	hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr));
97	hdr->flags = 0;
98	if (ep->tx_credit_flow_enabled && !ep->bundle_tx)
99		hdr->flags |= ATH10K_HTC_FLAG_NEED_CREDIT_UPDATE;
100
101	spin_lock_bh(&ep->htc->tx_lock);
102	hdr->seq_no = ep->seq_no++;
103	spin_unlock_bh(&ep->htc->tx_lock);
104}
105
106static int ath10k_htc_consume_credit(struct ath10k_htc_ep *ep,
107				     unsigned int len,
108				     bool consume)
109{
110	struct ath10k_htc *htc = ep->htc;
111	struct ath10k *ar = htc->ar;
112	enum ath10k_htc_ep_id eid = ep->eid;
113	int credits, ret = 0;
114
115	if (!ep->tx_credit_flow_enabled)
116		return 0;
117
118	credits = DIV_ROUND_UP(len, ep->tx_credit_size);
119	spin_lock_bh(&htc->tx_lock);
120
121	if (ep->tx_credits < credits) {
122		ath10k_dbg(ar, ATH10K_DBG_HTC,
123			   "htc insufficient credits ep %d required %d available %d consume %d\n",
124			   eid, credits, ep->tx_credits, consume);
125		ret = -EAGAIN;
126		goto unlock;
127	}
128
129	if (consume) {
130		ep->tx_credits -= credits;
131		ath10k_dbg(ar, ATH10K_DBG_HTC,
132			   "htc ep %d consumed %d credits total %d\n",
133			   eid, credits, ep->tx_credits);
134	}
135
136unlock:
137	spin_unlock_bh(&htc->tx_lock);
138	return ret;
139}
140
141static void ath10k_htc_release_credit(struct ath10k_htc_ep *ep, unsigned int len)
142{
143	struct ath10k_htc *htc = ep->htc;
144	struct ath10k *ar = htc->ar;
145	enum ath10k_htc_ep_id eid = ep->eid;
146	int credits;
147
148	if (!ep->tx_credit_flow_enabled)
149		return;
150
151	credits = DIV_ROUND_UP(len, ep->tx_credit_size);
152	spin_lock_bh(&htc->tx_lock);
153	ep->tx_credits += credits;
154	ath10k_dbg(ar, ATH10K_DBG_HTC,
155		   "htc ep %d reverted %d credits back total %d\n",
156		   eid, credits, ep->tx_credits);
157	spin_unlock_bh(&htc->tx_lock);
158
159	if (ep->ep_ops.ep_tx_credits)
160		ep->ep_ops.ep_tx_credits(htc->ar);
161}
162
163int ath10k_htc_send(struct ath10k_htc *htc,
164		    enum ath10k_htc_ep_id eid,
165		    struct sk_buff *skb)
166{
167	struct ath10k *ar = htc->ar;
168	struct ath10k_htc_ep *ep = &htc->endpoint[eid];
169	struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
170	struct ath10k_hif_sg_item sg_item;
171	struct device *dev = htc->ar->dev;
172	int ret;
173	unsigned int skb_len;
174
175	if (htc->ar->state == ATH10K_STATE_WEDGED)
176		return -ECOMM;
177
178	if (eid >= ATH10K_HTC_EP_COUNT) {
179		ath10k_warn(ar, "Invalid endpoint id: %d\n", eid);
180		return -ENOENT;
181	}
182
183	skb_push(skb, sizeof(struct ath10k_htc_hdr));
184
185	skb_len = skb->len;
186	ret = ath10k_htc_consume_credit(ep, skb_len, true);
187	if (ret)
188		goto err_pull;
189
190	ath10k_htc_prepare_tx_skb(ep, skb);
191
192	skb_cb->eid = eid;
193	if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL) {
194		skb_cb->paddr = dma_map_single(dev, skb->data, skb->len,
195					       DMA_TO_DEVICE);
196		ret = dma_mapping_error(dev, skb_cb->paddr);
197		if (ret) {
198			ret = -EIO;
199			goto err_credits;
200		}
201	}
202
203	sg_item.transfer_id = ep->eid;
204	sg_item.transfer_context = skb;
205	sg_item.vaddr = skb->data;
206	sg_item.paddr = skb_cb->paddr;
207	sg_item.len = skb->len;
208
209	ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
210	if (ret)
211		goto err_unmap;
212
213	return 0;
214
215err_unmap:
216	if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
217		dma_unmap_single(dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
218err_credits:
219	ath10k_htc_release_credit(ep, skb_len);
220err_pull:
221	skb_pull(skb, sizeof(struct ath10k_htc_hdr));
222	return ret;
223}
224
225void ath10k_htc_tx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
226{
227	struct ath10k_htc *htc = &ar->htc;
228	struct ath10k_skb_cb *skb_cb;
229	struct ath10k_htc_ep *ep;
230
231	if (WARN_ON_ONCE(!skb))
232		return;
233
234	skb_cb = ATH10K_SKB_CB(skb);
235	ep = &htc->endpoint[skb_cb->eid];
236
237	ath10k_htc_notify_tx_completion(ep, skb);
238	/* the skb now belongs to the completion handler */
239}
240EXPORT_SYMBOL(ath10k_htc_tx_completion_handler);
241
242/***********/
243/* Receive */
244/***********/
245
246static void
247ath10k_htc_process_credit_report(struct ath10k_htc *htc,
248				 const struct ath10k_htc_credit_report *report,
249				 int len,
250				 enum ath10k_htc_ep_id eid)
251{
252	struct ath10k *ar = htc->ar;
253	struct ath10k_htc_ep *ep;
254	int i, n_reports;
255
256	if (len % sizeof(*report))
257		ath10k_warn(ar, "Uneven credit report len %d", len);
258
259	n_reports = len / sizeof(*report);
260
261	spin_lock_bh(&htc->tx_lock);
262	for (i = 0; i < n_reports; i++, report++) {
263		if (report->eid >= ATH10K_HTC_EP_COUNT)
264			break;
265
266		ep = &htc->endpoint[report->eid];
267		ep->tx_credits += report->credits;
268
269		ath10k_dbg(ar, ATH10K_DBG_HTC, "htc ep %d got %d credits (total %d)\n",
270			   report->eid, report->credits, ep->tx_credits);
271
272		if (ep->ep_ops.ep_tx_credits) {
273			spin_unlock_bh(&htc->tx_lock);
274			ep->ep_ops.ep_tx_credits(htc->ar);
275			spin_lock_bh(&htc->tx_lock);
276		}
277	}
278	spin_unlock_bh(&htc->tx_lock);
279}
280
281static int
282ath10k_htc_process_lookahead(struct ath10k_htc *htc,
283			     const struct ath10k_htc_lookahead_report *report,
284			     int len,
285			     enum ath10k_htc_ep_id eid,
286			     void *next_lookaheads,
287			     int *next_lookaheads_len)
288{
289	struct ath10k *ar = htc->ar;
290
291	/* Invalid lookahead flags are actually transmitted by
292	 * the target in the HTC control message.
293	 * Since this will happen at every boot we silently ignore
294	 * the lookahead in this case
295	 */
296	if (report->pre_valid != ((~report->post_valid) & 0xFF))
297		return 0;
298
299	if (next_lookaheads && next_lookaheads_len) {
300		ath10k_dbg(ar, ATH10K_DBG_HTC,
301			   "htc rx lookahead found pre_valid 0x%x post_valid 0x%x\n",
302			   report->pre_valid, report->post_valid);
303
304		/* look ahead bytes are valid, copy them over */
305		memcpy((u8 *)next_lookaheads, report->lookahead, 4);
306
307		*next_lookaheads_len = 1;
308	}
309
310	return 0;
311}
312
313static int
314ath10k_htc_process_lookahead_bundle(struct ath10k_htc *htc,
315				    const struct ath10k_htc_lookahead_bundle *report,
316				    int len,
317				    enum ath10k_htc_ep_id eid,
318				    void *next_lookaheads,
319				    int *next_lookaheads_len)
320{
321	struct ath10k *ar = htc->ar;
322	int bundle_cnt = len / sizeof(*report);
323
324	if (!bundle_cnt || (bundle_cnt > htc->max_msgs_per_htc_bundle)) {
325		ath10k_warn(ar, "Invalid lookahead bundle count: %d\n",
326			    bundle_cnt);
327		return -EINVAL;
328	}
329
330	if (next_lookaheads && next_lookaheads_len) {
331		int i;
332
333		for (i = 0; i < bundle_cnt; i++) {
334			memcpy(((u8 *)next_lookaheads) + 4 * i,
335			       report->lookahead, 4);
336			report++;
337		}
338
339		*next_lookaheads_len = bundle_cnt;
340	}
341
342	return 0;
343}
344
345int ath10k_htc_process_trailer(struct ath10k_htc *htc,
346			       u8 *buffer,
347			       int length,
348			       enum ath10k_htc_ep_id src_eid,
349			       void *next_lookaheads,
350			       int *next_lookaheads_len)
351{
352	struct ath10k_htc_lookahead_bundle *bundle;
353	struct ath10k *ar = htc->ar;
354	int status = 0;
355	struct ath10k_htc_record *record;
356	u8 *orig_buffer;
357	int orig_length;
358	size_t len;
359
360	orig_buffer = buffer;
361	orig_length = length;
362
363	while (length > 0) {
364		record = (struct ath10k_htc_record *)buffer;
365
366		if (length < sizeof(record->hdr)) {
367			status = -EINVAL;
368			break;
369		}
370
371		if (record->hdr.len > length) {
372			/* no room left in buffer for record */
373			ath10k_warn(ar, "Invalid record length: %d\n",
374				    record->hdr.len);
375			status = -EINVAL;
376			break;
377		}
378
379		switch (record->hdr.id) {
380		case ATH10K_HTC_RECORD_CREDITS:
381			len = sizeof(struct ath10k_htc_credit_report);
382			if (record->hdr.len < len) {
383				ath10k_warn(ar, "Credit report too long\n");
384				status = -EINVAL;
385				break;
386			}
387			ath10k_htc_process_credit_report(htc,
388							 record->credit_report,
389							 record->hdr.len,
390							 src_eid);
391			break;
392		case ATH10K_HTC_RECORD_LOOKAHEAD:
393			len = sizeof(struct ath10k_htc_lookahead_report);
394			if (record->hdr.len < len) {
395				ath10k_warn(ar, "Lookahead report too long\n");
396				status = -EINVAL;
397				break;
398			}
399			status = ath10k_htc_process_lookahead(htc,
400							      record->lookahead_report,
401							      record->hdr.len,
402							      src_eid,
403							      next_lookaheads,
404							      next_lookaheads_len);
405			break;
406		case ATH10K_HTC_RECORD_LOOKAHEAD_BUNDLE:
407			bundle = record->lookahead_bundle;
408			status = ath10k_htc_process_lookahead_bundle(htc,
409								     bundle,
410								     record->hdr.len,
411								     src_eid,
412								     next_lookaheads,
413								     next_lookaheads_len);
414			break;
415		default:
416			ath10k_warn(ar, "Unhandled record: id:%d length:%d\n",
417				    record->hdr.id, record->hdr.len);
418			break;
419		}
420
421		if (status)
422			break;
423
424		/* multiple records may be present in a trailer */
425		buffer += sizeof(record->hdr) + record->hdr.len;
426		length -= sizeof(record->hdr) + record->hdr.len;
427	}
428
429	if (status)
430		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc rx bad trailer", "",
431				orig_buffer, orig_length);
432
433	return status;
434}
435EXPORT_SYMBOL(ath10k_htc_process_trailer);
436
437void ath10k_htc_rx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
438{
439	int status = 0;
440	struct ath10k_htc *htc = &ar->htc;
441	struct ath10k_htc_hdr *hdr;
442	struct ath10k_htc_ep *ep;
443	u16 payload_len;
444	u32 trailer_len = 0;
445	size_t min_len;
446	u8 eid;
447	bool trailer_present;
448
449	hdr = (struct ath10k_htc_hdr *)skb->data;
450	skb_pull(skb, sizeof(*hdr));
451
452	eid = hdr->eid;
453
454	if (eid >= ATH10K_HTC_EP_COUNT) {
455		ath10k_warn(ar, "HTC Rx: invalid eid %d\n", eid);
456		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad header", "",
457				hdr, sizeof(*hdr));
458		goto out;
459	}
460
461	ep = &htc->endpoint[eid];
462	if (ep->service_id == ATH10K_HTC_SVC_ID_UNUSED) {
463		ath10k_warn(ar, "htc rx endpoint %d is not connected\n", eid);
464		goto out;
465	}
466
467	payload_len = __le16_to_cpu(hdr->len);
468
469	if (payload_len + sizeof(*hdr) > ATH10K_HTC_MAX_LEN) {
470		ath10k_warn(ar, "HTC rx frame too long, len: %zu\n",
471			    payload_len + sizeof(*hdr));
472		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len", "",
473				hdr, sizeof(*hdr));
474		goto out;
475	}
476
477	if (skb->len < payload_len) {
478		ath10k_dbg(ar, ATH10K_DBG_HTC,
479			   "HTC Rx: insufficient length, got %d, expected %d\n",
480			   skb->len, payload_len);
481		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len",
482				"", hdr, sizeof(*hdr));
483		goto out;
484	}
485
486	/* get flags to check for trailer */
487	trailer_present = hdr->flags & ATH10K_HTC_FLAG_TRAILER_PRESENT;
488	if (trailer_present) {
489		u8 *trailer;
490
491		trailer_len = hdr->trailer_len;
492		min_len = sizeof(struct ath10k_ath10k_htc_record_hdr);
493
494		if ((trailer_len < min_len) ||
495		    (trailer_len > payload_len)) {
496			ath10k_warn(ar, "Invalid trailer length: %d\n",
497				    trailer_len);
498			goto out;
499		}
500
501		trailer = (u8 *)hdr;
502		trailer += sizeof(*hdr);
503		trailer += payload_len;
504		trailer -= trailer_len;
505		status = ath10k_htc_process_trailer(htc, trailer,
506						    trailer_len, hdr->eid,
507						    NULL, NULL);
508		if (status)
509			goto out;
510
511		skb_trim(skb, skb->len - trailer_len);
512	}
513
514	if (((int)payload_len - (int)trailer_len) <= 0)
515		/* zero length packet with trailer data, just drop these */
516		goto out;
517
518	ath10k_dbg(ar, ATH10K_DBG_HTC, "htc rx completion ep %d skb %pK\n",
519		   eid, skb);
520	ep->ep_ops.ep_rx_complete(ar, skb);
521
522	/* skb is now owned by the rx completion handler */
523	skb = NULL;
524out:
525	kfree_skb(skb);
526}
527EXPORT_SYMBOL(ath10k_htc_rx_completion_handler);
528
529static void ath10k_htc_control_rx_complete(struct ath10k *ar,
530					   struct sk_buff *skb)
531{
532	struct ath10k_htc *htc = &ar->htc;
533	struct ath10k_htc_msg *msg = (struct ath10k_htc_msg *)skb->data;
534
535	switch (__le16_to_cpu(msg->hdr.message_id)) {
536	case ATH10K_HTC_MSG_READY_ID:
537	case ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID:
538		/* handle HTC control message */
539		if (completion_done(&htc->ctl_resp)) {
540			/* this is a fatal error, target should not be
541			 * sending unsolicited messages on the ep 0
542			 */
543			ath10k_warn(ar, "HTC rx ctrl still processing\n");
544			complete(&htc->ctl_resp);
545			goto out;
546		}
547
548		htc->control_resp_len =
549			min_t(int, skb->len,
550			      ATH10K_HTC_MAX_CTRL_MSG_LEN);
551
552		memcpy(htc->control_resp_buffer, skb->data,
553		       htc->control_resp_len);
554
555		complete(&htc->ctl_resp);
556		break;
557	case ATH10K_HTC_MSG_SEND_SUSPEND_COMPLETE:
558		htc->htc_ops.target_send_suspend_complete(ar);
559		break;
560	default:
561		ath10k_warn(ar, "ignoring unsolicited htc ep0 event\n");
562		break;
563	}
564
565out:
566	kfree_skb(skb);
567}
568
569/***************/
570/* Init/Deinit */
571/***************/
572
573static const char *htc_service_name(enum ath10k_htc_svc_id id)
574{
575	switch (id) {
576	case ATH10K_HTC_SVC_ID_RESERVED:
577		return "Reserved";
578	case ATH10K_HTC_SVC_ID_RSVD_CTRL:
579		return "Control";
580	case ATH10K_HTC_SVC_ID_WMI_CONTROL:
581		return "WMI";
582	case ATH10K_HTC_SVC_ID_WMI_DATA_BE:
583		return "DATA BE";
584	case ATH10K_HTC_SVC_ID_WMI_DATA_BK:
585		return "DATA BK";
586	case ATH10K_HTC_SVC_ID_WMI_DATA_VI:
587		return "DATA VI";
588	case ATH10K_HTC_SVC_ID_WMI_DATA_VO:
589		return "DATA VO";
590	case ATH10K_HTC_SVC_ID_NMI_CONTROL:
591		return "NMI Control";
592	case ATH10K_HTC_SVC_ID_NMI_DATA:
593		return "NMI Data";
594	case ATH10K_HTC_SVC_ID_HTT_DATA_MSG:
595		return "HTT Data";
596	case ATH10K_HTC_SVC_ID_HTT_DATA2_MSG:
597		return "HTT Data";
598	case ATH10K_HTC_SVC_ID_HTT_DATA3_MSG:
599		return "HTT Data";
600	case ATH10K_HTC_SVC_ID_TEST_RAW_STREAMS:
601		return "RAW";
602	case ATH10K_HTC_SVC_ID_HTT_LOG_MSG:
603		return "PKTLOG";
604	}
605
606	return "Unknown";
607}
608
609static void ath10k_htc_reset_endpoint_states(struct ath10k_htc *htc)
610{
611	struct ath10k_htc_ep *ep;
612	int i;
613
614	for (i = ATH10K_HTC_EP_0; i < ATH10K_HTC_EP_COUNT; i++) {
615		ep = &htc->endpoint[i];
616		ep->service_id = ATH10K_HTC_SVC_ID_UNUSED;
617		ep->max_ep_message_len = 0;
618		ep->max_tx_queue_depth = 0;
619		ep->eid = i;
620		ep->htc = htc;
621		ep->tx_credit_flow_enabled = true;
622	}
623}
624
625static u8 ath10k_htc_get_credit_allocation(struct ath10k_htc *htc,
626					   u16 service_id)
627{
628	u8 allocation = 0;
629
630	/* The WMI control service is the only service with flow control.
631	 * Let it have all transmit credits.
632	 */
633	if (service_id == ATH10K_HTC_SVC_ID_WMI_CONTROL)
634		allocation = htc->total_transmit_credits;
635
636	return allocation;
637}
638
639static int ath10k_htc_send_bundle(struct ath10k_htc_ep *ep,
640				  struct sk_buff *bundle_skb,
641				  struct sk_buff_head *tx_save_head)
642{
643	struct ath10k_hif_sg_item sg_item;
644	struct ath10k_htc *htc = ep->htc;
645	struct ath10k *ar = htc->ar;
646	struct sk_buff *skb;
647	int ret, cn = 0;
648	unsigned int skb_len;
649
650	ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle skb len %d\n", bundle_skb->len);
651	skb_len = bundle_skb->len;
652	ret = ath10k_htc_consume_credit(ep, skb_len, true);
653
654	if (!ret) {
655		sg_item.transfer_id = ep->eid;
656		sg_item.transfer_context = bundle_skb;
657		sg_item.vaddr = bundle_skb->data;
658		sg_item.len = bundle_skb->len;
659
660		ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
661		if (ret)
662			ath10k_htc_release_credit(ep, skb_len);
663	}
664
665	if (ret)
666		dev_kfree_skb_any(bundle_skb);
667
668	for (cn = 0; (skb = skb_dequeue_tail(tx_save_head)); cn++) {
669		if (ret) {
670			skb_pull(skb, sizeof(struct ath10k_htc_hdr));
671			skb_queue_head(&ep->tx_req_head, skb);
672		} else {
673			skb_queue_tail(&ep->tx_complete_head, skb);
674		}
675	}
676
677	if (!ret)
678		queue_work(ar->workqueue_tx_complete, &ar->tx_complete_work);
679
680	ath10k_dbg(ar, ATH10K_DBG_HTC,
681		   "bundle tx status %d eid %d req count %d count %d len %d\n",
682		   ret, ep->eid, skb_queue_len(&ep->tx_req_head), cn, skb_len);
683	return ret;
684}
685
686static void ath10k_htc_send_one_skb(struct ath10k_htc_ep *ep, struct sk_buff *skb)
687{
688	struct ath10k_htc *htc = ep->htc;
689	struct ath10k *ar = htc->ar;
690	int ret;
691
692	ret = ath10k_htc_send(htc, ep->eid, skb);
693
694	if (ret)
695		skb_queue_head(&ep->tx_req_head, skb);
696
697	ath10k_dbg(ar, ATH10K_DBG_HTC, "tx one status %d eid %d len %d pending count %d\n",
698		   ret, ep->eid, skb->len, skb_queue_len(&ep->tx_req_head));
699}
700
701static int ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep *ep)
702{
703	struct ath10k_htc *htc = ep->htc;
704	struct sk_buff *bundle_skb, *skb;
705	struct sk_buff_head tx_save_head;
706	struct ath10k_htc_hdr *hdr;
707	u8 *bundle_buf;
708	int ret = 0, credit_pad, credit_remainder, trans_len, bundles_left = 0;
709
710	if (htc->ar->state == ATH10K_STATE_WEDGED)
711		return -ECOMM;
712
713	if (ep->tx_credit_flow_enabled &&
714	    ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE)
715		return 0;
716
717	bundles_left = ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
718	bundle_skb = dev_alloc_skb(bundles_left);
719
720	if (!bundle_skb)
721		return -ENOMEM;
722
723	bundle_buf = bundle_skb->data;
724	skb_queue_head_init(&tx_save_head);
725
726	while (true) {
727		skb = skb_dequeue(&ep->tx_req_head);
728		if (!skb)
729			break;
730
731		credit_pad = 0;
732		trans_len = skb->len + sizeof(*hdr);
733		credit_remainder = trans_len % ep->tx_credit_size;
734
735		if (credit_remainder != 0) {
736			credit_pad = ep->tx_credit_size - credit_remainder;
737			trans_len += credit_pad;
738		}
739
740		ret = ath10k_htc_consume_credit(ep,
741						bundle_buf + trans_len - bundle_skb->data,
742						false);
743		if (ret) {
744			skb_queue_head(&ep->tx_req_head, skb);
745			break;
746		}
747
748		if (bundles_left < trans_len) {
749			bundle_skb->len = bundle_buf - bundle_skb->data;
750			ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
751
752			if (ret) {
753				skb_queue_head(&ep->tx_req_head, skb);
754				return ret;
755			}
756
757			if (skb_queue_len(&ep->tx_req_head) == 0) {
758				ath10k_htc_send_one_skb(ep, skb);
759				return ret;
760			}
761
762			if (ep->tx_credit_flow_enabled &&
763			    ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE) {
764				skb_queue_head(&ep->tx_req_head, skb);
765				return 0;
766			}
767
768			bundles_left =
769				ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
770			bundle_skb = dev_alloc_skb(bundles_left);
771
772			if (!bundle_skb) {
773				skb_queue_head(&ep->tx_req_head, skb);
774				return -ENOMEM;
775			}
776			bundle_buf = bundle_skb->data;
777			skb_queue_head_init(&tx_save_head);
778		}
779
780		skb_push(skb, sizeof(struct ath10k_htc_hdr));
781		ath10k_htc_prepare_tx_skb(ep, skb);
782
783		memcpy(bundle_buf, skb->data, skb->len);
784		hdr = (struct ath10k_htc_hdr *)bundle_buf;
785		hdr->flags |= ATH10K_HTC_FLAG_SEND_BUNDLE;
786		hdr->pad_len = __cpu_to_le16(credit_pad);
787		bundle_buf += trans_len;
788		bundles_left -= trans_len;
789		skb_queue_tail(&tx_save_head, skb);
790	}
791
792	if (bundle_buf != bundle_skb->data) {
793		bundle_skb->len = bundle_buf - bundle_skb->data;
794		ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
795	} else {
796		dev_kfree_skb_any(bundle_skb);
797	}
798
799	return ret;
800}
801
802static void ath10k_htc_bundle_tx_work(struct work_struct *work)
803{
804	struct ath10k *ar = container_of(work, struct ath10k, bundle_tx_work);
805	struct ath10k_htc_ep *ep;
806	struct sk_buff *skb;
807	int i;
808
809	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
810		ep = &ar->htc.endpoint[i];
811
812		if (!ep->bundle_tx)
813			continue;
814
815		ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx work eid %d count %d\n",
816			   ep->eid, skb_queue_len(&ep->tx_req_head));
817
818		if (skb_queue_len(&ep->tx_req_head) >=
819		    ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE) {
820			ath10k_htc_send_bundle_skbs(ep);
821		} else {
822			skb = skb_dequeue(&ep->tx_req_head);
823
824			if (!skb)
825				continue;
826			ath10k_htc_send_one_skb(ep, skb);
827		}
828	}
829}
830
831static void ath10k_htc_tx_complete_work(struct work_struct *work)
832{
833	struct ath10k *ar = container_of(work, struct ath10k, tx_complete_work);
834	struct ath10k_htc_ep *ep;
835	enum ath10k_htc_ep_id eid;
836	struct sk_buff *skb;
837	int i;
838
839	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
840		ep = &ar->htc.endpoint[i];
841		eid = ep->eid;
842		if (ep->bundle_tx && eid == ar->htt.eid) {
843			ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx complete eid %d pending complete count%d\n",
844				   ep->eid, skb_queue_len(&ep->tx_complete_head));
845
846			while (true) {
847				skb = skb_dequeue(&ep->tx_complete_head);
848				if (!skb)
849					break;
850				ath10k_htc_notify_tx_completion(ep, skb);
851			}
852		}
853	}
854}
855
856int ath10k_htc_send_hl(struct ath10k_htc *htc,
857		       enum ath10k_htc_ep_id eid,
858		       struct sk_buff *skb)
859{
860	struct ath10k_htc_ep *ep = &htc->endpoint[eid];
861	struct ath10k *ar = htc->ar;
862
863	if (sizeof(struct ath10k_htc_hdr) + skb->len > ep->tx_credit_size) {
864		ath10k_dbg(ar, ATH10K_DBG_HTC, "tx exceed max len %d\n", skb->len);
865		return -ENOMEM;
866	}
867
868	ath10k_dbg(ar, ATH10K_DBG_HTC, "htc send hl eid %d bundle %d tx count %d len %d\n",
869		   eid, ep->bundle_tx, skb_queue_len(&ep->tx_req_head), skb->len);
870
871	if (ep->bundle_tx) {
872		skb_queue_tail(&ep->tx_req_head, skb);
873		queue_work(ar->workqueue, &ar->bundle_tx_work);
874		return 0;
875	} else {
876		return ath10k_htc_send(htc, eid, skb);
877	}
878}
879
880void ath10k_htc_setup_tx_req(struct ath10k_htc_ep *ep)
881{
882	if (ep->htc->max_msgs_per_htc_bundle >= ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE &&
883	    !ep->bundle_tx) {
884		ep->bundle_tx = true;
885		skb_queue_head_init(&ep->tx_req_head);
886		skb_queue_head_init(&ep->tx_complete_head);
887	}
888}
889
890void ath10k_htc_stop_hl(struct ath10k *ar)
891{
892	struct ath10k_htc_ep *ep;
893	int i;
894
895	cancel_work_sync(&ar->bundle_tx_work);
896	cancel_work_sync(&ar->tx_complete_work);
897
898	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
899		ep = &ar->htc.endpoint[i];
900
901		if (!ep->bundle_tx)
902			continue;
903
904		ath10k_dbg(ar, ATH10K_DBG_HTC, "stop tx work eid %d count %d\n",
905			   ep->eid, skb_queue_len(&ep->tx_req_head));
906
907		skb_queue_purge(&ep->tx_req_head);
908	}
909}
910
911int ath10k_htc_wait_target(struct ath10k_htc *htc)
912{
913	struct ath10k *ar = htc->ar;
914	int i, status = 0;
915	unsigned long time_left;
916	struct ath10k_htc_msg *msg;
917	u16 message_id;
918
919	time_left = wait_for_completion_timeout(&htc->ctl_resp,
920						ATH10K_HTC_WAIT_TIMEOUT_HZ);
921	if (!time_left) {
922		/* Workaround: In some cases the PCI HIF doesn't
923		 * receive interrupt for the control response message
924		 * even if the buffer was completed. It is suspected
925		 * iomap writes unmasking PCI CE irqs aren't propagated
926		 * properly in KVM PCI-passthrough sometimes.
927		 */
928		ath10k_warn(ar, "failed to receive control response completion, polling..\n");
929
930		for (i = 0; i < CE_COUNT; i++)
931			ath10k_hif_send_complete_check(htc->ar, i, 1);
932
933		time_left =
934		wait_for_completion_timeout(&htc->ctl_resp,
935					    ATH10K_HTC_WAIT_TIMEOUT_HZ);
936
937		if (!time_left)
938			status = -ETIMEDOUT;
939	}
940
941	if (status < 0) {
942		ath10k_err(ar, "ctl_resp never came in (%d)\n", status);
943		return status;
944	}
945
946	if (htc->control_resp_len < sizeof(msg->hdr) + sizeof(msg->ready)) {
947		ath10k_err(ar, "Invalid HTC ready msg len:%d\n",
948			   htc->control_resp_len);
949		return -ECOMM;
950	}
951
952	msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
953	message_id   = __le16_to_cpu(msg->hdr.message_id);
954
955	if (message_id != ATH10K_HTC_MSG_READY_ID) {
956		ath10k_err(ar, "Invalid HTC ready msg: 0x%x\n", message_id);
957		return -ECOMM;
958	}
959
960	if (ar->hw_params.use_fw_tx_credits)
961		htc->total_transmit_credits = __le16_to_cpu(msg->ready.credit_count);
962	else
963		htc->total_transmit_credits = 1;
964
965	htc->target_credit_size = __le16_to_cpu(msg->ready.credit_size);
966
967	ath10k_dbg(ar, ATH10K_DBG_HTC,
968		   "Target ready! transmit resources: %d size:%d actual credits:%d\n",
969		   htc->total_transmit_credits,
970		   htc->target_credit_size,
971		   msg->ready.credit_count);
972
973	if ((htc->total_transmit_credits == 0) ||
974	    (htc->target_credit_size == 0)) {
975		ath10k_err(ar, "Invalid credit size received\n");
976		return -ECOMM;
977	}
978
979	/* The only way to determine if the ready message is an extended
980	 * message is from the size.
981	 */
982	if (htc->control_resp_len >=
983	    sizeof(msg->hdr) + sizeof(msg->ready_ext)) {
984		htc->alt_data_credit_size =
985			__le16_to_cpu(msg->ready_ext.reserved) &
986			ATH10K_HTC_MSG_READY_EXT_ALT_DATA_MASK;
987		htc->max_msgs_per_htc_bundle =
988			min_t(u8, msg->ready_ext.max_msgs_per_htc_bundle,
989			      HTC_HOST_MAX_MSG_PER_RX_BUNDLE);
990		ath10k_dbg(ar, ATH10K_DBG_HTC,
991			   "Extended ready message RX bundle size %d alt size %d\n",
992			   htc->max_msgs_per_htc_bundle,
993			   htc->alt_data_credit_size);
994	}
995
996	INIT_WORK(&ar->bundle_tx_work, ath10k_htc_bundle_tx_work);
997	INIT_WORK(&ar->tx_complete_work, ath10k_htc_tx_complete_work);
998
999	return 0;
1000}
1001
1002void ath10k_htc_change_tx_credit_flow(struct ath10k_htc *htc,
1003				      enum ath10k_htc_ep_id eid,
1004				      bool enable)
1005{
1006	struct ath10k *ar = htc->ar;
1007	struct ath10k_htc_ep *ep = &ar->htc.endpoint[eid];
1008
1009	ep->tx_credit_flow_enabled = enable;
1010}
1011
1012int ath10k_htc_connect_service(struct ath10k_htc *htc,
1013			       struct ath10k_htc_svc_conn_req *conn_req,
1014			       struct ath10k_htc_svc_conn_resp *conn_resp)
1015{
1016	struct ath10k *ar = htc->ar;
1017	struct ath10k_htc_msg *msg;
1018	struct ath10k_htc_conn_svc *req_msg;
1019	struct ath10k_htc_conn_svc_response resp_msg_dummy;
1020	struct ath10k_htc_conn_svc_response *resp_msg = &resp_msg_dummy;
1021	enum ath10k_htc_ep_id assigned_eid = ATH10K_HTC_EP_COUNT;
1022	struct ath10k_htc_ep *ep;
1023	struct sk_buff *skb;
1024	unsigned int max_msg_size = 0;
1025	int length, status;
1026	unsigned long time_left;
1027	bool disable_credit_flow_ctrl = false;
1028	u16 message_id, service_id, flags = 0;
1029	u8 tx_alloc = 0;
1030
1031	/* special case for HTC pseudo control service */
1032	if (conn_req->service_id == ATH10K_HTC_SVC_ID_RSVD_CTRL) {
1033		disable_credit_flow_ctrl = true;
1034		assigned_eid = ATH10K_HTC_EP_0;
1035		max_msg_size = ATH10K_HTC_MAX_CTRL_MSG_LEN;
1036		memset(&resp_msg_dummy, 0, sizeof(resp_msg_dummy));
1037		goto setup;
1038	}
1039
1040	tx_alloc = ath10k_htc_get_credit_allocation(htc,
1041						    conn_req->service_id);
1042	if (!tx_alloc)
1043		ath10k_dbg(ar, ATH10K_DBG_BOOT,
1044			   "boot htc service %s does not allocate target credits\n",
1045			   htc_service_name(conn_req->service_id));
1046
1047	skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1048	if (!skb) {
1049		ath10k_err(ar, "Failed to allocate HTC packet\n");
1050		return -ENOMEM;
1051	}
1052
1053	length = sizeof(msg->hdr) + sizeof(msg->connect_service);
1054	skb_put(skb, length);
1055	memset(skb->data, 0, length);
1056
1057	msg = (struct ath10k_htc_msg *)skb->data;
1058	msg->hdr.message_id =
1059		__cpu_to_le16(ATH10K_HTC_MSG_CONNECT_SERVICE_ID);
1060
1061	flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC);
1062
1063	/* Only enable credit flow control for WMI ctrl service */
1064	if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) {
1065		flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL;
1066		disable_credit_flow_ctrl = true;
1067	}
1068
1069	req_msg = &msg->connect_service;
1070	req_msg->flags = __cpu_to_le16(flags);
1071	req_msg->service_id = __cpu_to_le16(conn_req->service_id);
1072
1073	reinit_completion(&htc->ctl_resp);
1074
1075	status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1076	if (status) {
1077		kfree_skb(skb);
1078		return status;
1079	}
1080
1081	/* wait for response */
1082	time_left = wait_for_completion_timeout(&htc->ctl_resp,
1083						ATH10K_HTC_CONN_SVC_TIMEOUT_HZ);
1084	if (!time_left) {
1085		ath10k_err(ar, "Service connect timeout\n");
1086		return -ETIMEDOUT;
1087	}
1088
1089	/* we controlled the buffer creation, it's aligned */
1090	msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
1091	resp_msg = &msg->connect_service_response;
1092	message_id = __le16_to_cpu(msg->hdr.message_id);
1093	service_id = __le16_to_cpu(resp_msg->service_id);
1094
1095	if ((message_id != ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID) ||
1096	    (htc->control_resp_len < sizeof(msg->hdr) +
1097	     sizeof(msg->connect_service_response))) {
1098		ath10k_err(ar, "Invalid resp message ID 0x%x", message_id);
1099		return -EPROTO;
1100	}
1101
1102	ath10k_dbg(ar, ATH10K_DBG_HTC,
1103		   "HTC Service %s connect response: status: 0x%x, assigned ep: 0x%x\n",
1104		   htc_service_name(service_id),
1105		   resp_msg->status, resp_msg->eid);
1106
1107	conn_resp->connect_resp_code = resp_msg->status;
1108
1109	/* check response status */
1110	if (resp_msg->status != ATH10K_HTC_CONN_SVC_STATUS_SUCCESS) {
1111		ath10k_err(ar, "HTC Service %s connect request failed: 0x%x)\n",
1112			   htc_service_name(service_id),
1113			   resp_msg->status);
1114		return -EPROTO;
1115	}
1116
1117	assigned_eid = (enum ath10k_htc_ep_id)resp_msg->eid;
1118	max_msg_size = __le16_to_cpu(resp_msg->max_msg_size);
1119
1120setup:
1121
1122	if (assigned_eid >= ATH10K_HTC_EP_COUNT)
1123		return -EPROTO;
1124
1125	if (max_msg_size == 0)
1126		return -EPROTO;
1127
1128	ep = &htc->endpoint[assigned_eid];
1129	ep->eid = assigned_eid;
1130
1131	if (ep->service_id != ATH10K_HTC_SVC_ID_UNUSED)
1132		return -EPROTO;
1133
1134	/* return assigned endpoint to caller */
1135	conn_resp->eid = assigned_eid;
1136	conn_resp->max_msg_len = __le16_to_cpu(resp_msg->max_msg_size);
1137
1138	/* setup the endpoint */
1139	ep->service_id = conn_req->service_id;
1140	ep->max_tx_queue_depth = conn_req->max_send_queue_depth;
1141	ep->max_ep_message_len = __le16_to_cpu(resp_msg->max_msg_size);
1142	ep->tx_credits = tx_alloc;
1143	ep->tx_credit_size = htc->target_credit_size;
1144
1145	if (conn_req->service_id == ATH10K_HTC_SVC_ID_HTT_DATA_MSG &&
1146	    htc->alt_data_credit_size != 0)
1147		ep->tx_credit_size = htc->alt_data_credit_size;
1148
1149	/* copy all the callbacks */
1150	ep->ep_ops = conn_req->ep_ops;
1151
1152	status = ath10k_hif_map_service_to_pipe(htc->ar,
1153						ep->service_id,
1154						&ep->ul_pipe_id,
1155						&ep->dl_pipe_id);
1156	if (status) {
1157		ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC service id: %d\n",
1158			   ep->service_id);
1159		return status;
1160	}
1161
1162	ath10k_dbg(ar, ATH10K_DBG_BOOT,
1163		   "boot htc service '%s' ul pipe %d dl pipe %d eid %d ready\n",
1164		   htc_service_name(ep->service_id), ep->ul_pipe_id,
1165		   ep->dl_pipe_id, ep->eid);
1166
1167	if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) {
1168		ep->tx_credit_flow_enabled = false;
1169		ath10k_dbg(ar, ATH10K_DBG_BOOT,
1170			   "boot htc service '%s' eid %d TX flow control disabled\n",
1171			   htc_service_name(ep->service_id), assigned_eid);
1172	}
1173
1174	return status;
1175}
1176
1177struct sk_buff *ath10k_htc_alloc_skb(struct ath10k *ar, int size)
1178{
1179	struct sk_buff *skb;
1180
1181	skb = dev_alloc_skb(size + sizeof(struct ath10k_htc_hdr));
1182	if (!skb)
1183		return NULL;
1184
1185	skb_reserve(skb, sizeof(struct ath10k_htc_hdr));
1186
1187	/* FW/HTC requires 4-byte aligned streams */
1188	if (!IS_ALIGNED((unsigned long)skb->data, 4))
1189		ath10k_warn(ar, "Unaligned HTC tx skb\n");
1190
1191	return skb;
1192}
1193
1194static void ath10k_htc_pktlog_process_rx(struct ath10k *ar, struct sk_buff *skb)
1195{
1196	trace_ath10k_htt_pktlog(ar, skb->data, skb->len);
1197	dev_kfree_skb_any(skb);
1198}
1199
1200static int ath10k_htc_pktlog_connect(struct ath10k *ar)
1201{
1202	struct ath10k_htc_svc_conn_resp conn_resp;
1203	struct ath10k_htc_svc_conn_req conn_req;
1204	int status;
1205
1206	memset(&conn_req, 0, sizeof(conn_req));
1207	memset(&conn_resp, 0, sizeof(conn_resp));
1208
1209	conn_req.ep_ops.ep_tx_complete = NULL;
1210	conn_req.ep_ops.ep_rx_complete = ath10k_htc_pktlog_process_rx;
1211	conn_req.ep_ops.ep_tx_credits = NULL;
1212
1213	/* connect to control service */
1214	conn_req.service_id = ATH10K_HTC_SVC_ID_HTT_LOG_MSG;
1215	status = ath10k_htc_connect_service(&ar->htc, &conn_req, &conn_resp);
1216	if (status) {
1217		ath10k_warn(ar, "failed to connect to PKTLOG service: %d\n",
1218			    status);
1219		return status;
1220	}
1221
1222	return 0;
1223}
1224
1225static bool ath10k_htc_pktlog_svc_supported(struct ath10k *ar)
1226{
1227	u8 ul_pipe_id;
1228	u8 dl_pipe_id;
1229	int status;
1230
1231	status = ath10k_hif_map_service_to_pipe(ar, ATH10K_HTC_SVC_ID_HTT_LOG_MSG,
1232						&ul_pipe_id,
1233						&dl_pipe_id);
1234	if (status) {
1235		ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC pktlog service id: %d\n",
1236			   ATH10K_HTC_SVC_ID_HTT_LOG_MSG);
1237
1238		return false;
1239	}
1240
1241	return true;
1242}
1243
1244int ath10k_htc_start(struct ath10k_htc *htc)
1245{
1246	struct ath10k *ar = htc->ar;
1247	struct sk_buff *skb;
1248	int status = 0;
1249	struct ath10k_htc_msg *msg;
1250
1251	skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1252	if (!skb)
1253		return -ENOMEM;
1254
1255	skb_put(skb, sizeof(msg->hdr) + sizeof(msg->setup_complete_ext));
1256	memset(skb->data, 0, skb->len);
1257
1258	msg = (struct ath10k_htc_msg *)skb->data;
1259	msg->hdr.message_id =
1260		__cpu_to_le16(ATH10K_HTC_MSG_SETUP_COMPLETE_EX_ID);
1261
1262	if (ar->hif.bus == ATH10K_BUS_SDIO) {
1263		/* Extra setup params used by SDIO */
1264		msg->setup_complete_ext.flags =
1265			__cpu_to_le32(ATH10K_HTC_SETUP_COMPLETE_FLAGS_RX_BNDL_EN);
1266		msg->setup_complete_ext.max_msgs_per_bundled_recv =
1267			htc->max_msgs_per_htc_bundle;
1268	}
1269	ath10k_dbg(ar, ATH10K_DBG_HTC, "HTC is using TX credit flow control\n");
1270
1271	status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1272	if (status) {
1273		kfree_skb(skb);
1274		return status;
1275	}
1276
1277	if (ath10k_htc_pktlog_svc_supported(ar)) {
1278		status = ath10k_htc_pktlog_connect(ar);
1279		if (status) {
1280			ath10k_err(ar, "failed to connect to pktlog: %d\n", status);
1281			return status;
1282		}
1283	}
1284
1285	return 0;
1286}
1287
1288/* registered target arrival callback from the HIF layer */
1289int ath10k_htc_init(struct ath10k *ar)
1290{
1291	int status;
1292	struct ath10k_htc *htc = &ar->htc;
1293	struct ath10k_htc_svc_conn_req conn_req;
1294	struct ath10k_htc_svc_conn_resp conn_resp;
1295
1296	spin_lock_init(&htc->tx_lock);
1297
1298	ath10k_htc_reset_endpoint_states(htc);
1299
1300	htc->ar = ar;
1301
1302	/* setup our pseudo HTC control endpoint connection */
1303	memset(&conn_req, 0, sizeof(conn_req));
1304	memset(&conn_resp, 0, sizeof(conn_resp));
1305	conn_req.ep_ops.ep_tx_complete = ath10k_htc_control_tx_complete;
1306	conn_req.ep_ops.ep_rx_complete = ath10k_htc_control_rx_complete;
1307	conn_req.max_send_queue_depth = ATH10K_NUM_CONTROL_TX_BUFFERS;
1308	conn_req.service_id = ATH10K_HTC_SVC_ID_RSVD_CTRL;
1309
1310	/* connect fake service */
1311	status = ath10k_htc_connect_service(htc, &conn_req, &conn_resp);
1312	if (status) {
1313		ath10k_err(ar, "could not connect to htc service (%d)\n",
1314			   status);
1315		return status;
1316	}
1317
1318	init_completion(&htc->ctl_resp);
1319
1320	return 0;
1321}
1322