1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * DMA driver for STMicroelectronics STi FDMA controller
4 *
5 * Copyright (C) 2014 STMicroelectronics
6 *
7 * Author: Ludovic Barre <Ludovic.barre@st.com>
8 *	   Peter Griffin <peter.griffin@linaro.org>
9 */
10
11#include <linux/init.h>
12#include <linux/module.h>
13#include <linux/of.h>
14#include <linux/of_dma.h>
15#include <linux/platform_device.h>
16#include <linux/property.h>
17#include <linux/interrupt.h>
18#include <linux/remoteproc.h>
19#include <linux/slab.h>
20
21#include "st_fdma.h"
22
23static inline struct st_fdma_chan *to_st_fdma_chan(struct dma_chan *c)
24{
25	return container_of(c, struct st_fdma_chan, vchan.chan);
26}
27
28static struct st_fdma_desc *to_st_fdma_desc(struct virt_dma_desc *vd)
29{
30	return container_of(vd, struct st_fdma_desc, vdesc);
31}
32
33static int st_fdma_dreq_get(struct st_fdma_chan *fchan)
34{
35	struct st_fdma_dev *fdev = fchan->fdev;
36	u32 req_line_cfg = fchan->cfg.req_line;
37	u32 dreq_line;
38	int try = 0;
39
40	/*
41	 * dreq_mask is shared for n channels of fdma, so all accesses must be
42	 * atomic. if the dreq_mask is changed between ffz and set_bit,
43	 * we retry
44	 */
45	do {
46		if (fdev->dreq_mask == ~0L) {
47			dev_err(fdev->dev, "No req lines available\n");
48			return -EINVAL;
49		}
50
51		if (try || req_line_cfg >= ST_FDMA_NR_DREQS) {
52			dev_err(fdev->dev, "Invalid or used req line\n");
53			return -EINVAL;
54		} else {
55			dreq_line = req_line_cfg;
56		}
57
58		try++;
59	} while (test_and_set_bit(dreq_line, &fdev->dreq_mask));
60
61	dev_dbg(fdev->dev, "get dreq_line:%d mask:%#lx\n",
62		dreq_line, fdev->dreq_mask);
63
64	return dreq_line;
65}
66
67static void st_fdma_dreq_put(struct st_fdma_chan *fchan)
68{
69	struct st_fdma_dev *fdev = fchan->fdev;
70
71	dev_dbg(fdev->dev, "put dreq_line:%#x\n", fchan->dreq_line);
72	clear_bit(fchan->dreq_line, &fdev->dreq_mask);
73}
74
75static void st_fdma_xfer_desc(struct st_fdma_chan *fchan)
76{
77	struct virt_dma_desc *vdesc;
78	unsigned long nbytes, ch_cmd, cmd;
79
80	vdesc = vchan_next_desc(&fchan->vchan);
81	if (!vdesc)
82		return;
83
84	fchan->fdesc = to_st_fdma_desc(vdesc);
85	nbytes = fchan->fdesc->node[0].desc->nbytes;
86	cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id);
87	ch_cmd = fchan->fdesc->node[0].pdesc | FDMA_CH_CMD_STA_START;
88
89	/* start the channel for the descriptor */
90	fnode_write(fchan, nbytes, FDMA_CNTN_OFST);
91	fchan_write(fchan, ch_cmd, FDMA_CH_CMD_OFST);
92	writel(cmd,
93		fchan->fdev->slim_rproc->peri + FDMA_CMD_SET_OFST);
94
95	dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id);
96}
97
98static void st_fdma_ch_sta_update(struct st_fdma_chan *fchan,
99				  unsigned long int_sta)
100{
101	unsigned long ch_sta, ch_err;
102	int ch_id = fchan->vchan.chan.chan_id;
103	struct st_fdma_dev *fdev = fchan->fdev;
104
105	ch_sta = fchan_read(fchan, FDMA_CH_CMD_OFST);
106	ch_err = ch_sta & FDMA_CH_CMD_ERR_MASK;
107	ch_sta &= FDMA_CH_CMD_STA_MASK;
108
109	if (int_sta & FDMA_INT_STA_ERR) {
110		dev_warn(fdev->dev, "chan:%d, error:%ld\n", ch_id, ch_err);
111		fchan->status = DMA_ERROR;
112		return;
113	}
114
115	switch (ch_sta) {
116	case FDMA_CH_CMD_STA_PAUSED:
117		fchan->status = DMA_PAUSED;
118		break;
119
120	case FDMA_CH_CMD_STA_RUNNING:
121		fchan->status = DMA_IN_PROGRESS;
122		break;
123	}
124}
125
126static irqreturn_t st_fdma_irq_handler(int irq, void *dev_id)
127{
128	struct st_fdma_dev *fdev = dev_id;
129	irqreturn_t ret = IRQ_NONE;
130	struct st_fdma_chan *fchan = &fdev->chans[0];
131	unsigned long int_sta, clr;
132
133	int_sta = fdma_read(fdev, FDMA_INT_STA_OFST);
134	clr = int_sta;
135
136	for (; int_sta != 0 ; int_sta >>= 2, fchan++) {
137		if (!(int_sta & (FDMA_INT_STA_CH | FDMA_INT_STA_ERR)))
138			continue;
139
140		spin_lock(&fchan->vchan.lock);
141		st_fdma_ch_sta_update(fchan, int_sta);
142
143		if (fchan->fdesc) {
144			if (!fchan->fdesc->iscyclic) {
145				list_del(&fchan->fdesc->vdesc.node);
146				vchan_cookie_complete(&fchan->fdesc->vdesc);
147				fchan->fdesc = NULL;
148				fchan->status = DMA_COMPLETE;
149			} else {
150				vchan_cyclic_callback(&fchan->fdesc->vdesc);
151			}
152
153			/* Start the next descriptor (if available) */
154			if (!fchan->fdesc)
155				st_fdma_xfer_desc(fchan);
156		}
157
158		spin_unlock(&fchan->vchan.lock);
159		ret = IRQ_HANDLED;
160	}
161
162	fdma_write(fdev, clr, FDMA_INT_CLR_OFST);
163
164	return ret;
165}
166
167static struct dma_chan *st_fdma_of_xlate(struct of_phandle_args *dma_spec,
168					 struct of_dma *ofdma)
169{
170	struct st_fdma_dev *fdev = ofdma->of_dma_data;
171	struct dma_chan *chan;
172	struct st_fdma_chan *fchan;
173	int ret;
174
175	if (dma_spec->args_count < 1)
176		return ERR_PTR(-EINVAL);
177
178	if (fdev->dma_device.dev->of_node != dma_spec->np)
179		return ERR_PTR(-EINVAL);
180
181	ret = rproc_boot(fdev->slim_rproc->rproc);
182	if (ret == -ENOENT)
183		return ERR_PTR(-EPROBE_DEFER);
184	else if (ret)
185		return ERR_PTR(ret);
186
187	chan = dma_get_any_slave_channel(&fdev->dma_device);
188	if (!chan)
189		goto err_chan;
190
191	fchan = to_st_fdma_chan(chan);
192
193	fchan->cfg.of_node = dma_spec->np;
194	fchan->cfg.req_line = dma_spec->args[0];
195	fchan->cfg.req_ctrl = 0;
196	fchan->cfg.type = ST_FDMA_TYPE_FREE_RUN;
197
198	if (dma_spec->args_count > 1)
199		fchan->cfg.req_ctrl = dma_spec->args[1]
200			& FDMA_REQ_CTRL_CFG_MASK;
201
202	if (dma_spec->args_count > 2)
203		fchan->cfg.type = dma_spec->args[2];
204
205	if (fchan->cfg.type == ST_FDMA_TYPE_FREE_RUN) {
206		fchan->dreq_line = 0;
207	} else {
208		fchan->dreq_line = st_fdma_dreq_get(fchan);
209		if (IS_ERR_VALUE(fchan->dreq_line)) {
210			chan = ERR_PTR(fchan->dreq_line);
211			goto err_chan;
212		}
213	}
214
215	dev_dbg(fdev->dev, "xlate req_line:%d type:%d req_ctrl:%#lx\n",
216		fchan->cfg.req_line, fchan->cfg.type, fchan->cfg.req_ctrl);
217
218	return chan;
219
220err_chan:
221	rproc_shutdown(fdev->slim_rproc->rproc);
222	return chan;
223
224}
225
226static void st_fdma_free_desc(struct virt_dma_desc *vdesc)
227{
228	struct st_fdma_desc *fdesc;
229	int i;
230
231	fdesc = to_st_fdma_desc(vdesc);
232	for (i = 0; i < fdesc->n_nodes; i++)
233		dma_pool_free(fdesc->fchan->node_pool, fdesc->node[i].desc,
234			      fdesc->node[i].pdesc);
235	kfree(fdesc);
236}
237
238static struct st_fdma_desc *st_fdma_alloc_desc(struct st_fdma_chan *fchan,
239					       int sg_len)
240{
241	struct st_fdma_desc *fdesc;
242	int i;
243
244	fdesc = kzalloc(struct_size(fdesc, node, sg_len), GFP_NOWAIT);
245	if (!fdesc)
246		return NULL;
247
248	fdesc->fchan = fchan;
249	fdesc->n_nodes = sg_len;
250	for (i = 0; i < sg_len; i++) {
251		fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool,
252				GFP_NOWAIT, &fdesc->node[i].pdesc);
253		if (!fdesc->node[i].desc)
254			goto err;
255	}
256	return fdesc;
257
258err:
259	while (--i >= 0)
260		dma_pool_free(fchan->node_pool, fdesc->node[i].desc,
261			      fdesc->node[i].pdesc);
262	kfree(fdesc);
263	return NULL;
264}
265
266static int st_fdma_alloc_chan_res(struct dma_chan *chan)
267{
268	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
269
270	/* Create the dma pool for descriptor allocation */
271	fchan->node_pool = dma_pool_create(dev_name(&chan->dev->device),
272					    fchan->fdev->dev,
273					    sizeof(struct st_fdma_hw_node),
274					    __alignof__(struct st_fdma_hw_node),
275					    0);
276
277	if (!fchan->node_pool) {
278		dev_err(fchan->fdev->dev, "unable to allocate desc pool\n");
279		return -ENOMEM;
280	}
281
282	dev_dbg(fchan->fdev->dev, "alloc ch_id:%d type:%d\n",
283		fchan->vchan.chan.chan_id, fchan->cfg.type);
284
285	return 0;
286}
287
288static void st_fdma_free_chan_res(struct dma_chan *chan)
289{
290	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
291	struct rproc *rproc = fchan->fdev->slim_rproc->rproc;
292	unsigned long flags;
293
294	dev_dbg(fchan->fdev->dev, "%s: freeing chan:%d\n",
295		__func__, fchan->vchan.chan.chan_id);
296
297	if (fchan->cfg.type != ST_FDMA_TYPE_FREE_RUN)
298		st_fdma_dreq_put(fchan);
299
300	spin_lock_irqsave(&fchan->vchan.lock, flags);
301	fchan->fdesc = NULL;
302	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
303
304	dma_pool_destroy(fchan->node_pool);
305	fchan->node_pool = NULL;
306	memset(&fchan->cfg, 0, sizeof(struct st_fdma_cfg));
307
308	rproc_shutdown(rproc);
309}
310
311static struct dma_async_tx_descriptor *st_fdma_prep_dma_memcpy(
312	struct dma_chan *chan,	dma_addr_t dst, dma_addr_t src,
313	size_t len, unsigned long flags)
314{
315	struct st_fdma_chan *fchan;
316	struct st_fdma_desc *fdesc;
317	struct st_fdma_hw_node *hw_node;
318
319	if (!len)
320		return NULL;
321
322	fchan = to_st_fdma_chan(chan);
323
324	/* We only require a single descriptor */
325	fdesc = st_fdma_alloc_desc(fchan, 1);
326	if (!fdesc) {
327		dev_err(fchan->fdev->dev, "no memory for desc\n");
328		return NULL;
329	}
330
331	hw_node = fdesc->node[0].desc;
332	hw_node->next = 0;
333	hw_node->control = FDMA_NODE_CTRL_REQ_MAP_FREE_RUN;
334	hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
335	hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
336	hw_node->control |= FDMA_NODE_CTRL_INT_EON;
337	hw_node->nbytes = len;
338	hw_node->saddr = src;
339	hw_node->daddr = dst;
340	hw_node->generic.length = len;
341	hw_node->generic.sstride = 0;
342	hw_node->generic.dstride = 0;
343
344	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
345}
346
347static int config_reqctrl(struct st_fdma_chan *fchan,
348			  enum dma_transfer_direction direction)
349{
350	u32 maxburst = 0, addr = 0;
351	enum dma_slave_buswidth width;
352	int ch_id = fchan->vchan.chan.chan_id;
353	struct st_fdma_dev *fdev = fchan->fdev;
354
355	switch (direction) {
356
357	case DMA_DEV_TO_MEM:
358		fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_WNR;
359		maxburst = fchan->scfg.src_maxburst;
360		width = fchan->scfg.src_addr_width;
361		addr = fchan->scfg.src_addr;
362		break;
363
364	case DMA_MEM_TO_DEV:
365		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_WNR;
366		maxburst = fchan->scfg.dst_maxburst;
367		width = fchan->scfg.dst_addr_width;
368		addr = fchan->scfg.dst_addr;
369		break;
370
371	default:
372		return -EINVAL;
373	}
374
375	fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_OPCODE_MASK;
376
377	switch (width) {
378
379	case DMA_SLAVE_BUSWIDTH_1_BYTE:
380		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST1;
381		break;
382
383	case DMA_SLAVE_BUSWIDTH_2_BYTES:
384		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST2;
385		break;
386
387	case DMA_SLAVE_BUSWIDTH_4_BYTES:
388		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST4;
389		break;
390
391	case DMA_SLAVE_BUSWIDTH_8_BYTES:
392		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST8;
393		break;
394
395	default:
396		return -EINVAL;
397	}
398
399	fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_NUM_OPS_MASK;
400	fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_NUM_OPS(maxburst-1);
401	dreq_write(fchan, fchan->cfg.req_ctrl, FDMA_REQ_CTRL_OFST);
402
403	fchan->cfg.dev_addr = addr;
404	fchan->cfg.dir = direction;
405
406	dev_dbg(fdev->dev, "chan:%d config_reqctrl:%#x req_ctrl:%#lx\n",
407		ch_id, addr, fchan->cfg.req_ctrl);
408
409	return 0;
410}
411
412static void fill_hw_node(struct st_fdma_hw_node *hw_node,
413			struct st_fdma_chan *fchan,
414			enum dma_transfer_direction direction)
415{
416	if (direction == DMA_MEM_TO_DEV) {
417		hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
418		hw_node->control |= FDMA_NODE_CTRL_DST_STATIC;
419		hw_node->daddr = fchan->cfg.dev_addr;
420	} else {
421		hw_node->control |= FDMA_NODE_CTRL_SRC_STATIC;
422		hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
423		hw_node->saddr = fchan->cfg.dev_addr;
424	}
425
426	hw_node->generic.sstride = 0;
427	hw_node->generic.dstride = 0;
428}
429
430static inline struct st_fdma_chan *st_fdma_prep_common(struct dma_chan *chan,
431		size_t len, enum dma_transfer_direction direction)
432{
433	struct st_fdma_chan *fchan;
434
435	if (!chan || !len)
436		return NULL;
437
438	fchan = to_st_fdma_chan(chan);
439
440	if (!is_slave_direction(direction)) {
441		dev_err(fchan->fdev->dev, "bad direction?\n");
442		return NULL;
443	}
444
445	return fchan;
446}
447
448static struct dma_async_tx_descriptor *st_fdma_prep_dma_cyclic(
449		struct dma_chan *chan, dma_addr_t buf_addr, size_t len,
450		size_t period_len, enum dma_transfer_direction direction,
451		unsigned long flags)
452{
453	struct st_fdma_chan *fchan;
454	struct st_fdma_desc *fdesc;
455	int sg_len, i;
456
457	fchan = st_fdma_prep_common(chan, len, direction);
458	if (!fchan)
459		return NULL;
460
461	if (!period_len)
462		return NULL;
463
464	if (config_reqctrl(fchan, direction)) {
465		dev_err(fchan->fdev->dev, "bad width or direction\n");
466		return NULL;
467	}
468
469	/* the buffer length must be a multiple of period_len */
470	if (len % period_len != 0) {
471		dev_err(fchan->fdev->dev, "len is not multiple of period\n");
472		return NULL;
473	}
474
475	sg_len = len / period_len;
476	fdesc = st_fdma_alloc_desc(fchan, sg_len);
477	if (!fdesc) {
478		dev_err(fchan->fdev->dev, "no memory for desc\n");
479		return NULL;
480	}
481
482	fdesc->iscyclic = true;
483
484	for (i = 0; i < sg_len; i++) {
485		struct st_fdma_hw_node *hw_node = fdesc->node[i].desc;
486
487		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
488
489		hw_node->control =
490			FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);
491		hw_node->control |= FDMA_NODE_CTRL_INT_EON;
492
493		fill_hw_node(hw_node, fchan, direction);
494
495		if (direction == DMA_MEM_TO_DEV)
496			hw_node->saddr = buf_addr + (i * period_len);
497		else
498			hw_node->daddr = buf_addr + (i * period_len);
499
500		hw_node->nbytes = period_len;
501		hw_node->generic.length = period_len;
502	}
503
504	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
505}
506
507static struct dma_async_tx_descriptor *st_fdma_prep_slave_sg(
508		struct dma_chan *chan, struct scatterlist *sgl,
509		unsigned int sg_len, enum dma_transfer_direction direction,
510		unsigned long flags, void *context)
511{
512	struct st_fdma_chan *fchan;
513	struct st_fdma_desc *fdesc;
514	struct st_fdma_hw_node *hw_node;
515	struct scatterlist *sg;
516	int i;
517
518	fchan = st_fdma_prep_common(chan, sg_len, direction);
519	if (!fchan)
520		return NULL;
521
522	if (!sgl)
523		return NULL;
524
525	fdesc = st_fdma_alloc_desc(fchan, sg_len);
526	if (!fdesc) {
527		dev_err(fchan->fdev->dev, "no memory for desc\n");
528		return NULL;
529	}
530
531	fdesc->iscyclic = false;
532
533	for_each_sg(sgl, sg, sg_len, i) {
534		hw_node = fdesc->node[i].desc;
535
536		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
537		hw_node->control = FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);
538
539		fill_hw_node(hw_node, fchan, direction);
540
541		if (direction == DMA_MEM_TO_DEV)
542			hw_node->saddr = sg_dma_address(sg);
543		else
544			hw_node->daddr = sg_dma_address(sg);
545
546		hw_node->nbytes = sg_dma_len(sg);
547		hw_node->generic.length = sg_dma_len(sg);
548	}
549
550	/* interrupt at end of last node */
551	hw_node->control |= FDMA_NODE_CTRL_INT_EON;
552
553	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
554}
555
556static size_t st_fdma_desc_residue(struct st_fdma_chan *fchan,
557				   struct virt_dma_desc *vdesc,
558				   bool in_progress)
559{
560	struct st_fdma_desc *fdesc = fchan->fdesc;
561	size_t residue = 0;
562	dma_addr_t cur_addr = 0;
563	int i;
564
565	if (in_progress) {
566		cur_addr = fchan_read(fchan, FDMA_CH_CMD_OFST);
567		cur_addr &= FDMA_CH_CMD_DATA_MASK;
568	}
569
570	for (i = fchan->fdesc->n_nodes - 1 ; i >= 0; i--) {
571		if (cur_addr == fdesc->node[i].pdesc) {
572			residue += fnode_read(fchan, FDMA_CNTN_OFST);
573			break;
574		}
575		residue += fdesc->node[i].desc->nbytes;
576	}
577
578	return residue;
579}
580
581static enum dma_status st_fdma_tx_status(struct dma_chan *chan,
582					 dma_cookie_t cookie,
583					 struct dma_tx_state *txstate)
584{
585	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
586	struct virt_dma_desc *vd;
587	enum dma_status ret;
588	unsigned long flags;
589
590	ret = dma_cookie_status(chan, cookie, txstate);
591	if (ret == DMA_COMPLETE || !txstate)
592		return ret;
593
594	spin_lock_irqsave(&fchan->vchan.lock, flags);
595	vd = vchan_find_desc(&fchan->vchan, cookie);
596	if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie)
597		txstate->residue = st_fdma_desc_residue(fchan, vd, true);
598	else if (vd)
599		txstate->residue = st_fdma_desc_residue(fchan, vd, false);
600	else
601		txstate->residue = 0;
602
603	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
604
605	return ret;
606}
607
608static void st_fdma_issue_pending(struct dma_chan *chan)
609{
610	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
611	unsigned long flags;
612
613	spin_lock_irqsave(&fchan->vchan.lock, flags);
614
615	if (vchan_issue_pending(&fchan->vchan) && !fchan->fdesc)
616		st_fdma_xfer_desc(fchan);
617
618	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
619}
620
621static int st_fdma_pause(struct dma_chan *chan)
622{
623	unsigned long flags;
624	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
625	int ch_id = fchan->vchan.chan.chan_id;
626	unsigned long cmd = FDMA_CMD_PAUSE(ch_id);
627
628	dev_dbg(fchan->fdev->dev, "pause chan:%d\n", ch_id);
629
630	spin_lock_irqsave(&fchan->vchan.lock, flags);
631	if (fchan->fdesc)
632		fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
633	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
634
635	return 0;
636}
637
638static int st_fdma_resume(struct dma_chan *chan)
639{
640	unsigned long flags;
641	unsigned long val;
642	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
643	int ch_id = fchan->vchan.chan.chan_id;
644
645	dev_dbg(fchan->fdev->dev, "resume chan:%d\n", ch_id);
646
647	spin_lock_irqsave(&fchan->vchan.lock, flags);
648	if (fchan->fdesc) {
649		val = fchan_read(fchan, FDMA_CH_CMD_OFST);
650		val &= FDMA_CH_CMD_DATA_MASK;
651		fchan_write(fchan, val, FDMA_CH_CMD_OFST);
652	}
653	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
654
655	return 0;
656}
657
658static int st_fdma_terminate_all(struct dma_chan *chan)
659{
660	unsigned long flags;
661	LIST_HEAD(head);
662	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
663	int ch_id = fchan->vchan.chan.chan_id;
664	unsigned long cmd = FDMA_CMD_PAUSE(ch_id);
665
666	dev_dbg(fchan->fdev->dev, "terminate chan:%d\n", ch_id);
667
668	spin_lock_irqsave(&fchan->vchan.lock, flags);
669	fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
670	fchan->fdesc = NULL;
671	vchan_get_all_descriptors(&fchan->vchan, &head);
672	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
673	vchan_dma_desc_free_list(&fchan->vchan, &head);
674
675	return 0;
676}
677
678static int st_fdma_slave_config(struct dma_chan *chan,
679				struct dma_slave_config *slave_cfg)
680{
681	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
682
683	memcpy(&fchan->scfg, slave_cfg, sizeof(fchan->scfg));
684	return 0;
685}
686
687static const struct st_fdma_driverdata fdma_mpe31_stih407_11 = {
688	.name = "STiH407",
689	.id = 0,
690};
691
692static const struct st_fdma_driverdata fdma_mpe31_stih407_12 = {
693	.name = "STiH407",
694	.id = 1,
695};
696
697static const struct st_fdma_driverdata fdma_mpe31_stih407_13 = {
698	.name = "STiH407",
699	.id = 2,
700};
701
702static const struct of_device_id st_fdma_match[] = {
703	{ .compatible = "st,stih407-fdma-mpe31-11"
704	  , .data = &fdma_mpe31_stih407_11 },
705	{ .compatible = "st,stih407-fdma-mpe31-12"
706	  , .data = &fdma_mpe31_stih407_12 },
707	{ .compatible = "st,stih407-fdma-mpe31-13"
708	  , .data = &fdma_mpe31_stih407_13 },
709	{},
710};
711MODULE_DEVICE_TABLE(of, st_fdma_match);
712
713static int st_fdma_parse_dt(struct platform_device *pdev,
714			const struct st_fdma_driverdata *drvdata,
715			struct st_fdma_dev *fdev)
716{
717	snprintf(fdev->fw_name, FW_NAME_SIZE, "fdma_%s_%d.elf",
718		drvdata->name, drvdata->id);
719
720	return of_property_read_u32(pdev->dev.of_node, "dma-channels",
721				    &fdev->nr_channels);
722}
723#define FDMA_DMA_BUSWIDTHS	(BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
724				 BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
725				 BIT(DMA_SLAVE_BUSWIDTH_3_BYTES) | \
726				 BIT(DMA_SLAVE_BUSWIDTH_4_BYTES))
727
728static void st_fdma_free(struct st_fdma_dev *fdev)
729{
730	struct st_fdma_chan *fchan;
731	int i;
732
733	for (i = 0; i < fdev->nr_channels; i++) {
734		fchan = &fdev->chans[i];
735		list_del(&fchan->vchan.chan.device_node);
736		tasklet_kill(&fchan->vchan.task);
737	}
738}
739
740static int st_fdma_probe(struct platform_device *pdev)
741{
742	struct st_fdma_dev *fdev;
743	struct device_node *np = pdev->dev.of_node;
744	const struct st_fdma_driverdata *drvdata;
745	int ret, i;
746
747	drvdata = device_get_match_data(&pdev->dev);
748
749	fdev = devm_kzalloc(&pdev->dev, sizeof(*fdev), GFP_KERNEL);
750	if (!fdev)
751		return -ENOMEM;
752
753	ret = st_fdma_parse_dt(pdev, drvdata, fdev);
754	if (ret) {
755		dev_err(&pdev->dev, "unable to find platform data\n");
756		goto err;
757	}
758
759	fdev->chans = devm_kcalloc(&pdev->dev, fdev->nr_channels,
760				   sizeof(struct st_fdma_chan), GFP_KERNEL);
761	if (!fdev->chans)
762		return -ENOMEM;
763
764	fdev->dev = &pdev->dev;
765	fdev->drvdata = drvdata;
766	platform_set_drvdata(pdev, fdev);
767
768	fdev->irq = platform_get_irq(pdev, 0);
769	if (fdev->irq < 0)
770		return -EINVAL;
771
772	ret = devm_request_irq(&pdev->dev, fdev->irq, st_fdma_irq_handler, 0,
773			       dev_name(&pdev->dev), fdev);
774	if (ret) {
775		dev_err(&pdev->dev, "Failed to request irq (%d)\n", ret);
776		goto err;
777	}
778
779	fdev->slim_rproc = st_slim_rproc_alloc(pdev, fdev->fw_name);
780	if (IS_ERR(fdev->slim_rproc)) {
781		ret = PTR_ERR(fdev->slim_rproc);
782		dev_err(&pdev->dev, "slim_rproc_alloc failed (%d)\n", ret);
783		goto err;
784	}
785
786	/* Initialise list of FDMA channels */
787	INIT_LIST_HEAD(&fdev->dma_device.channels);
788	for (i = 0; i < fdev->nr_channels; i++) {
789		struct st_fdma_chan *fchan = &fdev->chans[i];
790
791		fchan->fdev = fdev;
792		fchan->vchan.desc_free = st_fdma_free_desc;
793		vchan_init(&fchan->vchan, &fdev->dma_device);
794	}
795
796	/* Initialise the FDMA dreq (reserve 0 & 31 for FDMA use) */
797	fdev->dreq_mask = BIT(0) | BIT(31);
798
799	dma_cap_set(DMA_SLAVE, fdev->dma_device.cap_mask);
800	dma_cap_set(DMA_CYCLIC, fdev->dma_device.cap_mask);
801	dma_cap_set(DMA_MEMCPY, fdev->dma_device.cap_mask);
802
803	fdev->dma_device.dev = &pdev->dev;
804	fdev->dma_device.device_alloc_chan_resources = st_fdma_alloc_chan_res;
805	fdev->dma_device.device_free_chan_resources = st_fdma_free_chan_res;
806	fdev->dma_device.device_prep_dma_cyclic	= st_fdma_prep_dma_cyclic;
807	fdev->dma_device.device_prep_slave_sg = st_fdma_prep_slave_sg;
808	fdev->dma_device.device_prep_dma_memcpy = st_fdma_prep_dma_memcpy;
809	fdev->dma_device.device_tx_status = st_fdma_tx_status;
810	fdev->dma_device.device_issue_pending = st_fdma_issue_pending;
811	fdev->dma_device.device_terminate_all = st_fdma_terminate_all;
812	fdev->dma_device.device_config = st_fdma_slave_config;
813	fdev->dma_device.device_pause = st_fdma_pause;
814	fdev->dma_device.device_resume = st_fdma_resume;
815
816	fdev->dma_device.src_addr_widths = FDMA_DMA_BUSWIDTHS;
817	fdev->dma_device.dst_addr_widths = FDMA_DMA_BUSWIDTHS;
818	fdev->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
819	fdev->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
820
821	ret = dmaenginem_async_device_register(&fdev->dma_device);
822	if (ret) {
823		dev_err(&pdev->dev,
824			"Failed to register DMA device (%d)\n", ret);
825		goto err_rproc;
826	}
827
828	ret = of_dma_controller_register(np, st_fdma_of_xlate, fdev);
829	if (ret) {
830		dev_err(&pdev->dev,
831			"Failed to register controller (%d)\n", ret);
832		goto err_rproc;
833	}
834
835	dev_info(&pdev->dev, "ST FDMA engine driver, irq:%d\n", fdev->irq);
836
837	return 0;
838
839err_rproc:
840	st_fdma_free(fdev);
841	st_slim_rproc_put(fdev->slim_rproc);
842err:
843	return ret;
844}
845
846static void st_fdma_remove(struct platform_device *pdev)
847{
848	struct st_fdma_dev *fdev = platform_get_drvdata(pdev);
849
850	devm_free_irq(&pdev->dev, fdev->irq, fdev);
851	st_slim_rproc_put(fdev->slim_rproc);
852	of_dma_controller_free(pdev->dev.of_node);
853}
854
855static struct platform_driver st_fdma_platform_driver = {
856	.driver = {
857		.name = DRIVER_NAME,
858		.of_match_table = st_fdma_match,
859	},
860	.probe = st_fdma_probe,
861	.remove_new = st_fdma_remove,
862};
863module_platform_driver(st_fdma_platform_driver);
864
865MODULE_LICENSE("GPL v2");
866MODULE_DESCRIPTION("STMicroelectronics FDMA engine driver");
867MODULE_AUTHOR("Ludovic.barre <Ludovic.barre@st.com>");
868MODULE_AUTHOR("Peter Griffin <peter.griffin@linaro.org>");
869MODULE_ALIAS("platform:" DRIVER_NAME);
870