1/* SPDX-License-Identifier: GPL-2.0+ */
2/*
3 * Copyright 2013-2014 Freescale Semiconductor, Inc.
4 * Copyright 2018 Angelo Dureghello <angelo@sysam.it>
5 */
6#ifndef _FSL_EDMA_COMMON_H_
7#define _FSL_EDMA_COMMON_H_
8
9#include <linux/dma-direction.h>
10#include <linux/platform_device.h>
11#include "virt-dma.h"
12
13#define EDMA_CR_EDBG		BIT(1)
14#define EDMA_CR_ERCA		BIT(2)
15#define EDMA_CR_ERGA		BIT(3)
16#define EDMA_CR_HOE		BIT(4)
17#define EDMA_CR_HALT		BIT(5)
18#define EDMA_CR_CLM		BIT(6)
19#define EDMA_CR_EMLM		BIT(7)
20#define EDMA_CR_ECX		BIT(16)
21#define EDMA_CR_CX		BIT(17)
22
23#define EDMA_SEEI_SEEI(x)	((x) & GENMASK(4, 0))
24#define EDMA_CEEI_CEEI(x)	((x) & GENMASK(4, 0))
25#define EDMA_CINT_CINT(x)	((x) & GENMASK(4, 0))
26#define EDMA_CERR_CERR(x)	((x) & GENMASK(4, 0))
27
28#define EDMA_TCD_ATTR_DSIZE(x)		(((x) & GENMASK(2, 0)))
29#define EDMA_TCD_ATTR_DMOD(x)		(((x) & GENMASK(4, 0)) << 3)
30#define EDMA_TCD_ATTR_SSIZE(x)		(((x) & GENMASK(2, 0)) << 8)
31#define EDMA_TCD_ATTR_SMOD(x)		(((x) & GENMASK(4, 0)) << 11)
32
33#define EDMA_TCD_ITER_MASK		GENMASK(14, 0)
34#define EDMA_TCD_CITER_CITER(x)		((x) & EDMA_TCD_ITER_MASK)
35#define EDMA_TCD_BITER_BITER(x)		((x) & EDMA_TCD_ITER_MASK)
36
37#define EDMA_TCD_CSR_START		BIT(0)
38#define EDMA_TCD_CSR_INT_MAJOR		BIT(1)
39#define EDMA_TCD_CSR_INT_HALF		BIT(2)
40#define EDMA_TCD_CSR_D_REQ		BIT(3)
41#define EDMA_TCD_CSR_E_SG		BIT(4)
42#define EDMA_TCD_CSR_E_LINK		BIT(5)
43#define EDMA_TCD_CSR_ACTIVE		BIT(6)
44#define EDMA_TCD_CSR_DONE		BIT(7)
45
46#define EDMA_V3_TCD_NBYTES_MLOFF_NBYTES(x) ((x) & GENMASK(9, 0))
47#define EDMA_V3_TCD_NBYTES_MLOFF(x)        (x << 10)
48#define EDMA_V3_TCD_NBYTES_DMLOE           (1 << 30)
49#define EDMA_V3_TCD_NBYTES_SMLOE           (1 << 31)
50
51#define EDMAMUX_CHCFG_DIS		0x0
52#define EDMAMUX_CHCFG_ENBL		0x80
53#define EDMAMUX_CHCFG_SOURCE(n)		((n) & 0x3F)
54
55#define DMAMUX_NR	2
56
57#define EDMA_TCD                0x1000
58
59#define FSL_EDMA_BUSWIDTHS	(BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
60				 BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
61				 BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) | \
62				 BIT(DMA_SLAVE_BUSWIDTH_8_BYTES))
63
64#define EDMA_V3_CH_SBR_RD          BIT(22)
65#define EDMA_V3_CH_SBR_WR          BIT(21)
66#define EDMA_V3_CH_CSR_ERQ         BIT(0)
67#define EDMA_V3_CH_CSR_EARQ        BIT(1)
68#define EDMA_V3_CH_CSR_EEI         BIT(2)
69#define EDMA_V3_CH_CSR_DONE        BIT(30)
70#define EDMA_V3_CH_CSR_ACTIVE      BIT(31)
71
72enum fsl_edma_pm_state {
73	RUNNING = 0,
74	SUSPENDED,
75};
76
77struct fsl_edma_hw_tcd {
78	__le32	saddr;
79	__le16	soff;
80	__le16	attr;
81	__le32	nbytes;
82	__le32	slast;
83	__le32	daddr;
84	__le16	doff;
85	__le16	citer;
86	__le32	dlast_sga;
87	__le16	csr;
88	__le16	biter;
89};
90
91struct fsl_edma_hw_tcd64 {
92	__le64  saddr;
93	__le16  soff;
94	__le16  attr;
95	__le32  nbytes;
96	__le64  slast;
97	__le64  daddr;
98	__le64  dlast_sga;
99	__le16  doff;
100	__le16  citer;
101	__le16  csr;
102	__le16  biter;
103} __packed;
104
105struct fsl_edma3_ch_reg {
106	__le32	ch_csr;
107	__le32	ch_es;
108	__le32	ch_int;
109	__le32	ch_sbr;
110	__le32	ch_pri;
111	__le32	ch_mux;
112	__le32  ch_mattr; /* edma4, reserved for edma3 */
113	__le32  ch_reserved;
114	union {
115		struct fsl_edma_hw_tcd tcd;
116		struct fsl_edma_hw_tcd64 tcd64;
117	};
118} __packed;
119
120/*
121 * These are iomem pointers, for both v32 and v64.
122 */
123struct edma_regs {
124	void __iomem *cr;
125	void __iomem *es;
126	void __iomem *erqh;
127	void __iomem *erql;	/* aka erq on v32 */
128	void __iomem *eeih;
129	void __iomem *eeil;	/* aka eei on v32 */
130	void __iomem *seei;
131	void __iomem *ceei;
132	void __iomem *serq;
133	void __iomem *cerq;
134	void __iomem *cint;
135	void __iomem *cerr;
136	void __iomem *ssrt;
137	void __iomem *cdne;
138	void __iomem *inth;
139	void __iomem *intl;
140	void __iomem *errh;
141	void __iomem *errl;
142};
143
144struct fsl_edma_sw_tcd {
145	dma_addr_t			ptcd;
146	void				*vtcd;
147};
148
149struct fsl_edma_chan {
150	struct virt_dma_chan		vchan;
151	enum dma_status			status;
152	enum fsl_edma_pm_state		pm_state;
153	bool				idle;
154	u32				slave_id;
155	struct fsl_edma_engine		*edma;
156	struct fsl_edma_desc		*edesc;
157	struct dma_slave_config		cfg;
158	u32				attr;
159	bool                            is_sw;
160	struct dma_pool			*tcd_pool;
161	dma_addr_t			dma_dev_addr;
162	u32				dma_dev_size;
163	enum dma_data_direction		dma_dir;
164	char				chan_name[32];
165	void __iomem			*tcd;
166	void __iomem			*mux_addr;
167	u32				real_count;
168	struct work_struct		issue_worker;
169	struct platform_device		*pdev;
170	struct device			*pd_dev;
171	u32				srcid;
172	struct clk			*clk;
173	int                             priority;
174	int				hw_chanid;
175	int				txirq;
176	bool				is_rxchan;
177	bool				is_remote;
178	bool				is_multi_fifo;
179};
180
181struct fsl_edma_desc {
182	struct virt_dma_desc		vdesc;
183	struct fsl_edma_chan		*echan;
184	bool				iscyclic;
185	enum dma_transfer_direction	dirn;
186	unsigned int			n_tcds;
187	struct fsl_edma_sw_tcd		tcd[];
188};
189
190#define FSL_EDMA_DRV_HAS_DMACLK		BIT(0)
191#define FSL_EDMA_DRV_MUX_SWAP		BIT(1)
192#define FSL_EDMA_DRV_CONFIG32		BIT(2)
193#define FSL_EDMA_DRV_WRAP_IO		BIT(3)
194#define FSL_EDMA_DRV_EDMA64		BIT(4)
195#define FSL_EDMA_DRV_HAS_PD		BIT(5)
196#define FSL_EDMA_DRV_HAS_CHCLK		BIT(6)
197#define FSL_EDMA_DRV_HAS_CHMUX		BIT(7)
198/* imx8 QM audio edma remote local swapped */
199#define FSL_EDMA_DRV_QUIRK_SWAPPED	BIT(8)
200/* control and status register is in tcd address space, edma3 reg layout */
201#define FSL_EDMA_DRV_SPLIT_REG		BIT(9)
202#define FSL_EDMA_DRV_BUS_8BYTE		BIT(10)
203#define FSL_EDMA_DRV_DEV_TO_DEV		BIT(11)
204#define FSL_EDMA_DRV_ALIGN_64BYTE	BIT(12)
205/* Need clean CHn_CSR DONE before enable TCD's ESG */
206#define FSL_EDMA_DRV_CLEAR_DONE_E_SG	BIT(13)
207/* Need clean CHn_CSR DONE before enable TCD's MAJORELINK */
208#define FSL_EDMA_DRV_CLEAR_DONE_E_LINK	BIT(14)
209#define FSL_EDMA_DRV_TCD64		BIT(15)
210
211#define FSL_EDMA_DRV_EDMA3	(FSL_EDMA_DRV_SPLIT_REG |	\
212				 FSL_EDMA_DRV_BUS_8BYTE |	\
213				 FSL_EDMA_DRV_DEV_TO_DEV |	\
214				 FSL_EDMA_DRV_ALIGN_64BYTE |	\
215				 FSL_EDMA_DRV_CLEAR_DONE_E_SG |	\
216				 FSL_EDMA_DRV_CLEAR_DONE_E_LINK)
217
218#define FSL_EDMA_DRV_EDMA4	(FSL_EDMA_DRV_SPLIT_REG |	\
219				 FSL_EDMA_DRV_BUS_8BYTE |	\
220				 FSL_EDMA_DRV_DEV_TO_DEV |	\
221				 FSL_EDMA_DRV_ALIGN_64BYTE |	\
222				 FSL_EDMA_DRV_CLEAR_DONE_E_LINK)
223
224struct fsl_edma_drvdata {
225	u32			dmamuxs; /* only used before v3 */
226	u32			chreg_off;
227	u32			chreg_space_sz;
228	u32			flags;
229	u32			mux_off;	/* channel mux register offset */
230	u32			mux_skip;	/* how much skip for each channel */
231	int			(*setup_irq)(struct platform_device *pdev,
232					     struct fsl_edma_engine *fsl_edma);
233};
234
235struct fsl_edma_engine {
236	struct dma_device	dma_dev;
237	void __iomem		*membase;
238	void __iomem		*muxbase[DMAMUX_NR];
239	struct clk		*muxclk[DMAMUX_NR];
240	struct clk		*dmaclk;
241	struct clk		*chclk;
242	struct mutex		fsl_edma_mutex;
243	const struct fsl_edma_drvdata *drvdata;
244	u32			n_chans;
245	int			txirq;
246	int			errirq;
247	bool			big_endian;
248	struct edma_regs	regs;
249	u64			chan_masked;
250	struct fsl_edma_chan	chans[] __counted_by(n_chans);
251};
252
253#define edma_read_tcdreg_c(chan, _tcd,  __name)				\
254(sizeof((_tcd)->__name) == sizeof(u64) ?				\
255	edma_readq(chan->edma, &(_tcd)->__name) :			\
256		((sizeof((_tcd)->__name) == sizeof(u32)) ?		\
257			edma_readl(chan->edma, &(_tcd)->__name) :	\
258			edma_readw(chan->edma, &(_tcd)->__name)		\
259		))
260
261#define edma_read_tcdreg(chan, __name)								\
262((fsl_edma_drvflags(chan) & FSL_EDMA_DRV_TCD64) ?						\
263	edma_read_tcdreg_c(chan, ((struct fsl_edma_hw_tcd64 __iomem *)chan->tcd), __name) :	\
264	edma_read_tcdreg_c(chan, ((struct fsl_edma_hw_tcd __iomem *)chan->tcd), __name)		\
265)
266
267#define edma_write_tcdreg_c(chan, _tcd, _val, __name)				\
268do {										\
269	switch (sizeof(_tcd->__name)) {						\
270	case sizeof(u64):							\
271		edma_writeq(chan->edma, (u64 __force)_val, &_tcd->__name);	\
272		break;								\
273	case sizeof(u32):							\
274		edma_writel(chan->edma, (u32 __force)_val, &_tcd->__name);	\
275		break;								\
276	case sizeof(u16):							\
277		edma_writew(chan->edma, (u16 __force)_val, &_tcd->__name);	\
278		break;								\
279	case sizeof(u8):							\
280		edma_writeb(chan->edma, (u8 __force)_val, &_tcd->__name);	\
281		break;								\
282	}									\
283} while (0)
284
285#define edma_write_tcdreg(chan, val, __name)							   \
286do {												   \
287	struct fsl_edma_hw_tcd64 __iomem *tcd64_r = (struct fsl_edma_hw_tcd64 __iomem *)chan->tcd; \
288	struct fsl_edma_hw_tcd __iomem *tcd_r = (struct fsl_edma_hw_tcd __iomem *)chan->tcd;	   \
289												   \
290	if (fsl_edma_drvflags(chan) & FSL_EDMA_DRV_TCD64)					   \
291		edma_write_tcdreg_c(chan, tcd64_r, val, __name);				   \
292	else											   \
293		edma_write_tcdreg_c(chan, tcd_r, val, __name);					   \
294} while (0)
295
296#define edma_cp_tcd_to_reg(chan, __tcd, __name)							   \
297do {	\
298	struct fsl_edma_hw_tcd64 __iomem *tcd64_r = (struct fsl_edma_hw_tcd64 __iomem *)chan->tcd; \
299	struct fsl_edma_hw_tcd __iomem *tcd_r = (struct fsl_edma_hw_tcd __iomem *)chan->tcd;	   \
300	struct fsl_edma_hw_tcd64 *tcd64_m = (struct fsl_edma_hw_tcd64 *)__tcd;			   \
301	struct fsl_edma_hw_tcd *tcd_m = (struct fsl_edma_hw_tcd *)__tcd;			   \
302												   \
303	if (fsl_edma_drvflags(chan) & FSL_EDMA_DRV_TCD64)					   \
304		edma_write_tcdreg_c(chan, tcd64_r,  tcd64_m->__name, __name);			   \
305	else											   \
306		edma_write_tcdreg_c(chan, tcd_r, tcd_m->__name, __name);			   \
307} while (0)
308
309#define edma_readl_chreg(chan, __name)				\
310	edma_readl(chan->edma,					\
311		   (void __iomem *)&(container_of(((__force void *)chan->tcd),\
312						  struct fsl_edma3_ch_reg, tcd)->__name))
313
314#define edma_writel_chreg(chan, val,  __name)			\
315	edma_writel(chan->edma, val,				\
316		   (void __iomem *)&(container_of(((__force void *)chan->tcd),\
317						  struct fsl_edma3_ch_reg, tcd)->__name))
318
319#define fsl_edma_get_tcd(_chan, _tcd, _field)			\
320(fsl_edma_drvflags(_chan) & FSL_EDMA_DRV_TCD64 ? (((struct fsl_edma_hw_tcd64 *)_tcd)->_field) : \
321						 (((struct fsl_edma_hw_tcd *)_tcd)->_field))
322
323#define fsl_edma_le_to_cpu(x)						\
324(sizeof(x) == sizeof(u64) ? le64_to_cpu((__force __le64)(x)) :		\
325	(sizeof(x) == sizeof(u32) ? le32_to_cpu((__force __le32)(x)) :	\
326				    le16_to_cpu((__force __le16)(x))))
327
328#define fsl_edma_get_tcd_to_cpu(_chan, _tcd, _field)				\
329(fsl_edma_drvflags(_chan) & FSL_EDMA_DRV_TCD64 ?				\
330	fsl_edma_le_to_cpu(((struct fsl_edma_hw_tcd64 *)_tcd)->_field) :	\
331	fsl_edma_le_to_cpu(((struct fsl_edma_hw_tcd *)_tcd)->_field))
332
333#define fsl_edma_set_tcd_to_le_c(_tcd, _val, _field)					\
334do {											\
335	switch (sizeof((_tcd)->_field)) {						\
336	case sizeof(u64):								\
337		*(__force __le64 *)(&((_tcd)->_field)) = cpu_to_le64(_val);		\
338		break;									\
339	case sizeof(u32):								\
340		*(__force __le32 *)(&((_tcd)->_field)) = cpu_to_le32(_val);		\
341		break;									\
342	case sizeof(u16):								\
343		*(__force __le16 *)(&((_tcd)->_field)) = cpu_to_le16(_val);		\
344		break;									\
345	}										\
346} while (0)
347
348#define fsl_edma_set_tcd_to_le(_chan, _tcd, _val, _field)	\
349do {								\
350	if (fsl_edma_drvflags(_chan) & FSL_EDMA_DRV_TCD64)	\
351		fsl_edma_set_tcd_to_le_c((struct fsl_edma_hw_tcd64 *)_tcd, _val, _field);	\
352	else											\
353		fsl_edma_set_tcd_to_le_c((struct fsl_edma_hw_tcd *)_tcd, _val, _field);		\
354} while (0)
355
356/*
357 * R/W functions for big- or little-endian registers:
358 * The eDMA controller's endian is independent of the CPU core's endian.
359 * For the big-endian IP module, the offset for 8-bit or 16-bit registers
360 * should also be swapped opposite to that in little-endian IP.
361 */
362static inline u64 edma_readq(struct fsl_edma_engine *edma, void __iomem *addr)
363{
364	u64 l, h;
365
366	if (edma->big_endian) {
367		l = ioread32be(addr);
368		h = ioread32be(addr + 4);
369	} else {
370		l = ioread32(addr);
371		h = ioread32(addr + 4);
372	}
373
374	return (h << 32) | l;
375}
376
377static inline u32 edma_readl(struct fsl_edma_engine *edma, void __iomem *addr)
378{
379	if (edma->big_endian)
380		return ioread32be(addr);
381	else
382		return ioread32(addr);
383}
384
385static inline u16 edma_readw(struct fsl_edma_engine *edma, void __iomem *addr)
386{
387	if (edma->big_endian)
388		return ioread16be(addr);
389	else
390		return ioread16(addr);
391}
392
393static inline void edma_writeb(struct fsl_edma_engine *edma,
394			       u8 val, void __iomem *addr)
395{
396	/* swap the reg offset for these in big-endian mode */
397	if (edma->big_endian)
398		iowrite8(val, (void __iomem *)((unsigned long)addr ^ 0x3));
399	else
400		iowrite8(val, addr);
401}
402
403static inline void edma_writew(struct fsl_edma_engine *edma,
404			       u16 val, void __iomem *addr)
405{
406	/* swap the reg offset for these in big-endian mode */
407	if (edma->big_endian)
408		iowrite16be(val, (void __iomem *)((unsigned long)addr ^ 0x2));
409	else
410		iowrite16(val, addr);
411}
412
413static inline void edma_writel(struct fsl_edma_engine *edma,
414			       u32 val, void __iomem *addr)
415{
416	if (edma->big_endian)
417		iowrite32be(val, addr);
418	else
419		iowrite32(val, addr);
420}
421
422static inline void edma_writeq(struct fsl_edma_engine *edma,
423			       u64 val, void __iomem *addr)
424{
425	if (edma->big_endian) {
426		iowrite32be(val & 0xFFFFFFFF, addr);
427		iowrite32be(val >> 32, addr + 4);
428	} else {
429		iowrite32(val & 0xFFFFFFFF, addr);
430		iowrite32(val >> 32, addr + 4);
431	}
432}
433
434static inline struct fsl_edma_chan *to_fsl_edma_chan(struct dma_chan *chan)
435{
436	return container_of(chan, struct fsl_edma_chan, vchan.chan);
437}
438
439static inline u32 fsl_edma_drvflags(struct fsl_edma_chan *fsl_chan)
440{
441	return fsl_chan->edma->drvdata->flags;
442}
443
444static inline struct fsl_edma_desc *to_fsl_edma_desc(struct virt_dma_desc *vd)
445{
446	return container_of(vd, struct fsl_edma_desc, vdesc);
447}
448
449static inline void fsl_edma_err_chan_handler(struct fsl_edma_chan *fsl_chan)
450{
451	fsl_chan->status = DMA_ERROR;
452	fsl_chan->idle = true;
453}
454
455void fsl_edma_tx_chan_handler(struct fsl_edma_chan *fsl_chan);
456void fsl_edma_disable_request(struct fsl_edma_chan *fsl_chan);
457void fsl_edma_chan_mux(struct fsl_edma_chan *fsl_chan,
458			unsigned int slot, bool enable);
459void fsl_edma_free_desc(struct virt_dma_desc *vdesc);
460int fsl_edma_terminate_all(struct dma_chan *chan);
461int fsl_edma_pause(struct dma_chan *chan);
462int fsl_edma_resume(struct dma_chan *chan);
463int fsl_edma_slave_config(struct dma_chan *chan,
464				 struct dma_slave_config *cfg);
465enum dma_status fsl_edma_tx_status(struct dma_chan *chan,
466		dma_cookie_t cookie, struct dma_tx_state *txstate);
467struct dma_async_tx_descriptor *fsl_edma_prep_dma_cyclic(
468		struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len,
469		size_t period_len, enum dma_transfer_direction direction,
470		unsigned long flags);
471struct dma_async_tx_descriptor *fsl_edma_prep_slave_sg(
472		struct dma_chan *chan, struct scatterlist *sgl,
473		unsigned int sg_len, enum dma_transfer_direction direction,
474		unsigned long flags, void *context);
475struct dma_async_tx_descriptor *fsl_edma_prep_memcpy(
476		struct dma_chan *chan, dma_addr_t dma_dst, dma_addr_t dma_src,
477		size_t len, unsigned long flags);
478void fsl_edma_xfer_desc(struct fsl_edma_chan *fsl_chan);
479void fsl_edma_issue_pending(struct dma_chan *chan);
480int fsl_edma_alloc_chan_resources(struct dma_chan *chan);
481void fsl_edma_free_chan_resources(struct dma_chan *chan);
482void fsl_edma_cleanup_vchan(struct dma_device *dmadev);
483void fsl_edma_setup_regs(struct fsl_edma_engine *edma);
484
485#endif /* _FSL_EDMA_COMMON_H_ */
486