1// SPDX-License-Identifier: GPL-2.0
2// PCI1xxxx SPI driver
3// Copyright (C) 2022 Microchip Technology Inc.
4// Authors: Tharun Kumar P <tharunkumar.pasumarthi@microchip.com>
5//          Kumaravel Thiagarajan <Kumaravel.Thiagarajan@microchip.com>
6
7
8#include <linux/bitfield.h>
9#include <linux/dma-mapping.h>
10#include <linux/iopoll.h>
11#include <linux/irq.h>
12#include <linux/module.h>
13#include <linux/msi.h>
14#include <linux/pci_regs.h>
15#include <linux/pci.h>
16#include <linux/spinlock.h>
17#include <linux/spi/spi.h>
18#include <linux/delay.h>
19
20#define DRV_NAME "spi-pci1xxxx"
21
22#define	SYS_FREQ_DEFAULT		(62500000)
23
24#define	PCI1XXXX_SPI_MAX_CLOCK_HZ	(30000000)
25#define	PCI1XXXX_SPI_CLK_20MHZ		(20000000)
26#define	PCI1XXXX_SPI_CLK_15MHZ		(15000000)
27#define	PCI1XXXX_SPI_CLK_12MHZ		(12000000)
28#define	PCI1XXXX_SPI_CLK_10MHZ		(10000000)
29#define	PCI1XXXX_SPI_MIN_CLOCK_HZ	(2000000)
30
31#define	PCI1XXXX_SPI_BUFFER_SIZE	(320)
32
33#define	SPI_MST_CTL_DEVSEL_MASK		(GENMASK(27, 25))
34#define	SPI_MST_CTL_CMD_LEN_MASK	(GENMASK(16, 8))
35#define	SPI_MST_CTL_SPEED_MASK		(GENMASK(7, 5))
36#define	SPI_MSI_VECTOR_SEL_MASK		(GENMASK(4, 4))
37
38#define	SPI_MST_CTL_FORCE_CE		(BIT(4))
39#define	SPI_MST_CTL_MODE_SEL		(BIT(2))
40#define	SPI_MST_CTL_GO			(BIT(0))
41
42#define SPI_PERI_ADDR_BASE		(0x160000)
43#define SPI_SYSTEM_ADDR_BASE		(0x2000)
44#define	SPI_MST1_ADDR_BASE		(0x800)
45
46#define DEV_REV_REG			(SPI_SYSTEM_ADDR_BASE + 0x00)
47#define SPI_SYSLOCK_REG			(SPI_SYSTEM_ADDR_BASE + 0xA0)
48#define SPI_CONFIG_PERI_ENABLE_REG	(SPI_SYSTEM_ADDR_BASE + 0x108)
49
50#define SPI_PERI_ENBLE_PF_MASK		(GENMASK(17, 16))
51#define DEV_REV_MASK			(GENMASK(7, 0))
52
53#define SPI_SYSLOCK			BIT(4)
54#define SPI0				(0)
55#define SPI1				(1)
56
57/* DMA Related Registers */
58#define SPI_DMA_ADDR_BASE		(0x1000)
59#define SPI_DMA_GLOBAL_WR_ENGINE_EN	(SPI_DMA_ADDR_BASE + 0x0C)
60#define SPI_DMA_WR_DOORBELL_REG		(SPI_DMA_ADDR_BASE + 0x10)
61#define SPI_DMA_GLOBAL_RD_ENGINE_EN	(SPI_DMA_ADDR_BASE + 0x2C)
62#define SPI_DMA_RD_DOORBELL_REG		(SPI_DMA_ADDR_BASE + 0x30)
63#define SPI_DMA_INTR_WR_STS		(SPI_DMA_ADDR_BASE + 0x4C)
64#define SPI_DMA_WR_INT_MASK		(SPI_DMA_ADDR_BASE + 0x54)
65#define SPI_DMA_INTR_WR_CLR		(SPI_DMA_ADDR_BASE + 0x58)
66#define SPI_DMA_ERR_WR_STS		(SPI_DMA_ADDR_BASE + 0x5C)
67#define SPI_DMA_INTR_IMWR_WDONE_LOW	(SPI_DMA_ADDR_BASE + 0x60)
68#define SPI_DMA_INTR_IMWR_WDONE_HIGH	(SPI_DMA_ADDR_BASE + 0x64)
69#define SPI_DMA_INTR_IMWR_WABORT_LOW	(SPI_DMA_ADDR_BASE + 0x68)
70#define SPI_DMA_INTR_IMWR_WABORT_HIGH	(SPI_DMA_ADDR_BASE + 0x6C)
71#define SPI_DMA_INTR_WR_IMWR_DATA	(SPI_DMA_ADDR_BASE + 0x70)
72#define SPI_DMA_INTR_RD_STS		(SPI_DMA_ADDR_BASE + 0xA0)
73#define SPI_DMA_RD_INT_MASK		(SPI_DMA_ADDR_BASE + 0xA8)
74#define SPI_DMA_INTR_RD_CLR		(SPI_DMA_ADDR_BASE + 0xAC)
75#define SPI_DMA_ERR_RD_STS		(SPI_DMA_ADDR_BASE + 0xB8)
76#define SPI_DMA_INTR_IMWR_RDONE_LOW	(SPI_DMA_ADDR_BASE + 0xCC)
77#define SPI_DMA_INTR_IMWR_RDONE_HIGH	(SPI_DMA_ADDR_BASE + 0xD0)
78#define SPI_DMA_INTR_IMWR_RABORT_LOW	(SPI_DMA_ADDR_BASE + 0xD4)
79#define SPI_DMA_INTR_IMWR_RABORT_HIGH	(SPI_DMA_ADDR_BASE + 0xD8)
80#define SPI_DMA_INTR_RD_IMWR_DATA	(SPI_DMA_ADDR_BASE + 0xDC)
81
82#define SPI_DMA_CH0_WR_BASE		(SPI_DMA_ADDR_BASE + 0x200)
83#define SPI_DMA_CH0_RD_BASE		(SPI_DMA_ADDR_BASE + 0x300)
84#define SPI_DMA_CH1_WR_BASE		(SPI_DMA_ADDR_BASE + 0x400)
85#define SPI_DMA_CH1_RD_BASE		(SPI_DMA_ADDR_BASE + 0x500)
86
87#define SPI_DMA_CH_CTL1_OFFSET		(0x00)
88#define SPI_DMA_CH_XFER_LEN_OFFSET	(0x08)
89#define SPI_DMA_CH_SAR_LO_OFFSET	(0x0C)
90#define SPI_DMA_CH_SAR_HI_OFFSET	(0x10)
91#define SPI_DMA_CH_DAR_LO_OFFSET	(0x14)
92#define SPI_DMA_CH_DAR_HI_OFFSET	(0x18)
93
94#define SPI_DMA_CH0_DONE_INT		BIT(0)
95#define SPI_DMA_CH1_DONE_INT		BIT(1)
96#define SPI_DMA_CH0_ABORT_INT		BIT(16)
97#define SPI_DMA_CH1_ABORT_INT		BIT(17)
98#define SPI_DMA_DONE_INT_MASK		(SPI_DMA_CH0_DONE_INT | SPI_DMA_CH1_DONE_INT)
99#define SPI_DMA_ABORT_INT_MASK		(SPI_DMA_CH0_ABORT_INT | SPI_DMA_CH1_ABORT_INT)
100#define DMA_CH_CONTROL_LIE		BIT(3)
101#define DMA_CH_CONTROL_RIE		BIT(4)
102#define DMA_INTR_EN			(DMA_CH_CONTROL_RIE | DMA_CH_CONTROL_LIE)
103
104/* x refers to SPI Host Controller HW instance id in the below macros - 0 or 1 */
105
106#define	SPI_MST_CMD_BUF_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x00)
107#define	SPI_MST_RSP_BUF_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x200)
108#define	SPI_MST_CTL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x400)
109#define	SPI_MST_EVENT_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x420)
110#define	SPI_MST_EVENT_MASK_REG_OFFSET(x)	(((x) * SPI_MST1_ADDR_BASE) + 0x424)
111#define	SPI_MST_PAD_CTL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x460)
112#define	SPIALERT_MST_DB_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x464)
113#define	SPIALERT_MST_VAL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x468)
114#define	SPI_PCI_CTRL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x480)
115
116#define PCI1XXXX_IRQ_FLAGS			(IRQF_NO_SUSPEND | IRQF_TRIGGER_NONE)
117#define SPI_MAX_DATA_LEN			320
118
119#define PCI1XXXX_SPI_TIMEOUT			(msecs_to_jiffies(100))
120#define SYSLOCK_RETRY_CNT			(1000)
121#define SPI_DMA_ENGINE_EN			(0x1)
122#define SPI_DMA_ENGINE_DIS			(0x0)
123
124#define SPI_INTR		BIT(8)
125#define SPI_FORCE_CE		BIT(4)
126
127#define SPI_CHIP_SEL_COUNT 7
128#define VENDOR_ID_MCHP 0x1055
129
130#define SPI_SUSPEND_CONFIG 0x101
131#define SPI_RESUME_CONFIG 0x203
132
133struct pci1xxxx_spi_internal {
134	u8 hw_inst;
135	u8 clkdiv;
136	int irq;
137	int mode;
138	bool spi_xfer_in_progress;
139	void *rx_buf;
140	bool dma_aborted_rd;
141	u32 bytes_recvd;
142	u32 tx_sgl_len;
143	u32 rx_sgl_len;
144	struct scatterlist *tx_sgl, *rx_sgl;
145	bool dma_aborted_wr;
146	struct completion spi_xfer_done;
147	struct spi_controller *spi_host;
148	struct pci1xxxx_spi *parent;
149	struct spi_transfer *xfer;
150	struct {
151		unsigned int dev_sel : 3;
152		unsigned int msi_vector_sel : 1;
153	} prev_val;
154};
155
156struct pci1xxxx_spi {
157	struct pci_dev *dev;
158	u8 total_hw_instances;
159	u8 dev_rev;
160	void __iomem *reg_base;
161	void __iomem *dma_offset_bar;
162	/* lock to safely access the DMA registers in isr */
163	spinlock_t dma_reg_lock;
164	bool can_dma;
165	struct pci1xxxx_spi_internal *spi_int[] __counted_by(total_hw_instances);
166};
167
168static const struct pci_device_id pci1xxxx_spi_pci_id_table[] = {
169	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
170	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
171	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
172	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
173	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
174	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
175	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
176	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
177	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
178	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
179	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
180	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
181	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
182	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
183	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
184	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
185	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
186	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
187	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
188	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
189	{ 0, }
190};
191
192MODULE_DEVICE_TABLE(pci, pci1xxxx_spi_pci_id_table);
193
194static int pci1xxxx_set_sys_lock(struct pci1xxxx_spi *par)
195{
196	writel(SPI_SYSLOCK, par->reg_base + SPI_SYSLOCK_REG);
197	return readl(par->reg_base + SPI_SYSLOCK_REG);
198}
199
200static int pci1xxxx_acquire_sys_lock(struct pci1xxxx_spi *par)
201{
202	u32 regval;
203
204	return readx_poll_timeout(pci1xxxx_set_sys_lock, par, regval,
205			   (regval & SPI_SYSLOCK), 100,
206			   SYSLOCK_RETRY_CNT * 100);
207}
208
209static void pci1xxxx_release_sys_lock(struct pci1xxxx_spi *par)
210{
211	writel(0x0, par->reg_base + SPI_SYSLOCK_REG);
212}
213
214static int pci1xxxx_check_spi_can_dma(struct pci1xxxx_spi *spi_bus, int irq)
215{
216	struct pci_dev *pdev = spi_bus->dev;
217	u32 pf_num;
218	u32 regval;
219	int ret;
220
221	/*
222	 * DEV REV Registers is a system register, HW Syslock bit
223	 * should be acquired before accessing the register
224	 */
225	ret = pci1xxxx_acquire_sys_lock(spi_bus);
226	if (ret) {
227		dev_err(&pdev->dev, "Error failed to acquire syslock\n");
228		return ret;
229	}
230
231	regval = readl(spi_bus->reg_base + DEV_REV_REG);
232	spi_bus->dev_rev = regval & DEV_REV_MASK;
233	if (spi_bus->dev_rev >= 0xC0) {
234		regval = readl(spi_bus->reg_base +
235			       SPI_CONFIG_PERI_ENABLE_REG);
236		pf_num = regval & SPI_PERI_ENBLE_PF_MASK;
237	}
238
239	pci1xxxx_release_sys_lock(spi_bus);
240
241	/*
242	 * DMA is supported only from C0 and SPI can use DMA only if
243	 * it is mapped to PF0
244	 */
245	if (spi_bus->dev_rev < 0xC0 || pf_num)
246		return -EOPNOTSUPP;
247
248	/*
249	 * DMA Supported only with MSI Interrupts
250	 * One of the SPI instance's MSI vector address and data
251	 * is used for DMA Interrupt
252	 */
253	if (!irq_get_msi_desc(irq)) {
254		dev_warn(&pdev->dev, "Error MSI Interrupt not supported, will operate in PIO mode\n");
255		return -EOPNOTSUPP;
256	}
257
258	spi_bus->dma_offset_bar = pcim_iomap(pdev, 2, pci_resource_len(pdev, 2));
259	if (!spi_bus->dma_offset_bar) {
260		dev_warn(&pdev->dev, "Error failed to map dma bar, will operate in PIO mode\n");
261		return -EOPNOTSUPP;
262	}
263
264	if (dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64))) {
265		dev_warn(&pdev->dev, "Error failed to set DMA mask, will operate in PIO mode\n");
266		pcim_iounmap(pdev, spi_bus->dma_offset_bar);
267		spi_bus->dma_offset_bar = NULL;
268		return -EOPNOTSUPP;
269	}
270
271	return 0;
272}
273
274static int pci1xxxx_spi_dma_init(struct pci1xxxx_spi *spi_bus, int irq)
275{
276	struct msi_msg msi;
277	int ret;
278
279	ret = pci1xxxx_check_spi_can_dma(spi_bus, irq);
280	if (ret)
281		return ret;
282
283	spin_lock_init(&spi_bus->dma_reg_lock);
284	get_cached_msi_msg(irq, &msi);
285	writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
286	writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
287	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_HIGH);
288	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_HIGH);
289	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_HIGH);
290	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_HIGH);
291	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_LOW);
292	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_LOW);
293	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_LOW);
294	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_LOW);
295	writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_WR_IMWR_DATA);
296	writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_RD_IMWR_DATA);
297	dma_set_max_seg_size(&spi_bus->dev->dev, PCI1XXXX_SPI_BUFFER_SIZE);
298	spi_bus->can_dma = true;
299	return 0;
300}
301
302static void pci1xxxx_spi_set_cs(struct spi_device *spi, bool enable)
303{
304	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi->controller);
305	struct pci1xxxx_spi *par = p->parent;
306	u32 regval;
307
308	/* Set the DEV_SEL bits of the SPI_MST_CTL_REG */
309	regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
310	if (!enable) {
311		regval |= SPI_FORCE_CE;
312		regval &= ~SPI_MST_CTL_DEVSEL_MASK;
313		regval |= (spi_get_chipselect(spi, 0) << 25);
314	} else {
315		regval &= ~SPI_FORCE_CE;
316	}
317	writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
318}
319
320static u8 pci1xxxx_get_clock_div(u32 hz)
321{
322	u8 val = 0;
323
324	if (hz >= PCI1XXXX_SPI_MAX_CLOCK_HZ)
325		val = 2;
326	else if ((hz < PCI1XXXX_SPI_MAX_CLOCK_HZ) && (hz >= PCI1XXXX_SPI_CLK_20MHZ))
327		val = 3;
328	else if ((hz < PCI1XXXX_SPI_CLK_20MHZ) && (hz >= PCI1XXXX_SPI_CLK_15MHZ))
329		val = 4;
330	else if ((hz < PCI1XXXX_SPI_CLK_15MHZ) && (hz >= PCI1XXXX_SPI_CLK_12MHZ))
331		val = 5;
332	else if ((hz < PCI1XXXX_SPI_CLK_12MHZ) && (hz >= PCI1XXXX_SPI_CLK_10MHZ))
333		val = 6;
334	else if ((hz < PCI1XXXX_SPI_CLK_10MHZ) && (hz >= PCI1XXXX_SPI_MIN_CLOCK_HZ))
335		val = 7;
336	else
337		val = 2;
338
339	return val;
340}
341
342static void pci1xxxx_spi_setup_dma_to_io(struct pci1xxxx_spi_internal *p,
343					 dma_addr_t dma_addr, u32 len)
344{
345	void __iomem *base;
346
347	if (!p->hw_inst)
348		base = p->parent->dma_offset_bar + SPI_DMA_CH0_RD_BASE;
349	else
350		base = p->parent->dma_offset_bar + SPI_DMA_CH1_RD_BASE;
351
352	writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
353	writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
354	writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_SAR_LO_OFFSET);
355	writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_SAR_HI_OFFSET);
356	/* Updated SPI Command Registers */
357	writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
358	       base + SPI_DMA_CH_DAR_LO_OFFSET);
359	writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
360	       base + SPI_DMA_CH_DAR_HI_OFFSET);
361}
362
363static void pci1xxxx_spi_setup_dma_from_io(struct pci1xxxx_spi_internal *p,
364					   dma_addr_t dma_addr, u32 len)
365{
366	void *base;
367
368	if (!p->hw_inst)
369		base = p->parent->dma_offset_bar + SPI_DMA_CH0_WR_BASE;
370	else
371		base = p->parent->dma_offset_bar + SPI_DMA_CH1_WR_BASE;
372
373	writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
374	writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
375	writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_DAR_LO_OFFSET);
376	writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_DAR_HI_OFFSET);
377	writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
378	       base + SPI_DMA_CH_SAR_LO_OFFSET);
379	writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
380	       base + SPI_DMA_CH_SAR_HI_OFFSET);
381}
382
383static void pci1xxxx_spi_setup(struct pci1xxxx_spi *par, u8 hw_inst, u32 mode,
384			       u8 clkdiv, u32 len)
385{
386	u32 regval;
387
388	regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
389	regval &= ~(SPI_MST_CTL_MODE_SEL | SPI_MST_CTL_CMD_LEN_MASK |
390		    SPI_MST_CTL_SPEED_MASK);
391
392	if (mode == SPI_MODE_3)
393		regval |= SPI_MST_CTL_MODE_SEL;
394
395	regval |= FIELD_PREP(SPI_MST_CTL_CMD_LEN_MASK, len);
396	regval |= FIELD_PREP(SPI_MST_CTL_SPEED_MASK, clkdiv);
397	writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
398}
399
400static void pci1xxxx_start_spi_xfer(struct pci1xxxx_spi_internal *p, u8 hw_inst)
401{
402	u32 regval;
403
404	regval = readl(p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
405	regval |= SPI_MST_CTL_GO;
406	writel(regval, p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
407}
408
409static int pci1xxxx_spi_transfer_with_io(struct spi_controller *spi_ctlr,
410					 struct spi_device *spi, struct spi_transfer *xfer)
411{
412	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
413	struct pci1xxxx_spi *par = p->parent;
414	int len, loop_iter, transfer_len;
415	unsigned long bytes_transfered;
416	unsigned long bytes_recvd;
417	unsigned long loop_count;
418	u8 *rx_buf, result;
419	const u8 *tx_buf;
420	u32 regval;
421	u8 clkdiv;
422
423	p->spi_xfer_in_progress = true;
424	p->bytes_recvd = 0;
425	clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz);
426	tx_buf = xfer->tx_buf;
427	rx_buf = xfer->rx_buf;
428	transfer_len = xfer->len;
429	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
430	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
431
432	if (tx_buf) {
433		bytes_transfered = 0;
434		bytes_recvd = 0;
435		loop_count = transfer_len / SPI_MAX_DATA_LEN;
436		if (transfer_len % SPI_MAX_DATA_LEN != 0)
437			loop_count += 1;
438
439		for (loop_iter = 0; loop_iter < loop_count; loop_iter++) {
440			len = SPI_MAX_DATA_LEN;
441			if ((transfer_len % SPI_MAX_DATA_LEN != 0) &&
442			    (loop_iter == loop_count - 1))
443				len = transfer_len % SPI_MAX_DATA_LEN;
444
445			reinit_completion(&p->spi_xfer_done);
446			memcpy_toio(par->reg_base + SPI_MST_CMD_BUF_OFFSET(p->hw_inst),
447				    &tx_buf[bytes_transfered], len);
448			bytes_transfered += len;
449			pci1xxxx_spi_setup(par, p->hw_inst, spi->mode, clkdiv, len);
450			pci1xxxx_start_spi_xfer(p, p->hw_inst);
451
452			/* Wait for DMA_TERM interrupt */
453			result = wait_for_completion_timeout(&p->spi_xfer_done,
454							     PCI1XXXX_SPI_TIMEOUT);
455			if (!result)
456				return -ETIMEDOUT;
457
458			if (rx_buf) {
459				memcpy_fromio(&rx_buf[bytes_recvd], par->reg_base +
460					      SPI_MST_RSP_BUF_OFFSET(p->hw_inst), len);
461				bytes_recvd += len;
462			}
463		}
464	}
465	p->spi_xfer_in_progress = false;
466
467	return 0;
468}
469
470static int pci1xxxx_spi_transfer_with_dma(struct spi_controller *spi_ctlr,
471					  struct spi_device *spi,
472					  struct spi_transfer *xfer)
473{
474	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
475	struct pci1xxxx_spi *par = p->parent;
476	dma_addr_t rx_dma_addr = 0;
477	dma_addr_t tx_dma_addr = 0;
478	int ret = 0;
479	u32 regval;
480
481	p->spi_xfer_in_progress = true;
482	p->tx_sgl = xfer->tx_sg.sgl;
483	p->rx_sgl = xfer->rx_sg.sgl;
484	p->rx_buf = xfer->rx_buf;
485	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
486	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
487
488	if (!xfer->tx_buf || !p->tx_sgl) {
489		ret = -EINVAL;
490		goto error;
491	}
492	p->xfer = xfer;
493	p->mode = spi->mode;
494	p->clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz);
495	p->bytes_recvd = 0;
496	p->rx_buf = xfer->rx_buf;
497	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
498	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
499
500	tx_dma_addr = sg_dma_address(p->tx_sgl);
501	rx_dma_addr = sg_dma_address(p->rx_sgl);
502	p->tx_sgl_len = sg_dma_len(p->tx_sgl);
503	p->rx_sgl_len = sg_dma_len(p->rx_sgl);
504	pci1xxxx_spi_setup(par, p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
505	pci1xxxx_spi_setup_dma_to_io(p, (tx_dma_addr), p->tx_sgl_len);
506	if (rx_dma_addr)
507		pci1xxxx_spi_setup_dma_from_io(p, rx_dma_addr, p->rx_sgl_len);
508	writel(p->hw_inst, par->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
509
510	reinit_completion(&p->spi_xfer_done);
511	/* Wait for DMA_TERM interrupt */
512	ret = wait_for_completion_timeout(&p->spi_xfer_done, PCI1XXXX_SPI_TIMEOUT);
513	if (!ret) {
514		ret = -ETIMEDOUT;
515		if (p->dma_aborted_rd) {
516			writel(SPI_DMA_ENGINE_DIS,
517			       par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
518			/*
519			 * DMA ENGINE reset takes time if any TLP
520			 * completeion in progress, should wait
521			 * till DMA Engine reset is completed.
522			 */
523			ret = readl_poll_timeout(par->dma_offset_bar +
524						 SPI_DMA_GLOBAL_RD_ENGINE_EN, regval,
525						 (regval == 0x0), 0, USEC_PER_MSEC);
526			if (ret) {
527				ret = -ECANCELED;
528				goto error;
529			}
530			writel(SPI_DMA_ENGINE_EN,
531			       par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
532			p->dma_aborted_rd = false;
533			ret = -ECANCELED;
534		}
535		if (p->dma_aborted_wr) {
536			writel(SPI_DMA_ENGINE_DIS,
537			       par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
538
539			/*
540			 * DMA ENGINE reset takes time if any TLP
541			 * completeion in progress, should wait
542			 * till DMA Engine reset is completed.
543			 */
544			ret = readl_poll_timeout(par->dma_offset_bar +
545						 SPI_DMA_GLOBAL_WR_ENGINE_EN, regval,
546						 (regval == 0x0), 0, USEC_PER_MSEC);
547			if (ret) {
548				ret = -ECANCELED;
549				goto error;
550			}
551
552			writel(SPI_DMA_ENGINE_EN,
553			       par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
554			p->dma_aborted_wr = false;
555			ret = -ECANCELED;
556		}
557		goto error;
558	}
559	ret = 0;
560
561error:
562	p->spi_xfer_in_progress = false;
563
564	return ret;
565}
566
567static int pci1xxxx_spi_transfer_one(struct spi_controller *spi_ctlr,
568				     struct spi_device *spi, struct spi_transfer *xfer)
569{
570	if (spi_ctlr->can_dma(spi_ctlr, spi, xfer) && spi_ctlr->cur_msg_mapped)
571		return pci1xxxx_spi_transfer_with_dma(spi_ctlr, spi, xfer);
572	else
573		return pci1xxxx_spi_transfer_with_io(spi_ctlr, spi, xfer);
574}
575
576static irqreturn_t pci1xxxx_spi_isr_io(int irq, void *dev)
577{
578	struct pci1xxxx_spi_internal *p = dev;
579	irqreturn_t spi_int_fired = IRQ_NONE;
580	u32 regval;
581
582	/* Clear the SPI GO_BIT Interrupt */
583	regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
584	if (regval & SPI_INTR) {
585		/* Clear xfer_done */
586		if (p->parent->can_dma && p->rx_buf)
587			writel(p->hw_inst, p->parent->dma_offset_bar +
588			       SPI_DMA_WR_DOORBELL_REG);
589		else
590			complete(&p->parent->spi_int[p->hw_inst]->spi_xfer_done);
591		spi_int_fired = IRQ_HANDLED;
592	}
593	writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
594	return spi_int_fired;
595}
596
597static void pci1xxxx_spi_setup_next_dma_transfer(struct pci1xxxx_spi_internal *p)
598{
599	dma_addr_t tx_dma_addr = 0;
600	dma_addr_t rx_dma_addr = 0;
601	u32 prev_len;
602
603	p->tx_sgl = sg_next(p->tx_sgl);
604	if (p->rx_sgl)
605		p->rx_sgl = sg_next(p->rx_sgl);
606	if (!p->tx_sgl) {
607		/* Clear xfer_done */
608		complete(&p->spi_xfer_done);
609	} else {
610		tx_dma_addr = sg_dma_address(p->tx_sgl);
611		prev_len = p->tx_sgl_len;
612		p->tx_sgl_len = sg_dma_len(p->tx_sgl);
613		if (prev_len != p->tx_sgl_len)
614			pci1xxxx_spi_setup(p->parent,
615					   p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
616		pci1xxxx_spi_setup_dma_to_io(p, tx_dma_addr, p->tx_sgl_len);
617		if (p->rx_sgl) {
618			rx_dma_addr = sg_dma_address(p->rx_sgl);
619			p->rx_sgl_len = sg_dma_len(p->rx_sgl);
620			pci1xxxx_spi_setup_dma_from_io(p, rx_dma_addr, p->rx_sgl_len);
621		}
622		writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
623	}
624}
625
626static irqreturn_t pci1xxxx_spi_isr_dma(int irq, void *dev)
627{
628	struct pci1xxxx_spi_internal *p = dev;
629	irqreturn_t spi_int_fired = IRQ_NONE;
630	unsigned long flags;
631	u32 regval;
632
633	spin_lock_irqsave(&p->parent->dma_reg_lock, flags);
634	/* Clear the DMA RD INT and start spi xfer*/
635	regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_RD_STS);
636	if (regval & SPI_DMA_DONE_INT_MASK) {
637		if (regval & SPI_DMA_CH0_DONE_INT)
638			pci1xxxx_start_spi_xfer(p, SPI0);
639		if (regval & SPI_DMA_CH1_DONE_INT)
640			pci1xxxx_start_spi_xfer(p, SPI1);
641		spi_int_fired = IRQ_HANDLED;
642	}
643	if (regval & SPI_DMA_ABORT_INT_MASK) {
644		p->dma_aborted_rd = true;
645		spi_int_fired = IRQ_HANDLED;
646	}
647	writel(regval, p->parent->dma_offset_bar + SPI_DMA_INTR_RD_CLR);
648
649	/* Clear the DMA WR INT */
650	regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_WR_STS);
651	if (regval & SPI_DMA_DONE_INT_MASK) {
652		if (regval & SPI_DMA_CH0_DONE_INT)
653			pci1xxxx_spi_setup_next_dma_transfer(p->parent->spi_int[SPI0]);
654
655		if (regval & SPI_DMA_CH1_DONE_INT)
656			pci1xxxx_spi_setup_next_dma_transfer(p->parent->spi_int[SPI1]);
657
658		spi_int_fired = IRQ_HANDLED;
659	}
660	if (regval & SPI_DMA_ABORT_INT_MASK) {
661		p->dma_aborted_wr = true;
662		spi_int_fired = IRQ_HANDLED;
663	}
664	writel(regval, p->parent->dma_offset_bar + SPI_DMA_INTR_WR_CLR);
665	spin_unlock_irqrestore(&p->parent->dma_reg_lock, flags);
666
667	/* Clear the SPI GO_BIT Interrupt */
668	regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
669	if (regval & SPI_INTR) {
670		writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_WR_DOORBELL_REG);
671		spi_int_fired = IRQ_HANDLED;
672	}
673	writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
674	return spi_int_fired;
675}
676
677static irqreturn_t pci1xxxx_spi_isr(int irq, void *dev)
678{
679	struct pci1xxxx_spi_internal *p = dev;
680
681	if (p->spi_host->can_dma(p->spi_host, NULL, p->xfer))
682		return pci1xxxx_spi_isr_dma(irq, dev);
683	else
684		return pci1xxxx_spi_isr_io(irq, dev);
685}
686
687static bool pci1xxxx_spi_can_dma(struct spi_controller *host,
688				 struct spi_device *spi,
689				 struct spi_transfer *xfer)
690{
691	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(host);
692	struct pci1xxxx_spi *par = p->parent;
693
694	return par->can_dma;
695}
696
697static int pci1xxxx_spi_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
698{
699	u8 hw_inst_cnt, iter, start, only_sec_inst;
700	struct pci1xxxx_spi_internal *spi_sub_ptr;
701	struct device *dev = &pdev->dev;
702	struct pci1xxxx_spi *spi_bus;
703	struct spi_controller *spi_host;
704	u32 regval;
705	int ret;
706
707	hw_inst_cnt = ent->driver_data & 0x0f;
708	start = (ent->driver_data & 0xf0) >> 4;
709	if (start == 1)
710		only_sec_inst = 1;
711	else
712		only_sec_inst = 0;
713
714	spi_bus = devm_kzalloc(&pdev->dev,
715			       struct_size(spi_bus, spi_int, hw_inst_cnt),
716			       GFP_KERNEL);
717	if (!spi_bus)
718		return -ENOMEM;
719
720	spi_bus->dev = pdev;
721	spi_bus->total_hw_instances = hw_inst_cnt;
722	pci_set_master(pdev);
723
724	for (iter = 0; iter < hw_inst_cnt; iter++) {
725		spi_bus->spi_int[iter] = devm_kzalloc(&pdev->dev,
726						      sizeof(struct pci1xxxx_spi_internal),
727						      GFP_KERNEL);
728		if (!spi_bus->spi_int[iter])
729			return -ENOMEM;
730		spi_sub_ptr = spi_bus->spi_int[iter];
731		spi_sub_ptr->spi_host = devm_spi_alloc_host(dev, sizeof(struct spi_controller));
732		if (!spi_sub_ptr->spi_host)
733			return -ENOMEM;
734
735		spi_sub_ptr->parent = spi_bus;
736		spi_sub_ptr->spi_xfer_in_progress = false;
737
738		if (!iter) {
739			ret = pcim_enable_device(pdev);
740			if (ret)
741				return -ENOMEM;
742
743			ret = pci_request_regions(pdev, DRV_NAME);
744			if (ret)
745				return -ENOMEM;
746
747			spi_bus->reg_base = pcim_iomap(pdev, 0, pci_resource_len(pdev, 0));
748			if (!spi_bus->reg_base) {
749				ret = -EINVAL;
750				goto error;
751			}
752
753			ret = pci_alloc_irq_vectors(pdev, hw_inst_cnt, hw_inst_cnt,
754						    PCI_IRQ_ALL_TYPES);
755			if (ret < 0) {
756				dev_err(&pdev->dev, "Error allocating MSI vectors\n");
757				goto error;
758			}
759
760			init_completion(&spi_sub_ptr->spi_xfer_done);
761			/* Initialize Interrupts - SPI_INT */
762			regval = readl(spi_bus->reg_base +
763				       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
764			regval &= ~SPI_INTR;
765			writel(regval, spi_bus->reg_base +
766			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
767			spi_sub_ptr->irq = pci_irq_vector(pdev, 0);
768
769			ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
770					       pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
771					       pci_name(pdev), spi_sub_ptr);
772			if (ret < 0) {
773				dev_err(&pdev->dev, "Unable to request irq : %d",
774					spi_sub_ptr->irq);
775				ret = -ENODEV;
776				goto error;
777			}
778
779			ret = pci1xxxx_spi_dma_init(spi_bus, spi_sub_ptr->irq);
780			if (ret && ret != -EOPNOTSUPP)
781				goto error;
782
783			/* This register is only applicable for 1st instance */
784			regval = readl(spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
785			if (!only_sec_inst)
786				regval |= (BIT(4));
787			else
788				regval &= ~(BIT(4));
789
790			writel(regval, spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
791		}
792
793		spi_sub_ptr->hw_inst = start++;
794
795		if (iter == 1) {
796			init_completion(&spi_sub_ptr->spi_xfer_done);
797			/* Initialize Interrupts - SPI_INT */
798			regval = readl(spi_bus->reg_base +
799			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
800			regval &= ~SPI_INTR;
801			writel(regval, spi_bus->reg_base +
802			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
803			spi_sub_ptr->irq = pci_irq_vector(pdev, iter);
804			ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
805					       pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
806					       pci_name(pdev), spi_sub_ptr);
807			if (ret < 0) {
808				dev_err(&pdev->dev, "Unable to request irq : %d",
809					spi_sub_ptr->irq);
810				ret = -ENODEV;
811				goto error;
812			}
813		}
814
815		spi_host = spi_sub_ptr->spi_host;
816		spi_host->num_chipselect = SPI_CHIP_SEL_COUNT;
817		spi_host->mode_bits = SPI_MODE_0 | SPI_MODE_3 | SPI_RX_DUAL |
818				      SPI_TX_DUAL | SPI_LOOP;
819		spi_host->can_dma = pci1xxxx_spi_can_dma;
820		spi_host->transfer_one = pci1xxxx_spi_transfer_one;
821
822		spi_host->set_cs = pci1xxxx_spi_set_cs;
823		spi_host->bits_per_word_mask = SPI_BPW_MASK(8);
824		spi_host->max_speed_hz = PCI1XXXX_SPI_MAX_CLOCK_HZ;
825		spi_host->min_speed_hz = PCI1XXXX_SPI_MIN_CLOCK_HZ;
826		spi_host->flags = SPI_CONTROLLER_MUST_TX;
827		spi_controller_set_devdata(spi_host, spi_sub_ptr);
828		ret = devm_spi_register_controller(dev, spi_host);
829		if (ret)
830			goto error;
831	}
832	pci_set_drvdata(pdev, spi_bus);
833
834	return 0;
835
836error:
837	pci_release_regions(pdev);
838	return ret;
839}
840
841static void store_restore_config(struct pci1xxxx_spi *spi_ptr,
842				 struct pci1xxxx_spi_internal *spi_sub_ptr,
843				 u8 inst, bool store)
844{
845	u32 regval;
846
847	if (store) {
848		regval = readl(spi_ptr->reg_base +
849			       SPI_MST_CTL_REG_OFFSET(spi_sub_ptr->hw_inst));
850		regval &= SPI_MST_CTL_DEVSEL_MASK;
851		spi_sub_ptr->prev_val.dev_sel = (regval >> 25) & 7;
852		regval = readl(spi_ptr->reg_base +
853			       SPI_PCI_CTRL_REG_OFFSET(spi_sub_ptr->hw_inst));
854		regval &= SPI_MSI_VECTOR_SEL_MASK;
855		spi_sub_ptr->prev_val.msi_vector_sel = (regval >> 4) & 1;
856	} else {
857		regval = readl(spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
858		regval &= ~SPI_MST_CTL_DEVSEL_MASK;
859		regval |= (spi_sub_ptr->prev_val.dev_sel << 25);
860		writel(regval,
861		       spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
862		writel((spi_sub_ptr->prev_val.msi_vector_sel << 4),
863			spi_ptr->reg_base + SPI_PCI_CTRL_REG_OFFSET(inst));
864	}
865}
866
867static int pci1xxxx_spi_resume(struct device *dev)
868{
869	struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
870	struct pci1xxxx_spi_internal *spi_sub_ptr;
871	u32 regval = SPI_RESUME_CONFIG;
872	u8 iter;
873
874	for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
875		spi_sub_ptr = spi_ptr->spi_int[iter];
876		spi_controller_resume(spi_sub_ptr->spi_host);
877		writel(regval, spi_ptr->reg_base +
878		       SPI_MST_EVENT_MASK_REG_OFFSET(iter));
879
880		/* Restore config at resume */
881		store_restore_config(spi_ptr, spi_sub_ptr, iter, 0);
882	}
883
884	return 0;
885}
886
887static int pci1xxxx_spi_suspend(struct device *dev)
888{
889	struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
890	struct pci1xxxx_spi_internal *spi_sub_ptr;
891	u32 reg1 = SPI_SUSPEND_CONFIG;
892	u8 iter;
893
894	for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
895		spi_sub_ptr = spi_ptr->spi_int[iter];
896
897		while (spi_sub_ptr->spi_xfer_in_progress)
898			msleep(20);
899
900		/* Store existing config before suspend */
901		store_restore_config(spi_ptr, spi_sub_ptr, iter, 1);
902		spi_controller_suspend(spi_sub_ptr->spi_host);
903		writel(reg1, spi_ptr->reg_base +
904		       SPI_MST_EVENT_MASK_REG_OFFSET(iter));
905	}
906
907	return 0;
908}
909
910static DEFINE_SIMPLE_DEV_PM_OPS(spi_pm_ops, pci1xxxx_spi_suspend,
911				pci1xxxx_spi_resume);
912
913static struct pci_driver pci1xxxx_spi_driver = {
914	.name		= DRV_NAME,
915	.id_table	= pci1xxxx_spi_pci_id_table,
916	.probe		= pci1xxxx_spi_probe,
917	.driver		=	{
918		.pm = pm_sleep_ptr(&spi_pm_ops),
919	},
920};
921
922module_pci_driver(pci1xxxx_spi_driver);
923
924MODULE_DESCRIPTION("Microchip Technology Inc. pci1xxxx SPI bus driver");
925MODULE_AUTHOR("Tharun Kumar P<tharunkumar.pasumarthi@microchip.com>");
926MODULE_AUTHOR("Kumaravel Thiagarajan<kumaravel.thiagarajan@microchip.com>");
927MODULE_LICENSE("GPL v2");
928