1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Cryptographic API.
4 *
5 * Support for ATMEL DES/TDES HW acceleration.
6 *
7 * Copyright (c) 2012 Eukr��a Electromatique - ATMEL
8 * Author: Nicolas Royer <nicolas@eukrea.com>
9 *
10 * Some ideas are from omap-aes.c drivers.
11 */
12
13
14#include <linux/kernel.h>
15#include <linux/module.h>
16#include <linux/slab.h>
17#include <linux/err.h>
18#include <linux/clk.h>
19#include <linux/io.h>
20#include <linux/hw_random.h>
21#include <linux/platform_device.h>
22
23#include <linux/device.h>
24#include <linux/dmaengine.h>
25#include <linux/init.h>
26#include <linux/errno.h>
27#include <linux/interrupt.h>
28#include <linux/irq.h>
29#include <linux/scatterlist.h>
30#include <linux/dma-mapping.h>
31#include <linux/mod_devicetable.h>
32#include <linux/delay.h>
33#include <linux/crypto.h>
34#include <crypto/scatterwalk.h>
35#include <crypto/algapi.h>
36#include <crypto/internal/des.h>
37#include <crypto/internal/skcipher.h>
38#include "atmel-tdes-regs.h"
39
40#define ATMEL_TDES_PRIORITY	300
41
42/* TDES flags  */
43/* Reserve bits [17:16], [13:12], [2:0] for AES Mode Register */
44#define TDES_FLAGS_ENCRYPT	TDES_MR_CYPHER_ENC
45#define TDES_FLAGS_OPMODE_MASK	(TDES_MR_OPMOD_MASK | TDES_MR_CFBS_MASK)
46#define TDES_FLAGS_ECB		TDES_MR_OPMOD_ECB
47#define TDES_FLAGS_CBC		TDES_MR_OPMOD_CBC
48
49#define TDES_FLAGS_MODE_MASK	(TDES_FLAGS_OPMODE_MASK | TDES_FLAGS_ENCRYPT)
50
51#define TDES_FLAGS_INIT		BIT(3)
52#define TDES_FLAGS_FAST		BIT(4)
53#define TDES_FLAGS_BUSY		BIT(5)
54#define TDES_FLAGS_DMA		BIT(6)
55
56#define ATMEL_TDES_QUEUE_LENGTH	50
57
58struct atmel_tdes_caps {
59	bool	has_dma;
60};
61
62struct atmel_tdes_dev;
63
64struct atmel_tdes_ctx {
65	struct atmel_tdes_dev *dd;
66
67	int		keylen;
68	u32		key[DES3_EDE_KEY_SIZE / sizeof(u32)];
69	unsigned long	flags;
70
71	u16		block_size;
72};
73
74struct atmel_tdes_reqctx {
75	unsigned long mode;
76	u8 lastc[DES_BLOCK_SIZE];
77};
78
79struct atmel_tdes_dma {
80	struct dma_chan			*chan;
81	struct dma_slave_config dma_conf;
82};
83
84struct atmel_tdes_dev {
85	struct list_head	list;
86	unsigned long		phys_base;
87	void __iomem		*io_base;
88
89	struct atmel_tdes_ctx	*ctx;
90	struct device		*dev;
91	struct clk			*iclk;
92	int					irq;
93
94	unsigned long		flags;
95
96	spinlock_t		lock;
97	struct crypto_queue	queue;
98
99	struct tasklet_struct	done_task;
100	struct tasklet_struct	queue_task;
101
102	struct skcipher_request	*req;
103	size_t				total;
104
105	struct scatterlist	*in_sg;
106	unsigned int		nb_in_sg;
107	size_t				in_offset;
108	struct scatterlist	*out_sg;
109	unsigned int		nb_out_sg;
110	size_t				out_offset;
111
112	size_t	buflen;
113	size_t	dma_size;
114
115	void	*buf_in;
116	int		dma_in;
117	dma_addr_t	dma_addr_in;
118	struct atmel_tdes_dma	dma_lch_in;
119
120	void	*buf_out;
121	int		dma_out;
122	dma_addr_t	dma_addr_out;
123	struct atmel_tdes_dma	dma_lch_out;
124
125	struct atmel_tdes_caps	caps;
126
127	u32	hw_version;
128};
129
130struct atmel_tdes_drv {
131	struct list_head	dev_list;
132	spinlock_t		lock;
133};
134
135static struct atmel_tdes_drv atmel_tdes = {
136	.dev_list = LIST_HEAD_INIT(atmel_tdes.dev_list),
137	.lock = __SPIN_LOCK_UNLOCKED(atmel_tdes.lock),
138};
139
140static int atmel_tdes_sg_copy(struct scatterlist **sg, size_t *offset,
141			void *buf, size_t buflen, size_t total, int out)
142{
143	size_t count, off = 0;
144
145	while (buflen && total) {
146		count = min((*sg)->length - *offset, total);
147		count = min(count, buflen);
148
149		if (!count)
150			return off;
151
152		scatterwalk_map_and_copy(buf + off, *sg, *offset, count, out);
153
154		off += count;
155		buflen -= count;
156		*offset += count;
157		total -= count;
158
159		if (*offset == (*sg)->length) {
160			*sg = sg_next(*sg);
161			if (*sg)
162				*offset = 0;
163			else
164				total = 0;
165		}
166	}
167
168	return off;
169}
170
171static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset)
172{
173	return readl_relaxed(dd->io_base + offset);
174}
175
176static inline void atmel_tdes_write(struct atmel_tdes_dev *dd,
177					u32 offset, u32 value)
178{
179	writel_relaxed(value, dd->io_base + offset);
180}
181
182static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset,
183			       const u32 *value, int count)
184{
185	for (; count--; value++, offset += 4)
186		atmel_tdes_write(dd, offset, *value);
187}
188
189static struct atmel_tdes_dev *atmel_tdes_dev_alloc(void)
190{
191	struct atmel_tdes_dev *tdes_dd;
192
193	spin_lock_bh(&atmel_tdes.lock);
194	/* One TDES IP per SoC. */
195	tdes_dd = list_first_entry_or_null(&atmel_tdes.dev_list,
196					   struct atmel_tdes_dev, list);
197	spin_unlock_bh(&atmel_tdes.lock);
198	return tdes_dd;
199}
200
201static int atmel_tdes_hw_init(struct atmel_tdes_dev *dd)
202{
203	int err;
204
205	err = clk_prepare_enable(dd->iclk);
206	if (err)
207		return err;
208
209	if (!(dd->flags & TDES_FLAGS_INIT)) {
210		atmel_tdes_write(dd, TDES_CR, TDES_CR_SWRST);
211		dd->flags |= TDES_FLAGS_INIT;
212	}
213
214	return 0;
215}
216
217static inline unsigned int atmel_tdes_get_version(struct atmel_tdes_dev *dd)
218{
219	return atmel_tdes_read(dd, TDES_HW_VERSION) & 0x00000fff;
220}
221
222static int atmel_tdes_hw_version_init(struct atmel_tdes_dev *dd)
223{
224	int err;
225
226	err = atmel_tdes_hw_init(dd);
227	if (err)
228		return err;
229
230	dd->hw_version = atmel_tdes_get_version(dd);
231
232	dev_info(dd->dev,
233			"version: 0x%x\n", dd->hw_version);
234
235	clk_disable_unprepare(dd->iclk);
236
237	return 0;
238}
239
240static void atmel_tdes_dma_callback(void *data)
241{
242	struct atmel_tdes_dev *dd = data;
243
244	/* dma_lch_out - completed */
245	tasklet_schedule(&dd->done_task);
246}
247
248static int atmel_tdes_write_ctrl(struct atmel_tdes_dev *dd)
249{
250	int err;
251	u32 valmr = TDES_MR_SMOD_PDC;
252
253	err = atmel_tdes_hw_init(dd);
254
255	if (err)
256		return err;
257
258	if (!dd->caps.has_dma)
259		atmel_tdes_write(dd, TDES_PTCR,
260			TDES_PTCR_TXTDIS | TDES_PTCR_RXTDIS);
261
262	/* MR register must be set before IV registers */
263	if (dd->ctx->keylen > (DES_KEY_SIZE << 1)) {
264		valmr |= TDES_MR_KEYMOD_3KEY;
265		valmr |= TDES_MR_TDESMOD_TDES;
266	} else if (dd->ctx->keylen > DES_KEY_SIZE) {
267		valmr |= TDES_MR_KEYMOD_2KEY;
268		valmr |= TDES_MR_TDESMOD_TDES;
269	} else {
270		valmr |= TDES_MR_TDESMOD_DES;
271	}
272
273	valmr |= dd->flags & TDES_FLAGS_MODE_MASK;
274
275	atmel_tdes_write(dd, TDES_MR, valmr);
276
277	atmel_tdes_write_n(dd, TDES_KEY1W1R, dd->ctx->key,
278						dd->ctx->keylen >> 2);
279
280	if (dd->req->iv && (valmr & TDES_MR_OPMOD_MASK) != TDES_MR_OPMOD_ECB)
281		atmel_tdes_write_n(dd, TDES_IV1R, (void *)dd->req->iv, 2);
282
283	return 0;
284}
285
286static int atmel_tdes_crypt_pdc_stop(struct atmel_tdes_dev *dd)
287{
288	int err = 0;
289	size_t count;
290
291	atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
292
293	if (dd->flags & TDES_FLAGS_FAST) {
294		dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
295		dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
296	} else {
297		dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
298					   dd->dma_size, DMA_FROM_DEVICE);
299
300		/* copy data */
301		count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
302				dd->buf_out, dd->buflen, dd->dma_size, 1);
303		if (count != dd->dma_size) {
304			err = -EINVAL;
305			dev_dbg(dd->dev, "not all data converted: %zu\n", count);
306		}
307	}
308
309	return err;
310}
311
312static int atmel_tdes_buff_init(struct atmel_tdes_dev *dd)
313{
314	int err = -ENOMEM;
315
316	dd->buf_in = (void *)__get_free_pages(GFP_KERNEL, 0);
317	dd->buf_out = (void *)__get_free_pages(GFP_KERNEL, 0);
318	dd->buflen = PAGE_SIZE;
319	dd->buflen &= ~(DES_BLOCK_SIZE - 1);
320
321	if (!dd->buf_in || !dd->buf_out) {
322		dev_dbg(dd->dev, "unable to alloc pages.\n");
323		goto err_alloc;
324	}
325
326	/* MAP here */
327	dd->dma_addr_in = dma_map_single(dd->dev, dd->buf_in,
328					dd->buflen, DMA_TO_DEVICE);
329	err = dma_mapping_error(dd->dev, dd->dma_addr_in);
330	if (err) {
331		dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen);
332		goto err_map_in;
333	}
334
335	dd->dma_addr_out = dma_map_single(dd->dev, dd->buf_out,
336					dd->buflen, DMA_FROM_DEVICE);
337	err = dma_mapping_error(dd->dev, dd->dma_addr_out);
338	if (err) {
339		dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen);
340		goto err_map_out;
341	}
342
343	return 0;
344
345err_map_out:
346	dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
347		DMA_TO_DEVICE);
348err_map_in:
349err_alloc:
350	free_page((unsigned long)dd->buf_out);
351	free_page((unsigned long)dd->buf_in);
352	return err;
353}
354
355static void atmel_tdes_buff_cleanup(struct atmel_tdes_dev *dd)
356{
357	dma_unmap_single(dd->dev, dd->dma_addr_out, dd->buflen,
358			 DMA_FROM_DEVICE);
359	dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
360		DMA_TO_DEVICE);
361	free_page((unsigned long)dd->buf_out);
362	free_page((unsigned long)dd->buf_in);
363}
364
365static int atmel_tdes_crypt_pdc(struct atmel_tdes_dev *dd,
366				dma_addr_t dma_addr_in,
367				dma_addr_t dma_addr_out, int length)
368{
369	int len32;
370
371	dd->dma_size = length;
372
373	if (!(dd->flags & TDES_FLAGS_FAST)) {
374		dma_sync_single_for_device(dd->dev, dma_addr_in, length,
375					   DMA_TO_DEVICE);
376	}
377
378	len32 = DIV_ROUND_UP(length, sizeof(u32));
379
380	atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
381	atmel_tdes_write(dd, TDES_TPR, dma_addr_in);
382	atmel_tdes_write(dd, TDES_TCR, len32);
383	atmel_tdes_write(dd, TDES_RPR, dma_addr_out);
384	atmel_tdes_write(dd, TDES_RCR, len32);
385
386	/* Enable Interrupt */
387	atmel_tdes_write(dd, TDES_IER, TDES_INT_ENDRX);
388
389	/* Start DMA transfer */
390	atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTEN | TDES_PTCR_RXTEN);
391
392	return 0;
393}
394
395static int atmel_tdes_crypt_dma(struct atmel_tdes_dev *dd,
396				dma_addr_t dma_addr_in,
397				dma_addr_t dma_addr_out, int length)
398{
399	struct scatterlist sg[2];
400	struct dma_async_tx_descriptor	*in_desc, *out_desc;
401	enum dma_slave_buswidth addr_width;
402
403	dd->dma_size = length;
404
405	if (!(dd->flags & TDES_FLAGS_FAST)) {
406		dma_sync_single_for_device(dd->dev, dma_addr_in, length,
407					   DMA_TO_DEVICE);
408	}
409
410	addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
411
412	dd->dma_lch_in.dma_conf.dst_addr_width = addr_width;
413	dd->dma_lch_out.dma_conf.src_addr_width = addr_width;
414
415	dmaengine_slave_config(dd->dma_lch_in.chan, &dd->dma_lch_in.dma_conf);
416	dmaengine_slave_config(dd->dma_lch_out.chan, &dd->dma_lch_out.dma_conf);
417
418	dd->flags |= TDES_FLAGS_DMA;
419
420	sg_init_table(&sg[0], 1);
421	sg_dma_address(&sg[0]) = dma_addr_in;
422	sg_dma_len(&sg[0]) = length;
423
424	sg_init_table(&sg[1], 1);
425	sg_dma_address(&sg[1]) = dma_addr_out;
426	sg_dma_len(&sg[1]) = length;
427
428	in_desc = dmaengine_prep_slave_sg(dd->dma_lch_in.chan, &sg[0],
429				1, DMA_MEM_TO_DEV,
430				DMA_PREP_INTERRUPT  |  DMA_CTRL_ACK);
431	if (!in_desc)
432		return -EINVAL;
433
434	out_desc = dmaengine_prep_slave_sg(dd->dma_lch_out.chan, &sg[1],
435				1, DMA_DEV_TO_MEM,
436				DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
437	if (!out_desc)
438		return -EINVAL;
439
440	out_desc->callback = atmel_tdes_dma_callback;
441	out_desc->callback_param = dd;
442
443	dmaengine_submit(out_desc);
444	dma_async_issue_pending(dd->dma_lch_out.chan);
445
446	dmaengine_submit(in_desc);
447	dma_async_issue_pending(dd->dma_lch_in.chan);
448
449	return 0;
450}
451
452static int atmel_tdes_crypt_start(struct atmel_tdes_dev *dd)
453{
454	int err, fast = 0, in, out;
455	size_t count;
456	dma_addr_t addr_in, addr_out;
457
458	if ((!dd->in_offset) && (!dd->out_offset)) {
459		/* check for alignment */
460		in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) &&
461			IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size);
462		out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) &&
463			IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size);
464		fast = in && out;
465
466		if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg))
467			fast = 0;
468	}
469
470
471	if (fast)  {
472		count = min_t(size_t, dd->total, sg_dma_len(dd->in_sg));
473		count = min_t(size_t, count, sg_dma_len(dd->out_sg));
474
475		err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
476		if (!err) {
477			dev_dbg(dd->dev, "dma_map_sg() error\n");
478			return -EINVAL;
479		}
480
481		err = dma_map_sg(dd->dev, dd->out_sg, 1,
482				DMA_FROM_DEVICE);
483		if (!err) {
484			dev_dbg(dd->dev, "dma_map_sg() error\n");
485			dma_unmap_sg(dd->dev, dd->in_sg, 1,
486				DMA_TO_DEVICE);
487			return -EINVAL;
488		}
489
490		addr_in = sg_dma_address(dd->in_sg);
491		addr_out = sg_dma_address(dd->out_sg);
492
493		dd->flags |= TDES_FLAGS_FAST;
494
495	} else {
496		/* use cache buffers */
497		count = atmel_tdes_sg_copy(&dd->in_sg, &dd->in_offset,
498				dd->buf_in, dd->buflen, dd->total, 0);
499
500		addr_in = dd->dma_addr_in;
501		addr_out = dd->dma_addr_out;
502
503		dd->flags &= ~TDES_FLAGS_FAST;
504	}
505
506	dd->total -= count;
507
508	if (dd->caps.has_dma)
509		err = atmel_tdes_crypt_dma(dd, addr_in, addr_out, count);
510	else
511		err = atmel_tdes_crypt_pdc(dd, addr_in, addr_out, count);
512
513	if (err && (dd->flags & TDES_FLAGS_FAST)) {
514		dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
515		dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE);
516	}
517
518	return err;
519}
520
521static void
522atmel_tdes_set_iv_as_last_ciphertext_block(struct atmel_tdes_dev *dd)
523{
524	struct skcipher_request *req = dd->req;
525	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
526	struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
527	unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
528
529	if (req->cryptlen < ivsize)
530		return;
531
532	if (rctx->mode & TDES_FLAGS_ENCRYPT)
533		scatterwalk_map_and_copy(req->iv, req->dst,
534					 req->cryptlen - ivsize, ivsize, 0);
535	else
536		memcpy(req->iv, rctx->lastc, ivsize);
537
538}
539
540static void atmel_tdes_finish_req(struct atmel_tdes_dev *dd, int err)
541{
542	struct skcipher_request *req = dd->req;
543	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
544
545	clk_disable_unprepare(dd->iclk);
546
547	dd->flags &= ~TDES_FLAGS_BUSY;
548
549	if (!err && (rctx->mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB)
550		atmel_tdes_set_iv_as_last_ciphertext_block(dd);
551
552	skcipher_request_complete(req, err);
553}
554
555static int atmel_tdes_handle_queue(struct atmel_tdes_dev *dd,
556			       struct skcipher_request *req)
557{
558	struct crypto_async_request *async_req, *backlog;
559	struct atmel_tdes_ctx *ctx;
560	struct atmel_tdes_reqctx *rctx;
561	unsigned long flags;
562	int err, ret = 0;
563
564	spin_lock_irqsave(&dd->lock, flags);
565	if (req)
566		ret = crypto_enqueue_request(&dd->queue, &req->base);
567	if (dd->flags & TDES_FLAGS_BUSY) {
568		spin_unlock_irqrestore(&dd->lock, flags);
569		return ret;
570	}
571	backlog = crypto_get_backlog(&dd->queue);
572	async_req = crypto_dequeue_request(&dd->queue);
573	if (async_req)
574		dd->flags |= TDES_FLAGS_BUSY;
575	spin_unlock_irqrestore(&dd->lock, flags);
576
577	if (!async_req)
578		return ret;
579
580	if (backlog)
581		crypto_request_complete(backlog, -EINPROGRESS);
582
583	req = skcipher_request_cast(async_req);
584
585	/* assign new request to device */
586	dd->req = req;
587	dd->total = req->cryptlen;
588	dd->in_offset = 0;
589	dd->in_sg = req->src;
590	dd->out_offset = 0;
591	dd->out_sg = req->dst;
592
593	rctx = skcipher_request_ctx(req);
594	ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req));
595	rctx->mode &= TDES_FLAGS_MODE_MASK;
596	dd->flags = (dd->flags & ~TDES_FLAGS_MODE_MASK) | rctx->mode;
597	dd->ctx = ctx;
598
599	err = atmel_tdes_write_ctrl(dd);
600	if (!err)
601		err = atmel_tdes_crypt_start(dd);
602	if (err) {
603		/* des_task will not finish it, so do it here */
604		atmel_tdes_finish_req(dd, err);
605		tasklet_schedule(&dd->queue_task);
606	}
607
608	return ret;
609}
610
611static int atmel_tdes_crypt_dma_stop(struct atmel_tdes_dev *dd)
612{
613	int err = -EINVAL;
614	size_t count;
615
616	if (dd->flags & TDES_FLAGS_DMA) {
617		err = 0;
618		if  (dd->flags & TDES_FLAGS_FAST) {
619			dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
620			dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
621		} else {
622			dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
623				dd->dma_size, DMA_FROM_DEVICE);
624
625			/* copy data */
626			count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
627				dd->buf_out, dd->buflen, dd->dma_size, 1);
628			if (count != dd->dma_size) {
629				err = -EINVAL;
630				dev_dbg(dd->dev, "not all data converted: %zu\n", count);
631			}
632		}
633	}
634	return err;
635}
636
637static int atmel_tdes_crypt(struct skcipher_request *req, unsigned long mode)
638{
639	struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
640	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(skcipher);
641	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
642	struct device *dev = ctx->dd->dev;
643
644	if (!req->cryptlen)
645		return 0;
646
647	if (!IS_ALIGNED(req->cryptlen, DES_BLOCK_SIZE)) {
648		dev_dbg(dev, "request size is not exact amount of DES blocks\n");
649		return -EINVAL;
650	}
651	ctx->block_size = DES_BLOCK_SIZE;
652
653	rctx->mode = mode;
654
655	if ((mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB &&
656	    !(mode & TDES_FLAGS_ENCRYPT)) {
657		unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
658
659		if (req->cryptlen >= ivsize)
660			scatterwalk_map_and_copy(rctx->lastc, req->src,
661						 req->cryptlen - ivsize,
662						 ivsize, 0);
663	}
664
665	return atmel_tdes_handle_queue(ctx->dd, req);
666}
667
668static int atmel_tdes_dma_init(struct atmel_tdes_dev *dd)
669{
670	int ret;
671
672	/* Try to grab 2 DMA channels */
673	dd->dma_lch_in.chan = dma_request_chan(dd->dev, "tx");
674	if (IS_ERR(dd->dma_lch_in.chan)) {
675		ret = PTR_ERR(dd->dma_lch_in.chan);
676		goto err_dma_in;
677	}
678
679	dd->dma_lch_in.dma_conf.dst_addr = dd->phys_base +
680		TDES_IDATA1R;
681	dd->dma_lch_in.dma_conf.src_maxburst = 1;
682	dd->dma_lch_in.dma_conf.src_addr_width =
683		DMA_SLAVE_BUSWIDTH_4_BYTES;
684	dd->dma_lch_in.dma_conf.dst_maxburst = 1;
685	dd->dma_lch_in.dma_conf.dst_addr_width =
686		DMA_SLAVE_BUSWIDTH_4_BYTES;
687	dd->dma_lch_in.dma_conf.device_fc = false;
688
689	dd->dma_lch_out.chan = dma_request_chan(dd->dev, "rx");
690	if (IS_ERR(dd->dma_lch_out.chan)) {
691		ret = PTR_ERR(dd->dma_lch_out.chan);
692		goto err_dma_out;
693	}
694
695	dd->dma_lch_out.dma_conf.src_addr = dd->phys_base +
696		TDES_ODATA1R;
697	dd->dma_lch_out.dma_conf.src_maxburst = 1;
698	dd->dma_lch_out.dma_conf.src_addr_width =
699		DMA_SLAVE_BUSWIDTH_4_BYTES;
700	dd->dma_lch_out.dma_conf.dst_maxburst = 1;
701	dd->dma_lch_out.dma_conf.dst_addr_width =
702		DMA_SLAVE_BUSWIDTH_4_BYTES;
703	dd->dma_lch_out.dma_conf.device_fc = false;
704
705	return 0;
706
707err_dma_out:
708	dma_release_channel(dd->dma_lch_in.chan);
709err_dma_in:
710	dev_err(dd->dev, "no DMA channel available\n");
711	return ret;
712}
713
714static void atmel_tdes_dma_cleanup(struct atmel_tdes_dev *dd)
715{
716	dma_release_channel(dd->dma_lch_in.chan);
717	dma_release_channel(dd->dma_lch_out.chan);
718}
719
720static int atmel_des_setkey(struct crypto_skcipher *tfm, const u8 *key,
721			   unsigned int keylen)
722{
723	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
724	int err;
725
726	err = verify_skcipher_des_key(tfm, key);
727	if (err)
728		return err;
729
730	memcpy(ctx->key, key, keylen);
731	ctx->keylen = keylen;
732
733	return 0;
734}
735
736static int atmel_tdes_setkey(struct crypto_skcipher *tfm, const u8 *key,
737			   unsigned int keylen)
738{
739	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
740	int err;
741
742	err = verify_skcipher_des3_key(tfm, key);
743	if (err)
744		return err;
745
746	memcpy(ctx->key, key, keylen);
747	ctx->keylen = keylen;
748
749	return 0;
750}
751
752static int atmel_tdes_ecb_encrypt(struct skcipher_request *req)
753{
754	return atmel_tdes_crypt(req, TDES_FLAGS_ECB | TDES_FLAGS_ENCRYPT);
755}
756
757static int atmel_tdes_ecb_decrypt(struct skcipher_request *req)
758{
759	return atmel_tdes_crypt(req, TDES_FLAGS_ECB);
760}
761
762static int atmel_tdes_cbc_encrypt(struct skcipher_request *req)
763{
764	return atmel_tdes_crypt(req, TDES_FLAGS_CBC | TDES_FLAGS_ENCRYPT);
765}
766
767static int atmel_tdes_cbc_decrypt(struct skcipher_request *req)
768{
769	return atmel_tdes_crypt(req, TDES_FLAGS_CBC);
770}
771
772static int atmel_tdes_init_tfm(struct crypto_skcipher *tfm)
773{
774	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
775
776	ctx->dd = atmel_tdes_dev_alloc();
777	if (!ctx->dd)
778		return -ENODEV;
779
780	crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_tdes_reqctx));
781
782	return 0;
783}
784
785static void atmel_tdes_skcipher_alg_init(struct skcipher_alg *alg)
786{
787	alg->base.cra_priority = ATMEL_TDES_PRIORITY;
788	alg->base.cra_flags = CRYPTO_ALG_ASYNC;
789	alg->base.cra_ctxsize = sizeof(struct atmel_tdes_ctx);
790	alg->base.cra_module = THIS_MODULE;
791
792	alg->init = atmel_tdes_init_tfm;
793}
794
795static struct skcipher_alg tdes_algs[] = {
796{
797	.base.cra_name		= "ecb(des)",
798	.base.cra_driver_name	= "atmel-ecb-des",
799	.base.cra_blocksize	= DES_BLOCK_SIZE,
800	.base.cra_alignmask	= 0x7,
801
802	.min_keysize		= DES_KEY_SIZE,
803	.max_keysize		= DES_KEY_SIZE,
804	.setkey			= atmel_des_setkey,
805	.encrypt		= atmel_tdes_ecb_encrypt,
806	.decrypt		= atmel_tdes_ecb_decrypt,
807},
808{
809	.base.cra_name		= "cbc(des)",
810	.base.cra_driver_name	= "atmel-cbc-des",
811	.base.cra_blocksize	= DES_BLOCK_SIZE,
812	.base.cra_alignmask	= 0x7,
813
814	.min_keysize		= DES_KEY_SIZE,
815	.max_keysize		= DES_KEY_SIZE,
816	.ivsize			= DES_BLOCK_SIZE,
817	.setkey			= atmel_des_setkey,
818	.encrypt		= atmel_tdes_cbc_encrypt,
819	.decrypt		= atmel_tdes_cbc_decrypt,
820},
821{
822	.base.cra_name		= "ecb(des3_ede)",
823	.base.cra_driver_name	= "atmel-ecb-tdes",
824	.base.cra_blocksize	= DES_BLOCK_SIZE,
825	.base.cra_alignmask	= 0x7,
826
827	.min_keysize		= DES3_EDE_KEY_SIZE,
828	.max_keysize		= DES3_EDE_KEY_SIZE,
829	.setkey			= atmel_tdes_setkey,
830	.encrypt		= atmel_tdes_ecb_encrypt,
831	.decrypt		= atmel_tdes_ecb_decrypt,
832},
833{
834	.base.cra_name		= "cbc(des3_ede)",
835	.base.cra_driver_name	= "atmel-cbc-tdes",
836	.base.cra_blocksize	= DES_BLOCK_SIZE,
837	.base.cra_alignmask	= 0x7,
838
839	.min_keysize		= DES3_EDE_KEY_SIZE,
840	.max_keysize		= DES3_EDE_KEY_SIZE,
841	.setkey			= atmel_tdes_setkey,
842	.encrypt		= atmel_tdes_cbc_encrypt,
843	.decrypt		= atmel_tdes_cbc_decrypt,
844	.ivsize			= DES_BLOCK_SIZE,
845},
846};
847
848static void atmel_tdes_queue_task(unsigned long data)
849{
850	struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *)data;
851
852	atmel_tdes_handle_queue(dd, NULL);
853}
854
855static void atmel_tdes_done_task(unsigned long data)
856{
857	struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *) data;
858	int err;
859
860	if (!(dd->flags & TDES_FLAGS_DMA))
861		err = atmel_tdes_crypt_pdc_stop(dd);
862	else
863		err = atmel_tdes_crypt_dma_stop(dd);
864
865	if (dd->total && !err) {
866		if (dd->flags & TDES_FLAGS_FAST) {
867			dd->in_sg = sg_next(dd->in_sg);
868			dd->out_sg = sg_next(dd->out_sg);
869			if (!dd->in_sg || !dd->out_sg)
870				err = -EINVAL;
871		}
872		if (!err)
873			err = atmel_tdes_crypt_start(dd);
874		if (!err)
875			return; /* DMA started. Not fininishing. */
876	}
877
878	atmel_tdes_finish_req(dd, err);
879	atmel_tdes_handle_queue(dd, NULL);
880}
881
882static irqreturn_t atmel_tdes_irq(int irq, void *dev_id)
883{
884	struct atmel_tdes_dev *tdes_dd = dev_id;
885	u32 reg;
886
887	reg = atmel_tdes_read(tdes_dd, TDES_ISR);
888	if (reg & atmel_tdes_read(tdes_dd, TDES_IMR)) {
889		atmel_tdes_write(tdes_dd, TDES_IDR, reg);
890		if (TDES_FLAGS_BUSY & tdes_dd->flags)
891			tasklet_schedule(&tdes_dd->done_task);
892		else
893			dev_warn(tdes_dd->dev, "TDES interrupt when no active requests.\n");
894		return IRQ_HANDLED;
895	}
896
897	return IRQ_NONE;
898}
899
900static void atmel_tdes_unregister_algs(struct atmel_tdes_dev *dd)
901{
902	int i;
903
904	for (i = 0; i < ARRAY_SIZE(tdes_algs); i++)
905		crypto_unregister_skcipher(&tdes_algs[i]);
906}
907
908static int atmel_tdes_register_algs(struct atmel_tdes_dev *dd)
909{
910	int err, i, j;
911
912	for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) {
913		atmel_tdes_skcipher_alg_init(&tdes_algs[i]);
914
915		err = crypto_register_skcipher(&tdes_algs[i]);
916		if (err)
917			goto err_tdes_algs;
918	}
919
920	return 0;
921
922err_tdes_algs:
923	for (j = 0; j < i; j++)
924		crypto_unregister_skcipher(&tdes_algs[j]);
925
926	return err;
927}
928
929static void atmel_tdes_get_cap(struct atmel_tdes_dev *dd)
930{
931
932	dd->caps.has_dma = 0;
933
934	/* keep only major version number */
935	switch (dd->hw_version & 0xf00) {
936	case 0x800:
937	case 0x700:
938		dd->caps.has_dma = 1;
939		break;
940	case 0x600:
941		break;
942	default:
943		dev_warn(dd->dev,
944				"Unmanaged tdes version, set minimum capabilities\n");
945		break;
946	}
947}
948
949static const struct of_device_id atmel_tdes_dt_ids[] = {
950	{ .compatible = "atmel,at91sam9g46-tdes" },
951	{ /* sentinel */ }
952};
953MODULE_DEVICE_TABLE(of, atmel_tdes_dt_ids);
954
955static int atmel_tdes_probe(struct platform_device *pdev)
956{
957	struct atmel_tdes_dev *tdes_dd;
958	struct device *dev = &pdev->dev;
959	struct resource *tdes_res;
960	int err;
961
962	tdes_dd = devm_kmalloc(&pdev->dev, sizeof(*tdes_dd), GFP_KERNEL);
963	if (!tdes_dd)
964		return -ENOMEM;
965
966	tdes_dd->dev = dev;
967
968	platform_set_drvdata(pdev, tdes_dd);
969
970	INIT_LIST_HEAD(&tdes_dd->list);
971	spin_lock_init(&tdes_dd->lock);
972
973	tasklet_init(&tdes_dd->done_task, atmel_tdes_done_task,
974					(unsigned long)tdes_dd);
975	tasklet_init(&tdes_dd->queue_task, atmel_tdes_queue_task,
976					(unsigned long)tdes_dd);
977
978	crypto_init_queue(&tdes_dd->queue, ATMEL_TDES_QUEUE_LENGTH);
979
980	tdes_dd->io_base = devm_platform_get_and_ioremap_resource(pdev, 0, &tdes_res);
981	if (IS_ERR(tdes_dd->io_base)) {
982		err = PTR_ERR(tdes_dd->io_base);
983		goto err_tasklet_kill;
984	}
985	tdes_dd->phys_base = tdes_res->start;
986
987	/* Get the IRQ */
988	tdes_dd->irq = platform_get_irq(pdev,  0);
989	if (tdes_dd->irq < 0) {
990		err = tdes_dd->irq;
991		goto err_tasklet_kill;
992	}
993
994	err = devm_request_irq(&pdev->dev, tdes_dd->irq, atmel_tdes_irq,
995			       IRQF_SHARED, "atmel-tdes", tdes_dd);
996	if (err) {
997		dev_err(dev, "unable to request tdes irq.\n");
998		goto err_tasklet_kill;
999	}
1000
1001	/* Initializing the clock */
1002	tdes_dd->iclk = devm_clk_get(&pdev->dev, "tdes_clk");
1003	if (IS_ERR(tdes_dd->iclk)) {
1004		dev_err(dev, "clock initialization failed.\n");
1005		err = PTR_ERR(tdes_dd->iclk);
1006		goto err_tasklet_kill;
1007	}
1008
1009	err = atmel_tdes_hw_version_init(tdes_dd);
1010	if (err)
1011		goto err_tasklet_kill;
1012
1013	atmel_tdes_get_cap(tdes_dd);
1014
1015	err = atmel_tdes_buff_init(tdes_dd);
1016	if (err)
1017		goto err_tasklet_kill;
1018
1019	if (tdes_dd->caps.has_dma) {
1020		err = atmel_tdes_dma_init(tdes_dd);
1021		if (err)
1022			goto err_buff_cleanup;
1023
1024		dev_info(dev, "using %s, %s for DMA transfers\n",
1025				dma_chan_name(tdes_dd->dma_lch_in.chan),
1026				dma_chan_name(tdes_dd->dma_lch_out.chan));
1027	}
1028
1029	spin_lock(&atmel_tdes.lock);
1030	list_add_tail(&tdes_dd->list, &atmel_tdes.dev_list);
1031	spin_unlock(&atmel_tdes.lock);
1032
1033	err = atmel_tdes_register_algs(tdes_dd);
1034	if (err)
1035		goto err_algs;
1036
1037	dev_info(dev, "Atmel DES/TDES\n");
1038
1039	return 0;
1040
1041err_algs:
1042	spin_lock(&atmel_tdes.lock);
1043	list_del(&tdes_dd->list);
1044	spin_unlock(&atmel_tdes.lock);
1045	if (tdes_dd->caps.has_dma)
1046		atmel_tdes_dma_cleanup(tdes_dd);
1047err_buff_cleanup:
1048	atmel_tdes_buff_cleanup(tdes_dd);
1049err_tasklet_kill:
1050	tasklet_kill(&tdes_dd->done_task);
1051	tasklet_kill(&tdes_dd->queue_task);
1052
1053	return err;
1054}
1055
1056static void atmel_tdes_remove(struct platform_device *pdev)
1057{
1058	struct atmel_tdes_dev *tdes_dd = platform_get_drvdata(pdev);
1059
1060	spin_lock(&atmel_tdes.lock);
1061	list_del(&tdes_dd->list);
1062	spin_unlock(&atmel_tdes.lock);
1063
1064	atmel_tdes_unregister_algs(tdes_dd);
1065
1066	tasklet_kill(&tdes_dd->done_task);
1067	tasklet_kill(&tdes_dd->queue_task);
1068
1069	if (tdes_dd->caps.has_dma)
1070		atmel_tdes_dma_cleanup(tdes_dd);
1071
1072	atmel_tdes_buff_cleanup(tdes_dd);
1073}
1074
1075static struct platform_driver atmel_tdes_driver = {
1076	.probe		= atmel_tdes_probe,
1077	.remove_new	= atmel_tdes_remove,
1078	.driver		= {
1079		.name	= "atmel_tdes",
1080		.of_match_table = atmel_tdes_dt_ids,
1081	},
1082};
1083
1084module_platform_driver(atmel_tdes_driver);
1085
1086MODULE_DESCRIPTION("Atmel DES/TDES hw acceleration support.");
1087MODULE_LICENSE("GPL v2");
1088MODULE_AUTHOR("Nicolas Royer - Eukr��a Electromatique");
1089