hifn7751.c revision 116924
1/* $FreeBSD: head/sys/dev/hifn/hifn7751.c 116924 2003-06-27 20:07:10Z sam $ */
2/*	$OpenBSD: hifn7751.c,v 1.120 2002/05/17 00:33:34 deraadt Exp $	*/
3
4/*
5 * Invertex AEON / Hifn 7751 driver
6 * Copyright (c) 1999 Invertex Inc. All rights reserved.
7 * Copyright (c) 1999 Theo de Raadt
8 * Copyright (c) 2000-2001 Network Security Technologies, Inc.
9 *			http://www.netsec.net
10 *
11 * This driver is based on a previous driver by Invertex, for which they
12 * requested:  Please send any comments, feedback, bug-fixes, or feature
13 * requests to software@invertex.com.
14 *
15 * Redistribution and use in source and binary forms, with or without
16 * modification, are permitted provided that the following conditions
17 * are met:
18 *
19 * 1. Redistributions of source code must retain the above copyright
20 *   notice, this list of conditions and the following disclaimer.
21 * 2. Redistributions in binary form must reproduce the above copyright
22 *   notice, this list of conditions and the following disclaimer in the
23 *   documentation and/or other materials provided with the distribution.
24 * 3. The name of the author may not be used to endorse or promote products
25 *   derived from this software without specific prior written permission.
26 *
27 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
28 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
29 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
30 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
31 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
32 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
33 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
34 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
35 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
36 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
37 *
38 * Effort sponsored in part by the Defense Advanced Research Projects
39 * Agency (DARPA) and Air Force Research Laboratory, Air Force
40 * Materiel Command, USAF, under agreement number F30602-01-2-0537.
41 *
42 */
43
44/*
45 * Driver for the Hifn 7751 encryption processor.
46 */
47#include "opt_hifn.h"
48
49#include <sys/param.h>
50#include <sys/systm.h>
51#include <sys/proc.h>
52#include <sys/errno.h>
53#include <sys/malloc.h>
54#include <sys/kernel.h>
55#include <sys/mbuf.h>
56#include <sys/lock.h>
57#include <sys/mutex.h>
58#include <sys/sysctl.h>
59
60#include <vm/vm.h>
61#include <vm/pmap.h>
62
63#include <machine/clock.h>
64#include <machine/bus.h>
65#include <machine/resource.h>
66#include <sys/bus.h>
67#include <sys/rman.h>
68
69#include <opencrypto/cryptodev.h>
70#include <sys/random.h>
71
72#include <pci/pcivar.h>
73#include <pci/pcireg.h>
74
75#ifdef HIFN_RNDTEST
76#include <dev/rndtest/rndtest.h>
77#endif
78#include <dev/hifn/hifn7751reg.h>
79#include <dev/hifn/hifn7751var.h>
80
81/*
82 * Prototypes and count for the pci_device structure
83 */
84static	int hifn_probe(device_t);
85static	int hifn_attach(device_t);
86static	int hifn_detach(device_t);
87static	int hifn_suspend(device_t);
88static	int hifn_resume(device_t);
89static	void hifn_shutdown(device_t);
90
91static device_method_t hifn_methods[] = {
92	/* Device interface */
93	DEVMETHOD(device_probe,		hifn_probe),
94	DEVMETHOD(device_attach,	hifn_attach),
95	DEVMETHOD(device_detach,	hifn_detach),
96	DEVMETHOD(device_suspend,	hifn_suspend),
97	DEVMETHOD(device_resume,	hifn_resume),
98	DEVMETHOD(device_shutdown,	hifn_shutdown),
99
100	/* bus interface */
101	DEVMETHOD(bus_print_child,	bus_generic_print_child),
102	DEVMETHOD(bus_driver_added,	bus_generic_driver_added),
103
104	{ 0, 0 }
105};
106static driver_t hifn_driver = {
107	"hifn",
108	hifn_methods,
109	sizeof (struct hifn_softc)
110};
111static devclass_t hifn_devclass;
112
113DRIVER_MODULE(hifn, pci, hifn_driver, hifn_devclass, 0, 0);
114MODULE_DEPEND(hifn, crypto, 1, 1, 1);
115#ifdef HIFN_RNDTEST
116MODULE_DEPEND(hifn, rndtest, 1, 1, 1);
117#endif
118
119static	void hifn_reset_board(struct hifn_softc *, int);
120static	void hifn_reset_puc(struct hifn_softc *);
121static	void hifn_puc_wait(struct hifn_softc *);
122static	int hifn_enable_crypto(struct hifn_softc *);
123static	void hifn_set_retry(struct hifn_softc *sc);
124static	void hifn_init_dma(struct hifn_softc *);
125static	void hifn_init_pci_registers(struct hifn_softc *);
126static	int hifn_sramsize(struct hifn_softc *);
127static	int hifn_dramsize(struct hifn_softc *);
128static	int hifn_ramtype(struct hifn_softc *);
129static	void hifn_sessions(struct hifn_softc *);
130static	void hifn_intr(void *);
131static	u_int hifn_write_command(struct hifn_command *, u_int8_t *);
132static	u_int32_t hifn_next_signature(u_int32_t a, u_int cnt);
133static	int hifn_newsession(void *, u_int32_t *, struct cryptoini *);
134static	int hifn_freesession(void *, u_int64_t);
135static	int hifn_process(void *, struct cryptop *, int);
136static	void hifn_callback(struct hifn_softc *, struct hifn_command *, u_int8_t *);
137static	int hifn_crypto(struct hifn_softc *, struct hifn_command *, struct cryptop *, int);
138static	int hifn_readramaddr(struct hifn_softc *, int, u_int8_t *);
139static	int hifn_writeramaddr(struct hifn_softc *, int, u_int8_t *);
140static	int hifn_dmamap_load_src(struct hifn_softc *, struct hifn_command *);
141static	int hifn_dmamap_load_dst(struct hifn_softc *, struct hifn_command *);
142static	int hifn_init_pubrng(struct hifn_softc *);
143static	void hifn_rng(void *);
144static	void hifn_tick(void *);
145static	void hifn_abort(struct hifn_softc *);
146static	void hifn_alloc_slot(struct hifn_softc *, int *, int *, int *, int *);
147
148static	void hifn_write_reg_0(struct hifn_softc *, bus_size_t, u_int32_t);
149static	void hifn_write_reg_1(struct hifn_softc *, bus_size_t, u_int32_t);
150
151static __inline__ u_int32_t
152READ_REG_0(struct hifn_softc *sc, bus_size_t reg)
153{
154    u_int32_t v = bus_space_read_4(sc->sc_st0, sc->sc_sh0, reg);
155    sc->sc_bar0_lastreg = (bus_size_t) -1;
156    return (v);
157}
158#define	WRITE_REG_0(sc, reg, val)	hifn_write_reg_0(sc, reg, val)
159
160static __inline__ u_int32_t
161READ_REG_1(struct hifn_softc *sc, bus_size_t reg)
162{
163    u_int32_t v = bus_space_read_4(sc->sc_st1, sc->sc_sh1, reg);
164    sc->sc_bar1_lastreg = (bus_size_t) -1;
165    return (v);
166}
167#define	WRITE_REG_1(sc, reg, val)	hifn_write_reg_1(sc, reg, val)
168
169SYSCTL_NODE(_hw, OID_AUTO, hifn, CTLFLAG_RD, 0, "Hifn driver parameters");
170
171#ifdef HIFN_DEBUG
172static	int hifn_debug = 0;
173SYSCTL_INT(_hw_hifn, OID_AUTO, debug, CTLFLAG_RW, &hifn_debug,
174	    0, "control debugging msgs");
175#endif
176
177static	struct hifn_stats hifnstats;
178SYSCTL_STRUCT(_hw_hifn, OID_AUTO, stats, CTLFLAG_RD, &hifnstats,
179	    hifn_stats, "driver statistics");
180static	int hifn_maxbatch = 1;
181SYSCTL_INT(_hw_hifn, OID_AUTO, maxbatch, CTLFLAG_RW, &hifn_maxbatch,
182	    0, "max ops to batch w/o interrupt");
183
184/*
185 * Probe for a supported device.  The PCI vendor and device
186 * IDs are used to detect devices we know how to handle.
187 */
188static int
189hifn_probe(device_t dev)
190{
191	if (pci_get_vendor(dev) == PCI_VENDOR_INVERTEX &&
192	    pci_get_device(dev) == PCI_PRODUCT_INVERTEX_AEON)
193		return (0);
194	if (pci_get_vendor(dev) == PCI_VENDOR_HIFN &&
195	    (pci_get_device(dev) == PCI_PRODUCT_HIFN_7751 ||
196	     pci_get_device(dev) == PCI_PRODUCT_HIFN_7951 ||
197	     pci_get_device(dev) == PCI_PRODUCT_HIFN_7811))
198		return (0);
199	if (pci_get_vendor(dev) == PCI_VENDOR_NETSEC &&
200	    pci_get_device(dev) == PCI_PRODUCT_NETSEC_7751)
201		return (0);
202	return (ENXIO);
203}
204
205static void
206hifn_dmamap_cb(void *arg, bus_dma_segment_t *segs, int nseg, int error)
207{
208	bus_addr_t *paddr = (bus_addr_t*) arg;
209	*paddr = segs->ds_addr;
210}
211
212static const char*
213hifn_partname(struct hifn_softc *sc)
214{
215	/* XXX sprintf numbers when not decoded */
216	switch (pci_get_vendor(sc->sc_dev)) {
217	case PCI_VENDOR_HIFN:
218		switch (pci_get_device(sc->sc_dev)) {
219		case PCI_PRODUCT_HIFN_6500:	return "Hifn 6500";
220		case PCI_PRODUCT_HIFN_7751:	return "Hifn 7751";
221		case PCI_PRODUCT_HIFN_7811:	return "Hifn 7811";
222		case PCI_PRODUCT_HIFN_7951:	return "Hifn 7951";
223		}
224		return "Hifn unknown-part";
225	case PCI_VENDOR_INVERTEX:
226		switch (pci_get_device(sc->sc_dev)) {
227		case PCI_PRODUCT_INVERTEX_AEON:	return "Invertex AEON";
228		}
229		return "Invertex unknown-part";
230	case PCI_VENDOR_NETSEC:
231		switch (pci_get_device(sc->sc_dev)) {
232		case PCI_PRODUCT_NETSEC_7751:	return "NetSec 7751";
233		}
234		return "NetSec unknown-part";
235	}
236	return "Unknown-vendor unknown-part";
237}
238
239static void
240default_harvest(struct rndtest_state *rsp, void *buf, u_int count)
241{
242	random_harvest(buf, count, count*NBBY, 0, RANDOM_PURE);
243}
244
245/*
246 * Attach an interface that successfully probed.
247 */
248static int
249hifn_attach(device_t dev)
250{
251	struct hifn_softc *sc = device_get_softc(dev);
252	u_int32_t cmd;
253	caddr_t kva;
254	int rseg, rid;
255	char rbase;
256	u_int16_t ena, rev;
257
258	KASSERT(sc != NULL, ("hifn_attach: null software carrier!"));
259	bzero(sc, sizeof (*sc));
260	sc->sc_dev = dev;
261
262	mtx_init(&sc->sc_mtx, device_get_nameunit(dev), "hifn driver", MTX_DEF);
263
264	/* XXX handle power management */
265
266	/*
267	 * The 7951 has a random number generator and
268	 * public key support; note this.
269	 */
270	if (pci_get_vendor(dev) == PCI_VENDOR_HIFN &&
271	    pci_get_device(dev) == PCI_PRODUCT_HIFN_7951)
272		sc->sc_flags = HIFN_HAS_RNG | HIFN_HAS_PUBLIC;
273	/*
274	 * The 7811 has a random number generator and
275	 * we also note it's identity 'cuz of some quirks.
276	 */
277	if (pci_get_vendor(dev) == PCI_VENDOR_HIFN &&
278	    pci_get_device(dev) == PCI_PRODUCT_HIFN_7811)
279		sc->sc_flags |= HIFN_IS_7811 | HIFN_HAS_RNG;
280
281	/*
282	 * Configure support for memory-mapped access to
283	 * registers and for DMA operations.
284	 */
285#define	PCIM_ENA	(PCIM_CMD_MEMEN|PCIM_CMD_BUSMASTEREN)
286	cmd = pci_read_config(dev, PCIR_COMMAND, 4);
287	cmd |= PCIM_ENA;
288	pci_write_config(dev, PCIR_COMMAND, cmd, 4);
289	cmd = pci_read_config(dev, PCIR_COMMAND, 4);
290	if ((cmd & PCIM_ENA) != PCIM_ENA) {
291		device_printf(dev, "failed to enable %s\n",
292			(cmd & PCIM_ENA) == 0 ?
293				"memory mapping & bus mastering" :
294			(cmd & PCIM_CMD_MEMEN) == 0 ?
295				"memory mapping" : "bus mastering");
296		goto fail_pci;
297	}
298#undef PCIM_ENA
299
300	/*
301	 * Setup PCI resources. Note that we record the bus
302	 * tag and handle for each register mapping, this is
303	 * used by the READ_REG_0, WRITE_REG_0, READ_REG_1,
304	 * and WRITE_REG_1 macros throughout the driver.
305	 */
306	rid = HIFN_BAR0;
307	sc->sc_bar0res = bus_alloc_resource(dev, SYS_RES_MEMORY, &rid,
308			 		    0, ~0, 1, RF_ACTIVE);
309	if (sc->sc_bar0res == NULL) {
310		device_printf(dev, "cannot map bar%d register space\n", 0);
311		goto fail_pci;
312	}
313	sc->sc_st0 = rman_get_bustag(sc->sc_bar0res);
314	sc->sc_sh0 = rman_get_bushandle(sc->sc_bar0res);
315	sc->sc_bar0_lastreg = (bus_size_t) -1;
316
317	rid = HIFN_BAR1;
318	sc->sc_bar1res = bus_alloc_resource(dev, SYS_RES_MEMORY, &rid,
319					    0, ~0, 1, RF_ACTIVE);
320	if (sc->sc_bar1res == NULL) {
321		device_printf(dev, "cannot map bar%d register space\n", 1);
322		goto fail_io0;
323	}
324	sc->sc_st1 = rman_get_bustag(sc->sc_bar1res);
325	sc->sc_sh1 = rman_get_bushandle(sc->sc_bar1res);
326	sc->sc_bar1_lastreg = (bus_size_t) -1;
327
328	hifn_set_retry(sc);
329
330	/*
331	 * Setup the area where the Hifn DMA's descriptors
332	 * and associated data structures.
333	 */
334	if (bus_dma_tag_create(NULL,			/* parent */
335			       1, 0,			/* alignment,boundary */
336			       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
337			       BUS_SPACE_MAXADDR,	/* highaddr */
338			       NULL, NULL,		/* filter, filterarg */
339			       HIFN_MAX_DMALEN,		/* maxsize */
340			       MAX_SCATTER,		/* nsegments */
341			       HIFN_MAX_SEGLEN,		/* maxsegsize */
342			       BUS_DMA_ALLOCNOW,	/* flags */
343			       &sc->sc_dmat)) {
344		device_printf(dev, "cannot allocate DMA tag\n");
345		goto fail_io1;
346	}
347	if (bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT, &sc->sc_dmamap)) {
348		device_printf(dev, "cannot create dma map\n");
349		bus_dma_tag_destroy(sc->sc_dmat);
350		goto fail_io1;
351	}
352	if (bus_dmamem_alloc(sc->sc_dmat, (void**) &kva, BUS_DMA_NOWAIT, &sc->sc_dmamap)) {
353		device_printf(dev, "cannot alloc dma buffer\n");
354		bus_dmamap_destroy(sc->sc_dmat, sc->sc_dmamap);
355		bus_dma_tag_destroy(sc->sc_dmat);
356		goto fail_io1;
357	}
358	if (bus_dmamap_load(sc->sc_dmat, sc->sc_dmamap, kva,
359			     sizeof (*sc->sc_dma),
360			     hifn_dmamap_cb, &sc->sc_dma_physaddr,
361			     BUS_DMA_NOWAIT)) {
362		device_printf(dev, "cannot load dma map\n");
363		bus_dmamem_free(sc->sc_dmat, kva, sc->sc_dmamap);
364		bus_dmamap_destroy(sc->sc_dmat, sc->sc_dmamap);
365		bus_dma_tag_destroy(sc->sc_dmat);
366		goto fail_io1;
367	}
368	sc->sc_dma = (struct hifn_dma *)kva;
369	bzero(sc->sc_dma, sizeof(*sc->sc_dma));
370
371	KASSERT(sc->sc_st0 != NULL, ("hifn_attach: null bar0 tag!"));
372	KASSERT(sc->sc_sh0 != NULL, ("hifn_attach: null bar0 handle!"));
373	KASSERT(sc->sc_st1 != NULL, ("hifn_attach: null bar1 tag!"));
374	KASSERT(sc->sc_sh1 != NULL, ("hifn_attach: null bar1 handle!"));
375
376	/*
377	 * Reset the board and do the ``secret handshake''
378	 * to enable the crypto support.  Then complete the
379	 * initialization procedure by setting up the interrupt
380	 * and hooking in to the system crypto support so we'll
381	 * get used for system services like the crypto device,
382	 * IPsec, RNG device, etc.
383	 */
384	hifn_reset_board(sc, 0);
385
386	if (hifn_enable_crypto(sc) != 0) {
387		device_printf(dev, "crypto enabling failed\n");
388		goto fail_mem;
389	}
390	hifn_reset_puc(sc);
391
392	hifn_init_dma(sc);
393	hifn_init_pci_registers(sc);
394
395	if (hifn_ramtype(sc))
396		goto fail_mem;
397
398	if (sc->sc_drammodel == 0)
399		hifn_sramsize(sc);
400	else
401		hifn_dramsize(sc);
402
403	/*
404	 * Workaround for NetSec 7751 rev A: half ram size because two
405	 * of the address lines were left floating
406	 */
407	if (pci_get_vendor(dev) == PCI_VENDOR_NETSEC &&
408	    pci_get_device(dev) == PCI_PRODUCT_NETSEC_7751 &&
409	    pci_get_revid(dev) == 0x61)	/*XXX???*/
410		sc->sc_ramsize >>= 1;
411
412	/*
413	 * Arrange the interrupt line.
414	 */
415	rid = 0;
416	sc->sc_irq = bus_alloc_resource(dev, SYS_RES_IRQ, &rid,
417					0, ~0, 1, RF_SHAREABLE|RF_ACTIVE);
418	if (sc->sc_irq == NULL) {
419		device_printf(dev, "could not map interrupt\n");
420		goto fail_mem;
421	}
422	/*
423	 * NB: Network code assumes we are blocked with splimp()
424	 *     so make sure the IRQ is marked appropriately.
425	 */
426	if (bus_setup_intr(dev, sc->sc_irq, INTR_TYPE_NET | INTR_MPSAFE,
427			   hifn_intr, sc, &sc->sc_intrhand)) {
428		device_printf(dev, "could not setup interrupt\n");
429		goto fail_intr2;
430	}
431
432	hifn_sessions(sc);
433
434	/*
435	 * NB: Keep only the low 16 bits; this masks the chip id
436	 *     from the 7951.
437	 */
438	rev = READ_REG_1(sc, HIFN_1_REVID) & 0xffff;
439
440	rseg = sc->sc_ramsize / 1024;
441	rbase = 'K';
442	if (sc->sc_ramsize >= (1024 * 1024)) {
443		rbase = 'M';
444		rseg /= 1024;
445	}
446	device_printf(sc->sc_dev, "%s, rev %u, %d%cB %cram, %u sessions\n",
447		hifn_partname(sc), rev,
448		rseg, rbase, sc->sc_drammodel ? 'd' : 's',
449		sc->sc_maxses);
450
451	sc->sc_cid = crypto_get_driverid(0);
452	if (sc->sc_cid < 0) {
453		device_printf(dev, "could not get crypto driver id\n");
454		goto fail_intr;
455	}
456
457	WRITE_REG_0(sc, HIFN_0_PUCNFG,
458	    READ_REG_0(sc, HIFN_0_PUCNFG) | HIFN_PUCNFG_CHIPID);
459	ena = READ_REG_0(sc, HIFN_0_PUSTAT) & HIFN_PUSTAT_CHIPENA;
460
461	switch (ena) {
462	case HIFN_PUSTAT_ENA_2:
463		crypto_register(sc->sc_cid, CRYPTO_3DES_CBC, 0, 0,
464		    hifn_newsession, hifn_freesession, hifn_process, sc);
465		crypto_register(sc->sc_cid, CRYPTO_ARC4, 0, 0,
466		    hifn_newsession, hifn_freesession, hifn_process, sc);
467		/*FALLTHROUGH*/
468	case HIFN_PUSTAT_ENA_1:
469		crypto_register(sc->sc_cid, CRYPTO_MD5, 0, 0,
470		    hifn_newsession, hifn_freesession, hifn_process, sc);
471		crypto_register(sc->sc_cid, CRYPTO_SHA1, 0, 0,
472		    hifn_newsession, hifn_freesession, hifn_process, sc);
473		crypto_register(sc->sc_cid, CRYPTO_MD5_HMAC, 0, 0,
474		    hifn_newsession, hifn_freesession, hifn_process, sc);
475		crypto_register(sc->sc_cid, CRYPTO_SHA1_HMAC, 0, 0,
476		    hifn_newsession, hifn_freesession, hifn_process, sc);
477		crypto_register(sc->sc_cid, CRYPTO_DES_CBC, 0, 0,
478		    hifn_newsession, hifn_freesession, hifn_process, sc);
479		break;
480	}
481
482	bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
483	    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
484
485	if (sc->sc_flags & (HIFN_HAS_PUBLIC | HIFN_HAS_RNG))
486		hifn_init_pubrng(sc);
487
488	/* NB: 1 means the callout runs w/o Giant locked */
489	callout_init(&sc->sc_tickto, 1);
490	callout_reset(&sc->sc_tickto, hz, hifn_tick, sc);
491
492	return (0);
493
494fail_intr:
495	bus_teardown_intr(dev, sc->sc_irq, sc->sc_intrhand);
496fail_intr2:
497	/* XXX don't store rid */
498	bus_release_resource(dev, SYS_RES_IRQ, 0, sc->sc_irq);
499fail_mem:
500	bus_dmamap_unload(sc->sc_dmat, sc->sc_dmamap);
501	bus_dmamem_free(sc->sc_dmat, sc->sc_dma, sc->sc_dmamap);
502	bus_dmamap_destroy(sc->sc_dmat, sc->sc_dmamap);
503	bus_dma_tag_destroy(sc->sc_dmat);
504
505	/* Turn off DMA polling */
506	WRITE_REG_1(sc, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
507	    HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);
508fail_io1:
509	bus_release_resource(dev, SYS_RES_MEMORY, HIFN_BAR1, sc->sc_bar1res);
510fail_io0:
511	bus_release_resource(dev, SYS_RES_MEMORY, HIFN_BAR0, sc->sc_bar0res);
512fail_pci:
513	mtx_destroy(&sc->sc_mtx);
514	return (ENXIO);
515}
516
517/*
518 * Detach an interface that successfully probed.
519 */
520static int
521hifn_detach(device_t dev)
522{
523	struct hifn_softc *sc = device_get_softc(dev);
524
525	KASSERT(sc != NULL, ("hifn_detach: null software carrier!"));
526
527	/* disable interrupts */
528	WRITE_REG_1(sc, HIFN_1_DMA_IER, 0);
529
530	/*XXX other resources */
531	callout_stop(&sc->sc_tickto);
532	callout_stop(&sc->sc_rngto);
533#ifdef HIFN_RNDTEST
534	if (sc->sc_rndtest)
535		rndtest_detach(sc->sc_rndtest);
536#endif
537
538	/* Turn off DMA polling */
539	WRITE_REG_1(sc, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
540	    HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);
541
542	crypto_unregister_all(sc->sc_cid);
543
544	bus_generic_detach(dev);	/*XXX should be no children, right? */
545
546	bus_teardown_intr(dev, sc->sc_irq, sc->sc_intrhand);
547	/* XXX don't store rid */
548	bus_release_resource(dev, SYS_RES_IRQ, 0, sc->sc_irq);
549
550	bus_dmamap_unload(sc->sc_dmat, sc->sc_dmamap);
551	bus_dmamem_free(sc->sc_dmat, sc->sc_dma, sc->sc_dmamap);
552	bus_dmamap_destroy(sc->sc_dmat, sc->sc_dmamap);
553	bus_dma_tag_destroy(sc->sc_dmat);
554
555	bus_release_resource(dev, SYS_RES_MEMORY, HIFN_BAR1, sc->sc_bar1res);
556	bus_release_resource(dev, SYS_RES_MEMORY, HIFN_BAR0, sc->sc_bar0res);
557
558	mtx_destroy(&sc->sc_mtx);
559
560	return (0);
561}
562
563/*
564 * Stop all chip I/O so that the kernel's probe routines don't
565 * get confused by errant DMAs when rebooting.
566 */
567static void
568hifn_shutdown(device_t dev)
569{
570#ifdef notyet
571	hifn_stop(device_get_softc(dev));
572#endif
573}
574
575/*
576 * Device suspend routine.  Stop the interface and save some PCI
577 * settings in case the BIOS doesn't restore them properly on
578 * resume.
579 */
580static int
581hifn_suspend(device_t dev)
582{
583	struct hifn_softc *sc = device_get_softc(dev);
584#ifdef notyet
585	int i;
586
587	hifn_stop(sc);
588	for (i = 0; i < 5; i++)
589		sc->saved_maps[i] = pci_read_config(dev, PCIR_MAPS + i * 4, 4);
590	sc->saved_biosaddr = pci_read_config(dev, PCIR_BIOS, 4);
591	sc->saved_intline = pci_read_config(dev, PCIR_INTLINE, 1);
592	sc->saved_cachelnsz = pci_read_config(dev, PCIR_CACHELNSZ, 1);
593	sc->saved_lattimer = pci_read_config(dev, PCIR_LATTIMER, 1);
594#endif
595	sc->sc_suspended = 1;
596
597	return (0);
598}
599
600/*
601 * Device resume routine.  Restore some PCI settings in case the BIOS
602 * doesn't, re-enable busmastering, and restart the interface if
603 * appropriate.
604 */
605static int
606hifn_resume(device_t dev)
607{
608	struct hifn_softc *sc = device_get_softc(dev);
609#ifdef notyet
610	int i;
611
612	/* better way to do this? */
613	for (i = 0; i < 5; i++)
614		pci_write_config(dev, PCIR_MAPS + i * 4, sc->saved_maps[i], 4);
615	pci_write_config(dev, PCIR_BIOS, sc->saved_biosaddr, 4);
616	pci_write_config(dev, PCIR_INTLINE, sc->saved_intline, 1);
617	pci_write_config(dev, PCIR_CACHELNSZ, sc->saved_cachelnsz, 1);
618	pci_write_config(dev, PCIR_LATTIMER, sc->saved_lattimer, 1);
619
620	/* reenable busmastering */
621	pci_enable_busmaster(dev);
622	pci_enable_io(dev, HIFN_RES);
623
624        /* reinitialize interface if necessary */
625        if (ifp->if_flags & IFF_UP)
626                rl_init(sc);
627#endif
628	sc->sc_suspended = 0;
629
630	return (0);
631}
632
633static int
634hifn_init_pubrng(struct hifn_softc *sc)
635{
636	u_int32_t r;
637	int i;
638
639#ifdef HIFN_RNDTEST
640	sc->sc_rndtest = rndtest_attach(sc->sc_dev);
641	if (sc->sc_rndtest)
642		sc->sc_harvest = rndtest_harvest;
643	else
644		sc->sc_harvest = default_harvest;
645#else
646	sc->sc_harvest = default_harvest;
647#endif
648	if ((sc->sc_flags & HIFN_IS_7811) == 0) {
649		/* Reset 7951 public key/rng engine */
650		WRITE_REG_1(sc, HIFN_1_PUB_RESET,
651		    READ_REG_1(sc, HIFN_1_PUB_RESET) | HIFN_PUBRST_RESET);
652
653		for (i = 0; i < 100; i++) {
654			DELAY(1000);
655			if ((READ_REG_1(sc, HIFN_1_PUB_RESET) &
656			    HIFN_PUBRST_RESET) == 0)
657				break;
658		}
659
660		if (i == 100) {
661			device_printf(sc->sc_dev, "public key init failed\n");
662			return (1);
663		}
664	}
665
666	/* Enable the rng, if available */
667	if (sc->sc_flags & HIFN_HAS_RNG) {
668		if (sc->sc_flags & HIFN_IS_7811) {
669			r = READ_REG_1(sc, HIFN_1_7811_RNGENA);
670			if (r & HIFN_7811_RNGENA_ENA) {
671				r &= ~HIFN_7811_RNGENA_ENA;
672				WRITE_REG_1(sc, HIFN_1_7811_RNGENA, r);
673			}
674			WRITE_REG_1(sc, HIFN_1_7811_RNGCFG,
675			    HIFN_7811_RNGCFG_DEFL);
676			r |= HIFN_7811_RNGENA_ENA;
677			WRITE_REG_1(sc, HIFN_1_7811_RNGENA, r);
678		} else
679			WRITE_REG_1(sc, HIFN_1_RNG_CONFIG,
680			    READ_REG_1(sc, HIFN_1_RNG_CONFIG) |
681			    HIFN_RNGCFG_ENA);
682
683		sc->sc_rngfirst = 1;
684		if (hz >= 100)
685			sc->sc_rnghz = hz / 100;
686		else
687			sc->sc_rnghz = 1;
688		/* NB: 1 means the callout runs w/o Giant locked */
689		callout_init(&sc->sc_rngto, 1);
690		callout_reset(&sc->sc_rngto, sc->sc_rnghz, hifn_rng, sc);
691	}
692
693	/* Enable public key engine, if available */
694	if (sc->sc_flags & HIFN_HAS_PUBLIC) {
695		WRITE_REG_1(sc, HIFN_1_PUB_IEN, HIFN_PUBIEN_DONE);
696		sc->sc_dmaier |= HIFN_DMAIER_PUBDONE;
697		WRITE_REG_1(sc, HIFN_1_DMA_IER, sc->sc_dmaier);
698	}
699
700	return (0);
701}
702
703static void
704hifn_rng(void *vsc)
705{
706#define	RANDOM_BITS(n)	(n)*sizeof (u_int32_t), (n)*sizeof (u_int32_t)*NBBY, 0
707	struct hifn_softc *sc = vsc;
708	u_int32_t sts, num[2];
709	int i;
710
711	if (sc->sc_flags & HIFN_IS_7811) {
712		for (i = 0; i < 5; i++) {
713			sts = READ_REG_1(sc, HIFN_1_7811_RNGSTS);
714			if (sts & HIFN_7811_RNGSTS_UFL) {
715				device_printf(sc->sc_dev,
716					      "RNG underflow: disabling\n");
717				return;
718			}
719			if ((sts & HIFN_7811_RNGSTS_RDY) == 0)
720				break;
721
722			/*
723			 * There are at least two words in the RNG FIFO
724			 * at this point.
725			 */
726			num[0] = READ_REG_1(sc, HIFN_1_7811_RNGDAT);
727			num[1] = READ_REG_1(sc, HIFN_1_7811_RNGDAT);
728			/* NB: discard first data read */
729			if (sc->sc_rngfirst)
730				sc->sc_rngfirst = 0;
731			else
732				(*sc->sc_harvest)(sc->sc_rndtest,
733					num, sizeof (num));
734		}
735	} else {
736		num[0] = READ_REG_1(sc, HIFN_1_RNG_DATA);
737
738		/* NB: discard first data read */
739		if (sc->sc_rngfirst)
740			sc->sc_rngfirst = 0;
741		else
742			(*sc->sc_harvest)(sc->sc_rndtest,
743				num, sizeof (num[0]));
744	}
745
746	callout_reset(&sc->sc_rngto, sc->sc_rnghz, hifn_rng, sc);
747#undef RANDOM_BITS
748}
749
750static void
751hifn_puc_wait(struct hifn_softc *sc)
752{
753	int i;
754
755	for (i = 5000; i > 0; i--) {
756		DELAY(1);
757		if (!(READ_REG_0(sc, HIFN_0_PUCTRL) & HIFN_PUCTRL_RESET))
758			break;
759	}
760	if (!i)
761		device_printf(sc->sc_dev, "proc unit did not reset\n");
762}
763
764/*
765 * Reset the processing unit.
766 */
767static void
768hifn_reset_puc(struct hifn_softc *sc)
769{
770	/* Reset processing unit */
771	WRITE_REG_0(sc, HIFN_0_PUCTRL, HIFN_PUCTRL_DMAENA);
772	hifn_puc_wait(sc);
773}
774
775/*
776 * Set the Retry and TRDY registers; note that we set them to
777 * zero because the 7811 locks up when forced to retry (section
778 * 3.6 of "Specification Update SU-0014-04".  Not clear if we
779 * should do this for all Hifn parts, but it doesn't seem to hurt.
780 */
781static void
782hifn_set_retry(struct hifn_softc *sc)
783{
784	/* NB: RETRY only responds to 8-bit reads/writes */
785	pci_write_config(sc->sc_dev, HIFN_RETRY_TIMEOUT, 0, 1);
786	pci_write_config(sc->sc_dev, HIFN_TRDY_TIMEOUT, 0, 4);
787}
788
789/*
790 * Resets the board.  Values in the regesters are left as is
791 * from the reset (i.e. initial values are assigned elsewhere).
792 */
793static void
794hifn_reset_board(struct hifn_softc *sc, int full)
795{
796	u_int32_t reg;
797
798	/*
799	 * Set polling in the DMA configuration register to zero.  0x7 avoids
800	 * resetting the board and zeros out the other fields.
801	 */
802	WRITE_REG_1(sc, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
803	    HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);
804
805	/*
806	 * Now that polling has been disabled, we have to wait 1 ms
807	 * before resetting the board.
808	 */
809	DELAY(1000);
810
811	/* Reset the DMA unit */
812	if (full) {
813		WRITE_REG_1(sc, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MODE);
814		DELAY(1000);
815	} else {
816		WRITE_REG_1(sc, HIFN_1_DMA_CNFG,
817		    HIFN_DMACNFG_MODE | HIFN_DMACNFG_MSTRESET);
818		hifn_reset_puc(sc);
819	}
820
821	KASSERT(sc->sc_dma != NULL, ("hifn_reset_board: null DMA tag!"));
822	bzero(sc->sc_dma, sizeof(*sc->sc_dma));
823
824	/* Bring dma unit out of reset */
825	WRITE_REG_1(sc, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
826	    HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);
827
828	hifn_puc_wait(sc);
829	hifn_set_retry(sc);
830
831	if (sc->sc_flags & HIFN_IS_7811) {
832		for (reg = 0; reg < 1000; reg++) {
833			if (READ_REG_1(sc, HIFN_1_7811_MIPSRST) &
834			    HIFN_MIPSRST_CRAMINIT)
835				break;
836			DELAY(1000);
837		}
838		if (reg == 1000)
839			printf(": cram init timeout\n");
840	}
841}
842
843static u_int32_t
844hifn_next_signature(u_int32_t a, u_int cnt)
845{
846	int i;
847	u_int32_t v;
848
849	for (i = 0; i < cnt; i++) {
850
851		/* get the parity */
852		v = a & 0x80080125;
853		v ^= v >> 16;
854		v ^= v >> 8;
855		v ^= v >> 4;
856		v ^= v >> 2;
857		v ^= v >> 1;
858
859		a = (v & 1) ^ (a << 1);
860	}
861
862	return a;
863}
864
865struct pci2id {
866	u_short		pci_vendor;
867	u_short		pci_prod;
868	char		card_id[13];
869};
870static struct pci2id pci2id[] = {
871	{
872		PCI_VENDOR_HIFN,
873		PCI_PRODUCT_HIFN_7951,
874		{ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
875		  0x00, 0x00, 0x00, 0x00, 0x00 }
876	}, {
877		PCI_VENDOR_NETSEC,
878		PCI_PRODUCT_NETSEC_7751,
879		{ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
880		  0x00, 0x00, 0x00, 0x00, 0x00 }
881	}, {
882		PCI_VENDOR_INVERTEX,
883		PCI_PRODUCT_INVERTEX_AEON,
884		{ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
885		  0x00, 0x00, 0x00, 0x00, 0x00 }
886	}, {
887		PCI_VENDOR_HIFN,
888		PCI_PRODUCT_HIFN_7811,
889		{ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
890		  0x00, 0x00, 0x00, 0x00, 0x00 }
891	}, {
892		/*
893		 * Other vendors share this PCI ID as well, such as
894		 * http://www.powercrypt.com, and obviously they also
895		 * use the same key.
896		 */
897		PCI_VENDOR_HIFN,
898		PCI_PRODUCT_HIFN_7751,
899		{ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
900		  0x00, 0x00, 0x00, 0x00, 0x00 }
901	},
902};
903
904/*
905 * Checks to see if crypto is already enabled.  If crypto isn't enable,
906 * "hifn_enable_crypto" is called to enable it.  The check is important,
907 * as enabling crypto twice will lock the board.
908 */
909static int
910hifn_enable_crypto(struct hifn_softc *sc)
911{
912	u_int32_t dmacfg, ramcfg, encl, addr, i;
913	char *offtbl = NULL;
914
915	for (i = 0; i < sizeof(pci2id)/sizeof(pci2id[0]); i++) {
916		if (pci2id[i].pci_vendor == pci_get_vendor(sc->sc_dev) &&
917		    pci2id[i].pci_prod == pci_get_device(sc->sc_dev)) {
918			offtbl = pci2id[i].card_id;
919			break;
920		}
921	}
922	if (offtbl == NULL) {
923		device_printf(sc->sc_dev, "Unknown card!\n");
924		return (1);
925	}
926
927	ramcfg = READ_REG_0(sc, HIFN_0_PUCNFG);
928	dmacfg = READ_REG_1(sc, HIFN_1_DMA_CNFG);
929
930	/*
931	 * The RAM config register's encrypt level bit needs to be set before
932	 * every read performed on the encryption level register.
933	 */
934	WRITE_REG_0(sc, HIFN_0_PUCNFG, ramcfg | HIFN_PUCNFG_CHIPID);
935
936	encl = READ_REG_0(sc, HIFN_0_PUSTAT) & HIFN_PUSTAT_CHIPENA;
937
938	/*
939	 * Make sure we don't re-unlock.  Two unlocks kills chip until the
940	 * next reboot.
941	 */
942	if (encl == HIFN_PUSTAT_ENA_1 || encl == HIFN_PUSTAT_ENA_2) {
943#ifdef HIFN_DEBUG
944		if (hifn_debug)
945			device_printf(sc->sc_dev,
946			    "Strong crypto already enabled!\n");
947#endif
948		goto report;
949	}
950
951	if (encl != 0 && encl != HIFN_PUSTAT_ENA_0) {
952#ifdef HIFN_DEBUG
953		if (hifn_debug)
954			device_printf(sc->sc_dev,
955			      "Unknown encryption level 0x%x\n", encl);
956#endif
957		return 1;
958	}
959
960	WRITE_REG_1(sc, HIFN_1_DMA_CNFG, HIFN_DMACNFG_UNLOCK |
961	    HIFN_DMACNFG_MSTRESET | HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);
962	DELAY(1000);
963	addr = READ_REG_1(sc, HIFN_UNLOCK_SECRET1);
964	DELAY(1000);
965	WRITE_REG_1(sc, HIFN_UNLOCK_SECRET2, 0);
966	DELAY(1000);
967
968	for (i = 0; i <= 12; i++) {
969		addr = hifn_next_signature(addr, offtbl[i] + 0x101);
970		WRITE_REG_1(sc, HIFN_UNLOCK_SECRET2, addr);
971
972		DELAY(1000);
973	}
974
975	WRITE_REG_0(sc, HIFN_0_PUCNFG, ramcfg | HIFN_PUCNFG_CHIPID);
976	encl = READ_REG_0(sc, HIFN_0_PUSTAT) & HIFN_PUSTAT_CHIPENA;
977
978#ifdef HIFN_DEBUG
979	if (hifn_debug) {
980		if (encl != HIFN_PUSTAT_ENA_1 && encl != HIFN_PUSTAT_ENA_2)
981			device_printf(sc->sc_dev, "Engine is permanently "
982				"locked until next system reset!\n");
983		else
984			device_printf(sc->sc_dev, "Engine enabled "
985				"successfully!\n");
986	}
987#endif
988
989report:
990	WRITE_REG_0(sc, HIFN_0_PUCNFG, ramcfg);
991	WRITE_REG_1(sc, HIFN_1_DMA_CNFG, dmacfg);
992
993	switch (encl) {
994	case HIFN_PUSTAT_ENA_1:
995	case HIFN_PUSTAT_ENA_2:
996		break;
997	case HIFN_PUSTAT_ENA_0:
998	default:
999		device_printf(sc->sc_dev, "disabled");
1000		break;
1001	}
1002
1003	return 0;
1004}
1005
1006/*
1007 * Give initial values to the registers listed in the "Register Space"
1008 * section of the HIFN Software Development reference manual.
1009 */
1010static void
1011hifn_init_pci_registers(struct hifn_softc *sc)
1012{
1013	/* write fixed values needed by the Initialization registers */
1014	WRITE_REG_0(sc, HIFN_0_PUCTRL, HIFN_PUCTRL_DMAENA);
1015	WRITE_REG_0(sc, HIFN_0_FIFOCNFG, HIFN_FIFOCNFG_THRESHOLD);
1016	WRITE_REG_0(sc, HIFN_0_PUIER, HIFN_PUIER_DSTOVER);
1017
1018	/* write all 4 ring address registers */
1019	WRITE_REG_1(sc, HIFN_1_DMA_CRAR, sc->sc_dma_physaddr +
1020	    offsetof(struct hifn_dma, cmdr[0]));
1021	WRITE_REG_1(sc, HIFN_1_DMA_SRAR, sc->sc_dma_physaddr +
1022	    offsetof(struct hifn_dma, srcr[0]));
1023	WRITE_REG_1(sc, HIFN_1_DMA_DRAR, sc->sc_dma_physaddr +
1024	    offsetof(struct hifn_dma, dstr[0]));
1025	WRITE_REG_1(sc, HIFN_1_DMA_RRAR, sc->sc_dma_physaddr +
1026	    offsetof(struct hifn_dma, resr[0]));
1027
1028	DELAY(2000);
1029
1030	/* write status register */
1031	WRITE_REG_1(sc, HIFN_1_DMA_CSR,
1032	    HIFN_DMACSR_D_CTRL_DIS | HIFN_DMACSR_R_CTRL_DIS |
1033	    HIFN_DMACSR_S_CTRL_DIS | HIFN_DMACSR_C_CTRL_DIS |
1034	    HIFN_DMACSR_D_ABORT | HIFN_DMACSR_D_DONE | HIFN_DMACSR_D_LAST |
1035	    HIFN_DMACSR_D_WAIT | HIFN_DMACSR_D_OVER |
1036	    HIFN_DMACSR_R_ABORT | HIFN_DMACSR_R_DONE | HIFN_DMACSR_R_LAST |
1037	    HIFN_DMACSR_R_WAIT | HIFN_DMACSR_R_OVER |
1038	    HIFN_DMACSR_S_ABORT | HIFN_DMACSR_S_DONE | HIFN_DMACSR_S_LAST |
1039	    HIFN_DMACSR_S_WAIT |
1040	    HIFN_DMACSR_C_ABORT | HIFN_DMACSR_C_DONE | HIFN_DMACSR_C_LAST |
1041	    HIFN_DMACSR_C_WAIT |
1042	    HIFN_DMACSR_ENGINE |
1043	    ((sc->sc_flags & HIFN_HAS_PUBLIC) ?
1044		HIFN_DMACSR_PUBDONE : 0) |
1045	    ((sc->sc_flags & HIFN_IS_7811) ?
1046		HIFN_DMACSR_ILLW | HIFN_DMACSR_ILLR : 0));
1047
1048	sc->sc_d_busy = sc->sc_r_busy = sc->sc_s_busy = sc->sc_c_busy = 0;
1049	sc->sc_dmaier |= HIFN_DMAIER_R_DONE | HIFN_DMAIER_C_ABORT |
1050	    HIFN_DMAIER_D_OVER | HIFN_DMAIER_R_OVER |
1051	    HIFN_DMAIER_S_ABORT | HIFN_DMAIER_D_ABORT | HIFN_DMAIER_R_ABORT |
1052	    ((sc->sc_flags & HIFN_IS_7811) ?
1053		HIFN_DMAIER_ILLW | HIFN_DMAIER_ILLR : 0);
1054	sc->sc_dmaier &= ~HIFN_DMAIER_C_WAIT;
1055	WRITE_REG_1(sc, HIFN_1_DMA_IER, sc->sc_dmaier);
1056
1057	WRITE_REG_0(sc, HIFN_0_PUCNFG, HIFN_PUCNFG_COMPSING |
1058	    HIFN_PUCNFG_DRFR_128 | HIFN_PUCNFG_TCALLPHASES |
1059	    HIFN_PUCNFG_TCDRVTOTEM | HIFN_PUCNFG_BUS32 |
1060	    (sc->sc_drammodel ? HIFN_PUCNFG_DRAM : HIFN_PUCNFG_SRAM));
1061
1062	WRITE_REG_0(sc, HIFN_0_PUISR, HIFN_PUISR_DSTOVER);
1063	WRITE_REG_1(sc, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
1064	    HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE | HIFN_DMACNFG_LAST |
1065	    ((HIFN_POLL_FREQUENCY << 16 ) & HIFN_DMACNFG_POLLFREQ) |
1066	    ((HIFN_POLL_SCALAR << 8) & HIFN_DMACNFG_POLLINVAL));
1067}
1068
1069/*
1070 * The maximum number of sessions supported by the card
1071 * is dependent on the amount of context ram, which
1072 * encryption algorithms are enabled, and how compression
1073 * is configured.  This should be configured before this
1074 * routine is called.
1075 */
1076static void
1077hifn_sessions(struct hifn_softc *sc)
1078{
1079	u_int32_t pucnfg;
1080	int ctxsize;
1081
1082	pucnfg = READ_REG_0(sc, HIFN_0_PUCNFG);
1083
1084	if (pucnfg & HIFN_PUCNFG_COMPSING) {
1085		if (pucnfg & HIFN_PUCNFG_ENCCNFG)
1086			ctxsize = 128;
1087		else
1088			ctxsize = 512;
1089		sc->sc_maxses = 1 +
1090		    ((sc->sc_ramsize - 32768) / ctxsize);
1091	} else
1092		sc->sc_maxses = sc->sc_ramsize / 16384;
1093
1094	if (sc->sc_maxses > 2048)
1095		sc->sc_maxses = 2048;
1096}
1097
1098/*
1099 * Determine ram type (sram or dram).  Board should be just out of a reset
1100 * state when this is called.
1101 */
1102static int
1103hifn_ramtype(struct hifn_softc *sc)
1104{
1105	u_int8_t data[8], dataexpect[8];
1106	int i;
1107
1108	for (i = 0; i < sizeof(data); i++)
1109		data[i] = dataexpect[i] = 0x55;
1110	if (hifn_writeramaddr(sc, 0, data))
1111		return (-1);
1112	if (hifn_readramaddr(sc, 0, data))
1113		return (-1);
1114	if (bcmp(data, dataexpect, sizeof(data)) != 0) {
1115		sc->sc_drammodel = 1;
1116		return (0);
1117	}
1118
1119	for (i = 0; i < sizeof(data); i++)
1120		data[i] = dataexpect[i] = 0xaa;
1121	if (hifn_writeramaddr(sc, 0, data))
1122		return (-1);
1123	if (hifn_readramaddr(sc, 0, data))
1124		return (-1);
1125	if (bcmp(data, dataexpect, sizeof(data)) != 0) {
1126		sc->sc_drammodel = 1;
1127		return (0);
1128	}
1129
1130	return (0);
1131}
1132
1133#define	HIFN_SRAM_MAX		(32 << 20)
1134#define	HIFN_SRAM_STEP_SIZE	16384
1135#define	HIFN_SRAM_GRANULARITY	(HIFN_SRAM_MAX / HIFN_SRAM_STEP_SIZE)
1136
1137static int
1138hifn_sramsize(struct hifn_softc *sc)
1139{
1140	u_int32_t a;
1141	u_int8_t data[8];
1142	u_int8_t dataexpect[sizeof(data)];
1143	int32_t i;
1144
1145	for (i = 0; i < sizeof(data); i++)
1146		data[i] = dataexpect[i] = i ^ 0x5a;
1147
1148	for (i = HIFN_SRAM_GRANULARITY - 1; i >= 0; i--) {
1149		a = i * HIFN_SRAM_STEP_SIZE;
1150		bcopy(&i, data, sizeof(i));
1151		hifn_writeramaddr(sc, a, data);
1152	}
1153
1154	for (i = 0; i < HIFN_SRAM_GRANULARITY; i++) {
1155		a = i * HIFN_SRAM_STEP_SIZE;
1156		bcopy(&i, dataexpect, sizeof(i));
1157		if (hifn_readramaddr(sc, a, data) < 0)
1158			return (0);
1159		if (bcmp(data, dataexpect, sizeof(data)) != 0)
1160			return (0);
1161		sc->sc_ramsize = a + HIFN_SRAM_STEP_SIZE;
1162	}
1163
1164	return (0);
1165}
1166
1167/*
1168 * XXX For dram boards, one should really try all of the
1169 * HIFN_PUCNFG_DSZ_*'s.  This just assumes that PUCNFG
1170 * is already set up correctly.
1171 */
1172static int
1173hifn_dramsize(struct hifn_softc *sc)
1174{
1175	u_int32_t cnfg;
1176
1177	cnfg = READ_REG_0(sc, HIFN_0_PUCNFG) &
1178	    HIFN_PUCNFG_DRAMMASK;
1179	sc->sc_ramsize = 1 << ((cnfg >> 13) + 18);
1180	return (0);
1181}
1182
1183static void
1184hifn_alloc_slot(struct hifn_softc *sc, int *cmdp, int *srcp, int *dstp, int *resp)
1185{
1186	struct hifn_dma *dma = sc->sc_dma;
1187
1188	if (dma->cmdi == HIFN_D_CMD_RSIZE) {
1189		dma->cmdi = 0;
1190		dma->cmdr[HIFN_D_CMD_RSIZE].l = htole32(HIFN_D_VALID |
1191		    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1192		HIFN_CMDR_SYNC(sc, HIFN_D_CMD_RSIZE,
1193		    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1194	}
1195	*cmdp = dma->cmdi++;
1196	dma->cmdk = dma->cmdi;
1197
1198	if (dma->srci == HIFN_D_SRC_RSIZE) {
1199		dma->srci = 0;
1200		dma->srcr[HIFN_D_SRC_RSIZE].l = htole32(HIFN_D_VALID |
1201		    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1202		HIFN_SRCR_SYNC(sc, HIFN_D_SRC_RSIZE,
1203		    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1204	}
1205	*srcp = dma->srci++;
1206	dma->srck = dma->srci;
1207
1208	if (dma->dsti == HIFN_D_DST_RSIZE) {
1209		dma->dsti = 0;
1210		dma->dstr[HIFN_D_DST_RSIZE].l = htole32(HIFN_D_VALID |
1211		    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1212		HIFN_DSTR_SYNC(sc, HIFN_D_DST_RSIZE,
1213		    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1214	}
1215	*dstp = dma->dsti++;
1216	dma->dstk = dma->dsti;
1217
1218	if (dma->resi == HIFN_D_RES_RSIZE) {
1219		dma->resi = 0;
1220		dma->resr[HIFN_D_RES_RSIZE].l = htole32(HIFN_D_VALID |
1221		    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1222		HIFN_RESR_SYNC(sc, HIFN_D_RES_RSIZE,
1223		    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1224	}
1225	*resp = dma->resi++;
1226	dma->resk = dma->resi;
1227}
1228
1229static int
1230hifn_writeramaddr(struct hifn_softc *sc, int addr, u_int8_t *data)
1231{
1232	struct hifn_dma *dma = sc->sc_dma;
1233	hifn_base_command_t wc;
1234	const u_int32_t masks = HIFN_D_VALID | HIFN_D_LAST | HIFN_D_MASKDONEIRQ;
1235	int r, cmdi, resi, srci, dsti;
1236
1237	wc.masks = htole16(3 << 13);
1238	wc.session_num = htole16(addr >> 14);
1239	wc.total_source_count = htole16(8);
1240	wc.total_dest_count = htole16(addr & 0x3fff);
1241
1242	hifn_alloc_slot(sc, &cmdi, &srci, &dsti, &resi);
1243
1244	WRITE_REG_1(sc, HIFN_1_DMA_CSR,
1245	    HIFN_DMACSR_C_CTRL_ENA | HIFN_DMACSR_S_CTRL_ENA |
1246	    HIFN_DMACSR_D_CTRL_ENA | HIFN_DMACSR_R_CTRL_ENA);
1247
1248	/* build write command */
1249	bzero(dma->command_bufs[cmdi], HIFN_MAX_COMMAND);
1250	*(hifn_base_command_t *)dma->command_bufs[cmdi] = wc;
1251	bcopy(data, &dma->test_src, sizeof(dma->test_src));
1252
1253	dma->srcr[srci].p = htole32(sc->sc_dma_physaddr
1254	    + offsetof(struct hifn_dma, test_src));
1255	dma->dstr[dsti].p = htole32(sc->sc_dma_physaddr
1256	    + offsetof(struct hifn_dma, test_dst));
1257
1258	dma->cmdr[cmdi].l = htole32(16 | masks);
1259	dma->srcr[srci].l = htole32(8 | masks);
1260	dma->dstr[dsti].l = htole32(4 | masks);
1261	dma->resr[resi].l = htole32(4 | masks);
1262
1263	bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
1264	    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1265
1266	for (r = 10000; r >= 0; r--) {
1267		DELAY(10);
1268		bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
1269		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
1270		if ((dma->resr[resi].l & htole32(HIFN_D_VALID)) == 0)
1271			break;
1272		bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
1273		    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1274	}
1275	if (r == 0) {
1276		device_printf(sc->sc_dev, "writeramaddr -- "
1277		    "result[%d](addr %d) still valid\n", resi, addr);
1278		r = -1;
1279		return (-1);
1280	} else
1281		r = 0;
1282
1283	WRITE_REG_1(sc, HIFN_1_DMA_CSR,
1284	    HIFN_DMACSR_C_CTRL_DIS | HIFN_DMACSR_S_CTRL_DIS |
1285	    HIFN_DMACSR_D_CTRL_DIS | HIFN_DMACSR_R_CTRL_DIS);
1286
1287	return (r);
1288}
1289
1290static int
1291hifn_readramaddr(struct hifn_softc *sc, int addr, u_int8_t *data)
1292{
1293	struct hifn_dma *dma = sc->sc_dma;
1294	hifn_base_command_t rc;
1295	const u_int32_t masks = HIFN_D_VALID | HIFN_D_LAST | HIFN_D_MASKDONEIRQ;
1296	int r, cmdi, srci, dsti, resi;
1297
1298	rc.masks = htole16(2 << 13);
1299	rc.session_num = htole16(addr >> 14);
1300	rc.total_source_count = htole16(addr & 0x3fff);
1301	rc.total_dest_count = htole16(8);
1302
1303	hifn_alloc_slot(sc, &cmdi, &srci, &dsti, &resi);
1304
1305	WRITE_REG_1(sc, HIFN_1_DMA_CSR,
1306	    HIFN_DMACSR_C_CTRL_ENA | HIFN_DMACSR_S_CTRL_ENA |
1307	    HIFN_DMACSR_D_CTRL_ENA | HIFN_DMACSR_R_CTRL_ENA);
1308
1309	bzero(dma->command_bufs[cmdi], HIFN_MAX_COMMAND);
1310	*(hifn_base_command_t *)dma->command_bufs[cmdi] = rc;
1311
1312	dma->srcr[srci].p = htole32(sc->sc_dma_physaddr +
1313	    offsetof(struct hifn_dma, test_src));
1314	dma->test_src = 0;
1315	dma->dstr[dsti].p =  htole32(sc->sc_dma_physaddr +
1316	    offsetof(struct hifn_dma, test_dst));
1317	dma->test_dst = 0;
1318	dma->cmdr[cmdi].l = htole32(8 | masks);
1319	dma->srcr[srci].l = htole32(8 | masks);
1320	dma->dstr[dsti].l = htole32(8 | masks);
1321	dma->resr[resi].l = htole32(HIFN_MAX_RESULT | masks);
1322
1323	bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
1324	    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1325
1326	for (r = 10000; r >= 0; r--) {
1327		DELAY(10);
1328		bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
1329		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
1330		if ((dma->resr[resi].l & htole32(HIFN_D_VALID)) == 0)
1331			break;
1332		bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
1333		    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1334	}
1335	if (r == 0) {
1336		device_printf(sc->sc_dev, "readramaddr -- "
1337		    "result[%d](addr %d) still valid\n", resi, addr);
1338		r = -1;
1339	} else {
1340		r = 0;
1341		bcopy(&dma->test_dst, data, sizeof(dma->test_dst));
1342	}
1343
1344	WRITE_REG_1(sc, HIFN_1_DMA_CSR,
1345	    HIFN_DMACSR_C_CTRL_DIS | HIFN_DMACSR_S_CTRL_DIS |
1346	    HIFN_DMACSR_D_CTRL_DIS | HIFN_DMACSR_R_CTRL_DIS);
1347
1348	return (r);
1349}
1350
1351/*
1352 * Initialize the descriptor rings.
1353 */
1354static void
1355hifn_init_dma(struct hifn_softc *sc)
1356{
1357	struct hifn_dma *dma = sc->sc_dma;
1358	int i;
1359
1360	hifn_set_retry(sc);
1361
1362	/* initialize static pointer values */
1363	for (i = 0; i < HIFN_D_CMD_RSIZE; i++)
1364		dma->cmdr[i].p = htole32(sc->sc_dma_physaddr +
1365		    offsetof(struct hifn_dma, command_bufs[i][0]));
1366	for (i = 0; i < HIFN_D_RES_RSIZE; i++)
1367		dma->resr[i].p = htole32(sc->sc_dma_physaddr +
1368		    offsetof(struct hifn_dma, result_bufs[i][0]));
1369
1370	dma->cmdr[HIFN_D_CMD_RSIZE].p =
1371	    htole32(sc->sc_dma_physaddr + offsetof(struct hifn_dma, cmdr[0]));
1372	dma->srcr[HIFN_D_SRC_RSIZE].p =
1373	    htole32(sc->sc_dma_physaddr + offsetof(struct hifn_dma, srcr[0]));
1374	dma->dstr[HIFN_D_DST_RSIZE].p =
1375	    htole32(sc->sc_dma_physaddr + offsetof(struct hifn_dma, dstr[0]));
1376	dma->resr[HIFN_D_RES_RSIZE].p =
1377	    htole32(sc->sc_dma_physaddr + offsetof(struct hifn_dma, resr[0]));
1378
1379	dma->cmdu = dma->srcu = dma->dstu = dma->resu = 0;
1380	dma->cmdi = dma->srci = dma->dsti = dma->resi = 0;
1381	dma->cmdk = dma->srck = dma->dstk = dma->resk = 0;
1382}
1383
1384/*
1385 * Writes out the raw command buffer space.  Returns the
1386 * command buffer size.
1387 */
1388static u_int
1389hifn_write_command(struct hifn_command *cmd, u_int8_t *buf)
1390{
1391	u_int8_t *buf_pos;
1392	hifn_base_command_t *base_cmd;
1393	hifn_mac_command_t *mac_cmd;
1394	hifn_crypt_command_t *cry_cmd;
1395	int using_mac, using_crypt, len;
1396	u_int32_t dlen, slen;
1397
1398	buf_pos = buf;
1399	using_mac = cmd->base_masks & HIFN_BASE_CMD_MAC;
1400	using_crypt = cmd->base_masks & HIFN_BASE_CMD_CRYPT;
1401
1402	base_cmd = (hifn_base_command_t *)buf_pos;
1403	base_cmd->masks = htole16(cmd->base_masks);
1404	slen = cmd->src_mapsize;
1405	if (cmd->sloplen)
1406		dlen = cmd->dst_mapsize - cmd->sloplen + sizeof(u_int32_t);
1407	else
1408		dlen = cmd->dst_mapsize;
1409	base_cmd->total_source_count = htole16(slen & HIFN_BASE_CMD_LENMASK_LO);
1410	base_cmd->total_dest_count = htole16(dlen & HIFN_BASE_CMD_LENMASK_LO);
1411	dlen >>= 16;
1412	slen >>= 16;
1413	base_cmd->session_num = htole16(cmd->session_num |
1414	    ((slen << HIFN_BASE_CMD_SRCLEN_S) & HIFN_BASE_CMD_SRCLEN_M) |
1415	    ((dlen << HIFN_BASE_CMD_DSTLEN_S) & HIFN_BASE_CMD_DSTLEN_M));
1416	buf_pos += sizeof(hifn_base_command_t);
1417
1418	if (using_mac) {
1419		mac_cmd = (hifn_mac_command_t *)buf_pos;
1420		dlen = cmd->maccrd->crd_len;
1421		mac_cmd->source_count = htole16(dlen & 0xffff);
1422		dlen >>= 16;
1423		mac_cmd->masks = htole16(cmd->mac_masks |
1424		    ((dlen << HIFN_MAC_CMD_SRCLEN_S) & HIFN_MAC_CMD_SRCLEN_M));
1425		mac_cmd->header_skip = htole16(cmd->maccrd->crd_skip);
1426		mac_cmd->reserved = 0;
1427		buf_pos += sizeof(hifn_mac_command_t);
1428	}
1429
1430	if (using_crypt) {
1431		cry_cmd = (hifn_crypt_command_t *)buf_pos;
1432		dlen = cmd->enccrd->crd_len;
1433		cry_cmd->source_count = htole16(dlen & 0xffff);
1434		dlen >>= 16;
1435		cry_cmd->masks = htole16(cmd->cry_masks |
1436		    ((dlen << HIFN_CRYPT_CMD_SRCLEN_S) & HIFN_CRYPT_CMD_SRCLEN_M));
1437		cry_cmd->header_skip = htole16(cmd->enccrd->crd_skip);
1438		cry_cmd->reserved = 0;
1439		buf_pos += sizeof(hifn_crypt_command_t);
1440	}
1441
1442	if (using_mac && cmd->mac_masks & HIFN_MAC_CMD_NEW_KEY) {
1443		bcopy(cmd->mac, buf_pos, HIFN_MAC_KEY_LENGTH);
1444		buf_pos += HIFN_MAC_KEY_LENGTH;
1445	}
1446
1447	if (using_crypt && cmd->cry_masks & HIFN_CRYPT_CMD_NEW_KEY) {
1448		switch (cmd->cry_masks & HIFN_CRYPT_CMD_ALG_MASK) {
1449		case HIFN_CRYPT_CMD_ALG_3DES:
1450			bcopy(cmd->ck, buf_pos, HIFN_3DES_KEY_LENGTH);
1451			buf_pos += HIFN_3DES_KEY_LENGTH;
1452			break;
1453		case HIFN_CRYPT_CMD_ALG_DES:
1454			bcopy(cmd->ck, buf_pos, HIFN_DES_KEY_LENGTH);
1455			buf_pos += cmd->cklen;
1456			break;
1457		case HIFN_CRYPT_CMD_ALG_RC4:
1458			len = 256;
1459			do {
1460				int clen;
1461
1462				clen = MIN(cmd->cklen, len);
1463				bcopy(cmd->ck, buf_pos, clen);
1464				len -= clen;
1465				buf_pos += clen;
1466			} while (len > 0);
1467			bzero(buf_pos, 4);
1468			buf_pos += 4;
1469			break;
1470		}
1471	}
1472
1473	if (using_crypt && cmd->cry_masks & HIFN_CRYPT_CMD_NEW_IV) {
1474		bcopy(cmd->iv, buf_pos, HIFN_IV_LENGTH);
1475		buf_pos += HIFN_IV_LENGTH;
1476	}
1477
1478	if ((cmd->base_masks & (HIFN_BASE_CMD_MAC|HIFN_BASE_CMD_CRYPT)) == 0) {
1479		bzero(buf_pos, 8);
1480		buf_pos += 8;
1481	}
1482
1483	return (buf_pos - buf);
1484}
1485
1486static int
1487hifn_dmamap_aligned(struct hifn_operand *op)
1488{
1489	int i;
1490
1491	for (i = 0; i < op->nsegs; i++) {
1492		if (op->segs[i].ds_addr & 3)
1493			return (0);
1494		if ((i != (op->nsegs - 1)) && (op->segs[i].ds_len & 3))
1495			return (0);
1496	}
1497	return (1);
1498}
1499
1500static int
1501hifn_dmamap_load_dst(struct hifn_softc *sc, struct hifn_command *cmd)
1502{
1503	struct hifn_dma *dma = sc->sc_dma;
1504	struct hifn_operand *dst = &cmd->dst;
1505	u_int32_t p, l;
1506	int idx, used = 0, i;
1507
1508	idx = dma->dsti;
1509	for (i = 0; i < dst->nsegs - 1; i++) {
1510		dma->dstr[idx].p = htole32(dst->segs[i].ds_addr);
1511		dma->dstr[idx].l = htole32(HIFN_D_VALID |
1512		    HIFN_D_MASKDONEIRQ | dst->segs[i].ds_len);
1513		HIFN_DSTR_SYNC(sc, idx,
1514		    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1515		used++;
1516
1517		if (++idx == HIFN_D_DST_RSIZE) {
1518			dma->dstr[idx].l = htole32(HIFN_D_VALID |
1519			    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1520			HIFN_DSTR_SYNC(sc, idx,
1521			    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1522			idx = 0;
1523		}
1524	}
1525
1526	if (cmd->sloplen == 0) {
1527		p = dst->segs[i].ds_addr;
1528		l = HIFN_D_VALID | HIFN_D_MASKDONEIRQ | HIFN_D_LAST |
1529		    dst->segs[i].ds_len;
1530	} else {
1531		p = sc->sc_dma_physaddr +
1532		    offsetof(struct hifn_dma, slop[cmd->slopidx]);
1533		l = HIFN_D_VALID | HIFN_D_MASKDONEIRQ | HIFN_D_LAST |
1534		    sizeof(u_int32_t);
1535
1536		if ((dst->segs[i].ds_len - cmd->sloplen) != 0) {
1537			dma->dstr[idx].p = htole32(dst->segs[i].ds_addr);
1538			dma->dstr[idx].l = htole32(HIFN_D_VALID |
1539			    HIFN_D_MASKDONEIRQ |
1540			    (dst->segs[i].ds_len - cmd->sloplen));
1541			HIFN_DSTR_SYNC(sc, idx,
1542			    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1543			used++;
1544
1545			if (++idx == HIFN_D_DST_RSIZE) {
1546				dma->dstr[idx].l = htole32(HIFN_D_VALID |
1547				    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1548				HIFN_DSTR_SYNC(sc, idx,
1549				    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1550				idx = 0;
1551			}
1552		}
1553	}
1554	dma->dstr[idx].p = htole32(p);
1555	dma->dstr[idx].l = htole32(l);
1556	HIFN_DSTR_SYNC(sc, idx, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1557	used++;
1558
1559	if (++idx == HIFN_D_DST_RSIZE) {
1560		dma->dstr[idx].l = htole32(HIFN_D_VALID | HIFN_D_JUMP |
1561		    HIFN_D_MASKDONEIRQ);
1562		HIFN_DSTR_SYNC(sc, idx,
1563		    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1564		idx = 0;
1565	}
1566
1567	dma->dsti = idx;
1568	dma->dstu += used;
1569	return (idx);
1570}
1571
1572static int
1573hifn_dmamap_load_src(struct hifn_softc *sc, struct hifn_command *cmd)
1574{
1575	struct hifn_dma *dma = sc->sc_dma;
1576	struct hifn_operand *src = &cmd->src;
1577	int idx, i;
1578	u_int32_t last = 0;
1579
1580	idx = dma->srci;
1581	for (i = 0; i < src->nsegs; i++) {
1582		if (i == src->nsegs - 1)
1583			last = HIFN_D_LAST;
1584
1585		dma->srcr[idx].p = htole32(src->segs[i].ds_addr);
1586		dma->srcr[idx].l = htole32(src->segs[i].ds_len |
1587		    HIFN_D_VALID | HIFN_D_MASKDONEIRQ | last);
1588		HIFN_SRCR_SYNC(sc, idx,
1589		    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1590
1591		if (++idx == HIFN_D_SRC_RSIZE) {
1592			dma->srcr[idx].l = htole32(HIFN_D_VALID |
1593			    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1594			HIFN_SRCR_SYNC(sc, HIFN_D_SRC_RSIZE,
1595			    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1596			idx = 0;
1597		}
1598	}
1599	dma->srci = idx;
1600	dma->srcu += src->nsegs;
1601	return (idx);
1602}
1603
1604static void
1605hifn_op_cb(void* arg, bus_dma_segment_t *seg, int nsegs, bus_size_t mapsize, int error)
1606{
1607	struct hifn_operand *op = arg;
1608
1609	KASSERT(nsegs <= MAX_SCATTER,
1610		("hifn_op_cb: too many DMA segments (%u > %u) "
1611		 "returned when mapping operand", nsegs, MAX_SCATTER));
1612	op->mapsize = mapsize;
1613	op->nsegs = nsegs;
1614	bcopy(seg, op->segs, nsegs * sizeof (seg[0]));
1615}
1616
1617static int
1618hifn_crypto(
1619	struct hifn_softc *sc,
1620	struct hifn_command *cmd,
1621	struct cryptop *crp,
1622	int hint)
1623{
1624	struct	hifn_dma *dma = sc->sc_dma;
1625	u_int32_t cmdlen;
1626	int cmdi, resi, err = 0;
1627
1628	/*
1629	 * need 1 cmd, and 1 res
1630	 *
1631	 * NB: check this first since it's easy.
1632	 */
1633	HIFN_LOCK(sc);
1634	if ((dma->cmdu + 1) > HIFN_D_CMD_RSIZE ||
1635	    (dma->resu + 1) > HIFN_D_RES_RSIZE) {
1636#ifdef HIFN_DEBUG
1637		if (hifn_debug) {
1638			device_printf(sc->sc_dev,
1639				"cmd/result exhaustion, cmdu %u resu %u\n",
1640				dma->cmdu, dma->resu);
1641		}
1642#endif
1643		hifnstats.hst_nomem_cr++;
1644		HIFN_UNLOCK(sc);
1645		return (ERESTART);
1646	}
1647
1648	if (bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT, &cmd->src_map)) {
1649		hifnstats.hst_nomem_map++;
1650		HIFN_UNLOCK(sc);
1651		return (ENOMEM);
1652	}
1653
1654	if (crp->crp_flags & CRYPTO_F_IMBUF) {
1655		if (bus_dmamap_load_mbuf(sc->sc_dmat, cmd->src_map,
1656		    cmd->src_m, hifn_op_cb, &cmd->src, BUS_DMA_NOWAIT)) {
1657			hifnstats.hst_nomem_load++;
1658			err = ENOMEM;
1659			goto err_srcmap1;
1660		}
1661	} else if (crp->crp_flags & CRYPTO_F_IOV) {
1662		if (bus_dmamap_load_uio(sc->sc_dmat, cmd->src_map,
1663		    cmd->src_io, hifn_op_cb, &cmd->src, BUS_DMA_NOWAIT)) {
1664			hifnstats.hst_nomem_load++;
1665			err = ENOMEM;
1666			goto err_srcmap1;
1667		}
1668	} else {
1669		err = EINVAL;
1670		goto err_srcmap1;
1671	}
1672
1673	if (hifn_dmamap_aligned(&cmd->src)) {
1674		cmd->sloplen = cmd->src_mapsize & 3;
1675		cmd->dst = cmd->src;
1676	} else {
1677		if (crp->crp_flags & CRYPTO_F_IOV) {
1678			err = EINVAL;
1679			goto err_srcmap;
1680		} else if (crp->crp_flags & CRYPTO_F_IMBUF) {
1681			int totlen, len;
1682			struct mbuf *m, *m0, *mlast;
1683
1684			KASSERT(cmd->dst_m == cmd->src_m,
1685				("hifn_crypto: dst_m initialized improperly"));
1686			hifnstats.hst_unaligned++;
1687			/*
1688			 * Source is not aligned on a longword boundary.
1689			 * Copy the data to insure alignment.  If we fail
1690			 * to allocate mbufs or clusters while doing this
1691			 * we return ERESTART so the operation is requeued
1692			 * at the crypto later, but only if there are
1693			 * ops already posted to the hardware; otherwise we
1694			 * have no guarantee that we'll be re-entered.
1695			 */
1696			totlen = cmd->src_mapsize;
1697			if (cmd->src_m->m_flags & M_PKTHDR) {
1698				len = MHLEN;
1699				MGETHDR(m0, M_DONTWAIT, MT_DATA);
1700				if (m0 && !m_dup_pkthdr(m0, cmd->src_m, M_DONTWAIT)) {
1701					m_free(m0);
1702					m0 = NULL;
1703				}
1704			} else {
1705				len = MLEN;
1706				MGET(m0, M_DONTWAIT, MT_DATA);
1707			}
1708			if (m0 == NULL) {
1709				hifnstats.hst_nomem_mbuf++;
1710				err = dma->cmdu ? ERESTART : ENOMEM;
1711				goto err_srcmap;
1712			}
1713			if (totlen >= MINCLSIZE) {
1714				MCLGET(m0, M_DONTWAIT);
1715				if ((m0->m_flags & M_EXT) == 0) {
1716					hifnstats.hst_nomem_mcl++;
1717					err = dma->cmdu ? ERESTART : ENOMEM;
1718					m_freem(m0);
1719					goto err_srcmap;
1720				}
1721				len = MCLBYTES;
1722			}
1723			totlen -= len;
1724			m0->m_pkthdr.len = m0->m_len = len;
1725			mlast = m0;
1726
1727			while (totlen > 0) {
1728				MGET(m, M_DONTWAIT, MT_DATA);
1729				if (m == NULL) {
1730					hifnstats.hst_nomem_mbuf++;
1731					err = dma->cmdu ? ERESTART : ENOMEM;
1732					m_freem(m0);
1733					goto err_srcmap;
1734				}
1735				len = MLEN;
1736				if (totlen >= MINCLSIZE) {
1737					MCLGET(m, M_DONTWAIT);
1738					if ((m->m_flags & M_EXT) == 0) {
1739						hifnstats.hst_nomem_mcl++;
1740						err = dma->cmdu ? ERESTART : ENOMEM;
1741						mlast->m_next = m;
1742						m_freem(m0);
1743						goto err_srcmap;
1744					}
1745					len = MCLBYTES;
1746				}
1747
1748				m->m_len = len;
1749				m0->m_pkthdr.len += len;
1750				totlen -= len;
1751
1752				mlast->m_next = m;
1753				mlast = m;
1754			}
1755			cmd->dst_m = m0;
1756		}
1757	}
1758
1759	if (cmd->dst_map == NULL) {
1760		if (bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT, &cmd->dst_map)) {
1761			hifnstats.hst_nomem_map++;
1762			err = ENOMEM;
1763			goto err_srcmap;
1764		}
1765		if (crp->crp_flags & CRYPTO_F_IMBUF) {
1766			if (bus_dmamap_load_mbuf(sc->sc_dmat, cmd->dst_map,
1767			    cmd->dst_m, hifn_op_cb, &cmd->dst, BUS_DMA_NOWAIT)) {
1768				hifnstats.hst_nomem_map++;
1769				err = ENOMEM;
1770				goto err_dstmap1;
1771			}
1772		} else if (crp->crp_flags & CRYPTO_F_IOV) {
1773			if (bus_dmamap_load_uio(sc->sc_dmat, cmd->dst_map,
1774			    cmd->dst_io, hifn_op_cb, &cmd->dst, BUS_DMA_NOWAIT)) {
1775				hifnstats.hst_nomem_load++;
1776				err = ENOMEM;
1777				goto err_dstmap1;
1778			}
1779		}
1780	}
1781
1782#ifdef HIFN_DEBUG
1783	if (hifn_debug) {
1784		device_printf(sc->sc_dev,
1785		    "Entering cmd: stat %8x ien %8x u %d/%d/%d/%d n %d/%d\n",
1786		    READ_REG_1(sc, HIFN_1_DMA_CSR),
1787		    READ_REG_1(sc, HIFN_1_DMA_IER),
1788		    dma->cmdu, dma->srcu, dma->dstu, dma->resu,
1789		    cmd->src_nsegs, cmd->dst_nsegs);
1790	}
1791#endif
1792
1793	if (cmd->src_map == cmd->dst_map) {
1794		bus_dmamap_sync(sc->sc_dmat, cmd->src_map,
1795		    BUS_DMASYNC_PREWRITE|BUS_DMASYNC_PREREAD);
1796	} else {
1797		bus_dmamap_sync(sc->sc_dmat, cmd->src_map,
1798		    BUS_DMASYNC_PREWRITE);
1799		bus_dmamap_sync(sc->sc_dmat, cmd->dst_map,
1800		    BUS_DMASYNC_PREREAD);
1801	}
1802
1803	/*
1804	 * need N src, and N dst
1805	 */
1806	if ((dma->srcu + cmd->src_nsegs) > HIFN_D_SRC_RSIZE ||
1807	    (dma->dstu + cmd->dst_nsegs + 1) > HIFN_D_DST_RSIZE) {
1808#ifdef HIFN_DEBUG
1809		if (hifn_debug) {
1810			device_printf(sc->sc_dev,
1811				"src/dst exhaustion, srcu %u+%u dstu %u+%u\n",
1812				dma->srcu, cmd->src_nsegs,
1813				dma->dstu, cmd->dst_nsegs);
1814		}
1815#endif
1816		hifnstats.hst_nomem_sd++;
1817		err = ERESTART;
1818		goto err_dstmap;
1819	}
1820
1821	if (dma->cmdi == HIFN_D_CMD_RSIZE) {
1822		dma->cmdi = 0;
1823		dma->cmdr[HIFN_D_CMD_RSIZE].l = htole32(HIFN_D_VALID |
1824		    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1825		HIFN_CMDR_SYNC(sc, HIFN_D_CMD_RSIZE,
1826		    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1827	}
1828	cmdi = dma->cmdi++;
1829	cmdlen = hifn_write_command(cmd, dma->command_bufs[cmdi]);
1830	HIFN_CMD_SYNC(sc, cmdi, BUS_DMASYNC_PREWRITE);
1831
1832	/* .p for command/result already set */
1833	dma->cmdr[cmdi].l = htole32(cmdlen | HIFN_D_VALID | HIFN_D_LAST |
1834	    HIFN_D_MASKDONEIRQ);
1835	HIFN_CMDR_SYNC(sc, cmdi,
1836	    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
1837	dma->cmdu++;
1838	if (sc->sc_c_busy == 0) {
1839		WRITE_REG_1(sc, HIFN_1_DMA_CSR, HIFN_DMACSR_C_CTRL_ENA);
1840		sc->sc_c_busy = 1;
1841	}
1842
1843	/*
1844	 * We don't worry about missing an interrupt (which a "command wait"
1845	 * interrupt salvages us from), unless there is more than one command
1846	 * in the queue.
1847	 */
1848	if (dma->cmdu > 1) {
1849		sc->sc_dmaier |= HIFN_DMAIER_C_WAIT;
1850		WRITE_REG_1(sc, HIFN_1_DMA_IER, sc->sc_dmaier);
1851	}
1852
1853	hifnstats.hst_ipackets++;
1854	hifnstats.hst_ibytes += cmd->src_mapsize;
1855
1856	hifn_dmamap_load_src(sc, cmd);
1857	if (sc->sc_s_busy == 0) {
1858		WRITE_REG_1(sc, HIFN_1_DMA_CSR, HIFN_DMACSR_S_CTRL_ENA);
1859		sc->sc_s_busy = 1;
1860	}
1861
1862	/*
1863	 * Unlike other descriptors, we don't mask done interrupt from
1864	 * result descriptor.
1865	 */
1866#ifdef HIFN_DEBUG
1867	if (hifn_debug)
1868		printf("load res\n");
1869#endif
1870	if (dma->resi == HIFN_D_RES_RSIZE) {
1871		dma->resi = 0;
1872		dma->resr[HIFN_D_RES_RSIZE].l = htole32(HIFN_D_VALID |
1873		    HIFN_D_JUMP | HIFN_D_MASKDONEIRQ);
1874		HIFN_RESR_SYNC(sc, HIFN_D_RES_RSIZE,
1875		    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1876	}
1877	resi = dma->resi++;
1878	KASSERT(dma->hifn_commands[resi] == NULL,
1879		("hifn_crypto: command slot %u busy", resi));
1880	dma->hifn_commands[resi] = cmd;
1881	HIFN_RES_SYNC(sc, resi, BUS_DMASYNC_PREREAD);
1882	if ((hint & CRYPTO_HINT_MORE) && sc->sc_curbatch < hifn_maxbatch) {
1883		dma->resr[resi].l = htole32(HIFN_MAX_RESULT |
1884		    HIFN_D_VALID | HIFN_D_LAST | HIFN_D_MASKDONEIRQ);
1885		sc->sc_curbatch++;
1886		if (sc->sc_curbatch > hifnstats.hst_maxbatch)
1887			hifnstats.hst_maxbatch = sc->sc_curbatch;
1888		hifnstats.hst_totbatch++;
1889	} else {
1890		dma->resr[resi].l = htole32(HIFN_MAX_RESULT |
1891		    HIFN_D_VALID | HIFN_D_LAST);
1892		sc->sc_curbatch = 0;
1893	}
1894	HIFN_RESR_SYNC(sc, resi,
1895	    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1896	dma->resu++;
1897	if (sc->sc_r_busy == 0) {
1898		WRITE_REG_1(sc, HIFN_1_DMA_CSR, HIFN_DMACSR_R_CTRL_ENA);
1899		sc->sc_r_busy = 1;
1900	}
1901
1902	if (cmd->sloplen)
1903		cmd->slopidx = resi;
1904
1905	hifn_dmamap_load_dst(sc, cmd);
1906
1907	if (sc->sc_d_busy == 0) {
1908		WRITE_REG_1(sc, HIFN_1_DMA_CSR, HIFN_DMACSR_D_CTRL_ENA);
1909		sc->sc_d_busy = 1;
1910	}
1911
1912#ifdef HIFN_DEBUG
1913	if (hifn_debug) {
1914		device_printf(sc->sc_dev, "command: stat %8x ier %8x\n",
1915		    READ_REG_1(sc, HIFN_1_DMA_CSR),
1916		    READ_REG_1(sc, HIFN_1_DMA_IER));
1917	}
1918#endif
1919
1920	sc->sc_active = 5;
1921	HIFN_UNLOCK(sc);
1922	KASSERT(err == 0, ("hifn_crypto: success with error %u", err));
1923	return (err);		/* success */
1924
1925err_dstmap:
1926	if (cmd->src_map != cmd->dst_map)
1927		bus_dmamap_unload(sc->sc_dmat, cmd->dst_map);
1928err_dstmap1:
1929	if (cmd->src_map != cmd->dst_map)
1930		bus_dmamap_destroy(sc->sc_dmat, cmd->dst_map);
1931err_srcmap:
1932	if (crp->crp_flags & CRYPTO_F_IMBUF) {
1933		if (cmd->src_m != cmd->dst_m)
1934			m_freem(cmd->dst_m);
1935	}
1936	bus_dmamap_unload(sc->sc_dmat, cmd->src_map);
1937err_srcmap1:
1938	bus_dmamap_destroy(sc->sc_dmat, cmd->src_map);
1939	HIFN_UNLOCK(sc);
1940	return (err);
1941}
1942
1943static void
1944hifn_tick(void* vsc)
1945{
1946	struct hifn_softc *sc = vsc;
1947
1948	HIFN_LOCK(sc);
1949	if (sc->sc_active == 0) {
1950		struct hifn_dma *dma = sc->sc_dma;
1951		u_int32_t r = 0;
1952
1953		if (dma->cmdu == 0 && sc->sc_c_busy) {
1954			sc->sc_c_busy = 0;
1955			r |= HIFN_DMACSR_C_CTRL_DIS;
1956		}
1957		if (dma->srcu == 0 && sc->sc_s_busy) {
1958			sc->sc_s_busy = 0;
1959			r |= HIFN_DMACSR_S_CTRL_DIS;
1960		}
1961		if (dma->dstu == 0 && sc->sc_d_busy) {
1962			sc->sc_d_busy = 0;
1963			r |= HIFN_DMACSR_D_CTRL_DIS;
1964		}
1965		if (dma->resu == 0 && sc->sc_r_busy) {
1966			sc->sc_r_busy = 0;
1967			r |= HIFN_DMACSR_R_CTRL_DIS;
1968		}
1969		if (r)
1970			WRITE_REG_1(sc, HIFN_1_DMA_CSR, r);
1971	} else
1972		sc->sc_active--;
1973	HIFN_UNLOCK(sc);
1974	callout_reset(&sc->sc_tickto, hz, hifn_tick, sc);
1975}
1976
1977static void
1978hifn_intr(void *arg)
1979{
1980	struct hifn_softc *sc = arg;
1981	struct hifn_dma *dma;
1982	u_int32_t dmacsr, restart;
1983	int i, u;
1984
1985	dmacsr = READ_REG_1(sc, HIFN_1_DMA_CSR);
1986
1987	/* Nothing in the DMA unit interrupted */
1988	if ((dmacsr & sc->sc_dmaier) == 0)
1989		return;
1990
1991	HIFN_LOCK(sc);
1992
1993	dma = sc->sc_dma;
1994
1995#ifdef HIFN_DEBUG
1996	if (hifn_debug) {
1997		device_printf(sc->sc_dev,
1998		    "irq: stat %08x ien %08x damier %08x i %d/%d/%d/%d k %d/%d/%d/%d u %d/%d/%d/%d\n",
1999		    dmacsr, READ_REG_1(sc, HIFN_1_DMA_IER), sc->sc_dmaier,
2000		    dma->cmdi, dma->srci, dma->dsti, dma->resi,
2001		    dma->cmdk, dma->srck, dma->dstk, dma->resk,
2002		    dma->cmdu, dma->srcu, dma->dstu, dma->resu);
2003	}
2004#endif
2005
2006	WRITE_REG_1(sc, HIFN_1_DMA_CSR, dmacsr & sc->sc_dmaier);
2007
2008	if ((sc->sc_flags & HIFN_HAS_PUBLIC) &&
2009	    (dmacsr & HIFN_DMACSR_PUBDONE))
2010		WRITE_REG_1(sc, HIFN_1_PUB_STATUS,
2011		    READ_REG_1(sc, HIFN_1_PUB_STATUS) | HIFN_PUBSTS_DONE);
2012
2013	restart = dmacsr & (HIFN_DMACSR_D_OVER | HIFN_DMACSR_R_OVER);
2014	if (restart)
2015		device_printf(sc->sc_dev, "overrun %x\n", dmacsr);
2016
2017	if (sc->sc_flags & HIFN_IS_7811) {
2018		if (dmacsr & HIFN_DMACSR_ILLR)
2019			device_printf(sc->sc_dev, "illegal read\n");
2020		if (dmacsr & HIFN_DMACSR_ILLW)
2021			device_printf(sc->sc_dev, "illegal write\n");
2022	}
2023
2024	restart = dmacsr & (HIFN_DMACSR_C_ABORT | HIFN_DMACSR_S_ABORT |
2025	    HIFN_DMACSR_D_ABORT | HIFN_DMACSR_R_ABORT);
2026	if (restart) {
2027		device_printf(sc->sc_dev, "abort, resetting.\n");
2028		hifnstats.hst_abort++;
2029		hifn_abort(sc);
2030		HIFN_UNLOCK(sc);
2031		return;
2032	}
2033
2034	if ((dmacsr & HIFN_DMACSR_C_WAIT) && (dma->cmdu == 0)) {
2035		/*
2036		 * If no slots to process and we receive a "waiting on
2037		 * command" interrupt, we disable the "waiting on command"
2038		 * (by clearing it).
2039		 */
2040		sc->sc_dmaier &= ~HIFN_DMAIER_C_WAIT;
2041		WRITE_REG_1(sc, HIFN_1_DMA_IER, sc->sc_dmaier);
2042	}
2043
2044	/* clear the rings */
2045	i = dma->resk; u = dma->resu;
2046	while (u != 0) {
2047		HIFN_RESR_SYNC(sc, i,
2048		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2049		if (dma->resr[i].l & htole32(HIFN_D_VALID)) {
2050			HIFN_RESR_SYNC(sc, i,
2051			    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2052			break;
2053		}
2054
2055		if (i != HIFN_D_RES_RSIZE) {
2056			struct hifn_command *cmd;
2057			u_int8_t *macbuf = NULL;
2058
2059			HIFN_RES_SYNC(sc, i, BUS_DMASYNC_POSTREAD);
2060			cmd = dma->hifn_commands[i];
2061			KASSERT(cmd != NULL,
2062				("hifn_intr: null command slot %u", i));
2063			dma->hifn_commands[i] = NULL;
2064
2065			if (cmd->base_masks & HIFN_BASE_CMD_MAC) {
2066				macbuf = dma->result_bufs[i];
2067				macbuf += 12;
2068			}
2069
2070			hifn_callback(sc, cmd, macbuf);
2071			hifnstats.hst_opackets++;
2072			u--;
2073		}
2074
2075		if (++i == (HIFN_D_RES_RSIZE + 1))
2076			i = 0;
2077	}
2078	dma->resk = i; dma->resu = u;
2079
2080	i = dma->srck; u = dma->srcu;
2081	while (u != 0) {
2082		if (i == HIFN_D_SRC_RSIZE)
2083			i = 0;
2084		HIFN_SRCR_SYNC(sc, i,
2085		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2086		if (dma->srcr[i].l & htole32(HIFN_D_VALID)) {
2087			HIFN_SRCR_SYNC(sc, i,
2088			    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2089			break;
2090		}
2091		i++, u--;
2092	}
2093	dma->srck = i; dma->srcu = u;
2094
2095	i = dma->cmdk; u = dma->cmdu;
2096	while (u != 0) {
2097		HIFN_CMDR_SYNC(sc, i,
2098		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2099		if (dma->cmdr[i].l & htole32(HIFN_D_VALID)) {
2100			HIFN_CMDR_SYNC(sc, i,
2101			    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2102			break;
2103		}
2104		if (i != HIFN_D_CMD_RSIZE) {
2105			u--;
2106			HIFN_CMD_SYNC(sc, i, BUS_DMASYNC_POSTWRITE);
2107		}
2108		if (++i == (HIFN_D_CMD_RSIZE + 1))
2109			i = 0;
2110	}
2111	dma->cmdk = i; dma->cmdu = u;
2112
2113	HIFN_UNLOCK(sc);
2114
2115	if (sc->sc_needwakeup) {		/* XXX check high watermark */
2116		int wakeup = sc->sc_needwakeup & (CRYPTO_SYMQ|CRYPTO_ASYMQ);
2117#ifdef HIFN_DEBUG
2118		if (hifn_debug)
2119			device_printf(sc->sc_dev,
2120				"wakeup crypto (%x) u %d/%d/%d/%d\n",
2121				sc->sc_needwakeup,
2122				dma->cmdu, dma->srcu, dma->dstu, dma->resu);
2123#endif
2124		sc->sc_needwakeup &= ~wakeup;
2125		crypto_unblock(sc->sc_cid, wakeup);
2126	}
2127}
2128
2129/*
2130 * Allocate a new 'session' and return an encoded session id.  'sidp'
2131 * contains our registration id, and should contain an encoded session
2132 * id on successful allocation.
2133 */
2134static int
2135hifn_newsession(void *arg, u_int32_t *sidp, struct cryptoini *cri)
2136{
2137	struct cryptoini *c;
2138	struct hifn_softc *sc = arg;
2139	int i, mac = 0, cry = 0;
2140
2141	KASSERT(sc != NULL, ("hifn_newsession: null softc"));
2142	if (sidp == NULL || cri == NULL || sc == NULL)
2143		return (EINVAL);
2144
2145	for (i = 0; i < sc->sc_maxses; i++)
2146		if (sc->sc_sessions[i].hs_state == HS_STATE_FREE)
2147			break;
2148	if (i == sc->sc_maxses)
2149		return (ENOMEM);
2150
2151	for (c = cri; c != NULL; c = c->cri_next) {
2152		switch (c->cri_alg) {
2153		case CRYPTO_MD5:
2154		case CRYPTO_SHA1:
2155		case CRYPTO_MD5_HMAC:
2156		case CRYPTO_SHA1_HMAC:
2157			if (mac)
2158				return (EINVAL);
2159			mac = 1;
2160			break;
2161		case CRYPTO_DES_CBC:
2162		case CRYPTO_3DES_CBC:
2163			/* XXX this may read fewer, does it matter? */
2164			read_random(sc->sc_sessions[i].hs_iv, HIFN_IV_LENGTH);
2165			/*FALLTHROUGH*/
2166		case CRYPTO_ARC4:
2167			if (cry)
2168				return (EINVAL);
2169			cry = 1;
2170			break;
2171		default:
2172			return (EINVAL);
2173		}
2174	}
2175	if (mac == 0 && cry == 0)
2176		return (EINVAL);
2177
2178	*sidp = HIFN_SID(device_get_unit(sc->sc_dev), i);
2179	sc->sc_sessions[i].hs_state = HS_STATE_USED;
2180
2181	return (0);
2182}
2183
2184/*
2185 * Deallocate a session.
2186 * XXX this routine should run a zero'd mac/encrypt key into context ram.
2187 * XXX to blow away any keys already stored there.
2188 */
2189static int
2190hifn_freesession(void *arg, u_int64_t tid)
2191{
2192	struct hifn_softc *sc = arg;
2193	int session;
2194	u_int32_t sid = CRYPTO_SESID2LID(tid);
2195
2196	KASSERT(sc != NULL, ("hifn_freesession: null softc"));
2197	if (sc == NULL)
2198		return (EINVAL);
2199
2200	session = HIFN_SESSION(sid);
2201	if (session >= sc->sc_maxses)
2202		return (EINVAL);
2203
2204	bzero(&sc->sc_sessions[session], sizeof(sc->sc_sessions[session]));
2205	return (0);
2206}
2207
2208static int
2209hifn_process(void *arg, struct cryptop *crp, int hint)
2210{
2211	struct hifn_softc *sc = arg;
2212	struct hifn_command *cmd = NULL;
2213	int session, err;
2214	struct cryptodesc *crd1, *crd2, *maccrd, *enccrd;
2215
2216	if (crp == NULL || crp->crp_callback == NULL) {
2217		hifnstats.hst_invalid++;
2218		return (EINVAL);
2219	}
2220	session = HIFN_SESSION(crp->crp_sid);
2221
2222	if (sc == NULL || session >= sc->sc_maxses) {
2223		err = EINVAL;
2224		goto errout;
2225	}
2226
2227	cmd = malloc(sizeof(struct hifn_command), M_DEVBUF, M_NOWAIT | M_ZERO);
2228	if (cmd == NULL) {
2229		hifnstats.hst_nomem++;
2230		err = ENOMEM;
2231		goto errout;
2232	}
2233
2234	if (crp->crp_flags & CRYPTO_F_IMBUF) {
2235		cmd->src_m = (struct mbuf *)crp->crp_buf;
2236		cmd->dst_m = (struct mbuf *)crp->crp_buf;
2237	} else if (crp->crp_flags & CRYPTO_F_IOV) {
2238		cmd->src_io = (struct uio *)crp->crp_buf;
2239		cmd->dst_io = (struct uio *)crp->crp_buf;
2240	} else {
2241		err = EINVAL;
2242		goto errout;	/* XXX we don't handle contiguous buffers! */
2243	}
2244
2245	crd1 = crp->crp_desc;
2246	if (crd1 == NULL) {
2247		err = EINVAL;
2248		goto errout;
2249	}
2250	crd2 = crd1->crd_next;
2251
2252	if (crd2 == NULL) {
2253		if (crd1->crd_alg == CRYPTO_MD5_HMAC ||
2254		    crd1->crd_alg == CRYPTO_SHA1_HMAC ||
2255		    crd1->crd_alg == CRYPTO_SHA1 ||
2256		    crd1->crd_alg == CRYPTO_MD5) {
2257			maccrd = crd1;
2258			enccrd = NULL;
2259		} else if (crd1->crd_alg == CRYPTO_DES_CBC ||
2260		    crd1->crd_alg == CRYPTO_3DES_CBC ||
2261		    crd1->crd_alg == CRYPTO_ARC4) {
2262			if ((crd1->crd_flags & CRD_F_ENCRYPT) == 0)
2263				cmd->base_masks |= HIFN_BASE_CMD_DECODE;
2264			maccrd = NULL;
2265			enccrd = crd1;
2266		} else {
2267			err = EINVAL;
2268			goto errout;
2269		}
2270	} else {
2271		if ((crd1->crd_alg == CRYPTO_MD5_HMAC ||
2272                     crd1->crd_alg == CRYPTO_SHA1_HMAC ||
2273                     crd1->crd_alg == CRYPTO_MD5 ||
2274                     crd1->crd_alg == CRYPTO_SHA1) &&
2275		    (crd2->crd_alg == CRYPTO_DES_CBC ||
2276		     crd2->crd_alg == CRYPTO_3DES_CBC ||
2277		     crd2->crd_alg == CRYPTO_ARC4) &&
2278		    ((crd2->crd_flags & CRD_F_ENCRYPT) == 0)) {
2279			cmd->base_masks = HIFN_BASE_CMD_DECODE;
2280			maccrd = crd1;
2281			enccrd = crd2;
2282		} else if ((crd1->crd_alg == CRYPTO_DES_CBC ||
2283		     crd1->crd_alg == CRYPTO_ARC4 ||
2284		     crd1->crd_alg == CRYPTO_3DES_CBC) &&
2285		    (crd2->crd_alg == CRYPTO_MD5_HMAC ||
2286                     crd2->crd_alg == CRYPTO_SHA1_HMAC ||
2287                     crd2->crd_alg == CRYPTO_MD5 ||
2288                     crd2->crd_alg == CRYPTO_SHA1) &&
2289		    (crd1->crd_flags & CRD_F_ENCRYPT)) {
2290			enccrd = crd1;
2291			maccrd = crd2;
2292		} else {
2293			/*
2294			 * We cannot order the 7751 as requested
2295			 */
2296			err = EINVAL;
2297			goto errout;
2298		}
2299	}
2300
2301	if (enccrd) {
2302		cmd->enccrd = enccrd;
2303		cmd->base_masks |= HIFN_BASE_CMD_CRYPT;
2304		switch (enccrd->crd_alg) {
2305		case CRYPTO_ARC4:
2306			cmd->cry_masks |= HIFN_CRYPT_CMD_ALG_RC4;
2307			if ((enccrd->crd_flags & CRD_F_ENCRYPT)
2308			    != sc->sc_sessions[session].hs_prev_op)
2309				sc->sc_sessions[session].hs_state =
2310				    HS_STATE_USED;
2311			break;
2312		case CRYPTO_DES_CBC:
2313			cmd->cry_masks |= HIFN_CRYPT_CMD_ALG_DES |
2314			    HIFN_CRYPT_CMD_MODE_CBC |
2315			    HIFN_CRYPT_CMD_NEW_IV;
2316			break;
2317		case CRYPTO_3DES_CBC:
2318			cmd->cry_masks |= HIFN_CRYPT_CMD_ALG_3DES |
2319			    HIFN_CRYPT_CMD_MODE_CBC |
2320			    HIFN_CRYPT_CMD_NEW_IV;
2321			break;
2322		default:
2323			err = EINVAL;
2324			goto errout;
2325		}
2326		if (enccrd->crd_alg != CRYPTO_ARC4) {
2327			if (enccrd->crd_flags & CRD_F_ENCRYPT) {
2328				if (enccrd->crd_flags & CRD_F_IV_EXPLICIT)
2329					bcopy(enccrd->crd_iv, cmd->iv,
2330					    HIFN_IV_LENGTH);
2331				else
2332					bcopy(sc->sc_sessions[session].hs_iv,
2333					    cmd->iv, HIFN_IV_LENGTH);
2334
2335				if ((enccrd->crd_flags & CRD_F_IV_PRESENT)
2336				    == 0) {
2337					if (crp->crp_flags & CRYPTO_F_IMBUF)
2338						m_copyback(cmd->src_m,
2339						    enccrd->crd_inject,
2340						    HIFN_IV_LENGTH, cmd->iv);
2341					else if (crp->crp_flags & CRYPTO_F_IOV)
2342						cuio_copyback(cmd->src_io,
2343						    enccrd->crd_inject,
2344						    HIFN_IV_LENGTH, cmd->iv);
2345				}
2346			} else {
2347				if (enccrd->crd_flags & CRD_F_IV_EXPLICIT)
2348					bcopy(enccrd->crd_iv, cmd->iv,
2349					    HIFN_IV_LENGTH);
2350				else if (crp->crp_flags & CRYPTO_F_IMBUF)
2351					m_copydata(cmd->src_m,
2352					    enccrd->crd_inject,
2353					    HIFN_IV_LENGTH, cmd->iv);
2354				else if (crp->crp_flags & CRYPTO_F_IOV)
2355					cuio_copydata(cmd->src_io,
2356					    enccrd->crd_inject,
2357					    HIFN_IV_LENGTH, cmd->iv);
2358			}
2359		}
2360
2361		cmd->ck = enccrd->crd_key;
2362		cmd->cklen = enccrd->crd_klen >> 3;
2363
2364		if (sc->sc_sessions[session].hs_state == HS_STATE_USED)
2365			cmd->cry_masks |= HIFN_CRYPT_CMD_NEW_KEY;
2366	}
2367
2368	if (maccrd) {
2369		cmd->maccrd = maccrd;
2370		cmd->base_masks |= HIFN_BASE_CMD_MAC;
2371
2372		switch (maccrd->crd_alg) {
2373		case CRYPTO_MD5:
2374			cmd->mac_masks |= HIFN_MAC_CMD_ALG_MD5 |
2375			    HIFN_MAC_CMD_RESULT | HIFN_MAC_CMD_MODE_HASH |
2376			    HIFN_MAC_CMD_POS_IPSEC;
2377                       break;
2378		case CRYPTO_MD5_HMAC:
2379			cmd->mac_masks |= HIFN_MAC_CMD_ALG_MD5 |
2380			    HIFN_MAC_CMD_RESULT | HIFN_MAC_CMD_MODE_HMAC |
2381			    HIFN_MAC_CMD_POS_IPSEC | HIFN_MAC_CMD_TRUNC;
2382			break;
2383		case CRYPTO_SHA1:
2384			cmd->mac_masks |= HIFN_MAC_CMD_ALG_SHA1 |
2385			    HIFN_MAC_CMD_RESULT | HIFN_MAC_CMD_MODE_HASH |
2386			    HIFN_MAC_CMD_POS_IPSEC;
2387			break;
2388		case CRYPTO_SHA1_HMAC:
2389			cmd->mac_masks |= HIFN_MAC_CMD_ALG_SHA1 |
2390			    HIFN_MAC_CMD_RESULT | HIFN_MAC_CMD_MODE_HMAC |
2391			    HIFN_MAC_CMD_POS_IPSEC | HIFN_MAC_CMD_TRUNC;
2392			break;
2393		}
2394
2395		if ((maccrd->crd_alg == CRYPTO_SHA1_HMAC ||
2396		     maccrd->crd_alg == CRYPTO_MD5_HMAC) &&
2397		    sc->sc_sessions[session].hs_state == HS_STATE_USED) {
2398			cmd->mac_masks |= HIFN_MAC_CMD_NEW_KEY;
2399			bcopy(maccrd->crd_key, cmd->mac, maccrd->crd_klen >> 3);
2400			bzero(cmd->mac + (maccrd->crd_klen >> 3),
2401			    HIFN_MAC_KEY_LENGTH - (maccrd->crd_klen >> 3));
2402		}
2403	}
2404
2405	cmd->crp = crp;
2406	cmd->session_num = session;
2407	cmd->softc = sc;
2408
2409	err = hifn_crypto(sc, cmd, crp, hint);
2410	if (!err) {
2411		if (enccrd)
2412			sc->sc_sessions[session].hs_prev_op =
2413				enccrd->crd_flags & CRD_F_ENCRYPT;
2414		if (sc->sc_sessions[session].hs_state == HS_STATE_USED)
2415			sc->sc_sessions[session].hs_state = HS_STATE_KEY;
2416		return 0;
2417	} else if (err == ERESTART) {
2418		/*
2419		 * There weren't enough resources to dispatch the request
2420		 * to the part.  Notify the caller so they'll requeue this
2421		 * request and resubmit it again soon.
2422		 */
2423#ifdef HIFN_DEBUG
2424		if (hifn_debug)
2425			device_printf(sc->sc_dev, "requeue request\n");
2426#endif
2427		free(cmd, M_DEVBUF);
2428		sc->sc_needwakeup |= CRYPTO_SYMQ;
2429		return (err);
2430	}
2431
2432errout:
2433	if (cmd != NULL)
2434		free(cmd, M_DEVBUF);
2435	if (err == EINVAL)
2436		hifnstats.hst_invalid++;
2437	else
2438		hifnstats.hst_nomem++;
2439	crp->crp_etype = err;
2440	crypto_done(crp);
2441	return (err);
2442}
2443
2444static void
2445hifn_abort(struct hifn_softc *sc)
2446{
2447	struct hifn_dma *dma = sc->sc_dma;
2448	struct hifn_command *cmd;
2449	struct cryptop *crp;
2450	int i, u;
2451
2452	i = dma->resk; u = dma->resu;
2453	while (u != 0) {
2454		cmd = dma->hifn_commands[i];
2455		KASSERT(cmd != NULL, ("hifn_abort: null command slot %u", i));
2456		dma->hifn_commands[i] = NULL;
2457		crp = cmd->crp;
2458
2459		if ((dma->resr[i].l & htole32(HIFN_D_VALID)) == 0) {
2460			/* Salvage what we can. */
2461			u_int8_t *macbuf;
2462
2463			if (cmd->base_masks & HIFN_BASE_CMD_MAC) {
2464				macbuf = dma->result_bufs[i];
2465				macbuf += 12;
2466			} else
2467				macbuf = NULL;
2468			hifnstats.hst_opackets++;
2469			hifn_callback(sc, cmd, macbuf);
2470		} else {
2471			if (cmd->src_map == cmd->dst_map) {
2472				bus_dmamap_sync(sc->sc_dmat, cmd->src_map,
2473				    BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
2474			} else {
2475				bus_dmamap_sync(sc->sc_dmat, cmd->src_map,
2476				    BUS_DMASYNC_POSTWRITE);
2477				bus_dmamap_sync(sc->sc_dmat, cmd->dst_map,
2478				    BUS_DMASYNC_POSTREAD);
2479			}
2480
2481			if (cmd->src_m != cmd->dst_m) {
2482				m_freem(cmd->src_m);
2483				crp->crp_buf = (caddr_t)cmd->dst_m;
2484			}
2485
2486			/* non-shared buffers cannot be restarted */
2487			if (cmd->src_map != cmd->dst_map) {
2488				/*
2489				 * XXX should be EAGAIN, delayed until
2490				 * after the reset.
2491				 */
2492				crp->crp_etype = ENOMEM;
2493				bus_dmamap_unload(sc->sc_dmat, cmd->dst_map);
2494				bus_dmamap_destroy(sc->sc_dmat, cmd->dst_map);
2495			} else
2496				crp->crp_etype = ENOMEM;
2497
2498			bus_dmamap_unload(sc->sc_dmat, cmd->src_map);
2499			bus_dmamap_destroy(sc->sc_dmat, cmd->src_map);
2500
2501			free(cmd, M_DEVBUF);
2502			if (crp->crp_etype != EAGAIN)
2503				crypto_done(crp);
2504		}
2505
2506		if (++i == HIFN_D_RES_RSIZE)
2507			i = 0;
2508		u--;
2509	}
2510	dma->resk = i; dma->resu = u;
2511
2512	/* Force upload of key next time */
2513	for (i = 0; i < sc->sc_maxses; i++)
2514		if (sc->sc_sessions[i].hs_state == HS_STATE_KEY)
2515			sc->sc_sessions[i].hs_state = HS_STATE_USED;
2516
2517	hifn_reset_board(sc, 1);
2518	hifn_init_dma(sc);
2519	hifn_init_pci_registers(sc);
2520}
2521
2522static void
2523hifn_callback(struct hifn_softc *sc, struct hifn_command *cmd, u_int8_t *macbuf)
2524{
2525	struct hifn_dma *dma = sc->sc_dma;
2526	struct cryptop *crp = cmd->crp;
2527	struct cryptodesc *crd;
2528	struct mbuf *m;
2529	int totlen, i, u;
2530
2531	if (cmd->src_map == cmd->dst_map) {
2532		bus_dmamap_sync(sc->sc_dmat, cmd->src_map,
2533		    BUS_DMASYNC_POSTWRITE | BUS_DMASYNC_POSTREAD);
2534	} else {
2535		bus_dmamap_sync(sc->sc_dmat, cmd->src_map,
2536		    BUS_DMASYNC_POSTWRITE);
2537		bus_dmamap_sync(sc->sc_dmat, cmd->dst_map,
2538		    BUS_DMASYNC_POSTREAD);
2539	}
2540
2541	if (crp->crp_flags & CRYPTO_F_IMBUF) {
2542		if (cmd->src_m != cmd->dst_m) {
2543			crp->crp_buf = (caddr_t)cmd->dst_m;
2544			totlen = cmd->src_mapsize;
2545			for (m = cmd->dst_m; m != NULL; m = m->m_next) {
2546				if (totlen < m->m_len) {
2547					m->m_len = totlen;
2548					totlen = 0;
2549				} else
2550					totlen -= m->m_len;
2551			}
2552			cmd->dst_m->m_pkthdr.len = cmd->src_m->m_pkthdr.len;
2553			m_freem(cmd->src_m);
2554		}
2555	}
2556
2557	if (cmd->sloplen != 0) {
2558		if (crp->crp_flags & CRYPTO_F_IMBUF)
2559			m_copyback((struct mbuf *)crp->crp_buf,
2560			    cmd->src_mapsize - cmd->sloplen,
2561			    cmd->sloplen, (caddr_t)&dma->slop[cmd->slopidx]);
2562		else if (crp->crp_flags & CRYPTO_F_IOV)
2563			cuio_copyback((struct uio *)crp->crp_buf,
2564			    cmd->src_mapsize - cmd->sloplen,
2565			    cmd->sloplen, (caddr_t)&dma->slop[cmd->slopidx]);
2566	}
2567
2568	i = dma->dstk; u = dma->dstu;
2569	while (u != 0) {
2570		if (i == HIFN_D_DST_RSIZE)
2571			i = 0;
2572		bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
2573		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2574		if (dma->dstr[i].l & htole32(HIFN_D_VALID)) {
2575			bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
2576			    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2577			break;
2578		}
2579		i++, u--;
2580	}
2581	dma->dstk = i; dma->dstu = u;
2582
2583	hifnstats.hst_obytes += cmd->dst_mapsize;
2584
2585	if ((cmd->base_masks & (HIFN_BASE_CMD_CRYPT | HIFN_BASE_CMD_DECODE)) ==
2586	    HIFN_BASE_CMD_CRYPT) {
2587		for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
2588			if (crd->crd_alg != CRYPTO_DES_CBC &&
2589			    crd->crd_alg != CRYPTO_3DES_CBC)
2590				continue;
2591			if (crp->crp_flags & CRYPTO_F_IMBUF)
2592				m_copydata((struct mbuf *)crp->crp_buf,
2593				    crd->crd_skip + crd->crd_len - HIFN_IV_LENGTH,
2594				    HIFN_IV_LENGTH,
2595				    cmd->softc->sc_sessions[cmd->session_num].hs_iv);
2596			else if (crp->crp_flags & CRYPTO_F_IOV) {
2597				cuio_copydata((struct uio *)crp->crp_buf,
2598				    crd->crd_skip + crd->crd_len - HIFN_IV_LENGTH,
2599				    HIFN_IV_LENGTH,
2600				    cmd->softc->sc_sessions[cmd->session_num].hs_iv);
2601			}
2602			break;
2603		}
2604	}
2605
2606	if (macbuf != NULL) {
2607		for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
2608                        int len;
2609
2610                        if (crd->crd_alg == CRYPTO_MD5)
2611				len = 16;
2612                        else if (crd->crd_alg == CRYPTO_SHA1)
2613				len = 20;
2614                        else if (crd->crd_alg == CRYPTO_MD5_HMAC ||
2615                            crd->crd_alg == CRYPTO_SHA1_HMAC)
2616				len = 12;
2617                        else
2618				continue;
2619
2620			if (crp->crp_flags & CRYPTO_F_IMBUF)
2621				m_copyback((struct mbuf *)crp->crp_buf,
2622                                   crd->crd_inject, len, macbuf);
2623			else if ((crp->crp_flags & CRYPTO_F_IOV) && crp->crp_mac)
2624				bcopy((caddr_t)macbuf, crp->crp_mac, len);
2625			break;
2626		}
2627	}
2628
2629	if (cmd->src_map != cmd->dst_map) {
2630		bus_dmamap_unload(sc->sc_dmat, cmd->dst_map);
2631		bus_dmamap_destroy(sc->sc_dmat, cmd->dst_map);
2632	}
2633	bus_dmamap_unload(sc->sc_dmat, cmd->src_map);
2634	bus_dmamap_destroy(sc->sc_dmat, cmd->src_map);
2635	free(cmd, M_DEVBUF);
2636	crypto_done(crp);
2637}
2638
2639/*
2640 * 7811 PB3 rev/2 parts lock-up on burst writes to Group 0
2641 * and Group 1 registers; avoid conditions that could create
2642 * burst writes by doing a read in between the writes.
2643 *
2644 * NB: The read we interpose is always to the same register;
2645 *     we do this because reading from an arbitrary (e.g. last)
2646 *     register may not always work.
2647 */
2648static void
2649hifn_write_reg_0(struct hifn_softc *sc, bus_size_t reg, u_int32_t val)
2650{
2651	if (sc->sc_flags & HIFN_IS_7811) {
2652		if (sc->sc_bar0_lastreg == reg - 4)
2653			bus_space_read_4(sc->sc_st0, sc->sc_sh0, HIFN_0_PUCNFG);
2654		sc->sc_bar0_lastreg = reg;
2655	}
2656	bus_space_write_4(sc->sc_st0, sc->sc_sh0, reg, val);
2657}
2658
2659static void
2660hifn_write_reg_1(struct hifn_softc *sc, bus_size_t reg, u_int32_t val)
2661{
2662	if (sc->sc_flags & HIFN_IS_7811) {
2663		if (sc->sc_bar1_lastreg == reg - 4)
2664			bus_space_read_4(sc->sc_st1, sc->sc_sh1, HIFN_1_REVID);
2665		sc->sc_bar1_lastreg = reg;
2666	}
2667	bus_space_write_4(sc->sc_st1, sc->sc_sh1, reg, val);
2668}
2669