1// SPDX-License-Identifier: GPL-2.0-only
2/* Glue code for SHA512 hashing optimized for sparc64 crypto opcodes.
3 *
4 * This is based largely upon crypto/sha512_generic.c
5 *
6 * Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com>
7 * Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk>
8 * Copyright (c) 2003 Kyle McMartin <kyle@debian.org>
9 */
10
11#define pr_fmt(fmt)	KBUILD_MODNAME ": " fmt
12
13#include <crypto/internal/hash.h>
14#include <linux/init.h>
15#include <linux/module.h>
16#include <linux/mm.h>
17#include <linux/types.h>
18#include <crypto/sha2.h>
19#include <crypto/sha512_base.h>
20
21#include <asm/pstate.h>
22#include <asm/elf.h>
23
24#include "opcodes.h"
25
26asmlinkage void sha512_sparc64_transform(u64 *digest, const char *data,
27					 unsigned int rounds);
28
29static void __sha512_sparc64_update(struct sha512_state *sctx, const u8 *data,
30				    unsigned int len, unsigned int partial)
31{
32	unsigned int done = 0;
33
34	if ((sctx->count[0] += len) < len)
35		sctx->count[1]++;
36	if (partial) {
37		done = SHA512_BLOCK_SIZE - partial;
38		memcpy(sctx->buf + partial, data, done);
39		sha512_sparc64_transform(sctx->state, sctx->buf, 1);
40	}
41	if (len - done >= SHA512_BLOCK_SIZE) {
42		const unsigned int rounds = (len - done) / SHA512_BLOCK_SIZE;
43
44		sha512_sparc64_transform(sctx->state, data + done, rounds);
45		done += rounds * SHA512_BLOCK_SIZE;
46	}
47
48	memcpy(sctx->buf, data + done, len - done);
49}
50
51static int sha512_sparc64_update(struct shash_desc *desc, const u8 *data,
52				 unsigned int len)
53{
54	struct sha512_state *sctx = shash_desc_ctx(desc);
55	unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE;
56
57	/* Handle the fast case right here */
58	if (partial + len < SHA512_BLOCK_SIZE) {
59		if ((sctx->count[0] += len) < len)
60			sctx->count[1]++;
61		memcpy(sctx->buf + partial, data, len);
62	} else
63		__sha512_sparc64_update(sctx, data, len, partial);
64
65	return 0;
66}
67
68static int sha512_sparc64_final(struct shash_desc *desc, u8 *out)
69{
70	struct sha512_state *sctx = shash_desc_ctx(desc);
71	unsigned int i, index, padlen;
72	__be64 *dst = (__be64 *)out;
73	__be64 bits[2];
74	static const u8 padding[SHA512_BLOCK_SIZE] = { 0x80, };
75
76	/* Save number of bits */
77	bits[1] = cpu_to_be64(sctx->count[0] << 3);
78	bits[0] = cpu_to_be64(sctx->count[1] << 3 | sctx->count[0] >> 61);
79
80	/* Pad out to 112 mod 128 and append length */
81	index = sctx->count[0] % SHA512_BLOCK_SIZE;
82	padlen = (index < 112) ? (112 - index) : ((SHA512_BLOCK_SIZE+112) - index);
83
84	/* We need to fill a whole block for __sha512_sparc64_update() */
85	if (padlen <= 112) {
86		if ((sctx->count[0] += padlen) < padlen)
87			sctx->count[1]++;
88		memcpy(sctx->buf + index, padding, padlen);
89	} else {
90		__sha512_sparc64_update(sctx, padding, padlen, index);
91	}
92	__sha512_sparc64_update(sctx, (const u8 *)&bits, sizeof(bits), 112);
93
94	/* Store state in digest */
95	for (i = 0; i < 8; i++)
96		dst[i] = cpu_to_be64(sctx->state[i]);
97
98	/* Wipe context */
99	memset(sctx, 0, sizeof(*sctx));
100
101	return 0;
102}
103
104static int sha384_sparc64_final(struct shash_desc *desc, u8 *hash)
105{
106	u8 D[64];
107
108	sha512_sparc64_final(desc, D);
109
110	memcpy(hash, D, 48);
111	memzero_explicit(D, 64);
112
113	return 0;
114}
115
116static struct shash_alg sha512 = {
117	.digestsize	=	SHA512_DIGEST_SIZE,
118	.init		=	sha512_base_init,
119	.update		=	sha512_sparc64_update,
120	.final		=	sha512_sparc64_final,
121	.descsize	=	sizeof(struct sha512_state),
122	.base		=	{
123		.cra_name	=	"sha512",
124		.cra_driver_name=	"sha512-sparc64",
125		.cra_priority	=	SPARC_CR_OPCODE_PRIORITY,
126		.cra_blocksize	=	SHA512_BLOCK_SIZE,
127		.cra_module	=	THIS_MODULE,
128	}
129};
130
131static struct shash_alg sha384 = {
132	.digestsize	=	SHA384_DIGEST_SIZE,
133	.init		=	sha384_base_init,
134	.update		=	sha512_sparc64_update,
135	.final		=	sha384_sparc64_final,
136	.descsize	=	sizeof(struct sha512_state),
137	.base		=	{
138		.cra_name	=	"sha384",
139		.cra_driver_name=	"sha384-sparc64",
140		.cra_priority	=	SPARC_CR_OPCODE_PRIORITY,
141		.cra_blocksize	=	SHA384_BLOCK_SIZE,
142		.cra_module	=	THIS_MODULE,
143	}
144};
145
146static bool __init sparc64_has_sha512_opcode(void)
147{
148	unsigned long cfr;
149
150	if (!(sparc64_elf_hwcap & HWCAP_SPARC_CRYPTO))
151		return false;
152
153	__asm__ __volatile__("rd %%asr26, %0" : "=r" (cfr));
154	if (!(cfr & CFR_SHA512))
155		return false;
156
157	return true;
158}
159
160static int __init sha512_sparc64_mod_init(void)
161{
162	if (sparc64_has_sha512_opcode()) {
163		int ret = crypto_register_shash(&sha384);
164		if (ret < 0)
165			return ret;
166
167		ret = crypto_register_shash(&sha512);
168		if (ret < 0) {
169			crypto_unregister_shash(&sha384);
170			return ret;
171		}
172
173		pr_info("Using sparc64 sha512 opcode optimized SHA-512/SHA-384 implementation\n");
174		return 0;
175	}
176	pr_info("sparc64 sha512 opcode not available.\n");
177	return -ENODEV;
178}
179
180static void __exit sha512_sparc64_mod_fini(void)
181{
182	crypto_unregister_shash(&sha384);
183	crypto_unregister_shash(&sha512);
184}
185
186module_init(sha512_sparc64_mod_init);
187module_exit(sha512_sparc64_mod_fini);
188
189MODULE_LICENSE("GPL");
190MODULE_DESCRIPTION("SHA-384 and SHA-512 Secure Hash Algorithm, sparc64 sha512 opcode accelerated");
191
192MODULE_ALIAS_CRYPTO("sha384");
193MODULE_ALIAS_CRYPTO("sha512");
194
195#include "crop_devid.c"
196