1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Asynchronous Compression operations
4 *
5 * Copyright (c) 2016, Intel Corporation
6 * Authors: Weigang Li <weigang.li@intel.com>
7 *          Giovanni Cabiddu <giovanni.cabiddu@intel.com>
8 */
9
10#include <crypto/internal/acompress.h>
11#include <linux/cryptouser.h>
12#include <linux/errno.h>
13#include <linux/kernel.h>
14#include <linux/module.h>
15#include <linux/seq_file.h>
16#include <linux/slab.h>
17#include <linux/string.h>
18#include <net/netlink.h>
19
20#include "compress.h"
21
22struct crypto_scomp;
23
24static const struct crypto_type crypto_acomp_type;
25
26static inline struct acomp_alg *__crypto_acomp_alg(struct crypto_alg *alg)
27{
28	return container_of(alg, struct acomp_alg, calg.base);
29}
30
31static inline struct acomp_alg *crypto_acomp_alg(struct crypto_acomp *tfm)
32{
33	return __crypto_acomp_alg(crypto_acomp_tfm(tfm)->__crt_alg);
34}
35
36static int __maybe_unused crypto_acomp_report(
37	struct sk_buff *skb, struct crypto_alg *alg)
38{
39	struct crypto_report_acomp racomp;
40
41	memset(&racomp, 0, sizeof(racomp));
42
43	strscpy(racomp.type, "acomp", sizeof(racomp.type));
44
45	return nla_put(skb, CRYPTOCFGA_REPORT_ACOMP, sizeof(racomp), &racomp);
46}
47
48static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
49	__maybe_unused;
50
51static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
52{
53	seq_puts(m, "type         : acomp\n");
54}
55
56static void crypto_acomp_exit_tfm(struct crypto_tfm *tfm)
57{
58	struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
59	struct acomp_alg *alg = crypto_acomp_alg(acomp);
60
61	alg->exit(acomp);
62}
63
64static int crypto_acomp_init_tfm(struct crypto_tfm *tfm)
65{
66	struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
67	struct acomp_alg *alg = crypto_acomp_alg(acomp);
68
69	if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
70		return crypto_init_scomp_ops_async(tfm);
71
72	acomp->compress = alg->compress;
73	acomp->decompress = alg->decompress;
74	acomp->dst_free = alg->dst_free;
75	acomp->reqsize = alg->reqsize;
76
77	if (alg->exit)
78		acomp->base.exit = crypto_acomp_exit_tfm;
79
80	if (alg->init)
81		return alg->init(acomp);
82
83	return 0;
84}
85
86static unsigned int crypto_acomp_extsize(struct crypto_alg *alg)
87{
88	int extsize = crypto_alg_extsize(alg);
89
90	if (alg->cra_type != &crypto_acomp_type)
91		extsize += sizeof(struct crypto_scomp *);
92
93	return extsize;
94}
95
96static inline int __crypto_acomp_report_stat(struct sk_buff *skb,
97					     struct crypto_alg *alg)
98{
99	struct comp_alg_common *calg = __crypto_comp_alg_common(alg);
100	struct crypto_istat_compress *istat = comp_get_stat(calg);
101	struct crypto_stat_compress racomp;
102
103	memset(&racomp, 0, sizeof(racomp));
104
105	strscpy(racomp.type, "acomp", sizeof(racomp.type));
106	racomp.stat_compress_cnt = atomic64_read(&istat->compress_cnt);
107	racomp.stat_compress_tlen = atomic64_read(&istat->compress_tlen);
108	racomp.stat_decompress_cnt =  atomic64_read(&istat->decompress_cnt);
109	racomp.stat_decompress_tlen = atomic64_read(&istat->decompress_tlen);
110	racomp.stat_err_cnt = atomic64_read(&istat->err_cnt);
111
112	return nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(racomp), &racomp);
113}
114
115#ifdef CONFIG_CRYPTO_STATS
116int crypto_acomp_report_stat(struct sk_buff *skb, struct crypto_alg *alg)
117{
118	return __crypto_acomp_report_stat(skb, alg);
119}
120#endif
121
122static const struct crypto_type crypto_acomp_type = {
123	.extsize = crypto_acomp_extsize,
124	.init_tfm = crypto_acomp_init_tfm,
125#ifdef CONFIG_PROC_FS
126	.show = crypto_acomp_show,
127#endif
128#if IS_ENABLED(CONFIG_CRYPTO_USER)
129	.report = crypto_acomp_report,
130#endif
131#ifdef CONFIG_CRYPTO_STATS
132	.report_stat = crypto_acomp_report_stat,
133#endif
134	.maskclear = ~CRYPTO_ALG_TYPE_MASK,
135	.maskset = CRYPTO_ALG_TYPE_ACOMPRESS_MASK,
136	.type = CRYPTO_ALG_TYPE_ACOMPRESS,
137	.tfmsize = offsetof(struct crypto_acomp, base),
138};
139
140struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
141					u32 mask)
142{
143	return crypto_alloc_tfm(alg_name, &crypto_acomp_type, type, mask);
144}
145EXPORT_SYMBOL_GPL(crypto_alloc_acomp);
146
147struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type,
148					u32 mask, int node)
149{
150	return crypto_alloc_tfm_node(alg_name, &crypto_acomp_type, type, mask,
151				node);
152}
153EXPORT_SYMBOL_GPL(crypto_alloc_acomp_node);
154
155struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp)
156{
157	struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
158	struct acomp_req *req;
159
160	req = __acomp_request_alloc(acomp);
161	if (req && (tfm->__crt_alg->cra_type != &crypto_acomp_type))
162		return crypto_acomp_scomp_alloc_ctx(req);
163
164	return req;
165}
166EXPORT_SYMBOL_GPL(acomp_request_alloc);
167
168void acomp_request_free(struct acomp_req *req)
169{
170	struct crypto_acomp *acomp = crypto_acomp_reqtfm(req);
171	struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
172
173	if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
174		crypto_acomp_scomp_free_ctx(req);
175
176	if (req->flags & CRYPTO_ACOMP_ALLOC_OUTPUT) {
177		acomp->dst_free(req->dst);
178		req->dst = NULL;
179	}
180
181	__acomp_request_free(req);
182}
183EXPORT_SYMBOL_GPL(acomp_request_free);
184
185void comp_prepare_alg(struct comp_alg_common *alg)
186{
187	struct crypto_istat_compress *istat = comp_get_stat(alg);
188	struct crypto_alg *base = &alg->base;
189
190	base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK;
191
192	if (IS_ENABLED(CONFIG_CRYPTO_STATS))
193		memset(istat, 0, sizeof(*istat));
194}
195
196int crypto_register_acomp(struct acomp_alg *alg)
197{
198	struct crypto_alg *base = &alg->calg.base;
199
200	comp_prepare_alg(&alg->calg);
201
202	base->cra_type = &crypto_acomp_type;
203	base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS;
204
205	return crypto_register_alg(base);
206}
207EXPORT_SYMBOL_GPL(crypto_register_acomp);
208
209void crypto_unregister_acomp(struct acomp_alg *alg)
210{
211	crypto_unregister_alg(&alg->base);
212}
213EXPORT_SYMBOL_GPL(crypto_unregister_acomp);
214
215int crypto_register_acomps(struct acomp_alg *algs, int count)
216{
217	int i, ret;
218
219	for (i = 0; i < count; i++) {
220		ret = crypto_register_acomp(&algs[i]);
221		if (ret)
222			goto err;
223	}
224
225	return 0;
226
227err:
228	for (--i; i >= 0; --i)
229		crypto_unregister_acomp(&algs[i]);
230
231	return ret;
232}
233EXPORT_SYMBOL_GPL(crypto_register_acomps);
234
235void crypto_unregister_acomps(struct acomp_alg *algs, int count)
236{
237	int i;
238
239	for (i = count - 1; i >= 0; --i)
240		crypto_unregister_acomp(&algs[i]);
241}
242EXPORT_SYMBOL_GPL(crypto_unregister_acomps);
243
244MODULE_LICENSE("GPL");
245MODULE_DESCRIPTION("Asynchronous compression type");
246