This source file includes following definitions.
- camellia_setkey
- ecb_encrypt
- ecb_decrypt
- cbc_encrypt
- cbc_decrypt
- ctr_crypt
- xts_encrypt
- xts_decrypt
- camellia_aesni_init
- camellia_aesni_fini
1
2
3
4
5
6
7
8 #include <asm/crypto/camellia.h>
9 #include <asm/crypto/glue_helper.h>
10 #include <crypto/algapi.h>
11 #include <crypto/internal/simd.h>
12 #include <crypto/xts.h>
13 #include <linux/crypto.h>
14 #include <linux/err.h>
15 #include <linux/module.h>
16 #include <linux/types.h>
17
18 #define CAMELLIA_AESNI_PARALLEL_BLOCKS 16
19 #define CAMELLIA_AESNI_AVX2_PARALLEL_BLOCKS 32
20
21
22 asmlinkage void camellia_ecb_enc_32way(struct camellia_ctx *ctx, u8 *dst,
23 const u8 *src);
24 asmlinkage void camellia_ecb_dec_32way(struct camellia_ctx *ctx, u8 *dst,
25 const u8 *src);
26
27 asmlinkage void camellia_cbc_dec_32way(struct camellia_ctx *ctx, u8 *dst,
28 const u8 *src);
29 asmlinkage void camellia_ctr_32way(struct camellia_ctx *ctx, u8 *dst,
30 const u8 *src, le128 *iv);
31
32 asmlinkage void camellia_xts_enc_32way(struct camellia_ctx *ctx, u8 *dst,
33 const u8 *src, le128 *iv);
34 asmlinkage void camellia_xts_dec_32way(struct camellia_ctx *ctx, u8 *dst,
35 const u8 *src, le128 *iv);
36
37 static const struct common_glue_ctx camellia_enc = {
38 .num_funcs = 4,
39 .fpu_blocks_limit = CAMELLIA_AESNI_PARALLEL_BLOCKS,
40
41 .funcs = { {
42 .num_blocks = CAMELLIA_AESNI_AVX2_PARALLEL_BLOCKS,
43 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_enc_32way) }
44 }, {
45 .num_blocks = CAMELLIA_AESNI_PARALLEL_BLOCKS,
46 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_enc_16way) }
47 }, {
48 .num_blocks = 2,
49 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk_2way) }
50 }, {
51 .num_blocks = 1,
52 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk) }
53 } }
54 };
55
56 static const struct common_glue_ctx camellia_ctr = {
57 .num_funcs = 4,
58 .fpu_blocks_limit = CAMELLIA_AESNI_PARALLEL_BLOCKS,
59
60 .funcs = { {
61 .num_blocks = CAMELLIA_AESNI_AVX2_PARALLEL_BLOCKS,
62 .fn_u = { .ctr = GLUE_CTR_FUNC_CAST(camellia_ctr_32way) }
63 }, {
64 .num_blocks = CAMELLIA_AESNI_PARALLEL_BLOCKS,
65 .fn_u = { .ctr = GLUE_CTR_FUNC_CAST(camellia_ctr_16way) }
66 }, {
67 .num_blocks = 2,
68 .fn_u = { .ctr = GLUE_CTR_FUNC_CAST(camellia_crypt_ctr_2way) }
69 }, {
70 .num_blocks = 1,
71 .fn_u = { .ctr = GLUE_CTR_FUNC_CAST(camellia_crypt_ctr) }
72 } }
73 };
74
75 static const struct common_glue_ctx camellia_enc_xts = {
76 .num_funcs = 3,
77 .fpu_blocks_limit = CAMELLIA_AESNI_PARALLEL_BLOCKS,
78
79 .funcs = { {
80 .num_blocks = CAMELLIA_AESNI_AVX2_PARALLEL_BLOCKS,
81 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_32way) }
82 }, {
83 .num_blocks = CAMELLIA_AESNI_PARALLEL_BLOCKS,
84 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_16way) }
85 }, {
86 .num_blocks = 1,
87 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc) }
88 } }
89 };
90
91 static const struct common_glue_ctx camellia_dec = {
92 .num_funcs = 4,
93 .fpu_blocks_limit = CAMELLIA_AESNI_PARALLEL_BLOCKS,
94
95 .funcs = { {
96 .num_blocks = CAMELLIA_AESNI_AVX2_PARALLEL_BLOCKS,
97 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_dec_32way) }
98 }, {
99 .num_blocks = CAMELLIA_AESNI_PARALLEL_BLOCKS,
100 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_dec_16way) }
101 }, {
102 .num_blocks = 2,
103 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk_2way) }
104 }, {
105 .num_blocks = 1,
106 .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk) }
107 } }
108 };
109
110 static const struct common_glue_ctx camellia_dec_cbc = {
111 .num_funcs = 4,
112 .fpu_blocks_limit = CAMELLIA_AESNI_PARALLEL_BLOCKS,
113
114 .funcs = { {
115 .num_blocks = CAMELLIA_AESNI_AVX2_PARALLEL_BLOCKS,
116 .fn_u = { .cbc = GLUE_CBC_FUNC_CAST(camellia_cbc_dec_32way) }
117 }, {
118 .num_blocks = CAMELLIA_AESNI_PARALLEL_BLOCKS,
119 .fn_u = { .cbc = GLUE_CBC_FUNC_CAST(camellia_cbc_dec_16way) }
120 }, {
121 .num_blocks = 2,
122 .fn_u = { .cbc = GLUE_CBC_FUNC_CAST(camellia_decrypt_cbc_2way) }
123 }, {
124 .num_blocks = 1,
125 .fn_u = { .cbc = GLUE_CBC_FUNC_CAST(camellia_dec_blk) }
126 } }
127 };
128
129 static const struct common_glue_ctx camellia_dec_xts = {
130 .num_funcs = 3,
131 .fpu_blocks_limit = CAMELLIA_AESNI_PARALLEL_BLOCKS,
132
133 .funcs = { {
134 .num_blocks = CAMELLIA_AESNI_AVX2_PARALLEL_BLOCKS,
135 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_32way) }
136 }, {
137 .num_blocks = CAMELLIA_AESNI_PARALLEL_BLOCKS,
138 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_16way) }
139 }, {
140 .num_blocks = 1,
141 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec) }
142 } }
143 };
144
145 static int camellia_setkey(struct crypto_skcipher *tfm, const u8 *key,
146 unsigned int keylen)
147 {
148 return __camellia_setkey(crypto_skcipher_ctx(tfm), key, keylen,
149 &tfm->base.crt_flags);
150 }
151
152 static int ecb_encrypt(struct skcipher_request *req)
153 {
154 return glue_ecb_req_128bit(&camellia_enc, req);
155 }
156
157 static int ecb_decrypt(struct skcipher_request *req)
158 {
159 return glue_ecb_req_128bit(&camellia_dec, req);
160 }
161
162 static int cbc_encrypt(struct skcipher_request *req)
163 {
164 return glue_cbc_encrypt_req_128bit(GLUE_FUNC_CAST(camellia_enc_blk),
165 req);
166 }
167
168 static int cbc_decrypt(struct skcipher_request *req)
169 {
170 return glue_cbc_decrypt_req_128bit(&camellia_dec_cbc, req);
171 }
172
173 static int ctr_crypt(struct skcipher_request *req)
174 {
175 return glue_ctr_req_128bit(&camellia_ctr, req);
176 }
177
178 static int xts_encrypt(struct skcipher_request *req)
179 {
180 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
181 struct camellia_xts_ctx *ctx = crypto_skcipher_ctx(tfm);
182
183 return glue_xts_req_128bit(&camellia_enc_xts, req,
184 XTS_TWEAK_CAST(camellia_enc_blk),
185 &ctx->tweak_ctx, &ctx->crypt_ctx, false);
186 }
187
188 static int xts_decrypt(struct skcipher_request *req)
189 {
190 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
191 struct camellia_xts_ctx *ctx = crypto_skcipher_ctx(tfm);
192
193 return glue_xts_req_128bit(&camellia_dec_xts, req,
194 XTS_TWEAK_CAST(camellia_enc_blk),
195 &ctx->tweak_ctx, &ctx->crypt_ctx, true);
196 }
197
198 static struct skcipher_alg camellia_algs[] = {
199 {
200 .base.cra_name = "__ecb(camellia)",
201 .base.cra_driver_name = "__ecb-camellia-aesni-avx2",
202 .base.cra_priority = 500,
203 .base.cra_flags = CRYPTO_ALG_INTERNAL,
204 .base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
205 .base.cra_ctxsize = sizeof(struct camellia_ctx),
206 .base.cra_module = THIS_MODULE,
207 .min_keysize = CAMELLIA_MIN_KEY_SIZE,
208 .max_keysize = CAMELLIA_MAX_KEY_SIZE,
209 .setkey = camellia_setkey,
210 .encrypt = ecb_encrypt,
211 .decrypt = ecb_decrypt,
212 }, {
213 .base.cra_name = "__cbc(camellia)",
214 .base.cra_driver_name = "__cbc-camellia-aesni-avx2",
215 .base.cra_priority = 500,
216 .base.cra_flags = CRYPTO_ALG_INTERNAL,
217 .base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
218 .base.cra_ctxsize = sizeof(struct camellia_ctx),
219 .base.cra_module = THIS_MODULE,
220 .min_keysize = CAMELLIA_MIN_KEY_SIZE,
221 .max_keysize = CAMELLIA_MAX_KEY_SIZE,
222 .ivsize = CAMELLIA_BLOCK_SIZE,
223 .setkey = camellia_setkey,
224 .encrypt = cbc_encrypt,
225 .decrypt = cbc_decrypt,
226 }, {
227 .base.cra_name = "__ctr(camellia)",
228 .base.cra_driver_name = "__ctr-camellia-aesni-avx2",
229 .base.cra_priority = 500,
230 .base.cra_flags = CRYPTO_ALG_INTERNAL,
231 .base.cra_blocksize = 1,
232 .base.cra_ctxsize = sizeof(struct camellia_ctx),
233 .base.cra_module = THIS_MODULE,
234 .min_keysize = CAMELLIA_MIN_KEY_SIZE,
235 .max_keysize = CAMELLIA_MAX_KEY_SIZE,
236 .ivsize = CAMELLIA_BLOCK_SIZE,
237 .chunksize = CAMELLIA_BLOCK_SIZE,
238 .setkey = camellia_setkey,
239 .encrypt = ctr_crypt,
240 .decrypt = ctr_crypt,
241 }, {
242 .base.cra_name = "__xts(camellia)",
243 .base.cra_driver_name = "__xts-camellia-aesni-avx2",
244 .base.cra_priority = 500,
245 .base.cra_flags = CRYPTO_ALG_INTERNAL,
246 .base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
247 .base.cra_ctxsize = sizeof(struct camellia_xts_ctx),
248 .base.cra_module = THIS_MODULE,
249 .min_keysize = 2 * CAMELLIA_MIN_KEY_SIZE,
250 .max_keysize = 2 * CAMELLIA_MAX_KEY_SIZE,
251 .ivsize = CAMELLIA_BLOCK_SIZE,
252 .setkey = xts_camellia_setkey,
253 .encrypt = xts_encrypt,
254 .decrypt = xts_decrypt,
255 },
256 };
257
258 static struct simd_skcipher_alg *camellia_simd_algs[ARRAY_SIZE(camellia_algs)];
259
260 static int __init camellia_aesni_init(void)
261 {
262 const char *feature_name;
263
264 if (!boot_cpu_has(X86_FEATURE_AVX) ||
265 !boot_cpu_has(X86_FEATURE_AVX2) ||
266 !boot_cpu_has(X86_FEATURE_AES) ||
267 !boot_cpu_has(X86_FEATURE_OSXSAVE)) {
268 pr_info("AVX2 or AES-NI instructions are not detected.\n");
269 return -ENODEV;
270 }
271
272 if (!cpu_has_xfeatures(XFEATURE_MASK_SSE | XFEATURE_MASK_YMM,
273 &feature_name)) {
274 pr_info("CPU feature '%s' is not supported.\n", feature_name);
275 return -ENODEV;
276 }
277
278 return simd_register_skciphers_compat(camellia_algs,
279 ARRAY_SIZE(camellia_algs),
280 camellia_simd_algs);
281 }
282
283 static void __exit camellia_aesni_fini(void)
284 {
285 simd_unregister_skciphers(camellia_algs, ARRAY_SIZE(camellia_algs),
286 camellia_simd_algs);
287 }
288
289 module_init(camellia_aesni_init);
290 module_exit(camellia_aesni_fini);
291
292 MODULE_LICENSE("GPL");
293 MODULE_DESCRIPTION("Camellia Cipher Algorithm, AES-NI/AVX2 optimized");
294 MODULE_ALIAS_CRYPTO("camellia");
295 MODULE_ALIAS_CRYPTO("camellia-asm");