Lines Matching refs:nbytes

49 static inline bool cast5_fpu_begin(bool fpu_enabled, unsigned int nbytes)  in cast5_fpu_begin()  argument
52 NULL, fpu_enabled, nbytes); in cast5_fpu_begin()
66 unsigned int nbytes; in ecb_crypt() local
75 while ((nbytes = walk->nbytes)) { in ecb_crypt()
79 fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes); in ecb_crypt()
82 if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) { in ecb_crypt()
88 nbytes -= bsize * CAST5_PARALLEL_BLOCKS; in ecb_crypt()
89 } while (nbytes >= bsize * CAST5_PARALLEL_BLOCKS); in ecb_crypt()
91 if (nbytes < bsize) in ecb_crypt()
103 nbytes -= bsize; in ecb_crypt()
104 } while (nbytes >= bsize); in ecb_crypt()
107 err = blkcipher_walk_done(desc, walk, nbytes); in ecb_crypt()
115 struct scatterlist *src, unsigned int nbytes) in ecb_encrypt() argument
119 blkcipher_walk_init(&walk, dst, src, nbytes); in ecb_encrypt()
124 struct scatterlist *src, unsigned int nbytes) in ecb_decrypt() argument
128 blkcipher_walk_init(&walk, dst, src, nbytes); in ecb_decrypt()
137 unsigned int nbytes = walk->nbytes; in __cbc_encrypt() local
149 nbytes -= bsize; in __cbc_encrypt()
150 } while (nbytes >= bsize); in __cbc_encrypt()
153 return nbytes; in __cbc_encrypt()
157 struct scatterlist *src, unsigned int nbytes) in cbc_encrypt() argument
162 blkcipher_walk_init(&walk, dst, src, nbytes); in cbc_encrypt()
165 while ((nbytes = walk.nbytes)) { in cbc_encrypt()
166 nbytes = __cbc_encrypt(desc, &walk); in cbc_encrypt()
167 err = blkcipher_walk_done(desc, &walk, nbytes); in cbc_encrypt()
178 unsigned int nbytes = walk->nbytes; in __cbc_decrypt() local
184 src += nbytes / bsize - 1; in __cbc_decrypt()
185 dst += nbytes / bsize - 1; in __cbc_decrypt()
190 if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) { in __cbc_decrypt()
192 nbytes -= bsize * (CAST5_PARALLEL_BLOCKS - 1); in __cbc_decrypt()
198 nbytes -= bsize; in __cbc_decrypt()
199 if (nbytes < bsize) in __cbc_decrypt()
205 } while (nbytes >= bsize * CAST5_PARALLEL_BLOCKS); in __cbc_decrypt()
212 nbytes -= bsize; in __cbc_decrypt()
213 if (nbytes < bsize) in __cbc_decrypt()
225 return nbytes; in __cbc_decrypt()
229 struct scatterlist *src, unsigned int nbytes) in cbc_decrypt() argument
235 blkcipher_walk_init(&walk, dst, src, nbytes); in cbc_decrypt()
239 while ((nbytes = walk.nbytes)) { in cbc_decrypt()
240 fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes); in cbc_decrypt()
241 nbytes = __cbc_decrypt(desc, &walk); in cbc_decrypt()
242 err = blkcipher_walk_done(desc, &walk, nbytes); in cbc_decrypt()
257 unsigned int nbytes = walk->nbytes; in ctr_crypt_final() local
260 crypto_xor(keystream, src, nbytes); in ctr_crypt_final()
261 memcpy(dst, keystream, nbytes); in ctr_crypt_final()
271 unsigned int nbytes = walk->nbytes; in __ctr_crypt() local
276 if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) { in __ctr_crypt()
283 nbytes -= bsize * CAST5_PARALLEL_BLOCKS; in __ctr_crypt()
284 } while (nbytes >= bsize * CAST5_PARALLEL_BLOCKS); in __ctr_crypt()
286 if (nbytes < bsize) in __ctr_crypt()
305 nbytes -= bsize; in __ctr_crypt()
306 } while (nbytes >= bsize); in __ctr_crypt()
309 return nbytes; in __ctr_crypt()
313 struct scatterlist *src, unsigned int nbytes) in ctr_crypt() argument
319 blkcipher_walk_init(&walk, dst, src, nbytes); in ctr_crypt()
323 while ((nbytes = walk.nbytes) >= CAST5_BLOCK_SIZE) { in ctr_crypt()
324 fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes); in ctr_crypt()
325 nbytes = __ctr_crypt(desc, &walk); in ctr_crypt()
326 err = blkcipher_walk_done(desc, &walk, nbytes); in ctr_crypt()
331 if (walk.nbytes) { in ctr_crypt()