Lines Matching refs:nbytes
48 static inline bool cast5_fpu_begin(bool fpu_enabled, unsigned int nbytes) in cast5_fpu_begin() argument
51 NULL, fpu_enabled, nbytes); in cast5_fpu_begin()
65 unsigned int nbytes; in ecb_crypt() local
74 while ((nbytes = walk->nbytes)) { in ecb_crypt()
78 fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes); in ecb_crypt()
81 if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) { in ecb_crypt()
87 nbytes -= bsize * CAST5_PARALLEL_BLOCKS; in ecb_crypt()
88 } while (nbytes >= bsize * CAST5_PARALLEL_BLOCKS); in ecb_crypt()
90 if (nbytes < bsize) in ecb_crypt()
102 nbytes -= bsize; in ecb_crypt()
103 } while (nbytes >= bsize); in ecb_crypt()
106 err = blkcipher_walk_done(desc, walk, nbytes); in ecb_crypt()
114 struct scatterlist *src, unsigned int nbytes) in ecb_encrypt() argument
118 blkcipher_walk_init(&walk, dst, src, nbytes); in ecb_encrypt()
123 struct scatterlist *src, unsigned int nbytes) in ecb_decrypt() argument
127 blkcipher_walk_init(&walk, dst, src, nbytes); in ecb_decrypt()
136 unsigned int nbytes = walk->nbytes; in __cbc_encrypt() local
148 nbytes -= bsize; in __cbc_encrypt()
149 } while (nbytes >= bsize); in __cbc_encrypt()
152 return nbytes; in __cbc_encrypt()
156 struct scatterlist *src, unsigned int nbytes) in cbc_encrypt() argument
161 blkcipher_walk_init(&walk, dst, src, nbytes); in cbc_encrypt()
164 while ((nbytes = walk.nbytes)) { in cbc_encrypt()
165 nbytes = __cbc_encrypt(desc, &walk); in cbc_encrypt()
166 err = blkcipher_walk_done(desc, &walk, nbytes); in cbc_encrypt()
177 unsigned int nbytes = walk->nbytes; in __cbc_decrypt() local
183 src += nbytes / bsize - 1; in __cbc_decrypt()
184 dst += nbytes / bsize - 1; in __cbc_decrypt()
189 if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) { in __cbc_decrypt()
191 nbytes -= bsize * (CAST5_PARALLEL_BLOCKS - 1); in __cbc_decrypt()
197 nbytes -= bsize; in __cbc_decrypt()
198 if (nbytes < bsize) in __cbc_decrypt()
204 } while (nbytes >= bsize * CAST5_PARALLEL_BLOCKS); in __cbc_decrypt()
211 nbytes -= bsize; in __cbc_decrypt()
212 if (nbytes < bsize) in __cbc_decrypt()
224 return nbytes; in __cbc_decrypt()
228 struct scatterlist *src, unsigned int nbytes) in cbc_decrypt() argument
234 blkcipher_walk_init(&walk, dst, src, nbytes); in cbc_decrypt()
238 while ((nbytes = walk.nbytes)) { in cbc_decrypt()
239 fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes); in cbc_decrypt()
240 nbytes = __cbc_decrypt(desc, &walk); in cbc_decrypt()
241 err = blkcipher_walk_done(desc, &walk, nbytes); in cbc_decrypt()
256 unsigned int nbytes = walk->nbytes; in ctr_crypt_final() local
259 crypto_xor(keystream, src, nbytes); in ctr_crypt_final()
260 memcpy(dst, keystream, nbytes); in ctr_crypt_final()
270 unsigned int nbytes = walk->nbytes; in __ctr_crypt() local
275 if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) { in __ctr_crypt()
282 nbytes -= bsize * CAST5_PARALLEL_BLOCKS; in __ctr_crypt()
283 } while (nbytes >= bsize * CAST5_PARALLEL_BLOCKS); in __ctr_crypt()
285 if (nbytes < bsize) in __ctr_crypt()
304 nbytes -= bsize; in __ctr_crypt()
305 } while (nbytes >= bsize); in __ctr_crypt()
308 return nbytes; in __ctr_crypt()
312 struct scatterlist *src, unsigned int nbytes) in ctr_crypt() argument
318 blkcipher_walk_init(&walk, dst, src, nbytes); in ctr_crypt()
322 while ((nbytes = walk.nbytes) >= CAST5_BLOCK_SIZE) { in ctr_crypt()
323 fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes); in ctr_crypt()
324 nbytes = __ctr_crypt(desc, &walk); in ctr_crypt()
325 err = blkcipher_walk_done(desc, &walk, nbytes); in ctr_crypt()
330 if (walk.nbytes) { in ctr_crypt()