Lines Matching refs:walk
170 struct blkcipher_walk walk; in ecb_encrypt() local
175 blkcipher_walk_init(&walk, dst, src, nbytes); in ecb_encrypt()
176 err = blkcipher_walk_virt(desc, &walk); in ecb_encrypt()
179 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt()
180 ce_aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_encrypt()
182 err = blkcipher_walk_done(desc, &walk, in ecb_encrypt()
183 walk.nbytes % AES_BLOCK_SIZE); in ecb_encrypt()
193 struct blkcipher_walk walk; in ecb_decrypt() local
198 blkcipher_walk_init(&walk, dst, src, nbytes); in ecb_decrypt()
199 err = blkcipher_walk_virt(desc, &walk); in ecb_decrypt()
202 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_decrypt()
203 ce_aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_decrypt()
205 err = blkcipher_walk_done(desc, &walk, in ecb_decrypt()
206 walk.nbytes % AES_BLOCK_SIZE); in ecb_decrypt()
216 struct blkcipher_walk walk; in cbc_encrypt() local
221 blkcipher_walk_init(&walk, dst, src, nbytes); in cbc_encrypt()
222 err = blkcipher_walk_virt(desc, &walk); in cbc_encrypt()
225 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in cbc_encrypt()
226 ce_aes_cbc_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cbc_encrypt()
228 walk.iv); in cbc_encrypt()
229 err = blkcipher_walk_done(desc, &walk, in cbc_encrypt()
230 walk.nbytes % AES_BLOCK_SIZE); in cbc_encrypt()
240 struct blkcipher_walk walk; in cbc_decrypt() local
245 blkcipher_walk_init(&walk, dst, src, nbytes); in cbc_decrypt()
246 err = blkcipher_walk_virt(desc, &walk); in cbc_decrypt()
249 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in cbc_decrypt()
250 ce_aes_cbc_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cbc_decrypt()
252 walk.iv); in cbc_decrypt()
253 err = blkcipher_walk_done(desc, &walk, in cbc_decrypt()
254 walk.nbytes % AES_BLOCK_SIZE); in cbc_decrypt()
264 struct blkcipher_walk walk; in ctr_encrypt() local
268 blkcipher_walk_init(&walk, dst, src, nbytes); in ctr_encrypt()
269 err = blkcipher_walk_virt_block(desc, &walk, AES_BLOCK_SIZE); in ctr_encrypt()
272 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ctr_encrypt()
273 ce_aes_ctr_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ctr_encrypt()
275 walk.iv); in ctr_encrypt()
277 if (nbytes && nbytes == walk.nbytes % AES_BLOCK_SIZE) in ctr_encrypt()
279 err = blkcipher_walk_done(desc, &walk, in ctr_encrypt()
280 walk.nbytes % AES_BLOCK_SIZE); in ctr_encrypt()
283 u8 *tdst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; in ctr_encrypt()
284 u8 *tsrc = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; in ctr_encrypt()
294 num_rounds(ctx), blocks, walk.iv); in ctr_encrypt()
296 err = blkcipher_walk_done(desc, &walk, 0); in ctr_encrypt()
308 struct blkcipher_walk walk; in xts_encrypt() local
312 blkcipher_walk_init(&walk, dst, src, nbytes); in xts_encrypt()
313 err = blkcipher_walk_virt(desc, &walk); in xts_encrypt()
316 for (first = 1; (blocks = (walk.nbytes / AES_BLOCK_SIZE)); first = 0) { in xts_encrypt()
317 ce_aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt()
319 walk.iv, (u8 *)ctx->key2.key_enc, first); in xts_encrypt()
320 err = blkcipher_walk_done(desc, &walk, in xts_encrypt()
321 walk.nbytes % AES_BLOCK_SIZE); in xts_encrypt()
333 struct blkcipher_walk walk; in xts_decrypt() local
337 blkcipher_walk_init(&walk, dst, src, nbytes); in xts_decrypt()
338 err = blkcipher_walk_virt(desc, &walk); in xts_decrypt()
341 for (first = 1; (blocks = (walk.nbytes / AES_BLOCK_SIZE)); first = 0) { in xts_decrypt()
342 ce_aes_xts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_decrypt()
344 walk.iv, (u8 *)ctx->key2.key_enc, first); in xts_decrypt()
345 err = blkcipher_walk_done(desc, &walk, in xts_decrypt()
346 walk.nbytes % AES_BLOCK_SIZE); in xts_decrypt()