AEGIS128_BLOCK_SIZE   48 arch/x86/crypto/aegis128-aesni-glue.c 	u8 bytes[AEGIS128_BLOCK_SIZE] __aligned(AEGIS128_BLOCK_ALIGN);
AEGIS128_BLOCK_SIZE   84 arch/x86/crypto/aegis128-aesni-glue.c 		if (pos + size >= AEGIS128_BLOCK_SIZE) {
AEGIS128_BLOCK_SIZE   86 arch/x86/crypto/aegis128-aesni-glue.c 				unsigned int fill = AEGIS128_BLOCK_SIZE - pos;
AEGIS128_BLOCK_SIZE   89 arch/x86/crypto/aegis128-aesni-glue.c 							 AEGIS128_BLOCK_SIZE,
AEGIS128_BLOCK_SIZE   98 arch/x86/crypto/aegis128-aesni-glue.c 			src += left & ~(AEGIS128_BLOCK_SIZE - 1);
AEGIS128_BLOCK_SIZE   99 arch/x86/crypto/aegis128-aesni-glue.c 			left &= AEGIS128_BLOCK_SIZE - 1;
AEGIS128_BLOCK_SIZE  112 arch/x86/crypto/aegis128-aesni-glue.c 		memset(buf.bytes + pos, 0, AEGIS128_BLOCK_SIZE - pos);
AEGIS128_BLOCK_SIZE  113 arch/x86/crypto/aegis128-aesni-glue.c 		crypto_aegis128_aesni_ad(state, AEGIS128_BLOCK_SIZE, buf.bytes);
AEGIS128_BLOCK_SIZE  121 arch/x86/crypto/aegis128-aesni-glue.c 	while (walk->nbytes >= AEGIS128_BLOCK_SIZE) {
AEGIS128_BLOCK_SIZE  123 arch/x86/crypto/aegis128-aesni-glue.c 				  round_down(walk->nbytes, AEGIS128_BLOCK_SIZE),
AEGIS128_BLOCK_SIZE  125 arch/x86/crypto/aegis128-aesni-glue.c 		skcipher_walk_done(walk, walk->nbytes % AEGIS128_BLOCK_SIZE);
AEGIS128_BLOCK_SIZE  251 arch/x86/crypto/aegis128-aesni-glue.c 	.chunksize = AEGIS128_BLOCK_SIZE,