AEGIS128_BLOCK_SIZE 48 arch/x86/crypto/aegis128-aesni-glue.c u8 bytes[AEGIS128_BLOCK_SIZE] __aligned(AEGIS128_BLOCK_ALIGN); AEGIS128_BLOCK_SIZE 84 arch/x86/crypto/aegis128-aesni-glue.c if (pos + size >= AEGIS128_BLOCK_SIZE) { AEGIS128_BLOCK_SIZE 86 arch/x86/crypto/aegis128-aesni-glue.c unsigned int fill = AEGIS128_BLOCK_SIZE - pos; AEGIS128_BLOCK_SIZE 89 arch/x86/crypto/aegis128-aesni-glue.c AEGIS128_BLOCK_SIZE, AEGIS128_BLOCK_SIZE 98 arch/x86/crypto/aegis128-aesni-glue.c src += left & ~(AEGIS128_BLOCK_SIZE - 1); AEGIS128_BLOCK_SIZE 99 arch/x86/crypto/aegis128-aesni-glue.c left &= AEGIS128_BLOCK_SIZE - 1; AEGIS128_BLOCK_SIZE 112 arch/x86/crypto/aegis128-aesni-glue.c memset(buf.bytes + pos, 0, AEGIS128_BLOCK_SIZE - pos); AEGIS128_BLOCK_SIZE 113 arch/x86/crypto/aegis128-aesni-glue.c crypto_aegis128_aesni_ad(state, AEGIS128_BLOCK_SIZE, buf.bytes); AEGIS128_BLOCK_SIZE 121 arch/x86/crypto/aegis128-aesni-glue.c while (walk->nbytes >= AEGIS128_BLOCK_SIZE) { AEGIS128_BLOCK_SIZE 123 arch/x86/crypto/aegis128-aesni-glue.c round_down(walk->nbytes, AEGIS128_BLOCK_SIZE), AEGIS128_BLOCK_SIZE 125 arch/x86/crypto/aegis128-aesni-glue.c skcipher_walk_done(walk, walk->nbytes % AEGIS128_BLOCK_SIZE); AEGIS128_BLOCK_SIZE 251 arch/x86/crypto/aegis128-aesni-glue.c .chunksize = AEGIS128_BLOCK_SIZE,