CHACHA_BLOCK_SIZE 42 arch/arm/crypto/chacha-neon-glue.c u8 buf[CHACHA_BLOCK_SIZE]; CHACHA_BLOCK_SIZE 44 arch/arm/crypto/chacha-neon-glue.c while (bytes >= CHACHA_BLOCK_SIZE * 4) { CHACHA_BLOCK_SIZE 46 arch/arm/crypto/chacha-neon-glue.c bytes -= CHACHA_BLOCK_SIZE * 4; CHACHA_BLOCK_SIZE 47 arch/arm/crypto/chacha-neon-glue.c src += CHACHA_BLOCK_SIZE * 4; CHACHA_BLOCK_SIZE 48 arch/arm/crypto/chacha-neon-glue.c dst += CHACHA_BLOCK_SIZE * 4; CHACHA_BLOCK_SIZE 51 arch/arm/crypto/chacha-neon-glue.c while (bytes >= CHACHA_BLOCK_SIZE) { CHACHA_BLOCK_SIZE 53 arch/arm/crypto/chacha-neon-glue.c bytes -= CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 54 arch/arm/crypto/chacha-neon-glue.c src += CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 55 arch/arm/crypto/chacha-neon-glue.c dst += CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 97 arch/arm/crypto/chacha-neon-glue.c if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable()) CHACHA_BLOCK_SIZE 111 arch/arm/crypto/chacha-neon-glue.c if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable()) CHACHA_BLOCK_SIZE 138 arch/arm/crypto/chacha-neon-glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 139 arch/arm/crypto/chacha-neon-glue.c .walksize = 4 * CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 154 arch/arm/crypto/chacha-neon-glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 155 arch/arm/crypto/chacha-neon-glue.c .walksize = 4 * CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 170 arch/arm/crypto/chacha-neon-glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 171 arch/arm/crypto/chacha-neon-glue.c .walksize = 4 * CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 43 arch/arm64/crypto/chacha-neon-glue.c int l = min(bytes, CHACHA_BLOCK_SIZE * 5); CHACHA_BLOCK_SIZE 45 arch/arm64/crypto/chacha-neon-glue.c if (l <= CHACHA_BLOCK_SIZE) { CHACHA_BLOCK_SIZE 46 arch/arm64/crypto/chacha-neon-glue.c u8 buf[CHACHA_BLOCK_SIZE]; CHACHA_BLOCK_SIZE 55 arch/arm64/crypto/chacha-neon-glue.c bytes -= CHACHA_BLOCK_SIZE * 5; CHACHA_BLOCK_SIZE 56 arch/arm64/crypto/chacha-neon-glue.c src += CHACHA_BLOCK_SIZE * 5; CHACHA_BLOCK_SIZE 57 arch/arm64/crypto/chacha-neon-glue.c dst += CHACHA_BLOCK_SIZE * 5; CHACHA_BLOCK_SIZE 94 arch/arm64/crypto/chacha-neon-glue.c if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable()) CHACHA_BLOCK_SIZE 108 arch/arm64/crypto/chacha-neon-glue.c if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable()) CHACHA_BLOCK_SIZE 135 arch/arm64/crypto/chacha-neon-glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 136 arch/arm64/crypto/chacha-neon-glue.c .walksize = 5 * CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 151 arch/arm64/crypto/chacha-neon-glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 152 arch/arm64/crypto/chacha-neon-glue.c .walksize = 5 * CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 167 arch/arm64/crypto/chacha-neon-glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 168 arch/arm64/crypto/chacha-neon-glue.c .walksize = 5 * CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 45 arch/x86/crypto/chacha_glue.c len = min(len, maxblocks * CHACHA_BLOCK_SIZE); CHACHA_BLOCK_SIZE 46 arch/x86/crypto/chacha_glue.c return round_up(len, CHACHA_BLOCK_SIZE) / CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 55 arch/x86/crypto/chacha_glue.c while (bytes >= CHACHA_BLOCK_SIZE * 8) { CHACHA_BLOCK_SIZE 58 arch/x86/crypto/chacha_glue.c bytes -= CHACHA_BLOCK_SIZE * 8; CHACHA_BLOCK_SIZE 59 arch/x86/crypto/chacha_glue.c src += CHACHA_BLOCK_SIZE * 8; CHACHA_BLOCK_SIZE 60 arch/x86/crypto/chacha_glue.c dst += CHACHA_BLOCK_SIZE * 8; CHACHA_BLOCK_SIZE 63 arch/x86/crypto/chacha_glue.c if (bytes > CHACHA_BLOCK_SIZE * 4) { CHACHA_BLOCK_SIZE 69 arch/x86/crypto/chacha_glue.c if (bytes > CHACHA_BLOCK_SIZE * 2) { CHACHA_BLOCK_SIZE 84 arch/x86/crypto/chacha_glue.c while (bytes >= CHACHA_BLOCK_SIZE * 8) { CHACHA_BLOCK_SIZE 86 arch/x86/crypto/chacha_glue.c bytes -= CHACHA_BLOCK_SIZE * 8; CHACHA_BLOCK_SIZE 87 arch/x86/crypto/chacha_glue.c src += CHACHA_BLOCK_SIZE * 8; CHACHA_BLOCK_SIZE 88 arch/x86/crypto/chacha_glue.c dst += CHACHA_BLOCK_SIZE * 8; CHACHA_BLOCK_SIZE 91 arch/x86/crypto/chacha_glue.c if (bytes > CHACHA_BLOCK_SIZE * 4) { CHACHA_BLOCK_SIZE 96 arch/x86/crypto/chacha_glue.c if (bytes > CHACHA_BLOCK_SIZE * 2) { CHACHA_BLOCK_SIZE 101 arch/x86/crypto/chacha_glue.c if (bytes > CHACHA_BLOCK_SIZE) { CHACHA_BLOCK_SIZE 108 arch/x86/crypto/chacha_glue.c while (bytes >= CHACHA_BLOCK_SIZE * 4) { CHACHA_BLOCK_SIZE 110 arch/x86/crypto/chacha_glue.c bytes -= CHACHA_BLOCK_SIZE * 4; CHACHA_BLOCK_SIZE 111 arch/x86/crypto/chacha_glue.c src += CHACHA_BLOCK_SIZE * 4; CHACHA_BLOCK_SIZE 112 arch/x86/crypto/chacha_glue.c dst += CHACHA_BLOCK_SIZE * 4; CHACHA_BLOCK_SIZE 115 arch/x86/crypto/chacha_glue.c if (bytes > CHACHA_BLOCK_SIZE) { CHACHA_BLOCK_SIZE 169 arch/x86/crypto/chacha_glue.c if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable()) CHACHA_BLOCK_SIZE 192 arch/x86/crypto/chacha_glue.c if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable()) CHACHA_BLOCK_SIZE 229 arch/x86/crypto/chacha_glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 244 arch/x86/crypto/chacha_glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 259 arch/x86/crypto/chacha_glue.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 382 crypto/adiantum.c if (round_up(stream_len, CHACHA_BLOCK_SIZE) <= req->cryptlen) CHACHA_BLOCK_SIZE 383 crypto/adiantum.c stream_len = round_up(stream_len, CHACHA_BLOCK_SIZE); CHACHA_BLOCK_SIZE 19 crypto/chacha_generic.c u8 stream[CHACHA_BLOCK_SIZE] __aligned(sizeof(long)); CHACHA_BLOCK_SIZE 21 crypto/chacha_generic.c while (bytes >= CHACHA_BLOCK_SIZE) { CHACHA_BLOCK_SIZE 23 crypto/chacha_generic.c crypto_xor_cpy(dst, src, stream, CHACHA_BLOCK_SIZE); CHACHA_BLOCK_SIZE 24 crypto/chacha_generic.c bytes -= CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 25 crypto/chacha_generic.c dst += CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 26 crypto/chacha_generic.c src += CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 49 crypto/chacha_generic.c nbytes = round_down(nbytes, CHACHA_BLOCK_SIZE); CHACHA_BLOCK_SIZE 153 crypto/chacha_generic.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 168 crypto/chacha_generic.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 183 crypto/chacha_generic.c .chunksize = CHACHA_BLOCK_SIZE, CHACHA_BLOCK_SIZE 504 drivers/char/random.c static void _extract_crng(struct crng_state *crng, __u8 out[CHACHA_BLOCK_SIZE]); CHACHA_BLOCK_SIZE 506 drivers/char/random.c __u8 tmp[CHACHA_BLOCK_SIZE], int used); CHACHA_BLOCK_SIZE 1005 drivers/char/random.c __u8 block[CHACHA_BLOCK_SIZE]; CHACHA_BLOCK_SIZE 1052 drivers/char/random.c __u8 out[CHACHA_BLOCK_SIZE]) CHACHA_BLOCK_SIZE 1069 drivers/char/random.c static void extract_crng(__u8 out[CHACHA_BLOCK_SIZE]) CHACHA_BLOCK_SIZE 1087 drivers/char/random.c __u8 tmp[CHACHA_BLOCK_SIZE], int used) CHACHA_BLOCK_SIZE 1094 drivers/char/random.c if (used + CHACHA_KEY_SIZE > CHACHA_BLOCK_SIZE) { CHACHA_BLOCK_SIZE 1106 drivers/char/random.c static void crng_backtrack_protect(__u8 tmp[CHACHA_BLOCK_SIZE], int used) CHACHA_BLOCK_SIZE 1121 drivers/char/random.c ssize_t ret = 0, i = CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 1122 drivers/char/random.c __u8 tmp[CHACHA_BLOCK_SIZE] __aligned(4); CHACHA_BLOCK_SIZE 1136 drivers/char/random.c i = min_t(int, nbytes, CHACHA_BLOCK_SIZE); CHACHA_BLOCK_SIZE 1707 drivers/char/random.c __u8 tmp[CHACHA_BLOCK_SIZE] __aligned(4); CHACHA_BLOCK_SIZE 1711 drivers/char/random.c while (nbytes >= CHACHA_BLOCK_SIZE) { CHACHA_BLOCK_SIZE 1713 drivers/char/random.c buf += CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 1714 drivers/char/random.c nbytes -= CHACHA_BLOCK_SIZE; CHACHA_BLOCK_SIZE 1722 drivers/char/random.c crng_backtrack_protect(tmp, CHACHA_BLOCK_SIZE); CHACHA_BLOCK_SIZE 2352 drivers/char/random.c u64 entropy_u64[CHACHA_BLOCK_SIZE / sizeof(u64)]; CHACHA_BLOCK_SIZE 2353 drivers/char/random.c u32 entropy_u32[CHACHA_BLOCK_SIZE / sizeof(u32)];