CHACHA_BLOCK_SIZE   42 arch/arm/crypto/chacha-neon-glue.c 	u8 buf[CHACHA_BLOCK_SIZE];
CHACHA_BLOCK_SIZE   44 arch/arm/crypto/chacha-neon-glue.c 	while (bytes >= CHACHA_BLOCK_SIZE * 4) {
CHACHA_BLOCK_SIZE   46 arch/arm/crypto/chacha-neon-glue.c 		bytes -= CHACHA_BLOCK_SIZE * 4;
CHACHA_BLOCK_SIZE   47 arch/arm/crypto/chacha-neon-glue.c 		src += CHACHA_BLOCK_SIZE * 4;
CHACHA_BLOCK_SIZE   48 arch/arm/crypto/chacha-neon-glue.c 		dst += CHACHA_BLOCK_SIZE * 4;
CHACHA_BLOCK_SIZE   51 arch/arm/crypto/chacha-neon-glue.c 	while (bytes >= CHACHA_BLOCK_SIZE) {
CHACHA_BLOCK_SIZE   53 arch/arm/crypto/chacha-neon-glue.c 		bytes -= CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE   54 arch/arm/crypto/chacha-neon-glue.c 		src += CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE   55 arch/arm/crypto/chacha-neon-glue.c 		dst += CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE   97 arch/arm/crypto/chacha-neon-glue.c 	if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable())
CHACHA_BLOCK_SIZE  111 arch/arm/crypto/chacha-neon-glue.c 	if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable())
CHACHA_BLOCK_SIZE  138 arch/arm/crypto/chacha-neon-glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  139 arch/arm/crypto/chacha-neon-glue.c 		.walksize		= 4 * CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  154 arch/arm/crypto/chacha-neon-glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  155 arch/arm/crypto/chacha-neon-glue.c 		.walksize		= 4 * CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  170 arch/arm/crypto/chacha-neon-glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  171 arch/arm/crypto/chacha-neon-glue.c 		.walksize		= 4 * CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE   43 arch/arm64/crypto/chacha-neon-glue.c 		int l = min(bytes, CHACHA_BLOCK_SIZE * 5);
CHACHA_BLOCK_SIZE   45 arch/arm64/crypto/chacha-neon-glue.c 		if (l <= CHACHA_BLOCK_SIZE) {
CHACHA_BLOCK_SIZE   46 arch/arm64/crypto/chacha-neon-glue.c 			u8 buf[CHACHA_BLOCK_SIZE];
CHACHA_BLOCK_SIZE   55 arch/arm64/crypto/chacha-neon-glue.c 		bytes -= CHACHA_BLOCK_SIZE * 5;
CHACHA_BLOCK_SIZE   56 arch/arm64/crypto/chacha-neon-glue.c 		src += CHACHA_BLOCK_SIZE * 5;
CHACHA_BLOCK_SIZE   57 arch/arm64/crypto/chacha-neon-glue.c 		dst += CHACHA_BLOCK_SIZE * 5;
CHACHA_BLOCK_SIZE   94 arch/arm64/crypto/chacha-neon-glue.c 	if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable())
CHACHA_BLOCK_SIZE  108 arch/arm64/crypto/chacha-neon-glue.c 	if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable())
CHACHA_BLOCK_SIZE  135 arch/arm64/crypto/chacha-neon-glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  136 arch/arm64/crypto/chacha-neon-glue.c 		.walksize		= 5 * CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  151 arch/arm64/crypto/chacha-neon-glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  152 arch/arm64/crypto/chacha-neon-glue.c 		.walksize		= 5 * CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  167 arch/arm64/crypto/chacha-neon-glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  168 arch/arm64/crypto/chacha-neon-glue.c 		.walksize		= 5 * CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE   45 arch/x86/crypto/chacha_glue.c 	len = min(len, maxblocks * CHACHA_BLOCK_SIZE);
CHACHA_BLOCK_SIZE   46 arch/x86/crypto/chacha_glue.c 	return round_up(len, CHACHA_BLOCK_SIZE) / CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE   55 arch/x86/crypto/chacha_glue.c 		while (bytes >= CHACHA_BLOCK_SIZE * 8) {
CHACHA_BLOCK_SIZE   58 arch/x86/crypto/chacha_glue.c 			bytes -= CHACHA_BLOCK_SIZE * 8;
CHACHA_BLOCK_SIZE   59 arch/x86/crypto/chacha_glue.c 			src += CHACHA_BLOCK_SIZE * 8;
CHACHA_BLOCK_SIZE   60 arch/x86/crypto/chacha_glue.c 			dst += CHACHA_BLOCK_SIZE * 8;
CHACHA_BLOCK_SIZE   63 arch/x86/crypto/chacha_glue.c 		if (bytes > CHACHA_BLOCK_SIZE * 4) {
CHACHA_BLOCK_SIZE   69 arch/x86/crypto/chacha_glue.c 		if (bytes > CHACHA_BLOCK_SIZE * 2) {
CHACHA_BLOCK_SIZE   84 arch/x86/crypto/chacha_glue.c 		while (bytes >= CHACHA_BLOCK_SIZE * 8) {
CHACHA_BLOCK_SIZE   86 arch/x86/crypto/chacha_glue.c 			bytes -= CHACHA_BLOCK_SIZE * 8;
CHACHA_BLOCK_SIZE   87 arch/x86/crypto/chacha_glue.c 			src += CHACHA_BLOCK_SIZE * 8;
CHACHA_BLOCK_SIZE   88 arch/x86/crypto/chacha_glue.c 			dst += CHACHA_BLOCK_SIZE * 8;
CHACHA_BLOCK_SIZE   91 arch/x86/crypto/chacha_glue.c 		if (bytes > CHACHA_BLOCK_SIZE * 4) {
CHACHA_BLOCK_SIZE   96 arch/x86/crypto/chacha_glue.c 		if (bytes > CHACHA_BLOCK_SIZE * 2) {
CHACHA_BLOCK_SIZE  101 arch/x86/crypto/chacha_glue.c 		if (bytes > CHACHA_BLOCK_SIZE) {
CHACHA_BLOCK_SIZE  108 arch/x86/crypto/chacha_glue.c 	while (bytes >= CHACHA_BLOCK_SIZE * 4) {
CHACHA_BLOCK_SIZE  110 arch/x86/crypto/chacha_glue.c 		bytes -= CHACHA_BLOCK_SIZE * 4;
CHACHA_BLOCK_SIZE  111 arch/x86/crypto/chacha_glue.c 		src += CHACHA_BLOCK_SIZE * 4;
CHACHA_BLOCK_SIZE  112 arch/x86/crypto/chacha_glue.c 		dst += CHACHA_BLOCK_SIZE * 4;
CHACHA_BLOCK_SIZE  115 arch/x86/crypto/chacha_glue.c 	if (bytes > CHACHA_BLOCK_SIZE) {
CHACHA_BLOCK_SIZE  169 arch/x86/crypto/chacha_glue.c 	if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable())
CHACHA_BLOCK_SIZE  192 arch/x86/crypto/chacha_glue.c 	if (req->cryptlen <= CHACHA_BLOCK_SIZE || !crypto_simd_usable())
CHACHA_BLOCK_SIZE  229 arch/x86/crypto/chacha_glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  244 arch/x86/crypto/chacha_glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  259 arch/x86/crypto/chacha_glue.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  382 crypto/adiantum.c 	if (round_up(stream_len, CHACHA_BLOCK_SIZE) <= req->cryptlen)
CHACHA_BLOCK_SIZE  383 crypto/adiantum.c 		stream_len = round_up(stream_len, CHACHA_BLOCK_SIZE);
CHACHA_BLOCK_SIZE   19 crypto/chacha_generic.c 	u8 stream[CHACHA_BLOCK_SIZE] __aligned(sizeof(long));
CHACHA_BLOCK_SIZE   21 crypto/chacha_generic.c 	while (bytes >= CHACHA_BLOCK_SIZE) {
CHACHA_BLOCK_SIZE   23 crypto/chacha_generic.c 		crypto_xor_cpy(dst, src, stream, CHACHA_BLOCK_SIZE);
CHACHA_BLOCK_SIZE   24 crypto/chacha_generic.c 		bytes -= CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE   25 crypto/chacha_generic.c 		dst += CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE   26 crypto/chacha_generic.c 		src += CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE   49 crypto/chacha_generic.c 			nbytes = round_down(nbytes, CHACHA_BLOCK_SIZE);
CHACHA_BLOCK_SIZE  153 crypto/chacha_generic.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  168 crypto/chacha_generic.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  183 crypto/chacha_generic.c 		.chunksize		= CHACHA_BLOCK_SIZE,
CHACHA_BLOCK_SIZE  504 drivers/char/random.c static void _extract_crng(struct crng_state *crng, __u8 out[CHACHA_BLOCK_SIZE]);
CHACHA_BLOCK_SIZE  506 drivers/char/random.c 				    __u8 tmp[CHACHA_BLOCK_SIZE], int used);
CHACHA_BLOCK_SIZE 1005 drivers/char/random.c 		__u8	block[CHACHA_BLOCK_SIZE];
CHACHA_BLOCK_SIZE 1052 drivers/char/random.c 			  __u8 out[CHACHA_BLOCK_SIZE])
CHACHA_BLOCK_SIZE 1069 drivers/char/random.c static void extract_crng(__u8 out[CHACHA_BLOCK_SIZE])
CHACHA_BLOCK_SIZE 1087 drivers/char/random.c 				    __u8 tmp[CHACHA_BLOCK_SIZE], int used)
CHACHA_BLOCK_SIZE 1094 drivers/char/random.c 	if (used + CHACHA_KEY_SIZE > CHACHA_BLOCK_SIZE) {
CHACHA_BLOCK_SIZE 1106 drivers/char/random.c static void crng_backtrack_protect(__u8 tmp[CHACHA_BLOCK_SIZE], int used)
CHACHA_BLOCK_SIZE 1121 drivers/char/random.c 	ssize_t ret = 0, i = CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE 1122 drivers/char/random.c 	__u8 tmp[CHACHA_BLOCK_SIZE] __aligned(4);
CHACHA_BLOCK_SIZE 1136 drivers/char/random.c 		i = min_t(int, nbytes, CHACHA_BLOCK_SIZE);
CHACHA_BLOCK_SIZE 1707 drivers/char/random.c 	__u8 tmp[CHACHA_BLOCK_SIZE] __aligned(4);
CHACHA_BLOCK_SIZE 1711 drivers/char/random.c 	while (nbytes >= CHACHA_BLOCK_SIZE) {
CHACHA_BLOCK_SIZE 1713 drivers/char/random.c 		buf += CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE 1714 drivers/char/random.c 		nbytes -= CHACHA_BLOCK_SIZE;
CHACHA_BLOCK_SIZE 1722 drivers/char/random.c 		crng_backtrack_protect(tmp, CHACHA_BLOCK_SIZE);
CHACHA_BLOCK_SIZE 2352 drivers/char/random.c 		u64 entropy_u64[CHACHA_BLOCK_SIZE / sizeof(u64)];
CHACHA_BLOCK_SIZE 2353 drivers/char/random.c 		u32 entropy_u32[CHACHA_BLOCK_SIZE / sizeof(u32)];