/linux-4.1.27/include/crypto/ |
H A D | sha256_base.h | 61 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; sha256_base_do_update() 65 if (unlikely((partial + len) >= SHA256_BLOCK_SIZE)) { sha256_base_do_update() 69 int p = SHA256_BLOCK_SIZE - partial; sha256_base_do_update() 78 blocks = len / SHA256_BLOCK_SIZE; sha256_base_do_update() 79 len %= SHA256_BLOCK_SIZE; sha256_base_do_update() 83 data += blocks * SHA256_BLOCK_SIZE; sha256_base_do_update() 96 const int bit_offset = SHA256_BLOCK_SIZE - sizeof(__be64); sha256_base_do_finalize() 99 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; sha256_base_do_finalize() 103 memset(sctx->buf + partial, 0x0, SHA256_BLOCK_SIZE - partial); sha256_base_do_finalize()
|
H A D | sha.h | 17 #define SHA256_BLOCK_SIZE 64 macro 76 u8 buf[SHA256_BLOCK_SIZE];
|
/linux-4.1.27/arch/sparc/crypto/ |
H A D | sha256_glue.c | 68 done = SHA256_BLOCK_SIZE - partial; __sha256_sparc64_update() 72 if (len - done >= SHA256_BLOCK_SIZE) { __sha256_sparc64_update() 73 const unsigned int rounds = (len - done) / SHA256_BLOCK_SIZE; __sha256_sparc64_update() 76 done += rounds * SHA256_BLOCK_SIZE; __sha256_sparc64_update() 86 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; sha256_sparc64_update() 89 if (partial + len < SHA256_BLOCK_SIZE) { sha256_sparc64_update() 104 static const u8 padding[SHA256_BLOCK_SIZE] = { 0x80, }; sha256_sparc64_final() 109 index = sctx->count % SHA256_BLOCK_SIZE; sha256_sparc64_final() 110 padlen = (index < 56) ? (56 - index) : ((SHA256_BLOCK_SIZE+56) - index); sha256_sparc64_final() 173 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/arch/arm/crypto/ |
H A D | sha2-ce-glue.c | 37 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) sha2_ce_update() 81 .cra_blocksize = SHA256_BLOCK_SIZE, 96 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | sha256_neon_glue.c | 38 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) sha256_update() 83 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | sha256_glue.c | 75 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/drivers/crypto/nx/ |
H A D | nx-sha256.c | 80 u64 buf_len = (sctx->count % SHA256_BLOCK_SIZE); nx_sha256_update() 85 * 1: < SHA256_BLOCK_SIZE: copy into state, return 0 nx_sha256_update() 86 * 2: >= SHA256_BLOCK_SIZE: process X blocks, copy in leftover nx_sha256_update() 88 total = (sctx->count % SHA256_BLOCK_SIZE) + len; nx_sha256_update() 89 if (total < SHA256_BLOCK_SIZE) { nx_sha256_update() 132 /* to_process: SHA256_BLOCK_SIZE aligned chunk to be nx_sha256_update() 141 to_process = to_process & ~(SHA256_BLOCK_SIZE - 1); nx_sha256_update() 176 } while (leftover >= SHA256_BLOCK_SIZE); nx_sha256_update() 209 if (sctx->count >= SHA256_BLOCK_SIZE) { nx_sha256_final() 222 len = sctx->count & (SHA256_BLOCK_SIZE - 1); nx_sha256_final() 226 if (len != (sctx->count & (SHA256_BLOCK_SIZE - 1))) { nx_sha256_final() 292 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/arch/mips/cavium-octeon/crypto/ |
H A D | octeon-sha256.c | 111 partial = sctx->count % SHA256_BLOCK_SIZE; __octeon_sha256_update() 116 if ((partial + len) >= SHA256_BLOCK_SIZE) { __octeon_sha256_update() 120 done + SHA256_BLOCK_SIZE); __octeon_sha256_update() 126 done += SHA256_BLOCK_SIZE; __octeon_sha256_update() 128 } while (done + SHA256_BLOCK_SIZE <= len); __octeon_sha256_update() 147 if ((sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) octeon_sha256_update() 243 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/arch/arm64/crypto/ |
H A D | sha2-ce-glue.c | 53 bool finalize = !sctx->sst.count && !(len % SHA256_BLOCK_SIZE); sha256_ce_finup() 99 .cra_blocksize = SHA256_BLOCK_SIZE, 114 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/arch/x86/crypto/ |
H A D | sha256_ssse3_glue.c | 64 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) sha256_ssse3_update() 112 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/drivers/crypto/ |
H A D | padlock-sha.c | 169 leftover = ((state.count - 1) & (SHA256_BLOCK_SIZE - 1)) + 1; padlock_sha256_finup() 170 space = SHA256_BLOCK_SIZE - leftover; padlock_sha256_finup() 284 .cra_blocksize = SHA256_BLOCK_SIZE, 412 if ((partial + len) >= SHA256_BLOCK_SIZE) { padlock_sha256_update_nano() 418 done + SHA256_BLOCK_SIZE); padlock_sha256_update_nano() 425 done += SHA256_BLOCK_SIZE; padlock_sha256_update_nano() 430 if (len - done >= SHA256_BLOCK_SIZE) { padlock_sha256_update_nano() 525 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | omap-sham.c | 424 d = SHA256_BLOCK_SIZE; get_block_size() 871 bs = SHA256_BLOCK_SIZE; omap_sham_init() 1453 .cra_blocksize = SHA256_BLOCK_SIZE, 1500 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | sahara.c | 38 #define SAHARA_MAX_SHA_BLOCK_SIZE SHA256_BLOCK_SIZE 1259 SHA_BUFFER_LEN + SHA256_BLOCK_SIZE); sahara_sha_cra_init() local 1361 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | atmel-sha.c | 294 ctx->block_size = SHA256_BLOCK_SIZE; atmel_sha_init() 1027 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | img-hash.c | 793 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | mxs-dcp.c | 869 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | talitos.c | 2242 .cra_blocksize = SHA256_BLOCK_SIZE, 2332 .cra_blocksize = SHA256_BLOCK_SIZE,
|
H A D | n2_core.c | 1342 .block_size = SHA256_BLOCK_SIZE },
|
/linux-4.1.27/drivers/crypto/qce/ |
H A D | sha.h | 23 #define QCE_SHA_MAX_BLOCKSIZE SHA256_BLOCK_SIZE
|
H A D | sha.c | 475 .blocksize = SHA256_BLOCK_SIZE, 493 .blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/arch/s390/crypto/ |
H A D | sha_common.c | 72 plen = (bsize > SHA256_BLOCK_SIZE) ? 16 : 8; s390_sha_final()
|
H A D | sha256_s390.c | 79 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/drivers/crypto/ccp/ |
H A D | ccp-crypto.h | 143 #define MAX_SHA_BLOCK_SIZE SHA256_BLOCK_SIZE
|
H A D | ccp-crypto-sha.c | 361 .block_size = SHA256_BLOCK_SIZE,
|
H A D | ccp-ops.c | 1513 block_size = SHA256_BLOCK_SIZE; ccp_run_sha_cmd()
|
/linux-4.1.27/crypto/ |
H A D | sha256_generic.c | 223 src += SHA256_BLOCK_SIZE; sha256_generic_block_fn() 259 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/arch/powerpc/crypto/ |
H A D | sha256-spe-glue.c | 234 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/include/linux/ |
H A D | ccp.h | 223 #define CCP_SHA_BLOCKSIZE SHA256_BLOCK_SIZE
|
/linux-4.1.27/drivers/crypto/ux500/hash/ |
H A D | hash_core.c | 1555 .cra_blocksize = SHA256_BLOCK_SIZE, 1603 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-4.1.27/drivers/crypto/caam/ |
H A D | caamhash.c | 1659 .blocksize = SHA256_BLOCK_SIZE,
|