xts 553 arch/s390/crypto/aes_s390.c u8 xts[16]; xts 564 arch/s390/crypto/aes_s390.c memset(pcc_param.xts, 0, sizeof(pcc_param.xts)); xts 570 arch/s390/crypto/aes_s390.c memcpy(xts_param.init, pcc_param.xts, 16); xts 444 arch/s390/crypto/paes_s390.c u8 xts[16]; xts 461 arch/s390/crypto/paes_s390.c memcpy(xts_param.init, pcc_param.xts, 16); xts 584 arch/x86/crypto/aesni-intel_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_enc8) } xts 587 arch/x86/crypto/aesni-intel_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_enc) } xts 597 arch/x86/crypto/aesni-intel_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_dec8) } xts 600 arch/x86/crypto/aesni-intel_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_dec) } xts 81 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_32way) } xts 84 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_16way) } xts 87 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc) } xts 135 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_32way) } xts 138 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_16way) } xts 141 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec) } xts 97 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_16way) } xts 100 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc) } xts 142 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_16way) } xts 145 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec) } xts 99 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_enc_8way) } xts 102 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_enc) } xts 138 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_dec_8way) } xts 141 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_dec) } xts 243 arch/x86/crypto/glue_helper.c gctx->funcs[i].fn_u.xts(ctx, dst, src, xts 79 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_16way) } xts 82 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_8way_avx) } xts 85 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc) } xts 127 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_16way) } xts 130 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_8way_avx) } xts 133 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec) } xts 131 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_8way_avx) } xts 134 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc) } xts 170 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_8way_avx) } xts 173 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec) } xts 128 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_enc_8way) } xts 131 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_enc) } xts 173 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_dec_8way) } xts 176 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_dec) } xts 32 arch/x86/include/asm/crypto/glue_helper.h common_glue_xts_func_t xts; xts 171 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.type = CCP_AES_TYPE_128; xts 172 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.action = (encrypt) ? CCP_AES_ACTION_ENCRYPT xts 174 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.unit_size = unit_size; xts 175 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.key = &ctx->u.aes.key_sg; xts 176 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.key_len = ctx->u.aes.key_len; xts 177 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.iv = &rctx->iv_sg; xts 178 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.iv_len = AES_BLOCK_SIZE; xts 179 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.src = req->src; xts 180 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.src_len = req->nbytes; xts 181 drivers/crypto/ccp/ccp-crypto-aes-xts.c rctx->cmd.u.xts.dst = req->dst; xts 185 drivers/crypto/ccp/ccp-dev-v3.c | (op->u.xts.action << REQ1_AES_ACTION_SHIFT) xts 186 drivers/crypto/ccp/ccp-dev-v3.c | (op->u.xts.unit_size << REQ1_XTS_AES_SIZE_SHIFT) xts 342 drivers/crypto/ccp/ccp-dev-v5.c CCP_XTS_TYPE(&function) = op->u.xts.type; xts 343 drivers/crypto/ccp/ccp-dev-v5.c CCP_XTS_ENCRYPT(&function) = op->u.xts.action; xts 344 drivers/crypto/ccp/ccp-dev-v5.c CCP_XTS_SIZE(&function) = op->u.xts.unit_size; xts 547 drivers/crypto/ccp/ccp-dev.h struct ccp_xts_aes_op xts; xts 1075 drivers/crypto/ccp/ccp-ops.c struct ccp_xts_aes_engine *xts = &cmd->u.xts; xts 1085 drivers/crypto/ccp/ccp-ops.c switch (xts->unit_size) { xts 1106 drivers/crypto/ccp/ccp-ops.c if (xts->key_len == AES_KEYSIZE_128) xts 1108 drivers/crypto/ccp/ccp-ops.c else if (xts->key_len == AES_KEYSIZE_256) xts 1113 drivers/crypto/ccp/ccp-ops.c if (!xts->final && (xts->src_len & (AES_BLOCK_SIZE - 1))) xts 1116 drivers/crypto/ccp/ccp-ops.c if (xts->iv_len != AES_BLOCK_SIZE) xts 1119 drivers/crypto/ccp/ccp-ops.c if (!xts->key || !xts->iv || !xts->src || !xts->dst) xts 1132 drivers/crypto/ccp/ccp-ops.c op.u.xts.type = aestype; xts 1133 drivers/crypto/ccp/ccp-ops.c op.u.xts.action = xts->action; xts 1134 drivers/crypto/ccp/ccp-ops.c op.u.xts.unit_size = xts->unit_size; xts 1156 drivers/crypto/ccp/ccp-ops.c ret = ccp_set_dm_area(&key, dm_offset, xts->key, 0, xts->key_len); xts 1159 drivers/crypto/ccp/ccp-ops.c ret = ccp_set_dm_area(&key, 0, xts->key, xts->key_len, xts->key_len); xts 1169 drivers/crypto/ccp/ccp-ops.c pad = dm_offset - xts->key_len; xts 1170 drivers/crypto/ccp/ccp-ops.c ret = ccp_set_dm_area(&key, pad, xts->key, 0, xts->key_len); xts 1173 drivers/crypto/ccp/ccp-ops.c ret = ccp_set_dm_area(&key, dm_offset + pad, xts->key, xts 1174 drivers/crypto/ccp/ccp-ops.c xts->key_len, xts->key_len); xts 1195 drivers/crypto/ccp/ccp-ops.c ret = ccp_set_dm_area(&ctx, 0, xts->iv, 0, xts->iv_len); xts 1209 drivers/crypto/ccp/ccp-ops.c if (sg_virt(xts->src) == sg_virt(xts->dst)) xts 1212 drivers/crypto/ccp/ccp-ops.c ret = ccp_init_data(&src, cmd_q, xts->src, xts->src_len, xts 1221 drivers/crypto/ccp/ccp-ops.c ret = ccp_init_data(&dst, cmd_q, xts->dst, xts->src_len, xts 1254 drivers/crypto/ccp/ccp-ops.c ccp_get_dm_area(&ctx, dm_offset, xts->iv, 0, xts->iv_len); xts 49 drivers/crypto/inside-secure/safexcel_cipher.c unsigned int key_len, xts; xts 465 drivers/crypto/inside-secure/safexcel_cipher.c switch (ctx->key_len >> ctx->xts) { xts 480 drivers/crypto/inside-secure/safexcel_cipher.c ctx->key_len >> ctx->xts); xts 2018 drivers/crypto/inside-secure/safexcel_cipher.c ctx->xts = 1; xts 652 include/linux/ccp.h struct ccp_xts_aes_engine xts;