XTS_BLOCK_SIZE    592 arch/s390/crypto/aes_s390.c 	if (unlikely(!xts_ctx->fc || (nbytes % XTS_BLOCK_SIZE) != 0))
XTS_BLOCK_SIZE    609 arch/s390/crypto/aes_s390.c 	if (unlikely(!xts_ctx->fc || (nbytes % XTS_BLOCK_SIZE) != 0))
XTS_BLOCK_SIZE    265 arch/x86/crypto/glue_helper.c 	const bool cts = (req->cryptlen % XTS_BLOCK_SIZE);
XTS_BLOCK_SIZE    273 arch/x86/crypto/glue_helper.c 	if (req->cryptlen < XTS_BLOCK_SIZE)
XTS_BLOCK_SIZE    279 arch/x86/crypto/glue_helper.c 		tail = req->cryptlen % XTS_BLOCK_SIZE + XTS_BLOCK_SIZE;
XTS_BLOCK_SIZE    321 arch/x86/crypto/glue_helper.c 			next_tweak = memcpy(b, req->iv, XTS_BLOCK_SIZE);
XTS_BLOCK_SIZE    327 arch/x86/crypto/glue_helper.c 		skcipher_request_set_crypt(&subreq, src, dst, XTS_BLOCK_SIZE,
XTS_BLOCK_SIZE    336 arch/x86/crypto/glue_helper.c 		scatterwalk_map_and_copy(b, dst, 0, XTS_BLOCK_SIZE, 0);
XTS_BLOCK_SIZE    337 arch/x86/crypto/glue_helper.c 		memcpy(b + 1, b, tail - XTS_BLOCK_SIZE);
XTS_BLOCK_SIZE    338 arch/x86/crypto/glue_helper.c 		scatterwalk_map_and_copy(b, src, XTS_BLOCK_SIZE,
XTS_BLOCK_SIZE    339 arch/x86/crypto/glue_helper.c 					 tail - XTS_BLOCK_SIZE, 0);
XTS_BLOCK_SIZE    342 arch/x86/crypto/glue_helper.c 		skcipher_request_set_crypt(&subreq, dst, dst, XTS_BLOCK_SIZE,
XTS_BLOCK_SIZE     91 crypto/xts.c   	const bool cts = (req->cryptlen % XTS_BLOCK_SIZE);
XTS_BLOCK_SIZE     92 crypto/xts.c   	const int bs = XTS_BLOCK_SIZE;
XTS_BLOCK_SIZE    114 crypto/xts.c   			    w.total - w.nbytes + avail < 2 * XTS_BLOCK_SIZE) {
XTS_BLOCK_SIZE    155 crypto/xts.c   		scatterwalk_map_and_copy(&b, rctx->tail, 0, XTS_BLOCK_SIZE, 0);
XTS_BLOCK_SIZE    157 crypto/xts.c   		scatterwalk_map_and_copy(&b, rctx->tail, 0, XTS_BLOCK_SIZE, 1);
XTS_BLOCK_SIZE    167 crypto/xts.c   	int offset = req->cryptlen & ~(XTS_BLOCK_SIZE - 1);
XTS_BLOCK_SIZE    170 crypto/xts.c   	int tail = req->cryptlen % XTS_BLOCK_SIZE;
XTS_BLOCK_SIZE    175 crypto/xts.c   				      offset - XTS_BLOCK_SIZE);
XTS_BLOCK_SIZE    177 crypto/xts.c   	scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE, 0);
XTS_BLOCK_SIZE    183 crypto/xts.c   	scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE + tail, 1);
XTS_BLOCK_SIZE    188 crypto/xts.c   				   XTS_BLOCK_SIZE, NULL);
XTS_BLOCK_SIZE    194 crypto/xts.c   	scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE, 0);
XTS_BLOCK_SIZE    196 crypto/xts.c   	scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE, 1);
XTS_BLOCK_SIZE    211 crypto/xts.c   		if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) {
XTS_BLOCK_SIZE    231 crypto/xts.c   		if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) {
XTS_BLOCK_SIZE    247 crypto/xts.c   	if (req->cryptlen < XTS_BLOCK_SIZE)
XTS_BLOCK_SIZE    253 crypto/xts.c   				   req->cryptlen & ~(XTS_BLOCK_SIZE - 1), NULL);
XTS_BLOCK_SIZE    272 crypto/xts.c   	if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0))
XTS_BLOCK_SIZE    289 crypto/xts.c   	if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0))
XTS_BLOCK_SIZE    386 crypto/xts.c   	if (alg->base.cra_blocksize != XTS_BLOCK_SIZE)
XTS_BLOCK_SIZE    425 crypto/xts.c   	inst->alg.base.cra_blocksize = XTS_BLOCK_SIZE;
XTS_BLOCK_SIZE    429 crypto/xts.c   	inst->alg.ivsize = XTS_BLOCK_SIZE;
XTS_BLOCK_SIZE   2025 drivers/crypto/inside-secure/safexcel_cipher.c 	if (req->cryptlen < XTS_BLOCK_SIZE)
XTS_BLOCK_SIZE   2033 drivers/crypto/inside-secure/safexcel_cipher.c 	if (req->cryptlen < XTS_BLOCK_SIZE)
XTS_BLOCK_SIZE   2049 drivers/crypto/inside-secure/safexcel_cipher.c 		.ivsize = XTS_BLOCK_SIZE,
XTS_BLOCK_SIZE   2056 drivers/crypto/inside-secure/safexcel_cipher.c 			.cra_blocksize = XTS_BLOCK_SIZE,
XTS_BLOCK_SIZE     90 drivers/crypto/vmx/aes_xts.c 	if (!crypto_simd_usable() || (req->cryptlen % XTS_BLOCK_SIZE) != 0) {