ecb 220 arch/s390/include/asm/kvm_host.h __u8 ecb; /* 0x0061 */ ecb 2986 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->ecb |= ECB_HOSTPROTINT; ecb 2988 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->ecb |= ECB_SRSI; ecb 2990 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->ecb |= ECB_TE; ecb 3917 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->ecb |= ECB_GS; ecb 23 arch/s390/kvm/kvm-s390.h #define IS_TE_ENABLED(vcpu) ((vcpu->arch.sie_block->ecb & ECB_TE)) ecb 67 arch/s390/kvm/priv.c vcpu->arch.sie_block->ecb |= ECB_GS; ecb 444 arch/s390/kvm/vsie.c const bool wants_tx = READ_ONCE(scb_o->ecb) & ECB_TE; ecb 445 arch/s390/kvm/vsie.c bool had_tx = scb_s->ecb & ECB_TE; ecb 452 arch/s390/kvm/vsie.c scb_s->ecb = 0; ecb 509 arch/s390/kvm/vsie.c scb_s->ecb |= scb_o->ecb & ECB_HOSTPROTINT; ecb 515 arch/s390/kvm/vsie.c scb_s->ecb |= ECB_TE; ecb 533 arch/s390/kvm/vsie.c scb_s->ecb |= scb_o->ecb & ECB_GS; ecb 622 arch/s390/kvm/vsie.c if (!rc && (scb_s->ecb & ECB_TE)) ecb 751 arch/s390/kvm/vsie.c if (gpa && (scb_s->ecb & ECB_TE)) { ecb 801 arch/s390/kvm/vsie.c if (((scb_s->ecb & ECB_GS) && !(scb_s->ecd & ECD_HOSTREGMGMT)) || ecb 43 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_enc_32way) } ecb 46 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_enc_16way) } ecb 49 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk_2way) } ecb 52 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk) } ecb 97 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_dec_32way) } ecb 100 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_dec_16way) } ecb 103 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk_2way) } ecb 106 arch/x86/crypto/camellia_aesni_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk) } ecb 65 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_enc_16way) } ecb 68 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk_2way) } ecb 71 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk) } ecb 110 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_ecb_dec_16way) } ecb 113 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk_2way) } ecb 116 arch/x86/crypto/camellia_aesni_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk) } ecb 1318 arch/x86/crypto/camellia_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk_2way) } ecb 1321 arch/x86/crypto/camellia_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_enc_blk) } ecb 1344 arch/x86/crypto/camellia_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk_2way) } ecb 1347 arch/x86/crypto/camellia_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(camellia_dec_blk) } ecb 73 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(cast6_ecb_enc_8way) } ecb 76 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__cast6_encrypt) } ecb 112 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(cast6_ecb_dec_8way) } ecb 115 arch/x86/crypto/cast6_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__cast6_decrypt) } ecb 49 arch/x86/crypto/glue_helper.c gctx->funcs[i].fn_u.ecb(ctx, dst, src); ecb 47 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_enc_16way) } ecb 50 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_enc_8way_avx) } ecb 53 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_encrypt) } ecb 95 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_dec_16way) } ecb 98 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_dec_8way_avx) } ecb 101 arch/x86/crypto/serpent_avx2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_decrypt) } ecb 105 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_enc_8way_avx) } ecb 108 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_encrypt) } ecb 144 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_dec_8way_avx) } ecb 147 arch/x86/crypto/serpent_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_decrypt) } ecb 82 arch/x86/crypto/serpent_sse2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_enc_blk_xway) } ecb 85 arch/x86/crypto/serpent_sse2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_encrypt) } ecb 108 arch/x86/crypto/serpent_sse2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_dec_blk_xway) } ecb 111 arch/x86/crypto/serpent_sse2_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_decrypt) } ecb 96 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_ecb_enc_8way) } ecb 99 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_enc_blk_3way) } ecb 102 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_enc_blk) } ecb 141 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_ecb_dec_8way) } ecb 144 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_dec_blk_3way) } ecb 147 arch/x86/crypto/twofish_avx_glue.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_dec_blk) } ecb 97 arch/x86/crypto/twofish_glue_3way.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_enc_blk_3way) } ecb 100 arch/x86/crypto/twofish_glue_3way.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_enc_blk) } ecb 110 arch/x86/crypto/twofish_glue_3way.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_enc_blk_ctr_3way) } ecb 113 arch/x86/crypto/twofish_glue_3way.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_enc_blk_ctr) } ecb 123 arch/x86/crypto/twofish_glue_3way.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_dec_blk_3way) } ecb 126 arch/x86/crypto/twofish_glue_3way.c .fn_u = { .ecb = GLUE_FUNC_CAST(twofish_dec_blk) } ecb 29 arch/x86/include/asm/crypto/glue_helper.h common_glue_func_t ecb; ecb 438 drivers/crypto/mediatek/mtk-aes.c goto ecb; ecb 459 drivers/crypto/mediatek/mtk-aes.c ecb: ecb 107 drivers/crypto/mxs-dcp.c unsigned int ecb:1; ecb 251 drivers/crypto/mxs-dcp.c if (rctx->ecb) ecb 307 drivers/crypto/mxs-dcp.c if (!rctx->ecb) { ecb 377 drivers/crypto/mxs-dcp.c if (!rctx->ecb) { ecb 448 drivers/crypto/mxs-dcp.c static int mxs_dcp_aes_enqueue(struct ablkcipher_request *req, int enc, int ecb) ecb 460 drivers/crypto/mxs-dcp.c rctx->ecb = ecb; ecb 77 drivers/scsi/aha1740.c struct ecb ecb[AHA1740_ECBS]; ecb 88 drivers/scsi/aha1740.c static inline struct ecb *ecb_dma_to_cpu (struct Scsi_Host *host, ecb 96 drivers/scsi/aha1740.c return (struct ecb *)(((char *) hdata->ecb) + (unsigned int) offset); ecb 104 drivers/scsi/aha1740.c offset = (char *) cpu - (char *) hdata->ecb; ecb 213 drivers/scsi/aha1740.c struct ecb *ecbptr; ecb 279 drivers/scsi/aha1740.c memset(ecbptr,0,sizeof(struct ecb)); ecb 353 drivers/scsi/aha1740.c if (!host->ecb[ecbno].cmdw) ecb 360 drivers/scsi/aha1740.c if (host->ecb[ecbno].cmdw) ecb 363 drivers/scsi/aha1740.c host->ecb[ecbno].cmdw = AHA1740CMD_INIT; /* SCSI Initiator Command ecb 373 drivers/scsi/aha1740.c host->ecb[ecbno].cdblen = SCpnt->cmd_len; /* SCSI Command ecb 383 drivers/scsi/aha1740.c memcpy(host->ecb[ecbno].cdb, cmd, SCpnt->cmd_len); ecb 403 drivers/scsi/aha1740.c host->ecb[ecbno].sg = 1; /* SCSI Initiator Command ecb 410 drivers/scsi/aha1740.c host->ecb[ecbno].datalen = nseg * sizeof(struct aha1740_chain); ecb 411 drivers/scsi/aha1740.c host->ecb[ecbno].dataptr = sg_dma; ecb 418 drivers/scsi/aha1740.c host->ecb[ecbno].datalen = 0; ecb 419 drivers/scsi/aha1740.c host->ecb[ecbno].dataptr = 0; ecb 421 drivers/scsi/aha1740.c host->ecb[ecbno].lun = SCpnt->device->lun; ecb 422 drivers/scsi/aha1740.c host->ecb[ecbno].ses = 1; /* Suppress underrun errors */ ecb 423 drivers/scsi/aha1740.c host->ecb[ecbno].dir = direction; ecb 424 drivers/scsi/aha1740.c host->ecb[ecbno].ars = 1; /* Yes, get the sense on an error */ ecb 425 drivers/scsi/aha1740.c host->ecb[ecbno].senselen = 12; ecb 426 drivers/scsi/aha1740.c host->ecb[ecbno].senseptr = ecb_cpu_to_dma (SCpnt->device->host, ecb 427 drivers/scsi/aha1740.c host->ecb[ecbno].sense); ecb 428 drivers/scsi/aha1740.c host->ecb[ecbno].statusptr = ecb_cpu_to_dma (SCpnt->device->host, ecb 429 drivers/scsi/aha1740.c host->ecb[ecbno].status); ecb 430 drivers/scsi/aha1740.c host->ecb[ecbno].done = done; ecb 431 drivers/scsi/aha1740.c host->ecb[ecbno].SCpnt = SCpnt; ecb 436 drivers/scsi/aha1740.c for (i = 0; i < sizeof(host->ecb[ecbno]) - 10; i++) ecb 437 drivers/scsi/aha1740.c printk("%02x ", ((unchar *)&host->ecb[ecbno])[i]); ecb 469 drivers/scsi/aha1740.c outl (ecb_cpu_to_dma (SCpnt->device->host, host->ecb + ecbno), ecb 590 drivers/scsi/aha1740.c host->ecb_dma_addr = dma_map_single (&edev->dev, host->ecb, ecb 591 drivers/scsi/aha1740.c sizeof (host->ecb), ecb 620 drivers/scsi/aha1740.c sizeof (host->ecb), DMA_BIDIRECTIONAL); ecb 638 drivers/scsi/aha1740.c sizeof (host->ecb), DMA_BIDIRECTIONAL);