icreq 1153 drivers/nvme/host/tcp.c struct nvme_tcp_icreq_pdu *icreq; icreq 1160 drivers/nvme/host/tcp.c icreq = kzalloc(sizeof(*icreq), GFP_KERNEL); icreq 1161 drivers/nvme/host/tcp.c if (!icreq) icreq 1170 drivers/nvme/host/tcp.c icreq->hdr.type = nvme_tcp_icreq; icreq 1171 drivers/nvme/host/tcp.c icreq->hdr.hlen = sizeof(*icreq); icreq 1172 drivers/nvme/host/tcp.c icreq->hdr.pdo = 0; icreq 1173 drivers/nvme/host/tcp.c icreq->hdr.plen = cpu_to_le32(icreq->hdr.hlen); icreq 1174 drivers/nvme/host/tcp.c icreq->pfv = cpu_to_le16(NVME_TCP_PFV_1_0); icreq 1175 drivers/nvme/host/tcp.c icreq->maxr2t = 0; /* single inflight r2t supported */ icreq 1176 drivers/nvme/host/tcp.c icreq->hpda = 0; /* no alignment constraint */ icreq 1178 drivers/nvme/host/tcp.c icreq->digest |= NVME_TCP_HDR_DIGEST_ENABLE; icreq 1180 drivers/nvme/host/tcp.c icreq->digest |= NVME_TCP_DATA_DIGEST_ENABLE; icreq 1182 drivers/nvme/host/tcp.c iov.iov_base = icreq; icreq 1183 drivers/nvme/host/tcp.c iov.iov_len = sizeof(*icreq); icreq 1245 drivers/nvme/host/tcp.c kfree(icreq); icreq 760 drivers/nvme/target/tcp.c struct nvme_tcp_icreq_pdu *icreq = &queue->pdu.icreq; icreq 766 drivers/nvme/target/tcp.c if (le32_to_cpu(icreq->hdr.plen) != sizeof(struct nvme_tcp_icreq_pdu)) { icreq 768 drivers/nvme/target/tcp.c le32_to_cpu(icreq->hdr.plen)); icreq 772 drivers/nvme/target/tcp.c if (icreq->pfv != NVME_TCP_PFV_1_0) { icreq 773 drivers/nvme/target/tcp.c pr_err("queue %d: bad pfv %d\n", queue->idx, icreq->pfv); icreq 777 drivers/nvme/target/tcp.c if (icreq->hpda != 0) { icreq 779 drivers/nvme/target/tcp.c icreq->hpda); icreq 783 drivers/nvme/target/tcp.c queue->hdr_digest = !!(icreq->digest & NVME_TCP_HDR_DIGEST_ENABLE); icreq 784 drivers/nvme/target/tcp.c queue->data_digest = !!(icreq->digest & NVME_TCP_DATA_DIGEST_ENABLE); icreq 294 include/linux/netfilter/nf_conntrack_pptp.h struct PptpInCallRequest icreq; icreq 181 include/linux/nvme-tcp.h struct nvme_tcp_icreq_pdu icreq; icreq 338 net/netfilter/nf_conntrack_pptp.c cid = pptpReq->icreq.callID;