iov_size 236 arch/ia64/hp/common/sba_iommu.c u32 iov_size; iov_size 1592 arch/ia64/hp/common/sba_iommu.c ioc->iov_size = ~ioc->imask + 1; iov_size 1596 arch/ia64/hp/common/sba_iommu.c ioc->iov_size >> 20); iov_size 1610 arch/ia64/hp/common/sba_iommu.c ioc->pdir_size = (ioc->iov_size / iovp_size) * PDIR_ENTRY_SIZE; iov_size 1637 arch/ia64/hp/common/sba_iommu.c ioc->iov_size/2 >> 20, ioc->ibase + ioc->iov_size/2); iov_size 1639 arch/ia64/hp/common/sba_iommu.c ((u64 *)ioc->pdir_base)[PDIR_INDEX(ioc->iov_size/2)] = ZX1_SBA_IOMMU_COOKIE; iov_size 1671 arch/ia64/hp/common/sba_iommu.c WRITE_REG(ioc->ibase | (get_iovp_order(ioc->iov_size) + iovp_shift), ioc->ioc_hpa + IOC_PCOM); iov_size 1826 arch/ia64/hp/common/sba_iommu.c hpa, ioc->iov_size >> 20, ioc->ibase); iov_size 3146 drivers/misc/vmw_vmci/vmci_queue_pair.c size_t iov_size, iov_size 3187 drivers/misc/vmw_vmci/vmci_queue_pair.c size_t iov_size, iov_size 3229 drivers/misc/vmw_vmci/vmci_queue_pair.c size_t iov_size, iov_size 777 drivers/vhost/vhost.c struct iovec iov[], int iov_size, int access); iov_size 2041 drivers/vhost/vhost.c struct iovec iov[], int iov_size, int access) iov_size 2052 drivers/vhost/vhost.c if (unlikely(ret >= iov_size)) { iov_size 2103 drivers/vhost/vhost.c struct iovec iov[], unsigned int iov_size, iov_size 2171 drivers/vhost/vhost.c iov_size - iov_count, access); iov_size 2209 drivers/vhost/vhost.c struct iovec iov[], unsigned int iov_size, iov_size 2293 drivers/vhost/vhost.c ret = get_indirect(vq, iov, iov_size, iov_size 2311 drivers/vhost/vhost.c iov_size - iov_count, access); iov_size 69 include/linux/vmw_vmci_api.h struct msghdr *msg, size_t iov_size, int mode); iov_size 71 include/linux/vmw_vmci_api.h struct msghdr *msg, size_t iov_size, int mode); iov_size 72 include/linux/vmw_vmci_api.h ssize_t vmci_qpair_peekv(struct vmci_qp *qpair, struct msghdr *msg, size_t iov_size,