s390_domain        96 arch/s390/include/asm/pci.h struct s390_domain;
s390_domain       160 arch/s390/include/asm/pci.h 	struct s390_domain *s390_domain; /* s390 IOMMU domain data */
s390_domain       556 arch/s390/pci/pci_dma.c 	WARN_ON(zdev->s390_domain);
s390_domain       620 arch/s390/pci/pci_dma.c 	WARN_ON(zdev->s390_domain);
s390_domain        37 drivers/iommu/s390-iommu.c static struct s390_domain *to_s390_domain(struct iommu_domain *dom)
s390_domain        39 drivers/iommu/s390-iommu.c 	return container_of(dom, struct s390_domain, domain);
s390_domain        56 drivers/iommu/s390-iommu.c 	struct s390_domain *s390_domain;
s390_domain        61 drivers/iommu/s390-iommu.c 	s390_domain = kzalloc(sizeof(*s390_domain), GFP_KERNEL);
s390_domain        62 drivers/iommu/s390-iommu.c 	if (!s390_domain)
s390_domain        65 drivers/iommu/s390-iommu.c 	s390_domain->dma_table = dma_alloc_cpu_table();
s390_domain        66 drivers/iommu/s390-iommu.c 	if (!s390_domain->dma_table) {
s390_domain        67 drivers/iommu/s390-iommu.c 		kfree(s390_domain);
s390_domain        71 drivers/iommu/s390-iommu.c 	spin_lock_init(&s390_domain->dma_table_lock);
s390_domain        72 drivers/iommu/s390-iommu.c 	spin_lock_init(&s390_domain->list_lock);
s390_domain        73 drivers/iommu/s390-iommu.c 	INIT_LIST_HEAD(&s390_domain->devices);
s390_domain        75 drivers/iommu/s390-iommu.c 	return &s390_domain->domain;
s390_domain        80 drivers/iommu/s390-iommu.c 	struct s390_domain *s390_domain = to_s390_domain(domain);
s390_domain        82 drivers/iommu/s390-iommu.c 	dma_cleanup_tables(s390_domain->dma_table);
s390_domain        83 drivers/iommu/s390-iommu.c 	kfree(s390_domain);
s390_domain        89 drivers/iommu/s390-iommu.c 	struct s390_domain *s390_domain = to_s390_domain(domain);
s390_domain       105 drivers/iommu/s390-iommu.c 	zdev->dma_table = s390_domain->dma_table;
s390_domain       111 drivers/iommu/s390-iommu.c 	spin_lock_irqsave(&s390_domain->list_lock, flags);
s390_domain       113 drivers/iommu/s390-iommu.c 	if (list_empty(&s390_domain->devices)) {
s390_domain       121 drivers/iommu/s390-iommu.c 		spin_unlock_irqrestore(&s390_domain->list_lock, flags);
s390_domain       125 drivers/iommu/s390-iommu.c 	zdev->s390_domain = s390_domain;
s390_domain       126 drivers/iommu/s390-iommu.c 	list_add(&domain_device->list, &s390_domain->devices);
s390_domain       127 drivers/iommu/s390-iommu.c 	spin_unlock_irqrestore(&s390_domain->list_lock, flags);
s390_domain       141 drivers/iommu/s390-iommu.c 	struct s390_domain *s390_domain = to_s390_domain(domain);
s390_domain       150 drivers/iommu/s390-iommu.c 	spin_lock_irqsave(&s390_domain->list_lock, flags);
s390_domain       151 drivers/iommu/s390-iommu.c 	list_for_each_entry_safe(domain_device, tmp, &s390_domain->devices,
s390_domain       160 drivers/iommu/s390-iommu.c 	spin_unlock_irqrestore(&s390_domain->list_lock, flags);
s390_domain       163 drivers/iommu/s390-iommu.c 		zdev->s390_domain = NULL;
s390_domain       199 drivers/iommu/s390-iommu.c 	if (zdev && zdev->s390_domain) {
s390_domain       209 drivers/iommu/s390-iommu.c static int s390_iommu_update_trans(struct s390_domain *s390_domain,
s390_domain       220 drivers/iommu/s390-iommu.c 	if (dma_addr < s390_domain->domain.geometry.aperture_start ||
s390_domain       221 drivers/iommu/s390-iommu.c 	    dma_addr + size > s390_domain->domain.geometry.aperture_end)
s390_domain       228 drivers/iommu/s390-iommu.c 	spin_lock_irqsave(&s390_domain->dma_table_lock, irq_flags);
s390_domain       230 drivers/iommu/s390-iommu.c 		entry = dma_walk_cpu_trans(s390_domain->dma_table, dma_addr);
s390_domain       240 drivers/iommu/s390-iommu.c 	spin_lock(&s390_domain->list_lock);
s390_domain       241 drivers/iommu/s390-iommu.c 	list_for_each_entry(domain_device, &s390_domain->devices, list) {
s390_domain       247 drivers/iommu/s390-iommu.c 	spin_unlock(&s390_domain->list_lock);
s390_domain       255 drivers/iommu/s390-iommu.c 			entry = dma_walk_cpu_trans(s390_domain->dma_table,
s390_domain       262 drivers/iommu/s390-iommu.c 	spin_unlock_irqrestore(&s390_domain->dma_table_lock, irq_flags);
s390_domain       270 drivers/iommu/s390-iommu.c 	struct s390_domain *s390_domain = to_s390_domain(domain);
s390_domain       279 drivers/iommu/s390-iommu.c 	rc = s390_iommu_update_trans(s390_domain, (unsigned long) paddr, iova,
s390_domain       288 drivers/iommu/s390-iommu.c 	struct s390_domain *s390_domain = to_s390_domain(domain);
s390_domain       300 drivers/iommu/s390-iommu.c 	rto = s390_domain->dma_table;
s390_domain       302 drivers/iommu/s390-iommu.c 	spin_lock_irqsave(&s390_domain->dma_table_lock, flags);
s390_domain       311 drivers/iommu/s390-iommu.c 	spin_unlock_irqrestore(&s390_domain->dma_table_lock, flags);
s390_domain       320 drivers/iommu/s390-iommu.c 	struct s390_domain *s390_domain = to_s390_domain(domain);
s390_domain       329 drivers/iommu/s390-iommu.c 	rc = s390_iommu_update_trans(s390_domain, (unsigned long) paddr, iova,