Lines Matching refs:s390_domain

22 struct s390_domain {  struct
35 static struct s390_domain *to_s390_domain(struct iommu_domain *dom) in to_s390_domain() argument
37 return container_of(dom, struct s390_domain, domain); in to_s390_domain()
54 struct s390_domain *s390_domain; in s390_domain_alloc() local
59 s390_domain = kzalloc(sizeof(*s390_domain), GFP_KERNEL); in s390_domain_alloc()
60 if (!s390_domain) in s390_domain_alloc()
63 s390_domain->dma_table = dma_alloc_cpu_table(); in s390_domain_alloc()
64 if (!s390_domain->dma_table) { in s390_domain_alloc()
65 kfree(s390_domain); in s390_domain_alloc()
69 spin_lock_init(&s390_domain->dma_table_lock); in s390_domain_alloc()
70 spin_lock_init(&s390_domain->list_lock); in s390_domain_alloc()
71 INIT_LIST_HEAD(&s390_domain->devices); in s390_domain_alloc()
73 return &s390_domain->domain; in s390_domain_alloc()
78 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_domain_free() local
80 dma_cleanup_tables(s390_domain->dma_table); in s390_domain_free()
81 kfree(s390_domain); in s390_domain_free()
87 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_iommu_attach_device() local
103 zdev->dma_table = s390_domain->dma_table; in s390_iommu_attach_device()
110 spin_lock_irqsave(&s390_domain->list_lock, flags); in s390_iommu_attach_device()
112 if (list_empty(&s390_domain->devices)) { in s390_iommu_attach_device()
120 spin_unlock_irqrestore(&s390_domain->list_lock, flags); in s390_iommu_attach_device()
124 zdev->s390_domain = s390_domain; in s390_iommu_attach_device()
125 list_add(&domain_device->list, &s390_domain->devices); in s390_iommu_attach_device()
126 spin_unlock_irqrestore(&s390_domain->list_lock, flags); in s390_iommu_attach_device()
140 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_iommu_detach_device() local
149 spin_lock_irqsave(&s390_domain->list_lock, flags); in s390_iommu_detach_device()
150 list_for_each_entry_safe(domain_device, tmp, &s390_domain->devices, in s390_iommu_detach_device()
159 spin_unlock_irqrestore(&s390_domain->list_lock, flags); in s390_iommu_detach_device()
162 zdev->s390_domain = NULL; in s390_iommu_detach_device()
202 if (zdev && zdev->s390_domain) { in s390_iommu_remove_device()
211 static int s390_iommu_update_trans(struct s390_domain *s390_domain, in s390_iommu_update_trans() argument
222 if (dma_addr < s390_domain->domain.geometry.aperture_start || in s390_iommu_update_trans()
223 dma_addr + size > s390_domain->domain.geometry.aperture_end) in s390_iommu_update_trans()
230 spin_lock_irqsave(&s390_domain->dma_table_lock, irq_flags); in s390_iommu_update_trans()
232 entry = dma_walk_cpu_trans(s390_domain->dma_table, dma_addr); in s390_iommu_update_trans()
242 spin_lock(&s390_domain->list_lock); in s390_iommu_update_trans()
243 list_for_each_entry(domain_device, &s390_domain->devices, list) { in s390_iommu_update_trans()
249 spin_unlock(&s390_domain->list_lock); in s390_iommu_update_trans()
257 entry = dma_walk_cpu_trans(s390_domain->dma_table, in s390_iommu_update_trans()
264 spin_unlock_irqrestore(&s390_domain->dma_table_lock, irq_flags); in s390_iommu_update_trans()
272 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_iommu_map() local
281 rc = s390_iommu_update_trans(s390_domain, (unsigned long) paddr, iova, in s390_iommu_map()
290 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_iommu_iova_to_phys() local
302 rto = s390_domain->dma_table; in s390_iommu_iova_to_phys()
304 spin_lock_irqsave(&s390_domain->dma_table_lock, flags); in s390_iommu_iova_to_phys()
313 spin_unlock_irqrestore(&s390_domain->dma_table_lock, flags); in s390_iommu_iova_to_phys()
321 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_iommu_unmap() local
330 rc = s390_iommu_update_trans(s390_domain, (unsigned long) paddr, iova, in s390_iommu_unmap()