stit 51 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit = container_of(head, stit 54 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_table_put(stit->tbl); stit 56 arch/powerpc/kvm/book3s_64_vio.c kfree(stit); stit 61 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit = container_of(kref, stit 64 arch/powerpc/kvm/book3s_64_vio.c list_del_rcu(&stit->next); stit 66 arch/powerpc/kvm/book3s_64_vio.c call_rcu(&stit->rcu, kvm_spapr_tce_iommu_table_free); stit 74 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit, *tmp; stit 83 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) { stit 85 arch/powerpc/kvm/book3s_64_vio.c if (table_group->tables[i] != stit->tbl) stit 88 arch/powerpc/kvm/book3s_64_vio.c kref_put(&stit->kref, kvm_spapr_tce_liobn_put); stit 102 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit; stit 147 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stit, &stt->iommu_tables, next) { stit 148 arch/powerpc/kvm/book3s_64_vio.c if (tbl != stit->tbl) stit 151 arch/powerpc/kvm/book3s_64_vio.c if (!kref_get_unless_zero(&stit->kref)) { stit 163 arch/powerpc/kvm/book3s_64_vio.c stit = kzalloc(sizeof(*stit), GFP_KERNEL); stit 164 arch/powerpc/kvm/book3s_64_vio.c if (!stit) { stit 169 arch/powerpc/kvm/book3s_64_vio.c stit->tbl = tbl; stit 170 arch/powerpc/kvm/book3s_64_vio.c kref_init(&stit->kref); stit 172 arch/powerpc/kvm/book3s_64_vio.c list_add_rcu(&stit->next, &stt->iommu_tables); stit 241 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit, *tmp; stit 248 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) { stit 249 arch/powerpc/kvm/book3s_64_vio.c WARN_ON(!kref_read(&stit->kref)); stit 251 arch/powerpc/kvm/book3s_64_vio.c if (kref_put(&stit->kref, kvm_spapr_tce_liobn_put)) stit 354 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit; stit 367 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stit, &stt->iommu_tables, next) { stit 370 arch/powerpc/kvm/book3s_64_vio.c long shift = stit->tbl->it_page_shift; stit 545 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit; stit 575 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stit 578 arch/powerpc/kvm/book3s_64_vio.c stit->tbl, entry); stit 580 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_iommu_map(vcpu->kvm, stt, stit->tbl, stit 583 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_kill(stit->tbl, entry, 1); stit 586 arch/powerpc/kvm/book3s_64_vio.c kvmppc_clear_tce(vcpu->kvm->mm, stit->tbl, entry); stit 609 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit; stit 671 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stit 673 arch/powerpc/kvm/book3s_64_vio.c stit->tbl, entry + i, ua, stit 677 arch/powerpc/kvm/book3s_64_vio.c kvmppc_clear_tce(vcpu->kvm->mm, stit->tbl, stit 687 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stit 688 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_kill(stit->tbl, entry, npages); stit 703 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_iommu_table *stit; stit 717 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stit 722 arch/powerpc/kvm/book3s_64_vio.c stit->tbl, entry + i); stit 731 arch/powerpc/kvm/book3s_64_vio.c kvmppc_clear_tce(vcpu->kvm->mm, stit->tbl, entry); stit 739 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stit 740 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_kill(stit->tbl, ioba >> stt->page_shift, npages); stit 110 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_iommu_table *stit; stit 123 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stit 126 arch/powerpc/kvm/book3s_64_vio_hv.c long shift = stit->tbl->it_page_shift; stit 391 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_iommu_table *stit; stit 420 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stit 423 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry); stit 426 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry, ua, dir); stit 428 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_kill_rm(stit->tbl, entry, 1); stit 431 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_clear_tce(vcpu->kvm, stit->tbl, entry); stit 486 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_iommu_table *stit; stit 574 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stit 576 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry + i, ua, stit 580 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_clear_tce(vcpu->kvm, stit->tbl, stit 590 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stit 591 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_kill_rm(stit->tbl, entry, npages); stit 606 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_iommu_table *stit; stit 624 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stit 629 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry + i); stit 638 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_clear_tce(vcpu->kvm, stit->tbl, entry); stit 646 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stit 647 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_kill_rm(stit->tbl, ioba >> stt->page_shift, npages);