stt 3913 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t stt:2; stt 3915 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t stt:2; stt 4328 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t stt:2; stt 4330 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t stt:2; stt 185 arch/powerpc/include/asm/kvm_ppc.h #define kvmppc_ioba_validate(stt, ioba, npages) \ stt 186 arch/powerpc/include/asm/kvm_ppc.h (iommu_tce_check_ioba((stt)->page_shift, (stt)->offset, \ stt 187 arch/powerpc/include/asm/kvm_ppc.h (stt)->size, (ioba), (npages)) ? \ stt 73 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt; stt 77 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stt, &kvm->arch.spapr_tce_tables, list) { stt 83 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) { stt 97 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt = NULL; stt 109 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stt, &kvm->arch.spapr_tce_tables, list) { stt 110 arch/powerpc/kvm/book3s_64_vio.c if (stt == f.file->private_data) { stt 131 arch/powerpc/kvm/book3s_64_vio.c if ((tbltmp->it_page_shift <= stt->page_shift) && stt 133 arch/powerpc/kvm/book3s_64_vio.c stt->offset << stt->page_shift) && stt 135 arch/powerpc/kvm/book3s_64_vio.c stt->size << stt->page_shift)) { stt 147 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stit, &stt->iommu_tables, next) { stt 172 arch/powerpc/kvm/book3s_64_vio.c list_add_rcu(&stit->next, &stt->iommu_tables); stt 179 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt = container_of(head, stt 181 arch/powerpc/kvm/book3s_64_vio.c unsigned long i, npages = kvmppc_tce_pages(stt->size); stt 184 arch/powerpc/kvm/book3s_64_vio.c if (stt->pages[i]) stt 185 arch/powerpc/kvm/book3s_64_vio.c __free_page(stt->pages[i]); stt 187 arch/powerpc/kvm/book3s_64_vio.c kfree(stt); stt 190 arch/powerpc/kvm/book3s_64_vio.c static struct page *kvm_spapr_get_tce_page(struct kvmppc_spapr_tce_table *stt, stt 193 arch/powerpc/kvm/book3s_64_vio.c struct page *page = stt->pages[sttpage]; stt 198 arch/powerpc/kvm/book3s_64_vio.c mutex_lock(&stt->alloc_lock); stt 199 arch/powerpc/kvm/book3s_64_vio.c page = stt->pages[sttpage]; stt 204 arch/powerpc/kvm/book3s_64_vio.c stt->pages[sttpage] = page; stt 206 arch/powerpc/kvm/book3s_64_vio.c mutex_unlock(&stt->alloc_lock); stt 213 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt = vmf->vma->vm_file->private_data; stt 216 arch/powerpc/kvm/book3s_64_vio.c if (vmf->pgoff >= kvmppc_tce_pages(stt->size)) stt 219 arch/powerpc/kvm/book3s_64_vio.c page = kvm_spapr_get_tce_page(stt, vmf->pgoff); stt 240 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt = filp->private_data; stt 242 arch/powerpc/kvm/book3s_64_vio.c struct kvm *kvm = stt->kvm; stt 245 arch/powerpc/kvm/book3s_64_vio.c list_del_rcu(&stt->list); stt 248 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) { stt 256 arch/powerpc/kvm/book3s_64_vio.c kvm_put_kvm(stt->kvm); stt 259 arch/powerpc/kvm/book3s_64_vio.c kvmppc_stt_pages(kvmppc_tce_pages(stt->size)), false); stt 260 arch/powerpc/kvm/book3s_64_vio.c call_rcu(&stt->rcu, release_spapr_tce_table); stt 273 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt = NULL; stt 288 arch/powerpc/kvm/book3s_64_vio.c stt = kzalloc(sizeof(*stt) + npages * sizeof(struct page *), stt 290 arch/powerpc/kvm/book3s_64_vio.c if (!stt) stt 293 arch/powerpc/kvm/book3s_64_vio.c stt->liobn = args->liobn; stt 294 arch/powerpc/kvm/book3s_64_vio.c stt->page_shift = args->page_shift; stt 295 arch/powerpc/kvm/book3s_64_vio.c stt->offset = args->offset; stt 296 arch/powerpc/kvm/book3s_64_vio.c stt->size = size; stt 297 arch/powerpc/kvm/book3s_64_vio.c stt->kvm = kvm; stt 298 arch/powerpc/kvm/book3s_64_vio.c mutex_init(&stt->alloc_lock); stt 299 arch/powerpc/kvm/book3s_64_vio.c INIT_LIST_HEAD_RCU(&stt->iommu_tables); stt 315 arch/powerpc/kvm/book3s_64_vio.c stt, O_RDWR | O_CLOEXEC); stt 318 arch/powerpc/kvm/book3s_64_vio.c list_add_rcu(&stt->list, &kvm->arch.spapr_tce_tables); stt 327 arch/powerpc/kvm/book3s_64_vio.c kfree(stt); stt 349 arch/powerpc/kvm/book3s_64_vio.c static long kvmppc_tce_validate(struct kvmppc_spapr_tce_table *stt, stt 361 arch/powerpc/kvm/book3s_64_vio.c if (iommu_tce_check_gpa(stt->page_shift, gpa)) stt 364 arch/powerpc/kvm/book3s_64_vio.c if (kvmppc_tce_to_ua(stt->kvm, tce, &ua)) stt 367 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_rcu(stit, &stt->iommu_tables, next) { stt 372 arch/powerpc/kvm/book3s_64_vio.c mem = mm_iommu_lookup(stt->kvm->mm, ua, 1ULL << shift); stt 388 arch/powerpc/kvm/book3s_64_vio.c static void kvmppc_tce_put(struct kvmppc_spapr_tce_table *stt, stt 395 arch/powerpc/kvm/book3s_64_vio.c idx -= stt->offset; stt 397 arch/powerpc/kvm/book3s_64_vio.c page = stt->pages[sttpage]; stt 404 arch/powerpc/kvm/book3s_64_vio.c page = kvm_spapr_get_tce_page(stt, sttpage); stt 465 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt, struct iommu_table *tbl, stt 469 arch/powerpc/kvm/book3s_64_vio.c unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); stt 520 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt, struct iommu_table *tbl, stt 525 arch/powerpc/kvm/book3s_64_vio.c unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); stt 543 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt; stt 552 arch/powerpc/kvm/book3s_64_vio.c stt = kvmppc_find_table(vcpu->kvm, liobn); stt 553 arch/powerpc/kvm/book3s_64_vio.c if (!stt) stt 556 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_ioba_validate(stt, ioba, 1); stt 562 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_validate(stt, tce); stt 573 arch/powerpc/kvm/book3s_64_vio.c entry = ioba >> stt->page_shift; stt 575 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stt 577 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_iommu_unmap(vcpu->kvm, stt, stt 580 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_iommu_map(vcpu->kvm, stt, stit->tbl, stt 591 arch/powerpc/kvm/book3s_64_vio.c kvmppc_tce_put(stt, entry, tce); stt 604 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt; stt 611 arch/powerpc/kvm/book3s_64_vio.c stt = kvmppc_find_table(vcpu->kvm, liobn); stt 612 arch/powerpc/kvm/book3s_64_vio.c if (!stt) stt 615 arch/powerpc/kvm/book3s_64_vio.c entry = ioba >> stt->page_shift; stt 626 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_ioba_validate(stt, ioba, npages); stt 644 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_validate(stt, tce); stt 671 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stt 672 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_iommu_map(vcpu->kvm, stt, stt 683 arch/powerpc/kvm/book3s_64_vio.c kvmppc_tce_put(stt, entry + i, tce); stt 687 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stt 701 arch/powerpc/kvm/book3s_64_vio.c struct kvmppc_spapr_tce_table *stt; stt 705 arch/powerpc/kvm/book3s_64_vio.c stt = kvmppc_find_table(vcpu->kvm, liobn); stt 706 arch/powerpc/kvm/book3s_64_vio.c if (!stt) stt 709 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_ioba_validate(stt, ioba, npages); stt 717 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stt 718 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry = ioba >> stt->page_shift; stt 721 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_iommu_unmap(vcpu->kvm, stt, stt 735 arch/powerpc/kvm/book3s_64_vio.c for (i = 0; i < npages; ++i, ioba += (1ULL << stt->page_shift)) stt 736 arch/powerpc/kvm/book3s_64_vio.c kvmppc_tce_put(stt, ioba >> stt->page_shift, tce_value); stt 739 arch/powerpc/kvm/book3s_64_vio.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stt 740 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_kill(stit->tbl, ioba >> stt->page_shift, npages); stt 67 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_table *stt; stt 69 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stt, &kvm->arch.spapr_tce_tables, list) stt 70 arch/powerpc/kvm/book3s_64_vio_hv.c if (stt->liobn == liobn) stt 71 arch/powerpc/kvm/book3s_64_vio_hv.c return stt; stt 105 arch/powerpc/kvm/book3s_64_vio_hv.c static long kvmppc_rm_tce_validate(struct kvmppc_spapr_tce_table *stt, stt 117 arch/powerpc/kvm/book3s_64_vio_hv.c if (iommu_tce_check_gpa(stt->page_shift, gpa)) stt 120 arch/powerpc/kvm/book3s_64_vio_hv.c if (kvmppc_rm_tce_to_ua(stt->kvm, tce, &ua, NULL)) stt 123 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stt 128 arch/powerpc/kvm/book3s_64_vio_hv.c mem = mm_iommu_lookup_rm(stt->kvm->mm, ua, 1ULL << shift); stt 171 arch/powerpc/kvm/book3s_64_vio_hv.c static void kvmppc_rm_tce_put(struct kvmppc_spapr_tce_table *stt, stt 177 arch/powerpc/kvm/book3s_64_vio_hv.c idx -= stt->offset; stt 178 arch/powerpc/kvm/book3s_64_vio_hv.c page = stt->pages[idx / TCES_PER_PAGE]; stt 195 arch/powerpc/kvm/book3s_64_vio_hv.c static long kvmppc_rm_ioba_validate(struct kvmppc_spapr_tce_table *stt, stt 199 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long ret = kvmppc_ioba_validate(stt, ioba, npages); stt 210 arch/powerpc/kvm/book3s_64_vio_hv.c idx = (ioba >> stt->page_shift) - stt->offset; stt 215 arch/powerpc/kvm/book3s_64_vio_hv.c if (!stt->pages[i]) stt 307 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_table *stt, struct iommu_table *tbl, stt 311 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); stt 366 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_table *stt, struct iommu_table *tbl, stt 371 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); stt 389 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_table *stt; stt 402 arch/powerpc/kvm/book3s_64_vio_hv.c stt = kvmppc_find_table(vcpu->kvm, liobn); stt 403 arch/powerpc/kvm/book3s_64_vio_hv.c if (!stt) stt 406 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_ioba_validate(stt, ioba, 1, tce == 0); stt 410 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_tce_validate(stt, tce); stt 418 arch/powerpc/kvm/book3s_64_vio_hv.c entry = ioba >> stt->page_shift; stt 420 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stt 422 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_tce_iommu_unmap(vcpu->kvm, stt, stt 425 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_tce_iommu_map(vcpu->kvm, stt, stt 436 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_tce_put(stt, entry, tce); stt 481 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_table *stt; stt 492 arch/powerpc/kvm/book3s_64_vio_hv.c stt = kvmppc_find_table(vcpu->kvm, liobn); stt 493 arch/powerpc/kvm/book3s_64_vio_hv.c if (!stt) stt 496 arch/powerpc/kvm/book3s_64_vio_hv.c entry = ioba >> stt->page_shift; stt 507 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_ioba_validate(stt, ioba, npages, false); stt 560 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_tce_validate(stt, tce); stt 574 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stt 575 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_tce_iommu_map(vcpu->kvm, stt, stt 586 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_tce_put(stt, entry + i, tce); stt 590 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stt 604 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_table *stt; stt 612 arch/powerpc/kvm/book3s_64_vio_hv.c stt = kvmppc_find_table(vcpu->kvm, liobn); stt 613 arch/powerpc/kvm/book3s_64_vio_hv.c if (!stt) stt 616 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_ioba_validate(stt, ioba, npages, tce_value == 0); stt 624 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) { stt 625 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry = ioba >> stt->page_shift; stt 628 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_tce_iommu_unmap(vcpu->kvm, stt, stt 642 arch/powerpc/kvm/book3s_64_vio_hv.c for (i = 0; i < npages; ++i, ioba += (1ULL << stt->page_shift)) stt 643 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_tce_put(stt, ioba >> stt->page_shift, tce_value); stt 646 arch/powerpc/kvm/book3s_64_vio_hv.c list_for_each_entry_lockless(stit, &stt->iommu_tables, next) stt 647 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_kill_rm(stit->tbl, ioba >> stt->page_shift, npages); stt 656 arch/powerpc/kvm/book3s_64_vio_hv.c struct kvmppc_spapr_tce_table *stt; stt 662 arch/powerpc/kvm/book3s_64_vio_hv.c stt = kvmppc_find_table(vcpu->kvm, liobn); stt 663 arch/powerpc/kvm/book3s_64_vio_hv.c if (!stt) stt 666 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_ioba_validate(stt, ioba, 1); stt 670 arch/powerpc/kvm/book3s_64_vio_hv.c idx = (ioba >> stt->page_shift) - stt->offset; stt 671 arch/powerpc/kvm/book3s_64_vio_hv.c page = stt->pages[idx / TCES_PER_PAGE]; stt 341 drivers/scsi/cxgbi/libcxgbi.c struct scsi_transport_template *stt) stt 358 drivers/scsi/cxgbi/libcxgbi.c shost->transportt = stt; stt 2690 drivers/scsi/cxgbi/libcxgbi.c struct scsi_transport_template **stt) stt 2692 drivers/scsi/cxgbi/libcxgbi.c *stt = iscsi_register_transport(itp); stt 2693 drivers/scsi/cxgbi/libcxgbi.c if (*stt == NULL) { stt 2700 drivers/scsi/cxgbi/libcxgbi.c itp->name, stt); stt 2706 drivers/scsi/cxgbi/libcxgbi.c struct scsi_transport_template **stt) stt 2708 drivers/scsi/cxgbi/libcxgbi.c if (*stt) { stt 2711 drivers/scsi/cxgbi/libcxgbi.c itp, itp->name, *stt); stt 2712 drivers/scsi/cxgbi/libcxgbi.c *stt = NULL; stt 578 drivers/scsi/libsas/sas_init.c struct scsi_transport_template *stt = sas_attach_transport(&sft); stt 581 drivers/scsi/libsas/sas_init.c if (!stt) stt 582 drivers/scsi/libsas/sas_init.c return stt; stt 584 drivers/scsi/libsas/sas_init.c i = to_sas_internal(stt); stt 586 drivers/scsi/libsas/sas_init.c stt->create_work_queue = 1; stt 587 drivers/scsi/libsas/sas_init.c stt->eh_strategy_handler = sas_scsi_recover_host; stt 589 drivers/scsi/libsas/sas_init.c return stt;