zdev 163 arch/s390/include/asm/pci.h static inline bool zdev_enabled(struct zpci_dev *zdev) zdev 165 arch/s390/include/asm/pci.h return (zdev->fh & (1UL << 31)) ? true : false; zdev 176 arch/s390/include/asm/pci.h void zpci_remove_device(struct zpci_dev *zdev); zdev 193 arch/s390/include/asm/pci.h int zpci_init_iommu(struct zpci_dev *zdev); zdev 194 arch/s390/include/asm/pci.h void zpci_destroy_iommu(struct zpci_dev *zdev); zdev 197 arch/s390/include/asm/pci.h static inline bool zpci_use_mio(struct zpci_dev *zdev) zdev 199 arch/s390/include/asm/pci.h return static_branch_likely(&have_mio) && zdev->mio_capable; zdev 217 arch/s390/include/asm/pci.h static inline int zpci_init_slot(struct zpci_dev *zdev) zdev 221 arch/s390/include/asm/pci.h static inline void zpci_exit_slot(struct zpci_dev *zdev) {} zdev 57 arch/s390/net/pnet.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 59 arch/s390/net/pnet.c memcpy(pnetids, zdev->util_str, sizeof(zdev->util_str)); zdev 60 arch/s390/net/pnet.c EBCASC(pnetids, sizeof(zdev->util_str)); zdev 60 arch/s390/pci/pci.c struct zpci_dev *tmp, *zdev = NULL; zdev 65 arch/s390/pci/pci.c zdev = tmp; zdev 70 arch/s390/pci/pci.c return zdev; zdev 75 arch/s390/pci/pci.c struct zpci_dev *tmp, *zdev; zdev 80 arch/s390/pci/pci.c list_for_each_entry_safe(zdev, tmp, &zpci_list, entry) { zdev 81 arch/s390/pci/pci.c if (zdev->state == ZPCI_FN_STATE_STANDBY && zdev 82 arch/s390/pci/pci.c !clp_get_state(zdev->fid, &state) && zdev 84 arch/s390/pci/pci.c list_move_tail(&zdev->entry, &remove); zdev 88 arch/s390/pci/pci.c list_for_each_entry_safe(zdev, tmp, &remove, entry) zdev 89 arch/s390/pci/pci.c zpci_remove_device(zdev); zdev 110 arch/s390/pci/pci.c int zpci_register_ioat(struct zpci_dev *zdev, u8 dmaas, zdev 113 arch/s390/pci/pci.c u64 req = ZPCI_CREATE_REQ(zdev->fh, dmaas, ZPCI_MOD_FC_REG_IOAT); zdev 125 arch/s390/pci/pci.c int zpci_unregister_ioat(struct zpci_dev *zdev, u8 dmaas) zdev 127 arch/s390/pci/pci.c u64 req = ZPCI_CREATE_REQ(zdev->fh, dmaas, ZPCI_MOD_FC_DEREG_IOAT); zdev 138 arch/s390/pci/pci.c int zpci_fmb_enable_device(struct zpci_dev *zdev) zdev 140 arch/s390/pci/pci.c u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_SET_MEASURE); zdev 144 arch/s390/pci/pci.c if (zdev->fmb || sizeof(*zdev->fmb) < zdev->fmb_length) zdev 147 arch/s390/pci/pci.c zdev->fmb = kmem_cache_zalloc(zdev_fmb_cache, GFP_KERNEL); zdev 148 arch/s390/pci/pci.c if (!zdev->fmb) zdev 150 arch/s390/pci/pci.c WARN_ON((u64) zdev->fmb & 0xf); zdev 153 arch/s390/pci/pci.c atomic64_set(&zdev->allocated_pages, 0); zdev 154 arch/s390/pci/pci.c atomic64_set(&zdev->mapped_pages, 0); zdev 155 arch/s390/pci/pci.c atomic64_set(&zdev->unmapped_pages, 0); zdev 157 arch/s390/pci/pci.c fib.fmb_addr = virt_to_phys(zdev->fmb); zdev 160 arch/s390/pci/pci.c kmem_cache_free(zdev_fmb_cache, zdev->fmb); zdev 161 arch/s390/pci/pci.c zdev->fmb = NULL; zdev 167 arch/s390/pci/pci.c int zpci_fmb_disable_device(struct zpci_dev *zdev) zdev 169 arch/s390/pci/pci.c u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_SET_MEASURE); zdev 173 arch/s390/pci/pci.c if (!zdev->fmb) zdev 182 arch/s390/pci/pci.c kmem_cache_free(zdev_fmb_cache, zdev->fmb); zdev 183 arch/s390/pci/pci.c zdev->fmb = NULL; zdev 188 arch/s390/pci/pci.c static int zpci_cfg_load(struct zpci_dev *zdev, int offset, u32 *val, u8 len) zdev 190 arch/s390/pci/pci.c u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len); zdev 204 arch/s390/pci/pci.c static int zpci_cfg_store(struct zpci_dev *zdev, int offset, u32 val, u8 len) zdev 206 arch/s390/pci/pci.c u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len); zdev 268 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 271 arch/s390/pci/pci.c idx = zdev->bars[bar].map_idx; zdev 275 arch/s390/pci/pci.c zpci_iomap_start[idx].fh = zdev->fh; zdev 287 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 290 arch/s390/pci/pci.c iova = ioremap((unsigned long) zdev->bars[bar].mio_wt, barsize); zdev 317 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 320 arch/s390/pci/pci.c iova = ioremap((unsigned long) zdev->bars[bar].mio_wb, barsize); zdev 374 arch/s390/pci/pci.c struct zpci_dev *zdev = get_zdev_by_bus(bus); zdev 377 arch/s390/pci/pci.c if (!zdev || devfn != ZPCI_DEVFN) zdev 380 arch/s390/pci/pci.c ret = zpci_cfg_load(zdev, where, val, size); zdev 388 arch/s390/pci/pci.c struct zpci_dev *zdev = get_zdev_by_bus(bus); zdev 391 arch/s390/pci/pci.c if (!zdev || devfn != ZPCI_DEVFN) zdev 394 arch/s390/pci/pci.c ret = zpci_cfg_store(zdev, where, val, size); zdev 415 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 424 arch/s390/pci/pci.c if (zpci_use_mio(zdev)) zdev 426 arch/s390/pci/pci.c (resource_size_t __force) zdev->bars[i].mio_wt; zdev 447 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 451 arch/s390/pci/pci.c if (zpci_use_mio(zdev)) zdev 463 arch/s390/pci/pci.c static int zpci_alloc_iomap(struct zpci_dev *zdev) zdev 478 arch/s390/pci/pci.c static void zpci_free_iomap(struct zpci_dev *zdev, int entry) zdev 486 arch/s390/pci/pci.c static struct resource *__alloc_res(struct zpci_dev *zdev, unsigned long start, zdev 498 arch/s390/pci/pci.c r->name = zdev->res_name; zdev 507 arch/s390/pci/pci.c static int zpci_setup_bus_resources(struct zpci_dev *zdev, zdev 514 arch/s390/pci/pci.c snprintf(zdev->res_name, sizeof(zdev->res_name), zdev 515 arch/s390/pci/pci.c "PCI Bus %04x:%02x", zdev->domain, ZPCI_BUS_NR); zdev 518 arch/s390/pci/pci.c if (!zdev->bars[i].size) zdev 520 arch/s390/pci/pci.c entry = zpci_alloc_iomap(zdev); zdev 523 arch/s390/pci/pci.c zdev->bars[i].map_idx = entry; zdev 527 arch/s390/pci/pci.c if (zdev->bars[i].val & 8) zdev 529 arch/s390/pci/pci.c if (zdev->bars[i].val & 4) zdev 532 arch/s390/pci/pci.c if (zpci_use_mio(zdev)) zdev 533 arch/s390/pci/pci.c addr = (unsigned long) zdev->bars[i].mio_wt; zdev 536 arch/s390/pci/pci.c size = 1UL << zdev->bars[i].size; zdev 538 arch/s390/pci/pci.c res = __alloc_res(zdev, addr, size, flags); zdev 540 arch/s390/pci/pci.c zpci_free_iomap(zdev, entry); zdev 543 arch/s390/pci/pci.c zdev->bars[i].res = res; zdev 550 arch/s390/pci/pci.c static void zpci_cleanup_bus_resources(struct zpci_dev *zdev) zdev 555 arch/s390/pci/pci.c if (!zdev->bars[i].size || !zdev->bars[i].res) zdev 558 arch/s390/pci/pci.c zpci_free_iomap(zdev, zdev->bars[i].map_idx); zdev 559 arch/s390/pci/pci.c release_resource(zdev->bars[i].res); zdev 560 arch/s390/pci/pci.c kfree(zdev->bars[i].res); zdev 593 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 595 arch/s390/pci/pci.c zpci_debug_init_device(zdev, dev_name(&pdev->dev)); zdev 596 arch/s390/pci/pci.c zpci_fmb_enable_device(zdev); zdev 603 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 605 arch/s390/pci/pci.c zpci_fmb_disable_device(zdev); zdev 606 arch/s390/pci/pci.c zpci_debug_exit_device(zdev); zdev 613 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 616 arch/s390/pci/pci.c if (zdev->state != ZPCI_FN_STATE_ONLINE) zdev 619 arch/s390/pci/pci.c ret = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES); zdev 624 arch/s390/pci/pci.c zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma, zdev 625 arch/s390/pci/pci.c (u64) zdev->dma_table); zdev 634 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 636 arch/s390/pci/pci.c if (zdev->state != ZPCI_FN_STATE_ONLINE) zdev 639 arch/s390/pci/pci.c zpci_unregister_ioat(zdev, 0); zdev 641 arch/s390/pci/pci.c return clp_disable_fh(zdev); zdev 652 arch/s390/pci/pci.c static int zpci_alloc_domain(struct zpci_dev *zdev) zdev 655 arch/s390/pci/pci.c zdev->domain = (u16) zdev->uid; zdev 656 arch/s390/pci/pci.c if (zdev->domain >= ZPCI_NR_DEVICES) zdev 660 arch/s390/pci/pci.c if (test_bit(zdev->domain, zpci_domain)) { zdev 664 arch/s390/pci/pci.c set_bit(zdev->domain, zpci_domain); zdev 670 arch/s390/pci/pci.c zdev->domain = find_first_zero_bit(zpci_domain, ZPCI_NR_DEVICES); zdev 671 arch/s390/pci/pci.c if (zdev->domain == ZPCI_NR_DEVICES) { zdev 675 arch/s390/pci/pci.c set_bit(zdev->domain, zpci_domain); zdev 680 arch/s390/pci/pci.c static void zpci_free_domain(struct zpci_dev *zdev) zdev 682 arch/s390/pci/pci.c if (zdev->domain >= ZPCI_NR_DEVICES) zdev 686 arch/s390/pci/pci.c clear_bit(zdev->domain, zpci_domain); zdev 692 arch/s390/pci/pci.c struct zpci_dev *zdev = get_zdev_by_bus(bus); zdev 694 arch/s390/pci/pci.c zpci_exit_slot(zdev); zdev 695 arch/s390/pci/pci.c zpci_cleanup_bus_resources(zdev); zdev 696 arch/s390/pci/pci.c zpci_destroy_iommu(zdev); zdev 697 arch/s390/pci/pci.c zpci_free_domain(zdev); zdev 700 arch/s390/pci/pci.c list_del(&zdev->entry); zdev 703 arch/s390/pci/pci.c zpci_dbg(3, "rem fid:%x\n", zdev->fid); zdev 704 arch/s390/pci/pci.c kfree(zdev); zdev 707 arch/s390/pci/pci.c static int zpci_scan_bus(struct zpci_dev *zdev) zdev 712 arch/s390/pci/pci.c ret = zpci_setup_bus_resources(zdev, &resources); zdev 716 arch/s390/pci/pci.c zdev->bus = pci_scan_root_bus(NULL, ZPCI_BUS_NR, &pci_root_ops, zdev 717 arch/s390/pci/pci.c zdev, &resources); zdev 718 arch/s390/pci/pci.c if (!zdev->bus) { zdev 722 arch/s390/pci/pci.c zdev->bus->max_bus_speed = zdev->max_bus_speed; zdev 723 arch/s390/pci/pci.c pci_bus_add_devices(zdev->bus); zdev 727 arch/s390/pci/pci.c zpci_cleanup_bus_resources(zdev); zdev 732 arch/s390/pci/pci.c int zpci_enable_device(struct zpci_dev *zdev) zdev 736 arch/s390/pci/pci.c rc = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES); zdev 740 arch/s390/pci/pci.c rc = zpci_dma_init_device(zdev); zdev 744 arch/s390/pci/pci.c zdev->state = ZPCI_FN_STATE_ONLINE; zdev 748 arch/s390/pci/pci.c clp_disable_fh(zdev); zdev 754 arch/s390/pci/pci.c int zpci_disable_device(struct zpci_dev *zdev) zdev 756 arch/s390/pci/pci.c zpci_dma_exit_device(zdev); zdev 757 arch/s390/pci/pci.c return clp_disable_fh(zdev); zdev 761 arch/s390/pci/pci.c int zpci_create_device(struct zpci_dev *zdev) zdev 765 arch/s390/pci/pci.c rc = zpci_alloc_domain(zdev); zdev 769 arch/s390/pci/pci.c rc = zpci_init_iommu(zdev); zdev 773 arch/s390/pci/pci.c mutex_init(&zdev->lock); zdev 774 arch/s390/pci/pci.c if (zdev->state == ZPCI_FN_STATE_CONFIGURED) { zdev 775 arch/s390/pci/pci.c rc = zpci_enable_device(zdev); zdev 779 arch/s390/pci/pci.c rc = zpci_scan_bus(zdev); zdev 784 arch/s390/pci/pci.c list_add_tail(&zdev->entry, &zpci_list); zdev 787 arch/s390/pci/pci.c zpci_init_slot(zdev); zdev 792 arch/s390/pci/pci.c if (zdev->state == ZPCI_FN_STATE_ONLINE) zdev 793 arch/s390/pci/pci.c zpci_disable_device(zdev); zdev 795 arch/s390/pci/pci.c zpci_destroy_iommu(zdev); zdev 797 arch/s390/pci/pci.c zpci_free_domain(zdev); zdev 802 arch/s390/pci/pci.c void zpci_remove_device(struct zpci_dev *zdev) zdev 804 arch/s390/pci/pci.c if (!zdev->bus) zdev 807 arch/s390/pci/pci.c pci_stop_root_bus(zdev->bus); zdev 808 arch/s390/pci/pci.c pci_remove_root_bus(zdev->bus); zdev 814 arch/s390/pci/pci.c struct zpci_dev *zdev = to_zpci(pdev); zdev 816 arch/s390/pci/pci.c return sclp_pci_report(report, zdev->fh, zdev->fid); zdev 97 arch/s390/pci/pci_clp.c static void clp_store_query_pci_fngrp(struct zpci_dev *zdev, zdev 100 arch/s390/pci/pci_clp.c zdev->tlb_refresh = response->refresh; zdev 101 arch/s390/pci/pci_clp.c zdev->dma_mask = response->dasm; zdev 102 arch/s390/pci/pci_clp.c zdev->msi_addr = response->msia; zdev 103 arch/s390/pci/pci_clp.c zdev->max_msi = response->noi; zdev 104 arch/s390/pci/pci_clp.c zdev->fmb_update = response->mui; zdev 108 arch/s390/pci/pci_clp.c zdev->max_bus_speed = PCIE_SPEED_5_0GT; zdev 111 arch/s390/pci/pci_clp.c zdev->max_bus_speed = PCI_SPEED_UNKNOWN; zdev 116 arch/s390/pci/pci_clp.c static int clp_query_pci_fngrp(struct zpci_dev *zdev, u8 pfgid) zdev 133 arch/s390/pci/pci_clp.c clp_store_query_pci_fngrp(zdev, &rrb->response); zdev 143 arch/s390/pci/pci_clp.c static int clp_store_query_pci_fn(struct zpci_dev *zdev, zdev 149 arch/s390/pci/pci_clp.c zdev->bars[i].val = le32_to_cpu(response->bar[i]); zdev 150 arch/s390/pci/pci_clp.c zdev->bars[i].size = response->bar_size[i]; zdev 152 arch/s390/pci/pci_clp.c zdev->start_dma = response->sdma; zdev 153 arch/s390/pci/pci_clp.c zdev->end_dma = response->edma; zdev 154 arch/s390/pci/pci_clp.c zdev->pchid = response->pchid; zdev 155 arch/s390/pci/pci_clp.c zdev->pfgid = response->pfgid; zdev 156 arch/s390/pci/pci_clp.c zdev->pft = response->pft; zdev 157 arch/s390/pci/pci_clp.c zdev->vfn = response->vfn; zdev 158 arch/s390/pci/pci_clp.c zdev->uid = response->uid; zdev 159 arch/s390/pci/pci_clp.c zdev->fmb_length = sizeof(u32) * response->fmb_len; zdev 161 arch/s390/pci/pci_clp.c memcpy(zdev->pfip, response->pfip, sizeof(zdev->pfip)); zdev 163 arch/s390/pci/pci_clp.c memcpy(zdev->util_str, response->util_str, zdev 164 arch/s390/pci/pci_clp.c sizeof(zdev->util_str)); zdev 166 arch/s390/pci/pci_clp.c zdev->mio_capable = response->mio_addr_avail; zdev 171 arch/s390/pci/pci_clp.c zdev->bars[i].mio_wb = (void __iomem *) response->mio.addr[i].wb; zdev 172 arch/s390/pci/pci_clp.c zdev->bars[i].mio_wt = (void __iomem *) response->mio.addr[i].wt; zdev 177 arch/s390/pci/pci_clp.c static int clp_query_pci_fn(struct zpci_dev *zdev, u32 fh) zdev 194 arch/s390/pci/pci_clp.c rc = clp_store_query_pci_fn(zdev, &rrb->response); zdev 197 arch/s390/pci/pci_clp.c rc = clp_query_pci_fngrp(zdev, rrb->response.pfgid); zdev 210 arch/s390/pci/pci_clp.c struct zpci_dev *zdev; zdev 214 arch/s390/pci/pci_clp.c zdev = kzalloc(sizeof(*zdev), GFP_KERNEL); zdev 215 arch/s390/pci/pci_clp.c if (!zdev) zdev 218 arch/s390/pci/pci_clp.c zdev->fh = fh; zdev 219 arch/s390/pci/pci_clp.c zdev->fid = fid; zdev 222 arch/s390/pci/pci_clp.c rc = clp_query_pci_fn(zdev, fh); zdev 227 arch/s390/pci/pci_clp.c zdev->state = ZPCI_FN_STATE_CONFIGURED; zdev 229 arch/s390/pci/pci_clp.c zdev->state = ZPCI_FN_STATE_STANDBY; zdev 231 arch/s390/pci/pci_clp.c rc = zpci_create_device(zdev); zdev 238 arch/s390/pci/pci_clp.c kfree(zdev); zdev 246 arch/s390/pci/pci_clp.c static int clp_set_pci_fn(struct zpci_dev *zdev, u8 nr_dma_as, u8 command) zdev 250 arch/s390/pci/pci_clp.c u32 fid = zdev->fid; zdev 261 arch/s390/pci/pci_clp.c rrb->request.fh = zdev->fh; zdev 280 arch/s390/pci/pci_clp.c zdev->fh = rrb->response.fh; zdev 290 arch/s390/pci/pci_clp.c int clp_enable_fh(struct zpci_dev *zdev, u8 nr_dma_as) zdev 294 arch/s390/pci/pci_clp.c rc = clp_set_pci_fn(zdev, nr_dma_as, CLP_SET_ENABLE_PCI_FN); zdev 295 arch/s390/pci/pci_clp.c zpci_dbg(3, "ena fid:%x, fh:%x, rc:%d\n", zdev->fid, zdev->fh, rc); zdev 299 arch/s390/pci/pci_clp.c if (zpci_use_mio(zdev)) { zdev 300 arch/s390/pci/pci_clp.c rc = clp_set_pci_fn(zdev, nr_dma_as, CLP_SET_ENABLE_MIO); zdev 302 arch/s390/pci/pci_clp.c zdev->fid, zdev->fh, rc); zdev 304 arch/s390/pci/pci_clp.c clp_disable_fh(zdev); zdev 310 arch/s390/pci/pci_clp.c int clp_disable_fh(struct zpci_dev *zdev) zdev 314 arch/s390/pci/pci_clp.c if (!zdev_enabled(zdev)) zdev 317 arch/s390/pci/pci_clp.c rc = clp_set_pci_fn(zdev, 0, CLP_SET_DISABLE_PCI_FN); zdev 318 arch/s390/pci/pci_clp.c zpci_dbg(3, "dis fid:%x, fh:%x, rc:%d\n", zdev->fid, zdev->fh, rc); zdev 362 arch/s390/pci/pci_clp.c struct zpci_dev *zdev; zdev 367 arch/s390/pci/pci_clp.c zdev = get_zdev_by_fid(entry->fid); zdev 368 arch/s390/pci/pci_clp.c if (!zdev) zdev 374 arch/s390/pci/pci_clp.c struct zpci_dev *zdev; zdev 383 arch/s390/pci/pci_clp.c zdev = get_zdev_by_fid(entry->fid); zdev 384 arch/s390/pci/pci_clp.c if (!zdev) zdev 387 arch/s390/pci/pci_clp.c zdev->fh = entry->fh; zdev 72 arch/s390/pci/pci_debug.c struct zpci_dev *zdev = m->private; zdev 73 arch/s390/pci/pci_debug.c atomic64_t *counter = &zdev->allocated_pages; zdev 83 arch/s390/pci/pci_debug.c struct zpci_dev *zdev = m->private; zdev 85 arch/s390/pci/pci_debug.c if (!zdev) zdev 88 arch/s390/pci/pci_debug.c mutex_lock(&zdev->lock); zdev 89 arch/s390/pci/pci_debug.c if (!zdev->fmb) { zdev 90 arch/s390/pci/pci_debug.c mutex_unlock(&zdev->lock); zdev 96 arch/s390/pci/pci_debug.c seq_printf(m, "Update interval: %u ms\n", zdev->fmb_update); zdev 97 arch/s390/pci/pci_debug.c seq_printf(m, "Samples: %u\n", zdev->fmb->samples); zdev 98 arch/s390/pci/pci_debug.c seq_printf(m, "Last update TOD: %Lx\n", zdev->fmb->last_update); zdev 101 arch/s390/pci/pci_debug.c &zdev->fmb->ld_ops); zdev 103 arch/s390/pci/pci_debug.c switch (zdev->fmb->format) { zdev 105 arch/s390/pci/pci_debug.c if (!(zdev->fmb->fmt_ind & ZPCI_FMB_DMA_COUNTER_VALID)) zdev 108 arch/s390/pci/pci_debug.c &zdev->fmb->fmt0.dma_rbytes); zdev 112 arch/s390/pci/pci_debug.c &zdev->fmb->fmt1.rx_bytes); zdev 116 arch/s390/pci/pci_debug.c &zdev->fmb->fmt2.consumed_work_units); zdev 120 arch/s390/pci/pci_debug.c &zdev->fmb->fmt3.tx_bytes); zdev 127 arch/s390/pci/pci_debug.c mutex_unlock(&zdev->lock); zdev 134 arch/s390/pci/pci_debug.c struct zpci_dev *zdev = ((struct seq_file *) file->private_data)->private; zdev 138 arch/s390/pci/pci_debug.c if (!zdev) zdev 145 arch/s390/pci/pci_debug.c mutex_lock(&zdev->lock); zdev 148 arch/s390/pci/pci_debug.c rc = zpci_fmb_disable_device(zdev); zdev 151 arch/s390/pci/pci_debug.c rc = zpci_fmb_enable_device(zdev); zdev 154 arch/s390/pci/pci_debug.c mutex_unlock(&zdev->lock); zdev 172 arch/s390/pci/pci_debug.c void zpci_debug_init_device(struct zpci_dev *zdev, const char *name) zdev 174 arch/s390/pci/pci_debug.c zdev->debugfs_dev = debugfs_create_dir(name, debugfs_root); zdev 177 arch/s390/pci/pci_debug.c zdev->debugfs_dev, zdev, &debugfs_pci_perf_fops); zdev 180 arch/s390/pci/pci_debug.c void zpci_debug_exit_device(struct zpci_dev *zdev) zdev 182 arch/s390/pci/pci_debug.c debugfs_remove_recursive(zdev->debugfs_dev); zdev 22 arch/s390/pci/pci_dma.c static int zpci_refresh_global(struct zpci_dev *zdev) zdev 24 arch/s390/pci/pci_dma.c return zpci_refresh_trans((u64) zdev->fh << 32, zdev->start_dma, zdev 25 arch/s390/pci/pci_dma.c zdev->iommu_pages * PAGE_SIZE); zdev 133 arch/s390/pci/pci_dma.c static int __dma_update_trans(struct zpci_dev *zdev, unsigned long pa, zdev 145 arch/s390/pci/pci_dma.c spin_lock_irqsave(&zdev->dma_table_lock, irq_flags); zdev 146 arch/s390/pci/pci_dma.c if (!zdev->dma_table) { zdev 152 arch/s390/pci/pci_dma.c entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr); zdev 168 arch/s390/pci/pci_dma.c entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr); zdev 175 arch/s390/pci/pci_dma.c spin_unlock_irqrestore(&zdev->dma_table_lock, irq_flags); zdev 179 arch/s390/pci/pci_dma.c static int __dma_purge_tlb(struct zpci_dev *zdev, dma_addr_t dma_addr, zdev 193 arch/s390/pci/pci_dma.c if (!zdev->tlb_refresh) zdev 200 arch/s390/pci/pci_dma.c ret = zpci_refresh_trans((u64) zdev->fh << 32, dma_addr, zdev 204 arch/s390/pci/pci_dma.c if (zpci_refresh_global(zdev)) zdev 207 arch/s390/pci/pci_dma.c spin_lock_irqsave(&zdev->iommu_bitmap_lock, irqflags); zdev 208 arch/s390/pci/pci_dma.c bitmap_andnot(zdev->iommu_bitmap, zdev->iommu_bitmap, zdev 209 arch/s390/pci/pci_dma.c zdev->lazy_bitmap, zdev->iommu_pages); zdev 210 arch/s390/pci/pci_dma.c bitmap_zero(zdev->lazy_bitmap, zdev->iommu_pages); zdev 211 arch/s390/pci/pci_dma.c spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, irqflags); zdev 218 arch/s390/pci/pci_dma.c static int dma_update_trans(struct zpci_dev *zdev, unsigned long pa, zdev 223 arch/s390/pci/pci_dma.c rc = __dma_update_trans(zdev, pa, dma_addr, size, flags); zdev 227 arch/s390/pci/pci_dma.c rc = __dma_purge_tlb(zdev, dma_addr, size, flags); zdev 229 arch/s390/pci/pci_dma.c __dma_update_trans(zdev, pa, dma_addr, size, ZPCI_PTE_INVALID); zdev 263 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 268 arch/s390/pci/pci_dma.c return iommu_area_alloc(zdev->iommu_bitmap, zdev->iommu_pages, zdev 269 arch/s390/pci/pci_dma.c start, size, zdev->start_dma >> PAGE_SHIFT, zdev 275 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 278 arch/s390/pci/pci_dma.c spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags); zdev 279 arch/s390/pci/pci_dma.c offset = __dma_alloc_iommu(dev, zdev->next_bit, size); zdev 283 arch/s390/pci/pci_dma.c if (zpci_refresh_global(zdev)) zdev 286 arch/s390/pci/pci_dma.c bitmap_andnot(zdev->iommu_bitmap, zdev->iommu_bitmap, zdev 287 arch/s390/pci/pci_dma.c zdev->lazy_bitmap, zdev->iommu_pages); zdev 288 arch/s390/pci/pci_dma.c bitmap_zero(zdev->lazy_bitmap, zdev->iommu_pages); zdev 295 arch/s390/pci/pci_dma.c zdev->next_bit = offset + size; zdev 296 arch/s390/pci/pci_dma.c spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, flags); zdev 298 arch/s390/pci/pci_dma.c return zdev->start_dma + offset * PAGE_SIZE; zdev 301 arch/s390/pci/pci_dma.c spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, flags); zdev 307 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 310 arch/s390/pci/pci_dma.c offset = (dma_addr - zdev->start_dma) >> PAGE_SHIFT; zdev 312 arch/s390/pci/pci_dma.c spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags); zdev 313 arch/s390/pci/pci_dma.c if (!zdev->iommu_bitmap) zdev 317 arch/s390/pci/pci_dma.c bitmap_clear(zdev->iommu_bitmap, offset, size); zdev 319 arch/s390/pci/pci_dma.c bitmap_set(zdev->lazy_bitmap, offset, size); zdev 322 arch/s390/pci/pci_dma.c spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, flags); zdev 340 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 361 arch/s390/pci/pci_dma.c ret = dma_update_trans(zdev, pa, dma_addr, size, flags); zdev 365 arch/s390/pci/pci_dma.c atomic64_add(nr_pages, &zdev->mapped_pages); zdev 380 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 385 arch/s390/pci/pci_dma.c ret = dma_update_trans(zdev, 0, dma_addr, npages * PAGE_SIZE, zdev 393 arch/s390/pci/pci_dma.c atomic64_add(npages, &zdev->unmapped_pages); zdev 401 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 418 arch/s390/pci/pci_dma.c atomic64_add(size / PAGE_SIZE, &zdev->allocated_pages); zdev 428 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 431 arch/s390/pci/pci_dma.c atomic64_sub(size / PAGE_SIZE, &zdev->allocated_pages); zdev 442 arch/s390/pci/pci_dma.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 459 arch/s390/pci/pci_dma.c ret = __dma_update_trans(zdev, pa, dma_addr, zdev 466 arch/s390/pci/pci_dma.c ret = __dma_purge_tlb(zdev, dma_addr_base, size, flags); zdev 471 arch/s390/pci/pci_dma.c atomic64_add(nr_pages, &zdev->mapped_pages); zdev 476 arch/s390/pci/pci_dma.c dma_update_trans(zdev, 0, dma_addr_base, dma_addr - dma_addr_base, zdev 547 arch/s390/pci/pci_dma.c int zpci_dma_init_device(struct zpci_dev *zdev) zdev 556 arch/s390/pci/pci_dma.c WARN_ON(zdev->s390_domain); zdev 558 arch/s390/pci/pci_dma.c spin_lock_init(&zdev->iommu_bitmap_lock); zdev 559 arch/s390/pci/pci_dma.c spin_lock_init(&zdev->dma_table_lock); zdev 561 arch/s390/pci/pci_dma.c zdev->dma_table = dma_alloc_cpu_table(); zdev 562 arch/s390/pci/pci_dma.c if (!zdev->dma_table) { zdev 576 arch/s390/pci/pci_dma.c zdev->start_dma = PAGE_ALIGN(zdev->start_dma); zdev 577 arch/s390/pci/pci_dma.c zdev->iommu_size = min3((u64) high_memory, zdev 578 arch/s390/pci/pci_dma.c ZPCI_TABLE_SIZE_RT - zdev->start_dma, zdev 579 arch/s390/pci/pci_dma.c zdev->end_dma - zdev->start_dma + 1); zdev 580 arch/s390/pci/pci_dma.c zdev->end_dma = zdev->start_dma + zdev->iommu_size - 1; zdev 581 arch/s390/pci/pci_dma.c zdev->iommu_pages = zdev->iommu_size >> PAGE_SHIFT; zdev 582 arch/s390/pci/pci_dma.c zdev->iommu_bitmap = vzalloc(zdev->iommu_pages / 8); zdev 583 arch/s390/pci/pci_dma.c if (!zdev->iommu_bitmap) { zdev 588 arch/s390/pci/pci_dma.c zdev->lazy_bitmap = vzalloc(zdev->iommu_pages / 8); zdev 589 arch/s390/pci/pci_dma.c if (!zdev->lazy_bitmap) { zdev 595 arch/s390/pci/pci_dma.c rc = zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma, zdev 596 arch/s390/pci/pci_dma.c (u64) zdev->dma_table); zdev 602 arch/s390/pci/pci_dma.c vfree(zdev->iommu_bitmap); zdev 603 arch/s390/pci/pci_dma.c zdev->iommu_bitmap = NULL; zdev 604 arch/s390/pci/pci_dma.c vfree(zdev->lazy_bitmap); zdev 605 arch/s390/pci/pci_dma.c zdev->lazy_bitmap = NULL; zdev 607 arch/s390/pci/pci_dma.c dma_free_cpu_table(zdev->dma_table); zdev 608 arch/s390/pci/pci_dma.c zdev->dma_table = NULL; zdev 613 arch/s390/pci/pci_dma.c void zpci_dma_exit_device(struct zpci_dev *zdev) zdev 620 arch/s390/pci/pci_dma.c WARN_ON(zdev->s390_domain); zdev 622 arch/s390/pci/pci_dma.c if (zpci_unregister_ioat(zdev, 0)) zdev 625 arch/s390/pci/pci_dma.c dma_cleanup_tables(zdev->dma_table); zdev 626 arch/s390/pci/pci_dma.c zdev->dma_table = NULL; zdev 627 arch/s390/pci/pci_dma.c vfree(zdev->iommu_bitmap); zdev 628 arch/s390/pci/pci_dma.c zdev->iommu_bitmap = NULL; zdev 629 arch/s390/pci/pci_dma.c vfree(zdev->lazy_bitmap); zdev 630 arch/s390/pci/pci_dma.c zdev->lazy_bitmap = NULL; zdev 632 arch/s390/pci/pci_dma.c zdev->next_bit = 0; zdev 49 arch/s390/pci/pci_event.c struct zpci_dev *zdev = get_zdev_by_fid(ccdf->fid); zdev 55 arch/s390/pci/pci_event.c if (zdev) zdev 56 arch/s390/pci/pci_event.c pdev = pci_get_slot(zdev->bus, ZPCI_DEVFN); zdev 76 arch/s390/pci/pci_event.c struct zpci_dev *zdev = get_zdev_by_fid(ccdf->fid); zdev 81 arch/s390/pci/pci_event.c if (zdev) zdev 82 arch/s390/pci/pci_event.c pdev = pci_get_slot(zdev->bus, ZPCI_DEVFN); zdev 91 arch/s390/pci/pci_event.c if (!zdev) { zdev 95 arch/s390/pci/pci_event.c zdev = get_zdev_by_fid(ccdf->fid); zdev 97 arch/s390/pci/pci_event.c if (!zdev || zdev->state != ZPCI_FN_STATE_STANDBY) zdev 99 arch/s390/pci/pci_event.c zdev->state = ZPCI_FN_STATE_CONFIGURED; zdev 100 arch/s390/pci/pci_event.c zdev->fh = ccdf->fh; zdev 101 arch/s390/pci/pci_event.c ret = zpci_enable_device(zdev); zdev 105 arch/s390/pci/pci_event.c pci_rescan_bus(zdev->bus); zdev 109 arch/s390/pci/pci_event.c if (!zdev) zdev 113 arch/s390/pci/pci_event.c if (!zdev) zdev 118 arch/s390/pci/pci_event.c ret = zpci_disable_device(zdev); zdev 122 arch/s390/pci/pci_event.c ret = sclp_pci_deconfigure(zdev->fid); zdev 123 arch/s390/pci/pci_event.c zpci_dbg(3, "deconf fid:%x, rc:%d\n", zdev->fid, ret); zdev 125 arch/s390/pci/pci_event.c zdev->state = ZPCI_FN_STATE_STANDBY; zdev 129 arch/s390/pci/pci_event.c if (!zdev) zdev 138 arch/s390/pci/pci_event.c zdev->fh = ccdf->fh; zdev 139 arch/s390/pci/pci_event.c zpci_disable_device(zdev); zdev 140 arch/s390/pci/pci_event.c zdev->state = ZPCI_FN_STATE_STANDBY; zdev 143 arch/s390/pci/pci_event.c zpci_remove_device(zdev); zdev 150 arch/s390/pci/pci_event.c if (!zdev) zdev 152 arch/s390/pci/pci_event.c zpci_remove_device(zdev); zdev 39 arch/s390/pci/pci_irq.c static int zpci_set_airq(struct zpci_dev *zdev) zdev 41 arch/s390/pci/pci_irq.c u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_REG_INT); zdev 47 arch/s390/pci/pci_irq.c fib.fmt0.noi = airq_iv_end(zdev->aibv); zdev 48 arch/s390/pci/pci_irq.c fib.fmt0.aibv = (unsigned long) zdev->aibv->vector; zdev 50 arch/s390/pci/pci_irq.c fib.fmt0.aisb = (unsigned long) zpci_sbv->vector + (zdev->aisb/64)*8; zdev 51 arch/s390/pci/pci_irq.c fib.fmt0.aisbo = zdev->aisb & 63; zdev 57 arch/s390/pci/pci_irq.c static int zpci_clear_airq(struct zpci_dev *zdev) zdev 59 arch/s390/pci/pci_irq.c u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_DEREG_INT); zdev 72 arch/s390/pci/pci_irq.c static int zpci_set_directed_irq(struct zpci_dev *zdev) zdev 74 arch/s390/pci/pci_irq.c u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_REG_INT_D); zdev 79 arch/s390/pci/pci_irq.c fib.fmt1.noi = zdev->msi_nr_irqs; zdev 80 arch/s390/pci/pci_irq.c fib.fmt1.dibvo = zdev->msi_first_bit; zdev 86 arch/s390/pci/pci_irq.c static int zpci_clear_directed_irq(struct zpci_dev *zdev) zdev 88 arch/s390/pci/pci_irq.c u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_DEREG_INT_D); zdev 236 arch/s390/pci/pci_irq.c struct zpci_dev *zdev = to_zpci(pdev); zdev 243 arch/s390/pci/pci_irq.c zdev->aisb = -1UL; zdev 244 arch/s390/pci/pci_irq.c zdev->msi_first_bit = -1U; zdev 247 arch/s390/pci/pci_irq.c msi_vecs = min_t(unsigned int, nvec, zdev->max_msi); zdev 259 arch/s390/pci/pci_irq.c zdev->aisb = bit; zdev 262 arch/s390/pci/pci_irq.c zdev->aibv = airq_iv_create(msi_vecs, AIRQ_IV_DATA | AIRQ_IV_BITLOCK); zdev 263 arch/s390/pci/pci_irq.c if (!zdev->aibv) zdev 267 arch/s390/pci/pci_irq.c zpci_ibv[bit] = zdev->aibv; zdev 290 arch/s390/pci/pci_irq.c msg.address_lo = zdev->msi_addr & 0xff0000ff; zdev 297 arch/s390/pci/pci_irq.c msg.address_lo = zdev->msi_addr & 0xffffffff; zdev 298 arch/s390/pci/pci_irq.c airq_iv_set_data(zdev->aibv, hwirq, irq); zdev 300 arch/s390/pci/pci_irq.c msg.address_hi = zdev->msi_addr >> 32; zdev 305 arch/s390/pci/pci_irq.c zdev->msi_first_bit = bit; zdev 306 arch/s390/pci/pci_irq.c zdev->msi_nr_irqs = msi_vecs; zdev 309 arch/s390/pci/pci_irq.c rc = zpci_set_directed_irq(zdev); zdev 311 arch/s390/pci/pci_irq.c rc = zpci_set_airq(zdev); zdev 320 arch/s390/pci/pci_irq.c struct zpci_dev *zdev = to_zpci(pdev); zdev 326 arch/s390/pci/pci_irq.c rc = zpci_clear_directed_irq(zdev); zdev 328 arch/s390/pci/pci_irq.c rc = zpci_clear_airq(zdev); zdev 348 arch/s390/pci/pci_irq.c if (zdev->aisb != -1UL) { zdev 349 arch/s390/pci/pci_irq.c zpci_ibv[zdev->aisb] = NULL; zdev 350 arch/s390/pci/pci_irq.c airq_iv_free_bit(zpci_sbv, zdev->aisb); zdev 351 arch/s390/pci/pci_irq.c zdev->aisb = -1UL; zdev 353 arch/s390/pci/pci_irq.c if (zdev->aibv) { zdev 354 arch/s390/pci/pci_irq.c airq_iv_release(zdev->aibv); zdev 355 arch/s390/pci/pci_irq.c zdev->aibv = NULL; zdev 358 arch/s390/pci/pci_irq.c if ((irq_delivery == DIRECTED) && zdev->msi_first_bit != -1U) zdev 359 arch/s390/pci/pci_irq.c airq_iv_free(zpci_ibv[0], zdev->msi_first_bit, zdev->msi_nr_irqs); zdev 24 arch/s390/pci/pci_sysfs.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); \ zdev 26 arch/s390/pci/pci_sysfs.c return sprintf(buf, fmt, zdev->member); \ zdev 45 arch/s390/pci/pci_sysfs.c struct zpci_dev *zdev = to_zpci(to_pci_dev(dev)); zdev 47 arch/s390/pci/pci_sysfs.c return sprintf(buf, zpci_use_mio(zdev) ? "1\n" : "0\n"); zdev 56 arch/s390/pci/pci_sysfs.c struct zpci_dev *zdev = to_zpci(pdev); zdev 84 arch/s390/pci/pci_sysfs.c ret = zpci_disable_device(zdev); zdev 88 arch/s390/pci/pci_sysfs.c ret = zpci_enable_device(zdev); zdev 91 arch/s390/pci/pci_sysfs.c pci_rescan_bus(zdev->bus); zdev 107 arch/s390/pci/pci_sysfs.c struct zpci_dev *zdev = to_zpci(pdev); zdev 109 arch/s390/pci/pci_sysfs.c return memory_read_from_buffer(buf, count, &off, zdev->util_str, zdev 110 arch/s390/pci/pci_sysfs.c sizeof(zdev->util_str)); zdev 121 arch/s390/pci/pci_sysfs.c struct zpci_dev *zdev = to_zpci(pdev); zdev 127 arch/s390/pci/pci_sysfs.c ret = sclp_pci_report(report, zdev->fh, zdev->fid); zdev 216 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_device *zdev; zdev 858 drivers/dma/xilinx/zynqmp_dma.c devm_free_irq(chan->zdev->dev, chan->irq, chan); zdev 870 drivers/dma/xilinx/zynqmp_dma.c static int zynqmp_dma_chan_probe(struct zynqmp_dma_device *zdev, zdev 878 drivers/dma/xilinx/zynqmp_dma.c chan = devm_kzalloc(zdev->dev, sizeof(*chan), GFP_KERNEL); zdev 881 drivers/dma/xilinx/zynqmp_dma.c chan->dev = zdev->dev; zdev 882 drivers/dma/xilinx/zynqmp_dma.c chan->zdev = zdev; zdev 900 drivers/dma/xilinx/zynqmp_dma.c dev_err(zdev->dev, "invalid bus-width value"); zdev 905 drivers/dma/xilinx/zynqmp_dma.c zdev->chan = chan; zdev 914 drivers/dma/xilinx/zynqmp_dma.c chan->common.device = &zdev->common; zdev 915 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&chan->common.device_node, &zdev->common.channels); zdev 941 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_device *zdev = ofdma->of_dma_data; zdev 943 drivers/dma/xilinx/zynqmp_dma.c return dma_get_slave_channel(&zdev->chan->common); zdev 985 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_device *zdev = dev_get_drvdata(dev); zdev 987 drivers/dma/xilinx/zynqmp_dma.c clk_disable_unprepare(zdev->clk_main); zdev 988 drivers/dma/xilinx/zynqmp_dma.c clk_disable_unprepare(zdev->clk_apb); zdev 1002 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_device *zdev = dev_get_drvdata(dev); zdev 1005 drivers/dma/xilinx/zynqmp_dma.c err = clk_prepare_enable(zdev->clk_main); zdev 1011 drivers/dma/xilinx/zynqmp_dma.c err = clk_prepare_enable(zdev->clk_apb); zdev 1014 drivers/dma/xilinx/zynqmp_dma.c clk_disable_unprepare(zdev->clk_main); zdev 1035 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_device *zdev; zdev 1039 drivers/dma/xilinx/zynqmp_dma.c zdev = devm_kzalloc(&pdev->dev, sizeof(*zdev), GFP_KERNEL); zdev 1040 drivers/dma/xilinx/zynqmp_dma.c if (!zdev) zdev 1043 drivers/dma/xilinx/zynqmp_dma.c zdev->dev = &pdev->dev; zdev 1044 drivers/dma/xilinx/zynqmp_dma.c INIT_LIST_HEAD(&zdev->common.channels); zdev 1047 drivers/dma/xilinx/zynqmp_dma.c dma_cap_set(DMA_MEMCPY, zdev->common.cap_mask); zdev 1049 drivers/dma/xilinx/zynqmp_dma.c p = &zdev->common; zdev 1059 drivers/dma/xilinx/zynqmp_dma.c zdev->clk_main = devm_clk_get(&pdev->dev, "clk_main"); zdev 1060 drivers/dma/xilinx/zynqmp_dma.c if (IS_ERR(zdev->clk_main)) { zdev 1062 drivers/dma/xilinx/zynqmp_dma.c return PTR_ERR(zdev->clk_main); zdev 1065 drivers/dma/xilinx/zynqmp_dma.c zdev->clk_apb = devm_clk_get(&pdev->dev, "clk_apb"); zdev 1066 drivers/dma/xilinx/zynqmp_dma.c if (IS_ERR(zdev->clk_apb)) { zdev 1068 drivers/dma/xilinx/zynqmp_dma.c return PTR_ERR(zdev->clk_apb); zdev 1071 drivers/dma/xilinx/zynqmp_dma.c platform_set_drvdata(pdev, zdev); zdev 1072 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_set_autosuspend_delay(zdev->dev, ZDMA_PM_TIMEOUT); zdev 1073 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_use_autosuspend(zdev->dev); zdev 1074 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_enable(zdev->dev); zdev 1075 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_get_sync(zdev->dev); zdev 1076 drivers/dma/xilinx/zynqmp_dma.c if (!pm_runtime_enabled(zdev->dev)) { zdev 1077 drivers/dma/xilinx/zynqmp_dma.c ret = zynqmp_dma_runtime_resume(zdev->dev); zdev 1082 drivers/dma/xilinx/zynqmp_dma.c ret = zynqmp_dma_chan_probe(zdev, pdev); zdev 1088 drivers/dma/xilinx/zynqmp_dma.c p->dst_addr_widths = BIT(zdev->chan->bus_width / 8); zdev 1089 drivers/dma/xilinx/zynqmp_dma.c p->src_addr_widths = BIT(zdev->chan->bus_width / 8); zdev 1091 drivers/dma/xilinx/zynqmp_dma.c dma_async_device_register(&zdev->common); zdev 1094 drivers/dma/xilinx/zynqmp_dma.c of_zynqmp_dma_xlate, zdev); zdev 1097 drivers/dma/xilinx/zynqmp_dma.c dma_async_device_unregister(&zdev->common); zdev 1101 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_mark_last_busy(zdev->dev); zdev 1102 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_put_sync_autosuspend(zdev->dev); zdev 1109 drivers/dma/xilinx/zynqmp_dma.c zynqmp_dma_chan_remove(zdev->chan); zdev 1111 drivers/dma/xilinx/zynqmp_dma.c if (!pm_runtime_enabled(zdev->dev)) zdev 1112 drivers/dma/xilinx/zynqmp_dma.c zynqmp_dma_runtime_suspend(zdev->dev); zdev 1113 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_disable(zdev->dev); zdev 1125 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_device *zdev = platform_get_drvdata(pdev); zdev 1128 drivers/dma/xilinx/zynqmp_dma.c dma_async_device_unregister(&zdev->common); zdev 1130 drivers/dma/xilinx/zynqmp_dma.c zynqmp_dma_chan_remove(zdev->chan); zdev 1131 drivers/dma/xilinx/zynqmp_dma.c pm_runtime_disable(zdev->dev); zdev 1132 drivers/dma/xilinx/zynqmp_dma.c if (!pm_runtime_enabled(zdev->dev)) zdev 1133 drivers/dma/xilinx/zynqmp_dma.c zynqmp_dma_runtime_suspend(zdev->dev); zdev 34 drivers/iommu/s390-iommu.c struct zpci_dev *zdev; zdev 90 drivers/iommu/s390-iommu.c struct zpci_dev *zdev = to_pci_dev(dev)->sysdata; zdev 95 drivers/iommu/s390-iommu.c if (!zdev) zdev 102 drivers/iommu/s390-iommu.c if (zdev->dma_table) zdev 103 drivers/iommu/s390-iommu.c zpci_dma_exit_device(zdev); zdev 105 drivers/iommu/s390-iommu.c zdev->dma_table = s390_domain->dma_table; zdev 106 drivers/iommu/s390-iommu.c rc = zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma, zdev 107 drivers/iommu/s390-iommu.c (u64) zdev->dma_table); zdev 114 drivers/iommu/s390-iommu.c domain->geometry.aperture_start = zdev->start_dma; zdev 115 drivers/iommu/s390-iommu.c domain->geometry.aperture_end = zdev->end_dma; zdev 118 drivers/iommu/s390-iommu.c } else if (domain->geometry.aperture_start != zdev->start_dma || zdev 119 drivers/iommu/s390-iommu.c domain->geometry.aperture_end != zdev->end_dma) { zdev 124 drivers/iommu/s390-iommu.c domain_device->zdev = zdev; zdev 125 drivers/iommu/s390-iommu.c zdev->s390_domain = s390_domain; zdev 132 drivers/iommu/s390-iommu.c zpci_dma_init_device(zdev); zdev 142 drivers/iommu/s390-iommu.c struct zpci_dev *zdev = to_pci_dev(dev)->sysdata; zdev 147 drivers/iommu/s390-iommu.c if (!zdev) zdev 153 drivers/iommu/s390-iommu.c if (domain_device->zdev == zdev) { zdev 163 drivers/iommu/s390-iommu.c zdev->s390_domain = NULL; zdev 164 drivers/iommu/s390-iommu.c zpci_unregister_ioat(zdev, 0); zdev 165 drivers/iommu/s390-iommu.c zpci_dma_init_device(zdev); zdev 172 drivers/iommu/s390-iommu.c struct zpci_dev *zdev = to_pci_dev(dev)->sysdata; zdev 178 drivers/iommu/s390-iommu.c iommu_device_link(&zdev->iommu_dev, dev); zdev 185 drivers/iommu/s390-iommu.c struct zpci_dev *zdev = to_pci_dev(dev)->sysdata; zdev 199 drivers/iommu/s390-iommu.c if (zdev && zdev->s390_domain) { zdev 205 drivers/iommu/s390-iommu.c iommu_device_unlink(&zdev->iommu_dev, dev); zdev 242 drivers/iommu/s390-iommu.c rc = zpci_refresh_trans((u64) domain_device->zdev->fh << 32, zdev 337 drivers/iommu/s390-iommu.c int zpci_init_iommu(struct zpci_dev *zdev) zdev 341 drivers/iommu/s390-iommu.c rc = iommu_device_sysfs_add(&zdev->iommu_dev, NULL, NULL, zdev 342 drivers/iommu/s390-iommu.c "s390-iommu.%08x", zdev->fid); zdev 346 drivers/iommu/s390-iommu.c iommu_device_set_ops(&zdev->iommu_dev, &s390_iommu_ops); zdev 348 drivers/iommu/s390-iommu.c rc = iommu_device_register(&zdev->iommu_dev); zdev 355 drivers/iommu/s390-iommu.c iommu_device_sysfs_remove(&zdev->iommu_dev); zdev 361 drivers/iommu/s390-iommu.c void zpci_destroy_iommu(struct zpci_dev *zdev) zdev 363 drivers/iommu/s390-iommu.c iommu_device_unregister(&zdev->iommu_dev); zdev 364 drivers/iommu/s390-iommu.c iommu_device_sysfs_remove(&zdev->iommu_dev); zdev 244 drivers/net/ethernet/8390/xsurf100.c static int xsurf100_probe(struct zorro_dev *zdev, zdev 252 drivers/net/ethernet/8390/xsurf100.c DEFINE_RES_MEM(zdev->resource.start + XS100_8390_BASE, zdev 267 drivers/net/ethernet/8390/xsurf100.c if (!request_mem_region(zdev->resource.start, 0x100, zdev->name)) { zdev 268 drivers/net/ethernet/8390/xsurf100.c dev_err(&zdev->dev, "cannot reserve X-Surf 100 control registers\n"); zdev 272 drivers/net/ethernet/8390/xsurf100.c if (!request_mem_region(zdev->resource.start + zdev 276 drivers/net/ethernet/8390/xsurf100.c dev_err(&zdev->dev, "cannot reserve 32-bit area\n"); zdev 291 drivers/net/ethernet/8390/xsurf100.c ax88796_data.base_regs = ioremap(zdev->resource.start, 0x100); zdev 295 drivers/net/ethernet/8390/xsurf100.c dev_err(&zdev->dev, "Cannot ioremap area %pR (registers)\n", zdev 296 drivers/net/ethernet/8390/xsurf100.c &zdev->resource); zdev 302 drivers/net/ethernet/8390/xsurf100.c ax88796_data.data_area = ioremap(zdev->resource.start + zdev 307 drivers/net/ethernet/8390/xsurf100.c dev_err(&zdev->dev, zdev 309 drivers/net/ethernet/8390/xsurf100.c &zdev->resource, XS100_8390_DATA32_BASE); zdev 318 drivers/net/ethernet/8390/xsurf100.c pdev = platform_device_register_resndata(&zdev->dev, "ax88796", zdev 319 drivers/net/ethernet/8390/xsurf100.c zdev->slotaddr, res, 2, zdev 324 drivers/net/ethernet/8390/xsurf100.c dev_err(&zdev->dev, "cannot register platform device\n"); zdev 329 drivers/net/ethernet/8390/xsurf100.c zorro_set_drvdata(zdev, pdev); zdev 341 drivers/net/ethernet/8390/xsurf100.c release_mem_region(zdev->resource.start + XS100_8390_DATA32_BASE, zdev 345 drivers/net/ethernet/8390/xsurf100.c release_mem_region(zdev->resource.start, 0x100); zdev 350 drivers/net/ethernet/8390/xsurf100.c static void xsurf100_remove(struct zorro_dev *zdev) zdev 352 drivers/net/ethernet/8390/xsurf100.c struct platform_device *pdev = zorro_get_drvdata(zdev); zdev 358 drivers/net/ethernet/8390/xsurf100.c release_mem_region(zdev->resource.start, 0x100); zdev 360 drivers/net/ethernet/8390/xsurf100.c release_mem_region(zdev->resource.start + XS100_8390_DATA32_BASE, zdev 36 drivers/pci/hotplug/s390_pci_hpc.c struct zpci_dev *zdev; zdev 46 drivers/pci/hotplug/s390_pci_hpc.c int ret = sclp_pci_configure(slot->zdev->fid); zdev 48 drivers/pci/hotplug/s390_pci_hpc.c zpci_dbg(3, "conf fid:%x, rc:%d\n", slot->zdev->fid, ret); zdev 50 drivers/pci/hotplug/s390_pci_hpc.c slot->zdev->state = ZPCI_FN_STATE_CONFIGURED; zdev 57 drivers/pci/hotplug/s390_pci_hpc.c int ret = sclp_pci_deconfigure(slot->zdev->fid); zdev 59 drivers/pci/hotplug/s390_pci_hpc.c zpci_dbg(3, "deconf fid:%x, rc:%d\n", slot->zdev->fid, ret); zdev 61 drivers/pci/hotplug/s390_pci_hpc.c slot->zdev->state = ZPCI_FN_STATE_STANDBY; zdev 71 drivers/pci/hotplug/s390_pci_hpc.c if (slot->zdev->state != ZPCI_FN_STATE_STANDBY) zdev 78 drivers/pci/hotplug/s390_pci_hpc.c rc = zpci_enable_device(slot->zdev); zdev 82 drivers/pci/hotplug/s390_pci_hpc.c pci_scan_slot(slot->zdev->bus, ZPCI_DEVFN); zdev 84 drivers/pci/hotplug/s390_pci_hpc.c pci_bus_add_devices(slot->zdev->bus); zdev 100 drivers/pci/hotplug/s390_pci_hpc.c if (!zpci_fn_configured(slot->zdev->state)) zdev 103 drivers/pci/hotplug/s390_pci_hpc.c pdev = pci_get_slot(slot->zdev->bus, ZPCI_DEVFN); zdev 109 drivers/pci/hotplug/s390_pci_hpc.c rc = zpci_disable_device(slot->zdev); zdev 120 drivers/pci/hotplug/s390_pci_hpc.c switch (slot->zdev->state) { zdev 145 drivers/pci/hotplug/s390_pci_hpc.c int zpci_init_slot(struct zpci_dev *zdev) zdev 151 drivers/pci/hotplug/s390_pci_hpc.c if (!zdev) zdev 158 drivers/pci/hotplug/s390_pci_hpc.c slot->zdev = zdev; zdev 161 drivers/pci/hotplug/s390_pci_hpc.c snprintf(name, SLOT_NAME_SIZE, "%08x", zdev->fid); zdev 162 drivers/pci/hotplug/s390_pci_hpc.c rc = pci_hp_register(&slot->hotplug_slot, zdev->bus, zdev 176 drivers/pci/hotplug/s390_pci_hpc.c void zpci_exit_slot(struct zpci_dev *zdev) zdev 182 drivers/pci/hotplug/s390_pci_hpc.c if (slot->zdev != zdev) zdev 127 drivers/s390/crypto/zcrypt_card.c struct zcrypt_card *zdev = zdev 129 drivers/s390/crypto/zcrypt_card.c zcrypt_card_free(zdev); zdev 214 drivers/s390/net/ism.h struct zpci_dev *zdev = to_zpci(ism->pdev); zdev 215 drivers/s390/net/ism.h u64 req = ZPCI_CREATE_REQ(zdev->fh, 2, 8); zdev 228 drivers/s390/net/ism.h struct zpci_dev *zdev = to_zpci(ism->pdev); zdev 229 drivers/s390/net/ism.h u64 req = ZPCI_CREATE_REQ(zdev->fh, 2, len); zdev 238 drivers/s390/net/ism.h struct zpci_dev *zdev = to_zpci(ism->pdev); zdev 239 drivers/s390/net/ism.h u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, size); zdev 1947 drivers/video/fbdev/cirrusfb.c struct zorro_dev *zdev = to_zorro_dev(info->device); zdev 1954 drivers/video/fbdev/cirrusfb.c zorro_release_device(zdev);