zdev              163 arch/s390/include/asm/pci.h static inline bool zdev_enabled(struct zpci_dev *zdev)
zdev              165 arch/s390/include/asm/pci.h 	return (zdev->fh & (1UL << 31)) ? true : false;
zdev              176 arch/s390/include/asm/pci.h void zpci_remove_device(struct zpci_dev *zdev);
zdev              193 arch/s390/include/asm/pci.h int zpci_init_iommu(struct zpci_dev *zdev);
zdev              194 arch/s390/include/asm/pci.h void zpci_destroy_iommu(struct zpci_dev *zdev);
zdev              197 arch/s390/include/asm/pci.h static inline bool zpci_use_mio(struct zpci_dev *zdev)
zdev              199 arch/s390/include/asm/pci.h 	return static_branch_likely(&have_mio) && zdev->mio_capable;
zdev              217 arch/s390/include/asm/pci.h static inline int zpci_init_slot(struct zpci_dev *zdev)
zdev              221 arch/s390/include/asm/pci.h static inline void zpci_exit_slot(struct zpci_dev *zdev) {}
zdev               57 arch/s390/net/pnet.c 		struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev               59 arch/s390/net/pnet.c 		memcpy(pnetids, zdev->util_str, sizeof(zdev->util_str));
zdev               60 arch/s390/net/pnet.c 		EBCASC(pnetids, sizeof(zdev->util_str));
zdev               60 arch/s390/pci/pci.c 	struct zpci_dev *tmp, *zdev = NULL;
zdev               65 arch/s390/pci/pci.c 			zdev = tmp;
zdev               70 arch/s390/pci/pci.c 	return zdev;
zdev               75 arch/s390/pci/pci.c 	struct zpci_dev *tmp, *zdev;
zdev               80 arch/s390/pci/pci.c 	list_for_each_entry_safe(zdev, tmp, &zpci_list, entry) {
zdev               81 arch/s390/pci/pci.c 		if (zdev->state == ZPCI_FN_STATE_STANDBY &&
zdev               82 arch/s390/pci/pci.c 		    !clp_get_state(zdev->fid, &state) &&
zdev               84 arch/s390/pci/pci.c 			list_move_tail(&zdev->entry, &remove);
zdev               88 arch/s390/pci/pci.c 	list_for_each_entry_safe(zdev, tmp, &remove, entry)
zdev               89 arch/s390/pci/pci.c 		zpci_remove_device(zdev);
zdev              110 arch/s390/pci/pci.c int zpci_register_ioat(struct zpci_dev *zdev, u8 dmaas,
zdev              113 arch/s390/pci/pci.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, dmaas, ZPCI_MOD_FC_REG_IOAT);
zdev              125 arch/s390/pci/pci.c int zpci_unregister_ioat(struct zpci_dev *zdev, u8 dmaas)
zdev              127 arch/s390/pci/pci.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, dmaas, ZPCI_MOD_FC_DEREG_IOAT);
zdev              138 arch/s390/pci/pci.c int zpci_fmb_enable_device(struct zpci_dev *zdev)
zdev              140 arch/s390/pci/pci.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_SET_MEASURE);
zdev              144 arch/s390/pci/pci.c 	if (zdev->fmb || sizeof(*zdev->fmb) < zdev->fmb_length)
zdev              147 arch/s390/pci/pci.c 	zdev->fmb = kmem_cache_zalloc(zdev_fmb_cache, GFP_KERNEL);
zdev              148 arch/s390/pci/pci.c 	if (!zdev->fmb)
zdev              150 arch/s390/pci/pci.c 	WARN_ON((u64) zdev->fmb & 0xf);
zdev              153 arch/s390/pci/pci.c 	atomic64_set(&zdev->allocated_pages, 0);
zdev              154 arch/s390/pci/pci.c 	atomic64_set(&zdev->mapped_pages, 0);
zdev              155 arch/s390/pci/pci.c 	atomic64_set(&zdev->unmapped_pages, 0);
zdev              157 arch/s390/pci/pci.c 	fib.fmb_addr = virt_to_phys(zdev->fmb);
zdev              160 arch/s390/pci/pci.c 		kmem_cache_free(zdev_fmb_cache, zdev->fmb);
zdev              161 arch/s390/pci/pci.c 		zdev->fmb = NULL;
zdev              167 arch/s390/pci/pci.c int zpci_fmb_disable_device(struct zpci_dev *zdev)
zdev              169 arch/s390/pci/pci.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_SET_MEASURE);
zdev              173 arch/s390/pci/pci.c 	if (!zdev->fmb)
zdev              182 arch/s390/pci/pci.c 		kmem_cache_free(zdev_fmb_cache, zdev->fmb);
zdev              183 arch/s390/pci/pci.c 		zdev->fmb = NULL;
zdev              188 arch/s390/pci/pci.c static int zpci_cfg_load(struct zpci_dev *zdev, int offset, u32 *val, u8 len)
zdev              190 arch/s390/pci/pci.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len);
zdev              204 arch/s390/pci/pci.c static int zpci_cfg_store(struct zpci_dev *zdev, int offset, u32 val, u8 len)
zdev              206 arch/s390/pci/pci.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len);
zdev              268 arch/s390/pci/pci.c 	struct zpci_dev *zdev =	to_zpci(pdev);
zdev              271 arch/s390/pci/pci.c 	idx = zdev->bars[bar].map_idx;
zdev              275 arch/s390/pci/pci.c 	zpci_iomap_start[idx].fh = zdev->fh;
zdev              287 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              290 arch/s390/pci/pci.c 	iova = ioremap((unsigned long) zdev->bars[bar].mio_wt, barsize);
zdev              317 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              320 arch/s390/pci/pci.c 	iova = ioremap((unsigned long) zdev->bars[bar].mio_wb, barsize);
zdev              374 arch/s390/pci/pci.c 	struct zpci_dev *zdev = get_zdev_by_bus(bus);
zdev              377 arch/s390/pci/pci.c 	if (!zdev || devfn != ZPCI_DEVFN)
zdev              380 arch/s390/pci/pci.c 		ret = zpci_cfg_load(zdev, where, val, size);
zdev              388 arch/s390/pci/pci.c 	struct zpci_dev *zdev = get_zdev_by_bus(bus);
zdev              391 arch/s390/pci/pci.c 	if (!zdev || devfn != ZPCI_DEVFN)
zdev              394 arch/s390/pci/pci.c 		ret = zpci_cfg_store(zdev, where, val, size);
zdev              415 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              424 arch/s390/pci/pci.c 		if (zpci_use_mio(zdev))
zdev              426 arch/s390/pci/pci.c 				(resource_size_t __force) zdev->bars[i].mio_wt;
zdev              447 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              451 arch/s390/pci/pci.c 	if (zpci_use_mio(zdev))
zdev              463 arch/s390/pci/pci.c static int zpci_alloc_iomap(struct zpci_dev *zdev)
zdev              478 arch/s390/pci/pci.c static void zpci_free_iomap(struct zpci_dev *zdev, int entry)
zdev              486 arch/s390/pci/pci.c static struct resource *__alloc_res(struct zpci_dev *zdev, unsigned long start,
zdev              498 arch/s390/pci/pci.c 	r->name = zdev->res_name;
zdev              507 arch/s390/pci/pci.c static int zpci_setup_bus_resources(struct zpci_dev *zdev,
zdev              514 arch/s390/pci/pci.c 	snprintf(zdev->res_name, sizeof(zdev->res_name),
zdev              515 arch/s390/pci/pci.c 		 "PCI Bus %04x:%02x", zdev->domain, ZPCI_BUS_NR);
zdev              518 arch/s390/pci/pci.c 		if (!zdev->bars[i].size)
zdev              520 arch/s390/pci/pci.c 		entry = zpci_alloc_iomap(zdev);
zdev              523 arch/s390/pci/pci.c 		zdev->bars[i].map_idx = entry;
zdev              527 arch/s390/pci/pci.c 		if (zdev->bars[i].val & 8)
zdev              529 arch/s390/pci/pci.c 		if (zdev->bars[i].val & 4)
zdev              532 arch/s390/pci/pci.c 		if (zpci_use_mio(zdev))
zdev              533 arch/s390/pci/pci.c 			addr = (unsigned long) zdev->bars[i].mio_wt;
zdev              536 arch/s390/pci/pci.c 		size = 1UL << zdev->bars[i].size;
zdev              538 arch/s390/pci/pci.c 		res = __alloc_res(zdev, addr, size, flags);
zdev              540 arch/s390/pci/pci.c 			zpci_free_iomap(zdev, entry);
zdev              543 arch/s390/pci/pci.c 		zdev->bars[i].res = res;
zdev              550 arch/s390/pci/pci.c static void zpci_cleanup_bus_resources(struct zpci_dev *zdev)
zdev              555 arch/s390/pci/pci.c 		if (!zdev->bars[i].size || !zdev->bars[i].res)
zdev              558 arch/s390/pci/pci.c 		zpci_free_iomap(zdev, zdev->bars[i].map_idx);
zdev              559 arch/s390/pci/pci.c 		release_resource(zdev->bars[i].res);
zdev              560 arch/s390/pci/pci.c 		kfree(zdev->bars[i].res);
zdev              593 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              595 arch/s390/pci/pci.c 	zpci_debug_init_device(zdev, dev_name(&pdev->dev));
zdev              596 arch/s390/pci/pci.c 	zpci_fmb_enable_device(zdev);
zdev              603 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              605 arch/s390/pci/pci.c 	zpci_fmb_disable_device(zdev);
zdev              606 arch/s390/pci/pci.c 	zpci_debug_exit_device(zdev);
zdev              613 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              616 arch/s390/pci/pci.c 	if (zdev->state != ZPCI_FN_STATE_ONLINE)
zdev              619 arch/s390/pci/pci.c 	ret = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES);
zdev              624 arch/s390/pci/pci.c 	zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma,
zdev              625 arch/s390/pci/pci.c 			   (u64) zdev->dma_table);
zdev              634 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              636 arch/s390/pci/pci.c 	if (zdev->state != ZPCI_FN_STATE_ONLINE)
zdev              639 arch/s390/pci/pci.c 	zpci_unregister_ioat(zdev, 0);
zdev              641 arch/s390/pci/pci.c 	return clp_disable_fh(zdev);
zdev              652 arch/s390/pci/pci.c static int zpci_alloc_domain(struct zpci_dev *zdev)
zdev              655 arch/s390/pci/pci.c 		zdev->domain = (u16) zdev->uid;
zdev              656 arch/s390/pci/pci.c 		if (zdev->domain >= ZPCI_NR_DEVICES)
zdev              660 arch/s390/pci/pci.c 		if (test_bit(zdev->domain, zpci_domain)) {
zdev              664 arch/s390/pci/pci.c 		set_bit(zdev->domain, zpci_domain);
zdev              670 arch/s390/pci/pci.c 	zdev->domain = find_first_zero_bit(zpci_domain, ZPCI_NR_DEVICES);
zdev              671 arch/s390/pci/pci.c 	if (zdev->domain == ZPCI_NR_DEVICES) {
zdev              675 arch/s390/pci/pci.c 	set_bit(zdev->domain, zpci_domain);
zdev              680 arch/s390/pci/pci.c static void zpci_free_domain(struct zpci_dev *zdev)
zdev              682 arch/s390/pci/pci.c 	if (zdev->domain >= ZPCI_NR_DEVICES)
zdev              686 arch/s390/pci/pci.c 	clear_bit(zdev->domain, zpci_domain);
zdev              692 arch/s390/pci/pci.c 	struct zpci_dev *zdev = get_zdev_by_bus(bus);
zdev              694 arch/s390/pci/pci.c 	zpci_exit_slot(zdev);
zdev              695 arch/s390/pci/pci.c 	zpci_cleanup_bus_resources(zdev);
zdev              696 arch/s390/pci/pci.c 	zpci_destroy_iommu(zdev);
zdev              697 arch/s390/pci/pci.c 	zpci_free_domain(zdev);
zdev              700 arch/s390/pci/pci.c 	list_del(&zdev->entry);
zdev              703 arch/s390/pci/pci.c 	zpci_dbg(3, "rem fid:%x\n", zdev->fid);
zdev              704 arch/s390/pci/pci.c 	kfree(zdev);
zdev              707 arch/s390/pci/pci.c static int zpci_scan_bus(struct zpci_dev *zdev)
zdev              712 arch/s390/pci/pci.c 	ret = zpci_setup_bus_resources(zdev, &resources);
zdev              716 arch/s390/pci/pci.c 	zdev->bus = pci_scan_root_bus(NULL, ZPCI_BUS_NR, &pci_root_ops,
zdev              717 arch/s390/pci/pci.c 				      zdev, &resources);
zdev              718 arch/s390/pci/pci.c 	if (!zdev->bus) {
zdev              722 arch/s390/pci/pci.c 	zdev->bus->max_bus_speed = zdev->max_bus_speed;
zdev              723 arch/s390/pci/pci.c 	pci_bus_add_devices(zdev->bus);
zdev              727 arch/s390/pci/pci.c 	zpci_cleanup_bus_resources(zdev);
zdev              732 arch/s390/pci/pci.c int zpci_enable_device(struct zpci_dev *zdev)
zdev              736 arch/s390/pci/pci.c 	rc = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES);
zdev              740 arch/s390/pci/pci.c 	rc = zpci_dma_init_device(zdev);
zdev              744 arch/s390/pci/pci.c 	zdev->state = ZPCI_FN_STATE_ONLINE;
zdev              748 arch/s390/pci/pci.c 	clp_disable_fh(zdev);
zdev              754 arch/s390/pci/pci.c int zpci_disable_device(struct zpci_dev *zdev)
zdev              756 arch/s390/pci/pci.c 	zpci_dma_exit_device(zdev);
zdev              757 arch/s390/pci/pci.c 	return clp_disable_fh(zdev);
zdev              761 arch/s390/pci/pci.c int zpci_create_device(struct zpci_dev *zdev)
zdev              765 arch/s390/pci/pci.c 	rc = zpci_alloc_domain(zdev);
zdev              769 arch/s390/pci/pci.c 	rc = zpci_init_iommu(zdev);
zdev              773 arch/s390/pci/pci.c 	mutex_init(&zdev->lock);
zdev              774 arch/s390/pci/pci.c 	if (zdev->state == ZPCI_FN_STATE_CONFIGURED) {
zdev              775 arch/s390/pci/pci.c 		rc = zpci_enable_device(zdev);
zdev              779 arch/s390/pci/pci.c 	rc = zpci_scan_bus(zdev);
zdev              784 arch/s390/pci/pci.c 	list_add_tail(&zdev->entry, &zpci_list);
zdev              787 arch/s390/pci/pci.c 	zpci_init_slot(zdev);
zdev              792 arch/s390/pci/pci.c 	if (zdev->state == ZPCI_FN_STATE_ONLINE)
zdev              793 arch/s390/pci/pci.c 		zpci_disable_device(zdev);
zdev              795 arch/s390/pci/pci.c 	zpci_destroy_iommu(zdev);
zdev              797 arch/s390/pci/pci.c 	zpci_free_domain(zdev);
zdev              802 arch/s390/pci/pci.c void zpci_remove_device(struct zpci_dev *zdev)
zdev              804 arch/s390/pci/pci.c 	if (!zdev->bus)
zdev              807 arch/s390/pci/pci.c 	pci_stop_root_bus(zdev->bus);
zdev              808 arch/s390/pci/pci.c 	pci_remove_root_bus(zdev->bus);
zdev              814 arch/s390/pci/pci.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              816 arch/s390/pci/pci.c 	return sclp_pci_report(report, zdev->fh, zdev->fid);
zdev               97 arch/s390/pci/pci_clp.c static void clp_store_query_pci_fngrp(struct zpci_dev *zdev,
zdev              100 arch/s390/pci/pci_clp.c 	zdev->tlb_refresh = response->refresh;
zdev              101 arch/s390/pci/pci_clp.c 	zdev->dma_mask = response->dasm;
zdev              102 arch/s390/pci/pci_clp.c 	zdev->msi_addr = response->msia;
zdev              103 arch/s390/pci/pci_clp.c 	zdev->max_msi = response->noi;
zdev              104 arch/s390/pci/pci_clp.c 	zdev->fmb_update = response->mui;
zdev              108 arch/s390/pci/pci_clp.c 		zdev->max_bus_speed = PCIE_SPEED_5_0GT;
zdev              111 arch/s390/pci/pci_clp.c 		zdev->max_bus_speed = PCI_SPEED_UNKNOWN;
zdev              116 arch/s390/pci/pci_clp.c static int clp_query_pci_fngrp(struct zpci_dev *zdev, u8 pfgid)
zdev              133 arch/s390/pci/pci_clp.c 		clp_store_query_pci_fngrp(zdev, &rrb->response);
zdev              143 arch/s390/pci/pci_clp.c static int clp_store_query_pci_fn(struct zpci_dev *zdev,
zdev              149 arch/s390/pci/pci_clp.c 		zdev->bars[i].val = le32_to_cpu(response->bar[i]);
zdev              150 arch/s390/pci/pci_clp.c 		zdev->bars[i].size = response->bar_size[i];
zdev              152 arch/s390/pci/pci_clp.c 	zdev->start_dma = response->sdma;
zdev              153 arch/s390/pci/pci_clp.c 	zdev->end_dma = response->edma;
zdev              154 arch/s390/pci/pci_clp.c 	zdev->pchid = response->pchid;
zdev              155 arch/s390/pci/pci_clp.c 	zdev->pfgid = response->pfgid;
zdev              156 arch/s390/pci/pci_clp.c 	zdev->pft = response->pft;
zdev              157 arch/s390/pci/pci_clp.c 	zdev->vfn = response->vfn;
zdev              158 arch/s390/pci/pci_clp.c 	zdev->uid = response->uid;
zdev              159 arch/s390/pci/pci_clp.c 	zdev->fmb_length = sizeof(u32) * response->fmb_len;
zdev              161 arch/s390/pci/pci_clp.c 	memcpy(zdev->pfip, response->pfip, sizeof(zdev->pfip));
zdev              163 arch/s390/pci/pci_clp.c 		memcpy(zdev->util_str, response->util_str,
zdev              164 arch/s390/pci/pci_clp.c 		       sizeof(zdev->util_str));
zdev              166 arch/s390/pci/pci_clp.c 	zdev->mio_capable = response->mio_addr_avail;
zdev              171 arch/s390/pci/pci_clp.c 		zdev->bars[i].mio_wb = (void __iomem *) response->mio.addr[i].wb;
zdev              172 arch/s390/pci/pci_clp.c 		zdev->bars[i].mio_wt = (void __iomem *) response->mio.addr[i].wt;
zdev              177 arch/s390/pci/pci_clp.c static int clp_query_pci_fn(struct zpci_dev *zdev, u32 fh)
zdev              194 arch/s390/pci/pci_clp.c 		rc = clp_store_query_pci_fn(zdev, &rrb->response);
zdev              197 arch/s390/pci/pci_clp.c 		rc = clp_query_pci_fngrp(zdev, rrb->response.pfgid);
zdev              210 arch/s390/pci/pci_clp.c 	struct zpci_dev *zdev;
zdev              214 arch/s390/pci/pci_clp.c 	zdev = kzalloc(sizeof(*zdev), GFP_KERNEL);
zdev              215 arch/s390/pci/pci_clp.c 	if (!zdev)
zdev              218 arch/s390/pci/pci_clp.c 	zdev->fh = fh;
zdev              219 arch/s390/pci/pci_clp.c 	zdev->fid = fid;
zdev              222 arch/s390/pci/pci_clp.c 	rc = clp_query_pci_fn(zdev, fh);
zdev              227 arch/s390/pci/pci_clp.c 		zdev->state = ZPCI_FN_STATE_CONFIGURED;
zdev              229 arch/s390/pci/pci_clp.c 		zdev->state = ZPCI_FN_STATE_STANDBY;
zdev              231 arch/s390/pci/pci_clp.c 	rc = zpci_create_device(zdev);
zdev              238 arch/s390/pci/pci_clp.c 	kfree(zdev);
zdev              246 arch/s390/pci/pci_clp.c static int clp_set_pci_fn(struct zpci_dev *zdev, u8 nr_dma_as, u8 command)
zdev              250 arch/s390/pci/pci_clp.c 	u32 fid = zdev->fid;
zdev              261 arch/s390/pci/pci_clp.c 		rrb->request.fh = zdev->fh;
zdev              280 arch/s390/pci/pci_clp.c 		zdev->fh = rrb->response.fh;
zdev              290 arch/s390/pci/pci_clp.c int clp_enable_fh(struct zpci_dev *zdev, u8 nr_dma_as)
zdev              294 arch/s390/pci/pci_clp.c 	rc = clp_set_pci_fn(zdev, nr_dma_as, CLP_SET_ENABLE_PCI_FN);
zdev              295 arch/s390/pci/pci_clp.c 	zpci_dbg(3, "ena fid:%x, fh:%x, rc:%d\n", zdev->fid, zdev->fh, rc);
zdev              299 arch/s390/pci/pci_clp.c 	if (zpci_use_mio(zdev)) {
zdev              300 arch/s390/pci/pci_clp.c 		rc = clp_set_pci_fn(zdev, nr_dma_as, CLP_SET_ENABLE_MIO);
zdev              302 arch/s390/pci/pci_clp.c 				zdev->fid, zdev->fh, rc);
zdev              304 arch/s390/pci/pci_clp.c 			clp_disable_fh(zdev);
zdev              310 arch/s390/pci/pci_clp.c int clp_disable_fh(struct zpci_dev *zdev)
zdev              314 arch/s390/pci/pci_clp.c 	if (!zdev_enabled(zdev))
zdev              317 arch/s390/pci/pci_clp.c 	rc = clp_set_pci_fn(zdev, 0, CLP_SET_DISABLE_PCI_FN);
zdev              318 arch/s390/pci/pci_clp.c 	zpci_dbg(3, "dis fid:%x, fh:%x, rc:%d\n", zdev->fid, zdev->fh, rc);
zdev              362 arch/s390/pci/pci_clp.c 	struct zpci_dev *zdev;
zdev              367 arch/s390/pci/pci_clp.c 	zdev = get_zdev_by_fid(entry->fid);
zdev              368 arch/s390/pci/pci_clp.c 	if (!zdev)
zdev              374 arch/s390/pci/pci_clp.c 	struct zpci_dev *zdev;
zdev              383 arch/s390/pci/pci_clp.c 	zdev = get_zdev_by_fid(entry->fid);
zdev              384 arch/s390/pci/pci_clp.c 	if (!zdev)
zdev              387 arch/s390/pci/pci_clp.c 	zdev->fh = entry->fh;
zdev               72 arch/s390/pci/pci_debug.c 	struct zpci_dev *zdev = m->private;
zdev               73 arch/s390/pci/pci_debug.c 	atomic64_t *counter = &zdev->allocated_pages;
zdev               83 arch/s390/pci/pci_debug.c 	struct zpci_dev *zdev = m->private;
zdev               85 arch/s390/pci/pci_debug.c 	if (!zdev)
zdev               88 arch/s390/pci/pci_debug.c 	mutex_lock(&zdev->lock);
zdev               89 arch/s390/pci/pci_debug.c 	if (!zdev->fmb) {
zdev               90 arch/s390/pci/pci_debug.c 		mutex_unlock(&zdev->lock);
zdev               96 arch/s390/pci/pci_debug.c 	seq_printf(m, "Update interval: %u ms\n", zdev->fmb_update);
zdev               97 arch/s390/pci/pci_debug.c 	seq_printf(m, "Samples: %u\n", zdev->fmb->samples);
zdev               98 arch/s390/pci/pci_debug.c 	seq_printf(m, "Last update TOD: %Lx\n", zdev->fmb->last_update);
zdev              101 arch/s390/pci/pci_debug.c 		     &zdev->fmb->ld_ops);
zdev              103 arch/s390/pci/pci_debug.c 	switch (zdev->fmb->format) {
zdev              105 arch/s390/pci/pci_debug.c 		if (!(zdev->fmb->fmt_ind & ZPCI_FMB_DMA_COUNTER_VALID))
zdev              108 arch/s390/pci/pci_debug.c 			     &zdev->fmb->fmt0.dma_rbytes);
zdev              112 arch/s390/pci/pci_debug.c 			     &zdev->fmb->fmt1.rx_bytes);
zdev              116 arch/s390/pci/pci_debug.c 			     &zdev->fmb->fmt2.consumed_work_units);
zdev              120 arch/s390/pci/pci_debug.c 			     &zdev->fmb->fmt3.tx_bytes);
zdev              127 arch/s390/pci/pci_debug.c 	mutex_unlock(&zdev->lock);
zdev              134 arch/s390/pci/pci_debug.c 	struct zpci_dev *zdev = ((struct seq_file *) file->private_data)->private;
zdev              138 arch/s390/pci/pci_debug.c 	if (!zdev)
zdev              145 arch/s390/pci/pci_debug.c 	mutex_lock(&zdev->lock);
zdev              148 arch/s390/pci/pci_debug.c 		rc = zpci_fmb_disable_device(zdev);
zdev              151 arch/s390/pci/pci_debug.c 		rc = zpci_fmb_enable_device(zdev);
zdev              154 arch/s390/pci/pci_debug.c 	mutex_unlock(&zdev->lock);
zdev              172 arch/s390/pci/pci_debug.c void zpci_debug_init_device(struct zpci_dev *zdev, const char *name)
zdev              174 arch/s390/pci/pci_debug.c 	zdev->debugfs_dev = debugfs_create_dir(name, debugfs_root);
zdev              177 arch/s390/pci/pci_debug.c 			    zdev->debugfs_dev, zdev, &debugfs_pci_perf_fops);
zdev              180 arch/s390/pci/pci_debug.c void zpci_debug_exit_device(struct zpci_dev *zdev)
zdev              182 arch/s390/pci/pci_debug.c 	debugfs_remove_recursive(zdev->debugfs_dev);
zdev               22 arch/s390/pci/pci_dma.c static int zpci_refresh_global(struct zpci_dev *zdev)
zdev               24 arch/s390/pci/pci_dma.c 	return zpci_refresh_trans((u64) zdev->fh << 32, zdev->start_dma,
zdev               25 arch/s390/pci/pci_dma.c 				  zdev->iommu_pages * PAGE_SIZE);
zdev              133 arch/s390/pci/pci_dma.c static int __dma_update_trans(struct zpci_dev *zdev, unsigned long pa,
zdev              145 arch/s390/pci/pci_dma.c 	spin_lock_irqsave(&zdev->dma_table_lock, irq_flags);
zdev              146 arch/s390/pci/pci_dma.c 	if (!zdev->dma_table) {
zdev              152 arch/s390/pci/pci_dma.c 		entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr);
zdev              168 arch/s390/pci/pci_dma.c 			entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr);
zdev              175 arch/s390/pci/pci_dma.c 	spin_unlock_irqrestore(&zdev->dma_table_lock, irq_flags);
zdev              179 arch/s390/pci/pci_dma.c static int __dma_purge_tlb(struct zpci_dev *zdev, dma_addr_t dma_addr,
zdev              193 arch/s390/pci/pci_dma.c 		if (!zdev->tlb_refresh)
zdev              200 arch/s390/pci/pci_dma.c 	ret = zpci_refresh_trans((u64) zdev->fh << 32, dma_addr,
zdev              204 arch/s390/pci/pci_dma.c 		if (zpci_refresh_global(zdev))
zdev              207 arch/s390/pci/pci_dma.c 		spin_lock_irqsave(&zdev->iommu_bitmap_lock, irqflags);
zdev              208 arch/s390/pci/pci_dma.c 		bitmap_andnot(zdev->iommu_bitmap, zdev->iommu_bitmap,
zdev              209 arch/s390/pci/pci_dma.c 			      zdev->lazy_bitmap, zdev->iommu_pages);
zdev              210 arch/s390/pci/pci_dma.c 		bitmap_zero(zdev->lazy_bitmap, zdev->iommu_pages);
zdev              211 arch/s390/pci/pci_dma.c 		spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, irqflags);
zdev              218 arch/s390/pci/pci_dma.c static int dma_update_trans(struct zpci_dev *zdev, unsigned long pa,
zdev              223 arch/s390/pci/pci_dma.c 	rc = __dma_update_trans(zdev, pa, dma_addr, size, flags);
zdev              227 arch/s390/pci/pci_dma.c 	rc = __dma_purge_tlb(zdev, dma_addr, size, flags);
zdev              229 arch/s390/pci/pci_dma.c 		__dma_update_trans(zdev, pa, dma_addr, size, ZPCI_PTE_INVALID);
zdev              263 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              268 arch/s390/pci/pci_dma.c 	return iommu_area_alloc(zdev->iommu_bitmap, zdev->iommu_pages,
zdev              269 arch/s390/pci/pci_dma.c 				start, size, zdev->start_dma >> PAGE_SHIFT,
zdev              275 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              278 arch/s390/pci/pci_dma.c 	spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags);
zdev              279 arch/s390/pci/pci_dma.c 	offset = __dma_alloc_iommu(dev, zdev->next_bit, size);
zdev              283 arch/s390/pci/pci_dma.c 			if (zpci_refresh_global(zdev))
zdev              286 arch/s390/pci/pci_dma.c 			bitmap_andnot(zdev->iommu_bitmap, zdev->iommu_bitmap,
zdev              287 arch/s390/pci/pci_dma.c 				      zdev->lazy_bitmap, zdev->iommu_pages);
zdev              288 arch/s390/pci/pci_dma.c 			bitmap_zero(zdev->lazy_bitmap, zdev->iommu_pages);
zdev              295 arch/s390/pci/pci_dma.c 	zdev->next_bit = offset + size;
zdev              296 arch/s390/pci/pci_dma.c 	spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, flags);
zdev              298 arch/s390/pci/pci_dma.c 	return zdev->start_dma + offset * PAGE_SIZE;
zdev              301 arch/s390/pci/pci_dma.c 	spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, flags);
zdev              307 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              310 arch/s390/pci/pci_dma.c 	offset = (dma_addr - zdev->start_dma) >> PAGE_SHIFT;
zdev              312 arch/s390/pci/pci_dma.c 	spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags);
zdev              313 arch/s390/pci/pci_dma.c 	if (!zdev->iommu_bitmap)
zdev              317 arch/s390/pci/pci_dma.c 		bitmap_clear(zdev->iommu_bitmap, offset, size);
zdev              319 arch/s390/pci/pci_dma.c 		bitmap_set(zdev->lazy_bitmap, offset, size);
zdev              322 arch/s390/pci/pci_dma.c 	spin_unlock_irqrestore(&zdev->iommu_bitmap_lock, flags);
zdev              340 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              361 arch/s390/pci/pci_dma.c 	ret = dma_update_trans(zdev, pa, dma_addr, size, flags);
zdev              365 arch/s390/pci/pci_dma.c 	atomic64_add(nr_pages, &zdev->mapped_pages);
zdev              380 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              385 arch/s390/pci/pci_dma.c 	ret = dma_update_trans(zdev, 0, dma_addr, npages * PAGE_SIZE,
zdev              393 arch/s390/pci/pci_dma.c 	atomic64_add(npages, &zdev->unmapped_pages);
zdev              401 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              418 arch/s390/pci/pci_dma.c 	atomic64_add(size / PAGE_SIZE, &zdev->allocated_pages);
zdev              428 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              431 arch/s390/pci/pci_dma.c 	atomic64_sub(size / PAGE_SIZE, &zdev->allocated_pages);
zdev              442 arch/s390/pci/pci_dma.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev              459 arch/s390/pci/pci_dma.c 		ret = __dma_update_trans(zdev, pa, dma_addr,
zdev              466 arch/s390/pci/pci_dma.c 	ret = __dma_purge_tlb(zdev, dma_addr_base, size, flags);
zdev              471 arch/s390/pci/pci_dma.c 	atomic64_add(nr_pages, &zdev->mapped_pages);
zdev              476 arch/s390/pci/pci_dma.c 	dma_update_trans(zdev, 0, dma_addr_base, dma_addr - dma_addr_base,
zdev              547 arch/s390/pci/pci_dma.c int zpci_dma_init_device(struct zpci_dev *zdev)
zdev              556 arch/s390/pci/pci_dma.c 	WARN_ON(zdev->s390_domain);
zdev              558 arch/s390/pci/pci_dma.c 	spin_lock_init(&zdev->iommu_bitmap_lock);
zdev              559 arch/s390/pci/pci_dma.c 	spin_lock_init(&zdev->dma_table_lock);
zdev              561 arch/s390/pci/pci_dma.c 	zdev->dma_table = dma_alloc_cpu_table();
zdev              562 arch/s390/pci/pci_dma.c 	if (!zdev->dma_table) {
zdev              576 arch/s390/pci/pci_dma.c 	zdev->start_dma = PAGE_ALIGN(zdev->start_dma);
zdev              577 arch/s390/pci/pci_dma.c 	zdev->iommu_size = min3((u64) high_memory,
zdev              578 arch/s390/pci/pci_dma.c 				ZPCI_TABLE_SIZE_RT - zdev->start_dma,
zdev              579 arch/s390/pci/pci_dma.c 				zdev->end_dma - zdev->start_dma + 1);
zdev              580 arch/s390/pci/pci_dma.c 	zdev->end_dma = zdev->start_dma + zdev->iommu_size - 1;
zdev              581 arch/s390/pci/pci_dma.c 	zdev->iommu_pages = zdev->iommu_size >> PAGE_SHIFT;
zdev              582 arch/s390/pci/pci_dma.c 	zdev->iommu_bitmap = vzalloc(zdev->iommu_pages / 8);
zdev              583 arch/s390/pci/pci_dma.c 	if (!zdev->iommu_bitmap) {
zdev              588 arch/s390/pci/pci_dma.c 		zdev->lazy_bitmap = vzalloc(zdev->iommu_pages / 8);
zdev              589 arch/s390/pci/pci_dma.c 		if (!zdev->lazy_bitmap) {
zdev              595 arch/s390/pci/pci_dma.c 	rc = zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma,
zdev              596 arch/s390/pci/pci_dma.c 				(u64) zdev->dma_table);
zdev              602 arch/s390/pci/pci_dma.c 	vfree(zdev->iommu_bitmap);
zdev              603 arch/s390/pci/pci_dma.c 	zdev->iommu_bitmap = NULL;
zdev              604 arch/s390/pci/pci_dma.c 	vfree(zdev->lazy_bitmap);
zdev              605 arch/s390/pci/pci_dma.c 	zdev->lazy_bitmap = NULL;
zdev              607 arch/s390/pci/pci_dma.c 	dma_free_cpu_table(zdev->dma_table);
zdev              608 arch/s390/pci/pci_dma.c 	zdev->dma_table = NULL;
zdev              613 arch/s390/pci/pci_dma.c void zpci_dma_exit_device(struct zpci_dev *zdev)
zdev              620 arch/s390/pci/pci_dma.c 	WARN_ON(zdev->s390_domain);
zdev              622 arch/s390/pci/pci_dma.c 	if (zpci_unregister_ioat(zdev, 0))
zdev              625 arch/s390/pci/pci_dma.c 	dma_cleanup_tables(zdev->dma_table);
zdev              626 arch/s390/pci/pci_dma.c 	zdev->dma_table = NULL;
zdev              627 arch/s390/pci/pci_dma.c 	vfree(zdev->iommu_bitmap);
zdev              628 arch/s390/pci/pci_dma.c 	zdev->iommu_bitmap = NULL;
zdev              629 arch/s390/pci/pci_dma.c 	vfree(zdev->lazy_bitmap);
zdev              630 arch/s390/pci/pci_dma.c 	zdev->lazy_bitmap = NULL;
zdev              632 arch/s390/pci/pci_dma.c 	zdev->next_bit = 0;
zdev               49 arch/s390/pci/pci_event.c 	struct zpci_dev *zdev = get_zdev_by_fid(ccdf->fid);
zdev               55 arch/s390/pci/pci_event.c 	if (zdev)
zdev               56 arch/s390/pci/pci_event.c 		pdev = pci_get_slot(zdev->bus, ZPCI_DEVFN);
zdev               76 arch/s390/pci/pci_event.c 	struct zpci_dev *zdev = get_zdev_by_fid(ccdf->fid);
zdev               81 arch/s390/pci/pci_event.c 	if (zdev)
zdev               82 arch/s390/pci/pci_event.c 		pdev = pci_get_slot(zdev->bus, ZPCI_DEVFN);
zdev               91 arch/s390/pci/pci_event.c 		if (!zdev) {
zdev               95 arch/s390/pci/pci_event.c 			zdev = get_zdev_by_fid(ccdf->fid);
zdev               97 arch/s390/pci/pci_event.c 		if (!zdev || zdev->state != ZPCI_FN_STATE_STANDBY)
zdev               99 arch/s390/pci/pci_event.c 		zdev->state = ZPCI_FN_STATE_CONFIGURED;
zdev              100 arch/s390/pci/pci_event.c 		zdev->fh = ccdf->fh;
zdev              101 arch/s390/pci/pci_event.c 		ret = zpci_enable_device(zdev);
zdev              105 arch/s390/pci/pci_event.c 		pci_rescan_bus(zdev->bus);
zdev              109 arch/s390/pci/pci_event.c 		if (!zdev)
zdev              113 arch/s390/pci/pci_event.c 		if (!zdev)
zdev              118 arch/s390/pci/pci_event.c 		ret = zpci_disable_device(zdev);
zdev              122 arch/s390/pci/pci_event.c 		ret = sclp_pci_deconfigure(zdev->fid);
zdev              123 arch/s390/pci/pci_event.c 		zpci_dbg(3, "deconf fid:%x, rc:%d\n", zdev->fid, ret);
zdev              125 arch/s390/pci/pci_event.c 			zdev->state = ZPCI_FN_STATE_STANDBY;
zdev              129 arch/s390/pci/pci_event.c 		if (!zdev)
zdev              138 arch/s390/pci/pci_event.c 		zdev->fh = ccdf->fh;
zdev              139 arch/s390/pci/pci_event.c 		zpci_disable_device(zdev);
zdev              140 arch/s390/pci/pci_event.c 		zdev->state = ZPCI_FN_STATE_STANDBY;
zdev              143 arch/s390/pci/pci_event.c 			zpci_remove_device(zdev);
zdev              150 arch/s390/pci/pci_event.c 		if (!zdev)
zdev              152 arch/s390/pci/pci_event.c 		zpci_remove_device(zdev);
zdev               39 arch/s390/pci/pci_irq.c static int zpci_set_airq(struct zpci_dev *zdev)
zdev               41 arch/s390/pci/pci_irq.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_REG_INT);
zdev               47 arch/s390/pci/pci_irq.c 	fib.fmt0.noi = airq_iv_end(zdev->aibv);
zdev               48 arch/s390/pci/pci_irq.c 	fib.fmt0.aibv = (unsigned long) zdev->aibv->vector;
zdev               50 arch/s390/pci/pci_irq.c 	fib.fmt0.aisb = (unsigned long) zpci_sbv->vector + (zdev->aisb/64)*8;
zdev               51 arch/s390/pci/pci_irq.c 	fib.fmt0.aisbo = zdev->aisb & 63;
zdev               57 arch/s390/pci/pci_irq.c static int zpci_clear_airq(struct zpci_dev *zdev)
zdev               59 arch/s390/pci/pci_irq.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_DEREG_INT);
zdev               72 arch/s390/pci/pci_irq.c static int zpci_set_directed_irq(struct zpci_dev *zdev)
zdev               74 arch/s390/pci/pci_irq.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_REG_INT_D);
zdev               79 arch/s390/pci/pci_irq.c 	fib.fmt1.noi = zdev->msi_nr_irqs;
zdev               80 arch/s390/pci/pci_irq.c 	fib.fmt1.dibvo = zdev->msi_first_bit;
zdev               86 arch/s390/pci/pci_irq.c static int zpci_clear_directed_irq(struct zpci_dev *zdev)
zdev               88 arch/s390/pci/pci_irq.c 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_DEREG_INT_D);
zdev              236 arch/s390/pci/pci_irq.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              243 arch/s390/pci/pci_irq.c 	zdev->aisb = -1UL;
zdev              244 arch/s390/pci/pci_irq.c 	zdev->msi_first_bit = -1U;
zdev              247 arch/s390/pci/pci_irq.c 	msi_vecs = min_t(unsigned int, nvec, zdev->max_msi);
zdev              259 arch/s390/pci/pci_irq.c 		zdev->aisb = bit;
zdev              262 arch/s390/pci/pci_irq.c 		zdev->aibv = airq_iv_create(msi_vecs, AIRQ_IV_DATA | AIRQ_IV_BITLOCK);
zdev              263 arch/s390/pci/pci_irq.c 		if (!zdev->aibv)
zdev              267 arch/s390/pci/pci_irq.c 		zpci_ibv[bit] = zdev->aibv;
zdev              290 arch/s390/pci/pci_irq.c 			msg.address_lo = zdev->msi_addr & 0xff0000ff;
zdev              297 arch/s390/pci/pci_irq.c 			msg.address_lo = zdev->msi_addr & 0xffffffff;
zdev              298 arch/s390/pci/pci_irq.c 			airq_iv_set_data(zdev->aibv, hwirq, irq);
zdev              300 arch/s390/pci/pci_irq.c 		msg.address_hi = zdev->msi_addr >> 32;
zdev              305 arch/s390/pci/pci_irq.c 	zdev->msi_first_bit = bit;
zdev              306 arch/s390/pci/pci_irq.c 	zdev->msi_nr_irqs = msi_vecs;
zdev              309 arch/s390/pci/pci_irq.c 		rc = zpci_set_directed_irq(zdev);
zdev              311 arch/s390/pci/pci_irq.c 		rc = zpci_set_airq(zdev);
zdev              320 arch/s390/pci/pci_irq.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              326 arch/s390/pci/pci_irq.c 		rc = zpci_clear_directed_irq(zdev);
zdev              328 arch/s390/pci/pci_irq.c 		rc = zpci_clear_airq(zdev);
zdev              348 arch/s390/pci/pci_irq.c 	if (zdev->aisb != -1UL) {
zdev              349 arch/s390/pci/pci_irq.c 		zpci_ibv[zdev->aisb] = NULL;
zdev              350 arch/s390/pci/pci_irq.c 		airq_iv_free_bit(zpci_sbv, zdev->aisb);
zdev              351 arch/s390/pci/pci_irq.c 		zdev->aisb = -1UL;
zdev              353 arch/s390/pci/pci_irq.c 	if (zdev->aibv) {
zdev              354 arch/s390/pci/pci_irq.c 		airq_iv_release(zdev->aibv);
zdev              355 arch/s390/pci/pci_irq.c 		zdev->aibv = NULL;
zdev              358 arch/s390/pci/pci_irq.c 	if ((irq_delivery == DIRECTED) && zdev->msi_first_bit != -1U)
zdev              359 arch/s390/pci/pci_irq.c 		airq_iv_free(zpci_ibv[0], zdev->msi_first_bit, zdev->msi_nr_irqs);
zdev               24 arch/s390/pci/pci_sysfs.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));		\
zdev               26 arch/s390/pci/pci_sysfs.c 	return sprintf(buf, fmt, zdev->member);				\
zdev               45 arch/s390/pci/pci_sysfs.c 	struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
zdev               47 arch/s390/pci/pci_sysfs.c 	return sprintf(buf, zpci_use_mio(zdev) ? "1\n" : "0\n");
zdev               56 arch/s390/pci/pci_sysfs.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev               84 arch/s390/pci/pci_sysfs.c 		ret = zpci_disable_device(zdev);
zdev               88 arch/s390/pci/pci_sysfs.c 		ret = zpci_enable_device(zdev);
zdev               91 arch/s390/pci/pci_sysfs.c 		pci_rescan_bus(zdev->bus);
zdev              107 arch/s390/pci/pci_sysfs.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              109 arch/s390/pci/pci_sysfs.c 	return memory_read_from_buffer(buf, count, &off, zdev->util_str,
zdev              110 arch/s390/pci/pci_sysfs.c 				       sizeof(zdev->util_str));
zdev              121 arch/s390/pci/pci_sysfs.c 	struct zpci_dev *zdev = to_zpci(pdev);
zdev              127 arch/s390/pci/pci_sysfs.c 	ret = sclp_pci_report(report, zdev->fh, zdev->fid);
zdev              216 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_device *zdev;
zdev              858 drivers/dma/xilinx/zynqmp_dma.c 		devm_free_irq(chan->zdev->dev, chan->irq, chan);
zdev              870 drivers/dma/xilinx/zynqmp_dma.c static int zynqmp_dma_chan_probe(struct zynqmp_dma_device *zdev,
zdev              878 drivers/dma/xilinx/zynqmp_dma.c 	chan = devm_kzalloc(zdev->dev, sizeof(*chan), GFP_KERNEL);
zdev              881 drivers/dma/xilinx/zynqmp_dma.c 	chan->dev = zdev->dev;
zdev              882 drivers/dma/xilinx/zynqmp_dma.c 	chan->zdev = zdev;
zdev              900 drivers/dma/xilinx/zynqmp_dma.c 		dev_err(zdev->dev, "invalid bus-width value");
zdev              905 drivers/dma/xilinx/zynqmp_dma.c 	zdev->chan = chan;
zdev              914 drivers/dma/xilinx/zynqmp_dma.c 	chan->common.device = &zdev->common;
zdev              915 drivers/dma/xilinx/zynqmp_dma.c 	list_add_tail(&chan->common.device_node, &zdev->common.channels);
zdev              941 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_device *zdev = ofdma->of_dma_data;
zdev              943 drivers/dma/xilinx/zynqmp_dma.c 	return dma_get_slave_channel(&zdev->chan->common);
zdev              985 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_device *zdev = dev_get_drvdata(dev);
zdev              987 drivers/dma/xilinx/zynqmp_dma.c 	clk_disable_unprepare(zdev->clk_main);
zdev              988 drivers/dma/xilinx/zynqmp_dma.c 	clk_disable_unprepare(zdev->clk_apb);
zdev             1002 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_device *zdev = dev_get_drvdata(dev);
zdev             1005 drivers/dma/xilinx/zynqmp_dma.c 	err = clk_prepare_enable(zdev->clk_main);
zdev             1011 drivers/dma/xilinx/zynqmp_dma.c 	err = clk_prepare_enable(zdev->clk_apb);
zdev             1014 drivers/dma/xilinx/zynqmp_dma.c 		clk_disable_unprepare(zdev->clk_main);
zdev             1035 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_device *zdev;
zdev             1039 drivers/dma/xilinx/zynqmp_dma.c 	zdev = devm_kzalloc(&pdev->dev, sizeof(*zdev), GFP_KERNEL);
zdev             1040 drivers/dma/xilinx/zynqmp_dma.c 	if (!zdev)
zdev             1043 drivers/dma/xilinx/zynqmp_dma.c 	zdev->dev = &pdev->dev;
zdev             1044 drivers/dma/xilinx/zynqmp_dma.c 	INIT_LIST_HEAD(&zdev->common.channels);
zdev             1047 drivers/dma/xilinx/zynqmp_dma.c 	dma_cap_set(DMA_MEMCPY, zdev->common.cap_mask);
zdev             1049 drivers/dma/xilinx/zynqmp_dma.c 	p = &zdev->common;
zdev             1059 drivers/dma/xilinx/zynqmp_dma.c 	zdev->clk_main = devm_clk_get(&pdev->dev, "clk_main");
zdev             1060 drivers/dma/xilinx/zynqmp_dma.c 	if (IS_ERR(zdev->clk_main)) {
zdev             1062 drivers/dma/xilinx/zynqmp_dma.c 		return PTR_ERR(zdev->clk_main);
zdev             1065 drivers/dma/xilinx/zynqmp_dma.c 	zdev->clk_apb = devm_clk_get(&pdev->dev, "clk_apb");
zdev             1066 drivers/dma/xilinx/zynqmp_dma.c 	if (IS_ERR(zdev->clk_apb)) {
zdev             1068 drivers/dma/xilinx/zynqmp_dma.c 		return PTR_ERR(zdev->clk_apb);
zdev             1071 drivers/dma/xilinx/zynqmp_dma.c 	platform_set_drvdata(pdev, zdev);
zdev             1072 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_set_autosuspend_delay(zdev->dev, ZDMA_PM_TIMEOUT);
zdev             1073 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_use_autosuspend(zdev->dev);
zdev             1074 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_enable(zdev->dev);
zdev             1075 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_get_sync(zdev->dev);
zdev             1076 drivers/dma/xilinx/zynqmp_dma.c 	if (!pm_runtime_enabled(zdev->dev)) {
zdev             1077 drivers/dma/xilinx/zynqmp_dma.c 		ret = zynqmp_dma_runtime_resume(zdev->dev);
zdev             1082 drivers/dma/xilinx/zynqmp_dma.c 	ret = zynqmp_dma_chan_probe(zdev, pdev);
zdev             1088 drivers/dma/xilinx/zynqmp_dma.c 	p->dst_addr_widths = BIT(zdev->chan->bus_width / 8);
zdev             1089 drivers/dma/xilinx/zynqmp_dma.c 	p->src_addr_widths = BIT(zdev->chan->bus_width / 8);
zdev             1091 drivers/dma/xilinx/zynqmp_dma.c 	dma_async_device_register(&zdev->common);
zdev             1094 drivers/dma/xilinx/zynqmp_dma.c 					 of_zynqmp_dma_xlate, zdev);
zdev             1097 drivers/dma/xilinx/zynqmp_dma.c 		dma_async_device_unregister(&zdev->common);
zdev             1101 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_mark_last_busy(zdev->dev);
zdev             1102 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_put_sync_autosuspend(zdev->dev);
zdev             1109 drivers/dma/xilinx/zynqmp_dma.c 	zynqmp_dma_chan_remove(zdev->chan);
zdev             1111 drivers/dma/xilinx/zynqmp_dma.c 	if (!pm_runtime_enabled(zdev->dev))
zdev             1112 drivers/dma/xilinx/zynqmp_dma.c 		zynqmp_dma_runtime_suspend(zdev->dev);
zdev             1113 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_disable(zdev->dev);
zdev             1125 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_device *zdev = platform_get_drvdata(pdev);
zdev             1128 drivers/dma/xilinx/zynqmp_dma.c 	dma_async_device_unregister(&zdev->common);
zdev             1130 drivers/dma/xilinx/zynqmp_dma.c 	zynqmp_dma_chan_remove(zdev->chan);
zdev             1131 drivers/dma/xilinx/zynqmp_dma.c 	pm_runtime_disable(zdev->dev);
zdev             1132 drivers/dma/xilinx/zynqmp_dma.c 	if (!pm_runtime_enabled(zdev->dev))
zdev             1133 drivers/dma/xilinx/zynqmp_dma.c 		zynqmp_dma_runtime_suspend(zdev->dev);
zdev               34 drivers/iommu/s390-iommu.c 	struct zpci_dev		*zdev;
zdev               90 drivers/iommu/s390-iommu.c 	struct zpci_dev *zdev = to_pci_dev(dev)->sysdata;
zdev               95 drivers/iommu/s390-iommu.c 	if (!zdev)
zdev              102 drivers/iommu/s390-iommu.c 	if (zdev->dma_table)
zdev              103 drivers/iommu/s390-iommu.c 		zpci_dma_exit_device(zdev);
zdev              105 drivers/iommu/s390-iommu.c 	zdev->dma_table = s390_domain->dma_table;
zdev              106 drivers/iommu/s390-iommu.c 	rc = zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma,
zdev              107 drivers/iommu/s390-iommu.c 				(u64) zdev->dma_table);
zdev              114 drivers/iommu/s390-iommu.c 		domain->geometry.aperture_start = zdev->start_dma;
zdev              115 drivers/iommu/s390-iommu.c 		domain->geometry.aperture_end = zdev->end_dma;
zdev              118 drivers/iommu/s390-iommu.c 	} else if (domain->geometry.aperture_start != zdev->start_dma ||
zdev              119 drivers/iommu/s390-iommu.c 		   domain->geometry.aperture_end != zdev->end_dma) {
zdev              124 drivers/iommu/s390-iommu.c 	domain_device->zdev = zdev;
zdev              125 drivers/iommu/s390-iommu.c 	zdev->s390_domain = s390_domain;
zdev              132 drivers/iommu/s390-iommu.c 	zpci_dma_init_device(zdev);
zdev              142 drivers/iommu/s390-iommu.c 	struct zpci_dev *zdev = to_pci_dev(dev)->sysdata;
zdev              147 drivers/iommu/s390-iommu.c 	if (!zdev)
zdev              153 drivers/iommu/s390-iommu.c 		if (domain_device->zdev == zdev) {
zdev              163 drivers/iommu/s390-iommu.c 		zdev->s390_domain = NULL;
zdev              164 drivers/iommu/s390-iommu.c 		zpci_unregister_ioat(zdev, 0);
zdev              165 drivers/iommu/s390-iommu.c 		zpci_dma_init_device(zdev);
zdev              172 drivers/iommu/s390-iommu.c 	struct zpci_dev *zdev = to_pci_dev(dev)->sysdata;
zdev              178 drivers/iommu/s390-iommu.c 	iommu_device_link(&zdev->iommu_dev, dev);
zdev              185 drivers/iommu/s390-iommu.c 	struct zpci_dev *zdev = to_pci_dev(dev)->sysdata;
zdev              199 drivers/iommu/s390-iommu.c 	if (zdev && zdev->s390_domain) {
zdev              205 drivers/iommu/s390-iommu.c 	iommu_device_unlink(&zdev->iommu_dev, dev);
zdev              242 drivers/iommu/s390-iommu.c 		rc = zpci_refresh_trans((u64) domain_device->zdev->fh << 32,
zdev              337 drivers/iommu/s390-iommu.c int zpci_init_iommu(struct zpci_dev *zdev)
zdev              341 drivers/iommu/s390-iommu.c 	rc = iommu_device_sysfs_add(&zdev->iommu_dev, NULL, NULL,
zdev              342 drivers/iommu/s390-iommu.c 				    "s390-iommu.%08x", zdev->fid);
zdev              346 drivers/iommu/s390-iommu.c 	iommu_device_set_ops(&zdev->iommu_dev, &s390_iommu_ops);
zdev              348 drivers/iommu/s390-iommu.c 	rc = iommu_device_register(&zdev->iommu_dev);
zdev              355 drivers/iommu/s390-iommu.c 	iommu_device_sysfs_remove(&zdev->iommu_dev);
zdev              361 drivers/iommu/s390-iommu.c void zpci_destroy_iommu(struct zpci_dev *zdev)
zdev              363 drivers/iommu/s390-iommu.c 	iommu_device_unregister(&zdev->iommu_dev);
zdev              364 drivers/iommu/s390-iommu.c 	iommu_device_sysfs_remove(&zdev->iommu_dev);
zdev              244 drivers/net/ethernet/8390/xsurf100.c static int xsurf100_probe(struct zorro_dev *zdev,
zdev              252 drivers/net/ethernet/8390/xsurf100.c 		DEFINE_RES_MEM(zdev->resource.start + XS100_8390_BASE,
zdev              267 drivers/net/ethernet/8390/xsurf100.c 	if (!request_mem_region(zdev->resource.start, 0x100, zdev->name)) {
zdev              268 drivers/net/ethernet/8390/xsurf100.c 		dev_err(&zdev->dev, "cannot reserve X-Surf 100 control registers\n");
zdev              272 drivers/net/ethernet/8390/xsurf100.c 	if (!request_mem_region(zdev->resource.start +
zdev              276 drivers/net/ethernet/8390/xsurf100.c 		dev_err(&zdev->dev, "cannot reserve 32-bit area\n");
zdev              291 drivers/net/ethernet/8390/xsurf100.c 	ax88796_data.base_regs = ioremap(zdev->resource.start, 0x100);
zdev              295 drivers/net/ethernet/8390/xsurf100.c 		dev_err(&zdev->dev, "Cannot ioremap area %pR (registers)\n",
zdev              296 drivers/net/ethernet/8390/xsurf100.c 			&zdev->resource);
zdev              302 drivers/net/ethernet/8390/xsurf100.c 	ax88796_data.data_area = ioremap(zdev->resource.start +
zdev              307 drivers/net/ethernet/8390/xsurf100.c 		dev_err(&zdev->dev,
zdev              309 drivers/net/ethernet/8390/xsurf100.c 			&zdev->resource,  XS100_8390_DATA32_BASE);
zdev              318 drivers/net/ethernet/8390/xsurf100.c 	pdev = platform_device_register_resndata(&zdev->dev, "ax88796",
zdev              319 drivers/net/ethernet/8390/xsurf100.c 						 zdev->slotaddr, res, 2,
zdev              324 drivers/net/ethernet/8390/xsurf100.c 		dev_err(&zdev->dev, "cannot register platform device\n");
zdev              329 drivers/net/ethernet/8390/xsurf100.c 	zorro_set_drvdata(zdev, pdev);
zdev              341 drivers/net/ethernet/8390/xsurf100.c 	release_mem_region(zdev->resource.start + XS100_8390_DATA32_BASE,
zdev              345 drivers/net/ethernet/8390/xsurf100.c 	release_mem_region(zdev->resource.start, 0x100);
zdev              350 drivers/net/ethernet/8390/xsurf100.c static void xsurf100_remove(struct zorro_dev *zdev)
zdev              352 drivers/net/ethernet/8390/xsurf100.c 	struct platform_device *pdev = zorro_get_drvdata(zdev);
zdev              358 drivers/net/ethernet/8390/xsurf100.c 	release_mem_region(zdev->resource.start, 0x100);
zdev              360 drivers/net/ethernet/8390/xsurf100.c 	release_mem_region(zdev->resource.start + XS100_8390_DATA32_BASE,
zdev               36 drivers/pci/hotplug/s390_pci_hpc.c 	struct zpci_dev *zdev;
zdev               46 drivers/pci/hotplug/s390_pci_hpc.c 	int ret = sclp_pci_configure(slot->zdev->fid);
zdev               48 drivers/pci/hotplug/s390_pci_hpc.c 	zpci_dbg(3, "conf fid:%x, rc:%d\n", slot->zdev->fid, ret);
zdev               50 drivers/pci/hotplug/s390_pci_hpc.c 		slot->zdev->state = ZPCI_FN_STATE_CONFIGURED;
zdev               57 drivers/pci/hotplug/s390_pci_hpc.c 	int ret = sclp_pci_deconfigure(slot->zdev->fid);
zdev               59 drivers/pci/hotplug/s390_pci_hpc.c 	zpci_dbg(3, "deconf fid:%x, rc:%d\n", slot->zdev->fid, ret);
zdev               61 drivers/pci/hotplug/s390_pci_hpc.c 		slot->zdev->state = ZPCI_FN_STATE_STANDBY;
zdev               71 drivers/pci/hotplug/s390_pci_hpc.c 	if (slot->zdev->state != ZPCI_FN_STATE_STANDBY)
zdev               78 drivers/pci/hotplug/s390_pci_hpc.c 	rc = zpci_enable_device(slot->zdev);
zdev               82 drivers/pci/hotplug/s390_pci_hpc.c 	pci_scan_slot(slot->zdev->bus, ZPCI_DEVFN);
zdev               84 drivers/pci/hotplug/s390_pci_hpc.c 	pci_bus_add_devices(slot->zdev->bus);
zdev              100 drivers/pci/hotplug/s390_pci_hpc.c 	if (!zpci_fn_configured(slot->zdev->state))
zdev              103 drivers/pci/hotplug/s390_pci_hpc.c 	pdev = pci_get_slot(slot->zdev->bus, ZPCI_DEVFN);
zdev              109 drivers/pci/hotplug/s390_pci_hpc.c 	rc = zpci_disable_device(slot->zdev);
zdev              120 drivers/pci/hotplug/s390_pci_hpc.c 	switch (slot->zdev->state) {
zdev              145 drivers/pci/hotplug/s390_pci_hpc.c int zpci_init_slot(struct zpci_dev *zdev)
zdev              151 drivers/pci/hotplug/s390_pci_hpc.c 	if (!zdev)
zdev              158 drivers/pci/hotplug/s390_pci_hpc.c 	slot->zdev = zdev;
zdev              161 drivers/pci/hotplug/s390_pci_hpc.c 	snprintf(name, SLOT_NAME_SIZE, "%08x", zdev->fid);
zdev              162 drivers/pci/hotplug/s390_pci_hpc.c 	rc = pci_hp_register(&slot->hotplug_slot, zdev->bus,
zdev              176 drivers/pci/hotplug/s390_pci_hpc.c void zpci_exit_slot(struct zpci_dev *zdev)
zdev              182 drivers/pci/hotplug/s390_pci_hpc.c 		if (slot->zdev != zdev)
zdev              127 drivers/s390/crypto/zcrypt_card.c 	struct zcrypt_card *zdev =
zdev              129 drivers/s390/crypto/zcrypt_card.c 	zcrypt_card_free(zdev);
zdev              214 drivers/s390/net/ism.h 	struct zpci_dev *zdev = to_zpci(ism->pdev);
zdev              215 drivers/s390/net/ism.h 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 2, 8);
zdev              228 drivers/s390/net/ism.h 	struct zpci_dev *zdev = to_zpci(ism->pdev);
zdev              229 drivers/s390/net/ism.h 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 2, len);
zdev              238 drivers/s390/net/ism.h 	struct zpci_dev *zdev = to_zpci(ism->pdev);
zdev              239 drivers/s390/net/ism.h 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, size);
zdev             1947 drivers/video/fbdev/cirrusfb.c 	struct zorro_dev *zdev = to_zorro_dev(info->device);
zdev             1954 drivers/video/fbdev/cirrusfb.c 	zorro_release_device(zdev);