tdev 70 arch/mips/dec/tc.c void __init tc_device_get_irq(struct tc_dev *tdev) tdev 72 arch/mips/dec/tc.c switch (tdev->slot) { tdev 74 arch/mips/dec/tc.c tdev->interrupt = dec_interrupt[DEC_IRQ_TC0]; tdev 77 arch/mips/dec/tc.c tdev->interrupt = dec_interrupt[DEC_IRQ_TC1]; tdev 80 arch/mips/dec/tc.c tdev->interrupt = dec_interrupt[DEC_IRQ_TC2]; tdev 86 arch/mips/dec/tc.c tdev->interrupt = dec_interrupt[DEC_IRQ_TC5]; tdev 89 arch/mips/dec/tc.c tdev->interrupt = dec_interrupt[DEC_IRQ_TC6]; tdev 92 arch/mips/dec/tc.c tdev->interrupt = -1; tdev 920 arch/mips/txx9/generic/setup.c struct txx9_sramc_dev *tdev; tdev 922 arch/mips/txx9/generic/setup.c tdev = container_of(dev, struct txx9_sramc_dev, dev); tdev 923 arch/mips/txx9/generic/setup.c kfree(tdev); tdev 63 drivers/ata/libata-acpi.c NULL : ACPI_HANDLE(&dev->tdev); tdev 73 drivers/ata/libata-acpi.c struct ata_device *tdev; tdev 76 drivers/ata/libata-acpi.c ata_for_each_dev(tdev, tlink, ALL) tdev 77 drivers/ata/libata-acpi.c tdev->flags |= ATA_DFLAG_DETACH; tdev 188 drivers/ata/libata-acpi.c acpi_preset_companion(&ap->tdev, host_companion, ap->port_no); tdev 193 drivers/ata/libata-acpi.c adev = ACPI_COMPANION(&ap->tdev); tdev 209 drivers/ata/libata-acpi.c struct acpi_device *port_companion = ACPI_COMPANION(&ap->tdev); tdev 234 drivers/ata/libata-acpi.c acpi_preset_companion(&dev->tdev, parent, adr); tdev 235 drivers/ata/libata-acpi.c adev = ACPI_COMPANION(&dev->tdev); tdev 269 drivers/ata/libata-acpi.c if (ACPI_HANDLE(&ap->tdev) && gtm) tdev 293 drivers/ata/libata-acpi.c acpi_handle handle = ACPI_HANDLE(&ap->tdev); tdev 368 drivers/ata/libata-acpi.c status = acpi_evaluate_object(ACPI_HANDLE(&ap->tdev), "_STM", tdev 844 drivers/ata/libata-acpi.c if (ACPI_HANDLE(&ap->tdev) && gtm) { tdev 889 drivers/ata/libata-acpi.c return acpi_pm_device_sleep_state(&dev->tdev, NULL, d_max_in); tdev 926 drivers/ata/libata-acpi.c port_handle = ACPI_HANDLE(&ap->tdev); tdev 428 drivers/ata/libata-eh.c struct ata_device *tdev; tdev 432 drivers/ata/libata-eh.c ata_for_each_dev(tdev, link, ALL) tdev 433 drivers/ata/libata-eh.c ehi->dev_action[tdev->devno] &= ~action; tdev 440 drivers/ata/libata-eh.c ata_for_each_dev(tdev, link, ALL) tdev 441 drivers/ata/libata-eh.c ehi->dev_action[tdev->devno] |= tdev 4556 drivers/ata/libata-scsi.c rc = scsi_add_host_with_dma(shost, &ap->tdev, ap->host->dev); tdev 69 drivers/ata/libata-transport.c container_of((d), struct ata_device, tdev) tdev 74 drivers/ata/libata-transport.c container_of((d), struct ata_link, tdev) tdev 79 drivers/ata/libata-transport.c container_of((d), struct ata_port, tdev) tdev 259 drivers/ata/libata-transport.c struct device *dev = &ap->tdev; tdev 283 drivers/ata/libata-transport.c struct device *dev = &ap->tdev; tdev 385 drivers/ata/libata-transport.c struct device *dev = &link->tdev; tdev 409 drivers/ata/libata-transport.c struct device *dev = &link->tdev; tdev 415 drivers/ata/libata-transport.c dev->parent = &ap->tdev; tdev 629 drivers/ata/libata-transport.c transport_destroy_device(&dev->tdev); tdev 630 drivers/ata/libata-transport.c put_device(&dev->tdev); tdev 641 drivers/ata/libata-transport.c struct device *dev = &ata_dev->tdev; tdev 660 drivers/ata/libata-transport.c struct device *dev = &ata_dev->tdev; tdev 666 drivers/ata/libata-transport.c dev->parent = &link->tdev; tdev 192 drivers/ata/libata-zpodd.c acpi_pm_set_device_wakeup(&dev->tdev, true); tdev 201 drivers/ata/libata-zpodd.c acpi_pm_set_device_wakeup(&dev->tdev, false); tdev 267 drivers/ata/libata-zpodd.c struct acpi_device *adev = ACPI_COMPANION(&dev->tdev); tdev 287 drivers/ata/libata-zpodd.c dev_pm_qos_expose_flags(&dev->tdev, 0); tdev 88 drivers/ata/libata.h #define to_ata_port(d) container_of(d, struct ata_port, tdev) tdev 41 drivers/ata/pata_acpi.c if (ACPI_HANDLE(&ap->tdev) == NULL || ata_acpi_gtm(ap, &acpi->gtm) < 0) tdev 197 drivers/ata/pata_acpi.c if (ACPI_HANDLE(&ap->tdev) == NULL) tdev 331 drivers/dma/mmp_tdma.c struct mmp_tdma_device *tdev = dev_id; tdev 336 drivers/dma/mmp_tdma.c struct mmp_tdma_chan *tdmac = tdev->tdmac[i]; tdev 553 drivers/dma/mmp_tdma.c static int mmp_tdma_chan_init(struct mmp_tdma_device *tdev, tdev 560 drivers/dma/mmp_tdma.c dev_err(tdev->dev, "too many channels for device!\n"); tdev 565 drivers/dma/mmp_tdma.c tdmac = devm_kzalloc(tdev->dev, sizeof(*tdmac), GFP_KERNEL); tdev 571 drivers/dma/mmp_tdma.c tdmac->dev = tdev->dev; tdev 572 drivers/dma/mmp_tdma.c tdmac->chan.device = &tdev->device; tdev 575 drivers/dma/mmp_tdma.c tdmac->reg_base = tdev->base + idx * 4; tdev 578 drivers/dma/mmp_tdma.c tdev->tdmac[tdmac->idx] = tdmac; tdev 583 drivers/dma/mmp_tdma.c &tdev->device.channels); tdev 604 drivers/dma/mmp_tdma.c struct mmp_tdma_device *tdev = ofdma->of_dma_data; tdev 605 drivers/dma/mmp_tdma.c dma_cap_mask_t mask = tdev->device.cap_mask; tdev 631 drivers/dma/mmp_tdma.c struct mmp_tdma_device *tdev; tdev 645 drivers/dma/mmp_tdma.c tdev = devm_kzalloc(&pdev->dev, sizeof(*tdev), GFP_KERNEL); tdev 646 drivers/dma/mmp_tdma.c if (!tdev) tdev 649 drivers/dma/mmp_tdma.c tdev->dev = &pdev->dev; tdev 657 drivers/dma/mmp_tdma.c tdev->base = devm_ioremap_resource(&pdev->dev, iores); tdev 658 drivers/dma/mmp_tdma.c if (IS_ERR(tdev->base)) tdev 659 drivers/dma/mmp_tdma.c return PTR_ERR(tdev->base); tdev 661 drivers/dma/mmp_tdma.c INIT_LIST_HEAD(&tdev->device.channels); tdev 675 drivers/dma/mmp_tdma.c mmp_tdma_int_handler, 0, "tdma", tdev); tdev 683 drivers/dma/mmp_tdma.c ret = mmp_tdma_chan_init(tdev, i, irq, type, pool); tdev 688 drivers/dma/mmp_tdma.c dma_cap_set(DMA_SLAVE, tdev->device.cap_mask); tdev 689 drivers/dma/mmp_tdma.c dma_cap_set(DMA_CYCLIC, tdev->device.cap_mask); tdev 690 drivers/dma/mmp_tdma.c tdev->device.dev = &pdev->dev; tdev 691 drivers/dma/mmp_tdma.c tdev->device.device_alloc_chan_resources = tdev 693 drivers/dma/mmp_tdma.c tdev->device.device_free_chan_resources = tdev 695 drivers/dma/mmp_tdma.c tdev->device.device_prep_dma_cyclic = mmp_tdma_prep_dma_cyclic; tdev 696 drivers/dma/mmp_tdma.c tdev->device.device_tx_status = mmp_tdma_tx_status; tdev 697 drivers/dma/mmp_tdma.c tdev->device.device_issue_pending = mmp_tdma_issue_pending; tdev 698 drivers/dma/mmp_tdma.c tdev->device.device_config = mmp_tdma_config; tdev 699 drivers/dma/mmp_tdma.c tdev->device.device_pause = mmp_tdma_pause_chan; tdev 700 drivers/dma/mmp_tdma.c tdev->device.device_resume = mmp_tdma_resume_chan; tdev 701 drivers/dma/mmp_tdma.c tdev->device.device_terminate_all = mmp_tdma_terminate_all; tdev 702 drivers/dma/mmp_tdma.c tdev->device.copy_align = DMAENGINE_ALIGN_8_BYTES; tdev 705 drivers/dma/mmp_tdma.c platform_set_drvdata(pdev, tdev); tdev 707 drivers/dma/mmp_tdma.c ret = dmaenginem_async_device_register(&tdev->device); tdev 709 drivers/dma/mmp_tdma.c dev_err(tdev->device.dev, "unable to register\n"); tdev 715 drivers/dma/mmp_tdma.c mmp_tdma_xlate, tdev); tdev 717 drivers/dma/mmp_tdma.c dev_err(tdev->device.dev, tdev 723 drivers/dma/mmp_tdma.c dev_info(tdev->device.dev, "initialized\n"); tdev 31 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_power_up(struct nvkm_device_tegra *tdev) tdev 35 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->vdd) { tdev 36 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = regulator_enable(tdev->vdd); tdev 41 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = clk_prepare_enable(tdev->clk); tdev 44 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->clk_ref) { tdev 45 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = clk_prepare_enable(tdev->clk_ref); tdev 49 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = clk_prepare_enable(tdev->clk_pwr); tdev 52 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c clk_set_rate(tdev->clk_pwr, 204000000); tdev 55 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c reset_control_assert(tdev->rst); tdev 58 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (!tdev->pdev->dev.pm_domain) { tdev 65 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c reset_control_deassert(tdev->rst); tdev 71 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c clk_disable_unprepare(tdev->clk_pwr); tdev 73 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->clk_ref) tdev 74 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c clk_disable_unprepare(tdev->clk_ref); tdev 76 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c clk_disable_unprepare(tdev->clk); tdev 78 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->vdd) tdev 79 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c regulator_disable(tdev->vdd); tdev 85 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_power_down(struct nvkm_device_tegra *tdev) tdev 89 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c clk_disable_unprepare(tdev->clk_pwr); tdev 90 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->clk_ref) tdev 91 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c clk_disable_unprepare(tdev->clk_ref); tdev 92 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c clk_disable_unprepare(tdev->clk); tdev 95 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->vdd) { tdev 96 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = regulator_disable(tdev->vdd); tdev 105 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_probe_iommu(struct nvkm_device_tegra *tdev) tdev 108 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct device *dev = &tdev->pdev->dev; tdev 121 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (!tdev->func->iommu_bit) tdev 124 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c mutex_init(&tdev->iommu.mutex); tdev 127 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->iommu.domain = iommu_domain_alloc(&platform_bus_type); tdev 128 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (!tdev->iommu.domain) tdev 136 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c pgsize_bitmap = tdev->iommu.domain->ops->pgsize_bitmap; tdev 138 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->iommu.pgshift = PAGE_SHIFT; tdev 140 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->iommu.pgshift = fls(pgsize_bitmap & ~PAGE_MASK); tdev 141 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->iommu.pgshift == 0) { tdev 145 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->iommu.pgshift -= 1; tdev 148 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = iommu_attach_device(tdev->iommu.domain, dev); tdev 152 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = nvkm_mm_init(&tdev->iommu.mm, 0, 0, tdev 153 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c (1ULL << tdev->func->iommu_bit) >> tdev 154 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->iommu.pgshift, 1); tdev 162 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c iommu_detach_device(tdev->iommu.domain, dev); tdev 165 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c iommu_domain_free(tdev->iommu.domain); tdev 168 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->iommu.domain = NULL; tdev 169 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->iommu.pgshift = 0; tdev 175 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_remove_iommu(struct nvkm_device_tegra *tdev) tdev 178 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->iommu.domain) { tdev 179 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_mm_fini(&tdev->iommu.mm); tdev 180 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c iommu_detach_device(tdev->iommu.domain, tdev->device.dev); tdev 181 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c iommu_domain_free(tdev->iommu.domain); tdev 195 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct nvkm_device_tegra *tdev = nvkm_device_tegra(device); tdev 196 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c return platform_get_resource(tdev->pdev, IORESOURCE_MEM, bar); tdev 216 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct nvkm_device_tegra *tdev = arg; tdev 217 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct nvkm_device *device = &tdev->device; tdev 228 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct nvkm_device_tegra *tdev = nvkm_device_tegra(device); tdev 229 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (tdev->irq) { tdev 230 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c free_irq(tdev->irq, tdev); tdev 231 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->irq = 0; tdev 238 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct nvkm_device_tegra *tdev = nvkm_device_tegra(device); tdev 241 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c irq = platform_get_irq_byname(tdev->pdev, "stall"); tdev 246 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c IRQF_SHARED, "nvkm", tdev); tdev 250 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->irq = irq; tdev 257 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct nvkm_device_tegra *tdev = nvkm_device_tegra(device); tdev 258 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_power_down(tdev); tdev 259 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_remove_iommu(tdev); tdev 260 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c return tdev; tdev 281 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c struct nvkm_device_tegra *tdev; tdev 284 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (!(tdev = kzalloc(sizeof(*tdev), GFP_KERNEL))) tdev 287 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->func = func; tdev 288 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->pdev = pdev; tdev 291 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->vdd = devm_regulator_get(&pdev->dev, "vdd"); tdev 292 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (IS_ERR(tdev->vdd)) { tdev 293 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = PTR_ERR(tdev->vdd); tdev 298 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->rst = devm_reset_control_get(&pdev->dev, "gpu"); tdev 299 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (IS_ERR(tdev->rst)) { tdev 300 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = PTR_ERR(tdev->rst); tdev 304 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->clk = devm_clk_get(&pdev->dev, "gpu"); tdev 305 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (IS_ERR(tdev->clk)) { tdev 306 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = PTR_ERR(tdev->clk); tdev 311 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->clk_ref = devm_clk_get(&pdev->dev, "ref"); tdev 312 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (IS_ERR(tdev->clk_ref)) { tdev 313 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = PTR_ERR(tdev->clk_ref); tdev 317 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->clk_pwr = devm_clk_get(&pdev->dev, "pwr"); tdev 318 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c if (IS_ERR(tdev->clk_pwr)) { tdev 319 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = PTR_ERR(tdev->clk_pwr); tdev 326 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = dma_set_mask(&pdev->dev, DMA_BIT_MASK(tdev->func->iommu_bit)); tdev 330 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_probe_iommu(tdev); tdev 332 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c ret = nvkm_device_tegra_power_up(tdev); tdev 336 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->gpu_speedo = tegra_sku_info.gpu_speedo_value; tdev 337 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c tdev->gpu_speedo_id = tegra_sku_info.gpu_speedo_id; tdev 341 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c &tdev->device); tdev 345 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c *pdevice = &tdev->device; tdev 350 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_power_down(tdev); tdev 352 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c nvkm_device_tegra_remove_iommu(tdev); tdev 354 drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c kfree(tdev); tdev 618 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk20a.c struct nvkm_device_tegra *tdev = device->func->tegra(device); tdev 629 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk20a.c clk->parent_rate = clk_get_rate(tdev->clk); tdev 1019 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gm20b.c struct nvkm_device_tegra *tdev = device->func->tegra(device); tdev 1026 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gm20b.c if (tdev->gpu_speedo_id == 0) tdev 574 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct nvkm_device_tegra *tdev = device->func->tegra(device); tdev 588 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (tdev->iommu.domain) { tdev 589 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c imem->mm_mutex = &tdev->iommu.mutex; tdev 590 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c imem->mm = &tdev->iommu.mm; tdev 591 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c imem->domain = tdev->iommu.domain; tdev 592 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c imem->iommu_pgshift = tdev->iommu.pgshift; tdev 593 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c imem->iommu_bit = tdev->func->iommu_bit; tdev 151 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gk20a.c struct nvkm_device_tegra *tdev = device->func->tegra(device); tdev 156 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gk20a.c uv = regulator_get_voltage(tdev->vdd); tdev 159 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gk20a.c volt->vdd = tdev->vdd; tdev 165 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gk20a.c gk20a_volt_calc_voltage(&coefs[i], tdev->gpu_speedo), tdev 69 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gm20b.c struct nvkm_device_tegra *tdev = device->func->tegra(device); tdev 73 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gm20b.c if (tdev->gpu_speedo_id >= ARRAY_SIZE(speedo_to_vmin)) { tdev 75 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gm20b.c tdev->gpu_speedo_id); tdev 84 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gm20b.c vmin = speedo_to_vmin[tdev->gpu_speedo_id]; tdev 86 drivers/gpu/drm/nouveau/nvkm/subdev/volt/gm20b.c if (tdev->gpu_speedo_id >= 1) tdev 70 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev; tdev 166 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev = tfile->tdev; tdev 176 drivers/gpu/drm/vmwgfx/ttm_object.c spin_lock(&tdev->object_lock); tdev 177 drivers/gpu/drm/vmwgfx/ttm_object.c ret = idr_alloc(&tdev->idr, base, 1, 0, GFP_NOWAIT); tdev 178 drivers/gpu/drm/vmwgfx/ttm_object.c spin_unlock(&tdev->object_lock); tdev 192 drivers/gpu/drm/vmwgfx/ttm_object.c spin_lock(&tdev->object_lock); tdev 193 drivers/gpu/drm/vmwgfx/ttm_object.c idr_remove(&tdev->idr, base->handle); tdev 194 drivers/gpu/drm/vmwgfx/ttm_object.c spin_unlock(&tdev->object_lock); tdev 202 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev = base->tfile->tdev; tdev 204 drivers/gpu/drm/vmwgfx/ttm_object.c spin_lock(&tdev->object_lock); tdev 205 drivers/gpu/drm/vmwgfx/ttm_object.c idr_remove(&tdev->idr, base->handle); tdev 206 drivers/gpu/drm/vmwgfx/ttm_object.c spin_unlock(&tdev->object_lock); tdev 285 drivers/gpu/drm/vmwgfx/ttm_object.c ttm_base_object_lookup_for_ref(struct ttm_object_device *tdev, uint32_t key) tdev 290 drivers/gpu/drm/vmwgfx/ttm_object.c base = idr_find(&tdev->idr, key); tdev 352 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_mem_global *mem_glob = tfile->tdev->mem_glob; tdev 427 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_mem_global *mem_glob = tfile->tdev->mem_glob; tdev 491 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_file *ttm_object_file_init(struct ttm_object_device *tdev, tdev 503 drivers/gpu/drm/vmwgfx/ttm_object.c tfile->tdev = tdev; tdev 530 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev = kmalloc(sizeof(*tdev), GFP_KERNEL); tdev 533 drivers/gpu/drm/vmwgfx/ttm_object.c if (unlikely(tdev == NULL)) tdev 536 drivers/gpu/drm/vmwgfx/ttm_object.c tdev->mem_glob = mem_glob; tdev 537 drivers/gpu/drm/vmwgfx/ttm_object.c spin_lock_init(&tdev->object_lock); tdev 538 drivers/gpu/drm/vmwgfx/ttm_object.c atomic_set(&tdev->object_count, 0); tdev 539 drivers/gpu/drm/vmwgfx/ttm_object.c ret = drm_ht_create(&tdev->object_hash, hash_order); tdev 543 drivers/gpu/drm/vmwgfx/ttm_object.c idr_init(&tdev->idr); tdev 544 drivers/gpu/drm/vmwgfx/ttm_object.c tdev->ops = *ops; tdev 545 drivers/gpu/drm/vmwgfx/ttm_object.c tdev->dmabuf_release = tdev->ops.release; tdev 546 drivers/gpu/drm/vmwgfx/ttm_object.c tdev->ops.release = ttm_prime_dmabuf_release; tdev 547 drivers/gpu/drm/vmwgfx/ttm_object.c tdev->dma_buf_size = ttm_round_pot(sizeof(struct dma_buf)) + tdev 549 drivers/gpu/drm/vmwgfx/ttm_object.c return tdev; tdev 552 drivers/gpu/drm/vmwgfx/ttm_object.c kfree(tdev); tdev 558 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev = *p_tdev; tdev 562 drivers/gpu/drm/vmwgfx/ttm_object.c WARN_ON_ONCE(!idr_is_empty(&tdev->idr)); tdev 563 drivers/gpu/drm/vmwgfx/ttm_object.c idr_destroy(&tdev->idr); tdev 564 drivers/gpu/drm/vmwgfx/ttm_object.c drm_ht_remove(&tdev->object_hash); tdev 566 drivers/gpu/drm/vmwgfx/ttm_object.c kfree(tdev); tdev 625 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev = base->tfile->tdev; tdev 627 drivers/gpu/drm/vmwgfx/ttm_object.c if (tdev->dmabuf_release) tdev 628 drivers/gpu/drm/vmwgfx/ttm_object.c tdev->dmabuf_release(dma_buf); tdev 633 drivers/gpu/drm/vmwgfx/ttm_object.c ttm_mem_global_free(tdev->mem_glob, tdev->dma_buf_size); tdev 651 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev = tfile->tdev; tdev 661 drivers/gpu/drm/vmwgfx/ttm_object.c if (dma_buf->ops != &tdev->ops) tdev 687 drivers/gpu/drm/vmwgfx/ttm_object.c struct ttm_object_device *tdev = tfile->tdev; tdev 719 drivers/gpu/drm/vmwgfx/ttm_object.c exp_info.ops = &tdev->ops; tdev 727 drivers/gpu/drm/vmwgfx/ttm_object.c ret = ttm_mem_global_alloc(tdev->mem_glob, tdev->dma_buf_size, tdev 737 drivers/gpu/drm/vmwgfx/ttm_object.c ttm_mem_global_free(tdev->mem_glob, tdev 738 drivers/gpu/drm/vmwgfx/ttm_object.c tdev->dma_buf_size); tdev 211 drivers/gpu/drm/vmwgfx/ttm_object.h ttm_base_object_lookup_for_ref(struct ttm_object_device *tdev, uint32_t key); tdev 280 drivers/gpu/drm/vmwgfx/ttm_object.h *tdev, tdev 789 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c dev_priv->tdev = ttm_object_device_init(&ttm_mem_glob, 12, tdev 792 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c if (unlikely(dev_priv->tdev == NULL)) { tdev 948 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c ttm_object_device_release(&dev_priv->tdev); tdev 998 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c ttm_object_device_release(&dev_priv->tdev); tdev 1028 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c vmw_fp->tfile = ttm_object_file_init(dev_priv->tdev, 10); tdev 511 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h struct ttm_object_device *tdev; tdev 926 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c base = ttm_base_object_lookup_for_ref(dev_priv->tdev, handle); tdev 61 drivers/infiniband/hw/cxgb3/cxio_hal.c static struct cxio_rdev *cxio_hal_find_rdev_by_t3cdev(struct t3cdev *tdev) tdev 66 drivers/infiniband/hw/cxgb3/cxio_hal.c if (rdev->t3cdev_p == tdev) tdev 196 drivers/infiniband/hw/cxgb3/cxio_hal.h int iwch_cxgb3_ofld_send(struct t3cdev *tdev, struct sk_buff *skb); tdev 135 drivers/infiniband/hw/cxgb3/iwch.c static void open_rnic_dev(struct t3cdev *tdev) tdev 139 drivers/infiniband/hw/cxgb3/iwch.c pr_debug("%s t3cdev %p\n", __func__, tdev); tdev 147 drivers/infiniband/hw/cxgb3/iwch.c rnicp->rdev.t3cdev_p = tdev; tdev 165 drivers/infiniband/hw/cxgb3/iwch.c close_rnic_dev(tdev); tdev 172 drivers/infiniband/hw/cxgb3/iwch.c static void close_rnic_dev(struct t3cdev *tdev) tdev 175 drivers/infiniband/hw/cxgb3/iwch.c pr_debug("%s t3cdev %p\n", __func__, tdev); tdev 178 drivers/infiniband/hw/cxgb3/iwch.c if (dev->rdev.t3cdev_p == tdev) { tdev 195 drivers/infiniband/hw/cxgb3/iwch.c static void iwch_event_handler(struct t3cdev *tdev, u32 evt, u32 port_id) tdev 197 drivers/infiniband/hw/cxgb3/iwch.c struct cxio_rdev *rdev = tdev->ulp; tdev 137 drivers/infiniband/hw/cxgb3/iwch_cm.c static int iwch_l2t_send(struct t3cdev *tdev, struct sk_buff *skb, struct l2t_entry *l2e) tdev 142 drivers/infiniband/hw/cxgb3/iwch_cm.c rdev = (struct cxio_rdev *)tdev->ulp; tdev 147 drivers/infiniband/hw/cxgb3/iwch_cm.c error = l2t_send(tdev, skb, l2e); tdev 153 drivers/infiniband/hw/cxgb3/iwch_cm.c int iwch_cxgb3_ofld_send(struct t3cdev *tdev, struct sk_buff *skb) tdev 158 drivers/infiniband/hw/cxgb3/iwch_cm.c rdev = (struct cxio_rdev *)tdev->ulp; tdev 163 drivers/infiniband/hw/cxgb3/iwch_cm.c error = cxgb3_ofld_send(tdev, skb); tdev 169 drivers/infiniband/hw/cxgb3/iwch_cm.c static void release_tid(struct t3cdev *tdev, u32 hwtid, struct sk_buff *skb) tdev 180 drivers/infiniband/hw/cxgb3/iwch_cm.c iwch_cxgb3_ofld_send(tdev, skb); tdev 202 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_cxgb3_ofld_send(ep->com.tdev, skb); tdev 223 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_cxgb3_ofld_send(ep->com.tdev, skb); tdev 229 drivers/infiniband/hw/cxgb3/iwch_cm.c ep->emss = T3C_DATA(ep->com.tdev)->mtus[G_TCPOPT_MSS(opt)] - 40; tdev 286 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_remove_tid(ep->com.tdev, (void *)ep, ep->hwtid); tdev 288 drivers/infiniband/hw/cxgb3/iwch_cm.c l2t_release(ep->com.tdev, ep->l2t); tdev 403 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_l2t_send(ep->com.tdev, skb, ep->l2t); tdev 423 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_l2t_send(ep->com.tdev, skb, ep->l2t); tdev 441 drivers/infiniband/hw/cxgb3/iwch_cm.c mtu_idx = find_best_mtu(T3C_DATA(ep->com.tdev), dst_mtu(ep->dst)); tdev 467 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_l2t_send(ep->com.tdev, skb, ep->l2t); tdev 524 drivers/infiniband/hw/cxgb3/iwch_cm.c iwch_l2t_send(ep->com.tdev, skb, ep->l2t); tdev 575 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_l2t_send(ep->com.tdev, skb, ep->l2t); tdev 626 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_l2t_send(ep->com.tdev, skb, ep->l2t); tdev 629 drivers/infiniband/hw/cxgb3/iwch_cm.c static int act_establish(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 641 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_insert_tid(ep->com.tdev, &t3c_client, ep, tid); tdev 649 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_free_atid(ep->com.tdev, ep->atid); tdev 806 drivers/infiniband/hw/cxgb3/iwch_cm.c iwch_cxgb3_ofld_send(ep->com.tdev, skb); tdev 1040 drivers/infiniband/hw/cxgb3/iwch_cm.c static int rx_data(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1085 drivers/infiniband/hw/cxgb3/iwch_cm.c static int tx_ack(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1132 drivers/infiniband/hw/cxgb3/iwch_cm.c static int abort_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1176 drivers/infiniband/hw/cxgb3/iwch_cm.c static int act_open_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1185 drivers/infiniband/hw/cxgb3/iwch_cm.c if (ep->com.tdev->type != T3A && act_open_has_tid(rpl->status)) tdev 1186 drivers/infiniband/hw/cxgb3/iwch_cm.c release_tid(ep->com.tdev, GET_TID(rpl), NULL); tdev 1187 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_free_atid(ep->com.tdev, ep->atid); tdev 1189 drivers/infiniband/hw/cxgb3/iwch_cm.c l2t_release(ep->com.tdev, ep->l2t); tdev 1219 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_cxgb3_ofld_send(ep->com.tdev, skb); tdev 1222 drivers/infiniband/hw/cxgb3/iwch_cm.c static int pass_open_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1252 drivers/infiniband/hw/cxgb3/iwch_cm.c return iwch_cxgb3_ofld_send(ep->com.tdev, skb); tdev 1255 drivers/infiniband/hw/cxgb3/iwch_cm.c static int close_listsrv_rpl(struct t3cdev *tdev, struct sk_buff *skb, tdev 1279 drivers/infiniband/hw/cxgb3/iwch_cm.c mtu_idx = find_best_mtu(T3C_DATA(ep->com.tdev), dst_mtu(ep->dst)); tdev 1301 drivers/infiniband/hw/cxgb3/iwch_cm.c iwch_l2t_send(ep->com.tdev, skb, ep->l2t); tdev 1306 drivers/infiniband/hw/cxgb3/iwch_cm.c static void reject_cr(struct t3cdev *tdev, u32 hwtid, __be32 peer_ip, tdev 1309 drivers/infiniband/hw/cxgb3/iwch_cm.c pr_debug("%s t3cdev %p tid %u peer_ip %x\n", __func__, tdev, hwtid, tdev 1315 drivers/infiniband/hw/cxgb3/iwch_cm.c if (tdev->type != T3A) tdev 1316 drivers/infiniband/hw/cxgb3/iwch_cm.c release_tid(tdev, hwtid, skb); tdev 1330 drivers/infiniband/hw/cxgb3/iwch_cm.c iwch_cxgb3_ofld_send(tdev, skb); tdev 1334 drivers/infiniband/hw/cxgb3/iwch_cm.c static int pass_accept_req(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1356 drivers/infiniband/hw/cxgb3/iwch_cm.c if (tdev->ctl(tdev, GET_IFF_FROM_MAC, &tim) < 0 || !tim.dev) { tdev 1362 drivers/infiniband/hw/cxgb3/iwch_cm.c rt = find_route(tdev, tdev 1372 drivers/infiniband/hw/cxgb3/iwch_cm.c l2t = t3_l2t_get(tdev, dst, NULL, &req->peer_ip); tdev 1381 drivers/infiniband/hw/cxgb3/iwch_cm.c l2t_release(tdev, l2t); tdev 1386 drivers/infiniband/hw/cxgb3/iwch_cm.c child_ep->com.tdev = tdev; tdev 1401 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_insert_tid(tdev, &t3c_client, child_ep, hwtid); tdev 1405 drivers/infiniband/hw/cxgb3/iwch_cm.c reject_cr(tdev, hwtid, req->peer_ip, skb); tdev 1410 drivers/infiniband/hw/cxgb3/iwch_cm.c static int pass_establish(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1428 drivers/infiniband/hw/cxgb3/iwch_cm.c static int peer_close(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1519 drivers/infiniband/hw/cxgb3/iwch_cm.c static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1533 drivers/infiniband/hw/cxgb3/iwch_cm.c t3_l2t_send_event(ep->com.tdev, ep->l2t); tdev 1620 drivers/infiniband/hw/cxgb3/iwch_cm.c iwch_cxgb3_ofld_send(ep->com.tdev, rpl_skb); tdev 1627 drivers/infiniband/hw/cxgb3/iwch_cm.c static int close_con_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1681 drivers/infiniband/hw/cxgb3/iwch_cm.c static int terminate(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1698 drivers/infiniband/hw/cxgb3/iwch_cm.c static int ec_status(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 1911 drivers/infiniband/hw/cxgb3/iwch_cm.c ep->com.tdev = h->rdev.t3cdev_p; tdev 1940 drivers/infiniband/hw/cxgb3/iwch_cm.c ep->l2t = t3_l2t_get(ep->com.tdev, ep->dst, NULL, tdev 1964 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_free_atid(ep->com.tdev, ep->atid); tdev 1993 drivers/infiniband/hw/cxgb3/iwch_cm.c ep->com.tdev = h->rdev.t3cdev_p; tdev 2023 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_free_stid(ep->com.tdev, ep->stid); tdev 2047 drivers/infiniband/hw/cxgb3/iwch_cm.c cxgb3_free_stid(ep->com.tdev, ep->stid); tdev 2061 drivers/infiniband/hw/cxgb3/iwch_cm.c struct t3cdev *tdev; tdev 2069 drivers/infiniband/hw/cxgb3/iwch_cm.c tdev = (struct t3cdev *)ep->com.tdev; tdev 2070 drivers/infiniband/hw/cxgb3/iwch_cm.c rdev = (struct cxio_rdev *)tdev->ulp; tdev 2137 drivers/infiniband/hw/cxgb3/iwch_cm.c l2t_release(ep->com.tdev, ep->l2t); tdev 2170 drivers/infiniband/hw/cxgb3/iwch_cm.c struct t3cdev *tdev; tdev 2175 drivers/infiniband/hw/cxgb3/iwch_cm.c tdev = *((struct t3cdev **) (skb->cb + sizeof(void *))); tdev 2176 drivers/infiniband/hw/cxgb3/iwch_cm.c ret = work_handlers[G_OPCODE(ntohl((__force __be32)skb->csum))](tdev, skb, ep); tdev 2189 drivers/infiniband/hw/cxgb3/iwch_cm.c static int sched(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 2199 drivers/infiniband/hw/cxgb3/iwch_cm.c *((struct t3cdev **) (skb->cb + sizeof(void *))) = tdev; tdev 2209 drivers/infiniband/hw/cxgb3/iwch_cm.c static int set_tcb_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 157 drivers/infiniband/hw/cxgb3/iwch_cm.h struct t3cdev *tdev; tdev 50 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = filp->private_data; tdev 52 drivers/mailbox/mailbox-test.c if (!tdev->tx_channel) { tdev 53 drivers/mailbox/mailbox-test.c dev_err(tdev->dev, "Channel cannot do Tx\n"); tdev 58 drivers/mailbox/mailbox-test.c dev_err(tdev->dev, tdev 65 drivers/mailbox/mailbox-test.c if (!tdev->signal) { tdev 66 drivers/mailbox/mailbox-test.c tdev->signal = kzalloc(MBOX_MAX_SIG_LEN, GFP_KERNEL); tdev 67 drivers/mailbox/mailbox-test.c if (!tdev->signal) tdev 71 drivers/mailbox/mailbox-test.c if (copy_from_user(tdev->signal, userbuf, count)) { tdev 72 drivers/mailbox/mailbox-test.c kfree(tdev->signal); tdev 73 drivers/mailbox/mailbox-test.c tdev->signal = NULL; tdev 88 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = filp->private_data; tdev 90 drivers/mailbox/mailbox-test.c return fasync_helper(fd, filp, on, &tdev->async_queue); tdev 97 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = filp->private_data; tdev 101 drivers/mailbox/mailbox-test.c if (!tdev->tx_channel) { tdev 102 drivers/mailbox/mailbox-test.c dev_err(tdev->dev, "Channel cannot do Tx\n"); tdev 107 drivers/mailbox/mailbox-test.c dev_err(tdev->dev, tdev 113 drivers/mailbox/mailbox-test.c tdev->message = kzalloc(MBOX_MAX_MSG_LEN, GFP_KERNEL); tdev 114 drivers/mailbox/mailbox-test.c if (!tdev->message) tdev 117 drivers/mailbox/mailbox-test.c ret = copy_from_user(tdev->message, userbuf, count); tdev 127 drivers/mailbox/mailbox-test.c if (tdev->tx_mmio && tdev->signal) { tdev 129 drivers/mailbox/mailbox-test.c tdev->signal, MBOX_MAX_SIG_LEN); tdev 131 drivers/mailbox/mailbox-test.c data = tdev->signal; tdev 133 drivers/mailbox/mailbox-test.c data = tdev->message; tdev 136 drivers/mailbox/mailbox-test.c tdev->message, MBOX_MAX_MSG_LEN); tdev 138 drivers/mailbox/mailbox-test.c ret = mbox_send_message(tdev->tx_channel, data); tdev 140 drivers/mailbox/mailbox-test.c dev_err(tdev->dev, "Failed to send message via mailbox\n"); tdev 143 drivers/mailbox/mailbox-test.c kfree(tdev->signal); tdev 144 drivers/mailbox/mailbox-test.c kfree(tdev->message); tdev 145 drivers/mailbox/mailbox-test.c tdev->signal = NULL; tdev 150 drivers/mailbox/mailbox-test.c static bool mbox_test_message_data_ready(struct mbox_test_device *tdev) tdev 155 drivers/mailbox/mailbox-test.c spin_lock_irqsave(&tdev->lock, flags); tdev 157 drivers/mailbox/mailbox-test.c spin_unlock_irqrestore(&tdev->lock, flags); tdev 165 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = filp->private_data; tdev 177 drivers/mailbox/mailbox-test.c if (!tdev->rx_channel) { tdev 184 drivers/mailbox/mailbox-test.c add_wait_queue(&tdev->waitq, &wait); tdev 189 drivers/mailbox/mailbox-test.c if (mbox_test_message_data_ready(tdev)) tdev 205 drivers/mailbox/mailbox-test.c spin_lock_irqsave(&tdev->lock, flags); tdev 207 drivers/mailbox/mailbox-test.c ptr = tdev->rx_buffer; tdev 220 drivers/mailbox/mailbox-test.c memset(tdev->rx_buffer, 0, MBOX_MAX_MSG_LEN); tdev 223 drivers/mailbox/mailbox-test.c spin_unlock_irqrestore(&tdev->lock, flags); tdev 228 drivers/mailbox/mailbox-test.c remove_wait_queue(&tdev->waitq, &wait); tdev 237 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = filp->private_data; tdev 239 drivers/mailbox/mailbox-test.c poll_wait(filp, &tdev->waitq, wait); tdev 241 drivers/mailbox/mailbox-test.c if (mbox_test_message_data_ready(tdev)) tdev 256 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev) tdev 261 drivers/mailbox/mailbox-test.c tdev->root_debugfs_dir = debugfs_create_dir(dev_name(&pdev->dev), NULL); tdev 262 drivers/mailbox/mailbox-test.c if (!tdev->root_debugfs_dir) { tdev 267 drivers/mailbox/mailbox-test.c debugfs_create_file("message", 0600, tdev->root_debugfs_dir, tdev 268 drivers/mailbox/mailbox-test.c tdev, &mbox_test_message_ops); tdev 270 drivers/mailbox/mailbox-test.c debugfs_create_file("signal", 0200, tdev->root_debugfs_dir, tdev 271 drivers/mailbox/mailbox-test.c tdev, &mbox_test_signal_ops); tdev 278 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = dev_get_drvdata(client->dev); tdev 281 drivers/mailbox/mailbox-test.c spin_lock_irqsave(&tdev->lock, flags); tdev 282 drivers/mailbox/mailbox-test.c if (tdev->rx_mmio) { tdev 283 drivers/mailbox/mailbox-test.c memcpy_fromio(tdev->rx_buffer, tdev->rx_mmio, MBOX_MAX_MSG_LEN); tdev 285 drivers/mailbox/mailbox-test.c tdev->rx_buffer, MBOX_MAX_MSG_LEN); tdev 289 drivers/mailbox/mailbox-test.c memcpy(tdev->rx_buffer, message, MBOX_MAX_MSG_LEN); tdev 292 drivers/mailbox/mailbox-test.c spin_unlock_irqrestore(&tdev->lock, flags); tdev 294 drivers/mailbox/mailbox-test.c wake_up_interruptible(&tdev->waitq); tdev 296 drivers/mailbox/mailbox-test.c kill_fasync(&tdev->async_queue, SIGIO, POLL_IN); tdev 301 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = dev_get_drvdata(client->dev); tdev 303 drivers/mailbox/mailbox-test.c if (tdev->tx_mmio) { tdev 304 drivers/mailbox/mailbox-test.c if (tdev->signal) tdev 305 drivers/mailbox/mailbox-test.c memcpy_toio(tdev->tx_mmio, tdev->message, MBOX_MAX_MSG_LEN); tdev 307 drivers/mailbox/mailbox-test.c memcpy_toio(tdev->tx_mmio, message, MBOX_MAX_MSG_LEN); tdev 351 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev; tdev 356 drivers/mailbox/mailbox-test.c tdev = devm_kzalloc(&pdev->dev, sizeof(*tdev), GFP_KERNEL); tdev 357 drivers/mailbox/mailbox-test.c if (!tdev) tdev 362 drivers/mailbox/mailbox-test.c tdev->tx_mmio = devm_ioremap_resource(&pdev->dev, res); tdev 363 drivers/mailbox/mailbox-test.c if (PTR_ERR(tdev->tx_mmio) == -EBUSY) { tdev 366 drivers/mailbox/mailbox-test.c tdev->tx_mmio = devm_ioremap(&pdev->dev, res->start, size); tdev 367 drivers/mailbox/mailbox-test.c } else if (IS_ERR(tdev->tx_mmio)) { tdev 368 drivers/mailbox/mailbox-test.c tdev->tx_mmio = NULL; tdev 373 drivers/mailbox/mailbox-test.c tdev->rx_mmio = devm_ioremap_resource(&pdev->dev, res); tdev 374 drivers/mailbox/mailbox-test.c if (PTR_ERR(tdev->rx_mmio) == -EBUSY) { tdev 376 drivers/mailbox/mailbox-test.c tdev->rx_mmio = devm_ioremap(&pdev->dev, res->start, size); tdev 377 drivers/mailbox/mailbox-test.c } else if (IS_ERR(tdev->rx_mmio)) { tdev 378 drivers/mailbox/mailbox-test.c tdev->rx_mmio = tdev->tx_mmio; tdev 381 drivers/mailbox/mailbox-test.c tdev->tx_channel = mbox_test_request_channel(pdev, "tx"); tdev 382 drivers/mailbox/mailbox-test.c tdev->rx_channel = mbox_test_request_channel(pdev, "rx"); tdev 384 drivers/mailbox/mailbox-test.c if (!tdev->tx_channel && !tdev->rx_channel) tdev 388 drivers/mailbox/mailbox-test.c if (!tdev->rx_channel && (tdev->rx_mmio != tdev->tx_mmio)) tdev 389 drivers/mailbox/mailbox-test.c tdev->rx_channel = tdev->tx_channel; tdev 391 drivers/mailbox/mailbox-test.c tdev->dev = &pdev->dev; tdev 392 drivers/mailbox/mailbox-test.c platform_set_drvdata(pdev, tdev); tdev 394 drivers/mailbox/mailbox-test.c spin_lock_init(&tdev->lock); tdev 396 drivers/mailbox/mailbox-test.c if (tdev->rx_channel) { tdev 397 drivers/mailbox/mailbox-test.c tdev->rx_buffer = devm_kzalloc(&pdev->dev, tdev 399 drivers/mailbox/mailbox-test.c if (!tdev->rx_buffer) tdev 403 drivers/mailbox/mailbox-test.c ret = mbox_test_add_debugfs(pdev, tdev); tdev 407 drivers/mailbox/mailbox-test.c init_waitqueue_head(&tdev->waitq); tdev 415 drivers/mailbox/mailbox-test.c struct mbox_test_device *tdev = platform_get_drvdata(pdev); tdev 417 drivers/mailbox/mailbox-test.c debugfs_remove_recursive(tdev->root_debugfs_dir); tdev 419 drivers/mailbox/mailbox-test.c if (tdev->tx_channel) tdev 420 drivers/mailbox/mailbox-test.c mbox_free_channel(tdev->tx_channel); tdev 421 drivers/mailbox/mailbox-test.c if (tdev->rx_channel) tdev 422 drivers/mailbox/mailbox-test.c mbox_free_channel(tdev->rx_channel); tdev 221 drivers/net/ethernet/chelsio/cxgb3/adapter.h struct t3cdev tdev; tdev 299 drivers/net/ethernet/chelsio/cxgb3/adapter.h #define tdev2adap(d) container_of(d, struct adapter, tdev) tdev 306 drivers/net/ethernet/chelsio/cxgb3/adapter.h int t3_offload_tx(struct t3cdev *tdev, struct sk_buff *skb); tdev 893 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c static inline int offload_tx(struct t3cdev *tdev, struct sk_buff *skb) tdev 898 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c ret = t3_offload_tx(tdev, skb); tdev 920 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c offload_tx(&adapter->tdev, skb); tdev 1352 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct t3cdev *tdev = dev2t3cdev(dev); tdev 1363 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c tdev->lldev = adapter->port[0]; tdev 1375 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c if (sysfs_create_group(&tdev->lldev->dev.kobj, &offload_attr_group)) tdev 1379 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_add_clients(tdev); tdev 1386 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_set_dummy_ops(tdev); tdev 1391 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c static int offload_close(struct t3cdev *tdev) tdev 1393 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct adapter *adapter = tdev2adap(tdev); tdev 1394 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct t3c_data *td = T3C_DATA(tdev); tdev 1400 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_remove_clients(tdev); tdev 1402 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c sysfs_remove_group(&tdev->lldev->dev.kobj, &offload_attr_group); tdev 1407 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c tdev->lldev = NULL; tdev 1408 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_set_dummy_ops(tdev); tdev 1446 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_event_notify(&adapter->tdev, OFFLOAD_PORT_UP, pi->port_id); tdev 1479 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_event_notify(&adapter->tdev, OFFLOAD_PORT_DOWN, pi->port_id); tdev 2802 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_event_notify(&adapter->tdev, OFFLOAD_DB_FULL, 0); tdev 2810 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_event_notify(&adapter->tdev, OFFLOAD_DB_EMPTY, 0); tdev 2820 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_event_notify(&adapter->tdev, OFFLOAD_DB_DROP, 0); tdev 2905 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_event_notify(&adapter->tdev, OFFLOAD_STATUS_DOWN, 0); tdev 2906 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c offload_close(&adapter->tdev); tdev 2971 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cxgb3_event_notify(&adapter->tdev, OFFLOAD_STATUS_UP, 0); tdev 3431 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c offload_close(&adapter->tdev); tdev 70 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c static inline int offload_activated(struct t3cdev *tdev) tdev 72 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c const struct adapter *adapter = tdev2adap(tdev); tdev 86 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev; tdev 92 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c list_for_each_entry(tdev, &ofld_dev_list, ofld_dev_list) { tdev 93 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c if (offload_activated(tdev)) tdev 94 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c client->add(tdev); tdev 111 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev; tdev 117 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c list_for_each_entry(tdev, &ofld_dev_list, ofld_dev_list) { tdev 118 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c if (offload_activated(tdev)) tdev 119 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c client->remove(tdev); tdev 133 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void cxgb3_add_clients(struct t3cdev *tdev) tdev 140 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c client->add(tdev); tdev 152 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void cxgb3_remove_clients(struct t3cdev *tdev) tdev 159 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c client->remove(tdev); tdev 164 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void cxgb3_event_notify(struct t3cdev *tdev, u32 event, u32 port) tdev 171 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c client->event_handler(tdev, event, port); tdev 373 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c static int cxgb_offload_ctl(struct t3cdev *tdev, unsigned int req, void *data) tdev 375 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct adapter *adapter = tdev2adap(tdev); tdev 501 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void *cxgb3_free_atid(struct t3cdev *tdev, int atid) tdev 503 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct tid_info *t = &(T3C_DATA(tdev))->tid_maps; tdev 521 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void cxgb3_free_stid(struct t3cdev *tdev, int stid) tdev 523 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct tid_info *t = &(T3C_DATA(tdev))->tid_maps; tdev 535 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void cxgb3_insert_tid(struct t3cdev *tdev, struct cxgb3_client *client, tdev 538 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct tid_info *t = &(T3C_DATA(tdev))->tid_maps; tdev 565 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev = td->dev; tdev 586 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c cxgb3_ofld_send(tdev, skb); tdev 604 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void cxgb3_queue_tid_release(struct t3cdev *tdev, unsigned int tid) tdev 606 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3c_data *td = T3C_DATA(tdev); tdev 627 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c void cxgb3_remove_tid(struct t3cdev *tdev, void *ctx, unsigned int tid) tdev 629 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct tid_info *t = &(T3C_DATA(tdev))->tid_maps; tdev 632 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c if (tdev->type == T3A) tdev 640 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c cxgb3_ofld_send(tdev, skb); tdev 643 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c cxgb3_queue_tid_release(tdev, tid); tdev 650 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c int cxgb3_alloc_atid(struct t3cdev *tdev, struct cxgb3_client *client, tdev 654 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct tid_info *t = &(T3C_DATA(tdev))->tid_maps; tdev 674 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c int cxgb3_alloc_stid(struct t3cdev *tdev, struct cxgb3_client *client, tdev 678 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct tid_info *t = &(T3C_DATA(tdev))->tid_maps; tdev 1081 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev = dev2t3cdev(dev); tdev 1083 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c BUG_ON(!tdev); tdev 1084 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c t3_l2t_update(tdev, neigh); tdev 1088 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c static void set_l2t_ix(struct t3cdev *tdev, u32 tid, struct l2t_entry *e) tdev 1107 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c tdev->send(tdev, skb); tdev 1116 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev; tdev 1126 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c tdev = dev2t3cdev(dev); tdev 1127 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c BUG_ON(!tdev); tdev 1130 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c e = t3_l2t_get(tdev, new, dev, daddr); tdev 1137 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c ti = &(T3C_DATA(tdev))->tid_maps; tdev 1145 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c l2t_hold(L2DATA(tdev), e); tdev 1147 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c set_l2t_ix(tdev, tid, e); tdev 1151 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c l2t_release(tdev, e); tdev 1219 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *dev = &adapter->tdev; tdev 1290 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev = &adapter->tdev; tdev 1291 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3c_data *t = T3C_DATA(tdev); tdev 1299 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c T3C_DATA(tdev) = NULL; tdev 1301 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c d = L2DATA(tdev); tdev 1303 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c RCU_INIT_POINTER(tdev->l2opt, NULL); tdev 1309 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c static inline void register_tdev(struct t3cdev *tdev) tdev 1314 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c snprintf(tdev->name, sizeof(tdev->name), "ofld_dev%d", unit++); tdev 1315 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c list_add_tail(&tdev->ofld_dev_list, &ofld_dev_list); tdev 1319 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c static inline void unregister_tdev(struct t3cdev *tdev) tdev 1322 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c list_del(&tdev->ofld_dev_list); tdev 1347 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev = &adapter->tdev; tdev 1349 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c INIT_LIST_HEAD(&tdev->ofld_dev_list); tdev 1351 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c cxgb3_set_dummy_ops(tdev); tdev 1352 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c tdev->send = t3_offload_tx; tdev 1353 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c tdev->ctl = cxgb_offload_ctl; tdev 1354 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c tdev->type = adap2type(adapter); tdev 1356 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c register_tdev(tdev); tdev 1361 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c struct t3cdev *tdev = &adapter->tdev; tdev 1363 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c tdev->recv = NULL; tdev 1364 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c tdev->neigh_update = NULL; tdev 1366 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c unregister_tdev(tdev); tdev 65 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h void cxgb3_add_clients(struct t3cdev *tdev); tdev 66 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h void cxgb3_remove_clients(struct t3cdev *tdev); tdev 67 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h void cxgb3_event_notify(struct t3cdev *tdev, u32 event, u32 port); tdev 90 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h void (*event_handler)(struct t3cdev *tdev, u32 event, u32 port); tdev 1832 drivers/net/ethernet/chelsio/cxgb3/sge.c int t3_offload_tx(struct t3cdev *tdev, struct sk_buff *skb) tdev 1834 drivers/net/ethernet/chelsio/cxgb3/sge.c struct adapter *adap = tdev2adap(tdev); tdev 1874 drivers/net/ethernet/chelsio/cxgb3/sge.c static inline void deliver_partial_bundle(struct t3cdev *tdev, tdev 1880 drivers/net/ethernet/chelsio/cxgb3/sge.c tdev->recv(tdev, skbs, n); tdev 1928 drivers/net/ethernet/chelsio/cxgb3/sge.c adapter->tdev.recv(&adapter->tdev, skbs, tdev 1939 drivers/net/ethernet/chelsio/cxgb3/sge.c deliver_partial_bundle(&adapter->tdev, q, skbs, ngathered); tdev 1956 drivers/net/ethernet/chelsio/cxgb3/sge.c static inline int rx_offload(struct t3cdev *tdev, struct sge_rspq *rq, tdev 1967 drivers/net/ethernet/chelsio/cxgb3/sge.c tdev->recv(tdev, rx_gather, RX_BUNDLE_SIZE); tdev 2437 drivers/net/ethernet/chelsio/cxgb3/sge.c ngathered = rx_offload(&adap->tdev, q, skb, tdev 2448 drivers/net/ethernet/chelsio/cxgb3/sge.c deliver_partial_bundle(&adap->tdev, q, offload_skbs, ngathered); tdev 1025 drivers/net/ethernet/neterion/s2io.c struct pci_dev *tdev = NULL; tdev 1026 drivers/net/ethernet/neterion/s2io.c for_each_pci_dev(tdev) { tdev 1027 drivers/net/ethernet/neterion/s2io.c if (tdev->vendor == NEC_VENID && tdev->device == NEC_DEVID) { tdev 1028 drivers/net/ethernet/neterion/s2io.c if (tdev->bus == s2io_pdev->bus->parent) { tdev 1029 drivers/net/ethernet/neterion/s2io.c pci_dev_put(tdev); tdev 1285 drivers/net/fddi/defza.c struct tc_dev *tdev = to_tc_dev(bdev); tdev 1312 drivers/net/fddi/defza.c start = tdev->resource.start; tdev 1313 drivers/net/fddi/defza.c len = tdev->resource.end - start + 1; tdev 1339 drivers/net/fddi/defza.c dev->irq = tdev->interrupt; tdev 1342 drivers/net/fddi/defza.c fp->name, (long long)tdev->resource.start, dev->irq); tdev 1517 drivers/net/fddi/defza.c struct tc_dev *tdev = to_tc_dev(bdev); tdev 1530 drivers/net/fddi/defza.c start = tdev->resource.start; tdev 1531 drivers/net/fddi/defza.c len = tdev->resource.end - start + 1; tdev 147 drivers/net/ppp/pptp.c struct net_device *tdev; tdev 162 drivers/net/ppp/pptp.c tdev = rt->dst.dev; tdev 164 drivers/net/ppp/pptp.c max_headroom = LL_RESERVED_SPACE(tdev) + sizeof(*iph) + sizeof(*hdr) + 2; tdev 1100 drivers/net/vxlan.c struct net_device *tdev; tdev 1105 drivers/net/vxlan.c tdev = __dev_get_by_index(net, *ifindex); tdev 1106 drivers/net/vxlan.c if (!tdev) tdev 322 drivers/pci/vpd.c struct pci_dev *tdev = pci_get_slot(dev->bus, tdev 326 drivers/pci/vpd.c if (!tdev) tdev 329 drivers/pci/vpd.c ret = pci_read_vpd(tdev, pos, count, arg); tdev 330 drivers/pci/vpd.c pci_dev_put(tdev); tdev 337 drivers/pci/vpd.c struct pci_dev *tdev = pci_get_slot(dev->bus, tdev 341 drivers/pci/vpd.c if (!tdev) tdev 344 drivers/pci/vpd.c ret = pci_write_vpd(tdev, pos, count, arg); tdev 345 drivers/pci/vpd.c pci_dev_put(tdev); tdev 351 drivers/pci/vpd.c struct pci_dev *tdev = pci_get_slot(dev->bus, tdev 355 drivers/pci/vpd.c if (!tdev) tdev 358 drivers/pci/vpd.c ret = pci_set_vpd_size(tdev, len); tdev 359 drivers/pci/vpd.c pci_dev_put(tdev); tdev 169 drivers/pnp/resource.c struct pnp_dev *tdev; tdev 207 drivers/pnp/resource.c pnp_for_each_dev(tdev) { tdev 208 drivers/pnp/resource.c if (tdev == dev) tdev 211 drivers/pnp/resource.c (tres = pnp_get_resource(tdev, IORESOURCE_IO, i)); tdev 232 drivers/pnp/resource.c struct pnp_dev *tdev; tdev 270 drivers/pnp/resource.c pnp_for_each_dev(tdev) { tdev 271 drivers/pnp/resource.c if (tdev == dev) tdev 274 drivers/pnp/resource.c (tres = pnp_get_resource(tdev, IORESOURCE_MEM, i)); tdev 355 drivers/pnp/resource.c struct pnp_dev *tdev; tdev 397 drivers/pnp/resource.c pnp_for_each_dev(tdev) { tdev 398 drivers/pnp/resource.c if (tdev == dev) tdev 401 drivers/pnp/resource.c (tres = pnp_get_resource(tdev, IORESOURCE_IRQ, i)); tdev 419 drivers/pnp/resource.c struct pnp_dev *tdev; tdev 456 drivers/pnp/resource.c pnp_for_each_dev(tdev) { tdev 457 drivers/pnp/resource.c if (tdev == dev) tdev 460 drivers/pnp/resource.c (tres = pnp_get_resource(tdev, IORESOURCE_DMA, i)); tdev 96 drivers/s390/char/tape_core.c struct tape_device *tdev; tdev 98 drivers/s390/char/tape_core.c tdev = dev_get_drvdata(dev); tdev 99 drivers/s390/char/tape_core.c return scnprintf(buf, PAGE_SIZE, "%i\n", tdev->medium_state); tdev 108 drivers/s390/char/tape_core.c struct tape_device *tdev; tdev 110 drivers/s390/char/tape_core.c tdev = dev_get_drvdata(dev); tdev 111 drivers/s390/char/tape_core.c return scnprintf(buf, PAGE_SIZE, "%i\n", tdev->first_minor); tdev 120 drivers/s390/char/tape_core.c struct tape_device *tdev; tdev 122 drivers/s390/char/tape_core.c tdev = dev_get_drvdata(dev); tdev 123 drivers/s390/char/tape_core.c return scnprintf(buf, PAGE_SIZE, "%s\n", (tdev->first_minor < 0) ? tdev 124 drivers/s390/char/tape_core.c "OFFLINE" : tape_state_verbose[tdev->tape_state]); tdev 133 drivers/s390/char/tape_core.c struct tape_device *tdev; tdev 136 drivers/s390/char/tape_core.c tdev = dev_get_drvdata(dev); tdev 137 drivers/s390/char/tape_core.c if (tdev->first_minor < 0) tdev 140 drivers/s390/char/tape_core.c spin_lock_irq(get_ccwdev_lock(tdev->cdev)); tdev 141 drivers/s390/char/tape_core.c if (list_empty(&tdev->req_queue)) tdev 146 drivers/s390/char/tape_core.c req = list_entry(tdev->req_queue.next, struct tape_request, tdev 150 drivers/s390/char/tape_core.c spin_unlock_irq(get_ccwdev_lock(tdev->cdev)); tdev 160 drivers/s390/char/tape_core.c struct tape_device *tdev; tdev 162 drivers/s390/char/tape_core.c tdev = dev_get_drvdata(dev); tdev 164 drivers/s390/char/tape_core.c return scnprintf(buf, PAGE_SIZE, "%i\n", tdev->char_data.block_size); tdev 227 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c static void abort_arp_failure(struct t3cdev *tdev, struct sk_buff *skb) tdev 233 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c tdev, GET_TID(req), skb); tdev 235 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c cxgb3_ofld_send(tdev, skb); tdev 475 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c static int do_act_establish(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 571 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c static int do_act_open_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) tdev 584 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c cxgb3_queue_tid_release(tdev, GET_TID(rpl)); tdev 1215 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c struct t3cdev *tdev = (struct t3cdev *)cdev->lldev; tdev 1222 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c err = tdev->ctl(tdev, ULP_ISCSI_GET_PARAMS, &uinfo); tdev 1246 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c err = cxgbi_ddp_ppm_setup(&tdev->ulp_iscsi, cdev, &tformat, tdev 1256 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c err = tdev->ctl(tdev, ULP_ISCSI_SET_PARAMS, &uinfo); tdev 60 drivers/tc/tc-driver.c struct tc_dev *tdev) tdev 66 drivers/tc/tc-driver.c if (strcmp(tdev->name, id->name) == 0 && tdev 67 drivers/tc/tc-driver.c strcmp(tdev->vendor, id->vendor) == 0) tdev 88 drivers/tc/tc-driver.c struct tc_dev *tdev = to_tc_dev(dev); tdev 92 drivers/tc/tc-driver.c id = tc_match_device(tdrv, tdev); tdev 42 drivers/tc/tc.c struct tc_dev *tdev; tdev 85 drivers/tc/tc.c tdev = kzalloc(sizeof(*tdev), GFP_KERNEL); tdev 86 drivers/tc/tc.c if (!tdev) { tdev 90 drivers/tc/tc.c dev_set_name(&tdev->dev, "tc%x", slot); tdev 91 drivers/tc/tc.c tdev->bus = tbus; tdev 92 drivers/tc/tc.c tdev->dev.parent = &tbus->dev; tdev 93 drivers/tc/tc.c tdev->dev.bus = &tc_bus_type; tdev 94 drivers/tc/tc.c tdev->slot = slot; tdev 97 drivers/tc/tc.c tdev->dma_mask = DMA_BIT_MASK(34); tdev 98 drivers/tc/tc.c tdev->dev.dma_mask = &tdev->dma_mask; tdev 99 drivers/tc/tc.c tdev->dev.coherent_dma_mask = DMA_BIT_MASK(34); tdev 102 drivers/tc/tc.c tdev->firmware[i] = tdev 104 drivers/tc/tc.c tdev->vendor[i] = tdev 106 drivers/tc/tc.c tdev->name[i] = tdev 109 drivers/tc/tc.c tdev->firmware[8] = 0; tdev 110 drivers/tc/tc.c tdev->vendor[8] = 0; tdev 111 drivers/tc/tc.c tdev->name[8] = 0; tdev 113 drivers/tc/tc.c pr_info("%s: %s %s %s\n", dev_name(&tdev->dev), tdev->vendor, tdev 114 drivers/tc/tc.c tdev->name, tdev->firmware); tdev 119 drivers/tc/tc.c tdev->resource.start = slotaddr; tdev 120 drivers/tc/tc.c tdev->resource.end = slotaddr + devsize - 1; tdev 122 drivers/tc/tc.c tdev->resource.start = extslotaddr; tdev 123 drivers/tc/tc.c tdev->resource.end = extslotaddr + devsize - 1; tdev 127 drivers/tc/tc.c dev_name(&tdev->dev), (long)(devsize >> 20), tdev 129 drivers/tc/tc.c kfree(tdev); tdev 132 drivers/tc/tc.c tdev->resource.name = tdev->name; tdev 133 drivers/tc/tc.c tdev->resource.flags = IORESOURCE_MEM; tdev 135 drivers/tc/tc.c tc_device_get_irq(tdev); tdev 137 drivers/tc/tc.c if (device_register(&tdev->dev)) { tdev 138 drivers/tc/tc.c put_device(&tdev->dev); tdev 141 drivers/tc/tc.c list_add_tail(&tdev->node, &tbus->devices); tdev 91 drivers/thermal/uniphier_thermal.c static int uniphier_tm_initialize_sensor(struct uniphier_tm_dev *tdev) tdev 93 drivers/thermal/uniphier_thermal.c struct regmap *map = tdev->regmap; tdev 99 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->block_base + PVTCTLEN, tdev 109 drivers/thermal/uniphier_thermal.c ret = regmap_read(map, tdev->data->map_base + TMODCOEF, &val); tdev 114 drivers/thermal/uniphier_thermal.c ret = of_property_read_u32_array(tdev->dev->of_node, tdev 121 drivers/thermal/uniphier_thermal.c regmap_write(map, tdev->data->tmod_setup_addr, tdev 127 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->block_base + PVTCTLMODE, tdev 131 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->block_base + EMONREPEAT, tdev 136 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->map_base + PVTCTLSEL, tdev 142 drivers/thermal/uniphier_thermal.c static void uniphier_tm_set_alert(struct uniphier_tm_dev *tdev, u32 ch, tdev 145 drivers/thermal/uniphier_thermal.c struct regmap *map = tdev->regmap; tdev 148 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->map_base + SETALERT0 + (ch << 2), tdev 154 drivers/thermal/uniphier_thermal.c static void uniphier_tm_enable_sensor(struct uniphier_tm_dev *tdev) tdev 156 drivers/thermal/uniphier_thermal.c struct regmap *map = tdev->regmap; tdev 161 drivers/thermal/uniphier_thermal.c if (tdev->alert_en[i]) tdev 165 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->map_base + PMALERTINTCTL, tdev 169 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->block_base + PVTCTLEN, tdev 175 drivers/thermal/uniphier_thermal.c static void uniphier_tm_disable_sensor(struct uniphier_tm_dev *tdev) tdev 177 drivers/thermal/uniphier_thermal.c struct regmap *map = tdev->regmap; tdev 180 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->map_base + PMALERTINTCTL, tdev 184 drivers/thermal/uniphier_thermal.c regmap_write_bits(map, tdev->data->block_base + PVTCTLEN, tdev 192 drivers/thermal/uniphier_thermal.c struct uniphier_tm_dev *tdev = data; tdev 193 drivers/thermal/uniphier_thermal.c struct regmap *map = tdev->regmap; tdev 197 drivers/thermal/uniphier_thermal.c ret = regmap_read(map, tdev->data->map_base + TMOD, &temp); tdev 211 drivers/thermal/uniphier_thermal.c static void uniphier_tm_irq_clear(struct uniphier_tm_dev *tdev) tdev 222 drivers/thermal/uniphier_thermal.c regmap_write_bits(tdev->regmap, tdev 223 drivers/thermal/uniphier_thermal.c tdev->data->map_base + PMALERTINTCTL, mask, bits); tdev 228 drivers/thermal/uniphier_thermal.c struct uniphier_tm_dev *tdev = _tdev; tdev 231 drivers/thermal/uniphier_thermal.c uniphier_tm_irq_clear(tdev); tdev 238 drivers/thermal/uniphier_thermal.c struct uniphier_tm_dev *tdev = _tdev; tdev 240 drivers/thermal/uniphier_thermal.c thermal_zone_device_update(tdev->tz_dev, THERMAL_EVENT_UNSPECIFIED); tdev 250 drivers/thermal/uniphier_thermal.c struct uniphier_tm_dev *tdev; tdev 254 drivers/thermal/uniphier_thermal.c tdev = devm_kzalloc(dev, sizeof(*tdev), GFP_KERNEL); tdev 255 drivers/thermal/uniphier_thermal.c if (!tdev) tdev 257 drivers/thermal/uniphier_thermal.c tdev->dev = dev; tdev 259 drivers/thermal/uniphier_thermal.c tdev->data = of_device_get_match_data(dev); tdev 260 drivers/thermal/uniphier_thermal.c if (WARN_ON(!tdev->data)) tdev 276 drivers/thermal/uniphier_thermal.c tdev->regmap = regmap; tdev 278 drivers/thermal/uniphier_thermal.c ret = uniphier_tm_initialize_sensor(tdev); tdev 286 drivers/thermal/uniphier_thermal.c 0, "thermal", tdev); tdev 290 drivers/thermal/uniphier_thermal.c platform_set_drvdata(pdev, tdev); tdev 292 drivers/thermal/uniphier_thermal.c tdev->tz_dev = devm_thermal_zone_of_sensor_register(dev, 0, tdev, tdev 294 drivers/thermal/uniphier_thermal.c if (IS_ERR(tdev->tz_dev)) { tdev 296 drivers/thermal/uniphier_thermal.c return PTR_ERR(tdev->tz_dev); tdev 300 drivers/thermal/uniphier_thermal.c trips = of_thermal_get_trip_points(tdev->tz_dev); tdev 301 drivers/thermal/uniphier_thermal.c ntrips = of_thermal_get_ntrips(tdev->tz_dev); tdev 312 drivers/thermal/uniphier_thermal.c uniphier_tm_set_alert(tdev, i, trips[i].temperature); tdev 313 drivers/thermal/uniphier_thermal.c tdev->alert_en[i] = true; tdev 321 drivers/thermal/uniphier_thermal.c uniphier_tm_enable_sensor(tdev); tdev 328 drivers/thermal/uniphier_thermal.c struct uniphier_tm_dev *tdev = platform_get_drvdata(pdev); tdev 331 drivers/thermal/uniphier_thermal.c uniphier_tm_disable_sensor(tdev); tdev 118 drivers/usb/misc/usbtest.c #define ERROR(tdev, fmt, args...) \ tdev 119 drivers/usb/misc/usbtest.c dev_err(&(tdev)->intf->dev , fmt , ## args) tdev 120 drivers/usb/misc/usbtest.c #define WARNING(tdev, fmt, args...) \ tdev 121 drivers/usb/misc/usbtest.c dev_warn(&(tdev)->intf->dev , fmt , ## args) tdev 384 drivers/usb/misc/usbtest.c static int check_guard_bytes(struct usbtest_dev *tdev, struct urb *urb) tdev 392 drivers/usb/misc/usbtest.c ERROR(tdev, "guard byte[%d] %d (not %d)\n", tdev 400 drivers/usb/misc/usbtest.c static int simple_check_buf(struct usbtest_dev *tdev, struct urb *urb) tdev 408 drivers/usb/misc/usbtest.c int ret = check_guard_bytes(tdev, urb); tdev 433 drivers/usb/misc/usbtest.c ERROR(tdev, "buf[%d] = %d (not %d)\n", i, *buf, expected); tdev 455 drivers/usb/misc/usbtest.c struct usbtest_dev *tdev, tdev 491 drivers/usb/misc/usbtest.c retval = simple_check_buf(tdev, urb); tdev 601 drivers/usb/misc/usbtest.c struct usbtest_dev *tdev, tdev 609 drivers/usb/misc/usbtest.c struct usb_device *udev = testdev_to_usbdev(tdev); tdev 644 drivers/usb/misc/usbtest.c ERROR(tdev, "perform_sglist failed, " tdev 704 drivers/usb/misc/usbtest.c static int is_good_config(struct usbtest_dev *tdev, int len) tdev 710 drivers/usb/misc/usbtest.c config = (struct usb_config_descriptor *) tdev->buf; tdev 716 drivers/usb/misc/usbtest.c ERROR(tdev, "bogus config descriptor length\n"); tdev 721 drivers/usb/misc/usbtest.c ERROR(tdev, "high bit of config attributes not set\n"); tdev 725 drivers/usb/misc/usbtest.c ERROR(tdev, "reserved config bits set\n"); tdev 737 drivers/usb/misc/usbtest.c ERROR(tdev, "bogus config descriptor read size\n"); tdev 741 drivers/usb/misc/usbtest.c static int is_good_ext(struct usbtest_dev *tdev, u8 *buf) tdev 749 drivers/usb/misc/usbtest.c ERROR(tdev, "bogus usb 2.0 extension descriptor length\n"); tdev 756 drivers/usb/misc/usbtest.c ERROR(tdev, "reserved bits set\n"); tdev 763 drivers/usb/misc/usbtest.c static int is_good_ss_cap(struct usbtest_dev *tdev, u8 *buf) tdev 770 drivers/usb/misc/usbtest.c ERROR(tdev, "bogus superspeed device capability descriptor length\n"); tdev 779 drivers/usb/misc/usbtest.c ERROR(tdev, "reserved bits set in bmAttributes\n"); tdev 785 drivers/usb/misc/usbtest.c ERROR(tdev, "reserved bits set in wSpeedSupported\n"); tdev 792 drivers/usb/misc/usbtest.c static int is_good_con_id(struct usbtest_dev *tdev, u8 *buf) tdev 799 drivers/usb/misc/usbtest.c ERROR(tdev, "bogus container id descriptor length\n"); tdev 804 drivers/usb/misc/usbtest.c ERROR(tdev, "reserved bits set\n"); tdev 1629 drivers/usb/misc/usbtest.c static int verify_not_halted(struct usbtest_dev *tdev, int ep, struct urb *urb) tdev 1637 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x couldn't get no-halt status, %d\n", tdev 1642 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x bogus status: %04x != 0\n", ep, status); tdev 1645 drivers/usb/misc/usbtest.c retval = simple_io(tdev, urb, 1, 0, 0, __func__); tdev 1651 drivers/usb/misc/usbtest.c static int verify_halted(struct usbtest_dev *tdev, int ep, struct urb *urb) tdev 1659 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x couldn't get halt status, %d\n", tdev 1664 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x bogus status: %04x != 1\n", ep, status); tdev 1667 drivers/usb/misc/usbtest.c retval = simple_io(tdev, urb, 1, 0, -EPIPE, __func__); tdev 1670 drivers/usb/misc/usbtest.c retval = simple_io(tdev, urb, 1, 0, -EPIPE, "verify_still_halted"); tdev 1676 drivers/usb/misc/usbtest.c static int test_halt(struct usbtest_dev *tdev, int ep, struct urb *urb) tdev 1681 drivers/usb/misc/usbtest.c retval = verify_not_halted(tdev, ep, urb); tdev 1691 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x couldn't set halt, %d\n", ep, retval); tdev 1694 drivers/usb/misc/usbtest.c retval = verify_halted(tdev, ep, urb); tdev 1701 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x couldn't clear halt, %d\n", tdev 1710 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x couldn't clear halt, %d\n", ep, retval); tdev 1713 drivers/usb/misc/usbtest.c retval = verify_not_halted(tdev, ep, urb); tdev 1722 drivers/usb/misc/usbtest.c static int test_toggle_sync(struct usbtest_dev *tdev, int ep, struct urb *urb) tdev 1729 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x couldn't clear halt, %d\n", ep, retval); tdev 1734 drivers/usb/misc/usbtest.c retval = simple_io(tdev, urb, 1, 0, 0, __func__); tdev 1741 drivers/usb/misc/usbtest.c ERROR(tdev, "ep %02x couldn't clear halt, %d\n", ep, retval); tdev 1746 drivers/usb/misc/usbtest.c retval = simple_io(tdev, urb, 1, 0, 0, __func__); tdev 2151 drivers/usb/misc/usbtest.c struct usbtest_dev *tdev, tdev 2159 drivers/usb/misc/usbtest.c struct urb *urb = usbtest_alloc_urb(testdev_to_usbdev(tdev), tdev 2165 drivers/usb/misc/usbtest.c retval = simple_io(tdev, urb, iterations, 0, 0, label); tdev 340 drivers/usb/serial/ti_usb_3410_5052.c static int ti_command_out_sync(struct ti_device *tdev, __u8 command, tdev 342 drivers/usb/serial/ti_usb_3410_5052.c static int ti_command_in_sync(struct ti_device *tdev, __u8 command, tdev 345 drivers/usb/serial/ti_usb_3410_5052.c static int ti_write_byte(struct usb_serial_port *port, struct ti_device *tdev, tdev 348 drivers/usb/serial/ti_usb_3410_5052.c static int ti_download_firmware(struct ti_device *tdev); tdev 515 drivers/usb/serial/ti_usb_3410_5052.c struct ti_device *tdev; tdev 528 drivers/usb/serial/ti_usb_3410_5052.c tdev = kzalloc(sizeof(struct ti_device), GFP_KERNEL); tdev 529 drivers/usb/serial/ti_usb_3410_5052.c if (!tdev) tdev 532 drivers/usb/serial/ti_usb_3410_5052.c mutex_init(&tdev->td_open_close_lock); tdev 533 drivers/usb/serial/ti_usb_3410_5052.c tdev->td_serial = serial; tdev 534 drivers/usb/serial/ti_usb_3410_5052.c usb_set_serial_data(serial, tdev); tdev 538 drivers/usb/serial/ti_usb_3410_5052.c tdev->td_is_3410 = 1; tdev 540 drivers/usb/serial/ti_usb_3410_5052.c tdev->td_is_3410 ? "3410" : "5052"); tdev 548 drivers/usb/serial/ti_usb_3410_5052.c tdev->td_rs485_only = true; tdev 558 drivers/usb/serial/ti_usb_3410_5052.c status = ti_download_firmware(tdev); tdev 564 drivers/usb/serial/ti_usb_3410_5052.c if (tdev->td_is_3410) { tdev 590 drivers/usb/serial/ti_usb_3410_5052.c kfree(tdev); tdev 598 drivers/usb/serial/ti_usb_3410_5052.c struct ti_device *tdev = usb_get_serial_data(serial); tdev 600 drivers/usb/serial/ti_usb_3410_5052.c kfree(tdev); tdev 645 drivers/usb/serial/ti_usb_3410_5052.c struct ti_device *tdev; tdev 657 drivers/usb/serial/ti_usb_3410_5052.c tdev = tport->tp_tdev; tdev 660 drivers/usb/serial/ti_usb_3410_5052.c if (mutex_lock_interruptible(&tdev->td_open_close_lock)) tdev 669 drivers/usb/serial/ti_usb_3410_5052.c if (tdev->td_open_port_count == 0) { tdev 671 drivers/usb/serial/ti_usb_3410_5052.c urb = tdev->td_serial->port[0]->interrupt_in_urb; tdev 677 drivers/usb/serial/ti_usb_3410_5052.c urb->context = tdev; tdev 688 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_OPEN_PORT, tdev 696 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_START_PORT, tdev 704 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_PURGE_PORT, tdev 711 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_PURGE_PORT, tdev 727 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_OPEN_PORT, tdev 735 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_START_PORT, tdev 760 drivers/usb/serial/ti_usb_3410_5052.c ++tdev->td_open_port_count; tdev 765 drivers/usb/serial/ti_usb_3410_5052.c if (tdev->td_open_port_count == 0) tdev 768 drivers/usb/serial/ti_usb_3410_5052.c mutex_unlock(&tdev->td_open_close_lock); tdev 775 drivers/usb/serial/ti_usb_3410_5052.c struct ti_device *tdev; tdev 781 drivers/usb/serial/ti_usb_3410_5052.c tdev = usb_get_serial_data(port->serial); tdev 795 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_CLOSE_PORT, tdev 802 drivers/usb/serial/ti_usb_3410_5052.c mutex_lock(&tdev->td_open_close_lock); tdev 803 drivers/usb/serial/ti_usb_3410_5052.c --tdev->td_open_port_count; tdev 804 drivers/usb/serial/ti_usb_3410_5052.c if (tdev->td_open_port_count == 0) { tdev 808 drivers/usb/serial/ti_usb_3410_5052.c mutex_unlock(&tdev->td_open_close_lock); tdev 1110 drivers/usb/serial/ti_usb_3410_5052.c struct ti_device *tdev = urb->context; tdev 1112 drivers/usb/serial/ti_usb_3410_5052.c struct usb_serial *serial = tdev->td_serial; tdev 1369 drivers/usb/serial/ti_usb_3410_5052.c struct ti_device *tdev = tport->tp_tdev; tdev 1379 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_in_sync(tdev, TI_GET_PORT_STATUS, tdev 1507 drivers/usb/serial/ti_usb_3410_5052.c static int ti_command_out_sync(struct ti_device *tdev, __u8 command, tdev 1512 drivers/usb/serial/ti_usb_3410_5052.c status = usb_control_msg(tdev->td_serial->dev, tdev 1513 drivers/usb/serial/ti_usb_3410_5052.c usb_sndctrlpipe(tdev->td_serial->dev, 0), command, tdev 1524 drivers/usb/serial/ti_usb_3410_5052.c static int ti_command_in_sync(struct ti_device *tdev, __u8 command, tdev 1529 drivers/usb/serial/ti_usb_3410_5052.c status = usb_control_msg(tdev->td_serial->dev, tdev 1530 drivers/usb/serial/ti_usb_3410_5052.c usb_rcvctrlpipe(tdev->td_serial->dev, 0), command, tdev 1544 drivers/usb/serial/ti_usb_3410_5052.c struct ti_device *tdev, unsigned long addr, tdev 1567 drivers/usb/serial/ti_usb_3410_5052.c status = ti_command_out_sync(tdev, TI_WRITE_DATA, TI_RAM_PORT, 0, tdev 1606 drivers/usb/serial/ti_usb_3410_5052.c static int ti_download_firmware(struct ti_device *tdev) tdev 1611 drivers/usb/serial/ti_usb_3410_5052.c struct usb_device *dev = tdev->td_serial->dev; tdev 1613 drivers/usb/serial/ti_usb_3410_5052.c tdev->td_serial->port[0]->bulk_out_endpointAddress); tdev 1657 drivers/usb/serial/ti_usb_3410_5052.c if (tdev->td_is_3410) tdev 161 drivers/video/fbdev/pmag-aa-fb.c struct tc_dev *tdev = to_tc_dev(dev); tdev 180 drivers/video/fbdev/pmag-aa-fb.c start = tdev->resource.start; tdev 181 drivers/video/fbdev/pmag-aa-fb.c len = tdev->resource.end - start + 1; tdev 250 drivers/video/fbdev/pmag-aa-fb.c struct tc_dev *tdev = to_tc_dev(dev); tdev 259 drivers/video/fbdev/pmag-aa-fb.c start = tdev->resource.start; tdev 260 drivers/video/fbdev/pmag-aa-fb.c len = tdev->resource.end - start + 1; tdev 146 drivers/video/fbdev/pmag-ba-fb.c struct tc_dev *tdev = to_tc_dev(dev); tdev 172 drivers/video/fbdev/pmag-ba-fb.c start = tdev->resource.start; tdev 173 drivers/video/fbdev/pmag-ba-fb.c len = tdev->resource.end - start + 1; tdev 238 drivers/video/fbdev/pmag-ba-fb.c struct tc_dev *tdev = to_tc_dev(dev); tdev 247 drivers/video/fbdev/pmag-ba-fb.c start = tdev->resource.start; tdev 248 drivers/video/fbdev/pmag-ba-fb.c len = tdev->resource.end - start + 1; tdev 251 drivers/video/fbdev/pmagb-b-fb.c struct tc_dev *tdev = to_tc_dev(dev); tdev 279 drivers/video/fbdev/pmagb-b-fb.c start = tdev->resource.start; tdev 280 drivers/video/fbdev/pmagb-b-fb.c len = tdev->resource.end - start + 1; tdev 356 drivers/video/fbdev/pmagb-b-fb.c struct tc_dev *tdev = to_tc_dev(dev); tdev 365 drivers/video/fbdev/pmagb-b-fb.c start = tdev->resource.start; tdev 366 drivers/video/fbdev/pmagb-b-fb.c len = tdev->resource.end - start + 1; tdev 683 include/linux/libata.h struct device tdev; tdev 783 include/linux/libata.h struct device tdev; tdev 853 include/linux/libata.h struct device tdev; tdev 140 include/linux/tc.h extern void tc_device_get_irq(struct tc_dev *tdev); tdev 279 net/ipv4/ip_tunnel.c struct net_device *tdev = NULL; tdev 300 net/ipv4/ip_tunnel.c tdev = rt->dst.dev; tdev 309 net/ipv4/ip_tunnel.c if (!tdev && tunnel->parms.link) tdev 310 net/ipv4/ip_tunnel.c tdev = __dev_get_by_index(tunnel->net, tunnel->parms.link); tdev 312 net/ipv4/ip_tunnel.c if (tdev) { tdev 313 net/ipv4/ip_tunnel.c hlen = tdev->hard_header_len + tdev->needed_headroom; tdev 314 net/ipv4/ip_tunnel.c mtu = min(tdev->mtu, IP_MAX_MTU); tdev 205 net/ipv4/ip_vti.c struct net_device *tdev; /* Device to other host */ tdev 259 net/ipv4/ip_vti.c tdev = dst->dev; tdev 261 net/ipv4/ip_vti.c if (tdev == dev) { tdev 1042 net/ipv6/ip6_tunnel.c struct net_device *tdev; tdev 1119 net/ipv6/ip6_tunnel.c tdev = dst->dev; tdev 1121 net/ipv6/ip6_tunnel.c if (tdev == dev) { tdev 1158 net/ipv6/ip6_tunnel.c max_headroom += LL_RESERVED_SPACE(tdev); tdev 446 net/ipv6/ip6_vti.c struct net_device *tdev; tdev 498 net/ipv6/ip6_vti.c tdev = dst->dev; tdev 500 net/ipv6/ip6_vti.c if (tdev == dev) { tdev 649 net/ipv6/ip6_vti.c struct net_device *tdev = NULL; tdev 677 net/ipv6/ip6_vti.c tdev = rt->dst.dev; tdev 681 net/ipv6/ip6_vti.c if (!tdev && p->link) tdev 682 net/ipv6/ip6_vti.c tdev = __dev_get_by_index(t->net, p->link); tdev 684 net/ipv6/ip6_vti.c if (tdev) tdev 685 net/ipv6/ip6_vti.c mtu = tdev->mtu - sizeof(struct ipv6hdr); tdev 825 net/ipv6/sit.c struct net_device *tdev; /* Device to other host */ tdev 919 net/ipv6/sit.c tdev = rt->dst.dev; tdev 921 net/ipv6/sit.c if (tdev == dev) { tdev 968 net/ipv6/sit.c max_headroom = LL_RESERVED_SPACE(tdev) + t_hlen; tdev 1061 net/ipv6/sit.c struct net_device *tdev = NULL; tdev 1079 net/ipv6/sit.c tdev = rt->dst.dev; tdev 1085 net/ipv6/sit.c if (!tdev && tunnel->parms.link) tdev 1086 net/ipv6/sit.c tdev = __dev_get_by_index(tunnel->net, tunnel->parms.link); tdev 1088 net/ipv6/sit.c if (tdev && !netif_is_l3_master(tdev)) { tdev 1091 net/ipv6/sit.c dev->hard_header_len = tdev->hard_header_len + sizeof(struct iphdr); tdev 1092 net/ipv6/sit.c dev->mtu = tdev->mtu - t_hlen; tdev 1140 net/netfilter/ipvs/ip_vs_xmit.c struct net_device *tdev; /* Device to other host */ tdev 1165 net/netfilter/ipvs/ip_vs_xmit.c tdev = rt->dst.dev; tdev 1170 net/netfilter/ipvs/ip_vs_xmit.c max_headroom = LL_RESERVED_SPACE(tdev) + sizeof(struct iphdr); tdev 1290 net/netfilter/ipvs/ip_vs_xmit.c struct net_device *tdev; /* Device to other host */ tdev 1315 net/netfilter/ipvs/ip_vs_xmit.c tdev = rt->dst.dev; tdev 1320 net/netfilter/ipvs/ip_vs_xmit.c max_headroom = LL_RESERVED_SPACE(tdev) + sizeof(struct ipv6hdr); tdev 266 net/xfrm/xfrm_interface.c struct net_device *tdev; tdev 286 net/xfrm/xfrm_interface.c tdev = dst->dev; tdev 288 net/xfrm/xfrm_interface.c if (tdev == dev) { tdev 315 net/xfrm/xfrm_interface.c skb->dev = tdev;