udl 21 drivers/gpu/drm/udl/udl_connector.c struct udl_device *udl = data; udl 29 drivers/gpu/drm/udl/udl_connector.c ret = usb_control_msg(udl->udev, udl 30 drivers/gpu/drm/udl/udl_connector.c usb_rcvctrlpipe(udl->udev, 0), udl 61 drivers/gpu/drm/udl/udl_connector.c struct udl_device *udl = connector->dev->dev_private; udl 62 drivers/gpu/drm/udl/udl_connector.c if (!udl->sku_pixel_limit) udl 65 drivers/gpu/drm/udl/udl_connector.c if (mode->vdisplay * mode->hdisplay > udl->sku_pixel_limit) udl 74 drivers/gpu/drm/udl/udl_connector.c struct udl_device *udl = connector->dev->dev_private; udl 86 drivers/gpu/drm/udl/udl_connector.c udl_connector->edid = drm_do_get_edid(connector, udl_get_edid_block, udl); udl 68 drivers/gpu/drm/udl/udl_dmabuf.c struct udl_device *udl = dev->dev_private; udl 105 drivers/gpu/drm/udl/udl_dmabuf.c mutex_lock(&udl->gem_lock); udl 130 drivers/gpu/drm/udl/udl_dmabuf.c mutex_unlock(&udl->gem_lock); udl 89 drivers/gpu/drm/udl/udl_drv.c struct udl_device *udl; udl 92 drivers/gpu/drm/udl/udl_drv.c udl = kzalloc(sizeof(*udl), GFP_KERNEL); udl 93 drivers/gpu/drm/udl/udl_drv.c if (!udl) udl 96 drivers/gpu/drm/udl/udl_drv.c r = drm_dev_init(&udl->drm, &driver, &interface->dev); udl 98 drivers/gpu/drm/udl/udl_drv.c kfree(udl); udl 102 drivers/gpu/drm/udl/udl_drv.c udl->udev = udev; udl 103 drivers/gpu/drm/udl/udl_drv.c udl->drm.dev_private = udl; udl 105 drivers/gpu/drm/udl/udl_drv.c r = udl_init(udl); udl 107 drivers/gpu/drm/udl/udl_drv.c drm_dev_fini(&udl->drm); udl 108 drivers/gpu/drm/udl/udl_drv.c kfree(udl); udl 112 drivers/gpu/drm/udl/udl_drv.c usb_set_intfdata(interface, udl); udl 113 drivers/gpu/drm/udl/udl_drv.c return udl; udl 120 drivers/gpu/drm/udl/udl_drv.c struct udl_device *udl; udl 122 drivers/gpu/drm/udl/udl_drv.c udl = udl_driver_create(interface); udl 123 drivers/gpu/drm/udl/udl_drv.c if (IS_ERR(udl)) udl 124 drivers/gpu/drm/udl/udl_drv.c return PTR_ERR(udl); udl 126 drivers/gpu/drm/udl/udl_drv.c r = drm_dev_register(&udl->drm, 0); udl 130 drivers/gpu/drm/udl/udl_drv.c DRM_INFO("Initialized udl on minor %d\n", udl->drm.primary->index); udl 135 drivers/gpu/drm/udl/udl_drv.c drm_dev_put(&udl->drm); udl 110 drivers/gpu/drm/udl/udl_drv.h int udl_init(struct udl_device *udl); udl 81 drivers/gpu/drm/udl/udl_fb.c struct udl_device *udl = to_udl(dev); udl 149 drivers/gpu/drm/udl/udl_fb.c atomic_add(bytes_sent, &udl->bytes_sent); udl 150 drivers/gpu/drm/udl/udl_fb.c atomic_add(bytes_identical, &udl->bytes_identical); udl 151 drivers/gpu/drm/udl/udl_fb.c atomic_add((width * height) << log_bpp, &udl->bytes_rendered); udl 155 drivers/gpu/drm/udl/udl_fb.c &udl->cpu_kcycles_used); udl 209 drivers/gpu/drm/udl/udl_fb.c struct udl_device *udl = to_udl(dev); udl 212 drivers/gpu/drm/udl/udl_fb.c if (drm_dev_is_unplugged(&udl->drm)) udl 436 drivers/gpu/drm/udl/udl_fb.c struct udl_device *udl = to_udl(dev); udl 445 drivers/gpu/drm/udl/udl_fb.c udl->fbdev = ufbdev; udl 475 drivers/gpu/drm/udl/udl_fb.c struct udl_device *udl = to_udl(dev); udl 476 drivers/gpu/drm/udl/udl_fb.c if (!udl->fbdev) udl 479 drivers/gpu/drm/udl/udl_fb.c udl_fbdev_destroy(dev, udl->fbdev); udl 480 drivers/gpu/drm/udl/udl_fb.c kfree(udl->fbdev); udl 481 drivers/gpu/drm/udl/udl_fb.c udl->fbdev = NULL; udl 486 drivers/gpu/drm/udl/udl_fb.c struct udl_device *udl = to_udl(dev); udl 488 drivers/gpu/drm/udl/udl_fb.c if (!udl->fbdev) udl 491 drivers/gpu/drm/udl/udl_fb.c ufbdev = udl->fbdev; udl 206 drivers/gpu/drm/udl/udl_gem.c struct udl_device *udl = to_udl(dev); udl 209 drivers/gpu/drm/udl/udl_gem.c mutex_lock(&udl->gem_lock); udl 229 drivers/gpu/drm/udl/udl_gem.c mutex_unlock(&udl->gem_lock); udl 32 drivers/gpu/drm/udl/udl_main.c struct udl_device *udl = to_udl(dev); udl 75 drivers/gpu/drm/udl/udl_main.c udl->sku_pixel_limit = max_area; udl 99 drivers/gpu/drm/udl/udl_main.c static int udl_select_std_channel(struct udl_device *udl) udl 112 drivers/gpu/drm/udl/udl_main.c ret = usb_control_msg(udl->udev, udl 113 drivers/gpu/drm/udl/udl_main.c usb_sndctrlpipe(udl->udev, 0), udl 133 drivers/gpu/drm/udl/udl_main.c struct udl_device *udl = unode->dev; udl 143 drivers/gpu/drm/udl/udl_main.c atomic_set(&udl->lost_pixels, 1); udl 147 drivers/gpu/drm/udl/udl_main.c urb->transfer_buffer_length = udl->urbs.size; /* reset to actual */ udl 149 drivers/gpu/drm/udl/udl_main.c spin_lock_irqsave(&udl->urbs.lock, flags); udl 150 drivers/gpu/drm/udl/udl_main.c list_add_tail(&unode->entry, &udl->urbs.list); udl 151 drivers/gpu/drm/udl/udl_main.c udl->urbs.available++; udl 152 drivers/gpu/drm/udl/udl_main.c spin_unlock_irqrestore(&udl->urbs.lock, flags); udl 163 drivers/gpu/drm/udl/udl_main.c up(&udl->urbs.limit_sem); udl 168 drivers/gpu/drm/udl/udl_main.c struct udl_device *udl = to_udl(dev); udl 169 drivers/gpu/drm/udl/udl_main.c int count = udl->urbs.count; udl 178 drivers/gpu/drm/udl/udl_main.c down(&udl->urbs.limit_sem); udl 180 drivers/gpu/drm/udl/udl_main.c spin_lock_irq(&udl->urbs.lock); udl 182 drivers/gpu/drm/udl/udl_main.c node = udl->urbs.list.next; /* have reserved one with sem */ udl 185 drivers/gpu/drm/udl/udl_main.c spin_unlock_irq(&udl->urbs.lock); udl 191 drivers/gpu/drm/udl/udl_main.c usb_free_coherent(urb->dev, udl->urbs.size, udl 196 drivers/gpu/drm/udl/udl_main.c udl->urbs.count = 0; udl 201 drivers/gpu/drm/udl/udl_main.c struct udl_device *udl = to_udl(dev); udl 207 drivers/gpu/drm/udl/udl_main.c spin_lock_init(&udl->urbs.lock); udl 210 drivers/gpu/drm/udl/udl_main.c udl->urbs.size = size; udl 211 drivers/gpu/drm/udl/udl_main.c INIT_LIST_HEAD(&udl->urbs.list); udl 213 drivers/gpu/drm/udl/udl_main.c sema_init(&udl->urbs.limit_sem, 0); udl 214 drivers/gpu/drm/udl/udl_main.c udl->urbs.count = 0; udl 215 drivers/gpu/drm/udl/udl_main.c udl->urbs.available = 0; udl 217 drivers/gpu/drm/udl/udl_main.c while (udl->urbs.count * size < wanted_size) { udl 221 drivers/gpu/drm/udl/udl_main.c unode->dev = udl; udl 233 drivers/gpu/drm/udl/udl_main.c buf = usb_alloc_coherent(udl->udev, size, GFP_KERNEL, udl 247 drivers/gpu/drm/udl/udl_main.c usb_fill_bulk_urb(urb, udl->udev, usb_sndbulkpipe(udl->udev, 1), udl 251 drivers/gpu/drm/udl/udl_main.c list_add_tail(&unode->entry, &udl->urbs.list); udl 253 drivers/gpu/drm/udl/udl_main.c up(&udl->urbs.limit_sem); udl 254 drivers/gpu/drm/udl/udl_main.c udl->urbs.count++; udl 255 drivers/gpu/drm/udl/udl_main.c udl->urbs.available++; udl 258 drivers/gpu/drm/udl/udl_main.c DRM_DEBUG("allocated %d %d byte urbs\n", udl->urbs.count, (int) size); udl 260 drivers/gpu/drm/udl/udl_main.c return udl->urbs.count; udl 265 drivers/gpu/drm/udl/udl_main.c struct udl_device *udl = to_udl(dev); udl 272 drivers/gpu/drm/udl/udl_main.c ret = down_timeout(&udl->urbs.limit_sem, GET_URB_TIMEOUT); udl 274 drivers/gpu/drm/udl/udl_main.c atomic_set(&udl->lost_pixels, 1); udl 276 drivers/gpu/drm/udl/udl_main.c ret, udl->urbs.available); udl 280 drivers/gpu/drm/udl/udl_main.c spin_lock_irq(&udl->urbs.lock); udl 282 drivers/gpu/drm/udl/udl_main.c BUG_ON(list_empty(&udl->urbs.list)); /* reserved one with limit_sem */ udl 283 drivers/gpu/drm/udl/udl_main.c entry = udl->urbs.list.next; udl 285 drivers/gpu/drm/udl/udl_main.c udl->urbs.available--; udl 287 drivers/gpu/drm/udl/udl_main.c spin_unlock_irq(&udl->urbs.lock); udl 298 drivers/gpu/drm/udl/udl_main.c struct udl_device *udl = to_udl(dev); udl 301 drivers/gpu/drm/udl/udl_main.c BUG_ON(len > udl->urbs.size); udl 307 drivers/gpu/drm/udl/udl_main.c atomic_set(&udl->lost_pixels, 1); udl 313 drivers/gpu/drm/udl/udl_main.c int udl_init(struct udl_device *udl) udl 315 drivers/gpu/drm/udl/udl_main.c struct drm_device *dev = &udl->drm; udl 320 drivers/gpu/drm/udl/udl_main.c mutex_init(&udl->gem_lock); udl 322 drivers/gpu/drm/udl/udl_main.c if (!udl_parse_vendor_descriptor(dev, udl->udev)) { udl 328 drivers/gpu/drm/udl/udl_main.c if (udl_select_std_channel(udl)) udl 350 drivers/gpu/drm/udl/udl_main.c if (udl->urbs.count) udl 364 drivers/gpu/drm/udl/udl_main.c struct udl_device *udl = to_udl(dev); udl 368 drivers/gpu/drm/udl/udl_main.c if (udl->urbs.count) udl 231 drivers/gpu/drm/udl/udl_modeset.c struct udl_device *udl = dev->dev_private; udl 242 drivers/gpu/drm/udl/udl_modeset.c memcpy(buf, udl->mode_buf, udl->mode_buf_len); udl 243 drivers/gpu/drm/udl/udl_modeset.c retval = udl_submit_urb(dev, urb, udl->mode_buf_len); udl 244 drivers/gpu/drm/udl/udl_modeset.c DRM_DEBUG("write mode info %d\n", udl->mode_buf_len); udl 252 drivers/gpu/drm/udl/udl_modeset.c struct udl_device *udl = dev->dev_private; udl 271 drivers/gpu/drm/udl/udl_modeset.c if (udl->mode_buf_len == 0) { udl 305 drivers/gpu/drm/udl/udl_modeset.c struct udl_device *udl = dev->dev_private; udl 310 drivers/gpu/drm/udl/udl_modeset.c udl->crtc = crtc; udl 312 drivers/gpu/drm/udl/udl_modeset.c buf = (char *)udl->mode_buf; udl 340 drivers/gpu/drm/udl/udl_modeset.c udl->mode_buf_len = wrptr - buf; udl 455 drivers/gpu/drm/udl/udl_modeset.c struct udl_device *udl = dev->dev_private; udl 458 drivers/gpu/drm/udl/udl_modeset.c if (!udl->crtc || !udl->crtc->primary->fb) udl 460 drivers/gpu/drm/udl/udl_modeset.c udl_crtc_commit(udl->crtc); udl 461 drivers/gpu/drm/udl/udl_modeset.c ufb = to_udl_fb(udl->crtc->primary->fb); udl 26 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c ddp->udl = NULL; udl 61 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c if (!ddp->udl) udl 113 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c dma_pool_free(ddp->pool, ddp->udl, ddp->udp); udl 195 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c ddp->udl = dma_pool_alloc(ddp_pool->pool, GFP_KERNEL, &ddp->udp); udl 196 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c if (!ddp->udl) { udl 233 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c ddp->udl[j] = (u64)(addr - thisoff); udl 255 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c ddp->udl[j] = (u64)(fcoe->extra_ddp_buffer_dma); udl 319 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c dma_pool_free(ddp->pool, ddp->udl, ddp->udp); udl 424 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c if (!ddp->udl) udl 41 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.h u64 *udl;