udl                21 drivers/gpu/drm/udl/udl_connector.c 	struct udl_device *udl = data;
udl                29 drivers/gpu/drm/udl/udl_connector.c 		ret = usb_control_msg(udl->udev,
udl                30 drivers/gpu/drm/udl/udl_connector.c 				      usb_rcvctrlpipe(udl->udev, 0),
udl                61 drivers/gpu/drm/udl/udl_connector.c 	struct udl_device *udl = connector->dev->dev_private;
udl                62 drivers/gpu/drm/udl/udl_connector.c 	if (!udl->sku_pixel_limit)
udl                65 drivers/gpu/drm/udl/udl_connector.c 	if (mode->vdisplay * mode->hdisplay > udl->sku_pixel_limit)
udl                74 drivers/gpu/drm/udl/udl_connector.c 	struct udl_device *udl = connector->dev->dev_private;
udl                86 drivers/gpu/drm/udl/udl_connector.c 	udl_connector->edid = drm_do_get_edid(connector, udl_get_edid_block, udl);
udl                68 drivers/gpu/drm/udl/udl_dmabuf.c 	struct udl_device *udl = dev->dev_private;
udl               105 drivers/gpu/drm/udl/udl_dmabuf.c 	mutex_lock(&udl->gem_lock);
udl               130 drivers/gpu/drm/udl/udl_dmabuf.c 	mutex_unlock(&udl->gem_lock);
udl                89 drivers/gpu/drm/udl/udl_drv.c 	struct udl_device *udl;
udl                92 drivers/gpu/drm/udl/udl_drv.c 	udl = kzalloc(sizeof(*udl), GFP_KERNEL);
udl                93 drivers/gpu/drm/udl/udl_drv.c 	if (!udl)
udl                96 drivers/gpu/drm/udl/udl_drv.c 	r = drm_dev_init(&udl->drm, &driver, &interface->dev);
udl                98 drivers/gpu/drm/udl/udl_drv.c 		kfree(udl);
udl               102 drivers/gpu/drm/udl/udl_drv.c 	udl->udev = udev;
udl               103 drivers/gpu/drm/udl/udl_drv.c 	udl->drm.dev_private = udl;
udl               105 drivers/gpu/drm/udl/udl_drv.c 	r = udl_init(udl);
udl               107 drivers/gpu/drm/udl/udl_drv.c 		drm_dev_fini(&udl->drm);
udl               108 drivers/gpu/drm/udl/udl_drv.c 		kfree(udl);
udl               112 drivers/gpu/drm/udl/udl_drv.c 	usb_set_intfdata(interface, udl);
udl               113 drivers/gpu/drm/udl/udl_drv.c 	return udl;
udl               120 drivers/gpu/drm/udl/udl_drv.c 	struct udl_device *udl;
udl               122 drivers/gpu/drm/udl/udl_drv.c 	udl = udl_driver_create(interface);
udl               123 drivers/gpu/drm/udl/udl_drv.c 	if (IS_ERR(udl))
udl               124 drivers/gpu/drm/udl/udl_drv.c 		return PTR_ERR(udl);
udl               126 drivers/gpu/drm/udl/udl_drv.c 	r = drm_dev_register(&udl->drm, 0);
udl               130 drivers/gpu/drm/udl/udl_drv.c 	DRM_INFO("Initialized udl on minor %d\n", udl->drm.primary->index);
udl               135 drivers/gpu/drm/udl/udl_drv.c 	drm_dev_put(&udl->drm);
udl               110 drivers/gpu/drm/udl/udl_drv.h int udl_init(struct udl_device *udl);
udl                81 drivers/gpu/drm/udl/udl_fb.c 	struct udl_device *udl = to_udl(dev);
udl               149 drivers/gpu/drm/udl/udl_fb.c 	atomic_add(bytes_sent, &udl->bytes_sent);
udl               150 drivers/gpu/drm/udl/udl_fb.c 	atomic_add(bytes_identical, &udl->bytes_identical);
udl               151 drivers/gpu/drm/udl/udl_fb.c 	atomic_add((width * height) << log_bpp, &udl->bytes_rendered);
udl               155 drivers/gpu/drm/udl/udl_fb.c 		   &udl->cpu_kcycles_used);
udl               209 drivers/gpu/drm/udl/udl_fb.c 	struct udl_device *udl = to_udl(dev);
udl               212 drivers/gpu/drm/udl/udl_fb.c 	if (drm_dev_is_unplugged(&udl->drm))
udl               436 drivers/gpu/drm/udl/udl_fb.c 	struct udl_device *udl = to_udl(dev);
udl               445 drivers/gpu/drm/udl/udl_fb.c 	udl->fbdev = ufbdev;
udl               475 drivers/gpu/drm/udl/udl_fb.c 	struct udl_device *udl = to_udl(dev);
udl               476 drivers/gpu/drm/udl/udl_fb.c 	if (!udl->fbdev)
udl               479 drivers/gpu/drm/udl/udl_fb.c 	udl_fbdev_destroy(dev, udl->fbdev);
udl               480 drivers/gpu/drm/udl/udl_fb.c 	kfree(udl->fbdev);
udl               481 drivers/gpu/drm/udl/udl_fb.c 	udl->fbdev = NULL;
udl               486 drivers/gpu/drm/udl/udl_fb.c 	struct udl_device *udl = to_udl(dev);
udl               488 drivers/gpu/drm/udl/udl_fb.c 	if (!udl->fbdev)
udl               491 drivers/gpu/drm/udl/udl_fb.c 	ufbdev = udl->fbdev;
udl               206 drivers/gpu/drm/udl/udl_gem.c 	struct udl_device *udl = to_udl(dev);
udl               209 drivers/gpu/drm/udl/udl_gem.c 	mutex_lock(&udl->gem_lock);
udl               229 drivers/gpu/drm/udl/udl_gem.c 	mutex_unlock(&udl->gem_lock);
udl                32 drivers/gpu/drm/udl/udl_main.c 	struct udl_device *udl = to_udl(dev);
udl                75 drivers/gpu/drm/udl/udl_main.c 				udl->sku_pixel_limit = max_area;
udl                99 drivers/gpu/drm/udl/udl_main.c static int udl_select_std_channel(struct udl_device *udl)
udl               112 drivers/gpu/drm/udl/udl_main.c 	ret = usb_control_msg(udl->udev,
udl               113 drivers/gpu/drm/udl/udl_main.c 			      usb_sndctrlpipe(udl->udev, 0),
udl               133 drivers/gpu/drm/udl/udl_main.c 	struct udl_device *udl = unode->dev;
udl               143 drivers/gpu/drm/udl/udl_main.c 			atomic_set(&udl->lost_pixels, 1);
udl               147 drivers/gpu/drm/udl/udl_main.c 	urb->transfer_buffer_length = udl->urbs.size; /* reset to actual */
udl               149 drivers/gpu/drm/udl/udl_main.c 	spin_lock_irqsave(&udl->urbs.lock, flags);
udl               150 drivers/gpu/drm/udl/udl_main.c 	list_add_tail(&unode->entry, &udl->urbs.list);
udl               151 drivers/gpu/drm/udl/udl_main.c 	udl->urbs.available++;
udl               152 drivers/gpu/drm/udl/udl_main.c 	spin_unlock_irqrestore(&udl->urbs.lock, flags);
udl               163 drivers/gpu/drm/udl/udl_main.c 		up(&udl->urbs.limit_sem);
udl               168 drivers/gpu/drm/udl/udl_main.c 	struct udl_device *udl = to_udl(dev);
udl               169 drivers/gpu/drm/udl/udl_main.c 	int count = udl->urbs.count;
udl               178 drivers/gpu/drm/udl/udl_main.c 		down(&udl->urbs.limit_sem);
udl               180 drivers/gpu/drm/udl/udl_main.c 		spin_lock_irq(&udl->urbs.lock);
udl               182 drivers/gpu/drm/udl/udl_main.c 		node = udl->urbs.list.next; /* have reserved one with sem */
udl               185 drivers/gpu/drm/udl/udl_main.c 		spin_unlock_irq(&udl->urbs.lock);
udl               191 drivers/gpu/drm/udl/udl_main.c 		usb_free_coherent(urb->dev, udl->urbs.size,
udl               196 drivers/gpu/drm/udl/udl_main.c 	udl->urbs.count = 0;
udl               201 drivers/gpu/drm/udl/udl_main.c 	struct udl_device *udl = to_udl(dev);
udl               207 drivers/gpu/drm/udl/udl_main.c 	spin_lock_init(&udl->urbs.lock);
udl               210 drivers/gpu/drm/udl/udl_main.c 	udl->urbs.size = size;
udl               211 drivers/gpu/drm/udl/udl_main.c 	INIT_LIST_HEAD(&udl->urbs.list);
udl               213 drivers/gpu/drm/udl/udl_main.c 	sema_init(&udl->urbs.limit_sem, 0);
udl               214 drivers/gpu/drm/udl/udl_main.c 	udl->urbs.count = 0;
udl               215 drivers/gpu/drm/udl/udl_main.c 	udl->urbs.available = 0;
udl               217 drivers/gpu/drm/udl/udl_main.c 	while (udl->urbs.count * size < wanted_size) {
udl               221 drivers/gpu/drm/udl/udl_main.c 		unode->dev = udl;
udl               233 drivers/gpu/drm/udl/udl_main.c 		buf = usb_alloc_coherent(udl->udev, size, GFP_KERNEL,
udl               247 drivers/gpu/drm/udl/udl_main.c 		usb_fill_bulk_urb(urb, udl->udev, usb_sndbulkpipe(udl->udev, 1),
udl               251 drivers/gpu/drm/udl/udl_main.c 		list_add_tail(&unode->entry, &udl->urbs.list);
udl               253 drivers/gpu/drm/udl/udl_main.c 		up(&udl->urbs.limit_sem);
udl               254 drivers/gpu/drm/udl/udl_main.c 		udl->urbs.count++;
udl               255 drivers/gpu/drm/udl/udl_main.c 		udl->urbs.available++;
udl               258 drivers/gpu/drm/udl/udl_main.c 	DRM_DEBUG("allocated %d %d byte urbs\n", udl->urbs.count, (int) size);
udl               260 drivers/gpu/drm/udl/udl_main.c 	return udl->urbs.count;
udl               265 drivers/gpu/drm/udl/udl_main.c 	struct udl_device *udl = to_udl(dev);
udl               272 drivers/gpu/drm/udl/udl_main.c 	ret = down_timeout(&udl->urbs.limit_sem, GET_URB_TIMEOUT);
udl               274 drivers/gpu/drm/udl/udl_main.c 		atomic_set(&udl->lost_pixels, 1);
udl               276 drivers/gpu/drm/udl/udl_main.c 		       ret, udl->urbs.available);
udl               280 drivers/gpu/drm/udl/udl_main.c 	spin_lock_irq(&udl->urbs.lock);
udl               282 drivers/gpu/drm/udl/udl_main.c 	BUG_ON(list_empty(&udl->urbs.list)); /* reserved one with limit_sem */
udl               283 drivers/gpu/drm/udl/udl_main.c 	entry = udl->urbs.list.next;
udl               285 drivers/gpu/drm/udl/udl_main.c 	udl->urbs.available--;
udl               287 drivers/gpu/drm/udl/udl_main.c 	spin_unlock_irq(&udl->urbs.lock);
udl               298 drivers/gpu/drm/udl/udl_main.c 	struct udl_device *udl = to_udl(dev);
udl               301 drivers/gpu/drm/udl/udl_main.c 	BUG_ON(len > udl->urbs.size);
udl               307 drivers/gpu/drm/udl/udl_main.c 		atomic_set(&udl->lost_pixels, 1);
udl               313 drivers/gpu/drm/udl/udl_main.c int udl_init(struct udl_device *udl)
udl               315 drivers/gpu/drm/udl/udl_main.c 	struct drm_device *dev = &udl->drm;
udl               320 drivers/gpu/drm/udl/udl_main.c 	mutex_init(&udl->gem_lock);
udl               322 drivers/gpu/drm/udl/udl_main.c 	if (!udl_parse_vendor_descriptor(dev, udl->udev)) {
udl               328 drivers/gpu/drm/udl/udl_main.c 	if (udl_select_std_channel(udl))
udl               350 drivers/gpu/drm/udl/udl_main.c 	if (udl->urbs.count)
udl               364 drivers/gpu/drm/udl/udl_main.c 	struct udl_device *udl = to_udl(dev);
udl               368 drivers/gpu/drm/udl/udl_main.c 	if (udl->urbs.count)
udl               231 drivers/gpu/drm/udl/udl_modeset.c 	struct udl_device *udl = dev->dev_private;
udl               242 drivers/gpu/drm/udl/udl_modeset.c 	memcpy(buf, udl->mode_buf, udl->mode_buf_len);
udl               243 drivers/gpu/drm/udl/udl_modeset.c 	retval = udl_submit_urb(dev, urb, udl->mode_buf_len);
udl               244 drivers/gpu/drm/udl/udl_modeset.c 	DRM_DEBUG("write mode info %d\n", udl->mode_buf_len);
udl               252 drivers/gpu/drm/udl/udl_modeset.c 	struct udl_device *udl = dev->dev_private;
udl               271 drivers/gpu/drm/udl/udl_modeset.c 		if (udl->mode_buf_len == 0) {
udl               305 drivers/gpu/drm/udl/udl_modeset.c 	struct udl_device *udl = dev->dev_private;
udl               310 drivers/gpu/drm/udl/udl_modeset.c 	udl->crtc = crtc;
udl               312 drivers/gpu/drm/udl/udl_modeset.c 	buf = (char *)udl->mode_buf;
udl               340 drivers/gpu/drm/udl/udl_modeset.c 	udl->mode_buf_len = wrptr - buf;
udl               455 drivers/gpu/drm/udl/udl_modeset.c 	struct udl_device *udl = dev->dev_private;
udl               458 drivers/gpu/drm/udl/udl_modeset.c 	if (!udl->crtc || !udl->crtc->primary->fb)
udl               460 drivers/gpu/drm/udl/udl_modeset.c 	udl_crtc_commit(udl->crtc);
udl               461 drivers/gpu/drm/udl/udl_modeset.c 	ufb = to_udl_fb(udl->crtc->primary->fb);
udl                26 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	ddp->udl = NULL;
udl                61 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	if (!ddp->udl)
udl               113 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 		dma_pool_free(ddp->pool, ddp->udl, ddp->udp);
udl               195 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	ddp->udl = dma_pool_alloc(ddp_pool->pool, GFP_KERNEL, &ddp->udp);
udl               196 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	if (!ddp->udl) {
udl               233 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 			ddp->udl[j] = (u64)(addr - thisoff);
udl               255 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 		ddp->udl[j] = (u64)(fcoe->extra_ddp_buffer_dma);
udl               319 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	dma_pool_free(ddp->pool, ddp->udl, ddp->udp);
udl               424 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	if (!ddp->udl)
udl                41 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.h 	u64 *udl;