Lines Matching refs:entry
42 struct drm_map_list *entry; in drm_find_matching_map() local
43 list_for_each_entry(entry, &dev->maplist, head) { in drm_find_matching_map()
52 if (!entry->map || in drm_find_matching_map()
53 map->type != entry->map->type || in drm_find_matching_map()
54 entry->master != dev->primary->master) in drm_find_matching_map()
60 return entry; in drm_find_matching_map()
63 if ((entry->map->offset & 0xffffffff) == in drm_find_matching_map()
65 return entry; in drm_find_matching_map()
69 if (entry->map->offset == map->offset) in drm_find_matching_map()
70 return entry; in drm_find_matching_map()
257 struct drm_agp_mem *entry; in drm_addmap_core() local
286 list_for_each_entry(entry, &dev->agp->memory, head) { in drm_addmap_core()
287 if ((map->offset >= entry->bound) && in drm_addmap_core()
288 (map->offset + map->size <= entry->bound + entry->pages * PAGE_SIZE)) { in drm_addmap_core()
560 struct drm_buf_entry * entry) in drm_cleanup_buf_error() argument
564 if (entry->seg_count) { in drm_cleanup_buf_error()
565 for (i = 0; i < entry->seg_count; i++) { in drm_cleanup_buf_error()
566 if (entry->seglist[i]) { in drm_cleanup_buf_error()
567 drm_pci_free(dev, entry->seglist[i]); in drm_cleanup_buf_error()
570 kfree(entry->seglist); in drm_cleanup_buf_error()
572 entry->seg_count = 0; in drm_cleanup_buf_error()
575 if (entry->buf_count) { in drm_cleanup_buf_error()
576 for (i = 0; i < entry->buf_count; i++) { in drm_cleanup_buf_error()
577 kfree(entry->buflist[i].dev_private); in drm_cleanup_buf_error()
579 kfree(entry->buflist); in drm_cleanup_buf_error()
581 entry->buf_count = 0; in drm_cleanup_buf_error()
601 struct drm_buf_entry *entry; in drm_legacy_addbufs_agp() local
664 entry = &dma->bufs[order]; in drm_legacy_addbufs_agp()
665 if (entry->buf_count) { in drm_legacy_addbufs_agp()
677 entry->buflist = kzalloc(count * sizeof(*entry->buflist), GFP_KERNEL); in drm_legacy_addbufs_agp()
678 if (!entry->buflist) { in drm_legacy_addbufs_agp()
684 entry->buf_size = size; in drm_legacy_addbufs_agp()
685 entry->page_order = page_order; in drm_legacy_addbufs_agp()
689 while (entry->buf_count < count) { in drm_legacy_addbufs_agp()
690 buf = &entry->buflist[entry->buf_count]; in drm_legacy_addbufs_agp()
691 buf->idx = dma->buf_count + entry->buf_count; in drm_legacy_addbufs_agp()
708 entry->buf_count = count; in drm_legacy_addbufs_agp()
709 drm_cleanup_buf_error(dev, entry); in drm_legacy_addbufs_agp()
715 DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); in drm_legacy_addbufs_agp()
718 entry->buf_count++; in drm_legacy_addbufs_agp()
725 (dma->buf_count + entry->buf_count) * in drm_legacy_addbufs_agp()
729 drm_cleanup_buf_error(dev, entry); in drm_legacy_addbufs_agp()
736 for (i = 0; i < entry->buf_count; i++) { in drm_legacy_addbufs_agp()
737 dma->buflist[i + dma->buf_count] = &entry->buflist[i]; in drm_legacy_addbufs_agp()
740 dma->buf_count += entry->buf_count; in drm_legacy_addbufs_agp()
741 dma->seg_count += entry->seg_count; in drm_legacy_addbufs_agp()
746 DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count); in drm_legacy_addbufs_agp()
750 request->count = entry->buf_count; in drm_legacy_addbufs_agp()
770 struct drm_buf_entry *entry; in drm_legacy_addbufs_pci() local
814 entry = &dma->bufs[order]; in drm_legacy_addbufs_pci()
815 if (entry->buf_count) { in drm_legacy_addbufs_pci()
827 entry->buflist = kzalloc(count * sizeof(*entry->buflist), GFP_KERNEL); in drm_legacy_addbufs_pci()
828 if (!entry->buflist) { in drm_legacy_addbufs_pci()
834 entry->seglist = kzalloc(count * sizeof(*entry->seglist), GFP_KERNEL); in drm_legacy_addbufs_pci()
835 if (!entry->seglist) { in drm_legacy_addbufs_pci()
836 kfree(entry->buflist); in drm_legacy_addbufs_pci()
848 kfree(entry->buflist); in drm_legacy_addbufs_pci()
849 kfree(entry->seglist); in drm_legacy_addbufs_pci()
859 entry->buf_size = size; in drm_legacy_addbufs_pci()
860 entry->page_order = page_order; in drm_legacy_addbufs_pci()
864 while (entry->buf_count < count) { in drm_legacy_addbufs_pci()
870 entry->buf_count = count; in drm_legacy_addbufs_pci()
871 entry->seg_count = count; in drm_legacy_addbufs_pci()
872 drm_cleanup_buf_error(dev, entry); in drm_legacy_addbufs_pci()
878 entry->seglist[entry->seg_count++] = dmah; in drm_legacy_addbufs_pci()
887 offset + size <= total && entry->buf_count < count; in drm_legacy_addbufs_pci()
888 offset += alignment, ++entry->buf_count) { in drm_legacy_addbufs_pci()
889 buf = &entry->buflist[entry->buf_count]; in drm_legacy_addbufs_pci()
890 buf->idx = dma->buf_count + entry->buf_count; in drm_legacy_addbufs_pci()
907 entry->buf_count = count; in drm_legacy_addbufs_pci()
908 entry->seg_count = count; in drm_legacy_addbufs_pci()
909 drm_cleanup_buf_error(dev, entry); in drm_legacy_addbufs_pci()
917 entry->buf_count, buf->address); in drm_legacy_addbufs_pci()
923 (dma->buf_count + entry->buf_count) * in drm_legacy_addbufs_pci()
927 drm_cleanup_buf_error(dev, entry); in drm_legacy_addbufs_pci()
935 for (i = 0; i < entry->buf_count; i++) { in drm_legacy_addbufs_pci()
936 dma->buflist[i + dma->buf_count] = &entry->buflist[i]; in drm_legacy_addbufs_pci()
947 dma->buf_count += entry->buf_count; in drm_legacy_addbufs_pci()
948 dma->seg_count += entry->seg_count; in drm_legacy_addbufs_pci()
949 dma->page_count += entry->seg_count << page_order; in drm_legacy_addbufs_pci()
950 dma->byte_count += PAGE_SIZE * (entry->seg_count << page_order); in drm_legacy_addbufs_pci()
954 request->count = entry->buf_count; in drm_legacy_addbufs_pci()
970 struct drm_buf_entry *entry; in drm_legacy_addbufs_sg() local
1025 entry = &dma->bufs[order]; in drm_legacy_addbufs_sg()
1026 if (entry->buf_count) { in drm_legacy_addbufs_sg()
1038 entry->buflist = kzalloc(count * sizeof(*entry->buflist), in drm_legacy_addbufs_sg()
1040 if (!entry->buflist) { in drm_legacy_addbufs_sg()
1046 entry->buf_size = size; in drm_legacy_addbufs_sg()
1047 entry->page_order = page_order; in drm_legacy_addbufs_sg()
1051 while (entry->buf_count < count) { in drm_legacy_addbufs_sg()
1052 buf = &entry->buflist[entry->buf_count]; in drm_legacy_addbufs_sg()
1053 buf->idx = dma->buf_count + entry->buf_count; in drm_legacy_addbufs_sg()
1071 entry->buf_count = count; in drm_legacy_addbufs_sg()
1072 drm_cleanup_buf_error(dev, entry); in drm_legacy_addbufs_sg()
1078 DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); in drm_legacy_addbufs_sg()
1081 entry->buf_count++; in drm_legacy_addbufs_sg()
1088 (dma->buf_count + entry->buf_count) * in drm_legacy_addbufs_sg()
1092 drm_cleanup_buf_error(dev, entry); in drm_legacy_addbufs_sg()
1099 for (i = 0; i < entry->buf_count; i++) { in drm_legacy_addbufs_sg()
1100 dma->buflist[i + dma->buf_count] = &entry->buflist[i]; in drm_legacy_addbufs_sg()
1103 dma->buf_count += entry->buf_count; in drm_legacy_addbufs_sg()
1104 dma->seg_count += entry->seg_count; in drm_legacy_addbufs_sg()
1109 DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count); in drm_legacy_addbufs_sg()
1113 request->count = entry->buf_count; in drm_legacy_addbufs_sg()
1267 struct drm_buf_entry *entry; in drm_legacy_markbufs() local
1283 entry = &dma->bufs[order]; in drm_legacy_markbufs()
1285 if (request->low_mark < 0 || request->low_mark > entry->buf_count) in drm_legacy_markbufs()
1287 if (request->high_mark < 0 || request->high_mark > entry->buf_count) in drm_legacy_markbufs()
1290 entry->low_mark = request->low_mark; in drm_legacy_markbufs()
1291 entry->high_mark = request->high_mark; in drm_legacy_markbufs()
1463 struct drm_map_list *entry; in drm_legacy_getsarea() local
1465 list_for_each_entry(entry, &dev->maplist, head) { in drm_legacy_getsarea()
1466 if (entry->map && entry->map->type == _DRM_SHM && in drm_legacy_getsarea()
1467 (entry->map->flags & _DRM_CONTAINS_LOCK)) { in drm_legacy_getsarea()
1468 return entry->map; in drm_legacy_getsarea()