/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
D | gk20a.c | 55 struct gk20a_instmem *imem; member 138 struct device *dev = node->base.imem->base.subdev.device->dev; in gk20a_instobj_cpu_map_dma() 170 gk20a_instmem_vaddr_gc(struct gk20a_instmem *imem, const u64 size) in gk20a_instmem_vaddr_gc() argument 172 while (imem->vaddr_use + size > imem->vaddr_max) { in gk20a_instmem_vaddr_gc() 176 if (list_empty(&imem->vaddr_lru)) in gk20a_instmem_vaddr_gc() 179 obj = list_first_entry(&imem->vaddr_lru, struct gk20a_instobj, in gk20a_instmem_vaddr_gc() 184 imem->vaddr_use -= nvkm_memory_size(&obj->memory); in gk20a_instmem_vaddr_gc() 185 nvkm_debug(&imem->base.subdev, "(GC) vaddr used: %x/%x\n", in gk20a_instmem_vaddr_gc() 186 imem->vaddr_use, imem->vaddr_max); in gk20a_instmem_vaddr_gc() 195 struct gk20a_instmem *imem = node->imem; in gk20a_instobj_acquire() local [all …]
|
D | nv40.c | 44 struct nv40_instmem *imem; member 70 return iobj->imem->iomem + iobj->node->offset; in nv40_instobj_acquire() 82 return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_rd32() 89 iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_wr32() 96 mutex_lock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor() 97 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv40_instobj_dtor() 98 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor() 118 struct nv40_instmem *imem = nv40_instmem(base); in nv40_instobj_new() local 127 iobj->imem = imem; in nv40_instobj_new() 129 mutex_lock(&imem->base.subdev.mutex); in nv40_instobj_new() [all …]
|
D | nv04.c | 42 struct nv04_instmem *imem; member 68 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_acquire() 81 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_rd32() 89 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_wr32() 97 mutex_lock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor() 98 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv04_instobj_dtor() 99 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor() 119 struct nv04_instmem *imem = nv04_instmem(base); in nv04_instobj_new() local 128 iobj->imem = imem; in nv04_instobj_new() 130 mutex_lock(&imem->base.subdev.mutex); in nv04_instobj_new() [all …]
|
D | base.c | 37 struct nvkm_instmem *imem; member 68 nvkm_bar_flush(iobj->imem->subdev.device->bar); in nvkm_instobj_release() 100 spin_lock(&iobj->imem->lock); in nvkm_instobj_dtor() 102 spin_unlock(&iobj->imem->lock); in nvkm_instobj_dtor() 174 nvkm_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, in nvkm_instobj_new() argument 182 ret = imem->func->memory_new(imem, size, align, zero, &memory); in nvkm_instobj_new() 186 if (!imem->func->persistent) { in nvkm_instobj_new() 194 iobj->imem = imem; in nvkm_instobj_new() 195 spin_lock(&iobj->imem->lock); in nvkm_instobj_new() 196 list_add_tail(&iobj->head, &imem->list); in nvkm_instobj_new() [all …]
|
D | nv50.c | 46 struct nv50_instmem *imem; member 74 struct nvkm_subdev *subdev = &iobj->imem->base.subdev; in nv50_instobj_boot() 101 struct nv50_instmem *imem = nv50_instobj(memory)->imem; in nv50_instobj_release() local 102 spin_unlock_irqrestore(&imem->lock, imem->lock_flags); in nv50_instobj_release() 109 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_acquire() local 110 struct nvkm_bar *bar = imem->base.subdev.device->bar; in nv50_instobj_acquire() 119 spin_lock_irqsave(&imem->lock, flags); in nv50_instobj_acquire() 120 imem->lock_flags = flags; in nv50_instobj_acquire() 128 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_rd32() local 129 struct nvkm_device *device = imem->base.subdev.device; in nv50_instobj_rd32() [all …]
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | dmanv40.c | 62 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_fini() local 76 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_fini() 77 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, 0x00000000); in nv40_fifo_dma_engine_fini() 78 nvkm_done(imem->ramfc); in nv40_fifo_dma_engine_fini() 92 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_init() local 107 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_init() 108 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, inst); in nv40_fifo_dma_engine_init() 109 nvkm_done(imem->ramfc); in nv40_fifo_dma_engine_init() 144 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv40_fifo_dma_object_ctor() local 160 hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, in nv40_fifo_dma_object_ctor() [all …]
|
D | dmanv04.c | 38 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_dtor() local 39 nvkm_ramht_remove(imem->ramht, cookie); in nv04_fifo_dma_object_dtor() 47 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_ctor() local 63 hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, in nv04_fifo_dma_object_ctor() 75 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_fifo_dma_fini() 138 struct nvkm_instmem *imem = fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_dtor() local 141 nvkm_kmap(imem->ramfc); in nv04_fifo_dma_dtor() 143 nvkm_wo32(imem->ramfc, chan->ramfc + c->ctxp, 0x00000000); in nv04_fifo_dma_dtor() 145 nvkm_done(imem->ramfc); in nv04_fifo_dma_dtor() 169 struct nvkm_instmem *imem = device->imem; in nv04_fifo_dma_new() local [all …]
|
D | dmanv10.c | 45 struct nvkm_instmem *imem = device->imem; in nv10_fifo_dma_new() local 75 nvkm_kmap(imem->ramfc); in nv10_fifo_dma_new() 76 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv10_fifo_dma_new() 77 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv10_fifo_dma_new() 78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv10_fifo_dma_new() 79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv10_fifo_dma_new() 86 nvkm_done(imem->ramfc); in nv10_fifo_dma_new()
|
D | dmanv17.c | 45 struct nvkm_instmem *imem = device->imem; in nv17_fifo_dma_new() local 76 nvkm_kmap(imem->ramfc); in nv17_fifo_dma_new() 77 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv17_fifo_dma_new() 78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv17_fifo_dma_new() 79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv17_fifo_dma_new() 80 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv17_fifo_dma_new() 87 nvkm_done(imem->ramfc); in nv17_fifo_dma_new()
|
D | nv17.c | 55 struct nvkm_instmem *imem = device->imem; in nv17_fifo_init() local 56 struct nvkm_ramht *ramht = imem->ramht; in nv17_fifo_init() 57 struct nvkm_memory *ramro = imem->ramro; in nv17_fifo_init() 58 struct nvkm_memory *ramfc = imem->ramfc; in nv17_fifo_init()
|
D | nv40.c | 65 struct nvkm_instmem *imem = device->imem; in nv40_fifo_init() local 66 struct nvkm_ramht *ramht = imem->ramht; in nv40_fifo_init() 67 struct nvkm_memory *ramro = imem->ramro; in nv40_fifo_init() 68 struct nvkm_memory *ramfc = imem->ramfc; in nv40_fifo_init()
|
D | nv04.c | 301 struct nvkm_instmem *imem = device->imem; in nv04_fifo_init() local 302 struct nvkm_ramht *ramht = imem->ramht; in nv04_fifo_init() 303 struct nvkm_memory *ramro = imem->ramro; in nv04_fifo_init() 304 struct nvkm_memory *ramfc = imem->ramfc; in nv04_fifo_init()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/mpeg/ |
D | nv40.c | 33 struct nvkm_instmem *imem = device->imem; in nv40_mpeg_mthd_dma() local 35 u32 dma0 = nvkm_instmem_rd32(imem, inst + 0); in nv40_mpeg_mthd_dma() 36 u32 dma1 = nvkm_instmem_rd32(imem, inst + 4); in nv40_mpeg_mthd_dma() 37 u32 dma2 = nvkm_instmem_rd32(imem, inst + 8); in nv40_mpeg_mthd_dma()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/core/ |
D | memory.c | 51 struct nvkm_instmem *imem = device->imem; in nvkm_memory_new() local 55 if (unlikely(target != NVKM_MEM_TARGET_INST || !imem)) in nvkm_memory_new() 58 ret = nvkm_instobj_new(imem, size, align, zero, &memory); in nvkm_memory_new()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/device/ |
D | base.c | 85 .imem = nv04_instmem_new, 106 .imem = nv04_instmem_new, 128 .imem = nv04_instmem_new, 148 .imem = nv04_instmem_new, 170 .imem = nv04_instmem_new, 192 .imem = nv04_instmem_new, 214 .imem = nv04_instmem_new, 236 .imem = nv04_instmem_new, 258 .imem = nv04_instmem_new, 280 .imem = nv04_instmem_new, [all …]
|
D | user.c | 47 struct nvkm_instmem *imem = device->imem; in nvkm_udevice_info() local 115 if (imem && args->v0.ram_size > 0) in nvkm_udevice_info() 116 args->v0.ram_user = args->v0.ram_user - imem->reserved; in nvkm_udevice_info()
|
/linux-4.4.14/arch/blackfin/mach-common/ |
D | cache-c.c | 29 u32 imem = bfin_read_IMEM_CONTROL(); in blackfin_invalidate_entire_icache() local 30 bfin_write_IMEM_CONTROL(imem & ~0x4); in blackfin_invalidate_entire_icache() 32 bfin_write_IMEM_CONTROL(imem); in blackfin_invalidate_entire_icache()
|
/linux-4.4.14/arch/ia64/sn/kernel/ |
D | bte_error.c | 43 ii_imem_u_t imem; /* II IMEM Register */ in shub1_bte_error_handler() local 109 imem.ii_imem_regval = REMOTE_HUB_L(nasid, IIO_IMEM); in shub1_bte_error_handler() 110 imem.ii_imem_fld_s.i_b0_esd = imem.ii_imem_fld_s.i_b1_esd = 1; in shub1_bte_error_handler() 111 REMOTE_HUB_S(nasid, IIO_IMEM, imem.ii_imem_regval); in shub1_bte_error_handler()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/include/nvkm/core/ |
D | device.h | 112 struct nvkm_instmem *imem; member 177 int (*imem )(struct nvkm_device *, int idx, struct nvkm_instmem **); member
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/dma/ |
D | user.c | 68 struct nvkm_instmem *instmem = device->imem; in nvkm_dmaobj_ctor()
|