Home
last modified time | relevance | path

Searched refs:imem (Results 1 – 20 of 20) sorted by relevance

/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
Dgk20a.c55 struct gk20a_instmem *imem; member
138 struct device *dev = node->base.imem->base.subdev.device->dev; in gk20a_instobj_cpu_map_dma()
170 gk20a_instmem_vaddr_gc(struct gk20a_instmem *imem, const u64 size) in gk20a_instmem_vaddr_gc() argument
172 while (imem->vaddr_use + size > imem->vaddr_max) { in gk20a_instmem_vaddr_gc()
176 if (list_empty(&imem->vaddr_lru)) in gk20a_instmem_vaddr_gc()
179 obj = list_first_entry(&imem->vaddr_lru, struct gk20a_instobj, in gk20a_instmem_vaddr_gc()
184 imem->vaddr_use -= nvkm_memory_size(&obj->memory); in gk20a_instmem_vaddr_gc()
185 nvkm_debug(&imem->base.subdev, "(GC) vaddr used: %x/%x\n", in gk20a_instmem_vaddr_gc()
186 imem->vaddr_use, imem->vaddr_max); in gk20a_instmem_vaddr_gc()
195 struct gk20a_instmem *imem = node->imem; in gk20a_instobj_acquire() local
[all …]
Dnv40.c44 struct nv40_instmem *imem; member
70 return iobj->imem->iomem + iobj->node->offset; in nv40_instobj_acquire()
82 return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_rd32()
89 iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_wr32()
96 mutex_lock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor()
97 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv40_instobj_dtor()
98 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor()
118 struct nv40_instmem *imem = nv40_instmem(base); in nv40_instobj_new() local
127 iobj->imem = imem; in nv40_instobj_new()
129 mutex_lock(&imem->base.subdev.mutex); in nv40_instobj_new()
[all …]
Dnv04.c42 struct nv04_instmem *imem; member
68 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_acquire()
81 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_rd32()
89 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_wr32()
97 mutex_lock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor()
98 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv04_instobj_dtor()
99 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor()
119 struct nv04_instmem *imem = nv04_instmem(base); in nv04_instobj_new() local
128 iobj->imem = imem; in nv04_instobj_new()
130 mutex_lock(&imem->base.subdev.mutex); in nv04_instobj_new()
[all …]
Dbase.c37 struct nvkm_instmem *imem; member
68 nvkm_bar_flush(iobj->imem->subdev.device->bar); in nvkm_instobj_release()
100 spin_lock(&iobj->imem->lock); in nvkm_instobj_dtor()
102 spin_unlock(&iobj->imem->lock); in nvkm_instobj_dtor()
174 nvkm_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, in nvkm_instobj_new() argument
182 ret = imem->func->memory_new(imem, size, align, zero, &memory); in nvkm_instobj_new()
186 if (!imem->func->persistent) { in nvkm_instobj_new()
194 iobj->imem = imem; in nvkm_instobj_new()
195 spin_lock(&iobj->imem->lock); in nvkm_instobj_new()
196 list_add_tail(&iobj->head, &imem->list); in nvkm_instobj_new()
[all …]
Dnv50.c46 struct nv50_instmem *imem; member
74 struct nvkm_subdev *subdev = &iobj->imem->base.subdev; in nv50_instobj_boot()
101 struct nv50_instmem *imem = nv50_instobj(memory)->imem; in nv50_instobj_release() local
102 spin_unlock_irqrestore(&imem->lock, imem->lock_flags); in nv50_instobj_release()
109 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_acquire() local
110 struct nvkm_bar *bar = imem->base.subdev.device->bar; in nv50_instobj_acquire()
119 spin_lock_irqsave(&imem->lock, flags); in nv50_instobj_acquire()
120 imem->lock_flags = flags; in nv50_instobj_acquire()
128 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_rd32() local
129 struct nvkm_device *device = imem->base.subdev.device; in nv50_instobj_rd32()
[all …]
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Ddmanv40.c62 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_fini() local
76 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_fini()
77 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, 0x00000000); in nv40_fifo_dma_engine_fini()
78 nvkm_done(imem->ramfc); in nv40_fifo_dma_engine_fini()
92 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_init() local
107 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_init()
108 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, inst); in nv40_fifo_dma_engine_init()
109 nvkm_done(imem->ramfc); in nv40_fifo_dma_engine_init()
144 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv40_fifo_dma_object_ctor() local
160 hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, in nv40_fifo_dma_object_ctor()
[all …]
Ddmanv04.c38 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_dtor() local
39 nvkm_ramht_remove(imem->ramht, cookie); in nv04_fifo_dma_object_dtor()
47 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_ctor() local
63 hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, in nv04_fifo_dma_object_ctor()
75 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_fifo_dma_fini()
138 struct nvkm_instmem *imem = fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_dtor() local
141 nvkm_kmap(imem->ramfc); in nv04_fifo_dma_dtor()
143 nvkm_wo32(imem->ramfc, chan->ramfc + c->ctxp, 0x00000000); in nv04_fifo_dma_dtor()
145 nvkm_done(imem->ramfc); in nv04_fifo_dma_dtor()
169 struct nvkm_instmem *imem = device->imem; in nv04_fifo_dma_new() local
[all …]
Ddmanv10.c45 struct nvkm_instmem *imem = device->imem; in nv10_fifo_dma_new() local
75 nvkm_kmap(imem->ramfc); in nv10_fifo_dma_new()
76 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv10_fifo_dma_new()
77 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv10_fifo_dma_new()
78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv10_fifo_dma_new()
79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv10_fifo_dma_new()
86 nvkm_done(imem->ramfc); in nv10_fifo_dma_new()
Ddmanv17.c45 struct nvkm_instmem *imem = device->imem; in nv17_fifo_dma_new() local
76 nvkm_kmap(imem->ramfc); in nv17_fifo_dma_new()
77 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv17_fifo_dma_new()
78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv17_fifo_dma_new()
79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv17_fifo_dma_new()
80 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv17_fifo_dma_new()
87 nvkm_done(imem->ramfc); in nv17_fifo_dma_new()
Dnv17.c55 struct nvkm_instmem *imem = device->imem; in nv17_fifo_init() local
56 struct nvkm_ramht *ramht = imem->ramht; in nv17_fifo_init()
57 struct nvkm_memory *ramro = imem->ramro; in nv17_fifo_init()
58 struct nvkm_memory *ramfc = imem->ramfc; in nv17_fifo_init()
Dnv40.c65 struct nvkm_instmem *imem = device->imem; in nv40_fifo_init() local
66 struct nvkm_ramht *ramht = imem->ramht; in nv40_fifo_init()
67 struct nvkm_memory *ramro = imem->ramro; in nv40_fifo_init()
68 struct nvkm_memory *ramfc = imem->ramfc; in nv40_fifo_init()
Dnv04.c301 struct nvkm_instmem *imem = device->imem; in nv04_fifo_init() local
302 struct nvkm_ramht *ramht = imem->ramht; in nv04_fifo_init()
303 struct nvkm_memory *ramro = imem->ramro; in nv04_fifo_init()
304 struct nvkm_memory *ramfc = imem->ramfc; in nv04_fifo_init()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/mpeg/
Dnv40.c33 struct nvkm_instmem *imem = device->imem; in nv40_mpeg_mthd_dma() local
35 u32 dma0 = nvkm_instmem_rd32(imem, inst + 0); in nv40_mpeg_mthd_dma()
36 u32 dma1 = nvkm_instmem_rd32(imem, inst + 4); in nv40_mpeg_mthd_dma()
37 u32 dma2 = nvkm_instmem_rd32(imem, inst + 8); in nv40_mpeg_mthd_dma()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/core/
Dmemory.c51 struct nvkm_instmem *imem = device->imem; in nvkm_memory_new() local
55 if (unlikely(target != NVKM_MEM_TARGET_INST || !imem)) in nvkm_memory_new()
58 ret = nvkm_instobj_new(imem, size, align, zero, &memory); in nvkm_memory_new()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/device/
Dbase.c85 .imem = nv04_instmem_new,
106 .imem = nv04_instmem_new,
128 .imem = nv04_instmem_new,
148 .imem = nv04_instmem_new,
170 .imem = nv04_instmem_new,
192 .imem = nv04_instmem_new,
214 .imem = nv04_instmem_new,
236 .imem = nv04_instmem_new,
258 .imem = nv04_instmem_new,
280 .imem = nv04_instmem_new,
[all …]
Duser.c47 struct nvkm_instmem *imem = device->imem; in nvkm_udevice_info() local
115 if (imem && args->v0.ram_size > 0) in nvkm_udevice_info()
116 args->v0.ram_user = args->v0.ram_user - imem->reserved; in nvkm_udevice_info()
/linux-4.4.14/arch/blackfin/mach-common/
Dcache-c.c29 u32 imem = bfin_read_IMEM_CONTROL(); in blackfin_invalidate_entire_icache() local
30 bfin_write_IMEM_CONTROL(imem & ~0x4); in blackfin_invalidate_entire_icache()
32 bfin_write_IMEM_CONTROL(imem); in blackfin_invalidate_entire_icache()
/linux-4.4.14/arch/ia64/sn/kernel/
Dbte_error.c43 ii_imem_u_t imem; /* II IMEM Register */ in shub1_bte_error_handler() local
109 imem.ii_imem_regval = REMOTE_HUB_L(nasid, IIO_IMEM); in shub1_bte_error_handler()
110 imem.ii_imem_fld_s.i_b0_esd = imem.ii_imem_fld_s.i_b1_esd = 1; in shub1_bte_error_handler()
111 REMOTE_HUB_S(nasid, IIO_IMEM, imem.ii_imem_regval); in shub1_bte_error_handler()
/linux-4.4.14/drivers/gpu/drm/nouveau/include/nvkm/core/
Ddevice.h112 struct nvkm_instmem *imem; member
177 int (*imem )(struct nvkm_device *, int idx, struct nvkm_instmem **); member
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/dma/
Duser.c68 struct nvkm_instmem *instmem = device->imem; in nvkm_dmaobj_ctor()