This source file includes following definitions.
- nvkm_fifo_chan_child_fini
- nvkm_fifo_chan_child_init
- nvkm_fifo_chan_child_del
- nvkm_fifo_chan_child_new
- nvkm_fifo_chan_child_get
- nvkm_fifo_chan_ntfy
- nvkm_fifo_chan_map
- nvkm_fifo_chan_rd32
- nvkm_fifo_chan_wr32
- nvkm_fifo_chan_fini
- nvkm_fifo_chan_init
- nvkm_fifo_chan_dtor
- nvkm_fifo_chan_ctor
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24 #include "chan.h"
25
26 #include <core/client.h>
27 #include <core/gpuobj.h>
28 #include <core/oproxy.h>
29 #include <subdev/mmu.h>
30 #include <engine/dma.h>
31
32 struct nvkm_fifo_chan_object {
33 struct nvkm_oproxy oproxy;
34 struct nvkm_fifo_chan *chan;
35 int hash;
36 };
37
38 static int
39 nvkm_fifo_chan_child_fini(struct nvkm_oproxy *base, bool suspend)
40 {
41 struct nvkm_fifo_chan_object *object =
42 container_of(base, typeof(*object), oproxy);
43 struct nvkm_engine *engine = object->oproxy.object->engine;
44 struct nvkm_fifo_chan *chan = object->chan;
45 struct nvkm_fifo_engn *engn = &chan->engn[engine->subdev.index];
46 const char *name = nvkm_subdev_name[engine->subdev.index];
47 int ret = 0;
48
49 if (--engn->usecount)
50 return 0;
51
52 if (chan->func->engine_fini) {
53 ret = chan->func->engine_fini(chan, engine, suspend);
54 if (ret) {
55 nvif_error(&chan->object,
56 "detach %s failed, %d\n", name, ret);
57 return ret;
58 }
59 }
60
61 if (engn->object) {
62 ret = nvkm_object_fini(engn->object, suspend);
63 if (ret && suspend)
64 return ret;
65 }
66
67 nvif_trace(&chan->object, "detached %s\n", name);
68 return ret;
69 }
70
71 static int
72 nvkm_fifo_chan_child_init(struct nvkm_oproxy *base)
73 {
74 struct nvkm_fifo_chan_object *object =
75 container_of(base, typeof(*object), oproxy);
76 struct nvkm_engine *engine = object->oproxy.object->engine;
77 struct nvkm_fifo_chan *chan = object->chan;
78 struct nvkm_fifo_engn *engn = &chan->engn[engine->subdev.index];
79 const char *name = nvkm_subdev_name[engine->subdev.index];
80 int ret;
81
82 if (engn->usecount++)
83 return 0;
84
85 if (engn->object) {
86 ret = nvkm_object_init(engn->object);
87 if (ret)
88 return ret;
89 }
90
91 if (chan->func->engine_init) {
92 ret = chan->func->engine_init(chan, engine);
93 if (ret) {
94 nvif_error(&chan->object,
95 "attach %s failed, %d\n", name, ret);
96 return ret;
97 }
98 }
99
100 nvif_trace(&chan->object, "attached %s\n", name);
101 return 0;
102 }
103
104 static void
105 nvkm_fifo_chan_child_del(struct nvkm_oproxy *base)
106 {
107 struct nvkm_fifo_chan_object *object =
108 container_of(base, typeof(*object), oproxy);
109 struct nvkm_engine *engine = object->oproxy.base.engine;
110 struct nvkm_fifo_chan *chan = object->chan;
111 struct nvkm_fifo_engn *engn = &chan->engn[engine->subdev.index];
112
113 if (chan->func->object_dtor)
114 chan->func->object_dtor(chan, object->hash);
115
116 if (!--engn->refcount) {
117 if (chan->func->engine_dtor)
118 chan->func->engine_dtor(chan, engine);
119 nvkm_object_del(&engn->object);
120 if (chan->vmm)
121 atomic_dec(&chan->vmm->engref[engine->subdev.index]);
122 }
123 }
124
125 static const struct nvkm_oproxy_func
126 nvkm_fifo_chan_child_func = {
127 .dtor[0] = nvkm_fifo_chan_child_del,
128 .init[0] = nvkm_fifo_chan_child_init,
129 .fini[0] = nvkm_fifo_chan_child_fini,
130 };
131
132 static int
133 nvkm_fifo_chan_child_new(const struct nvkm_oclass *oclass, void *data, u32 size,
134 struct nvkm_object **pobject)
135 {
136 struct nvkm_engine *engine = oclass->engine;
137 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(oclass->parent);
138 struct nvkm_fifo_engn *engn = &chan->engn[engine->subdev.index];
139 struct nvkm_fifo_chan_object *object;
140 int ret = 0;
141
142 if (!(object = kzalloc(sizeof(*object), GFP_KERNEL)))
143 return -ENOMEM;
144 nvkm_oproxy_ctor(&nvkm_fifo_chan_child_func, oclass, &object->oproxy);
145 object->chan = chan;
146 *pobject = &object->oproxy.base;
147
148 if (!engn->refcount++) {
149 struct nvkm_oclass cclass = {
150 .client = oclass->client,
151 .engine = oclass->engine,
152 };
153
154 if (chan->vmm)
155 atomic_inc(&chan->vmm->engref[engine->subdev.index]);
156
157 if (engine->func->fifo.cclass) {
158 ret = engine->func->fifo.cclass(chan, &cclass,
159 &engn->object);
160 } else
161 if (engine->func->cclass) {
162 ret = nvkm_object_new_(engine->func->cclass, &cclass,
163 NULL, 0, &engn->object);
164 }
165 if (ret)
166 return ret;
167
168 if (chan->func->engine_ctor) {
169 ret = chan->func->engine_ctor(chan, oclass->engine,
170 engn->object);
171 if (ret)
172 return ret;
173 }
174 }
175
176 ret = oclass->base.ctor(&(const struct nvkm_oclass) {
177 .base = oclass->base,
178 .engn = oclass->engn,
179 .handle = oclass->handle,
180 .object = oclass->object,
181 .client = oclass->client,
182 .parent = engn->object ?
183 engn->object :
184 oclass->parent,
185 .engine = engine,
186 }, data, size, &object->oproxy.object);
187 if (ret)
188 return ret;
189
190 if (chan->func->object_ctor) {
191 object->hash =
192 chan->func->object_ctor(chan, object->oproxy.object);
193 if (object->hash < 0)
194 return object->hash;
195 }
196
197 return 0;
198 }
199
200 static int
201 nvkm_fifo_chan_child_get(struct nvkm_object *object, int index,
202 struct nvkm_oclass *oclass)
203 {
204 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
205 struct nvkm_fifo *fifo = chan->fifo;
206 struct nvkm_device *device = fifo->engine.subdev.device;
207 struct nvkm_engine *engine;
208 u64 mask = chan->engines;
209 int ret, i, c;
210
211 for (; c = 0, i = __ffs64(mask), mask; mask &= ~(1ULL << i)) {
212 if (!(engine = nvkm_device_engine(device, i)))
213 continue;
214 oclass->engine = engine;
215 oclass->base.oclass = 0;
216
217 if (engine->func->fifo.sclass) {
218 ret = engine->func->fifo.sclass(oclass, index);
219 if (oclass->base.oclass) {
220 if (!oclass->base.ctor)
221 oclass->base.ctor = nvkm_object_new;
222 oclass->ctor = nvkm_fifo_chan_child_new;
223 return 0;
224 }
225
226 index -= ret;
227 continue;
228 }
229
230 while (engine->func->sclass[c].oclass) {
231 if (c++ == index) {
232 oclass->base = engine->func->sclass[index];
233 if (!oclass->base.ctor)
234 oclass->base.ctor = nvkm_object_new;
235 oclass->ctor = nvkm_fifo_chan_child_new;
236 return 0;
237 }
238 }
239 index -= c;
240 }
241
242 return -EINVAL;
243 }
244
245 static int
246 nvkm_fifo_chan_ntfy(struct nvkm_object *object, u32 type,
247 struct nvkm_event **pevent)
248 {
249 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
250 if (chan->func->ntfy)
251 return chan->func->ntfy(chan, type, pevent);
252 return -ENODEV;
253 }
254
255 static int
256 nvkm_fifo_chan_map(struct nvkm_object *object, void *argv, u32 argc,
257 enum nvkm_object_map *type, u64 *addr, u64 *size)
258 {
259 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
260 *type = NVKM_OBJECT_MAP_IO;
261 *addr = chan->addr;
262 *size = chan->size;
263 return 0;
264 }
265
266 static int
267 nvkm_fifo_chan_rd32(struct nvkm_object *object, u64 addr, u32 *data)
268 {
269 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
270 if (unlikely(!chan->user)) {
271 chan->user = ioremap(chan->addr, chan->size);
272 if (!chan->user)
273 return -ENOMEM;
274 }
275 if (unlikely(addr + 4 > chan->size))
276 return -EINVAL;
277 *data = ioread32_native(chan->user + addr);
278 return 0;
279 }
280
281 static int
282 nvkm_fifo_chan_wr32(struct nvkm_object *object, u64 addr, u32 data)
283 {
284 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
285 if (unlikely(!chan->user)) {
286 chan->user = ioremap(chan->addr, chan->size);
287 if (!chan->user)
288 return -ENOMEM;
289 }
290 if (unlikely(addr + 4 > chan->size))
291 return -EINVAL;
292 iowrite32_native(data, chan->user + addr);
293 return 0;
294 }
295
296 static int
297 nvkm_fifo_chan_fini(struct nvkm_object *object, bool suspend)
298 {
299 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
300 chan->func->fini(chan);
301 return 0;
302 }
303
304 static int
305 nvkm_fifo_chan_init(struct nvkm_object *object)
306 {
307 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
308 chan->func->init(chan);
309 return 0;
310 }
311
312 static void *
313 nvkm_fifo_chan_dtor(struct nvkm_object *object)
314 {
315 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
316 struct nvkm_fifo *fifo = chan->fifo;
317 void *data = chan->func->dtor(chan);
318 unsigned long flags;
319
320 spin_lock_irqsave(&fifo->lock, flags);
321 if (!list_empty(&chan->head)) {
322 __clear_bit(chan->chid, fifo->mask);
323 list_del(&chan->head);
324 }
325 spin_unlock_irqrestore(&fifo->lock, flags);
326
327 if (chan->user)
328 iounmap(chan->user);
329
330 if (chan->vmm) {
331 nvkm_vmm_part(chan->vmm, chan->inst->memory);
332 nvkm_vmm_unref(&chan->vmm);
333 }
334
335 nvkm_gpuobj_del(&chan->push);
336 nvkm_gpuobj_del(&chan->inst);
337 return data;
338 }
339
340 static const struct nvkm_object_func
341 nvkm_fifo_chan_func = {
342 .dtor = nvkm_fifo_chan_dtor,
343 .init = nvkm_fifo_chan_init,
344 .fini = nvkm_fifo_chan_fini,
345 .ntfy = nvkm_fifo_chan_ntfy,
346 .map = nvkm_fifo_chan_map,
347 .rd32 = nvkm_fifo_chan_rd32,
348 .wr32 = nvkm_fifo_chan_wr32,
349 .sclass = nvkm_fifo_chan_child_get,
350 };
351
352 int
353 nvkm_fifo_chan_ctor(const struct nvkm_fifo_chan_func *func,
354 struct nvkm_fifo *fifo, u32 size, u32 align, bool zero,
355 u64 hvmm, u64 push, u64 engines, int bar, u32 base,
356 u32 user, const struct nvkm_oclass *oclass,
357 struct nvkm_fifo_chan *chan)
358 {
359 struct nvkm_client *client = oclass->client;
360 struct nvkm_device *device = fifo->engine.subdev.device;
361 struct nvkm_dmaobj *dmaobj;
362 unsigned long flags;
363 int ret;
364
365 nvkm_object_ctor(&nvkm_fifo_chan_func, oclass, &chan->object);
366 chan->func = func;
367 chan->fifo = fifo;
368 chan->engines = engines;
369 INIT_LIST_HEAD(&chan->head);
370
371
372 ret = nvkm_gpuobj_new(device, size, align, zero, NULL, &chan->inst);
373 if (ret)
374 return ret;
375
376
377 if (push) {
378 dmaobj = nvkm_dmaobj_search(client, push);
379 if (IS_ERR(dmaobj))
380 return PTR_ERR(dmaobj);
381
382 ret = nvkm_object_bind(&dmaobj->object, chan->inst, -16,
383 &chan->push);
384 if (ret)
385 return ret;
386 }
387
388
389 if (hvmm) {
390 struct nvkm_vmm *vmm = nvkm_uvmm_search(client, hvmm);
391 if (IS_ERR(vmm))
392 return PTR_ERR(vmm);
393
394 if (vmm->mmu != device->mmu)
395 return -EINVAL;
396
397 ret = nvkm_vmm_join(vmm, chan->inst->memory);
398 if (ret)
399 return ret;
400
401 chan->vmm = nvkm_vmm_ref(vmm);
402 }
403
404
405 spin_lock_irqsave(&fifo->lock, flags);
406 chan->chid = find_first_zero_bit(fifo->mask, NVKM_FIFO_CHID_NR);
407 if (chan->chid >= NVKM_FIFO_CHID_NR) {
408 spin_unlock_irqrestore(&fifo->lock, flags);
409 return -ENOSPC;
410 }
411 list_add(&chan->head, &fifo->chan);
412 __set_bit(chan->chid, fifo->mask);
413 spin_unlock_irqrestore(&fifo->lock, flags);
414
415
416 chan->addr = device->func->resource_addr(device, bar) +
417 base + user * chan->chid;
418 chan->size = user;
419
420 nvkm_fifo_cevent(fifo);
421 return 0;
422 }