1 /*
2  * Copyright 2012 Red Hat Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  * Authors: Ben Skeggs
23  */
24 #include <core/engctx.h>
25 #include <core/engine.h>
26 #include <core/client.h>
27 
28 static inline int
nvkm_engctx_exists(struct nvkm_object * parent,struct nvkm_engine * engine,void ** pobject)29 nvkm_engctx_exists(struct nvkm_object *parent,
30 		   struct nvkm_engine *engine, void **pobject)
31 {
32 	struct nvkm_engctx *engctx;
33 	struct nvkm_object *parctx;
34 
35 	list_for_each_entry(engctx, &engine->contexts, head) {
36 		parctx = nv_pclass(nv_object(engctx), NV_PARENT_CLASS);
37 		if (parctx == parent) {
38 			atomic_inc(&nv_object(engctx)->refcount);
39 			*pobject = engctx;
40 			return 1;
41 		}
42 	}
43 
44 	return 0;
45 }
46 
47 int
nvkm_engctx_create_(struct nvkm_object * parent,struct nvkm_object * engobj,struct nvkm_oclass * oclass,struct nvkm_object * pargpu,u32 size,u32 align,u32 flags,int length,void ** pobject)48 nvkm_engctx_create_(struct nvkm_object *parent, struct nvkm_object *engobj,
49 		    struct nvkm_oclass *oclass, struct nvkm_object *pargpu,
50 		    u32 size, u32 align, u32 flags, int length, void **pobject)
51 {
52 	struct nvkm_client *client = nvkm_client(parent);
53 	struct nvkm_engine *engine = nv_engine(engobj);
54 	struct nvkm_object *engctx;
55 	unsigned long save;
56 	int ret;
57 
58 	/* check if this engine already has a context for the parent object,
59 	 * and reference it instead of creating a new one
60 	 */
61 	spin_lock_irqsave(&engine->lock, save);
62 	ret = nvkm_engctx_exists(parent, engine, pobject);
63 	spin_unlock_irqrestore(&engine->lock, save);
64 	if (ret)
65 		return ret;
66 
67 	/* create the new context, supports creating both raw objects and
68 	 * objects backed by instance memory
69 	 */
70 	if (size) {
71 		ret = nvkm_gpuobj_create_(parent, engobj, oclass,
72 					  NV_ENGCTX_CLASS, pargpu, size,
73 					  align, flags, length, pobject);
74 	} else {
75 		ret = nvkm_object_create_(parent, engobj, oclass,
76 					  NV_ENGCTX_CLASS, length, pobject);
77 	}
78 
79 	engctx = *pobject;
80 	if (ret)
81 		return ret;
82 
83 	/* must take the lock again and re-check a context doesn't already
84 	 * exist (in case of a race) - the lock had to be dropped before as
85 	 * it's not possible to allocate the object with it held.
86 	 */
87 	spin_lock_irqsave(&engine->lock, save);
88 	ret = nvkm_engctx_exists(parent, engine, pobject);
89 	if (ret) {
90 		spin_unlock_irqrestore(&engine->lock, save);
91 		nvkm_object_ref(NULL, &engctx);
92 		return ret;
93 	}
94 
95 	if (client->vm)
96 		atomic_inc(&client->vm->engref[nv_engidx(engine)]);
97 	list_add(&nv_engctx(engctx)->head, &engine->contexts);
98 	nv_engctx(engctx)->addr = ~0ULL;
99 	spin_unlock_irqrestore(&engine->lock, save);
100 	return 0;
101 }
102 
103 void
nvkm_engctx_destroy(struct nvkm_engctx * engctx)104 nvkm_engctx_destroy(struct nvkm_engctx *engctx)
105 {
106 	struct nvkm_engine *engine = engctx->gpuobj.object.engine;
107 	struct nvkm_client *client = nvkm_client(engctx);
108 	unsigned long save;
109 
110 	nvkm_gpuobj_unmap(&engctx->vma);
111 	spin_lock_irqsave(&engine->lock, save);
112 	list_del(&engctx->head);
113 	spin_unlock_irqrestore(&engine->lock, save);
114 
115 	if (client->vm)
116 		atomic_dec(&client->vm->engref[nv_engidx(engine)]);
117 
118 	if (engctx->gpuobj.size)
119 		nvkm_gpuobj_destroy(&engctx->gpuobj);
120 	else
121 		nvkm_object_destroy(&engctx->gpuobj.object);
122 }
123 
124 int
nvkm_engctx_init(struct nvkm_engctx * engctx)125 nvkm_engctx_init(struct nvkm_engctx *engctx)
126 {
127 	struct nvkm_object *object = nv_object(engctx);
128 	struct nvkm_subdev *subdev = nv_subdev(object->engine);
129 	struct nvkm_object *parent;
130 	struct nvkm_subdev *pardev;
131 	int ret;
132 
133 	ret = nvkm_gpuobj_init(&engctx->gpuobj);
134 	if (ret)
135 		return ret;
136 
137 	parent = nv_pclass(object->parent, NV_PARENT_CLASS);
138 	pardev = nv_subdev(parent->engine);
139 	if (nv_parent(parent)->context_attach) {
140 		mutex_lock(&pardev->mutex);
141 		ret = nv_parent(parent)->context_attach(parent, object);
142 		mutex_unlock(&pardev->mutex);
143 	}
144 
145 	if (ret) {
146 		nv_error(parent, "failed to attach %s context, %d\n",
147 			 subdev->name, ret);
148 		return ret;
149 	}
150 
151 	nv_debug(parent, "attached %s context\n", subdev->name);
152 	return 0;
153 }
154 
155 int
nvkm_engctx_fini(struct nvkm_engctx * engctx,bool suspend)156 nvkm_engctx_fini(struct nvkm_engctx *engctx, bool suspend)
157 {
158 	struct nvkm_object *object = nv_object(engctx);
159 	struct nvkm_subdev *subdev = nv_subdev(object->engine);
160 	struct nvkm_object *parent;
161 	struct nvkm_subdev *pardev;
162 	int ret = 0;
163 
164 	parent = nv_pclass(object->parent, NV_PARENT_CLASS);
165 	pardev = nv_subdev(parent->engine);
166 	if (nv_parent(parent)->context_detach) {
167 		mutex_lock(&pardev->mutex);
168 		ret = nv_parent(parent)->context_detach(parent, suspend, object);
169 		mutex_unlock(&pardev->mutex);
170 	}
171 
172 	if (ret) {
173 		nv_error(parent, "failed to detach %s context, %d\n",
174 			 subdev->name, ret);
175 		return ret;
176 	}
177 
178 	nv_debug(parent, "detached %s context\n", subdev->name);
179 	return nvkm_gpuobj_fini(&engctx->gpuobj, suspend);
180 }
181 
182 int
_nvkm_engctx_ctor(struct nvkm_object * parent,struct nvkm_object * engine,struct nvkm_oclass * oclass,void * data,u32 size,struct nvkm_object ** pobject)183 _nvkm_engctx_ctor(struct nvkm_object *parent, struct nvkm_object *engine,
184 		  struct nvkm_oclass *oclass, void *data, u32 size,
185 		  struct nvkm_object **pobject)
186 {
187 	struct nvkm_engctx *engctx;
188 	int ret;
189 
190 	ret = nvkm_engctx_create(parent, engine, oclass, NULL, 256, 256,
191 				 NVOBJ_FLAG_ZERO_ALLOC, &engctx);
192 	*pobject = nv_object(engctx);
193 	return ret;
194 }
195 
196 void
_nvkm_engctx_dtor(struct nvkm_object * object)197 _nvkm_engctx_dtor(struct nvkm_object *object)
198 {
199 	nvkm_engctx_destroy(nv_engctx(object));
200 }
201 
202 int
_nvkm_engctx_init(struct nvkm_object * object)203 _nvkm_engctx_init(struct nvkm_object *object)
204 {
205 	return nvkm_engctx_init(nv_engctx(object));
206 }
207 
208 int
_nvkm_engctx_fini(struct nvkm_object * object,bool suspend)209 _nvkm_engctx_fini(struct nvkm_object *object, bool suspend)
210 {
211 	return nvkm_engctx_fini(nv_engctx(object), suspend);
212 }
213 
214 struct nvkm_object *
nvkm_engctx_get(struct nvkm_engine * engine,u64 addr)215 nvkm_engctx_get(struct nvkm_engine *engine, u64 addr)
216 {
217 	struct nvkm_engctx *engctx;
218 	unsigned long flags;
219 
220 	spin_lock_irqsave(&engine->lock, flags);
221 	list_for_each_entry(engctx, &engine->contexts, head) {
222 		if (engctx->addr == addr) {
223 			engctx->save = flags;
224 			return nv_object(engctx);
225 		}
226 	}
227 	spin_unlock_irqrestore(&engine->lock, flags);
228 	return NULL;
229 }
230 
231 void
nvkm_engctx_put(struct nvkm_object * object)232 nvkm_engctx_put(struct nvkm_object *object)
233 {
234 	if (object) {
235 		struct nvkm_engine *engine = nv_engine(object->engine);
236 		struct nvkm_engctx *engctx = nv_engctx(object);
237 		spin_unlock_irqrestore(&engine->lock, engctx->save);
238 	}
239 }
240