Lines Matching refs:subdev

22 #define gk20a_pmu(p) container_of((p), struct gk20a_pmu, base.subdev)
54 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_target()
62 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_get_cur_state()
73 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_get_target_state()
88 nvkm_trace(&pmu->base.subdev, "cur level = %d, new level = %d\n", in gk20a_pmu_dvfs_get_target_state()
103 struct nvkm_device *device = pmu->base.subdev.device; in gk20a_pmu_dvfs_get_dev_status()
112 struct nvkm_device *device = pmu->base.subdev.device; in gk20a_pmu_dvfs_reset_dev_status()
124 struct nvkm_subdev *subdev = &pmu->base.subdev; in gk20a_pmu_dvfs_work() local
125 struct nvkm_device *device = subdev->device; in gk20a_pmu_dvfs_work()
141 nvkm_warn(subdev, "failed to get device status\n"); in gk20a_pmu_dvfs_work()
150 nvkm_trace(subdev, "utilization = %d %%, avg_load = %d %%\n", in gk20a_pmu_dvfs_work()
155 nvkm_warn(subdev, "failed to get current state\n"); in gk20a_pmu_dvfs_work()
160 nvkm_trace(subdev, "set new state to %d\n", state); in gk20a_pmu_dvfs_work()
170 gk20a_pmu_fini(struct nvkm_subdev *subdev, bool suspend) in gk20a_pmu_fini() argument
172 struct gk20a_pmu *pmu = gk20a_pmu(subdev); in gk20a_pmu_fini()
173 nvkm_timer_alarm_cancel(subdev->device->timer, &pmu->alarm); in gk20a_pmu_fini()
178 gk20a_pmu_dtor(struct nvkm_subdev *subdev) in gk20a_pmu_dtor() argument
180 return gk20a_pmu(subdev); in gk20a_pmu_dtor()
184 gk20a_pmu_init(struct nvkm_subdev *subdev) in gk20a_pmu_init() argument
186 struct gk20a_pmu *pmu = gk20a_pmu(subdev); in gk20a_pmu_init()
187 struct nvkm_device *device = pmu->base.subdev.device; in gk20a_pmu_init()
223 nvkm_subdev_ctor(&gk20a_pmu, device, index, 0, &pmu->base.subdev); in gk20a_pmu_new()