Lines Matching refs:vbl_ctrl
125 struct msm_vblank_ctrl *vbl_ctrl = container_of(work, in vblank_ctrl_worker() local
127 struct msm_drm_private *priv = container_of(vbl_ctrl, in vblank_ctrl_worker()
133 spin_lock_irqsave(&vbl_ctrl->lock, flags); in vblank_ctrl_worker()
134 list_for_each_entry_safe(vbl_ev, tmp, &vbl_ctrl->event_list, node) { in vblank_ctrl_worker()
136 spin_unlock_irqrestore(&vbl_ctrl->lock, flags); in vblank_ctrl_worker()
147 spin_lock_irqsave(&vbl_ctrl->lock, flags); in vblank_ctrl_worker()
150 spin_unlock_irqrestore(&vbl_ctrl->lock, flags); in vblank_ctrl_worker()
156 struct msm_vblank_ctrl *vbl_ctrl = &priv->vblank_ctrl; in vblank_ctrl_queue_work() local
167 spin_lock_irqsave(&vbl_ctrl->lock, flags); in vblank_ctrl_queue_work()
168 list_add_tail(&vbl_ev->node, &vbl_ctrl->event_list); in vblank_ctrl_queue_work()
169 spin_unlock_irqrestore(&vbl_ctrl->lock, flags); in vblank_ctrl_queue_work()
171 queue_work(priv->wq, &vbl_ctrl->work); in vblank_ctrl_queue_work()
185 struct msm_vblank_ctrl *vbl_ctrl = &priv->vblank_ctrl; in msm_unload() local
192 cancel_work_sync(&vbl_ctrl->work); in msm_unload()
193 list_for_each_entry_safe(vbl_ev, tmp, &vbl_ctrl->event_list, node) { in msm_unload()