Lines Matching refs:lp
1906 struct preempted_vcore_list *lp = &per_cpu(preempted_vcores, cpu); in init_vcore_lists() local
1907 spin_lock_init(&lp->lock); in init_vcore_lists()
1908 INIT_LIST_HEAD(&lp->list); in init_vcore_lists()
1914 struct preempted_vcore_list *lp = this_cpu_ptr(&preempted_vcores); in kvmppc_vcore_preempt() local
1919 spin_lock(&lp->lock); in kvmppc_vcore_preempt()
1920 list_add_tail(&vc->preempt_list, &lp->list); in kvmppc_vcore_preempt()
1921 spin_unlock(&lp->lock); in kvmppc_vcore_preempt()
1930 struct preempted_vcore_list *lp; in kvmppc_vcore_end_preempt() local
1934 lp = &per_cpu(preempted_vcores, vc->pcpu); in kvmppc_vcore_end_preempt()
1935 spin_lock(&lp->lock); in kvmppc_vcore_end_preempt()
1937 spin_unlock(&lp->lock); in kvmppc_vcore_end_preempt()
2182 struct preempted_vcore_list *lp = this_cpu_ptr(&preempted_vcores); in collect_piggybacks() local
2185 spin_lock(&lp->lock); in collect_piggybacks()
2186 list_for_each_entry_safe(pvc, vcnext, &lp->list, preempt_list) { in collect_piggybacks()
2208 spin_unlock(&lp->lock); in collect_piggybacks()