zapped 2795 arch/x86/kvm/mmu.c int i, zapped = 0; zapped 2808 arch/x86/kvm/mmu.c zapped++; zapped 2812 arch/x86/kvm/mmu.c return zapped; zapped 315 kernel/locking/lockdep.c struct list_head zapped; zapped 972 kernel/locking/lockdep.c if (in_list(&class->lock_entry, &pf->zapped)) zapped 1095 kernel/locking/lockdep.c INIT_LIST_HEAD(&delayed_free.pf[0].zapped); zapped 1096 kernel/locking/lockdep.c INIT_LIST_HEAD(&delayed_free.pf[1].zapped); zapped 4872 kernel/locking/lockdep.c list_move_tail(&class->lock_entry, &pf->zapped); zapped 4926 kernel/locking/lockdep.c if (list_empty(&pf->zapped)) zapped 4947 kernel/locking/lockdep.c list_for_each_entry(class, &pf->zapped, lock_entry) zapped 4950 kernel/locking/lockdep.c list_splice_init(&pf->zapped, &free_lock_classes); zapped 601 net/qrtr/qrtr.c const struct sockaddr_qrtr *addr, int zapped) zapped 609 net/qrtr/qrtr.c if (!zapped && addr->sq_port == ipc->us.sq_port) zapped 618 net/qrtr/qrtr.c if (!zapped)