zapped           2795 arch/x86/kvm/mmu.c 	int i, zapped = 0;
zapped           2808 arch/x86/kvm/mmu.c 			zapped++;
zapped           2812 arch/x86/kvm/mmu.c 	return zapped;
zapped            315 kernel/locking/lockdep.c 	struct list_head zapped;
zapped            972 kernel/locking/lockdep.c 		if (in_list(&class->lock_entry, &pf->zapped))
zapped           1095 kernel/locking/lockdep.c 	INIT_LIST_HEAD(&delayed_free.pf[0].zapped);
zapped           1096 kernel/locking/lockdep.c 	INIT_LIST_HEAD(&delayed_free.pf[1].zapped);
zapped           4872 kernel/locking/lockdep.c 		list_move_tail(&class->lock_entry, &pf->zapped);
zapped           4926 kernel/locking/lockdep.c 	if (list_empty(&pf->zapped))
zapped           4947 kernel/locking/lockdep.c 	list_for_each_entry(class, &pf->zapped, lock_entry)
zapped           4950 kernel/locking/lockdep.c 	list_splice_init(&pf->zapped, &free_lock_classes);
zapped            601 net/qrtr/qrtr.c 		       const struct sockaddr_qrtr *addr, int zapped)
zapped            609 net/qrtr/qrtr.c 	if (!zapped && addr->sq_port == ipc->us.sq_port)
zapped            618 net/qrtr/qrtr.c 	if (!zapped)