sk_lock            65 include/net/llc.h 	spinlock_t sk_lock;
sk_lock           364 include/net/sock.h 	socket_lock_t		sk_lock;
sk_lock          1493 include/net/sock.h 	if (sk->sk_lock.owned) {
sk_lock          1494 include/net/sock.h 		sk->sk_lock.owned = 0;
sk_lock          1497 include/net/sock.h 		mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_);
sk_lock          1510 include/net/sock.h 	sk->sk_lock.owned = 0;						\
sk_lock          1511 include/net/sock.h 	init_waitqueue_head(&sk->sk_lock.wq);				\
sk_lock          1512 include/net/sock.h 	spin_lock_init(&(sk)->sk_lock.slock);				\
sk_lock          1513 include/net/sock.h 	debug_check_no_locks_freed((void *)&(sk)->sk_lock,		\
sk_lock          1514 include/net/sock.h 			sizeof((sk)->sk_lock));				\
sk_lock          1515 include/net/sock.h 	lockdep_set_class_and_name(&(sk)->sk_lock.slock,		\
sk_lock          1517 include/net/sock.h 	lockdep_init_map(&(sk)->sk_lock.dep_map, (name), (key), 0);	\
sk_lock          1523 include/net/sock.h 	return lockdep_is_held(&sk->sk_lock) ||
sk_lock          1524 include/net/sock.h 	       lockdep_is_held(&sk->sk_lock.slock);
sk_lock          1539 include/net/sock.h #define bh_lock_sock(__sk)	spin_lock(&((__sk)->sk_lock.slock))
sk_lock          1541 include/net/sock.h 				spin_lock_nested(&((__sk)->sk_lock.slock), \
sk_lock          1543 include/net/sock.h #define bh_unlock_sock(__sk)	spin_unlock(&((__sk)->sk_lock.slock))
sk_lock          1559 include/net/sock.h 		spin_unlock_bh(&sk->sk_lock.slock);
sk_lock          1586 include/net/sock.h 	return sk->sk_lock.owned;
sk_lock          1591 include/net/sock.h 	return sk->sk_lock.owned;
sk_lock          1599 include/net/sock.h 	return !sk->sk_lock.owned && !spin_is_locked(&sk->sk_lock.slock);
sk_lock           472 net/ceph/messenger.c 	lockdep_set_class(&sock->sk->sk_lock, &socket_class);
sk_lock           520 net/core/sock.c 		mutex_acquire(&sk->sk_lock.dep_map, 0, 1, _RET_IP_);
sk_lock           524 net/core/sock.c 		mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_);
sk_lock          2409 net/core/sock.c 	__releases(&sk->sk_lock.slock)
sk_lock          2410 net/core/sock.c 	__acquires(&sk->sk_lock.slock)
sk_lock          2415 net/core/sock.c 		prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait,
sk_lock          2417 net/core/sock.c 		spin_unlock_bh(&sk->sk_lock.slock);
sk_lock          2419 net/core/sock.c 		spin_lock_bh(&sk->sk_lock.slock);
sk_lock          2423 net/core/sock.c 	finish_wait(&sk->sk_lock.wq, &wait);
sk_lock          2427 net/core/sock.c 	__releases(&sk->sk_lock.slock)
sk_lock          2428 net/core/sock.c 	__acquires(&sk->sk_lock.slock)
sk_lock          2435 net/core/sock.c 		spin_unlock_bh(&sk->sk_lock.slock);
sk_lock          2449 net/core/sock.c 		spin_lock_bh(&sk->sk_lock.slock);
sk_lock          2461 net/core/sock.c 	spin_lock_bh(&sk->sk_lock.slock);
sk_lock          2463 net/core/sock.c 	spin_unlock_bh(&sk->sk_lock.slock);
sk_lock          2941 net/core/sock.c 	spin_lock_bh(&sk->sk_lock.slock);
sk_lock          2942 net/core/sock.c 	if (sk->sk_lock.owned)
sk_lock          2944 net/core/sock.c 	sk->sk_lock.owned = 1;
sk_lock          2945 net/core/sock.c 	spin_unlock(&sk->sk_lock.slock);
sk_lock          2949 net/core/sock.c 	mutex_acquire(&sk->sk_lock.dep_map, subclass, 0, _RET_IP_);
sk_lock          2956 net/core/sock.c 	spin_lock_bh(&sk->sk_lock.slock);
sk_lock          2967 net/core/sock.c 	if (waitqueue_active(&sk->sk_lock.wq))
sk_lock          2968 net/core/sock.c 		wake_up(&sk->sk_lock.wq);
sk_lock          2969 net/core/sock.c 	spin_unlock_bh(&sk->sk_lock.slock);
sk_lock          2989 net/core/sock.c 	spin_lock_bh(&sk->sk_lock.slock);
sk_lock          2991 net/core/sock.c 	if (!sk->sk_lock.owned)
sk_lock          2998 net/core/sock.c 	sk->sk_lock.owned = 1;
sk_lock          2999 net/core/sock.c 	spin_unlock(&sk->sk_lock.slock);
sk_lock          3003 net/core/sock.c 	mutex_acquire(&sk->sk_lock.dep_map, 0, 0, _RET_IP_);
sk_lock            80 net/core/sock_map.c 	__acquires(&sk->sk_lock.slock)
sk_lock            88 net/core/sock_map.c 	__releases(&sk->sk_lock.slock)
sk_lock           214 net/ipv4/icmp.c 	if (unlikely(!spin_trylock(&sk->sk_lock.slock))) {
sk_lock           225 net/ipv4/icmp.c 	spin_unlock(&sk->sk_lock.slock);
sk_lock           879 net/ipv4/inet_connection_sock.c 	__releases(&sk->sk_lock.slock)
sk_lock           117 net/ipv6/icmp.c 	if (unlikely(!spin_trylock(&sk->sk_lock.slock))) {
sk_lock           129 net/ipv6/icmp.c 	spin_unlock(&sk->sk_lock.slock);
sk_lock          1535 net/l2tp/l2tp_core.c 	lockdep_set_class_and_name(&sk->sk_lock.slock, &l2tp_socket_class,
sk_lock           119 net/l2tp/l2tp_ppp.c 	struct mutex		sk_lock;	/* Protects .sk */
sk_lock           458 net/l2tp/l2tp_ppp.c 		mutex_lock(&ps->sk_lock);
sk_lock           460 net/l2tp/l2tp_ppp.c 						     lockdep_is_held(&ps->sk_lock));
sk_lock           462 net/l2tp/l2tp_ppp.c 		mutex_unlock(&ps->sk_lock);
sk_lock           555 net/l2tp/l2tp_ppp.c 	mutex_init(&ps->sk_lock);
sk_lock           767 net/l2tp/l2tp_ppp.c 		mutex_lock(&ps->sk_lock);
sk_lock           769 net/l2tp/l2tp_ppp.c 					      lockdep_is_held(&ps->sk_lock)) ||
sk_lock           771 net/l2tp/l2tp_ppp.c 			mutex_unlock(&ps->sk_lock);
sk_lock           790 net/l2tp/l2tp_ppp.c 		mutex_lock(&ps->sk_lock);
sk_lock           793 net/l2tp/l2tp_ppp.c 			mutex_unlock(&ps->sk_lock);
sk_lock           824 net/l2tp/l2tp_ppp.c 		mutex_unlock(&ps->sk_lock);
sk_lock           832 net/l2tp/l2tp_ppp.c 	mutex_unlock(&ps->sk_lock);
sk_lock           703 net/llc/llc_conn.c 	spin_lock_bh(&sap->sk_lock);
sk_lock           708 net/llc/llc_conn.c 	spin_unlock_bh(&sap->sk_lock);
sk_lock           723 net/llc/llc_conn.c 	spin_lock_bh(&sap->sk_lock);
sk_lock           727 net/llc/llc_conn.c 	spin_unlock_bh(&sap->sk_lock);
sk_lock            41 net/llc/llc_core.c 		spin_lock_init(&sap->sk_lock);
sk_lock            41 net/llc/llc_proc.c 		spin_lock_bh(&sap->sk_lock);
sk_lock            52 net/llc/llc_proc.c 		spin_unlock_bh(&sap->sk_lock);
sk_lock           102 net/llc/llc_proc.c 	spin_unlock_bh(&sap->sk_lock);
sk_lock           104 net/llc/llc_proc.c 		spin_lock_bh(&sap->sk_lock);
sk_lock           108 net/llc/llc_proc.c 		spin_unlock_bh(&sap->sk_lock);
sk_lock           121 net/llc/llc_proc.c 		spin_unlock_bh(&sap->sk_lock);
sk_lock           198 net/llc/llc_proc.c 		   !!sk->sk_backlog.tail, !!sk->sk_lock.owned);
sk_lock           396 net/llc/llc_sap.c 	spin_lock_bh(&sap->sk_lock);
sk_lock           412 net/llc/llc_sap.c 	spin_unlock_bh(&sap->sk_lock);
sk_lock           463 net/rxrpc/call_accept.c 	__releases(&rx->sk.sk_lock.slock)
sk_lock           222 net/rxrpc/call_object.c 	__releases(&rx->sk.sk_lock.slock)
sk_lock           597 net/rxrpc/sendmsg.c 	__releases(&rx->sk.sk_lock.slock)
sk_lock           639 net/rxrpc/sendmsg.c 	__releases(&rx->sk.sk_lock.slock)
sk_lock          2310 net/tipc/socket.c 			if (likely(spin_trylock_bh(&sk->sk_lock.slock))) {
sk_lock          2312 net/tipc/socket.c 				spin_unlock_bh(&sk->sk_lock.slock);