sk_lock 65 include/net/llc.h spinlock_t sk_lock; sk_lock 364 include/net/sock.h socket_lock_t sk_lock; sk_lock 1493 include/net/sock.h if (sk->sk_lock.owned) { sk_lock 1494 include/net/sock.h sk->sk_lock.owned = 0; sk_lock 1497 include/net/sock.h mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_); sk_lock 1510 include/net/sock.h sk->sk_lock.owned = 0; \ sk_lock 1511 include/net/sock.h init_waitqueue_head(&sk->sk_lock.wq); \ sk_lock 1512 include/net/sock.h spin_lock_init(&(sk)->sk_lock.slock); \ sk_lock 1513 include/net/sock.h debug_check_no_locks_freed((void *)&(sk)->sk_lock, \ sk_lock 1514 include/net/sock.h sizeof((sk)->sk_lock)); \ sk_lock 1515 include/net/sock.h lockdep_set_class_and_name(&(sk)->sk_lock.slock, \ sk_lock 1517 include/net/sock.h lockdep_init_map(&(sk)->sk_lock.dep_map, (name), (key), 0); \ sk_lock 1523 include/net/sock.h return lockdep_is_held(&sk->sk_lock) || sk_lock 1524 include/net/sock.h lockdep_is_held(&sk->sk_lock.slock); sk_lock 1539 include/net/sock.h #define bh_lock_sock(__sk) spin_lock(&((__sk)->sk_lock.slock)) sk_lock 1541 include/net/sock.h spin_lock_nested(&((__sk)->sk_lock.slock), \ sk_lock 1543 include/net/sock.h #define bh_unlock_sock(__sk) spin_unlock(&((__sk)->sk_lock.slock)) sk_lock 1559 include/net/sock.h spin_unlock_bh(&sk->sk_lock.slock); sk_lock 1586 include/net/sock.h return sk->sk_lock.owned; sk_lock 1591 include/net/sock.h return sk->sk_lock.owned; sk_lock 1599 include/net/sock.h return !sk->sk_lock.owned && !spin_is_locked(&sk->sk_lock.slock); sk_lock 472 net/ceph/messenger.c lockdep_set_class(&sock->sk->sk_lock, &socket_class); sk_lock 520 net/core/sock.c mutex_acquire(&sk->sk_lock.dep_map, 0, 1, _RET_IP_); sk_lock 524 net/core/sock.c mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_); sk_lock 2409 net/core/sock.c __releases(&sk->sk_lock.slock) sk_lock 2410 net/core/sock.c __acquires(&sk->sk_lock.slock) sk_lock 2415 net/core/sock.c prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait, sk_lock 2417 net/core/sock.c spin_unlock_bh(&sk->sk_lock.slock); sk_lock 2419 net/core/sock.c spin_lock_bh(&sk->sk_lock.slock); sk_lock 2423 net/core/sock.c finish_wait(&sk->sk_lock.wq, &wait); sk_lock 2427 net/core/sock.c __releases(&sk->sk_lock.slock) sk_lock 2428 net/core/sock.c __acquires(&sk->sk_lock.slock) sk_lock 2435 net/core/sock.c spin_unlock_bh(&sk->sk_lock.slock); sk_lock 2449 net/core/sock.c spin_lock_bh(&sk->sk_lock.slock); sk_lock 2461 net/core/sock.c spin_lock_bh(&sk->sk_lock.slock); sk_lock 2463 net/core/sock.c spin_unlock_bh(&sk->sk_lock.slock); sk_lock 2941 net/core/sock.c spin_lock_bh(&sk->sk_lock.slock); sk_lock 2942 net/core/sock.c if (sk->sk_lock.owned) sk_lock 2944 net/core/sock.c sk->sk_lock.owned = 1; sk_lock 2945 net/core/sock.c spin_unlock(&sk->sk_lock.slock); sk_lock 2949 net/core/sock.c mutex_acquire(&sk->sk_lock.dep_map, subclass, 0, _RET_IP_); sk_lock 2956 net/core/sock.c spin_lock_bh(&sk->sk_lock.slock); sk_lock 2967 net/core/sock.c if (waitqueue_active(&sk->sk_lock.wq)) sk_lock 2968 net/core/sock.c wake_up(&sk->sk_lock.wq); sk_lock 2969 net/core/sock.c spin_unlock_bh(&sk->sk_lock.slock); sk_lock 2989 net/core/sock.c spin_lock_bh(&sk->sk_lock.slock); sk_lock 2991 net/core/sock.c if (!sk->sk_lock.owned) sk_lock 2998 net/core/sock.c sk->sk_lock.owned = 1; sk_lock 2999 net/core/sock.c spin_unlock(&sk->sk_lock.slock); sk_lock 3003 net/core/sock.c mutex_acquire(&sk->sk_lock.dep_map, 0, 0, _RET_IP_); sk_lock 80 net/core/sock_map.c __acquires(&sk->sk_lock.slock) sk_lock 88 net/core/sock_map.c __releases(&sk->sk_lock.slock) sk_lock 214 net/ipv4/icmp.c if (unlikely(!spin_trylock(&sk->sk_lock.slock))) { sk_lock 225 net/ipv4/icmp.c spin_unlock(&sk->sk_lock.slock); sk_lock 879 net/ipv4/inet_connection_sock.c __releases(&sk->sk_lock.slock) sk_lock 117 net/ipv6/icmp.c if (unlikely(!spin_trylock(&sk->sk_lock.slock))) { sk_lock 129 net/ipv6/icmp.c spin_unlock(&sk->sk_lock.slock); sk_lock 1535 net/l2tp/l2tp_core.c lockdep_set_class_and_name(&sk->sk_lock.slock, &l2tp_socket_class, sk_lock 119 net/l2tp/l2tp_ppp.c struct mutex sk_lock; /* Protects .sk */ sk_lock 458 net/l2tp/l2tp_ppp.c mutex_lock(&ps->sk_lock); sk_lock 460 net/l2tp/l2tp_ppp.c lockdep_is_held(&ps->sk_lock)); sk_lock 462 net/l2tp/l2tp_ppp.c mutex_unlock(&ps->sk_lock); sk_lock 555 net/l2tp/l2tp_ppp.c mutex_init(&ps->sk_lock); sk_lock 767 net/l2tp/l2tp_ppp.c mutex_lock(&ps->sk_lock); sk_lock 769 net/l2tp/l2tp_ppp.c lockdep_is_held(&ps->sk_lock)) || sk_lock 771 net/l2tp/l2tp_ppp.c mutex_unlock(&ps->sk_lock); sk_lock 790 net/l2tp/l2tp_ppp.c mutex_lock(&ps->sk_lock); sk_lock 793 net/l2tp/l2tp_ppp.c mutex_unlock(&ps->sk_lock); sk_lock 824 net/l2tp/l2tp_ppp.c mutex_unlock(&ps->sk_lock); sk_lock 832 net/l2tp/l2tp_ppp.c mutex_unlock(&ps->sk_lock); sk_lock 703 net/llc/llc_conn.c spin_lock_bh(&sap->sk_lock); sk_lock 708 net/llc/llc_conn.c spin_unlock_bh(&sap->sk_lock); sk_lock 723 net/llc/llc_conn.c spin_lock_bh(&sap->sk_lock); sk_lock 727 net/llc/llc_conn.c spin_unlock_bh(&sap->sk_lock); sk_lock 41 net/llc/llc_core.c spin_lock_init(&sap->sk_lock); sk_lock 41 net/llc/llc_proc.c spin_lock_bh(&sap->sk_lock); sk_lock 52 net/llc/llc_proc.c spin_unlock_bh(&sap->sk_lock); sk_lock 102 net/llc/llc_proc.c spin_unlock_bh(&sap->sk_lock); sk_lock 104 net/llc/llc_proc.c spin_lock_bh(&sap->sk_lock); sk_lock 108 net/llc/llc_proc.c spin_unlock_bh(&sap->sk_lock); sk_lock 121 net/llc/llc_proc.c spin_unlock_bh(&sap->sk_lock); sk_lock 198 net/llc/llc_proc.c !!sk->sk_backlog.tail, !!sk->sk_lock.owned); sk_lock 396 net/llc/llc_sap.c spin_lock_bh(&sap->sk_lock); sk_lock 412 net/llc/llc_sap.c spin_unlock_bh(&sap->sk_lock); sk_lock 463 net/rxrpc/call_accept.c __releases(&rx->sk.sk_lock.slock) sk_lock 222 net/rxrpc/call_object.c __releases(&rx->sk.sk_lock.slock) sk_lock 597 net/rxrpc/sendmsg.c __releases(&rx->sk.sk_lock.slock) sk_lock 639 net/rxrpc/sendmsg.c __releases(&rx->sk.sk_lock.slock) sk_lock 2310 net/tipc/socket.c if (likely(spin_trylock_bh(&sk->sk_lock.slock))) { sk_lock 2312 net/tipc/socket.c spin_unlock_bh(&sk->sk_lock.slock);