nlocks 445 fs/ceph/locks.c __le32 nlocks; nlocks 447 fs/ceph/locks.c nlocks = cpu_to_le32(num_fcntl_locks); nlocks 448 fs/ceph/locks.c err = ceph_pagelist_append(pagelist, &nlocks, sizeof(nlocks)); nlocks 459 fs/ceph/locks.c nlocks = cpu_to_le32(num_flock_locks); nlocks 460 fs/ceph/locks.c err = ceph_pagelist_append(pagelist, &nlocks, sizeof(nlocks)); nlocks 346 kernel/locking/test-ww_mutex.c int nlocks; nlocks 381 kernel/locking/test-ww_mutex.c const int nlocks = stress->nlocks; nlocks 386 kernel/locking/test-ww_mutex.c order = get_random_order(nlocks); nlocks 397 kernel/locking/test-ww_mutex.c for (n = 0; n < nlocks; n++) { nlocks 446 kernel/locking/test-ww_mutex.c order = get_random_order(stress->nlocks); nlocks 450 kernel/locking/test-ww_mutex.c for (n = 0; n < stress->nlocks; n++) { nlocks 500 kernel/locking/test-ww_mutex.c const int nlocks = stress->nlocks; nlocks 501 kernel/locking/test-ww_mutex.c struct ww_mutex *lock = stress->locks + (get_random_int() % nlocks); nlocks 524 kernel/locking/test-ww_mutex.c static int stress(int nlocks, int nthreads, unsigned int flags) nlocks 529 kernel/locking/test-ww_mutex.c locks = kmalloc_array(nlocks, sizeof(*locks), GFP_KERNEL); nlocks 533 kernel/locking/test-ww_mutex.c for (n = 0; n < nlocks; n++) nlocks 565 kernel/locking/test-ww_mutex.c stress->nlocks = nlocks; nlocks 574 kernel/locking/test-ww_mutex.c for (n = 0; n < nlocks; n++)