nlocks            445 fs/ceph/locks.c 	__le32 nlocks;
nlocks            447 fs/ceph/locks.c 	nlocks = cpu_to_le32(num_fcntl_locks);
nlocks            448 fs/ceph/locks.c 	err = ceph_pagelist_append(pagelist, &nlocks, sizeof(nlocks));
nlocks            459 fs/ceph/locks.c 	nlocks = cpu_to_le32(num_flock_locks);
nlocks            460 fs/ceph/locks.c 	err = ceph_pagelist_append(pagelist, &nlocks, sizeof(nlocks));
nlocks            346 kernel/locking/test-ww_mutex.c 	int nlocks;
nlocks            381 kernel/locking/test-ww_mutex.c 	const int nlocks = stress->nlocks;
nlocks            386 kernel/locking/test-ww_mutex.c 	order = get_random_order(nlocks);
nlocks            397 kernel/locking/test-ww_mutex.c 		for (n = 0; n < nlocks; n++) {
nlocks            446 kernel/locking/test-ww_mutex.c 	order = get_random_order(stress->nlocks);
nlocks            450 kernel/locking/test-ww_mutex.c 	for (n = 0; n < stress->nlocks; n++) {
nlocks            500 kernel/locking/test-ww_mutex.c 	const int nlocks = stress->nlocks;
nlocks            501 kernel/locking/test-ww_mutex.c 	struct ww_mutex *lock = stress->locks + (get_random_int() % nlocks);
nlocks            524 kernel/locking/test-ww_mutex.c static int stress(int nlocks, int nthreads, unsigned int flags)
nlocks            529 kernel/locking/test-ww_mutex.c 	locks = kmalloc_array(nlocks, sizeof(*locks), GFP_KERNEL);
nlocks            533 kernel/locking/test-ww_mutex.c 	for (n = 0; n < nlocks; n++)
nlocks            565 kernel/locking/test-ww_mutex.c 		stress->nlocks = nlocks;
nlocks            574 kernel/locking/test-ww_mutex.c 	for (n = 0; n < nlocks; n++)