num_locks         380 drivers/hwspinlock/hwspinlock_core.c 	if (id < 0 || id >= hwlock->bank->num_locks) {
num_locks         487 drivers/hwspinlock/hwspinlock_core.c 		const struct hwspinlock_ops *ops, int base_id, int num_locks)
num_locks         492 drivers/hwspinlock/hwspinlock_core.c 	if (!bank || !ops || !dev || !num_locks || !ops->trylock ||
num_locks         501 drivers/hwspinlock/hwspinlock_core.c 	bank->num_locks = num_locks;
num_locks         503 drivers/hwspinlock/hwspinlock_core.c 	for (i = 0; i < num_locks; i++) {
num_locks         539 drivers/hwspinlock/hwspinlock_core.c 	for (i = 0; i < bank->num_locks; i++) {
num_locks         615 drivers/hwspinlock/hwspinlock_core.c 			      int base_id, int num_locks)
num_locks         624 drivers/hwspinlock/hwspinlock_core.c 	ret = hwspin_lock_register(bank, dev, ops, base_id, num_locks);
num_locks          58 drivers/hwspinlock/hwspinlock_internal.h 	int num_locks;
num_locks          81 drivers/hwspinlock/omap_hwspinlock.c 	int num_locks, i, ret;
num_locks         125 drivers/hwspinlock/omap_hwspinlock.c 	num_locks = i * 32; /* actual number of locks in this device */
num_locks         127 drivers/hwspinlock/omap_hwspinlock.c 	bank = kzalloc(struct_size(bank, lock, num_locks), GFP_KERNEL);
num_locks         135 drivers/hwspinlock/omap_hwspinlock.c 	for (i = 0, hwlock = &bank->lock[0]; i < num_locks; i++, hwlock++)
num_locks         139 drivers/hwspinlock/omap_hwspinlock.c 						base_id, num_locks);
num_locks         144 drivers/hwspinlock/omap_hwspinlock.c 		num_locks);
num_locks          93 drivers/hwspinlock/u8500_hsem.c 	int i, ret, num_locks = U8500_MAX_SEMAPHORE;
num_locks         114 drivers/hwspinlock/u8500_hsem.c 	bank = kzalloc(struct_size(bank, lock, num_locks), GFP_KERNEL);
num_locks         122 drivers/hwspinlock/u8500_hsem.c 	for (i = 0, hwlock = &bank->lock[0]; i < num_locks; i++, hwlock++)
num_locks         129 drivers/hwspinlock/u8500_hsem.c 						pdata->base_id, num_locks);
num_locks        2764 fs/ceph/caps.c 		flags = atomic_read(&fi->num_locks) ? CHECK_FILELOCK : 0;
num_locks          38 fs/ceph/locks.c 	atomic_inc(&fi->num_locks);
num_locks          46 fs/ceph/locks.c 	atomic_dec(&fi->num_locks);
num_locks         717 fs/ceph/super.h 	atomic_t num_locks;
num_locks        1065 fs/cifs/cifsglob.h 			atomic_t num_locks;
num_locks        2497 fs/cifs/cifssmb.c 	cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
num_locks        2567 fs/cifs/cifssmb.c 	cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
num_locks         636 fs/cifs/smb1ops.c 	atomic_set(&tcon->stats.cifs_stats.num_locks, 0);
num_locks         655 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_locks),
num_locks        4824 fs/cifs/smb2pdu.c 	cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
num_locks         557 fs/ocfs2/dlm/dlmcommon.h 	u8 num_locks;    // locks sent in this structure
num_locks        1116 fs/ocfs2/dlm/dlmrecovery.c 	BUG_ON(mres->num_locks > DLM_MAX_MIGRATABLE_LOCKS);
num_locks        1117 fs/ocfs2/dlm/dlmrecovery.c 	if (!mres->num_locks)
num_locks        1133 fs/ocfs2/dlm/dlmrecovery.c 				 struct_size(mres, ml, mres->num_locks),
num_locks        1173 fs/ocfs2/dlm/dlmrecovery.c 	mres->num_locks = 0;
num_locks        1220 fs/ocfs2/dlm/dlmrecovery.c 	int lock_num = mres->num_locks;
num_locks        1233 fs/ocfs2/dlm/dlmrecovery.c 	mres->num_locks++;
num_locks        1235 fs/ocfs2/dlm/dlmrecovery.c 	if (mres->num_locks == DLM_MAX_MIGRATABLE_LOCKS)
num_locks        1811 fs/ocfs2/dlm/dlmrecovery.c 	mlog(0, "running %d locks for this lockres\n", mres->num_locks);
num_locks        1812 fs/ocfs2/dlm/dlmrecovery.c 	for (i=0; i<mres->num_locks; i++) {
num_locks        1817 fs/ocfs2/dlm/dlmrecovery.c 			BUG_ON(mres->num_locks != 1);
num_locks          59 include/linux/hwspinlock.h 		const struct hwspinlock_ops *ops, int base_id, int num_locks);
num_locks          80 include/linux/hwspinlock.h 			      int base_id, int num_locks);