num_locks 380 drivers/hwspinlock/hwspinlock_core.c if (id < 0 || id >= hwlock->bank->num_locks) { num_locks 487 drivers/hwspinlock/hwspinlock_core.c const struct hwspinlock_ops *ops, int base_id, int num_locks) num_locks 492 drivers/hwspinlock/hwspinlock_core.c if (!bank || !ops || !dev || !num_locks || !ops->trylock || num_locks 501 drivers/hwspinlock/hwspinlock_core.c bank->num_locks = num_locks; num_locks 503 drivers/hwspinlock/hwspinlock_core.c for (i = 0; i < num_locks; i++) { num_locks 539 drivers/hwspinlock/hwspinlock_core.c for (i = 0; i < bank->num_locks; i++) { num_locks 615 drivers/hwspinlock/hwspinlock_core.c int base_id, int num_locks) num_locks 624 drivers/hwspinlock/hwspinlock_core.c ret = hwspin_lock_register(bank, dev, ops, base_id, num_locks); num_locks 58 drivers/hwspinlock/hwspinlock_internal.h int num_locks; num_locks 81 drivers/hwspinlock/omap_hwspinlock.c int num_locks, i, ret; num_locks 125 drivers/hwspinlock/omap_hwspinlock.c num_locks = i * 32; /* actual number of locks in this device */ num_locks 127 drivers/hwspinlock/omap_hwspinlock.c bank = kzalloc(struct_size(bank, lock, num_locks), GFP_KERNEL); num_locks 135 drivers/hwspinlock/omap_hwspinlock.c for (i = 0, hwlock = &bank->lock[0]; i < num_locks; i++, hwlock++) num_locks 139 drivers/hwspinlock/omap_hwspinlock.c base_id, num_locks); num_locks 144 drivers/hwspinlock/omap_hwspinlock.c num_locks); num_locks 93 drivers/hwspinlock/u8500_hsem.c int i, ret, num_locks = U8500_MAX_SEMAPHORE; num_locks 114 drivers/hwspinlock/u8500_hsem.c bank = kzalloc(struct_size(bank, lock, num_locks), GFP_KERNEL); num_locks 122 drivers/hwspinlock/u8500_hsem.c for (i = 0, hwlock = &bank->lock[0]; i < num_locks; i++, hwlock++) num_locks 129 drivers/hwspinlock/u8500_hsem.c pdata->base_id, num_locks); num_locks 2764 fs/ceph/caps.c flags = atomic_read(&fi->num_locks) ? CHECK_FILELOCK : 0; num_locks 38 fs/ceph/locks.c atomic_inc(&fi->num_locks); num_locks 46 fs/ceph/locks.c atomic_dec(&fi->num_locks); num_locks 717 fs/ceph/super.h atomic_t num_locks; num_locks 1065 fs/cifs/cifsglob.h atomic_t num_locks; num_locks 2497 fs/cifs/cifssmb.c cifs_stats_inc(&tcon->stats.cifs_stats.num_locks); num_locks 2567 fs/cifs/cifssmb.c cifs_stats_inc(&tcon->stats.cifs_stats.num_locks); num_locks 636 fs/cifs/smb1ops.c atomic_set(&tcon->stats.cifs_stats.num_locks, 0); num_locks 655 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_locks), num_locks 4824 fs/cifs/smb2pdu.c cifs_stats_inc(&tcon->stats.cifs_stats.num_locks); num_locks 557 fs/ocfs2/dlm/dlmcommon.h u8 num_locks; // locks sent in this structure num_locks 1116 fs/ocfs2/dlm/dlmrecovery.c BUG_ON(mres->num_locks > DLM_MAX_MIGRATABLE_LOCKS); num_locks 1117 fs/ocfs2/dlm/dlmrecovery.c if (!mres->num_locks) num_locks 1133 fs/ocfs2/dlm/dlmrecovery.c struct_size(mres, ml, mres->num_locks), num_locks 1173 fs/ocfs2/dlm/dlmrecovery.c mres->num_locks = 0; num_locks 1220 fs/ocfs2/dlm/dlmrecovery.c int lock_num = mres->num_locks; num_locks 1233 fs/ocfs2/dlm/dlmrecovery.c mres->num_locks++; num_locks 1235 fs/ocfs2/dlm/dlmrecovery.c if (mres->num_locks == DLM_MAX_MIGRATABLE_LOCKS) num_locks 1811 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "running %d locks for this lockres\n", mres->num_locks); num_locks 1812 fs/ocfs2/dlm/dlmrecovery.c for (i=0; i<mres->num_locks; i++) { num_locks 1817 fs/ocfs2/dlm/dlmrecovery.c BUG_ON(mres->num_locks != 1); num_locks 59 include/linux/hwspinlock.h const struct hwspinlock_ops *ops, int base_id, int num_locks); num_locks 80 include/linux/hwspinlock.h int base_id, int num_locks);