ml 44 arch/arm/vfp/vfp.h static inline void add128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) ml 51 arch/arm/vfp/vfp.h : "0" (nl), "1" (nh), "r" (ml), "r" (mh) ml 57 arch/arm/vfp/vfp.h static inline void sub128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) ml 64 arch/arm/vfp/vfp.h : "0" (nl), "1" (nh), "r" (ml), "r" (mh) ml 72 arch/arm/vfp/vfp.h u32 nh, nl, mh, ml; ml 76 arch/arm/vfp/vfp.h ml = m; ml 77 arch/arm/vfp/vfp.h rl = (u64)nl * ml; ml 80 arch/arm/vfp/vfp.h rma = (u64)nh * ml; ml 112 arch/arm/vfp/vfp.h u64 mh, ml, remh, reml, termh, terml, z; ml 126 arch/arm/vfp/vfp.h ml = m << 32; ml 129 arch/arm/vfp/vfp.h add128(&remh, &reml, remh, reml, mh, ml); ml 421 arch/mips/include/asm/octeon/cvmx-pci-defs.h uint32_t ml:8; ml 429 arch/mips/include/asm/octeon/cvmx-pci-defs.h uint32_t ml:8; ml 87 arch/mips/include/asm/txx9/tx3927.h endian_def_b4(ml, mg, ip, il); ml 154 arch/sh/kernel/cpu/sh2a/fpu.c unsigned long long ml, int n) ml 158 arch/sh/kernel/cpu/sh2a/fpu.c return (mh << (64 - n)) | (ml >> n); ml 164 arch/sh/kernel/cpu/sh2a/fpu.c unsigned long long mh, ml, nh, nl; ml 175 arch/sh/kernel/cpu/sh2a/fpu.c mult64(ix, iy, &mh, &ml); ml 177 arch/sh/kernel/cpu/sh2a/fpu.c nl = ml; ml 188 arch/sh/kernel/cpu/sh2a/fpu.c ix = (rshift64(mh, ml, w - 52) & 0x000fffffffffffffLL) ml 191 arch/sh/kernel/cpu/sh2a/fpu.c ix = rshift64(mh, ml, w - 51 - exp) & 0x000fffffffffffffLL; ml 403 arch/sparc/kernel/chmc.c const void *ml; ml 452 arch/sparc/kernel/chmc.c ml = of_get_property(op->dev.of_node, "memory-layout", &p->layout_len); ml 453 arch/sparc/kernel/chmc.c if (!ml) { ml 462 arch/sparc/kernel/chmc.c memcpy(&p->layout, ml, p->layout_len); ml 227 crypto/vmac.c #define poly_step(ah, al, kh, kl, mh, ml) \ ml 246 crypto/vmac.c ADD128(ah, al, mh, ml); \ ml 275 crypto/vmac.c const u64 *mh, const u64 *ml) ml 303 crypto/vmac.c p += (u64)(((u32 *)ml)[INDEX_LOW]); ml 312 crypto/vmac.c p += (u64)(((u32 *)ml)[INDEX_HIGH]); ml 333 crypto/vmac.c #define poly_step(ah, al, kh, kl, mh, ml) \ ml 334 crypto/vmac.c poly_step_func(&(ah), &(al), &(kh), &(kl), &(mh), &(ml)) ml 1241 drivers/iio/industrialio-buffer.c int ml = indio_dev->masklength; ml 1244 drivers/iio/industrialio-buffer.c ml = max(ml, channels[i].scan_index + 1); ml 1245 drivers/iio/industrialio-buffer.c indio_dev->masklength = ml; ml 112 drivers/input/ff-memless.c static void ml_schedule_timer(struct ml_device *ml) ml 125 drivers/input/ff-memless.c state = &ml->states[i]; ml 142 drivers/input/ff-memless.c del_timer(&ml->timer); ml 145 drivers/input/ff-memless.c mod_timer(&ml->timer, earliest); ml 326 drivers/input/ff-memless.c static int ml_get_combo_effect(struct ml_device *ml, ml 341 drivers/input/ff-memless.c state = &ml->states[i]; ml 355 drivers/input/ff-memless.c effect_type = get_compatible_type(ml->dev->ff, effect->type); ml 383 drivers/input/ff-memless.c ml_combine_effects(combo_effect, state, ml->gain); ml 390 drivers/input/ff-memless.c static void ml_play_effects(struct ml_device *ml) ml 397 drivers/input/ff-memless.c while (ml_get_combo_effect(ml, handled_bm, &effect)) ml 398 drivers/input/ff-memless.c ml->play_effect(ml->dev, ml->private, &effect); ml 400 drivers/input/ff-memless.c ml_schedule_timer(ml); ml 405 drivers/input/ff-memless.c struct ml_device *ml = from_timer(ml, t, timer); ml 406 drivers/input/ff-memless.c struct input_dev *dev = ml->dev; ml 412 drivers/input/ff-memless.c ml_play_effects(ml); ml 421 drivers/input/ff-memless.c struct ml_device *ml = dev->ff->private; ml 424 drivers/input/ff-memless.c ml->gain = gain; ml 427 drivers/input/ff-memless.c __clear_bit(FF_EFFECT_PLAYING, &ml->states[i].flags); ml 429 drivers/input/ff-memless.c ml_play_effects(ml); ml 437 drivers/input/ff-memless.c struct ml_device *ml = dev->ff->private; ml 438 drivers/input/ff-memless.c struct ml_effect_state *state = &ml->states[effect_id]; ml 460 drivers/input/ff-memless.c ml_play_effects(ml); ml 468 drivers/input/ff-memless.c struct ml_device *ml = dev->ff->private; ml 469 drivers/input/ff-memless.c struct ml_effect_state *state = &ml->states[effect->id]; ml 480 drivers/input/ff-memless.c ml_schedule_timer(ml); ml 490 drivers/input/ff-memless.c struct ml_device *ml = ff->private; ml 499 drivers/input/ff-memless.c del_timer_sync(&ml->timer); ml 501 drivers/input/ff-memless.c kfree(ml->private); ml 513 drivers/input/ff-memless.c struct ml_device *ml; ml 518 drivers/input/ff-memless.c ml = kzalloc(sizeof(struct ml_device), GFP_KERNEL); ml 519 drivers/input/ff-memless.c if (!ml) ml 522 drivers/input/ff-memless.c ml->dev = dev; ml 523 drivers/input/ff-memless.c ml->private = data; ml 524 drivers/input/ff-memless.c ml->play_effect = play_effect; ml 525 drivers/input/ff-memless.c ml->gain = 0xffff; ml 526 drivers/input/ff-memless.c timer_setup(&ml->timer, ml_effect_timer, 0); ml 532 drivers/input/ff-memless.c kfree(ml); ml 537 drivers/input/ff-memless.c ff->private = ml; ml 552 drivers/input/ff-memless.c ml->states[i].effect = &ff->effects[i]; ml 54 fs/ocfs2/dlm/dlmast.c if (lock->ml.highest_blocked == LKM_IVMODE) ml 56 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.highest_blocked == LKM_NLMODE); ml 63 fs/ocfs2/dlm/dlmast.c if (lock->ml.type == LKM_EXMODE) ml 66 fs/ocfs2/dlm/dlmast.c else if (lock->ml.type == LKM_NLMODE) ml 69 fs/ocfs2/dlm/dlmast.c else if (lock->ml.highest_blocked != LKM_EXMODE) ml 91 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 92 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 93 fs/ocfs2/dlm/dlmast.c lock->ast_pending, lock->ml.type); ml 99 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 100 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie))); ml 110 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 111 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie))); ml 114 fs/ocfs2/dlm/dlmast.c lock->ml.highest_blocked = LKM_IVMODE; ml 157 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 158 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie))); ml 190 fs/ocfs2/dlm/dlmast.c lock->ml.node == dlm->node_num ? "master" : ml 215 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 216 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie))); ml 219 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num); ml 235 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 236 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie))); ml 239 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node == dlm->node_num); ml 255 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num); ml 259 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 260 fs/ocfs2/dlm/dlmast.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 360 fs/ocfs2/dlm/dlmast.c if (lock->ml.cookie == cookie) ml 372 fs/ocfs2/dlm/dlmast.c if (lock->ml.cookie == cookie) { ml 399 fs/ocfs2/dlm/dlmast.c lock->ml.type, lock->ml.convert_type); ml 401 fs/ocfs2/dlm/dlmast.c if (lock->ml.convert_type != LKM_IVMODE) { ml 402 fs/ocfs2/dlm/dlmast.c lock->ml.type = lock->ml.convert_type; ml 403 fs/ocfs2/dlm/dlmast.c lock->ml.convert_type = LKM_IVMODE; ml 444 fs/ocfs2/dlm/dlmast.c res->lockname.len, res->lockname.name, lock->ml.node, msg_type, ml 453 fs/ocfs2/dlm/dlmast.c past.cookie = lock->ml.cookie; ml 465 fs/ocfs2/dlm/dlmast.c lock->ml.node, &status); ml 469 fs/ocfs2/dlm/dlmast.c lock->ml.node); ml 473 fs/ocfs2/dlm/dlmast.c "node is dead!\n", lock->ml.node); ml 477 fs/ocfs2/dlm/dlmast.c "DLM_MIGRATING!\n", lock->ml.node); ml 481 fs/ocfs2/dlm/dlmast.c lock->ml.node, status); ml 233 fs/ocfs2/dlm/dlmcommon.h struct dlm_mig_lockres_priv ml; ml 342 fs/ocfs2/dlm/dlmcommon.h struct dlm_migratable_lock ml; ml 567 fs/ocfs2/dlm/dlmcommon.h struct dlm_migratable_lock ml[0]; // 16 bytes each, begins at byte 112 ml 116 fs/ocfs2/dlm/dlmconvert.c lock->ml.type, lock->ml.convert_type, type); ml 121 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.convert_type != LKM_IVMODE) { ml 137 fs/ocfs2/dlm/dlmconvert.c switch (lock->ml.type) { ml 141 fs/ocfs2/dlm/dlmconvert.c dlm_lock_mode_name(lock->ml.type), ml 151 fs/ocfs2/dlm/dlmconvert.c dlm_lock_mode_name(lock->ml.type), ml 157 fs/ocfs2/dlm/dlmconvert.c dlm_lock_mode_name(lock->ml.type), ml 167 fs/ocfs2/dlm/dlmconvert.c if (type <= lock->ml.type) ml 175 fs/ocfs2/dlm/dlmconvert.c if (!dlm_lock_compatible(tmplock->ml.type, type)) ml 180 fs/ocfs2/dlm/dlmconvert.c if (!dlm_lock_compatible(tmplock->ml.type, type)) ml 183 fs/ocfs2/dlm/dlmconvert.c if (!dlm_lock_compatible(tmplock->ml.convert_type, type)) ml 194 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.node == dlm->node_num) ml 196 fs/ocfs2/dlm/dlmconvert.c lock->ml.type = type; ml 214 fs/ocfs2/dlm/dlmconvert.c lock->ml.type, type); ml 221 fs/ocfs2/dlm/dlmconvert.c lock->ml.convert_type = type; ml 240 fs/ocfs2/dlm/dlmconvert.c lock->ml.convert_type = LKM_IVMODE; ml 257 fs/ocfs2/dlm/dlmconvert.c mlog(0, "type=%d, convert_type=%d, busy=%d\n", lock->ml.type, ml 258 fs/ocfs2/dlm/dlmconvert.c lock->ml.convert_type, res->state & DLM_LOCK_RES_IN_PROGRESS); ml 271 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.convert_type != LKM_IVMODE) { ml 275 fs/ocfs2/dlm/dlmconvert.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 276 fs/ocfs2/dlm/dlmconvert.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 277 fs/ocfs2/dlm/dlmconvert.c lock->ml.convert_type); ml 282 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.type == type && lock->ml.convert_type == LKM_IVMODE) { ml 287 fs/ocfs2/dlm/dlmconvert.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 288 fs/ocfs2/dlm/dlmconvert.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 289 fs/ocfs2/dlm/dlmconvert.c lock->ml.type, lock->ml.convert_type); ml 299 fs/ocfs2/dlm/dlmconvert.c lock->ml.convert_type = type; ml 302 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.type == LKM_EXMODE) { ml 306 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.convert_type == LKM_NLMODE) ml 371 fs/ocfs2/dlm/dlmconvert.c convert.cookie = lock->ml.cookie; ml 485 fs/ocfs2/dlm/dlmconvert.c if (tmp_lock->ml.cookie == cnv->cookie && ml 486 fs/ocfs2/dlm/dlmconvert.c tmp_lock->ml.node == cnv->node_idx) { ml 66 fs/ocfs2/dlm/dlmdebug.c lock->ml.type, lock->ml.convert_type, lock->ml.node, ml 67 fs/ocfs2/dlm/dlmdebug.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 68 fs/ocfs2/dlm/dlmdebug.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 473 fs/ocfs2/dlm/dlmdebug.c list_type, lock->ml.type, lock->ml.convert_type, ml 474 fs/ocfs2/dlm/dlmdebug.c lock->ml.node, ml 475 fs/ocfs2/dlm/dlmdebug.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 476 fs/ocfs2/dlm/dlmdebug.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 81 fs/ocfs2/dlm/dlmlock.c if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) ml 86 fs/ocfs2/dlm/dlmlock.c if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) ml 88 fs/ocfs2/dlm/dlmlock.c if (!dlm_lock_compatible(tmplock->ml.convert_type, ml 89 fs/ocfs2/dlm/dlmlock.c lock->ml.type)) ml 110 fs/ocfs2/dlm/dlmlock.c mlog(0, "type=%d\n", lock->ml.type); ml 117 fs/ocfs2/dlm/dlmlock.c lock->ml.node != dlm->node_num) { ml 146 fs/ocfs2/dlm/dlmlock.c lock->ml.node); ml 157 fs/ocfs2/dlm/dlmlock.c lock->ml.node); ml 207 fs/ocfs2/dlm/dlmlock.c lock->ml.type, res->lockname.len, ml 295 fs/ocfs2/dlm/dlmlock.c create.requested_type = lock->ml.type; ml 296 fs/ocfs2/dlm/dlmlock.c create.cookie = lock->ml.cookie; ml 385 fs/ocfs2/dlm/dlmlock.c newlock->ml.type = type; ml 386 fs/ocfs2/dlm/dlmlock.c newlock->ml.convert_type = LKM_IVMODE; ml 387 fs/ocfs2/dlm/dlmlock.c newlock->ml.highest_blocked = LKM_IVMODE; ml 388 fs/ocfs2/dlm/dlmlock.c newlock->ml.node = node; ml 389 fs/ocfs2/dlm/dlmlock.c newlock->ml.pad1 = 0; ml 390 fs/ocfs2/dlm/dlmlock.c newlock->ml.list = 0; ml 391 fs/ocfs2/dlm/dlmlock.c newlock->ml.flags = 0; ml 395 fs/ocfs2/dlm/dlmlock.c newlock->ml.cookie = cpu_to_be64(cookie); ml 2510 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) { ml 2514 fs/ocfs2/dlm/dlmmaster.c cookie = be64_to_cpu(lock->ml.cookie); ml 2940 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) { ml 2942 fs/ocfs2/dlm/dlmmaster.c lock->ml.node); ml 2949 fs/ocfs2/dlm/dlmmaster.c lock->ml.node); ml 2997 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node == dlm->node_num) ml 2999 fs/ocfs2/dlm/dlmmaster.c if (test_bit(lock->ml.node, dlm->exit_domain_map)) ml 3001 fs/ocfs2/dlm/dlmmaster.c nodenum = lock->ml.node; ml 1058 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { ml 1133 fs/ocfs2/dlm/dlmrecovery.c struct_size(mres, ml, mres->num_locks), ml 1192 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.type != LKM_EXMODE && lock->ml.type != LKM_PRMODE) ml 1206 fs/ocfs2/dlm/dlmrecovery.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 1207 fs/ocfs2/dlm/dlmrecovery.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 1209 fs/ocfs2/dlm/dlmrecovery.c lock->ml.node); ml 1219 fs/ocfs2/dlm/dlmrecovery.c struct dlm_migratable_lock *ml; ml 1222 fs/ocfs2/dlm/dlmrecovery.c ml = &(mres->ml[lock_num]); ml 1223 fs/ocfs2/dlm/dlmrecovery.c ml->cookie = lock->ml.cookie; ml 1224 fs/ocfs2/dlm/dlmrecovery.c ml->type = lock->ml.type; ml 1225 fs/ocfs2/dlm/dlmrecovery.c ml->convert_type = lock->ml.convert_type; ml 1226 fs/ocfs2/dlm/dlmrecovery.c ml->highest_blocked = lock->ml.highest_blocked; ml 1227 fs/ocfs2/dlm/dlmrecovery.c ml->list = queue; ml 1229 fs/ocfs2/dlm/dlmrecovery.c ml->flags = lock->lksb->flags; ml 1232 fs/ocfs2/dlm/dlmrecovery.c ml->node = lock->ml.node; ml 1245 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.cookie = 0; ml 1246 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.type = LKM_IVMODE; ml 1247 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.convert_type = LKM_IVMODE; ml 1248 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.highest_blocked = LKM_IVMODE; ml 1250 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.node = dlm->node_num; ml 1255 fs/ocfs2/dlm/dlmrecovery.c struct dlm_migratable_lock *ml, ml 1258 fs/ocfs2/dlm/dlmrecovery.c if (unlikely(ml->cookie == 0 && ml 1259 fs/ocfs2/dlm/dlmrecovery.c ml->type == LKM_IVMODE && ml 1260 fs/ocfs2/dlm/dlmrecovery.c ml->convert_type == LKM_IVMODE && ml 1261 fs/ocfs2/dlm/dlmrecovery.c ml->highest_blocked == LKM_IVMODE && ml 1262 fs/ocfs2/dlm/dlmrecovery.c ml->list == DLM_BLOCKED_LIST)) { ml 1263 fs/ocfs2/dlm/dlmrecovery.c *nodenum = ml->node; ml 1518 fs/ocfs2/dlm/dlmrecovery.c item->u.ml.lockres = res; /* already have a ref */ ml 1519 fs/ocfs2/dlm/dlmrecovery.c item->u.ml.real_master = real_master; ml 1520 fs/ocfs2/dlm/dlmrecovery.c item->u.ml.extra_ref = extra_refs; ml 1554 fs/ocfs2/dlm/dlmrecovery.c res = item->u.ml.lockres; ml 1555 fs/ocfs2/dlm/dlmrecovery.c real_master = item->u.ml.real_master; ml 1556 fs/ocfs2/dlm/dlmrecovery.c extra_ref = item->u.ml.extra_ref; ml 1800 fs/ocfs2/dlm/dlmrecovery.c struct dlm_migratable_lock *ml; ml 1813 fs/ocfs2/dlm/dlmrecovery.c ml = &(mres->ml[i]); ml 1815 fs/ocfs2/dlm/dlmrecovery.c if (dlm_is_dummy_lock(dlm, ml, &from)) { ml 1826 fs/ocfs2/dlm/dlmrecovery.c BUG_ON(ml->highest_blocked != LKM_IVMODE); ml 1830 fs/ocfs2/dlm/dlmrecovery.c queue = dlm_list_num_to_pointer(res, ml->list); ml 1836 fs/ocfs2/dlm/dlmrecovery.c if (ml->node == dlm->node_num) { ml 1847 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.cookie == ml->cookie) ml 1858 fs/ocfs2/dlm/dlmrecovery.c c = ml->cookie; ml 1865 fs/ocfs2/dlm/dlmrecovery.c ml->node, ml->list, ml->flags, ml->type, ml 1866 fs/ocfs2/dlm/dlmrecovery.c ml->convert_type, ml->highest_blocked); ml 1871 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node != ml->node) { ml 1872 fs/ocfs2/dlm/dlmrecovery.c c = lock->ml.cookie; ml 1878 fs/ocfs2/dlm/dlmrecovery.c lock->ml.node); ml 1879 fs/ocfs2/dlm/dlmrecovery.c c = ml->cookie; ml 1885 fs/ocfs2/dlm/dlmrecovery.c ml->node, ml->list, ml->flags, ml->type, ml 1886 fs/ocfs2/dlm/dlmrecovery.c ml->convert_type, ml->highest_blocked); ml 1892 fs/ocfs2/dlm/dlmrecovery.c c = ml->cookie; ml 1897 fs/ocfs2/dlm/dlmrecovery.c j, ml->list, res->lockname.len, ml 1917 fs/ocfs2/dlm/dlmrecovery.c newlock = dlm_new_lock(ml->type, ml->node, ml 1918 fs/ocfs2/dlm/dlmrecovery.c be64_to_cpu(ml->cookie), NULL); ml 1926 fs/ocfs2/dlm/dlmrecovery.c if (ml->convert_type != LKM_IVMODE) { ml 1928 fs/ocfs2/dlm/dlmrecovery.c newlock->ml.convert_type = ml->convert_type; ml 1930 fs/ocfs2/dlm/dlmrecovery.c lksb->flags |= (ml->flags & ml 1933 fs/ocfs2/dlm/dlmrecovery.c if (ml->type == LKM_NLMODE) ml 1940 fs/ocfs2/dlm/dlmrecovery.c if (ml->list == DLM_BLOCKED_LIST) ml 1957 fs/ocfs2/dlm/dlmrecovery.c BUG_ON(ml->type != LKM_EXMODE && ml 1958 fs/ocfs2/dlm/dlmrecovery.c ml->type != LKM_PRMODE); ml 1960 fs/ocfs2/dlm/dlmrecovery.c (ml->type == LKM_EXMODE || ml 1966 fs/ocfs2/dlm/dlmrecovery.c res->lockname.name, ml->type); ml 2001 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.cookie == ml->cookie) { ml 2002 fs/ocfs2/dlm/dlmrecovery.c c = lock->ml.cookie; ml 2011 fs/ocfs2/dlm/dlmrecovery.c ml->type, ml->convert_type, ml->node, ml 2012 fs/ocfs2/dlm/dlmrecovery.c dlm_get_lock_cookie_node(be64_to_cpu(ml->cookie)), ml 2013 fs/ocfs2/dlm/dlmrecovery.c dlm_get_lock_cookie_seq(be64_to_cpu(ml->cookie)), ml 2014 fs/ocfs2/dlm/dlmrecovery.c ml->list); ml 2024 fs/ocfs2/dlm/dlmrecovery.c ml->list == DLM_CONVERTING_LIST && ml 2025 fs/ocfs2/dlm/dlmrecovery.c newlock->ml.type > ml 2026 fs/ocfs2/dlm/dlmrecovery.c newlock->ml.convert_type) { ml 2034 fs/ocfs2/dlm/dlmrecovery.c res->lockname.len, res->lockname.name, ml->node); ml 2035 fs/ocfs2/dlm/dlmrecovery.c dlm_lockres_set_refmap_bit(dlm, res, ml->node); ml 2210 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.type != LKM_EXMODE && ml 2211 fs/ocfs2/dlm/dlmrecovery.c lock->ml.type != LKM_PRMODE) ml 2213 fs/ocfs2/dlm/dlmrecovery.c } else if (lock->ml.type == LKM_EXMODE) ml 2244 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == search_node) { ml 2279 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { ml 2288 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { ml 2297 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { ml 2364 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { ml 42 fs/ocfs2/dlm/dlmthread.c #define dlm_lock_is_remote(dlm, lock) ((lock)->ml.node != (dlm)->node_num) ml 369 fs/ocfs2/dlm/dlmthread.c if (target->ml.convert_type == LKM_IVMODE) { ml 377 fs/ocfs2/dlm/dlmthread.c if (!dlm_lock_compatible(lock->ml.type, ml 378 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type)) { ml 381 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked == LKM_IVMODE) { ml 386 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked < target->ml.convert_type) ml 387 fs/ocfs2/dlm/dlmthread.c lock->ml.highest_blocked = ml 388 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type; ml 395 fs/ocfs2/dlm/dlmthread.c if (!dlm_lock_compatible(lock->ml.type, ml 396 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type)) { ml 398 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked == LKM_IVMODE) { ml 402 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked < target->ml.convert_type) ml 403 fs/ocfs2/dlm/dlmthread.c lock->ml.highest_blocked = ml 404 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type; ml 411 fs/ocfs2/dlm/dlmthread.c BUG_ON(target->ml.highest_blocked != LKM_IVMODE); ml 416 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_node(be64_to_cpu(target->ml.cookie)), ml 417 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_seq(be64_to_cpu(target->ml.cookie)), ml 418 fs/ocfs2/dlm/dlmthread.c target->ml.type, ml 419 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type, target->ml.node); ml 421 fs/ocfs2/dlm/dlmthread.c target->ml.type = target->ml.convert_type; ml 422 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type = LKM_IVMODE; ml 444 fs/ocfs2/dlm/dlmthread.c if (!dlm_lock_compatible(lock->ml.type, target->ml.type)) { ml 446 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked == LKM_IVMODE) { ml 450 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked < target->ml.type) ml 451 fs/ocfs2/dlm/dlmthread.c lock->ml.highest_blocked = target->ml.type; ml 458 fs/ocfs2/dlm/dlmthread.c if (!dlm_lock_compatible(lock->ml.type, target->ml.type)) { ml 460 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked == LKM_IVMODE) { ml 464 fs/ocfs2/dlm/dlmthread.c if (lock->ml.highest_blocked < target->ml.type) ml 465 fs/ocfs2/dlm/dlmthread.c lock->ml.highest_blocked = target->ml.type; ml 473 fs/ocfs2/dlm/dlmthread.c BUG_ON(target->ml.highest_blocked != LKM_IVMODE); ml 478 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_node(be64_to_cpu(target->ml.cookie)), ml 479 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_seq(be64_to_cpu(target->ml.cookie)), ml 480 fs/ocfs2/dlm/dlmthread.c target->ml.type, target->ml.node); ml 590 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 591 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 592 fs/ocfs2/dlm/dlmthread.c lock->ml.type, lock->ml.node); ml 601 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) { ml 636 fs/ocfs2/dlm/dlmthread.c BUG_ON(lock->ml.highest_blocked <= LKM_IVMODE); ml 637 fs/ocfs2/dlm/dlmthread.c hi = lock->ml.highest_blocked; ml 638 fs/ocfs2/dlm/dlmthread.c lock->ml.highest_blocked = LKM_IVMODE; ml 649 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 650 fs/ocfs2/dlm/dlmthread.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 651 fs/ocfs2/dlm/dlmthread.c hi, lock->ml.node); ml 653 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) { ml 220 fs/ocfs2/dlm/dlmunlock.c lock->ml.convert_type = LKM_IVMODE; ml 229 fs/ocfs2/dlm/dlmunlock.c BUG_ON(lock->ml.convert_type != LKM_IVMODE); ml 231 fs/ocfs2/dlm/dlmunlock.c BUG_ON(lock->ml.convert_type == LKM_IVMODE); ml 252 fs/ocfs2/dlm/dlmunlock.c dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), ml 253 fs/ocfs2/dlm/dlmunlock.c dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), ml 279 fs/ocfs2/dlm/dlmunlock.c lock->ml.convert_type = LKM_IVMODE; ml 338 fs/ocfs2/dlm/dlmunlock.c unlock.cookie = lock->ml.cookie; ml 467 fs/ocfs2/dlm/dlmunlock.c if (lock->ml.cookie == unlock->cookie && ml 468 fs/ocfs2/dlm/dlmunlock.c lock->ml.node == unlock->node_idx) { ml 488 fs/ocfs2/dlm/dlmunlock.c lock->ml.type != LKM_EXMODE) ml 629 fs/ocfs2/dlm/dlmunlock.c if (flags & LKM_VALBLK && lock->ml.type != LKM_EXMODE) ml 330 init/initramfs.c int ml = maybe_link(); ml 331 init/initramfs.c if (ml >= 0) { ml 333 init/initramfs.c if (ml != 1) ml 602 lib/inflate.c unsigned ml, md; /* masks for bl and bd bits */ ml 613 lib/inflate.c ml = mask_bits[bl]; /* precompute masks for speed */ ml 618 lib/inflate.c if ((e = (t = tl + ((unsigned)b & ml))->e) > 16) ml 116 lib/lz4/lz4hc_compress.c size_t ml = 0; ml 128 lib/lz4/lz4hc_compress.c if (*(match + ml) == *(ip + ml) ml 133 lib/lz4/lz4hc_compress.c if (mlt > ml) { ml 134 lib/lz4/lz4hc_compress.c ml = mlt; ml 155 lib/lz4/lz4hc_compress.c if (mlt > ml) { ml 157 lib/lz4/lz4hc_compress.c ml = mlt; ml 165 lib/lz4/lz4hc_compress.c return (int)ml; ml 358 lib/lz4/lz4hc_compress.c int ml, ml2, ml3, ml0; ml 379 lib/lz4/lz4hc_compress.c ml = LZ4HC_InsertAndFindBestMatch(ctx, ip, ml 381 lib/lz4/lz4hc_compress.c if (!ml) { ml 389 lib/lz4/lz4hc_compress.c ml0 = ml; ml 392 lib/lz4/lz4hc_compress.c if (ip + ml < mflimit) ml 394 lib/lz4/lz4hc_compress.c ip + ml - 2, ip + 0, ml 395 lib/lz4/lz4hc_compress.c matchlimit, ml, &ref2, ml 398 lib/lz4/lz4hc_compress.c ml2 = ml; ml 400 lib/lz4/lz4hc_compress.c if (ml2 == ml) { ml 403 lib/lz4/lz4hc_compress.c &anchor, ml, ref, limit, oend)) ml 413 lib/lz4/lz4hc_compress.c ml = ml0; ml 420 lib/lz4/lz4hc_compress.c ml = ml2; ml 434 lib/lz4/lz4hc_compress.c int new_ml = ml; ml 465 lib/lz4/lz4hc_compress.c if (start2 < ip + ml) ml 466 lib/lz4/lz4hc_compress.c ml = (int)(start2 - ip); ml 469 lib/lz4/lz4hc_compress.c ml, ref, limit, oend)) ml 478 lib/lz4/lz4hc_compress.c if (start3 < ip + ml + 3) { ml 480 lib/lz4/lz4hc_compress.c if (start3 >= (ip + ml)) { ml 485 lib/lz4/lz4hc_compress.c if (start2 < ip + ml) { ml 486 lib/lz4/lz4hc_compress.c int correction = (int)(ip + ml - start2); ml 499 lib/lz4/lz4hc_compress.c ml, ref, limit, oend)) ml 503 lib/lz4/lz4hc_compress.c ml = ml3; ml 522 lib/lz4/lz4hc_compress.c if (start2 < ip + ml) { ml 526 lib/lz4/lz4hc_compress.c if (ml > OPTIMAL_ML) ml 527 lib/lz4/lz4hc_compress.c ml = OPTIMAL_ML; ml 528 lib/lz4/lz4hc_compress.c if (ip + ml > start2 + ml2 - MINMATCH) ml 529 lib/lz4/lz4hc_compress.c ml = (int)(start2 - ip) + ml2 - MINMATCH; ml 530 lib/lz4/lz4hc_compress.c correction = ml - (int)(start2 - ip); ml 537 lib/lz4/lz4hc_compress.c ml = (int)(start2 - ip); ml 539 lib/lz4/lz4hc_compress.c if (LZ4HC_encodeSequence(&ip, &op, &anchor, ml, ml 545 lib/lz4/lz4hc_compress.c ml = ml2; ml 1835 lib/zstd/compress.c size_t ml = EQUAL_READ32 - 1; ml 1845 lib/zstd/compress.c if (match[ml] == ip[ml]) /* potentially better */ ml 1854 lib/zstd/compress.c if (currMl > ml) { ml 1855 lib/zstd/compress.c ml = currMl; ml 1866 lib/zstd/compress.c return ml; ml 609 net/openvswitch/flow_table.c struct list_head *ml; ml 611 net/openvswitch/flow_table.c list_for_each(ml, &tbl->mask_list) { ml 613 net/openvswitch/flow_table.c m = container_of(ml, struct sw_flow_mask, list); ml 3539 net/packet/af_packet.c struct packet_mclist *ml; ml 3541 net/packet/af_packet.c while ((ml = *mlp) != NULL) { ml 3542 net/packet/af_packet.c if (ml->ifindex == dev->ifindex) { ml 3543 net/packet/af_packet.c packet_dev_mc(dev, ml, -1); ml 3544 net/packet/af_packet.c *mlp = ml->next; ml 3545 net/packet/af_packet.c kfree(ml); ml 3547 net/packet/af_packet.c mlp = &ml->next; ml 3554 net/packet/af_packet.c struct packet_mclist *ml, *i; ml 3575 net/packet/af_packet.c for (ml = po->mclist; ml; ml = ml->next) { ml 3576 net/packet/af_packet.c if (ml->ifindex == mreq->mr_ifindex && ml 3577 net/packet/af_packet.c ml->type == mreq->mr_type && ml 3578 net/packet/af_packet.c ml->alen == mreq->mr_alen && ml 3579 net/packet/af_packet.c memcmp(ml->addr, mreq->mr_address, ml->alen) == 0) { ml 3580 net/packet/af_packet.c ml->count++; ml 3608 net/packet/af_packet.c struct packet_mclist *ml, **mlp; ml 3612 net/packet/af_packet.c for (mlp = &pkt_sk(sk)->mclist; (ml = *mlp) != NULL; mlp = &ml->next) { ml 3613 net/packet/af_packet.c if (ml->ifindex == mreq->mr_ifindex && ml 3614 net/packet/af_packet.c ml->type == mreq->mr_type && ml 3615 net/packet/af_packet.c ml->alen == mreq->mr_alen && ml 3616 net/packet/af_packet.c memcmp(ml->addr, mreq->mr_address, ml->alen) == 0) { ml 3617 net/packet/af_packet.c if (--ml->count == 0) { ml 3619 net/packet/af_packet.c *mlp = ml->next; ml 3620 net/packet/af_packet.c dev = __dev_get_by_index(sock_net(sk), ml->ifindex); ml 3622 net/packet/af_packet.c packet_dev_mc(dev, ml, -1); ml 3623 net/packet/af_packet.c kfree(ml); ml 3635 net/packet/af_packet.c struct packet_mclist *ml; ml 3641 net/packet/af_packet.c while ((ml = po->mclist) != NULL) { ml 3644 net/packet/af_packet.c po->mclist = ml->next; ml 3645 net/packet/af_packet.c dev = __dev_get_by_index(sock_net(sk), ml->ifindex); ml 3647 net/packet/af_packet.c packet_dev_mc(dev, ml, -1); ml 3648 net/packet/af_packet.c kfree(ml); ml 41 net/packet/diag.c struct packet_mclist *ml; ml 48 net/packet/diag.c for (ml = po->mclist; ml; ml = ml->next) { ml 58 net/packet/diag.c dml->pdmc_index = ml->ifindex; ml 59 net/packet/diag.c dml->pdmc_type = ml->type; ml 60 net/packet/diag.c dml->pdmc_alen = ml->alen; ml 61 net/packet/diag.c dml->pdmc_count = ml->count; ml 62 net/packet/diag.c BUILD_BUG_ON(sizeof(dml->pdmc_addr) != sizeof(ml->addr)); ml 63 net/packet/diag.c memcpy(dml->pdmc_addr, ml->addr, sizeof(ml->addr)); ml 46 sound/drivers/vx/vx_mixer.c u8 ml; ml 56 sound/drivers/vx/vx_mixer.c u8 ml; ml 64 sound/drivers/vx/vx_mixer.c #define SET_CDC_DATA_REG(di,r) ((di).b.ml = (u8) (r))