mlog 106 arch/powerpc/platforms/pseries/ras.c static inline u8 rtas_mc_error_sub_type(const struct pseries_mc_errorlog *mlog) mlog 108 arch/powerpc/platforms/pseries/ras.c switch (mlog->error_type) { mlog 110 arch/powerpc/platforms/pseries/ras.c return (mlog->sub_err_type & 0x07); mlog 114 arch/powerpc/platforms/pseries/ras.c return (mlog->sub_err_type & 0x03); mlog 896 fs/ocfs2/alloc.c mlog(ML_ERROR, "Checksum failed for extent block %llu\n", mlog 2584 fs/ocfs2/alloc.c mlog(ML_ERROR, mlog 5193 fs/ocfs2/alloc.c mlog(ML_ERROR, "Owner %llu tried to set %d flags on an " mlog 5201 fs/ocfs2/alloc.c mlog(ML_ERROR, "Owner %llu tried to clear %d flags on an " mlog 5480 fs/ocfs2/alloc.c mlog(ML_ERROR, "Owner %llu: Invalid record truncate: (%u, %u) " mlog 6036 fs/ocfs2/alloc.c mlog(ML_ERROR, "Could not get bitmap inode!\n"); mlog 6157 fs/ocfs2/alloc.c mlog(ML_ERROR, "Could not get load truncate log inode!\n"); mlog 6257 fs/ocfs2/alloc.c mlog(ML_ERROR, "Asked to recover my own truncate log!\n"); mlog 6698 fs/ocfs2/alloc.c mlog(0, "Reusing block(%llu) from " mlog 7311 fs/ocfs2/alloc.c mlog(ML_ERROR, "Inode %lu has an empty " mlog 7683 fs/ocfs2/alloc.c mlog(ML_NOTICE, "Wait for trim on device (%s) to " mlog 7698 fs/ocfs2/alloc.c mlog(ML_NOTICE, "The same trim on device (%s) was " mlog 61 fs/ocfs2/aops.c mlog(ML_ERROR, "block offset > PATH_MAX: %llu", mlog 76 fs/ocfs2/aops.c mlog(ML_ERROR, "block offset is outside the allocated size: " mlog 89 fs/ocfs2/aops.c mlog(ML_ERROR, "couldn't getblock for symlink!\n"); mlog 101 fs/ocfs2/aops.c mlog(ML_ERROR, "couldn't kmap!\n"); mlog 150 fs/ocfs2/aops.c mlog(ML_NOTICE, "get_block on system inode 0x%p (%lu)\n", mlog 162 fs/ocfs2/aops.c mlog(ML_ERROR, "Error %d from get_blocks(0x%p, %llu, 1, " mlog 197 fs/ocfs2/aops.c mlog(ML_ERROR, mlog 202 fs/ocfs2/aops.c mlog(ML_ERROR, "Size %llu, clusters %u\n", (unsigned long long)i_size_read(inode), OCFS2_I(inode)->ip_clusters); mlog 499 fs/ocfs2/aops.c mlog(ML_ERROR, "get_blocks() failed, block = %llu\n", mlog 1188 fs/ocfs2/aops.c mlog(ML_ERROR, "Get physical blkno failed for inode %llu, " mlog 2174 fs/ocfs2/aops.c mlog(0, "get block of %lu at %llu:%u req %u\n", mlog 296 fs/ocfs2/blockcheck.c mlog(ML_NOTICE, "Block check count has wrapped\n"); mlog 312 fs/ocfs2/blockcheck.c mlog(ML_NOTICE, "Checksum failure count has wrapped\n"); mlog 328 fs/ocfs2/blockcheck.c mlog(ML_NOTICE, "ECC recovery count has wrapped\n"); mlog 402 fs/ocfs2/blockcheck.c mlog(ML_ERROR, mlog 417 fs/ocfs2/blockcheck.c mlog(ML_ERROR, "Fixed CRC32 failed: stored: 0x%x, computed 0x%x\n", mlog 515 fs/ocfs2/blockcheck.c mlog(ML_ERROR, mlog 548 fs/ocfs2/blockcheck.c mlog(ML_ERROR, "Fixed CRC32 failed: stored: %u, computed %u\n", mlog 129 fs/ocfs2/buffer_head_io.c mlog(ML_ERROR, mlog 139 fs/ocfs2/buffer_head_io.c mlog(ML_ERROR, mlog 219 fs/ocfs2/buffer_head_io.c mlog(ML_ERROR, "asked to read %d blocks!\n", nr); mlog 308 fs/ocfs2/buffer_head_io.c mlog(ML_ERROR, "block %llu had the JBD bit set " mlog 294 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Heartbeat write timeout to device %s after %u " mlog 308 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Number of regions %d, failed regions %d\n", mlog 328 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Queue write timeout for %u ms\n", mlog 417 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "send NEGO_APPROVE msg to node %d\n", i); mlog 421 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "send NEGO_APPROVE msg to node %d fail %d\n", mlog 432 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "send NEGO_TIMEOUT msg to node %d fail %d\n", mlog 449 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "got nego timeout message from bad node.\n"); mlog 497 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "IO Error %d\n", bio->bi_status); mlog 527 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Could not alloc slots BIO!\n"); mlog 547 fs/ocfs2/cluster/heartbeat.c mlog(ML_HB_BIO, "page %d, vec_len = %u, vec_start = %u\n", mlog 644 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Dump slot information: seq = 0x%llx, node = %u, " mlog 698 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "%s (%s): expected(%u:0x%llx, 0x%llx), " mlog 735 fs/ocfs2/cluster/heartbeat.c mlog(ML_HB_BIO, "our node generation = 0x%llx, cksum = 0x%x\n", mlog 747 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "calling funcs %p\n", f); mlog 772 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Node %s event for %d\n", mlog 805 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Queue node %s event for node %d\n", mlog 824 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Shutdown, node %d leaves region\n", mlog 931 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Node %d has written a bad crc to %s\n", mlog 956 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Node %d changed generation (0x%llx " mlog 964 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Slot %d gen 0x%llx cksum 0x%x " mlog 979 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Node %d (id 0x%llx) joined my region\n", mlog 986 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "o2hb: Add node %d to live nodes " mlog 1011 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Node %d on device %s has a dead count " mlog 1029 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Node %d left my region\n", mlog 1037 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "o2hb: Remove node %d from live " mlog 1109 fs/ocfs2/cluster/heartbeat.c mlog(ML_NOTICE, "o2hb: No configured nodes found!\n"); mlog 1154 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Write error %d on device \"%s\"\n", mlog 1205 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT|ML_KTHREAD, "hb thread running\n"); mlog 1212 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Node has been deleted, ret = %d\n", ret); mlog 1235 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, mlog 1270 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT|ML_KTHREAD, "o2hb thread exiting\n"); mlog 1505 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "hb region release (%s)\n", reg->hr_dev_name); mlog 1661 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "hr_start_block = %llu, hr_blocks = %u\n", mlog 1663 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "hr_block_bytes = %u, hr_block_bits = %u\n", mlog 1665 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "hr_timeout_ms = %u\n", reg->hr_timeout_ms); mlog 1666 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "dead threshold = %u\n", o2hb_dead_threshold); mlog 1695 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, "Going to require %u pages to cover %u blocks " mlog 1817 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, mlog 2299 fs/ocfs2/cluster/heartbeat.c mlog(ML_CLUSTER, "Pin region %s\n", uuid); mlog 2305 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "Pin region %s fails with %d\n", mlog 2345 fs/ocfs2/cluster/heartbeat.c mlog(ML_CLUSTER, "Unpin region %s\n", uuid); mlog 2443 fs/ocfs2/cluster/heartbeat.c mlog(ML_CLUSTER, "returning %d on behalf of %p for funcs %p\n", mlog 2454 fs/ocfs2/cluster/heartbeat.c mlog(ML_CLUSTER, "on behalf of %p for funcs %p\n", mlog 2480 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, mlog 2496 fs/ocfs2/cluster/heartbeat.c mlog(ML_HEARTBEAT, mlog 2514 fs/ocfs2/cluster/heartbeat.c mlog(ML_ERROR, "stopping heartbeat on all active regions.\n"); mlog 2538 fs/ocfs2/cluster/heartbeat.c mlog(0, "Region: %s\n", config_item_name(®->hr_item)); mlog 173 fs/ocfs2/cluster/masklog.h mlog(mask, fmt, ##__VA_ARGS__); \ mlog 181 fs/ocfs2/cluster/masklog.h mlog(ML_ERROR, "status = %lld\n", (long long)_st); \ mlog 187 fs/ocfs2/cluster/masklog.h mlog(ML_ERROR, "bug expression: " #cond "\n"); \ mlog 188 fs/ocfs2/cluster/masklog.h mlog(ML_ERROR, fmt, ##args); \ mlog 464 fs/ocfs2/cluster/nodemanager.c mlog(ML_NOTICE, mlog 471 fs/ocfs2/cluster/nodemanager.c mlog(ML_NOTICE, "o2net: idle timeout must be larger " mlog 501 fs/ocfs2/cluster/nodemanager.c mlog(ML_NOTICE, mlog 508 fs/ocfs2/cluster/nodemanager.c mlog(ML_NOTICE, "o2net: keepalive delay must be " mlog 599 fs/ocfs2/cluster/nodemanager.c mlog(ML_CLUSTER, "o2nm: Registering node %s\n", name); mlog 636 fs/ocfs2/cluster/nodemanager.c mlog(ML_CLUSTER, "o2nm: Unregistered node %s\n", mlog 104 fs/ocfs2/cluster/quorum.c mlog(0, "heartbeating: %d, connected: %d, " mlog 117 fs/ocfs2/cluster/quorum.c mlog(ML_ERROR, "fencing this node because it is " mlog 131 fs/ocfs2/cluster/quorum.c mlog(ML_ERROR, "fencing this node because it is " mlog 140 fs/ocfs2/cluster/quorum.c mlog(ML_ERROR, "fencing this node because it is " mlog 154 fs/ocfs2/cluster/quorum.c mlog(ML_NOTICE, "not fencing this node, heartbeating: %d, " mlog 172 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds); mlog 181 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1); mlog 209 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); mlog 234 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); mlog 252 fs/ocfs2/cluster/quorum.c mlog(0, "node %u\n", node); mlog 277 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected); mlog 309 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected); mlog 75 fs/ocfs2/cluster/tcp.c mlog(ML_MSG, "[mag %u len %u typ %u stat %d sys_stat %d " \ mlog 85 fs/ocfs2/cluster/tcp.c mlog(ML_SOCKET, "[sc %p refs %d sock %p node %u page %p " \ mlog 362 fs/ocfs2/cluster/tcp.c mlog(0, "completed %d messages for node %u\n", num_kills, mlog 512 fs/ocfs2/cluster/tcp.c mlog(ML_CONN, "node %u sc: %p -> %p, valid %u -> %u, err %d -> %d\n", mlog 559 fs/ocfs2/cluster/tcp.c mlog(ML_CONN, "queueing conn attempt in %lu jiffies\n", delay); mlog 808 fs/ocfs2/cluster/tcp.c mlog(0, "max_len for message handler out of range: %u\n", mlog 815 fs/ocfs2/cluster/tcp.c mlog(0, "no message type provided: %u, %p\n", msg_type, func); mlog 821 fs/ocfs2/cluster/tcp.c mlog(0, "no message handler provided: %u, %p\n", mlog 852 fs/ocfs2/cluster/tcp.c mlog(ML_TCP, "registered handler func %p type %u key %08x\n", mlog 876 fs/ocfs2/cluster/tcp.c mlog(ML_TCP, "unregistering handler func %p type %u key %08x\n", mlog 923 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "sendmsg returned %d instead of %zu\n", ret, total); mlog 927 fs/ocfs2/cluster/tcp.c mlog(0, "returning error: %d\n", ret); mlog 948 fs/ocfs2/cluster/tcp.c mlog(0, "sendpage of size %zu to " SC_NODEF_FMT mlog 953 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "sendpage of size %zu to " SC_NODEF_FMT mlog 1031 fs/ocfs2/cluster/tcp.c mlog(0, "attempt to tx without o2netd running\n"); mlog 1037 fs/ocfs2/cluster/tcp.c mlog(0, "bad kvec array length\n"); mlog 1044 fs/ocfs2/cluster/tcp.c mlog(0, "total payload len %zu too large\n", caller_bytes); mlog 1067 fs/ocfs2/cluster/tcp.c mlog(0, "failed to %zu element kvec!\n", veclen); mlog 1074 fs/ocfs2/cluster/tcp.c mlog(0, "failed to allocate a o2net_msg!\n"); mlog 1102 fs/ocfs2/cluster/tcp.c mlog(0, "error returned from o2net_send_tcp_msg=%d\n", ret); mlog 1119 fs/ocfs2/cluster/tcp.c mlog(0, "woken, returning system status %d, user status %d\n", mlog 1209 fs/ocfs2/cluster/tcp.c mlog(ML_TCP, "couldn't find handler for type %u key %08x\n", mlog 1240 fs/ocfs2/cluster/tcp.c mlog(0, "sending handler status %d, syserr %d returned %d\n", mlog 1481 fs/ocfs2/cluster/tcp.c mlog(ML_MSG, "sc sending handshake with ver %llu id %llx\n", mlog 1615 fs/ocfs2/cluster/tcp.c mlog(0, "couldn't allocate sc\n"); mlog 1622 fs/ocfs2/cluster/tcp.c mlog(0, "can't create socket: %d\n", ret); mlog 1636 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "bind failed with %d at address %pI4\n", mlog 1643 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "setting TCP_NODELAY failed with %d\n", ret); mlog 1649 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "set TCP_USER_TIMEOUT failed with %d\n", ret); mlog 1841 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "setting TCP_NODELAY failed with %d\n", ret); mlog 1847 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "set TCP_USER_TIMEOUT failed with %d\n", ret); mlog 1883 fs/ocfs2/cluster/tcp.c mlog(ML_CONN, "attempt to connect from node '%s' at " mlog 2071 fs/ocfs2/cluster/tcp.c mlog(ML_KTHREAD, "starting o2net thread...\n"); mlog 2074 fs/ocfs2/cluster/tcp.c mlog(ML_ERROR, "unable to launch o2net thread\n"); mlog 2114 fs/ocfs2/cluster/tcp.c mlog(ML_KTHREAD, "waiting for o2net thread to exit....\n"); mlog 406 fs/ocfs2/dcache.c mlog(ML_ERROR, "Dentry is missing cluster lock. " mlog 318 fs/ocfs2/dir.c mlog(ML_ERROR, "bad entry in directory #%llu: %s - " mlog 447 fs/ocfs2/dir.c mlog(ML_ERROR, "Checksum failed for dinode %llu\n", mlog 576 fs/ocfs2/dir.c mlog(ML_ERROR, mlog 619 fs/ocfs2/dir.c mlog(ML_ERROR, mlog 721 fs/ocfs2/dir.c mlog(ML_ERROR, "reading directory %llu, " mlog 1271 fs/ocfs2/dir.c mlog(ML_ERROR, "Dir %llu: Bad dx_entry ptr idx %d, (%p, %p)\n", mlog 1748 fs/ocfs2/dir.c mlog(ML_ERROR, "Unable to read inode block for dir %llu\n", mlog 2137 fs/ocfs2/dir.c mlog(ML_ERROR, "bad directory (dir #%llu) - no `.' or `..'\n", mlog 3721 fs/ocfs2/dir.c mlog(ML_ERROR, "DX Dir: %llu, Asked to rebalance empty leaf: " mlog 80 fs/ocfs2/dlm/dlmapi.h mlog(ML_ERROR, "dlm status = %s\n", dlm_errname((st))); \ mlog 88 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "%s: res %.*s, lock %u:%llu, " mlog 97 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, lock %u:%llu, AST getting flushed\n", mlog 108 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, lock %u:%llu, Cancelling BAST\n", mlog 155 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, lock %u:%llu, BAST getting flushed\n", mlog 189 fs/ocfs2/dlm/dlmast.c mlog(0, "getting lvb from lockres for %s node\n", mlog 213 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, lock %u:%llu, Local AST\n", dlm->name, mlog 233 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, lock %u:%llu, Remote AST\n", dlm->name, mlog 257 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, lock %u:%llu, Local BAST, blocked %d\n", mlog 299 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "Invalid name length (%d) in proxy ast " mlog 306 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "Both PUT and GET lvb specified, (0x%x)\n", mlog 312 fs/ocfs2/dlm/dlmast.c mlog(0, "lvb: %s\n", flags & LKM_PUT_LVB ? "put lvb" : mlog 315 fs/ocfs2/dlm/dlmast.c mlog(0, "type=%d, blocked_type=%d\n", past->type, past->blocked_type); mlog 319 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "Unknown ast type! %d, cookie=%u:%llu" mlog 330 fs/ocfs2/dlm/dlmast.c mlog(0, "Got %sast for unknown lockres! cookie=%u:%llu, " mlog 342 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s\n", dlm->name, res->lockname.len, mlog 347 fs/ocfs2/dlm/dlmast.c mlog(0, "Responding with DLM_RECOVERING!\n"); mlog 352 fs/ocfs2/dlm/dlmast.c mlog(0, "Responding with DLM_MIGRATING!\n"); mlog 379 fs/ocfs2/dlm/dlmast.c mlog(0, "Got %sast for unknown lock! cookie=%u:%llu, name=%.*s, " mlog 395 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, lock %u:%llu, Granted type %d => %d\n", mlog 443 fs/ocfs2/dlm/dlmast.c mlog(0, "%s: res %.*s, to %u, type %d, blocked_type %d\n", dlm->name, mlog 467 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "%s: res %.*s, error %d send AST to node %u\n", mlog 472 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "sent AST to node %u, it thinks this " mlog 476 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "sent AST to node %u, it returned " mlog 480 fs/ocfs2/dlm/dlmast.c mlog(ML_ERROR, "AST to node %u returned %d!\n", mlog 115 fs/ocfs2/dlm/dlmconvert.c mlog(0, "type=%d, convert_type=%d, new convert_type=%d\n", mlog 122 fs/ocfs2/dlm/dlmconvert.c mlog(ML_ERROR, "attempted to convert a lock with a lock " mlog 130 fs/ocfs2/dlm/dlmconvert.c mlog(ML_ERROR, "attempted to convert a lock not on grant " mlog 140 fs/ocfs2/dlm/dlmconvert.c mlog(0, "will set lvb: converting %s->%s\n", mlog 149 fs/ocfs2/dlm/dlmconvert.c mlog(0, "will fetch new value into " mlog 155 fs/ocfs2/dlm/dlmconvert.c mlog(0, "will NOT fetch new value " mlog 190 fs/ocfs2/dlm/dlmconvert.c mlog(0, "res %.*s, granting %s lock\n", res->lockname.len, mlog 195 fs/ocfs2/dlm/dlmconvert.c mlog(0, "doing in-place convert for nonlocal lock\n"); mlog 212 fs/ocfs2/dlm/dlmconvert.c mlog(0, "failed to convert NOQUEUE lock %.*s from " mlog 218 fs/ocfs2/dlm/dlmconvert.c mlog(0, "res %.*s, queueing...\n", res->lockname.len, mlog 257 fs/ocfs2/dlm/dlmconvert.c mlog(0, "type=%d, convert_type=%d, busy=%d\n", lock->ml.type, mlog 262 fs/ocfs2/dlm/dlmconvert.c mlog(0, "bailing out early since res is RECOVERING " mlog 273 fs/ocfs2/dlm/dlmconvert.c mlog(ML_ERROR, "converting a remote lock that is already " mlog 283 fs/ocfs2/dlm/dlmconvert.c mlog(0, "last convert request returned DLM_RECOVERING, but " mlog 331 fs/ocfs2/dlm/dlmconvert.c mlog(0, "%s: res %.*s, owner died and lock has been moved back " mlog 366 fs/ocfs2/dlm/dlmconvert.c mlog(0, "%.*s\n", res->lockname.len, res->lockname.name); mlog 392 fs/ocfs2/dlm/dlmconvert.c mlog(0, "node %u returned DLM_RECOVERING from convert " mlog 395 fs/ocfs2/dlm/dlmconvert.c mlog(0, "node %u returned DLM_MIGRATING from convert " mlog 398 fs/ocfs2/dlm/dlmconvert.c mlog(0, "node %u returned DLM_FORWARD from convert " mlog 403 fs/ocfs2/dlm/dlmconvert.c mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " mlog 413 fs/ocfs2/dlm/dlmconvert.c mlog(0, "node %u died so returning DLM_RECOVERING " mlog 462 fs/ocfs2/dlm/dlmconvert.c mlog(ML_ERROR, "both PUT and GET lvb specified\n"); mlog 467 fs/ocfs2/dlm/dlmconvert.c mlog(0, "lvb: %s\n", flags & LKM_PUT_LVB ? "put lvb" : mlog 495 fs/ocfs2/dlm/dlmconvert.c mlog(ML_ERROR, "did not find lock to convert on grant queue! " mlog 84 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Allocated DLM hash pagevec; %d pages (%lu expected), %lu buckets per page\n", mlog 152 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: Unhash res %.*s\n", dlm->name, res->lockname.len, mlog 171 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: Hash res %.*s\n", dlm->name, res->lockname.len, mlog 183 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%.*s\n", len, name); mlog 215 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%.*s\n", len, name); mlog 323 fs/ocfs2/dlm/dlmdomain.c mlog(0, "freeing memory from domain %s\n", dlm->name); mlog 413 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Migrating locks from domain %s\n", dlm->name); mlog 451 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: perhaps there are more lock resources " mlog 455 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: we won't do dlm recovery after migrating " mlog 467 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: %d lock resources in hash last pass\n", mlog 471 fs/ocfs2/dlm/dlmdomain.c mlog(0, "DONE Migrating locks from domain %s\n", dlm->name); mlog 497 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: Node %u sent a begin exit domain message\n", dlm->name, node); mlog 517 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Node %d is joining, we wait on it.\n", mlog 553 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%p %u %p", msg, len, data); mlog 582 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: Sending domain exit message %u to node %u\n", dlm->name, mlog 591 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Error %d sending domain exit message %u " mlog 656 fs/ocfs2/dlm/dlmdomain.c mlog(ML_NOTICE, "Error %d sending domain exit message " mlog 698 fs/ocfs2/dlm/dlmdomain.c mlog(0, "shutting down domain %s\n", dlm->name); mlog 707 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: more migration to do\n", dlm->name); mlog 712 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Following lockres' are still on the " mlog 736 fs/ocfs2/dlm/dlmdomain.c mlog(0, mlog 746 fs/ocfs2/dlm/dlmdomain.c mlog(ML_NOTICE, mlog 804 fs/ocfs2/dlm/dlmdomain.c mlog(0, "node %u wants to join domain %s\n", query->node_idx, mlog 813 fs/ocfs2/dlm/dlmdomain.c mlog(0, "node %u is not in our live map yet\n", mlog 836 fs/ocfs2/dlm/dlmdomain.c mlog(0, "disallow join as node %u does not " mlog 864 fs/ocfs2/dlm/dlmdomain.c mlog(0, "node %u trying to join, but recovery " mlog 868 fs/ocfs2/dlm/dlmdomain.c mlog(0, "node %u trying to join, but it " mlog 872 fs/ocfs2/dlm/dlmdomain.c mlog(0, "node %u trying to join, but it " mlog 905 fs/ocfs2/dlm/dlmdomain.c mlog(0, "We respond with %u\n", packet.code); mlog 919 fs/ocfs2/dlm/dlmdomain.c mlog(0, "node %u asserts join on domain %s\n", assert->node_idx, mlog 934 fs/ocfs2/dlm/dlmdomain.c mlog(0, "dlm recovery is ongoing, disallow join\n"); mlog 969 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Domain %s: Joining node %d has global " mlog 978 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Domain %s: Local node %d has global " mlog 987 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Region %.*s\n", O2HB_MAX_REGION_NAME_LEN, r); mlog 1008 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Domain %s: Region '%.*s' registered " mlog 1031 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Domain %s: Region '%.*s' registered " mlog 1070 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Region %.*s\n", O2HB_MAX_REGION_NAME_LEN, p); mlog 1078 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Sending regions to node %d\n", i); mlog 1086 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Region mismatch %d, node %d\n", mlog 1107 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Node %u queries hb regions on domain %s\n", qr->qr_node, mlog 1120 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Node %d queried hb regions on domain %s " mlog 1127 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Node %d queried hb regions on domain %s " mlog 1136 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Node %d queried hb regions on domain %s " mlog 1164 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Node %3d, %pI4:%u\n", qn->qn_nodes[j].ni_nodenum, mlog 1192 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Domain %s: Node %d (%pI4:%u) " mlog 1200 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Domain %s: Node %d (%pI4:%u) " mlog 1239 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Node %3d, %pI4:%u\n", node->nd_num, mlog 1256 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Sending nodeinfo to node %d\n", i); mlog 1264 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "node mismatch %d, node %d\n", ret, i); mlog 1283 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Node %u queries nodes on domain %s\n", qn->qn_nodenum, mlog 1289 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Node %d queried nodes on domain %s before " mlog 1297 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Node %d queried nodes on domain %s but " mlog 1306 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Node %d queried nodes on domain %s " mlog 1331 fs/ocfs2/dlm/dlmdomain.c mlog(0, "node %u cancels join on domain %s\n", cancel->node_idx, mlog 1367 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " mlog 1387 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, mlog 1402 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Error return %d cancelling join on " mlog 1423 fs/ocfs2/dlm/dlmdomain.c mlog(0, "querying node %d\n", node); mlog 1438 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " mlog 1459 fs/ocfs2/dlm/dlmdomain.c mlog(ML_NOTICE, mlog 1475 fs/ocfs2/dlm/dlmdomain.c mlog(0, mlog 1486 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "invalid response %d from node %u\n", mlog 1494 fs/ocfs2/dlm/dlmdomain.c mlog(0, "status %d, node %d response is %d\n", status, node, mlog 1508 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Sending join assert to node %u\n", node); mlog 1519 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " mlog 1551 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Error return %d asserting " mlog 1574 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Latest response of disallow -- should restart\n"); mlog 1586 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Node maps changed -- should restart\n"); mlog 1597 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%p", dlm); mlog 1641 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Yay, done querying nodes!\n"); mlog 1700 fs/ocfs2/dlm/dlmdomain.c mlog(0, "returning %d\n", status); mlog 1715 fs/ocfs2/dlm/dlmdomain.c mlog(0, "registering handlers.\n"); mlog 1863 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Join domain %s\n", dlm->name); mlog 1908 fs/ocfs2/dlm/dlmdomain.c mlog(ML_NOTICE, "Timed out joining dlm domain " mlog 1924 fs/ocfs2/dlm/dlmdomain.c mlog(0, "backoff %d\n", backoff); mlog 2011 fs/ocfs2/dlm/dlmdomain.c mlog(0, "dlm->recovery_map=%p, &(dlm->recovery_map[0])=%p\n", mlog 2052 fs/ocfs2/dlm/dlmdomain.c mlog(0, "context init: refcount %u\n", mlog 2113 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "domain name length too long\n"); mlog 2117 fs/ocfs2/dlm/dlmdomain.c mlog(0, "register called for domain \"%s\"\n", domain); mlog 2134 fs/ocfs2/dlm/dlmdomain.c mlog(0, "This ctxt is not joined yet!\n"); mlog 2143 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, mlog 2319 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Could not create o2dlm_mle slabcache\n"); mlog 2325 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Could not create o2dlm_lockres and " mlog 2332 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Count not create o2dlm_lock slabcache\n"); mlog 2338 fs/ocfs2/dlm/dlmdomain.c mlog(ML_ERROR, "Unable to register network handlers\n"); mlog 110 fs/ocfs2/dlm/dlmlock.c mlog(0, "type=%d\n", lock->ml.type); mlog 127 fs/ocfs2/dlm/dlmlock.c mlog(0, "I can grant this lock right away\n"); mlog 144 fs/ocfs2/dlm/dlmlock.c mlog(0, "%s: returning DLM_NORMAL to " mlog 155 fs/ocfs2/dlm/dlmlock.c mlog(0, "%s: returning NOTQUEUED to " mlog 206 fs/ocfs2/dlm/dlmlock.c mlog(0, "type=%d, lockres %.*s, flags = 0x%x\n", mlog 242 fs/ocfs2/dlm/dlmlock.c mlog(0, "%s: recovery lock was owned by " mlog 263 fs/ocfs2/dlm/dlmlock.c mlog(0, "%s: $RECOVERY lock for this node (%u) is " mlog 306 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "%s: res %.*s, Stale lockres no longer " mlog 314 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "%s: res %.*s, Error %d send CREATE LOCK to " mlog 351 fs/ocfs2/dlm/dlmlock.c mlog(0, "freeing kernel-allocated lksb\n"); mlog 373 fs/ocfs2/dlm/dlmlock.c mlog(0, "removing lock's lockres reference\n"); mlog 463 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "Domain %s not fully joined, but node %u is " mlog 489 fs/ocfs2/dlm/dlmlock.c mlog(0, "set DLM_LKSB_GET_LVB flag\n"); mlog 504 fs/ocfs2/dlm/dlmlock.c mlog(0, "lockres recovering/migrating/in-progress\n"); mlog 536 fs/ocfs2/dlm/dlmlock.c mlog(0, "This node's cookie will now wrap!\n"); mlog 580 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "strange LOCAL convert request!\n"); mlog 590 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "NULL lock pointer in convert " mlog 597 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "NULL lockres pointer in convert " mlog 611 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "new args: lksb=%p, ast=%p, bast=%p, " mlog 613 fs/ocfs2/dlm/dlmlock.c mlog(ML_ERROR, "orig args: lksb=%p, ast=%p, bast=%p, " mlog 631 fs/ocfs2/dlm/dlmlock.c mlog(0, "retrying convert with migration/recovery/" mlog 670 fs/ocfs2/dlm/dlmlock.c mlog(0, "type=%d, flags = 0x%x\n", mode, flags); mlog 671 fs/ocfs2/dlm/dlmlock.c mlog(0, "creating lock: lock=%p res=%p\n", lock, res); mlog 680 fs/ocfs2/dlm/dlmlock.c mlog(0, "LKM_VALBLK passed by caller\n"); mlog 224 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "bad mle: %p\n", mle); mlog 367 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u already removed from nodemap!\n", idx); mlog 381 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u already in node map!\n", idx); mlog 416 fs/ocfs2/dlm/dlmmaster.c mlog(0, "Releasing mle for %.*s, type %d\n", mle->mnamelen, mle->mname, mlog 478 fs/ocfs2/dlm/dlmmaster.c mlog(0, "destroying lockres %.*s\n", res->lockname.len, mlog 490 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, mlog 608 fs/ocfs2/dlm/dlmmaster.c mlog(0, "res %.*s, set node %u, %ps()\n", res->lockname.len, mlog 619 fs/ocfs2/dlm/dlmmaster.c mlog(0, "res %.*s, clr node %u, %ps()\n", res->lockname.len, mlog 630 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, inflight++: now %u, %ps()\n", dlm->name, mlog 651 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, inflight--: now %u, %ps()\n", dlm->name, mlog 663 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: inflight assert worker++: now %u\n", mlog 674 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: inflight assert worker--: now %u\n", mlog 722 fs/ocfs2/dlm/dlmmaster.c mlog(0, "get lockres %s (len %d)\n", lockid, namelen); mlog 773 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Resource %.*s not " mlog 786 fs/ocfs2/dlm/dlmmaster.c mlog(0, "allocating a new resource\n"); mlog 797 fs/ocfs2/dlm/dlmmaster.c mlog(0, "no lockres found, allocated our own: %p\n", res); mlog 820 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "master entry for nonexistent lock!\n"); mlog 836 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: late on %s\n", mlog 868 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, At least one node (%d) " mlog 900 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: Recovery map is not empty, but must " mlog 905 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: waiting 500ms for heartbeat state " mlog 919 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, At least one node (%d) " mlog 949 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, Requests only up to %u but " mlog 960 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, Node map changed, redo the master " mlog 964 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s: res %.*s, Spinning on " mlog 975 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, Mastered by %u\n", dlm->name, res->lockname.len, mlog 1020 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: owner is suddenly %u\n", dlm->name, mlog 1029 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "link to %u went down?: %d\n", res->owner, ret); mlog 1049 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: %.*s: node map changed, restarting\n", mlog 1054 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: status change: old=%d new=%d\n", mlog 1064 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: restart lock mastery succeeded, " mlog 1070 fs/ocfs2/dlm/dlmmaster.c mlog(0, "map not changed and voting not done " mlog 1110 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: waiting again\n", dlm->name, mlog 1114 fs/ocfs2/dlm/dlmmaster.c mlog(0, "done waiting, master is %u\n", res->owner); mlog 1122 fs/ocfs2/dlm/dlmmaster.c mlog(0, "about to master %.*s here, this=%u\n", mlog 1218 fs/ocfs2/dlm/dlmmaster.c mlog(0, "something happened such that the " mlog 1230 fs/ocfs2/dlm/dlmmaster.c mlog(ML_NOTICE, "node %d up while restarting\n", node); mlog 1233 fs/ocfs2/dlm/dlmmaster.c mlog(0, "sending request to new node\n"); mlog 1237 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "node down! %d\n", node); mlog 1246 fs/ocfs2/dlm/dlmmaster.c mlog(0, "expected master %u died" mlog 1253 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s:still " mlog 1268 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: no " mlog 1329 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "TCP stack not ready!\n"); mlog 1332 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "bad args passed to o2net!\n"); mlog 1335 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "out of memory while trying to send " mlog 1343 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "unhandled error!"); mlog 1348 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "link to %d went down!\n", to); mlog 1358 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u is the master, response=YES\n", to); mlog 1359 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: master node %u now knows I have a " mlog 1365 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u not master, response=NO\n", to); mlog 1369 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u not master, response=MAYBE\n", to); mlog 1374 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u hit an error, resending\n", to); mlog 1379 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "bad response! %u\n", response); mlog 1457 fs/ocfs2/dlm/dlmmaster.c mlog(0, "returning DLM_MASTER_RESP_ERROR since res is " mlog 1493 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "lock with no owner should be " mlog 1502 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "no mle found for this lock!\n"); mlog 1512 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u is master, but trying to migrate to " mlog 1515 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "no owner on lockres, but this " mlog 1587 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "no lockres, but an mle with this node as master!\n"); mlog 1593 fs/ocfs2/dlm/dlmmaster.c mlog(0, "migration mle was found (%u->%u)\n", mlog 1617 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%u is the owner of %.*s, cleaning everyone else\n", mlog 1623 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "failed to dispatch assert master work\n"); mlog 1679 fs/ocfs2/dlm/dlmmaster.c mlog(0, "sending assert master to %d (%.*s)\n", to, mlog 1690 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Error %d when sending message %u (key " mlog 1694 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "unhandled error=%d!\n", tmpret); mlog 1698 fs/ocfs2/dlm/dlmmaster.c mlog(0, "link to %d went down!\n", to); mlog 1704 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR,"during assert master of %.*s to %u, " mlog 1720 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%.*s: very strange, " mlog 1726 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%.*s: node %u create mles on other " mlog 1732 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%.*s: node %u has a reference to this " mlog 1783 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid name length!"); mlog 1790 fs/ocfs2/dlm/dlmmaster.c mlog(0, "assert_master with flags: %u\n", flags); mlog 1796 fs/ocfs2/dlm/dlmmaster.c mlog(0, "just got an assert_master from %u, but no " mlog 1804 fs/ocfs2/dlm/dlmmaster.c mlog(0, "no bits set in the maybe_map, but %u " mlog 1809 fs/ocfs2/dlm/dlmmaster.c mlog(0, "master %u was found, %u should " mlog 1816 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%u is the lowest node, " mlog 1825 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: got cleanup assert" mlog 1830 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: got unrelated assert" mlog 1849 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%u asserting but %.*s is " mlog 1856 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "DIE! Mastery assert from %u, " mlog 1867 fs/ocfs2/dlm/dlmmaster.c mlog(0, "owner %u re-asserting on " mlog 1872 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "got assert_master from " mlog 1879 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "got assert from %u, but lock " mlog 1889 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "got assert from %u, but " mlog 1933 fs/ocfs2/dlm/dlmmaster.c mlog(0, "finishing off migration of lockres %.*s, " mlog 1968 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s:%.*s: got assert master from %u " mlog 1987 fs/ocfs2/dlm/dlmmaster.c mlog(0, "assert_master from %u, but current " mlog 2004 fs/ocfs2/dlm/dlmmaster.c mlog(0, "need to tell master to reassert\n"); mlog 2008 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "strange, got assert from %u, MASTER " mlog 2016 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: got assert from %u, need a ref\n", mlog 2023 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Bad message received from another node. Dumping state " mlog 2070 fs/ocfs2/dlm/dlmmaster.c mlog(0, "IGNORE HIGHER: %.*s\n", res->lockname.len, mlog 2126 fs/ocfs2/dlm/dlmmaster.c mlog(0, "Someone asked us to assert mastery, but we're " mlog 2137 fs/ocfs2/dlm/dlmmaster.c mlog(0, "worker about to master %.*s here, this=%u\n", mlog 2154 fs/ocfs2/dlm/dlmmaster.c mlog(0, "finished with dlm_assert_master_worker\n"); mlog 2197 fs/ocfs2/dlm/dlmmaster.c mlog(ML_NOTICE, "%s: node %u has not seen " mlog 2205 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: reco lock master is %u\n", dlm->name, mlog 2236 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s: res %.*s, error %d send DEREF to node %u\n", mlog 2240 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s: res %.*s, DEREF to node %u got %d\n", mlog 2274 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid name length!"); mlog 2278 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid node number: %u\n", node); mlog 2288 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s:%.*s: bad lockres name\n", mlog 2310 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s:%.*s: node %u trying to drop ref " mlog 2366 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid name length!"); mlog 2370 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid node number: %u\n", node); mlog 2380 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s:%.*s: bad lockres name\n", mlog 2389 fs/ocfs2/dlm/dlmmaster.c mlog(ML_NOTICE, "%s:%.*s: node %u sends deref done " mlog 2430 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s: res %.*s, error %d send DEREF DONE " mlog 2435 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s: res %.*s, DEREF to node %u got %d\n", mlog 2464 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s node %u ref dropped in dispatch\n", mlog 2468 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s:%.*s: node %u trying to drop ref " mlog 2515 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: Not migratable res %.*s, lock %u:%llu on " mlog 2531 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, Migratable\n", dlm->name, res->lockname.len, mlog 2562 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: Migrating %.*s to node %u\n", dlm->name, namelen, name, mlog 2599 fs/ocfs2/dlm/dlmmaster.c mlog(0, "another process is already migrating it\n"); mlog 2609 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "tried to migrate %.*s to %u, but " mlog 2656 fs/ocfs2/dlm/dlmmaster.c mlog(0, "migration to node %u failed with %d\n", mlog 2694 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: timed out during migration\n", mlog 2699 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: expected migration " mlog 2715 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: caught signal during migration\n", mlog 2749 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: Migrating %.*s to %u, returns %d\n", dlm->name, namelen, mlog 2787 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: res %.*s, Migrate to node %u failed with %d\n", mlog 2841 fs/ocfs2/dlm/dlmmaster.c mlog(0, "dlm_mark_lockres_migrating: %.*s, from %u to %u\n", mlog 2866 fs/ocfs2/dlm/dlmmaster.c mlog(0, "about to wait on migration_wq, dirty=%s\n", mlog 2876 fs/ocfs2/dlm/dlmmaster.c mlog(0, "woken again: migrating? %s, dead? %s\n", mlog 2880 fs/ocfs2/dlm/dlmmaster.c mlog(0, "all is well: migrating? %s, dead? %s\n", mlog 2885 fs/ocfs2/dlm/dlmmaster.c mlog(0, "trying again...\n"); mlog 2893 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "aha. migration target %u just went down\n", mlog 2941 fs/ocfs2/dlm/dlmmaster.c mlog(0, "putting lock for node %u\n", mlog 2967 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: node %u had a ref to this " mlog 3063 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "%s: res %.*s, Error %d send " mlog 3067 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "unhandled error=%d!\n", ret); mlog 3073 fs/ocfs2/dlm/dlmmaster.c mlog(0, "migrate request (node %u) returned %d!\n", mlog 3080 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: need ref for node %u\n", mlog 3092 fs/ocfs2/dlm/dlmmaster.c mlog(0, "returning ret=%d\n", ret); mlog 3140 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Got a migrate request, but the " mlog 3207 fs/ocfs2/dlm/dlmmaster.c mlog(0, "tried to migrate %.*s, but some " mlog 3214 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "migration error mle: " mlog 3233 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s:%.*s: master=%u, newmaster=%u, " mlog 3313 fs/ocfs2/dlm/dlmmaster.c mlog(0, "mle found, but dead node %u would not have been " mlog 3321 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u was expected master\n", dead_node); mlog 3340 fs/ocfs2/dlm/dlmmaster.c mlog(0, "dlm=%s, dead node=%u\n", dlm->name, dead_node); mlog 3383 fs/ocfs2/dlm/dlmmaster.c mlog(ML_NOTICE, "%s: target %u died during " mlog 3395 fs/ocfs2/dlm/dlmmaster.c mlog(0, "%s: node %u died during migration from " mlog 3435 fs/ocfs2/dlm/dlmmaster.c mlog(0, "now time to do a migrate request to other nodes\n"); mlog 3443 fs/ocfs2/dlm/dlmmaster.c mlog(0, "doing assert master of %.*s to all except the original node\n", mlog 3457 fs/ocfs2/dlm/dlmmaster.c mlog(0, "doing assert master of %.*s back to %u\n", mlog 3462 fs/ocfs2/dlm/dlmmaster.c mlog(0, "assert master to original master failed " mlog 3556 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "bad mle: %p\n", mle); mlog 107 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: changing dead_node from %u to %u\n", mlog 116 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: changing new_master from %u to %u\n", mlog 153 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: work thread has %d work items\n", dlm->name, tot); mlog 190 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "starting dlm recovery thread...\n"); mlog 206 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "waiting for dlm recovery thread to exit\n"); mlog 242 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "%s(%d): recovery info, state=%s, dead=%u, master=%u\n", mlog 275 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "%s: reco state, node %u, state=%s\n", mlog 279 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "%s: lockres %.*s on recovering list\n", mlog 292 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm thread running for %s...\n", dlm->name); mlog 310 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "quitting DLM recovery thread\n"); mlog 400 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: reco thread %d in recovery: " mlog 444 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: no need do recovery after migrating all " mlog 453 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "new master %u died while recovering %u!\n", mlog 470 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "dead_node %u no longer in recovery map!\n", mlog 481 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s(%d):recovery thread found node %u in the recovery map!\n", mlog 504 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "another node will master this recovery session.\n"); mlog 523 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: Error %d remastering locks for node %u, " mlog 530 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "DONE mastering recovery of %s:%u here(this=%u)!\n", mlog 556 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: failed to alloc recovery area, " mlog 568 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: Requesting lock info from node %u\n", dlm->name, mlog 591 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "waited 1 sec for %u, " mlog 597 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u returned " mlog 615 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u died after requesting " mlog 623 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "now receiving recovery data from " mlog 628 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "already receiving recovery data from " mlog 633 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "already DONE receiving recovery data " mlog 641 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: Done requesting all lock info\n", dlm->name); mlog 652 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "checking recovery state of node %u\n", mlog 657 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "bad ndata state for " mlog 663 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u died after " mlog 670 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u still in state %s\n", mlog 677 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u state is done\n", mlog 681 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u state is finalize\n", mlog 688 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "pass #%d, all_nodes_done?: %s\n", ++pass, mlog 704 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "all nodes are done! send finalize\n"); mlog 713 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "should be done with recovery!\n"); mlog 715 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "finishing recovery of %s at %lu, " mlog 795 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "\n"); mlog 798 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm_request_all_locks: dead node is %u, sending request " mlog 811 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: Error %d send LOCK_REQUEST to node %u " mlog 834 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: node %u sent dead_node=%u, but local " mlog 887 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: recovery worker started, dead=%u, master=%u\n", mlog 895 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "%s: will not send recovery state, " mlog 901 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "%s: reco state invalid: reco(dead=%u, " mlog 925 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: node %u went down while sending " mlog 941 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: node %u went down while sending " mlog 959 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "sending DATA DONE message to %u, " mlog 966 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: Error %d send RECO_DATA_DONE to node %u " mlog 989 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "got DATA DONE: dead_node=%u, reco.dead_node=%u, " mlog 1008 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "bad ndata state for node %u:" mlog 1019 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u is DONE sending " mlog 1035 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "failed to find recovery node data for node " mlog 1039 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "leaving reco data done handler, ret=%d\n", ret); mlog 1059 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "AHA! there was " mlog 1076 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "found lockres owned by dead node while " mlog 1081 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "found UNKNOWN owner while doing recovery " mlog 1126 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: sending mig lockres (%s) to %u\n", mlog 1138 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: res %.*s, Error %d send MIG_LOCKRES to " mlog 1150 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "node %u told me to kill " mlog 1204 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "Mismatched lvb in lock cookie=%u:%llu, name=%.*s, " mlog 1281 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "sending to %u\n", send_to); mlog 1286 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "argh. lockres has %d locks. this will " mlog 1315 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: sending dummy lock to %u, %s\n", mlog 1328 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: dlm_send_mig_lockres_msg returned %d\n", mlog 1332 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u went down while sending %s " mlog 1372 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "Domain %s not joined! " mlog 1388 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s message received from node %u\n", mlog 1392 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "all done flag. all lockres data received!\n"); mlog 1412 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node is attempting to migrate " mlog 1429 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "lock %.*s is already migrating\n", mlog 1434 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "node is attempting to migrate " mlog 1504 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "recovery has passed me a lockres with an " mlog 1564 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm_lockres_master_requery ret=%d\n", mlog 1569 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "lockres %.*s not claimed. " mlog 1576 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "master needs to respond to sender " mlog 1587 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm_process_recovery_data returned %d\n", ret); mlog 1589 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm_process_recovery_data succeeded\n"); mlog 1660 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "lock master is %u\n", *real_master); mlog 1684 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "Error %d when sending message %u (key " mlog 1695 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u responded to master requery with %u\n", mlog 1811 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "running %d locks for this lockres\n", mres->num_locks); mlog 1818 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: dummy lock for %u\n", mlog 1859 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "Could not find local lock " mlog 1873 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "Mismatched node# in lock " mlog 1880 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "Migrate lock cookie %u:%llu, " mlog 1893 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "Lock cookie %u:%llu was on list %u " mlog 1912 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "just reordered a local lock!\n"); mlog 1963 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s:%.*s: received bad " mlog 2003 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s:%.*s: %u:%llu: lock already " mlog 2009 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "sent lock: type=%d, conv=%d, " mlog 2032 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: added lock for node %u, " mlog 2039 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "done running all the locks\n"); mlog 2064 fs/ocfs2/dlm/dlmrecovery.c mlog(0, mlog 2081 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node died with convert pending " mlog 2089 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node died with lock pending " mlog 2110 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node died with unlock pending " mlog 2121 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node died with cancel pending " mlog 2148 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: res %.*s, Changing owner from %u to %u\n", mlog 2193 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: res %.*s, Changing owner from %u to %u\n", mlog 2255 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "clearing %.*s lvb, dead node %u had EX\n", mlog 2307 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: freed %u locks for dead node %u, " mlog 2311 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s:%.*s: freed %u locks for dead node %u, " mlog 2319 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: dead node %u had a ref, but had " mlog 2365 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "AHA! there was " mlog 2397 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: owned by " mlog 2416 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s:%.*s: dead node %u had a ref, but had " mlog 2434 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: recovery master %d just died\n", mlog 2440 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: dead master %d had reached " mlog 2449 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "Clearing join state for node %u\n", idx); mlog 2455 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "for domain %s, node %d is already dead. " mlog 2465 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u already removed from domain!\n", idx); mlog 2478 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u being removed from domain map!\n", idx); mlog 2528 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "ast for recovery lock fired!, this=%u, dlm=%s\n", mlog 2534 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "bast for recovery lock fired!, this=%u, dlm=%s\n", mlog 2539 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "unlockast for recovery lock fired!\n"); mlog 2560 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "starting recovery of %s at %lu, dead=%u, this=%u\n", mlog 2569 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: dlmlock($RECOVERY) returned %d, lksb=%d\n", mlog 2573 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm=%s dlmlock says I got it (this=%u)\n", mlog 2579 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: got reco EX lock, but %u will " mlog 2590 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: got reco EX lock, but " mlog 2593 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: new master is %u " mlog 2605 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: dead=%u, this=%u, sending " mlog 2623 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "got DLM_DENIED, trying LKM_CANCEL\n"); mlog 2633 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "dlmunlock returned %d\n", ret); mlog 2636 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm=%s dlmlock says another node got it (this=%u)\n", mlog 2645 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: reco master taking awhile\n", mlog 2650 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: reco master %u is ready to recover %u\n", mlog 2654 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "dlm=%s dlmlock says master node died (this=%u)\n", mlog 2661 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: got %s from dlmlock($RECOVERY), " mlog 2670 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "recovery lock not found\n"); mlog 2686 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: dead node is %u\n", dlm->name, dead_node); mlog 2701 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "not sending begin reco to dead node " mlog 2706 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "not sending begin reco to self\n"); mlog 2711 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "attempting to send begin reco msg to %d\n", mlog 2721 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "%s: node %u was down when sending " mlog 2732 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: trying to start recovery of node " mlog 2745 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "begin reco of dlm %s to node %u " mlog 2753 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "recovery lock not found\n"); mlog 2777 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u wants to recover node %u (%u:%u) " mlog 2787 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u wants to recover node %u (%u:%u)\n", mlog 2796 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: new_master %u died, changing " mlog 2800 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: new_master %u NOT DEAD, changing " mlog 2807 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_NOTICE, "%s: dead_node previously set to %u, " mlog 2814 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "recovery master %u sees %u as dead, but this " mlog 2819 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%u not in domain/live_nodes map " mlog 2832 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: recovery started by node %u, for %u (%u:%u)\n", mlog 2850 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "finishing recovery for node %s:%u, " mlog 2872 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "Error %d when sending message %u (key " mlog 2879 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "node %u went down after this " mlog 2911 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: node %u finalizing recovery stage%d of " mlog 2918 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "node %u sent recovery finalize msg, but node " mlog 2924 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "node %u sent recovery finalize msg for dead " mlog 2934 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: received finalize1 from " mlog 2946 fs/ocfs2/dlm/dlmrecovery.c mlog(ML_ERROR, "%s: received finalize2 from " mlog 2960 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "%s: recovery done, reco master was %u, dead now %u, master now %u\n", mlog 118 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: Adding res %.*s to purge list\n", mlog 127 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: Removing res %.*s from purge list\n", mlog 161 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: Removing res %.*s from purgelist\n", mlog 169 fs/ocfs2/dlm/dlmthread.c mlog(ML_ERROR, "%s: res %.*s in use after deref\n", mlog 181 fs/ocfs2/dlm/dlmthread.c mlog(ML_ERROR, "%s: Resource %.*s not on the Tracking list\n", mlog 205 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: Purging res %.*s, master %d\n", dlm->name, mlog 210 fs/ocfs2/dlm/dlmthread.c mlog(ML_NOTICE, "%s: res %.*s already in DLM_LOCK_RES_DROPPING_REF state\n", mlog 237 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: Removing res %.*s from purgelist, master %d\n", mlog 245 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: deref %.*s in progress\n", mlog 252 fs/ocfs2/dlm/dlmthread.c mlog(ML_ERROR, "%s: res %.*s in use after deref\n", mlog 264 fs/ocfs2/dlm/dlmthread.c mlog(ML_ERROR, "Resource %.*s not on the Tracking list\n", mlog 320 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s is in use or being remastered, " mlog 365 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s has locks on the convert queue\n", dlm->name, mlog 370 fs/ocfs2/dlm/dlmthread.c mlog(ML_ERROR, "%s: res %.*s converting lock to invalid mode\n", mlog 413 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s, AST for Converting lock %u:%llu, type " mlog 475 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s, AST for Blocked lock %u:%llu, type %d, " mlog 532 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s\n", dlm->name, res->lockname.len, mlog 540 fs/ocfs2/dlm/dlmthread.c mlog(0, "Starting dlm_thread...\n"); mlog 556 fs/ocfs2/dlm/dlmthread.c mlog(ML_KTHREAD, "Waiting for dlm thread to exit\n"); mlog 587 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s, Flush AST for lock %u:%llu, type %d, " mlog 613 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s, AST queued while flushing last " mlog 646 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s, Flush BAST for lock %u:%llu, " mlog 665 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s, BAST queued while flushing last " mlog 691 fs/ocfs2/dlm/dlmthread.c mlog(0, "dlm thread running for %s...\n", dlm->name); mlog 735 fs/ocfs2/dlm/dlmthread.c mlog(ML_ERROR, "%s: inprog %d, mig %d, reco %d," mlog 755 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: res %.*s, inprogress, delay list " mlog 791 fs/ocfs2/dlm/dlmthread.c mlog(0, "%s: Throttling dlm thread\n", mlog 812 fs/ocfs2/dlm/dlmthread.c mlog(0, "quitting DLM thread\n"); mlog 96 fs/ocfs2/dlm/dlmunlock.c mlog(0, "master_node = %d, valblk = %d\n", master_node, mlog 110 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "lockres %.*s: Someone is calling dlmunlock " mlog 119 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "lockres in progress!\n"); mlog 186 fs/ocfs2/dlm/dlmunlock.c mlog(0, "%s:%.*s: clearing actions, %s\n", mlog 218 fs/ocfs2/dlm/dlmunlock.c mlog(0, "clearing convert_type at %smaster node\n", mlog 251 fs/ocfs2/dlm/dlmunlock.c mlog(0, "lock %u:%llu should be gone now! refs=%d\n", mlog 323 fs/ocfs2/dlm/dlmunlock.c mlog(0, "%.*s\n", res->lockname.len, res->lockname.name); mlog 329 fs/ocfs2/dlm/dlmunlock.c mlog(0, "%s:%.*s: this node became the master due to a " mlog 357 fs/ocfs2/dlm/dlmunlock.c mlog(0, "master was in-progress. retry\n"); mlog 360 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " mlog 408 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "bad args! GET_LVB specified on unlock!\n"); mlog 413 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "bad args! cannot modify lvb on a CANCEL " mlog 419 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "Invalid name length in unlock handler!\n"); mlog 429 fs/ocfs2/dlm/dlmunlock.c mlog(0, "lvb: %s\n", flags & LKM_PUT_LVB ? "put lvb" : "none"); mlog 436 fs/ocfs2/dlm/dlmunlock.c mlog(0, "returning DLM_FORWARD -- res no longer exists\n"); mlog 446 fs/ocfs2/dlm/dlmunlock.c mlog(0, "returning DLM_RECOVERING\n"); mlog 453 fs/ocfs2/dlm/dlmunlock.c mlog(0, "returning DLM_MIGRATING\n"); mlog 460 fs/ocfs2/dlm/dlmunlock.c mlog(0, "returning DLM_FORWARD -- not master\n"); mlog 501 fs/ocfs2/dlm/dlmunlock.c mlog(0, "lockres is in progress\n"); mlog 511 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "failed to find lock to unlock! " mlog 553 fs/ocfs2/dlm/dlmunlock.c mlog(ML_ERROR, "lock to cancel is not on any list!\n"); mlog 606 fs/ocfs2/dlm/dlmunlock.c mlog(0, "VALBLK given with CANCEL: ignoring VALBLK\n"); mlog 625 fs/ocfs2/dlm/dlmunlock.c mlog(0, "lock=%p res=%p\n", lock, res); mlog 636 fs/ocfs2/dlm/dlmunlock.c mlog(0, "done calling dlmunlock_master: returned %d, " mlog 641 fs/ocfs2/dlm/dlmunlock.c mlog(0, "done calling dlmunlock_remote: returned %d, " mlog 660 fs/ocfs2/dlm/dlmunlock.c mlog(0, "retrying unlock due to pending recovery/" mlog 666 fs/ocfs2/dlm/dlmunlock.c mlog(0, "calling unlockast(%p, %d)\n", data, status); mlog 685 fs/ocfs2/dlm/dlmunlock.c mlog(0, "kicking the thread\n"); mlog 694 fs/ocfs2/dlm/dlmunlock.c mlog(0, "returning status=%d!\n", status); mlog 128 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "open called on inode %lu, flags 0x%x\n", inode->i_ino, mlog 175 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "close called on inode %lu\n", inode->i_ino); mlog 234 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "inode %lu, count = %zu, *ppos = %llu\n", mlog 268 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "read %zd bytes\n", readlen); mlog 281 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "inode %lu, count = %zu, *ppos = %llu\n", mlog 309 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "wrote %zu bytes\n", count); mlog 347 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "inode %lu\n", inode->i_ino); mlog 359 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "we're a directory, ip->ip_conn = 0x%p\n", ip->ip_conn); mlog 452 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "mkdir %.*s\n", domain->len, domain->name); mlog 457 fs/ocfs2/dlmfs/dlmfs.c mlog(ML_ERROR, "invalid domain name for directory.\n"); mlog 473 fs/ocfs2/dlmfs/dlmfs.c mlog(ML_ERROR, "Error %d could not register domain \"%.*s\"\n", mlog 499 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "create %.*s\n", name->len, name->name); mlog 506 fs/ocfs2/dlmfs/dlmfs.c mlog(ML_ERROR, "invalid lock name, %.*s\n", name->len, mlog 530 fs/ocfs2/dlmfs/dlmfs.c mlog(0, "unlink inode %lu\n", inode->i_ino); mlog 536 fs/ocfs2/dlmfs/dlmfs.c mlog(ML_ERROR, "unlink %pd, error %d from destroy\n", mlog 94 fs/ocfs2/dlmfs/userdlm.c mlog(ML_ERROR, "Dlm error %d while calling %s on " \ mlog 118 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "AST fired for lockres %.*s, level %d => %d\n", mlog 126 fs/ocfs2/dlmfs/userdlm.c mlog(ML_ERROR, "lksb status value of %u on lockres %.*s\n", mlog 205 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "BAST fired for lockres %.*s, blocking %d, level %d\n", mlog 223 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "UNLOCK AST fired for lockres %.*s, flags 0x%x\n", mlog 227 fs/ocfs2/dlmfs/userdlm.c mlog(ML_ERROR, "dlm returns status %d\n", status); mlog 293 fs/ocfs2/dlmfs/userdlm.c mlog(0, "lockres %.*s\n", lockres->l_namelen, lockres->l_name); mlog 311 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s USER_LOCK_BLOCKED\n", mlog 318 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s USER_LOCK_IN_TEARDOWN\n", mlog 326 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s USER_LOCK_IN_CANCEL\n", mlog 348 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s, EX/PR Holders %u,%u\n", mlog 357 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s, EX Holders %u\n", mlog 367 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s, downconvert %d => %d\n", mlog 422 fs/ocfs2/dlmfs/userdlm.c mlog(ML_ERROR, "lockres %.*s: invalid request!\n", mlog 428 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s, level %d, flags = 0x%x\n", mlog 521 fs/ocfs2/dlmfs/userdlm.c mlog(ML_ERROR, "lockres %.*s: invalid request!\n", mlog 599 fs/ocfs2/dlmfs/userdlm.c mlog(ML_BASTS, "lockres %.*s\n", lockres->l_namelen, lockres->l_name); mlog 119 fs/ocfs2/dlmglue.c mlog(level, "LVB information for %s (called from %s:%u):\n", mlog 121 fs/ocfs2/dlmglue.c mlog(level, "version: %u, clusters: %u, generation: 0x%x\n", mlog 124 fs/ocfs2/dlmglue.c mlog(level, "size: %llu, uid %u, gid %u, mode 0x%x\n", mlog 128 fs/ocfs2/dlmglue.c mlog(level, "nlink %u, atime_packed 0x%llx, ctime_packed 0x%llx, " mlog 360 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "DLM error %d while calling %s on resource %s\n", \ mlog 363 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "DLM error %d while calling %s on resource %.*s%08x\n", \ mlog 402 fs/ocfs2/dlmglue.c mlog(0, "built lock resource with name: %s\n", name); mlog 410 fs/ocfs2/dlmglue.c mlog(0, "Add tracking for lockres %s\n", res->l_name); mlog 992 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, block %d, level %d, l_block %d, dwn %d\n", mlog 998 fs/ocfs2/dlmglue.c mlog(0, "needs_downconvert = %d\n", needs_downconvert); mlog 1117 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "BAST fired for lockres %s, blocking %d, level %d, " mlog 1156 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "lockres %s: lksb status value of %d!\n", mlog 1162 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "AST fired for lockres %s, action %d, unlock %d, " mlog 1178 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "lockres %s: AST fired with invalid action: %u, " mlog 1210 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "UNLOCK AST fired for lockres %s, action = %d\n", mlog 1215 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "Dlm passes error %d for lock %s, " mlog 1224 fs/ocfs2/dlmglue.c mlog(0, "Cancel convert success for %s\n", lockres->l_name); mlog 1312 fs/ocfs2/dlmglue.c mlog(0, "lock %s, level = %d, flags = %u\n", lockres->l_name, level, mlog 1340 fs/ocfs2/dlmglue.c mlog(0, "lock %s, return from ocfs2_dlm_lock\n", lockres->l_name); mlog 1553 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "lockres %s has action %u pending\n", mlog 1572 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, convert from %d to %d\n", mlog 1594 fs/ocfs2/dlmglue.c mlog(0, "lock %s, successful return from ocfs2_dlm_lock\n", mlog 1724 fs/ocfs2/dlmglue.c mlog(0, "Inode %llu\n", (unsigned long long)OCFS2_I(inode)->ip_blkno); mlog 1764 fs/ocfs2/dlmglue.c mlog(0, "inode %llu take %s RW lock\n", mlog 1788 fs/ocfs2/dlmglue.c mlog(0, "inode %llu try to take %s RW lock\n", mlog 1809 fs/ocfs2/dlmglue.c mlog(0, "inode %llu drop %s RW lock\n", mlog 1826 fs/ocfs2/dlmglue.c mlog(0, "inode %llu take PRMODE open lock\n", mlog 1848 fs/ocfs2/dlmglue.c mlog(0, "inode %llu try to take %s open lock\n", mlog 1885 fs/ocfs2/dlmglue.c mlog(0, "inode %llu drop open lock\n", mlog 1938 fs/ocfs2/dlmglue.c mlog(0, "Cancel returning %d. flags: 0x%lx, level: %d, act: %d\n", ret, mlog 1982 fs/ocfs2/dlmglue.c mlog(ML_ERROR, mlog 2059 fs/ocfs2/dlmglue.c mlog(0, "Lock: \"%s\" ex: %d, trylock: %d, returns: %d\n", mlog 2082 fs/ocfs2/dlmglue.c mlog(0, "Unlock: \"%s\" flags: 0x%lx, level: %d, act: %d\n", mlog 2283 fs/ocfs2/dlmglue.c mlog(0, "status %d\n", status); mlog 2318 fs/ocfs2/dlmglue.c mlog(0, "Orphaned inode %llu was deleted while we " mlog 2337 fs/ocfs2/dlmglue.c mlog(0, "Trusting LVB on inode %llu\n", mlog 2422 fs/ocfs2/dlmglue.c mlog(0, "inode %llu, take %s META lock\n", mlog 2608 fs/ocfs2/dlmglue.c mlog(0, "inode %llu drop %s META lock\n", mlog 2693 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "Recursive locking is not permitted to " mlog 2864 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "lock on nfs sync lock failed %d\n", status); mlog 3035 fs/ocfs2/dlmglue.c mlog(0, "End of list found, %p\n", ret); mlog 3334 fs/ocfs2/dlmglue.c mlog(ML_ERROR, mlog 3405 fs/ocfs2/dlmglue.c mlog(0, "waiting on busy lock \"%s\": flags = %lx, action = " mlog 3428 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "destroying busy lock: \"%s\"\n", mlog 3431 fs/ocfs2/dlmglue.c mlog(0, "destroying blocked lock: \"%s\"\n", lockres->l_name); mlog 3449 fs/ocfs2/dlmglue.c mlog(0, "lock %s\n", lockres->l_name); mlog 3454 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "lockres flags: %lu\n", lockres->l_flags); mlog 3458 fs/ocfs2/dlmglue.c mlog(0, "lock %s, successful return from ocfs2_dlm_unlock\n", mlog 3524 fs/ocfs2/dlmglue.c mlog(0, "Waiting on lockres %s\n", lockres->l_name); mlog 3593 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "lockres %s, lvl %d <= %d, blcklst %d, mask %d, " mlog 3605 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, level %d => %d, blocking %d\n", mlog 3623 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, level %d => %d\n", lockres->l_name, mlog 3667 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, skip convert\n", lockres->l_name); mlog 3682 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s\n", lockres->l_name); mlog 3699 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s\n", lockres->l_name); mlog 3754 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, ReQ: Pending\n", mlog 3789 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, Aborting dc\n", lockres->l_name); mlog 3800 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, ReQ: EX/PR Holders %u,%u\n", mlog 3810 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, ReQ: EX Holders %u\n", mlog 3821 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, ReQ: Lock Refreshing\n", mlog 3830 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, ReQ: Checkpointing\n", mlog 3852 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, UNBLOCK_STOP_POST\n", mlog 3861 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, block=%d:%d, level=%d:%d, " mlog 3914 fs/ocfs2/dlmglue.c mlog(0, "generation: %u\n", oi->ip_dir_lock_gen); mlog 3931 fs/ocfs2/dlmglue.c mlog(ML_ERROR, "Could not sync inode %llu for downconvert!", mlog 4059 fs/ocfs2/dlmglue.c mlog(0, "extra_ref = %d\n", extra_ref); mlog 4080 fs/ocfs2/dlmglue.c mlog(0, "d_delete(%pd);\n", dentry); mlog 4276 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s blocked\n", lockres->l_name); mlog 4299 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s, requeue = %s.\n", lockres->l_name, mlog 4319 fs/ocfs2/dlmglue.c mlog(ML_BASTS, "lockres %s won't be scheduled: flags 0x%lx\n", mlog 4409 fs/ocfs2/dlmglue.c mlog(0, "downconvert_thread: awoken\n"); mlog 66 fs/ocfs2/export.c mlog(ML_ERROR, "getting nfs sync lock(EX) failed %d\n", status); mlog 80 fs/ocfs2/export.c mlog(ML_ERROR, "test inode bit failed %d\n", status); mlog 144 fs/ocfs2/export.c mlog(ML_ERROR, "getting nfs sync lock(EX) failed %d\n", status); mlog 168 fs/ocfs2/export.c mlog(ML_ERROR, "test inode bit failed %d\n", status); mlog 991 fs/ocfs2/extent_map.c mlog(ML_ERROR, mlog 1192 fs/ocfs2/file.c mlog(ML_ERROR, "Another case of recursive locking:\n"); mlog 1354 fs/ocfs2/file.c mlog(ML_ERROR, "Another case of recursive locking:\n"); mlog 222 fs/ocfs2/filecheck.c mlog(ML_NOTICE, mlog 484 fs/ocfs2/filecheck.c mlog(ML_NOTICE, mlog 112 fs/ocfs2/heartbeat.c mlog(ML_ERROR, "bit=%d map->num_nodes=%d\n", bit, map->num_nodes); mlog 315 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 854 fs/ocfs2/inode.c mlog(ML_ERROR, "Skipping delete of root inode.\n"); mlog 873 fs/ocfs2/inode.c mlog(ML_ERROR, "Skipping delete of system file %llu\n", mlog 934 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1034 fs/ocfs2/inode.c mlog(ML_ERROR, "getting nfs sync lock(PR) failed %d\n", status); mlog 1380 fs/ocfs2/inode.c mlog(ML_ERROR, "Checksum failed for dinode %llu\n", mlog 1446 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1453 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1462 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1471 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1481 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1506 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1514 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1536 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1546 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 1555 fs/ocfs2/inode.c mlog(ML_ERROR, mlog 350 fs/ocfs2/ioctl.c mlog(ML_ERROR, "unable to get alloc inode in " mlog 467 fs/ocfs2/ioctl.c mlog(ML_ERROR, "Can't read the group descriptor # " mlog 644 fs/ocfs2/ioctl.c mlog(ML_ERROR, "unable to get global_bitmap inode\n"); mlog 555 fs/ocfs2/journal.c mlog(ML_ERROR, mlog 655 fs/ocfs2/journal.c mlog(ML_ERROR, "giving me a buffer that's not uptodate!\n"); mlog 656 fs/ocfs2/journal.c mlog(ML_ERROR, "b_blocknr=%llu, b_state=0x%lx\n", mlog 699 fs/ocfs2/journal.c mlog(ML_ERROR, "Unknown access type!\n"); mlog 706 fs/ocfs2/journal.c mlog(ML_ERROR, "Error %d getting %d access to buffer!\n", mlog 786 fs/ocfs2/journal.c mlog(ML_ERROR, "jbd2_journal_dirty_metadata failed. " mlog 838 fs/ocfs2/journal.c mlog(ML_ERROR, "access error (bad inode)\n"); mlog 854 fs/ocfs2/journal.c mlog(ML_ERROR, "Could not get lock on journal!\n"); mlog 862 fs/ocfs2/journal.c mlog(ML_ERROR, "Journal file size (%lld) is too small!\n", mlog 875 fs/ocfs2/journal.c mlog(ML_ERROR, "Linux journal layer error\n"); mlog 1042 fs/ocfs2/journal.c mlog(ML_ERROR, "File system error %d recorded in " mlog 1044 fs/ocfs2/journal.c mlog(ML_ERROR, "File system on device %s needs checking.\n", mlog 1063 fs/ocfs2/journal.c mlog(ML_ERROR, "Failed to load journal!\n"); mlog 1090 fs/ocfs2/journal.c mlog(ML_ERROR, "unable to launch ocfs2commit thread, " mlog 1440 fs/ocfs2/journal.c mlog(ML_ERROR, mlog 1444 fs/ocfs2/journal.c mlog(ML_ERROR, "Volume requires unmount.\n"); mlog 1615 fs/ocfs2/journal.c mlog(ML_ERROR, "Could not lock journal!\n"); mlog 1649 fs/ocfs2/journal.c mlog(ML_ERROR, "Linux journal layer error\n"); mlog 1784 fs/ocfs2/journal.c mlog(ML_ERROR, "access error\n"); mlog 1789 fs/ocfs2/journal.c mlog(ML_ERROR, "access error (bad inode)\n"); mlog 2283 fs/ocfs2/journal.c mlog(0, "mount error, exiting!\n"); mlog 2313 fs/ocfs2/journal.c mlog(ML_ERROR, "status = %d, journal is " mlog 2324 fs/ocfs2/journal.c mlog(ML_KTHREAD, mlog 283 fs/ocfs2/localalloc.c mlog(ML_NOTICE, "Requested local alloc window %d is larger " mlog 312 fs/ocfs2/localalloc.c mlog(ML_ERROR, "Invalid local alloc inode, %llu\n", mlog 320 fs/ocfs2/localalloc.c mlog(ML_ERROR, "Local alloc size is invalid (la_size = %u)\n", mlog 335 fs/ocfs2/localalloc.c mlog(ML_ERROR, "inconsistent detected, clean journal with" mlog 187 fs/ocfs2/move_extents.c mlog(0, "reserve metadata_blocks: %d, data_clusters: %u, credits: %d\n", mlog 305 fs/ocfs2/move_extents.c mlog(0, "len_claimed: %u, len: %u\n", new_len, *len); mlog 314 fs/ocfs2/move_extents.c mlog(0, "cpos: %u, phys_cpos: %u, new_phys_cpos: %u\n", cpos, mlog 454 fs/ocfs2/move_extents.c mlog(0, "find the victim group: #%llu, " mlog 525 fs/ocfs2/move_extents.c mlog(0, "extents get ready to be moved to #%llu block\n", mlog 570 fs/ocfs2/move_extents.c mlog(0, "found phys_cpos: %u to fit the wanted moving.\n", *phys_cpos); mlog 638 fs/ocfs2/move_extents.c mlog(ML_ERROR, "unable to get global_bitmap inode\n"); mlog 821 fs/ocfs2/move_extents.c mlog(0, "Inode: %llu, start: %llu, len: %llu, cstart: %u, clen: %u, " mlog 864 fs/ocfs2/move_extents.c mlog(0, "#Defrag: cpos: %u, phys_cpos: %u, " mlog 192 fs/ocfs2/namei.c mlog(ML_ERROR, "new_inode failed!\n"); mlog 718 fs/ocfs2/namei.c mlog(ML_ERROR, "%s: no old_dir_bh!\n", osb->uuid_str); mlog 1085 fs/ocfs2/namei.c mlog(ML_NOTICE, "max lookup times reached, filesystem " mlog 1305 fs/ocfs2/namei.c mlog(ML_ERROR, "no old_dir_bh!\n"); mlog 1617 fs/ocfs2/namei.c mlog(ML_ERROR, "need to change nlink for old dir " mlog 1950 fs/ocfs2/namei.c mlog(ML_ERROR, mlog 230 fs/ocfs2/quota_global.c mlog(ML_ERROR, "Quota write (off=%llu, len=%llu) cancelled " mlog 353 fs/ocfs2/quota_global.c mlog(ML_ERROR, "failed to get global quota inode (type=%d)\n", mlog 386 fs/ocfs2/quota_global.c mlog(ML_ERROR, "Cannot read global quota info (%d).\n", mlog 437 fs/ocfs2/quota_global.c mlog(ML_ERROR, "Cannot write global quota info structure\n"); mlog 499 fs/ocfs2/quota_global.c mlog(ML_ERROR, "Short read from global quota file " mlog 568 fs/ocfs2/quota_global.c mlog(ML_ERROR, "Failed to lock quota info, losing quota write" mlog 178 fs/ocfs2/quota_local.c mlog(ML_ERROR, "failed to read quota file header (type=%d)\n", mlog 184 fs/ocfs2/quota_local.c mlog(ML_ERROR, "quota file magic does not match (%u != %u)," mlog 190 fs/ocfs2/quota_local.c mlog(ML_ERROR, "quota file version does not match (%u != %u)," mlog 202 fs/ocfs2/quota_local.c mlog(ML_ERROR, "cannot get global quota file inode " mlog 210 fs/ocfs2/quota_local.c mlog(ML_ERROR, "failed to read global quota file header " mlog 216 fs/ocfs2/quota_local.c mlog(ML_ERROR, "global quota file magic does not match " mlog 222 fs/ocfs2/quota_local.c mlog(ML_ERROR, "global quota file version does not match " mlog 430 fs/ocfs2/quota_local.c mlog(ML_ERROR, "failed to read quota file info header " mlog 503 fs/ocfs2/quota_local.c mlog(ML_ERROR, "Failed to get quota structure " mlog 629 fs/ocfs2/quota_local.c mlog(ML_ERROR, "failed to read quota file info header " mlog 698 fs/ocfs2/quota_local.c mlog(ML_ERROR, "failed to allocate memory for ocfs2 quota" mlog 724 fs/ocfs2/quota_local.c mlog(ML_ERROR, "failed to read quota file info header " mlog 831 fs/ocfs2/quota_local.c mlog(ML_ERROR, "releasing quota file with used " mlog 950 fs/ocfs2/quota_local.c mlog(ML_ERROR, "Did not find empty entry in chunk %d with %u" mlog 91 fs/ocfs2/refcounttree.c mlog(ML_ERROR, "Checksum failed for refcount block %llu\n", mlog 234 fs/ocfs2/refcounttree.c mlog(ML_ERROR, "Duplicate refcount block %llu found!\n", mlog 85 fs/ocfs2/reservations.c mlog(ML_NOTICE, "Dumping resmap for device %s. Bitmap length: %u\n", mlog 92 fs/ocfs2/reservations.c mlog(ML_NOTICE, "start: %u\tend: %u\tlen: %u\tlast_start: %u" mlog 101 fs/ocfs2/reservations.c mlog(ML_NOTICE, "%d reservations found. LRU follows\n", i); mlog 105 fs/ocfs2/reservations.c mlog(ML_NOTICE, "LRU(%d) start: %u\tend: %u\tlen: %u\t" mlog 125 fs/ocfs2/reservations.c mlog(ML_ERROR, mlog 148 fs/ocfs2/reservations.c mlog(ML_ERROR, "reservation %d has bad start off!\n", mlog 154 fs/ocfs2/reservations.c mlog(ML_ERROR, "reservation %d has no length!\n", mlog 160 fs/ocfs2/reservations.c mlog(ML_ERROR, "reservation %d has invalid range!\n", mlog 166 fs/ocfs2/reservations.c mlog(ML_ERROR, "reservation %d extends past bitmap!\n", mlog 335 fs/ocfs2/reservations.c mlog(ML_ERROR, "Duplicate reservation window!\n"); mlog 538 fs/ocfs2/reservations.c mlog(ML_ERROR, "goal: %u next_resv: start %u len %u\n", mlog 307 fs/ocfs2/resize.c mlog(ML_ERROR, "The disk is too old and small. " mlog 387 fs/ocfs2/resize.c mlog(ML_ERROR, "Group descriptor # %llu has bad chain %u " mlog 392 fs/ocfs2/resize.c mlog(ML_ERROR, "Group descriptor # %llu has bit count %u but " mlog 397 fs/ocfs2/resize.c mlog(ML_ERROR, "Group descriptor # %llu has free bit count %u " mlog 422 fs/ocfs2/resize.c mlog(ML_ERROR, "add a group which is in the current volume.\n"); mlog 424 fs/ocfs2/resize.c mlog(ML_ERROR, "input chain exceeds the limit.\n"); mlog 426 fs/ocfs2/resize.c mlog(ML_ERROR, mlog 429 fs/ocfs2/resize.c mlog(ML_ERROR, "add group's clusters overflow.\n"); mlog 431 fs/ocfs2/resize.c mlog(ML_ERROR, "the cluster exceeds the maximum of a group\n"); mlog 433 fs/ocfs2/resize.c mlog(ML_ERROR, "the free cluster exceeds the total clusters\n"); mlog 435 fs/ocfs2/resize.c mlog(ML_ERROR, mlog 438 fs/ocfs2/resize.c mlog(ML_ERROR, "group blkno is invalid\n"); mlog 440 fs/ocfs2/resize.c mlog(ML_ERROR, "group descriptor check failed.\n"); mlog 488 fs/ocfs2/resize.c mlog(ML_ERROR, "The disk is too old and small." mlog 496 fs/ocfs2/resize.c mlog(ML_ERROR, "Can't read the group descriptor # %llu " mlog 223 fs/ocfs2/slot_map.c mlog(ML_ERROR, mlog 470 fs/ocfs2/slot_map.c mlog(ML_ERROR, "no free slots available!\n"); mlog 154 fs/ocfs2/suballoc.c mlog(ML_ERROR, fmt, ##__VA_ARGS__); \ mlog 257 fs/ocfs2/suballoc.c mlog(ML_ERROR, mlog 2327 fs/ocfs2/suballoc.c mlog(ML_ERROR, "minimum allocation requested %u exceeds " mlog 2711 fs/ocfs2/suballoc.c mlog(ML_ERROR, "read block %llu failed %d\n", mlog 2718 fs/ocfs2/suballoc.c mlog(ML_ERROR, "invalid inode %llu requested\n", mlog 2726 fs/ocfs2/suballoc.c mlog(ML_ERROR, "inode %llu has invalid suballoc slot %u\n", mlog 2772 fs/ocfs2/suballoc.c mlog(ML_ERROR, "suballoc bit %u out of range of %u\n", mlog 2784 fs/ocfs2/suballoc.c mlog(ML_ERROR, "read group %llu failed %d\n", mlog 2826 fs/ocfs2/suballoc.c mlog(ML_ERROR, "get alloc slot and bit failed %d\n", status); mlog 2837 fs/ocfs2/suballoc.c mlog(ML_ERROR, "unable to get alloc inode in slot %u\n", mlog 2847 fs/ocfs2/suballoc.c mlog(ML_ERROR, "lock on alloc inode on slot %u failed %d\n", mlog 2855 fs/ocfs2/suballoc.c mlog(ML_ERROR, "test suballoc bit failed %d\n", status); mlog 465 fs/ocfs2/super.c mlog(ML_ERROR, "Unable to load system inode %d, " mlog 494 fs/ocfs2/super.c mlog(ML_ERROR, "status=%d, sysfile=%d, slot=%d\n", mlog 626 fs/ocfs2/super.c mlog(ML_ERROR, "Cannot change heartbeat mode on remount\n"); mlog 633 fs/ocfs2/super.c mlog(ML_ERROR, "Cannot change data mode on remount\n"); mlog 642 fs/ocfs2/super.c mlog(ML_ERROR, "Cannot enable inode64 on remount\n"); mlog 658 fs/ocfs2/super.c mlog(ML_ERROR, "Remount on readonly device is forbidden.\n"); mlog 668 fs/ocfs2/super.c mlog(ML_ERROR, "Cannot remount RDWR " mlog 675 fs/ocfs2/super.c mlog(ML_ERROR, "Cannot remount RDWR because " mlog 739 fs/ocfs2/super.c mlog(ML_ERROR, "Hardware sector size too large: %d (max=%d)\n", mlog 757 fs/ocfs2/super.c mlog(ML_ERROR, "incompatible version: %u.%u\n", mlog 763 fs/ocfs2/super.c mlog(ML_ERROR, "incompatible volume signature: %8s\n", mlog 770 fs/ocfs2/super.c mlog(ML_ERROR, "This is an ocfs v1 filesystem which must be " mlog 816 fs/ocfs2/super.c mlog(ML_ERROR, "Cannot heartbeat on a locally " mlog 821 fs/ocfs2/super.c mlog(ML_ERROR, "Userspace stack expected, but " mlog 829 fs/ocfs2/super.c mlog(ML_ERROR, "Mismatching o2cb heartbeat modes\n"); mlog 837 fs/ocfs2/super.c mlog(ML_ERROR, "Heartbeat has to be started to mount " mlog 855 fs/ocfs2/super.c mlog(ML_ERROR, mlog 864 fs/ocfs2/super.c mlog(ML_ERROR, mlog 901 fs/ocfs2/super.c mlog(ML_ERROR, "Failed to suspend/unsuspend quotas on " mlog 993 fs/ocfs2/super.c mlog(ML_ERROR, "superblock probe failed!\n"); mlog 1037 fs/ocfs2/super.c mlog(ML_ERROR, "Readonly device detected but readonly " mlog 1046 fs/ocfs2/super.c mlog(ML_ERROR, "Local heartbeat specified on readonly " mlog 1054 fs/ocfs2/super.c mlog(ML_ERROR, "Recovery required on readonly " mlog 1118 fs/ocfs2/super.c mlog(ML_ERROR, "Unable to create device kset %s.\n", sb->s_id); mlog 1126 fs/ocfs2/super.c mlog(ML_ERROR, "Unable to create filecheck sysfs directory at " mlog 1210 fs/ocfs2/super.c mlog(ML_ERROR, "User quotas were requested, but this " mlog 1217 fs/ocfs2/super.c mlog(ML_ERROR, "Group quotas were requested, but this " mlog 1223 fs/ocfs2/super.c mlog(ML_ERROR, "ACL support requested but extended attributes " mlog 1382 fs/ocfs2/super.c mlog(ML_ERROR, mlog 1449 fs/ocfs2/super.c mlog(ML_ERROR, mlog 1463 fs/ocfs2/super.c mlog(ML_ERROR, "Invalid heartbeat mount options\n"); mlog 1634 fs/ocfs2/super.c mlog(ML_ERROR, "failed to get bitmap inode\n"); mlog 1761 fs/ocfs2/super.c mlog(ML_ERROR, "unable to set blocksize\n"); mlog 1801 fs/ocfs2/super.c mlog(ML_ERROR, "couldn't mount because cluster name on" mlog 1982 fs/ocfs2/super.c mlog(ML_ERROR, "The journal cannot address the entire volume. " mlog 2067 fs/ocfs2/super.c mlog(ML_ERROR, "Invalid number of node slots (%u)\n", mlog 2077 fs/ocfs2/super.c mlog(ML_ERROR, "Unable to initialize recovery state\n"); mlog 2105 fs/ocfs2/super.c mlog(ML_ERROR, "unable to alloc vol label\n"); mlog 2139 fs/ocfs2/super.c mlog(ML_ERROR, "couldn't mount because of unsupported " mlog 2145 fs/ocfs2/super.c mlog(ML_ERROR, "couldn't mount RDWR because of " mlog 2158 fs/ocfs2/super.c mlog(ML_ERROR, mlog 2187 fs/ocfs2/super.c mlog(ML_ERROR, "unable to alloc journal\n"); mlog 2213 fs/ocfs2/super.c mlog(ML_ERROR, "Volume has invalid cluster size (%d)\n", mlog 2225 fs/ocfs2/super.c mlog(ML_ERROR, "Volume too large " mlog 2233 fs/ocfs2/super.c mlog(ML_ERROR, "Out of memory trying to setup our uuid.\n"); mlog 2328 fs/ocfs2/super.c mlog(ML_ERROR, "found superblock with incorrect block " mlog 2336 fs/ocfs2/super.c mlog(ML_ERROR, "found superblock with bad version: " mlog 2343 fs/ocfs2/super.c mlog(ML_ERROR, "bad block number on superblock: " mlog 2349 fs/ocfs2/super.c mlog(ML_ERROR, "bad cluster size found: %u\n", mlog 2352 fs/ocfs2/super.c mlog(ML_ERROR, "bad root_blkno: 0\n"); mlog 2354 fs/ocfs2/super.c mlog(ML_ERROR, "bad system_dir_blkno: 0\n"); mlog 2356 fs/ocfs2/super.c mlog(ML_ERROR, mlog 2385 fs/ocfs2/super.c mlog(ML_ERROR, "Could not initialize journal!\n"); mlog 2414 fs/ocfs2/super.c mlog(ML_ERROR, "ocfs2 journal load failed! %d\n", status); mlog 186 fs/ocfs2/uptodate.c mlog(ML_ERROR, "Owner %llu, count = %u, purged = %u\n", mlog 328 fs/ocfs2/uptodate.c mlog(ML_ERROR, "Duplicate block %llu cached!\n", mlog 1992 fs/ocfs2/xattr.c mlog(ML_ERROR, mlog 1999 fs/ocfs2/xattr.c mlog(ML_ERROR, mlog 2007 fs/ocfs2/xattr.c mlog(ML_ERROR, mlog 5530 fs/ocfs2/xattr.c mlog(ML_ERROR, "Too much hash collision in xattr bucket %llu, "