/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/fs/ocfs2/dlm/ |
H A D | dlmrecovery.c | 125 mlog(0, "%s: changing dead_node from %u to %u\n", 134 mlog(0, "%s: changing new_master from %u to %u\n", 171 mlog(0, "%s: work thread has %d work items\n", dlm->name, tot); 208 mlog(0, "starting dlm recovery thread...\n"); 224 mlog(0, "waiting for dlm recovery thread to exit\n"); 260 mlog(ML_NOTICE, "%s(%d): recovery info, state=%s, dead=%u, master=%u\n", 293 mlog(ML_NOTICE, "%s: reco state, node %u, state=%s\n", 297 mlog(ML_NOTICE, "%s: lockres %.*s on recovering list\n", 310 mlog(0, "dlm thread running for %s...\n", dlm->name); 328 mlog( [all...] |
H A D | dlmast.c | 100 mlog(ML_ERROR, "ast list not empty!! pending=%d, newlevel=%d\n", 105 mlog(0, "lock has an ast getting flushed right now\n"); 114 mlog(0, "%s: cancelling bast for %.*s\n", 158 mlog(0, "lock has a bast getting flushed right now\n"); 191 mlog(0, "getting lvb from lockres for %s node\n", 292 mlog(ML_ERROR, "Invalid name length (%d) in proxy ast " 299 mlog(ML_ERROR, "Both PUT and GET lvb specified, (0x%x)\n", 305 mlog(0, "lvb: %s\n", flags & LKM_PUT_LVB ? "put lvb" : 308 mlog(0, "type=%d, blocked_type=%d\n", past->type, past->blocked_type); 312 mlog(ML_ERRO [all...] |
H A D | dlmconvert.c | 138 mlog(ML_ERROR, "attempted to convert a lock with a lock " 146 mlog(ML_ERROR, "attempted to convert a lock not on grant " 156 mlog(0, "will set lvb: converting %s->%s\n", 165 mlog(0, "will fetch new value into " 171 mlog(0, "will NOT fetch new value " 208 mlog(0, "res %.*s, granting %s lock\n", res->lockname.len, 213 mlog(0, "doing in-place convert for nonlocal lock\n"); 224 mlog(0, "failed to convert NOQUEUE lock %.*s from " 230 mlog(0, "res %.*s, queueing...\n", res->lockname.len, 269 mlog( [all...] |
H A D | dlmunlock.c | 110 mlog(0, "master_node = %d, valblk = %d\n", master_node, 124 mlog(ML_ERROR, "lockres %.*s: Someone is calling dlmunlock " 133 mlog(ML_ERROR, "lockres in progress!\n"); 198 mlog(0, "%s:%.*s: clearing actions, %s\n", 226 mlog(0, "clearing convert_type at %smaster node\n", 248 mlog(0, "lock %u:%llu should be gone now! refs=%d\n", 326 mlog(0, "%s:%.*s: this node became the master due to a " 354 mlog(0, "master was in-progress. retry\n"); 357 mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " 403 mlog(ML_ERRO [all...] |
H A D | dlmlock.c | 145 mlog(0, "I can grant this lock right away\n"); 162 mlog(0, "%s: returning DLM_NORMAL to " 173 mlog(0, "%s: returning NOTQUEUED to " 228 mlog(0, "lockres %.*s, flags = 0x%x\n", res->lockname.len, 257 mlog(0, "%s: recovery lock was owned by " 278 mlog(0, "%s: $RECOVERY lock for this node (%u) is " 324 mlog(ML_ERROR, "%s:%.*s: BUG. this is a stale lockres " 332 mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " 337 mlog(0, "node %u died so returning DLM_RECOVERING " 372 mlog( [all...] |
H A D | dlmmaster.c | 239 mlog(ML_ERROR, "bad mle: %p\n", mle); 385 mlog(0, "node %u already removed from nodemap!\n", idx); 399 mlog(0, "node %u already in node map!\n", idx); 437 mlog(0, "Releasing mle for %.*s, type %d\n", mle->mnamelen, mle->mname, 499 mlog(0, "destroying lockres %.*s\n", res->lockname.len, 506 mlog(ML_ERROR, "Resource %.*s not on the Tracking list\n", 521 mlog(ML_ERROR, 650 mlog(0, "%s:%.*s: inflight++: now %u\n", 664 mlog(0, "%s:%.*s: inflight--: now %u\n", 708 mlog( [all...] |
H A D | dlmthread.c | 132 mlog(0, "putting lockres %.*s:%p onto purge list\n", 141 mlog(0, "removing lockres %.*s:%p from purge list, owner=%u\n", 175 mlog(0, "purging lockres %.*s, master = %d\n", res->lockname.len, 196 mlog(0, "%s:%.*s: dlm_deref_lockres returned %d\n", 203 mlog(0, "removing lockres %.*s:%p from purgelist, " 212 mlog(ML_ERROR, "found lockres %s:%.*s: in use after deref\n", 269 mlog(0, "lockres %s:%.*s: is in use or " 299 //mlog(0, "res->lockname.len=%d\n", res->lockname.len); 300 //mlog(0, "res->lockname.name=%p\n", res->lockname.name); 301 //mlog( [all...] |
H A D | dlmdomain.c | 94 mlog(0, "Allocated DLM hash pagevec; %d pages (%lu expected), %lu buckets per page\n", 326 mlog(0, "freeing memory from domain %s\n", dlm->name); 421 mlog(0, "Migrating locks from domain %s\n", dlm->name); 451 mlog(0, "%s: touched %d lockreses in bucket %d " 460 mlog(0, "%s: %d lock resources in hash last pass\n", 464 mlog(0, "DONE Migrating locks from domain %s\n", dlm->name); 488 mlog(0, "Node %d is joining, we wait on it.\n", 553 mlog(0, "Asking node %u if we can leave the domain %s me = %u\n", 563 mlog(ML_ERROR, "Error %d when sending message %u (key 0x%x) to " 565 mlog( [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/fs/ocfs2/cluster/ |
H A D | masklog.h | 84 /* NOTE: If you add a flag, you need to also update mlog.c! */ 203 #define mlog(mask, fmt, args...) do { \ macro 220 mlog(ML_ERROR, "status = %lld\n", (long long)_st); \ 225 mlog(ML_ENTRY, "ENTRY:" fmt , ##args); \ 229 mlog(ML_ENTRY, "ENTRY:\n"); \ 238 mlog(ML_EXIT, "EXIT: %lu\n", (unsigned long) (st)); \ 240 mlog(ML_EXIT, "EXIT: %ld\n", (signed long) (st)); \ 244 mlog(ML_EXIT, "EXIT: %u\n", (unsigned int) (st)); \ 248 mlog(ML_EXIT, "EXIT: %d\n", (signed int) (st)); \ 250 mlog(ML_EXI [all...] |
H A D | quorum.c | 117 mlog(0, "heartbeating: %d, connected: %d, " 130 mlog(ML_ERROR, "fencing this node because it is " 144 mlog(ML_ERROR, "fencing this node because it is " 153 mlog(ML_ERROR, "fencing this node because it is " 176 mlog(0, "node %u, %d total\n", node, qs->qs_holds); 185 mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1); 213 mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); 238 mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); 256 mlog(0, "node %u\n", node); 281 mlog( [all...] |
H A D | tcp.c | 87 mlog(ML_MSG, "[mag %u len %u typ %u stat %d sys_stat %d " \ 97 mlog(ML_SOCKET, "[sc %p refs %d sock %p node %u page %p " \ 328 mlog(0, "completed %d messages for node %u\n", num_kills, 468 mlog(ML_CONN, "node %u sc: %p -> %p, valid %u -> %u, err %d -> %d\n", 514 mlog(ML_CONN, "queueing conn attempt in %lu jiffies\n", delay); 762 mlog(0, "max_len for message handler out of range: %u\n", 769 mlog(0, "no message type provided: %u, %p\n", msg_type, func); 775 mlog(0, "no message handler provided: %u, %p\n", 806 mlog(ML_TCP, "registered handler func %p type %u key %08x\n", 832 mlog(ML_TC [all...] |
H A D | heartbeat.c | 172 mlog(ML_ERROR, "Heartbeat write timeout to device %s after %u " 180 mlog(ML_HEARTBEAT, "Queue write timeout for %u ms\n", 233 mlog(ML_ERROR, "IO Error %d\n", error); 262 mlog(ML_ERROR, "Could not alloc slots BIO!\n"); 281 mlog(ML_HB_BIO, "page %d, vec_len = %u, vec_start = %u\n", 375 mlog(ML_ERROR, "Dump slot information: seq = 0x%llx, node = %u, " 443 mlog(ML_HB_BIO, "our node generation = 0x%llx, cksum = 0x%x\n", 457 mlog(ML_HEARTBEAT, "calling funcs %p\n", f); 489 mlog(ML_HEARTBEAT, "Node %s event for %d\n", 520 mlog(ML_HEARTBEA [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/fs/ocfs2/ |
H A D | buffer_head_io.c | 110 mlog(ML_BH_IO, "No buffers will be read!\n"); 126 mlog(ML_BH_IO, 136 mlog(ML_ERROR, 145 mlog(ML_ERROR, 203 mlog(ML_ERROR, "asked to read %d blocks!\n", nr); 210 mlog(ML_BH_IO, "No buffers will be read!\n"); 254 mlog(ML_UPTODATE, 265 mlog(ML_BH_IO, "trying to sync read a jbd " 275 mlog(ML_BH_IO, "asking me to sync read a dirty " 292 mlog(ML_ERRO [all...] |
H A D | export.c | 62 mlog(0, "nfs wants inode with blkno: 0\n"); 81 mlog(ML_ERROR, "getting nfs sync lock(EX) failed %d\n", status); 93 mlog(0, "test inode bit failed %d\n", status); 96 mlog(ML_ERROR, "test inode bit failed %d\n", status); 103 mlog(0, "inode %llu suballoc bit is clear\n", 117 mlog(0, "stale inode ino: %llu generation: %u\n", 133 mlog(0, "stale inode ino: %llu generation: %u\n", 160 mlog(0, "find parent of directory %llu\n", 205 mlog(ML_ERROR, "fh buffer is too small for encoding\n"); 213 mlog( [all...] |
H A D | reservations.c | 93 mlog(ML_NOTICE, "Dumping resmap for device %s. Bitmap length: %u\n", 100 mlog(ML_NOTICE, "start: %u\tend: %u\tlen: %u\tlast_start: %u" 109 mlog(ML_NOTICE, "%d reservations found. LRU follows\n", i); 113 mlog(ML_NOTICE, "LRU(%d) start: %u\tend: %u\tlen: %u\t" 133 mlog(ML_ERROR, 156 mlog(ML_ERROR, "reservation %d has bad start off!\n", 162 mlog(ML_ERROR, "reservation %d has no length!\n", 168 mlog(ML_ERROR, "reservation %d has invalid range!\n", 174 mlog(ML_ERROR, "reservation %d extends past bitmap!\n", 324 mlog( [all...] |
H A D | uptodate.c | 155 mlog(0, "Purge item %llu\n", 183 mlog(0, "Purge %u %s items from Owner %llu\n", to_purge, 201 mlog(ML_ERROR, "Owner %llu, count = %u, purged = %u\n", 252 mlog(0, "Owner %llu, query block %llu (inline = %u)\n", 264 mlog(0, "index = %d, item = %p\n", index, item); 309 mlog(0, "block %llu takes position %u\n", (unsigned long long) block, 327 mlog(0, "Insert block %llu num = %u\n", (unsigned long long) block, 341 mlog(ML_ERROR, "Duplicate block %llu cached!\n", 392 mlog(0, "Expanded %llu to a tree cache: flags 0x%x, num = %u\n", 408 mlog( [all...] |
H A D | journal.c | 313 mlog(0, "No transactions for me to flush!\n"); 332 mlog(0, "commit_thread: flushed transaction %lu (%u handles)\n", 431 mlog(0, "Trying to extend transaction by %d blocks\n", nblocks); 444 mlog(0, 530 mlog(ML_ERROR, 635 mlog(ML_ERROR, "giving me a buffer that's not uptodate!\n"); 636 mlog(ML_ERROR, "b_blocknr=%llu\n", 662 mlog(ML_ERROR, "Unknown access type!\n"); 669 mlog(ML_ERROR, "Error %d getting %d access to buffer!\n", 794 mlog(ML_ERRO [all...] |
H A D | resize.c | 320 mlog(ML_ERROR, "The disk is too old and small. " 345 mlog(0, "extend the last group at %llu, new clusters = %d\n", 400 mlog(ML_ERROR, "Group descriptor # %llu has bad chain %u " 405 mlog(ML_ERROR, "Group descriptor # %llu has bit count %u but " 410 mlog(ML_ERROR, "Group descriptor # %llu has free bit count %u " 435 mlog(ML_ERROR, "add a group which is in the current volume.\n"); 437 mlog(ML_ERROR, "input chain exceeds the limit.\n"); 439 mlog(ML_ERROR, 442 mlog(ML_ERROR, "add group's clusters overflow.\n"); 444 mlog(ML_ERRO [all...] |
H A D | dcache.c | 56 mlog(0, "negative dentry: %.*s\n", dentry->d_name.len, 70 mlog(0, "inode (%llu) deleted, returning false\n", 81 mlog(0, "Inode %llu orphaned, returning false " 93 mlog(0, "Inode %llu doesn't have dentry lock, " 160 mlog(0, "dentry found: %.*s\n", 217 mlog(0, "Attach \"%.*s\", parent %llu, fsdata: %p\n", 257 mlog(0, "Found: %s\n", dl->dl_lockres.l_name); 432 mlog(ML_ERROR, "Dentry is missing cluster lock. "
|
H A D | heartbeat.c | 69 mlog(0, "ocfs2: node down event for %d\n", node_num); 126 mlog(ML_ERROR, "bit=%d map->num_nodes=%d\n", bit, map->num_nodes);
|
H A D | super.c | 466 mlog(ML_ERROR, "Unable to load system inode %d, " 496 mlog(ML_ERROR, "status=%d, sysfile=%d, slot=%d\n", 622 mlog(ML_ERROR, "Cannot change heartbeat mode on remount\n"); 629 mlog(ML_ERROR, "Cannot change data mode on remount\n"); 638 mlog(ML_ERROR, "Cannot enable inode64 on remount\n"); 654 mlog(ML_ERROR, "Remount on readonly device is forbidden.\n"); 660 mlog(0, "Going to ro mode.\n"); 664 mlog(0, "Making ro filesystem writeable.\n"); 667 mlog(ML_ERROR, "Cannot remount RDWR " 674 mlog(ML_ERRO [all...] |
H A D | inode.c | 156 mlog(0, "Inode was not in inode cache, reading it.\n"); 173 mlog(0, "returning inode with number %llu\n", 296 mlog(ML_ERROR, 310 mlog(0, "local alloc inode: i_ino=%lu\n", inode->i_ino); 316 mlog(0, "superblock inode: i_ino=%lu\n", inode->i_ino); 400 mlog(ML_ERROR, "bad inode\n"); 407 mlog(ML_ERROR, "bad inode args\n"); 698 mlog(0, "Recovery is happening on orphan dir %d, will skip " 821 mlog(ML_ERROR, "Skipping delete of root inode.\n"); 831 mlog( [all...] |
H A D | localalloc.c | 204 mlog(0, "requested: %dM, max: %uM, default: %uM\n", 279 mlog(0, "state=%d, bits=%llu, la_bits=%d, ret=%d\n", 300 mlog(ML_NOTICE, "Requested local alloc window %d is larger " 329 mlog(ML_ERROR, "Invalid local alloc inode, %llu\n", 337 mlog(ML_ERROR, "Local alloc size is invalid (la_size = %u)\n", 352 mlog(ML_ERROR, "Local alloc hasn't been recovered!\n" 367 mlog(0, "Local alloc window bits = %d\n", osb->local_alloc_bits); 716 mlog(0, "Calling in_range for max block %llu\n", 732 mlog(0, "bits=%d, slot=%d, ret=%d\n", bits_wanted, osb->slot_num, 826 mlog( [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/fs/ocfs2/dlmfs/ |
H A D | userdlm.c | 107 mlog(ML_ERROR, "Dlm error %d while calling %s on " \ 131 mlog(ML_BASTS, "AST fired for lockres %.*s, level %d => %d\n", 139 mlog(ML_ERROR, "lksb status value of %u on lockres %.*s\n", 218 mlog(ML_BASTS, "BAST fired for lockres %.*s, blocking %d, level %d\n", 236 mlog(ML_BASTS, "UNLOCK AST fired for lockres %.*s, flags 0x%x\n", 240 mlog(ML_ERROR, "dlm returns status %d\n", status); 306 mlog(0, "lockres %.*s\n", lockres->l_namelen, lockres->l_name); 324 mlog(ML_BASTS, "lockres %.*s USER_LOCK_BLOCKED\n", 331 mlog(ML_BASTS, "lockres %.*s USER_LOCK_IN_TEARDOWN\n", 339 mlog(ML_BAST [all...] |
H A D | dlmfs.c | 143 mlog(0, "open called on inode %lu, flags 0x%x\n", inode->i_ino, 190 mlog(0, "close called on inode %lu\n", inode->i_ino); 250 mlog(0, "inode %lu, count = %zu, *ppos = %llu\n", 284 mlog(0, "read %zd bytes\n", readlen); 298 mlog(0, "inode %lu, count = %zu, *ppos = %llu\n", 328 mlog(0, "wrote %zd bytes\n", writelen); 366 mlog(0, "inode %lu\n", inode->i_ino); 378 mlog(0, "we're a directory, ip->ip_conn = 0x%p\n", ip->ip_conn); 491 mlog(0, "mkdir %.*s\n", domain->len, domain->name); 496 mlog(ML_ERRO [all...] |