// More complete version of sharding/sync6.js // Test that distributed lock forcing does not result in inconsistencies, using a // fast timeout. // Note that this test will always have random factors, since we can't control the // thread scheduling. var bitbucket = _isWindows() ? "NUL" : "/dev/null"; test = new SyncCCTest("sync6", {logpath: bitbucket, logappend: ""}); // Startup another process to handle our commands to the cluster, mostly so it's // easier to read. var commandConn = startMongodTest( 30000 + 4, "syncCommander", false, {}); //{ logpath : bitbucket } )//{verbose : ""} ) // { logpath : MongoRunner.dataDir + "/syncCommander/mongod.log" } ); // Up the log level for this test commandConn.getDB( "admin" ).runCommand( { setParameter : 1, logLevel : 0 } ); // Have lots of threads, so use larger i // Can't test too many, we get socket exceptions... possibly due to the // javascript console. // TODO: Figure out our max bounds here - use less threads now to avoid pinger starvation issues. for ( var t = 0; t < 4; t++ ) { for ( var i = 4; i < 5; i++ ) { // Our force time is 6 seconds - slightly diff from sync6 to ensure exact time not important var takeoverMS = 6000; // Generate valid sleep and skew for this timeout var threadSleepWithLock = takeoverMS / 2; var configServerTimeSkew = [ 0, 0, 0 ]; for ( var h = 0; h < 3; h++ ) { // Skew by 1/30th the takeover time either way, at max configServerTimeSkew[h] = ( i + h ) % Math.floor( takeoverMS / 60 ); // Make skew pos or neg configServerTimeSkew[h] *= ( ( i + h ) % 2 ) ? -1 : 1; } // Build command command = { _testDistLockWithSkew : 1 }; // Basic test parameters command["lockName"] = "TimeSkewFailNewTest_lock_" + i; command["host"] = test.url; command["seed"] = i; command["numThreads"] = ( i % 50 ) + 1; // Critical values so we're sure of correct operation command["takeoverMS"] = takeoverMS; command["wait"] = 4 * takeoverMS; // so we must force the lock command["skewHosts"] = configServerTimeSkew; command["threadWait"] = threadSleepWithLock; // Less critical test params // 1/3 of threads will not release the lock command["hangThreads"] = 3; // Amount of time to wait before trying lock again command["threadSleep"] = 1; // ( ( i + 1 ) * 100 ) % (takeoverMS / 4) // Amount of total clock skew possible between locking threads (processes) // This can be large now. command["skewRange"] = ( command["takeoverMS"] * 3 ) * 60 * 1000; // Double-check our sleep, host skew, and takeoverMS values again // At maximum, our threads must sleep only half the lock timeout time. assert( command["threadWait"] <= command["takeoverMS"] / 2 ); for ( var h = 0; h < command["skewHosts"].length; h++ ) { // At maximum, our config server time skew needs to be less than 1/30th // the total time skew (1/60th either way). assert( Math.abs( command["skewHosts"][h] ) <= ( command["takeoverMS"] / 60 ) ); } result = commandConn.getDB( "admin" ).runCommand( command ); printjson( result ); printjson( command ); assert( result.ok, "Skewed threads did not increment correctly." ); } } stopMongoProgram( 30004 ); test.stop();