0
0
mirror of https://github.com/mongodb/mongo.git synced 2024-11-28 16:24:56 +01:00
mongodb/jstests/sharding/shard3.js
2014-10-13 19:59:21 -04:00

173 lines
5.8 KiB
JavaScript

// shard3.js
// Include helpers for analyzing explain output.
load("jstests/libs/analyze_plan.js");
s = new ShardingTest( "shard3" , 2 , 1 , 2 , { enableBalancer : 1 } );
s2 = s._mongos[1];
db = s.getDB( "test" )
s.adminCommand( { enablesharding : "test" } );
s.adminCommand( { shardcollection : "test.foo" , key : { num : 1 } } );
assert( sh.getBalancerState() , "A1" )
sh.setBalancerState( false )
assert( ! sh.getBalancerState() , "A2" )
sh.setBalancerState( true )
assert( sh.getBalancerState() , "A3" )
sh.setBalancerState( false )
assert( ! sh.getBalancerState() , "A4" )
s.config.databases.find().forEach( printjson )
a = s.getDB( "test" ).foo;
b = s2.getDB( "test" ).foo;
primary = s.getServer( "test" ).getDB( "test" ).foo;
secondary = s.getOther( primary.name ).getDB( "test" ).foo;
a.save( { num : 1 } );
a.save( { num : 2 } );
a.save( { num : 3 } );
assert.eq( 3 , a.find().toArray().length , "normal A" );
assert.eq( 3 , b.find().toArray().length , "other A" );
assert.eq( 3 , primary.count() , "p1" )
assert.eq( 0 , secondary.count() , "s1" )
assert.eq( 1 , s.onNumShards( "foo" ) , "on 1 shards" );
s.adminCommand( { split : "test.foo" , middle : { num : 2 } } );
s.adminCommand( { movechunk : "test.foo" , find : { num : 3 } , to : s.getOther( s.getServer( "test" ) ).name, _waitForDelete : true } );
assert( primary.find().toArray().length > 0 , "blah 1" );
assert( secondary.find().toArray().length > 0 , "blah 2" );
assert.eq( 3 , primary.find().itcount() + secondary.find().itcount() , "blah 3" )
assert.eq( 3 , a.find().toArray().length , "normal B" );
assert.eq( 3 , b.find().toArray().length , "other B" );
printjson( primary._db._adminCommand( "shardingState" ) );
// --- filtering ---
function doCounts( name , total , onlyItCounts ){
total = total || ( primary.count() + secondary.count() );
if ( ! onlyItCounts )
assert.eq( total , a.count() , name + " count" );
assert.eq( total , a.find().sort( { n : 1 } ).itcount() , name + " itcount - sort n" );
assert.eq( total , a.find().itcount() , name + " itcount" );
assert.eq( total , a.find().sort( { _id : 1 } ).itcount() , name + " itcount - sort _id" );
return total;
}
var total = doCounts( "before wrong save" )
assert.writeOK(secondary.insert( { _id : 111 , num : -3 } ));
doCounts( "after wrong save" , total , true )
e = a.find().explain("executionStats").executionStats;
assert.eq( 3 , e.nReturned , "ex1" )
assert.eq( 0 , e.totalKeysExamined , "ex2" )
assert.eq( 4 , e.totalDocsExamined , "ex3" )
var chunkSkips = 0;
for (var shard in e.executionStages.shards) {
var theShard = e.executionStages.shards[shard];
chunkSkips += getChunkSkips(theShard.executionStages);
}
assert.eq( 1 , chunkSkips , "ex4" )
// SERVER-4612
// make sure idhack obeys chunks
x = a.findOne( { _id : 111 } )
assert( ! x , "idhack didn't obey chunk boundaries " + tojson(x) );
// --- move all to 1 ---
print( "MOVE ALL TO 1" );
assert.eq( 2 , s.onNumShards( "foo" ) , "on 2 shards" );
s.printCollectionInfo( "test.foo" );
assert( a.findOne( { num : 1 } ) )
assert( b.findOne( { num : 1 } ) )
print( "GOING TO MOVE" );
assert( a.findOne( { num : 1 } ) , "pre move 1" )
s.printCollectionInfo( "test.foo" );
myto = s.getOther( s.getServer( "test" ) ).name
print( "counts before move: " + tojson( s.shardCounts( "foo" ) ) );
s.adminCommand( { movechunk : "test.foo" , find : { num : 1 } , to : myto, _waitForDelete : true } )
print( "counts after move: " + tojson( s.shardCounts( "foo" ) ) );
s.printCollectionInfo( "test.foo" );
assert.eq( 1 , s.onNumShards( "foo" ) , "on 1 shard again" );
assert( a.findOne( { num : 1 } ) , "post move 1" )
assert( b.findOne( { num : 1 } ) , "post move 2" )
print( "*** drop" );
s.printCollectionInfo( "test.foo" , "before drop" );
a.drop();
s.printCollectionInfo( "test.foo" , "after drop" );
assert.eq( 0 , a.count() , "a count after drop" )
assert.eq( 0 , b.count() , "b count after drop" )
s.printCollectionInfo( "test.foo" , "after counts" );
assert.eq( 0 , primary.count() , "p count after drop" )
assert.eq( 0 , secondary.count() , "s count after drop" )
print( "*** dropDatabase setup" )
s.printShardingStatus()
s.adminCommand( { shardcollection : "test.foo" , key : { num : 1 } } );
a.save( { num : 2 } );
a.save( { num : 3 } );
s.adminCommand( { split : "test.foo" , middle : { num : 2 } } );
s.adminCommand( { movechunk : "test.foo" , find : { num : 3 } , to : s.getOther( s.getServer( "test" ) ).name, _waitForDelete : true } );
s.printShardingStatus();
s.printCollectionInfo( "test.foo" , "after dropDatabase setup" );
doCounts( "after dropDatabase setup2" )
s.printCollectionInfo( "test.foo" , "after dropDatabase setup3" );
print( "*** ready to call dropDatabase" )
res = s.getDB( "test" ).dropDatabase();
assert.eq( 1 , res.ok , "dropDatabase failed : " + tojson( res ) );
// Waiting for SERVER-2253
assert.eq( 0 , s.config.databases.count( { _id: "test" } ) , "database 'test' was dropped but still appears in configDB" );
s.printShardingStatus();
s.printCollectionInfo( "test.foo" , "after dropDatabase call 1" );
assert.eq( 0 , doCounts( "after dropDatabase called" ) )
// ---- retry commands SERVER-1471 ----
s.adminCommand( { enablesharding : "test2" } );
s.adminCommand( { shardcollection : "test2.foo" , key : { num : 1 } } );
dba = s.getDB( "test2" );
dbb = s2.getDB( "test2" );
dba.foo.save( { num : 1 } );
dba.foo.save( { num : 2 } );
dba.foo.save( { num : 3 } );
assert.eq( 1 , s.onNumShards( "foo" , "test2" ) , "B on 1 shards" );
assert.eq( 3 , dba.foo.count() , "Ba" );
assert.eq( 3 , dbb.foo.count() , "Bb" );
s.adminCommand( { split : "test2.foo" , middle : { num : 2 } } );
s.adminCommand( { movechunk : "test2.foo" , find : { num : 3 } , to : s.getOther( s.getServer( "test2" ) ).name, _waitForDelete : true } );
assert.eq( 2 , s.onNumShards( "foo" , "test2" ) , "B on 2 shards" );
x = dba.foo.stats()
printjson( x )
y = dbb.foo.stats()
printjson( y )
s.stop();