mirror of
https://github.com/mongodb/mongo.git
synced 2024-11-24 00:17:37 +01:00
2e6ca86ecd
GitOrigin-RevId: 8e890a8f96c9cb0c7eb3318c0d7e92256077ab49
69 lines
2.8 KiB
JavaScript
69 lines
2.8 KiB
JavaScript
import {ReplSetTest} from "jstests/libs/replsettest.js";
|
|
import {ShardTransitionUtil} from "jstests/libs/shard_transition_util.js";
|
|
import {ShardingTest} from "jstests/libs/shardingtest.js";
|
|
|
|
var s = new ShardingTest({shards: 2, other: {enableBalancer: true}});
|
|
var config = s.s0.getDB('config');
|
|
|
|
assert.commandWorked(
|
|
s.s0.adminCommand({enableSharding: 'needToMove', primaryShard: s.shard0.shardName}));
|
|
|
|
// Returns an error when trying to remove a shard that doesn't exist.
|
|
assert.commandFailedWithCode(s.s0.adminCommand({removeshard: "shardz"}), ErrorCodes.ShardNotFound);
|
|
|
|
var topologyTime0 = config.shards.findOne({_id: s.shard0.shardName}).topologyTime;
|
|
var topologyTime1 = config.shards.findOne({_id: s.shard1.shardName}).topologyTime;
|
|
assert.gt(topologyTime1, topologyTime0);
|
|
|
|
// removeShard is not permited on shard0 (the configShard) if configShard is enabled, so we want
|
|
// to use transitionToDedicatedConfigServer instead
|
|
var removeShardOrTransitionToDedicated =
|
|
TestData.configShard ? "transitionToDedicatedConfigServer" : "removeShard";
|
|
|
|
// First remove puts in draining mode, the second tells me a db needs to move, the third
|
|
// actually removes
|
|
assert.commandWorked(s.s0.adminCommand({[removeShardOrTransitionToDedicated]: s.shard0.shardName}));
|
|
|
|
// Can't make all shards in the cluster draining
|
|
assert.commandFailedWithCode(s.s0.adminCommand({removeshard: s.shard1.shardName}),
|
|
ErrorCodes.IllegalOperation);
|
|
|
|
var removeResult = assert.commandWorked(
|
|
s.s0.adminCommand({[removeShardOrTransitionToDedicated]: s.shard0.shardName}));
|
|
assert.eq(removeResult.dbsToMove, ['needToMove'], "didn't show db to move");
|
|
assert(removeResult.note !== undefined);
|
|
|
|
s.s0.getDB('needToMove').dropDatabase();
|
|
|
|
// Ensure the balancer moves the config.system.sessions collection chunks out of the shard being
|
|
// removed
|
|
s.awaitBalancerRound();
|
|
|
|
if (TestData.configShard) {
|
|
// A config shard can't be removed until all range deletions have finished.
|
|
ShardTransitionUtil.waitForRangeDeletions(s.s);
|
|
}
|
|
|
|
removeResult = assert.commandWorked(
|
|
s.s0.adminCommand({[removeShardOrTransitionToDedicated]: s.shard0.shardName}));
|
|
assert.eq('completed', removeResult.state, 'Shard was not removed: ' + tojson(removeResult));
|
|
|
|
var existingShards = config.shards.find({}).toArray();
|
|
assert.eq(
|
|
1, existingShards.length, "Removed server still appears in count: " + tojson(existingShards));
|
|
|
|
var topologyTime2 = existingShards[0].topologyTime;
|
|
assert.gt(topologyTime2, topologyTime1);
|
|
|
|
assert.commandFailed(s.s0.adminCommand({removeshard: s.shard1.shardName}));
|
|
|
|
// Should create a shard0002 shard
|
|
const rs = new ReplSetTest({nodes: 1});
|
|
rs.startSet({shardsvr: ""});
|
|
rs.initiate();
|
|
assert.commandWorked(s.s0.adminCommand({addshard: rs.getURL()}));
|
|
assert.eq(2, s.config.shards.count(), "new server does not appear in count");
|
|
|
|
rs.stopSet();
|
|
s.stop();
|