shardElectionTimeoutFactor(1).customRaftPolicyImplementation(null));
JavaTestKit.shutdownActorSystem(leaderSystem, null, true);
+ Cluster.get(followerSystem).leave(MEMBER_1_ADDRESS);
followerTestKit.waitUntilNoLeader(followerDistributedDataStore.getActorContext(), CARS);
}
});
+ MemberNode.verifyRaftState(followerDistributedDataStore, "people", new RaftStateVerifier() {
+ @Override
+ public void verify(OnDemandRaftState raftState) {
+ assertEquals("getLastApplied", 0, raftState.getLastApplied());
+ }
+ });
+
// Prepare, ready and canCommit a WO tx that writes to 2 shards. This will become the current tx in
// the leader shard.
customRaftPolicyImplementation(DisableElectionsRaftPolicy.class.getName()).
shardElectionTimeoutFactor(10));
+ Cluster.get(followerSystem).leave(MEMBER_1_ADDRESS);
leaderTestKit.waitUntilNoLeader(leaderDistributedDataStore.getActorContext(), "cars");
// Submit all tx's - the messages should get queued for retry.
JavaTestKit.shutdownActorSystem(leaderSystem, null, true);
+ Cluster.get(followerSystem).leave(MEMBER_1_ADDRESS);
+
Uninterruptibles.sleepUninterruptibly(100, TimeUnit.MILLISECONDS);
sendDatastoreContextUpdate(followerDistributedDataStore, followerDatastoreContextBuilder.
JavaTestKit.shutdownActorSystem(leaderSystem, null, true);
+ Cluster.get(followerSystem).leave(MEMBER_1_ADDRESS);
+
sendDatastoreContextUpdate(followerDistributedDataStore, followerDatastoreContextBuilder.
operationTimeoutInMillis(500).shardElectionTimeoutFactor(1).customRaftPolicyImplementation(null));