Bump versions to 4.0.0-SNAPSHOT
[controller.git] / opendaylight / md-sal / sal-distributed-datastore / src / main / java / org / opendaylight / controller / cluster / sharding / ShardedDataTreeActor.java
index e8f3f708607322706e617c5d1bd28426f058c9af..52c3d25faa18ef1de438d06e0c468f05998bf1a0 100644 (file)
@@ -5,10 +5,9 @@
  * terms of the Eclipse Public License v1.0 which accompanies this distribution,
  * and is available at http://www.eclipse.org/legal/epl-v10.html
  */
-
 package org.opendaylight.controller.cluster.sharding;
 
-import static akka.actor.ActorRef.noSender;
+import static java.util.Objects.requireNonNull;
 
 import akka.actor.ActorRef;
 import akka.actor.ActorSelection;
@@ -16,9 +15,7 @@ import akka.actor.ActorSystem;
 import akka.actor.PoisonPill;
 import akka.actor.Props;
 import akka.actor.Status;
-import akka.actor.Status.Failure;
 import akka.actor.Status.Success;
-import akka.cluster.Cluster;
 import akka.cluster.ClusterEvent;
 import akka.cluster.ClusterEvent.MemberExited;
 import akka.cluster.ClusterEvent.MemberRemoved;
@@ -27,46 +24,36 @@ import akka.cluster.ClusterEvent.MemberWeaklyUp;
 import akka.cluster.ClusterEvent.ReachableMember;
 import akka.cluster.ClusterEvent.UnreachableMember;
 import akka.cluster.Member;
-import akka.cluster.ddata.DistributedData;
-import akka.cluster.ddata.ORMap;
-import akka.cluster.ddata.Replicator;
-import akka.cluster.ddata.Replicator.Changed;
-import akka.cluster.ddata.Replicator.Subscribe;
-import akka.cluster.ddata.Replicator.Update;
 import akka.dispatch.OnComplete;
 import akka.pattern.Patterns;
 import akka.util.Timeout;
-import com.google.common.base.Optional;
-import com.google.common.base.Preconditions;
-import com.google.common.collect.Sets;
-import com.google.common.collect.Sets.SetView;
 import java.util.ArrayList;
 import java.util.Collection;
+import java.util.Collections;
 import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
+import java.util.Optional;
 import java.util.concurrent.CompletableFuture;
 import java.util.concurrent.TimeUnit;
-import java.util.function.Function;
-import java.util.stream.Collectors;
-import javax.annotation.Nullable;
 import org.opendaylight.controller.cluster.access.concepts.MemberName;
 import org.opendaylight.controller.cluster.common.actor.AbstractUntypedPersistentActor;
 import org.opendaylight.controller.cluster.datastore.ClusterWrapper;
-import org.opendaylight.controller.cluster.datastore.DistributedDataStore;
+import org.opendaylight.controller.cluster.datastore.DistributedDataStoreInterface;
 import org.opendaylight.controller.cluster.datastore.config.PrefixShardConfiguration;
-import org.opendaylight.controller.cluster.datastore.utils.ActorContext;
+import org.opendaylight.controller.cluster.datastore.utils.ActorUtils;
 import org.opendaylight.controller.cluster.datastore.utils.ClusterUtils;
 import org.opendaylight.controller.cluster.raft.client.messages.FindLeader;
 import org.opendaylight.controller.cluster.raft.client.messages.FindLeaderReply;
-import org.opendaylight.controller.cluster.sharding.messages.CreatePrefixShard;
+import org.opendaylight.controller.cluster.sharding.messages.LookupPrefixShard;
 import org.opendaylight.controller.cluster.sharding.messages.NotifyProducerCreated;
 import org.opendaylight.controller.cluster.sharding.messages.NotifyProducerRemoved;
 import org.opendaylight.controller.cluster.sharding.messages.PrefixShardCreated;
+import org.opendaylight.controller.cluster.sharding.messages.PrefixShardRemovalLookup;
 import org.opendaylight.controller.cluster.sharding.messages.PrefixShardRemoved;
 import org.opendaylight.controller.cluster.sharding.messages.ProducerCreated;
 import org.opendaylight.controller.cluster.sharding.messages.ProducerRemoved;
-import org.opendaylight.controller.cluster.sharding.messages.RemovePrefixShard;
+import org.opendaylight.controller.cluster.sharding.messages.StartConfigShardLookup;
 import org.opendaylight.mdsal.common.api.LogicalDatastoreType;
 import org.opendaylight.mdsal.dom.api.DOMDataTreeIdentifier;
 import org.opendaylight.mdsal.dom.api.DOMDataTreeProducer;
@@ -87,6 +74,7 @@ import scala.concurrent.duration.FiniteDuration;
  * Actor that tracks currently open producers/shards on remote nodes and handles notifications of remote
  * nodes of newly open producers/shards on the local node.
  */
+@Deprecated(forRemoval = true)
 public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
 
     private static final Logger LOG = LoggerFactory.getLogger(ShardedDataTreeActor.class);
@@ -95,26 +83,19 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
     private static final Timeout DEFAULT_ASK_TIMEOUT = new Timeout(15, TimeUnit.SECONDS);
 
     static final FiniteDuration SHARD_LOOKUP_TASK_INTERVAL = new FiniteDuration(1L, TimeUnit.SECONDS);
-    static final int LOOKUP_TASK_MAX_RETRIES = 100;
 
     private final DistributedShardedDOMDataTree shardingService;
     private final ActorSystem actorSystem;
     private final ClusterWrapper clusterWrapper;
     // helper actorContext used only for static calls to executeAsync etc
     // for calls that need specific actor context tied to a datastore use the one provided in the DistributedDataStore
-    private final ActorContext actorContext;
+    private final ActorUtils actorUtils;
     private final ShardingServiceAddressResolver resolver;
-    private final DistributedDataStore distributedConfigDatastore;
-    private final DistributedDataStore distributedOperDatastore;
+    private final DistributedDataStoreInterface distributedConfigDatastore;
+    private final DistributedDataStoreInterface distributedOperDatastore;
+    private final int lookupTaskMaxRetries;
 
     private final Map<DOMDataTreeIdentifier, ActorProducerRegistration> idToProducer = new HashMap<>();
-    private final Map<DOMDataTreeIdentifier, ShardFrontendRegistration> idToShardRegistration = new HashMap<>();
-
-    private final Cluster cluster;
-    private final ActorRef replicator;
-
-    private ORMap<PrefixShardConfiguration> currentData = ORMap.create();
-    private Map<DOMDataTreeIdentifier, PrefixShardConfiguration> currentConfiguration = new HashMap<>();
 
     ShardedDataTreeActor(final ShardedDataTreeActorCreator builder) {
         LOG.debug("Creating ShardedDataTreeActor on {}", builder.getClusterWrapper().getCurrentMemberName());
@@ -124,31 +105,26 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         clusterWrapper = builder.getClusterWrapper();
         distributedConfigDatastore = builder.getDistributedConfigDatastore();
         distributedOperDatastore = builder.getDistributedOperDatastore();
-        actorContext = distributedConfigDatastore.getActorContext();
+        lookupTaskMaxRetries = builder.getLookupTaskMaxRetries();
+        actorUtils = distributedConfigDatastore.getActorUtils();
         resolver = new ShardingServiceAddressResolver(
                 DistributedShardedDOMDataTree.ACTOR_ID, clusterWrapper.getCurrentMemberName());
 
         clusterWrapper.subscribeToMemberEvents(self());
-        cluster = Cluster.get(actorSystem);
-
-        replicator = DistributedData.get(context().system()).replicator();
     }
 
     @Override
     public void preStart() {
-        final Subscribe<ORMap<PrefixShardConfiguration>> subscribe =
-                new Subscribe<>(ClusterUtils.CONFIGURATION_KEY, self());
-        replicator.tell(subscribe, noSender());
     }
 
     @Override
-    protected void handleRecover(final Object message) throws Exception {
+    protected void handleRecover(final Object message) {
         LOG.debug("Received a recover message {}", message);
     }
 
     @Override
-    protected void handleCommand(final Object message) throws Exception {
-        LOG.debug("Received {}", message);
+    protected void handleCommand(final Object message) {
+        LOG.debug("{} : Received {}", clusterWrapper.getCurrentMemberName(), message);
         if (message instanceof ClusterEvent.MemberUp) {
             memberUp((ClusterEvent.MemberUp) message);
         } else if (message instanceof ClusterEvent.MemberWeaklyUp) {
@@ -161,8 +137,6 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
             memberUnreachable((ClusterEvent.UnreachableMember) message);
         } else if (message instanceof ClusterEvent.ReachableMember) {
             memberReachable((ClusterEvent.ReachableMember) message);
-        } else if (message instanceof Changed) {
-            onConfigChanged((Changed) message);
         } else if (message instanceof ProducerCreated) {
             onProducerCreated((ProducerCreated) message);
         } else if (message instanceof NotifyProducerCreated) {
@@ -173,51 +147,17 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
             onNotifyProducerRemoved((NotifyProducerRemoved) message);
         } else if (message instanceof PrefixShardCreated) {
             onPrefixShardCreated((PrefixShardCreated) message);
-        } else if (message instanceof CreatePrefixShard) {
-            onCreatePrefixShard((CreatePrefixShard) message);
-        } else if (message instanceof RemovePrefixShard) {
-            onRemovePrefixShard((RemovePrefixShard) message);
+        } else if (message instanceof LookupPrefixShard) {
+            onLookupPrefixShard((LookupPrefixShard) message);
+        } else if (message instanceof PrefixShardRemovalLookup) {
+            onPrefixShardRemovalLookup((PrefixShardRemovalLookup) message);
         } else if (message instanceof PrefixShardRemoved) {
             onPrefixShardRemoved((PrefixShardRemoved) message);
+        } else if (message instanceof StartConfigShardLookup) {
+            onStartConfigShardLookup((StartConfigShardLookup) message);
         }
     }
 
-    private void onConfigChanged(final Changed<ORMap<PrefixShardConfiguration>> change) {
-        LOG.debug("member : {}, Received configuration changed: {}", clusterWrapper.getCurrentMemberName(), change);
-
-        currentData = change.dataValue();
-        final Map<String, PrefixShardConfiguration> changedConfig = change.dataValue().getEntries();
-
-        LOG.debug("Changed set {}", changedConfig);
-
-        try {
-            final Map<DOMDataTreeIdentifier, PrefixShardConfiguration> newConfig =
-                    changedConfig.values().stream().collect(
-                            Collectors.toMap(PrefixShardConfiguration::getPrefix, Function.identity()));
-            resolveConfig(newConfig);
-        } catch (final IllegalStateException e) {
-            LOG.error("Failed, ", e);
-        }
-
-    }
-
-    private void resolveConfig(final Map<DOMDataTreeIdentifier, PrefixShardConfiguration> newConfig) {
-
-        // get the removed configurations
-        final SetView<DOMDataTreeIdentifier> deleted =
-                Sets.difference(currentConfiguration.keySet(), newConfig.keySet());
-        shardingService.resolveShardRemovals(deleted);
-
-        // get the added configurations
-        final SetView<DOMDataTreeIdentifier> additions =
-                Sets.difference(newConfig.keySet(), currentConfiguration.keySet());
-        shardingService.resolveShardAdditions(additions);
-        // we can ignore those that existed previously since the potential changes in replicas will be handled by
-        // shard manager.
-
-        currentConfiguration = new HashMap<>(newConfig);
-    }
-
     @Override
     public String persistenceId() {
         return PERSISTENCE_ID;
@@ -278,7 +218,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
 
         // fastpath if we have no peers
         if (resolver.getShardingServicePeerActorAddresses().isEmpty()) {
-            getSender().tell(new Status.Success(null), noSender());
+            getSender().tell(new Status.Success(null), ActorRef.noSender());
         }
 
         final ActorRef sender = getSender();
@@ -290,7 +230,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
             final ActorSelection actorSelection = actorSystem.actorSelection(address);
             futures.add(
                     FutureConverters.toJava(
-                            actorContext.executeOperationAsync(
+                            actorUtils.executeOperationAsync(
                                     actorSelection, new NotifyProducerCreated(subtrees), DEFAULT_ASK_TIMEOUT))
                     .toCompletableFuture());
         }
@@ -298,12 +238,12 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         final CompletableFuture<Void> combinedFuture = CompletableFuture.allOf(
                 futures.toArray(new CompletableFuture[futures.size()]));
 
-        combinedFuture.thenRun(() -> {
-            sender.tell(new Status.Success(null), noSender());
-        }).exceptionally(throwable -> {
-            sender.tell(new Status.Failure(throwable), self());
-            return null;
-        });
+        combinedFuture
+                .thenRun(() -> sender.tell(new Success(null), ActorRef.noSender()))
+                .exceptionally(throwable -> {
+                    sender.tell(new Status.Failure(throwable), self());
+                    return null;
+                });
     }
 
     private void onNotifyProducerCreated(final NotifyProducerCreated message) {
@@ -330,7 +270,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
             final ActorSelection selection = actorSystem.actorSelection(address);
 
             futures.add(FutureConverters.toJava(
-                    actorContext.executeOperationAsync(selection, new NotifyProducerRemoved(message.getSubtrees())))
+                    actorUtils.executeOperationAsync(selection, new NotifyProducerRemoved(message.getSubtrees())))
                     .toCompletableFuture());
         }
 
@@ -354,79 +294,48 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         final ActorProducerRegistration registration = idToProducer.remove(message.getSubtrees().iterator().next());
         if (registration == null) {
             LOG.warn("The notification contained a path on which no producer is registered, throwing away");
-            getSender().tell(new Status.Success(null), noSender());
+            getSender().tell(new Status.Success(null), ActorRef.noSender());
             return;
         }
 
         try {
             registration.close();
-            getSender().tell(new Status.Success(null), noSender());
+            getSender().tell(new Status.Success(null), ActorRef.noSender());
         } catch (final DOMDataTreeProducerException e) {
             LOG.error("Unable to close producer", e);
-            getSender().tell(new Status.Failure(e), noSender());
+            getSender().tell(new Status.Failure(e), ActorRef.noSender());
         }
     }
 
     @SuppressWarnings("checkstyle:IllegalCatch")
-    private void onCreatePrefixShard(final CreatePrefixShard message) {
-        LOG.debug("Member: {}, Received CreatePrefixShard: {}", clusterWrapper.getCurrentMemberName(), message);
+    private void onLookupPrefixShard(final LookupPrefixShard message) {
+        LOG.debug("Member: {}, Received LookupPrefixShard: {}", clusterWrapper.getCurrentMemberName(), message);
 
-        final PrefixShardConfiguration configuration = message.getConfiguration();
+        final DOMDataTreeIdentifier prefix = message.getPrefix();
 
-        final Update<ORMap<PrefixShardConfiguration>> update =
-                new Update<>(ClusterUtils.CONFIGURATION_KEY, currentData, Replicator.writeLocal(),
-                    map -> map.put(cluster, configuration.toDataMapKey(), configuration));
-
-        replicator.tell(update, self());
-
-        final ActorContext context =
-                configuration.getPrefix().getDatastoreType() == LogicalDatastoreType.CONFIGURATION
-                        ? distributedConfigDatastore.getActorContext() : distributedOperDatastore.getActorContext();
+        final ActorUtils utils = prefix.getDatastoreType() == LogicalDatastoreType.CONFIGURATION
+                        ? distributedConfigDatastore.getActorUtils() : distributedOperDatastore.getActorUtils();
 
         // schedule a notification task for the reply
         actorSystem.scheduler().scheduleOnce(SHARD_LOOKUP_TASK_INTERVAL,
                 new ShardCreationLookupTask(actorSystem, getSender(), clusterWrapper,
-                        context, shardingService, configuration.getPrefix()),
-                actorSystem.dispatcher());
+                        utils, shardingService, prefix, lookupTaskMaxRetries), actorSystem.dispatcher());
     }
 
     private void onPrefixShardCreated(final PrefixShardCreated message) {
         LOG.debug("Member: {}, Received PrefixShardCreated: {}", clusterWrapper.getCurrentMemberName(), message);
 
-        final Collection<String> addresses = resolver.getShardingServicePeerActorAddresses();
-        final ActorRef sender = getSender();
+        final PrefixShardConfiguration config = message.getConfiguration();
 
-        final List<CompletableFuture<Object>> futures = new ArrayList<>();
-
-        for (final String address : addresses) {
-            final ActorSelection actorSelection = actorSystem.actorSelection(address);
-            futures.add(FutureConverters.toJava(actorContext.executeOperationAsync(actorSelection,
-                    new CreatePrefixShard(message.getConfiguration()))).toCompletableFuture());
-        }
-
-        final CompletableFuture<Void> combinedFuture =
-                CompletableFuture.allOf(futures.toArray(new CompletableFuture[futures.size()]));
-
-        combinedFuture.thenRun(() -> {
-            sender.tell(new Status.Success(null), self());
-        }).exceptionally(throwable -> {
-            sender.tell(new Status.Failure(throwable), self());
-            return null;
-        });
+        shardingService.resolveShardAdditions(Collections.singleton(config.getPrefix()));
     }
 
-    private void onRemovePrefixShard(final RemovePrefixShard message) {
-        LOG.debug("Member: {}, Received RemovePrefixShard: {}", clusterWrapper.getCurrentMemberName(), message);
-
-        //TODO the removal message should have the configuration or some other way to get to the key
-        final Update<ORMap<PrefixShardConfiguration>> removal =
-                new Update<>(ClusterUtils.CONFIGURATION_KEY, currentData, Replicator.writeLocal(),
-                    map -> map.remove(cluster, "prefix=" + message.getPrefix()));
-        replicator.tell(removal, self());
+    private void onPrefixShardRemovalLookup(final PrefixShardRemovalLookup message) {
+        LOG.debug("Member: {}, Received PrefixShardRemovalLookup: {}", clusterWrapper.getCurrentMemberName(), message);
 
         final ShardRemovalLookupTask removalTask =
                 new ShardRemovalLookupTask(actorSystem, getSender(),
-                        actorContext, message.getPrefix());
+                        actorUtils, message.getPrefix(), lookupTaskMaxRetries);
 
         actorSystem.scheduler().scheduleOnce(SHARD_LOOKUP_TASK_INTERVAL, removalTask, actorSystem.dispatcher());
     }
@@ -434,15 +343,21 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
     private void onPrefixShardRemoved(final PrefixShardRemoved message) {
         LOG.debug("Received PrefixShardRemoved: {}", message);
 
-        final ShardFrontendRegistration registration = idToShardRegistration.get(message.getPrefix());
+        shardingService.resolveShardRemovals(Collections.singleton(message.getPrefix()));
+    }
 
-        if (registration == null) {
-            LOG.warn("Received shard removed for {}, but not shard registered at this prefix all registrations: {}",
-                    message.getPrefix(), idToShardRegistration);
-            return;
-        }
+    private void onStartConfigShardLookup(final StartConfigShardLookup message) {
+        LOG.debug("Received StartConfigShardLookup: {}", message);
 
-        registration.close();
+        final ActorUtils context =
+                message.getType().equals(LogicalDatastoreType.CONFIGURATION)
+                        ? distributedConfigDatastore.getActorUtils() : distributedOperDatastore.getActorUtils();
+
+        // schedule a notification task for the reply
+        actorSystem.scheduler().scheduleOnce(SHARD_LOOKUP_TASK_INTERVAL,
+                new ConfigShardLookupTask(
+                        actorSystem, getSender(), context, message, lookupTaskMaxRetries),
+                actorSystem.dispatcher());
     }
 
     private static MemberName memberToName(final Member member) {
@@ -486,39 +401,6 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         }
     }
 
-    private abstract static class LookupTask implements Runnable {
-
-        private final ActorRef replyTo;
-        private int retries = 0;
-
-        private LookupTask(final ActorRef replyTo) {
-            this.replyTo = replyTo;
-        }
-
-        abstract void reschedule(int retries);
-
-        void tryReschedule(@Nullable final Throwable throwable) {
-            if (retries <= LOOKUP_TASK_MAX_RETRIES) {
-                retries++;
-                reschedule(retries);
-            } else {
-                fail(throwable);
-            }
-        }
-
-        void fail(@Nullable final Throwable throwable) {
-            if (throwable == null) {
-                replyTo.tell(new Failure(
-                        new DOMDataTreeShardCreationFailedException("Unable to find the backend shard."
-                                + "Failing..")), noSender());
-            } else {
-                replyTo.tell(new Failure(
-                        new DOMDataTreeShardCreationFailedException("Unable to find the backend shard."
-                                + "Failing..", throwable)), noSender());
-            }
-        }
-    }
-
     /**
      * Handles the lookup step of cds shard creation once the configuration is updated.
      */
@@ -527,23 +409,26 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         private final ActorSystem system;
         private final ActorRef replyTo;
         private final ClusterWrapper clusterWrapper;
-        private final ActorContext context;
+        private final ActorUtils context;
         private final DistributedShardedDOMDataTree shardingService;
         private final DOMDataTreeIdentifier toLookup;
+        private final int lookupMaxRetries;
 
         ShardCreationLookupTask(final ActorSystem system,
                                 final ActorRef replyTo,
                                 final ClusterWrapper clusterWrapper,
-                                final ActorContext context,
+                                final ActorUtils context,
                                 final DistributedShardedDOMDataTree shardingService,
-                                final DOMDataTreeIdentifier toLookup) {
-            super(replyTo);
+                                final DOMDataTreeIdentifier toLookup,
+                                final int lookupMaxRetries) {
+            super(replyTo, lookupMaxRetries);
             this.system = system;
             this.replyTo = replyTo;
             this.clusterWrapper = clusterWrapper;
             this.context = context;
             this.shardingService = shardingService;
             this.toLookup = toLookup;
+            this.lookupMaxRetries = lookupMaxRetries;
         }
 
         @Override
@@ -553,7 +438,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
 
             localShardFuture.onComplete(new OnComplete<ActorRef>() {
                 @Override
-                public void onComplete(Throwable throwable, ActorRef actorRef) throws Throwable {
+                public void onComplete(final Throwable throwable, final ActorRef actorRef) {
                     if (throwable != null) {
                         tryReschedule(throwable);
                     } else {
@@ -562,7 +447,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
                         system.scheduler().scheduleOnce(
                                 SHARD_LOOKUP_TASK_INTERVAL,
                                 new ShardLeaderLookupTask(system, replyTo, context, clusterWrapper, actorRef,
-                                        shardingService, toLookup),
+                                        shardingService, toLookup, lookupMaxRetries),
                                 system.dispatcher());
                     }
                 }
@@ -570,7 +455,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         }
 
         @Override
-        void reschedule(int retries) {
+        void reschedule(final int retries) {
             LOG.debug("Local backend for shard[{}] not found, try: {}, rescheduling..", toLookup, retries);
             system.scheduler().scheduleOnce(
                     SHARD_LOOKUP_TASK_INTERVAL, ShardCreationLookupTask.this, system.dispatcher());
@@ -584,20 +469,22 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
 
         private final ActorSystem system;
         private final ActorRef replyTo;
-        private final ActorContext context;
+        private final ActorUtils context;
         private final ClusterWrapper clusterWrapper;
         private final ActorRef shard;
         private final DistributedShardedDOMDataTree shardingService;
         private final DOMDataTreeIdentifier toLookup;
+        private final int lookupMaxRetries;
 
         ShardLeaderLookupTask(final ActorSystem system,
                               final ActorRef replyTo,
-                              final ActorContext context,
+                              final ActorUtils context,
                               final ClusterWrapper clusterWrapper,
                               final ActorRef shard,
                               final DistributedShardedDOMDataTree shardingService,
-                              final DOMDataTreeIdentifier toLookup) {
-            super(replyTo);
+                              final DOMDataTreeIdentifier toLookup,
+                              final int lookupMaxRetries) {
+            super(replyTo, lookupMaxRetries);
             this.system = system;
             this.replyTo = replyTo;
             this.context = context;
@@ -605,6 +492,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
             this.shard = shard;
             this.shardingService = shardingService;
             this.toLookup = toLookup;
+            this.lookupMaxRetries = lookupMaxRetries;
         }
 
         @Override
@@ -612,21 +500,22 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
 
             final Future<Object> ask = Patterns.ask(shard, FindLeader.INSTANCE, context.getOperationTimeout());
 
-            ask.onComplete(new OnComplete<Object>() {
+            ask.onComplete(new OnComplete<>() {
                 @Override
-                public void onComplete(final Throwable throwable, final Object findLeaderReply) throws Throwable {
+                public void onComplete(final Throwable throwable, final Object findLeaderReply) {
                     if (throwable != null) {
                         tryReschedule(throwable);
                     } else {
                         final FindLeaderReply findLeader = (FindLeaderReply) findLeaderReply;
-                        final java.util.Optional<String> leaderActor = findLeader.getLeaderActor();
+                        final Optional<String> leaderActor = findLeader.getLeaderActor();
                         if (leaderActor.isPresent()) {
                             // leader is found, backend seems ready, check if the frontend is ready
                             LOG.debug("{} - Leader for shard[{}] backend ready, starting frontend lookup..",
                                     clusterWrapper.getCurrentMemberName(), toLookup);
                             system.scheduler().scheduleOnce(
                                     SHARD_LOOKUP_TASK_INTERVAL,
-                                    new FrontendLookupTask(system, replyTo, shardingService, toLookup),
+                                    new FrontendLookupTask(
+                                            system, replyTo, shardingService, toLookup, lookupMaxRetries),
                                     system.dispatcher());
                         } else {
                             tryReschedule(null);
@@ -638,7 +527,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         }
 
         @Override
-        void reschedule(int retries) {
+        void reschedule(final int retries) {
             LOG.debug("{} - Leader for shard[{}] backend not found on try: {}, retrying..",
                     clusterWrapper.getCurrentMemberName(), toLookup, retries);
             system.scheduler().scheduleOnce(
@@ -661,8 +550,9 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         FrontendLookupTask(final ActorSystem system,
                            final ActorRef replyTo,
                            final DistributedShardedDOMDataTree shardingService,
-                           final DOMDataTreeIdentifier toLookup) {
-            super(replyTo);
+                           final DOMDataTreeIdentifier toLookup,
+                           final int lookupMaxRetries) {
+            super(replyTo, lookupMaxRetries);
             this.system = system;
             this.replyTo = replyTo;
             this.shardingService = shardingService;
@@ -675,7 +565,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
                     shardingService.lookupShardFrontend(toLookup);
 
             if (entry != null && tableEntryIdCheck(entry, toLookup) && entry.getValue() != null) {
-                replyTo.tell(new Success(null), noSender());
+                replyTo.tell(new Success(null), ActorRef.noSender());
             } else {
                 tryReschedule(null);
             }
@@ -687,7 +577,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
                 return false;
             }
 
-            if (YangInstanceIdentifier.EMPTY.equals(prefix.getRootIdentifier())) {
+            if (YangInstanceIdentifier.empty().equals(prefix.getRootIdentifier())) {
                 return true;
             }
 
@@ -699,7 +589,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         }
 
         @Override
-        void reschedule(int retries) {
+        void reschedule(final int retries) {
             LOG.debug("Frontend for shard[{}] not found on try: {}, retrying..", toLookup, retries);
             system.scheduler().scheduleOnce(
                     SHARD_LOOKUP_TASK_INTERVAL, FrontendLookupTask.this, system.dispatcher());
@@ -714,14 +604,15 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
 
         private final ActorSystem system;
         private final ActorRef replyTo;
-        private final ActorContext context;
+        private final ActorUtils context;
         private final DOMDataTreeIdentifier toLookup;
 
         ShardRemovalLookupTask(final ActorSystem system,
                                final ActorRef replyTo,
-                               final ActorContext context,
-                               final DOMDataTreeIdentifier toLookup) {
-            super(replyTo);
+                               final ActorUtils context,
+                               final DOMDataTreeIdentifier toLookup,
+                               final int lookupMaxRetries) {
+            super(replyTo, lookupMaxRetries);
             this.system = system;
             this.replyTo = replyTo;
             this.context = context;
@@ -735,12 +626,12 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
 
             localShardFuture.onComplete(new OnComplete<ActorRef>() {
                 @Override
-                public void onComplete(Throwable throwable, ActorRef actorRef) throws Throwable {
+                public void onComplete(final Throwable throwable, final ActorRef actorRef) {
                     if (throwable != null) {
                         //TODO Shouldn't we check why findLocalShard failed?
                         LOG.debug("Backend shard[{}] removal lookup successful notifying the registration future",
                                 toLookup);
-                        replyTo.tell(new Success(null), noSender());
+                        replyTo.tell(new Success(null), ActorRef.noSender());
                     } else {
                         tryReschedule(null);
                     }
@@ -749,7 +640,7 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         }
 
         @Override
-        void reschedule(int retries) {
+        void reschedule(final int retries) {
             LOG.debug("Backend shard[{}] removal lookup failed, shard is still present, try: {}, rescheduling..",
                     toLookup, retries);
             system.scheduler().scheduleOnce(
@@ -757,13 +648,114 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
         }
     }
 
+    /**
+     * Task for handling the lookup of the backend for the configuration shard.
+     */
+    private static class ConfigShardLookupTask extends LookupTask {
+
+        private final ActorSystem system;
+        private final ActorRef replyTo;
+        private final ActorUtils context;
+
+        ConfigShardLookupTask(final ActorSystem system,
+                              final ActorRef replyTo,
+                              final ActorUtils context,
+                              final StartConfigShardLookup message,
+                              final int lookupMaxRetries) {
+            super(replyTo, lookupMaxRetries);
+            this.system = system;
+            this.replyTo = replyTo;
+            this.context = context;
+        }
+
+        @Override
+        void reschedule(final int retries) {
+            LOG.debug("Local backend for prefix configuration shard not found, try: {}, rescheduling..", retries);
+            system.scheduler().scheduleOnce(
+                    SHARD_LOOKUP_TASK_INTERVAL, ConfigShardLookupTask.this, system.dispatcher());
+        }
+
+        @Override
+        public void run() {
+            final Optional<ActorRef> localShard =
+                    context.findLocalShard(ClusterUtils.PREFIX_CONFIG_SHARD_ID);
+
+            if (!localShard.isPresent()) {
+                tryReschedule(null);
+            } else {
+                LOG.debug("Local backend for prefix configuration shard lookup successful");
+                replyTo.tell(new Status.Success(null), ActorRef.noSender());
+            }
+        }
+    }
+
+    /**
+     * Task for handling the readiness state of the config shard. Reports success once the leader is elected.
+     */
+    private static class ConfigShardReadinessTask extends LookupTask {
+
+        private final ActorSystem system;
+        private final ActorRef replyTo;
+        private final ActorUtils context;
+        private final ClusterWrapper clusterWrapper;
+        private final ActorRef shard;
+
+        ConfigShardReadinessTask(final ActorSystem system,
+                                 final ActorRef replyTo,
+                                 final ActorUtils context,
+                                 final ClusterWrapper clusterWrapper,
+                                 final ActorRef shard,
+                                 final int lookupMaxRetries) {
+            super(replyTo, lookupMaxRetries);
+            this.system = system;
+            this.replyTo = replyTo;
+            this.context = context;
+            this.clusterWrapper = clusterWrapper;
+            this.shard = shard;
+        }
+
+        @Override
+        void reschedule(final int retries) {
+            LOG.debug("{} - Leader for config shard not found on try: {}, retrying..",
+                    clusterWrapper.getCurrentMemberName(), retries);
+            system.scheduler().scheduleOnce(
+                    SHARD_LOOKUP_TASK_INTERVAL, ConfigShardReadinessTask.this, system.dispatcher());
+        }
+
+        @Override
+        public void run() {
+            final Future<Object> ask = Patterns.ask(shard, FindLeader.INSTANCE, context.getOperationTimeout());
+
+            ask.onComplete(new OnComplete<>() {
+                @Override
+                public void onComplete(final Throwable throwable, final Object findLeaderReply) {
+                    if (throwable != null) {
+                        tryReschedule(throwable);
+                    } else {
+                        final FindLeaderReply findLeader = (FindLeaderReply) findLeaderReply;
+                        final Optional<String> leaderActor = findLeader.getLeaderActor();
+                        if (leaderActor.isPresent()) {
+                            // leader is found, backend seems ready, check if the frontend is ready
+                            LOG.debug("{} - Leader for config shard is ready. Ending lookup.",
+                                    clusterWrapper.getCurrentMemberName());
+                            replyTo.tell(new Status.Success(null), ActorRef.noSender());
+                        } else {
+                            tryReschedule(null);
+                        }
+                    }
+                }
+            }, system.dispatcher());
+        }
+    }
+
     public static class ShardedDataTreeActorCreator {
 
         private DistributedShardedDOMDataTree shardingService;
-        private DistributedDataStore distributedConfigDatastore;
-        private DistributedDataStore distributedOperDatastore;
+        private DistributedDataStoreInterface distributedConfigDatastore;
+        private DistributedDataStoreInterface distributedOperDatastore;
         private ActorSystem actorSystem;
         private ClusterWrapper cluster;
+        private int maxRetries;
 
         public DistributedShardedDOMDataTree getShardingService() {
             return shardingService;
@@ -783,8 +775,8 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
             return this;
         }
 
-        public ShardedDataTreeActorCreator setClusterWrapper(final ClusterWrapper cluster) {
-            this.cluster = cluster;
+        public ShardedDataTreeActorCreator setClusterWrapper(final ClusterWrapper clusterWrapper) {
+            this.cluster = clusterWrapper;
             return this;
         }
 
@@ -792,32 +784,41 @@ public class ShardedDataTreeActor extends AbstractUntypedPersistentActor {
             return cluster;
         }
 
-        public DistributedDataStore getDistributedConfigDatastore() {
+        public DistributedDataStoreInterface getDistributedConfigDatastore() {
             return distributedConfigDatastore;
         }
 
         public ShardedDataTreeActorCreator setDistributedConfigDatastore(
-                final DistributedDataStore distributedConfigDatastore) {
+                final DistributedDataStoreInterface distributedConfigDatastore) {
             this.distributedConfigDatastore = distributedConfigDatastore;
             return this;
         }
 
-        public DistributedDataStore getDistributedOperDatastore() {
+        public DistributedDataStoreInterface getDistributedOperDatastore() {
             return distributedOperDatastore;
         }
 
         public ShardedDataTreeActorCreator setDistributedOperDatastore(
-                final DistributedDataStore distributedOperDatastore) {
+                final DistributedDataStoreInterface distributedOperDatastore) {
             this.distributedOperDatastore = distributedOperDatastore;
             return this;
         }
 
+        public ShardedDataTreeActorCreator setLookupTaskMaxRetries(final int newMaxRetries) {
+            this.maxRetries = newMaxRetries;
+            return this;
+        }
+
+        public int getLookupTaskMaxRetries() {
+            return maxRetries;
+        }
+
         private void verify() {
-            Preconditions.checkNotNull(shardingService);
-            Preconditions.checkNotNull(actorSystem);
-            Preconditions.checkNotNull(cluster);
-            Preconditions.checkNotNull(distributedConfigDatastore);
-            Preconditions.checkNotNull(distributedOperDatastore);
+            requireNonNull(shardingService);
+            requireNonNull(actorSystem);
+            requireNonNull(cluster);
+            requireNonNull(distributedConfigDatastore);
+            requireNonNull(distributedOperDatastore);
         }
 
         public Props props() {