X-Git-Url: https://git.opendaylight.org/gerrit/gitweb?a=blobdiff_plain;f=opendaylight%2Fmd-sal%2Fsal-distributed-datastore%2Fsrc%2Fmain%2Fjava%2Forg%2Fopendaylight%2Fcontroller%2Fcluster%2Fdatastore%2FDistributedDataStore.java;h=3029ef7e399a4db99c6ed2ad18a7e7701cb2f8ee;hb=09b8efdb40105cd4cd3c21c9a9aea2c6687972be;hp=107c959112fbc34398ae48b144752a8b28c99e41;hpb=1fbedd9f6b3ec25266c69ed0339a528fee533ca6;p=controller.git diff --git a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java index 107c959112..3029ef7e39 100644 --- a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java +++ b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java @@ -11,9 +11,13 @@ package org.opendaylight.controller.cluster.datastore; import akka.actor.ActorSystem; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; +import java.util.concurrent.CountDownLatch; +import java.util.concurrent.TimeUnit; import org.opendaylight.controller.cluster.datastore.identifiers.ShardManagerIdentifier; +import org.opendaylight.controller.cluster.datastore.jmx.mbeans.DatastoreConfigurationMXBeanImpl; import org.opendaylight.controller.cluster.datastore.shardstrategy.ShardStrategyFactory; import org.opendaylight.controller.cluster.datastore.utils.ActorContext; +import org.opendaylight.controller.cluster.datastore.utils.Dispatchers; import org.opendaylight.controller.md.sal.common.api.data.AsyncDataBroker; import org.opendaylight.controller.md.sal.common.api.data.AsyncDataChangeListener; import org.opendaylight.controller.sal.core.spi.data.DOMStore; @@ -32,12 +36,25 @@ import org.slf4j.LoggerFactory; /** * */ -public class DistributedDataStore implements DOMStore, SchemaContextListener, AutoCloseable { +public class DistributedDataStore implements DOMStore, SchemaContextListener, + DatastoreContextConfigAdminOverlay.Listener, AutoCloseable { private static final Logger LOG = LoggerFactory.getLogger(DistributedDataStore.class); - public static final int REGISTER_DATA_CHANGE_LISTENER_TIMEOUT_FACTOR = 24; // 24 times the usual operation timeout + private static final String UNKNOWN_TYPE = "unknown"; + + private static final long READY_WAIT_FACTOR = 3; private final ActorContext actorContext; + private final long waitTillReadyTimeInMillis; + + + private AutoCloseable closeable; + + private DatastoreConfigurationMXBeanImpl datastoreConfigMXBean; + + private CountDownLatch waitTillReadyCountDownLatch = new CountDownLatch(1); + + private final String type; public DistributedDataStore(ActorSystem actorSystem, ClusterWrapper cluster, Configuration configuration, DatastoreContext datastoreContext) { @@ -46,20 +63,39 @@ public class DistributedDataStore implements DOMStore, SchemaContextListener, Au Preconditions.checkNotNull(configuration, "configuration should not be null"); Preconditions.checkNotNull(datastoreContext, "datastoreContext should not be null"); - String type = datastoreContext.getDataStoreType(); + this.type = datastoreContext.getDataStoreType(); String shardManagerId = ShardManagerIdentifier.builder().type(type).build().toString(); LOG.info("Creating ShardManager : {}", shardManagerId); + String shardDispatcher = + new Dispatchers(actorSystem.dispatchers()).getDispatcherPath(Dispatchers.DispatcherType.Shard); + actorContext = new ActorContext(actorSystem, actorSystem.actorOf( - ShardManager.props(cluster, configuration, datastoreContext) - .withMailbox(ActorContext.MAILBOX), shardManagerId ), + ShardManager.props(cluster, configuration, datastoreContext, waitTillReadyCountDownLatch) + .withDispatcher(shardDispatcher).withMailbox(ActorContext.MAILBOX), shardManagerId ), cluster, configuration, datastoreContext); + + this.waitTillReadyTimeInMillis = + actorContext.getDatastoreContext().getShardLeaderElectionTimeout().duration().toMillis() * READY_WAIT_FACTOR; + + + datastoreConfigMXBean = new DatastoreConfigurationMXBeanImpl(datastoreContext.getDataStoreMXBeanType()); + datastoreConfigMXBean.setContext(datastoreContext); + datastoreConfigMXBean.registerMBean(); } public DistributedDataStore(ActorContext actorContext) { this.actorContext = Preconditions.checkNotNull(actorContext, "actorContext should not be null"); + this.type = UNKNOWN_TYPE; + this.waitTillReadyTimeInMillis = + actorContext.getDatastoreContext().getShardLeaderElectionTimeout().duration().toMillis() * READY_WAIT_FACTOR; + + } + + public void setCloseable(AutoCloseable closeable) { + this.closeable = closeable; } @SuppressWarnings("unchecked") @@ -111,7 +147,25 @@ public class DistributedDataStore implements DOMStore, SchemaContextListener, Au } @Override - public void close() throws Exception { + public void onDatastoreContextUpdated(DatastoreContext context) { + LOG.info("DatastoreContext updated for data store {}", actorContext.getDataStoreType()); + + actorContext.setDatastoreContext(context); + datastoreConfigMXBean.setContext(context); + } + + @Override + public void close() { + datastoreConfigMXBean.unregisterMBean(); + + if(closeable != null) { + try { + closeable.close(); + } catch (Exception e) { + LOG.debug("Error closing insance", e); + } + } + actorContext.shutdown(); } @@ -119,4 +173,21 @@ public class DistributedDataStore implements DOMStore, SchemaContextListener, Au ActorContext getActorContext() { return actorContext; } + + public void waitTillReady(){ + LOG.info("Beginning to wait for data store to become ready : {}", type); + + try { + waitTillReadyCountDownLatch.await(waitTillReadyTimeInMillis, TimeUnit.MILLISECONDS); + + LOG.debug("Data store {} is now ready", type); + } catch (InterruptedException e) { + LOG.error("Interrupted when trying to wait for shards to become leader in a reasonable amount of time - giving up"); + } + } + + @VisibleForTesting + public CountDownLatch getWaitTillReadyCountDownLatch() { + return waitTillReadyCountDownLatch; + } }