X-Git-Url: https://git.opendaylight.org/gerrit/gitweb?a=blobdiff_plain;f=opendaylight%2Fmd-sal%2Fsal-distributed-datastore%2Fsrc%2Fmain%2Fjava%2Forg%2Fopendaylight%2Fcontroller%2Fcluster%2Fdatastore%2FDistributedDataStore.java;h=c780881a2ffad1ed50695b7a38111068ec2f8e3f;hb=1a4a451d17d39b0dac45a70de9016bd4b0e6b249;hp=780f28f358ec327f5ec6ab9cace63f769695e55a;hpb=575cb53156868951bb1f11f0282c2e32d84e800a;p=controller.git diff --git a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java index 780f28f358..49f5388842 100644 --- a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java +++ b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/DistributedDataStore.java @@ -8,22 +8,28 @@ package org.opendaylight.controller.cluster.datastore; -import java.util.concurrent.Executors; - import akka.actor.ActorRef; import akka.actor.ActorSystem; - -import org.opendaylight.controller.cluster.datastore.messages.RegisterChangeListener; -import org.opendaylight.controller.cluster.datastore.messages.RegisterChangeListenerReply; -import org.opendaylight.controller.cluster.datastore.messages.UpdateSchemaContext; -import org.opendaylight.controller.cluster.datastore.shardstrategy.ShardStrategyFactory; +import com.google.common.annotations.VisibleForTesting; +import com.google.common.base.Preconditions; +import com.google.common.util.concurrent.Uninterruptibles; +import java.util.concurrent.CountDownLatch; +import java.util.concurrent.TimeUnit; +import org.opendaylight.controller.cluster.datastore.config.Configuration; +import org.opendaylight.controller.cluster.datastore.identifiers.ShardManagerIdentifier; +import org.opendaylight.controller.cluster.datastore.jmx.mbeans.DatastoreConfigurationMXBeanImpl; +import org.opendaylight.controller.cluster.datastore.jmx.mbeans.DatastoreInfoMXBeanImpl; import org.opendaylight.controller.cluster.datastore.utils.ActorContext; +import org.opendaylight.controller.cluster.datastore.utils.Dispatchers; +import org.opendaylight.controller.cluster.datastore.utils.PrimaryShardInfoFutureCache; import org.opendaylight.controller.md.sal.common.api.data.AsyncDataBroker; import org.opendaylight.controller.md.sal.common.api.data.AsyncDataChangeListener; +import org.opendaylight.controller.md.sal.dom.api.DOMDataTreeChangeListener; import org.opendaylight.controller.sal.core.spi.data.DOMStore; import org.opendaylight.controller.sal.core.spi.data.DOMStoreReadTransaction; import org.opendaylight.controller.sal.core.spi.data.DOMStoreReadWriteTransaction; import org.opendaylight.controller.sal.core.spi.data.DOMStoreTransactionChain; +import org.opendaylight.controller.sal.core.spi.data.DOMStoreTreeChangePublisher; import org.opendaylight.controller.sal.core.spi.data.DOMStoreWriteTransaction; import org.opendaylight.yangtools.concepts.ListenerRegistration; import org.opendaylight.yangtools.yang.data.api.YangInstanceIdentifier; @@ -33,114 +39,213 @@ import org.opendaylight.yangtools.yang.model.api.SchemaContextListener; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import com.google.common.util.concurrent.ListeningExecutorService; -import com.google.common.util.concurrent.MoreExecutors; - /** * */ -public class DistributedDataStore implements DOMStore, SchemaContextListener, AutoCloseable { +public class DistributedDataStore implements DOMStore, SchemaContextListener, + DatastoreContextConfigAdminOverlay.Listener, DOMStoreTreeChangePublisher, AutoCloseable { - private static final Logger - LOG = LoggerFactory.getLogger(DistributedDataStore.class); + private static final Logger LOG = LoggerFactory.getLogger(DistributedDataStore.class); + private static final String UNKNOWN_TYPE = "unknown"; - private static final int DEFAULT_EXECUTOR_POOL_SIZE = 10; + private static final long READY_WAIT_FACTOR = 3; - private final String type; private final ActorContext actorContext; + private final long waitTillReadyTimeInMillis; + + + private AutoCloseable closeable; + + private DatastoreConfigurationMXBeanImpl datastoreConfigMXBean; + + private DatastoreInfoMXBeanImpl datastoreInfoMXBean; + + private final CountDownLatch waitTillReadyCountDownLatch = new CountDownLatch(1); + + private final String type; - private SchemaContext schemaContext; + private final TransactionContextFactory txContextFactory; + public DistributedDataStore(ActorSystem actorSystem, ClusterWrapper cluster, + Configuration configuration, DatastoreContext datastoreContext) { + Preconditions.checkNotNull(actorSystem, "actorSystem should not be null"); + Preconditions.checkNotNull(cluster, "cluster should not be null"); + Preconditions.checkNotNull(configuration, "configuration should not be null"); + Preconditions.checkNotNull(datastoreContext, "datastoreContext should not be null"); + this.type = datastoreContext.getDataStoreType(); - /** - * Executor used to run FutureTask's - * - * This is typically used when we need to make a request to an actor and - * wait for it's response and the consumer needs to be provided a Future. - * - * FIXME : Make the thread pool size configurable. - */ - private final ListeningExecutorService executor = - MoreExecutors.listeningDecorator(Executors.newFixedThreadPool(DEFAULT_EXECUTOR_POOL_SIZE)); + String shardManagerId = ShardManagerIdentifier.builder().type(type).build().toString(); - public DistributedDataStore(ActorSystem actorSystem, String type, ClusterWrapper cluster, Configuration configuration) { - this(new ActorContext(actorSystem, actorSystem - .actorOf(ShardManager.props(type, cluster, configuration), - "shardmanager-" + type), cluster, configuration), type); + LOG.info("Creating ShardManager : {}", shardManagerId); + + String shardDispatcher = + new Dispatchers(actorSystem.dispatchers()).getDispatcherPath(Dispatchers.DispatcherType.Shard); + + PrimaryShardInfoFutureCache primaryShardInfoCache = new PrimaryShardInfoFutureCache(); + actorContext = new ActorContext(actorSystem, createShardManager(actorSystem, cluster, configuration, + datastoreContext, shardDispatcher, shardManagerId, primaryShardInfoCache), cluster, + configuration, datastoreContext, primaryShardInfoCache); + + this.waitTillReadyTimeInMillis = + actorContext.getDatastoreContext().getShardLeaderElectionTimeout().duration().toMillis() * READY_WAIT_FACTOR; + + this.txContextFactory = TransactionContextFactory.create(actorContext); + + datastoreConfigMXBean = new DatastoreConfigurationMXBeanImpl(datastoreContext.getDataStoreMXBeanType()); + datastoreConfigMXBean.setContext(datastoreContext); + datastoreConfigMXBean.registerMBean(); + + datastoreInfoMXBean = new DatastoreInfoMXBeanImpl(datastoreContext.getDataStoreMXBeanType(), actorContext); + datastoreInfoMXBean.registerMBean(); } - public DistributedDataStore(ActorContext actorContext, String type) { - this.type = type; - this.actorContext = actorContext; + @VisibleForTesting + DistributedDataStore(ActorContext actorContext) { + this.actorContext = Preconditions.checkNotNull(actorContext, "actorContext should not be null"); + this.txContextFactory = TransactionContextFactory.create(actorContext); + this.type = UNKNOWN_TYPE; + this.waitTillReadyTimeInMillis = + actorContext.getDatastoreContext().getShardLeaderElectionTimeout().duration().toMillis() * READY_WAIT_FACTOR; } + public void setCloseable(AutoCloseable closeable) { + this.closeable = closeable; + } + @SuppressWarnings("unchecked") @Override - public >> ListenerRegistration registerChangeListener( - YangInstanceIdentifier path, L listener, + public >> + ListenerRegistration registerChangeListener( + final YangInstanceIdentifier path, L listener, AsyncDataBroker.DataChangeScope scope) { - ActorRef dataChangeListenerActor = actorContext.getActorSystem().actorOf( - DataChangeListener.props(schemaContext,listener,path )); + Preconditions.checkNotNull(path, "path should not be null"); + Preconditions.checkNotNull(listener, "listener should not be null"); - String shardName = ShardStrategyFactory.getStrategy(path).findShard(path); + LOG.debug("Registering listener: {} for path: {} scope: {}", listener, path, scope); - Object result = actorContext.executeLocalShardOperation(shardName, - new RegisterChangeListener(path, dataChangeListenerActor.path(), - scope).toSerializable(), - ActorContext.ASK_DURATION - ); + String shardName = actorContext.getShardStrategyFactory().getStrategy(path).findShard(path); - if (result != null) { - RegisterChangeListenerReply reply = RegisterChangeListenerReply - .fromSerializable(actorContext.getActorSystem(), result); - return new DataChangeListenerRegistrationProxy(actorContext - .actorSelection(reply.getListenerRegistrationPath()), listener, - dataChangeListenerActor); - } + final DataChangeListenerRegistrationProxy listenerRegistrationProxy = + new DataChangeListenerRegistrationProxy(shardName, actorContext, listener); + listenerRegistrationProxy.init(path, scope); - LOG.debug( - "No local shard for shardName {} was found so returning a noop registration", - shardName); - return new NoOpDataChangeListenerRegistration(listener); + return listenerRegistrationProxy; } + @Override + public ListenerRegistration registerTreeChangeListener(YangInstanceIdentifier treeId, L listener) { + Preconditions.checkNotNull(treeId, "treeId should not be null"); + Preconditions.checkNotNull(listener, "listener should not be null"); + final String shardName = actorContext.getShardStrategyFactory().getStrategy(treeId).findShard(treeId); + LOG.debug("Registering tree listener: {} for tree: {} shard: {}", listener, treeId, shardName); + final DataTreeChangeListenerProxy listenerRegistrationProxy = + new DataTreeChangeListenerProxy(actorContext, listener); + listenerRegistrationProxy.init(shardName, treeId); + return listenerRegistrationProxy; + } @Override public DOMStoreTransactionChain createTransactionChain() { - return new TransactionChainProxy(actorContext, executor, schemaContext); + return txContextFactory.createTransactionChain(); } @Override public DOMStoreReadTransaction newReadOnlyTransaction() { - return new TransactionProxy(actorContext, TransactionProxy.TransactionType.READ_ONLY, - executor, schemaContext); + return new TransactionProxy(txContextFactory, TransactionType.READ_ONLY); } @Override public DOMStoreWriteTransaction newWriteOnlyTransaction() { - return new TransactionProxy(actorContext, TransactionProxy.TransactionType.WRITE_ONLY, - executor, schemaContext); + actorContext.acquireTxCreationPermit(); + return new TransactionProxy(txContextFactory, TransactionType.WRITE_ONLY); } @Override public DOMStoreReadWriteTransaction newReadWriteTransaction() { - return new TransactionProxy(actorContext, TransactionProxy.TransactionType.READ_WRITE, - executor, schemaContext); + actorContext.acquireTxCreationPermit(); + return new TransactionProxy(txContextFactory, TransactionType.READ_WRITE); } - @Override public void onGlobalContextUpdated(SchemaContext schemaContext) { - this.schemaContext = schemaContext; - actorContext.getShardManager().tell( - new UpdateSchemaContext(schemaContext), null); + @Override + public void onGlobalContextUpdated(SchemaContext schemaContext) { + actorContext.setSchemaContext(schemaContext); + } + + @Override + public void onDatastoreContextUpdated(DatastoreContext context) { + LOG.info("DatastoreContext updated for data store {}", actorContext.getDataStoreType()); + + actorContext.setDatastoreContext(context); + datastoreConfigMXBean.setContext(context); } - @Override public void close() throws Exception { + @Override + public void close() { + if (datastoreConfigMXBean != null) { + datastoreConfigMXBean.unregisterMBean(); + } + if (datastoreInfoMXBean != null) { + datastoreInfoMXBean.unregisterMBean(); + } + + if (closeable != null) { + try { + closeable.close(); + } catch (Exception e) { + LOG.debug("Error closing instance", e); + } + } + + txContextFactory.close(); actorContext.shutdown(); + } + + public ActorContext getActorContext() { + return actorContext; + } + + public void waitTillReady(){ + LOG.info("Beginning to wait for data store to become ready : {}", type); + + try { + if (waitTillReadyCountDownLatch.await(waitTillReadyTimeInMillis, TimeUnit.MILLISECONDS)) { + LOG.debug("Data store {} is now ready", type); + } else { + LOG.error("Shared leaders failed to settle in {} seconds, giving up", TimeUnit.MILLISECONDS.toSeconds(waitTillReadyTimeInMillis)); + } + } catch (InterruptedException e) { + LOG.error("Interrupted while waiting for shards to settle", e); + } + } + + private ActorRef createShardManager(ActorSystem actorSystem, ClusterWrapper cluster, Configuration configuration, + DatastoreContext datastoreContext, String shardDispatcher, String shardManagerId, + PrimaryShardInfoFutureCache primaryShardInfoCache){ + Exception lastException = null; + + for(int i=0;i<100;i++) { + try { + return actorSystem.actorOf( + ShardManager.props(cluster, configuration, datastoreContext, waitTillReadyCountDownLatch, + primaryShardInfoCache).withDispatcher(shardDispatcher).withMailbox( + ActorContext.MAILBOX), shardManagerId); + } catch (Exception e){ + lastException = e; + Uninterruptibles.sleepUninterruptibly(100, TimeUnit.MILLISECONDS); + LOG.debug(String.format("Could not create actor %s because of %s - waiting for sometime before retrying (retry count = %d)", shardManagerId, e.getMessage(), i)); + } + } + + throw new IllegalStateException("Failed to create Shard Manager", lastException); + } + @VisibleForTesting + public CountDownLatch getWaitTillReadyCountDownLatch() { + return waitTillReadyCountDownLatch; } }