X-Git-Url: https://git.opendaylight.org/gerrit/gitweb?p=controller.git;a=blobdiff_plain;f=opendaylight%2Fmd-sal%2Fsal-distributed-datastore%2Fsrc%2Fmain%2Fjava%2Forg%2Fopendaylight%2Fcontroller%2Fcluster%2Fdatastore%2FAbstractFrontendHistory.java;h=8563c3e913a87ea1de742b9a0b76b7ce3a1847e9;hp=1adca56af2680a32a5c4f68afe39bc1d0f79533d;hb=8ed5603b4f3503559f2f85137c42d7605ebfd3e9;hpb=db9a673c114febc785fbd324947ac2c3e3095d06 diff --git a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/AbstractFrontendHistory.java b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/AbstractFrontendHistory.java index 1adca56af2..8563c3e913 100644 --- a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/AbstractFrontendHistory.java +++ b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/AbstractFrontendHistory.java @@ -7,14 +7,25 @@ */ package org.opendaylight.controller.cluster.datastore; +import com.google.common.base.MoreObjects; import com.google.common.base.Preconditions; -import com.google.common.base.Ticker; +import com.google.common.collect.ImmutableMap; +import com.google.common.collect.Range; +import com.google.common.collect.RangeSet; +import com.google.common.primitives.UnsignedLong; import java.util.HashMap; import java.util.Map; import java.util.Optional; import javax.annotation.Nullable; +import org.opendaylight.controller.cluster.access.commands.AbstractReadTransactionRequest; +import org.opendaylight.controller.cluster.access.commands.ClosedTransactionException; import org.opendaylight.controller.cluster.access.commands.CommitLocalTransactionRequest; +import org.opendaylight.controller.cluster.access.commands.DeadTransactionException; +import org.opendaylight.controller.cluster.access.commands.IncrementTransactionSequenceRequest; +import org.opendaylight.controller.cluster.access.commands.LocalHistorySuccess; import org.opendaylight.controller.cluster.access.commands.OutOfOrderRequestException; +import org.opendaylight.controller.cluster.access.commands.TransactionPurgeRequest; +import org.opendaylight.controller.cluster.access.commands.TransactionPurgeResponse; import org.opendaylight.controller.cluster.access.commands.TransactionRequest; import org.opendaylight.controller.cluster.access.commands.TransactionSuccess; import org.opendaylight.controller.cluster.access.concepts.LocalHistoryIdentifier; @@ -34,15 +45,24 @@ import org.slf4j.LoggerFactory; */ abstract class AbstractFrontendHistory implements Identifiable { private static final Logger LOG = LoggerFactory.getLogger(AbstractFrontendHistory.class); - private static final OutOfOrderRequestException UNSEQUENCED_START = new OutOfOrderRequestException(0); private final Map transactions = new HashMap<>(); + private final RangeSet purgedTransactions; private final String persistenceId; - private final Ticker ticker; + private final ShardDataTree tree; - AbstractFrontendHistory(final String persistenceId, final Ticker ticker) { + /** + * Transactions closed by the previous leader. Boolean indicates whether the transaction was committed (true) or + * aborted (false). We only ever shrink these. + */ + private Map closedTransactions; + + AbstractFrontendHistory(final String persistenceId, final ShardDataTree tree, + final Map closedTransactions, final RangeSet purgedTransactions) { this.persistenceId = Preconditions.checkNotNull(persistenceId); - this.ticker = Preconditions.checkNotNull(ticker); + this.tree = Preconditions.checkNotNull(tree); + this.closedTransactions = Preconditions.checkNotNull(closedTransactions); + this.purgedTransactions = Preconditions.checkNotNull(purgedTransactions); } final String persistenceId() { @@ -50,33 +70,79 @@ abstract class AbstractFrontendHistory implements Identifiable handleTransactionRequest(final TransactionRequest request, final RequestEnvelope envelope, final long now) throws RequestException { + final TransactionIdentifier id = request.getTarget(); + final UnsignedLong ul = UnsignedLong.fromLongBits(id.getTransactionId()); - // FIXME: handle purging of transactions + if (request instanceof TransactionPurgeRequest) { + if (purgedTransactions.contains(ul)) { + // Retransmitted purge request: nothing to do + LOG.debug("{}: transaction {} already purged", persistenceId, id); + return new TransactionPurgeResponse(id, request.getSequence()); + } + + // We perform two lookups instead of a straight remove, because once the map becomes empty we switch it + // to an ImmutableMap, which does not allow remove(). + if (closedTransactions.containsKey(ul)) { + tree.purgeTransaction(id, () -> { + closedTransactions.remove(ul); + if (closedTransactions.isEmpty()) { + closedTransactions = ImmutableMap.of(); + } + + purgedTransactions.add(Range.closedOpen(ul, UnsignedLong.ONE.plus(ul))); + LOG.debug("{}: finished purging inherited transaction {}", persistenceId(), id); + envelope.sendSuccess(new TransactionPurgeResponse(id, request.getSequence()), readTime() - now); + }); + return null; + } + + final FrontendTransaction tx = transactions.get(id); + if (tx == null) { + // This should never happen because the purge callback removes the transaction and puts it into + // purged transactions in one go. If it does, we warn about the situation and + LOG.warn("{}: transaction {} not tracked in {}, but not present in active transactions", persistenceId, + id, purgedTransactions); + purgedTransactions.add(Range.closedOpen(ul, UnsignedLong.ONE.plus(ul))); + return new TransactionPurgeResponse(id, request.getSequence()); + } + + tree.purgeTransaction(id, () -> { + purgedTransactions.add(Range.closedOpen(ul, UnsignedLong.ONE.plus(ul))); + transactions.remove(id); + LOG.debug("{}: finished purging transaction {}", persistenceId(), id); + envelope.sendSuccess(new TransactionPurgeResponse(id, request.getSequence()), readTime() - now); + }); + return null; + } + + if (purgedTransactions.contains(ul)) { + LOG.warn("{}: Request {} is contained purged transactions {}", persistenceId, request, purgedTransactions); + throw new DeadTransactionException(purgedTransactions); + } + final Boolean closed = closedTransactions.get(ul); + if (closed != null) { + final boolean successful = closed.booleanValue(); + LOG.debug("{}: Request {} refers to a {} transaction", persistenceId, request, successful ? "successful" + : "failed"); + throw new ClosedTransactionException(successful); + } - final TransactionIdentifier id = request.getTarget(); FrontendTransaction tx = transactions.get(id); if (tx == null) { // The transaction does not exist and we are about to create it, check sequence number if (request.getSequence() != 0) { - LOG.debug("{}: no transaction state present, unexpected request {}", persistenceId(), request); - throw UNSEQUENCED_START; - } - - if (request instanceof CommitLocalTransactionRequest) { - tx = createReadyTransaction(id, ((CommitLocalTransactionRequest) request).getModification()); - LOG.debug("{}: allocated new ready transaction {}", persistenceId(), id); - } else { - tx = createOpenTransaction(id); - LOG.debug("{}: allocated new open transaction {}", persistenceId(), id); + LOG.warn("{}: no transaction state present, unexpected request {}", persistenceId(), request); + throw new OutOfOrderRequestException(0); } + tx = createTransaction(request, id); transactions.put(id, tx); - } else { + } else if (!(request instanceof IncrementTransactionSequenceRequest)) { final Optional> maybeReplay = tx.replaySequence(request.getSequence()); if (maybeReplay.isPresent()) { final TransactionSuccess replay = maybeReplay.get(); @@ -88,10 +154,55 @@ abstract class AbstractFrontendHistory implements Identifiable { + envelope.sendSuccess(new LocalHistorySuccess(getIdentifier(), sequence), readTime() - now); + }); + } + + void purge(final long sequence, final RequestEnvelope envelope, final long now) { + LOG.debug("{}: purging history {}", persistenceId(), getIdentifier()); + tree.purgeTransactionChain(getIdentifier(), () -> { + envelope.sendSuccess(new LocalHistorySuccess(getIdentifier(), sequence), readTime() - now); + }); + } + + private FrontendTransaction createTransaction(final TransactionRequest request, final TransactionIdentifier id) + throws RequestException { + if (request instanceof CommitLocalTransactionRequest) { + LOG.debug("{}: allocating new ready transaction {}", persistenceId(), id); + tree.getStats().incrementReadWriteTransactionCount(); + return createReadyTransaction(id, ((CommitLocalTransactionRequest) request).getModification()); + } + if (request instanceof AbstractReadTransactionRequest) { + if (((AbstractReadTransactionRequest) request).isSnapshotOnly()) { + LOG.debug("{}: allocating new open snapshot {}", persistenceId(), id); + tree.getStats().incrementReadOnlyTransactionCount(); + return createOpenSnapshot(id); + } + } + + LOG.debug("{}: allocating new open transaction {}", persistenceId(), id); + tree.getStats().incrementReadWriteTransactionCount(); + return createOpenTransaction(id); + } + + abstract FrontendTransaction createOpenSnapshot(TransactionIdentifier id) throws RequestException; + abstract FrontendTransaction createOpenTransaction(TransactionIdentifier id) throws RequestException; abstract FrontendTransaction createReadyTransaction(TransactionIdentifier id, DataTreeModification mod) throws RequestException; - abstract ShardDataTreeCohort createReadyCohort(final TransactionIdentifier id, final DataTreeModification mod); + abstract ShardDataTreeCohort createFailedCohort(TransactionIdentifier id, DataTreeModification mod, + Exception failure); + + abstract ShardDataTreeCohort createReadyCohort(TransactionIdentifier id, DataTreeModification mod); + + @Override + public String toString() { + return MoreObjects.toStringHelper(this).omitNullValues().add("identifier", getIdentifier()) + .add("persistenceId", persistenceId).add("transactions", transactions).toString(); + } }