import com.google.common.base.MoreObjects;
import com.google.common.base.Preconditions;
-import com.google.common.collect.Range;
-import com.google.common.collect.RangeSet;
-import com.google.common.collect.TreeRangeSet;
-import com.google.common.primitives.UnsignedLong;
import java.util.HashMap;
+import java.util.Iterator;
import java.util.Map;
import javax.annotation.Nullable;
import javax.annotation.concurrent.NotThreadSafe;
import org.opendaylight.controller.cluster.access.commands.DestroyLocalHistoryRequest;
import org.opendaylight.controller.cluster.access.commands.LocalHistoryRequest;
import org.opendaylight.controller.cluster.access.commands.LocalHistorySuccess;
-import org.opendaylight.controller.cluster.access.commands.OutOfOrderRequestException;
+import org.opendaylight.controller.cluster.access.commands.OutOfSequenceEnvelopeException;
import org.opendaylight.controller.cluster.access.commands.PurgeLocalHistoryRequest;
import org.opendaylight.controller.cluster.access.commands.TransactionRequest;
import org.opendaylight.controller.cluster.access.commands.TransactionSuccess;
import org.opendaylight.controller.cluster.access.concepts.RequestEnvelope;
import org.opendaylight.controller.cluster.access.concepts.RequestException;
import org.opendaylight.controller.cluster.access.concepts.UnsupportedRequestException;
+import org.opendaylight.controller.cluster.datastore.ShardDataTreeCohort.State;
+import org.opendaylight.controller.cluster.datastore.utils.UnsignedLongRangeSet;
import org.opendaylight.yangtools.concepts.Identifiable;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
private final Map<LocalHistoryIdentifier, LocalFrontendHistory> localHistories;
// RangeSet performs automatic merging, hence we keep minimal state tracking information
- private final RangeSet<UnsignedLong> purgedHistories;
+ private final UnsignedLongRangeSet purgedHistories;
// Used for all standalone transactions
private final AbstractFrontendHistory standaloneHistory;
// - per-RequestException throw counters
LeaderFrontendState(final String persistenceId, final ClientIdentifier clientId, final ShardDataTree tree) {
- this(persistenceId, clientId, tree, TreeRangeSet.create(), StandaloneFrontendHistory.create(persistenceId,
- clientId, tree), new HashMap<>());
+ this(persistenceId, clientId, tree, UnsignedLongRangeSet.create(),
+ StandaloneFrontendHistory.create(persistenceId, clientId, tree), new HashMap<>());
}
LeaderFrontendState(final String persistenceId, final ClientIdentifier clientId, final ShardDataTree tree,
- final RangeSet<UnsignedLong> purgedHistories, final AbstractFrontendHistory standaloneHistory,
+ final UnsignedLongRangeSet purgedHistories, final AbstractFrontendHistory standaloneHistory,
final Map<LocalHistoryIdentifier, LocalFrontendHistory> localHistories) {
this.persistenceId = Preconditions.checkNotNull(persistenceId);
this.clientId = Preconditions.checkNotNull(clientId);
return clientId;
}
- private void checkRequestSequence(final RequestEnvelope envelope) throws OutOfOrderRequestException {
+ private void checkRequestSequence(final RequestEnvelope envelope) throws OutOfSequenceEnvelopeException {
if (expectedTxSequence != envelope.getTxSequence()) {
- throw new OutOfOrderRequestException(expectedTxSequence);
+ throw new OutOfSequenceEnvelopeException(expectedTxSequence);
}
}
try {
if (request instanceof CreateLocalHistoryRequest) {
- return handleCreateHistory((CreateLocalHistoryRequest) request);
+ return handleCreateHistory((CreateLocalHistoryRequest) request, envelope, now);
} else if (request instanceof DestroyLocalHistoryRequest) {
return handleDestroyHistory((DestroyLocalHistoryRequest) request, envelope, now);
} else if (request instanceof PurgeLocalHistoryRequest) {
return handlePurgeHistory((PurgeLocalHistoryRequest)request, envelope, now);
} else {
+ LOG.warn("{}: rejecting unsupported request {}", persistenceId, request);
throw new UnsupportedRequestException(request);
}
} finally {
}
}
- private LocalHistorySuccess handleCreateHistory(final CreateLocalHistoryRequest request) throws RequestException {
- final LocalHistoryIdentifier id = request.getTarget();
- final AbstractFrontendHistory existing = localHistories.get(id);
+ private LocalHistorySuccess handleCreateHistory(final CreateLocalHistoryRequest request,
+ final RequestEnvelope envelope, final long now) throws RequestException {
+ final LocalHistoryIdentifier historyId = request.getTarget();
+ final AbstractFrontendHistory existing = localHistories.get(historyId);
if (existing != null) {
// History already exists: report success
- LOG.debug("{}: history {} already exists", persistenceId, id);
- return new LocalHistorySuccess(id, request.getSequence());
+ LOG.debug("{}: history {} already exists", persistenceId, historyId);
+ return new LocalHistorySuccess(historyId, request.getSequence());
}
// We have not found the history. Before we create it we need to check history ID sequencing so that we do not
// end up resurrecting a purged history.
- if (purgedHistories.contains(UnsignedLong.fromLongBits(id.getHistoryId()))) {
+ if (purgedHistories.contains(historyId.getHistoryId())) {
LOG.debug("{}: rejecting purged request {}", persistenceId, request);
- throw new DeadHistoryException(lastSeenHistory.longValue());
+ throw new DeadHistoryException(purgedHistories.toImmutable());
}
// Update last history we have seen
- if (lastSeenHistory != null && Long.compareUnsigned(lastSeenHistory, id.getHistoryId()) < 0) {
- lastSeenHistory = id.getHistoryId();
+ if (lastSeenHistory == null || Long.compareUnsigned(lastSeenHistory, historyId.getHistoryId()) < 0) {
+ lastSeenHistory = historyId.getHistoryId();
}
- localHistories.put(id, LocalFrontendHistory.create(persistenceId, tree, id));
- LOG.debug("{}: created history {}", persistenceId, id);
- return new LocalHistorySuccess(id, request.getSequence());
+ // We have to send the response only after persistence has completed
+ final ShardDataTreeTransactionChain chain = tree.ensureTransactionChain(historyId, () -> {
+ LOG.debug("{}: persisted history {}", persistenceId, historyId);
+ envelope.sendSuccess(new LocalHistorySuccess(historyId, request.getSequence()), tree.readTime() - now);
+ });
+
+ localHistories.put(historyId, LocalFrontendHistory.create(persistenceId, tree, chain));
+ LOG.debug("{}: created history {}", persistenceId, historyId);
+ return null;
}
private LocalHistorySuccess handleDestroyHistory(final DestroyLocalHistoryRequest request,
}
LOG.debug("{}: purging history {}", persistenceId, id);
- purgedHistories.add(Range.singleton(UnsignedLong.fromLongBits(id.getHistoryId())));
+ purgedHistories.add(id.getHistoryId());
existing.purge(request.getSequence(), envelope, now);
return null;
}
if (lhId.getHistoryId() != 0) {
history = localHistories.get(lhId);
if (history == null) {
- LOG.debug("{}: rejecting unknown history request {}", persistenceId, request);
+ if (purgedHistories.contains(lhId.getHistoryId())) {
+ LOG.warn("{}: rejecting request {} to purged history", persistenceId, request);
+ throw new DeadHistoryException(purgedHistories.toImmutable());
+ }
+
+ LOG.warn("{}: rejecting unknown history request {}", persistenceId, request);
throw new UnknownHistoryException(lastSeenHistory);
}
} else {
}
void retire() {
- // FIXME: flush all state
+ // Hunt down any transactions associated with this frontend
+ final Iterator<SimpleShardDataTreeCohort> it = tree.cohortIterator();
+ while (it.hasNext()) {
+ final SimpleShardDataTreeCohort cohort = it.next();
+ if (clientId.equals(cohort.getIdentifier().getHistoryId().getClientId())) {
+ if (cohort.getState() != State.COMMIT_PENDING) {
+ LOG.debug("{}: Retiring transaction {}", persistenceId, cohort.getIdentifier());
+ it.remove();
+ } else {
+ LOG.debug("{}: Transaction {} already committing, not retiring it", persistenceId,
+ cohort.getIdentifier());
+ }
+ }
+ }
+
+ // Clear out all transaction chains
+ localHistories.values().forEach(AbstractFrontendHistory::retire);
+ localHistories.clear();
+ standaloneHistory.retire();
}
@Override