X-Git-Url: https://git.opendaylight.org/gerrit/gitweb?p=controller.git;a=blobdiff_plain;f=opendaylight%2Fmd-sal%2Fsal-distributed-datastore%2Fsrc%2Fmain%2Fjava%2Forg%2Fopendaylight%2Fcontroller%2Fcluster%2Fdatastore%2FShardRecoveryCoordinator.java;h=a916ea6dead760219d536e0061ac01c35c99a879;hp=f3cbc8649dba9d44fe1c619d6533535c34045f02;hb=0acbc10a6c37b6faeaa4ce67701d90f18aa89bde;hpb=34c6032dd81c2d76720cce53478c38f5e5cdddc4 diff --git a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/ShardRecoveryCoordinator.java b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/ShardRecoveryCoordinator.java index f3cbc8649d..a916ea6dea 100644 --- a/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/ShardRecoveryCoordinator.java +++ b/opendaylight/md-sal/sal-distributed-datastore/src/main/java/org/opendaylight/controller/cluster/datastore/ShardRecoveryCoordinator.java @@ -8,19 +8,17 @@ package org.opendaylight.controller.cluster.datastore; import com.google.common.base.Preconditions; +import com.google.common.base.Throwables; import java.io.File; import java.io.IOException; -import org.opendaylight.controller.cluster.datastore.utils.DataTreeModificationOutput; +import org.opendaylight.controller.cluster.datastore.persisted.ShardDataTreeSnapshot; +import org.opendaylight.controller.cluster.datastore.persisted.ShardSnapshotState; import org.opendaylight.controller.cluster.datastore.utils.NormalizedNodeXMLOutput; -import org.opendaylight.controller.cluster.datastore.utils.PruningDataTreeModification; -import org.opendaylight.controller.cluster.datastore.utils.SerializationUtils; import org.opendaylight.controller.cluster.raft.RaftActorRecoveryCohort; +import org.opendaylight.controller.cluster.raft.persisted.Snapshot; +import org.opendaylight.controller.cluster.raft.persisted.Snapshot.State; import org.opendaylight.controller.cluster.raft.protobuff.client.messages.Payload; -import org.opendaylight.yangtools.yang.data.api.YangInstanceIdentifier; import org.opendaylight.yangtools.yang.data.api.schema.NormalizedNode; -import org.opendaylight.yangtools.yang.data.api.schema.tree.DataTreeCandidates; -import org.opendaylight.yangtools.yang.data.api.schema.tree.DataValidationFailedException; -import org.opendaylight.yangtools.yang.model.api.SchemaContext; import org.slf4j.Logger; /** @@ -36,95 +34,93 @@ class ShardRecoveryCoordinator implements RaftActorRecoveryCohort { private final ShardDataTree store; private final String shardName; private final Logger log; - private final SchemaContext schemaContext; - private PruningDataTreeModification transaction; - private int size; - private final byte[] restoreFromSnapshot; + private final Snapshot restoreFromSnapshot; - ShardRecoveryCoordinator(ShardDataTree store, SchemaContext schemaContext, byte[] restoreFromSnapshot, - String shardName, Logger log) { + private boolean open; + + ShardRecoveryCoordinator(final ShardDataTree store, final Snapshot restoreFromSnapshot, final String shardName, + final Logger log) { this.store = Preconditions.checkNotNull(store); + this.shardName = Preconditions.checkNotNull(shardName); + this.log = Preconditions.checkNotNull(log); + this.restoreFromSnapshot = restoreFromSnapshot; - this.shardName = shardName; - this.log = log; - this.schemaContext = schemaContext; } @Override - public void startLogRecoveryBatch(int maxBatchSize) { + public void startLogRecoveryBatch(final int maxBatchSize) { log.debug("{}: starting log recovery batch with max size {}", shardName, maxBatchSize); - transaction = new PruningDataTreeModification(store.newModification(), store.getDataTree(), schemaContext); - size = 0; + open = true; } @Override - public void appendRecoveredLogEntry(Payload payload) { - Preconditions.checkState(transaction != null, "call startLogRecovery before calling appendRecoveredLogEntry"); + @SuppressWarnings("checkstyle:IllegalCatch") + public void appendRecoveredLogEntry(final Payload payload) { + Preconditions.checkState(open, "call startLogRecovery before calling appendRecoveredLogEntry"); try { - if (payload instanceof DataTreeCandidatePayload) { - DataTreeCandidates.applyToModification(transaction, ((DataTreeCandidatePayload)payload).getCandidate()); - size++; - } else { - log.error("{}: Unknown payload {} received during recovery", shardName, payload); - } - } catch (IOException e) { - log.error("{}: Error extracting ModificationPayload", shardName, e); + store.applyRecoveryPayload(payload); + } catch (Exception e) { + log.error("{}: failed to apply payload {}", shardName, payload, e); + throw new IllegalStateException(String.format("%s: Failed to apply recovery payload %s", + shardName, payload), e); } } - private void commitTransaction(PruningDataTreeModification tx) throws DataValidationFailedException { - store.commit(tx.getResultingModification()); - } - /** * Applies the current batched log entries to the data store. */ @Override public void applyCurrentLogRecoveryBatch() { - Preconditions.checkState(transaction != null, "call startLogRecovery before calling applyCurrentLogRecoveryBatch"); + Preconditions.checkState(open, "call startLogRecovery before calling applyCurrentLogRecoveryBatch"); + open = false; + } - log.debug("{}: Applying current log recovery batch with size {}", shardName, size); - try { - commitTransaction(transaction); - } catch (Exception e) { - File file = new File(System.getProperty("karaf.data", "."), - "failed-recovery-batch-" + shardName + ".out"); - DataTreeModificationOutput.toFile(file, transaction.getResultingModification()); - throw new RuntimeException(String.format( - "%s: Failed to apply recovery batch. Modification data was written to file %s", - shardName, file), e); - } - transaction = null; + private File writeRoot(final String kind, final NormalizedNode node) { + final File file = new File(System.getProperty("karaf.data", "."), + "failed-recovery-" + kind + "-" + shardName + ".xml"); + NormalizedNodeXMLOutput.toFile(file, node); + return file; } /** * Applies a recovered snapshot to the data store. * - * @param snapshotBytes the serialized snapshot + * @param snapshotState the serialized snapshot */ @Override - public void applyRecoverySnapshot(final byte[] snapshotBytes) { + @SuppressWarnings("checkstyle:IllegalCatch") + public void applyRecoverySnapshot(final Snapshot.State snapshotState) { + if (!(snapshotState instanceof ShardSnapshotState)) { + log.debug("{}: applyRecoverySnapshot ignoring snapshot: {}", snapshotState); + } + log.debug("{}: Applying recovered snapshot", shardName); - final NormalizedNode node = SerializationUtils.deserializeNormalizedNode(snapshotBytes); - final PruningDataTreeModification tx = new PruningDataTreeModification(store.newModification(), - store.getDataTree(), schemaContext); - tx.write(YangInstanceIdentifier.EMPTY, node); + ShardDataTreeSnapshot shardSnapshot = ((ShardSnapshotState)snapshotState).getSnapshot(); try { - commitTransaction(tx); + store.applyRecoverySnapshot(shardSnapshot); } catch (Exception e) { - File file = new File(System.getProperty("karaf.data", "."), - "failed-recovery-snapshot-" + shardName + ".xml"); - NormalizedNodeXMLOutput.toFile(file, node); - throw new RuntimeException(String.format( - "%s: Failed to apply recovery snapshot. Node data was written to file %s", - shardName, file), e); + final File f = writeRoot("snapshot", shardSnapshot.getRootNode().orElse(null)); + throw new IllegalStateException(String.format( + "%s: Failed to apply recovery snapshot %s. Node data was written to file %s", + shardName, shardSnapshot, f), e); } } @Override - public byte[] getRestoreFromSnapshot() { + public Snapshot getRestoreFromSnapshot() { return restoreFromSnapshot; } + + @Override + @Deprecated + public State deserializePreCarbonSnapshot(final byte[] from) { + try { + return new ShardSnapshotState(ShardDataTreeSnapshot.deserializePreCarbon(from)); + } catch (IOException e) { + log.error("{}: failed to deserialize snapshot", shardName, e); + throw Throwables.propagate(e); + } + } }