package org.opendaylight.controller.cluster.datastore.persisted;
import com.google.common.annotations.Beta;
-import java.io.ByteArrayInputStream;
-import java.io.DataInputStream;
import java.io.IOException;
-import java.io.InputStream;
-import java.io.OutputStream;
+import java.io.ObjectInput;
+import java.io.ObjectOutput;
import java.util.Optional;
-import org.opendaylight.controller.cluster.datastore.node.utils.stream.SerializationUtils;
import org.opendaylight.yangtools.yang.data.api.schema.NormalizedNode;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* Abstract base class for snapshots of the ShardDataTree.
*/
@Beta
public abstract class ShardDataTreeSnapshot {
- private static final Logger LOG = LoggerFactory.getLogger(ShardDataTreeSnapshot.class);
-
ShardDataTreeSnapshot() {
// Hidden to prevent subclassing from outside of this package
}
- @Deprecated
- public static ShardDataTreeSnapshot deserialize(final byte[] bytes) throws IOException {
- /**
- * Unfortunately versions prior to Boron did not include any way to evolve the snapshot format and contained
- * only the raw data stored in the datastore. Furthermore utilities involved do not check if the array is
- * completely consumed, which has a nasty side-effect when coupled with the fact that PayloadVersion writes
- * a short value.
- *
- * Since our versions fit into a single byte, we end up writing the 0 as the first byte, which would be
- * interpreted as 'not present' by the old snapshot format, which uses writeBoolean/readBoolean. A further
- * complication is that readBoolean interprets any non-zero value as true, hence we cannot use a wild value
- * to cause it to fail.
- */
- if (isLegacyStream(bytes)) {
- return deserializeLegacy(bytes);
- }
+ public static ShardDataTreeSnapshot deserialize(final ObjectInput in) throws IOException {
+ final ShardDataTreeSnapshot ret = AbstractVersionedShardDataTreeSnapshot.versionedDeserialize(in);
- try {
- try (final InputStream is = new ByteArrayInputStream(bytes)) {
- return deserialize(is);
- }
- } catch (IOException e) {
- LOG.debug("Failed to deserialize versioned stream, attempting pre-Lithium ProtoBuf", e);
- return deserializeLegacy(bytes);
+ // Make sure we consume all bytes, otherwise something went very wrong
+ final int bytesLeft = in.available();
+ if (bytesLeft != 0) {
+ throw new IOException("Deserialization left " + bytesLeft + " in the buffer");
}
- }
- public static ShardDataTreeSnapshot deserialize(final InputStream is) throws IOException {
- try (final DataInputStream dis = new DataInputStream(is)) {
- final ShardDataTreeSnapshot ret = AbstractVersionedShardDataTreeSnapshot.deserialize(dis);
- // Make sure we consume all bytes, otherwise something went very wrong
- final int bytesLeft = dis.available();
- if (bytesLeft != 0) {
- throw new IOException("Deserialization left " + bytesLeft + " in the buffer");
- }
-
-
- return ret;
- }
+ return ret;
}
/**
*/
public abstract Optional<NormalizedNode<?, ?>> getRootNode();
- public abstract void serialize(final OutputStream os) throws IOException;
-
- @Deprecated
- private static boolean isLegacyStream(final byte[] bytes) {
- if (bytes.length < 2) {
- // Versioned streams have at least two bytes
- return true;
- }
-
- /*
- * The stream could potentially be a versioned stream. Here we rely on the signature marker available
- * in org.opendaylight.controller.cluster.datastore.node.utils.stream.TokenTypes.
- *
- * For an old stream to be this long, the first byte has to be non-zero and the second byte has to be 0xAB.
- *
- * For a versioned stream, that translates to at least version 427 -- giving us at least 421 further versions
- * before this check breaks.
- */
- return bytes[0] != 0 && bytes[1] == (byte)0xAB;
- }
-
- @Deprecated
- private static ShardDataTreeSnapshot deserializeLegacy(final byte[] bytes) {
- return new PreBoronShardDataTreeSnapshot(SerializationUtils.deserializeNormalizedNode(bytes));
- }
+ public abstract void serialize(ObjectOutput out) throws IOException;
}