*/
package io.atomix.storage.journal;
-import com.google.common.collect.Sets;
import io.atomix.storage.journal.index.JournalIndex;
import io.atomix.storage.journal.index.SparseJournalIndex;
import io.atomix.utils.serializer.Namespace;
import java.nio.file.Files;
import java.nio.file.StandardOpenOption;
import java.util.Set;
+import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.atomic.AtomicInteger;
import static com.google.common.base.MoreObjects.toStringHelper;
private final JournalIndex index;
private final Namespace namespace;
private final MappableJournalSegmentWriter<E> writer;
- private final Set<MappableJournalSegmentReader<E>> readers = Sets.newConcurrentHashSet();
+ private final Set<MappableJournalSegmentReader<E>> readers = ConcurrentHashMap.newKeySet();
private final AtomicInteger references = new AtomicInteger();
private boolean open = true;
import java.util.Map;
import java.util.SortedMap;
import java.util.TreeMap;
+import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.ConcurrentNavigableMap;
import java.util.concurrent.ConcurrentSkipListMap;
-import com.google.common.collect.Sets;
import io.atomix.utils.serializer.Namespace;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
private volatile long commitIndex;
private final ConcurrentNavigableMap<Long, JournalSegment<E>> segments = new ConcurrentSkipListMap<>();
- private final Collection<SegmentedJournalReader<E>> readers = Sets.newConcurrentHashSet();
+ private final Collection<SegmentedJournalReader<E>> readers = ConcurrentHashMap.newKeySet();
private JournalSegment<E> currentSegment;
private volatile boolean open = true;