package org.opendaylight.controller.statisticsmanager.internal;
import java.util.ArrayList;
+import java.util.Collections;
+import java.util.EnumSet;
import java.util.HashMap;
+import java.util.HashSet;
import java.util.List;
import java.util.Map;
+import java.util.Map.Entry;
+import java.util.Set;
+import java.util.concurrent.ConcurrentHashMap;
+import java.util.concurrent.ConcurrentMap;
+import java.util.concurrent.CountDownLatch;
+import java.util.concurrent.ExecutorService;
+import java.util.concurrent.Executors;
+import java.util.concurrent.TimeUnit;
+import java.util.concurrent.atomic.AtomicInteger;
+import org.opendaylight.controller.clustering.services.CacheConfigException;
+import org.opendaylight.controller.clustering.services.CacheExistException;
+import org.opendaylight.controller.clustering.services.ICacheUpdateAware;
+import org.opendaylight.controller.clustering.services.IClusterContainerServices;
+import org.opendaylight.controller.clustering.services.IClusterServices;
+import org.opendaylight.controller.connectionmanager.IConnectionManager;
import org.opendaylight.controller.forwardingrulesmanager.FlowEntry;
+import org.opendaylight.controller.sal.connection.ConnectionLocality;
+import org.opendaylight.controller.sal.core.IContainer;
import org.opendaylight.controller.sal.core.Node;
import org.opendaylight.controller.sal.core.NodeConnector;
import org.opendaylight.controller.sal.core.NodeTable;
+import org.opendaylight.controller.sal.core.Property;
+import org.opendaylight.controller.sal.core.UpdateType;
import org.opendaylight.controller.sal.flowprogrammer.Flow;
+import org.opendaylight.controller.sal.inventory.IListenInventoryUpdates;
import org.opendaylight.controller.sal.reader.FlowOnNode;
import org.opendaylight.controller.sal.reader.IReadService;
+import org.opendaylight.controller.sal.reader.IReadServiceListener;
import org.opendaylight.controller.sal.reader.NodeConnectorStatistics;
import org.opendaylight.controller.sal.reader.NodeDescription;
import org.opendaylight.controller.sal.reader.NodeTableStatistics;
+import org.opendaylight.controller.sal.utils.ServiceHelper;
import org.opendaylight.controller.statisticsmanager.IStatisticsManager;
+import org.opendaylight.controller.switchmanager.ISwitchManager;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
/**
- * The class which implements the methods for retrieving
- * the network nodes statistics.
+ * The class caches latest network nodes statistics as notified by reader
+ * services and provides API to retrieve them.
*/
-public class StatisticsManager implements IStatisticsManager {
- private static final Logger log = LoggerFactory
- .getLogger(StatisticsManager.class);
+public class StatisticsManager implements IStatisticsManager, IReadServiceListener, IListenInventoryUpdates,
+ ICacheUpdateAware<Object,Object> {
+ private static final Logger log = LoggerFactory.getLogger(StatisticsManager.class);
+ private IContainer container;
+ private IClusterContainerServices clusterContainerService;
private IReadService reader;
+ private IConnectionManager connectionManager;
+ //statistics caches
+ private ConcurrentMap<Node, List<FlowOnNode>> flowStatistics;
+ private ConcurrentMap<Node, List<NodeConnectorStatistics>> nodeConnectorStatistics;
+ private ConcurrentMap<Node, List<NodeTableStatistics>> tableStatistics;
+ private ConcurrentMap<Node, NodeDescription> descriptionStatistics;
- public StatisticsManager() {
+ // data structure for latches
+ // this is not a cluster cache
+ private ConcurrentMap<Node, CountDownLatch> latches = new ConcurrentHashMap<Node, CountDownLatch>();
+ // 30 seconds is the timeout.
+ // the value of this can be tweaked based on performance tests.
+ private static long latchTimeout = 30;
+ // cache for flow stats refresh triggers
+ // an entry added to this map triggers the statistics manager
+ // to which the node is connected to get the latest flow stats from that node
+ // this is a cluster cache
+ private ConcurrentMap<Integer, Node> triggers;
+
+ // use an atomic integer for the triggers key
+ private AtomicInteger triggerKey = new AtomicInteger();
+
+ // single thread executor for the triggers
+ private ExecutorService triggerExecutor;
+
+ static final String TRIGGERS_CACHE = "statisticsmanager.triggers";
+ static final String FLOW_STATISTICS_CACHE = "statisticsmanager.flowStatistics";
+
+ private void nonClusterObjectCreate() {
+ flowStatistics = new ConcurrentHashMap<Node, List<FlowOnNode>>();
+ nodeConnectorStatistics = new ConcurrentHashMap<Node, List<NodeConnectorStatistics>>();
+ tableStatistics = new ConcurrentHashMap<Node, List<NodeTableStatistics>>();
+ descriptionStatistics = new ConcurrentHashMap<Node, NodeDescription>();
+ triggers = new ConcurrentHashMap<Integer, Node>();
+ }
+
+ private void allocateCaches() {
+ if (clusterContainerService == null) {
+ nonClusterObjectCreate();
+ log.error("Clustering service unavailable. Allocated non-cluster statistics manager cache.");
+ return;
+ }
+
+ try {
+ clusterContainerService.createCache(FLOW_STATISTICS_CACHE,
+ EnumSet.of(IClusterServices.cacheMode.TRANSACTIONAL));
+ clusterContainerService.createCache("statisticsmanager.nodeConnectorStatistics",
+ EnumSet.of(IClusterServices.cacheMode.TRANSACTIONAL));
+ clusterContainerService.createCache("statisticsmanager.tableStatistics",
+ EnumSet.of(IClusterServices.cacheMode.TRANSACTIONAL));
+ clusterContainerService.createCache("statisticsmanager.descriptionStatistics",
+ EnumSet.of(IClusterServices.cacheMode.TRANSACTIONAL));
+ clusterContainerService.createCache(TRIGGERS_CACHE,
+ EnumSet.of(IClusterServices.cacheMode.NON_TRANSACTIONAL, IClusterServices.cacheMode.ASYNC));
+ } catch (CacheConfigException cce) {
+ log.error("Statistics cache configuration invalid - check cache mode");
+ } catch (CacheExistException ce) {
+ log.debug("Skipping statistics cache creation - already present");
+ }
+ }
+ @SuppressWarnings({ "unchecked" })
+ private void retrieveCaches() {
+ ConcurrentMap<?, ?> map;
+
+ if (this.clusterContainerService == null) {
+ log.warn("Can't retrieve statistics manager cache, Clustering service unavailable.");
+ return;
+ }
+
+ log.debug("Statistics Manager - retrieveCaches for Container {}", container);
+
+ map = clusterContainerService.getCache(FLOW_STATISTICS_CACHE);
+ if (map != null) {
+ this.flowStatistics = (ConcurrentMap<Node, List<FlowOnNode>>) map;
+ } else {
+ log.error("Cache allocation failed for statisticsmanager.flowStatistics in container {}", container.getName());
+ }
+
+ map = clusterContainerService.getCache("statisticsmanager.nodeConnectorStatistics");
+ if (map != null) {
+ this.nodeConnectorStatistics = (ConcurrentMap<Node, List<NodeConnectorStatistics>>) map;
+ } else {
+ log.error("Cache allocation failed for statisticsmanager.nodeConnectorStatistics in container {}", container.getName());
+ }
+
+ map = clusterContainerService.getCache("statisticsmanager.tableStatistics");
+ if (map != null) {
+ this.tableStatistics = (ConcurrentMap<Node, List<NodeTableStatistics>>) map;
+ } else {
+ log.error("Cache allocation failed for statisticsmanager.tableStatistics in container {}", container.getName());
+ }
+
+ map = clusterContainerService.getCache("statisticsmanager.descriptionStatistics");
+ if (map != null) {
+ this.descriptionStatistics = (ConcurrentMap<Node, NodeDescription>) map;
+ } else {
+ log.error("Cache allocation failed for statisticsmanager.descriptionStatistics in container {}", container.getName());
+ }
+
+ map = clusterContainerService.getCache(TRIGGERS_CACHE);
+ if (map != null) {
+ this.triggers = (ConcurrentMap<Integer, Node>) map;
+ } else {
+ log.error("Cache allocation failed for " + TRIGGERS_CACHE +" in container {}", container.getName());
+ }
}
/**
*/
void init() {
log.debug("INIT called!");
+ allocateCaches();
+ retrieveCaches();
+
}
/**
*/
void start() {
log.debug("START called!");
+ this.triggerExecutor = Executors.newSingleThreadExecutor();
+ }
+
+ /**
+ * Function called after registering the service in OSGi service registry.
+ */
+ void started(){
+ // Retrieve current statistics so we don't have to wait for next refresh
+ ISwitchManager switchManager = (ISwitchManager) ServiceHelper.getInstance(
+ ISwitchManager.class, container.getName(), this);
+ if ((reader != null) && (switchManager != null)) {
+ Set<Node> nodeSet = switchManager.getNodes();
+ for (Node node : nodeSet) {
+ List<FlowOnNode> flows = reader.readAllFlows(node);
+ if (flows != null) {
+ flowStatistics.put(node, flows);
+ }
+ NodeDescription descr = reader.readDescription(node);
+ if (descr != null) {
+ descriptionStatistics.put(node, descr);
+ }
+ List<NodeTableStatistics> tableStats = reader.readNodeTable(node);
+ if (tableStats != null) {
+ tableStatistics.put(node, tableStats);
+ }
+ List<NodeConnectorStatistics> ncStats = reader.readNodeConnectors(node);
+ if (ncStats != null) {
+ nodeConnectorStatistics.put(node, ncStats);
+ }
+ }
+
+ } else {
+ log.trace("Failed to retrieve current statistics. Statistics will not be immediately available!");
+ }
}
/**
*/
void stop() {
log.debug("STOP called!");
+ this.triggerExecutor.shutdownNow();
+ }
+
+ void setClusterContainerService(IClusterContainerServices s) {
+ log.debug("Cluster Service set for Statistics Mgr");
+ this.clusterContainerService = s;
+ }
+
+ void unsetClusterContainerService(IClusterContainerServices s) {
+ if (this.clusterContainerService == s) {
+ log.debug("Cluster Service removed for Statistics Mgr!");
+ this.clusterContainerService = null;
+ }
+ }
+ void setIContainer(IContainer c){
+ container = c;
+ }
+ public void unsetIContainer(IContainer s) {
+ if (this.container == s) {
+ this.container = null;
+ }
}
public void setReaderService(IReadService service) {
}
public void unsetReaderService(IReadService service) {
- log.debug("Got a service UNset request");
+ log.debug("Got a service UNset request {}", service);
this.reader = null;
}
@Override
public List<FlowOnNode> getFlows(Node node) {
- return reader.readAllFlows(node);
+ if (node == null) {
+ return Collections.emptyList();
+ }
+
+ List<FlowOnNode> flowList = new ArrayList<FlowOnNode>();
+ List<FlowOnNode> cachedList = flowStatistics.get(node);
+ if (cachedList != null){
+ flowList.addAll(cachedList);
+ }
+ return flowList;
+ }
+
+ /**
+ * {@inheritDoc}
+ */
+ @Override
+ public List<FlowOnNode> getFlowsNoCache(Node node) {
+ if (node == null) {
+ return Collections.emptyList();
+ }
+ // check if the node is local to this controller
+ ConnectionLocality locality = ConnectionLocality.LOCAL;
+ if(this.connectionManager != null) {
+ locality = this.connectionManager.getLocalityStatus(node);
+ }
+ if (locality == ConnectionLocality.NOT_LOCAL) {
+ // send a trigger to all and wait for either a response or timeout
+ CountDownLatch newLatch = new CountDownLatch(1);
+ CountDownLatch oldLatch = this.latches.putIfAbsent(node, newLatch);
+ this.triggers.put(this.triggerKey.incrementAndGet(), node);
+ try {
+ boolean retStatus;
+ if(oldLatch != null) {
+ retStatus = oldLatch.await(StatisticsManager.latchTimeout, TimeUnit.SECONDS);
+ } else {
+ retStatus = newLatch.await(StatisticsManager.latchTimeout, TimeUnit.SECONDS);
+ }
+ // log the return code as it will give us, if
+ // the latch timed out.
+ log.debug("latch timed out {}", !retStatus);
+ } catch (InterruptedException e) {
+ // log the error and move on
+ log.warn("Waiting for statistics response interrupted", e);
+ // restore the interrupt status
+ // its a good practice to restore the interrupt status
+ // if you are not propagating the InterruptedException
+ Thread.currentThread().interrupt();
+ }
+ // now that the wait is over
+ // remove the latch entry
+ this.latches.remove(node);
+ } else {
+ // the node is local.
+ // call the read service
+ if (this.reader != null) {
+ List<FlowOnNode> flows = reader.nonCachedReadAllFlows(node);
+ if (flows != null) {
+ nodeFlowStatisticsUpdated(node, flows);
+ }
+ }
+ }
+ // at this point we are ready to return the cached value.
+ // this cached value will be up to date with a very high probability
+ // due to what we have done previously ie:- send a trigger for cache update
+ // or refreshed the cache if the node is local.
+ return getFlows(node);
}
@Override
- public Map<Node, List<FlowOnNode>> getFlowStatisticsForFlowList(
- List<FlowEntry> flowList) {
- Map<Node, List<FlowOnNode>> map = new HashMap<Node, List<FlowOnNode>>();
- if (flowList != null) {
- for (FlowEntry entry : flowList) {
- Node node = entry.getNode();
- Flow flow = entry.getFlow();
- List<FlowOnNode> list = (map.containsKey(node)) ? map.get(node)
- : new ArrayList<FlowOnNode>();
- list.add(reader.readFlow(node, flow));
- map.put(node, list);
+ public Map<Node, List<FlowOnNode>> getFlowStatisticsForFlowList(List<FlowEntry> flowList) {
+ Map<Node, List<FlowOnNode>> statMapOutput = new HashMap<Node, List<FlowOnNode>>();
+
+ if (flowList == null || flowList.isEmpty()){
+ return statMapOutput;
+ }
+
+ Node node;
+ // Index FlowEntries' flows by node so we don't traverse entire flow list for each flowEntry
+ Map<Node, Set<Flow>> index = new HashMap<Node, Set<Flow>>();
+ for (FlowEntry flowEntry : flowList) {
+ node = flowEntry.getNode();
+ Set<Flow> set = (index.containsKey(node) ? index.get(node) : new HashSet<Flow>());
+ set.add(flowEntry.getFlow());
+ index.put(node, set);
+ }
+
+ // Iterate over flows per indexed node and add to output
+ for (Entry<Node, Set<Flow>> indexEntry : index.entrySet()) {
+ node = indexEntry.getKey();
+ List<FlowOnNode> flowsPerNode = flowStatistics.get(node);
+
+ if (flowsPerNode != null && !flowsPerNode.isEmpty()){
+ List<FlowOnNode> filteredFlows = statMapOutput.containsKey(node) ?
+ statMapOutput.get(node) : new ArrayList<FlowOnNode>();
+
+ for (FlowOnNode flowOnNode : flowsPerNode) {
+ if (indexEntry.getValue().contains(flowOnNode.getFlow())) {
+ filteredFlows.add(flowOnNode);
+ }
+ }
+ statMapOutput.put(node, filteredFlows);
}
}
- return map;
+ return statMapOutput;
}
@Override
public int getFlowsNumber(Node node) {
- return reader.readAllFlows(node).size();
+ List<FlowOnNode> l;
+ if (node == null || (l = flowStatistics.get(node)) == null){
+ return -1;
+ }
+ return l.size();
}
@Override
public NodeDescription getNodeDescription(Node node) {
- return reader.readDescription(node);
+ if (node == null){
+ return null;
+ }
+ NodeDescription nd = descriptionStatistics.get(node);
+ return nd != null? nd.clone() : null;
}
@Override
- public NodeConnectorStatistics getNodeConnectorStatistics(
- NodeConnector nodeConnector) {
- return reader.readNodeConnector(nodeConnector);
+ public NodeConnectorStatistics getNodeConnectorStatistics(NodeConnector nodeConnector) {
+ if (nodeConnector == null){
+ return null;
+ }
+
+ List<NodeConnectorStatistics> statList = nodeConnectorStatistics.get(nodeConnector.getNode());
+ if (statList != null){
+ for (NodeConnectorStatistics stat : statList) {
+ if (stat.getNodeConnector().equals(nodeConnector)){
+ return stat;
+ }
+ }
+ }
+ return null;
}
@Override
public List<NodeConnectorStatistics> getNodeConnectorStatistics(Node node) {
- return reader.readNodeConnectors(node);
+ if (node == null){
+ return Collections.emptyList();
+ }
+
+ List<NodeConnectorStatistics> statList = new ArrayList<NodeConnectorStatistics>();
+ List<NodeConnectorStatistics> cachedList = nodeConnectorStatistics.get(node);
+ if (cachedList != null) {
+ statList.addAll(cachedList);
+ }
+ return statList;
}
@Override
public NodeTableStatistics getNodeTableStatistics(NodeTable nodeTable) {
- return reader.readNodeTable(nodeTable);
+ if (nodeTable == null){
+ return null;
+ }
+ List<NodeTableStatistics> statList = tableStatistics.get(nodeTable.getNode());
+ if (statList != null){
+ for (NodeTableStatistics stat : statList) {
+ if (stat.getNodeTable().getID().equals(nodeTable.getID())){
+ return stat;
+ }
+ }
+ }
+ return null;
}
@Override
public List<NodeTableStatistics> getNodeTableStatistics(Node node){
- return reader.readNodeTable(node);
+ if (node == null){
+ return Collections.emptyList();
+ }
+ List<NodeTableStatistics> statList = new ArrayList<NodeTableStatistics>();
+ List<NodeTableStatistics> cachedList = tableStatistics.get(node);
+ if (cachedList != null) {
+ statList.addAll(cachedList);
+ }
+ return statList;
+ }
+
+ @Override
+ public void nodeFlowStatisticsUpdated(Node node, List<FlowOnNode> flowStatsList) {
+ // No equality check because duration fields change constantly
+ this.flowStatistics.put(node, flowStatsList);
+ }
+
+ @Override
+ public void nodeConnectorStatisticsUpdated(Node node, List<NodeConnectorStatistics> ncStatsList) {
+ List<NodeConnectorStatistics> currentStat = this.nodeConnectorStatistics.get(node);
+ if (! ncStatsList.equals(currentStat)){
+ this.nodeConnectorStatistics.put(node, ncStatsList);
+ }
+ }
+
+ @Override
+ public void nodeTableStatisticsUpdated(Node node, List<NodeTableStatistics> tableStatsList) {
+ List<NodeTableStatistics> currentStat = this.tableStatistics.get(node);
+ if (! tableStatsList.equals(currentStat)) {
+ this.tableStatistics.put(node, tableStatsList);
+ }
+ }
+
+ @Override
+ public void descriptionStatisticsUpdated(Node node, NodeDescription nodeDescription) {
+ NodeDescription currentDesc = this.descriptionStatistics.get(node);
+ if (! nodeDescription.equals(currentDesc)){
+ this.descriptionStatistics.put(node, nodeDescription);
+ }
+ }
+
+ @Override
+ public void updateNode(Node node, UpdateType type, Set<Property> props) {
+ // If node is removed, clean up stats mappings
+ if (type == UpdateType.REMOVED) {
+ flowStatistics.remove(node);
+ nodeConnectorStatistics.remove(node);
+ tableStatistics.remove(node);
+ descriptionStatistics.remove(node);
+ }
+ }
+
+ @Override
+ public void updateNodeConnector(NodeConnector nodeConnector, UpdateType type, Set<Property> props) {
+ // Not interested in this update
+ }
+
+ public void unsetIConnectionManager(IConnectionManager s) {
+ if (s == this.connectionManager) {
+ this.connectionManager = null;
+ }
+ }
+
+ public void setIConnectionManager(IConnectionManager s) {
+ this.connectionManager = s;
+ }
+
+ @Override
+ public void entryCreated(Object key, String cacheName, boolean originLocal) {
+ /*
+ * Do nothing
+ */
+ }
+
+ @Override
+ public void entryUpdated(Object key, Object new_value, String cacheName, boolean originLocal) {
+ if (originLocal) {
+ /*
+ * Local updates are of no interest
+ */
+ return;
+ }
+ if (cacheName.equals(TRIGGERS_CACHE)) {
+ log.trace("Got a trigger for key {} : value {}", key, new_value);
+ final Node n = (Node) new_value;
+ // check if the node is local to this controller
+ ConnectionLocality locality = ConnectionLocality.NOT_LOCAL;
+ if(this.connectionManager != null) {
+ locality = this.connectionManager.getLocalityStatus(n);
+ }
+ if (locality == ConnectionLocality.LOCAL) {
+ log.trace("trigger for node {} processes locally", n);
+ // delete the trigger and proceed with handling the trigger
+ this.triggers.remove(key);
+ // this is a potentially long running task
+ // off load it from the listener thread
+ Runnable r = new Runnable() {
+ @Override
+ public void run() {
+ // the node is local.
+ // call the read service
+ if (reader != null) {
+ List<FlowOnNode> flows = reader.nonCachedReadAllFlows(n);
+ if (flows != null) {
+ flowStatistics.put(n, flows);
+ }
+ }
+ }
+ };
+ // submit the runnable for execution
+ if(this.triggerExecutor != null) {
+ this.triggerExecutor.execute(r);
+ }
+ }
+ } else if (cacheName.equals(FLOW_STATISTICS_CACHE)) {
+ // flow statistics cache updated
+ // get the node
+ log.trace("Got a flow statistics cache update for key {}", key);
+ // this is a short running task
+ // no need of off loading from the listener thread
+ final Node n = (Node) key;
+ // check if an outstanding trigger exists for this node
+ CountDownLatch l = this.latches.get(n);
+ if(l != null) {
+ // someone was waiting for this update
+ // let him know
+ l.countDown();
+ }
+ }
+ }
+
+ @Override
+ public void entryDeleted(Object key, String cacheName, boolean originLocal) {
+ /*
+ * Do nothing
+ */
}
}