Step 1: Move vm scripts to the right place
[integration/test.git] / test / tools / odl-mdsal-clustering-tests / clustering-performance-test / flow_config_blaster.py
index 52b53eb1e6b93f34a315e385608f8c8b05fd9561..d7e5bb9932a7c59a6232bed388e0dd111a0fc86b 100755 (executable)
@@ -10,6 +10,7 @@ import argparse
 import time
 import threading
 import re
+import copy
 
 import requests
 import netaddr
@@ -51,27 +52,139 @@ class FlowConfigBlaster(object):
     getheaders = {'Accept': 'application/json'}
 
     FLWURL = "restconf/config/opendaylight-inventory:nodes/node/openflow:%d/table/0/flow/%d"
+    TBLURL = "restconf/config/opendaylight-inventory:nodes/node/openflow:%d/table/0"
     INVURL = 'restconf/operational/opendaylight-inventory:nodes'
-
-    ok_total = 0
+    TIMEOUT = 10
 
     flows = {}
 
-    def __init__(self, host, port, ncycles, nthreads, nnodes, nflows, startflow, auth, json_template):
+    # The "built-in" flow template
+    flow_mode_template = {
+        u'flow': [
+            {
+                u'hard-timeout': 65000,
+                u'idle-timeout': 65000,
+                u'cookie_mask': 4294967295,
+                u'flow-name': u'FLOW-NAME-TEMPLATE',
+                u'priority': 2,
+                u'strict': False,
+                u'cookie': 0,
+                u'table_id': 0,
+                u'installHw': False,
+                u'id': u'FLOW-ID-TEMPLATE',
+                u'match': {
+                    u'ipv4-destination': u'0.0.0.0/32',
+                    u'ethernet-match': {
+                        u'ethernet-type': {
+                            u'type': 2048
+                        }
+                    }
+                },
+                u'instructions': {
+                    u'instruction': [
+                        {
+                            u'order': 0,
+                            u'apply-actions': {
+                                u'action': [
+                                    {
+                                        u'drop-action': {},
+                                        u'order': 0
+                                    }
+                                ]
+                            }
+                        }
+                    ]
+                }
+            }
+        ]
+    }
+
+    class FcbStats(object):
+        """
+        FlowConfigBlaster Statistics: a class that stores and further processes
+        statistics collected by Blaster worker threads during their execution.
+        """
+        def __init__(self):
+            self.ok_rqst_rate = Counter(0.0)
+            self.total_rqst_rate = Counter(0.0)
+            self.ok_flow_rate = Counter(0.0)
+            self.total_flow_rate = Counter(0.0)
+            self.ok_rqsts = Counter(0)
+            self.total_rqsts = Counter(0)
+            self.ok_flows = Counter(0)
+            self.total_flows = Counter(0)
+
+        def process_stats(self, rqst_stats, flow_stats, elapsed_time):
+            """
+            Calculates the stats for RESTCONF request and flow programming
+            throughput, and aggregates statistics across all Blaster threads.
+            """
+            ok_rqsts = rqst_stats[200] + rqst_stats[204]
+            total_rqsts = sum(rqst_stats.values())
+            ok_flows = flow_stats[200] + flow_stats[204]
+            total_flows = sum(flow_stats.values())
+
+            ok_rqst_rate = ok_rqsts / elapsed_time
+            total_rqst_rate = total_rqsts / elapsed_time
+            ok_flow_rate = ok_flows / elapsed_time
+            total_flow_rate = total_flows / elapsed_time
+
+            self.ok_rqsts.increment(ok_rqsts)
+            self.total_rqsts.increment(total_rqsts)
+            self.ok_flows.increment(ok_flows)
+            self.total_flows.increment(total_flows)
+
+            self.ok_rqst_rate.increment(ok_rqst_rate)
+            self.total_rqst_rate.increment(total_rqst_rate)
+            self.ok_flow_rate.increment(ok_flow_rate)
+            self.total_flow_rate.increment(total_flow_rate)
+
+            return ok_rqst_rate, total_rqst_rate, ok_flow_rate, total_flow_rate
+
+        def get_ok_rqst_rate(self):
+            return self.ok_rqst_rate.value
+
+        def get_total_rqst_rate(self):
+            return self.total_rqst_rate.value
+
+        def get_ok_flow_rate(self):
+            return self.ok_flow_rate.value
+
+        def get_total_flow_rate(self):
+            return self.total_flow_rate.value
+
+        def get_ok_rqsts(self):
+            return self.ok_rqsts.value
+
+        def get_total_rqsts(self):
+            return self.total_rqsts.value
+
+        def get_ok_flows(self):
+            return self.ok_flows.value
+
+        def get_total_flows(self):
+            return self.total_flows.value
+
+    def __init__(self, host, port, ncycles, nthreads, fpr, nnodes, nflows, startflow, auth, flow_mod_template=None):
         self.host = host
         self.port = port
         self.ncycles = ncycles
         self.nthreads = nthreads
+        self.fpr = fpr
         self.nnodes = nnodes
         self.nflows = nflows
         self.startflow = startflow
         self.auth = auth
 
-        self.json_template = json_template
-        self.url_template = 'http://' + self.host + ":" + self.port + '/' + self.FLWURL
+        if flow_mod_template:
+            self.flow_mode_template = flow_mod_template
+
+        self.post_url_template = 'http://%s:' + self.port + '/' + self.TBLURL
+        self.del_url_template = 'http://%s:' + self.port + '/' + self.FLWURL
 
-        self.ok_rate = Counter(0.0)
-        self.total_rate = Counter(0.0)
+        self.stats = self.FcbStats()
+        self.total_ok_flows = 0
+        self.total_ok_rqsts = 0
 
         self.ip_addr = Counter(int(netaddr.IPAddress('10.0.0.1')) + startflow)
 
@@ -84,16 +197,22 @@ class FlowConfigBlaster(object):
 
     def get_num_nodes(self, session):
         """
-        Determines the number of OF nodes in the connected mininet network. If mininet is not connected, the default
-        number of flows is 16
+        Determines the number of OF nodes in the connected mininet network. If
+        mininet is not connected, the default number of flows is set to 16.
+        :param session: 'requests' session which to use to query the controller
+                        for openflow nodes
+        :return: None
         """
-        inventory_url = 'http://' + self.host + ":" + self.port + '/' + self.INVURL
+        hosts = self.host.split(",")
+        host = hosts[0]
+        inventory_url = 'http://' + host + ":" + self.port + '/' + self.INVURL
         nodes = self.nnodes
 
         if not self.auth:
-            r = session.get(inventory_url, headers=self.getheaders, stream=False)
+            r = session.get(inventory_url, headers=self.getheaders, stream=False, timeout=self.TIMEOUT)
         else:
-            r = session.get(inventory_url, headers=self.getheaders, stream=False, auth=('admin', 'admin'))
+            r = session.get(inventory_url, headers=self.getheaders, stream=False, auth=('admin', 'admin'),
+                            timeout=self.TIMEOUT)
 
         if r.status_code == 200:
             try:
@@ -109,29 +228,83 @@ class FlowConfigBlaster(object):
 
         return nodes
 
-    def add_flow(self, session, node, flow_id, ipaddr):
+    def create_flow_from_template(self, flow_id, ipaddr, node_id):
         """
-        Adds a single flow to the config data store via REST
+        Create a new flow instance from the flow template specified during
+        FlowConfigBlaster instantiation. Flow templates are json-compatible
+        dictionaries that MUST contain elements for flow cookie, flow name,
+        flow id and the destination IPv4 address in the flow match field.
+        :param flow_id: Id for the new flow to create
+        :param ipaddr: IP Address to put into the flow's match
+        :return: The newly created flow instance
         """
-        flow_data = self.json_template % (flow_id, 'TestFlow-%d' % flow_id, 65000, str(flow_id), 65000,
-                                          str(netaddr.IPAddress(ipaddr)))
-        # print flow_data
-        flow_url = self.url_template % (node, flow_id)
+        flow = copy.deepcopy(self.flow_mode_template['flow'][0])
+        flow['cookie'] = flow_id
+        flow['flow-name'] = 'TestFlow-%d' % flow_id
+        flow['id'] = str(flow_id)
+        flow['match']['ipv4-destination'] = '%s/32' % str(netaddr.IPAddress(ipaddr))
+        return flow
+
+    def post_flows(self, session, node, flow_list, flow_count):
+        """
+        Performs a RESTCONF post of flows passed in the 'flow_list' parameters
+        :param session: 'requests' session on which to perform the POST
+        :param node: The ID of the openflow node to which to post the flows
+        :param flow_list: List of flows (in dictionary form) to POST
+        :return: status code from the POST operation
+        """
+        flow_data = self.convert_to_json(flow_list, node)
+
+        hosts = self.host.split(",")
+        host = hosts[flow_count % len(hosts)]
+        flow_url = self.assemble_post_url(host, node)
         # print flow_url
 
         if not self.auth:
-            r = session.put(flow_url, data=flow_data, headers=self.putheaders, stream=False)
+            r = session.post(flow_url, data=flow_data, headers=self.putheaders, stream=False, timeout=self.TIMEOUT)
         else:
-            r = session.put(flow_url, data=flow_data, headers=self.putheaders, stream=False, auth=('admin', 'admin'))
+            r = session.post(flow_url, data=flow_data, headers=self.putheaders, stream=False, auth=('admin', 'admin'),
+                             timeout=self.TIMEOUT)
 
         return r.status_code
 
-    def add_flows(self, start_flow, tid):
+    def assemble_post_url(self, host, node):
+        """
+        Creates url pointing to config dataStore: /nodes/node/<node-id>/table/<table-id>
+        :param host: ip address or host name pointing to controller
+        :param node: id of node (without protocol prefix and colon)
+        :return: url suitable for sending a flow to controller via POST method
+        """
+        return self.post_url_template % (host, node)
+
+    def convert_to_json(self, flow_list, node_id=None):
         """
-        Adds flows into the ODL config space. This function is executed by a worker thread
+        Dumps flows to json form.
+        :param flow_list: list of flows in json friendly structure
+        :param node_id: node identifier of corresponding node
+        :return: string containing plain json
         """
+        fmod = dict(self.flow_mode_template)
+        fmod['flow'] = flow_list
+        flow_data = json.dumps(fmod)
+        # print flow_data
+        return flow_data
 
-        add_res = {200: 0}
+    def add_flows(self, start_flow_id, tid):
+        """
+        Adds flows into the ODL config data store. This function is executed by
+        a worker thread (the Blaster thread). The number of flows created and
+        the batch size (i.e. how many flows will be put into a RESTCONF request)
+        are determined by control parameters initialized when FlowConfigBlaster
+        is created.
+        :param start_flow_id - the ID of the first flow. Each Blaster thread
+                               programs a different set of flows
+        :param tid: Thread ID - used to id the Blaster thread when statistics
+                                for the thread are printed out
+        :return: None
+        """
+        rqst_stats = {200: 0, 204: 0}
+        flow_stats = {200: 0, 204: 0}
 
         s = requests.Session()
 
@@ -140,31 +313,36 @@ class FlowConfigBlaster(object):
         with self.print_lock:
             print '    Thread %d:\n        Adding %d flows on %d nodes' % (tid, self.nflows, n_nodes)
 
+        nflows = 0
         with Timer() as t:
-            for flow in range(self.nflows):
+            while nflows < self.nflows:
                 node_id = randrange(1, n_nodes + 1)
-                flow_id = tid * (self.ncycles * self.nflows) + flow + start_flow + self.startflow
-                self.flows[tid][flow_id] = node_id
-                sts = self.add_flow(s, node_id, flow_id, self.ip_addr.increment())
+                flow_list = []
+                for i in range(self.fpr):
+                    flow_id = tid * (self.ncycles * self.nflows) + nflows + start_flow_id + self.startflow
+                    self.flows[tid][flow_id] = node_id
+                    flow_list.append(self.create_flow_from_template(flow_id, self.ip_addr.increment(), node_id))
+                    nflows += 1
+                    if nflows >= self.nflows:
+                        break
+                sts = self.post_flows(s, node_id, flow_list, nflows)
                 try:
-                    add_res[sts] += 1
+                    rqst_stats[sts] += 1
+                    flow_stats[sts] += len(flow_list)
                 except KeyError:
-                    add_res[sts] = 1
+                    rqst_stats[sts] = 1
+                    flow_stats[sts] = len(flow_list)
 
-        add_time = t.secs
-        add_ok_rate = add_res[200] / add_time
-        add_total_rate = sum(add_res.values()) / add_time
-
-        self.ok_rate.increment(add_ok_rate)
-        self.total_rate.increment(add_total_rate)
+        ok_rps, total_rps, ok_fps, total_fps = self.stats.process_stats(rqst_stats, flow_stats, t.secs)
 
         with self.print_lock:
-            print '    Thread %d: ' % tid
-            print '        Add time: %.2f,' % add_time
-            print '        Add success rate:  %.2f, Add total rate: %.2f' % (add_ok_rate, add_total_rate)
-            print '        Add Results: ',
-            print add_res
-            self.ok_total += add_res[200]
+            print '\n    Thread %d results (ADD): ' % tid
+            print '        Elapsed time: %.2fs,' % t.secs
+            print '        Requests/s: %.2f OK, %.2f Total' % (ok_rps, total_rps)
+            print '        Flows/s:    %.2f OK, %.2f Total' % (ok_fps, total_fps)
+            print '        Stats ({Requests}, {Flows}): ',
+            print rqst_stats,
+            print flow_stats
             self.threads_done += 1
 
         s.close()
@@ -172,25 +350,39 @@ class FlowConfigBlaster(object):
         with self.cond:
             self.cond.notifyAll()
 
-    def delete_flow(self, session, node, flow_id):
+    def delete_flow(self, session, node, flow_id, flow_count):
         """
-        Deletes a single flow from the ODL config data store via REST
+        Deletes a single flow from the ODL config data store using RESTCONF
+        :param session: 'requests' session on which to perform the POST
+        :param node: Id of the openflow node from which to delete the flow
+        :param flow_id: ID of the to-be-deleted flow
+        :return: status code from the DELETE operation
         """
-        flow_url = self.url_template % (node, flow_id)
+
+        hosts = self.host.split(",")
+        host = hosts[flow_count % len(hosts)]
+        flow_url = self.del_url_template % (host, node, flow_id)
+        # print flow_url
 
         if not self.auth:
-            r = session.delete(flow_url, headers=self.getheaders)
+            r = session.delete(flow_url, headers=self.getheaders, timeout=self.TIMEOUT)
         else:
-            r = session.delete(flow_url, headers=self.getheaders, auth=('admin', 'admin'))
+            r = session.delete(flow_url, headers=self.getheaders, auth=('admin', 'admin'), timeout=self.TIMEOUT)
 
         return r.status_code
 
     def delete_flows(self, start_flow, tid):
         """
-        Deletes flow from the ODL config space that have been added using the 'add_flows()' function. This function is
-        executed by a worker thread
+        Deletes flow from the ODL config space that have been added using the
+        'add_flows()' function. This function is executed by a worker thread
+        :param start_flow - the ID of the first flow. Each Blaster thread
+                               deletes a different set of flows
+        :param tid: Thread ID - used to id the Blaster thread when statistics
+                                for the thread are printed out
+        :return:
         """
-        del_res = {200: 0}
+
+        rqst_stats = {200: 0, 204: 0}
 
         s = requests.Session()
         n_nodes = self.get_num_nodes(s)
@@ -201,26 +393,21 @@ class FlowConfigBlaster(object):
         with Timer() as t:
             for flow in range(self.nflows):
                 flow_id = tid * (self.ncycles * self.nflows) + flow + start_flow + self.startflow
-                sts = self.delete_flow(s, self.flows[tid][flow_id], flow_id)
+                sts = self.delete_flow(s, self.flows[tid][flow_id], flow_id, flow)
                 try:
-                    del_res[sts] += 1
+                    rqst_stats[sts] += 1
                 except KeyError:
-                    del_res[sts] = 1
+                    rqst_stats[sts] = 1
 
-        del_time = t.secs
-
-        del_ok_rate = del_res[200] / del_time
-        del_total_rate = sum(del_res.values()) / del_time
-
-        self.ok_rate.increment(del_ok_rate)
-        self.total_rate.increment(del_total_rate)
+        ok_rps, total_rps, ok_fps, total_fps = self.stats.process_stats(rqst_stats, rqst_stats, t.secs)
 
         with self.print_lock:
-            print '    Thread %d: ' % tid
-            print '        Delete time: %.2f,' % del_time
-            print '        Delete success rate:  %.2f, Delete total rate: %.2f' % (del_ok_rate, del_total_rate)
-            print '        Delete Results: ',
-            print del_res
+            print '\n    Thread %d results (DELETE): ' % tid
+            print '        Elapsed time: %.2fs,' % t.secs
+            print '        Requests/s:  %.2f OK,  %.2f Total' % (ok_rps, total_rps)
+            print '        Flows/s:     %.2f OK,  %.2f Total' % (ok_fps, total_fps)
+            print '        Stats ({Requests})',
+            print rqst_stats
             self.threads_done += 1
 
         s.close()
@@ -230,11 +417,18 @@ class FlowConfigBlaster(object):
 
     def run_cycle(self, function):
         """
-        Runs an add or delete cycle. Starts a number of worker threads that each add a bunch of flows. Work is done
-        in context of the worker threads.
+        Runs a flow-add or flow-delete test cycle. Each test consists of a
+        <cycles> test cycles, where <threads> worker (Blaster) threads are
+        started in each test cycle. Each Blaster thread programs <flows>
+        OpenFlow flows into the controller using the controller's RESTCONF API.
+        :param function: Add or delete, determines what test will be executed.
+        :return: None
         """
+        self.total_ok_flows = 0
+        self.total_ok_rqsts = 0
 
         for c in range(self.ncycles):
+            self.stats = self.FcbStats()
             with self.print_lock:
                 print '\nCycle %d:' % c
 
@@ -246,33 +440,40 @@ class FlowConfigBlaster(object):
 
             # Wait for all threads to finish and measure the execution time
             with Timer() as t:
-                while self.threads_done < self.nthreads:
-                    with self.cond:
-                        self.cond.wait()
+                for thread in threads:
+                    thread.join()
 
             with self.print_lock:
-                print '    Total success rate: %.2f, Total rate: %.2f' % (
-                    self.ok_rate.value, self.total_rate.value)
-                measured_rate = (self.nthreads * self.nflows) / t.secs
-                print '    Measured rate:      %.2f (%.2f%% of Total success rate)' % \
-                      (measured_rate, measured_rate / self.total_rate.value * 100)
-                print '    Measured time:      %.2fs' % t.secs
+                print '\n*** Test summary:'
+                print '    Elapsed time:    %.2fs' % t.secs
+                print '    Peak requests/s: %.2f OK, %.2f Total' % (
+                    self.stats.get_ok_rqst_rate(), self.stats.get_total_rqst_rate())
+                print '    Peak flows/s:    %.2f OK, %.2f Total' % (
+                    self.stats.get_ok_flow_rate(), self.stats.get_total_flow_rate())
+                print '    Avg. requests/s: %.2f OK, %.2f Total (%.2f%% of peak total)' % (
+                    self.stats.get_ok_rqsts() / t.secs,
+                    self.stats.get_total_rqsts() / t.secs,
+                    (self.stats.get_total_rqsts() / t.secs * 100) / self.stats.get_total_rqst_rate())
+                print '    Avg. flows/s:    %.2f OK, %.2f Total (%.2f%% of peak total)' % (
+                    self.stats.get_ok_flows() / t.secs,
+                    self.stats.get_total_flows() / t.secs,
+                    (self.stats.get_total_flows() / t.secs * 100) / self.stats.get_total_flow_rate())
+
+                self.total_ok_flows += self.stats.get_ok_flows()
+                self.total_ok_rqsts += self.stats.get_ok_rqsts()
                 self.threads_done = 0
 
-            self.ok_rate.value = 0
-            self.total_rate.value = 0
-
     def add_blaster(self):
         self.run_cycle(self.add_flows)
 
     def delete_blaster(self):
         self.run_cycle(self.delete_flows)
 
-    def get_total_flows(self):
-        return sum(len(self.flows[key]) for key in self.flows.keys())
-
     def get_ok_flows(self):
-        return self.ok_total
+        return self.total_ok_flows
+
+    def get_ok_rqsts(self):
+        return self.total_ok_rqsts
 
 
 def get_json_from_file(filename):
@@ -282,108 +483,150 @@ def get_json_from_file(filename):
     :return: The json flow template (string)
     """
     with open(filename, 'r') as f:
-        read_data = f.read()
-    return read_data
-
-
-if __name__ == "__main__":
-    JSON_FLOW_MOD1 = '''{
-        "flow-node-inventory:flow": [
-            {
-                "flow-node-inventory:cookie": %d,
-                "flow-node-inventory:cookie_mask": 4294967295,
-                "flow-node-inventory:flow-name": "%s",
-                "flow-node-inventory:hard-timeout": %d,
-                "flow-node-inventory:id": "%s",
-                "flow-node-inventory:idle-timeout": %d,
-                "flow-node-inventory:installHw": false,
-                "flow-node-inventory:instructions": {
-                    "flow-node-inventory:instruction": [
-                        {
-                            "flow-node-inventory:apply-actions": {
-                                "flow-node-inventory:action": [
-                                    {
-                                        "flow-node-inventory:drop-action": {},
-                                        "flow-node-inventory:order": 0
-                                    }
-                                ]
-                            },
-                            "flow-node-inventory:order": 0
-                        }
-                    ]
-                },
-                "flow-node-inventory:match": {
-                    "flow-node-inventory:ipv4-destination": "%s/32",
-                    "flow-node-inventory:ethernet-match": {
-                        "flow-node-inventory:ethernet-type": {
-                            "flow-node-inventory:type": 2048
+        try:
+            ft = json.load(f)
+            keys = ft['flow'][0].keys()
+            if (u'cookie' in keys) and (u'flow-name' in keys) and (u'id' in keys) and (u'match' in keys):
+                if u'ipv4-destination' in ft[u'flow'][0]['match'].keys():
+                    print 'File "%s" ok to use as flow template' % filename
+                    return ft
+        except ValueError:
+            print 'JSON parsing of file %s failed' % filename
+            pass
+
+    return None
+
+###############################################################################
+# This is an example of what the content of a JSON flow mode template should
+# look like. Cut & paste to create a custom template. "id" and "ipv4-destination"
+# MUST be unique if multiple flows will be programmed in the same test. It's
+# also beneficial to have unique "cookie" and "flow-name" attributes for easier
+# identification of the flow.
+###############################################################################
+example_flow_mod_json = '''{
+    "flow": [
+        {
+            "id": "38",
+            "cookie": 38,
+            "instructions": {
+                "instruction": [
+                    {
+                        "order": 0,
+                        "apply-actions": {
+                            "action": [
+                                {
+                                    "order": 0,
+                                    "drop-action": { }
+                                }
+                            ]
                         }
                     }
+                ]
+            },
+            "hard-timeout": 65000,
+            "match": {
+                "ethernet-match": {
+                    "ethernet-type": {
+                        "type": 2048
+                    }
                 },
-                "flow-node-inventory:priority": 2,
-                "flow-node-inventory:strict": false,
-                "flow-node-inventory:table_id": 0
-            }
-        ]
-    }'''
-
-    parser = argparse.ArgumentParser(description='Flow programming performance test: First adds and then deletes flows '
-                                                 'into the config tree, as specified by optional parameters.')
-
-    parser.add_argument('--host', default='127.0.0.1',
-                        help='Host where odl controller is running (default is 127.0.0.1)')
-    parser.add_argument('--port', default='8181',
-                        help='Port on which odl\'s RESTCONF is listening (default is 8181)')
-    parser.add_argument('--cycles', type=int, default=1,
-                        help='Number of flow add/delete cycles; default 1. Both Flow Adds and Flow Deletes are '
-                             'performed in cycles. <THREADS> worker threads are started in each cycle and the cycle '
-                             'ends when all threads finish. Another cycle is started when the previous cycle finished.')
-    parser.add_argument('--threads', type=int, default=1,
-                        help='Number of request worker threads to start in each cycle; default=1. '
-                             'Each thread will add/delete <FLOWS> flows.')
-    parser.add_argument('--flows', type=int, default=10,
-                        help='Number of flows that will be added/deleted by each worker thread in each cycle; '
-                             'default 10')
-    parser.add_argument('--nodes', type=int, default=16,
-                        help='Number of nodes if mininet is not connected; default=16. If mininet is connected, '
-                             'flows will be evenly distributed (programmed) into connected nodes.')
-    parser.add_argument('--delay', type=int, default=0,
-                        help='Time (in seconds) to wait between the add and delete cycles; default=0')
-    parser.add_argument('--delete', dest='delete', action='store_true', default=True,
-                        help='Delete all added flows one by one, benchmark delete '
-                             'performance.')
-    parser.add_argument('--no-delete', dest='delete', action='store_false',
-                        help='Do not perform the delete cycle.')
-    parser.add_argument('--auth', dest='auth', action='store_true', default=False,
-                        help="Use the ODL default username/password 'admin'/'admin' to authenticate access to REST; "
-                             'default: no authentication')
-    parser.add_argument('--startflow', type=int, default=0,
-                        help='The starting Flow ID; default=0')
-    parser.add_argument('--file', default='',
-                        help='File from which to read the JSON flow template; default: no file, use a built in '
-                             'template.')
+                "ipv4-destination": "10.0.0.38/32"
+            },
+            "flow-name": "TestFlow-8",
+            "strict": false,
+            "cookie_mask": 4294967295,
+            "priority": 2,
+            "table_id": 0,
+            "idle-timeout": 65000,
+            "installHw": false
+        }
+
+    ]
+}'''
+
+
+def create_arguments_parser():
+    """
+    Shorthand to arg parser on library level in order to access and eventually enhance in ancestors.
+    :return: argument parser supporting config blaster arguments and parameters
+    """
+    my_parser = argparse.ArgumentParser(description='Flow programming performance test: First adds and then'
+                                                    ' deletes flows into the config tree, as specified by'
+                                                    ' optional parameters.')
+
+    my_parser.add_argument('--host', default='127.0.0.1',
+                           help='Host where odl controller is running (default is 127.0.0.1).  '
+                                'Specify a comma-separated list of hosts to perform round-robin load-balancing.')
+    my_parser.add_argument('--port', default='8181',
+                           help='Port on which odl\'s RESTCONF is listening (default is 8181)')
+    my_parser.add_argument('--cycles', type=int, default=1,
+                           help='Number of flow add/delete cycles; default 1. Both Flow Adds and Flow Deletes are '
+                                'performed in cycles. <THREADS> worker threads are started in each cycle and the cycle '
+                                'ends when all threads finish. Another cycle is started when the previous cycle '
+                                'finished.')
+    my_parser.add_argument('--threads', type=int, default=1,
+                           help='Number of request worker threads to start in each cycle; default=1. '
+                                'Each thread will add/delete <FLOWS> flows.')
+    my_parser.add_argument('--flows', type=int, default=10,
+                           help='Number of flows that will be added/deleted by each worker thread in each cycle; '
+                                'default 10')
+    my_parser.add_argument('--fpr', type=int, default=1,
+                           help='Flows-per-Request - number of flows (batch size) sent in each HTTP request; '
+                                'default 1')
+    my_parser.add_argument('--nodes', type=int, default=16,
+                           help='Number of nodes if mininet is not connected; default=16. If mininet is connected, '
+                                'flows will be evenly distributed (programmed) into connected nodes.')
+    my_parser.add_argument('--delay', type=int, default=0,
+                           help='Time (in seconds) to wait between the add and delete cycles; default=0')
+    my_parser.add_argument('--delete', dest='delete', action='store_true', default=True,
+                           help='Delete all added flows one by one, benchmark delete '
+                                'performance.')
+    my_parser.add_argument('--no-delete', dest='delete', action='store_false',
+                           help='Do not perform the delete cycle.')
+    my_parser.add_argument('--auth', dest='auth', action='store_true', default=False,
+                           help="Use the ODL default username/password 'admin'/'admin' to authenticate access to REST; "
+                                'default: no authentication')
+    my_parser.add_argument('--startflow', type=int, default=0,
+                           help='The starting Flow ID; default=0')
+    my_parser.add_argument('--file', default='',
+                           help='File from which to read the JSON flow template; default: no file, use a built in '
+                                'template.')
+    return my_parser
+
 
+if __name__ == "__main__":
+    ############################################################################
+    # This program executes the base performance test. The test adds flows into
+    # the controller's config space. This function is basically the CLI frontend
+    # to the FlowConfigBlaster class and drives its main functions: adding and
+    # deleting flows from the controller's config data store
+    ############################################################################
+
+    parser = create_arguments_parser()
     in_args = parser.parse_args()
 
     if in_args.file != '':
         flow_template = get_json_from_file(in_args.file)
     else:
-        flow_template = JSON_FLOW_MOD1
+        flow_template = None
 
-    fct = FlowConfigBlaster(in_args.host, in_args.port, in_args.cycles, in_args.threads, in_args.nodes,
-                            in_args.flows, in_args.startflow, in_args.auth, flow_template)
+    fct = FlowConfigBlaster(in_args.host, in_args.port, in_args.cycles, in_args.threads, in_args.fpr, in_args.nodes,
+                            in_args.flows, in_args.startflow, in_args.auth)
 
-    # Run through <cycles>, where <threads> are started in each cycle and <flows> are added from each thread
+    # Run through <cycles>, where <threads> are started in each cycle and
+    # <flows> are added from each thread
     fct.add_blaster()
 
-    print '\n*** Total flows added: %s' % fct.get_total_flows()
-    print '    HTTP[OK] results:  %d\n' % fct.get_ok_flows()
+    print '\n*** Total flows added: %s' % fct.get_ok_flows()
+    print '    HTTP[OK] results:  %d\n' % fct.get_ok_rqsts()
 
     if in_args.delay > 0:
         print '*** Waiting for %d seconds before the delete cycle ***\n' % in_args.delay
         time.sleep(in_args.delay)
 
-    # Run through <cycles>, where <threads> are started in each cycle and <flows> previously added in an add cycle are
-    # deleted in each thread
+    # Run through <cycles>, where <threads> are started in each cycle and
+    # <flows> previously added in an add cycle are deleted in each thread
     if in_args.delete:
         fct.delete_blaster()
+        print '\n*** Total flows deleted: %s' % fct.get_ok_flows()
+        print '    HTTP[OK] results:    %d\n' % fct.get_ok_rqsts()