X-Git-Url: https://git.opendaylight.org/gerrit/gitweb?a=blobdiff_plain;f=test%2Ftools%2Fodl-mdsal-clustering-tests%2Fclustering-performance-test%2Fflow_config_blaster.py;h=d7e5bb9932a7c59a6232bed388e0dd111a0fc86b;hb=072f6e3a8d1bdf8f4c663843589c22d93ba07791;hp=52b53eb1e6b93f34a315e385608f8c8b05fd9561;hpb=3a6e3cfd6c0fa0eb4f6e03f8a8b92bc42191a358;p=integration%2Ftest.git diff --git a/test/tools/odl-mdsal-clustering-tests/clustering-performance-test/flow_config_blaster.py b/test/tools/odl-mdsal-clustering-tests/clustering-performance-test/flow_config_blaster.py index 52b53eb1e6..d7e5bb9932 100755 --- a/test/tools/odl-mdsal-clustering-tests/clustering-performance-test/flow_config_blaster.py +++ b/test/tools/odl-mdsal-clustering-tests/clustering-performance-test/flow_config_blaster.py @@ -10,6 +10,7 @@ import argparse import time import threading import re +import copy import requests import netaddr @@ -51,27 +52,139 @@ class FlowConfigBlaster(object): getheaders = {'Accept': 'application/json'} FLWURL = "restconf/config/opendaylight-inventory:nodes/node/openflow:%d/table/0/flow/%d" + TBLURL = "restconf/config/opendaylight-inventory:nodes/node/openflow:%d/table/0" INVURL = 'restconf/operational/opendaylight-inventory:nodes' - - ok_total = 0 + TIMEOUT = 10 flows = {} - def __init__(self, host, port, ncycles, nthreads, nnodes, nflows, startflow, auth, json_template): + # The "built-in" flow template + flow_mode_template = { + u'flow': [ + { + u'hard-timeout': 65000, + u'idle-timeout': 65000, + u'cookie_mask': 4294967295, + u'flow-name': u'FLOW-NAME-TEMPLATE', + u'priority': 2, + u'strict': False, + u'cookie': 0, + u'table_id': 0, + u'installHw': False, + u'id': u'FLOW-ID-TEMPLATE', + u'match': { + u'ipv4-destination': u'0.0.0.0/32', + u'ethernet-match': { + u'ethernet-type': { + u'type': 2048 + } + } + }, + u'instructions': { + u'instruction': [ + { + u'order': 0, + u'apply-actions': { + u'action': [ + { + u'drop-action': {}, + u'order': 0 + } + ] + } + } + ] + } + } + ] + } + + class FcbStats(object): + """ + FlowConfigBlaster Statistics: a class that stores and further processes + statistics collected by Blaster worker threads during their execution. + """ + def __init__(self): + self.ok_rqst_rate = Counter(0.0) + self.total_rqst_rate = Counter(0.0) + self.ok_flow_rate = Counter(0.0) + self.total_flow_rate = Counter(0.0) + self.ok_rqsts = Counter(0) + self.total_rqsts = Counter(0) + self.ok_flows = Counter(0) + self.total_flows = Counter(0) + + def process_stats(self, rqst_stats, flow_stats, elapsed_time): + """ + Calculates the stats for RESTCONF request and flow programming + throughput, and aggregates statistics across all Blaster threads. + """ + ok_rqsts = rqst_stats[200] + rqst_stats[204] + total_rqsts = sum(rqst_stats.values()) + ok_flows = flow_stats[200] + flow_stats[204] + total_flows = sum(flow_stats.values()) + + ok_rqst_rate = ok_rqsts / elapsed_time + total_rqst_rate = total_rqsts / elapsed_time + ok_flow_rate = ok_flows / elapsed_time + total_flow_rate = total_flows / elapsed_time + + self.ok_rqsts.increment(ok_rqsts) + self.total_rqsts.increment(total_rqsts) + self.ok_flows.increment(ok_flows) + self.total_flows.increment(total_flows) + + self.ok_rqst_rate.increment(ok_rqst_rate) + self.total_rqst_rate.increment(total_rqst_rate) + self.ok_flow_rate.increment(ok_flow_rate) + self.total_flow_rate.increment(total_flow_rate) + + return ok_rqst_rate, total_rqst_rate, ok_flow_rate, total_flow_rate + + def get_ok_rqst_rate(self): + return self.ok_rqst_rate.value + + def get_total_rqst_rate(self): + return self.total_rqst_rate.value + + def get_ok_flow_rate(self): + return self.ok_flow_rate.value + + def get_total_flow_rate(self): + return self.total_flow_rate.value + + def get_ok_rqsts(self): + return self.ok_rqsts.value + + def get_total_rqsts(self): + return self.total_rqsts.value + + def get_ok_flows(self): + return self.ok_flows.value + + def get_total_flows(self): + return self.total_flows.value + + def __init__(self, host, port, ncycles, nthreads, fpr, nnodes, nflows, startflow, auth, flow_mod_template=None): self.host = host self.port = port self.ncycles = ncycles self.nthreads = nthreads + self.fpr = fpr self.nnodes = nnodes self.nflows = nflows self.startflow = startflow self.auth = auth - self.json_template = json_template - self.url_template = 'http://' + self.host + ":" + self.port + '/' + self.FLWURL + if flow_mod_template: + self.flow_mode_template = flow_mod_template + + self.post_url_template = 'http://%s:' + self.port + '/' + self.TBLURL + self.del_url_template = 'http://%s:' + self.port + '/' + self.FLWURL - self.ok_rate = Counter(0.0) - self.total_rate = Counter(0.0) + self.stats = self.FcbStats() + self.total_ok_flows = 0 + self.total_ok_rqsts = 0 self.ip_addr = Counter(int(netaddr.IPAddress('10.0.0.1')) + startflow) @@ -84,16 +197,22 @@ class FlowConfigBlaster(object): def get_num_nodes(self, session): """ - Determines the number of OF nodes in the connected mininet network. If mininet is not connected, the default - number of flows is 16 + Determines the number of OF nodes in the connected mininet network. If + mininet is not connected, the default number of flows is set to 16. + :param session: 'requests' session which to use to query the controller + for openflow nodes + :return: None """ - inventory_url = 'http://' + self.host + ":" + self.port + '/' + self.INVURL + hosts = self.host.split(",") + host = hosts[0] + inventory_url = 'http://' + host + ":" + self.port + '/' + self.INVURL nodes = self.nnodes if not self.auth: - r = session.get(inventory_url, headers=self.getheaders, stream=False) + r = session.get(inventory_url, headers=self.getheaders, stream=False, timeout=self.TIMEOUT) else: - r = session.get(inventory_url, headers=self.getheaders, stream=False, auth=('admin', 'admin')) + r = session.get(inventory_url, headers=self.getheaders, stream=False, auth=('admin', 'admin'), + timeout=self.TIMEOUT) if r.status_code == 200: try: @@ -109,29 +228,83 @@ class FlowConfigBlaster(object): return nodes - def add_flow(self, session, node, flow_id, ipaddr): + def create_flow_from_template(self, flow_id, ipaddr, node_id): """ - Adds a single flow to the config data store via REST + Create a new flow instance from the flow template specified during + FlowConfigBlaster instantiation. Flow templates are json-compatible + dictionaries that MUST contain elements for flow cookie, flow name, + flow id and the destination IPv4 address in the flow match field. + :param flow_id: Id for the new flow to create + :param ipaddr: IP Address to put into the flow's match + :return: The newly created flow instance """ - flow_data = self.json_template % (flow_id, 'TestFlow-%d' % flow_id, 65000, str(flow_id), 65000, - str(netaddr.IPAddress(ipaddr))) - # print flow_data - flow_url = self.url_template % (node, flow_id) + flow = copy.deepcopy(self.flow_mode_template['flow'][0]) + flow['cookie'] = flow_id + flow['flow-name'] = 'TestFlow-%d' % flow_id + flow['id'] = str(flow_id) + flow['match']['ipv4-destination'] = '%s/32' % str(netaddr.IPAddress(ipaddr)) + return flow + + def post_flows(self, session, node, flow_list, flow_count): + """ + Performs a RESTCONF post of flows passed in the 'flow_list' parameters + :param session: 'requests' session on which to perform the POST + :param node: The ID of the openflow node to which to post the flows + :param flow_list: List of flows (in dictionary form) to POST + :return: status code from the POST operation + """ + flow_data = self.convert_to_json(flow_list, node) + + hosts = self.host.split(",") + host = hosts[flow_count % len(hosts)] + flow_url = self.assemble_post_url(host, node) # print flow_url if not self.auth: - r = session.put(flow_url, data=flow_data, headers=self.putheaders, stream=False) + r = session.post(flow_url, data=flow_data, headers=self.putheaders, stream=False, timeout=self.TIMEOUT) else: - r = session.put(flow_url, data=flow_data, headers=self.putheaders, stream=False, auth=('admin', 'admin')) + r = session.post(flow_url, data=flow_data, headers=self.putheaders, stream=False, auth=('admin', 'admin'), + timeout=self.TIMEOUT) return r.status_code - def add_flows(self, start_flow, tid): + def assemble_post_url(self, host, node): + """ + Creates url pointing to config dataStore: /nodes/node//table/ + :param host: ip address or host name pointing to controller + :param node: id of node (without protocol prefix and colon) + :return: url suitable for sending a flow to controller via POST method + """ + return self.post_url_template % (host, node) + + def convert_to_json(self, flow_list, node_id=None): """ - Adds flows into the ODL config space. This function is executed by a worker thread + Dumps flows to json form. + :param flow_list: list of flows in json friendly structure + :param node_id: node identifier of corresponding node + :return: string containing plain json """ + fmod = dict(self.flow_mode_template) + fmod['flow'] = flow_list + flow_data = json.dumps(fmod) + # print flow_data + return flow_data - add_res = {200: 0} + def add_flows(self, start_flow_id, tid): + """ + Adds flows into the ODL config data store. This function is executed by + a worker thread (the Blaster thread). The number of flows created and + the batch size (i.e. how many flows will be put into a RESTCONF request) + are determined by control parameters initialized when FlowConfigBlaster + is created. + :param start_flow_id - the ID of the first flow. Each Blaster thread + programs a different set of flows + :param tid: Thread ID - used to id the Blaster thread when statistics + for the thread are printed out + :return: None + """ + rqst_stats = {200: 0, 204: 0} + flow_stats = {200: 0, 204: 0} s = requests.Session() @@ -140,31 +313,36 @@ class FlowConfigBlaster(object): with self.print_lock: print ' Thread %d:\n Adding %d flows on %d nodes' % (tid, self.nflows, n_nodes) + nflows = 0 with Timer() as t: - for flow in range(self.nflows): + while nflows < self.nflows: node_id = randrange(1, n_nodes + 1) - flow_id = tid * (self.ncycles * self.nflows) + flow + start_flow + self.startflow - self.flows[tid][flow_id] = node_id - sts = self.add_flow(s, node_id, flow_id, self.ip_addr.increment()) + flow_list = [] + for i in range(self.fpr): + flow_id = tid * (self.ncycles * self.nflows) + nflows + start_flow_id + self.startflow + self.flows[tid][flow_id] = node_id + flow_list.append(self.create_flow_from_template(flow_id, self.ip_addr.increment(), node_id)) + nflows += 1 + if nflows >= self.nflows: + break + sts = self.post_flows(s, node_id, flow_list, nflows) try: - add_res[sts] += 1 + rqst_stats[sts] += 1 + flow_stats[sts] += len(flow_list) except KeyError: - add_res[sts] = 1 + rqst_stats[sts] = 1 + flow_stats[sts] = len(flow_list) - add_time = t.secs - add_ok_rate = add_res[200] / add_time - add_total_rate = sum(add_res.values()) / add_time - - self.ok_rate.increment(add_ok_rate) - self.total_rate.increment(add_total_rate) + ok_rps, total_rps, ok_fps, total_fps = self.stats.process_stats(rqst_stats, flow_stats, t.secs) with self.print_lock: - print ' Thread %d: ' % tid - print ' Add time: %.2f,' % add_time - print ' Add success rate: %.2f, Add total rate: %.2f' % (add_ok_rate, add_total_rate) - print ' Add Results: ', - print add_res - self.ok_total += add_res[200] + print '\n Thread %d results (ADD): ' % tid + print ' Elapsed time: %.2fs,' % t.secs + print ' Requests/s: %.2f OK, %.2f Total' % (ok_rps, total_rps) + print ' Flows/s: %.2f OK, %.2f Total' % (ok_fps, total_fps) + print ' Stats ({Requests}, {Flows}): ', + print rqst_stats, + print flow_stats self.threads_done += 1 s.close() @@ -172,25 +350,39 @@ class FlowConfigBlaster(object): with self.cond: self.cond.notifyAll() - def delete_flow(self, session, node, flow_id): + def delete_flow(self, session, node, flow_id, flow_count): """ - Deletes a single flow from the ODL config data store via REST + Deletes a single flow from the ODL config data store using RESTCONF + :param session: 'requests' session on which to perform the POST + :param node: Id of the openflow node from which to delete the flow + :param flow_id: ID of the to-be-deleted flow + :return: status code from the DELETE operation """ - flow_url = self.url_template % (node, flow_id) + + hosts = self.host.split(",") + host = hosts[flow_count % len(hosts)] + flow_url = self.del_url_template % (host, node, flow_id) + # print flow_url if not self.auth: - r = session.delete(flow_url, headers=self.getheaders) + r = session.delete(flow_url, headers=self.getheaders, timeout=self.TIMEOUT) else: - r = session.delete(flow_url, headers=self.getheaders, auth=('admin', 'admin')) + r = session.delete(flow_url, headers=self.getheaders, auth=('admin', 'admin'), timeout=self.TIMEOUT) return r.status_code def delete_flows(self, start_flow, tid): """ - Deletes flow from the ODL config space that have been added using the 'add_flows()' function. This function is - executed by a worker thread + Deletes flow from the ODL config space that have been added using the + 'add_flows()' function. This function is executed by a worker thread + :param start_flow - the ID of the first flow. Each Blaster thread + deletes a different set of flows + :param tid: Thread ID - used to id the Blaster thread when statistics + for the thread are printed out + :return: """ - del_res = {200: 0} + + rqst_stats = {200: 0, 204: 0} s = requests.Session() n_nodes = self.get_num_nodes(s) @@ -201,26 +393,21 @@ class FlowConfigBlaster(object): with Timer() as t: for flow in range(self.nflows): flow_id = tid * (self.ncycles * self.nflows) + flow + start_flow + self.startflow - sts = self.delete_flow(s, self.flows[tid][flow_id], flow_id) + sts = self.delete_flow(s, self.flows[tid][flow_id], flow_id, flow) try: - del_res[sts] += 1 + rqst_stats[sts] += 1 except KeyError: - del_res[sts] = 1 + rqst_stats[sts] = 1 - del_time = t.secs - - del_ok_rate = del_res[200] / del_time - del_total_rate = sum(del_res.values()) / del_time - - self.ok_rate.increment(del_ok_rate) - self.total_rate.increment(del_total_rate) + ok_rps, total_rps, ok_fps, total_fps = self.stats.process_stats(rqst_stats, rqst_stats, t.secs) with self.print_lock: - print ' Thread %d: ' % tid - print ' Delete time: %.2f,' % del_time - print ' Delete success rate: %.2f, Delete total rate: %.2f' % (del_ok_rate, del_total_rate) - print ' Delete Results: ', - print del_res + print '\n Thread %d results (DELETE): ' % tid + print ' Elapsed time: %.2fs,' % t.secs + print ' Requests/s: %.2f OK, %.2f Total' % (ok_rps, total_rps) + print ' Flows/s: %.2f OK, %.2f Total' % (ok_fps, total_fps) + print ' Stats ({Requests})', + print rqst_stats self.threads_done += 1 s.close() @@ -230,11 +417,18 @@ class FlowConfigBlaster(object): def run_cycle(self, function): """ - Runs an add or delete cycle. Starts a number of worker threads that each add a bunch of flows. Work is done - in context of the worker threads. + Runs a flow-add or flow-delete test cycle. Each test consists of a + test cycles, where worker (Blaster) threads are + started in each test cycle. Each Blaster thread programs + OpenFlow flows into the controller using the controller's RESTCONF API. + :param function: Add or delete, determines what test will be executed. + :return: None """ + self.total_ok_flows = 0 + self.total_ok_rqsts = 0 for c in range(self.ncycles): + self.stats = self.FcbStats() with self.print_lock: print '\nCycle %d:' % c @@ -246,33 +440,40 @@ class FlowConfigBlaster(object): # Wait for all threads to finish and measure the execution time with Timer() as t: - while self.threads_done < self.nthreads: - with self.cond: - self.cond.wait() + for thread in threads: + thread.join() with self.print_lock: - print ' Total success rate: %.2f, Total rate: %.2f' % ( - self.ok_rate.value, self.total_rate.value) - measured_rate = (self.nthreads * self.nflows) / t.secs - print ' Measured rate: %.2f (%.2f%% of Total success rate)' % \ - (measured_rate, measured_rate / self.total_rate.value * 100) - print ' Measured time: %.2fs' % t.secs + print '\n*** Test summary:' + print ' Elapsed time: %.2fs' % t.secs + print ' Peak requests/s: %.2f OK, %.2f Total' % ( + self.stats.get_ok_rqst_rate(), self.stats.get_total_rqst_rate()) + print ' Peak flows/s: %.2f OK, %.2f Total' % ( + self.stats.get_ok_flow_rate(), self.stats.get_total_flow_rate()) + print ' Avg. requests/s: %.2f OK, %.2f Total (%.2f%% of peak total)' % ( + self.stats.get_ok_rqsts() / t.secs, + self.stats.get_total_rqsts() / t.secs, + (self.stats.get_total_rqsts() / t.secs * 100) / self.stats.get_total_rqst_rate()) + print ' Avg. flows/s: %.2f OK, %.2f Total (%.2f%% of peak total)' % ( + self.stats.get_ok_flows() / t.secs, + self.stats.get_total_flows() / t.secs, + (self.stats.get_total_flows() / t.secs * 100) / self.stats.get_total_flow_rate()) + + self.total_ok_flows += self.stats.get_ok_flows() + self.total_ok_rqsts += self.stats.get_ok_rqsts() self.threads_done = 0 - self.ok_rate.value = 0 - self.total_rate.value = 0 - def add_blaster(self): self.run_cycle(self.add_flows) def delete_blaster(self): self.run_cycle(self.delete_flows) - def get_total_flows(self): - return sum(len(self.flows[key]) for key in self.flows.keys()) - def get_ok_flows(self): - return self.ok_total + return self.total_ok_flows + + def get_ok_rqsts(self): + return self.total_ok_rqsts def get_json_from_file(filename): @@ -282,108 +483,150 @@ def get_json_from_file(filename): :return: The json flow template (string) """ with open(filename, 'r') as f: - read_data = f.read() - return read_data - - -if __name__ == "__main__": - JSON_FLOW_MOD1 = '''{ - "flow-node-inventory:flow": [ - { - "flow-node-inventory:cookie": %d, - "flow-node-inventory:cookie_mask": 4294967295, - "flow-node-inventory:flow-name": "%s", - "flow-node-inventory:hard-timeout": %d, - "flow-node-inventory:id": "%s", - "flow-node-inventory:idle-timeout": %d, - "flow-node-inventory:installHw": false, - "flow-node-inventory:instructions": { - "flow-node-inventory:instruction": [ - { - "flow-node-inventory:apply-actions": { - "flow-node-inventory:action": [ - { - "flow-node-inventory:drop-action": {}, - "flow-node-inventory:order": 0 - } - ] - }, - "flow-node-inventory:order": 0 - } - ] - }, - "flow-node-inventory:match": { - "flow-node-inventory:ipv4-destination": "%s/32", - "flow-node-inventory:ethernet-match": { - "flow-node-inventory:ethernet-type": { - "flow-node-inventory:type": 2048 + try: + ft = json.load(f) + keys = ft['flow'][0].keys() + if (u'cookie' in keys) and (u'flow-name' in keys) and (u'id' in keys) and (u'match' in keys): + if u'ipv4-destination' in ft[u'flow'][0]['match'].keys(): + print 'File "%s" ok to use as flow template' % filename + return ft + except ValueError: + print 'JSON parsing of file %s failed' % filename + pass + + return None + +############################################################################### +# This is an example of what the content of a JSON flow mode template should +# look like. Cut & paste to create a custom template. "id" and "ipv4-destination" +# MUST be unique if multiple flows will be programmed in the same test. It's +# also beneficial to have unique "cookie" and "flow-name" attributes for easier +# identification of the flow. +############################################################################### +example_flow_mod_json = '''{ + "flow": [ + { + "id": "38", + "cookie": 38, + "instructions": { + "instruction": [ + { + "order": 0, + "apply-actions": { + "action": [ + { + "order": 0, + "drop-action": { } + } + ] } } + ] + }, + "hard-timeout": 65000, + "match": { + "ethernet-match": { + "ethernet-type": { + "type": 2048 + } }, - "flow-node-inventory:priority": 2, - "flow-node-inventory:strict": false, - "flow-node-inventory:table_id": 0 - } - ] - }''' - - parser = argparse.ArgumentParser(description='Flow programming performance test: First adds and then deletes flows ' - 'into the config tree, as specified by optional parameters.') - - parser.add_argument('--host', default='127.0.0.1', - help='Host where odl controller is running (default is 127.0.0.1)') - parser.add_argument('--port', default='8181', - help='Port on which odl\'s RESTCONF is listening (default is 8181)') - parser.add_argument('--cycles', type=int, default=1, - help='Number of flow add/delete cycles; default 1. Both Flow Adds and Flow Deletes are ' - 'performed in cycles. worker threads are started in each cycle and the cycle ' - 'ends when all threads finish. Another cycle is started when the previous cycle finished.') - parser.add_argument('--threads', type=int, default=1, - help='Number of request worker threads to start in each cycle; default=1. ' - 'Each thread will add/delete flows.') - parser.add_argument('--flows', type=int, default=10, - help='Number of flows that will be added/deleted by each worker thread in each cycle; ' - 'default 10') - parser.add_argument('--nodes', type=int, default=16, - help='Number of nodes if mininet is not connected; default=16. If mininet is connected, ' - 'flows will be evenly distributed (programmed) into connected nodes.') - parser.add_argument('--delay', type=int, default=0, - help='Time (in seconds) to wait between the add and delete cycles; default=0') - parser.add_argument('--delete', dest='delete', action='store_true', default=True, - help='Delete all added flows one by one, benchmark delete ' - 'performance.') - parser.add_argument('--no-delete', dest='delete', action='store_false', - help='Do not perform the delete cycle.') - parser.add_argument('--auth', dest='auth', action='store_true', default=False, - help="Use the ODL default username/password 'admin'/'admin' to authenticate access to REST; " - 'default: no authentication') - parser.add_argument('--startflow', type=int, default=0, - help='The starting Flow ID; default=0') - parser.add_argument('--file', default='', - help='File from which to read the JSON flow template; default: no file, use a built in ' - 'template.') + "ipv4-destination": "10.0.0.38/32" + }, + "flow-name": "TestFlow-8", + "strict": false, + "cookie_mask": 4294967295, + "priority": 2, + "table_id": 0, + "idle-timeout": 65000, + "installHw": false + } + + ] +}''' + + +def create_arguments_parser(): + """ + Shorthand to arg parser on library level in order to access and eventually enhance in ancestors. + :return: argument parser supporting config blaster arguments and parameters + """ + my_parser = argparse.ArgumentParser(description='Flow programming performance test: First adds and then' + ' deletes flows into the config tree, as specified by' + ' optional parameters.') + + my_parser.add_argument('--host', default='127.0.0.1', + help='Host where odl controller is running (default is 127.0.0.1). ' + 'Specify a comma-separated list of hosts to perform round-robin load-balancing.') + my_parser.add_argument('--port', default='8181', + help='Port on which odl\'s RESTCONF is listening (default is 8181)') + my_parser.add_argument('--cycles', type=int, default=1, + help='Number of flow add/delete cycles; default 1. Both Flow Adds and Flow Deletes are ' + 'performed in cycles. worker threads are started in each cycle and the cycle ' + 'ends when all threads finish. Another cycle is started when the previous cycle ' + 'finished.') + my_parser.add_argument('--threads', type=int, default=1, + help='Number of request worker threads to start in each cycle; default=1. ' + 'Each thread will add/delete flows.') + my_parser.add_argument('--flows', type=int, default=10, + help='Number of flows that will be added/deleted by each worker thread in each cycle; ' + 'default 10') + my_parser.add_argument('--fpr', type=int, default=1, + help='Flows-per-Request - number of flows (batch size) sent in each HTTP request; ' + 'default 1') + my_parser.add_argument('--nodes', type=int, default=16, + help='Number of nodes if mininet is not connected; default=16. If mininet is connected, ' + 'flows will be evenly distributed (programmed) into connected nodes.') + my_parser.add_argument('--delay', type=int, default=0, + help='Time (in seconds) to wait between the add and delete cycles; default=0') + my_parser.add_argument('--delete', dest='delete', action='store_true', default=True, + help='Delete all added flows one by one, benchmark delete ' + 'performance.') + my_parser.add_argument('--no-delete', dest='delete', action='store_false', + help='Do not perform the delete cycle.') + my_parser.add_argument('--auth', dest='auth', action='store_true', default=False, + help="Use the ODL default username/password 'admin'/'admin' to authenticate access to REST; " + 'default: no authentication') + my_parser.add_argument('--startflow', type=int, default=0, + help='The starting Flow ID; default=0') + my_parser.add_argument('--file', default='', + help='File from which to read the JSON flow template; default: no file, use a built in ' + 'template.') + return my_parser + +if __name__ == "__main__": + ############################################################################ + # This program executes the base performance test. The test adds flows into + # the controller's config space. This function is basically the CLI frontend + # to the FlowConfigBlaster class and drives its main functions: adding and + # deleting flows from the controller's config data store + ############################################################################ + + parser = create_arguments_parser() in_args = parser.parse_args() if in_args.file != '': flow_template = get_json_from_file(in_args.file) else: - flow_template = JSON_FLOW_MOD1 + flow_template = None - fct = FlowConfigBlaster(in_args.host, in_args.port, in_args.cycles, in_args.threads, in_args.nodes, - in_args.flows, in_args.startflow, in_args.auth, flow_template) + fct = FlowConfigBlaster(in_args.host, in_args.port, in_args.cycles, in_args.threads, in_args.fpr, in_args.nodes, + in_args.flows, in_args.startflow, in_args.auth) - # Run through , where are started in each cycle and are added from each thread + # Run through , where are started in each cycle and + # are added from each thread fct.add_blaster() - print '\n*** Total flows added: %s' % fct.get_total_flows() - print ' HTTP[OK] results: %d\n' % fct.get_ok_flows() + print '\n*** Total flows added: %s' % fct.get_ok_flows() + print ' HTTP[OK] results: %d\n' % fct.get_ok_rqsts() if in_args.delay > 0: print '*** Waiting for %d seconds before the delete cycle ***\n' % in_args.delay time.sleep(in_args.delay) - # Run through , where are started in each cycle and previously added in an add cycle are - # deleted in each thread + # Run through , where are started in each cycle and + # previously added in an add cycle are deleted in each thread if in_args.delete: fct.delete_blaster() + print '\n*** Total flows deleted: %s' % fct.get_ok_flows() + print ' HTTP[OK] results: %d\n' % fct.get_ok_rqsts()