| """ |
| Distributed Cloud Emulator (dcemulator) |
| (c) 2015 by Manuel Peuster <manuel.peuster@upb.de> |
| """ |
| import logging |
| |
| import site |
| import time |
| from subprocess import Popen |
| import os |
| import re |
| |
| |
| |
| from mininet.net import Dockernet |
| from mininet.node import Controller, DefaultController, OVSSwitch, OVSKernelSwitch, Docker, RemoteController |
| from mininet.cli import CLI |
| from mininet.link import TCLink |
| import networkx as nx |
| from emuvim.dcemulator.monitoring import DCNetworkMonitor |
| from emuvim.dcemulator.node import Datacenter, EmulatorCompute |
| from emuvim.dcemulator.resourcemodel import ResourceModelRegistrar |
| |
| class DCNetwork(Dockernet): |
| """ |
| Wraps the original Mininet/Dockernet class and provides |
| methods to add data centers, switches, etc. |
| |
| This class is used by topology definition scripts. |
| """ |
| |
| def __init__(self, controller=RemoteController, monitor=False, |
| dc_emulation_max_cpu=1.0, # fraction of overall CPU time for emulation |
| dc_emulation_max_mem=512, # emulation max mem in MB |
| **kwargs): |
| """ |
| Create an extended version of a Dockernet network |
| :param dc_emulation_max_cpu: max. CPU time used by containers in data centers |
| :param kwargs: path through for Mininet parameters |
| :return: |
| """ |
| self.dcs = {} |
| |
| # call original Docker.__init__ and setup default controller |
| Dockernet.__init__( |
| self, switch=OVSKernelSwitch, **kwargs) |
| |
| # Ryu management |
| self.ryu_process = None |
| if controller == RemoteController: |
| # start Ryu controller |
| self.startRyu() |
| |
| # add the specified controller |
| self.addController('c0', controller=controller) |
| |
| # graph of the complete DC network |
| self.DCNetwork_graph = nx.MultiDiGraph() |
| |
| # monitoring agent |
| if monitor: |
| self.monitor_agent = DCNetworkMonitor(self) |
| else: |
| self.monitor_agent = None |
| |
| # initialize resource model registrar |
| self.rm_registrar = ResourceModelRegistrar( |
| dc_emulation_max_cpu, dc_emulation_max_mem) |
| |
| def addDatacenter(self, label, metadata={}, resource_log_path=None): |
| """ |
| Create and add a logical cloud data center to the network. |
| """ |
| if label in self.dcs: |
| raise Exception("Data center label already exists: %s" % label) |
| dc = Datacenter(label, metadata=metadata, resource_log_path=resource_log_path) |
| dc.net = self # set reference to network |
| self.dcs[label] = dc |
| dc.create() # finally create the data center in our Mininet instance |
| logging.info("added data center: %s" % label) |
| return dc |
| |
| def addLink(self, node1, node2, **params): |
| """ |
| Able to handle Datacenter objects as link |
| end points. |
| """ |
| assert node1 is not None |
| assert node2 is not None |
| logging.debug("addLink: n1=%s n2=%s" % (str(node1), str(node2))) |
| # ensure type of node1 |
| if isinstance( node1, basestring ): |
| if node1 in self.dcs: |
| node1 = self.dcs[node1].switch |
| if isinstance( node1, Datacenter ): |
| node1 = node1.switch |
| # ensure type of node2 |
| if isinstance( node2, basestring ): |
| if node2 in self.dcs: |
| node2 = self.dcs[node2].switch |
| if isinstance( node2, Datacenter ): |
| node2 = node2.switch |
| # try to give containers a default IP |
| if isinstance( node1, Docker ): |
| if "params1" not in params: |
| params["params1"] = {} |
| if "ip" not in params["params1"]: |
| params["params1"]["ip"] = self.getNextIp() |
| if isinstance( node2, Docker ): |
| if "params2" not in params: |
| params["params2"] = {} |
| if "ip" not in params["params2"]: |
| params["params2"]["ip"] = self.getNextIp() |
| # ensure that we allow TCLinks between data centers |
| # TODO this is not optimal, we use cls=Link for containers and TCLink for data centers |
| # see Dockernet issue: https://github.com/mpeuster/dockernet/issues/3 |
| if "cls" not in params: |
| params["cls"] = TCLink |
| |
| link = Dockernet.addLink(self, node1, node2, **params) |
| |
| # try to give container interfaces a default id |
| node1_port_id = node1.ports[link.intf1] |
| if isinstance(node1, Docker): |
| if "id" in params["params1"]: |
| node1_port_id = params["params1"]["id"] |
| |
| node2_port_id = node2.ports[link.intf2] |
| if isinstance(node2, Docker): |
| if "id" in params["params2"]: |
| node2_port_id = params["params2"]["id"] |
| |
| |
| |
| # add edge and assigned port number to graph in both directions between node1 and node2 |
| # port_id: id given in descriptor (if available, otherwise same as port) |
| # port: portnumber assigned by Dockernet |
| |
| attr_dict = {} |
| # possible weight metrics allowed by TClink class: |
| weight_metrics = ['bw', 'delay', 'jitter', 'loss'] |
| edge_attributes = [p for p in params if p in weight_metrics] |
| for attr in edge_attributes: |
| # if delay: strip ms (need number as weight in graph) |
| match = re.search('([0-9]*\.?[0-9]+)', params[attr]) |
| if match: |
| attr_number = match.group(1) |
| else: |
| attr_number = None |
| attr_dict[attr] = attr_number |
| |
| |
| attr_dict2 = {'src_port_id': node1_port_id, 'src_port': node1.ports[link.intf1], |
| 'dst_port_id': node2_port_id, 'dst_port': node2.ports[link.intf2]} |
| attr_dict2.update(attr_dict) |
| self.DCNetwork_graph.add_edge(node1.name, node2.name, attr_dict=attr_dict2) |
| |
| attr_dict2 = {'src_port_id': node2_port_id, 'src_port': node2.ports[link.intf2], |
| 'dst_port_id': node1_port_id, 'dst_port': node1.ports[link.intf1]} |
| attr_dict2.update(attr_dict) |
| self.DCNetwork_graph.add_edge(node2.name, node1.name, attr_dict=attr_dict2) |
| |
| return link |
| |
| def addDocker( self, label, **params ): |
| """ |
| Wrapper for addDocker method to use custom container class. |
| """ |
| self.DCNetwork_graph.add_node(label) |
| return Dockernet.addDocker(self, label, cls=EmulatorCompute, **params) |
| |
| def removeDocker( self, label, **params ): |
| """ |
| Wrapper for removeDocker method to update graph. |
| """ |
| self.DCNetwork_graph.remove_node(label) |
| return Dockernet.removeDocker(self, label, **params) |
| |
| def addSwitch( self, name, add_to_graph=True, **params ): |
| """ |
| Wrapper for addSwitch method to store switch also in graph. |
| """ |
| if add_to_graph: |
| self.DCNetwork_graph.add_node(name) |
| return Dockernet.addSwitch(self, name, protocols='OpenFlow10,OpenFlow12,OpenFlow13', **params) |
| |
| def getAllContainers(self): |
| """ |
| Returns a list with all containers within all data centers. |
| """ |
| all_containers = [] |
| for dc in self.dcs.itervalues(): |
| all_containers += dc.listCompute() |
| return all_containers |
| |
| def start(self): |
| # start |
| for dc in self.dcs.itervalues(): |
| dc.start() |
| Dockernet.start(self) |
| |
| def stop(self): |
| |
| # stop the monitor agent |
| if self.monitor_agent is not None: |
| self.monitor_agent.stop() |
| |
| # stop emulator net |
| Dockernet.stop(self) |
| |
| # stop Ryu controller |
| self.stopRyu() |
| |
| |
| def CLI(self): |
| CLI(self) |
| |
| # to remove chain do setChain( src, dst, cmd='del-flows') |
| def setChain(self, vnf_src_name, vnf_dst_name, vnf_src_interface=None, vnf_dst_interface=None, cmd='add-flow', weight=None): |
| |
| #check if port is specified (vnf:port) |
| if vnf_src_interface is None: |
| # take first interface by default |
| connected_sw = self.DCNetwork_graph.neighbors(vnf_src_name)[0] |
| link_dict = self.DCNetwork_graph[vnf_src_name][connected_sw] |
| vnf_src_interface = link_dict[0]['src_port_id'] |
| #logging.info('vnf_src_if: {0}'.format(vnf_src_interface)) |
| |
| for connected_sw in self.DCNetwork_graph.neighbors(vnf_src_name): |
| link_dict = self.DCNetwork_graph[vnf_src_name][connected_sw] |
| for link in link_dict: |
| #logging.info("here1: {0},{1}".format(link_dict[link],vnf_src_interface)) |
| if link_dict[link]['src_port_id'] == vnf_src_interface: |
| # found the right link and connected switch |
| #logging.info("conn_sw: {2},{0},{1}".format(link_dict[link]['src_port_id'], vnf_src_interface, connected_sw)) |
| src_sw = connected_sw |
| |
| src_sw_inport = link_dict[link]['dst_port'] |
| break |
| |
| if vnf_dst_interface is None: |
| # take first interface by default |
| connected_sw = self.DCNetwork_graph.neighbors(vnf_dst_name)[0] |
| link_dict = self.DCNetwork_graph[connected_sw][vnf_dst_name] |
| vnf_dst_interface = link_dict[0]['dst_port_id'] |
| |
| vnf_dst_name = vnf_dst_name.split(':')[0] |
| for connected_sw in self.DCNetwork_graph.neighbors(vnf_dst_name): |
| link_dict = self.DCNetwork_graph[connected_sw][vnf_dst_name] |
| for link in link_dict: |
| if link_dict[link]['dst_port_id'] == vnf_dst_interface: |
| # found the right link and connected switch |
| dst_sw = connected_sw |
| dst_sw_outport = link_dict[link]['src_port'] |
| break |
| |
| |
| # get shortest path |
| #path = nx.shortest_path(self.DCNetwork_graph, vnf_src_name, vnf_dst_name) |
| try: |
| # returns the first found shortest path |
| # if all shortest paths are wanted, use: all_shortest_paths |
| path = nx.shortest_path(self.DCNetwork_graph, src_sw, dst_sw, weight=weight) |
| except: |
| logging.info("No path could be found between {0} and {1}".format(vnf_src_name, vnf_dst_name)) |
| return "No path could be found between {0} and {1}".format(vnf_src_name, vnf_dst_name) |
| |
| logging.info("Path between {0} and {1}: {2}".format(vnf_src_name, vnf_dst_name, path)) |
| |
| #current_hop = vnf_src_name |
| current_hop = src_sw |
| switch_inport = src_sw_inport |
| |
| for i in range(0,len(path)): |
| current_node = self.getNodeByName(current_hop) |
| if path.index(current_hop) < len(path)-1: |
| next_hop = path[path.index(current_hop)+1] |
| else: |
| #last switch reached |
| next_hop = vnf_dst_name |
| |
| next_node = self.getNodeByName(next_hop) |
| |
| if next_hop == vnf_dst_name: |
| switch_outport = dst_sw_outport |
| logging.info("end node reached: {0}".format(vnf_dst_name)) |
| elif not isinstance( next_node, OVSSwitch ): |
| logging.info("Next node: {0} is not a switch".format(next_hop)) |
| return "Next node: {0} is not a switch".format(next_hop) |
| else: |
| # take first link between switches by default |
| index_edge_out = 0 |
| switch_outport = self.DCNetwork_graph[current_hop][next_hop][index_edge_out]['src_port'] |
| |
| |
| #logging.info("add flow in switch: {0} in_port: {1} out_port: {2}".format(current_node.name, switch_inport, switch_outport)) |
| # set of entry via ovs-ofctl |
| # TODO use rest API of ryu to set flow entries to correct dpid |
| # TODO this only sets port in to out, no match, so this will give trouble when multiple services are deployed... |
| # TODO need multiple matches to do this (VLAN tags) |
| if isinstance( current_node, OVSSwitch ): |
| match = 'in_port=%s' % switch_inport |
| |
| if cmd=='add-flow': |
| action = 'action=%s' % switch_outport |
| s = ',' |
| ofcmd = s.join([match,action]) |
| elif cmd=='del-flows': |
| ofcmd = match |
| else: |
| ofcmd='' |
| |
| current_node.dpctl(cmd, ofcmd) |
| logging.info("add flow in switch: {0} in_port: {1} out_port: {2}".format(current_node.name, switch_inport, |
| switch_outport)) |
| # take first link between switches by default |
| if isinstance( next_node, OVSSwitch ): |
| switch_inport = self.DCNetwork_graph[current_hop][next_hop][0]['dst_port'] |
| current_hop = next_hop |
| |
| return "path added between {0} and {1}".format(vnf_src_name, vnf_dst_name) |
| #return "destination node: {0} not reached".format(vnf_dst_name) |
| |
| # start Ryu Openflow controller as Remote Controller for the DCNetwork |
| def startRyu(self): |
| # start Ryu controller with rest-API |
| python_install_path = site.getsitepackages()[0] |
| ryu_path = python_install_path + '/ryu/app/simple_switch_13.py' |
| ryu_path2 = python_install_path + '/ryu/app/ofctl_rest.py' |
| # change the default Openflow controller port to 6653 (official IANA-assigned port number), as used by Mininet |
| # Ryu still uses 6633 as default |
| ryu_option = '--ofp-tcp-listen-port' |
| ryu_of_port = '6653' |
| ryu_cmd = 'ryu-manager' |
| FNULL = open("/tmp/ryu.log", 'w') |
| #self.ryu_process = Popen([ryu_cmd, ryu_path, ryu_path2, ryu_option, ryu_of_port], stdout=FNULL, stderr=FNULL) |
| # no learning switch |
| self.ryu_process = Popen([ryu_cmd, ryu_path2, ryu_option, ryu_of_port], stdout=FNULL, stderr=FNULL) |
| time.sleep(1) |
| |
| def stopRyu(self): |
| if self.ryu_process is not None: |
| self.ryu_process.terminate() |
| self.ryu_process.kill() |
| |