X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=src%2Femuvim%2Fdcemulator%2Fnet.py;h=da546aba58f5575f5b26715dbc6126aeaa9c1581;hb=57551f0497178f51b4cb67b3e21a949cdda8505d;hp=e01e9502a9fd67b05cfe2c65b6d4564c6c9b3aa8;hpb=9524ad314883db5445b27dfd2c61084a7e7329a2;p=osm%2Fvim-emu.git diff --git a/src/emuvim/dcemulator/net.py b/src/emuvim/dcemulator/net.py index e01e950..da546ab 100755 --- a/src/emuvim/dcemulator/net.py +++ b/src/emuvim/dcemulator/net.py @@ -12,11 +12,11 @@ from mininet.net import Dockernet from mininet.node import Controller, OVSSwitch, OVSKernelSwitch, Switch, Docker, Host, RemoteController from mininet.cli import CLI from mininet.log import setLogLevel, info, debug -from mininet.link import TCLink, Link +from mininet.link import TCLink import networkx as nx from emuvim.dcemulator.monitoring import DCNetworkMonitor - from emuvim.dcemulator.node import Datacenter, EmulatorCompute +from emuvim.dcemulator.resourcemodel import ResourceModelRegistrar class DCNetwork(Dockernet): @@ -27,14 +27,20 @@ class DCNetwork(Dockernet): This class is used by topology definition scripts. """ - def __init__(self, **kwargs): + def __init__(self, dc_emulation_max_cpu=1.0, **kwargs): + """ + Create an extended version of a Dockernet network + :param dc_emulation_max_cpu: max. CPU time used by containers in data centers + :param kwargs: path through for Mininet parameters + :return: + """ self.dcs = {} - # create a Mininet/Dockernet network + # call original Docker.__init__ and setup default controller - #Dockernet.__init__( - # self, controller=RemoteController, switch=OVSKernelSwitch, **kwargs) Dockernet.__init__( self, controller=RemoteController, switch=OVSKernelSwitch, **kwargs) + + # ass a remote controller to be able to use Ryu self.addController('c0', controller=RemoteController) # graph of the complete DC network @@ -46,6 +52,8 @@ class DCNetwork(Dockernet): # start Ryu controller self.startRyu() + # initialize resource model registrar + self.rm_registrar = ResourceModelRegistrar(dc_emulation_max_cpu) def addDatacenter(self, label, metadata={}): """ @@ -82,17 +90,22 @@ class DCNetwork(Dockernet): node2 = node2.switch # try to give containers a default IP if isinstance( node1, Docker ): - if not "params1" in params: + if "params1" not in params: params["params1"] = {} - if not "ip" in params["params1"]: + if "ip" not in params["params1"]: params["params1"]["ip"] = self.getNextIp() if isinstance( node2, Docker ): - if not "params2" in params: + if "params2" not in params: params["params2"] = {} - if not "ip" in params["params2"]: + if "ip" not in params["params2"]: params["params2"]["ip"] = self.getNextIp() + # ensure that we allow TCLinks between data centers + # TODO this is not optimal, we use cls=Link for containers and TCLink for data centers + # see Dockernet issue: https://github.com/mpeuster/dockernet/issues/3 + if "cls" not in params: + params["cls"] = TCLink - link = Dockernet.addLink(self, node1, node2, **params) # TODO we need TCLinks with user defined performance here + link = Dockernet.addLink(self, node1, node2, **params) # add edge and assigned port number to graph in both directions between node1 and node2 self.DCNetwork_graph.add_edge(node1.name, node2.name, \