X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=src%2Femuvim%2Fdcemulator%2Fnode.py;h=b29d6343368f6d88f68c3422ff823f388219b6f5;hb=0d98d75450be28bba00ebabcd9041b4ae541f964;hp=d9971c89a5575c21eb8d5e14073f1fbe6b571178;hpb=57551f0497178f51b4cb67b3e21a949cdda8505d;p=osm%2Fvim-emu.git diff --git a/src/emuvim/dcemulator/node.py b/src/emuvim/dcemulator/node.py index d9971c8..b29d634 100755 --- a/src/emuvim/dcemulator/node.py +++ b/src/emuvim/dcemulator/node.py @@ -1,10 +1,39 @@ """ -Distributed Cloud Emulator (dcemulator) -(c) 2015 by Manuel Peuster +Copyright (c) 2015 SONATA-NFV and Paderborn University +ALL RIGHTS RESERVED. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. + +Neither the name of the SONATA-NFV [, ANY ADDITIONAL AFFILIATION] +nor the names of its contributors may be used to endorse or promote +products derived from this software without specific prior written +permission. + +This work has been performed in the framework of the SONATA project, +funded by the European Commission under Grant number 671517 through +the Horizon 2020 and 5G-PPP programmes. The authors would like to +acknowledge the contributions of their colleagues of the SONATA +partner consortium (www.sonata-nfv.eu). """ from mininet.node import Docker from mininet.link import Link +from emuvim.dcemulator.resourcemodel import NotEnoughResourcesAvailable import logging +import time +import json + +LOG = logging.getLogger("dcemulator") +LOG.setLevel(logging.DEBUG) DCDPID_BASE = 1000 # start of switch dpid's used for data center switches @@ -13,7 +42,7 @@ DCDPID_BASE = 1000 # start of switch dpid's used for data center switches class EmulatorCompute(Docker): """ Emulator specific compute node class. - Inherits from Dockernet's Docker host class. + Inherits from Containernet's Docker host class. Represents a single container connected to a (logical) data center. We can add emulator specific helper functions to it. @@ -21,10 +50,9 @@ class EmulatorCompute(Docker): def __init__( self, name, dimage, **kwargs): - logging.debug("Create EmulatorCompute instance: %s" % name) self.datacenter = kwargs.get("datacenter") # pointer to current DC self.flavor_name = kwargs.get("flavor_name") - + LOG.debug("Starting compute instance %r in data center %r" % (name, str(self.datacenter))) # call original Docker.__init__ Docker.__init__(self, name, dimage, **kwargs) @@ -34,7 +62,7 @@ class EmulatorCompute(Docker): this compute instance is connected to. """ # format list of tuples (name, Ip, MAC, isUp, status) - return [(str(i), i.IP(), i.MAC(), i.isUp(), i.status()) + return [{'intf_name':str(i), 'ip':i.IP(), 'mac':i.MAC(), 'up':i.isUp(), 'status':i.status()} for i in self.intfList()] def getStatus(self): @@ -44,7 +72,9 @@ class EmulatorCompute(Docker): status = {} status["name"] = self.name status["network"] = self.getNetworkStatus() + status["docker_network"] = self.dcinfo['NetworkSettings']['IPAddress'] status["image"] = self.dimage + status["flavor_name"] = self.flavor_name status["cpu_quota"] = self.cpu_quota status["cpu_period"] = self.cpu_period status["cpu_shares"] = self.cpu_shares @@ -68,7 +98,7 @@ class Datacenter(object): DC_COUNTER = 1 - def __init__(self, label, metadata={}): + def __init__(self, label, metadata={}, resource_log_path=None): self.net = None # DCNetwork to which we belong # each node (DC) has a short internal name used by Mininet # this is caused by Mininets naming limitations for swtiches etc. @@ -78,6 +108,8 @@ class Datacenter(object): self.label = label # dict to store arbitrary metadata (e.g. latitude and longitude) self.metadata = metadata + # path to which resource information should be logged (e.g. for experiments). None = no logging + self.resource_log_path = resource_log_path # first prototype assumes one "bigswitch" per DC self.switch = None # keep track of running containers @@ -103,7 +135,7 @@ class Datacenter(object): """ self.switch = self.net.addSwitch( "%s.s1" % self.name, dpid=hex(self._get_next_dc_dpid())[2:]) - logging.debug("created data center switch: %s" % str(self.switch)) + LOG.debug("created data center switch: %s" % str(self.switch)) def start(self): pass @@ -125,7 +157,7 @@ class Datacenter(object): raise Exception("Container with name %s already exists." % name) # set default parameter if image is None: - image = "ubuntu" + image = "ubuntu:trusty" if network is None: network = {} # {"ip": "10.0.0.254/8"} if isinstance(network, dict): @@ -134,22 +166,33 @@ class Datacenter(object): if len(network) < 1: network.append({}) - # allocate in resource resource model and compute resource limits for new container - if self._resource_model is not None: - # TODO pass resource limits to new container (cf. Dockernet API) Issue #47 - (cpu_limit, mem_limit, disk_limit) = alloc = self._resource_model.allocate(name, flavor_name) - logging.info("Allocation result: %r" % str(alloc)) # create the container d = self.net.addDocker( "%s" % (name), dimage=image, dcmd=command, datacenter=self, - flavor_name=flavor_name) + flavor_name=flavor_name + ) + + # apply resource limits to container if a resource model is defined + if self._resource_model is not None: + try: + self._resource_model.allocate(d) + self._resource_model.write_allocation_log(d, self.resource_log_path) + except NotEnoughResourcesAvailable as ex: + LOG.warning("Allocation of container %r was blocked by resource model." % name) + LOG.info(ex.message) + # ensure that we remove the container + self.net.removeDocker(name) + return None + # connect all given networks + # if no --net option is given, network = [{}], so 1 empty dict in the list + # this results in 1 default interface with a default ip address for nw in network: - # TODO we cannot use TCLink here (see: https://github.com/mpeuster/dockernet/issues/3) - self.net.addLink(d, self.switch, params1=nw, cls=Link) + # TODO we cannot use TCLink here (see: https://github.com/mpeuster/containernet/issues/3) + self.net.addLink(d, self.switch, params1=nw, cls=Link, intfName1=nw.get('id')) # do bookkeeping self.containers[name] = d return d # we might use UUIDs for naming later on @@ -161,13 +204,25 @@ class Datacenter(object): assert name is not None if name not in self.containers: raise Exception("Container with name %s not found." % name) + LOG.debug("Stopping compute instance %r in data center %r" % (name, str(self))) + + # stop the monitored metrics + if self.net.monitor_agent is not None: + self.net.monitor_agent.stop_metric(name) + + # call resource model and free resources + if self._resource_model is not None: + self._resource_model.free(self.containers[name]) + self._resource_model.write_free_log(self.containers[name], self.resource_log_path) + + # remove links self.net.removeLink( link=None, node1=self.containers[name], node2=self.switch) + + # remove container self.net.removeDocker("%s" % (name)) del self.containers[name] - # call resource model and free resources - if self._resource_model is not None: - self._resource_model.free(name) + return True def listCompute(self): @@ -190,9 +245,14 @@ class Datacenter(object): } def assignResourceModel(self, rm): + """ + Assign a resource model to this DC. + :param rm: a BaseResourceModel object + :return: + """ if self._resource_model is not None: raise Exception("There is already an resource model assigned to this DC.") self._resource_model = rm self.net.rm_registrar.register(self, rm) - logging.info("Assigned RM: %r to DC: %r" % (rm, self)) + LOG.info("Assigned RM: %r to DC: %r" % (rm, self))