X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=vimconn_vmware.py;h=3e81f5133a7d350b448abbcf4c9934089ead7f84;hb=0a3fc475690c26c63fb61269315dc17e05584b55;hp=19382d67ccfeaec16cb9f448cb323b6432e3d7de;hpb=fda5f7c30615df1edd2f982c18cba17f1200bba4;p=osm%2FRO.git diff --git a/vimconn_vmware.py b/vimconn_vmware.py index 19382d67..3e81f513 100644 --- a/vimconn_vmware.py +++ b/vimconn_vmware.py @@ -177,6 +177,9 @@ class vimconnector(vimconn.vimconnector): self.admin_password = None self.admin_user = None self.org_name = "" + self.nsx_manager = None + self.nsx_user = None + self.nsx_password = None if tenant_name is not None: orgnameandtenant = tenant_name.split(":") @@ -197,6 +200,13 @@ class vimconnector(vimconn.vimconnector): except KeyError: raise vimconn.vimconnException(message="Error admin username or admin password is empty.") + try: + self.nsx_manager = config['nsx_manager'] + self.nsx_user = config['nsx_user'] + self.nsx_password = config['nsx_password'] + except KeyError: + raise vimconn.vimconnException(message="Error: nsx manager or nsx user or nsx password is empty in Config") + self.org_uuid = None self.vca = None @@ -1072,6 +1082,43 @@ class vimconnector(vimconn.vimconnector): return self.get_catalogid(catalog_md5_name, vca.get_catalogs()) + def get_image_list(self, filter_dict={}): + '''Obtain tenant images from VIM + Filter_dict can be: + name: image name + id: image uuid + checksum: image checksum + location: image path + Returns the image list of dictionaries: + [{}, ...] + List can be empty + ''' + vca = self.connect() + if not vca: + raise vimconn.vimconnConnectionException("self.connect() is failed.") + try: + image_list = [] + catalogs = vca.get_catalogs() + if len(catalogs) == 0: + return image_list + else: + for catalog in catalogs: + catalog_uuid = catalog.get_id().split(":")[3] + name = catalog.name + filtered_dict = {} + if filter_dict.get("name") and filter_dict["name"] != name: + continue + if filter_dict.get("id") and filter_dict["id"] != catalog_uuid: + continue + filtered_dict ["name"] = name + filtered_dict ["id"] = catalog_uuid + image_list.append(filtered_dict) + + self.logger.debug("List of already created catalog items: {}".format(image_list)) + return image_list + except Exception as exp: + raise vimconn.vimconnException("Exception occured while retriving catalog items {}".format(exp)) + def get_vappid(self, vdc=None, vapp_name=None): """ Method takes vdc object and vApp name and returns vapp uuid or None @@ -1615,6 +1662,40 @@ class vimconnector(vimconn.vimconnector): """ self.logger.debug("Client requesting refresh vm status for {} ".format(vm_list)) + + mac_ip_addr={} + rheaders = {'Content-Type': 'application/xml'} + iso_edges = ['edge-2','edge-3','edge-6','edge-7','edge-8','edge-9','edge-10'] + + try: + for edge in iso_edges: + nsx_api_url = '/api/4.0/edges/'+ edge +'/dhcp/leaseInfo' + self.logger.debug("refresh_vms_status: NSX Manager url: {}".format(nsx_api_url)) + + resp = requests.get(self.nsx_manager + nsx_api_url, + auth = (self.nsx_user, self.nsx_password), + verify = False, headers = rheaders) + + if resp.status_code == requests.codes.ok: + dhcp_leases = XmlElementTree.fromstring(resp.text) + for child in dhcp_leases: + if child.tag == 'dhcpLeaseInfo': + dhcpLeaseInfo = child + for leaseInfo in dhcpLeaseInfo: + for elem in leaseInfo: + if (elem.tag)=='macAddress': + mac_addr = elem.text + if (elem.tag)=='ipAddress': + ip_addr = elem.text + if (mac_addr) is not None: + mac_ip_addr[mac_addr]= ip_addr + self.logger.debug("NSX Manager DHCP Lease info: mac_ip_addr : {}".format(mac_ip_addr)) + else: + self.logger.debug("Error occurred while getting DHCP lease info from NSX Manager: {}".format(resp.content)) + except KeyError: + self.logger.debug("Error in response from NSX Manager {}".format(KeyError.message)) + self.logger.debug(traceback.format_exc()) + vca = self.connect() if not vca: raise vimconn.vimconnConnectionException("self.connect() is failed.") @@ -1644,6 +1725,10 @@ class vimconnector(vimconn.vimconnector): for vapp_network in vm_app_networks: for vm_network in vapp_network: if vm_network['name'] == vmname: + #Assign IP Address based on MAC Address in NSX DHCP lease info + for mac_adres,ip_adres in mac_ip_addr.iteritems(): + if mac_adres == vm_network['mac']: + vm_network['ip']=ip_adres interface = {"mac_address": vm_network['mac'], "vim_net_id": self.get_network_id_by_name(vm_network['network_name']), "vim_interface_id": self.get_network_id_by_name(vm_network['network_name']), @@ -1685,15 +1770,30 @@ class vimconnector(vimconn.vimconnector): the_vapp = vca.get_vapp(vdc, vapp_name) # TODO fix all status if "start" in action_dict: - if action_dict["start"] == "rebuild": - the_vapp.deploy(powerOn=True) + vm_info = the_vapp.get_vms_details() + vm_status = vm_info[0]['status'] + self.logger.info("Power on vApp: vm_status:{} {}".format(type(vm_status),vm_status)) + if vm_status == "Suspended" or vm_status == "Powered off": + power_on_task = the_vapp.poweron() + if power_on_task is not None and type(power_on_task) is GenericTask: + result = vca.block_until_completed(power_on_task) + if result: + self.logger.info("action_vminstance: Powered on vApp: {}".format(vapp_name)) + else: + self.logger.info("action_vminstance: Failed to power on vApp: {}".format(vapp_name)) + else: + self.logger.info("action_vminstance: Wait for vApp {} to power on".format(vapp_name)) + elif "rebuild" in action_dict: + self.logger.info("action_vminstance: Rebuilding vApp: {}".format(vapp_name)) + power_on_task = the_vapp.deploy(powerOn=True) + if type(power_on_task) is GenericTask: + result = vca.block_until_completed(power_on_task) + if result: + self.logger.info("action_vminstance: Rebuilt vApp: {}".format(vapp_name)) + else: + self.logger.info("action_vminstance: Failed to rebuild vApp: {}".format(vapp_name)) else: - vm_info = the_vapp.get_vms_details() - vm_status = vm_info[0]['status'] - if vm_status == "Suspended": - the_vapp.poweron() - elif vm_status.status == "Powered off": - the_vapp.poweron() + self.logger.info("action_vminstance: Wait for vApp rebuild {} to power on".format(vapp_name)) elif "pause" in action_dict: pass ## server.pause() @@ -1701,7 +1801,15 @@ class vimconnector(vimconn.vimconnector): pass ## server.resume() elif "shutoff" in action_dict or "shutdown" in action_dict: - the_vapp.shutdown() + power_off_task = the_vapp.undeploy(action='powerOff') + if type(power_off_task) is GenericTask: + result = vca.block_until_completed(power_off_task) + if result: + self.logger.info("action_vminstance: Powered off vApp: {}".format(vapp_name)) + else: + self.logger.info("action_vminstance: Failed to power off vApp: {}".format(vapp_name)) + else: + self.logger.info("action_vminstance: Wait for vApp {} to power off".format(vapp_name)) elif "forceOff" in action_dict: the_vapp.reset() elif "terminate" in action_dict: