--- /dev/null
+#!/usr/bin/env python
+# -*- coding: utf-8 -*-
+
+##
+# Copyright 2015 Telefónica Investigación y Desarrollo, S.A.U.
+# This file is part of openvim
+# All Rights Reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact with: nfvlabs@tid.es
+##
+
+'''
+This module interact with the openvim database,
+It implements general table management
+and complex writings 'transactional' sures,
+that is, or all is changed or nothing
+'''
+
+__author__="Alfonso Tierno"
+__date__ ="$10-jul-2014 12:07:15$"
+
+import MySQLdb as mdb
+import uuid as myUuid
+import auxiliary_functions as af
+import json
+import logging
+from netaddr import IPNetwork, IPSet, IPRange, all_matching_cidrs
+
+HTTP_Bad_Request = 400
+HTTP_Unauthorized = 401
+HTTP_Not_Found = 404
+HTTP_Method_Not_Allowed = 405
+HTTP_Request_Timeout = 408
+HTTP_Conflict = 409
+HTTP_Service_Unavailable = 503
+HTTP_Internal_Server_Error = 500
+
+
+class vim_db():
+ def __init__(self, vlan_range, logger_name= None, debug=None):
+ '''vlan_range must be a tuple (vlan_ini, vlan_end) with available vlan values for networks
+ every dataplane network contain a unique value, regardless of it is used or not
+ '''
+ #initialization
+ self.net_vlan_range = vlan_range
+ self.net_vlan_usedlist = None
+ self.net_vlan_lastused = self.net_vlan_range[0] -1
+ self.debug=debug
+ if logger_name:
+ self.logger_name = logger_name
+ else:
+ self.logger_name = 'openvim.db'
+ self.logger = logging.getLogger(self.logger_name)
+ if debug:
+ self.logger.setLevel( getattr(logging, debug) )
+
+
+ def connect(self, host=None, user=None, passwd=None, database=None):
+ '''Connect to the concrete data base.
+ The first time a valid host, user, passwd and database must be provided,
+ Following calls can skip this parameters
+ '''
+ try:
+ if host is not None: self.host = host
+ if user is not None: self.user = user
+ if passwd is not None: self.passwd = passwd
+ if database is not None: self.database = database
+
+ self.con = mdb.connect(self.host, self.user, self.passwd, self.database)
+ self.logger.debug("connected to DB %s at %s@%s", self.database,self.user, self.host)
+ return 0
+ except mdb.Error as e:
+ self.logger.error("Cannot connect to DB %s at %s@%s Error %d: %s", self.database, self.user, self.host, e.args[0], e.args[1])
+ return -1
+
+ def get_db_version(self):
+ ''' Obtain the database schema version.
+ Return: (negative, text) if error or version 0.0 where schema_version table is missing
+ (version_int, version_text) if ok
+ '''
+ cmd = "SELECT version_int,version,openvim_ver FROM schema_version"
+ for retry_ in range(0,2):
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ rows = self.cur.fetchall()
+ highest_version_int=0
+ highest_version=""
+ #print rows
+ for row in rows: #look for the latest version
+ if row[0]>highest_version_int:
+ highest_version_int, highest_version = row[0:2]
+ return highest_version_int, highest_version
+ except (mdb.Error, AttributeError) as e:
+ self.logger.error("get_db_version DB Exception %d: %s. Command %s",e.args[0], e.args[1], cmd)
+ r,c = self.format_error(e)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def disconnect(self):
+ '''disconnect from the data base'''
+ try:
+ self.con.close()
+ del self.con
+ except mdb.Error as e:
+ self.logger.error("while disconnecting from DB: Error %d: %s",e.args[0], e.args[1])
+ return -1
+ except AttributeError as e: #self.con not defined
+ if e[0][-5:] == "'con'": return -1, "Database internal error, no connection."
+ else: raise
+
+ def format_error(self, e, func, cmd, command=None, extra=None):
+ '''Creates a text error base on the produced exception
+ Params:
+ e: mdb exception
+ func: name of the function that makes the call, for logging purposes
+ cmd: database command that produce the exception
+ command: if the intention is update or delete
+ extra: extra information to add to some commands
+ Return
+ HTTP error in negative, formatted error text
+ '''
+
+ self.logger.error("%s DB Exception %s. Command %s",func, str(e), cmd)
+ if type(e[0]) is str:
+ if e[0][-5:] == "'con'": return -HTTP_Internal_Server_Error, "DB Exception, no connection."
+ else: raise
+ if e.args[0]==2006 or e.args[0]==2013 : #MySQL server has gone away (((or))) Exception 2013: Lost connection to MySQL server during query
+ #reconnect
+ self.connect()
+ return -HTTP_Request_Timeout,"Database reconnection. Try Again"
+ fk=e.args[1].find("foreign key constraint fails")
+ if fk>=0:
+ if command=="update": return -HTTP_Bad_Request, "tenant_id %s not found." % extra
+ elif command=="delete": return -HTTP_Bad_Request, "Resource is not free. There are %s that prevent its deletion." % extra
+ de = e.args[1].find("Duplicate entry")
+ fk = e.args[1].find("for key")
+ uk = e.args[1].find("Unknown column")
+ wc = e.args[1].find("in 'where clause'")
+ fl = e.args[1].find("in 'field list'")
+ #print de, fk, uk, wc,fl
+ if de>=0:
+ if fk>=0: #error 1062
+ return -HTTP_Conflict, "Value %s already in use for %s" % (e.args[1][de+15:fk], e.args[1][fk+7:])
+ if uk>=0:
+ if wc>=0:
+ return -HTTP_Bad_Request, "Field %s cannot be used for filtering" % e.args[1][uk+14:wc]
+ if fl>=0:
+ return -HTTP_Bad_Request, "Field %s does not exist" % e.args[1][uk+14:wc]
+ return -HTTP_Internal_Server_Error, "Database internal Error %d: %s" % (e.args[0], e.args[1])
+
+ def __data2db_format(self, data):
+ '''convert data to database format. If data is None it return the 'Null' text,
+ otherwise it return the text surrounded by quotes ensuring internal quotes are escaped'''
+ if data==None:
+ return 'Null'
+ out=str(data)
+ if "'" not in out:
+ return "'" + out + "'"
+ elif '"' not in out:
+ return '"' + out + '"'
+ else:
+ return json.dumps(out)
+
+ def __get_used_net_vlan(self):
+ #get used from database if needed
+ try:
+ cmd = "SELECT vlan FROM nets WHERE vlan>='%s' ORDER BY vlan LIMIT 25" % self.net_vlan_lastused
+ with self.con:
+ self.cur = self.con.cursor()
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ vlan_tuple = self.cur.fetchall()
+ #convert a tuple of tuples in a list of numbers
+ self.net_vlan_usedlist = []
+ for k in vlan_tuple:
+ self.net_vlan_usedlist.append(k[0])
+ return 0
+ except (mdb.Error, AttributeError) as e:
+ return self.format_error(e, "get_free_net_vlan", cmd)
+
+ def get_free_net_vlan(self):
+ '''obtain a vlan not used in any net'''
+
+ while True:
+ self.logger.debug("net_vlan_lastused:%d net_vlan_range:%d-%d net_vlan_usedlist:%s",
+ self.net_vlan_lastused, self.net_vlan_range[0], self.net_vlan_range[1], str(self.net_vlan_usedlist))
+ self.net_vlan_lastused += 1
+ if self.net_vlan_lastused == self.net_vlan_range[1]:
+ #start from the begining
+ self.net_vlan_lastused = self.net_vlan_range[0]
+ self.net_vlan_usedlist = None
+ if self.net_vlan_usedlist is None \
+ or (len(self.net_vlan_usedlist)>0 and self.net_vlan_lastused >= self.net_vlan_usedlist[-1] and len(self.net_vlan_usedlist)==25):
+ r = self.__get_used_net_vlan()
+ if r<0: return r
+ self.logger.debug("new net_vlan_usedlist %s", str(self.net_vlan_usedlist))
+ if self.net_vlan_lastused in self.net_vlan_usedlist:
+ continue
+ else:
+ return self.net_vlan_lastused
+
+ def get_table(self, **sql_dict):
+ ''' Obtain rows from a table.
+ Atribure sql_dir: dictionary with the following key: value
+ 'SELECT': [list of fields to retrieve] (by default all)
+ 'FROM': string of table name (Mandatory)
+ 'WHERE': dict of key:values, translated to key=value AND ... (Optional)
+ 'WHERE_NOT': dict of key:values, translated to key!=value AND ... (Optional)
+ 'WHERE_OR': dict of key:values, translated to key=value OR ... (Optional)
+ 'WHERE_AND_OR: str 'AND' or 'OR'(by default) mark the priority to 'WHERE AND (WHERE_OR)' or (WHERE) OR WHERE_OR' (Optional)
+ 'LIMIT': limit of number of rows (Optional)
+ 'DISTINCT': make a select distinct to remove repeated elements
+ Return: a list with dictionarys at each row
+ '''
+ #print sql_dict
+ select_ = "SELECT "
+ if sql_dict.get("DISTINCT"):
+ select_ += "DISTINCT "
+ select_ += ("*" if not sql_dict.get('SELECT') else ",".join(map(str,sql_dict['SELECT'])) )
+ #print 'select_', select_
+ from_ = "FROM " + str(sql_dict['FROM'])
+ #print 'from_', from_
+
+ where_and = None
+ where_or = None
+ w = sql_dict.get('WHERE')
+ if w:
+ where_and = " AND ".join(map( lambda x: str(x) + (" is Null" if w[x] is None else "='"+str(w[x])+"'"), w.keys()) )
+ w = sql_dict.get('WHERE_NOT')
+ if w:
+ where_and_not = " AND ".join(map( lambda x: str(x) + (" is not Null" if w[x] is None else "!='"+str(w[x])+"'"), w.keys()) )
+ if where_and:
+ where_and += " AND " + where_and_not
+ else:
+ where_and = where_and_not
+ w = sql_dict.get('WHERE_OR')
+ if w:
+ where_or = " OR ".join(map( lambda x: str(x) + (" is Null" if w[x] is None else "='"+str(w[x])+"'"), w.keys()) )
+
+ if where_and!=None and where_or!=None:
+ if sql_dict.get("WHERE_AND_OR") == "AND":
+ where_ = "WHERE " + where_and + " AND (" + where_or + ")"
+ else:
+ where_ = "WHERE (" + where_and + ") OR " + where_or
+ elif where_and!=None and where_or==None:
+ where_ = "WHERE " + where_and
+ elif where_and==None and where_or!=None:
+ where_ = "WHERE " + where_or
+ else:
+ where_ = ""
+ #print 'where_', where_
+ limit_ = "LIMIT " + str(sql_dict['LIMIT']) if sql_dict.get("LIMIT") else ""
+ #print 'limit_', limit_
+ cmd = " ".join( (select_, from_, where_, limit_) )
+ for retry_ in range(0,2):
+ try:
+ with self.con:
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ rows = self.cur.fetchall()
+ return self.cur.rowcount, rows
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "get_table", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def new_tenant(self, tenant_dict):
+ ''' Add one row into a table.
+ Attribure
+ tenant_dict: dictionary with the key: value to insert
+ It checks presence of uuid and add one automatically otherwise
+ Return: (result, uuid) where result can be 0 if error, or 1 if ok
+ '''
+ for retry_ in range(0,2):
+ cmd=""
+ inserted=-1
+ try:
+ #create uuid if not provided
+ if 'uuid' not in tenant_dict:
+ uuid = tenant_dict['uuid'] = str(myUuid.uuid1()) # create_uuid
+ else:
+ uuid = str(tenant_dict['uuid'])
+ #obtain tenant_id for logs
+ tenant_id = uuid
+ with self.con:
+ self.cur = self.con.cursor()
+ #inserting new uuid
+ cmd = "INSERT INTO uuids (uuid, used_at) VALUES ('%s','tenants')" % uuid
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #insert tenant
+ cmd= "INSERT INTO tenants (" + \
+ ",".join(map(str, tenant_dict.keys() )) + ") VALUES(" + \
+ ",".join(map(lambda x: "Null" if x is None else "'"+str(x)+"'",tenant_dict.values() )) + ")"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ inserted = self.cur.rowcount
+ ##inserting new log
+ #del tenant_dict['uuid'] # not interested for the log
+ #cmd = "INSERT INTO logs (related,level,tenant_id,uuid,description) VALUES ('tenants','debug','%s','%s',\"new tenant %s\")" % (uuid, tenant_id, str(tenant_dict))
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+ #commit transaction
+ self.cur.close()
+ if inserted == 0: return 0, uuid
+ with self.con:
+ self.cur = self.con.cursor()
+ #adding public flavors
+ cmd = "INSERT INTO tenants_flavors(flavor_id,tenant_id) SELECT uuid as flavor_id,'"+ tenant_id + "' FROM flavors WHERE public = 'yes'"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ self.logger.debug("attached public flavors: %s", str(self.cur.rowcount))
+ #rows = self.cur.fetchall()
+ #for row in rows:
+ # cmd = "INSERT INTO tenants_flavors(flavor_id,tenant_id) VALUES('%s','%s')" % (row[0], tenant_id)
+ # self.cur.execute(cmd )
+ #adding public images
+ cmd = "INSERT INTO tenants_images(image_id,tenant_id) SELECT uuid as image_id,'"+ tenant_id + "' FROM images WHERE public = 'yes'"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ self.logger.debug("attached public images: %s", str(self.cur.rowcount))
+ return 1, uuid
+ except (mdb.Error, AttributeError) as e:
+ if inserted==1:
+ self.logger.warning("new_tenant DB Exception %d: %s. Command %s",e.args[0], e.args[1], cmd)
+ return 1, uuid
+ else:
+ r,c = self.format_error(e, "new_tenant", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def new_row(self, table, INSERT, add_uuid=False, log=False):
+ ''' Add one row into a table.
+ Atribure
+ INSERT: dictionary with the key: value to insert
+ table: table where to insert
+ add_uuid: if True, it will crated an uuid key entry at INSERT if not provided
+ It checks presence of uuid and add one automatically otherwise
+ Return: (result, uuid) where result can be 0 if error, or 1 if ok
+ '''
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ if add_uuid:
+ #create uuid if not provided
+ if 'uuid' not in INSERT:
+ uuid = INSERT['uuid'] = str(myUuid.uuid1()) # create_uuid
+ else:
+ uuid = str(INSERT['uuid'])
+ else:
+ uuid=None
+ with self.con:
+ self.cur = self.con.cursor()
+ if add_uuid:
+ #inserting new uuid
+ cmd = "INSERT INTO uuids (uuid, used_at) VALUES ('%s','%s')" % (uuid, table)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #insertion
+ cmd= "INSERT INTO " + table +" (" + \
+ ",".join(map(str, INSERT.keys() )) + ") VALUES(" + \
+ ",".join(map(lambda x: 'Null' if x is None else "'"+str(x)+"'", INSERT.values() )) + ")"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ nb_rows = self.cur.rowcount
+ #inserting new log
+ #if nb_rows > 0 and log:
+ # if add_uuid: del INSERT['uuid']
+ # #obtain tenant_id for logs
+ # if 'tenant_id' in INSERT:
+ # tenant_id = INSERT['tenant_id']
+ # del INSERT['tenant_id']
+ # elif table == 'tenants':
+ # tenant_id = uuid
+ # else:
+ # tenant_id = None
+ # if uuid is None: uuid_k = uuid_v = ""
+ # else: uuid_k=",uuid"; uuid_v=",'" + str(uuid) + "'"
+ # if tenant_id is None: tenant_k = tenant_v = ""
+ # else: tenant_k=",tenant_id"; tenant_v=",'" + str(tenant_id) + "'"
+ # cmd = "INSERT INTO logs (related,level%s%s,description) VALUES ('%s','debug'%s%s,\"new %s %s\")" \
+ # % (uuid_k, tenant_k, table, uuid_v, tenant_v, table[:-1], str(INSERT))
+ # self.logger.debug(cmd)
+ # self.cur.execute(cmd)
+ return nb_rows, uuid
+
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "new_row", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def __remove_quotes(self, data):
+ '''remove single quotes ' of any string content of data dictionary'''
+ for k,v in data.items():
+ if type(v) == str:
+ if "'" in v:
+ data[k] = data[k].replace("'","_")
+
+ def _update_rows_internal(self, table, UPDATE, WHERE={}):
+ cmd= "UPDATE " + table +" SET " + \
+ ",".join(map(lambda x: str(x)+'='+ self.__data2db_format(UPDATE[x]), UPDATE.keys() ));
+ if WHERE:
+ cmd += " WHERE " + " and ".join(map(lambda x: str(x)+ (' is Null' if WHERE[x] is None else"='"+str(WHERE[x])+"'" ), WHERE.keys() ))
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ nb_rows = self.cur.rowcount
+ return nb_rows, None
+
+ def update_rows(self, table, UPDATE, WHERE={}, log=False):
+ ''' Update one or several rows into a table.
+ Atributes
+ UPDATE: dictionary with the key-new_value pairs to change
+ table: table to be modified
+ WHERE: dictionary to filter target rows, key-value
+ log: if true, a log entry is added at logs table
+ Return: (result, None) where result indicates the number of updated files
+ '''
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ #gettting uuid
+ uuid = WHERE.get('uuid')
+
+ with self.con:
+ self.cur = self.con.cursor()
+ cmd= "UPDATE " + table +" SET " + \
+ ",".join(map(lambda x: str(x)+'='+ self.__data2db_format(UPDATE[x]), UPDATE.keys() ));
+ if WHERE:
+ cmd += " WHERE " + " and ".join(map(lambda x: str(x)+ (' is Null' if WHERE[x] is None else"='"+str(WHERE[x])+"'" ), WHERE.keys() ))
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ nb_rows = self.cur.rowcount
+ #if nb_rows > 0 and log:
+ # #inserting new log
+ # if uuid is None: uuid_k = uuid_v = ""
+ # else: uuid_k=",uuid"; uuid_v=",'" + str(uuid) + "'"
+ # cmd = "INSERT INTO logs (related,level%s,description) VALUES ('%s','debug'%s,\"updating %d entry %s\")" \
+ # % (uuid_k, table, uuid_v, nb_rows, (str(UPDATE)).replace('"','-') )
+ # self.logger.debug(cmd)
+ # self.cur.execute(cmd)
+ return nb_rows, uuid
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "update_rows", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def get_host(self, host_id):
+ if af.check_valid_uuid(host_id):
+ where_filter="uuid='" + host_id + "'"
+ else:
+ where_filter="name='" + host_id + "'"
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ #get HOST
+ cmd = "SELECT uuid, user, name, ip_name, description, ranking, admin_state_up, DATE_FORMAT(created_at,'%Y-%m-%dT%H:%i:%s') as created_at \
+ FROM hosts WHERE " + where_filter
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount == 0 :
+ return 0, "host '" + str(host_id) +"'not found."
+ elif self.cur.rowcount > 1 :
+ return 0, "host '" + str(host_id) +"' matches more than one result."
+ host = self.cur.fetchone()
+ host_id = host['uuid']
+ #get numa
+ cmd = "SELECT id, numa_socket, hugepages, memory, admin_state_up FROM numas WHERE host_id = '" + str(host_id) + "'"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ host['numas'] = self.cur.fetchall()
+ for numa in host['numas']:
+ #print "SELECT core_id, instance_id, status, thread_id, v_thread_id FROM resources_core WHERE numa_id = '" + str(numa['id']) + "'"
+ #get cores
+ cmd = "SELECT core_id, instance_id, status, thread_id, v_thread_id FROM resources_core WHERE numa_id = '" + str(numa['id']) + "'"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ numa['cores'] = self.cur.fetchall()
+ for core in numa['cores']:
+ if core['instance_id'] == None: del core['instance_id'], core['v_thread_id']
+ if core['status'] == 'ok': del core['status']
+ #get used memory
+ cmd = "SELECT sum(consumed) as hugepages_consumed FROM resources_mem WHERE numa_id = '" + str(numa['id']) + "' GROUP BY numa_id"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ used = self.cur.fetchone()
+ used_= int(used['hugepages_consumed']) if used != None else 0
+ numa['hugepages_consumed'] = used_
+ #get ports
+ #cmd = "CALL GetPortsFromNuma(%s)'" % str(numa['id'])
+ #self.cur.callproc('GetPortsFromNuma', (numa['id'],) )
+ #every time a Procedure is launched you need to close and open the cursor
+ #under Error 2014: Commands out of sync; you can't run this command now
+ #self.cur.close()
+ #self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ cmd="SELECT Mbps, pci, status, Mbps_used, instance_id, if(id=root_id,'PF','VF') as type_,\
+ switch_port, switch_dpid, mac, source_name\
+ FROM resources_port WHERE numa_id=%d ORDER BY root_id, type_ DESC" % (numa['id'])
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ifaces = self.cur.fetchall()
+ #The SQL query will ensure to have SRIOV interfaces from a port first
+ sriovs=[]
+ Mpbs_consumed = 0
+ numa['interfaces'] = []
+ for iface in ifaces:
+ if not iface["instance_id"]:
+ del iface["instance_id"]
+ if iface['status'] == 'ok':
+ del iface['status']
+ Mpbs_consumed += int(iface["Mbps_used"])
+ del iface["Mbps_used"]
+ if iface["type_"]=='PF':
+ if not iface["switch_dpid"]:
+ del iface["switch_dpid"]
+ if not iface["switch_port"]:
+ del iface["switch_port"]
+ if sriovs:
+ iface["sriovs"] = sriovs
+ if Mpbs_consumed:
+ iface["Mpbs_consumed"] = Mpbs_consumed
+ del iface["type_"]
+ numa['interfaces'].append(iface)
+ sriovs=[]
+ Mpbs_consumed = 0
+ else: #VF, SRIOV
+ del iface["switch_port"]
+ del iface["switch_dpid"]
+ del iface["type_"]
+ del iface["Mbps"]
+ sriovs.append(iface)
+
+ #delete internal field
+ del numa['id']
+ return 1, host
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "get_host", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def new_uuid(self):
+ max_retries=10
+ while max_retries>0:
+ uuid = str( myUuid.uuid1() )
+ if self.check_uuid(uuid)[0] == 0:
+ return uuid
+ max_retries-=1
+ return uuid
+
+ def check_uuid(self, uuid):
+ '''check in the database if this uuid is already present'''
+ try:
+ cmd = "SELECT * FROM uuids where uuid='" + str(uuid) + "'"
+ with self.con:
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ rows = self.cur.fetchall()
+ return self.cur.rowcount, rows
+ except (mdb.Error, AttributeError) as e:
+ return self.format_error(e, "check_uuid", cmd)
+
+ def __get_next_ids(self):
+ '''get next auto increment index of all table in the database'''
+ self.cur.execute("SELECT table_name,AUTO_INCREMENT FROM information_schema.tables WHERE AUTO_INCREMENT IS NOT NULL AND table_schema = DATABASE()")
+ rows = self.cur.fetchall()
+ return self.cur.rowcount, dict(rows)
+
+ def edit_host(self, host_id, host_dict):
+ #get next port index
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+
+ #update table host
+ numa_list = host_dict.pop('numas', () )
+ if host_dict:
+ self._update_rows_internal("hosts", host_dict, {"uuid": host_id})
+
+ where = {"host_id": host_id}
+ for numa_dict in numa_list:
+ where["numa_socket"] = str(numa_dict.pop('numa_socket'))
+ interface_list = numa_dict.pop('interfaces', () )
+ if numa_dict:
+ self._update_rows_internal("numas", numa_dict, where)
+ for interface in interface_list:
+ source_name = str(interface.pop("source_name") )
+ if interface:
+ #get interface id from resources_port
+ cmd= "SELECT rp.id as id FROM resources_port as rp join numas as n on n.id=rp.numa_id join hosts as h on h.uuid=n.host_id " +\
+ "WHERE host_id='%s' and rp.source_name='%s'" %(host_id, source_name)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ row = self.cur.fetchone()
+ if self.cur.rowcount<=0:
+ return -HTTP_Bad_Request, "Interface source_name='%s' from numa_socket='%s' not found" % (source_name, str(where["numa_socket"]))
+ interface_id = row[0]
+ self._update_rows_internal("resources_port", interface, {"root_id": interface_id})
+ return self.get_host(host_id)
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "edit_host", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def new_host(self, host_dict):
+ #get next port index
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+
+ result, next_ids = self.__get_next_ids()
+ #print "next_ids: " + str(next_ids)
+ if result <= 0: return result, "Internal DataBase error getting next id of tables"
+
+ #create uuid if not provided
+ if 'uuid' not in host_dict:
+ uuid = host_dict['uuid'] = str(myUuid.uuid1()) # create_uuid
+ else: #check uuid is valid
+ uuid = str(host_dict['uuid'])
+ # result, data = self.check_uuid(uuid)
+ # if (result == 1):
+ # return -1, "UUID '%s' already in use" % uuid
+
+ #inserting new uuid
+ cmd = "INSERT INTO uuids (uuid, used_at) VALUES ('%s','hosts')" % uuid
+ self.logger.debug(cmd)
+ result = self.cur.execute(cmd)
+
+ #insert in table host
+ numa_list = host_dict.pop('numas', [])
+ #get nonhupages and nonisolated cpus
+ host_dict['RAM']=0
+ host_dict['cpus']=0
+ for numa in numa_list:
+ mem_numa = numa.get('memory', 0) - numa.get('hugepages',0)
+ if mem_numa>0:
+ host_dict['RAM'] += mem_numa
+ for core in numa.get("cores", []):
+ if "status" in core and core["status"]=="noteligible":
+ host_dict['cpus']+=1
+ host_dict['RAM']*=1024 # from GB to MB
+
+ keys = ",".join(host_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", host_dict.values() ) )
+ cmd = "INSERT INTO hosts (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ result = self.cur.execute(cmd)
+ #if result != 1: return -1, "Database Error while inserting at hosts table"
+
+ #insert numas
+ nb_numas = nb_cores = nb_ifaces = 0
+ for numa_dict in numa_list:
+ nb_numas += 1
+ interface_list = numa_dict.pop('interfaces', [])
+ core_list = numa_dict.pop('cores', [])
+ numa_dict['id'] = next_ids['numas']; next_ids['numas'] += 1
+ numa_dict['host_id'] = uuid
+ keys = ",".join(numa_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", numa_dict.values() ) )
+ cmd = "INSERT INTO numas (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ result = self.cur.execute(cmd)
+
+ #insert cores
+ for core_dict in core_list:
+ nb_cores += 1
+ core_dict['numa_id'] = numa_dict['id']
+ keys = ",".join(core_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", core_dict.values() ) )
+ cmd = "INSERT INTO resources_core (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ result = self.cur.execute(cmd)
+
+ #insert ports
+ for port_dict in interface_list:
+ nb_ifaces += 1
+ sriov_list = port_dict.pop('sriovs', [])
+ port_dict['numa_id'] = numa_dict['id']
+ port_dict['id'] = port_dict['root_id'] = next_ids['resources_port']
+ next_ids['resources_port'] += 1
+ switch_port = port_dict.get('switch_port', None)
+ switch_dpid = port_dict.get('switch_dpid', None)
+ keys = ",".join(port_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", port_dict.values() ) )
+ cmd = "INSERT INTO resources_port (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ result = self.cur.execute(cmd)
+
+ #insert sriovs into port table
+ for sriov_dict in sriov_list:
+ sriov_dict['switch_port'] = switch_port
+ sriov_dict['switch_dpid'] = switch_dpid
+ sriov_dict['numa_id'] = port_dict['numa_id']
+ sriov_dict['Mbps'] = port_dict['Mbps']
+ sriov_dict['root_id'] = port_dict['id']
+ sriov_dict['id'] = next_ids['resources_port']
+ if "vlan" in sriov_dict:
+ del sriov_dict["vlan"]
+ next_ids['resources_port'] += 1
+ keys = ",".join(sriov_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", sriov_dict.values() ) )
+ cmd = "INSERT INTO resources_port (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ result = self.cur.execute(cmd)
+
+ #inserting new log
+ #cmd = "INSERT INTO logs (related,level,uuid,description) VALUES ('hosts','debug','%s','new host: %d numas, %d theads, %d ifaces')" % (uuid, nb_numas, nb_cores, nb_ifaces)
+ #self.logger.debug(cmd)
+ #result = self.cur.execute(cmd)
+
+ #inseted ok
+ with self.con:
+ self.cur = self.con.cursor()
+ self.logger.debug("callproc('UpdateSwitchPort', () )")
+ self.cur.callproc('UpdateSwitchPort', () )
+
+ self.logger.debug("getting host '%s'",str(host_dict['uuid']))
+ return self.get_host(host_dict['uuid'])
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "new_host", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def new_flavor(self, flavor_dict, tenant_id ):
+ '''Add new flavor into the database. Create uuid if not provided
+ Atributes
+ flavor_dict: flavor dictionary with the key: value to insert. Must be valid flavors columns
+ tenant_id: if not 'any', it matches this flavor/tenant inserting at tenants_flavors table
+ Return: (result, data) where result can be
+ negative: error at inserting. data contain text
+ 1, inserted, data contain inserted uuid flavor
+ '''
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+
+ #create uuid if not provided
+ if 'uuid' not in flavor_dict:
+ uuid = flavor_dict['uuid'] = str(myUuid.uuid1()) # create_uuid
+ else: #check uuid is valid
+ uuid = str(flavor_dict['uuid'])
+ # result, data = self.check_uuid(uuid)
+ # if (result == 1):
+ # return -1, "UUID '%s' already in use" % uuid
+
+ #inserting new uuid
+ cmd = "INSERT INTO uuids (uuid, used_at) VALUES ('%s','flavors')" % uuid
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+ #insert in table flavor
+ keys = ",".join(flavor_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", flavor_dict.values() ) )
+ cmd = "INSERT INTO flavors (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #if result != 1: return -1, "Database Error while inserting at flavors table"
+
+ #insert tenants_flavors
+ if tenant_id != 'any':
+ cmd = "INSERT INTO tenants_flavors (tenant_id,flavor_id) VALUES ('%s','%s')" % (tenant_id, uuid)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+ #inserting new log
+ #del flavor_dict['uuid']
+ #if 'extended' in flavor_dict: del flavor_dict['extended'] #remove two many information
+ #cmd = "INSERT INTO logs (related,level,uuid, tenant_id, description) VALUES ('flavors','debug','%s','%s',\"new flavor: %s\")" \
+ # % (uuid, tenant_id, str(flavor_dict))
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+
+ #inseted ok
+ return 1, uuid
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "new_flavor", cmd, "update", tenant_id)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def new_image(self, image_dict, tenant_id):
+ '''Add new image into the database. Create uuid if not provided
+ Atributes
+ image_dict: image dictionary with the key: value to insert. Must be valid images columns
+ tenant_id: if not 'any', it matches this image/tenant inserting at tenants_images table
+ Return: (result, data) where result can be
+ negative: error at inserting. data contain text
+ 1, inserted, data contain inserted uuid image
+ '''
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+
+ #create uuid if not provided
+ if 'uuid' not in image_dict:
+ uuid = image_dict['uuid'] = str(myUuid.uuid1()) # create_uuid
+ else: #check uuid is valid
+ uuid = str(image_dict['uuid'])
+ # result, data = self.check_uuid(uuid)
+ # if (result == 1):
+ # return -1, "UUID '%s' already in use" % uuid
+
+ #inserting new uuid
+ cmd = "INSERT INTO uuids (uuid, used_at) VALUES ('%s','images')" % uuid
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+ #insert in table image
+ keys = ",".join(image_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", image_dict.values() ) )
+ cmd = "INSERT INTO images (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #if result != 1: return -1, "Database Error while inserting at images table"
+
+ #insert tenants_images
+ if tenant_id != 'any':
+ cmd = "INSERT INTO tenants_images (tenant_id,image_id) VALUES ('%s','%s')" % (tenant_id, uuid)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+ ##inserting new log
+ #cmd = "INSERT INTO logs (related,level,uuid, tenant_id, description) VALUES ('images','debug','%s','%s',\"new image: %s path: %s\")" % (uuid, tenant_id, image_dict['name'], image_dict['path'])
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+
+ #inseted ok
+ return 1, uuid
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "new_image", cmd, "update", tenant_id)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def delete_image_flavor(self, item_type, item_id, tenant_id):
+ '''deletes an image or flavor from database
+ item_type must be a 'image' or 'flavor'
+ item_id is the uuid
+ tenant_id is the asociated tenant, can be 'any' with means all
+ If tenan_id is not any, it deletes from tenants_images/flavors,
+ which means this image/flavor is used by this tenant, and if success,
+ it tries to delete from images/flavors in case this is not public,
+ that only will success if image is private and not used by other tenants
+ If tenant_id is any, it tries to delete from both tables at the same transaction
+ so that image/flavor is completely deleted from all tenants or nothing
+ '''
+ for retry_ in range(0,2):
+ deleted = -1
+ deleted_item = -1
+ result = (-HTTP_Internal_Server_Error, "internal error")
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+ cmd = "DELETE FROM tenants_%ss WHERE %s_id = '%s'" % (item_type, item_type, item_id)
+ if tenant_id != 'any':
+ cmd += " AND tenant_id = '%s'" % tenant_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ deleted = self.cur.rowcount
+ if tenant_id == 'any': #delete from images/flavors in the SAME transaction
+ cmd = "DELETE FROM %ss WHERE uuid = '%s'" % (item_type, item_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ deleted = self.cur.rowcount
+ if deleted>=1:
+ #delete uuid
+ cmd = "DELETE FROM uuids WHERE uuid = '%s'" % item_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ##inserting new log
+ #cmd = "INSERT INTO logs (related,level,uuid,tenant_id,description) \
+ # VALUES ('%ss','debug','%s','%s','delete %s completely')" % \
+ # (item_type, item_id, tenant_id, item_type)
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+ return deleted, "%s '%s' completely deleted" % (item_type, item_id)
+ return 0, "%s '%s' not found" % (item_type, item_id)
+
+ if deleted == 1:
+ ##inserting new log
+ #cmd = "INSERT INTO logs (related,level,uuid,tenant_id,description) \
+ # VALUES ('%ss','debug','%s','%s','delete %s reference for this tenant')" % \
+ # (item_type, item_id, tenant_id, item_type)
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+
+ #commit transaction
+ self.cur.close()
+ #if tenant!=any delete from images/flavors in OTHER transaction. If fails is because dependencies so that not return error
+ if deleted==1:
+ with self.con:
+ self.cur = self.con.cursor()
+
+ #delete image/flavor if not public
+ cmd = "DELETE FROM %ss WHERE uuid = '%s' AND public = 'no'" % (item_type, item_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ deleted_item = self.cur.rowcount
+ if deleted_item == 1:
+ #delete uuid
+ cmd = "DELETE FROM uuids WHERE uuid = '%s'" % item_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ##inserting new log
+ #cmd = "INSERT INTO logs (related,level,uuid,tenant_id,description) \
+ # VALUES ('%ss','debug','%s','%s','delete %s completely')" % \
+ # (item_type, item_id, tenant_id, item_type)
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+ except (mdb.Error, AttributeError) as e:
+ #print "delete_%s DB Exception %d: %s" % (item_type, e.args[0], e.args[1])
+ if deleted <0:
+ result = self.format_error(e, "delete_"+item_type, cmd, "delete", "servers")
+ finally:
+ if deleted==1:
+ return 1, "%s '%s' from tenant '%s' %sdeleted" % \
+ (item_type, item_id, tenant_id, "completely " if deleted_item==1 else "")
+ elif deleted==0:
+ return 0, "%s '%s' from tenant '%s' not found" % (item_type, item_id, tenant_id)
+ else:
+ if result[0]!=-HTTP_Request_Timeout or retry_==1: return result
+
+ def delete_row(self, table, uuid):
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ #delete host
+ self.cur = self.con.cursor()
+ cmd = "DELETE FROM %s WHERE uuid = '%s'" % (table, uuid)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ deleted = self.cur.rowcount
+ if deleted == 1:
+ #delete uuid
+ if table == 'tenants': tenant_str=uuid
+ else: tenant_str='Null'
+ self.cur = self.con.cursor()
+ cmd = "DELETE FROM uuids WHERE uuid = '%s'" % uuid
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ##inserting new log
+ #cmd = "INSERT INTO logs (related,level,uuid,tenant_id,description) VALUES ('%s','debug','%s','%s','delete %s')" % (table, uuid, tenant_str, table[:-1])
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+ return deleted, table[:-1] + " '%s' %s" %(uuid, "deleted" if deleted==1 else "not found")
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "delete_row", cmd, "delete", 'instances' if table=='hosts' or table=='tenants' else 'dependencies')
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def delete_row_by_key(self, table, key, value):
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ #delete host
+ self.cur = self.con.cursor()
+ cmd = "DELETE FROM %s" % (table)
+ if key!=None:
+ if value!=None:
+ cmd += " WHERE %s = '%s'" % (key, value)
+ else:
+ cmd += " WHERE %s is null" % (key)
+ else: #delete all
+ pass
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ deleted = self.cur.rowcount
+ if deleted < 1:
+ return -1, 'Not found'
+ #delete uuid
+ return 0, deleted
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "delete_row_by_key", cmd, "delete", 'instances' if table=='hosts' or table=='tenants' else 'dependencies')
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def delete_row_by_dict(self, **sql_dict):
+ ''' Deletes rows from a table.
+ Attribute sql_dir: dictionary with the following key: value
+ 'FROM': string of table name (Mandatory)
+ 'WHERE': dict of key:values, translated to key=value AND ... (Optional)
+ 'WHERE_NOT': dict of key:values, translated to key<>value AND ... (Optional)
+ 'WHERE_NOTNULL': (list or tuple of items that must not be null in a where ... (Optional)
+ 'LIMIT': limit of number of rows (Optional)
+ Return: the (number of items deleted, descriptive test) if ok; (negative, descriptive text) if error
+ '''
+ #print sql_dict
+ from_ = "FROM " + str(sql_dict['FROM'])
+ #print 'from_', from_
+ if 'WHERE' in sql_dict and len(sql_dict['WHERE']) > 0:
+ w=sql_dict['WHERE']
+ where_ = "WHERE " + " AND ".join(map( lambda x: str(x) + (" is Null" if w[x] is None else "='"+str(w[x])+"'"), w.keys()) )
+ else: where_ = ""
+ if 'WHERE_NOT' in sql_dict and len(sql_dict['WHERE_NOT']) > 0:
+ w=sql_dict['WHERE_NOT']
+ where_2 = " AND ".join(map( lambda x: str(x) + (" is not Null" if w[x] is None else "<>'"+str(w[x])+"'"), w.keys()) )
+ if len(where_)==0: where_ = "WHERE " + where_2
+ else: where_ = where_ + " AND " + where_2
+ if 'WHERE_NOTNULL' in sql_dict and len(sql_dict['WHERE_NOTNULL']) > 0:
+ w=sql_dict['WHERE_NOTNULL']
+ where_2 = " AND ".join(map( lambda x: str(x) + " is not Null", w) )
+ if len(where_)==0: where_ = "WHERE " + where_2
+ else: where_ = where_ + " AND " + where_2
+ #print 'where_', where_
+ limit_ = "LIMIT " + str(sql_dict['LIMIT']) if 'LIMIT' in sql_dict else ""
+ #print 'limit_', limit_
+ cmd = " ".join( ("DELETE", from_, where_, limit_) )
+ self.logger.debug(cmd)
+ for retry_ in range(0,2):
+ try:
+ with self.con:
+ #delete host
+ self.cur = self.con.cursor()
+ self.cur.execute(cmd)
+ deleted = self.cur.rowcount
+ return deleted, "%d deleted from %s" % (deleted, sql_dict['FROM'][:-1] )
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "delete_row_by_dict", cmd, "delete", 'dependencies')
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+
+ def get_instance(self, instance_id):
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ #get INSTANCE
+ cmd = "SELECT uuid, name, description, progress, host_id, flavor_id, image_id, status, last_error, "\
+ "tenant_id, ram, vcpus, created_at FROM instances WHERE uuid='{}'".format(instance_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount == 0 : return 0, "instance '" + str(instance_id) +"'not found."
+ instance = self.cur.fetchone()
+ #get networks
+ cmd = "SELECT uuid as iface_id, net_id, mac as mac_address, ip_address, name, Mbps as bandwidth, "\
+ "vpci, model FROM ports WHERE (type='instance:bridge' or type='instance:ovs') AND "\
+ "instance_id= '{}'".format(instance_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount > 0 :
+ instance['networks'] = self.cur.fetchall()
+
+ #get extended
+ extended = {}
+ #get devices
+ cmd = "SELECT type, vpci, image_id, xml,dev FROM instance_devices WHERE instance_id = '%s' " % str(instance_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount > 0 :
+ extended['devices'] = self.cur.fetchall()
+ #get numas
+ numas = []
+ cmd = "SELECT id, numa_socket as source FROM numas WHERE host_id = '" + str(instance['host_id']) + "'"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ host_numas = self.cur.fetchall()
+ #print 'host_numas', host_numas
+ for k in host_numas:
+ numa_id = str(k['id'])
+ numa_dict ={}
+ #get memory
+ cmd = "SELECT consumed FROM resources_mem WHERE instance_id = '%s' AND numa_id = '%s'" % ( instance_id, numa_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount > 0:
+ mem_dict = self.cur.fetchone()
+ numa_dict['memory'] = mem_dict['consumed']
+ #get full cores
+ cursor2 = self.con.cursor()
+ cmd = "SELECT core_id, paired, MIN(v_thread_id) as v1, MAX(v_thread_id) as v2, COUNT(instance_id) as nb, MIN(thread_id) as t1, MAX(thread_id) as t2 FROM resources_core WHERE instance_id = '%s' AND numa_id = '%s' GROUP BY core_id,paired" % ( str(instance_id), numa_id)
+ self.logger.debug(cmd)
+ cursor2.execute(cmd)
+ core_list = []; core_source = []
+ paired_list = []; paired_source = []
+ thread_list = []; thread_source = []
+ if cursor2.rowcount > 0:
+ cores = cursor2.fetchall()
+ for core in cores:
+ if core[4] == 2: #number of used threads from core
+ if core[3] == core[2]: #only one thread asigned to VM, so completely core
+ core_list.append(core[2])
+ core_source.append(core[5])
+ elif core[1] == 'Y':
+ paired_list.append(core[2:4])
+ paired_source.append(core[5:7])
+ else:
+ thread_list.extend(core[2:4])
+ thread_source.extend(core[5:7])
+
+ else:
+ thread_list.append(core[2])
+ thread_source.append(core[5])
+ if len(core_list) > 0:
+ numa_dict['cores'] = len(core_list)
+ numa_dict['cores-id'] = core_list
+ numa_dict['cores-source'] = core_source
+ if len(paired_list) > 0:
+ numa_dict['paired-threads'] = len(paired_list)
+ numa_dict['paired-threads-id'] = paired_list
+ numa_dict['paired-threads-source'] = paired_source
+ if len(thread_list) > 0:
+ numa_dict['threads'] = len(thread_list)
+ numa_dict['threads-id'] = thread_list
+ numa_dict['threads-source'] = thread_source
+
+ #get dedicated ports and SRIOV
+ cmd = "SELECT port_id as iface_id, p.vlan as vlan, p.mac as mac_address, net_id, if(model='PF',\
+ 'yes',if(model='VF','no','yes:sriov')) as dedicated, rp.Mbps as bandwidth, name, vpci, \
+ pci as source \
+ FROM resources_port as rp join ports as p on port_id=uuid WHERE p.instance_id = '%s' AND numa_id = '%s' and p.type='instance:data'" % (instance_id, numa_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount > 0:
+ numa_dict['interfaces'] = self.cur.fetchall()
+ #print 'interfaces', numa_dict
+
+ if len(numa_dict) > 0 :
+ numa_dict['source'] = k['source'] #numa socket
+ numas.append(numa_dict)
+
+ if len(numas) > 0 : extended['numas'] = numas
+ if len(extended) > 0 : instance['extended'] = extended
+ af.DeleteNone(instance)
+ return 1, instance
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "get_instance", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def get_numas(self, requirements, prefered_host_id=None, only_of_ports=True):
+ '''Obtain a valid NUMA/HOST for deployment a VM
+ requirements: contain requirement regarding:
+ requirements['ram']: Non huge page memory in MB; 0 to skip
+ requirements['vcpus']: Non isolated cpus; 0 to skip
+ requirements['numa']: Requiremets to be fixed in ONE Numa node
+ requirements['numa']['memory']: Huge page memory in GB at ; 0 for any
+ requirements['numa']['proc_req_type']: Type of processor, cores or threads
+ requirements['numa']['proc_req_nb']: Number of isolated cpus
+ requirements['numa']['port_list']: Physical NIC ports list ; [] for any
+ requirements['numa']['sriov_list']: Virtual function NIC ports list ; [] for any
+ prefered_host_id: if not None return this host if it match
+ only_of_ports: if True only those ports conected to the openflow (of) are valid,
+ that is, with switch_port information filled; if False, all NIC ports are valid.
+ Return a valid numa and host
+ '''
+
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+# #Find numas of prefered host
+# prefered_numas = ()
+# if prefered_host_id != None:
+# self.cur = self.con.cursor()
+# self.cur.execute("SELECT id FROM numas WHERE host_id='%s'" + prefered_host_id)
+# prefered_numas = self.cur.fetchall()
+# self.cur.close()
+
+ #Find valid host for the ram and vcpus
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ cmd = "CALL GetHostByMemCpu(%s, %s)" % (str(requirements['ram']), str(requirements['vcpus']))
+ self.logger.debug(cmd)
+ self.cur.callproc('GetHostByMemCpu', (str(requirements['ram']), str(requirements['vcpus'])) )
+ valid_hosts = self.cur.fetchall()
+ self.cur.close()
+ self.cur = self.con.cursor()
+ match_found = False
+ if len(valid_hosts)<=0:
+ error_text = 'No room at data center. Cannot find a host with %s MB memory and %s cpus available' % (str(requirements['ram']), str(requirements['vcpus']))
+ #self.logger.debug(error_text)
+ return -1, error_text
+
+ #elif req_numa != None:
+ #Find valid numa nodes for memory requirements
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ cmd = "CALL GetNumaByMemory(%s)" % str(requirements['numa']['memory'])
+ self.logger.debug(cmd)
+ self.cur.callproc('GetNumaByMemory', (requirements['numa']['memory'],) )
+ valid_for_memory = self.cur.fetchall()
+ self.cur.close()
+ self.cur = self.con.cursor()
+ if len(valid_for_memory)<=0:
+ error_text = 'No room at data center. Cannot find a host with %s GB Hugepages memory available' % str(requirements['numa']['memory'])
+ #self.logger.debug(error_text)
+ return -1, error_text
+
+ #Find valid numa nodes for processor requirements
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ if requirements['numa']['proc_req_type'] == 'threads':
+ cpu_requirement_text='cpu-threads'
+ cmd = "CALL GetNumaByThread(%s)" % str(requirements['numa']['proc_req_nb'])
+ self.logger.debug(cmd)
+ self.cur.callproc('GetNumaByThread', (requirements['numa']['proc_req_nb'],) )
+ else:
+ cpu_requirement_text='cpu-cores'
+ cmd = "CALL GetNumaByCore(%s)" % str(requirements['numa']['proc_req_nb'])
+ self.logger.debug(cmd)
+ self.cur.callproc('GetNumaByCore', (requirements['numa']['proc_req_nb'],) )
+ valid_for_processor = self.cur.fetchall()
+ self.cur.close()
+ self.cur = self.con.cursor()
+ if len(valid_for_processor)<=0:
+ error_text = 'No room at data center. Cannot find a host with %s %s available' % (str(requirements['numa']['proc_req_nb']),cpu_requirement_text)
+ #self.logger.debug(error_text)
+ return -1, error_text
+
+ #Find the numa nodes that comply for memory and processor requirements
+ #sorting from less to more memory capacity
+ valid_numas = []
+ for m_numa in valid_for_memory:
+ numa_valid_for_processor = False
+ for p_numa in valid_for_processor:
+ if m_numa['numa_id'] == p_numa['numa_id']:
+ numa_valid_for_processor = True
+ break
+ numa_valid_for_host = False
+ prefered_numa = False
+ for p_host in valid_hosts:
+ if m_numa['host_id'] == p_host['uuid']:
+ numa_valid_for_host = True
+ if p_host['uuid'] == prefered_host_id:
+ prefered_numa = True
+ break
+ if numa_valid_for_host and numa_valid_for_processor:
+ if prefered_numa:
+ valid_numas.insert(0, m_numa['numa_id'])
+ else:
+ valid_numas.append(m_numa['numa_id'])
+ if len(valid_numas)<=0:
+ error_text = 'No room at data center. Cannot find a host with %s MB hugepages memory and %s %s available in the same numa' %\
+ (requirements['numa']['memory'], str(requirements['numa']['proc_req_nb']),cpu_requirement_text)
+ #self.logger.debug(error_text)
+ return -1, error_text
+
+ # print 'Valid numas list: '+str(valid_numas)
+
+ #Find valid numa nodes for interfaces requirements
+ #For each valid numa we will obtain the number of available ports and check if these are valid
+ match_found = False
+ for numa_id in valid_numas:
+ # print 'Checking '+str(numa_id)
+ match_found = False
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ if only_of_ports:
+ cmd="CALL GetAvailablePorts(%s)" % str(numa_id)
+ self.logger.debug(cmd)
+ self.cur.callproc('GetAvailablePorts', (numa_id,) )
+ else:
+ cmd="CALL GetAllAvailablePorts(%s)" % str(numa_id)
+ self.logger.debug(cmd)
+ self.cur.callproc('GetAllAvailablePorts', (numa_id,) )
+ available_ports = self.cur.fetchall()
+ self.cur.close()
+ self.cur = self.con.cursor()
+
+ #Set/reset reservations
+ for port in available_ports:
+ port['Mbps_reserved'] = 0
+ port['SRIOV_reserved'] = 0
+
+ #Try to allocate physical ports
+ physical_ports_found = True
+ for iface in requirements['numa']['port_list']:
+ # print '\t\tchecking iface: '+str(iface)
+ portFound = False
+ for port in available_ports:
+ # print '\t\t\tfor port: '+str(port)
+ #If the port is not empty continue
+ if port['Mbps_free'] != port['Mbps'] or port['Mbps_reserved'] != 0:
+ # print '\t\t\t\t Not empty port'
+ continue;
+ #If the port speed is not enough continue
+ if port['Mbps'] < iface['bandwidth']:
+ # print '\t\t\t\t Not enough speed'
+ continue;
+
+ #Otherwise this is a valid port
+ port['Mbps_reserved'] = port['Mbps']
+ port['SRIOV_reserved'] = 0
+ iface['port_id'] = port['port_id']
+ iface['vlan'] = None
+ iface['mac'] = port['mac']
+ iface['switch_port'] = port['switch_port']
+ # print '\t\t\t\t Dedicated port found '+str(port['port_id'])
+ portFound = True
+ break;
+
+ #if all ports have been checked and no match has been found
+ #this is not a valid numa
+ if not portFound:
+ # print '\t\t\t\t\tAll ports have been checked and no match has been found for numa '+str(numa_id)+'\n\n'
+ physical_ports_found = False
+ break
+
+ #if there is no match continue checking the following numa
+ if not physical_ports_found:
+ continue
+
+ #Try to allocate SR-IOVs
+ sriov_ports_found = True
+ for iface in requirements['numa']['sriov_list']:
+ # print '\t\tchecking iface: '+str(iface)
+ portFound = False
+ for port in available_ports:
+ # print '\t\t\tfor port: '+str(port)
+ #If there are not available SR-IOVs continue
+ if port['availableSRIOV'] - port['SRIOV_reserved'] <= 0:
+ # print '\t\t\t\t Not enough SR-IOV'
+ continue;
+ #If the port free speed is not enough continue
+ if port['Mbps_free'] - port['Mbps_reserved'] < iface['bandwidth']:
+ # print '\t\t\t\t Not enough speed'
+ continue;
+
+ #Otherwise this is a valid port
+ port['Mbps_reserved'] += iface['bandwidth']
+ port['SRIOV_reserved'] += 1
+ # print '\t\t\t\t SR-IOV found '+str(port['port_id'])
+ iface['port_id'] = port['port_id']
+ iface['vlan'] = None
+ iface['mac'] = port['mac']
+ iface['switch_port'] = port['switch_port']
+ portFound = True
+ break;
+
+ #if all ports have been checked and no match has been found
+ #this is not a valid numa
+ if not portFound:
+ # print '\t\t\t\t\tAll ports have been checked and no match has been found for numa '+str(numa_id)+'\n\n'
+ sriov_ports_found = False
+ break
+
+ #if there is no match continue checking the following numa
+ if not sriov_ports_found:
+ continue
+
+
+ if sriov_ports_found and physical_ports_found:
+ match_found = True
+ break
+
+ if not match_found:
+ error_text = 'No room at data center. Cannot find a host with the required hugepages, vcpus and interfaces'
+ #self.logger.debug(error_text)
+ return -1, error_text
+
+ #self.logger.debug('Full match found in numa %s', str(numa_id))
+
+ for numa in valid_for_processor:
+ if numa_id==numa['numa_id']:
+ host_id=numa['host_id']
+ break
+ return 0, {'numa_id':numa_id, 'host_id': host_id, }
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "get_numas", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def new_instance(self, instance_dict, nets, ports_to_free):
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+
+ #create uuid if not provided
+ if 'uuid' not in instance_dict:
+ uuid = instance_dict['uuid'] = str(myUuid.uuid1()) # create_uuid
+ else: #check uuid is valid
+ uuid = str(instance_dict['uuid'])
+
+
+ #inserting new uuid
+ cmd = "INSERT INTO uuids (uuid, root_uuid, used_at) VALUES ('%s','%s', 'instances')" % (uuid, uuid)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+ #insert in table instance
+ extended = instance_dict.pop('extended', None);
+ bridgedifaces = instance_dict.pop('bridged-ifaces', () );
+
+ keys = ",".join(instance_dict.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", instance_dict.values() ) )
+ cmd = "INSERT INTO instances (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #if result != 1: return -1, "Database Error while inserting at instances table"
+
+ #insert resources
+ nb_bridge_ifaces = nb_cores = nb_ifaces = nb_numas = 0
+ #insert bridged_ifaces
+
+ for iface in bridgedifaces:
+ #generate and insert a iface uuid
+ if 'enable_dhcp' in iface and iface['enable_dhcp']:
+ dhcp_first_ip = iface["dhcp_first_ip"]
+ del iface["dhcp_first_ip"]
+ dhcp_last_ip = iface["dhcp_last_ip"]
+ del iface["dhcp_last_ip"]
+ dhcp_cidr = iface["cidr"]
+ del iface["cidr"]
+ del iface["enable_dhcp"]
+ used_dhcp_ips = self._get_dhcp_ip_used_list(iface["net_id"])
+ iface["ip_address"] = self.get_free_ip_from_range(dhcp_first_ip, dhcp_last_ip,
+ dhcp_cidr, used_dhcp_ips)
+
+ iface['uuid'] = str(myUuid.uuid1()) # create_uuid
+ cmd = "INSERT INTO uuids (uuid, root_uuid, used_at) VALUES ('%s','%s', 'ports')" % (iface['uuid'], uuid)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #insert iface
+ iface['instance_id'] = uuid
+ # iface['type'] = 'instance:bridge'
+ if 'name' not in iface: iface['name']="br"+str(nb_bridge_ifaces)
+ iface['Mbps']=iface.pop('bandwidth', None)
+ if 'mac_address' not in iface:
+ iface['mac'] = af.gen_random_mac()
+ else:
+ iface['mac'] = iface['mac_address']
+ del iface['mac_address']
+ #iface['mac']=iface.pop('mac_address', None) #for leaving mac generation to libvirt
+ keys = ",".join(iface.keys())
+ values = ",".join( map(lambda x: "Null" if x is None else "'"+str(x)+"'", iface.values() ) )
+ cmd = "INSERT INTO ports (" + keys + ") VALUES (" + values + ")"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ nb_bridge_ifaces += 1
+
+ if extended is not None:
+ if 'numas' not in extended or extended['numas'] is None: extended['numas'] = ()
+ for numa in extended['numas']:
+ nb_numas += 1
+ #cores
+ if 'cores' not in numa or numa['cores'] is None: numa['cores'] = ()
+ for core in numa['cores']:
+ nb_cores += 1
+ cmd = "UPDATE resources_core SET instance_id='%s'%s%s WHERE id='%s'" \
+ % (uuid, \
+ (",v_thread_id='" + str(core['vthread']) + "'") if 'vthread' in core else '', \
+ (",paired='" + core['paired'] + "'") if 'paired' in core else '', \
+ core['id'] )
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #interfaces
+ if 'interfaces' not in numa or numa['interfaces'] is None: numa['interfaces'] = ()
+ for iface in numa['interfaces']:
+ #generate and insert an uuid; iface[id]=iface_uuid; iface[uuid]= net_id
+ iface['id'] = str(myUuid.uuid1()) # create_uuid
+ cmd = "INSERT INTO uuids (uuid, root_uuid, used_at) VALUES ('%s','%s', 'ports')" % (iface['id'], uuid)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ nb_ifaces += 1
+ mbps_=("'"+str(iface['Mbps_used'])+"'") if 'Mbps_used' in iface and iface['Mbps_used'] is not None else "Mbps"
+ if iface["dedicated"]=="yes":
+ iface_model="PF"
+ elif iface["dedicated"]=="yes:sriov":
+ iface_model="VFnotShared"
+ elif iface["dedicated"]=="no":
+ iface_model="VF"
+ #else error
+ INSERT=(iface['mac_address'], iface['switch_port'], iface.get('vlan',None), 'instance:data', iface['Mbps_used'], iface['id'],
+ uuid, instance_dict['tenant_id'], iface.get('name',None), iface.get('vpci',None), iface.get('uuid',None), iface_model )
+ cmd = "INSERT INTO ports (mac,switch_port,vlan,type,Mbps,uuid,instance_id,tenant_id,name,vpci,net_id, model) " + \
+ " VALUES (" + ",".join(map(lambda x: 'Null' if x is None else "'"+str(x)+"'", INSERT )) + ")"
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if 'uuid' in iface:
+ nets.append(iface['uuid'])
+
+ #discover if this port is not used by anyone
+ cmd = "SELECT source_name, mac FROM ( SELECT root_id, count(instance_id) as used FROM resources_port" \
+ " WHERE root_id=(SELECT root_id from resources_port WHERE id='%s')"\
+ " GROUP BY root_id ) AS A JOIN resources_port as B ON A.root_id=B.id AND A.used=0" % iface['port_id']
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ports_to_free += self.cur.fetchall()
+
+ cmd = "UPDATE resources_port SET instance_id='%s', port_id='%s',Mbps_used=%s WHERE id='%s'" \
+ % (uuid, iface['id'], mbps_, iface['port_id'])
+ #if Mbps_used not suply, set the same value of 'Mpbs', that is the total
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ #memory
+ if 'memory' in numa and numa['memory'] is not None and numa['memory']>0:
+ cmd = "INSERT INTO resources_mem (numa_id, instance_id, consumed) VALUES ('%s','%s','%s')" % (numa['numa_id'], uuid, numa['memory'])
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if 'devices' not in extended or extended['devices'] is None: extended['devices'] = ()
+ for device in extended['devices']:
+ if 'vpci' in device: vpci = "'" + device['vpci'] + "'"
+ else: vpci = 'Null'
+ if 'image_id' in device: image_id = "'" + device['image_id'] + "'"
+ else: image_id = 'Null'
+ if 'xml' in device: xml = "'" + device['xml'] + "'"
+ else: xml = 'Null'
+ if 'dev' in device: dev = "'" + device['dev'] + "'"
+ else: dev = 'Null'
+ cmd = "INSERT INTO instance_devices (type, instance_id, image_id, vpci, xml, dev) VALUES ('%s','%s', %s, %s, %s, %s)" % \
+ (device['type'], uuid, image_id, vpci, xml, dev)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ##inserting new log
+ #cmd = "INSERT INTO logs (related,level,uuid,description) VALUES ('instances','debug','%s','new instance: %d numas, %d theads, %d ifaces %d bridge_ifaces')" % (uuid, nb_numas, nb_cores, nb_ifaces, nb_bridge_ifaces)
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+
+ #inseted ok
+ return 1, uuid
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "new_instance", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def get_free_ip_from_range(self, first_ip, last_ip, cidr, ip_used_list):
+ """
+ Calculate a free IP from a range given
+ :param first_ip: First dhcp ip range
+ :param last_ip: Last dhcp ip range
+ :param cidr: net cidr
+ :param ip_used_list: contain all used ips to avoid ip collisions
+ :return:
+ """
+
+ ip_tools = IPNetwork(cidr)
+ cidr_len = ip_tools.prefixlen
+ ips = IPNetwork(first_ip + '/' + str(cidr_len))
+ ip_used_list.append(str(ips[0])) # first ip
+ ip_used_list.append(str(ips[1])) # gw ip
+ ip_used_list.append(str(ips[-1])) # broadcast ip
+ for vm_ip in ips:
+ if str(vm_ip) not in ip_used_list:
+ return vm_ip
+
+ return None
+
+ def _get_dhcp_ip_used_list(self, net_id):
+ """
+ REtreive from DB all ips already used by the dhcp server for a given net
+ :param net_id:
+ :return:
+ """
+ WHERE={'type': 'instance:ovs', 'net_id': net_id}
+ for retry_ in range(0, 2):
+ cmd = ""
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ select_ = "SELECT uuid, ip_address FROM ports "
+
+ if WHERE is None or len(WHERE) == 0:
+ where_ = ""
+ else:
+ where_ = "WHERE " + " AND ".join(
+ map(lambda x: str(x) + (" is Null" if WHERE[x] is None else "='" + str(WHERE[x]) + "'"),
+ WHERE.keys()))
+ limit_ = "LIMIT 100"
+ cmd = " ".join((select_, where_, limit_))
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ports = self.cur.fetchall()
+ ip_address_list = []
+ for port in ports:
+ ip_address_list.append(port['ip_address'])
+
+ return ip_address_list
+
+
+ def delete_instance(self, instance_id, tenant_id, net_dataplane_list, ports_to_free, net_ovs_list, logcause="requested by http"):
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor()
+ #get INSTANCE
+ cmd = "SELECT uuid FROM instances WHERE uuid='%s' AND tenant_id='%s'" % (instance_id, tenant_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount == 0 : return 0, "instance %s not found in tenant %s" % (instance_id, tenant_id)
+
+ #delete bridged ifaces, instace_devices, resources_mem; done by database: it is automatic by Database; FOREIGN KEY DELETE CASCADE
+
+ #get nets afected
+ cmd = "SELECT DISTINCT net_id from ports WHERE instance_id = '%s' AND net_id is not Null AND type='instance:data'" % instance_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ net_list__ = self.cur.fetchall()
+ for net in net_list__:
+ net_dataplane_list.append(net[0])
+
+ # get ovs manangement nets
+ cmd = "SELECT DISTINCT net_id, vlan, ip_address, mac FROM ports WHERE instance_id='{}' AND net_id is not Null AND "\
+ "type='instance:ovs'".format(instance_id)
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ net_ovs_list += self.cur.fetchall()
+
+ #get dataplane interfaces releases by this VM; both PF and VF with no other VF
+ cmd="SELECT source_name, mac FROM (SELECT root_id, count(instance_id) as used FROM resources_port WHERE instance_id='%s' GROUP BY root_id ) AS A" % instance_id \
+ + " JOIN (SELECT root_id, count(instance_id) as used FROM resources_port GROUP BY root_id) AS B ON A.root_id=B.root_id AND A.used=B.used"\
+ + " JOIN resources_port as C ON A.root_id=C.id"
+# cmd = "SELECT DISTINCT root_id FROM resources_port WHERE instance_id = '%s'" % instance_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ports_to_free += self.cur.fetchall()
+
+ #update resources port
+ cmd = "UPDATE resources_port SET instance_id=Null, port_id=Null, Mbps_used='0' WHERE instance_id = '%s'" % instance_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+# #filter dataplane ports used by this VM that now are free
+# for port in ports_list__:
+# cmd = "SELECT mac, count(instance_id) FROM resources_port WHERE root_id = '%s'" % port[0]
+# self.logger.debug(cmd)
+# self.cur.execute(cmd)
+# mac_list__ = self.cur.fetchone()
+# if mac_list__ and mac_list__[1]==0:
+# ports_to_free.append(mac_list__[0])
+
+
+ #update resources core
+ cmd = "UPDATE resources_core SET instance_id=Null, v_thread_id=Null, paired='N' WHERE instance_id = '%s'" % instance_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+ #delete all related uuids
+ cmd = "DELETE FROM uuids WHERE root_uuid='%s'" % instance_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+
+ ##insert log
+ #cmd = "INSERT INTO logs (related,level,uuid,description) VALUES ('instances','debug','%s','delete instance %s')" % (instance_id, logcause)
+ #self.logger.debug(cmd)
+ #self.cur.execute(cmd)
+
+ #delete instance
+ cmd = "DELETE FROM instances WHERE uuid='%s' AND tenant_id='%s'" % (instance_id, tenant_id)
+ self.cur.execute(cmd)
+ return 1, "instance %s from tenant %s DELETED" % (instance_id, tenant_id)
+
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "delete_instance", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def get_ports(self, WHERE):
+ ''' Obtain ports using the WHERE filtering.
+ Attributes:
+ 'where_': dict of key:values, translated to key=value AND ... (Optional)
+ Return: a list with dictionarys at each row
+ '''
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ select_ = "SELECT uuid,'ACTIVE' as status,admin_state_up,name,net_id,\
+ tenant_id,type,mac,vlan,switch_port,instance_id,Mbps FROM ports "
+
+ if WHERE is None or len(WHERE) == 0: where_ = ""
+ else:
+ where_ = "WHERE " + " AND ".join(map( lambda x: str(x) + (" is Null" if WHERE[x] is None else "='"+str(WHERE[x])+"'"), WHERE.keys()) )
+ limit_ = "LIMIT 100"
+ cmd = " ".join( (select_, where_, limit_) )
+ # print "SELECT multiple de instance_ifaces, iface_uuid, external_ports" #print cmd
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ ports = self.cur.fetchall()
+ if self.cur.rowcount>0: af.DeleteNone(ports)
+ return self.cur.rowcount, ports
+ # return self.get_table(FROM=from_, SELECT=select_,WHERE=where_,LIMIT=100)
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "get_ports", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+
+ def check_target_net(self, net_id, tenant_id, port_type):
+ '''check if valid attachement of a port into a target net
+ Attributes:
+ net_id: target net uuid
+ tenant_id: client where tenant belongs. Not used in this version
+ port_type: string with the option 'instance:bridge', 'instance:data', 'external'
+ Return:
+ (0,net_dict) if ok, where net_dict contain 'uuid','type','vlan', ...
+ (negative,string-error) if error
+ '''
+ for retry_ in range(0,2):
+ cmd=""
+ try:
+ with self.con:
+ self.cur = self.con.cursor(mdb.cursors.DictCursor)
+ cmd = "SELECT * FROM nets WHERE uuid='%s'" % net_id
+ self.logger.debug(cmd)
+ self.cur.execute(cmd)
+ if self.cur.rowcount == 0 : return -1, "network_id %s does not match any net" % net_id
+ net = self.cur.fetchone()
+ break
+
+ except (mdb.Error, AttributeError) as e:
+ r,c = self.format_error(e, "check_target_net", cmd)
+ if r!=-HTTP_Request_Timeout or retry_==1: return r,c
+ #check permissions
+ if tenant_id is not None and tenant_id is not "admin":
+ if net['tenant_id']==tenant_id and net['shared']=='false':
+ return -1, "needed admin privileges to attach to the net %s" % net_id
+ #check types
+ if (net['type'] in ('ptp','data') and port_type not in ('instance:data','external')) or \
+ (net['type'] in ('bridge_data','bridge_man') and port_type not in ('instance:bridge', 'instance:ovs')):
+ return -1, "Cannot attach a port of type %s into a net of type %s" % (port_type, net['type'])
+ if net['type'] == 'ptp':
+ #look how many
+ nb_ports, data = self.get_ports( {'net_id':net_id} )
+ if nb_ports<0:
+ return -1, data
+ else:
+ if net['provider']:
+ nb_ports +=1
+ if nb_ports >=2:
+ return -1, "net of type p2p already contain two ports attached. No room for another"
+
+ return 0, net
+
+if __name__ == "__main__":
+ print "Hello World"