X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=osm_common%2Fmsgkafka.py;h=5caa5b1becb86a78142a4ac50d53596aa64e630e;hb=3d82ba2cc4c3ebb340a88400aaa4a8d3683806a1;hp=767fff6c30bba3703e27afc093f77f9b10ea94f0;hpb=48b78e112662d927cad376d739e43fb94f108ff3;p=osm%2Fcommon.git diff --git a/osm_common/msgkafka.py b/osm_common/msgkafka.py index 767fff6..5caa5b1 100644 --- a/osm_common/msgkafka.py +++ b/osm_common/msgkafka.py @@ -21,13 +21,15 @@ from aiokafka import AIOKafkaProducer from aiokafka.errors import KafkaError from osm_common.msgbase import MsgBase, MsgException -__author__ = "Alfonso Tierno , " \ - "Guillermo Calvino " +__author__ = ( + "Alfonso Tierno , " + "Guillermo Calvino " +) class MsgKafka(MsgBase): - def __init__(self, logger_name='msg'): - self.logger = logging.getLogger(logger_name) + def __init__(self, logger_name="msg", lock=False): + super().__init__(logger_name, lock) self.host = None self.port = None self.consumer = None @@ -42,7 +44,7 @@ class MsgKafka(MsgBase): self.logger = logging.getLogger(config["logger_name"]) self.host = config["host"] self.port = config["port"] - self.loop = asyncio.get_event_loop() + self.loop = config.get("loop") or asyncio.get_event_loop() self.broker = str(self.host) + ":" + str(self.port) self.group_id = config.get("group_id") @@ -64,11 +66,19 @@ class MsgKafka(MsgBase): :param msg: message content, can be string or dictionary :return: None or raises MsgException on failing """ - try: - self.loop.run_until_complete(self.aiowrite(topic=topic, key=key, msg=msg)) - - except Exception as e: - raise MsgException("Error writing {} topic: {}".format(topic, str(e))) + retry = 2 # Try two times + while retry: + try: + self.loop.run_until_complete( + self.aiowrite(topic=topic, key=key, msg=msg) + ) + break + except Exception as e: + retry -= 1 + if retry == 0: + raise MsgException( + "Error writing {} topic: {}".format(topic, str(e)) + ) def read(self, topic): """ @@ -84,49 +94,101 @@ class MsgKafka(MsgBase): raise MsgException("Error reading {} topic: {}".format(topic, str(e))) async def aiowrite(self, topic, key, msg, loop=None): + """ + Asyncio write + :param topic: str kafka topic + :param key: str kafka key + :param msg: str or dictionary kafka message + :param loop: asyncio loop. To be DEPRECATED! in near future!!! loop must be provided inside config at connect + :return: None + """ if not loop: loop = self.loop try: - self.producer = AIOKafkaProducer(loop=loop, key_serializer=str.encode, value_serializer=str.encode, - bootstrap_servers=self.broker) + self.producer = AIOKafkaProducer( + loop=loop, + key_serializer=str.encode, + value_serializer=str.encode, + bootstrap_servers=self.broker, + ) await self.producer.start() - await self.producer.send(topic=topic, key=key, value=yaml.safe_dump(msg, default_flow_style=True)) + await self.producer.send( + topic=topic, key=key, value=yaml.safe_dump(msg, default_flow_style=True) + ) except Exception as e: - raise MsgException("Error publishing topic '{}', key '{}': {}".format(topic, key, e)) + raise MsgException( + "Error publishing topic '{}', key '{}': {}".format(topic, key, e) + ) finally: await self.producer.stop() - async def aioread(self, topic, loop=None, callback=None, aiocallback=None, **kwargs): + async def aioread( + self, + topic, + loop=None, + callback=None, + aiocallback=None, + group_id=None, + from_beginning=None, + **kwargs + ): """ - Asyncio read from one or several topics. It blocks. + Asyncio read from one or several topics. :param topic: can be str: single topic; or str list: several topics - :param loop: asyncio loop + :param loop: asyncio loop. To be DEPRECATED! in near future!!! loop must be provided inside config at connect :param callback: synchronous callback function that will handle the message in kafka bus :param aiocallback: async callback function that will handle the message in kafka bus + :param group_id: kafka group_id to use. Can be False (set group_id to None), None (use general group_id provided + at connect inside config), or a group_id string + :param from_beginning: if True, messages will be obtained from beginning instead of only new ones. + If group_id is supplied, only the not processed messages by other worker are obtained. + If group_id is None, all messages stored at kafka are obtained. :param kwargs: optional keyword arguments for callback function :return: If no callback defined, it returns (topic, key, message) """ if not loop: loop = self.loop + if group_id is False: + group_id = None + elif group_id is None: + group_id = self.group_id try: if isinstance(topic, (list, tuple)): topic_list = topic else: topic_list = (topic,) - - self.consumer = AIOKafkaConsumer(loop=loop, bootstrap_servers=self.broker, group_id=self.group_id) + self.consumer = AIOKafkaConsumer( + loop=loop, + bootstrap_servers=self.broker, + group_id=group_id, + auto_offset_reset="earliest" if from_beginning else "latest", + ) await self.consumer.start() self.consumer.subscribe(topic_list) async for message in self.consumer: if callback: - callback(message.topic, yaml.load(message.key), yaml.load(message.value), **kwargs) + callback( + message.topic, + yaml.safe_load(message.key), + yaml.safe_load(message.value), + **kwargs + ) elif aiocallback: - await aiocallback(message.topic, yaml.load(message.key), yaml.load(message.value), **kwargs) + await aiocallback( + message.topic, + yaml.safe_load(message.key), + yaml.safe_load(message.value), + **kwargs + ) else: - return message.topic, yaml.load(message.key), yaml.load(message.value) + return ( + message.topic, + yaml.safe_load(message.key), + yaml.safe_load(message.value), + ) except KafkaError as e: raise MsgException(str(e)) finally: