File size: 16,132 Bytes
5980447
1
2
{"repo": "BiznetGIO/RESTKnot", "pull_number": 126, "instance_id": "BiznetGIO__RESTKnot-126", "issue_numbers": "", "base_commit": "068b0616f0a7b839449138cbebf6f53471b2b2da", "patch": "diff --git a/agent/dnsagent/cli.py b/agent/dnsagent/cli.py\ndeleted file mode 100644\n--- a/agent/dnsagent/cli.py\n+++ /dev/null\n@@ -1,69 +0,0 @@\n-\"\"\"\n-Usage:\n-  dnsagent <command> [<args>...]\n-\n-Options:\n-  -h, --help                             display this help and exit\n-  -v, --version                          Print version information and quit\n-\n-Commands:\n-  zone            Zone Record Configuration Zone Command\n-  command         CZone onfiguration Command\n-  start           Starting Agent\n-\n-Run 'dnsagent COMMAND --help' for more information on a command.\n-\"\"\"\n-import logging\n-import os\n-import sys\n-from inspect import getmembers, isclass\n-from logging.handlers import RotatingFileHandler\n-\n-from docopt import docopt\n-import dnsagent.clis\n-from dnsagent import __version__ as VERSION\n-\n-\n-def configure_logger():\n-    stdout_handler = logging.StreamHandler(sys.stdout)\n-    stdout_format = logging.Formatter(\n-        \"[%(asctime)s - %(levelname)s - %(filename)s:%(lineno)s - %(funcName)20s() ] %(message)s\"\n-    )\n-    stdout_handler.setFormatter(stdout_format)\n-    stdout_handler.setLevel(logging.INFO)\n-\n-    root = logging.getLogger()\n-    root.addHandler(stdout_handler)\n-    root.setLevel(logging.DEBUG)\n-\n-\n-def main():\n-    \"\"\"Main CLI entrypoint.\"\"\"\n-    configure_logger()\n-\n-    options = docopt(__doc__, version=VERSION, options_first=True)\n-    command_name = \"\"\n-    args = \"\"\n-    command_class = \"\"\n-\n-    command_name = options.pop(\"<command>\")\n-    args = options.pop(\"<args>\")\n-\n-    if args is None:\n-        args = {}\n-\n-    try:\n-        module = getattr(dnsagent.clis, command_name)\n-        dnsagent.clis = getmembers(module, isclass)\n-        command_class = [\n-            command[1] for command in dnsagent.clis if command[0] != \"Base\"\n-        ][0]\n-    except AttributeError as e:\n-        raise ValueError(f\"{e}\")\n-\n-    command = command_class(options, args)\n-    command.execute()\n-\n-\n-if __name__ == \"__main__\":\n-    main()\ndiff --git a/agent/dnsagent/clis/__init__.py b/agent/dnsagent/clis/__init__.py\ndeleted file mode 100644\n--- a/agent/dnsagent/clis/__init__.py\n+++ /dev/null\n@@ -1 +0,0 @@\n-from .start import *\ndiff --git a/agent/dnsagent/clis/base.py b/agent/dnsagent/clis/base.py\ndeleted file mode 100644\n--- a/agent/dnsagent/clis/base.py\n+++ /dev/null\n@@ -1,19 +0,0 @@\n-from docopt import docopt\n-\n-\n-class Base(object):\n-    \"\"\"Base class for the commands\"\"\"\n-\n-    def __init__(self, options, command_args):\n-        \"\"\"\n-        Initialize the commands.\n-\n-        :param command_args: arguments of the command\n-        \"\"\"\n-        self.options = options\n-        self.args = docopt(self.__doc__, argv=command_args)\n-\n-    def execute(self):\n-        \"\"\"Execute the commands\"\"\"\n-\n-        raise NotImplementedError\ndiff --git a/agent/dnsagent/clis/start.py b/agent/dnsagent/clis/start.py\ndeleted file mode 100644\n--- a/agent/dnsagent/clis/start.py\n+++ /dev/null\n@@ -1,62 +0,0 @@\n-import os\n-import logging\n-\n-from dnsagent.clis.base import Base\n-from dnsagent.libs import kafka as kafka_lib\n-from dnsagent.libs import knot as knot_lib\n-\n-\n-logger = logging.getLogger(__name__)\n-\n-\n-class Start(Base):\n-    \"\"\"\n-        usage:\n-            start\n-\n-        Command :\n-\n-        Options:\n-        -h --help                             Print usage\n-    \"\"\"\n-\n-    def connect_kafka(self):\n-        broker_host = os.environ.get(\"RESTKNOT_KAFKA_BROKER\")\n-        broker_port = os.environ.get(\"RESTKNOT_KAFKA_PORTS\")\n-        broker = f\"{broker_host}:{broker_port}\"\n-        topic = os.environ.get(\"RESTKNOT_KAFKA_TOPIC\")\n-\n-        if (broker_host and broker_port) is None:\n-            logger.info(\"Can't find kafka host and port\")\n-            exit()\n-\n-        try:\n-            logger.info(\"Connecting to broker : \" + broker)\n-            consumer = kafka_lib.get_kafka_consumer(broker, topic)\n-            return consumer\n-        except Exception as e:\n-            logger.info(f\"Can't Connect to broker: {e}\")\n-            exit()\n-\n-    def take_message(self, consumer):\n-        agent_type = os.environ.get(\"RESTKNOT_AGENT_TYPE\")\n-\n-        try:\n-            for message in consumer:\n-                message = message.value\n-\n-                agent_type_msg = message[\"agent\"][\"agent_type\"]\n-                if agent_type in agent_type_msg:\n-\n-                    knot_queries = message[\"knot\"]\n-                    for query in knot_queries:\n-                        knot_lib.execute(query)\n-\n-            consumer.close()\n-\n-        except KeyboardInterrupt:\n-            print(\"Stopping dnsagent. Press Ctrl+C again to exit\")\n-\n-    def execute(self):\n-        consumer = self.connect_kafka()\n-        self.take_message(consumer)\ndiff --git a/agent/dnsagent/libs/kafka.py b/agent/dnsagent/libs/kafka.py\ndeleted file mode 100644\n--- a/agent/dnsagent/libs/kafka.py\n+++ /dev/null\n@@ -1,16 +0,0 @@\n-from kafka import KafkaConsumer\n-from json import loads\n-\n-\n-def get_kafka_consumer(broker, topic):\n-    try:\n-        consumer = KafkaConsumer(\n-            topic,\n-            bootstrap_servers=[broker],\n-            auto_offset_reset=\"earliest\",\n-            enable_auto_commit=True,\n-            value_deserializer=lambda x: loads(x.decode(\"utf-8\")),\n-        )\n-        return consumer\n-    except Exception as e:\n-        raise ValueError(f\"{e}\")\ndiff --git a/agent/dnsagent/start.py b/agent/dnsagent/start.py\nnew file mode 100644\n--- /dev/null\n+++ b/agent/dnsagent/start.py\n@@ -0,0 +1,74 @@\n+import os\n+import logging\n+import json\n+import sys\n+\n+from confluent_kafka import Consumer, KafkaException\n+\n+from dnsagent.libs import knot as knot_lib\n+\n+\n+def consume():\n+    brokers = os.environ.get(\"RESTKNOT_KAFKA_BROKERS\")\n+    topic = os.environ.get(\"RESTKNOT_KAFKA_TOPIC\")\n+    group_id = os.environ.get(\"RESTKNOT_KAFKA_GROUP_ID\")\n+    agent_type = os.environ.get(\"RESTKNOT_AGENT_TYPE\")\n+\n+    conf = {\n+        \"bootstrap.servers\": brokers,\n+        \"group.id\": group_id,\n+        \"auto.offset.reset\": \"earliest\",\n+        \"enable.auto.commit\": True,\n+    }\n+\n+    def print_assignment(consumer, partitions):\n+        print(\"Consumer assigned to:\", partitions)\n+\n+    consumer = Consumer(conf)\n+    consumer.subscribe([topic], on_assign=print_assignment)\n+\n+    try:\n+        while True:\n+            message = consumer.poll(timeout=1.0)\n+            if message is None:\n+                continue\n+            if message.error():\n+                raise KafkaException(message.error())\n+\n+            message = message.value()\n+            message = json.loads(message.decode(\"utf-8\"))\n+\n+            agent_type_msg = message[\"agent\"][\"agent_type\"]\n+            if agent_type in agent_type_msg:\n+\n+                knot_queries = message[\"knot\"]\n+                for query in knot_queries:\n+                    knot_lib.execute(query)\n+\n+    except KeyboardInterrupt:\n+        print(\" dnsagent stopped. Aborted by user\")\n+    finally:\n+        # Close down consumer to commit final offsets.\n+        consumer.close()\n+\n+\n+def configure_logger():\n+    stdout_handler = logging.StreamHandler(sys.stdout)\n+    stdout_format = logging.Formatter(\n+        \"[%(asctime)s - %(levelname)s - %(filename)s:%(lineno)s - %(funcName)20s() ] %(message)s\"\n+    )\n+    stdout_handler.setFormatter(stdout_format)\n+    stdout_handler.setLevel(logging.INFO)\n+\n+    root = logging.getLogger()\n+    root.addHandler(stdout_handler)\n+    root.setLevel(logging.DEBUG)\n+\n+\n+def main():\n+    configure_logger()\n+    consume()\n+\n+\n+if __name__ == \"__main__\":\n+    main()\ndiff --git a/agent/setup.py b/agent/setup.py\n--- a/agent/setup.py\n+++ b/agent/setup.py\n@@ -36,5 +36,5 @@\n     include_package_data=True,\n     packages=[\"dnsagent\"],\n     install_requires=requirements,\n-    entry_points={\"console_scripts\": [\"dnsagent = dnsagent.cli:main\"]},\n+    entry_points={\"console_scripts\": [\"dnsagent = dnsagent.start:main\"]},\n )\ndiff --git a/api/app/controllers/api/__init__.py b/api/app/controllers/api/__init__.py\n--- a/api/app/controllers/api/__init__.py\n+++ b/api/app/controllers/api/__init__.py\n@@ -9,6 +9,7 @@\n     GetDomainDataId,\n )\n from .health import HealthCheck\n+from .meta import MetaVersion, MetaConfig\n from .record import GetRecordData, GetRecordDataId, RecordAdd, RecordDelete, RecordEdit\n from .ttl import GetTtlData, GetTtlDataId, TtlAdd, TtlDelete, TtlEdit\n from .type_ import GetTypeData, GetTypeDataId, TypeAdd, TypeDelete, TypeEdit\n@@ -19,6 +20,8 @@\n \n \n api.add_resource(HealthCheck, \"/health\")\n+api.add_resource(MetaVersion, \"/meta/version\")\n+api.add_resource(MetaConfig, \"/meta/config\")\n \n api.add_resource(GetRecordData, \"/record/list\")\n api.add_resource(GetRecordDataId, \"/record/list/<record_id>\")\ndiff --git a/api/app/controllers/api/health.py b/api/app/controllers/api/health.py\n--- a/api/app/controllers/api/health.py\n+++ b/api/app/controllers/api/health.py\n@@ -1,12 +1,11 @@\n from flask_restful import Resource\n \n from app.vendors.rest import response\n-from app.helpers import helpers\n \n \n class HealthCheck(Resource):\n     def get(self):\n-        build = helpers.read_version(\"requirements.txt\", \"build-version.txt\")\n-\n-        data = {\"status\": \"running\", \"build\": build}\n+        data = {\n+            \"status\": \"running\",\n+        }\n         return response(200, data=data, message=\"OK\")\ndiff --git a/api/app/controllers/api/meta.py b/api/app/controllers/api/meta.py\nnew file mode 100644\n--- /dev/null\n+++ b/api/app/controllers/api/meta.py\n@@ -0,0 +1,24 @@\n+from flask_restful import Resource\n+\n+from app.vendors.rest import response\n+from app.helpers import helpers\n+from app.middlewares import auth\n+\n+\n+class MetaVersion(Resource):\n+    def get(self):\n+        build = helpers.read_version(\"requirements.txt\", \"build-version.txt\")\n+\n+        data = {\"build\": build}\n+        return response(200, data=data, message=\"OK\")\n+\n+\n+class MetaConfig(Resource):\n+    @auth.auth_required\n+    def get(self):\n+        config = helpers.get_config()\n+        brokers = config[\"brokers\"]\n+        clusters = config[\"knot_servers\"]\n+\n+        data = {\"knot_servers\": clusters, \"brokers\": brokers}\n+        return response(200, data=data, message=\"OK\")\ndiff --git a/api/app/helpers/command.py b/api/app/helpers/command.py\n--- a/api/app/helpers/command.py\n+++ b/api/app/helpers/command.py\n@@ -1,10 +1,7 @@\n import json\n-import os\n-import pathlib\n-\n-import yaml\n \n from app.helpers import producer\n+from app.helpers import helpers\n from app.models import model\n \n \n@@ -139,30 +136,14 @@ def set_default_zone(record_ids):\n     producer.send(message)\n \n \n-def cluster_file():\n-    \"\"\"Return cluster file path.\"\"\"\n-    path = os.environ.get(\"RESTKNOT_CLUSTER_FILE\")\n-    if not path:\n-        current_path = pathlib.Path(__file__)\n-        path = current_path.parents[2].joinpath(\"servers.yml\")\n-\n-    is_exists = os.path.exists(path)\n-    if is_exists:\n-        return path\n-    else:\n-        raise ValueError(f\"Clustering File Not Found: {path}\")\n-\n-\n-def get_clusters():\n-    \"\"\"Return cluster file content.\"\"\"\n-    file_ = cluster_file()\n-    clusters = yaml.safe_load(open(file_))\n-    return clusters\n-\n-\n def delegate(zone, zone_id, command, agent_type):\n     \"\"\"Send delegation config command with JSON structure to broker.\"\"\"\n-    clusters = get_clusters()\n+    config = helpers.get_config()\n+    try:\n+        clusters = config[\"knot_servers\"]\n+    except KeyError:\n+        raise ValueError(\"Can't Knot server list in config\")\n+\n     cluster = clusters[agent_type]\n \n     # default for master\ndiff --git a/api/app/helpers/helpers.py b/api/app/helpers/helpers.py\n--- a/api/app/helpers/helpers.py\n+++ b/api/app/helpers/helpers.py\n@@ -1,7 +1,10 @@\n import datetime\n import pathlib\n+import os\n from functools import wraps\n \n+import yaml\n+\n from app.helpers import producer\n from app.vendors.rest import response\n \n@@ -97,3 +100,24 @@ def read_version(other_file_name, filename):\n         version = \"__UNKNOWN__\"\n \n     return version\n+\n+\n+def config_file():\n+    \"\"\"Return config file path.\"\"\"\n+    path = os.environ.get(\"RESTKNOT_CONFIG_FILE\")\n+    if not path:\n+        current_path = pathlib.Path(__file__)\n+        path = current_path.parents[2].joinpath(\"config.yml\")\n+\n+    is_exists = os.path.exists(path)\n+    if is_exists:\n+        return path\n+    else:\n+        raise ValueError(f\"Config File Not Found: {path}\")\n+\n+\n+def get_config():\n+    \"\"\"Return config file content.\"\"\"\n+    file_ = config_file()\n+    config = yaml.safe_load(open(file_))\n+    return config\ndiff --git a/api/app/helpers/producer.py b/api/app/helpers/producer.py\n--- a/api/app/helpers/producer.py\n+++ b/api/app/helpers/producer.py\n@@ -2,33 +2,43 @@\n import os\n \n from flask import current_app\n-from kafka import KafkaProducer\n+from confluent_kafka import Producer\n+\n+from app.helpers import helpers\n \n \n def kafka_producer():\n     \"\"\"Create Kafka producer.\"\"\"\n-    host = os.environ.get(\"KAFKA_HOST\")\n-    port = os.environ.get(\"KAFKA_PORT\")\n-    broker = f\"{host}:{port}\"\n-\n-    producer = KafkaProducer(\n-        bootstrap_servers=[broker],\n-        value_serializer=lambda m: json.dumps(m).encode(\"utf-8\"),\n-    )\n+    config = helpers.get_config()\n+    try:\n+        brokers = config[\"brokers\"]\n+    except KeyError:\n+        raise ValueError(\"Can't find brokers list in config\")\n+\n+    brokers = \",\".join(brokers)\n+    conf = {\"bootstrap.servers\": brokers}\n+    producer = Producer(**conf)\n     return producer\n \n \n+def _delivery_report(err, msg):\n+    if err is not None:\n+        raise ValueError(f\"Message delivery failed: {err}\")\n+\n+\n def send(message):\n     \"\"\"Send given message to Kafka broker.\"\"\"\n     producer = None\n     try:\n         producer = kafka_producer()\n         topic = os.environ.get(\"RESTKNOT_KAFKA_TOPIC\")\n-        producer.send(topic, message)\n-        producer.flush()\n+        encoded_message = json.dumps(message).encode(\"utf-8\")\n+        producer.produce(topic, encoded_message, callback=_delivery_report)\n     except Exception as e:\n         current_app.logger.error(f\"{e}\")\n         raise ValueError(f\"{e}\")\n-    finally:\n-        if producer:\n-            producer.close()\n+\n+    # Serve delivery callback queue.\n+    producer.poll(0)\n+    # Wait until all messages have been delivered\n+    producer.flush()\ndiff --git a/api/gunicorn.conf.py b/api/gunicorn.conf.py\nnew file mode 100644\n--- /dev/null\n+++ b/api/gunicorn.conf.py\n@@ -0,0 +1,13 @@\n+import os\n+import multiprocessing\n+\n+\n+def max_workers():\n+    return multiprocessing.cpu_count() * 2 + 1\n+\n+\n+host = os.environ.get(\"APP_HOST\", \"0.0.0.0\")\n+port = os.environ.get(\"APP_PORT\", \"8000\")\n+\n+bind = f\"{host}:{port}\"\n+workers = max_workers()\n", "test_patch": "diff --git a/api/tests/integration/conftest.py b/api/tests/integration/conftest.py\n--- a/api/tests/integration/conftest.py\n+++ b/api/tests/integration/conftest.py\n@@ -19,7 +19,7 @@ def clean_users():\n @pytest.fixture\n def client():\n     current_path = pathlib.Path(__file__)\n-    dotenv_path = current_path.parents[2].joinpath(\".env.example\")\n+    dotenv_path = current_path.parents[2].joinpath(\".example.env\")\n     load_dotenv(dotenv_path)\n \n     app = create_app()\n", "problem_statement": "", "hints_text": "", "created_at": "2021-04-14T10:12:08Z"}