File size: 16,132 Bytes
5980447 |
1 2 |
{"repo": "BiznetGIO/RESTKnot", "pull_number": 126, "instance_id": "BiznetGIO__RESTKnot-126", "issue_numbers": "", "base_commit": "068b0616f0a7b839449138cbebf6f53471b2b2da", "patch": "diff --git a/agent/dnsagent/cli.py b/agent/dnsagent/cli.py\ndeleted file mode 100644\n--- a/agent/dnsagent/cli.py\n+++ /dev/null\n@@ -1,69 +0,0 @@\n-\"\"\"\n-Usage:\n- dnsagent <command> [<args>...]\n-\n-Options:\n- -h, --help display this help and exit\n- -v, --version Print version information and quit\n-\n-Commands:\n- zone Zone Record Configuration Zone Command\n- command CZone onfiguration Command\n- start Starting Agent\n-\n-Run 'dnsagent COMMAND --help' for more information on a command.\n-\"\"\"\n-import logging\n-import os\n-import sys\n-from inspect import getmembers, isclass\n-from logging.handlers import RotatingFileHandler\n-\n-from docopt import docopt\n-import dnsagent.clis\n-from dnsagent import __version__ as VERSION\n-\n-\n-def configure_logger():\n- stdout_handler = logging.StreamHandler(sys.stdout)\n- stdout_format = logging.Formatter(\n- \"[%(asctime)s - %(levelname)s - %(filename)s:%(lineno)s - %(funcName)20s() ] %(message)s\"\n- )\n- stdout_handler.setFormatter(stdout_format)\n- stdout_handler.setLevel(logging.INFO)\n-\n- root = logging.getLogger()\n- root.addHandler(stdout_handler)\n- root.setLevel(logging.DEBUG)\n-\n-\n-def main():\n- \"\"\"Main CLI entrypoint.\"\"\"\n- configure_logger()\n-\n- options = docopt(__doc__, version=VERSION, options_first=True)\n- command_name = \"\"\n- args = \"\"\n- command_class = \"\"\n-\n- command_name = options.pop(\"<command>\")\n- args = options.pop(\"<args>\")\n-\n- if args is None:\n- args = {}\n-\n- try:\n- module = getattr(dnsagent.clis, command_name)\n- dnsagent.clis = getmembers(module, isclass)\n- command_class = [\n- command[1] for command in dnsagent.clis if command[0] != \"Base\"\n- ][0]\n- except AttributeError as e:\n- raise ValueError(f\"{e}\")\n-\n- command = command_class(options, args)\n- command.execute()\n-\n-\n-if __name__ == \"__main__\":\n- main()\ndiff --git a/agent/dnsagent/clis/__init__.py b/agent/dnsagent/clis/__init__.py\ndeleted file mode 100644\n--- a/agent/dnsagent/clis/__init__.py\n+++ /dev/null\n@@ -1 +0,0 @@\n-from .start import *\ndiff --git a/agent/dnsagent/clis/base.py b/agent/dnsagent/clis/base.py\ndeleted file mode 100644\n--- a/agent/dnsagent/clis/base.py\n+++ /dev/null\n@@ -1,19 +0,0 @@\n-from docopt import docopt\n-\n-\n-class Base(object):\n- \"\"\"Base class for the commands\"\"\"\n-\n- def __init__(self, options, command_args):\n- \"\"\"\n- Initialize the commands.\n-\n- :param command_args: arguments of the command\n- \"\"\"\n- self.options = options\n- self.args = docopt(self.__doc__, argv=command_args)\n-\n- def execute(self):\n- \"\"\"Execute the commands\"\"\"\n-\n- raise NotImplementedError\ndiff --git a/agent/dnsagent/clis/start.py b/agent/dnsagent/clis/start.py\ndeleted file mode 100644\n--- a/agent/dnsagent/clis/start.py\n+++ /dev/null\n@@ -1,62 +0,0 @@\n-import os\n-import logging\n-\n-from dnsagent.clis.base import Base\n-from dnsagent.libs import kafka as kafka_lib\n-from dnsagent.libs import knot as knot_lib\n-\n-\n-logger = logging.getLogger(__name__)\n-\n-\n-class Start(Base):\n- \"\"\"\n- usage:\n- start\n-\n- Command :\n-\n- Options:\n- -h --help Print usage\n- \"\"\"\n-\n- def connect_kafka(self):\n- broker_host = os.environ.get(\"RESTKNOT_KAFKA_BROKER\")\n- broker_port = os.environ.get(\"RESTKNOT_KAFKA_PORTS\")\n- broker = f\"{broker_host}:{broker_port}\"\n- topic = os.environ.get(\"RESTKNOT_KAFKA_TOPIC\")\n-\n- if (broker_host and broker_port) is None:\n- logger.info(\"Can't find kafka host and port\")\n- exit()\n-\n- try:\n- logger.info(\"Connecting to broker : \" + broker)\n- consumer = kafka_lib.get_kafka_consumer(broker, topic)\n- return consumer\n- except Exception as e:\n- logger.info(f\"Can't Connect to broker: {e}\")\n- exit()\n-\n- def take_message(self, consumer):\n- agent_type = os.environ.get(\"RESTKNOT_AGENT_TYPE\")\n-\n- try:\n- for message in consumer:\n- message = message.value\n-\n- agent_type_msg = message[\"agent\"][\"agent_type\"]\n- if agent_type in agent_type_msg:\n-\n- knot_queries = message[\"knot\"]\n- for query in knot_queries:\n- knot_lib.execute(query)\n-\n- consumer.close()\n-\n- except KeyboardInterrupt:\n- print(\"Stopping dnsagent. Press Ctrl+C again to exit\")\n-\n- def execute(self):\n- consumer = self.connect_kafka()\n- self.take_message(consumer)\ndiff --git a/agent/dnsagent/libs/kafka.py b/agent/dnsagent/libs/kafka.py\ndeleted file mode 100644\n--- a/agent/dnsagent/libs/kafka.py\n+++ /dev/null\n@@ -1,16 +0,0 @@\n-from kafka import KafkaConsumer\n-from json import loads\n-\n-\n-def get_kafka_consumer(broker, topic):\n- try:\n- consumer = KafkaConsumer(\n- topic,\n- bootstrap_servers=[broker],\n- auto_offset_reset=\"earliest\",\n- enable_auto_commit=True,\n- value_deserializer=lambda x: loads(x.decode(\"utf-8\")),\n- )\n- return consumer\n- except Exception as e:\n- raise ValueError(f\"{e}\")\ndiff --git a/agent/dnsagent/start.py b/agent/dnsagent/start.py\nnew file mode 100644\n--- /dev/null\n+++ b/agent/dnsagent/start.py\n@@ -0,0 +1,74 @@\n+import os\n+import logging\n+import json\n+import sys\n+\n+from confluent_kafka import Consumer, KafkaException\n+\n+from dnsagent.libs import knot as knot_lib\n+\n+\n+def consume():\n+ brokers = os.environ.get(\"RESTKNOT_KAFKA_BROKERS\")\n+ topic = os.environ.get(\"RESTKNOT_KAFKA_TOPIC\")\n+ group_id = os.environ.get(\"RESTKNOT_KAFKA_GROUP_ID\")\n+ agent_type = os.environ.get(\"RESTKNOT_AGENT_TYPE\")\n+\n+ conf = {\n+ \"bootstrap.servers\": brokers,\n+ \"group.id\": group_id,\n+ \"auto.offset.reset\": \"earliest\",\n+ \"enable.auto.commit\": True,\n+ }\n+\n+ def print_assignment(consumer, partitions):\n+ print(\"Consumer assigned to:\", partitions)\n+\n+ consumer = Consumer(conf)\n+ consumer.subscribe([topic], on_assign=print_assignment)\n+\n+ try:\n+ while True:\n+ message = consumer.poll(timeout=1.0)\n+ if message is None:\n+ continue\n+ if message.error():\n+ raise KafkaException(message.error())\n+\n+ message = message.value()\n+ message = json.loads(message.decode(\"utf-8\"))\n+\n+ agent_type_msg = message[\"agent\"][\"agent_type\"]\n+ if agent_type in agent_type_msg:\n+\n+ knot_queries = message[\"knot\"]\n+ for query in knot_queries:\n+ knot_lib.execute(query)\n+\n+ except KeyboardInterrupt:\n+ print(\" dnsagent stopped. Aborted by user\")\n+ finally:\n+ # Close down consumer to commit final offsets.\n+ consumer.close()\n+\n+\n+def configure_logger():\n+ stdout_handler = logging.StreamHandler(sys.stdout)\n+ stdout_format = logging.Formatter(\n+ \"[%(asctime)s - %(levelname)s - %(filename)s:%(lineno)s - %(funcName)20s() ] %(message)s\"\n+ )\n+ stdout_handler.setFormatter(stdout_format)\n+ stdout_handler.setLevel(logging.INFO)\n+\n+ root = logging.getLogger()\n+ root.addHandler(stdout_handler)\n+ root.setLevel(logging.DEBUG)\n+\n+\n+def main():\n+ configure_logger()\n+ consume()\n+\n+\n+if __name__ == \"__main__\":\n+ main()\ndiff --git a/agent/setup.py b/agent/setup.py\n--- a/agent/setup.py\n+++ b/agent/setup.py\n@@ -36,5 +36,5 @@\n include_package_data=True,\n packages=[\"dnsagent\"],\n install_requires=requirements,\n- entry_points={\"console_scripts\": [\"dnsagent = dnsagent.cli:main\"]},\n+ entry_points={\"console_scripts\": [\"dnsagent = dnsagent.start:main\"]},\n )\ndiff --git a/api/app/controllers/api/__init__.py b/api/app/controllers/api/__init__.py\n--- a/api/app/controllers/api/__init__.py\n+++ b/api/app/controllers/api/__init__.py\n@@ -9,6 +9,7 @@\n GetDomainDataId,\n )\n from .health import HealthCheck\n+from .meta import MetaVersion, MetaConfig\n from .record import GetRecordData, GetRecordDataId, RecordAdd, RecordDelete, RecordEdit\n from .ttl import GetTtlData, GetTtlDataId, TtlAdd, TtlDelete, TtlEdit\n from .type_ import GetTypeData, GetTypeDataId, TypeAdd, TypeDelete, TypeEdit\n@@ -19,6 +20,8 @@\n \n \n api.add_resource(HealthCheck, \"/health\")\n+api.add_resource(MetaVersion, \"/meta/version\")\n+api.add_resource(MetaConfig, \"/meta/config\")\n \n api.add_resource(GetRecordData, \"/record/list\")\n api.add_resource(GetRecordDataId, \"/record/list/<record_id>\")\ndiff --git a/api/app/controllers/api/health.py b/api/app/controllers/api/health.py\n--- a/api/app/controllers/api/health.py\n+++ b/api/app/controllers/api/health.py\n@@ -1,12 +1,11 @@\n from flask_restful import Resource\n \n from app.vendors.rest import response\n-from app.helpers import helpers\n \n \n class HealthCheck(Resource):\n def get(self):\n- build = helpers.read_version(\"requirements.txt\", \"build-version.txt\")\n-\n- data = {\"status\": \"running\", \"build\": build}\n+ data = {\n+ \"status\": \"running\",\n+ }\n return response(200, data=data, message=\"OK\")\ndiff --git a/api/app/controllers/api/meta.py b/api/app/controllers/api/meta.py\nnew file mode 100644\n--- /dev/null\n+++ b/api/app/controllers/api/meta.py\n@@ -0,0 +1,24 @@\n+from flask_restful import Resource\n+\n+from app.vendors.rest import response\n+from app.helpers import helpers\n+from app.middlewares import auth\n+\n+\n+class MetaVersion(Resource):\n+ def get(self):\n+ build = helpers.read_version(\"requirements.txt\", \"build-version.txt\")\n+\n+ data = {\"build\": build}\n+ return response(200, data=data, message=\"OK\")\n+\n+\n+class MetaConfig(Resource):\n+ @auth.auth_required\n+ def get(self):\n+ config = helpers.get_config()\n+ brokers = config[\"brokers\"]\n+ clusters = config[\"knot_servers\"]\n+\n+ data = {\"knot_servers\": clusters, \"brokers\": brokers}\n+ return response(200, data=data, message=\"OK\")\ndiff --git a/api/app/helpers/command.py b/api/app/helpers/command.py\n--- a/api/app/helpers/command.py\n+++ b/api/app/helpers/command.py\n@@ -1,10 +1,7 @@\n import json\n-import os\n-import pathlib\n-\n-import yaml\n \n from app.helpers import producer\n+from app.helpers import helpers\n from app.models import model\n \n \n@@ -139,30 +136,14 @@ def set_default_zone(record_ids):\n producer.send(message)\n \n \n-def cluster_file():\n- \"\"\"Return cluster file path.\"\"\"\n- path = os.environ.get(\"RESTKNOT_CLUSTER_FILE\")\n- if not path:\n- current_path = pathlib.Path(__file__)\n- path = current_path.parents[2].joinpath(\"servers.yml\")\n-\n- is_exists = os.path.exists(path)\n- if is_exists:\n- return path\n- else:\n- raise ValueError(f\"Clustering File Not Found: {path}\")\n-\n-\n-def get_clusters():\n- \"\"\"Return cluster file content.\"\"\"\n- file_ = cluster_file()\n- clusters = yaml.safe_load(open(file_))\n- return clusters\n-\n-\n def delegate(zone, zone_id, command, agent_type):\n \"\"\"Send delegation config command with JSON structure to broker.\"\"\"\n- clusters = get_clusters()\n+ config = helpers.get_config()\n+ try:\n+ clusters = config[\"knot_servers\"]\n+ except KeyError:\n+ raise ValueError(\"Can't Knot server list in config\")\n+\n cluster = clusters[agent_type]\n \n # default for master\ndiff --git a/api/app/helpers/helpers.py b/api/app/helpers/helpers.py\n--- a/api/app/helpers/helpers.py\n+++ b/api/app/helpers/helpers.py\n@@ -1,7 +1,10 @@\n import datetime\n import pathlib\n+import os\n from functools import wraps\n \n+import yaml\n+\n from app.helpers import producer\n from app.vendors.rest import response\n \n@@ -97,3 +100,24 @@ def read_version(other_file_name, filename):\n version = \"__UNKNOWN__\"\n \n return version\n+\n+\n+def config_file():\n+ \"\"\"Return config file path.\"\"\"\n+ path = os.environ.get(\"RESTKNOT_CONFIG_FILE\")\n+ if not path:\n+ current_path = pathlib.Path(__file__)\n+ path = current_path.parents[2].joinpath(\"config.yml\")\n+\n+ is_exists = os.path.exists(path)\n+ if is_exists:\n+ return path\n+ else:\n+ raise ValueError(f\"Config File Not Found: {path}\")\n+\n+\n+def get_config():\n+ \"\"\"Return config file content.\"\"\"\n+ file_ = config_file()\n+ config = yaml.safe_load(open(file_))\n+ return config\ndiff --git a/api/app/helpers/producer.py b/api/app/helpers/producer.py\n--- a/api/app/helpers/producer.py\n+++ b/api/app/helpers/producer.py\n@@ -2,33 +2,43 @@\n import os\n \n from flask import current_app\n-from kafka import KafkaProducer\n+from confluent_kafka import Producer\n+\n+from app.helpers import helpers\n \n \n def kafka_producer():\n \"\"\"Create Kafka producer.\"\"\"\n- host = os.environ.get(\"KAFKA_HOST\")\n- port = os.environ.get(\"KAFKA_PORT\")\n- broker = f\"{host}:{port}\"\n-\n- producer = KafkaProducer(\n- bootstrap_servers=[broker],\n- value_serializer=lambda m: json.dumps(m).encode(\"utf-8\"),\n- )\n+ config = helpers.get_config()\n+ try:\n+ brokers = config[\"brokers\"]\n+ except KeyError:\n+ raise ValueError(\"Can't find brokers list in config\")\n+\n+ brokers = \",\".join(brokers)\n+ conf = {\"bootstrap.servers\": brokers}\n+ producer = Producer(**conf)\n return producer\n \n \n+def _delivery_report(err, msg):\n+ if err is not None:\n+ raise ValueError(f\"Message delivery failed: {err}\")\n+\n+\n def send(message):\n \"\"\"Send given message to Kafka broker.\"\"\"\n producer = None\n try:\n producer = kafka_producer()\n topic = os.environ.get(\"RESTKNOT_KAFKA_TOPIC\")\n- producer.send(topic, message)\n- producer.flush()\n+ encoded_message = json.dumps(message).encode(\"utf-8\")\n+ producer.produce(topic, encoded_message, callback=_delivery_report)\n except Exception as e:\n current_app.logger.error(f\"{e}\")\n raise ValueError(f\"{e}\")\n- finally:\n- if producer:\n- producer.close()\n+\n+ # Serve delivery callback queue.\n+ producer.poll(0)\n+ # Wait until all messages have been delivered\n+ producer.flush()\ndiff --git a/api/gunicorn.conf.py b/api/gunicorn.conf.py\nnew file mode 100644\n--- /dev/null\n+++ b/api/gunicorn.conf.py\n@@ -0,0 +1,13 @@\n+import os\n+import multiprocessing\n+\n+\n+def max_workers():\n+ return multiprocessing.cpu_count() * 2 + 1\n+\n+\n+host = os.environ.get(\"APP_HOST\", \"0.0.0.0\")\n+port = os.environ.get(\"APP_PORT\", \"8000\")\n+\n+bind = f\"{host}:{port}\"\n+workers = max_workers()\n", "test_patch": "diff --git a/api/tests/integration/conftest.py b/api/tests/integration/conftest.py\n--- a/api/tests/integration/conftest.py\n+++ b/api/tests/integration/conftest.py\n@@ -19,7 +19,7 @@ def clean_users():\n @pytest.fixture\n def client():\n current_path = pathlib.Path(__file__)\n- dotenv_path = current_path.parents[2].joinpath(\".env.example\")\n+ dotenv_path = current_path.parents[2].joinpath(\".example.env\")\n load_dotenv(dotenv_path)\n \n app = create_app()\n", "problem_statement": "", "hints_text": "", "created_at": "2021-04-14T10:12:08Z"}
|