Delete test.json
Browse files
test.json
DELETED
|
@@ -1,31 +0,0 @@
|
|
| 1 |
-
[
|
| 2 |
-
{
|
| 3 |
-
"repo": "mixpanel/mixpanel-python",
|
| 4 |
-
"pull_number": 64,
|
| 5 |
-
"instance_id": "mixpanel__mixpanel-python-64",
|
| 6 |
-
"issue_numbers": [
|
| 7 |
-
"63"
|
| 8 |
-
],
|
| 9 |
-
"base_commit": "40c98e0b285898384cc4aa6cc803d8d0f46f6218",
|
| 10 |
-
"patch": "diff --git a/mixpanel/__init__.py b/mixpanel/__init__.py\n--- a/mixpanel/__init__.py\n+++ b/mixpanel/__init__.py\n@@ -345,6 +345,7 @@ def send(self, endpoint, json_message, api_key=None):\n :param endpoint: the Mixpanel API endpoint appropriate for the message\n :type endpoint: \"events\" | \"people\" | \"imports\"\n :param str json_message: a JSON message formatted for the endpoint\n+ :param str api_key: your Mixpanel project's API key\n :raises MixpanelException: if the endpoint doesn't exist, the server is\n unreachable, or the message cannot be processed\n \"\"\"\n@@ -412,6 +413,7 @@ def __init__(self, max_size=50, events_url=None, people_url=None, import_url=Non\n 'imports': [],\n }\n self._max_size = min(50, max_size)\n+ self._api_key = None\n \n def send(self, endpoint, json_message, api_key=None):\n \"\"\"Record an event or profile update.\n@@ -424,16 +426,22 @@ def send(self, endpoint, json_message, api_key=None):\n :param endpoint: the Mixpanel API endpoint appropriate for the message\n :type endpoint: \"events\" | \"people\" | \"imports\"\n :param str json_message: a JSON message formatted for the endpoint\n+ :param str api_key: your Mixpanel project's API key\n :raises MixpanelException: if the endpoint doesn't exist, the server is\n unreachable, or any buffered message cannot be processed\n+\n+ .. versionadded:: 4.3.2\n+ The *api_key* parameter.\n \"\"\"\n if endpoint not in self._buffers:\n raise MixpanelException('No such endpoint \"{0}\". Valid endpoints are one of {1}'.format(endpoint, self._buffers.keys()))\n \n buf = self._buffers[endpoint]\n buf.append(json_message)\n+ if api_key is not None:\n+ self._api_key = api_key\n if len(buf) >= self._max_size:\n- self._flush_endpoint(endpoint, api_key)\n+ self._flush_endpoint(endpoint)\n \n def flush(self):\n \"\"\"Immediately send all buffered messages to Mixpanel.\n@@ -444,13 +452,13 @@ def flush(self):\n for endpoint in self._buffers.keys():\n self._flush_endpoint(endpoint)\n \n- def _flush_endpoint(self, endpoint, api_key=None):\n+ def _flush_endpoint(self, endpoint):\n buf = self._buffers[endpoint]\n while buf:\n batch = buf[:self._max_size]\n batch_json = '[{0}]'.format(','.join(batch))\n try:\n- self._consumer.send(endpoint, batch_json, api_key)\n+ self._consumer.send(endpoint, batch_json, self._api_key)\n except MixpanelException as orig_e:\n mp_e = MixpanelException(orig_e)\n mp_e.message = batch_json\n",
|
| 11 |
-
"created_at": "2016-12-22T00:07:05Z"
|
| 12 |
-
},
|
| 13 |
-
{
|
| 14 |
-
"repo": "NVIDIA/NeMo",
|
| 15 |
-
"pull_number": 162,
|
| 16 |
-
"instance_id": "NVIDIA__NeMo-162",
|
| 17 |
-
"issue_numbers": "",
|
| 18 |
-
"base_commit": "f5f09838b96ab48f40d97c100fbcfc5b7f1ac59e",
|
| 19 |
-
"patch": "diff --git a/collections/nemo_nlp/nemo_nlp/data/data_layers.py b/collections/nemo_nlp/nemo_nlp/data/data_layers.py\n--- a/collections/nemo_nlp/nemo_nlp/data/data_layers.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/data_layers.py\n@@ -683,9 +683,9 @@ def _collate_fn(self, x):\n [np.stack(x, axis=0) for x in components]\n src_ids = torch.Tensor(src_ids).long().to(self._device)\n src_segment_ids = torch.Tensor(src_segment_ids).long().to(self._device)\n- src_mask = torch.Tensor(src_mask).float().to(self._device)\n+ src_mask = torch.Tensor(src_mask).long().to(self._device)\n tgt_ids = torch.Tensor(tgt_ids).long().to(self._device)\n- tgt_mask = torch.Tensor(tgt_mask).float().to(self._device)\n+ tgt_mask = torch.Tensor(tgt_mask).long().to(self._device)\n sent_ids = torch.Tensor(sent_ids).long().to(self._device)\n return src_ids, src_segment_ids, src_mask, tgt_ids, tgt_mask, sent_ids\n \ndiff --git a/collections/nemo_nlp/nemo_nlp/data/datasets/bert_pretraining.py b/collections/nemo_nlp/nemo_nlp/data/datasets/bert_pretraining.py\n--- a/collections/nemo_nlp/nemo_nlp/data/datasets/bert_pretraining.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/datasets/bert_pretraining.py\n@@ -249,7 +249,7 @@ def truncate_seq_pair(a, b, max_num_tokens):\n \n input_ids, output_mask = self.mask_ids(output_ids)\n \n- input_mask = np.zeros(self.max_seq_length, dtype=np.float32)\n+ input_mask = np.zeros(self.max_seq_length, dtype=np.long)\n input_mask[:len(input_ids)] = 1\n \n input_type_ids = np.zeros(self.max_seq_length, dtype=np.int)\n@@ -263,7 +263,7 @@ def truncate_seq_pair(a, b, max_num_tokens):\n \n # TODO: wrap the return value with () for consistent style.\n return np.array(input_ids), input_type_ids,\\\n- np.array(input_mask, dtype=np.float32), np.array(output_ids),\\\n+ np.array(input_mask, dtype=np.long), np.array(output_ids),\\\n np.array(output_mask, dtype=np.float32), is_next\n \n def mask_ids(self, ids):\ndiff --git a/collections/nemo_nlp/nemo_nlp/data/datasets/glue.py b/collections/nemo_nlp/nemo_nlp/data/datasets/glue.py\n--- a/collections/nemo_nlp/nemo_nlp/data/datasets/glue.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/datasets/glue.py\n@@ -55,7 +55,7 @@ def __getitem__(self, idx):\n feature = self.features[idx]\n return (np.array(feature.input_ids),\n np.array(feature.segment_ids),\n- np.array(feature.input_mask, dtype=np.float32),\n+ np.array(feature.input_mask, dtype=np.long),\n np.array(feature.label_id))\n \n \ndiff --git a/collections/nemo_nlp/nemo_nlp/data/datasets/joint_intent_slot.py b/collections/nemo_nlp/nemo_nlp/data/datasets/joint_intent_slot.py\n--- a/collections/nemo_nlp/nemo_nlp/data/datasets/joint_intent_slot.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/datasets/joint_intent_slot.py\n@@ -214,7 +214,7 @@ def __len__(self):\n def __getitem__(self, idx):\n return (np.array(self.all_input_ids[idx]),\n np.array(self.all_segment_ids[idx]),\n- np.array(self.all_input_mask[idx]),\n+ np.array(self.all_input_mask[idx], dtype=np.long),\n np.array(self.all_loss_mask[idx]),\n np.array(self.all_subtokens_mask[idx]),\n self.all_intents[idx],\n@@ -263,6 +263,6 @@ def __len__(self):\n def __getitem__(self, idx):\n return (np.array(self.all_input_ids[idx]),\n np.array(self.all_segment_ids[idx]),\n- np.array(self.all_input_mask[idx], dtype=np.float32),\n+ np.array(self.all_input_mask[idx], dtype=np.long),\n np.array(self.all_loss_mask[idx]),\n np.array(self.all_subtokens_mask[idx]))\ndiff --git a/collections/nemo_nlp/nemo_nlp/data/datasets/punctuation_capitalization.py b/collections/nemo_nlp/nemo_nlp/data/datasets/punctuation_capitalization.py\n--- a/collections/nemo_nlp/nemo_nlp/data/datasets/punctuation_capitalization.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/datasets/punctuation_capitalization.py\n@@ -386,7 +386,7 @@ def __len__(self):\n def __getitem__(self, idx):\n return (np.array(self.all_input_ids[idx]),\n np.array(self.all_segment_ids[idx]),\n- np.array(self.all_input_mask[idx], dtype=np.float32),\n+ np.array(self.all_input_mask[idx], dtype=np.long),\n np.array(self.all_loss_mask[idx]),\n np.array(self.all_subtokens_mask[idx]),\n np.array(self.punct_all_labels[idx]),\ndiff --git a/collections/nemo_nlp/nemo_nlp/data/datasets/sentence_classification.py b/collections/nemo_nlp/nemo_nlp/data/datasets/sentence_classification.py\n--- a/collections/nemo_nlp/nemo_nlp/data/datasets/sentence_classification.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/datasets/sentence_classification.py\n@@ -115,7 +115,7 @@ def __getitem__(self, idx):\n \n return (np.array(feature.input_ids),\n np.array(feature.segment_ids),\n- np.array(feature.input_mask, dtype=np.float32),\n+ np.array(feature.input_mask, dtype=np.long),\n feature.sent_label)\n \n def convert_sequences_to_features(self,\ndiff --git a/collections/nemo_nlp/nemo_nlp/data/datasets/token_classification.py b/collections/nemo_nlp/nemo_nlp/data/datasets/token_classification.py\n--- a/collections/nemo_nlp/nemo_nlp/data/datasets/token_classification.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/datasets/token_classification.py\n@@ -333,7 +333,7 @@ def __len__(self):\n def __getitem__(self, idx):\n return (np.array(self.all_input_ids[idx]),\n np.array(self.all_segment_ids[idx]),\n- np.array(self.all_input_mask[idx], dtype=np.float32),\n+ np.array(self.all_input_mask[idx], dtype=np.long),\n np.array(self.all_loss_mask[idx]),\n np.array(self.all_subtokens_mask[idx]),\n np.array(self.all_labels[idx]))\n@@ -377,6 +377,6 @@ def __len__(self):\n def __getitem__(self, idx):\n return (np.array(self.all_input_ids[idx]),\n np.array(self.all_segment_ids[idx]),\n- np.array(self.all_input_mask[idx], dtype=np.float32),\n+ np.array(self.all_input_mask[idx], dtype=np.long),\n np.array(self.all_loss_mask[idx]),\n np.array(self.all_subtokens_mask[idx]))\ndiff --git a/collections/nemo_nlp/nemo_nlp/data/tokenizers/bert_tokenizer.py b/collections/nemo_nlp/nemo_nlp/data/tokenizers/bert_tokenizer.py\n--- a/collections/nemo_nlp/nemo_nlp/data/tokenizers/bert_tokenizer.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/tokenizers/bert_tokenizer.py\n@@ -1,5 +1,5 @@\n from .tokenizer_spec import TokenizerSpec\n-from pytorch_transformers import BertTokenizer\n+from transformers import BertTokenizer\n import re\n \n \ndiff --git a/collections/nemo_nlp/nemo_nlp/data/tokenizers/gpt2_tokenizer.py b/collections/nemo_nlp/nemo_nlp/data/tokenizers/gpt2_tokenizer.py\n--- a/collections/nemo_nlp/nemo_nlp/data/tokenizers/gpt2_tokenizer.py\n+++ b/collections/nemo_nlp/nemo_nlp/data/tokenizers/gpt2_tokenizer.py\n@@ -1,5 +1,5 @@\n from .tokenizer_spec import TokenizerSpec\n-from pytorch_transformers import GPT2Tokenizer\n+from transformers import GPT2Tokenizer\n \n \n class NemoGPT2Tokenizer(TokenizerSpec):\ndiff --git a/collections/nemo_nlp/nemo_nlp/huggingface/bert.py b/collections/nemo_nlp/nemo_nlp/huggingface/bert.py\n--- a/collections/nemo_nlp/nemo_nlp/huggingface/bert.py\n+++ b/collections/nemo_nlp/nemo_nlp/huggingface/bert.py\n@@ -1,10 +1,10 @@\n # Copyright (c) 2019 NVIDIA Corporation\n from typing import Optional, List\n \n-from pytorch_transformers import (BertConfig,\n- BertModel,\n- BERT_PRETRAINED_MODEL_ARCHIVE_MAP,\n- BERT_PRETRAINED_CONFIG_ARCHIVE_MAP)\n+from transformers import (BertConfig,\n+ BertModel,\n+ BERT_PRETRAINED_MODEL_ARCHIVE_MAP,\n+ BERT_PRETRAINED_CONFIG_ARCHIVE_MAP)\n \n from nemo.backends.pytorch.nm import TrainableNM\n from nemo.core.neural_modules import PretrainedModelInfo\n@@ -18,7 +18,7 @@\n class BERT(TrainableNM):\n \"\"\"\n BERT wraps around the Huggingface implementation of BERT from their\n- pytorch-transformers repository for easy use within NeMo.\n+ transformers repository for easy use within NeMo.\n \n Args:\n pretrained_model_name (str): If using a pretrained model, this should\ndiff --git a/collections/nemo_nlp/setup.py b/collections/nemo_nlp/setup.py\n--- a/collections/nemo_nlp/setup.py\n+++ b/collections/nemo_nlp/setup.py\n@@ -25,7 +25,7 @@\n 'python-dateutil<2.8.1,>=2.1',\n 'boto3',\n 'unidecode',\n- 'pytorch-transformers',\n+ 'transformers',\n 'matplotlib',\n 'h5py',\n 'youtokentome'\ndiff --git a/examples/nlp/joint_intent_slot_infer.py b/examples/nlp/joint_intent_slot_infer.py\n--- a/examples/nlp/joint_intent_slot_infer.py\n+++ b/examples/nlp/joint_intent_slot_infer.py\n@@ -2,7 +2,7 @@\n import os\n \n import numpy as np\n-from pytorch_transformers import BertTokenizer\n+from transformers import BertTokenizer\n from sklearn.metrics import confusion_matrix, classification_report\n \n import nemo\ndiff --git a/examples/nlp/joint_intent_slot_infer_b1.py b/examples/nlp/joint_intent_slot_infer_b1.py\n--- a/examples/nlp/joint_intent_slot_infer_b1.py\n+++ b/examples/nlp/joint_intent_slot_infer_b1.py\n@@ -1,7 +1,7 @@\n import argparse\n \n import numpy as np\n-from pytorch_transformers import BertTokenizer\n+from transformers import BertTokenizer\n \n import nemo\n import nemo_nlp\ndiff --git a/examples/nlp/joint_intent_slot_with_bert.py b/examples/nlp/joint_intent_slot_with_bert.py\n--- a/examples/nlp/joint_intent_slot_with_bert.py\n+++ b/examples/nlp/joint_intent_slot_with_bert.py\n@@ -3,7 +3,7 @@\n import os\n \n import numpy as np\n-from pytorch_transformers import BertTokenizer\n+from transformers import BertTokenizer\n \n import nemo\n from nemo.utils.lr_policies import get_lr_policy\ndiff --git a/examples/nlp/sentence_classification_with_bert.py b/examples/nlp/sentence_classification_with_bert.py\n--- a/examples/nlp/sentence_classification_with_bert.py\n+++ b/examples/nlp/sentence_classification_with_bert.py\n@@ -2,7 +2,7 @@\n import math\n \n import numpy as np\n-from pytorch_transformers import BertTokenizer\n+from transformers import BertTokenizer\n from torch import nn\n import torch\n \ndiff --git a/nemo/nemo/backends/pytorch/nm.py b/nemo/nemo/backends/pytorch/nm.py\n--- a/nemo/nemo/backends/pytorch/nm.py\n+++ b/nemo/nemo/backends/pytorch/nm.py\n@@ -36,7 +36,7 @@ def __init__(self, **kwargs):\n nn.Module.__init__(self) # For PyTorch API\n self._device = get_cuda_device(self.placement)\n \n- def __call__(self, force_pt=False, *input, **kwargs):\n+ def __call__(self, *input, force_pt=False, **kwargs):\n pt_call = len(input) > 0 or force_pt\n if pt_call:\n return nn.Module.__call__(self, *input, **kwargs)\ndiff --git a/scripts/get_decoder_params_from_bert.py b/scripts/get_decoder_params_from_bert.py\n--- a/scripts/get_decoder_params_from_bert.py\n+++ b/scripts/get_decoder_params_from_bert.py\n@@ -1,6 +1,6 @@\n import torch\n-from pytorch_transformers import BERT_PRETRAINED_MODEL_ARCHIVE_MAP\n-from pytorch_transformers.file_utils import cached_path\n+from transformers import BERT_PRETRAINED_MODEL_ARCHIVE_MAP\n+from transformers.file_utils import cached_path\n import argparse\n \n state_dict_mappings = {\n",
|
| 20 |
-
"created_at": "2019-12-03T01:19:14Z"
|
| 21 |
-
},
|
| 22 |
-
{
|
| 23 |
-
"repo": "mlsecproject/combine",
|
| 24 |
-
"pull_number": 103,
|
| 25 |
-
"instance_id": "mlsecproject__combine-103",
|
| 26 |
-
"issue_numbers": "",
|
| 27 |
-
"base_commit": "d662493af9f6ee7bc36c5509c277f93980009bec",
|
| 28 |
-
"patch": "diff --git a/baler.py b/baler.py\n--- a/baler.py\n+++ b/baler.py\n@@ -1,21 +1,21 @@\n import ConfigParser\n-import csv\n import datetime as dt\n import gzip\n import json\n+import logging\n import os\n-import sys\n+import re\n import requests\n+import sys\n import time\n-import re\n-from Queue import Queue\n+import unicodecsv\n import threading\n from logger import get_logger\n-import logging\n-\n+from Queue import Queue\n \n logger = get_logger('baler')\n \n+\n def tiq_output(reg_file, enr_file):\n config = ConfigParser.SafeConfigParser()\n cfg_success = config.read('combine.cfg')\n@@ -43,8 +43,8 @@ def tiq_output(reg_file, enr_file):\n outbound_data = [row for row in reg_data if row[2] == 'outbound']\n \n try:\n- bale_reg_csvgz(inbound_data, os.path.join(tiq_dir, 'raw', 'public_inbound', today+'.csv.gz'))\n- bale_reg_csvgz(outbound_data, os.path.join(tiq_dir, 'raw', 'public_outbound', today+'.csv.gz'))\n+ bale_reg_csvgz(inbound_data, os.path.join(tiq_dir, 'raw', 'public_inbound', today + '.csv.gz'))\n+ bale_reg_csvgz(outbound_data, os.path.join(tiq_dir, 'raw', 'public_outbound', today + '.csv.gz'))\n except:\n pass\n \n@@ -52,8 +52,8 @@ def tiq_output(reg_file, enr_file):\n outbound_data = [row for row in enr_data if row[2] == 'outbound']\n \n try:\n- bale_enr_csvgz(inbound_data, os.path.join(tiq_dir, 'enriched', 'public_inbound', today+'.csv.gz'))\n- bale_enr_csvgz(outbound_data, os.path.join(tiq_dir, 'enriched', 'public_outbound', today+'.csv.gz'))\n+ bale_enr_csvgz(inbound_data, os.path.join(tiq_dir, 'enriched', 'public_inbound', today + '.csv.gz'))\n+ bale_enr_csvgz(outbound_data, os.path.join(tiq_dir, 'enriched', 'public_outbound', today + '.csv.gz'))\n except:\n pass\n \n@@ -64,7 +64,7 @@ def bale_reg_csvgz(harvest, output_file):\n \"\"\" bale the data as a gziped csv file\"\"\"\n logger.info('Output regular data as GZip CSV to %s' % output_file)\n with gzip.open(output_file, 'wb') as csv_file:\n- bale_writer = csv.writer(csv_file, quoting=csv.QUOTE_ALL)\n+ bale_writer = unicodecsv.writer(csv_file, quoting=unicodecsv.QUOTE_ALL)\n \n # header row\n bale_writer.writerow(('entity', 'type', 'direction', 'source', 'notes', 'date'))\n@@ -75,7 +75,7 @@ def bale_reg_csv(harvest, output_file):\n \"\"\" bale the data as a csv file\"\"\"\n logger.info('Output regular data as CSV to %s' % output_file)\n with open(output_file, 'wb') as csv_file:\n- bale_writer = csv.writer(csv_file, quoting=csv.QUOTE_ALL)\n+ bale_writer = unicodecsv.writer(csv_file, quoting=unicodecsv.QUOTE_ALL)\n \n # header row\n bale_writer.writerow(('entity', 'type', 'direction', 'source', 'notes', 'date'))\n@@ -86,112 +86,117 @@ def bale_enr_csv(harvest, output_file):\n \"\"\" output the data as an enriched csv file\"\"\"\n logger.info('Output enriched data as CSV to %s' % output_file)\n with open(output_file, 'wb') as csv_file:\n- bale_writer = csv.writer(csv_file, quoting=csv.QUOTE_ALL)\n+ bale_writer = unicodecsv.writer(csv_file, quoting=unicodecsv.QUOTE_ALL)\n \n # header row\n bale_writer.writerow(('entity', 'type', 'direction', 'source', 'notes', 'date', 'asnumber', 'asname', 'country', 'host', 'rhost'))\n bale_writer.writerows(harvest)\n \n+\n def bale_enr_csvgz(harvest, output_file):\n \"\"\" output the data as an enriched gziped csv file\"\"\"\n logger.info('Output enriched data as GZip CSV to %s' % output_file)\n with gzip.open(output_file, 'wb') as csv_file:\n- bale_writer = csv.writer(csv_file, quoting=csv.QUOTE_ALL)\n+ bale_writer = unicodecsv.writer(csv_file, quoting=unicodecsv.QUOTE_ALL)\n \n # header row\n bale_writer.writerow(('entity', 'type', 'direction', 'source', 'notes', 'date', 'asnumber', 'asname', 'country', 'host', 'rhost'))\n bale_writer.writerows(harvest)\n \n-def bale_CRITs_indicator(base_url,data,indicator_que):\n+\n+def bale_CRITs_indicator(base_url, data, indicator_que):\n \"\"\" One thread of adding indicators to CRITs\"\"\"\n while not indicator_que.empty():\n- indicator=indicator_que.get()\n+ indicator = indicator_que.get()\n if indicator[1] == 'IPv4':\n # using the IP API\n- url=base_url+'ips/'\n- data['add_indicator']=\"true\"\n- data['ip']=indicator[0]\n- data['ip_type']='Address - ipv4-addr'\n- data['reference']=indicator[3]\n+ url = base_url + 'ips/'\n+ data['add_indicator'] = \"true\"\n+ data['ip'] = indicator[0]\n+ data['ip_type'] = 'Address - ipv4-addr'\n+ data['reference'] = indicator[3]\n # getting the source automatically:\n- source=re.findall(r'\\/\\/(.*?)\\/',data['reference'])\n+ source = re.findall(r'\\/\\/(.*?)\\/', data['reference'])\n if source:\n- data['source']=source[0]\n- res = requests.post(url,data=data,verify=False)\n- if not res.status_code in [201,200,400]:\n+ data['source'] = source[0]\n+ res = requests.post(url, data=data, verify=False)\n+ if not res.status_code in [201, 200, 400]:\n logger.info(\"Issues with adding: %s\" % data['ip'])\n elif indicator[1] == \"FQDN\":\n # using the Domain API\n- url=base_url+'domains/'\n- data['add_indicator']=\"true\"\n- data['domain']=indicator[0]\n- data['reference']=indicator[3]\n+ url = base_url + 'domains/'\n+ data['add_indicator'] = \"true\"\n+ data['domain'] = indicator[0]\n+ data['reference'] = indicator[3]\n # getting the source automatically:\n- source=re.findall(r'\\/\\/(.*?)\\/',data['reference'])\n+ source = re.findall(r'\\/\\/(.*?)\\/', data['reference'])\n if source:\n- data['source']=source[0]\n- res = requests.post(url,data=data,verify=False)\n- if not res.status_code in [201,200,400]:\n+ data['source'] = source[0]\n+ res = requests.post(url, data=data, verify=False)\n+ if not res.status_code in [201, 200, 400]:\n logger.info(\"Issues with adding: %s\" % data['domain'])\n else:\n- logger.info(\"don't yet know what to do with: %s[%s]\" % (indicator[1],indicator[0]))\n+ logger.info(\"don't yet know what to do with: %s[%s]\" % (indicator[1], indicator[0]))\n+\n \n-def bale_CRITs(harvest,filename):\n+def bale_CRITs(harvest, filename):\n \"\"\" taking the output from combine and pushing it to the CRITs web API\"\"\"\n # checking the minimum requirements for parameters\n # it would be nice to have some metadata on the feeds that can be imported in the intel library:\n # -> confidence\n # -> type of feed (bot vs spam vs ddos, you get the picture)\n- data={'confidence':'medium'}\n- start_time=time.time()\n+ data = {'confidence': 'medium'}\n+ start_time = time.time()\n config = ConfigParser.SafeConfigParser()\n cfg_success = config.read('combine.cfg')\n if not cfg_success:\n logger.error('tiq_output: Could not read combine.cfg.\\n')\n logger.error('HINT: edit combine-example.cfg and save as combine.cfg.\\n')\n return\n- if config.has_option('Baler','crits_username'):\n- data['username']=config.get('Baler', 'crits_username')\n+ if config.has_option('Baler', 'crits_username'):\n+ data['username'] = config.get('Baler', 'crits_username')\n else:\n raise 'Please check the combine.cnf file for the crits_username field in the [Baler] section'\n- if config.has_option('Baler','crits_api_key'):\n- data['api_key']=config.get('Baler', 'crits_api_key')\n+ if config.has_option('Baler', 'crits_api_key'):\n+ data['api_key'] = config.get('Baler', 'crits_api_key')\n else:\n raise 'Please check the combine.cnf file for the crits_api_key field in the [Baler] section'\n- if config.has_option('Baler','crits_campaign'):\n- data['campaign']=config.get('Baler', 'crits_campaign')\n+ if config.has_option('Baler', 'crits_campaign'):\n+ data['campaign'] = config.get('Baler', 'crits_campaign')\n else:\n logger.info('Lacking a campaign name, we will default to \"combine.\" Errors might ensue if it does not exist in CRITs')\n- data['campaign']='combine'\n- if config.has_option('Baler','crits_url'):\n- base_url=config.get('Baler','crits_url')\n+ data['campaign'] = 'combine'\n+ if config.has_option('Baler', 'crits_url'):\n+ base_url = config.get('Baler', 'crits_url')\n else:\n raise 'Please check the combine.cnf file for the crits_url field in the [Baler] section'\n- if config.has_option('Baler','crits_maxThreads'):\n- maxThreads=int(config.get('Baler', 'crits_maxThreads'))\n+ if config.has_option('Baler', 'crits_maxThreads'):\n+ maxThreads = int(config.get('Baler', 'crits_maxThreads'))\n else:\n logger.info('No number of maximum Threads has been given, defaulting to 10')\n- maxThreads=10\n+ maxThreads = 10\n \n- data['source']='Combine'\n- data['method']='trawl'\n+ data['source'] = 'Combine'\n+ data['method'] = 'trawl'\n \n # initializing the Queue to the list of indicators in the harvest\n- ioc_queue=Queue()\n+ ioc_queue = Queue()\n for indicator in harvest:\n ioc_queue.put(indicator)\n- total_iocs=ioc_queue.qsize()\n+ total_iocs = ioc_queue.qsize()\n \n for x in range(maxThreads):\n- th=threading.Thread(target=bale_CRITs_indicator, args=(base_url,data,ioc_queue))\n+ th = threading.Thread(target=bale_CRITs_indicator, args=(base_url, data, ioc_queue))\n th.start()\n \n for x in threading.enumerate():\n- if x.name==\"MainThread\":\n+ if x.name == \"MainThread\":\n continue\n x.join()\n \n- logger.info('Output %d indicators to CRITs using %d threads. Operation tool %d seconds\\n' % (total_iocs,maxThreads,time.time()-start_time))\n+ logger.info('Output %d indicators to CRITs using %d threads. Operation tool %d seconds\\n' %\n+ (total_iocs, maxThreads, time.time() - start_time))\n+\n \n def bale(input_file, output_file, output_format, is_regular):\n config = ConfigParser.SafeConfigParser()\n@@ -203,13 +208,13 @@ def bale(input_file, output_file, output_format, is_regular):\n \n logger.info('Reading processed data from %s' % input_file)\n with open(input_file, 'rb') as f:\n- harvest = json.load(f)\n+ harvest = json.load(f, encoding='utf8')\n \n # TODO: also need plugins here (cf. #23)\n if is_regular:\n- format_funcs = {'csv': bale_reg_csv,'crits':bale_CRITs}\n+ format_funcs = {'csv': bale_reg_csv, 'crits': bale_CRITs}\n else:\n- format_funcs = {'csv': bale_enr_csv,'crits':bale_CRITs}\n+ format_funcs = {'csv': bale_enr_csv, 'crits': bale_CRITs}\n format_funcs[output_format](harvest, output_file)\n \n if __name__ == \"__main__\":\ndiff --git a/winnower.py b/winnower.py\n--- a/winnower.py\n+++ b/winnower.py\n@@ -9,30 +9,36 @@\n import sys\n \n from netaddr import IPAddress, IPRange, IPSet\n+from sortedcontainers import SortedDict\n \n from logger import get_logger\n import logging\n \n logger = get_logger('winnower')\n \n+# from http://en.wikipedia.org/wiki/Reserved_IP_addresses:\n+reserved_ranges = IPSet(['0.0.0.0/8', '100.64.0.0/10', '127.0.0.0/8', '192.88.99.0/24',\n+ '198.18.0.0/15', '198.51.100.0/24', '203.0.113.0/24', '233.252.0.0/24'])\n+gi_org = SortedDict()\n+\n \n def load_gi_org(filename):\n- gi_org = {}\n with open(filename, 'rb') as f:\n org_reader = csv.DictReader(f, fieldnames=['start', 'end', 'org'])\n for row in org_reader:\n- gi_org[IPRange(row['start'], row['end'])] = row['org']\n+ gi_org[row['start']] = (IPRange(row['start'], row['end']), unicode(row['org'], errors='replace'))\n+\n return gi_org\n \n \n-def org_by_addr(address, org_data):\n+def org_by_addr(address):\n as_num = None\n as_name = None\n- for iprange in org_data:\n- if address in iprange:\n- as_num, sep, as_name = org_data[iprange].partition(' ')\n- as_num = as_num.replace(\"AS\", \"\") # Making sure the variable only has the number\n- break\n+ gi_index = gi_org.bisect(str(int(address)))\n+ gi_net = gi_org[gi_org.iloc[gi_index - 1]]\n+ if address in gi_net[0]:\n+ as_num, sep, as_name = gi_net[1].partition(' ')\n+ as_num = as_num.replace(\"AS\", \"\") # Making sure the variable only has the number\n return as_num, as_name\n \n \n@@ -46,8 +52,8 @@ def maxhits(dns_records):\n return hostname\n \n \n-def enrich_IPv4(address, org_data, geo_data, dnsdb=None):\n- as_num, as_name = org_by_addr(address, org_data)\n+def enrich_IPv4(address, geo_data, dnsdb=None):\n+ as_num, as_name = org_by_addr(address)\n country = geo_data.country_code_by_addr('%s' % address)\n if dnsdb:\n hostname = maxhits(dnsdb.query_rdata_ip('%s' % address))\n@@ -73,12 +79,9 @@ def filter_date(records, date):\n \n \n def reserved(address):\n- # from http://en.wikipedia.org/wiki/Reserved_IP_addresses:\n- ranges = IPSet(['0.0.0.0/8', '100.64.0.0/10', '127.0.0.0/8', '192.88.99.0/24',\n- '198.18.0.0/15', '198.51.100.0/24', '203.0.113.0/24', '233.252.0.0/24'])\n a_reserved = address.is_reserved()\n a_private = address.is_private()\n- a_inr = address in ranges\n+ a_inr = address in reserved_ranges\n if a_reserved or a_private or a_inr:\n return True\n else:\n@@ -138,7 +141,7 @@ def winnow(in_file, out_file, enr_file):\n \n # TODO: make these locations configurable?\n logger.info('Loading GeoIP data')\n- org_data = load_gi_org('data/GeoIPASNum2.csv')\n+ gi_org = load_gi_org('data/GeoIPASNum2.csv')\n geo_data = pygeoip.GeoIP('data/GeoIP.dat', pygeoip.MEMORY_CACHE)\n \n wheat = []\n@@ -147,23 +150,21 @@ def winnow(in_file, out_file, enr_file):\n logger.info('Beginning winnowing process')\n for each in crop:\n (addr, addr_type, direction, source, note, date) = each\n- # TODO: enrich DNS indicators as well\n if addr_type == 'IPv4' and is_ipv4(addr):\n- logger.info('Enriching %s' % addr)\n+ #logger.info('Enriching %s' % addr)\n ipaddr = IPAddress(addr)\n if not reserved(ipaddr):\n wheat.append(each)\n if enrich_ip:\n- e_data = (addr, addr_type, direction, source, note, date) + enrich_IPv4(ipaddr, org_data, geo_data, dnsdb)\n+ e_data = (addr, addr_type, direction, source, note, date) + enrich_IPv4(ipaddr, geo_data, dnsdb)\n enriched.append(e_data)\n else:\n- e_data = (addr, addr_type, direction, source, note, date) + enrich_IPv4(ipaddr, org_data, geo_data)\n+ e_data = (addr, addr_type, direction, source, note, date) + enrich_IPv4(ipaddr, geo_data)\n enriched.append(e_data)\n else:\n logger.error('Found invalid address: %s from: %s' % (addr, source))\n elif addr_type == 'FQDN' and is_fqdn(addr):\n- # TODO: validate these (cf. https://github.com/mlsecproject/combine/issues/15 )\n- logger.info('Enriching %s' % addr)\n+ #logger.info('Enriching %s' % addr)\n wheat.append(each)\n if enrich_dns and dnsdb:\n e_data = (addr, addr_type, direction, source, note, date, enrich_FQDN(addr, date, dnsdb))\n@@ -173,10 +174,12 @@ def winnow(in_file, out_file, enr_file):\n \n logger.info('Dumping results')\n with open(out_file, 'wb') as f:\n- json.dump(wheat, f, indent=2)\n+ w_data = json.dumps(wheat, indent=2, ensure_ascii=False).encode('utf8')\n+ f.write(w_data)\n \n with open(enr_file, 'wb') as f:\n- json.dump(enriched, f, indent=2)\n+ e_data = json.dumps(enriched, indent=2, ensure_ascii=False).encode('utf8')\n+ f.write(e_data)\n \n \n if __name__ == \"__main__\":\n",
|
| 29 |
-
"created_at": "2014-12-26T18:31:08Z"
|
| 30 |
-
}
|
| 31 |
-
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|