text
stringlengths
0
1.05M
meta
dict
__author__ = 'cpaulson' import pyKriging from pyKriging.krige import kriging from pyKriging.samplingplan import samplingplan # The Kriging model starts by defining a sampling plan, we use an optimal Latin Hypercube here sp = samplingplan(2) X = sp.optimallhc(15) # Next, we define the problem we would like to solve testfun = pyKriging.testfunctions().paulson1 y = testfun(X) # We can choose between a ga and a pso here optimizer = 'ga' # Now that we have our initial data, we can create an instance of a kriging model print 'Setting up the Kriging Model' k = kriging(X, y, testfunction=testfun, name='simple_ei', testPoints=300) k.train(optimizer=optimizer) k.snapshot() # Add 10 points based on model error reduction for i in range(5): newpoints = k.infill(1, method='error') for point in newpoints: print 'Adding point {}'.format(point) k.addPoint(point, testfun(point)[0]) k.train(optimizer=optimizer) k.snapshot() # Infill ten points based on the expected improvement criterion for i in range(5): newpoints = k.infill(1, method='ei') for point in newpoints: print 'Adding point {}'.format(point) k.addPoint(point, testfun(point)[0]) k.train(optimizer=optimizer) k.snapshot() # And plot the results print 'Now plotting final results...' k.plot()
{ "repo_name": "DailyActie/Surrogate-Model", "path": "01-codes/pyKriging-master/examples/2D_simple_train_expected_improvement.py", "copies": "1", "size": "1321", "license": "mit", "hash": 7022152080519651000, "line_mean": 29.7209302326, "line_max": 94, "alpha_frac": 0.7161241484, "autogenerated": false, "ratio": 3.29426433915212, "config_test": true, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.451038848755212, "avg_score": null, "num_lines": null }
__author__ = 'cpaulson' import numpy as np from numpy.matlib import ones, eye from pyKriging import kriging class coKriging(): def __init__(self, Xc, yc, Xe, ye): # Create the data arrays self.Xc = np.atleast_2d(Xc).T self.yc = yc self.nc = self.Xc.shape[0] self.Xe = np.atleast_2d(Xe).T self.ye = ye self.ne = self.Xe.shape[0] # rho regression parameter self.rho = 1.9961 self.reorder_data() # self.traincheap() self.k = self.Xc.shape[1] # if self.Xe.shape[1] != self.Xc.shape[1]: # print 'Xc and Xe must have the same number of design variables. Fatal error -- Exiting...' # exit() # Configure the hyperparameter arrays self.thetad = np.ones(self.k) self.thetac = None # self.thetac = self.kc.theta self.pd = np.ones(self.k) * 2. # self.pc = self.kc.pl self.pc = np.ones(self.k) * 2. # Matrix Operations self.one = ones([self.ne + self.nc, 1]) self.y = [self.yc, self.ye] print 'here1' def reorder_data(self): xe = [] ye = [] xc = [] yc = [] Xd = [] yd = [] for enu, entry in enumerate(self.Xc): if entry in self.Xe: print 'Found this value in XE!!' for enu1, test in enumerate(self.Xe): # if entry[0] == test[0] and entry[1] == test[1]: if entry == test: xe.append(test.tolist()) ye.append(self.ye[enu1].tolist()) xc.append(entry.tolist()) yc.append(self.yc[enu].tolist()) Xd.append(entry.tolist()) yd.append(self.ye[enu1].tolist() - self.rho * self.yc[enu].tolist()) break else: xc.insert(0, entry.tolist()) yc.insert(0, self.yc[enu].tolist()) self.Xe = np.array(xe) self.ye = np.array(ye) self.Xc = np.array(xc) self.yc = np.array(yc) self.Xd = np.array(Xd) self.yd = np.atleast_2d(np.array(yd)) def updateData(self): self.nc = self.Xc.shape[0] self.ne = self.Xe.shape[0] self.distanceXc() self.distanceXe() self.distanceXcXe() def traincheap(self): self.kc = kriging(self.Xc, self.yc) self.kc.train() print def distanceXc(self): self.distanceXc = np.zeros((self.nc, self.nc, self.k)) for i in range(self.nc): for j in xrange(i + 1, self.nc): self.distanceXc[i][j] = np.abs((self.Xc[i] - self.Xc[j])) def distanceXe(self): self.distanceXe = np.zeros((self.ne, self.ne, self.k)) for i in range(self.ne): for j in xrange(i + 1, self.ne): self.distanceXe[i][j] = np.abs((self.Xe[i] - self.Xe[j])) def distanceXcXe(self): self.distanceXcXe = np.zeros((self.nc, self.ne, self.k)) for i in range(self.nc): for j in xrange(self.ne): self.distanceXcXe[i][j] = np.abs((self.Xc[i] - self.Xe[j])) def updatePsi(self): self.PsicXc = np.zeros((self.nc, self.nc), dtype=np.float) self.PsicXe = np.zeros((self.ne, self.ne), dtype=np.float) self.PsicXcXe = np.zeros((self.nc, self.ne), dtype=np.float) # # print self.thetac # print self.pc # print self.distanceXc newPsicXc = np.exp(-np.sum(self.thetac * np.power(self.distanceXc, self.pc), axis=2)) print newPsicXc[0] self.PsicXc = np.triu(newPsicXc, 1) self.PsicXc = self.PsicXc + self.PsicXc.T + np.mat(eye(self.nc)) + np.multiply(np.mat(eye(self.nc)), np.spacing(1)) self.UPsicXc = np.linalg.cholesky(self.PsicXc) self.UPsicXc = self.UPsicXc.T print self.PsicXc[0] print self.UPsicXc exit() newPsicXe = np.exp(-np.sum(self.thetac * np.power(self.distanceXe, self.pc), axis=2)) self.PsicXe = np.triu(newPsicXe, 1) self.PsiXe = self.PsicXe + self.PsicXe.T + np.mat(eye(self.ne)) + np.multiply(np.mat(eye(self.ne)), np.spacing(1)) self.UPsicXe = np.linalg.cholesky(self.PsicXe) self.UPsicXe = self.UPsicXe.T newPsiXeXc = np.exp(-np.sum(self.thetad * np.power(self.distanceXcXe, self.pd), axis=2)) self.PsicXcXe = np.triu(newPsiXeXc, 1) def neglnlikehood(self): a = np.linalg.solve(self.UPsicXc.T, np.matrix(self.yc).T) b = np.linalg.solve(self.UPsicXc, a) c = ones([self.nc, 1]).T * b d = np.linalg.solve(self.UPsicXc.T, ones([self.nc, 1])) e = np.linalg.solve(self.UPsicXc, d) f = ones([self.nc, 1]).T * e self.muc = c / f # This only works if yc is transposed, then its a scalar under two layers of arrays. Correct? Not sure print 'y', self.yd.T a = np.linalg.solve(self.UPsicXe.T, self.yd) print 'a', a b = np.linalg.solve(self.UPsicXe, a) print 'b', b c = ones([self.ne, 1]) * b print 'c', c d = np.linalg.solve(self.UPsicXe.T, ones([self.ne, 1], dtype=float)) print d e = np.linalg.solve(self.UPsicXe, d) print e f = ones([self.ne, 1]).T * e print f self.mud = c / f a = np.linalg.solve(self.UPsicXc.T, (self.yc - ones([self.nc, 1]) * self.muc)) / self.nc b = np.linalg.solve(self.UPsicXc, a) self.SigmaSqrc = (self.yc - ones([self.nc, 1]) * self.muc).T * b print self.ne print self.mud print self.UPsicXe.T a = np.linalg.solve(self.UPsicXe.T, (self.yd - ones([self.ne, 1]) * self.mud)) / self.ne b = np.linalg.solve(self.UPsicXe, a) self.SigmaSqrd = (self.yd - ones([self.ne, 1]) * self.mud).T * b self.C = np.array([self.SigmaSqrc * self.PsicXc, self.rho * self.SigmaSqrc * self.PsicXcXe, self.rho * self.SigmaSqrc * self.PsicXeXc, np.power(self.rho, 2) * self.SigmaSqrc * self.PsicXe + self.SigmaSqrd * self.PsidXe]) np.reshape(c, [2, 2]) self.UC = np.linalg.cholesky(self.C) # self.mu=(self.one.T *(self.UC\(self.UC.T\y)))/(one'*(ModelInfo.UC\(ModelInfo.UC'\one))); def fc(X): return np.power(X[:, 0], 2) + X[:, 0] + np.power(X[:, 1], 2) + X[:, 1] def fe(X): return np.power(X[:, 0], 2) + np.power(X[:, 1], 2) if __name__ == '__main__': import samplingplan import random sp = samplingplan.samplingplan(2) X = sp.optimallhc(20) Xe = np.array(random.sample(X, 6)) yc = fc(X) ye = fe(Xe) ck = coKriging(X, yc, Xe, ye) ck.updateData() ck.updatePsi() ck.neglnlikehood()
{ "repo_name": "DailyActie/Surrogate-Model", "path": "01-codes/pyKriging-master/pyKriging/coKriging.py", "copies": "1", "size": "7053", "license": "mit", "hash": 3295755510962204700, "line_mean": 32.1126760563, "line_max": 112, "alpha_frac": 0.5185027648, "autogenerated": false, "ratio": 2.9060568603213843, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.8921071855450546, "avg_score": 0.0006975539341677225, "num_lines": 213 }
__author__ = 'cpaulson' import numpy as np from numpy.matlib import rand,zeros,ones,empty,eye from pyKriging import kriging class coKriging(): def __init__(self, Xc, yc, Xe, ye): # Create the data arrays self.Xc = np.atleast_2d(Xc).T self.yc = yc self.nc = self.Xc.shape[0] self.Xe = np.atleast_2d(Xe).T self.ye = ye self.ne = self.Xe.shape[0] # rho regression parameter self.rho = 1.9961 self.reorder_data() # self.traincheap() self.k = self.Xc.shape[1] # if self.Xe.shape[1] != self.Xc.shape[1]: # print 'Xc and Xe must have the same number of design variables. Fatal error -- Exiting...' # exit() # Configure the hyperparameter arrays self.thetad = np.ones(self.k) self.thetac = None # self.thetac = self.kc.theta self.pd = np.ones(self.k) * 2. # self.pc = self.kc.pl self.pc = np.ones(self.k) * 2. # Matrix Operations self.one=ones([self.ne+self.nc,1]) self.y=[self.yc, self.ye] print('here1') def reorder_data(self): xe = [] ye = [] xc = [] yc = [] Xd = [] yd = [] for enu,entry in enumerate(self.Xc): if entry in self.Xe: print('Found this value in XE!!') for enu1,test in enumerate(self.Xe): # if entry[0] == test[0] and entry[1] == test[1]: if entry == test: xe.append(test.tolist()) ye.append(self.ye[enu1].tolist()) xc.append(entry.tolist()) yc.append(self.yc[enu].tolist()) Xd.append(entry.tolist()) yd.append(self.ye[enu1].tolist() - self.rho * self.yc[enu].tolist()) break else: xc.insert(0,entry.tolist()) yc.insert(0,self.yc[enu].tolist()) self.Xe = np.array(xe) self.ye = np.array(ye) self.Xc = np.array(xc) self.yc = np.array(yc) self.Xd = np.array(Xd) self.yd = np.atleast_2d(np.array(yd)) def updateData(self): self.nc = self.Xc.shape[0] self.ne = self.Xe.shape[0] self.distanceXc() self.distanceXe() self.distanceXcXe() def traincheap(self): self.kc = kriging(self.Xc, self.yc) self.kc.train() print() def distanceXc(self): self.distanceXc = np.zeros((self.nc,self.nc, self.k)) for i in range( self.nc ): for j in range(i+1,self.nc): self.distanceXc[i][j] = np.abs((self.Xc[i]-self.Xc[j])) def distanceXe(self): self.distanceXe = np.zeros((self.ne,self.ne, self.k)) for i in range( self.ne ): for j in range(i+1,self.ne): self.distanceXe[i][j] = np.abs((self.Xe[i]-self.Xe[j])) def distanceXcXe(self): self.distanceXcXe = np.zeros((self.nc,self.ne, self.k)) for i in range( self.nc ): for j in range(self.ne): self.distanceXcXe[i][j] = np.abs((self.Xc[i]-self.Xe[j])) def updatePsi(self): self.PsicXc = np.zeros((self.nc,self.nc), dtype=np.float) self.PsicXe = np.zeros((self.ne,self.ne), dtype=np.float) self.PsicXcXe = np.zeros((self.nc,self.ne), dtype=np.float) # # print self.thetac # print self.pc # print self.distanceXc newPsicXc = np.exp(-np.sum(self.thetac*np.power(self.distanceXc,self.pc), axis=2)) print(newPsicXc[0]) self.PsicXc = np.triu(newPsicXc,1) self.PsicXc = self.PsicXc + self.PsicXc.T + np.mat(eye(self.nc))+np.multiply(np.mat(eye(self.nc)),np.spacing(1)) self.UPsicXc = np.linalg.cholesky(self.PsicXc) self.UPsicXc = self.UPsicXc.T print(self.PsicXc[0]) print(self.UPsicXc) exit() newPsicXe = np.exp(-np.sum(self.thetac*np.power(self.distanceXe,self.pc), axis=2)) self.PsicXe = np.triu(newPsicXe,1) self.PsiXe = self.PsicXe + self.PsicXe.T + np.mat(eye(self.ne))+np.multiply(np.mat(eye(self.ne)),np.spacing(1)) self.UPsicXe = np.linalg.cholesky(self.PsicXe) self.UPsicXe = self.UPsicXe.T newPsiXeXc = np.exp(-np.sum(self.thetad*np.power(self.distanceXcXe,self.pd), axis=2)) self.PsicXcXe = np.triu(newPsiXeXc,1) def neglnlikehood(self): a = np.linalg.solve(self.UPsicXc.T, np.matrix(self.yc).T) b = np.linalg.solve( self.UPsicXc, a ) c = ones([self.nc,1]).T * b d = np.linalg.solve(self.UPsicXc.T, ones([self.nc,1])) e = np.linalg.solve(self.UPsicXc, d) f = ones([self.nc,1]).T * e self.muc = c/f # This only works if yc is transposed, then its a scalar under two layers of arrays. Correct? Not sure print('y',self.yd.T) a = np.linalg.solve(self.UPsicXe.T, self.yd) print('a',a) b = np.linalg.solve(self.UPsicXe, a) print('b', b) c = ones([self.ne,1]) * b print('c', c) d = np.linalg.solve(self.UPsicXe.T, ones([self.ne,1], dtype=float)) print(d) e = np.linalg.solve(self.UPsicXe, d) print(e) f = ones([self.ne,1]).T * e print(f) self.mud= c/f a = np.linalg.solve(self.UPsicXc.T,(self.yc-ones([self.nc,1])*self.muc))/self.nc b = np.linalg.solve(self.UPsicXc, a) self.SigmaSqrc=(self.yc-ones([self.nc,1])*self.muc).T* b print(self.ne) print(self.mud) print(self.UPsicXe.T) a = np.linalg.solve(self.UPsicXe.T,(self.yd-ones([self.ne,1])*self.mud))/self.ne b = np.linalg.solve(self.UPsicXe, a) self.SigmaSqrd=(self.yd-ones([self.ne,1])*self.mud).T* b self.C=np.array([self.SigmaSqrc*self.PsicXc, self.rho*self.SigmaSqrc*self.PsicXcXe, self.rho*self.SigmaSqrc*self.PsicXeXc, np.power(self.rho,2)*self.SigmaSqrc*self.PsicXe+self.SigmaSqrd*self.PsidXe]) np.reshape(c,[2,2]) self.UC = np.linalg.cholesky(self.C) # self.mu=(self.one.T *(self.UC\(self.UC.T\y)))/(one'*(ModelInfo.UC\(ModelInfo.UC'\one))); def fc(X): return np.power(X[:,0], 2) + X[:,0] + np.power(X[:,1], 2) + X[:,1] def fe(X): return np.power(X[:,0], 2) + np.power(X[:,1], 2) if __name__=='__main__': from . import samplingplan import random sp = samplingplan.samplingplan(2) X = sp.optimallhc(20) Xe = np.array( random.sample(X, 6) ) yc = fc(X) ye = fe(Xe) ck = coKriging(X, yc, Xe, ye) ck.updateData() ck.updatePsi() ck.neglnlikehood()
{ "repo_name": "capaulson/pyKriging", "path": "pyKriging/coKriging.py", "copies": "1", "size": "6760", "license": "mit", "hash": -7312813251629930000, "line_mean": 30.0091743119, "line_max": 207, "alpha_frac": 0.5431952663, "autogenerated": false, "ratio": 2.7739023389413213, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.38170976052413214, "avg_score": null, "num_lines": null }
__author__ = 'cpierret' import io import struct _bpk = bytes(b'PK\003\004') _dirpk = bytes(b'PK\001\002') _firstHeader = { '[': 'Content_Types].xml', 'x': 'l/', 'd': 'ocProps/', '_': 'rels/.rels' } _excelOffsetStringDict = { 0: b'\x09\x04\x06\x00\x00\x00\x10\x00', 2080: b"Microsoft Excel 5.0 Worksheet", 2080: b"Foglio di lavoro Microsoft Exce", 2114: b"Biff5", 2121: b"Biff5" } _msoffice = b'\xD0\xCF\x11\xE0\xA1\xB1\x1A\xE1' _workbook = b'W\x00o\x00r\x00k\x00b\x00o\x00o\x00k\x00' _unpacker = struct.Struct('<HII') def isExcel(file): """ Detects if a file is in an Excel 2007+ (MS OpenXml format) or another Excel version binary file (OLE CF format). Adapted and improved from libmagic rule files: msdos and msooxml False negatives: - MS Office HTML Excel files - Some files generated by non Microsoft 3rd party applications may also not be recognized - encrypted Excel files :param file: the path to the file to be tested :return: True if Excel file detected, False otherwise (may still be a valid Excel file) """ try: with open(file, "rb") as f: bytes = f.read(4) if bytes != _bpk: # Check for formats prior to Excel 2007 for offset, magic in _excelOffsetStringDict.iteritems(): f.seek(offset) if f.read(len(magic)) == magic: return True f.seek(0) if f.read(len(_msoffice)) == _msoffice: f.seek(0x480) if f.read(len(_workbook)) == _workbook: return True f.seek(-500,io.SEEK_END) if f.read(500).find(_workbook) != -1: return True return False # possibly OpenXML Excel2007+, parse zip central directory statinfo = os.stat(file) filesize = min(statinfo.st_size,65534) blocksize = 2114 bytes = b'' f.seek(-blocksize, io.SEEK_END) while filesize>=blocksize: bytes = bytearray(f.read(blocksize)) + bytes position = bytes.find(b'PK\x05\x06') if position != -1: position + 12 direntries, dirsize, diroffset = _unpacker.unpack(bytes[position+10:position+20]) if diroffset+dirsize > statinfo.st_size: return False f.seek(diroffset) directory = bytearray(f.read(dirsize)) pos = 0 for i in xrange(direntries): if directory[pos+0:pos+4] != _dirpk: return False curs = pos+0x1b sizes = (directory[curs]*256+directory[curs+1],directory[curs+2]*256+directory[curs+3],directory[curs+4]*256+directory[curs+5]) if directory[pos+0x2E:pos+0x2E+sizes[0]] == "xl/workbook.xml": return True pos = pos + 0x2E+ sum(sizes) if pos + 0x2E > len(directory): return False return False blocksize *= 2 f.seek(-blocksize) filesize -= blocksize return False return False except IOError as e: return False if __name__ == '__main__': import sys, os _xlsExt = ['.xls','.xlsx','.xlsm','.xlam'] root = 'c:\\' print "Finding all files with an Excel extension that are not detected as an Excel file in "+root for path, subdirs, files in os.walk(root): for name in files: if name[0] != '~' and name[0] != '$' and os.path.splitext(name)[1].lower() in _xlsExt: filename = os.path.join(path, name) if not isExcel(filename): print filename
{ "repo_name": "cpierret/pyxl-magic", "path": "pyxl-magic.py", "copies": "1", "size": "4051", "license": "apache-2.0", "hash": 5605617627699429000, "line_mean": 39.1089108911, "line_max": 151, "alpha_frac": 0.5104912367, "autogenerated": false, "ratio": 3.8325449385052033, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9823258605826556, "avg_score": 0.0039555138757295, "num_lines": 101 }
__author__ = 'cpt' """Initial extraction of sas queries from wrdsapi. Idea to make into a general class to generate queries to be passed through wrdsapi. """ from . import utility as wrds_util import os import re def wrds_sas_script(download_path, dataset, year, month=0, day=0, rows=[]): """Generates a .sas file. To be executed on the WRDS server to produce the desired dataset. e.g. sample request. DATA new_data; SET crsp.dsf (where = ((year(date) between 2008 and 2008) and (month(date) between 2 and 2) and (day(date) between 2 and 2))); IF (1<= _N_<= 10000000); proc export data = new_data outfile = "~/crsp_dsf20080202rows1to10000000.tsv" dbms = tab replace; putnames = yes; run; :param download_path: path for local sas script. :param dataset: :param year: :param month: :param day: :param rows: :return [sas_file, output_file, dataset]: """ ystr = '' + ('_' + str(year)) * (year != 'all') mstr = '' + (month != 0)*('0'*(month < 10) + str(month)) dstr = '' + (day != 0)*('0'*(day < 10) + str(day)) ymdstr = ystr + mstr + dstr sas_file = 'wrds_export_' + re.sub('\.', '_', dataset) if rows: row_str = 'rows' + str(rows[0]) + 'to' + str(rows[1]) sas_file += ymdstr + row_str else: sas_file += ymdstr sas_file += '.sas' [dataset, output_file] = \ wrds_util.fix_input_name(dataset, year, month, day, rows) with open(os.path.join(download_path, sas_file), 'wb') as fd: fd.write('DATA new_data;\n') fd.write('\tSET ' + dataset) if year != 'all': where_query = ' (where = (' year_query = ('(year(' + wrds_util.wrds_datevar(dataset) + ')' + ' between ' + str(year) + ' and ' + str(year) + ')') where_query += year_query if month != 0: month_query = (' and (month(' + wrds_util.wrds_datevar(dataset) + ') between ' + str(month) + ' and ' + str(month)+')') where_query += month_query if day != 0: day_query = (' and (day(' + wrds_util.wrds_datevar(dataset) + ') between ' + str(day) + ' and ' + str(day) + ')') where_query += day_query where_query += '));\n' fd.write(where_query) else: fd.write(';\n') if rows: row_query = ('\tIF (' + str(rows[0]) + '<= _N_<= ' + str(rows[1]) + '); \n') fd.write(row_query) fd.write('\n') fd.write('proc export data = new_data\n') fd.write(('\toutfile = "~/' + output_file + '" \n' + '\tdbms = tab \n' + '\treplace; \n' + '\tputnames = yes; \n' + 'run; \n')) return [sas_file, output_file, dataset]
{ "repo_name": "Craig-PT/pywrds", "path": "pywrds/sas_query.py", "copies": "1", "size": "2980", "license": "bsd-3-clause", "hash": 5161933224125479000, "line_mean": 31.3913043478, "line_max": 79, "alpha_frac": 0.4872483221, "autogenerated": false, "ratio": 3.3596392333709133, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.43468875554709135, "avg_score": null, "num_lines": null }
__author__ = 'cpt' import os import json import re import datetime import time import math import shutil import paramiko from _wrds_db_descriptors import WRDS_DOMAIN, _GET_ALL, FIRST_DATES, \ FIRST_DATE_GUESSES, AUTOEXEC_TEXT, WRDS_USER_QUOTA from pywrds import sshlib from pywrds import utility as wrds_util from . import sas_query class WrdsSession(object): """ Class to hold a WRDS session. Extracts user info and initialises SSH connection. """ def __init__(self): """ """ # if username is not None: # self._username = username # Read in the user info from file and assign variables. this_file = os.path.abspath(__file__) self.user_path = os.path.join(this_file.split('pywrds')[0], 'pywrds') self.user_info_filename = os.path.join(self.user_path, 'user_info.txt') if os.path.exists(self.user_info_filename): with open(self.user_info_filename, 'r') as f: # r instead of rb for Python3 # compatibility # content = f.read() content = content.replace(u'\xe2\x80\x9c', u'"') content = content.replace(u'\xe2\x80\x9d', u'"') try: self.user_info = json.loads(content) except ValueError: print ('WrdsSession warning: user_info.txt file does not ' + 'conform to json format. Please address this ' + 'and reload ectools.') else: print ('WrdsSession warning: Please create a user_info.txt ' + 'file conforming to the format given in the ' + 'user_info_example.txt file.') self.download_path = os.path.join(self.user_path, 'output') if 'download_path' in self.user_info: self.download_path = self.user_info['download_path'] self.wrds_institution = [] if 'wrds_institution' in self.user_info.keys(): self.wrds_institution = self.user_info['wrds_institution'] self.wrds_username =[] if 'wrds_username' in self.user_info.keys(): self.wrds_username = self.user_info['wrds_username'] if 'last_wrds_download' not in self.user_info.keys(): self.user_info['last_wrds_download'] = {} self.last_wrds_download = self.user_info['last_wrds_download'] self.now = time.localtime() [self.this_year, self.this_month, self.today] = \ [self.now.tm_year, self.now.tm_mon, self.now.tm_mday] # Initialise SSH Client. Only works with key authentication setup. # TODO: Generalise login to cases without key authentication try: [self.ssh, self.sftp] = \ sshlib.getSSH(ssh=None, sftp=None, domain=WRDS_DOMAIN, username=self.wrds_username) except: raise Warning("Need to implement login without key authentication") def get_ymd_range(self, min_date, dataset, weekdays=1): """Gets a list of tuples [year, month, date] over which to iterate in wrds_loop. Some datasets include very large files and need to be queried at a monthly or daily frequency to prevent giant files from causing problems on the server. :param min_date: :param dataset: :param weekdays: :return ymdrange: """ [min_year, min_month, min_day] = self.min_ymd(min_date, dataset) ymdrange = [] years = xrange(min_year, self.now.tm_year+1) for year in years: frequency = wrds_util.get_loop_frequency(dataset, year) if frequency == 'Y': new_ymd = [year, 0, 0] ymdrange.append(new_ymd) elif frequency == 'M': new_ymd = [[year, x, 0] for x in range(1, 13)] ymdrange = ymdrange + new_ymd elif frequency == 'D': new_ymd = [[year, x, y] for x in range(1, 13) for y in range( 1, 32)] new_ymd = wrds_util.fix_weekdays(new_ymd, weekdays) ymdrange = ymdrange + new_ymd ymdrange = [x for x in ymdrange if x <= [self.this_year, self.this_month, self.today]] ymdrange = [x for x in ymdrange if x >= [min_year, min_month, min_day]] return ymdrange def update_user_info(self, n_files, new_files, fname, dataset, year, month=0, day=0): """update_user_info(n_files, new_files, fname, dataset, year, month=0, day=0) amends the user_info file to reflect the most recent download dates for wrds files. return """ if new_files > 0: n_files = n_files + new_files if 'last_wrds_download' not in self.user_info.keys(): self.user_info['last_wrds_download'] = {} self.user_info['last_wrds_download'][dataset] = \ year*10000 + month*100 + day with open(self.user_info_filename, 'wb') as fd: fd.write(json.dumps(self.user_info, indent=4)) else: print ('Could not retrieve: ' + fname) return def min_ymd(self, min_date, dataset): """Finds (year,month,day) at which to start wrds_loop when downloading the entirety of a dataset. It checks user_info to find what files have already been downloaded. :param min_date: :param dataset: :return [min_year, min_month, min_day]: """ if dataset in _GET_ALL: return [-1, -1, -1] if 'last_wrds_download' not in self.user_info: self.user_info['last_wrds_download'] = {} if dataset not in self.user_info['last_wrds_download']: if dataset in FIRST_DATES: self.user_info['last_wrds_download'][dataset] = FIRST_DATES[ dataset] else: self.user_info['last_wrds_download'][dataset] = 18000000 if not isinstance(min_date, (int, float)): min_date = 0 if min_date == 0: min_date = self.user_info['last_wrds_download'][dataset] min_date = str(min_date) if not min_date.isdigit() or len(min_date) != 8: min_date = 0 print ('user_info["last_wrds_download"]["' + dataset + '"]=' + min_date + ' error, should be an eight digit integer.') min_year = int(float(min_date[:4])) min_month = int(float(min_date[4:6])) min_day = int(float(min_date[6:])) if min_month == min_day == 0: min_year += 1 elif min_day == 0: min_month += 1 if min_month == 13: min_month = 1 min_year += 1 else: min_day += 1 try: wday = datetime.date(min_month, min_month, min_day).weekday() except: min_day = 1 min_month += 1 if min_month == 13: min_month = 1 min_year += 1 if min_date != 0: if min_date < 1880: min_day = 0 min_month = 0 min_year = 1880 print ('Setting min_year = 1880. This will result in ' +'many empty data files and unnecessary looping. ' +'This can be prevented by a) inputting a higher ' +'min_date or b) finding the first date at which ' +'this dataset is available on WRDS and letting ' +'Brock know so he can update the code appropriately.') elif min_date < 2050: min_day = 0 min_month = 0 min_year = int(min_date) elif 188000 < min_date < 1880000: min_month = min_date%100 min_year = (min_date - (min_date%100))/100 elif min_date < 20500000: min_day = min_date%100 min_month = (min_date%10000 - min_day)/100 min_year = (min_date - (min_date%10000))/10000 if min_date == 0: if dataset in FIRST_DATES.keys(): min_day = FIRST_DATES[dataset]%100 min_month = ((FIRST_DATES[dataset] - min_day)%10000)/100 min_year = (FIRST_DATES[dataset] - 100*min_month - min_day) / 10000 elif any(re.search(x, dataset) for x in FIRST_DATE_GUESSES.keys()): key = [x for x in FIRST_DATE_GUESSES.keys() if re.search(x, dataset)][0] if dataset in FIRST_DATE_GUESSES.keys(): key = dataset if FIRST_DATE_GUESSES[key] == -1: return [-1, -1, -1] min_day = FIRST_DATE_GUESSES[key]%100 min_month = ((FIRST_DATE_GUESSES[key]-min_day)%10000)/100 min_year = (FIRST_DATE_GUESSES[key]-100*min_month-min_day)/10000 else: min_day = 0 min_month = 0 min_year = 1880 return [min_year, min_month, min_day] def setup_wrds_key(self): """setup_wrds_key() sets up a key-based authentication on the wrds server, so that the user can log in without a password going forward. return [ssh, sftp] """ if not self.wrds_username: print('setup_wrds_key() cannot run until wrds_username is ' +'specified in the user_info.txt file.') return [None, None] institution = self.get_wrds_institution() return [self.ssh, self.sftp] def get_wrds_institution(self): """Returns the institution associated with the user's account on the wrds server. :return institution_path: """ if not self.sftp: return None try: wrds_path = self.sftp.normalize(path='.') except IOError: print ('sftp cannot resolve a path on the wrds server') return None institution_path = re.sub('/home/', '', wrds_path).split('/')[0] if self.wrds_institution != institution_path: if not self.wrds_institution: self.wrds_institution = institution_path self.user_info['wrds_institution'] = self.wrds_institution with open(self.user_info_filename, 'wb') as fd: fd.write(json.dumps(self.user_info, indent=4)) else: print ('user_info["wrds_institution"] does not ' + 'match the directory "' + institution_path + '" ' + 'found on the wrds server. ' + 'This mismatch may cause errors ' + 'in the download process.') return institution_path def find_wrds(self, filename): """Query WRDS for a list of tables available from dataset_name. E.g. setting dataset_name = 'crsp' returns a file with a list of names including "dsf" (daily stock file) and "msf" (monthly stock file). :param filename: :return: [file_list, ssh, sftp] """ tic = time.time() local_sas_file = os.path.join(self.download_path, 'wrds_dicts.sas') with open(local_sas_file, 'wb') as fd: fd.write('\tproc sql;\n') fd.write('\tselect memname\n') # optional: "select distinct memname" # fd.write('\tfrom dictionary.tables\n') fd.write('\twhere libname = "' + filename.upper() + '";\n') fd.write('\tquit;\n') for fname in ['wrds_dicts.sas', 'wrds_dicts.lst', 'wrds_dicts.log']: try: self.sftp.remove(fname) except KeyboardInterrupt: raise KeyboardInterrupt except: # TODO: Handle case when file doesn't exist explicitly. pass [put_success] = self._try_put(local_sas_file, 'wrds_dicts.sas') sas_command = 'sas -noterminal wrds_dicts.sas' [stdin, stdout, stderr] = self.ssh.exec_command(sas_command) exit_status = -1 while exit_status == -1: time.sleep(10) exit_status = stdout.channel.recv_exit_status() local_path = os.path.join(self.download_path, filename + '_dicts.lst') remote_path = ('/home/' + self.wrds_institution + '/' + self.wrds_username + '/wrds_dicts.lst') remote_files = self._try_listdir('.') remote_list = remote_files.keys() if exit_status in [0, 1] and 'wrds_dicts.lst' in remote_list: [get_success, dt] = self._try_get(local_path, remote_path) else: print('find_wrds did not generate a wrds_dicts.lst ' + 'file for input: ' + repr(filename)) try: self.sftp.remove('wrds_dicts.sas') except (IOError, EOFError, paramiko.SSHException): pass os.remove(local_sas_file) flist = [] if os.path.exists(local_path): with open(local_path, 'rb') as fd: flist = fd.read().splitlines() flist = [x.strip() for x in flist] flist = [x for x in flist if x != ''] dash_line = [x for x in range(len(flist)) if flist[x].strip('- ') == ''] if dash_line: dnum = dash_line[0] flist = flist[dnum:] return [flist] def get_wrds(self, dataset, Y, M=0, D=0, recombine=1): """Remotely download a file from the WRDS server. For example, the command x = get_wrds('crsp.msf', 2010, 6) will log in to the WRDS server, issue a query to generate a tab-separated(*) file containing the entire CRSP Monthly Stock File dataset for June 2010, then download that file to your download_path (which you can edit in the user information section above). The output x is a pair [indicator,elapsed_time] where indicator is a one if the download was successful, zero otherwise. The arguments Y, M, D stand for Year, Month, Day, respectively. Ommitting the month argument get_wrds(dataset_name, year) will retrieve a single file for the entire year. (*) Tab-separated files (tsv) tend to work slightly better than comma-separated files (csv) because sometimes company names have commas e.g. Company Name, Inc. :param dataset: :param Y: :param M: :param D: :param recombine: :return [n_files, total_rows, time_elapsed]: """ keep_going = 1 [startrow, n_files, total_rows, tic] = [1, 0, 0, time.time()] rows_per_file = wrds_util.rows_per_file_adjusted(dataset) [dset2, outfile] = wrds_util.fix_input_name(dataset, Y, M, D, []) # Check if output file in local dir, if not send request. if os.path.exists(os.path.join(self.download_path, outfile)): keep_going = 0 while keep_going: R = [startrow, startrow - 1 + rows_per_file] [dset2, outfile] = wrds_util.fix_input_name(dataset, Y, M, D, R) if not os.path.exists(os.path.join(self.download_path, outfile)): [keep_going, dt] = self._get_wrds_chunk(dataset, Y, M, D, R) if keep_going > 0: n_files += 1 if os.path.exists(os.path.join(self.download_path, outfile)): log_lines = wrds_util.get_n_lines_from_log( outfile, dname=self.download_path) n_lines = wrds_util.get_n_lines(os.path.join( self.download_path, outfile)) if log_lines > n_lines: print('get_wrds error: file "%s" has %s lines, but %s ' 'were expected.', (outfile, str(n_lines), str(log_lines))) keep_going = 0 total_rows += n_lines if n_lines < rows_per_file: keep_going = 0 if log_lines == n_lines < rows_per_file: keep_going = 0 if not (log_lines == -1 or log_lines == n_lines): print('get_wrds warning: ' +'log_lines = '+str(log_lines)) if startrow == 1: subfrom = 'rows1to' + str(rows_per_file) newname = re.sub(subfrom, '', outfile) newp2f = os.path.join(self.download_path, newname) oldp2f = os.path.join(self.download_path, outfile) os.rename(oldp2f, newp2f) else: subfrom = 'to' + str(R[-1]) subto = 'to' + str(R[0] - 1 + n_lines) newname = re.sub(subfrom, subto, outfile) oldp2f = os.path.join(self.download_path, outfile) newp2f = os.path.join(self.download_path, newname) os.rename(oldp2f, newp2f) if recombine == 1: subfrom = 'rows[0-9]*to[0-9]*\.tsv' recombine_name = re.sub(subfrom, '', outfile) wrds_util.recombine_files(recombine_name, dname=self.download_path) else: startrow += rows_per_file newname = outfile else: keep_going = 0 return [n_files, total_rows, time.time()-tic] def _get_wrds_chunk(self, dataset, Y, M=0, D=0, R=[]): """Helper fn to manage server data storage limits. Some files requested by get_wrds are too large to fit in a user's allotted space on the wrds server. For these files, get_wrds will split the request into multiple smaller requests to retrieve multiple files and run each of them through _get_wrds_chunk. If the argument "recombine" is set to its default value of 1, these files will be recombined once the loop completes. :param dataset: :param Y: :param M: :param D: :param R: :return [success, time_elapsed]: """ tic = time.time() [sas_file, outfile, dataset] = \ sas_query.wrds_sas_script(self.download_path, dataset, Y, M, D, R) log_file = re.sub('\.sas$', '.log', sas_file) put_success = self._put_sas_file(outfile, sas_file) exit_status = self._sas_step(sas_file, outfile) exit_status = self._handle_sas_failure(exit_status, outfile, log_file) if exit_status in [0, 1]: remote_files = self._try_listdir('.') file_list = remote_files.keys() if outfile not in file_list: print('exit_status in [0, 1] suggests SAS succeeded, but the ' 'desired output_file %s is not present in the file ' 'list:', outfile) print(file_list) else: remote_size = self._wait_for_sas_file_completion(outfile) [get_success, dt] = self._retrieve_file(outfile, remote_size) local_size \ = wrds_util.wait_for_retrieve_completion(outfile, get_success) compare_success = \ self._compare_local_to_remote(outfile, remote_size, local_size) got_log = self._get_log_file(log_file, sas_file) checkfile = os.path.join(self.download_path, outfile) if os.path.exists(checkfile) or exit_status == 0: return [1, time.time()-tic] return [0, time.time()-tic] def _rename_after_download(self): return NotImplementedError def wrds_loop(self, dataset, min_date=0, recombine=1): """Executes get_wrds(database_name,...) over all years and months for which data is available for the specified data set. File separated into chunks for downloading will be recombined into their original forms if recombine is set to its default value 1. :param dataset: :param min_date: :param recombine: :return [n_files, time_elapsed]: """ tic = time.time() [n_files, n_lines, n_lines0] = [0, 0, 0] [min_year, min_month, min_day] = self.min_ymd(min_date, dataset) flist = os.listdir(self.download_path) if [min_year, min_month, min_day] == [-1, -1, -1]: Y = 'all' get_output = self.get_wrds(dataset, Y, M=0, D=0, recombine=recombine) [new_files, total_lines, dt] = self.get_output if new_files > 0: n_files += 1 return [n_files, time.time()-tic] for ymd in self.get_ymd_range(min_date, dataset, 1): [Y, M, D] = ymd [dset2, outfile] = wrds_util.fix_input_name(dataset, Y, M, D, []) if outfile in flist: continue get_output = self.get_wrds(dataset, Y, M=M, D=D, recombine=recombine) [new_files, total_lines, dt] = self.get_output n_files += new_files self.update_user_info(n_files, new_files, fname=outfile, dataset=dataset, year=Y, month=M, day=D) return [n_files, time.time()-tic] def _put_sas_file(self, outfile, sas_file): """Puts sas_file in home directory on wrds server, checks autoexec and removes existing run and log scripts. Checks autoexec file present, and adds if not. Removes all previous sas and log files with the wrds_export prefix and any result files with the rows<x>to<y>.tsv format. 1. Removes old files which may interfere with the new files. Assumes export files in format wrds_export_<outfile>.sas, 2. Checks enough space in user account on wrds server to run sas_file. 3. Checks necessary autoexec.sas files are present in the directory. :param outfile: :param sas_file: :return put_success (bool): """ remote_files = self._try_listdir('.') initial_files = remote_files.values() # 1. Removes old files, both .sas and .log files with wrds_export prefix old_export_files = \ [x for x in initial_files if re.search('wrds_export.*sas$', x.filename) or re.search('wrds_export.*log$', x.filename) or x.filename == sas_file] for old_file in old_export_files: try: self.sftp.remove(old_file.filename) except (IOError, EOFError, paramiko.SSHException): pass initial_files.remove(old_file) # Catch the row fragment of any output files, e.g. rows1to1000000.tsv. pattern = '[0-9]*rows[0-9]+to[0-9]+\.tsv$' old_outfiles = [x for x in initial_files if re.sub(pattern, '', x.filename) == re.sub(pattern, '', outfile)] for old_file in old_outfiles: try: self.sftp.remove(old_file.filename) except (IOError, EOFError, paramiko.SSHException): pass initial_files.remove(old_file) # TODO: see if the file is something want before deleting it. # 2. Check available space on remote. file_sizes = [initial_file.st_size for initial_file in initial_files] total_file_size = sum(file_sizes) if total_file_size > WRDS_USER_QUOTA: MBs = int(float(total_file_size)/1000000) print('You are using approximately ' + str(MBs) + ' megabytes of' ' your 1 GB quota on the WRDS server. This may cause ' 'WrdsSession.get_wrds to operate incorrectly. The files ' 'present are: ') print([x.filename for x in initial_files]) # 3. Check necessary autoexec.sas files are present on remote. auto_names = ['autoexec.sas', '.autoexecsas'] autoexecs = [x.filename for x in initial_files if x.filename in auto_names] if autoexecs == ['.autoexecsas']: # if 'autoexec.sas' is not present, the sas program will fail a # backup copy is stored by default in .autoexecsas ssh_command = 'cp .autoexecsas autoexec.sas' [exec_succes, stdin, stdout, stderr] = self._try_exec(ssh_command) elif autoexecs == ['autoexec.sas']: ssh_command = 'cp autoexec.sas .autoexecsas' [exec_succes, stdin, stdout, stderr] = self._try_exec(ssh_command) elif autoexecs == []: with open('autoexec.sas', 'wb') as fd: fd.write(AUTOEXEC_TEXT) local_path = 'autoexec.sas' remote_path = 'autoexec.sas' [put_success] = self._try_put(local_path, remote_path) ssh_command = 'cp autoexec.sas .autoexecsas' [exec_succes, stdin, stdout, stderr] = self._try_exec(ssh_command) os.remove('autoexec.sas') local_path = os.path.join(self.download_path, sas_file) remote_path = sas_file return self._try_put(local_path, remote_path) def _sas_step(self, sas_file, outfile): """Wraps running of sas command (_run_sas_command). TODO: Retrying and re-initializing the network connection if necessary. :param sas_file: :param outfile: :return exit_status: """ [sas_completion, n_sas_trys, max_sas_trys] = [0, 0, 3] while sas_completion == 0 and n_sas_trys < max_sas_trys: exit_status = self._run_sas_command(sas_file, outfile) n_sas_trys += 1 sas_completion = 1 if exit_status in [42, 104]: # 42 = network read failed, 104 = connection reset by peer # TODO: Deal with reinitiating a session - this will break. sas_completion = 0 if not self.sftp: return exit_status remote_files = self._try_listdir('.') if outfile in remote_files.keys(): exit_status = 0 sas_completion = 1 elif 'log_file' in remote_files.keys(): exit_status = -1 sas_completion = 1 return exit_status def _run_sas_command(self, sas_file, outfile): """Executes sas_file on wrds server. Waits for return of exit status. :param sas_file: :param outfile: :return exit_status: """ sas_command = ('sas -noterminal ' + sas_file) [stdin, stdout, stderr] = self.ssh.exec_command(sas_command) [exit_status, exit_status2, waited, maxwait] = [-1, -1, 0, 1200] while exit_status == -1 and waited < maxwait: time.sleep(10) waited += 10 exit_status = stdout.channel.recv_exit_status() if waited >= maxwait: print('get_wrds stopped waiting for SAS completion at step 1: ' + outfile) return exit_status def _handle_sas_failure(self, exit_status, outfile, log_file): """Checks sas exit status returned by wrds server and responds appropriately to any statuses other than success. :param exit_status: :param outfile: :param log_file: :return exit_status: """ real_failure = 1 remote_files = self._try_listdir('.') if exit_status == 2 and log_file in remote_files.keys(): with self.sftp.file(log_file) as fd: logcontent = fd.read() if re.search('error: file .* does not exist.', logcontent, flags=re.I): real_failure = 0 if exit_status not in [0, 1] and real_failure == 1: # 1 is "SAS system issued warnings", non-fatal # if outfile in remote_files.keys(): print('SAS is apparently returning an incorrect exit status: ' + str(exit_status) + ', ' + outfile + '. ectools is ' + 'downloading the file for user inspection.') remote_path = outfile local_path = os.path.join(self.download_path, outfile) [get_success, dt] = self._try_get(local_path, remote_path) if get_success == 0: print('File download failure.') else: print('get_wrds failed on file "' + outfile + '"\n' + 'exit_status = ' + str(exit_status) + '\n' + 'For ' 'details, see log file "' + log_file + '"') return exit_status def _wait_for_sas_file_completion(self, outfile): """Checks size of outfile on the wrds server within get_wrds. Until it observes two successive measurements with the same file size, it infers that the sas script is still writing the file. :param outfile: :return remote_size: """ [remote_size, remote_size_delayed, mtime, total_wait, max_wait] \ = [0, 1, time.time(), 0, 1200] while self.sftp and ((total_wait < max_wait) and remote_size != remote_size_delayed): remote_size = remote_size_delayed time.sleep(10) total_wait += 10 try: output_stat = self.sftp.stat(outfile) remote_size_delayed = output_stat.st_size mtime = output_stat.st_mtime except (IOError, EOFError, paramiko.SSHException): raise NotImplementedError # TODO: ssh reconnect if total_wait >= max_wait: print('get_wrds stopped waiting for SAS completion at step 2', + remote_size + remote_size_delayed + mtime) remote_size = 0 # should i remove the file in this case? return remote_size def _retrieve_file(self, outfile, remote_size): """Retrieves the outfile produced on the wrds server in get_wrds, including correct handling of several common network errors. :param outfile: :param remote_size: :return get_success: """ tic = time.time() if remote_size == 0: return [0, time.time()-tic] if remote_size >= 10**7: # skip messages for small files print('starting retrieve_file: ' + outfile + ' (' + repr( remote_size) + ') bytes') vfs = os.statvfs(self.download_path) free_local_space = vfs.f_bavail * vfs.f_frsize if remote_size > free_local_space: print('get_wrds cannot download file ' + outfile + ', only ' + str(free_local_space)+' bytes available on drive for ' + str(remote_size) + '-byte file.') return [0, time.time()-tic] remote_path = ('/home/' + self.wrds_institution + '/' + self.wrds_username + '/' + outfile) write_file = '.' + outfile + '--writing' local_path = os.path.join(os.path.expanduser('~'), write_file) [get_success, dt] = self._try_get(local_path, remote_path) print('retrieve_file: ' + repr(outfile) + ' ('+repr(remote_size) + ' bytes) ' + ' time elapsed=' + repr(time.time()-tic)) return [get_success, time.time()-tic] def _compare_local_to_remote(self, outfile, remote_size, local_size): """Compares the size of the file "outfile" downloaded (local_size) to the size of the file as listed on the server (remote_size) to check download completed properly. :param outfile: :param remote_size: :param local_size: :return compare_success (bool): """ compare_success = 0 write_file = '.' + outfile + '--writing' local_path = os.path.join(os.path.expanduser('~'), write_file) if remote_size == local_size != 0: [exec_succes, stdin, stdout, stderr] = \ self._try_exec('rm ' + outfile) to_path = os.path.join(self.download_path, outfile) shutil.move(local_path, to_path) compare_success = 1 elif local_size != 0: print(['remote_size != local_size', outfile, remote_size, local_size]) log_size = math.log(local_size, 2) if log_size == int(log_size): print('The error appears to involve ' +'the download stopping at 2^' + repr(log_size) + ' bytes.') error_file = '.' + outfile + '--size_error' from_file = os.path.join(os.path.expanduser('~'), error_file) to_file = os.path.join(self.download_path, outfile) shutil.move(from_file, to_file) compare_success = 0 return compare_success def _get_log_file(self, log_file, sas_file): """Attempts to retrieve SAS log file generated by _get_wrds_chunk from the WRDS server. Also removes the sas_file from the local directory, though strictly speaking this belongs in a separate function. :param log_file: :param sas_file: :return success (bool): """ success = 0 remote_path = ('/home/' + self.wrds_institution + '/' + self.wrds_username + '/' + log_file) local_path = os.path.join(self.download_path, log_file) [success, dt] = \ self._try_get(local_path, remote_path) [exec_succes, stdin, stdout, stderr] = self._try_exec('rm ' + sas_file) [exec_succes, stdin, stdout, stderr] = self._try_exec('rm wrds_export*') saspath = os.path.join(self.download_path, sas_file) if os.path.exists(saspath): os.remove(saspath) return [success] def _try_put(self, local_path, remote_path, domain=None, username=None, ports=[22]): """Transfers file from local_path to remote_path using the sftp client. TODO: Reinitiating the ssh connection if needbe. :param local_path: :param remote_path: :param domain: :param username: :param ports: :return success: """ local_stat = os.stat(local_path) [success, n_tries, max_tries] = [0, 0, 3] while not success and n_tries < max_tries: try: remote_attrs = self.sftp.put(local_path, remote_path) # Check file transferred is same as local version if remote_attrs.st_size == local_stat.st_size: success = 1 except KeyboardInterrupt: try: self.sftp.remove(remote_path) except: pass raise KeyboardInterrupt except (IOError, EOFError, paramiko.SSHException): # TODO: Handle sftp error, try to reconnect. try: self.sftp.remove(remote_path) except (IOError, EOFError, paramiko.SSHException): pass n_tries += 1 return [success] def _try_get(self, local_path, remote_path, domain=None, username=None, ports=[22]): """Tries three times to download file from remote_path to local_path using the sftp client. TODO: If a connection error occurs, it is re-established. Does *not* check that the remote file exists, that the local_path is not already in use, or that there is enough space free on the local disk to complete the download. :param local_path: :param remote_path: :param domain: :param username: :param ports: :return [success (bool), time_elapsed]: """ tic = time.time() [success, n_tries, max_tries] = [0, 0, 3] while not success and n_tries < max_tries: try: self.sftp.get(remotepath=remote_path, localpath=local_path) success = 1 except (paramiko.SSHException, paramiko.SFTPError, IOError, EOFError): if os.path.exists(local_path): os.remove(local_path) # TODO: Handle sftp error, try to reconnect. n_tries += 1 except KeyboardInterrupt: if os.path.exists(local_path): os.remove(local_path) raise KeyboardInterrupt return [success, time.time()-tic] def _try_exec(self, command, domain=None, username=None, ports=[22]): """ :param command: :param domain: :param username: :param ports: :return [success, stdin, stdout, stderr]: """ [success, n_tries, max_tries] = [0, 0, 3] [stdin, stdout, stderr] = [None, None, None] while not success and n_tries < max_tries: try: [stdin, stdout, stderr] = self.ssh.exec_command(command) success = 1 except (IOError, EOFError, paramiko.SSHException): # TODO: Handle sftp error, try to reconnect. n_tries += 1 return [success, stdin, stdout, stderr] def _try_listdir(self, remote_dir, domain=None, username=None, ports=[22]): """Tries three times to get a a list of files and their attributes from the directory remote_dir on the remote server. TODO: reinitiating the ssh connection if needbe. :param remote_dir: :param domain: :param username: :param ports: :return (dict): {filename: [attributes]} across files in the remote directory """ remote_list = [] [success, n_tries, max_tries] = [0, 0, 3] while not success and n_tries < max_tries: try: remote_list = self.sftp.listdir_attr(remote_dir) success = 1 except (IOError, EOFError, paramiko.SSHException): # TODO: Handle sftp error n_tries += 1 return {x.filename: x for x in remote_list}
{ "repo_name": "Craig-PT/pywrds", "path": "pywrds/wrdsapi.py", "copies": "1", "size": "38936", "license": "bsd-3-clause", "hash": -4342997565028721000, "line_mean": 39.5583333333, "line_max": 88, "alpha_frac": 0.535006164, "autogenerated": false, "ratio": 3.932134922237932, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9956108839167084, "avg_score": 0.0022064494141695507, "num_lines": 960 }
__author__ = 'cpt' """ Utility functions for WRDS manipulations. Generally, taking outside of wrdsapi any static methods. """ import re import datetime import os import time def rows_per_file_adjusted(dataset): """ Chooses a number of rows to query to ensure that the files produced do not approach the 1 GB server limit on WRDS. For most datasets, 10^7 rows in a file is not a problem. For optionm.opprcd, this number is dropped to 10^6. To date optionm.opprcd is the only dataset for which this has consistently been necessary. This is subject to change with further use cases. :param dataset: :return rows_per_file: """ rows_per_file = 10**7 if dataset.replace('.', '_') == 'optionm_opprcd': rows_per_file = 10**6 elif dataset.replace('.', '_') == 'optionm_optionmnames': rows_per_file = 10**6 return rows_per_file def get_loop_frequency(dataset, year): """Finds the best frequency at which to query the server for the given dataset so as to avoid producing problematically large files. :param dataset: :param year: :return frequency: """ frequency = 'Y' if dataset.startswith('optionm'): if year < 2008: frequency = 'M' else: frequency = 'D' elif re.search('det_xepsus', dataset, flags=re.I): if year > 2005: frequency = 'M' elif re.search('det_xepsint', dataset, flags=re.I): if year > 2003: frequency = 'M' elif re.search('taq', dataset, flags=re.I): frequency = 'D' return frequency def fix_weekdays(ymds, weekdays=1): """Takes a set of [year, month, date] tuples "ymds" and removes those which are not valid days, e.g. June 31, February 30. If weekdays is set to its default value of 1, it also removes Saturday and Sundays. :param ymds: :param weekdays: :return ymds: """ ymds2 = [] for [y, m, d] in ymds: try: wday = datetime.date(y, m, d).weekday() except ValueError: wday = -1 if weekdays == 1 and wday in range(5): # weekdays==1 --> only keep weekdays # ymds2.append([y, m, d]) elif weekdays == 0 and wday != -1: # weekdays==0 --> keey any valid day # ymds2.append([y, m, d]) return ymds2 def fix_input_name(dataset, year, month, day, rows=[]): """Adjusts the user-supplied dataset name to use the same upper/lower case conventions as WRDS does. :param dataset: :param year: :param month: :param day: :param rows: :return [dataset, output_file]: """ [Y, M, D, R] = [year, month, day, rows] if year != 'all': ystr = '_' * (dataset[-1].isdigit()) + str(Y) mstr = '' + (M != 0) * ('0' * (month < 10) + str(M)) dstr = (D != 0)*('0'*(D < 10) + str(D)) ymdstr = ystr + mstr + dstr + '.tsv' output_file = re.sub('\.', '_', dataset) + ymdstr else: output_file = re.sub('\.', '_', dataset) + '.tsv' if dataset.lower() == 'optionm.opprcd': dataset += str(year) elif dataset.lower() in ['taq.cq', 'taq.ct']: dataset = re.sub('cq', 'CQ', dataset) dataset = re.sub('ct', 'CT', dataset) ystr = '_' + str(Y) mstr = '' + (M != 0)*('0'*(M < 10) + str(M)) dstr = '' + (D != 0)*('0'*(D < 10) + str(D)) ymdstr = ystr + mstr + dstr dataset += ymdstr elif dataset.lower() in ['taq.mast', 'taq.div']: ymdstr = '_' + str(Y) + (M != 0)*('0'*(M < 10) + str(M)) dataset += ymdstr elif dataset.lower() == 'taq.rgsh': ymdstr = str(100*Y + M)[2:] dataset = 'taq.RGSH' + ymdstr if R: rowstr = 'rows' + str(R[0]) + 'to' + str(R[1]) + '.tsv' output_file = re.sub('.tsv$', '', output_file) + rowstr return [dataset, output_file] def wrds_datevar(filename): """Different datasets in WRDS use different names for their date-variables. Function returns the right date variable for each dataset. This may need periodic updating. Crowdsourcing is welcome. :param filename: :return date_var: """ if filename in ['tfn.s12', 'tfn.s34']: return 'fdate' if re.search('^crsp', filename): return 'date' if re.search('^comp', filename): return 'DATADATE' if re.search('^optionm\.opprcd', filename): return 'date' if re.search('^optionm', filename): return 'effect_date' if re.search('^ibes', filename): return 'anndats' return 'date' def wait_for_retrieve_completion(outfile, get_success, max_wait=1200): """Checks size of downloaded outfile until two successive give the same result. Until this point, it infers that the download is still in progress. :param outfile: :param get_success: :param max_wait: :return: local_size """ if get_success == 0: return 0 [total_wait, local_size, local_size_delayed, mtime2] = \ [0, 0, 1, time.time()] write_file = '.' + outfile + '--writing' local_path = os.path.join(os.path.expanduser('~'), write_file) while total_wait < max_wait and local_size != local_size_delayed: local_size = local_size_delayed time.sleep(5) total_wait += 5 local_stat = os.stat(local_path) local_size_delayed = local_stat.st_size mtime2 = local_stat.st_mtime if total_wait >= max_wait: print('get_wrds stopped waiting for SAS completion at step 3, %s, %s, ' '%s', (local_size, local_size_delayed, mtime2)) local_size = 0 return local_size def get_n_lines(path2file): """Returns number of lines for a text files located at path2file. :param path2file: :return n_lines: """ with open(path2file, 'rb') as fd: fsize = os.stat(fd.name).st_size n_lines = 0 first_line = fd.readline().split('\t') while fd.tell() < fsize: fline = fd.readline() n_lines += 1 return n_lines def get_n_lines_from_log(outfile, dname): """Reads SAS log file created in get_wrds to find the number of lines which the wrds server says should be in a downloaded file "outfile". This number can then be checked against the number actually found in the file. :param outfile: :param dname: :return logfile_lines: """ log_lines = -1 sasfile = 'wrds_export_' + re.sub('\.tsv$', '.log', outfile) if not os.path.exists(os.path.join(dname, sasfile)): partial_fname = re.sub('[0-9]*rows.*', '', sasfile) sasfile2 = partial_fname+'_'+re.sub(partial_fname, '', sasfile) if os.path.exists(os.path.join(dname, sasfile2)): sasfile = sasfile2 all_fname = re.sub('rows', '_allrows', sasfile) if os.path.exists(os.path.join(dname, all_fname)): sasfile = all_fname if os.path.exists(os.path.join(dname, sasfile)): with open(os.path.join(dname, sasfile)) as fd: fsize = os.stat(fd.name).st_size while fd.tell() < fsize: fline = fd.readline() if re.search('^[0-9]* records created in ', fline): log_lines = re.split(' records created in ', fline)[0] log_lines = int(float(log_lines)) break pattern0 = ('NOTE: The data set WORK\.NEW_DATA ' +'has [0-9]* observations') if re.search(pattern0, fline): pattern01 = 'NOTE: The data set WORK\.NEW_DATA has ' pattern02 = ' observations' split_log = re.split(pattern02, fline)[0] log_lines = re.split(pattern01, split_log)[-1] log_lines = int(float(log_lines)) break pattern1 = 'NOTE: [0-9]* records were written to the file' if re.search(pattern1, fline): split_log = re.split('NOTE: ', fline)[-1] log_lines = re.split('records', split_log)[0] log_lines = int(float(log_lines)) break # The numbers given by the pattern below are often # # one row lower than the numbers given by the above # # patterns, the latter being the desired answer. # # This code is kept as an option to re-implement # # should their arise cases where none of the other # # patterns are found. # #pattern2 = 'NOTE: There were [0-9]* observations read' #if re.search(pattern2,fline): # split_log = re.split(' observations read',fline)[0] # log_lines = re.split('NOTE: There were ',split_log)[-1] # log_lines = int(float(log_lines)) # break return log_lines def _recombine_ready(fname, dname=None, suppress=0): """Checks files downloaded by get_wrds to see if the loop has completed successfully and the files are ready to be be recombined. If dname == None, the directory defaults to os.getcwd(). :param fname: :param dname: :param suppress: :return isready (bool): """ if not dname: dname = os.getcwd() isready = 1 fname0 = re.sub('rows[0-9][0-9]*to[0-9][0-9]*\.tsv', '', fname) if os.path.exists(os.path.join(dname, fname + '.tsv')): isready = 0 rows_per_file = rows_per_file_adjusted(fname0) flist0 = os.listdir(dname) flist0 = [x for x in flist0 if x.endswith('.tsv')] flist0 = [x for x in flist0 if re.search(fname0, x)] fdict = {x: x.split('rows')[-1] for x in flist0} fdict = {x: re.split('_?to_?',fdict[x])[0] for x in fdict} fdict = {x: float(fdict[x]) for x in fdict if fdict[x].isdigit()} flist = [[fdict[x], x] for x in fdict] if isready and flist == []: isready = 0 if suppress == 0: print('recombine_ready: No such files found: ' + fname) numlist = [x[0] for x in sorted(flist)] missing_nums = [x for x in numlist if x != 1] missing_nums = [x for x in missing_nums if x-rows_per_file not in numlist] if isready and missing_nums != []: isready = 0 if suppress == 0: print('recombine_ready: ' + fname + ' missing_nums ' + repr(missing_nums+numlist)) end_nums = [re.sub('\.tsv$', '', x[1]) for x in flist] end_nums = [re.split('to', x)[-1] for x in end_nums] end_nums = [float(x) for x in end_nums] if isready and end_nums != [] and max(end_nums)%rows_per_file == 0: max_num = int(max(end_nums)) flist2 = [x[1] for x in flist if x[1].endswith(repr(max_num) + '.tsv')] if len(flist2) == 1: outfile = flist2[0] n_lines = get_n_lines(os.path.join(dname, outfile)) log_n_lines = get_n_lines_from_log(outfile, dname) if n_lines != log_n_lines: isready = 0 print('recombine_ready: '+outfile +' n_lines!=log_n_lines: ' +repr([n_lines, log_n_lines])) else: isready = 0 if suppress == 0: print('recombine_ready: ' + fname + ' appears incomplete: ' + repr(max(end_nums))) return isready def recombine_files(fname, dname=None, suppress=0): """Reads the files downloaded by get_wrds and combines them back into the single file of interest. If dname==None, the directory defaults to os.getcwd(). :param fname: :param dname: :param suppress: :return num_combined_files: """ if not dname: dname = os.getcwd() combined_files = 0 if not _recombine_ready(fname, dname, suppress): return combined_files fname0 = re.sub('rows[0-9][0-9]*to[0-9][0-9]*\.tsv', '', fname) rows_per_file = rows_per_file_adjusted(fname0) flist0 = [x for x in os.listdir(dname) if re.search(fname0, x)] flist0 = [x for x in flist0 if x.endswith('.tsv')] fdict = {x: x.split('rows')[-1] for x in flist0} fdict = {x: re.split('_?to_?',fdict[x])[0] for x in fdict} fdict = {x: float(fdict[x]) for x in fdict if fdict[x].isdigit()} flist = [[fdict[x], x] for x in fdict] flist = [x[1] for x in sorted(flist)] with open(os.path.join(dname, flist[-1]), 'rb') as fd: fsize = os.stat(fd.name).st_size nlines = 0 while fd.tell() > fsize: fd.readline() nlines += 1 if nlines >= rows_per_file: print([fname, flist[-1], 'len(flines)=' + repr(nlines), 'should_be=' + repr(rows_per_file)]) return combined_files with open(os.path.join(dname, fname0+'.tsv'), 'wb') as fd: headers = [] found_problem = 0 for fname1 in flist: fd1 = open(os.path.join(dname, fname1), 'rb') fsize1 = os.stat(fd1.name).st_size headers1 = fd1.readline().strip('\r\n') if headers == []: headers = headers1 fd.write(headers1 + '\n') if headers1 != headers: print('Problem with header matching:' + fname1) found_problem = 1 if found_problem == 0: try: while fd1.tell() < fsize1: fd.write(fd1.readline().strip('\r\n') + '\n') fd1.close() except KeyboardInterrupt: fd1.close() fd.close() os.remove(fd.name) raise KeyboardInterrupt combined_files += 1 if found_problem == 0: for fname1 in flist: os.remove(os.path.join(dname, fname1)) return combined_files
{ "repo_name": "Craig-PT/pywrds", "path": "pywrds/utility.py", "copies": "1", "size": "14052", "license": "bsd-3-clause", "hash": 4018626762907152000, "line_mean": 32.7788461538, "line_max": 79, "alpha_frac": 0.5450469684, "autogenerated": false, "ratio": 3.4730598121601584, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4518106780560158, "avg_score": null, "num_lines": null }
TICKER = 1 WAVE_VELOCITY = 10 from Tkinter import * top = Tk() top.title('Doppler effect') ca = Canvas(bd=0,highlightthickness=0,bg="white") ca.pack(expand=YES,fill=BOTH) ca.config(width=640,height=480) class Wave: def __init__(self, ca, vel, x, y): self.life = 0 self.canvas = ca self.vel = vel self.pos = [x,y] self.__radius = 0 self.__index = self.__get_index() self.update() def __get_index(self): return self.canvas.create_oval(*self.get_coords(),outline="#00aacc") def get_coords(self): return [ self.pos[0]-self.__radius, # x1 self.pos[1]-self.__radius, # y1 self.pos[0]+self.__radius, # x2 self.pos[1]+self.__radius ] def update(self): self.canvas.coords(self.__index, *self.get_coords()) self.__radius += self.vel self.life += 1 if self.life <= 100: self.canvas.after(TICKER,self.update) else: self.canvas.delete(self.__index) class Particle: def __init__(self, **kws): self.canvas = kws.get('canvas') self.kick = True self.pos = kws.get('pos',[]) self.freq = kws.get('freq', 0) self.vel = kws.get('vel',[0,0]) self.__radius = 5 self.__index = self.__get_index() self.update() def get_coords(self): return [ self.pos[0]-self.__radius, self.pos[1]-self.__radius, self.pos[0]+self.__radius, self.pos[1]+self.__radius ] def __get_index(self): return self.canvas.create_oval(*self.get_coords(),fill='red') def update(self): self.canvas.update_idletasks() self.pos[0] += self.vel[0] self.pos[1] += self.vel[1] self.canvas.coords(self.__index, *self.get_coords()) self.canvas.after(TICKER, self.update) Wave(self.canvas, WAVE_VELOCITY, self.pos[0], self.pos[1]) if not self.kick: if self.pos[0] >= self.canvas.winfo_width(): self.pos[0] = 0 if self.pos[0] < 0: self.pos[0] = self.canvas.winfo_width() if self.pos[1] >= self.canvas.winfo_height(): self.pos[1] = 0 if self.pos[1] < 0: self.pos[1] = self.canvas.winfo_height() else: if self.pos[0] >= self.canvas.winfo_width(): self.vel[0] *= -1 if self.pos[0] < 0: self.vel[0] *= -1 if self.pos[1] >= self.canvas.winfo_height(): self.vel[1] *= -1 if self.pos[1] < 0: self.vel[1] *= -1 Particle(pos=[100,240],canvas=ca,vel=[5,5]) top.bind('<Escape>', lambda e : top.destroy(), '+') top.mainloop()
{ "repo_name": "cptx032/miniprojects", "path": "doppler.py", "copies": "1", "size": "2402", "license": "unlicense", "hash": -1530097305278883300, "line_mean": 25.7, "line_max": 70, "alpha_frac": 0.6236469609, "autogenerated": false, "ratio": 2.4839710444674252, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.8368983458990393, "avg_score": 0.04772690927540618, "num_lines": 90 }
__author__ = 'CQC' # -*- coding:utf-8 -*- import urllib import urllib2 import re import thread import time #糗事百科爬虫类 class QSBK: #初始化方法,定义一些变量 def __init__(self): self.pageIndex = 1 self.user_agent = 'Mozilla/4.0 (compatible; MSIE 5.5; Windows NT)' #初始化headers self.headers = { 'User-Agent' : self.user_agent } #存放段子的变量,每一个元素是每一页的段子们 self.stories = [] #存放程序是否继续运行的变量 self.enable = False #传入某一页的索引获得页面代码 def getPage(self,pageIndex): try: url = 'http://www.qiushibaike.com/hot/page/' + str(pageIndex) #构建请求的request request = urllib2.Request(url,headers = self.headers) #利用urlopen获取页面代码 response = urllib2.urlopen(request) #将页面转化为UTF-8编码 pageCode = response.read().decode('utf-8') return pageCode except urllib2.URLError, e: if hasattr(e,"reason"): print u"连接糗事百科失败,错误原因",e.reason return None #传入某一页代码,返回本页不带图片的段子列表 def getPageItems(self,pageIndex): pageCode = self.getPage(pageIndex) if not pageCode: print "页面加载失败...." return None pattern = re.compile('<div.*?class="author.*?>.*?<a.*?</a>.*?<a.*?>(.*?)</a>.*?<div.*?class'+ '="content".*?title="(.*?)">(.*?)</div>(.*?)<div class="stats.*?class="number">(.*?)</i>',re.S) items = re.findall(pattern,pageCode) #用来存储每页的段子们 pageStories = [] #遍历正则表达式匹配的信息 for item in items: #是否含有图片 haveImg = re.search("img",item[3]) #如果不含有图片,把它加入list中 if not haveImg: #item[0]是一个段子的发布者,item[1]是发布时间,item[2]是内容,item[4]是点赞数 pageStories.append([item[0].strip(),item[1].strip(),item[2].strip(),item[4].strip()]) return pageStories #加载并提取页面的内容,加入到列表中 def loadPage(self): #如果当前未看的页数少于2页,则加载新一页 if self.enable == True: if len(self.stories) < 2: #获取新一页 pageStories = self.getPageItems(self.pageIndex) #将该页的段子存放到全局list中 if pageStories: self.stories.append(pageStories) #获取完之后页码索引加一,表示下次读取下一页 self.pageIndex += 1 #调用该方法,每次敲回车打印输出一个段子 def getOneStory(self,pageStories,page): #遍历一页的段子 for story in pageStories: #等待用户输入 input = raw_input() #每当输入回车一次,判断一下是否要加载新页面 self.loadPage() #如果输入Q则程序结束 if input == "Q": self.enable = False return print u"第%d页\t发布人:%s\t发布时间:%s\n%s\n赞:%s\n" %(page,story[0],story[1],story[2],story[3]) #开始方法 def start(self): print u"正在读取糗事百科,按回车查看新段子,Q退出" #使变量为True,程序可以正常运行 self.enable = True #先加载一页内容 self.loadPage() #局部变量,控制当前读到了第几页 nowPage = 0 while self.enable: if len(self.stories)>0: #从全局list中获取一页的段子 pageStories = self.stories[0] #当前读到的页数加一 nowPage += 1 #将全局list中第一个元素删除,因为已经取出 del self.stories[0] #输出该页的段子 self.getOneStory(pageStories,nowPage) spider = QSBK() spider.start()
{ "repo_name": "zhangmhao/crawl-house", "path": "prototype/spider-qiushibaike.py", "copies": "1", "size": "4279", "license": "mit", "hash": -2761944601421572600, "line_mean": 29.8545454545, "line_max": 120, "alpha_frac": 0.5204833481, "autogenerated": false, "ratio": 2.311307901907357, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.8213464504841448, "avg_score": 0.023665349033181817, "num_lines": 110 }
__author__ = 'CQC' # -*- coding:utf-8 -*- import urllib import urllib2 import re #处理页面标签类 class Tool: #去除img标签,7位长空格 removeImg = re.compile('<img.*?>| {7}|') #删除超链接标签 removeAddr = re.compile('<a.*?>|</a>') #把换行的标签换为\n replaceLine = re.compile('<tr>|<div>|</div>|</p>') #将表格制表<td>替换为\t replaceTD= re.compile('<td>') #把段落开头换为\n加空两格 replacePara = re.compile('<p.*?>') #将换行符或双换行符替换为\n replaceBR = re.compile('<br><br>|<br>') #将其余标签剔除 removeExtraTag = re.compile('<.*?>') def replace(self,x): x = re.sub(self.removeImg,"",x) x = re.sub(self.removeAddr,"",x) x = re.sub(self.replaceLine,"\n",x) x = re.sub(self.replaceTD,"\t",x) x = re.sub(self.replacePara,"\n ",x) x = re.sub(self.replaceBR,"\n",x) x = re.sub(self.removeExtraTag,"",x) #strip()将前后多余内容删除 return x.strip() #百度贴吧爬虫类 class BDTB: #初始化,传入基地址,是否只看楼主的参数 def __init__(self,baseUrl,seeLZ): self.baseURL = baseUrl self.seeLZ = '?see_lz='+str(seeLZ) self.tool = Tool() #传入页码,获取该页帖子的代码 def getPage(self,pageNum): try: url = self.baseURL+ self.seeLZ + '&pn=' + str(pageNum) request = urllib2.Request(url) response = urllib2.urlopen(request) return response.read().decode('utf-8') except urllib2.URLError, e: if hasattr(e,"reason"): print u"连接百度贴吧失败,错误原因",e.reason return None #获取帖子标题 def getTitle(self): page = self.getPage(1) pattern = re.compile('<h1 class="core_title_txt.*?>(.*?)</h1>',re.S) result = re.search(pattern,page) if result: #print result.group(1) #测试输出 return result.group(1).strip() else: return None #获取帖子一共有多少页 def getPageNum(self): page = self.getPage(1) pattern = re.compile('<li class="l_reply_num.*?</span>.*?<span.*?>(.*?)</span>',re.S) result = re.search(pattern,page) if result: #print result.group(1) #测试输出 return result.group(1).strip() else: return None #获取每一层楼的内容,传入页面内容 def getContent(self,page): pattern = re.compile('<div id="post_content_.*?>(.*?)</div>',re.S) items = re.findall(pattern,page) #for item in items: # print item print self.tool.replace(items[1]) baseURL = 'http://tieba.baidu.com/p/3138733512' bdtb = BDTB(baseURL,1) bdtb.getContent(bdtb.getPage(1))
{ "repo_name": "zhangmhao/crawl-house", "path": "prototype/spider-baiduTieba.py", "copies": "1", "size": "2901", "license": "mit", "hash": 1816679850483613200, "line_mean": 28.1590909091, "line_max": 93, "alpha_frac": 0.546588694, "autogenerated": false, "ratio": 2.4639769452449567, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.3510565639244957, "avg_score": null, "num_lines": null }
__author__ = 'CQC' # -*- coding:utf-8 -*- import urllib import urllib2 import cookielib import re import webbrowser import tool #模拟登录淘宝类 class Taobao: #初始化方法 def __init__(self): #登录的URL self.loginURL = "https://login.taobao.com/member/login.jhtml" #代理IP地址,防止自己的IP被封禁 self.proxyURL = 'http://120.193.146.97:843' #登录POST数据时发送的头部信息 self.loginHeaders = { 'Host':'login.taobao.com', 'User-Agent' : 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:35.0) Gecko/20100101 Firefox/35.0', 'Referer' : 'https://login.taobao.com/member/login.jhtml', 'Content-Type': 'application/x-www-form-urlencoded', 'Connection' : 'Keep-Alive' } #用户名 self.username = 'cqcre' #ua字符串,经过淘宝ua算法计算得出,包含了时间戳,浏览器,屏幕分辨率,随机数,鼠标移动,鼠标点击,其实还有键盘输入记录,鼠标移动的记录、点击的记录等等的信息 self.ua = '191UW5TcyMNYQwiAiwTR3tCf0J/QnhEcUpkMmQ=|Um5Ockt0TXdPc011TXVKdyE=|U2xMHDJ+H2QJZwBxX39Rb1d5WXcrSixAJ1kjDVsN|VGhXd1llXGNaYFhkWmJaYl1gV2pIdUtyTXRKfkN4Qn1FeEF6R31TBQ==|VWldfS0TMw8xDjYWKhAwHiUdOA9wCDEVaxgkATdcNU8iDFoM|VmNDbUMV|V2NDbUMV|WGRYeCgGZhtmH2VScVI2UT5fORtmD2gCawwuRSJHZAFsCWMOdVYyVTpbPR99HWAFYVMpUDUFORshHiQdJR0jAT0JPQc/BDoFPgooFDZtVBR5Fn9VOwt2EWhCOVQ4WSJPJFkHXhgoSDVIMRgnHyFqQ3xEezceIRkmahRqFDZLIkUvRiEDaA9qQ3xEezcZORc5bzk=|WWdHFy0TMw8vEy0UIQE0ADgYJBohGjoAOw4uEiwXLAw2DThuOA==|WmBAED5+KnIbdRh1GXgFQSZbGFdrUm1UblZqVGxQa1ZiTGxQcEp1I3U=|W2NDEz19KXENZwJjHkY7Ui9OJQsre09zSWlXY1oMLBExHzERLxsuE0UT|XGZGFjh4LHQdcx5zH34DRyBdHlFtVGtSaFBsUmpWbVBkSmpXd05zTnMlcw==|XWdHFzl5LXUJYwZnGkI/VitKIQ8vEzMKNws3YTc=|XmdaZ0d6WmVFeUB8XGJaYEB4TGxWbk5yTndXa0tyT29Ta0t1QGBeZDI=' #密码,在这里不能输入真实密码,淘宝对此密码进行了加密处理,256位,此处为加密后的密码 self.password2 = '7511aa6854629e45de220d29174f1066537a73420ef6dbb5b46f202396703a2d56b0312df8769d886e6ca63d587fdbb99ee73927e8c07d9c88cd02182e1a21edc13fb8e140a4a2a4b53bf38484bd0e08199e03eb9bf7b365a5c673c03407d812b91394f0d3c7564042e3f2b11d156aeea37ad6460118914125ab8f8ac466f' self.post = post = { 'ua':self.ua, 'TPL_checkcode':'', 'CtrlVersion': '1,0,0,7', 'TPL_password':'', 'TPL_redirect_url':'http://i.taobao.com/my_taobao.htm?nekot=udm8087E1424147022443', 'TPL_username':self.username, 'loginsite':'0', 'newlogin':'0', 'from':'tb', 'fc':'default', 'style':'default', 'css_style':'', 'tid':'XOR_1_000000000000000000000000000000_625C4720470A0A050976770A', 'support':'000001', 'loginType':'4', 'minititle':'', 'minipara':'', 'umto':'NaN', 'pstrong':'3', 'llnick':'', 'sign':'', 'need_sign':'', 'isIgnore':'', 'full_redirect':'', 'popid':'', 'callback':'', 'guf':'', 'not_duplite_str':'', 'need_user_id':'', 'poy':'', 'gvfdcname':'10', 'gvfdcre':'', 'from_encoding ':'', 'sub':'', 'TPL_password_2':self.password2, 'loginASR':'1', 'loginASRSuc':'1', 'allp':'', 'oslanguage':'zh-CN', 'sr':'1366*768', 'osVer':'windows|6.1', 'naviVer':'firefox|35' } #将POST的数据进行编码转换 self.postData = urllib.urlencode(self.post) #设置代理 self.proxy = urllib2.ProxyHandler({'http':self.proxyURL}) #设置cookie self.cookie = cookielib.LWPCookieJar() #设置cookie处理器 self.cookieHandler = urllib2.HTTPCookieProcessor(self.cookie) #设置登录时用到的opener,它的open方法相当于urllib2.urlopen self.opener = urllib2.build_opener(self.cookieHandler,self.proxy,urllib2.HTTPHandler) #赋值J_HToken self.J_HToken = '' #登录成功时,需要的Cookie self.newCookie = cookielib.CookieJar() #登陆成功时,需要的一个新的opener self.newOpener = urllib2.build_opener(urllib2.HTTPCookieProcessor(self.newCookie)) #引入工具类 self.tool = tool.Tool() #得到是否需要输入验证码,这次请求的相应有时会不同,有时需要验证有时不需要 def needCheckCode(self): #第一次登录获取验证码尝试,构建request request = urllib2.Request(self.loginURL,self.postData,self.loginHeaders) #得到第一次登录尝试的相应 response = self.opener.open(request) #获取其中的内容 content = response.read().decode('gbk') #获取状态吗 status = response.getcode() #状态码为200,获取成功 if status == 200: print u"获取请求成功" #\u8bf7\u8f93\u5165\u9a8c\u8bc1\u7801这六个字是请输入验证码的utf-8编码 pattern = re.compile(u'\u8bf7\u8f93\u5165\u9a8c\u8bc1\u7801',re.S) result = re.search(pattern,content) #如果找到该字符,代表需要输入验证码 if result: print u"此次安全验证异常,您需要输入验证码" return content #否则不需要 else: #返回结果直接带有J_HToken字样,表明直接验证通过 tokenPattern = re.compile('id="J_HToken" value="(.*?)"') tokenMatch = re.search(tokenPattern,content) if tokenMatch: self.J_HToken = tokenMatch.group(1) print u"此次安全验证通过,您这次不需要输入验证码" return False else: print u"获取请求失败" return None #得到验证码图片 def getCheckCode(self,page): #得到验证码的图片 pattern = re.compile('<img id="J_StandardCode_m.*?data-src="(.*?)"',re.S) #匹配的结果 matchResult = re.search(pattern,page) #已经匹配得到内容,并且验证码图片链接不为空 if matchResult and matchResult.group(1): return matchResult.group(1) else: print u"没有找到验证码内容" return False #输入验证码,重新请求,如果验证成功,则返回J_HToken def loginWithCheckCode(self): #提示用户输入验证码 checkcode = raw_input('请输入验证码:') #将验证码重新添加到post的数据中 self.post['TPL_checkcode'] = checkcode #对post数据重新进行编码 self.postData = urllib.urlencode(self.post) try: #再次构建请求,加入验证码之后的第二次登录尝试 request = urllib2.Request(self.loginURL,self.postData,self.loginHeaders) #得到第一次登录尝试的相应 response = self.opener.open(request) #获取其中的内容 content = response.read().decode('gbk') #检测验证码错误的正则表达式,\u9a8c\u8bc1\u7801\u9519\u8bef 是验证码错误五个字的编码 pattern = re.compile(u'\u9a8c\u8bc1\u7801\u9519\u8bef',re.S) result = re.search(pattern,content) #如果返回页面包括了,验证码错误五个字 if result: print u"验证码输入错误" return False else: #返回结果直接带有J_HToken字样,说明验证码输入成功,成功跳转到了获取HToken的界面 tokenPattern = re.compile('id="J_HToken" value="(.*?)"') tokenMatch = re.search(tokenPattern,content) #如果匹配成功,找到了J_HToken if tokenMatch: print u"验证码输入正确" self.J_HToken = tokenMatch.group(1) return tokenMatch.group(1) else: #匹配失败,J_Token获取失败 print u"J_Token获取失败" return False except urllib2.HTTPError, e: print u"连接服务器出错,错误原因",e.reason return False #通过token获得st def getSTbyToken(self,token): tokenURL = 'https://passport.alipay.com/mini_apply_st.js?site=0&token=%s&callback=stCallback6' % token request = urllib2.Request(tokenURL) response = urllib2.urlopen(request) #处理st,获得用户淘宝主页的登录地址 pattern = re.compile('{"st":"(.*?)"}',re.S) result = re.search(pattern,response.read()) #如果成功匹配 if result: print u"成功获取st码" #获取st的值 st = result.group(1) return st else: print u"未匹配到st" return False #利用st码进行登录,获取重定向网址 def loginByST(self,st,username): stURL = 'https://login.taobao.com/member/vst.htm?st=%s&TPL_username=%s' % (st,username) headers = { 'User-Agent' : 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:35.0) Gecko/20100101 Firefox/35.0', 'Host':'login.taobao.com', 'Connection' : 'Keep-Alive' } request = urllib2.Request(stURL,headers = headers) response = self.newOpener.open(request) content = response.read().decode('gbk') #检测结果,看是否登录成功 pattern = re.compile('top.location = "(.*?)"',re.S) match = re.search(pattern,content) if match: print u"登录网址成功" location = match.group(1) return True else: print "登录失败" return False #获得已买到的宝贝页面 def getGoodsPage(self,pageIndex): goodsURL = 'http://buyer.trade.taobao.com/trade/itemlist/listBoughtItems.htm?action=itemlist/QueryAction&event_submit_do_query=1' + '&pageNum=' + str(pageIndex) response = self.newOpener.open(goodsURL) page = response.read().decode('gbk') return page #获取所有已买到的宝贝信息 def getAllGoods(self,pageNum): print u"获取到的商品列表如下" for x in range(1,int(pageNum)+1): page = self.getGoodsPage(x) self.tool.getGoodsInfo(page) #程序运行主干 def main(self): #是否需要验证码,是则得到页面内容,不是则返回False needResult = self.needCheckCode() #请求获取失败,得到的结果是None if not needResult ==None: if not needResult == False: print u"您需要手动输入验证码" checkCode = self.getCheckCode(needResult) #得到了验证码的链接 if not checkCode == False: print u"验证码获取成功" print u"请在浏览器中输入您看到的验证码" webbrowser.open_new_tab(checkCode) self.loginWithCheckCode() #验证码链接为空,无效验证码 else: print u"验证码获取失败,请重试" else: print u"不需要输入验证码" else: print u"请求登录页面失败,无法确认是否需要验证码" #判断token是否正常获取到 if not self.J_HToken: print "获取Token失败,请重试" return #获取st码 st = self.getSTbyToken(self.J_HToken) #利用st进行登录 result = self.loginByST(st,self.username) if result: #获得所有宝贝的页面 page = self.getGoodsPage(1) pageNum = self.tool.getPageNum(page) self.getAllGoods(pageNum) else: print u"登录失败" taobao = Taobao() taobao.main()
{ "repo_name": "zhangmhao/crawl-house", "path": "prototype/spider-taoballist.py", "copies": "1", "size": "12386", "license": "mit", "hash": 7391967095079775000, "line_mean": 35.7333333333, "line_max": 772, "alpha_frac": 0.5762323271, "autogenerated": false, "ratio": 2.3108167770419428, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.3387049104141943, "avg_score": null, "num_lines": null }
####AUTHOR/CREDIT: O'Brien, Oliver (o.obrien@ucl.ac.uk) #Paramters: startID (in order of the london.csv file - not necc numerical order!), maxID import csv import os import sys def routing(): locations = [] maxID = int(sys.argv[3]) csvreader = csv.reader(open('london.csv', 'r')) for row in csvreader: locations.append([row[0], row[3], row[4]]) started = False for start in locations: startID = int(start[0]) if int(sys.argv[1]) == startID: started = True if int(sys.argv[2]) == startID: started = False if started == True: print "***************" + str(startID) + "***************" for end in locations: endID = int(end[0]) if startID != endID and startID <= maxID and endID <= maxID: fname = str(startID) + "_" + str(endID) + ".txt" if os.path.exists("../routes/results_londonr6/" + fname) == False: print "Trying " + str(startID) + " to " + str(endID) cmd = "../routes/routino-2.7.2/web/bin/router --transport=bicycle --quiet --lon1=" + start[2] + " --lat1=" + start[1] + " --lon2=" + end[2] + " --lat2=" + end[1] + " --shortest --output-text-all" os.system(cmd) os.rename("shortest-all.txt", "../routes/results_londonr6/" + fname)
{ "repo_name": "oscarechobravo/ftc2050-toolchain", "path": "routino.py", "copies": "1", "size": "1343", "license": "apache-2.0", "hash": -8411698330467659000, "line_mean": 35.2972972973, "line_max": 203, "alpha_frac": 0.5368577811, "autogenerated": false, "ratio": 3.1378504672897196, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9152857373636067, "avg_score": 0.004370174950730325, "num_lines": 37 }
__author__ = 'cristian' import MySQLdb import MySQLdb.cursors class MySql: _host = '' _user = '' _password = '' _database = '' _con = None error = '' last_id = 0 num_rows = 0 def __init__(self, mysql_host, mysql_user, mysql_pass, mysql_db): self._host = mysql_host self._user = mysql_user self._password = mysql_pass self._database = mysql_db # mysql connection function def open(self): self.error = '' try: con = MySQLdb.connect(self._host, self._user, self._password, self._database, cursorclass=MySQLdb.cursors.DictCursor) except MySQLdb.Error, e: try: con = None self.error = "MySQL Error [%d]: %s" % (e.args[0], e.args[1]) except IndexError: con = None self.error = "MySQL Error: %s" % str(e) return con # mysql query function def query(self, query): self.error = '' self.last_id = 0 self.num_rows = 0 if self._con is None: con = self.open() if self.error: return False self._con = con if self._con.open is False: con = self.open() if self.error: return False self._con = con with self._con: try: curs = self._con.cursor() curs.execute(query) rows = curs.fetchall() if curs.lastrowid: self.last_id = curs.lastrowid self.num_rows = curs.rowcount except MySQLdb.Error, e: try: rows = False self.error = "MySQL Error [%d]: %s" % (e.args[0], e.args[1]) except IndexError: rows = False self.error = "MySQL Error: %s" % str(e) return rows
{ "repo_name": "ceakki/ddns-server", "path": "dns/mysql.py", "copies": "1", "size": "2003", "license": "mit", "hash": 3517876505594465300, "line_mean": 23.4390243902, "line_max": 89, "alpha_frac": 0.4583125312, "autogenerated": false, "ratio": 4.207983193277311, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5166295724477311, "avg_score": null, "num_lines": null }
__author__ = 'croxis' from datetime import datetime from flask import flash, redirect, render_template, request, url_for from flask.ext.login import current_user, login_required, login_user, logout_user from werkzeug.contrib.atom import AtomFeed from . import user from ..models import User from .forms import LoginForm, UserEditForm, UserRegisterForm @user.route('/login', methods=['GET', 'POST']) #@oid.loginhandler def login(): """Login by password or preset list of openid providers. New openid will register new user.""" form = LoginForm() user = None if form.validate_on_submit(): try: user = User.objects.get(name=form.username.data) except: flash('Name and password do not match') if user is not None and user.verify_password(form.password.data): login_user(user, form.remember_me.data) return redirect(request.args.get('next') or url_for('main.index')) else: flash('Name and password do not match') return render_template('login.html', title='Login', form=form) @user.route('/logout') @login_required def logout(): logout_user() flash('You have been logged out') return redirect(url_for('main.index')) @user.route('/profile/edit', methods=['GET', 'POST']) @login_required def profile_edit(): form = UserEditForm() if form.validate_on_submit() and current_user.verify_password(form.old_password.data): if form.email.data: current_user.email = form.email.data if form.password.data: current_user.password = form.password.data current_user.save() flash('Your profile has been updated') return redirect(url_for('user.profile')) form.email.data = current_user.email return render_template('user.html', title=current_user.name, form=form) @user.route('/register', methods=['GET', 'POST']) def register(): """Provide openid provider list here as well. First user is admin""" form = UserRegisterForm() if form.validate_on_submit(): user = User(name=form.username.data, email=form.email.data, registered=datetime.utcnow()) user.password = form.password.data # Has to be separate due to setter property if not User.objects: user.level = 'admin' user.save() flash('You can now login. Email authentication will be added later.') return redirect(url_for('user.login')) return render_template('register_user.html', title='New User!', form=form) @user.route('/<name>/feed') def user_feed(name): user_agent = request.headers.get('User-Agent') user = User.objects.get_or_404(name=name) feed = AtomFeed(name + " following these mods.", feed_url=request.url, url=request.host_url) if user.subscribed_mods: #TODO: Try and make the sorting done via mongo instead mods = list(user.subscribed_mods) mods.sort(key=lambda r: r.updated_timestamp, reverse=True) for mod in mods: feed.add(mod.name, mod.downloads[-1].change_log, content_type='html', author=mod.downloads[-1].author.name, url=url_for('mod.mod_page', name=mod.name, _external=True), updated=mod.updated_timestamp) return feed.get_response() @user.route('/<name>') def user_page(name): user_agent = request.headers.get('User-Agent') user = User.objects.get_or_404(name=name) return render_template('user.html', title=user.name, user=user)
{ "repo_name": "croxis/kmr", "path": "app/user/views.py", "copies": "1", "size": "3597", "license": "mit", "hash": 15622878305283932, "line_mean": 37.2765957447, "line_max": 98, "alpha_frac": 0.6427578538, "autogenerated": false, "ratio": 3.8103813559322033, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4953139209732203, "avg_score": null, "num_lines": null }
__author__ = 'croxis' from datetime import datetime import mimetypes import random from flask import abort, flash, make_response, redirect, render_template, request, url_for from flask.ext.login import current_user, login_required from . import mod from .forms import ModDownloadUpdateForm, ModEditForm, ModRegisterForm, ModUploadForm from ..models import Mod, ModDownload, ModDownloadStat, User, UserDownloadStat, mongo @mod.route('/register', methods=['GET', 'POST']) @login_required def register_mod(): # Require a file to prevent the site from being cluttered with "in planning" mods that never have a release # I'm looking at you bukkit! form = ModRegisterForm() request.data if form.validate_on_submit(): mod = Mod( name=form.name.data, install_instructions=form.install_instructions.data, summary=form.summary.data, description=form.description.data, registered_timestamp=datetime.utcnow(), updated_timestamp=datetime.utcnow(), random=random.random() ) #mod.images.append(form.image_path.data) #return str((form.image_path.data, request.files['image_path'])) '''image = request.files['image_path'] print(type(image), dir(image)) f = mongo.GridFSProxy() fi = mongo.ImageGridFsProxy() f.put(image) fi.put(image) print(type(f), dir(f)) print(type(fi), dir(fi)) mod.images.append(f) #mod.images.append(fi)''' f = mongo.GridFSProxy() f.put(request.files['image_path'], content_type=mimetypes.guess_type(form.image_path.data.filename)) mod.images.append(f) mod.save() current_user.owned_mods.append(mod) current_user.save() mod_download = ModDownload( change_log=form.change_log.data, mod_version=form.mod_version.data, ksp_version=form.ksp_version.data, upload_timestamp=datetime.utcnow(), author=User.objects(name=current_user.name).first() ) # Which should take precedence, local or remote hosting? # TODO: Validate zip file as a zip file and ksp mod. Validate in form? mod_download.zip_file.put(request.files['file_path'], content_type='application/zip') mod.downloads.append(mod_download) mod.save() return redirect(url_for('mod.mod_page', name=mod.name)) return render_template('register_mod.html', title='New Mod!', form=form) @mod.route('/list') def list(): query = request.args.get('query') page = request.args.get('page', 1, type=int) pagination = Mod.objects.paginate(page=page, per_page=25) mods = pagination.items return render_template('list.html', title='All Mods.', mods=mods, pagination=pagination) @mod.route('/<name>/upload', methods=['GET', 'POST']) @login_required def mod_upload(name): mod = Mod.objects.get_or_404(name=name) if mod not in current_user.owned_mods or not current_user.is_admin(): flash("You do not own this mod") return redirect(url_for('mod.mod_page', name=name)) form = ModUploadForm() if form.validate_on_submit(): mod_download = ModDownload( change_log=form.change_log.data, mod_version=form.mod_version.data, ksp_version=form.ksp_version.data, upload_timestamp=datetime.utcnow(), author=User.objects(name=current_user.name).first() ) # Which should take precedence, local or remote hosting? # TODO: Validate zip file as a zip file and ksp mod. Validate in form? mod_download.zip_file.put(request.files['file_path'], content_type='application/zip') mod.downloads.append(mod_download) mod.save() return redirect(url_for('mod.mod_page', name=mod.name)) return render_template('upload_mod.html', title='Upload Mod!', name=name, form=form) @mod.route('/<name>/<version>/delete') @login_required def mod_download_delete(name, version): user_agent = request.headers.get('User-Agent') mod = Mod.objects.get_or_404(name=name) if mod not in current_user.owned_mods or not current_user.is_admin(): flash("You do not own this mod") return redirect(url_for('mod.mod_page', name=mod.name)) for download in mod.downloads: if version == download.mod_version or (version == 'None' and download.mod_version is None): mod.downloads.remove(download) mod.save() flash(name + " " + version + " deleted.") return redirect(url_for('mod.mod_edit', name=mod.name)) @mod.route('/<name>/<version>/download') def mod_download_version(name, version): user_agent = request.headers.get('User-Agent') mod = Mod.objects.get_or_404(name=name) for download in mod.downloads: if download.mod_version == version: response = make_response(download.zip_file.read()) response.mimetype = download.zip_file.content_type response.headers["Content-Disposition"] = ("attachment; filename=" + name.replace(" ", "_") + "-" + version + ".zip") mod_stat = ModDownloadStat(timestamp=datetime.utcnow()) if current_user.is_authenticated(): mod_stat.user = User.objects(name=current_user.name).first() user_stat = UserDownloadStat( mod=mod, version=version, timestamp=datetime.utcnow() ) current_user.downloaded.append(user_stat) current_user.save() download.downloaded.append(mod_stat) mod.save() return response abort(404) @mod.route('/<name>/<version>/edit', methods=['GET', 'POST']) @login_required def mod_download_edit(name, version): user_agent = request.headers.get('User-Agent') mod = Mod.objects.get_or_404(name=name) if mod not in current_user.owned_mods or not current_user.is_admin(): flash("You do not own this mod") return redirect(url_for('mod.mod_page', name=mod.name)) for download in mod.downloads: if download.mod_version == version: form = ModDownloadUpdateForm() if form.validate_on_submit(): download.mod_version = form.mod_version.data download.ksp_version = form.ksp_version.data download.change_log = form.change_log.data mod.save() print(url_for('mod.mod_edit', name=name)) return redirect(url_for('mod.mod_edit', name=name)) form.mod_version.data = download.mod_version form.ksp_version.data = download.ksp_version form.change_log.data = download.change_log return render_template('edit_download_mod.html', title='Edit file!', name=name, form=form) abort(404) @mod.route('/<name>/delete') @login_required def mod_delete(name): user_agent = request.headers.get('User-Agent') mod = Mod.objects.get_or_404(name=name) if mod not in current_user.owned_mods: flash("You do not own this mod") return redirect(url_for('mod.mod_page', name=mod.name)) mod.delete() flash(name + " deleted.") return redirect(url_for('main.index')) @mod.route('/<name>/download') def mod_download_latest(name): """As far as I can read it is not possible to query and return embedded documents in mongo. Consider making ModDownload Document""" user_agent = request.headers.get('User-Agent') mod = Mod.objects.get_or_404(name=name) # TODO: This is fugly. Consider making ModDownload a document d = mod.get_latest_download() response = make_response(d.zip_file.read()) response.mimetype = d.zip_file.content_type response.headers["Content-Disposition"] = "attachment; filename=" + mod.name.replace(" ", "_") + "-" + d.mod_version + ".zip" mod_stat = ModDownloadStat(timestamp=datetime.utcnow()) if current_user.is_authenticated(): mod_stat.user = User.objects(name=current_user.name).first() user_stat = UserDownloadStat( mod=mod, version=d.mod_version, timestamp=datetime.utcnow() ) current_user.downloaded.append(user_stat) current_user.save() d.downloaded.append(mod_stat) mod.save() return response @mod.route('/<name>/edit', methods=['GET', 'POST']) @login_required def mod_edit(name): user_agent = request.headers.get('User-Agent') mod = Mod.objects.get_or_404(name=name) if mod not in current_user.owned_mods or not current_user.is_admin(): flash("You do not own this mod") return redirect(url_for('mod.mod_page', name=mod.name)) form = ModEditForm() if form.validate_on_submit(): mod.name = form.name.data mod.summary = form.summary.data mod.description = form.description.data mod.install_instructions = form.install_instructions.data mod.save() return redirect(url_for('mod.mod_page', name=name)) form.name.data = mod.name form.summary.data = mod.summary form.description.data = mod.description form.install_instructions.data = mod.install_instructions return render_template('mod_edit.html', title='Edit: ' + mod.name, mod=mod, form=form) @mod.route('/<name>') def mod_page(name): user_agent = request.headers.get('User-Agent') mod = Mod.objects.get_or_404(name=name) return render_template('mod.html', title=mod.name, mod=mod)
{ "repo_name": "croxis/kmr", "path": "app/mod/views.py", "copies": "1", "size": "9657", "license": "mit", "hash": 3099745709020508000, "line_mean": 40.4506437768, "line_max": 135, "alpha_frac": 0.6301128715, "autogenerated": false, "ratio": 3.7113758647194466, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.48414887362194464, "avg_score": null, "num_lines": null }
__author__ = 'croxis' from flask_wtf import Form from flask_wtf.file import FileField, FileAllowed, FileRequired from wtforms import SubmitField, TextField, TextAreaField, ValidationError from wtforms.validators import Optional, Required from flask.ext.pagedown.fields import PageDownField from ..models import Mod class ModDownloadUpdateForm(Form): mod_version = TextField('Mod Version', validators=[Required()]) ksp_version = TextField('KSP Version', validators=[Required()]) change_log = PageDownField(validators=[Required()]) submit = SubmitField('Update') class ModEditForm(Form): name = TextField('Mod Name', validators=[Required()]) summary = TextAreaField('Summary -- a short description of your mod', validators=[Required()]) description = PageDownField('Extended Description', validators=[Required()]) install_instructions = PageDownField('Install Instructions', validators=[Required()]) submit = SubmitField('Update') class ModUploadForm(Form): mod_version = TextField('Mod Version', validators=[Required()]) ksp_version = TextField('KSP Version', validators=[Required()]) file_path = FileField('Zip File', validators=[Required(), FileRequired(), FileAllowed(['zip'], 'Zip files only!')]) change_log = PageDownField(validators=[Required()]) submit = SubmitField('Upload') class ModRegisterForm(Form): # TODO: Make a form that properly inherits from both. CSRF token missing when I try to do so name = TextField('Mod Name', validators=[Required()]) image_path = FileField('Mod Image File', validators=[Optional(), FileRequired(), FileAllowed(['png', 'jpg', 'gif'], 'Image files only!')]) summary = TextAreaField('Summary -- a short description of your mod', validators=[Required()]) description = PageDownField('Extended Description', validators=[Required()]) install_instructions = PageDownField('Install Instructions', validators=[Required()]) mod_version = TextField('Mod Version', validators=[Required()]) ksp_version = TextField('KSP Version', validators=[Required()]) file_path = FileField('Zip File', validators=[Required(), FileRequired(), FileAllowed(['zip'], 'Zip files only!')]) change_log = PageDownField(validators=[Required()]) submit = SubmitField('Upload') def validate_name(self, field): if Mod.objects(name=field.data).first(): raise ValidationError("Mod name already in use.")
{ "repo_name": "croxis/kmr", "path": "app/mod/forms.py", "copies": "1", "size": "2440", "license": "mit", "hash": 2784830012214971400, "line_mean": 47.82, "line_max": 142, "alpha_frac": 0.7155737705, "autogenerated": false, "ratio": 4.326241134751773, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0023495599016414744, "num_lines": 50 }
__author__ = 'croxis' from io import BytesIO import random import re from PIL import Image, ImageDraw, ImageFont import requests import lib.transforms as transforms import lib.utils as utils from lib.manalib import Manatext from . import magic_image from . import img_manager try: import textwrap import nltk.data sent_tokenizer = nltk.data.load('tokenizers/punkt/english.pickle') # This crazy thing is actually invoked as an unpass, so newlines are still # encoded. def sentencecase(s): s = s.replace(utils.x_marker, utils.reserved_marker) lines = s.split(utils.newline) clines = [] for line in lines: if line: sentences = sent_tokenizer.tokenize(line) clines += [' '.join([sent.capitalize() for sent in sentences])] return utils.newline.join(clines).replace(utils.reserved_marker, utils.x_marker) except ImportError: def sentencecase(s): return s.capitalize() def get_fonts(): return dict( font_title=ImageFont.truetype("fonts/beleren-bold_P1.01.ttf", size=18), font_type=ImageFont.truetype("fonts/beleren-bold_P1.01.ttf", size=16), font=ImageFont.truetype("fonts/mplantin.ttf", size=18)) def draw_costs(image, draw, fonts, card): cost = get_cost(card) w, h = img_manager.get_icon('white').size x_offset = 0 for x in range(0, cost['white']): image.paste(img_manager.get_icon('white'), (321 - x_offset, 42 - h // 2), img_manager.get_icon('white')) x_offset += 23 for x in range(0, cost['blue']): image.paste(img_manager.get_icon('blue'), (321 - x_offset, 42 - h // 2), img_manager.get_icon('blue')) x_offset += 23 for x in range(0, cost['black']): image.paste(img_manager.get_icon('black'), (321 - x_offset, 42 - h // 2), img_manager.get_icon('blue')) x_offset += 23 for x in range(0, cost['green']): image.paste(img_manager.get_icon('green'), (321 - x_offset, 42 - h // 2), img_manager.get_icon('blue')) x_offset += 23 for x in range(0, cost['red']): image.paste(img_manager.get_icon('red'), (321 - x_offset, 42 - h // 2), img_manager.get_icon('blue')) x_offset += 23 if cost['colorless']: colorless_mana = img_manager.get_icon('colorless') draw_colorless = ImageDraw.Draw(colorless_mana) w, h = draw_colorless.textsize(str(cost['colorless'])) W, H = colorless_mana.size draw_colorless.text(((W - w) // 2 - 2, (H - h) // 2 - 5), str(cost['colorless']), fill=(0, 0, 0, 255), font=fonts['font_title']) image.paste(colorless_mana, (321 - x_offset, 36 - h // 2), colorless_mana) colorless_mana.close() def draw_title(image, draw, fonts, card): w, h = draw.textsize(card.name.title()) draw.text((35, 38 - h // 2), # card.name.format(gatherer=True), card.name.title(), fill=(0, 0, 0, 255), font=fonts['font_title']) def draw_types(image, draw, fonts, card): typeline = "" if card.supertypes: typeline += ' '.join(card.supertypes).title() + ' ' typeline += ' '.join(card.types).title() if card.subtypes: typeline += ' - ' + ' '.join(card.subtypes).title() w, h = draw.textsize(typeline) draw.text((35, 304 - h // 2), typeline, fill=(0, 0, 0, 255), font=fonts['font_type']) def get_card_text(card): # Card texts # card_text = card.text.format() mtext = card.text.text mtext = transforms.text_unpass_1_choice(mtext, delimit=True) mtext = transforms.text_unpass_2_counters(mtext) mtext = transforms.text_unpass_3_uncast(mtext) mtext = transforms.text_unpass_4_unary(mtext) mtext = transforms.text_unpass_5_symbols(mtext, for_forum=False) mtext = sentencecase(mtext) # We will do step 5 ourselves to keep capitalization mtext = transforms.text_unpass_6_cardname(mtext, card.name.title()) mtext = transforms.text_unpass_7_newlines(mtext) mtext = transforms.text_unpass_8_unicode(mtext) new_text = Manatext('') new_text.text = mtext new_text.costs = card.text.costs card_text = new_text.format() return card_text def draw_card_text(image, draw, fonts, card): lines = textwrap.wrap(get_card_text(card), 37, replace_whitespace=False) y_offset = 0 for line in lines: for sub_line in line.split('\n'): x_offset = 0 rg = re.compile('(\\{.*?\\})', re.IGNORECASE | re.DOTALL) for subsub_line in rg.split(sub_line): if subsub_line: x = 36 + x_offset y = 335 + y_offset - 3 if rg.match(subsub_line): if '{w}' in subsub_line.lower(): image.paste(img_manager.get_icon_text('white'), (x, y), img_manager.get_icon_text('blue')) x_offset += 21 elif '{b}' in subsub_line.lower(): image.paste(img_manager.get_icon_text('black'), (36 + x_offset, 335 + y_offset - 3), img_manager.get_icon_text('blue')) x_offset += 21 elif '{u}' in subsub_line.lower(): image.paste(img_manager.get_icon_text('blue'), (36 + x_offset, 335 + y_offset - 3), img_manager.get_icon_text('blue')) x_offset += 21 elif '{r}' in subsub_line.lower(): image.paste(img_manager.get_icon_text('red'), (36 + x_offset, 335 + y_offset - 3), img_manager.get_icon_text('blue')) x_offset += 21 elif '{g}' in subsub_line.lower(): image.paste(img_manager.get_icon_text('green'), (36 + x_offset, 335 + y_offset - 3), img_manager.get_icon_text('blue')) x_offset += 21 elif '{t}' in subsub_line.lower(): image.paste(img_manager.get_icon_text('tap'), (36 + x_offset, 335 + y_offset - 3), img_manager.get_icon_text('tap')) x_offset += 21 else: try: int(subsub_line[1]) colorless_mana = img_manager.get_icon_text( 'colorless') draw_colorless = ImageDraw.Draw(colorless_mana) w, h = draw_colorless.textsize( str(subsub_line[1])) draw_colorless.text( ((18 - w) // 2 - 2, (18 - h) // 2 - 4), str(subsub_line[1]), fill=(0, 0, 0, 255), font=fonts['font_title']) image.paste(colorless_mana, (x, y), colorless_mana) colorless_mana.close() x_offset += 21 except: pass else: draw.text((35 + x_offset, 335 + y_offset), subsub_line, fill=(0, 0, 0, 255), font=fonts['font']) x_offset += fonts['font'].getsize(subsub_line)[0] y_offset += 19 def draw_card_copywrite(image, draw, fonts, card): draw.text((60, 484), "Copy, right?", fill=(0, 0, 0, 255), font=fonts['font']) def draw_power_toughness(image, draw, fonts, card): if not card.pt: return power = str(card.pt_p.count('^')) toughness = str(card.pt_t.count('^')) c = card.cost.colors if len(c) == '': c = 'a' if len(c) > 1: c = 'm' c = c.lower() if not c: c = 'a' pt_image = Image.open('app/card_parts/magic-new.mse-style/' + c + 'pt.jpg') image.paste(pt_image, (271, 461)) draw.text((295, 470), power + " / " + toughness, fill=(0, 0, 0, 255), font=fonts['font_title']) def draw_rarity(image, draw, fonts, card): pass def create_card_img(card, google): background_color = get_background_color(card) image = img_manager.get_background(background_color) fonts = get_fonts() draw = ImageDraw.Draw(image) draw_costs(image, draw, fonts, card) draw_title(image, draw, fonts, card) draw_types(image, draw, fonts, card) draw_card_text(image, draw, fonts, card) draw_card_copywrite(image, draw, fonts, card) draw_power_toughness(image, draw, fonts, card) draw_rarity(image, draw, fonts, card) art, w, h = get_card_art(card, google) draw_card_art(image, draw, fonts, card, art, w, h) return image def draw_card_art(image, draw, fonts, card, art, w, h): image.paste(art, ((image.size[0] - w) // 2, 175 - h // 2)) def get_cost(card): cost = {} cost['colorless'] = 0 cost['white'] = card.cost.format().lower().count('w') cost['blue'] = card.cost.format().lower().count('u') cost['black'] = card.cost.format().lower().count('b') cost['red'] = card.cost.format().lower().count('r') cost['green'] = card.cost.format().lower().count('g') rg = re.compile('(\\d+)', re.IGNORECASE | re.DOTALL) m = rg.search(card.cost.format()) if m: cost['colorless'] = int(m.group(1)) return cost def get_background_color(card): colors = card.cost.get_colors() if colors == "": return 'artifact' if len(colors) > 1: return 'multicolor' if colors == "W": return 'white' if colors == "U": return 'blue' if colors == "B": return 'black' if colors == 'R': return 'red' if colors == 'G': return 'green' return None def get_card_art(card, google): if google: google_result = google_card_art(card) if google_result != None: return google_result return get_default_card_art(card) def get_default_card_art(card): art = img_manager.default_portrait art = art.crop((0, 0, 311, 228)) w, h = art.size return (art, w, h) def google_card_art(card): terms = magic_image.find_search_terms(card) random.shuffle(terms) img_url = None for term in terms[:5]: color = term[-1] query = "+".join(term[:-1]) if color == 'u': color = 'blue' img_url = magic_image.fetch(query + '+"fantasy"+paintings+-card', color) if img_url: break if img_url: with BytesIO(requests.get(img_url).content) as reader: reader.seek(0) try: art = Image.open(reader) art.thumbnail((311, 311)) art = art.crop((0, 0, 311, 229)) w, h = art.size return (art, w, h) except OSError: print("Unable to handle this kind of image.") return None
{ "repo_name": "croxis/mtgai", "path": "app/card_visual.py", "copies": "1", "size": "12141", "license": "mit", "hash": 1580799061882881800, "line_mean": 35.1339285714, "line_max": 79, "alpha_frac": 0.4854624825, "autogenerated": false, "ratio": 3.750695088044486, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4736157570544486, "avg_score": null, "num_lines": null }
__author__ = 'croxis' import os.path from PIL import Image COLORS = {'white': 'w', 'black': 'b', 'blue': 'u', 'green': 'g', 'red': 'r', 'colorless': 'c', 'artifact': 'a', 'multicolor': 'm'} C = {v: k for k, v in COLORS.items()} class ImageManager: def __init__(self, root_path='app/card_parts/', card_path='magic-new.mse-style/', icon_path='magic-mana-beveled.mse-symbol-font/', icon_text_path='magic-mana-small.mse-symbol-font/'): self.card_path = os.path.join(root_path, card_path) self.icon_path = os.path.join(root_path, icon_path) self.icon_text_path = os.path.join(root_path, icon_text_path) self.cards = {} self.icons = {} self.icons_text = {} for color in COLORS: self.cards[color] = Image.open(self.card_path + COLORS[color] + 'card.jpg') for name in ['w', 'b', 'u', 'g', 'r']: self.icons[C[name]] = Image.open(self.icon_path + 'mana_' + name + '.png') self.icons['tap'] = Image.open(self.icon_path + 'mana_t.png') self.icons['colorless'] = Image.open(self.icon_path + 'mana_circle.png') for icon in self.icons.values(): icon.thumbnail((22, 22)) for name in ['w', 'b', 'u', 'g', 'r']: self.icons_text[C[name]] = Image.open(self.icon_text_path + 'mana_' + name + '.png') self.icons_text['tap'] = Image.open(self.icon_text_path + 'mana_t.png') self.icons_text['colorless'] = Image.open(self.icon_text_path + 'mana_circle.png') for icon in self.icons_text.values(): icon.thumbnail((18, 18)) self.default_portrait = Image.open(os.path.join(root_path, 'Magic_the_gathering_pentagon.png')) self.default_portrait.thumbnail((311, 311)) def get_background(self, color): return self.cards[color].copy() def get_icon(self, name): if name == 'colorless': return self.icons[name].copy() return self.icons[name] def get_icon_text(self, name): if name == 'colorless': return self.icons_text[name].copy() return self.icons_text[name]
{ "repo_name": "croxis/mtgai", "path": "app/image_manager.py", "copies": "1", "size": "2261", "license": "mit", "hash": 6680155407458835000, "line_mean": 31.3, "line_max": 103, "alpha_frac": 0.5444493587, "autogenerated": false, "ratio": 3.2025495750708215, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.42469989337708214, "avg_score": null, "num_lines": null }
__author__ = 'croxis' from datetime import datetime import mimetypes import random from flask import abort, current_app, jsonify, make_response, render_template, request from flask.ext.login import current_user from . import main from ..models import Mod, mongo @main.route('/') def index(): user_agent = request.headers.get('User-Agent') # Idea: If tags are used maybe the random sample can be done via one or two tags # Also return newest version of ksp try: rand_mods = random.sample(list(Mod.objects), 4) except ValueError: rand_mods = list(Mod.objects) updated_mods = Mod.objects.order_by('-updated_timestamp')[:5] subscribed_updated = [] if current_user.is_authenticated(): for mod in current_user.subscribed_mods: for download in mod.downloads: u = False for user_download in current_user.downloaded: if user_download.timestamp < download.upload_timestamp: u = True break if u: subscribed_updated.append((mod, download)) return render_template( 'index.html', current_time=datetime.utcnow(), current_user=current_user, rand_mods=rand_mods, subscribed_updates=subscribed_updated, updated_mods=updated_mods ) @main.route('/images/<mod>/<index>') def images(mod, index): img = Mod.objects.get_or_404(name=mod).images[int(index)] response = make_response(img.read()) response.mimetype = img.content_type[0] return response @main.route('/search_results') def search_results(): #TODO: Unhardcode pymongo db query_string = request.args.get('search_string') if query_string: results = mongo.connection[current_app.config['MONGODB_SETTINGS']['DB']].command( 'text', 'mod', search=query_string, limit=10 ) sending_results = [] for result in results['results']: sending_results.append( { 'image_url': '/images/'+result['obj']['name']+'/0', 'name': result['obj']['name'], 'summary': result['obj']['summary'], 'updated_timestamp': result['obj']['updated_timestamp'] } ) return jsonify(results=sending_results) return jsonify(results=[]) @main.route('/search') def search(): query = request.args.get('query') return render_template('search.html', title='Search!') @main.route('/toggle_following') def toggle_following(): if current_user.is_authenticated(): mod_name = request.args.get('mod_name') mod = Mod.objects.get_or_404(name=mod_name) following = True if mod in current_user.subscribed_mods: current_user.subscribed_mods.remove(mod) following = False else: current_user.subscribed_mods.append(mod) current_user.save() return jsonify(following=following) abort(404)
{ "repo_name": "croxis/kmr", "path": "app/main/views.py", "copies": "1", "size": "3100", "license": "mit", "hash": -7740487037489334000, "line_mean": 30.6428571429, "line_max": 89, "alpha_frac": 0.5938709677, "autogenerated": false, "ratio": 4.133333333333334, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0017179564909669185, "num_lines": 98 }
__author__ = 'croxis' from flask.ext.login import AnonymousUserMixin, LoginManager, UserMixin from flask.ext.mongoengine import MongoEngine from werkzeug.security import generate_password_hash, check_password_hash import bcrypt # Break the cycle of circular dependency mongo = MongoEngine() login = LoginManager() @login.user_loader def load_user(user_id): """Returns a user or None based on hashed session id. Do not use user name""" return User.objects(id=user_id).first() class Mod(mongo.Document): name = mongo.StringField(required=True, unique=True, default='') dependencies = mongo.ListField(mongo.ReferenceField('self')) description = mongo.StringField(required=True) downloads = mongo.ListField(mongo.EmbeddedDocumentField('ModDownload')) #images = mongo.ListField(mongo.ImageField(thumbnail_size=(100, 100, True))) images = mongo.ListField(mongo.FileField()) install_instructions = mongo.StringField(required=True, default='') random = mongo.FloatField(required=True, default=0) registered_timestamp = mongo.DateTimeField(required=True) summary = mongo.StringField(required=True) tags = mongo.ListField(mongo.ReferenceField('Tag')) updated_timestamp = mongo.DateTimeField(required=True) def get_latest_download(self): """Returns the latest download of the mod.""" if not self.downloads: return d = self.downloads[0] for download in self.downloads: if download.upload_timestamp > d.upload_timestamp: d = download return d def get_total_downloads(self): i = 0 for download in self.downloads: i += len(download.downloaded) return i class ModDownload(mongo.EmbeddedDocument): """Specific mod download. It looks like it is not possible to query mongo and return just an Embedded Document. Consider making ModDownload a Document with a reference to the Mod? Author's note: I chose to store the file in the database using gridfs as convenience for the prototype. I have not researched into seeing if this is A Good Idea™. My guess is that uploading the the file system and serving directly via nginx will probably scale better. It might also be worth refactoring this to a generic file class with moddownload, image files, and craft files inheriting. Apparently nginx can be compiled to serve gridfs as well. """ mod_version = mongo.StringField(required=True) ksp_version = mongo.StringField(required=True) url = mongo.StringField() zip_file = mongo.FileField() upload_timestamp = mongo.DateTimeField(required=True) author = mongo.ReferenceField('User', required=True) change_log = mongo.StringField(default='') downloaded = mongo.ListField(mongo.EmbeddedDocumentField('ModDownloadStat')) class ModDownloadStat(mongo.EmbeddedDocument): user = mongo.ReferenceField('User') timestamp = mongo.DateTimeField() class User(mongo.Document, UserMixin): """Flask-Login requires an user class. We indulge.""" name = mongo.StringField(required=True, unique=True, default='') email = mongo.StringField(required=True) # Not email field due to new tld password_hash = mongo.StringField() level = mongo.StringField(default='standard') subscribed_mods = mongo.ListField(mongo.ReferenceField(Mod)) owned_mods = mongo.ListField(mongo.ReferenceField(Mod)) downloaded = mongo.ListField(mongo.EmbeddedDocumentField('UserDownloadStat')) registered = mongo.DateTimeField() @property def password(self): raise AttributeError('password is not a readable attribute') @password.setter def password(self, password): self.password_hash = bcrypt.hashpw(password, bcrypt.gensalt(log_rounds=12)) def verify_password(self, password): return bcrypt.checkpw(password, self.password_hash) def is_admin(self): if self.level.lower() == 'admin': return True return False class AnonymousUser(AnonymousUserMixin): def is_admin(self): return False class UserDownloadStat(mongo.EmbeddedDocument): """Entry in a list of mods, versions, and when a user downloaded a file.""" mod = mongo.ReferenceField(Mod) version = mongo.StringField() timestamp = mongo.DateTimeField() class Tag(mongo.Document): """User defined tags""" name = mongo.StringField(required=True) login.anonymous_user = AnonymousUser
{ "repo_name": "croxis/kmr", "path": "app/models.py", "copies": "1", "size": "4495", "license": "mit", "hash": -2229914786500480800, "line_mean": 34.952, "line_max": 118, "alpha_frac": 0.7122190073, "autogenerated": false, "ratio": 4.191231343283582, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5403450350583582, "avg_score": null, "num_lines": null }
__author__ = 'croxis' from flask import abort, flash, jsonify, make_response, redirect, render_template, request, url_for from flask.ext.openid import OpenID # OpenID support broken until fix: https://github.com/mitsuhiko/flask-openid/pull/34 from flask.ext.pagedown.fields import PageDownField from flask_wtf import Form from flask_wtf.file import FileField, FileAllowed, FileRequired from werkzeug.security import generate_password_hash, check_password_hash from wtforms import BooleanField, PasswordField, SubmitField, TextField, TextAreaField, ValidationError from wtforms.validators import Email, EqualTo, Optional, Required, URL import mimetypes import random from datetime import datetime '''@oid.after_login def after_login(resp): if not resp.email: flash('Invalid login. Please try again.') return redirect(url_for('login')) print("test")''' ######## # Routes ######## if __name__=='__main__': #Mod.drop_collection() #User.drop_collection() #app.run(debug=True, host='0.0.0.0') pass
{ "repo_name": "croxis/kmr", "path": "app/main/kmr.py", "copies": "1", "size": "1047", "license": "mit", "hash": 3628862929456666600, "line_mean": 24.5609756098, "line_max": 121, "alpha_frac": 0.7258834766, "autogenerated": false, "ratio": 3.6228373702422147, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.980238503311287, "avg_score": 0.009267162745868824, "num_lines": 41 }
__author__ = 'croxis' from flask_wtf import Form from flask_wtf.file import FileField from wtforms import TextField, TextAreaField, ValidationError from wtforms.validators import Required from flask.ext.pagedown.fields import PageDownField from ..models import Mod, User '''class OIDLoginForm(Form): """OID user is currently broken at the moment""" openid = TextField('openid', validators = [Required()]) remember_me = BooleanField('remember_me', default = False)''' class ModDownloadForm(Form): mod_version = TextField('Mod Version', validators=[Required()]) ksp_version = TextField('KSP Version', validators=[Required()]) file_path = FileField('Zip File', validators=[Required()]) #submit = SubmitField('Upload') class ModForm(Form): name = TextField('Mod Name', validators=[Required()]) summery = TextAreaField('Summery -- a short description of your mod', validators=[Required()]) description = TextAreaField('Extended Description', validators=[Required()]) def validate_name(self, field): if Mod.objects(name=field.data).first(): raise ValidationError("Mod name already in use.")
{ "repo_name": "croxis/kmr", "path": "app/main/forms.py", "copies": "1", "size": "1159", "license": "mit", "hash": 4696400715043045000, "line_mean": 34.1212121212, "line_max": 98, "alpha_frac": 0.7100949094, "autogenerated": false, "ratio": 4.066666666666666, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.03182498384518587, "num_lines": 33 }
__author__ = 'croxis' from flask_wtf import Form from wtforms import BooleanField, PasswordField, SubmitField, TextField, ValidationError from wtforms.validators import Email, EqualTo, Required from ..models import User class LoginForm(Form): username = TextField('User Name', validators=[Required()]) password = PasswordField('Password', validators=[Required()]) remember_me = BooleanField('Stay logged in') submit = SubmitField('Log In') class UserEditForm(Form): email = TextField('Email', validators=[Email()]) old_password = PasswordField('Current Password', validators=[Required()]) password = PasswordField('New Password', validators=[EqualTo('password2', message='Passwords must match')]) password2 = PasswordField('Confirm new password') submit = SubmitField('Update') class UserRegisterForm(Form): username = TextField('User Name', validators=[Required()]) email = TextField('Email', validators=[Required(), Email()]) password = PasswordField('Password', validators=[Required(), EqualTo('password2', message='Passwords must match')]) password2 = PasswordField('Confirm password', validators=[Required()]) submit = SubmitField('Register') def validate_username(self, field): if User.objects(name=field.data).first(): raise ValidationError("User name already in use.")
{ "repo_name": "croxis/kmr", "path": "app/user/forms.py", "copies": "1", "size": "1365", "license": "mit", "hash": -3902717611878918700, "line_mean": 39.1764705882, "line_max": 119, "alpha_frac": 0.7157509158, "autogenerated": false, "ratio": 4.431818181818182, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5647569097618181, "avg_score": null, "num_lines": null }
__author__ = 'croxis' """ Python source: https://gist.github.com/petrklus/b1f427accdf7438606a6 Original pseudo code: Set Temperature = Temperature \ 100 Calculate Red: If Temperature <= 66 Then Red = 255 Else Red = Temperature - 60 Red = 329.698727446 * (Red ^ -0.1332047592) If Red < 0 Then Red = 0 If Red > 255 Then Red = 255 End If Calculate Green: If Temperature <= 66 Then Green = Temperature Green = 99.4708025861 * Ln(Green) - 161.1195681661 If Green < 0 Then Green = 0 If Green > 255 Then Green = 255 Else Green = Temperature - 60 Green = 288.1221695283 * (Green ^ -0.0755148492) If Green < 0 Then Green = 0 If Green > 255 Then Green = 255 End If Calculate Blue: If Temperature >= 66 Then Blue = 255 Else If Temperature <= 19 Then Blue = 0 Else Blue = Temperature - 10 Blue = 138.5177312231 * Ln(Blue) - 305.0447927307 If Blue < 0 Then Blue = 0 If Blue > 255 Then Blue = 255 End If End If """ import math from panda3d.core import Vec3 def convert_K_to_RGB_float(temperature): red, green, blue = convert_K_to_RGB(temperature) return red/255.0, green/255.0, blue/255.0 def convert_K_to_RGB(colour_temperature): """ Converts from K to RGB, algorithm courtesy of http://www.tannerhelland.com/4435/convert-temperature-rgb-algorithm-code/ """ # range check if colour_temperature < 1000: colour_temperature = 1000 elif colour_temperature > 40000: colour_temperature = 40000 tmp_internal = colour_temperature / 100.0 # red if tmp_internal <= 66: red = 255 else: tmp_red = 329.698727446 * math.pow(tmp_internal - 60, -0.1332047592) if tmp_red < 0: red = 0 elif tmp_red > 255: red = 255 else: red = tmp_red # green if tmp_internal <= 66: tmp_green = 99.4708025861 * math.log(tmp_internal) - 161.1195681661 if tmp_green < 0: green = 0 elif tmp_green > 255: green = 255 else: green = tmp_green else: tmp_green = 288.1221695283 * math.pow(tmp_internal - 60, -0.0755148492) if tmp_green < 0: green = 0 elif tmp_green > 255: green = 255 else: green = tmp_green # blue if tmp_internal >= 66: blue = 255 elif tmp_internal <= 19: blue = 0 else: tmp_blue = 138.5177312231 * math.log( tmp_internal - 10) - 305.0447927307 if tmp_blue < 0: blue = 0 elif tmp_blue > 255: blue = 255 else: blue = tmp_blue return Vec3(red, green, blue)
{ "repo_name": "croxis/Panda-Core-Technology", "path": "Planets/utils/blackbody.py", "copies": "1", "size": "2901", "license": "mit", "hash": 7705592892069374000, "line_mean": 23.593220339, "line_max": 79, "alpha_frac": 0.5429162358, "autogenerated": false, "ratio": 3.5814814814814815, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.46243977172814815, "avg_score": null, "num_lines": null }
__author__ = 'croxis' import sandbox from direct.directnotify.DirectNotify import DirectNotify log = DirectNotify().newCategory("SpaceDrive-ClientNet") #PROPOSAL! {server entity id: client entity id} and reverse lookup dict too class ClientNetworkSystem(sandbox.UDPNetworkSystem): def init2(self): self.packetCount = 0 self.accept('login', self.sendLogin) self.accept('requestStations', self.requestStations) self.accept('requestThrottle', self.requestThrottle) self.accept('requestCreateShip', self.requestCreateShip) self.accept('requestTarget', self.requestTarget) def process_packet(self, msgID, remotePacketCount, ack, acks, hashID, serialized, address): #If not in our protocol range then we just reject if msgID < 0 or msgID > 200: return data = protocol.readProto(msgID, serialized) if msgID == protocol.CONFIRM_STATIONS: sandbox.send('shipUpdate', [data, True]) sandbox.send('setShipID', [data]) sandbox.send('makeStationUI', [data]) elif msgID == protocol.PLAYER_SHIPS: sandbox.send('shipUpdates', [data]) sandbox.send('shipSelectScreen', [data]) elif msgID == protocol.POS_PHYS_UPDATE: sandbox.send('shipUpdates', [data]) elif msgID == protocol.SHIP_CLASSES: sandbox.send('shipClassList', [data]) def sendLogin(self, serverAddress): self.serverAddress = serverAddress datagram = self.generateGenericPacket(protocol.LOGIN) universals.log.debug("sending login") self.send(datagram) def requestCreateShip(self, shipName, className): datagram = protocol.requestCreateShip(shipName, className) self.send(datagram) def requestStations(self, shipid, stations): datagram = protocol.requestStations(shipid, stations) self.send(datagram) def requestThrottle(self, throttle, heading): datagram = protocol.requestThrottle(throttle, heading) self.send(datagram) def requestTarget(self, targetID): datagram = protocol.requestTurretTarget(targetID) self.send(datagram) def send(self, datagram): self.send_data(datagram, self.serverAddress) class ServerComponent: """Theoretical component for server generated and sent entities""" serverEntityID = 0 lastServerUpdate = 0
{ "repo_name": "croxis/SpaceDrive", "path": "spacedrive/networking/client_networking.py", "copies": "1", "size": "2496", "license": "mit", "hash": 3542728348685860400, "line_mean": 35.2537313433, "line_max": 95, "alpha_frac": 0.6574519231, "autogenerated": false, "ratio": 4.146179401993355, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5303631325093355, "avg_score": null, "num_lines": null }
__author__ = 'croxis' """ Python source: https://gist.github.com/petrklus/b1f427accdf7438606a6 Original pseudo code: Set Temperature = Temperature \ 100 Calculate Red: If Temperature <= 66 Then Red = 255 Else Red = Temperature - 60 Red = 329.698727446 * (Red ^ -0.1332047592) If Red < 0 Then Red = 0 If Red > 255 Then Red = 255 End If Calculate Green: If Temperature <= 66 Then Green = Temperature Green = 99.4708025861 * Ln(Green) - 161.1195681661 If Green < 0 Then Green = 0 If Green > 255 Then Green = 255 Else Green = Temperature - 60 Green = 288.1221695283 * (Green ^ -0.0755148492) If Green < 0 Then Green = 0 If Green > 255 Then Green = 255 End If Calculate Blue: If Temperature >= 66 Then Blue = 255 Else If Temperature <= 19 Then Blue = 0 Else Blue = Temperature - 10 Blue = 138.5177312231 * Ln(Blue) - 305.0447927307 If Blue < 0 Then Blue = 0 If Blue > 255 Then Blue = 255 End If End If """ import math from panda3d.core import Vec3 def convert_K_to_RGB_float(temperature): red, green, blue = convert_K_to_RGB(temperature) return red/255.0, green/255.0, blue/255.0 def convert_K_to_RGB(colour_temperature): """ Converts from K to RGB, algorithm courtesy of http://www.tannerhelland.com/4435/convert-temperature-rgb-algorithm-code/ """ # range check if colour_temperature < 1000: colour_temperature = 1000 elif colour_temperature > 40000: colour_temperature = 40000 tmp_internal = colour_temperature / 100.0 # red if tmp_internal <= 66: red = 255 else: tmp_red = 329.698727446 * math.pow(tmp_internal - 60, -0.1332047592) if tmp_red < 0: red = 0 elif tmp_red > 255: red = 255 else: red = tmp_red # green if tmp_internal <= 66: tmp_green = 99.4708025861 * math.log(tmp_internal) - 161.1195681661 if tmp_green < 0: green = 0 elif tmp_green > 255: green = 255 else: green = tmp_green else: tmp_green = 288.1221695283 * math.pow(tmp_internal - 60, -0.0755148492) if tmp_green < 0: green = 0 elif tmp_green > 255: green = 255 else: green = tmp_green # blue if tmp_internal >= 66: blue = 255 elif tmp_internal <= 19: blue = 0 else: tmp_blue = 138.5177312231 * math.log( tmp_internal - 10) - 305.0447927307 if tmp_blue < 0: blue = 0 elif tmp_blue > 255: blue = 255 else: blue = tmp_blue return Vec3(red, green, blue)
{ "repo_name": "croxis/SpaceDrive", "path": "spacedrive/utils/blackbody.py", "copies": "1", "size": "3018", "license": "mit", "hash": -7012207628510967000, "line_mean": 23.593220339, "line_max": 79, "alpha_frac": 0.5218687873, "autogenerated": false, "ratio": 3.7121771217712176, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4734045909071217, "avg_score": null, "num_lines": null }
import requests import optparse import sys import threading import time import Queue import re from progressbar import ProgressBar header = { "Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8", 'Accept-Encoding': 'gzip, deflate, compress', 'Cache-Control': 'max-age=0', 'Connection': 'keep-alive', "Accept-Language": "zh-cn,zh;q=0.8,en-us;q=0.5,en;q=0.3", "User-Agent": "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:35.0) Gecko/20100101 Firefox/35.0" } class DirScan: def __init__(self, target, threads_num, ext): self.target = target.strip() self.threads_num = threads_num self.ext = ext self.lock = threading.Lock() #outfile self.__load_dir_dict() self.errorpage = r'无法加载模块|[nN]ot [fF]ound|不存在|未找到|Error|Welcome to nginx!|404|美容|百姓大事件|功能入口' self.regex = re.compile(self.errorpage) def __load_dir_dict(self): self.queue = Queue.Queue() ext = self.ext target = self.target hostuser = target.split('.') hostuser = hostuser[len(hostuser)-2] #print hostuser bak = ['/'+hostuser+'.rar','/'+hostuser+'.zip','/'+hostuser+hostuser+'.rar','/'+hostuser+'.rar','/'+hostuser+'.tar.gz','/'+hostuser+'.tar','/'+hostuser+'123.zip','/'+hostuser+'123.tar.gz','/'+hostuser+hostuser+'.zip','/'+hostuser+hostuser+'.tar.gz','/'+hostuser+hostuser+'.tar','/'+hostuser+'.bak'] for j in range(len(bak)): BAK = bak[j] self.queue.put(BAK) with open('mulu.txt') as f: for line in f: mulu = line.replace('$ext$',ext).strip() if mulu: #print mulu self.queue.put(mulu) def _scan(self): #print "[*]%s Scaning...." % self.target try: while self.queue.qsize() > 0: sub = self.queue.get(timeout=1.0) #try: #print sub domain = self.target + sub #print domain r = requests.get(domain, headers = header, allow_redirects=False, timeout=5) code = r.status_code text = r.content lens = len(text) if code == 200 and not self.regex.findall(text) and lens != 0 : try: title = re.findall(r"<title>(.+?)</title>",text) print "[*] %s =======> 200 (Title:%s)\n" %(domain, title[0]), except Exception,e: print "[*] %s =======> 200\n" %domain, except Exception,e: pass def run(self): self.start_time = time.time() t_sequ = [] t_pro = threading.Thread(target=self.progress, name="progress") t_pro.setDaemon(True) t_pro.start() for i in range(self.threads_num): t = threading.Thread(target=self._scan, name=str(i)) t.setDaemon(True) t.start() while self.thread_count > 0: time.sleep(0.01) if __name__ == '__main__': #parser = optparse.OptionParser('usage: %prog [options] http://www.c-chicken.cc') #parser.add_option('-t', '--threads', dest='threads_num', # default=10, type='int', # help='Number of threads. default = 10') # parser.add_option('-e', '--ext', dest='ext', default='php', # type='string', help='You want to Scan WebScript. default is php') # parser.add_option('-o', '--output', dest='output', default=None, # type='string', help='Output file name. default is {target}.txt') # (options, args) = parser.parse_args() #if len(args) < 1: # parser.print_help() # sys.exit(0) f = open('target.txt','r') hosts = f.read().split() for i in range(len(hosts)): Dict = 'http://'+hosts[i] d = DirScan(target=Dict,threads_num=10,ext='php') d.run()
{ "repo_name": "RicterZ/moescan", "path": "tools/DirScan/pldirscan.py", "copies": "1", "size": "4027", "license": "mit", "hash": -2029998308000329200, "line_mean": 34.8648648649, "line_max": 307, "alpha_frac": 0.5342878674, "autogenerated": false, "ratio": 3.3037344398340247, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.43380223072340246, "avg_score": null, "num_lines": null }
__author__ = 'CSPF' from ..config import db class User(db.Model): __tablename__ = "users" id = db.Column("id",db.Integer,primary_key=True,autoincrement=True) username = db.Column('username',db.String(20),unique=True,index=True) password = db.Column('password',db.String(50)) def __init__(self,username,password): self.username = username self.password = password def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return unicode(self.id) def create(self): row = self.query.filter_by(username=self.username).first() if row: return False else: db.session.add(self) db.session.commit() return True def change_password(self,password): if self.query.filter_by(username=self.username).update(dict(password=self.password)): db.session.commit() return True else: return False def __repr__(self): return '<User %r>' % self.username
{ "repo_name": "breakthesec/DodoVulnerableBank", "path": "DodoRESTAPI/app/model/Users.py", "copies": "1", "size": "1139", "license": "mit", "hash": 6843217516838618000, "line_mean": 24.3111111111, "line_max": 93, "alpha_frac": 0.5899912204, "autogenerated": false, "ratio": 3.9411764705882355, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5031167690988235, "avg_score": null, "num_lines": null }
__author__ = 'CSPF' from ..config import app from .. model.Transaction import Transaction from .. model.Account import Account from flask import request,jsonify @app.route("/transaction/statement/<int:customer_id>",methods=['POST',"GET"]) def view_statement(customer_id): transactions = Transaction.query.filter_by(customer_id=customer_id).all() if not transactions: return "", 204 else: statement = {} for transaction in transactions: statement.update({transaction.id: {"amount":transaction.amount,"remarks":transaction.remarks, "details":transaction.transaction_details,"sender_account_number":transaction.sender_account_number, "receiver_account_number":transaction.receiver_account_number } }) return jsonify(statement) @app.route("/transaction/transfer",methods=['POST',"GET"]) def transfer_amount(): customer_id = request.form['customer_id'] to_account = request.form['to_account'] amount = request.form['amount'] account = Account.query.filter_by(id=customer_id).first() if account is None: return jsonify({"success":False}) else: account.debit(amount) receiver_account= Account.query.filter_by(account_number=to_account).first() if receiver_account: receiver_account.credit(amount) transaction = Transaction(account.id,amount,account.account_number,to_account, remarks="Transferred", transaction_details="Transferred from "+account.account_number+" to "+to_account) transaction.insert() return jsonify({"success":True})
{ "repo_name": "breakthesec/DodoVulnerableBank", "path": "DodoRESTAPI/app/controller/transaction.py", "copies": "1", "size": "1850", "license": "mit", "hash": -2944468790624310000, "line_mean": 44.3, "line_max": 134, "alpha_frac": 0.5994594595, "autogenerated": false, "ratio": 4.579207920792079, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.567866738029208, "avg_score": null, "num_lines": null }
__author__ = 'CSPF' from ..config import db class Account(db.Model): __tablename__ = "accounts" id = db.Column("id",db.Integer,primary_key=True) branch = db.Column('branch',db.String(50)) account_number = db.Column('account_number',db.String(50)) balance = db.Column('balance',db.Integer) def __init__(self,id,account_number,balance=0,branch="Arkham"): self.id = id self.account_number = account_number self.balance = balance self.branch = branch def get_id(self): return unicode(self.id) def create(self): row = self.query.filter_by(id=self.id).first() if row: return False else: db.session.add(self) db.session.commit() return True def debit(self,debit_amount): new_amount = self.balance - int(debit_amount) if self.query.filter_by(id=self.id).update(dict(balance=new_amount)): db.session.commit() print "New Balance : "+str(self.balance)+"\n" def credit(self,debit_amount): new_amount = self.balance + int(debit_amount) if self.query.filter_by(id=self.id).update(dict(balance=new_amount)): db.session.commit() print "New Balance : "+str(self.balance)+"\n" def __repr__(self): return '<Account %r>' % self.account_number
{ "repo_name": "breakthesec/DodoVulnerableBank", "path": "DodoRESTAPI/app/model/Account.py", "copies": "1", "size": "1420", "license": "mit", "hash": -7613404034541715000, "line_mean": 30.2727272727, "line_max": 77, "alpha_frac": 0.5725352113, "autogenerated": false, "ratio": 3.641025641025641, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.47135608523256406, "avg_score": null, "num_lines": null }
__author__ = 'CSPF' from ..config import db class Transaction(db.Model): __tablename__ = "transaction" id = db.Column("id", db.Integer, primary_key=True, autoincrement=True) customer_id = db.Column("customer_id", db.Integer) amount = db.Column('amount', db.Integer) remarks = db.Column('remarks', db.String(100)) transaction_details = db.Column('transaction_details', db.String(200)) sender_account_number = db.Column('sender_accoutn_number', db.String(50)) receiver_account_number = db.Column('receiver_account_number', db.String(50)) def __init__(self, customer_id, amount, sender_account_number, receiver_account_number, remarks="", transaction_details=""): self.customer_id = customer_id self.amount = amount self.sender_account_number = sender_account_number self.receiver_account_number = receiver_account_number self.remarks = remarks self.transaction_details = transaction_details def get_id(self): return unicode(self.id) def insert(self): db.session.add(self) db.session.commit() return True def __repr__(self): return '<Transaction %r>' % self.id
{ "repo_name": "breakthesec/DodoVulnerableBank", "path": "DodoRESTAPI/app/model/Transaction.py", "copies": "1", "size": "1250", "license": "mit", "hash": 481864319006227140, "line_mean": 35.8787878788, "line_max": 103, "alpha_frac": 0.6328, "autogenerated": false, "ratio": 3.8580246913580245, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.49908246913580245, "avg_score": null, "num_lines": null }
__author__ = 'CSPF' import os from OpenSSL import crypto import socket, ssl from utils import get_home_dir class SSLCertifiacate: def __init__(self,host=None): self.host = host ssl_dir = os.path.join(get_home_dir(), "ssl") self.key_path = os.path.join(ssl_dir, "dodo.key") self.cert_path = os.path.join(ssl_dir, "dodo.crt") if not os.path.exists(ssl_dir): os.makedirs(ssl_dir) def generate(self): if not self.host: self.host = socket.gethostname() print "SSL Host used for Certificate Generation: "+self.host key = crypto.PKey() key.generate_key(crypto.TYPE_RSA, 2048) cert = crypto.X509() cert.get_subject().C = "IN" cert.get_subject().ST = "TN" cert.get_subject().L = "dodo" cert.get_subject().O = "dodo" cert.get_subject().OU = "dodo" cert.get_subject().CN = self.host cert.set_serial_number(1111) cert.gmtime_adj_notBefore(0) cert.gmtime_adj_notAfter(10 * 365 * 24 * 60 * 60) cert.set_issuer(cert.get_subject()) cert.set_pubkey(key) cert.sign(key, "sha1") with open(self.cert_path, "w") as f: f.write(crypto.dump_certificate(crypto.FILETYPE_PEM, cert)) with open(self.key_path, "w") as f: f.write(crypto.dump_privatekey(crypto.FILETYPE_PEM, key)) def get(self): if not os.path.exists(self.cert_path) or not os.path.exists(self.cert_path): self.generate() # context = ssl.SSLContext(ssl.PROTOCOL_TLSv1_2) #For some reason android not able to communicate with stupid python on TLSv1_2 context = ssl.SSLContext(ssl.PROTOCOL_SSLv23) context.load_cert_chain(self.cert_path, self.key_path) return context
{ "repo_name": "breakthesec/DodoVulnerableBank", "path": "DodoRESTAPI/app/digital_certificate.py", "copies": "1", "size": "1868", "license": "mit", "hash": -2807401258273301000, "line_mean": 33.9230769231, "line_max": 135, "alpha_frac": 0.5867237687, "autogenerated": false, "ratio": 3.341681574239714, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.44284053429397136, "avg_score": null, "num_lines": null }
_author_ = 'ctippur' import boto3 import os from aws_requests_auth.aws_auth import AWSRequestsAuth class credsaws(): ## init session = boto3.session.Session() #es_host='search-enrichment-service-ebin6h6q5xc7ehx2snhwbtyl3m.us-west-2.es.amazonaws.com' #es_host=os.environ['es_endpoint'] def __init__(self): credentials=self.session.get_credentials().get_frozen_credentials() self.awsauth = AWSRequestsAuth( aws_access_key=credentials.access_key, aws_secret_access_key=credentials.secret_key, aws_token=credentials.token, aws_host=self.es_host, aws_region=self.session.region_name, aws_service='es' ) def getAuth(self): return self.awsauth def getESHost(self): return self.es_host class credslocal(object): ## init session = boto3.session.Session( region_name='us-east-1', aws_access_key_id='example_key_id', aws_secret_access_key='my_super_secret_key' ) awsauth = None def __init__(self, service): print ("In init") print (boto3.__version__) self.session.resource(service) self.awsauth = AWSRequestsAuth( aws_access_key="credentials.access_key", aws_secret_access_key="credentials.secret_key", aws_token="credentials.token", aws_host="self.es_host", aws_region="us-east-1", aws_service=service ) def getAuth(self): return self.awsauth def getSession(self): return self.session
{ "repo_name": "ctippur/swaggeralliance", "path": "src/python/flask/swagger_server/lib/aws/Creds.py", "copies": "1", "size": "1614", "license": "apache-2.0", "hash": -4570439810450338300, "line_mean": 28.3454545455, "line_max": 94, "alpha_frac": 0.6059479554, "autogenerated": false, "ratio": 3.45610278372591, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9374977765008567, "avg_score": 0.03741459482346852, "num_lines": 55 }
__author__ = 'Călin Sălăgean' class Person(): __id_class = 0 def __init__(self, first_name = None, last_name = None, address = None, id = None): ''' Person class constructor :param name: :param address: :param event: :return: Class instance ''' self.validate(first_name, last_name, address) self.__first_name, self.__last_name, self.__address = first_name, last_name, address self.__id = self.__set_id(id) @staticmethod def validate(first_name, last_name, address): ''' Validation method for Person class :param name: :param address: :param event: :return: None :raise AttributeError: ''' if not first_name or not last_name: raise AttributeError("Please provide full name") if not address: raise AttributeError("Please provide an address") # TODO: Implement validating existing events def __set_id(self, id): ''' Increments the class _id attribute :param id: :return: None ''' if id is None: instance_id = Person.__id_class else: instance_id = id Person.__id_class += 1 return instance_id def get_id(self): ''' Returns person ID :return id: ''' return self.__id def get_name(self): ''' Returns person name :return name: ''' return self.__first_name + ' ' + self.__last_name def get_address(self): ''' Returns person address :return address: ''' return self.__address def update(self, first_name = None, last_name = None, address = None): ''' Update person instance attributes :param first_name: :param last_name: :param address: :return: None ''' if first_name is None: first_name = self.__first_name if last_name is None: last_name = self.__last_name if address is None: address = self.__address self.validate(first_name, last_name, address) self.__first_name, self.__last_name, self.__address = first_name, last_name, address def get_serialization(self): ''' Return JSON serialization :return: JSON Object ''' dict = { 'id': self.__id, 'first_name': self.__first_name, 'last_name': self.__last_name, 'address': self.__address, } return dict @staticmethod def set_class_id(id): ''' Set class _id attribute :param id: :return: None ''' Person.__id_class = id
{ "repo_name": "dooma/Events", "path": "events/models/person.py", "copies": "1", "size": "2816", "license": "mit", "hash": 2847937633247938000, "line_mean": 24.5818181818, "line_max": 92, "alpha_frac": 0.5151084252, "autogenerated": false, "ratio": 4.307810107197549, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5322918532397549, "avg_score": null, "num_lines": null }
__author__ = 'Călin Sălăgean' from datetime import datetime import re class Event(): __id_class = 0 def __init__(self, date = None, time = None, description = None, id = None): ''' Event class contructor :param date: :param time: :param description: :return: Instance variable ''' self.validate(date, time, description) self.__date, self.__time, self.__description, self.__id = date, time, description, self.__set_id(id) def validate(self, date, time, description): ''' Raise exceptions if parameters doesn't have right format :param date: :param time: :param description: :return: None ''' if not date or not time or not description: raise AttributeError("Please provide all fields") try: datetime.strptime(date, '%d/%m/%Y') except ValueError: raise ValueError("Date doesn't have the right format") time = re.sub('\W', '', time) if (len(time) != 4 and len(time) != 2) or not time.isdigit(): raise ValueError("Time doesn't have the right format") def __set_id(self, id): ''' Increments class _id attribute :return: None ''' if id is None: instance_id = Event.__id_class else: instance_id = id Event.__id_class += 1 return instance_id def get_id(self): ''' Returns event id :return: id ''' return self.__id def get_date(self): ''' Returns event date :return: date ''' return self.__date def get_time(self): ''' Returns event time :return: time ''' return self.__time def get_description(self): ''' Returns event description :return: description ''' return self.__description def update(self, date = None, time = None, description = None): ''' Updates event instance :param date: :param time: :param description: :return: ''' if date is None: date = self.__date if time is None: time = self.__time if description is None: description = self.__description self.validate(date, time, description) self.__date, self.__time, self.__description = date, time, description def get_serialization(self): ''' Returns JSON object :return: json ''' dict = { 'id': self.get_id(), 'date': self.get_date(), 'time': self.get_time(), 'description': self.get_description() } return dict @staticmethod def set_class_id(id): ''' Sets class id :param id: :return: ''' Event.__id_class = id
{ "repo_name": "dooma/Events", "path": "events/models/event.py", "copies": "1", "size": "2981", "license": "mit", "hash": -1009934390293902100, "line_mean": 23.825, "line_max": 108, "alpha_frac": 0.5080591001, "autogenerated": false, "ratio": 4.444776119402985, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0027998827659700353, "num_lines": 120 }
__author__ = 'Călin Sălăgean' from events.controllers.event import EventController from events.controllers.person import PersonController class Menu(): text = """ 1. Adaugare persoana 2. Afisare persoane 3. Afisare persoana selectata 4. Stergere persaoana 5. Editare persoana 6. Inscriere persoana la eveniment 7. Adaugare eveniment 8. Afisare evenimente 9. Afisare eveniment selectat 10. Sterge eveniment 11. Editare eveniment 12. Cautare persoane 13. Cautare evenimente 14. Lista de evenimente dupa persoana 15. Lista de persoane inscrise la evenimente ordonate dupa numarul de evenimente 16. Listeaza primele 20% evenimente importante (Alta tasta). Iesire """ def __init__(self): self.menu_text = Menu.text self.person_controller = PersonController() self.event_controller = EventController() self.command_menu = { '1': self.add_person, '2': self.show_all_persons, '3': self.show_person, '4': self.delete_person, '5': self.edit_person, '6': self.associate_person, '7': self.add_event, '8': self.show_all_events, '9': self.show_event, '10': self.delete_event, '11': self.edit_event, '12': self.search_people, '13': self.search_events, '14': self.events_list_by_person, '15': self.people_most_events, '16': self.get_top_events } def print_menu(self): print(self.menu_text) def execute_command(self): self.print_menu() command = input("Introduceti comanda ") try: self.command_menu[command]() self.execute_command() except: print("Iesire") def add_person(self): first_name = input("Introduceti prenume ") last_name = input("Introduceti nume de familie ") address = input("Introduceti adresa ") print(self.person_controller.insert(first_name, last_name, address)) def show_all_persons(self): print(self.person_controller.index()) def show_person(self): id = input("Introduceti id-ul ") print(self.person_controller.show(id)) def delete_person(self): id = input("Introduceti id-ul ") print(self.person_controller.delete(id)) def edit_person(self): id = input("Introduceti id-ul ") first_name = input("Introduceti prenume ") last_name = input("Introduceti nume de familie ") address = input("Introduceti adresa ") print(self.person_controller.update(id, first_name, last_name, address)) def associate_person(self): person_id = input("Introduceti ID persoana ") event_id = input("Introduceti ID event ") data = input("Introduceti data ") print(self.person_controller.associate(person_id, event_id, data)) def add_event(self): date = input("Introduceti data ") time = input("Introduceti ora ") description = input("Introduceti descriere ") print(self.event_controller.insert(date, time, description)) def show_event(self): id = input("Introduceti id-ul ") print(self.event_controller.show(id)) def show_all_events(self): print(self.event_controller.index()) def delete_event(self): id = input("Introduceti id-ul ") print(self.event_controller.delete(id)) def edit_event(self): id = input("Introduceti id-ul ") date = input("Introduceti data ") time = input("Introduceti ora ") description = input("Introduceti descriere ") print(self.event_controller.update(id, date, time, description)) def search_people(self): term = input("Introduceti termenul de cautare ") print(self.person_controller.search(term)) def search_events(self): term = input("Introduceti termenul de cautare ") print(self.event_controller.search(term)) def events_list_by_person(self): person_id = input("Introduceti ID ") print(self.person_controller.get_events(person_id)) def people_most_events(self): print(self.person_controller.get_top_persons()) def get_top_events(self): print(self.event_controller.get_top_events()) menu = Menu() menu.execute_command()
{ "repo_name": "dooma/Events", "path": "start.py", "copies": "1", "size": "4508", "license": "mit", "hash": 1671441266844625400, "line_mean": 29.8630136986, "line_max": 88, "alpha_frac": 0.6057713651, "autogenerated": false, "ratio": 3.533333333333333, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.962970766684497, "avg_score": 0.0018794063176726969, "num_lines": 146 }
__author__ = 'Călin Sălăgean' from events.models.event import Event from events.repositories.event import EventRepository from events.repositories.person_event import PersonEventRepository from events.repositories.person import PersonRepository class EventController(): def __init__(self): ''' EventController constructor :return: object ''' self.repository = EventRepository() self.person_event_repository = PersonEventRepository() self.person_repository = PersonRepository() def index(self): ''' Return all events in one string ready to display :return: string ''' events = self.repository.get_all() output = "" for event in events: output += str(event.get_id()) + "\t" + event.get_date() + "\t" + event.get_time() + "\t" + event.get_description() + "\n" return output def show(self, id): ''' Return a found event ready to display :param id: :return: string ''' try: id = int(id) except ValueError: return "Please enter an integer for ID" try: event = self.repository.get(id) return str(event.get_id()) + "\t" + event.get_date() + "\t" + event.get_time() + "\t" + event.get_description() + "\n" except (ValueError, AttributeError) as e: return e def insert(self, date = None, time = None, description = None): ''' Insert an event :param date: :param time: :param description: :return: string ''' try: event = Event(date, time, description) except (AttributeError, ValueError) as e: return e self.repository.insert(event) return "Event inserted successfully" def update(self, id, date = None, time = None, description = None): ''' Update an event :param id: :param date: :param time: :param description: :return: string ''' try: id = int(id) except ValueError: return "Please insert a valid ID" try: event = self.repository.get(id) except ValueError as e: return e if date == '': date = None if time == '': time = None if description == '': description = None try: event.update(date, time, description) self.repository.update(event) except (ValueError, AttributeError) as e: return e return "Event updated successfully!" def delete(self, id): ''' Delete an event :param id: :return: string ''' try: id = int(id) except ValueError: return "Please enter an integer for ID" try: event = self.repository.get(id) self.repository.delete(event) except (ValueError, AttributeError) as e: return e return "Event was deleted successfully" def get_top_events(self): ''' Show top events with most visitors :return: string ''' relations = self.get_instantiated_relation() occurences = self.determine_occurences(relations) relations.sort( key=lambda rel: EventController.find_dict2(occurences, rel[0].get_id())[rel[0].get_id()], reverse=True ) # Unique relations unique_relations = [] actual_event = -1 for rel in relations: if rel[0].get_id() != actual_event: actual_event = rel[0].get_id() unique_relations.append(rel) output = "" for rel in unique_relations: number_events = EventController.find_dict2(occurences, rel[0].get_id())[rel[0].get_id()] output += str(rel[0].get_id()) + "\t" + rel[0].get_description() + "\tparticipa \t" + str(number_events) + " persoane\n" return output def get_instantiated_relation(self, person_id = None, event_id = None): ''' Instantiate a Repository Many-to-Many relation :param person_id: :param event_id: :return: array ''' if person_id is None and event_id is None: person_events = self.person_event_repository.get_all() else: person_events = self.person_event_repository.get_by_id(person_id, event_id) result = [] for relation in person_events: person = self.person_repository.get(relation.get_person_id()) event = self.repository.get(relation.get_event_id()) date = relation.get_date() if date is None: date = "Doesn't exist" result.append((event, person, date)) return result @staticmethod def determine_occurences(relations): ''' Find occurences in multiple tuples of realtions :param relations: :return: array ''' occurences = [] for relation in relations: event_id = relation[0].get_id() try: occurence = EventController.find_dict2(occurences, event_id) except: occurences.append({ event_id: 0 }) occurence = occurences[-1] finally: occurence[event_id] += 1 return occurences @staticmethod def find_dict(array, id): ''' Find dictionary with given id :param array: :param id: :return: dictionary :raise: ValueError if id is not found ''' for elem in array: try: elem[id] return elem except KeyError: continue raise ValueError @staticmethod def find_dict2(array, id): ''' Find dictionary with given id :param array: :param id: :return: dictionary :raise: ValueError if id is not found ''' if not len(array): raise ValueError try: array[0][id] return array[0] finally: return find_dict2(array[1:], id)
{ "repo_name": "dooma/Events", "path": "events/controllers/event.py", "copies": "1", "size": "6395", "license": "mit", "hash": -3859223540747952600, "line_mean": 26.3205128205, "line_max": 133, "alpha_frac": 0.5292553191, "autogenerated": false, "ratio": 4.3394433129667345, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5368698632066734, "avg_score": null, "num_lines": null }
__author__ = 'Călin Sălăgean' from events.models.person import Person from utils.IO import IO class PersonRepository(): def __init__(self, filename = 'person.json'): ''' PersonRepository class constructor :param filename: Default value 'person.json' :return: ''' self.__defaultfile = filename Person.set_class_id(self.get_max_id()) def insert(self, person): ''' Add a person object into disk storage :param event: :return: None ''' operation = IO(self.__defaultfile) try: people = operation.get() except ValueError: people = [] people.append(person.get_serialization()) operation.set(people) def update(self, person): ''' Updates a person into disk storage :param updated_event: :return: ''' operation = IO(self.__defaultfile) people = operation.get() for per in people: if per['id'] == person.get_id(): people[people.index(per)] = person.get_serialization() operation.set(people) def delete(self, person): ''' Deletes a person from disk storage :param deleted_event: :return: ''' operation = IO(self.__defaultfile) people = operation.get() for per in people: if per['id'] == person.get_id(): people.remove(per) operation.set(people) def get(self, id): ''' Returns a person with provided ID from disk storage :param id: :return: Event instance ''' for per in self.get_all(): if per.get_id() == id: return per raise ValueError('Person not found!') def get_all(self): ''' Returns all persons from disk storage :return: ''' operation = IO(self.__defaultfile) people = [] for person in operation.get(): person_instance = Person(person['first_name'], person['last_name'], person['address'], person['id']) people.append(person_instance) return people def get_max_id(self): ''' Returns the maximum ID found on disk storage :return: ''' people = self.get_all() max = -1 for person in people: if max < person.get_id(): max = person.get_id() return max + 1
{ "repo_name": "dooma/Events", "path": "events/repositories/person.py", "copies": "1", "size": "2520", "license": "mit", "hash": -5382964545277476000, "line_mean": 24.18, "line_max": 112, "alpha_frac": 0.5252284466, "autogenerated": false, "ratio": 4.423550087873462, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.001327656414382078, "num_lines": 100 }
__author__ = 'Călin Sălăgean' from events.repositories.event import EventRepository from events.repositories.person import PersonRepository from datetime import datetime class PersonEvent(): def __init__(self, person_id = None, event_id = None, date = None): ''' PersonEvent constructor :param person_id: :param event_id: :param date: :return: object ''' self.validate(person_id, event_id, date) self.__person_id, self.__event_id, self.__date = person_id, event_id, date @staticmethod def validate(person_id, event_id, date): ''' Validates PersonEvent params :param person_id: :param event_id: :param date: :return: None :raise AttributeError: :raise ValueError: ''' if person_id is None or event_id is None: raise AttributeError('Please provide all fields') event_repository = EventRepository() person_repository = PersonRepository() try: event = event_repository.get(event_id) person = person_repository.get(person_id) except ValueError: raise ValueError('Invalid IDs! Please provide valid IDs') if date is not None: try: datetime.strptime(date, '%d/%m/%Y') except ValueError: raise ValueError("Date doesn't have the right format") def update(self, person_id = None, event_id = None, date = None): ''' Updates realation between person and event :param person_id: :param event_id: :param date: :return: None ''' if person_id is None: person_id = self.__person_id if event_id is None: event_id = self.__event_id if date is None: date = self.__date self.validate(person_id, event_id, date) self.__person_id, self.__event_id, self.__date = person_id, event_id, date def get_person_id(self): ''' Returns person ID :return person_id: ''' return self.__person_id def get_event_id(self): ''' Returns event ID :return event_id: ''' return self.__event_id def get_date(self): ''' Returns date when relation was instantiated :return date: ''' return self.__date def get_serialization(self): ''' Returns a JSON serialization :return: JSON Object ''' dict = { 'person_id': self.__person_id, 'event_id': self.__event_id, 'date': self.__date } return dict
{ "repo_name": "dooma/Events", "path": "events/models/person_event.py", "copies": "1", "size": "2731", "license": "mit", "hash": -7596357108319026000, "line_mean": 26.5656565657, "line_max": 82, "alpha_frac": 0.5450879765, "autogenerated": false, "ratio": 4.302839116719243, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0029635821328909476, "num_lines": 99 }
__author__ = 'Călin Sălăgean' from events.repositories.person_event import PersonEvent, PersonEventRepository from events.repositories.person import PersonRepository from events.repositories.event import EventRepository from events.models.person_event import PersonEvent import re class PersonController(): def __init__(self): ''' PersonController constructor :return: object ''' self.repository = PersonRepository() self.person_event_repository = PersonEventRepository() self.event_repository = EventRepository() def index(self): ''' Return all persons :return: string ''' output = "" for person in self.repository.get_all(): output += str(person.get_id()) + "\t" + person.get_name() + "\t" + person.get_address() + "\n" return output def show(self, id): ''' Return a found person :param id: :return: string ''' try: id = int(id) except ValueError: return "Please enter an integer for ID" try: person = self.repository.get(id) return str(person.get_id()) + "\t" + person.get_name() + "\t" + person.get_address() + "\n" except (ValueError, AttributeError) as e: return e def insert(self, name = None, address = None, event = None): ''' Insert a person :param date: :param time: :param description: :return: string ''' try: person = Person(name, address, event) self.repository.insert(person) except (ValueError, AttributeError) as e: return e return "Inserted successfully" def update(self, id, first_name = None, last_name = None, address = None): ''' Update an event :param id: :param date: :param time: :param description: :return: string ''' try: id = int(id) except ValueError: return "Please insert a valid ID" try: person = self.repository.get(id) except ValueError as e: return e if first_name == '': first_name = None if last_name == '': last_name = None if address == '': address = None try: person.update(first_name, last_name, address) self.repository.update(person) except (ValueError, AttributeError) as e: return e return "Person updated successfully!" def delete(self, id): ''' Delete an event from instances array :param id: :return: string ''' try: id = int(id) except ValueError: return "Please enter an integer for ID" try: person = self.repository.get(id) self.repository.delete(person) except (ValueError, AttributeError) as e: return e return "Person was deleted successfully" def associate(self, person_id, event_id, data): ''' Associate a person with an event :param person_id: :param event_id: :param data: :return: string ''' try: person_id = int(person_id) event_id = int(event_id) except ValueError: return "Please enter an integer for ID" try: person_event = PersonEvent(person_id, event_id, data) self.person_event_repository.insert(person_event) except (ValueError, AttributeError) as e: return e return "Association created successfully" def search(self, term): ''' Search people by term :param term: :return: string ''' found_people = [] for person in self.repository.get_all(): person_criteria = person.get_name() + person.get_address() if (re.match(term.lower(), person_criteria.lower()) is not None): found_people.append(person) output = "" for person in found_people: output += str(person.get_id()) + "\t" + person.get_name() + "\t" + person.get_address() + "\n" return output def get_events(self, id): ''' Return events that a person has participated in :param id: :return: string ''' try: id = int(id) except ValueError: return "Please enter an integer for ID" output = "" try: relations = self.get_instantiated_relation(id) except: print("The id doesn't exist") relations.sort(key=lambda rel: (rel[0].get_description(), rel[0].get_date())) for relation in relations: event = relation[0] person = relation[1] date = relation[2] output += str(event.get_id()) + "\t" + event.get_description() + "\t Inscreire in data: " + str(date) + "\n" return output def get_top_persons(self): ''' Return top people with most visits :return: string ''' relations = self.get_instantiated_relation() occurences = self.determine_occurences(relations) relations.sort( key=lambda rel: PersonController.find_dict(occurences, rel[1].get_id())[rel[1].get_id()], reverse=True ) # Unique relations unique_relations = [] actual_person = -1 for rel in relations: if rel[1].get_id() != actual_person: actual_person = rel[1].get_id() unique_relations.append(rel) output = "" for rel in unique_relations: number_events = PersonController.find_dict(occurences, rel[1].get_id())[rel[1].get_id()] output += rel[1].get_name() + "\tparticipa la\t" + str(number_events) + "\n" return output def get_instantiated_relation(self, person_id = None, event_id = None): ''' Reinstantiate a relation between person and event from repository :param person_id: :param event_id: :return: string ''' if person_id is None and event_id is None: person_events = self.person_event_repository.get_all() else: person_events = self.person_event_repository.get_by_id(person_id, event_id) result = [] for relation in person_events: event = self.event_repository.get(relation.get_event_id()) person = self.repository.get(relation.get_person_id()) date = relation.get_date() if date is None: date = "Doesn't exist" result.append((event, person, date)) return result @staticmethod def determine_occurences(relations): ''' Find occurences in multiple tuples of realtions :param relations: :return: array ''' occurences = [] for relation in relations: person_id = relation[1].get_id() try: occurence = PersonController.find_dict(occurences, person_id) except: occurences.append({ person_id: 0 }) occurence = occurences[-1] finally: occurence[person_id] += 1 return occurences @staticmethod def find_dict(array, id): ''' Find dictionary with given id :param array: :param id: :return: dictionary :raise: ValueError if id is not found ''' for elem in array: try: elem[id] return elem except KeyError: continue raise ValueError
{ "repo_name": "dooma/Events", "path": "events/controllers/person.py", "copies": "1", "size": "7945", "license": "mit", "hash": -2761423835787932000, "line_mean": 27.1666666667, "line_max": 120, "alpha_frac": 0.5333669101, "autogenerated": false, "ratio": 4.318651441000544, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5352018351100544, "avg_score": null, "num_lines": null }
__author__ = 'Călin Sălăgean' from utils.IO import IO from events.models.event import Event class EventRepository(): def __init__(self, filename = 'event.json'): ''' EventRepository class constructor :param filename: Default value 'event.json' :return: ''' self.__datafile = filename Event.set_class_id(self.get_max_id()) def insert(self, event): ''' Add an event object into disk storage :param event: :return: None ''' operation = IO(self.__datafile) try: data = operation.get() except ValueError: data = [] data.append(event.get_serialization()) operation.set(data) def update(self, updated_event): ''' Updates an event into disk storage :param updated_event: :return: ''' operation = IO(self.__datafile) events = operation.get() for event in events: if event['id'] == updated_event.get_id(): events[events.index(event)] = updated_event.get_serialization() break operation.set(events) def delete(self, deleted_event): ''' Deletes an event from disk storage :param deleted_event: :return: ''' operation = IO(self.__datafile) events = operation.get() for event in events: if event['id'] == deleted_event.get_id(): events.remove(event) break operation.set(events) def get(self, id): ''' Returns an event with provided ID from disk storage :param id: :return: Event instance ''' operation = IO(self.__datafile) for event in operation.get(): if event['id'] == id: return Event(event['date'], event['time'], event['description'], event['id']) raise ValueError('Event not found!') def get_all(self): ''' Returns all events from disk storage :return: ''' operation = IO(self.__datafile) events = [] for event in operation.get(): instance = Event(event['date'], event['time'], event['description'], event['id']) events.append(instance) return events def get_max_id(self): ''' Returns the maximum ID found on disk storage :return: ''' events = self.get_all() max = -1 for event in events: if max < event.get_id(): max = event.get_id() return max + 1
{ "repo_name": "dooma/Events", "path": "events/repositories/event.py", "copies": "1", "size": "2655", "license": "mit", "hash": -5491324509669188000, "line_mean": 25.0098039216, "line_max": 93, "alpha_frac": 0.5214932127, "autogenerated": false, "ratio": 4.47972972972973, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0014465438012729271, "num_lines": 102 }
__author__ = 'Călin Sălăgean' from utils.IO import IO from events.models.person_event import PersonEvent class PersonEventRepository(): def __init__(self, filename = None): self.__defaultfile = filename or 'person_event.json' def insert(self, person_event): operation = IO(self.__defaultfile) try: person_events = operation.get() except ValueError: person_events = [] person_events.append(person_event.get_serialization()) operation.set(person_events) def get_all(self): operation = IO(self.__defaultfile) person_events = [] for person_event in operation.get(): person_events.append(PersonEvent(person_event['person_id'], person_event['event_id'], person_event['date'])) return person_events def get_by_id(self, person_id = None, event_id = None): operation = IO(self.__defaultfile) person_events = [] for person_event in operation.get(): if person_id is not None and event_id is not None: if person_event['person_id'] == person_id and person_event['event_id'] == event_id: return [PersonEvent(person_id, event_id, person_event['date'])] elif person_id is None: if person_event['event_id'] == event_id: person_events.append(PersonEvent(person_event['person_id'], event_id, person_event['date'])) else: if person_event['person_id'] == person_id: person_events.append(PersonEvent(person_id, person_event['event_id'], person_event['date'])) return person_events
{ "repo_name": "dooma/Events", "path": "events/repositories/person_event.py", "copies": "1", "size": "1682", "license": "mit", "hash": 2032470505474927000, "line_mean": 34.7446808511, "line_max": 120, "alpha_frac": 0.5979749851, "autogenerated": false, "ratio": 3.9786729857819907, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.004921017967252904, "num_lines": 47 }
__author__ = 'Călin Sălăgean' import os, json class IO(): def __init__(self, filename = None): ''' IO Constructor :param filename: Default None :return: object ''' self.validate(filename) self.__filename = filename def get(self): ''' Returns parsed file content :return: dictionary :raise: ValueError if content is invalid ''' file = open(IO.filepath(self.__filename)) try: content = json.loads(file.read()) return content except: raise ValueError('The content of ' + self.__filename + ' is not valid!') finally: file.close() def set(self, data): ''' Writes data into file as JSON content :param data: dictionary (JSON format) :return: None ''' temporary_file_prefix = 'tmp_' temporary_file = open(IO.filepath(temporary_file_prefix + self.__filename), 'w') json.dump(data, temporary_file) temporary_file.close() if temporary_file.closed: os.remove(IO.filepath(self.__filename)) os.rename(temporary_file.name, IO.filepath(self.__filename)) @staticmethod def filepath(filename): ''' Static method that creates relative path for given filename :param filename: :return: string ''' data_directory = 'data/' return os.path.abspath(data_directory + filename) @staticmethod def validate(filename): ''' Validates filename. If file does not exist, it will create the file :param filename: :return: :raise: AttributeError if filename is invalid ''' if filename == None or not len(str(filename)) or filename.split('.')[1] != 'json': raise AttributeError('Please provide a valid file name!') if not os.path.exists(IO.filepath(filename)): open(IO.filepath(filename), 'w')
{ "repo_name": "dooma/Events", "path": "utils/IO.py", "copies": "1", "size": "2023", "license": "mit", "hash": -5437359181193553000, "line_mean": 28.2898550725, "line_max": 90, "alpha_frac": 0.5658415842, "autogenerated": false, "ratio": 4.478935698447893, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.004707532996841141, "num_lines": 69 }
__author__ = 'Călin Sălăgean' import unittest from events.models.event import Event class TestEvent(unittest.TestCase): def test_initialization(self): event = Event('14/01/2015', '14:00', 'Colectiv') self.assertIsInstance(event, Event) def test_date_input_validation(self): with self.assertRaises(AttributeError): event = Event(time = '14:00', description = 'Colectiv') def test_time_input_validation(self): with self.assertRaises(AttributeError): event = Event(date = '14/01/2015', description = 'Colectiv') def test_valid_properties(self): event = Event('14/01/2015', '14:00', 'Colectiv') self.assertEqual(event.date, '14/01/2015') self.assertEqual(event.time, '14:00') self.assertEqual(event.description, 'Colectiv') def test_date_format_validation(self): with self.assertRaisesRegexp(ValueError, "Date doesn't have the right format"): event = Event('14012015', '14:00', 'Colectiv') def test_time_format_validation(self): with self.assertRaisesRegex(ValueError, "Time doesn't have the right format"): event = Event('14/01/2015', '1400', 'Colectiv')
{ "repo_name": "dooma/Events", "path": "spec/models/test_event.py", "copies": "1", "size": "1213", "license": "mit", "hash": 4103978191900530000, "line_mean": 38.064516129, "line_max": 87, "alpha_frac": 0.652892562, "autogenerated": false, "ratio": 3.5798816568047336, "config_test": true, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4732774218804734, "avg_score": null, "num_lines": null }
__author__ = 'Călin Sălăgean' import unittest from events.models.person_event import PersonEvent class TestPersonEvent(unittest.TestCase): def test_instance(self): with self.assertRaises(ValueError): person_event = PersonEvent(1, 2) self.assertIsInstance(person_event, PersonEvent) def test_validation(self): with self.assertRaisesRegex(AttributeError, 'Please provide all fields'): person_event = PersonEvent() person_event = PersonEvent(person_id=1) person_event = PersonEvent(event_id=1) with self.assertRaisesRegex(ValueError, 'Invalid IDs! Please provide valid IDs'): person_event = PersonEvent(1, 1) person_event = PersonEvent(25, 25) def test_update(self): person_event = PersonEvent(25, 25) person_event.update(25,84) self.assertEqual(person_event.get_person_id(), 25) self.assertEqual(person_event.get_event_id(), 84) def test_get_serialization(self): person_event = PersonEvent(25, 25) serialization = person_event.get_serialization() self.assertEqual(serialization['person_id'], 25) self.assertEqual(serialization['event_id'], 25)
{ "repo_name": "dooma/Events", "path": "spec/models/test_person_event.py", "copies": "1", "size": "1239", "license": "mit", "hash": -7105311679982183000, "line_mean": 32.4324324324, "line_max": 89, "alpha_frac": 0.6618122977, "autogenerated": false, "ratio": 4.012987012987013, "config_test": true, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5174799310687013, "avg_score": null, "num_lines": null }
__author__ = 'Călin Sălăgean' import unittest, os from utils.IO import IO from events.repositories.event import EventRepository from events.models.event import Event class TestEventRepository(unittest.TestCase): def test_insert(self): io = IO('test.json') io.set([]) event = Event('14/01/2015', '14:00', 'Colectiv') EventRepository('test.json').insert(event) content = io.get() self.assertIs(type(content), list) stored_event = content[0] self.assertEqual(stored_event['date'], '14/01/2015') self.assertEqual(stored_event['time'], '14:00') self.assertEqual(stored_event['description'], 'Colectiv') def test_get_all(self): io = IO('test.json') io.set([]) repo = EventRepository('test.json') content = repo.get_all() self.assertEqual(content, []) event1 = Event('14/01/2015', '14:00', 'Colectiv') event2 = Event('14/01/2015', '14:00', 'Colectiv') repo.insert(event1) repo.insert(event2) content = repo.get_all() self.assertIs(type(content), list) self.assertEqual(len(content), 2) stored_event1 = content[0] stored_event2 = content[1] self.assertEqual(stored_event1.get_date(), '14/01/2015') self.assertEqual(stored_event1.get_time(), '14:00') self.assertEqual(stored_event1.get_description(), 'Colectiv') self.assertEqual(stored_event2.get_date(), '14/01/2015') self.assertEqual(stored_event2.get_time(), '14:00') self.assertEqual(stored_event2.get_description(), 'Colectiv') def test_get(self): Event.set_class_id(0) io = IO('test.json') io.set([]) repo = EventRepository('test.json') event1 = Event('14/01/2015', '14:00', 'Colectiv') event2 = Event('14/01/2015', '14:00', 'Colectiv') repo.insert(event1) repo.insert(event2) content = repo.get(0) self.assertIs(type(content), Event) self.assertEqual(content.get_date(), '14/01/2015') self.assertEqual(content.get_time(), '14:00') self.assertEqual(content.get_description(), 'Colectiv') with self.assertRaisesRegex(ValueError, 'Event not found!'): content = repo.get(10) def test_update(self): Event.set_class_id(0) io = IO('test.json') io.set([]) repo = EventRepository('test.json') event1 = Event('14/01/2015', '14:00', 'Colectiv') event2 = Event('14/01/2015', '14:00', 'Colectiv') repo.insert(event1) repo.insert(event2) event = repo.get(0) self.assertIs(type(event), Event) event.update('15/01/2015', '15:00', 'Untold festival') repo.update(event) updated_event = repo.get(0) self.assertEqual(updated_event.get_date(), '15/01/2015') self.assertEqual(updated_event.get_time(), '15:00') self.assertEqual(updated_event.get_description(), 'Untold festival') def test_get_max_id(self): Event.set_class_id(0) io = IO('test.json') io.set([]) repo = EventRepository('test.json') self.assertEqual(repo.get_max_id(), 0) content = repo.get_all() self.assertEqual(content, []) event1 = Event('14/01/2015', '14:00', 'Colectiv') event2 = Event('14/01/2015', '14:00', 'Colectiv') repo.insert(event1) repo.insert(event2) content = repo.get_all() self.assertEqual(repo.get_max_id(), 2) def test_delete(self): Event.set_class_id(0) io = IO('test.json') io.set([]) repo = EventRepository('test.json') event1 = Event('14/01/2015', '14:00', 'Colectiv') event2 = Event('14/01/2015', '14:00', 'Colectiv') repo.insert(event1) repo.insert(event2) event = repo.get(0) self.assertIs(type(event), Event) repo.delete(event) with self.assertRaisesRegex(ValueError, 'Event not found!'): updated_event = repo.get(0)
{ "repo_name": "dooma/Events", "path": "spec/repositories/test_event.py", "copies": "1", "size": "4099", "license": "mit", "hash": -3537187086741714000, "line_mean": 30.2748091603, "line_max": 76, "alpha_frac": 0.5842285156, "autogenerated": false, "ratio": 3.34094616639478, "config_test": true, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9421457882144557, "avg_score": 0.0007433599700443378, "num_lines": 131 }
__author__ = "cuongnb14@gmail.com" from sqlalchemy.ext.declarative import declarative_base from sqlalchemy import Column, Integer, String, Float from sqlalchemy.orm import relationship, backref from sqlalchemy import ForeignKey Base = declarative_base() class WebApp(Base): """Mapping with table apps""" __tablename__ = 'autoscaling_web_app' id = Column(Integer, primary_key=True) name = Column(String, unique=True) uuid = Column(String, unique=True) min_instances = Column(Integer) max_instances = Column(Integer) cpus = Column(Float) mem = Column(Float) policies = relationship("Policy", order_by="Policy.id", backref="app", cascade="all, delete, delete-orphan") class Policy(Base): """Mapping with table policies""" __tablename__ = 'autoscaling_policies' id = Column(Integer, primary_key=True) web_app_id = Column(String, ForeignKey('autoscaling_web_app.id')) metric_type = Column(String) upper_threshold = Column(Float) lower_threshold = Column(Float) instances_out = Column(Integer) instances_in = Column(Integer) scale_up_wait = Column(Integer) scale_down_wait = Column(Integer)
{ "repo_name": "cuongnb14/autoscaling-paas", "path": "project/autoscaling/autoscaling/model.py", "copies": "1", "size": "1175", "license": "mit", "hash": 3092280253413701600, "line_mean": 32.5714285714, "line_max": 112, "alpha_frac": 0.7004255319, "autogenerated": false, "ratio": 3.753993610223642, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9938869205334767, "avg_score": 0.0031099873577749688, "num_lines": 35 }
__author__ = 'cuongnb14@gmail.com' import logging class Scaler(object): """docstring for ClassName""" def __init__(self, marathon_client, app): self.logger = logging.getLogger("Scaler") self.marathon_client = marathon_client self.app = app def scale(self, delta): """sacle app_name (add or remove) delta intances @param int delta number intances add if (delta > 0) or remove if (delta < 0) """ current_instances = self.marathon_client.get_app("app-"+self.app.uuid).instances new_instance = current_instances + delta if(new_instance > self.app.max_instances): new_instance = self.app.max_instances if(new_instance < self.app.min_instances): new_instance = self.app.min_instances if(new_instance != current_instances): self.logger.info("scale to: %d", new_instance) self.marathon_client.scale_app("app-"+self.app.uuid, new_instance) else: self.logger.info("number instances to threshold, no scale!")
{ "repo_name": "cuongnb14/autoscaling-paas", "path": "project/autoscaling/autoscaling/scaler.py", "copies": "1", "size": "1070", "license": "mit", "hash": 8568042082940434000, "line_mean": 38.6296296296, "line_max": 88, "alpha_frac": 0.6224299065, "autogenerated": false, "ratio": 3.7152777777777777, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4837707684277778, "avg_score": null, "num_lines": null }
class dstat_plugin(dstat): """ EVENTS COMMENT. """ def __init__(self): self.name = 'latency' self.type = 'p' self.width = 4 self.scale = 34 def vars(self): ret = [] for name in glob.glob('/tmp/dstat/sla/client[0-9]*'): ret.append(os.path.basename(name)) ret.sort() return ret def nick(self): return [name.lower() for name in self.vars] def extract(self): for csla in self.vars: f = open('/tmp/dstat/sla/'+csla+'/latency', 'r') st = f.readline().rstrip() if(st == ""): self.val[csla] = 0 else: self.val[csla] = float(st) f.close() def check(self): for csla in glob.glob('/tmp/dstat/sla/client[0-9]*'): if not os.access(csla+'/latency', os.R_OK): raise Exception, 'Cannot access latency %s information' % os.path.basename(csla) # vim:ts=4:sw=4:et
{ "repo_name": "barzan/dbseer", "path": "middleware_old/dstat_for_server/plugins/dstat_client_latency.py", "copies": "4", "size": "1026", "license": "apache-2.0", "hash": -3124171144379731500, "line_mean": 24.65, "line_max": 96, "alpha_frac": 0.5019493177, "autogenerated": false, "ratio": 3.320388349514563, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5822337667214563, "avg_score": null, "num_lines": null }
#pads with spaces and returns string,so it can be divided exactly by a certain amount def sim_pad(MESSAGE, PAD_TO_MULT): MESSAGE = MESSAGE + "1" pad = (((len(MESSAGE)/PAD_TO_MULT)+1)*PAD_TO_MULT)-len(MESSAGE) for i in range(pad): MESSAGE = MESSAGE + " " return MESSAGE #pads a string so it can be divided exactly by a number but with a choosen charcater #also returns both the padded string and value of how many characters it has paded by, needed for unpadding #returns two values in a tuple def sim_pad_with_char(MESSAGE, PAD_TO_MULT, CHAR_TO_PAD_WITH): if len(CHAR_TO_PAD_WITH)== 1: pad = (((len(MESSAGE)/PAD_TO_MULT)+1)*PAD_TO_MULT)-len(MESSAGE) for i in range(pad): MESSAGE = MESSAGE + CHAR_TO_PAD_WITH return (MESSAGE, pad) else: raise ValueError('ValueError: legnth of CHAR_TO_PAD_WITH must only be 1') #pads a string with a certain character and certain amount def sim_pad_with_amount(MESSAGE, CHAR_TO_PAD_WITH, PAD_AMOUNT): for i in range(PAD_AMOUNT): MESSAGE = MESSAGE + CHAR_TO_PAD_WITH return (MESSAGE) #unpads a padded string def sim_unpad(PADDED_MES): PADDED_MES = PADDED_MES.rstrip()[:-1] return PADDED_MES #unpads a string that is padded with a certain character #requires how many extra characters the string has been padded with def sim_unpad_with_char(PADDED_MES, PAD_AMOUNT): PAD_AMOUNT = PAD_AMOUNT*-1 PADDED_MES = PADDED_MES[:PAD_AMOUNT] return PADDED_MES #unpad a string based on the amount it was padded by def sim_unpad_with_amount(PADDED_MES, PAD_AMOUNT): PAD_AMOUNT = PAD_AMOUNT*-1 PADDED_MES = PADDED_MES[:PAD_AMOUNT] return PADDED_MES
{ "repo_name": "Curiousteenager/simplepad", "path": "simplepad.py", "copies": "1", "size": "1915", "license": "mit", "hash": -2354845574991672000, "line_mean": 35.5490196078, "line_max": 134, "alpha_frac": 0.6804177546, "autogenerated": false, "ratio": 3.0837359098228663, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4264153664422866, "avg_score": null, "num_lines": null }
__author__ = 'Curtis Mackay' pe_index = 0; import itertools class PE(object): """ Data structure for storing PE dependency data """ newid = itertools.count().__next__ def __init__(self, operation, pe_inputs, pe_outputs, pe_id=None): self.operation = operation self.pe_inputs = pe_inputs self.pe_outputs = pe_outputs if pe_id is None: self.pe_id = PE.newid() else: self.pe_id = pe_id def __repr__(self): return "PE %s" % self.pe_id # return "PE inputs:%s, outputs:%s" % (self.pe_inputs, self.pe_outputs) def __str__(self): return "PE inputs:%s, outputs:%s" % (self.pe_inputs, self.pe_outputs) def compress_mapping(pe_list, net_map): for pe in pe_list: for i in range(len(pe.pe_inputs)): for net_set in net_map: if pe.pe_inputs[i] in net_set: pe.pe_inputs[i] = net_set[0] for i in range(len(pe.pe_outputs)): for net_set in net_map: if pe.pe_outputs[i] in net_set: pe.pe_outputs[i] = net_set[0] def create_dependency_list(pe_list): dep_list = list() for pe in pe_list: for pe_dep_candidate in pe_list: input_set = set(pe.pe_inputs) output_set = set(pe_dep_candidate.pe_outputs) if input_set & output_set: new_dependency = True for dep in dep_list: if dep[-1] is pe_dep_candidate: dep.append(pe) new_dependency = False if new_dependency: dep_list.append([pe_dep_candidate, pe]) # Add in any PE's that have no dependency for pe in pe_list: # optimized way to flatten list flattened_dep_list = [item for sublist in dep_list for item in sublist] if pe not in set(flattened_dep_list): dep_list.append([pe]) return dep_list
{ "repo_name": "crnmrn/SPADE-ARCH", "path": "compiler/dpda_compiler/parser/dependency_tracker.py", "copies": "1", "size": "1991", "license": "bsd-3-clause", "hash": 392705441356094300, "line_mean": 31.6557377049, "line_max": 79, "alpha_frac": 0.5364138624, "autogenerated": false, "ratio": 3.555357142857143, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9572226805638843, "avg_score": 0.003908839923660002, "num_lines": 61 }
__author__ = 'cutylewiwi' # -*- coding:utf-8 -*- import urllib import urllib2 import re class Group: def __init__(self, name, faculties): self.name = name self.faculties = faculties class GroupList: def __init__(self, baseUrl): self.baseUrl = baseUrl self.groups = [] def getPage(self, _url): try: url = _url request = urllib2.Request(url) response = urllib2.urlopen(request) # print response.read() return response except urllib2.URLError, e: if hasattr(e,"reason"): print u"Faild to get Prof List at University of Delaware",e.reason return None def getGroupList(self): page = self.getPage(self.baseUrl) regex = '<article>(.*?)</article>' myItems = re.findall(regex, page.read(), re.S) # print myItems tmpStr = myItems[0] regex = '<li>\n<h3.*?>(.*?)</h3>.*?</ul>.*?<p.*?>(.*?)</p>(.*?)</p>' myItems = re.findall(regex, tmpStr, re.S) print len(myItems) print myItems return for item in myItems: # print item GroupName = item[0] regex = '<a.*?>(.*?)</a>' GroupFaculties = re.findall(regex, item[1], re.S) self.groups.append(Group(GroupName, GroupFaculties)) # print ProfName # print ProfPhotoUrl # print ProfPUrl # print ProfTitle # print ProfArea # print ProfOffice # print " " # self.profs.append(Prof(ProfName, ProfPhotoUrl, ProfPUrl, ProfTitle, ProfArea, ProfOffice)) def outPutProf(self): result = "<?xml version=\"1.0\" ?>\n\t<institution>\n" self.getGroupList() for group in self.groups: result += "\t\t<research>\n" result += "\t\t\t<groupname>%s</groupname>\n" % (group.name) for faculty in group.faculties: tmpStr = faculty if tmpStr[:3] == "Dr.": tmpStr = tmpStr[4:] result += "\t\t\t<professorname>%s</professorname>\n" % (tmpStr) result += "\t\t</research>\n" result += "\t</institution>\n" # print result file = open("UTA_research.xml","w") file.writelines(result) baseURL = 'https://cse.uta.edu/research/research-areas.php' pl = GroupList(baseURL) # pl.getPage() pl.outPutProf() # pl.getGroupList()
{ "repo_name": "doge-search/webdoge", "path": "wuyudong/UTA/UTA_research.py", "copies": "1", "size": "2593", "license": "unlicense", "hash": 7706575861810142000, "line_mean": 30.012345679, "line_max": 104, "alpha_frac": 0.5106054763, "autogenerated": false, "ratio": 3.5963938973647713, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.46069993736647713, "avg_score": null, "num_lines": null }
__author__ = 'cutylewiwi' # -*- coding:utf-8 -*- import urllib import urllib2 import re class Prof: def __init__(self, name, photoUrl, pUrl, title, area, office, phone, email): self.name = name self.photoUrl = photoUrl self.pUrl = pUrl self.title = title self.area = area self.office = office self.phone = phone self.email = email class ProfList: def __init__(self, baseUrl): self.baseUrl = baseUrl self.profs = [] def getPage(self): try: url = self.baseUrl request = urllib2.Request(url) response = urllib2.urlopen(request) # print response.read() return response except urllib2.URLError, e: if hasattr(e,"reason"): print u"Faild to get Prof List at KSU",e.reason return None def getProfList(self): page = self.getPage() # regex = '<div class=\"wdn-grid-set\">.*?<a class=\"wdn-button\" title=\"Web page for.*?\" href=\"(.*?)\".*?<strong>(.*?)</strong>.*?<div class=\"gs-fac-rsch\">(.*?)(<br />)?</div>' # regex = '<div id="node-2964" class="node node-type-page build-mode-full clearfix">.*?<div class="content">.*?<div>(.*?)<div>.*?<div>.*?</div>' regex = '<table class="views-view-grid col-2" summary="Faculty of the CIS Department">(.*?)</table>' myItems = re.findall(regex, page.read(), re.S) # print myItems[0] # return for item in myItems: tmpStr = item regex = '<div class="picture">.*?<img.*?src="(.*?)".*?>.*?</div>' myTokens = re.findall(regex, tmpStr, re.S) # print myTokens ProfPhotoUrl = myTokens[0] regex = '<div class="views-field-value">.*?<a.*?>(.*?)</a>.*?</div>' myTokens = re.findall(regex, tmpStr, re.S) ProfName = myTokens[0] regex = '<div class="views-field-value-1">.*?<b>(.*?)</b>.*?</div>' myTokens = re.findall(regex, tmpStr, re.S) ProfTitle = myTokens[0] regex = '<div class="views-field-name">.*?<a.*?>(.*?)</a>.*?</div>' myTokens = re.findall(regex, tmpStr, re.S) ProfPUrl = myTokens[0] ProfArea = "" regex = '<div class="views-field-value-2">.*?<span class="field-content">(.*?)</span>.*?</div>' myTokens = re.findall(regex, tmpStr, re.S) ProfOffice = myTokens[0] regex = '<div class="views-field-value-3">.*?<span class="field-content">(.*?)</span>.*?</div>' myTokens = re.findall(regex, tmpStr, re.S) ProfPhone = myTokens[0] regex = '<div class="views-field-mail">.*?<a.*?>(.*?)</a>.*?</div>' myTokens = re.findall(regex, tmpStr, re.S) ProfEmail = myTokens[0] # print ProfName # print ProfPhotoUrl # print ProfPUrl # print ProfTitle # print ProfArea # print ProfOffice # print ProfPhone # print ProfEmail # print " " self.profs.append(Prof(ProfName, ProfPhotoUrl, ProfPUrl, ProfTitle, ProfArea, ProfOffice, ProfPhone, ProfEmail)) def outPutProf(self): result = "<?xml version=\"1.0\" ?>\n\t<institution>\n" self.getProfList() for prof in self.profs: result += "\t\t<professor>\n" result += "\t\t\t<name>%s</name>\n" % (prof.name) result += "\t\t\t<title>%s</title>\n" % (prof.title) result += "\t\t\t<office>%s</office>\n" % (prof.office) result += "\t\t\t<email>%s</email>\n" % (prof.email) result += "\t\t\t<phone>%s</phone>\n" % (prof.phone) result += "\t\t\t<website>%s</website>\n" % (prof.pUrl) result += "\t\t\t<image>%s</image>\n" % (prof.photoUrl) result += "\t\t</professor>\n" result += "\t</institution>\n" # print result fileName = "KSU.xml" outputDir = "result" file = open(fileName,"w") file.writelines(result) baseURL = 'https://www.cis.ksu.edu/people/current-faculty' pl = ProfList(baseURL) pl.outPutProf() # pl.getPage() # pl.getProfList()
{ "repo_name": "doge-search/webdoge", "path": "wuyudong/KSU/KSU.py", "copies": "1", "size": "4385", "license": "unlicense", "hash": 9212551889555674000, "line_mean": 38.2293577982, "line_max": 190, "alpha_frac": 0.5069555302, "autogenerated": false, "ratio": 3.404503105590062, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9397932742735269, "avg_score": 0.0027051786109585888, "num_lines": 109 }
__author__ = 'cutylewiwi' # -*- coding:utf-8 -*- import urllib import urllib2 import re class Prof: def __init__(self, name, photoUrl, pUrl, title, area, office): self.name = name self.photoUrl = photoUrl self.pUrl = pUrl self.title = title self.area = area self.office = office class ProfList: def __init__(self, baseUrl): self.baseUrl = baseUrl self.profs = [] def getPage(self): try: url = self.baseUrl request = urllib2.Request(url) response = urllib2.urlopen(request) # print response.read() return response except urllib2.URLError, e: if hasattr(e,"reason"): print u"Faild to get Prof List at University of Delaware",e.reason return None def getProfList(self): page = self.getPage() #regex = '<div class=\"one-third column\">.*?<li>(.*?)</li>.*?</div>' regex = '<div class="one-third column clear">.*?<h2>Full-Time</h2>.*?<ul>(.*?)</ul>.*?</div>' #regex = '<tr>.*?<img src=\"(.*?)\".*?</tr>' myItems = re.findall(regex, page.read(), re.S) tmpStr = myItems[0] # print tmpStr regex = '<li>.*?<a.*?href="(.*?)">(.*?)</a>.*?</li>' myItems = re.findall(regex, tmpStr, re.S) # print myItems # return for item in myItems: # print item ProfName = item[1] ProfPhotoUrl = "" ProfPUrl = "http://www.eecs.utk.edu/people/faculty" + item[0] ProfTitle = "" ProfArea = "" ProfOffice = "" # print ProfName # print ProfPhotoUrl # print ProfPUrl # print ProfTitle # print ProfArea # print ProfOffice # print " " self.profs.append(Prof(ProfName, ProfPhotoUrl, ProfPUrl, ProfTitle, ProfArea, ProfOffice)) def outPutProf(self): result = "<?xml version=\"1.0\" ?>\n\t<institution>\n" self.getProfList() for prof in self.profs: result += "\t\t<professor>\n" result += "\t\t\t<name>%s</name>\n" % (prof.name) result += "\t\t\t<title>%s</title>\n" % (prof.title) result += "\t\t\t<office>%s</office>\n" % (prof.office) result += "\t\t\t<email></email>\n" result += "\t\t\t<phone></phone>\n" result += "\t\t\t<website>%s</website>\n" % (prof.pUrl) result += "\t\t\t<image>%s</image>\n" % (prof.photoUrl) result += "\t\t</professor>\n" result += "\t</institution>\n" # print result file = open("UTK.xml","w") file.writelines(result) baseURL = 'http://www.eecs.utk.edu/people/faculty/' pl = ProfList(baseURL) # pl.getPage() pl.outPutProf() # pl.getProfList()
{ "repo_name": "doge-search/webdoge", "path": "wuyudong/UTK/UTK.py", "copies": "1", "size": "2980", "license": "unlicense", "hash": -5998241134554957000, "line_mean": 31.4831460674, "line_max": 102, "alpha_frac": 0.4946308725, "autogenerated": false, "ratio": 3.401826484018265, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.43964573565182646, "avg_score": null, "num_lines": null }
__author__ = 'cvdsouza' #Developed at : SEFCOM Labs by JJ SEO and Clinton Dsouza from bluetooth import * import threading class MyThread (threading.Thread): def __init__ (self, socket): self.socket = socket threading.Thread.__init__(self) def run (self): try: str = raw_input('Enter your input:') self.socket.send (str) except: self.socket.close() print "socket close" server_sock = BluetoothSocket(RFCOMM) server_sock.bind(("", PORT_ANY)) server_sock.listen(1) port = server_sock.getsockname()[1] uuid = "a60f35f0-b93a-11de-8a39-08002009c666" #uuid = "04c6093b-0000-1000-8000-00805f9b34fb" #uuid = "84f39d29-7d6d-437d-973b-fba39e49d4ee" advertise_service(server_sock, "SampleServer", service_id=uuid, service_classes=[uuid, SERIAL_PORT_CLASS], profiles=[SERIAL_PORT_PROFILE], # protocols = [ OBEX_UUID ] ) print("Waiting for connection on RFCOMM channel %d" % port) client_sock, client_info = server_sock.accept() print("Accepted connection from ", client_info) try: while True: MyThread(client_sock).start() data = client_sock.recv(1024) if len(data) == 0: break print("received [%s]" % data) with open("STS.json","a") as myfile: myfile.write('\n'+data) except IOError: pass print("disconnected") client_sock.close() server_sock.close() print("all done")
{ "repo_name": "jtromo/ASU-Thesis-RaspberryPiSmartKey", "path": "RaspberryPiSmartKey/.sync/Archive/PyBlueServer.py", "copies": "1", "size": "1412", "license": "apache-2.0", "hash": 2074202821976899000, "line_mean": 23.7719298246, "line_max": 92, "alpha_frac": 0.652266289, "autogenerated": false, "ratio": 2.966386554621849, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4118652843621849, "avg_score": null, "num_lines": null }
__author__ = 'cvl' def numberic_system(base_system): result = {} for csc_n in base_system: result[csc_n] = base_system.find(csc_n) return result class CSC(object): system = {} def __init__(self, alphabet): self.system = numberic_system(alphabet) def csc(self, sym): result = '' for s in sym: result += str(self.system[s]) return result def r_csc(self, num): for key in self.system.keys(): if self.system[key] == int(num): return key return 'out_of_range' def increment(self, csc_number): csc_len = len(csc_number) i = 0 while 1: if i > csc_len: csc_number += '0' if i == csc_len: csc_number += '0' break num = csc_number[i] if num in self.system.keys(): csc_result = self.r_csc(int(self.csc(num)) + 1) if csc_result != 'out_of_range': csc_number = csc_number[:i] + csc_result + csc_number[i + 1:] break else: csc_number = csc_number[:i] + '0' + csc_number[i + 1:] i += 1 else: csc_number = csc_number[:i] + '0' + csc_number[i + 1:] i += 1 return csc_number
{ "repo_name": "cludtk/freedomain", "path": "csc.py", "copies": "1", "size": "1394", "license": "mit", "hash": 4554861451716396000, "line_mean": 26.3333333333, "line_max": 81, "alpha_frac": 0.449784792, "autogenerated": false, "ratio": 3.7074468085106385, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.46572316005106384, "avg_score": null, "num_lines": null }
__author__ = 'cwhelan' # converts sam format to modil format # only works on data from one chromsome for now, so please split first! import sys current_read_id = "NA" read1_alignments = {} read2_alignments = {} def print_reads(read1_alignments, read2_alignments): for p1 in read1_alignments: for p2 in read2_alignments: if abs(p1-p2) <= 25000: a1 = read1_alignments[p1] a2 = read2_alignments[p2] if p1 < p2: left_read = a1 right_read = a2 else: left_read = a2 right_read = a1 left_fields = left_read.split("\t") right_fields = right_read.split("\t") left_flag = int(left_fields[1]) right_flag = int(right_fields[1]) if not left_flag & 0x10 and right_flag & 0x10: left_read_start = int(left_fields[3]) left_read_end = left_read_start + len(left_fields[9]) right_read_start = int(right_fields[3]) right_read_end = right_read_start + len(right_fields[9]) span_of_mapping = right_read_end - left_read_start print "\t".join(map(str, [left_read_end, right_read_start, left_fields[0], lib_insert_size - span_of_mapping, lib_insert_size, right_fields[0]])) lib_insert_size = int(sys.argv[1]) for line in sys.stdin: fields = line.split("\t") read_pair_name = fields[0] if read_pair_name != current_read_id: if current_read_id != "NA": print_reads(read1_alignments, read2_alignments) read1_alignments = {} read2_alignments = {} current_read_id = read_pair_name flag = int(fields[1]) pos = int(fields[3]) if flag & 0x40: read1_alignments[pos] = line else: read2_alignments[pos] = line print_reads(read1_alignments, read2_alignments)
{ "repo_name": "cwhelan/cloudbreak", "path": "src/main/scripts/nameSortedSamToMODIL.py", "copies": "1", "size": "1992", "license": "mit", "hash": -2189905813125169000, "line_mean": 36.5849056604, "line_max": 165, "alpha_frac": 0.546184739, "autogenerated": false, "ratio": 3.393526405451448, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9437362840591652, "avg_score": 0.0004696607719593218, "num_lines": 53 }
__author__ = 'CwT' from AbstractAnalysis import AbstractFlowAnalysis class GlobalAnalysis(AbstractFlowAnalysis): def __init__(self): self._globalvars = {} AbstractFlowAnalysis.__init__(self) @property def GlobalSet(self): return self._globalvars def flowTransfer(self, block): for instruction in block.insns: handler, var = instruction.isGlobal() if handler == 'register': if var not in self._globalvars.keys(): self._globalvars[var] = [] elif handler == 'check': for each in var: start = each.find('<$') name = each[start+2:-1] if name in self._globalvars.keys() and \ each not in self._globalvars[name]: self._globalvars[name].append(each) def IgnoreFunc(self): if self._graph._name == 'main': return True return False def printData(self): print(self._globalvars)
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/GlobalAnalysis.py", "copies": "1", "size": "1063", "license": "mit", "hash": 741299972539474000, "line_mean": 31.2121212121, "line_max": 63, "alpha_frac": 0.5305738476, "autogenerated": false, "ratio": 4.621739130434783, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0006887052341597796, "num_lines": 33 }
__author__ = 'CwT' from DexParse import * array = [0x32, 0xa, # if-eq v0, v0, 8 0x26, 0x3, 0x0, # fill-array-data v0, 3 0x0300, 0x2, 0x1, 0x0, 0x0 # fill-array-data-payload 0x0300, width, size, data[] ] array2 = [0x32, 0xd, # if-eq v0, v0, 8 0x1023, 0x06d7, # new-array v0, v1, typeid 0x26, 0x3, 0x0, # fill-array-data v0, 3 0x0300, 0x1, 0x2, 0x0, 0x0, # fill-array-data-payload 0x0300, width, size, data[] ] def addJunkopcode(dexfile, class_name, method_name): tmp_method = dexfile.getmethodItem(class_name, method_name) coderef = tmp_method["method"].coderef dexfile.verifyclass(tmp_method["defidx"]) index = dexfile.gettypeid("[B") if index < 0: index = dexfile.gettypeid("[C") if index < 0: pass print("find index: ", index) for i in range(0, len(array2)): coderef.insns.insert(i, array2[i]) num = coderef.insns_size * 2 # byte number coderef.insns[3] = index coderef.insns[9] = num & 0xffff coderef.insns[10] = (num >> 16) & 0xffff coderef.insns_size += 12 def modifyopcode(dexfile, class_name, method_name): tmp_method = dexfile.getmethodItem(class_name, method_name) coderef = tmp_method["method"].coderef coderef.insns[4] = 0x3a if __name__ == '__main__': dexfile = DexFile("classes.dex") # modifyopcode(dexfile, "Lcom/cc/test/MainActivity;", "onCreate") addJunkopcode(dexfile, "Lcom/cc/test/MainActivity;", "onCreateOptionsMenu") dexfile.copytofile("classescp.dex")
{ "repo_name": "CvvT/DexParse", "path": "AddJunk.py", "copies": "2", "size": "1572", "license": "apache-2.0", "hash": 1338007297107412500, "line_mean": 36.4285714286, "line_max": 92, "alpha_frac": 0.6164122137, "autogenerated": false, "ratio": 2.673469387755102, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.42898816014551017, "avg_score": null, "num_lines": null }
__author__ = 'CwT' from ForwardFlowAnalysis import ForwardFlowAnalysis class SSAliasAnalysis(ForwardFlowAnalysis): def __init__(self): self._AliasSetBefore = {} self._AliasSetAfter = {} ForwardFlowAnalysis.__init__(self) def flowTransfer(self, block): set = self.getAliasSet(block.Id) out = dict(set) vars = self._graph.getTypeMap() for instruction in block.insns: a, b = instruction.getAlias(vars) if a: self.eliminate(a, out) self.add(a, b, out) result = self.getAliasSet(block.Id, False) if len(result['Must']) == len(out['Must']) and \ len(result['May']) == len(out['May']): return False for entity in ['Must', 'May']: result[entity][:] = out[entity] return True def add(self, vara, varb, set): for entity in ['Must', 'May']: additems = [] for pair in set[entity]: if varb in pair: other = pair[0] if pair[0] != varb else pair[1] if (vara, other) not in set[entity] or (other, vara) not in set[entity]: additems.append((vara, other)) for pair in additems: set[entity].append(pair) if (vara, varb) not in set['Must'] or (varb, vara) not in set['Must']: set['Must'].append((vara, varb)) def eliminate(self, var, set): for entity in ['Must', 'May']: Godie = [] for pair in set[entity]: if var in pair: Godie.append(pair) for pair in Godie: set[entity].remove(pair) Godie[:] = [] # Merge: Intersection def mergeInto(self, sourceBlock, destBlock): sourceSet = self.getAliasSet(sourceBlock.Id) destSet = self.getAliasSet(destBlock.Id, False) for pair in destSet['Must']: reversed = (pair[1], pair[0]) if pair not in sourceSet['Must'] and reversed not in sourceSet['Must']: destSet['May'] += [pair] destSet['Must'].remove(pair) for pair in sourceSet['May']: reserved = (pair[1], pair[0]) if pair not in destSet and reserved not in destSet: destSet['May'].append(pair) def copy(self, sourceBlock, destBlock): # destFlow and sourceFLow must be set # copy to destSet = self.getAliasSet(destBlock.Id) srcSet = self.getAliasSet(sourceBlock.Id, False) for entity in ['Must', 'May']: destSet[entity][:] = srcSet[entity] def getAliasSet(self, Id, BeforeOrAfter=True): Target = self._AliasSetBefore if BeforeOrAfter else self._AliasSetAfter if Id not in Target.keys(): Target[Id] = dict({'Must': [], 'May': []}) return Target[Id] def printData(self): print "Function Alias:" for key, value in self._AliasSetAfter.items(): print(key) print(" Must Alias Pair") print " ", value['Must'] print(" May Alias Pair") print " ", value['May']
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/AliasAnalysis.py", "copies": "1", "size": "3222", "license": "mit", "hash": 1238581724217521700, "line_mean": 36.4651162791, "line_max": 92, "alpha_frac": 0.5301055245, "autogenerated": false, "ratio": 3.900726392251816, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.987005373519344, "avg_score": 0.01215563631167532, "num_lines": 86 }
__author__ = 'CwT' from ForwardFlowAnalysis import ForwardFlowAnalysis class SSAliasAnalysis(ForwardFlowAnalysis): def __init__(self): self._AliasSet = dict({'Must': [], 'May': []}) ForwardFlowAnalysis.__init__(self) @property def AliasSet(self): return self._AliasSet def flowTransfer(self, block): set = self._AliasSet vars = self._graph.getTypeMap() len_must, len_may = len(set['Must']), len(set['May']) for instruction in block.insns: a, b = instruction.getAlias(vars) if a: self.eliminate(a, set) self.add(a, b, set) if len(set['Must']) == len_must and \ len(set['May']) == len_may: return False return True def add(self, vara, varb, set): for entity in ['Must', 'May']: additems = [] for pair in set[entity]: if varb in pair: other = pair[0] if pair[0] != varb else pair[1] if (vara, other) not in set[entity] or (other, vara) not in set[entity]: additems.append((vara, other)) for pair in additems: set[entity].append(pair) if (vara, varb) not in set['Must'] or (varb, vara) not in set['Must']: set['Must'].append((vara, varb)) def eliminate(self, var, set): for entity in ['Must', 'May']: Godie = [] for pair in set[entity]: if var in pair: Godie.append(pair) for pair in Godie: set[entity].remove(pair) Godie[:] = [] # Merge: Intersection def mergeInto(self, sourceBlock, destBlock): pass def copy(self, sourceBlock, destBlock): pass def printData(self): print "Function Alias:" for key, value in self._AliasSet.items(): print(key) print(" Must Alias Pair") print " ", value print(" May Alias Pair") print " ", value
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/SSAliasAnalysis.py", "copies": "1", "size": "2103", "license": "mit", "hash": -6958093383351475000, "line_mean": 31.3538461538, "line_max": 92, "alpha_frac": 0.504041845, "autogenerated": false, "ratio": 3.938202247191011, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4942244092191011, "avg_score": null, "num_lines": null }
__author__ = 'CwT' from ForwardFlowAnalysis import ForwardFlowAnalysis class SSALiteralAnalysis(ForwardFlowAnalysis): def __init__(self): self._LiteralSet = {} ForwardFlowAnalysis.__init__(self) @property def varMaps(self): return self._LiteralSet def flowTransfer(self, block): set = self._LiteralSet changed = False for instruction in block.insns: name, types = instruction.getType(set) if name: if name not in set.keys() or \ type(types) != type(set[name]): changed = True set[name] = types return changed def getType(self, name): if name in self._LiteralSet: return self._LiteralSet[name] return None def mergeInto(self, sourceBlock, destBlock): pass def copy(self, sourceBlock, destBlock): pass def printData(self): print "Function Type:" for key, value in self._LiteralSet.items(): print key, value.type
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/SSALiteralAnalysis.py", "copies": "1", "size": "1081", "license": "mit", "hash": 6381821593756199000, "line_mean": 26.7179487179, "line_max": 55, "alpha_frac": 0.5689176688, "autogenerated": false, "ratio": 4.466942148760331, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.553585981756033, "avg_score": null, "num_lines": null }
__author__ = 'CwT' from Model import Node import Queue class InterProcTaintAnalysis(object): def __init__(self): self._graph = None self._FuncExist = {} self._FuncParams = {} self._SrcNodes = [] self._TargetNodes = [] self._Counter = 0 # Tainted Src Op and Var self.taintedSrcOp = ['Expr_ArrayDimFetch'] self.taintedSrcVar = ['<$_GET>', '<$_POST>', '<$_REQUEST>','<$_COOKIE>', '<$_SESSION>', '<$_FILES>', '<$_ENV>'] self.tgtToBeTaintedOp = ['Terminal_Echo','Expr_Print', 'Expr_FuncCall'] self.tgtToBeTaintedName = ['LITERAL(\'printf\')','LITERAL(\'print_r\')'] self.taintedSrcMysql = ['LITERAL(\'mysql_query\')', 'LITERAL(\'mysqli_query\')'] # InterProcAnalysis.__init__(self) def doAnalysis(self, SuperGraph): self._graph = SuperGraph self.work(self.MarkTaintedSource) self.work(self.MarkTaintTarget) for value in self._graph.subgraphs.values(): self.AnalysisFunction(value) self.SearchPath() def AnalysisFunction(self, func): workList = Queue.Queue() workList.put(func.Entry) while not workList.empty(): block = workList.get() changed = self.flowthrough(func, block) if changed: for item in block.successors: workList.put(item) def assignId(self): self._Counter += 2 return self._Counter def addNode(self, node): self._graph.addNode(node) def addEdge(self, srcFunc, srcVar, tgtFunc, tgtVar): return self._graph.addEdge(srcFunc, srcVar, tgtFunc, tgtVar) def addTargetNodeToEdge(self, srcFunc, srcVar, tgtNode, EdgeId): return self._graph.addTargetNodeToEdge(srcFunc, srcVar, tgtNode, EdgeId) def addSourceNodeToEdge(self, srcNode, tgtFunc, tgtVar, EdgeId): return self._graph.addSourceNodeToEdge(srcNode, tgtFunc, tgtVar, EdgeId) def printData(self): print str(self._graph) def work(self, handler): for value in self._graph.subgraphs.values(): for block in value.blocks.values(): handler(value, block) def MarkTaintedSource(self, function, block): for instruction in block.insns: line = instruction.Data op = line['op'] if 'var' in line.keys() and \ op in self.taintedSrcOp: for srcVar in self.taintedSrcVar: if srcVar in line['var']: # node = Node(line['var'], function.Name, Node.SOURCE) node = self.FindOrCreateNode(function.Name, line['var'], Node.SOURCE) self._SrcNodes.append(node) # add src node self.addNode(node) elif op == 'Expr_FuncCall': # and \ # line['name'] in self.taintedSrcMysql: self.addNode(Node(line['result'], function.Name, Node.RESULT)) elif op == 'Expr_Param': param = self.FindOrCreateNode(function.Name, line['result'], Node.PARAMETER) # param = Node(line['result'], function.Name, Node.PARAMETER) if function.Name not in self._FuncParams.keys(): self._FuncParams[function.Name] = [] self._FuncParams[function.Name].append(param) # add in order self.addNode(param) elif op == 'Terminal_Return': retValue = line['expr'] if retValue.startswith('Var#'): # node = Node(retValue, function.Name, Node.RETURN) node = self.FindOrCreateNode(function.Name, retValue, Node.RETURN) if function.Name not in self._FuncExist.keys(): self._FuncExist[function.Name] = [] self._FuncExist[function.Name].append(node) self.addNode(node) def MarkTaintTarget(self, function, block): for instruction in block.insns: line = instruction.Data op = line['op'] if op in self.tgtToBeTaintedOp: if 'name' in line.keys(): # Expr_FuncCall if line['name'] in self.tgtToBeTaintedName or \ line['name'] in self.taintedSrcMysql: for key in line.keys(): if line[key].startswith('args'): node = self.FindOrCreateNode(function.Name, line[key]) self._TargetNodes.append(node) # add target node else: # Terminal_Echo, Expr_Print node = self.FindOrCreateNode(function.Name, line['expr']) self._TargetNodes.append(node) # add target node def FindOrCreateNode(self, funcName, funcVar, type=Node.TARGET): return self._graph.FindOrCreateNode(funcName, funcVar, type) def flowthrough(self, function, block): changed = False for instruction in block.insns: # Mark tainted source line = instruction.Data # handle function call if self.handleFunctionCall(function, line): continue # Taint propagation srcSet, tgtSet = self.taintLine(line) for each in srcSet: for target in tgtSet: if self.addEdge(function.Name, each, function.Name, target): changed = True return changed def handleFunctionCall(self, function, line): if 'op' in line.keys() and line['op'] == 'Expr_FuncCall': funcName = line['name'][9:-2] + "():" func = self._graph.getFuncByName(funcName) if func: # function exists id = self.assignId() for key in line.keys(): if key.startswith('args'): # Note edge: main:var -> test():arg[0] number = int(key[5:-1]) self.addTargetNodeToEdge(function.Name, line[key], self._FuncParams[funcName][number], id) # TO-DO add backward edge if funcName in self._FuncExist.keys(): for existNode in self._FuncExist[funcName]: # positive number for caller, and negative number for callee self.addSourceNodeToEdge(existNode, function.Name, line['result'], -id) return True return False def taintLine(self, line): srcSet = set() tgtSet = set() if line.has_key('op'): op = line['op'] if op == 'Expr_ArrayDimFetch': srcSet.add(line['var']) srcSet.add(line['dim']) tgtSet.add(line['result']) elif op == 'Expr_Assign': srcSet.add(line['expr']) tgtSet.add(line['var']) elif op in ['Expr_BinaryOp_Smaller', 'Expr_BinaryOp_Equal', 'Expr_BinaryOp_Greater', 'Expr_BinaryOp_SmallerOrEqual', 'Expr_BinaryOp_GreaterOrEqual']: pass elif op in ['Expr_BinaryOp_BitwiseAnd', 'Expr_BinaryOp_BitwiseOr', 'Expr_BinaryOp_Xor']: pass elif op.startswith('Expr_BinaryOp'): srcSet.add(line['left']) srcSet.add(line['right']) tgtSet.add(line['result']) elif op.startswith('Expr_Cast'): srcSet.add(line['expr']) tgtSet.add(line['result']) elif op == 'Expr_FuncCall': for key in line.keys(): if key.startswith('args'): srcSet.add(line[key]) tgtSet.add(line['result']) elif op == 'Phi': for key in line.keys(): if key.startswith('chose'): srcSet.add(line[key]) tgtSet.add(line['result']) elif op in ['Stmt_Jump', 'Stmt_JumpIf']: pass elif op.startswith('Terminal'): pass elif op == 'Expr_Print': srcSet.add(line['expr']) tgtSet.add(line['result']) else: if line.has_key('result'): for key in line.keys(): if key != 'result' and key != 'op' and line[key].startswith('Var#'): srcSet.add(line[key]) tgtSet.add(line['result']) if None in srcSet: srcSet.remove(None) if None in tgtSet: tgtSet.remove(None) return srcSet, tgtSet # Calculate reachability by depth-first search def SearchPath(self): for src in self._SrcNodes: targets = [] self.DFSearch(src, [], [], targets) if len(targets) > 0: print src.Name for each in targets: print "\t", each.Name def DFSearch(self, node, CallString, visited, targets): if (node.Type & Node.TARGET) and len(CallString) == 0: targets.append(node) return True for successor in node.successors: if successor in visited: continue visited.append(successor) subnode = successor[0] edgeId = successor[1] if edgeId == 0: # intra-procedure self.DFSearch(subnode, CallString, visited, targets) elif edgeId > 0: # from caller to callee # call into CallString.append(edgeId) self.DFSearch(subnode, CallString, visited, targets) elif edgeId < 0: # return from callee to caller if CallString[-1] + edgeId == 0: # legal del CallString[-1] self.DFSearch(subnode, CallString, visited, targets) visited.remove(successor)
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/cfl/InterProcTaintAnalysis.py", "copies": "1", "size": "9996", "license": "mit", "hash": -9168881039073787000, "line_mean": 42.0862068966, "line_max": 161, "alpha_frac": 0.5269107643, "autogenerated": false, "ratio": 4.166736140058358, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5193646904358358, "avg_score": null, "num_lines": null }
__author__ = 'CwT' from type.Type import Literal, Normal from type.Type import Array as TArray class Graph(object): def __init__(self, data): self._func = {} for key, value in data.items(): self._func[key] = Function(key, value) @property def Entry(self): return self._func['main'] @property def functions(self): return self._func def Method(self, name): return self._func[name] def register(self, name, visitor): for func in self._func.values(): func.register(name, visitor) def doAnalyse(self): for func in self._func.values(): func.doAnalyse() def getFuncByName(self, name): if name in self._func.keys(): return self._func[name] else: return None class Function(object): def __init__(self, funcName, data): self._name = funcName self._blocks = {} for key, value in data.items(): self._blocks[key] = BasicBlock(key, value) self.createGraph() self._visitors = {} self._vName = [] @property def Name(self): return self._name @property def blocks(self): return self._blocks @property def Entry(self): return self._blocks['Block#1'] def createGraph(self): for key, value in self._blocks.items(): for each in value.parents: block = self._blocks[each] value.predecessors.append(block) block.successors.append(value) def register(self, name, visitor): self._vName.append(name) self._visitors[name] = visitor() def doAnalyse(self): for name in self._vName: visitor = self._visitors[name] visitor.doAnalysis(self) visitor.printData() def getTypeMap(self): if 'Literal' in self._vName: return self._visitors['Literal'].varMaps return None def getAliasPair(self): if 'Alias' in self._vName: return self._visitors['Alias'].AliasSet class BasicBlock(object): def __init__(self, key, cont): self._id = key self._predecessors = [] self._successors = [] self._code = [] self._parents = [] if cont.has_key('Parent'): self._parents = cont['Parent'] for item in cont['code']: self._code.append(Instruction(item)) @property def parents(self): return self._parents @property def predecessors(self): return self._predecessors @property def successors(self): return self._successors @property def insns(self): return self._code @property def Id(self): return self._id class Instruction(object): def __init__(self, data): self._data = data @property def Data(self): return self._data def isLiteral(self, expr, varMaps): if expr.startswith('LITERAL'): return True elif expr.startswith('Var#') and expr in varMaps.keys(): return isinstance(varMaps[expr], Literal) else: return False def getType(self, varMaps): op = self._data['op'] if op == 'Expr_Assign': expr = self._data['expr'] if expr.startswith('LITERAL'): return self._data['var'], Literal(self._data['expr']) elif expr.startswith('Var#') and expr in varMaps.keys(): return self._data['var'], varMaps[expr] elif op.startswith('Expr_Binary'): if self.isLiteral(self._data['left'], varMaps) or \ self.isLiteral(self._data['right'], varMaps): return self._data['result'], Literal() else: return self._data['result'], Normal() elif op.startswith('Expr_ArrayDim'): return self._data['var'], TArray() elif op == 'Expr_Array': return self._data['result'], TArray() return None, None def getAlias(self, varMaps): op = self._data['op'] if op == 'Expr_AssignRef': return self._data['var'], self._data['expr'] elif op == 'Expr_Assign': expr = self._data['expr'] if expr in varMaps.keys() and isinstance(varMaps[expr], TArray): return self._data['var'], expr return None, None def isGlobal(self): op = self._data['op'] if op == 'Terminal_GlobalVar': return 'register', self._data['var'][9:-2] else: return 'check', filter(lambda x: x.find('<$') != -1, self.getVars()) def findVarbyName(self, name): return filter(lambda x: x.find('<$') != -1 and x[x.find('<$')+2:-1] == name, self.getVars()) def getVars(self): ret = [] if 'var' in self._data.keys(): ret.append(self._data['var']) if 'result' in self._data.keys(): ret.append(self._data['result']) if 'expr' in self._data.keys(): ret.append(self._data['expr']) return ret
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/Graph.py", "copies": "1", "size": "5159", "license": "mit", "hash": 7992706562260699000, "line_mean": 27.5027624309, "line_max": 100, "alpha_frac": 0.5406086451, "autogenerated": false, "ratio": 4.021044427123928, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5061653072223928, "avg_score": null, "num_lines": null }
__author__ = 'CwT' import struct import Util class OATHdr: def __init__(self): self.offset = 0 self.magic = [] self.version = [] self.checksum = 0 self.dexfileCount = 0 self.size = 0 self.keyvalueSize = 0 def readfd(self, file, offset): self.offset = offset file.seek(offset) value = struct.unpack("4s4sIIIIIIIIIIIIIIIIIII", file.read(21*4)) self.magic = value[0] self.version = value[1] self.dexfileCount = value[5] self.keyvalueSize = value[20] def getHeaderSize(self): return self.keyvalueSize + 21 * 4 def getDexListStart(self): return self.offset + self.getHeaderSize() class DexMeta: def __init__(self): self.name = None self.dexOffset = 0 self.classCount = 0 def readfd(self, file, oatFile): filenameLen = struct.unpack("I", file.read(4))[0] self.name = struct.unpack(str(filenameLen)+"s", file.read(filenameLen))[0] value = struct.unpack("II", file.read(4*2)) self.dexOffset = value[1] file.seek(oatFile.offset + self.dexOffset + 96) self.classCount = struct.unpack("I", file.read(4))[0] def getMetaSize(self): return 4 + len(self.name) + 4*2 + 4*self.classCount class OATfile: def __init__(self): self.offset = 0 self.oatHdr = OATHdr() def readfd(self, file, offset): self.offset = offset self.oatHdr.readfd(file, offset) def getDexFiles(self, file): offset = self.oatHdr.getDexListStart() dexFiles = [] for i in range(self.oatHdr.dexfileCount): file.seek(offset) dexMeta = DexMeta() dexMeta.readfd(file, self) offset += dexMeta.getMetaSize() dexFiles.append(dexMeta) # print hex(offset) return dexFiles
{ "repo_name": "CvvT/ElfParse", "path": "OatParse.py", "copies": "1", "size": "1908", "license": "apache-2.0", "hash": -2010881188756968000, "line_mean": 27.4776119403, "line_max": 82, "alpha_frac": 0.5780922432, "autogenerated": false, "ratio": 3.359154929577465, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9426042439130724, "avg_score": 0.002240946729348096, "num_lines": 67 }
__author__ = 'CwT' import struct def elf_hash(funcname): h = 0 for x in funcname: h = (h << 4) + ord(x) g = h & 0xf0000000 h ^= g h ^= g >> 24 print() return h class Ehdr: def __init__(self): self.ident = [] self.type = 0 # 0: unknow; 1: relocation 2: execute 3: share self.machine = 0 self.version = 0 # version: 1 self.entry = 0 self.phoff = 0 # program header table offset self.shoff = 0 # section header table offset self.flags = 0 self.ehsize = 0 # size of elf header self.phentsize = 0 # size of each program table entry self.phnum = 0 # number of program table entries self.shentsize = 0 # size of each section table entry self.shnum = 0 # number of section table entries self.shstrndx = 0 # section header string index self.size = 52 # bytes self.shtable = None self.phtable = None self.dyntable = None def readfromfd(self, file): file.seek(0) value = struct.unpack("16sHHIIIIIHHHHHH", file.read(self.size)) self.ident = value[0] self.type = value[1] self.machine = value[2] self.version = value[3] self.entry = value[4] self.phoff = value[5] self.shoff = value[6] self.flags = value[7] self.ehsize = value[8] self.phentsize = value[9] self.phnum = value[10] self.shentsize = value[11] self.shnum = value[12] self.shstrndx = value[13] def init_table(self): self.shtable = ShTable() self.shtable.readfromfd(file, self.shoff, self.shnum, self.shstrndx) self.phtable = PhTable() self.phtable.readfromfd(file, self.phoff, self.phnum) tmp = self.phtable.getSegment("PT_DYNAMIC") self.dyntable = DynTable(tmp.offset, tmp.filesz) self.dyntable.readfromfd(file) def printf(self): for x in self.ident: print hex(ord(x)), print '' print(self.type) print(self.machine) print(self.version) print(self.entry) print(self.phoff) print(self.shoff) print(self.flags) print(self.ehsize) print(self.phentsize) print(self.phnum) print(self.shentsize) print(self.shnum) print(self.shstrndx) class Shder: def __init__(self): self.name = 0 # index of section header string table section self.type = 0 self.flags = 0 self.addr = 0 self.offset = 0 # offset from the beginning of the file self.size = 0 self.link = 0 self.info = 0 self.addralign = 0 self.entsize = 0 self.size = 40 def readfromfd(self, file): value = struct.unpack("IIIIIIIIII", file.read(self.size)) self.name = value[0] self.type = value[1] self.flags = value[2] self.addr = value[3] self.offset = value[4] self.size = value[5] self.link = value[6] self.info = value[7] self.addralign = value[8] self.entsize = value[9] class ShTable: def __init__(self): self.shtable = [] self.off = 0 self.num = 0 self.shstrtable = [] self.strtab = [] self.symtab = [] self.hashtable = None def getstrtablefromfd(self, file, str_index): sh = self.shtable[str_index] file.seek(sh.offset) for i in range(0, sh.size): self.shstrtable.append(struct.unpack("c", file.read(1))[0]) def readfromfd(self, file, shoff, shnum, str_index): file.seek(shoff) self.off = shoff self.num = shnum for i in range(0, shnum): shder = Shder() shder.readfromfd(file) self.shtable.append(shder) self.getstrtablefromfd(file, str_index) self.getSymstr(file) self.getSymtab(file) self.gethash(file) def getshname(self, index): start = index while 1: if self.shstrtable[index] == '\0': break index += 1 return "".join(self.shstrtable[start:index]) def findsection(self, secName): for i in range(0, self.num): if self.getshname(self.shtable[i].name) == secName: return self.shtable[i] return None def getSymstr(self, file): tmp = self.findsection(".dynstr") if tmp is not None: file.seek(tmp.offset) for i in range(0, tmp.size): self.strtab.append(struct.unpack("c", file.read(1))[0]) else: print("error in getSymstr") def getSymtab(self, file): tmp = self.findsection(".dynsym") if tmp is not None: len = tmp.size / tmp.entsize file.seek(tmp.offset) for i in range(0, len): sym = Symbol() sym.readfromfd(file) self.symtab.append(sym) else: print("error in get symtab") def gethash(self, file): tmp = self.findsection(".hash") if tmp is not None: self.hashtable = Hash() self.hashtable.readfromfd(file, tmp.offset) else: print("error in gethash") def getnamebyindex(self, index): start = index while 1: if self.strtab[index] == '\0': break index += 1 return "".join(self.strtab[start:index]) def getfuncinfo(self, funcname): index = elf_hash(funcname) % self.hashtable.nbucket index = self.hashtable.getbucket(index) if index >= len(self.symtab): print("error, get symbol table first") exit() str_index = self.symtab[index].name if self.getnamebyindex(str_index) != funcname: while 1: index = self.hashtable.getchain(index) if index == 0: print("did not find the func %s" %funcname) break str_index = self.symtab[index].name if self.getnamebyindex(str_index) == funcname: break if index == 0: return None return self.symtab[index] def printf(self): for i in range(0, self.num): print(self.getshname(self.shtable[i].name)) def printsym(self): for i in range(0, len(self.symtab)): print(self.getnamebyindex(self.symtab[i].name)) class Phdr: TYPE = {'PT_NULL': 0, 'PT_LOAD': 1, 'PT_DYNAMIC': 2, 'PT_INTERP': 3, 'PT_NOTE': 4, 'PT_SHLIB': 5, 'PT_PHDR': 6, } def __init__(self): self.type = 0 self.offset = 0 self.vaddr = 0 # virtual address self.paddr = 0 # physical address self.filesz = 0 # size in file self.memsz = 0 # size in memory self.flags = 0 self.align = 0 self.size = 8 * 4 def readfromfd(self, file): value = struct.unpack("IIIIIIII", file.read(self.size)) self.type = value[0] self.offset = value[1] self.vaddr = value[2] self.paddr = value[3] self.filesz = value[4] self.memsz = value[5] self.flags = value[6] self.align = value[7] class PhTable: def __init__(self): self.phtable = [] self.num = 0 self.off = 0 def readfromfd(self, file, offset, phnum): file.seek(offset) self.off = offset self.num = phnum for i in range(0, self.num): phdr = Phdr() phdr.readfromfd(file) self.phtable.append(phdr) def getSegment(self, segName): if segName not in Phdr.TYPE.keys(): print("unknow segment name") return None for i in range(0, self.num): if self.phtable[i].type == Phdr.TYPE[segName]: print("find %s" % (segName)) return self.phtable[i] class Symbol: def __init__(self): self.name = 0 self.value = 0 self.size = 0 self.info = 0 self.other = 0 self.shndx = 0 self.size = 4 * 3 + 2 * 1 + 2 def readfromfd(self, file): value = struct.unpack("IIIBBH", file.read(self.size)) self.name = value[0] self.value = value[1] self.size = value[2] self.info = value[3] self.other = value[4] self.shndx = value[5] class Dynamic: TAG = {0: 'DT_NULL', 1: 'DT_NEEDED', 2: 'DT_PLTRELSZ', 3: 'DT_PLTGOT', 4: 'DT_HASH', 5: 'DT_STRTAB', 6: 'DT_SYMTAB', 7: 'DT_RELA', 8: 'DT_REALSZ', 9: 'DT_RELAENT', 10: 'DT_STRSZ', 11: 'DT_SYMENT', 12: 'DT_INIT', 13: 'DT_FINI', 14: 'DT_SONAME', 15: 'DT_RPATH', 16: 'DT_SYMBOLIC', 17:'DT_REL', 18: 'DT_RELSZ', 19: 'DT_RELENT', 20: 'DT_PLTREL', 21: 'DT_DEBUG', 22: 'DT_TEXTREL', 23: 'DT_JMPREL', 25: 'DT_INIT_ARRAY', 26: 'DT_FINI_ARRAY', 27: 'DT_INIT_ARRAYSZ', 28: 'DT_FINI_ARRAYSZ', } def __init__(self): self.tag = 0 self.union = 0 self.size = 8 def readfromfd(self, file): value = struct.unpack("iI", file.read(self.size)) self.tag = value[0] self.union = value[1] def printf(self): if self.tag in Dynamic.TAG.keys(): print("tag is: ", Dynamic.TAG[self.tag]) else: print("tag is: ", self.tag) print("union is: ", self.union) print("size is: ", self.size) class DynTable: TAG = {'DT_NULL': 0, 'DT_NEEDED': 1, 'DT_PLTRELSZ': 2, 'DT_PLTGOT': 3, 'DT_HASH': 4, 'DT_STRTAB': 5, 'DT_SYMTAB': 6, 'DT_RELA': 7, 'DT_REALSZ': 8, 'DT_RELAENT': 9, 'DT_STRSZ': 10, 'DT_SYMENT': 11, 'DT_INIT': 12, 'DT_FINI': 13, 'DT_SONAME': 14, 'DT_RPATH': 15, 'DT_SYMBOLIC': 16, 'DT_REL': 17, 'DT_RELSZ': 18, 'DT_RELENT': 19, 'DT_PLTREL': 20, 'DT_DEBUG': 21, 'DT_TEXTREL': 22, 'DT_JMPREL': 23, 'DT_INIT_ARRAY': 25, 'DT_FINI_ARRAY': 26, 'DT_INIT_ARRAYSZ': 27, ' DT_FINI_ARRAYSZ': 28, } def __init__(self, offset, size): self.off = offset self.size = size self.num = size / 8 self.dyn = [] self.strtab = [] self.symtab_off = 0 self.symtab = [] self.hashtable = None def readfromfd(self, file): file.seek(self.off) for i in range(0, self.num): dyn = Dynamic() dyn.readfromfd(file) self.dyn.append(dyn) self.getStrtable(file) self.getHashtable(file) self.symtab_off = self.getsecBytag("DT_SYMTAB").union # self.getSymboltable(file) def getsecBytag(self, tag): for i in range(0, self.num): if self.dyn[i].tag == DynTable.TAG[tag]: return self.dyn[i] print("did not find %s" % tag) return None def getStrtable(self, file): offset = self.getsecBytag("DT_STRTAB").union size = self.getsecBytag("DT_STRSZ").union file.seek(offset) for i in range(0, size): self.strtab.append(struct.unpack("c", file.read(1))[0]) def getHashtable(self, file): offset = self.getsecBytag("DT_HASH").union self.hashtable = Hash() self.hashtable.readfromfd(file, offset) def getSymboltable(self, file, filesz): offset = self.getsecBytag("DT_SYMTAB").union num = filesz / self.getsecBytag("DT_SYMENT").union # size of each symbol entry file.seek(offset) for i in range(0, num): symbol = Symbol() symbol.readfromfd(file) self.symtab.append(symbol) def getnamebyindex(self, index): start = index while 1: if self.strtab[index] == '\0': break index += 1 return "".join(self.strtab[start:index]) def getfuncinfo(self, funcname, file): index = elf_hash(funcname) % self.hashtable.nbucket index = self.hashtable.getbucket(index) file.seek(self.symtab_off + index * 16) # str_index = self.symtab[index].name str_index = struct.unpack("I", file.read(4))[0] if self.getnamebyindex(str_index) != funcname: while 1: index = self.hashtable.getchain(index) print(index, "index") if index == 0: print("did not find the func %s" %funcname) break file.seek(self.symtab_off + index * 16) str_index = struct.unpack("I", file.read(4))[0] if self.getnamebyindex(str_index) == funcname: break if index == 0: return None func = Symbol() file.seek(self.symtab_off + index * 16) func.readfromfd(file) return func def printSymbol(self): for i in range(0, len(self.symtab)): print(self.getnamebyindex(self.symtab[i].name)) class Hash: def __init__(self): self.nbucket = 0 self.nchain = 0 self.bucket = [] self.chain = [] def readfromfd(self, file, offset): file.seek(offset) value = struct.unpack("II", file.read(8)) self.nbucket = value[0] self.nchain = value[1] for i in range(0, self.nbucket): self.bucket.append(struct.unpack("I", file.read(4))[0]) for i in range(0, self.nchain): self.chain.append(struct.unpack("I", file.read(4))[0]) def getbucket(self, index): if index >= self.nbucket: print("index %d out of range %d" % (index, self.nbucket)) return -1 return self.bucket[index] def getchain(self, index): return self.chain[index] if __name__ == '__main__': file = open("libtest.so", "rb") elfheader = Ehdr() elfheader.readfromfd(file) elfheader.printf() file.close()
{ "repo_name": "List3nt0/ElfParse", "path": "ElfParse.py", "copies": "3", "size": "14115", "license": "apache-2.0", "hash": -185202154555888580, "line_mean": 31.299771167, "line_max": 101, "alpha_frac": 0.5307828551, "autogenerated": false, "ratio": 3.3345145287030475, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5365297383803048, "avg_score": null, "num_lines": null }
__author__ = 'CwT' class CFLGraph(object): def __init__(self, SuperGraph): self._supGraph = SuperGraph self._nodes = {} @property def Entry(self): return self._supGraph.Entry @property def subgraphs(self): return self._supGraph.functions @property def AllNodes(self): return self._nodes def getFuncByName(self, name): return self._supGraph.getFuncByName(name) def addNode(self, node): # Add node if not exist # Update type if exist if node.Name not in self._nodes.keys(): self._nodes[node.Name] = node self._nodes[node.Name].update(node.Type) def addEdge(self, srcFunc, srcVar, tgtFunc, tgtVar): ''' :param srcFunc: :param srcVar: :param tgtFunc: :param tgtVar: :return: If add an edge successfully ''' source = srcFunc + ":" + srcVar target = tgtFunc + ":" + tgtVar ret = False # source must be added before if source in self._nodes.keys(): if target not in self._nodes.keys(): # target could be added dynamically self.addNode(Node(tgtVar, tgtFunc)) ret = self._nodes[source].addEdge(self._nodes[target]) return ret def addTargetNodeToEdge(self, srcFunc, srcVar, tgtNode, EdgeId): source = srcFunc + ":" + srcVar ret = False if source in self._nodes.keys(): ret = self._nodes[source].addEdge(tgtNode, EdgeId) return ret def addSourceNodeToEdge(self, srcNode, tgtFunc, tgtVar, EdgeId): target = tgtFunc + ":" + tgtVar if target not in self._nodes.keys(): self.addNode(Node(tgtVar, tgtFunc)) return srcNode.addEdge(self._nodes[target], EdgeId) def isTainted(self, funcName, varName): if (funcName + ":" + varName) in self._nodes.keys(): return True return False def FindOrCreateNode(self, funcName, funcVar, type): target = funcName + ":" + funcVar if target not in self._nodes.keys(): self._nodes[target] = Node(funcVar, funcName, type) else: self._nodes[target].update(type) return self._nodes[target] def __str__(self): ret = '' for key, value in self._nodes.items(): ret += (key + ":\n") ret += str(value) return ret class Node(object): SOURCE = 1 TARGET = 2 MIDDLE = 4 PARAMETER = 8 RETURN = 16 RESULT = 32 def __init__(self, name, functionname, type=MIDDLE): ''' :param name: :param functionname: :param type: including Middle, Source, Target :return: ''' self._varName = name self._func = functionname self._edges = [] # point to other nodes self._type = type @property def Name(self): return self._func + ":" + self._varName @property def FuncName(self): return self._func @property def VarName(self): return self._varName @property def Type(self): return self._type @property def successors(self): return self._edges def addEdge(self, node, EdgeId=0): tuple = (node, EdgeId) if tuple not in self._edges: self._edges.append(tuple) return True return False def update(self, type): self._type |= type def __str__(self): ret = '\t' for each in self._edges: ret += (each[0].Name + ", ") return ret + "\n"
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/cfl/Model.py", "copies": "1", "size": "3639", "license": "mit", "hash": 5028505321046524000, "line_mean": 25.3695652174, "line_max": 87, "alpha_frac": 0.5504259412, "autogenerated": false, "ratio": 3.95114006514658, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9961601140294757, "avg_score": 0.007992973210364514, "num_lines": 138 }
__author__ = 'CwT' from DexParse import * if __name__ == '__main__': dexfile = DexFile("classes.dex") # @Params: classname, acces_flag(public, private...), parent_class_name, Filename classitem = dexfile.addclass("Lcom/cc/test/DexParse;", Access_Flag["public"], "Ljava/lang/Object;", "DexParse.java") classdata = ClassdataItem(None, 2) dexfile.addclassData(classdata) classdata.addstaticfield(dexfile.addfield(classitem.classIdx, "int", "IntFromCwt"), Access_Flag['public'] | Access_Flag["static"]) classdata.addinstancefield(dexfile.addfield(classitem.classIdx, "boolean", "BoolFromCwt"), Access_Flag["private"]) param_list = [] # param_list.append("[Lcom/lang/String;") # param_list.append("boolean") # prepare for parameter list code = CodeItem(None, 2) # @param register_size, in_size, out_size, tries_size, debug_off, # insns_size, insns_list, debugref, tries_list, handler insns = [] tries = [] insns.append(0xe) # return void debug = DebugInfo(None, 2) names_list = [] debug_list = [0] debug.adddebugitem(0, 0, names_list, debug_list) dexfile.adddebug(debug) code.addcode(1, 1, 0, 0, 0, len(insns), insns, debug, tries, None) dexfile.addcode(code) method_idx = dexfile.addmethod(classitem.classIdx, param_list, "void", "main") classdata.adddirectmethod(method_idx, Access_Flag['public'] | Access_Flag['static'], code) classdata.commit() classitem.addclassdefref(None, None, classdata, None) tmp_method = dexfile.getmethodItem("Lcom/cc/test/MainActivity;", "onCreate") coderef = tmp_method['method'].coderef coderef.insns.insert(0, 0) coderef.insns.insert(0, method_idx) coderef.insns.insert(0, 0x71) coderef.insns_size += 3 dexfile.copytofile("classescp.dex")
{ "repo_name": "peterdocter/DexParse", "path": "AddClass.py", "copies": "2", "size": "1874", "license": "apache-2.0", "hash": -1306748393954928400, "line_mean": 42.5813953488, "line_max": 94, "alpha_frac": 0.6531483458, "autogenerated": false, "ratio": 3.1077943615257047, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4760942707325705, "avg_score": null, "num_lines": null }
__author__ = 'CwT' from ElfParse import * if __name__ == '__main__': file = open("libbaiduprotect.so", 'r+b') file.seek(0) elfheader = Ehdr() elfheader.readfromfd(file) elfheader.init_table() addr = elfheader.shtable.findsection(".text").offset if addr == -1: print("find section .text failed") exit(0) # symbol = elfheader.shtable.getfuncinfo("native_hello") # symbol = elfheader.dyntable.getfuncinfo("native_hello", file) # print("func addr :", symbol.value - 1) # i don't know why must minus one, it's the beginning of the text # file.seek(symbol.value - 1) # size = symbol.size # content = [] # for i in range(0, size-1): # content.append(~struct.unpack("B", file.read(1))[0] & 0xff) # file.seek(symbol.value - 1) # # print(struct.unpack("B", file.read(1))[0]) # for i in range(0, size-1): # # print(struct.pack("B", content[i]), ) # file.write(struct.pack("B", content[i])) # elfheader.shtable.printsym() elfheader.printf() file.close()
{ "repo_name": "List3nt0/ElfParse", "path": "Encrypt.py", "copies": "3", "size": "1065", "license": "apache-2.0", "hash": -5129427311128843000, "line_mean": 34.5, "line_max": 111, "alpha_frac": 0.5962441315, "autogenerated": false, "ratio": 2.974860335195531, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5071104466695531, "avg_score": null, "num_lines": null }
__author__ = 'CwT' from idaapi import * from idautils import * from idc import * import re def getDword(addr): return Dword(addr) def getByte(addr): return Byte(addr) def getWord(addr): return Word(addr) class DexorJar: def __init__(self): self.pfileName = 0 self.isDex = 0 self.okaytoFree = 0 # do not care self.pRawDexFile = 0 self.pJarFile = 0 self.pDexMemory = 0 # do not care def dump(self, addr): self.pfileName = getDword(addr) self.isDex = getByte(addr + 4) self.pRawDexFile = getDword(addr + 8) self.pJarFile = getDword(addr + 12) def printf(self): str = "" baseaddr = self.pfileName onebyte = getByte(baseaddr) while onebyte != 0: str += chr(onebyte) baseaddr += 1 onebyte = getByte(baseaddr) print("filename is:", str) if self.isDex > 0: pass # print("it's a dex file, addr: ", hex(self.pRawDexFile)) else: # print("it's a jar file, addr: ", hex(self.pJarFile)) jarfile = JarFile() jarfile.dump(cookie.pJarFile) jarfile.printf() dvmaddr = jarfile.pDvmDex class JarFile: def __init__(self): self.archive = None # do not care self.pcacheFileName = 0 self.pDvmDex = 0 def dump(self, addr): self.pcacheFileName = getDword(addr + 36) self.pDvmDex = getDword(addr + 40) def printf(self): str = "" baseaddr = self.pcacheFileName one = getByte(baseaddr) while one != 0: str += chr(one) baseaddr += 1 one = getByte(baseaddr) print("cache file name is : ", str) # print("DvmDex addr is :", hex(self.pDvmDex)) class RawDexFile: def __init__(self): self.pcacheFileName = 0 self.pDvmDex = 0 def dump(self, addr): self.pcacheFileName = getDword(addr) self.pDvmDex = getDword(addr+4) def printf(self): str = "" baseaddr = self.pcacheFileName if baseaddr == 0: print "cache file name is null" return one = getByte(baseaddr) while one != 0: str += chr(one) baseaddr += 1 one = getByte(baseaddr) print("cache file name is : ", str) # print("DvmDex addr is :", hex(self.pDvmDex)) class DvmDex: def __init__(self): self.pDexFile = 0 self.pHeader = 0 # it is a clone of dex file # just for now def dump(self, addr): self.pDexFile = getDword(addr) self.pHeader = getDword(addr + 4) def printf(self): # i wanna see the diff between the pDexFile.dexfile and pheader print("dexfile addr is: ", hex(self.pDexFile)) # print("header addr is: ", hex(self.pHeader)) experiment = True if experiment: gDvm = 0 offset = 0 for i in range(10): insn = GetDisasm(here()+i*2) match = re.search(r'PC[\s]*;[\s]*dword_(?P<addr>[0-9A-Z]+)', insn) if match is not None: address = "0x" + match.group('addr') gDvm = Dword(int(address, 16)) match = re.search(r'[R[\d]+,#(?P<off>0x[\dA-F]+)]', insn) if match is not None: offset = match.group('off') offset = int(offset, 16) target = gDvm + offset else: target = int(0x40DDB654) # find it in dvminternalnativeshutdown--->dvmHashTableFree print "target:", hex(target) userDex = getDword(target) size = getDword(userDex) entry = getDword(userDex+12) print "Size:", size print "Entry:", hex(entry) for i in range(size): hash = getDword(entry+8*i) item = getDword(entry+8*i+4) if hash == item and hash != 0: cookie = DexorJar() cookie.dump(hash) cookie.printf() if cookie.isDex == 0: jarfile = JarFile() jarfile.dump(cookie.pJarFile) jarfile.printf() dvmaddr = jarfile.pDvmDex else: rawDex = RawDexFile() rawDex.dump(cookie.pRawDexFile) rawDex.printf() dvmaddr = rawDex.pDvmDex dvmDex = DvmDex() dvmDex.dump(dvmaddr) dvmDex.printf()
{ "repo_name": "CvvT/dumpDex", "path": "findcookie.py", "copies": "1", "size": "4305", "license": "apache-2.0", "hash": 5947876427478314000, "line_mean": 27.3223684211, "line_max": 88, "alpha_frac": 0.5479674797, "autogenerated": false, "ratio": 3.2055100521221145, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.42534775318221146, "avg_score": null, "num_lines": null }
__author__ = 'CwT' from idaapi import * from idautils import * from idc import * import struct import os import hashlib Access_Flag = {'public': 1, 'private': 2, 'protected': 4, 'static': 8, 'final': 0x10, 'synchronized': 0x20, 'volatile': 0x40, 'bridge': 0x40, 'transient': 0x80, 'varargs': 0x80, 'native': 0x100, 'interface': 0x200, 'abstract': 0x400, 'strictfp': 0x800, 'synthetic': 0x1000, 'annotation': 0x2000, 'enum': 0x4000, 'constructor': 0x10000, 'declared_synchronized': 0x20000} TypeDescriptor = {'void': 'V', 'boolean': 'Z', 'byte': 'B', 'short': 'S', 'char': 'C', 'int': 'I', 'long': 'J', 'float': 'F', 'double': 'D', 'boolean[]': '[Z', 'byte[]': '[B', 'short[]': '[S', 'char[]': '[C', 'int[]': 'I', 'long[]': '[J', 'float[]': '[F', 'double[]': 'D'} ShortyDescriptor = {'void': 'V', 'boolean': 'Z', 'byte': 'B', 'short': 'S', 'char': 'C', 'int': 'I', 'long': 'J', 'float': 'F', 'double': 'D'} ACSII = {'1': 1, '2': 2, '3': 3, '4': 4, '5': 5, '6': 6, '7': 7, '8': 8, '9': 9, '0': 0, 'a': 10, 'b': 11, 'c': 12, 'd': 13, 'e': 14, 'f': 15} def checksum(f, len): a = 1 b = 0 f.seek(12) print("file size is :", len) for i in range(12, len): onebyte = struct.unpack("B", f.read(1))[0] a = (a + onebyte) % 65521 b = (b + a) % 65521 return b << 16 | a def get_file_sha1(f): f.seek(32) # skip magic, checksum, sha sha = hashlib.sha1() while True: data = f.read(1024) if not data: break sha.update(data) return sha.hexdigest() def getDword(addr): return Dword(addr) def getByte(addr): return Byte(addr) def getWord(addr): return Word(addr) def rightshift(value, n): mask = 0x80000000 check = value & mask if check != mask: return value >> n else: submask = mask for loop in range(0, n): submask = (submask | (mask >> loop)) strdata = struct.pack("I", submask | (value >> n)) ret = struct.unpack("i", strdata)[0] return ret def readunsignedleb128(addr): res = getByte(addr) len = 1 if res > 0x7f: cur = getByte(addr + 1) res = (res & 0x7f) | ((cur & 0x7f) << 7) len = 2 if cur > 0x7f: cur = getByte(addr + 2) res |= (cur & 0x7f) << 14 len = 3 if cur > 0x7f: cur = getByte(addr + 3) res |= (cur & 0x7f) << 21 len = 4 if cur > 0x7f: cur = getByte(addr + 4) res |= cur << 28 len = 5 return res, len def readsignedleb128(addr): res = getByte(addr) len = 1 if res <= 0x7f: res = rightshift((res << 25), 25) else: cur = getByte(addr + 1) res = (res & 0x7f) | ((cur & 0x7f) << 7) len = 2 if cur <= 0x7f: res = rightshift((res << 18), 18) else: cur = getByte(addr + 2) res |= (cur & 0x7f) << 14 len = 3 if cur <= 0x7f: res = rightshift((res << 11), 11) else: cur = getByte(addr + 3) res |= (cur & 0x7f) << 21 len = 4 if cur <= 0x7f: res = rightshift((res << 4), 4) else: cur = getByte(addr + 4) res |= cur << 28 len = 5 return res, len def writesignedleb128(num, file): if num >= 0: writeunsignedleb128(num, file) else: mask = 0x80000000 for i in range(0, 32): tmp = num & mask mask >>= 1 if tmp == 0: break loop = 32 - i + 1 while loop > 7: cur = num & 0x7f | 0x80 num >>= 7 file.write(struct.pack("B", cur)) loop -= 7 cur = num & 0x7f file.write(struct.pack("B", cur)) def signedleb128forlen(num): if num >= 0: return unsignedleb128forlen(num) else: mask = 0x80000000 for i in range(0, 32): tmp = num & mask mask >>= 1 if tmp == 0: break loop = 32 - i + 1 if loop % 7 == 0: return loop / 7 else: return loop / 7 + 1 def writeunsignedleb128(num, file): if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7F | 0x80 file.write(struct.pack("B", cur)) num >>= 7 if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7f | 0x80 file.write(struct.pack("B", cur)) num >>= 7 if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7f | 0x80 file.write(struct.pack("B", cur)) num >>= 7 if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7f | 0x80 file.write(struct.pack("B", cur)) num >>= 7 file.write(struct.pack("B", num)) def unsignedleb128forlen(num): len = 1 temp = num while num > 0x7f: len += 1 num >>= 7 if len > 5: print("error for unsignedleb128forlen", temp) return len def readunsignedleb128p1(addr): res, len = readunsignedleb128(addr) return res - 1, len def writeunsignedleb128p1(num, file): writeunsignedleb128(num+1, file) def unsignedleb128p1forlen(num): return unsignedleb128forlen(num+1) def getutf8str(addr): string = [] while 1: onebyte = getByte(addr) addr += 1 if onebyte == 0: break string.append(onebyte) return bytearray(string).decode("utf-8") def getstr(bytes): return bytearray(bytes).decode("utf-8") class DexorJar: def __init__(self): self.pfileName = 0 self.isDex = 0 self.okaytoFree = 0 # do not care self.pRawDexFile = 0 self.pJarFile = 0 self.pDexMemory = 0 # do not care def dump(self, addr): self.pfileName = getDword(addr) self.isDex = getByte(addr + 4) self.pRawDexFile = getDword(addr + 8) self.pJarFile = getDword(addr + 12) def printf(self): str = "" baseaddr = self.pfileName onebyte = getByte(baseaddr) while onebyte != 0: str += chr(onebyte) baseaddr += 1 onebyte = getByte(baseaddr) print("filename is:", str) if self.isDex > 0: print("it's a dex file, addr: ", hex(self.pRawDexFile)) else: print("it's a jar file, addr: ", hex(self.pJarFile)) class JarFile: def __init__(self): self.archive = None # do not care self.pcacheFileName = 0 self.pDvmDex = 0 def dump(self, addr): self.pcacheFileName = getDword(addr + 36) self.pDvmDex = getDword(addr + 40) def printf(self): str = "" baseaddr = self.pcacheFileName one = getByte(baseaddr) while one != 0: str += chr(one) baseaddr += 1 one = getByte(baseaddr) print("cache file name is : ", str) print("DvmDex addr is :", hex(self.pDvmDex)) class DvmDex: def __init__(self): self.pDexFile = 0 self.pHeader = 0 # it is a clone of dex file # just for now def dump(self, addr): self.pDexFile = getDword(addr) self.pHeader = getDword(addr + 4) def printf(self): # i wanna see the diff between the pDexFile.dexfile and pheader print("dexfile addr is: ", hex(self.pDexFile)) print("header addr is: ", hex(self.pHeader)) global baseAddr class DexFile: def __init__(self): self.pOptHeader = 0 self.pHeader = 0 self.pStringIds = 0 self.pTypeIds = 0 self.pFieldIds = 0 self.pMethodIds = 0 self.pProtoIds = 0 self.pClassDefs = 0 self.pLinkData = 0 self.baseAddr = 0 self.dexheader = DexHeader() self.dexmaplist = DexMapList() def dump(self, addr): global baseAddr self.pOptHeader = getDword(addr) self.pHeader = getDword(addr + 4) self.pStringIds = getDword(addr + 8) self.pTypeIds = getDword(addr + 12) self.pFieldIds = getDword(addr + 16) self.pMethodIds = getDword(addr + 20) self.pProtoIds = getDword(addr + 24) self.pProtoIds = getDword(addr + 28) self.pLinkData = getDword(addr + 32) self.baseAddr = getDword(addr + 44) baseAddr = self.baseAddr self.dexheader.dump(self.pHeader) self.dexmaplist.dump(self.dexheader.mapOff) self.dexmaplist.dexmapitem[0].item.append(self.dexheader) print("end build dex and start to get reference") self.dexmaplist.getreference() def makeoffset(self): off = self.dexmaplist.makeoff() align = off % 4 if align != 0: off += (4 - align) self.dexheader.makeoffset(self.dexmaplist.dexmapitem) self.dexheader.fileSize = off self.dexheader.dataSize = off - self.dexheader.mapOff def copytofile(self, filename): if os.path.exists(filename): os.remove(filename) file = open(filename, 'wb+') file.seek(0, 0) self.makeoffset() self.dexmaplist.copy(file) rest = self.dexheader.fileSize -file.tell() for i in range(0, rest): file.write(struct.pack("B", 0)) file_sha = get_file_sha1(file) tmp = bytes(file_sha) i = 0 file.seek(12) while i < 40: num = (ACSII[tmp[i]] << 4) + ACSII[tmp[i+1]] file.write(struct.pack("B", num)) i += 2 csum = checksum(file, self.dexheader.fileSize) print("checksum:", hex(csum), "file size:", self.dexheader.fileSize) file.seek(8) file.write(struct.pack("I", csum)) file.close() def printf(self): print("dex head addr: ", hex(self.pHeader)) print("dex head addr: ", hex(self.baseAddr)) class DexHeader: def __init__(self): self.magic = [] self.checksum = 0 self.signature = [] self.fileSize = 0 self.headerSize = 0 self.endianTag = 0 self.linkSize = 0 self.linkOff = 0 self.mapOff = 0 self.stringIdsSize = 0 self.stringIdsOff = 0 self.typeIdsSize = 0 self.typeIdsOff = 0 self.protoIdsSize = 0 self.protoIdsOff = 0 self.fieldIdsSize = 0 self.fieldIdsOff = 0 self.methodIdsSize = 0 self.methodIdsOff = 0 self.classDefsSize = 0 self.classDefsOff = 0 self.dataSize = 0 self.dataOff = 0 def dump(self, addr): len = 0 while len < 8: self.magic.append(getByte(addr + len)) len += 1 self.checksum = getDword(addr + 8) len = 0 while len < 20: self.signature.append(getByte(addr + 12 + len)) len += 1 self.fileSize = getDword(addr + 32) self.headerSize = getDword(addr + 36) self.endianTag = getDword(addr + 40) self.linkSize = getDword(addr + 44) self.linkOff = getDword(addr + 48) self.mapOff = getDword(addr + 52) self.stringIdsSize = getDword(addr + 56) self.stringIdsOff = getDword(addr + 60) self.typeIdsSize = getDword(addr + 64) self.typeIdsOff = getDword(addr + 68) self.protoIdsSize = getDword(addr + 72) self.protoIdsOff = getDword(addr + 76) self.fieldIdsSize = getDword(addr + 80) self.fieldIdsOff = getDword(addr + 88) self.methodIdsSize = getDword(addr + 92) self.methodIdsOff = getDword(addr + 96) self.classDefsSize = getDword(addr + 100) self.classDefsOff = getDword(addr + 104) self.dataSize = getDword(addr + 108) self.dataOff = getDword(addr + 112) def makeoffset(self, dexmaplist): self.stringIdsSize = dexmaplist[1].size self.stringIdsOff = dexmaplist[1].offset self.typeIdsSize = dexmaplist[2].size self.typeIdsOff = dexmaplist[2].offset self.protoIdsSize = dexmaplist[3].size self.protoIdsOff = dexmaplist[3].offset self.fieldIdsSize = dexmaplist[4].size self.fieldIdsOff = dexmaplist[4].offset self.methodIdsSize = dexmaplist[5].size self.methodIdsOff = dexmaplist[5].offset self.classDefsSize = dexmaplist[6].size self.classDefsOff = dexmaplist[6].offset self.dataOff = dexmaplist[0x1000].offset self.dataSize = 0 self.mapOff = dexmaplist[0x1000].offset self.fileSize = 0 def copytofile(self, file): file.seek(0, 0) len = 0 while len < 8: file.write(struct.pack("B", self.magic[len])) len += 1 file.write(struct.pack("I", self.checksum)) len = 0 while len < 20: file.write(struct.pack("B", self.signature[len])) len += 1 file.write(struct.pack("I", self.fileSize)) file.write(struct.pack("I", self.headerSize)) file.write(struct.pack("I", self.endianTag)) file.write(struct.pack("I", self.linkSize)) file.write(struct.pack("I", self.linkOff)) file.write(struct.pack("I", self.mapOff)) file.write(struct.pack("I", self.stringIdsSize)) file.write(struct.pack("I", self.stringIdsOff)) file.write(struct.pack("I", self.typeIdsSize)) file.write(struct.pack("I", self.typeIdsOff)) file.write(struct.pack("I", self.protoIdsSize)) file.write(struct.pack("I", self.protoIdsOff)) file.write(struct.pack("I", self.fieldIdsSize)) file.write(struct.pack("I", self.fieldIdsOff)) file.write(struct.pack("I", self.methodIdsSize)) file.write(struct.pack("I", self.methodIdsOff)) file.write(struct.pack("I", self.classDefsSize)) file.write(struct.pack("I", self.classDefsOff)) file.write(struct.pack("I", self.dataSize)) file.write(struct.pack("I", self.dataOff)) class DexMapList: Seq = (0, 1, 2, 3, 4, 5, 6, 0x1000, 0x1001, 0x1002, 0x1003, 0x2001, 0x2000, 0x2002, 0x2003, 0x2004, 0x2005, 0x2006) def __init__(self): self.start = 0 self.size = 0 self.dexmapitem = {} def dump(self, offset): self.start = offset self.size = getDword(offset + baseAddr) mapitem = [] for i in range(0, self.size): item = DexMapItem() item.dump(offset + baseAddr + 4 + i * 12) mapitem.append(item) item.printf() for i in range(0, self.size): mapitem[i].setitem(self.dexmapitem) self.dexmapitem[mapitem[i].type] = mapitem[i] def copy(self, file): for i in range(0, len(DexMapList.Seq)): index = DexMapList.Seq[i] if index in self.dexmapitem.keys(): print(index, "start at:", file.tell()) if index != 0x1000: self.dexmapitem[index].copytofile(file) else: self.copytofile(file) def copytofile(self, file): print("output map list", file.tell()) file.seek(self.start, 0) file.write(struct.pack("I", self.size)) for i in range(0, len(DexMapList.Seq)): index = DexMapList.Seq[i] if index in self.dexmapitem.keys(): # print(self.dexmapitem[index].type) file.write(struct.pack("H", self.dexmapitem[index].type)) file.write(struct.pack("H", self.dexmapitem[index].unused)) file.write(struct.pack("I", self.dexmapitem[index].size)) file.write(struct.pack("I", self.dexmapitem[index].offset)) def makeoff(self): off = 0 for i in range(0, len(DexMapList.Seq)): index = DexMapList.Seq[i] if index in self.dexmapitem.keys(): align = off % 4 if align != 0: off += (4 - align) if index != 0x1000: off = self.dexmapitem[index].makeoffset(off) else: off = self.makeoffset(off) return off def makeoffset(self, off): self.start = off off += (4 + self.size * 12) self.dexmapitem[0x1000].offset = self.start return off def getreference(self): self.dexmapitem[1].getref(self.dexmapitem) print("string id get ref done") self.dexmapitem[3].getref(self.dexmapitem) print("proto id get ref done") self.dexmapitem[6].getref(self.dexmapitem) print("class def get ref done") if 0x1002 in self.dexmapitem.keys(): self.dexmapitem[0x1002].getref(self.dexmapitem) print("annotation set ref get ref done") if 0x1003 in self.dexmapitem.keys(): self.dexmapitem[0x1003].getref(self.dexmapitem) print("annotation set get ref done") # self.dexmapitem[0x2000].getref(self.dexmapitem) self.dexmapitem[0x2001].getref(self.dexmapitem) print("code item get ref done") if 0x2006 in self.dexmapitem.keys(): self.dexmapitem[0x2006].getref(self.dexmapitem) print("annotation dir item get ref done") def getrefbystr(self, str): return self.dexmapitem[0x2002].getrefbystr(str) def printf(self, index): print ("DexMapList:") print ("size: ", self.size) for i in self.dexmapitem: self.dexmapitem[i].printf(index) class DexMapItem: Constant = {0: 'TYPE_HEADER_ITEM', 1: 'TYPE_STRING_ID_ITEM', 2: 'TYPE_TYPE_ID_ITEM', 3: 'TYPE_PROTO_ID_ITEM', 4: 'TYPE_FIELD_ID_ITEM', 5: 'TYPE_METHOD_ID_ITEM', 6: 'TYPE_CLASS_DEF_ITEM', 0x1000: 'TYPE_MAP_LIST', 0x1001: 'TYPE_TYPE_LIST', 0x1002: 'TYPE_ANNOTATION_SET_REF_LIST', 0x1003: 'TYPE_ANNOTATION_SET_ITEM', 0x2000: 'TYPE_CLASS_DATA_ITEM', 0x2001: 'TYPE_CODE_ITEM', 0x2002: 'TYPE_STRING_DATA_ITEM', 0x2003: 'TYPE_DEBUG_INFO_ITEM', 0x2004: 'TYPE_ANNOTATION_ITEM', 0x2005: 'TYPE_ENCODED_ARRAY_ITEM', 0x2006: 'TYPE_ANNOTATIONS_DIRECTORY_ITEM'} def __init__(self): self.type = 0 self.unused = 0 self.size = 0 self.offset = 0 self.item = [] self.len = 0 def dump(self, addr): self.type = getWord(addr) self.unused = getWord(addr + 2) self.size = getDword(addr + 4) self.offset = getDword(addr + 8) def copytofile(self, file): file.seek(self.offset, 0) if self.type <= 0x2006: align = file.tell() % 4 if align != 0: for i in range(0, 4-align): file.write(struct.pack("B", 0)) print("copytofile:", DexMapItem.Constant[self.type], file.tell()) for i in range(0, self.size): if self.type == 0x2000: print("index, offset", i, hex(self.item[i].start), self.item[i].static_field_size, self.item[i].instance_fields_size, self.item[i].direct_methods_size, self.item[i].virtual_methods_size) self.item[i].copytofile(file) # if self.type == 0x2002: # print("for debug", i, getstr(self.item[i].str)) def printf(self): print ("type: ", DexMapItem.Constant[self.type]) print ("size: ", self.size) print ("offset: ", hex(self.offset), hex(self.offset + baseAddr)) # if self.type == index: # for i in range(0, self.size): # self.item[i].printf() # print () def setitem(self, dexmapitem): self.printf() addr = baseAddr + self.offset for i in range(0, self.size): if self.type == 1: # string dexstringid = DexStringID() dexstringid.dump(addr + 4 * i) self.item.append(dexstringid) elif self.type == 2: dextypeid = DexTypeID() dextypeid.dump(addr + 4 * i, dexmapitem[1].item) self.item.append(dextypeid) # make sure has already build string table elif self.type == 3: dexprotoid = DexProtoId() dexprotoid.dump(addr + 12 * i, dexmapitem[1].item, dexmapitem[2].item) self.item.append(dexprotoid) elif self.type == 4: dexfieldid = DexFieldId() dexfieldid.dump(addr + 8 * i, dexmapitem[1].item, dexmapitem[2].item) self.item.append(dexfieldid) elif self.type == 5: dexmethodid = DexMethodId() dexmethodid.dump(addr + 8 * i, dexmapitem[1].item, dexmapitem[2].item) self.item.append(dexmethodid) elif self.type == 6: dexclassdef = DexClassDef() dexclassdef.dump(addr + 32 * i, dexmapitem[1].item, dexmapitem[2].item) self.item.append(dexclassdef) elif self.type == 0x1001: # TYPE_TYPE_LIST typeitem = TypeItem() typeitem.dump(addr, dexmapitem[2].item) addr += typeitem.len self.item.append(typeitem) elif self.type == 0x1002: # TYPE_ANNOTATION_SET_REF_LIST annoitem = AnnotationsetrefList() annoitem.dump(addr) addr += annoitem.len self.item.append(annoitem) elif self.type == 0x1003: # TYPE_ANNOTATION_SET_ITEM annoitem = AnnotationsetItem() annoitem.dump(addr) addr += annoitem.len self.item.append(annoitem) elif self.type == 0x2000: # TYPE_CLASS_DATA_ITEM classitem = ClassdataItem() classitem.dump(addr, dexmapitem[0x2001].item) addr += classitem.len self.item.append(classitem) # elif self.type == 0x2001: # TYPE_CODE_ITEM # codeitem = CodeItem() # codeitem.dump(addr) # addr += codeitem.len # self.item.append(codeitem) elif self.type == 0x2002: # TYPE_STRING_DATA_ITEM stringdata = StringData() stringdata.dump(addr) addr += stringdata.len self.item.append(stringdata) elif self.type == 0x2003: # TYPE_DEBUG_INFO_ITEM debuginfo = DebugInfo() debuginfo.dump(addr) addr += debuginfo.len self.item.append(debuginfo) elif self.type == 0x2004: # TYPE_ANNOTATION_ITEM item = AnnotationItem() item.dump(addr) addr += item.len self.item.append(item) elif self.type == 0x2005: # TYPE_ENCODED_ARRAY_ITEM arrayitem = EncodedArrayItem() arrayitem.dump(addr) addr += arrayitem.len self.item.append(arrayitem) elif self.type == 0x2006: # TYPE_ANNOTATIONS_DIRECTORY_ITEM dirItem = AnnotationsDirItem() dirItem.dump(addr) addr += dirItem.len self.item.append(dirItem) def makeoffset(self, off): if self.type < 0x2000 or self.type == 0x2001 or self.type == 0x2006: align = off % 4 if align != 0: off += (4 - align) self.offset = off if self.type == 0: # header self.len = 112 elif self.type == 1: # string id self.len = 4 * self.size elif self.type == 2: # type id self.len = 4 * self.size elif self.type == 3: # proto id self.len = 12 * self.size elif self.type == 4: # field id self.len = 8 * self.size elif self.type == 5: # method id self.len = 8 * self.size elif self.type == 6: # class def self.len = 32 * self.size elif self.type == 0x1000: # map list, resolve specially in dexmaplist class pass elif 0x1001 <= self.type <= 0x2006: # type list, annotation ref set list, annotation set item... for i in range(0, self.size): off = self.item[i].makeoffset(off) # if self.type == 0x2002: # print("for debug", i, off) self.len = off - self.offset if self.type == 0x2000: print("the off is:", off) if self.type <= 6: return off + self.len else: return off def getref(self, dexmaplist): for i in range(0, self.size): self.item[i].getreference(dexmaplist) def getreference(self, offset): # offset if offset == 0: return None i = 0 for i in range(0, self.size): if self.item[i].start == offset + baseAddr: return self.item[i] # if i >= self.size: # os._exit(offset) print("failed : don not find the refernce") return None def getrefbystr(self, str): # for modify the string data if self.type == 0x2002: for i in range(0, self.size): if getstr(self.item[i].str) == str: return self.item[i] else: print("error occur here", self.type) return None def getindexbyname(self, str): # search for type id item for i in range(0, self.size): if self.item[i].str == str: print("find index of", DexMapItem.Constant[self.type], str) return i print("did not find it in", DexMapItem.Constant[self.type]) return -1 def getindexbyproto(self, short_idx, return_type_idx, param_list, length): # called by item, index of 3 for i in range(0, self.size): if short_idx == self.item[i].shortyIdx and return_type_idx == self.item[i].returnTypeIdx: if self.item[i].ref is not None: if self.item[i].ref.equal(param_list, length): return i return -1 class DexStringID: def __init__(self): self.stringDataOff = 0 self.size = 0 self.str = "" self.ref = None def dump(self, addr): self.stringDataOff = getDword(addr) self.size, len = readunsignedleb128(self.stringDataOff + baseAddr) self.str = getutf8str(self.stringDataOff + len + baseAddr) def copytofile(self, file): # self.stringDataoff = self.ref.start file.write(struct.pack("I", self.ref.start)) def getreference(self, dexmaplist): self.ref = dexmaplist[0x2002].getreference(self.stringDataOff) # if self.ref is not None: # self.ref.printf() def printf(self): print ("size: ", self.size, " str: ", self.str, "dataof: ", self.stringDataOff) class DexTypeID: def __init__(self): self.descriptorIdx = 0 self.str = "" def dump(self, addr, str_table): self.descriptorIdx = getDword(addr) self.str = str_table[self.descriptorIdx].str def copytofile(self, file): file.write(struct.pack("I", self.descriptorIdx)) def printf(self): print ("type id: ", self.str) class DexProtoId: def __init__(self): self.shortyIdx = 0 self.returnTypeIdx = 0 self.parametersOff = 0 self.name = "" self.returnstr = "" self.ref = None def dump(self, addr, str_table, type_table): self.shortyIdx = getDword(addr) self.returnTypeIdx = getDword(addr + 4) self.parametersOff = getDword(addr + 8) self.name = str_table[self.shortyIdx].str self.returnstr = type_table[self.returnTypeIdx].str def copytofile(self, file): file.write(struct.pack("I", self.shortyIdx)) file.write(struct.pack("I", self.returnTypeIdx)) if self.ref is not None: file.write(struct.pack("I", self.ref.start)) else: file.write(struct.pack("I", 0)) def getreference(self, dexmaplist): self.ref = dexmaplist[0x1001].getreference(self.parametersOff) def printf(self): print ("return Type:", self.returnstr) print ("methodname:", self.name) if self.ref is not None: self.ref.printf() class DexFieldId: def __init__(self): self.classIdx = 0 self.typeIdx = 0 self.nameIdx = 0 self.classstr = "" self.typestr = "" self.name = "" def dump(self, addr, str_table, type_table): self.classIdx = getWord(addr) self.typeIdx = getWord(addr + 2) self.nameIdx = getDword(addr + 4) self.classstr = type_table[self.classIdx].str self.typestr = type_table[self.typeIdx].str self.name = str_table[self.nameIdx].str def copytofile(self, file): file.write(struct.pack("H", self.classIdx)) file.write(struct.pack("H", self.typeIdx)) file.write(struct.pack("I", self.nameIdx)) def printf(self): print ("classstr:", self.classstr) print ("typestr:", self.typestr) print ("name:", self.name) print () class DexMethodId: def __init__(self): self.classIdx = 0 self.protoIdx = 0 self.nameIdx = 0 self.classstr = "" self.name = "" def dump(self, addr, str_table, type_table): self.classIdx = getWord(addr) self.protoIdx = getWord(addr + 2) self.nameIdx = getWord(addr + 4) self.classstr = type_table[self.classIdx].str self.name = str_table[self.nameIdx].str def copytofile(self, file): file.write(struct.pack("H", self.classIdx)) file.write(struct.pack("H", self.protoIdx)) file.write(struct.pack("I", self.nameIdx)) def printf(self): print ("classstr:", self.classstr) print ("name:", self.name) print () class DexClassDef: def __init__(self): self.classIdx = 0 self.accessFlags = 0 self.superclassIdx = 0 self.interfacesOff = 0 self.sourceFileIdx = 0 self.annotationsOff = 0 self.classDataOff = 0 self.staticValuesOff = 0 self.classstr = "" self.superclassstr = "" self.sourceFilestr = "" self.interfacesRef = None self.annotationsRef = None self.classDataRef = None self.staticValuesRef = None def dump(self, addr, str_table, type_table): self.classIdx = getDword(addr) self.accessFlags = getDword(addr + 4) self.superclassIdx = getDword(addr + 8) self.interfacesOff = getDword(addr + 12) self.sourceFileIdx = getDword(addr + 16) self.annotationsOff = getDword(addr + 20) self.classDataOff = getDword(addr + 24) self.staticValuesOff = getDword(addr + 28) self.classstr = type_table[self.classIdx].str self.superclassstr = type_table[self.superclassIdx].str if self.sourceFileIdx == 0xFFFFFFFF: self.sourceFilestr = "NO_INDEX" else: self.sourceFilestr = str_table[self.sourceFileIdx].str # get class data reference by its name,e.g. Lcom/cc/test/MainActivity; def getclassdefref(self, str): if self.classstr == str and self.classDataOff > 0: return self.classDataRef return None def copytofile(self, file): file.write(struct.pack("I", self.classIdx)) file.write(struct.pack("I", self.accessFlags)) file.write(struct.pack("I", self.superclassIdx)) if self.interfacesRef is not None: file.write(struct.pack("I", self.interfacesRef.start)) # print(self.interfacesRef.start) else: file.write(struct.pack("I", 0)) file.write(struct.pack("I", self.sourceFileIdx)) if self.annotationsRef is not None: file.write(struct.pack("I", self.annotationsRef.start)) # print(self.annotationsRef.start) else: file.write(struct.pack("I", 0)) if self.classDataRef is not None: file.write(struct.pack("I", self.classDataRef.start)) else: file.write(struct.pack("I", 0)) if self.staticValuesRef is not None: file.write(struct.pack("I", self.staticValuesRef.start)) else: file.write(struct.pack("I", 0)) def getreference(self, dexmaplist): self.interfacesRef = dexmaplist[0x1001].getreference(self.interfacesOff) if 0x2006 in dexmaplist.keys(): self.annotationsRef = dexmaplist[0x2006].getreference(self.annotationsOff) self.classDataRef = dexmaplist[0x2000].getreference(self.classDataOff) if 0x2005 in dexmaplist.keys(): self.staticValuesRef = dexmaplist[0x2005].getreference(self.staticValuesOff) def printf(self): print ("classtype:", self.classIdx, self.classstr) print("access flag:", self.accessFlags) print ("superclasstype:", self.superclassIdx, self.superclassstr) print ("iterface off", self.interfacesOff) print("source file index", self.sourceFilestr) print("annotations off", self.annotationsOff) print("class data off", self.classDataOff) print("static values off", self.staticValuesOff) if self.interfacesRef is not None: self.interfacesRef.printf() if self.annotationsRef is not None: self.annotationsRef.printf() if self.classDataRef is not None: self.classDataRef.printf() if self.staticValuesRef is not None: self.staticValuesRef.printf() class TypeItem: # alignment: 4 bytes def __init__(self): self.start = 0 self.size = 0 self.list = [] self.str = [] self.len = 0 def dump(self, addr, type_table): self.start = addr self.size = getDword(addr) self.len = 4 + 2 * self.size for i in range(0, self.size): self.list.append(getWord(addr + 4 + 2 * i)) self.str.append(type_table[self.list[i]].str) if self.size % 2 == 1: getWord(addr + 4 + 2 * self.size) self.len += 2 def copytofile(self, file): file.write(struct.pack("I", self.size)) for i in range(0, self.size): file.write(struct.pack("H", self.list[i])) if self.size % 2 == 1: file.write(struct.pack("H", 0)) def equal(self, param_list, length): if length != self.size: return False for i in range(0, self.size): if param_list[i] != self.str[i]: return False return True def makeoffset(self, off): align = off % 4 if align != 0: off += (4 - align) self.len = 4 + 2 * self.size self.start = off return off + self.len def printf(self): for i in range(0, self.size): print (self.list[i], self.str[i]) # alignment: 4bytes class AnnotationsetItem: def __init__(self): self.start = 0 self.len = 0 self.size = 0 self.entries = [] self.ref = [] def dump(self, addr): self.start = addr self.size = getDword(addr) self.len = 4 + 4 * self.size for i in range(0, self.size): self.entries.append(getDword(addr + 4 + 4 * i)) def copytofile(self, file): file.write(struct.pack("I", self.size)) for i in range(0, self.size): file.write(struct.pack("I", self.ref[i].start)) def makeoffset(self, off): align = off % 4 if align != 0: off += (4 - align) self.start = off self.len = 4 + 4 * self.size return off + self.len def getreference(self, dexmaplist): for i in range(0, self.size): self.ref.append(dexmaplist[0x2004].getreference(self.entries[i])) def printf(self): print ("size: ", self.size) # alignment: 4bytes class AnnotationsetrefList: def __init__(self): self.start = 0 self.size = 0 self.list = [] # annotaions_off, offset of annotation_set_item self.ref = [] self.len = 0 def dump(self, addr): self.start = addr self.size = getDword(addr) self.len = 4 + 4 * self.size for i in range(0, self.size): self.list.append(getDword(addr + 4 + 4 * i)) def copytofile(self, file): file.write(struct.pack("I", self.size)) for i in range(0, self.size): if self.ref[i] is not None: file.write(struct.pack("I", self.ref[i].start)) else: file.write(struct.pack("I", 0)) def makeoffset(self, off): align = off % 4 if align != 0: off += (4 - align) self.start = off self.len = 4 + 4 * self.size return off + self.len def getreference(self, dexmaplist): for i in range(0, self.size): self.ref.append(dexmaplist[0x1003].getreference(self.list[i])) def printf(self): print ("size: ", self.size) # alignment:none class ClassdataItem: def __init__(self): self.start = 0 self.len = 0 self.static_field_size = 0 self.instance_fields_size = 0 self.direct_methods_size = 0 self.virtual_methods_size = 0 self.static_fields = [] self.instance_fields = [] self.direct_methods = [] self.virtual_methods = [] def dump(self, addr, code_table): self.start = addr self.static_field_size, length = readunsignedleb128(addr) self.len += length self.instance_fields_size, length = readunsignedleb128(addr + self.len) self.len += length self.direct_methods_size, length = readunsignedleb128(addr + self.len) self.len += length self.virtual_methods_size, length = readunsignedleb128(addr + self.len) self.len += length print("class item", self.static_field_size, self.instance_fields_size, self.direct_methods_size, self.virtual_methods_size) for i in range(0, self.static_field_size): field = Encodedfield() field.dump(addr + self.len) self.len += field.len self.static_fields.append(field) for i in range(0, self.instance_fields_size): field = Encodedfield() field.dump(addr + self.len) self.len += field.len self.instance_fields.append(field) for i in range(0, self.direct_methods_size): method = Encodedmethod() method.dump(addr + self.len, code_table) self.len += method.len self.direct_methods.append(method) for i in range(0, self.virtual_methods_size): method = Encodedmethod() method.dump(addr + self.len, code_table) self.len += method.len self.virtual_methods.append(method) def copytofile(self, file): file.seek(self.start) writeunsignedleb128(self.static_field_size, file) writeunsignedleb128(self.instance_fields_size, file) writeunsignedleb128(self.direct_methods_size, file) writeunsignedleb128(self.virtual_methods_size, file) for i in range(0, self.static_field_size): self.static_fields[i].copytofile(file) for i in range(0, self.instance_fields_size): self.instance_fields[i].copytofile(file) for i in range(0, self.direct_methods_size): self.direct_methods[i].copytofile(file) for i in range(0, self.virtual_methods_size): self.virtual_methods[i].copytofile(file) # besides adding refenrence, also need to set the correct index def getreference(self, dexmaplist): last = 0 for i in range(0, self.static_field_size): self.static_fields[i].field_idx = last + self.static_fields[i].field_idx_diff last = self.static_fields[i].field_idx last = 0 for i in range(0, self.instance_fields_size): self.instance_fields[i].field_idx = last + self.instance_fields[i].field_idx_diff last = self.instance_fields[i].field_idx last = 0 for i in range(0, self.direct_methods_size): self.direct_methods[i].getreference(dexmaplist) self.direct_methods[i].method_idx = last + self.direct_methods[i].method_idx_diff last = self.direct_methods[i].method_idx last = 0 for i in range(0, self.virtual_methods_size): self.virtual_methods[i].getreference(dexmaplist) self.virtual_methods[i].method_idx = last + self.virtual_methods[i].method_idx_diff last = self.virtual_methods[i].method_idx def makeoffset(self, off): self.start = off off += unsignedleb128forlen(self.static_field_size) off += unsignedleb128forlen(self.instance_fields_size) off += unsignedleb128forlen(self.direct_methods_size) off += unsignedleb128forlen(self.virtual_methods_size) for i in range(0, self.static_field_size): off = self.static_fields[i].makeoffset(off) for i in range(0, self.instance_fields_size): off = self.instance_fields[i].makeoffset(off) for i in range(0, self.direct_methods_size): off = self.direct_methods[i].makeoffset(off) for i in range(0, self.virtual_methods_size): off = self.virtual_methods[i].makeoffset(off) self.len = off - self.start return off def printf(self): print ("static field size: ", self.static_field_size) print ("instance fields size: ", self.instance_fields_size) print ("direct methods size: ", self.direct_methods_size) print ("virtual methods size: ", self.virtual_methods_size) for i in range(0, self.static_field_size): self.static_fields[i].printf() for i in range(0, self.instance_fields_size): self.instance_fields[i].printf() for i in range(0, self.direct_methods_size): self.direct_methods[i].printf() for i in range(0, self.virtual_methods_size): self.virtual_methods[i].printf() class Encodedfield: def __init__(self): self.start = 0 self.len = 0 self.field_idx_diff = 0 self.access_flags = 0 self.field_idx = 0 # need to set later def dump(self, addr): self.start = addr self.field_idx_diff, length = readunsignedleb128(addr) self.len += length self.access_flags, length = readunsignedleb128(addr + self.len) self.len += length def __lt__(self, other): # for sort return self.field_idx_diff < other.field_idx_diff def copytofile(self, file): writeunsignedleb128(self.field_idx_diff, file) writeunsignedleb128(self.access_flags, file) def makeoffset(self, off): self.start = off self.len += unsignedleb128forlen(self.field_idx_diff) self.len += unsignedleb128forlen(self.access_flags) return off + self.len def printf(self): print ("diff: ", self.field_idx_diff) print ("access: ", self.access_flags) class Encodedmethod: def __init__(self): self.start = 0 self.len = 0 self.method_idx_diff = 0 self.access_flags = 0 self.code_off = 0 self.method_idx = 0 self.coderef = None self.modified = 0 # if set this var, means that code_off will moodified to zero def dump(self, addr, code_table): self.start = addr self.method_idx_diff, length = readunsignedleb128(addr) self.len += length self.access_flags, length = readunsignedleb128(addr + self.len) self.len += length self.code_off, length = readunsignedleb128(addr + self.len) self.len += length if self.code_off != 0: self.coderef = CodeItem() self.coderef.dump(int(self.code_off + baseAddr) & 0xFFFFFFFF) code_table.append(self.coderef) def copytofile(self, file): writeunsignedleb128(self.method_idx_diff, file) writeunsignedleb128(self.access_flags, file) if self.modified == 1: writeunsignedleb128(0, file) elif self.coderef is not None: writeunsignedleb128(self.coderef.start, file) else: writeunsignedleb128(0, file) def makeoffset(self, off): self.start = off self.len += unsignedleb128forlen(self.method_idx_diff) self.len += unsignedleb128forlen(self.access_flags) if self.modified == 1: self.len += unsignedleb128forlen(0) elif self.coderef is not None: self.len += unsignedleb128forlen(self.coderef.start) else: self.len += unsignedleb128forlen(0) return off + self.len def getreference(self, dexmaplist): self.coderef = dexmaplist[0x2001].getreference(self.code_off) def printf(self): print ("method_idx_diff: ", self.method_idx_diff) print("method idx:", self.method_idx) print ("access: ", self.access_flags) print ("code off: ", self.code_off) # alignment: 4bytes class CodeItem: def __init__(self): self.start = 0 self.len = 0 self.register_size = 0 self.ins_size = 0 self.outs_size = 0 self.tries_size = 0 self.debug_info_off = 0 self.insns_size = 0 self.insns = [] self.debugRef = None self.padding = 0 self.tries = [] self.handler = None def dump(self, addr): self.start = addr self.register_size = getWord(addr) self.ins_size = getWord(addr + 2) self.outs_size = getWord(addr + 4) self.tries_size = getWord(addr + 6) self.debug_info_off = getDword(addr + 8) self.insns_size = getDword(addr + 12) self.len += 16 print(self.start, self.register_size, self.ins_size, self.outs_size, self.tries_size, self.debug_info_off, self.insns_size) for i in range(0, self.insns_size): self.insns.append(getWord(addr + self.len + 2 * i)) self.len += 2 * self.insns_size if self.tries_size != 0 and self.insns_size % 2 == 1: self.len += 2 for i in range(0, self.tries_size): tryitem = TryItem() tryitem.dump(addr + self.len + 8 * i) self.tries.append(tryitem) self.len += 8 * self.tries_size if self.tries_size != 0: self.handler = EncodedhandlerList() self.handler.dump(addr + self.len) self.len += self.handler.len align = self.len % 4 if align != 0: self.len += (4 - align) def copytofile(self, file): file.seek(self.start, 0) file.write(struct.pack("H", self.register_size)) file.write(struct.pack("H", self.ins_size)) file.write(struct.pack("H", self.outs_size)) file.write(struct.pack("H", self.tries_size)) if self.debugRef is not None: file.write(struct.pack("I", self.debugRef.start)) else: file.write(struct.pack("I", 0)) file.write(struct.pack("I", self.insns_size)) for i in range(0, self.insns_size): file.write(struct.pack("H", self.insns[i])) if self.tries_size != 0 and self.insns_size % 2 == 1: file.write(struct.pack("H", self.padding)) for i in range(0, self.tries_size): self.tries[i].copytofile(file) if self.tries_size != 0: self.handler.copytofile(file) align = file.tell() % 4 # for alignment if align != 0: for i in range(0, 4-align): file.write(struct.pack("B", 0)) # print("code item addr:", file.tell()) def makeoffset(self, off): align = off % 4 if align != 0: off += (4 - align) self.start = off off += (4 * 2 + 2 * 4) # 4 ushort and 2 uint off += (2 * self.insns_size) if self.tries_size != 0 and self.insns_size % 2 == 1: # for padding off += 2 for i in range(0, self.tries_size): off = self.tries[i].makeoffset(off) if self.tries_size != 0: off = self.handler.makeoffset(off) self.len = off - self.start return off def getreference(self, dexmaplist): self.debugRef = dexmaplist[0x2003].getreference(self.debug_info_off) def printf(self): print("registers_size:", self.register_size) print("ins_size, outs_size, tries_size:", self.ins_size, self.outs_size, self.tries_size) print("debug info of:", self.debug_info_off) print("insn_size:", self.insns_size) for i in range(0, self.insns_size): print(self.insns[i]) # tmp = Instruction.InstructionSet(self.insns) # tmp.printf() class TryItem: def __init__(self): self.start = 0 self.len = 8 self.start_addr = 0 self.insn_count = 0 self.handler_off = 0 def dump(self, addr): self.start = addr self.start_addr = getDword(addr) self.insn_count = getWord(addr + 4) self.handler_off = getWord(addr + 6) def copytofile(self, file): file.write(struct.pack("I", self.start_addr)) file.write(struct.pack("H", self.insn_count)) file.write(struct.pack("H", self.handler_off)) def makeoffset(self, off): self.start = off self.len = 4 + 2 + 2 return off + self.len def printf(self): print ("start_Addr: ", self.start_addr) print ("insn_count: ", self.insn_count) print ("handler_off: ", self.handler_off) print () class EncodedhandlerList: def __init__(self): self.start = 0 self.len = 0 self.size = 0 self.list = [] def dump(self, addr): self.start = addr self.size, length = readunsignedleb128(addr) self.len += length for i in range(0, self.size): handler = EncodedhandlerItem() handler.dump(addr + self.len) self.len += handler.len self.list.append(handler) def copytofile(self, file): file.seek(self.start, 0) writeunsignedleb128(self.size, file) for i in range(0, self.size): self.list[i].copytofile(file) def makeoffset(self, off): self.start = off off += unsignedleb128forlen(self.size) for i in range(0, self.size): off = self.list[i].makeoffset(off) return off class EncodedhandlerItem: def __init__(self): self.start = 0 self.len = 0 self.size = 0 self.handlers = [] self.catch_all_addr = 0 def dump(self, addr): self.start = addr self.size, length = readsignedleb128(addr) self.len += length for i in range(0, abs(self.size)): pair = EncodedTypeAddrPair() pair.dump(addr + self.len) self.len += pair.len self.handlers.append(pair) if self.size <= 0: self.catch_all_addr, length = readunsignedleb128(addr + self.len) self.len += length def copytofile(self, file): writesignedleb128(self.size, file) for i in range(0, abs(self.size)): self.handlers[i].copytofile(file) if self.size <= 0: writeunsignedleb128(self.catch_all_addr, file) def makeoffset(self, off): self.start = off off += signedleb128forlen(self.size) for i in range(0, abs(self.size)): off = self.handlers[i].makeoffset(off) if self.size <= 0: off += unsignedleb128forlen(self.catch_all_addr) self.len = off - self.start return off class EncodedTypeAddrPair: def __init__(self): self.type_idx = 0 self.addr = 0 self.len = 0 def dump(self, addr): self.type_idx, length = readunsignedleb128(addr) self.len += length self.addr, length = readunsignedleb128(addr + length) self.len += length def copytofile(self, file): writeunsignedleb128(self.type_idx, file) writeunsignedleb128(self.addr, file) def makeoffset(self, off): off += unsignedleb128forlen(self.type_idx) off += unsignedleb128forlen(self.addr) return off def printf(self): print ("type idx: ", self.type_idx) print ("addr: ", self.addr) print () class StringData: def __init__(self): self.start = 0 self.len = 0 self.size = 0 self.str = [] def dump(self, addr): self.start = addr self.size, length = readunsignedleb128(addr) self.len += length while 1: onebyte = getByte(addr + self.len) self.len += 1 if onebyte == 0: break self.str.append(onebyte) def copytofile(self, file): writeunsignedleb128(self.size, file) for i in range(0, len(self.str)): file.write(struct.pack("B", self.str[i])) file.write(struct.pack("B", 0)) def makeoffset(self, off): self.start = off self.len = len(self.str) + unsignedleb128forlen(self.size) return off + self.len + 1 # 1 byte for '\0' def modify(self, str): self.size = len(str) self.str = bytearray(str) def printf(self): print (getstr(self.str)) # alignment: none class DebugInfo: def __init__(self): self.start = 0 self.len = 0 self.line_start = 0 self.parameters_size = 0 self.parameter_names = [] self.debug = [] def dump(self, addr): self.start = addr self.line_start, length = readunsignedleb128(addr) self.len += length self.parameters_size, length = readunsignedleb128(addr + self.len) self.len += length for i in range(0, self.parameters_size): num, length = readunsignedleb128p1(addr + self.len) self.len += length self.parameter_names.append(num) while 1: onebyte = getByte(addr + self.len) self.len += 1 self.debug.append(onebyte) if onebyte == 0: break elif onebyte == 1 or onebyte == 5 or onebyte == 6: num, length = readunsignedleb128(addr + self.len) self.len += length self.debug.append(num) elif onebyte == 2: num, length = readsignedleb128(addr + self.len) self.len += length self.debug.append(num) elif onebyte == 3: num, length = readunsignedleb128(addr + self.len) self.len += length self.debug.append(num) num, length = readunsignedleb128p1(addr + self.len) self.len += length self.debug.append(num) num, length = readunsignedleb128p1(addr + self.len) self.len += length self.debug.append(num) elif onebyte == 4: num, length = readunsignedleb128(addr + self.len) self.len += length self.debug.append(num) num, length = readunsignedleb128p1(addr + self.len) self.len += length self.debug.append(num) num, length = readunsignedleb128p1(addr + self.len) self.len += length self.debug.append(num) num, length = readunsignedleb128p1(addr + self.len) self.len += length self.debug.append(num) elif onebyte == 9: num, length = readunsignedleb128p1(addr + self.len) self.len += length self.debug.append(num) def adddebugitem(self, linestart, paramsize, names_list, debug_list): self.line_start = linestart self.parameters_size = paramsize self.parameter_names = names_list self.debug = debug_list def copytofile(self, file): file.seek(self.start, 0) writeunsignedleb128(self.line_start, file) writeunsignedleb128(self.parameters_size, file) for i in range(0, self.parameters_size): # print(self.parameter_names[i]) # if i == self.parameters_size-1: # writeunsignedleb128p1alignshort(self.parameter_names[i], file) # else: writeunsignedleb128p1(self.parameter_names[i], file) index = 0 while 1: onebyte = self.debug[index] file.write(struct.pack("B", onebyte)) index += 1 if onebyte == 0: break elif onebyte == 1: writeunsignedleb128(self.debug[index], file) index += 1 elif onebyte == 2: writesignedleb128(self.debug[index], file) index += 1 elif onebyte == 3: writeunsignedleb128(self.debug[index], file) writeunsignedleb128p1(self.debug[index+1], file) writeunsignedleb128p1(self.debug[index+2], file) index += 3 elif onebyte == 4: writeunsignedleb128(self.debug[index], file) writeunsignedleb128p1(self.debug[index+1], file) writeunsignedleb128p1(self.debug[index+2], file) writeunsignedleb128p1(self.debug[index+3], file) index += 4 elif onebyte == 5: writeunsignedleb128(self.debug[index], file) index += 1 elif onebyte == 6: writeunsignedleb128(self.debug[index], file) index += 1 elif onebyte == 9: writeunsignedleb128p1(self.debug[index], file) index += 1 def printf(self): print(self.line_start, self.parameters_size) def makeoffset(self, off): self.start = off off += unsignedleb128forlen(self.line_start) off += unsignedleb128forlen(self.parameters_size) for i in range(0, self.parameters_size): off += unsignedleb128p1forlen(self.parameter_names[i]) index = 0 while 1: onebyte = self.debug[index] off += 1 index += 1 if onebyte == 0: break elif onebyte == 1: off += unsignedleb128forlen(self.debug[index]) index += 1 elif onebyte == 2: off += signedleb128forlen(self.debug[index]) index += 1 elif onebyte == 3: off += unsignedleb128forlen(self.debug[index]) off += unsignedleb128p1forlen(self.debug[index+1]) off += unsignedleb128p1forlen(self.debug[index+2]) index += 3 elif onebyte == 4: off += unsignedleb128forlen(self.debug[index]) off += unsignedleb128p1forlen(self.debug[index+1]) off += unsignedleb128p1forlen(self.debug[index+2]) off += unsignedleb128p1forlen(self.debug[index+3]) index += 4 elif onebyte == 5: off += unsignedleb128forlen(self.debug[index]) index += 1 elif onebyte == 6: off += unsignedleb128forlen(self.debug[index]) index += 1 elif onebyte == 9: off += unsignedleb128p1forlen(self.debug[index]) index += 1 self.len = off - self.start return off # alignment: none class AnnotationItem: Visibity = {0: 'VISIBITITY_BUILD', 1: 'VISIBILITY_RUNTIME', 2: 'VISIBILITY_SYSTEM'} def __init__(self): self.start = 0 self.len = 0 self.visibility = 0 self.annotation = EncodedAnnotation() def dump(self, addr): self.start = addr self.visibility = getByte(addr) # infile self.annotation.dump(addr + 1) self.len = self.annotation.len + 1 def copytofile(self, file): file.write(struct.pack("B", self.visibility)) self.annotation.copytofile(file) def makeoffset(self, off): self.start = off off += 1 off = self.annotation.makeoffset(off) self.len = off - self.start return off class EncodedAnnotation: def __init__(self): self.len = 0 self.type_idx = 0 self.size = 0 self.elements = [] def dump(self, addr): self.type_idx, length = readunsignedleb128(addr) self.len += length self.size, length = readunsignedleb128(addr + self.len) self.len += length self.elements = [] # annotation_element[size] for i in range(0, self.size): element = AnnotationElement() element.dump(addr + self.len) self.len += element.len self.elements.append(element) def copytofile(self, file): writeunsignedleb128(self.type_idx, file) writeunsignedleb128(self.size, file) for i in range(0, self.size): self.elements[i].copytofile(file) def makeoffset(self, off): off += unsignedleb128forlen(self.type_idx) off += unsignedleb128forlen(self.size) for i in range(0, self.size): off = self.elements[i].makeoffset(off) return off class AnnotationElement: def __init__(self): self.len = 0 self.name_idx = 0 self.value = EncodedValue() def dump(self, addr): self.name_idx, length = readunsignedleb128(addr) self.len += length self.value.dump(addr + self.len) self.len += self.value.len def copytofile(self, file): writeunsignedleb128(self.name_idx, file) self.value.copytofile(file) def makeoffset(self, off): off += unsignedleb128forlen(self.name_idx) off = self.value.makeoffset(off) return off class EncodedValue: def __init__(self): self.len = 0 self.onebyte = 0 self.type = 0 self.arg = 0 self.value = [] def dump(self, addr): self.onebyte = getByte(addr) self.type = self.onebyte & 0x1F self.arg = (self.onebyte >> 5) & 0x7 if self.type == 0x00: # print 'here 0x00 VALUE_BYTE in class : ' + str(curClass_idx) if self.arg != 0: print ("[-] Ca ,get error in VALUE_BYTE") self.value.append(getByte(addr + 1)) self.len = 2 elif self.type == 0x02: # print 'here 0x02 VALUE_SHORT in class : ' + str(curClass_idx) if self.arg >= 2: print ("[-] Ca ,get error in VALUE_SHORT at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x03: # print 'here 0x03 VALUE_CHAR in class : ' + str(curClass_idx) for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x04: # print 'here 0x04 VALUE_INT in class : ' + str(curClass_idx) if self.arg >= 4: print ("[-] Ca ,get error in VALUE_INT at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x06: # print 'here 0x06 VALUE_LONG in class : ' + str(curClass_idx) if self.arg >= 8: print ("[-] Ca ,get error in VALUE_LONG at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x10: # print 'here 0x10 VALUE_FLOAT in class : ' + str(curClass_idx) if self.arg >= 4: print ("[-] Ca ,get error in VALUE_FLOAT at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x11: # print 'here 0x11 VALUE_DOUBLE in class : ' + str(curClass_idx) if self.arg >= 8: print ("[-] Ca ,get error in VALUE_DOUBLE at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x17: # print 'here 0x17 VALUE_STRING in class : ' + str(curClass_idx) if self.arg >= 4: print ("[-] Ca ,get error in VALUE_STRING at class : ", hex(addr)) for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x18: # print 'here 0x18 VALUE_TYPE in class : ' + str(curClass_idx) if self.arg >= 4: print ("[-] Ca ,get error in VALUE_TYPE at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x19: # print 'here 0x19 VALUE_FIELD in class : ' + str(curClass_idx) if self.arg >= 4: print ("[-] Ca ,get error in VALUE_FIELD at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x1a: # print 'here 0x1a VALUE_METHOD in class : ' + str(curClass_idx) if self.arg >= 4: print ("[-] Ca ,get error in VALUE_METHOD at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x1b: # print 'here 0x1b VALUE_ENUM in class : ' + str(curClass_idx) if self.arg >= 4: print ("[-] Ca ,get error in VALUE_ENUM at class : ") for i in range(0, self.arg+1): self.value.append(getByte(addr + 1 + i)) self.len = self.arg + 2 elif self.type == 0x1c: # print 'here 0x1c VALUE_ARRAY in class : ' + str(curClass_idx) if self.arg != 0x00: print ("[-] Ca ,get error in VALUE_ARRAY") array = EncodedArray() array.dump(addr + 1) self.len = array.len + 1 self.value.append(array) elif self.type == 0x1d: # print 'here 0x1d VALUE_ANNOTATION in class : ' + str(curClass_idx) anno = EncodedAnnotation() anno.dump(addr + 1) self.len = anno.len + 1 self.value.append(anno) else: self.len = 1 # if case(0x1e): # print 'here 0x1e VALUE_NULL in class : ' + str(curClass_idx) # break # if case(0x1f): # print 'here 0x1f VALUE_BOOLEAN in class : ' + str(curClass_idx) # break def copytofile(self, file): file.write(struct.pack("B", self.onebyte)) if self.type <= 0x1b: for i in range(0, self.arg+1): file.write(struct.pack("B", self.value[i])) elif self.type == 0x1c: self.value[0].copytofile(file) elif self.type == 0x1d: self.value[0].copytofile(file) def makeoffset(self, off): off += 1 if self.type <= 0x1b: off += self.arg+1 elif self.type == 0x1c: off = self.value[0].makeoffset(off) elif self.type == 0x1d: off = self.value[0].makeoffset(off) return off def printf(self): print("encoded value :", self.type, self.arg) class EncodedArray: def __init__(self): self.size = 0 self.len = 0 self.values = [] def dump(self, addr): self.size, length = readunsignedleb128(addr) self.len += length for i in range(0, self.size): value = EncodedValue() value.dump(addr + self.len) self.len += value.len self.values.append(value) def copytofile(self, file): writeunsignedleb128(self.size, file) for i in range(0, self.size): self.values[i].copytofile(file) def makeoffset(self, off): off += unsignedleb128forlen(self.size) for i in range(0, self.size): off = self.values[i].makeoffset(off) return off def printf(self): print("encoded array size", self.size) # alignment: none class EncodedArrayItem: def __init__(self): self.start = 0 self.len = 0 self.value = EncodedArray() def dump(self, addr): self.start = addr self.len = 0 self.value.dump(addr) self.len = self.value.len def copytofile(self, file): self.value.copytofile(file) def makeoffset(self, off): # if self.start == 1096008: self.start = off off = self.value.makeoffset(off) self.len = off - self.start return off def printf(self): print("None for EncodedArrayItem by now") # alignment: 4 bytes class AnnotationsDirItem: def __init__(self): self.start = 0 self.len = 0 self.class_annotations_off = 0 self.fields_size = 0 self.annotated_methods_size = 0 self.annotate_parameters_size = 0 self.field_annotations = [] # field_annotation[size] self.method_annotations = [] self.parameter_annotations = [] self.class_annotations_ref = None def dump(self, addr): self.start = addr self.class_annotations_off = getDword(addr) # in file self.fields_size = getDword(addr + 4) # in file self.annotated_methods_size = getDword(addr + 8) # in file self.annotate_parameters_size = getDword(addr + 12) # in file self.len = 16 for i in range(0, self.fields_size): field = FieldAnnotation() field.dump(addr + self.len + 8 * i) self.field_annotations.append(field) self.len += 8 * self.fields_size for i in range(0, self.annotated_methods_size): method = MethodAnnotation() method.dump(addr + self.len + 8 * i) self.method_annotations.append(method) self.len += 8 * self.annotated_methods_size for i in range(0, self.annotate_parameters_size): param = ParamterAnnotation() param.dump(addr + 8 * i) self.parameter_annotations.append(param) self.len += 8 * self.annotate_parameters_size def copytofile(self, file): if self.class_annotations_ref is not None: file.write(struct.pack("I", self.class_annotations_ref.start)) else: file.write(struct.pack("I", self.class_annotations_off)) file.write(struct.pack("I", self.fields_size)) file.write(struct.pack("I", self.annotated_methods_size)) file.write(struct.pack("I", self.annotate_parameters_size)) for i in range(0, self.fields_size): self.field_annotations[i].copytofile(file) for i in range(0, self.annotated_methods_size): self.method_annotations[i].copytofile(file) for i in range(0, self.annotate_parameters_size): self.parameter_annotations[i].copytofile(file) def makeoffset(self, off): self.start = off off += 4 * 4 for i in range(0, self.fields_size): off = self.field_annotations[i].makeoffset(off) for i in range(0, self.annotated_methods_size): off = self.method_annotations[i].makeoffset(off) for i in range(0, self.annotate_parameters_size): off = self.parameter_annotations[i].makeoffset(off) self.len = off - self.start return off def getreference(self, dexmaplist): self.class_annotations_ref = dexmaplist[0x1003].getreference(self.class_annotations_off) for i in range(0, self.fields_size): self.field_annotations[i].getreference(dexmaplist) for i in range(0, self.annotated_methods_size): self.method_annotations[i].getreference(dexmaplist) for i in range(0, self.annotate_parameters_size): self.parameter_annotations[i].getreference(dexmaplist) def printf(self): print("None for AnnotationDirItem by now") class FieldAnnotation: def __init__(self): self.field_idx = 0 self.annotations_off = 0 self.annotations_off_ref = None def dump(self, addr): self.field_idx = getDword(addr) # in file self.annotations_off = getDword(addr + 4) # in file, offset of annotation_set_item self.annotations_off_ref = None def copytofile(self, file): file.write(struct.pack("I", self.field_idx)) file.write(struct.pack("I", self.annotations_off_ref.start)) def makeoffset(self, off): off += 4 * 2 return off def getreference(self, dexmaplist): self.annotations_off_ref = dexmaplist[0x1003].getreference(self.annotations_off) class MethodAnnotation: def __init__(self): self.method_idx = 0 self.annotations_off = 0 self.annotations_off_ref = None def dump(self, addr): self.method_idx = getDword(addr) # in file self.annotations_off = getDword(addr + 4) # in file self.annotations_off_ref = None def copytofile(self, file): file.write(struct.pack("I", self.method_idx)) file.write(struct.pack("I", self.annotations_off_ref.start)) def makeoffset(self, off): off += 4 * 2 return off def getreference(self, dexmaplist): self.annotations_off_ref = dexmaplist[0x1003].getreference(self.annotations_off) class ParamterAnnotation: def __init__(self): self.method_idx = 0 self.annotations_off = 0 self.annotations_off_ref = None def dump(self, addr): self.method_idx = getDword(addr) # in file self.annotations_off = getDword(addr + 4) # in file. offset of "annotation_set_ref_list" self.annotations_off_ref = None def copytofile(self, file): file.write(struct.pack("I", self.method_idx)) if self.annotations_off_ref is not None: file.write(struct.pack("I", self.annotations_off_ref.start)) else: file.write(struct.pack("I", 0)) def makeoffset(self, off): off += 4 * 2 return off def getreference(self, dexmaplist): self.annotations_off_ref = dexmaplist[0x1002].getreference(self.annotations_off) addr = 1558176632 print(hex(addr)) cookie = DexorJar() cookie.dump(addr) cookie.printf() if cookie.isDex == 0: jarfile = JarFile() jarfile.dump(cookie.pJarFile) jarfile.printf() dvmaddr = jarfile.pDvmDex else: print("not support yet") dvmDex = DvmDex() dvmDex.dump(dvmaddr) dvmDex.printf() dexfile = DexFile() dexfile.dump(dvmDex.pDexFile) print("begin copy to file:") dexfile.copytofile("dump.dex")
{ "repo_name": "keulraesik/dumpDex", "path": "dump.py", "copies": "5", "size": "76543", "license": "apache-2.0", "hash": 968704851888276900, "line_mean": 34.5683085502, "line_max": 206, "alpha_frac": 0.551859739, "autogenerated": false, "ratio": 3.5134031029101256, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0024213306412503685, "num_lines": 2152 }
__author__ = 'CwT' from idaapi import * from idautils import * from idc import * import struct global baseAddr def writefile(file, addr, len): for i in range(len): one = getByte(addr+i) file.write(struct.pack("B", one)) def getDword(addr): return Dword(addr) def getByte(addr): return Byte(addr) def getWord(addr): return Word(addr) def dexGetStringData(dexfile, offset): addr = dexfile.baseAddr + offset while getByte(addr) > 0x7f: # skip uleb len addr += 1 addr += 1 str = "" one = getByte(addr) while one != 0: str += chr(one) addr += 1 one = getByte(addr) return str def dexGetStringId(dexfile, idx): return getDword(dexfile.pStringIds+4*idx) def dexStringById(dexfile, idx): offset = dexGetStringId(dexfile, idx) return dexGetStringData(dexfile, offset) def dexGetTypeId(dexfile, idx): return getDword(dexfile.pTypeIds+4*idx) def dexStringByTypeIdx(dexfile, idx): return dexStringById(dexfile, dexGetTypeId(dexfile, idx)) def dexGetClassDescriptor(dexfile, classdef): return dexStringByTypeIdx(dexfile, classdef.classIdx) def slashtodot(str): ret = "" for i in str: if i == '/': ret += '.' elif i == ';': continue else: ret += i return ret def rightshift(value, n): mask = 0x80000000 check = value & mask if check != mask: return value >> n else: submask = mask for loop in range(0, n): submask = (submask | (mask >> loop)) strdata = struct.pack("I", submask | (value >> n)) ret = struct.unpack("i", strdata)[0] return ret def readunsignedleb128(addr): res = getByte(addr) len = 1 if res > 0x7f: cur = getByte(addr + 1) res = (res & 0x7f) | ((cur & 0x7f) << 7) len = 2 if cur > 0x7f: cur = getByte(addr + 2) res |= (cur & 0x7f) << 14 len = 3 if cur > 0x7f: cur = getByte(addr + 3) res |= (cur & 0x7f) << 21 len = 4 if cur > 0x7f: cur = getByte(addr + 4) res |= cur << 28 len = 5 return res, len def readsignedleb128(addr): res = getByte(addr) len = 1 if res <= 0x7f: res = rightshift((res << 25), 25) else: cur = getByte(addr + 1) res = (res & 0x7f) | ((cur & 0x7f) << 7) len = 2 if cur <= 0x7f: res = rightshift((res << 18), 18) else: cur = getByte(addr + 2) res |= (cur & 0x7f) << 14 len = 3 if cur <= 0x7f: res = rightshift((res << 11), 11) else: cur = getByte(addr + 3) res |= (cur & 0x7f) << 21 len = 4 if cur <= 0x7f: res = rightshift((res << 4), 4) else: cur = getByte(addr + 4) res |= cur << 28 len = 5 return res, len def writesignedleb128(num, file): if num >= 0: writeunsignedleb128(num, file) else: mask = 0x80000000 for i in range(0, 32): tmp = num & mask mask >>= 1 if tmp == 0: break loop = 32 - i + 1 while loop > 7: cur = num & 0x7f | 0x80 num >>= 7 file.write(struct.pack("B", cur)) loop -= 7 cur = num & 0x7f file.write(struct.pack("B", cur)) def signedleb128forlen(num): if num >= 0: return unsignedleb128forlen(num) else: mask = 0x80000000 for i in range(0, 32): tmp = num & mask mask >>= 1 if tmp == 0: break loop = 32 - i + 1 if loop % 7 == 0: return loop / 7 else: return loop / 7 + 1 def writeunsignedleb128(num, file): if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7F | 0x80 file.write(struct.pack("B", cur)) num >>= 7 if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7f | 0x80 file.write(struct.pack("B", cur)) num >>= 7 if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7f | 0x80 file.write(struct.pack("B", cur)) num >>= 7 if num <= 0x7f: file.write(struct.pack("B", num)) else: cur = num & 0x7f | 0x80 file.write(struct.pack("B", cur)) num >>= 7 file.write(struct.pack("B", num)) def unsignedleb128forlen(num): len = 1 temp = num while num > 0x7f: len += 1 num >>= 7 if len > 5: print("error for unsignedleb128forlen", temp) return len def readunsignedleb128p1(addr): res, len = readunsignedleb128(addr) return res - 1, len def writeunsignedleb128p1(num, file): writeunsignedleb128(num+1, file) def unsignedleb128p1forlen(num): return unsignedleb128forlen(num+1) class DvmDex: def __init__(self): self.pDexFile = 0 self.pHeader = 0 # it is a clone of dex file # just for now def dump(self, addr): self.pDexFile = getDword(addr) self.pHeader = getDword(addr + 4) def printf(self): # i wanna see the diff between the pDexFile.dexfile and pheader print("dexfile addr is: ", hex(self.pDexFile)) print("header addr is: ", hex(self.pHeader)) class DexFile: def __init__(self): self.pOptHeader = 0 self.pHeader = 0 self.pStringIds = 0 self.pTypeIds = 0 self.pFieldIds = 0 self.pMethodIds = 0 self.pProtoIds = 0 self.pClassDefs = 0 self.pLinkData = 0 self.baseAddr = 0 self.OptHeader = OptHeader() self.dexHeader = DexHeader() def dump(self, addr): global baseAddr self.pOptHeader = getDword(addr) self.pHeader = getDword(addr + 4) self.pStringIds = getDword(addr + 8) self.pTypeIds = getDword(addr + 12) self.pFieldIds = getDword(addr + 16) self.pMethodIds = getDword(addr + 20) self.pProtoIds = getDword(addr + 24) self.pClassDefs = getDword(addr + 28) self.pLinkData = getDword(addr + 32) self.baseAddr = getDword(addr + 44) baseAddr = self.baseAddr self.OptHeader.dump(self.pOptHeader) self.dexHeader.dump(self.pHeader) self.fixDexHeader() def fixDexHeader(self): self.dexHeader.stringIdsOff = self.pStringIds - self.pHeader self.dexHeader.typeIdsOff = self.pTypeIds - self.pHeader self.dexHeader.fieldIdsOff = self.pFieldIds - self.pHeader self.dexHeader.methodIdsOff = self.pMethodIds - self.pHeader self.dexHeader.protoIdsOff = self.pProtoIds - self.pHeader self.dexHeader.classDefsOff = self.pClassDefs - self.pHeader if self.dexHeader.dataOff == 0: self.dexHeader.dataOff = self.dexHeader.classDefsOff + self.dexHeader.classDefsSize*32 # We should figure out a new method to fix the data size # self.dexHeader.dataSize = 0x5DD28000 - self.baseAddr - self.dexHeader.dataOff def lookupClass(self, type): num_class_def = self.dexHeader.classDefsSize print "num class def:", num_class_def for i in range(num_class_def): classdef = DexClassDef() classdef.dump(self.pClassDefs+32*i) descriptor = dexGetClassDescriptor(self, classdef) if descriptor == type: print "Find the class", descriptor if classdef.classDataOff == 0: print "classDataOff is 0" return classdata = ClassdataItem() classdata.dump(int(self.baseAddr+classdef.classDataOff) & 0xffffffff) print "direct methods:", classdata.direct_methods_size for j in range(classdata.direct_methods_size): method = classdata.direct_methods[j] method.printf() print "virtual methods:", classdata.virtual_methods_size for j in range(classdata.virtual_methods_size): method = classdata.virtual_methods[j] method.printf() def printf(self): print("dex head addr: ", hex(self.pHeader)) print("dex head addr: ", hex(self.baseAddr)) class DexClassDef: def __init__(self): self.classIdx = 0 self.accessFlags = 0 self.superclassIdx = 0 self.interfacesOff = 0 self.sourceFileIdx = 0 self.annotationsOff = 0 self.classDataOff = 0 self.staticValuesOff = 0 def dump(self, addr): self.classIdx = getDword(addr) self.accessFlags = getDword(addr + 4) self.superclassIdx = getDword(addr + 8) self.interfacesOff = getDword(addr + 12) self.sourceFileIdx = getDword(addr + 16) self.annotationsOff = getDword(addr + 20) self.classDataOff = getDword(addr + 24) self.staticValuesOff = getDword(addr + 28) def copytofile(self, file): file.write(struct.pack("I", self.classIdx)) file.write(struct.pack("I", self.accessFlags)) file.write(struct.pack("I", self.superclassIdx)) file.write(struct.pack("I", self.interfacesOff)) file.write(struct.pack("I", self.sourceFileIdx)) file.write(struct.pack("I", self.annotationsOff)) file.write(struct.pack("I", self.classDataOff)) file.write(struct.pack("I", self.staticValuesOff)) class DexHeader: def __init__(self): self.magic = [] self.checksum = 0 self.signature = [] self.fileSize = 0 self.headerSize = 0 self.endianTag = 0 self.linkSize = 0 self.linkOff = 0 self.mapOff = 0 self.stringIdsSize = 0 self.stringIdsOff = 0 self.typeIdsSize = 0 self.typeIdsOff = 0 self.protoIdsSize = 0 self.protoIdsOff = 0 self.fieldIdsSize = 0 self.fieldIdsOff = 0 self.methodIdsSize = 0 self.methodIdsOff = 0 self.classDefsSize = 0 self.classDefsOff = 0 self.dataSize = 0 # have it self.dataOff = 0 # have it def dump(self, addr): len = 0 while len < 8: self.magic.append(getByte(addr + len)) len += 1 self.checksum = getDword(addr + 8) len = 0 while len < 20: self.signature.append(getByte(addr + 12 + len)) len += 1 self.fileSize = getDword(addr + 32) self.headerSize = getDword(addr + 36) self.endianTag = getDword(addr + 40) self.linkSize = getDword(addr + 44) self.linkOff = getDword(addr + 48) self.mapOff = getDword(addr + 52) self.stringIdsSize = getDword(addr + 56) self.stringIdsOff = getDword(addr + 60) self.typeIdsSize = getDword(addr + 64) self.typeIdsOff = getDword(addr + 68) self.protoIdsSize = getDword(addr + 72) self.protoIdsOff = getDword(addr + 76) self.fieldIdsSize = getDword(addr + 80) self.fieldIdsOff = getDword(addr + 84) self.methodIdsSize = getDword(addr + 88) self.methodIdsOff = getDword(addr + 92) self.classDefsSize = getDword(addr + 96) self.classDefsOff = getDword(addr + 100) self.dataSize = getDword(addr + 104) self.dataOff = getDword(addr + 108) def printf(self): print "string off", self.stringIdsOff print "type off", self.typeIdsOff print "proto off", self.protoIdsOff print "field off", self.fieldIdsOff print "method off", self.methodIdsOff print "classdef off", self.classDefsOff print "classdef size:", self.classDefsSize def copytofile(self, file): len = 0 while len < 8: file.write(struct.pack("B", self.magic[len])) len += 1 file.write(struct.pack("I", self.checksum)) len = 0 while len < 20: file.write(struct.pack("B", self.signature[len])) len += 1 file.write(struct.pack("I", self.fileSize)) file.write(struct.pack("I", self.headerSize)) file.write(struct.pack("I", self.endianTag)) file.write(struct.pack("I", self.linkSize)) file.write(struct.pack("I", self.linkOff)) file.write(struct.pack("I", self.mapOff)) file.write(struct.pack("I", self.stringIdsSize)) file.write(struct.pack("I", self.stringIdsOff)) file.write(struct.pack("I", self.typeIdsSize)) file.write(struct.pack("I", self.typeIdsOff)) file.write(struct.pack("I", self.protoIdsSize)) file.write(struct.pack("I", self.protoIdsOff)) file.write(struct.pack("I", self.fieldIdsSize)) file.write(struct.pack("I", self.fieldIdsOff)) file.write(struct.pack("I", self.methodIdsSize)) file.write(struct.pack("I", self.methodIdsOff)) file.write(struct.pack("I", self.classDefsSize)) file.write(struct.pack("I", self.classDefsOff)) file.write(struct.pack("I", self.dataSize)) file.write(struct.pack("I", self.dataOff)) class OptHeader: def __init__(self): self.magic = [] # take 8 bytes self.dexoffset = 0 self.dexLength = 0 self.depsOffset = 0 self.depsLength = 0 self.optOffset = 0 self.optLength = 0 self.flag = 0 self.checksum = 0 def dump(self, addr): if addr == 0: return len = 0 while len < 8: self.magic.append(getByte(addr + len)) len += 1 self.dexoffset = getDword(addr+8) self.dexLength = getDword(addr+12) self.depsOffset = getDword(addr+16) self.depsLength = getDword(addr+20) self.optOffset = getDword(addr+24) self.optLength = getDword(addr+28) self.flag = getDword(addr+32) self.checksum = getDword(addr+36) def copytofile(self, file): len = 0 while len < 8: file.write(struct.pack("B", self.magic[len])) len += 1 file.write(struct.pack("I", self.dexoffset)) file.write(struct.pack("I", self.dexLength)) file.write(struct.pack("I", self.depsOffset)) file.write(struct.pack("I", self.depsLength)) file.write(struct.pack("I", self.optOffset)) file.write(struct.pack("I", self.optLength)) file.write(struct.pack("I", self.flag)) file.write(struct.pack("I", self.checksum)) class ClassdataItem: def __init__(self): self.len = 0 self.static_field_size = 0 self.instance_fields_size = 0 self.direct_methods_size = 0 self.virtual_methods_size = 0 self.static_fields = [] self.instance_fields = [] self.direct_methods = [] self.virtual_methods = [] def dump(self, addr): self.static_field_size, length = readunsignedleb128(addr) self.len += length self.instance_fields_size, length = readunsignedleb128(addr + self.len) self.len += length self.direct_methods_size, length = readunsignedleb128(addr + self.len) self.len += length self.virtual_methods_size, length = readunsignedleb128(addr + self.len) self.len += length for i in range(0, self.static_field_size): field = Encodedfield() field.dump(addr + self.len) self.len += field.len self.static_fields.append(field) for i in range(0, self.instance_fields_size): field = Encodedfield() field.dump(addr + self.len) self.len += field.len self.instance_fields.append(field) for i in range(0, self.direct_methods_size): method = Encodedmethod() method.dump(addr + self.len) self.len += method.len self.direct_methods.append(method) for i in range(0, self.virtual_methods_size): method = Encodedmethod() method.dump(addr + self.len) self.len += method.len self.virtual_methods.append(method) def recallLength(self): self.len = 0 self.len += unsignedleb128forlen(self.static_field_size) self.len += unsignedleb128forlen(self.instance_fields_size) self.len += unsignedleb128forlen(self.direct_methods_size) self.len += unsignedleb128forlen(self.virtual_methods_size) for i in range(0, self.static_field_size): self.len += self.static_fields[i].len for i in range(0, self.instance_fields_size): self.len += self.instance_fields[i].len for i in range(0, self.direct_methods_size): self.len += self.direct_methods[i].recallLength() for i in range(0, self.virtual_methods_size): self.len += self.virtual_methods[i].recallLength() return self.len def copytofile(self, file): writeunsignedleb128(self.static_field_size, file) writeunsignedleb128(self.instance_fields_size, file) writeunsignedleb128(self.direct_methods_size, file) writeunsignedleb128(self.virtual_methods_size, file) for i in range(0, self.static_field_size): self.static_fields[i].copytofile(file) for i in range(0, self.instance_fields_size): self.instance_fields[i].copytofile(file) for i in range(0, self.direct_methods_size): self.direct_methods[i].copytofile(file) for i in range(0, self.virtual_methods_size): self.virtual_methods[i].copytofile(file) class Encodedfield: def __init__(self): self.len = 0 self.field_idx_diff = 0 self.access_flags = 0 self.field_idx = 0 # need to set later def dump(self, addr): self.field_idx_diff, length = readunsignedleb128(addr) self.len += length self.access_flags, length = readunsignedleb128(addr + self.len) self.len += length def copytofile(self, file): writeunsignedleb128(self.field_idx_diff, file) writeunsignedleb128(self.access_flags, file) class Encodedmethod: def __init__(self): self.len = 0 self.method_idx_diff = 0 self.access_flags = 0 self.code_off = 0 self.method_idx = 0 def dump(self, addr): self.method_idx_diff, length = readunsignedleb128(addr) self.len += length self.access_flags, length = readunsignedleb128(addr + self.len) self.len += length self.code_off, length = readunsignedleb128(addr + self.len) self.len += length def recallLength(self): self.len = 0 self.len += unsignedleb128forlen(self.method_idx_diff) self.len += unsignedleb128forlen(self.access_flags) self.len += unsignedleb128forlen(self.code_off) return self.len def copytofile(self, file): writeunsignedleb128(self.method_idx_diff, file) writeunsignedleb128(self.access_flags, file) writeunsignedleb128(self.code_off, file) def printf(self): print "code offset:", self.code_off print "access flag:", self.access_flags # alignment: 4bytes class CodeItem: def __init__(self): self.len = 0 self.register_size = 0 self.ins_size = 0 self.outs_size = 0 self.tries_size = 0 self.debug_info_off = 0 self.insns_size = 0 self.insns = [] self.debugRef = None self.padding = 0 self.tries = [] self.handler = None def dump(self, addr): self.register_size = getWord(addr) # 2 self.ins_size = getWord(addr + 2) # 0 self.outs_size = getWord(addr + 4) # 0x4187 self.tries_size = getWord(addr + 6) # 0x13 self.debug_info_off = getDword(addr + 8) # 0xD self.insns_size = getDword(addr + 12) # 0x22 self.len += 16 for i in range(0, self.insns_size): self.insns.append(getWord(addr + self.len + 2 * i)) self.len += 2 * self.insns_size if self.tries_size != 0 and self.insns_size % 2 == 1: self.len += 2 for i in range(0, self.tries_size): tryitem = TryItem() tryitem.dump(addr + self.len + 8 * i) self.tries.append(tryitem) self.len += 8 * self.tries_size if self.tries_size != 0: self.handler = EncodedhandlerList() self.handler.dump(addr + self.len) self.len += self.handler.len # align = self.len % 4 # if align != 0: # self.len += (4 - align) class TryItem: def __init__(self): self.start = 0 self.len = 8 self.start_addr = 0 self.insn_count = 0 self.handler_off = 0 def dump(self, addr): self.start = addr self.start_addr = getDword(addr) self.insn_count = getWord(addr + 4) self.handler_off = getWord(addr + 6) class EncodedhandlerList: def __init__(self): self.start = 0 self.len = 0 self.size = 0 self.list = [] def dump(self, addr): self.start = addr self.size, length = readunsignedleb128(addr) self.len += length for i in range(0, self.size): handler = EncodedhandlerItem() handler.dump(addr + self.len) self.len += handler.len self.list.append(handler) class EncodedhandlerItem: def __init__(self): self.start = 0 self.len = 0 self.size = 0 self.handlers = [] self.catch_all_addr = 0 def dump(self, addr): self.start = addr self.size, length = readsignedleb128(addr) self.len += length for i in range(0, abs(self.size)): pair = EncodedTypeAddrPair() pair.dump(addr + self.len) self.len += pair.len self.handlers.append(pair) if self.size <= 0: self.catch_all_addr, length = readunsignedleb128(addr + self.len) self.len += length class EncodedTypeAddrPair: def __init__(self): self.type_idx = 0 self.addr = 0 self.len = 0 def dump(self, addr): self.type_idx, length = readunsignedleb128(addr) self.len += length self.addr, length = readunsignedleb128(addr + length) self.len += length address = int(0x5d4e8020) # DexFile address dexfile = DexFile() dexfile.dump(address) dexfile.dexHeader.printf() dexfile.lookupClass("Lcom/baidu/lbsapi/auth/LBSAuthManagerListener;") # dexfile.copytofile()
{ "repo_name": "CvvT/dumpDex", "path": "lookup.py", "copies": "1", "size": "23228", "license": "apache-2.0", "hash": 9203349476013620000, "line_mean": 32.3256814921, "line_max": 98, "alpha_frac": 0.562166351, "autogenerated": false, "ratio": 3.4824587706146928, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.45446251216146927, "avg_score": null, "num_lines": null }
__author__ = 'CwT' import Queue class FlowAnalysis(object): def __init__(self): self.GenSet = {} self.KillSet = {} self._graph = None def doAnalysis(self, directGraph): self._graph = directGraph numofcomputation = 0 queue = Queue.Queue() # queue.put(directGraph.Entry) for key, value in directGraph.blocks.items(): queue.put(value) while not queue.empty(): numofcomputation += 1 block = queue.get() # meet # print "Processing... ", block.Id self.meetFlow(block) # transfer function changed = self.flowthrough(block) if changed: for item in block.successors: queue.put(item) def meetFlow(self, block): IN = self.getInFlow(block) copied = False for item in IN: if not copied: self.copy(item, block) copied = True else: self.mergeInto(item, block) def flowthrough(self, block): # TO-DO Transfer return self.flowTransfer(block) def getInFlow(self, block): return block.predecessors if self.isForward else block.successors def getOutFlow(self, block): return block.successors if self.isForward else block.predecessors def isForward(self): return def mergeInto(self, sourceBlock, destBlock): return def copy(self, sourceBlock, destBlock): return
{ "repo_name": "Panalyzer/Panalyzer", "path": "src/FlowAnalysis.py", "copies": "1", "size": "1554", "license": "mit", "hash": -274345365934147400, "line_mean": 25.8103448276, "line_max": 73, "alpha_frac": 0.5566280566, "autogenerated": false, "ratio": 4.304709141274238, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5361337197874239, "avg_score": null, "num_lines": null }
__author__ = 'CwT' OPCODE = {0: "nop", 1: "move", 2: "move/from16", 5: "move-wide/from16", 7: "move-object", 8: "move-object/from16", 0xa: "move-result", 0xb: "move-result-wide", 0xc: "move-result-object", 0xd: "move-exception", 0xe: "return-void", 0xf: "return", 0x10: "return-wide", 0x11: "return-object", 0x12: "const/4", 0x13: "const/16", 0x1a: "const-string", 0x1c: "const-class", 0x1f: "check-cast", 0x22: "new-instance", 0x23: "new-array", 0x28: "goto", 0x32: "if-eq", 0x33: "if-ne", 0x34: "if-lt", 0x35: "if-ge", 0x36: "if-gt", 0x37: "if-le", 0x38: "if-eqz", 0x39: "if-nez", 0x3a: "if-ltz", 0x3b: "if-gez", 0x3c: "if-gtz", 0x3d: "if-lez", 0x44: "aget", 0x45: "aget-wide", 0x46: "aget-object", 0x47: "aget-boolean", 0x48: "aget-byte", 0x49: "aget-char", 0x4a: "aget-short", 0x4b: "aput", 0x4c: "aput-wide", 0x4d: "aput-object", 0x4e: "aput-boolean", 0x4f: "aput-byte", 0x50: "aput-char", 0x51: "aput-short", 0x52: "iget", 0x53: "iget-wide", 0x54: "iget-object", 0x55: "iget-boolean", 0x56: "iget-byte", 0x57: "iget-char", 0x58: "iget-short", 0x59: "iput", 0x5a: "iput-wide", 0x5b: "iput-object", 0x5c: "iput-boolean", 0x5d: "iput-byte", 0x5e: "iput-char", 0x5f: "iput-short", 0x60: "sget", 0x61: "sget-wide", 0x62: "sget-object", 0x63: "sget-boolean", 0x64: "sget-byte", 0x65: "sget-char", 0x66: "sget-short", 0x67: "sput", 0x68: "sput-wide", 0x69: "sput-object", 0x6a: "sput-boolean", 0x6b: "sput-byte", 0x6c: "sput-char", 0x6d: "sput-short", 0x6e: "invoke-virtual", 0x6f: "invoke-super", 0x70: "invoke-direct", 0x71: "invoke-static", 0x72: "invoke-interface", 0x74: "invoke-virtual/range", 0x75: "invoke-super/range", 0x76: "invoke-direct/range", 0x77: "invoke-static/range", 0x78: "invoke-interface/range", 0x7b: "neg-int", 0x7c: "not-int", 0x7d: "neg-long", 0x7e: "not-long", 0x7f: "neg-float", 0x80: "neg-double", 0x81: "int-to-long", 0x82: "int-to-float", 0x83: "int-to-double", 0x84: "long-to-int", 0x85: "long-to-float", 0x86: "long-to-double", 0x87: "float-to-int", 0x88: "float-to-long", 0x89: "float-to-double", 0x8a: "double-to-int", 0x8b: "double-to-long", 0x8c: "double-to-float", 0x8d: "int-to-byte", 0x8e: "int-to-char", 0x8f: "int-to-short", 0x90: "add-int", 0x91: "sub-int", 0x92: "mul-int", 0x93: "div-int", 0x94: "rem-int", 0x95: "and-int", 0x96: "or-int", 0x97: "xor-int", 0x98: "shl-int", 0x99: "shr-int", 0x9a: "ushr-int", 0x9b: "add-long", 0x9c: "sub-long", 0x9d: "mul-long", 0x9e: "div-long", 0x9f: "rem-long", 0xa0: "and-long", 0xa1: "or-long", 0xa2: "xor-long", 0xa3: "shl-long", 0xa4: "shr-long", 0xa5: "ushr-long", 0xa6: "add-float", 0xa7: "sub-float", 0xa8: "mul-float", 0xa9: "div-float", 0xaa: "rem-float", 0xab: "add-double", 0xac: "sub-double", 0xad: "mul-double", 0xae: "div-double", 0xaf: "rem-double", 0xb0: "add-int/2addr", 0xb1: "sub-int/2addr", 0xb2: "mul-int/2addr", 0xb3: "div-int/2addr", 0xb4: "rem-int/2addr", 0xb5: "and-int/2addr", 0xb6: "or-int/2addr", 0xb7: "xor-int/2addr", 0xb8: "shl-int/2addr", 0xb9: "shr-int/2addr", 0xba: "ushr-int/2addr", 0xbb: "add-long/2addr", 0xbc: "sub-long/2addr", 0xbd: "mul-long/2addr", 0xbe: "div-long/2addr", 0xbf: "rem-long/2addr", 0xc0: "and-long/2addr", 0xc1: "or-long/2addr", 0xc2: "xor-long/2addr", 0xc3: "shl-long/2addr", 0xc4: "shr-long/2addr", 0xc5: "ushr-long/2addr", 0xc6: "add-float/2addr", 0xc7: "sub-float/2addr", 0xc8: "mul-float/2addr", 0xc9: "div-float/2addr", 0xca: "rem-float/2addr", 0xcb: "add-double/2addr", 0xcc: "sub-double/2addr", 0xcd: "mul-double/2addr", 0xce: "div-double/2addr", 0xcf: "rem-double/2addr", 0xd0: "add-int/lit16", 0xd1: "rsub-int", 0xd2: "mul-int/lit16", 0xd3: "div-int/lit16", 0xd4: "rem-int/lit16", 0xd5: "and-int/lit16", 0xd6: "or-int/lit16", 0xd7: "xor-int/lit16", 0xd8: "add-int/lit8", 0xd9: "rsub-int/lit8", 0xda: "mul-int/lit8", 0xdb: "div-int/lit8", 0xdc: "rem-int/lit8", 0xdd: "and-int/lit8", 0xde: "or-int/lit8", 0xdf: "xor-int/lit8", 0xe0: "shl-int/lit8", 0xe1: "shl-int/lit8", 0xe2: "ushr-int/lit8", } class Instruction: def __init__(self): self.insns = [] self.str_ins = "" self.dst = -1 self.src = -1 self.target = -1 def init(self, param_insns, index): start = index one = param_insns[index] index += 1 opcode = one & 0xff one >>= 8 if opcode in (0, 0xe): self.str_ins = OPCODE[opcode] elif opcode == 1: index = self.parse(param_insns, index, one, opcode, 0) elif opcode == 2: index = self.parse(param_insns, index, one, opcode, 1) elif opcode == 5: index = self.parse(param_insns, index, one, opcode, 1) elif opcode == 7: index = self.parse(param_insns, index, one, opcode, 0) elif opcode == 8: index = self.parse(param_insns, index, one, opcode, 1) elif opcode in range(0xa, 0xe): index = self.parse(param_insns, index, one, opcode, 2) elif opcode in range(0xf, 0x12): index = self.parse(param_insns, index, one, opcode, 8) elif opcode == 0x12: index = self.parse(param_insns, index, one, opcode, 0) elif opcode == 0x13: index = self.parse(param_insns, index, one, opcode, 1) elif opcode == 0x1a: index = self.parse(param_insns, index, one, opcode, 1) elif opcode == 0x1c: index = self.parse(param_insns, index, one, opcode, 1) elif opcode == 0x1f: index = self.parse(param_insns, index, one, opcode, 1) elif opcode == 0x22: index = self.parse(param_insns, index, one, opcode, 1) elif opcode == 0x23: index = self.parse(param_insns, index, one, opcode, 4) elif opcode == 0x28: index = self.parse(param_insns, index, one, opcode, 2) elif opcode in range(0x32, 0x38): index = self.parse(param_insns, index, one, opcode, 4) elif opcode in range(0x38, 0x3e): index = self.parse(param_insns, index, one, opcode, 1) elif opcode in range(0x44, 0x52): index = self.parse(param_insns, index, one, opcode, 3) elif opcode in range(0x52, 0x60): index = self.parse(param_insns, index, one, opcode, 4) elif opcode in range(0x60, 0x6e): index = self.parse(param_insns, index, one, opcode, 1) elif opcode in range(0x6e, 0x73): index = self.parse(param_insns, index, one, opcode, 5) elif opcode in range(0x7b, 0x90): index = self.parse(param_insns, index, one, opcode, 0) elif opcode in range(0x90, 0xb0): index = self.parse(param_insns, index, one, opcode, 7) elif opcode in range(0xb0, 0xd0): index = self.parse(param_insns, index, one, opcode, 0) elif opcode in range(0xd0, 0xd8): index = self.parse(param_insns, index, one, opcode, 4) elif opcode in range(0xd8, 0xe3): index = self.parse(param_insns, index, one, opcode, 7) else: print("ERROR: didn't classified the opcode", opcode) for i in range(start, index): self.insns.append(param_insns[i]) return index def parse(self, param_insns, index, one, opcode, kind): if kind == 0: # opcode vx, vy self.src = (one >> 4) & 0xf self.dst = one & 0xf self.str_ins = OPCODE[opcode] + " " + str(self.dst) + ", " + str(self.src) elif kind == 1: # opcode vx/vxx, vyy(id) self.dst = one self.src = int(param_insns[index]) index += 1 self.str_ins = OPCODE[opcode] + " " + str(self.dst) + ", " + str(self.src) elif kind == 2: # opcode vx/vxx self.dst = one self.str_ins = OPCODE[opcode] + " " + str(self.dst) elif kind == 3: # opcode vx, vy, vz self.dst = one one = int(param_insns[index]) self.src = one & 0xff self.target = (one >> 8) & 0xff self.str_ins = OPCODE[opcode] + " " + str(self.dst) + ", " + str(self.src) + ", " + str(self.target) index += 1 elif kind == 4: # opcode vx, vy, id(lit) self.dst = one & 0xf self.src = (one >> 4) & 0xf self.target = int(param_insns[index]) index += 1 self.str_ins = OPCODE[opcode] + " " + str(self.dst) + ", " + str(self.src) + ", " + str(self.target) elif kind == 5: # invoke-kind {vC, vD, vE, vF, vG}, meth@BBBB self.src = (one >> 4) & 0xf # use for count parameter self.target = int(param_insns[index]) param = int(param_insns[index+1]) index += 2 self.str_ins = OPCODE[opcode] + " {" if self.src > 0: self.str_ins += str(param & 0xf) if self.src > 1: self.str_ins += ", " + str((param >> 4) & 0xf) if self.src > 2: self.str_ins += ", " + str((param >> 8) & 0xf) if self.src > 3: self.str_ins += ", " + str((param >> 12) & 0xf) if self.src > 4: self.str_ins += ", " + str(one & 0xf) self.str_ins += "}, " + str(self.target) elif kind == 6: # invoke-kind/range {vCCCC .. vNNNN}, meth@BBBB self.src = one # use for count parameter self.target = int(param_insns[index]) param = int(param_insns[index+1]) index += 2 self.str_ins = OPCODE[opcode] + " {" if self.src > 0: self.str_ins += str(param) + " ... " + str(param + self.src - 1) self.str_ins += "}, " + str(self.target) elif kind == 7: # opcode vXX, vYY, vZZ self.dst = one self.src = param_insns[index] & 0xff self.target = (param_insns[index] >> 8) & 0xff self.str_ins = OPCODE[opcode] + " " + str(self.dst) + ", " + str(self.src) + ", " + str(self.target) elif kind == 8: # opcode vxx self.dst = one self.str_ins = OPCODE[opcode] + " " + str(self.dst) else: print("ERROR: didn't classified this kind") return index # notice that: each instruction store as short(little) class InstructionSet: def __init__(self, insns): length = len(insns) index = 0 self.set = [] while index < length: tmp = Instruction() index = tmp.init(insns, index) self.set.append(tmp) print(tmp.str_ins, hex(tmp.insns[0] & 0xff), len(tmp.insns)) def printf(self): length = len(self.set) for i in range(0, length): print(self.set[i].str_ins)
{ "repo_name": "CvvT/DexParse", "path": "Instruction.py", "copies": "2", "size": "11175", "license": "apache-2.0", "hash": -876353120084651600, "line_mean": 54.5970149254, "line_max": 132, "alpha_frac": 0.5356599553, "autogenerated": false, "ratio": 2.753141167775314, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9276713716786559, "avg_score": 0.002417481257750962, "num_lines": 201 }
__author__ = 'cyankw' import re import urllib.request from multiprocessing import Pool import time strtime = time.clock() #计时器启动 def getHtml (url): #用于读取网页源代码 page = urllib.request.urlopen(url) html = page.read() return html global year global keywd def getCVE(html): #对读取的网页源代码使用正则表达式进行筛选 reg = r'/\bCVE-%s-\b\d{1,4}'%(year) CVEs = re.compile(reg) CVElist = re.findall(CVEs,html) Xmax = len(CVElist) x=0 CVElist2=[] while x < Xmax: #去除每个被正则表达式匹配出的结果的第一个字符“/” CVElist2.append(CVElist[x][1:]) x=x+1 return CVElist2 def getKWD(html): #关键词词频统计,统计网页中keywd出现次数,(每出现一次,就记录一次) reg = r'/\b<em>\b%s\b</em>\b'%(keywd) KWDs = re.compile(reg) KWDlist = re.findall(KWDs,html) #print(KWDlist) return KWDlist if __name__=='__main__': i=1 getCVE2=[] print('Vulnerabilities Detector Started') year = input('Year ') times = int(input('Result Number ')) poolnum = int(input('Pool Numbers ')) print('clawler started...') p = Pool(poolnum) #使用进程池调用的进程数量 rangmax = times m=0 gd=0 numb = 0 while numb<times: html = getHtml("http://cve.scap.org.cn/cve_list.php?action=cvss&floor=9.5&ceil=10&p=%u"%(i)).decode('utf-8') #从指定网页读取源代码 i = i + 1 getCVE2 = list(set(getCVE(html))) #使用getCVE()函数进行数据筛选,去除重复项目,并记录结果到列表getCVE2中 #print(getCVE2) f=0 ln=len(getCVE2) while f<ln : if numb == times: break getKY = [] keywd = getCVE2[f] #从二级列表getCVE2中获取到的次级元素中迭代当前元素(列表)中的子元素 html = getHtml("http://www.baidu.com/baidu?wd=%s" % (keywd)).decode('utf-8') #将指定关键词进行百度搜索 getKY.append((getKWD(html))) #使用getKWD()进行词频统计 url = "http://www.baidu.com/baidu?wd=%s" %(keywd)#用于打印固定文本 #print(getKY) if len(getKY) >= 5: #对词频结果给出其价值 GG='EXCELLENT!!!! %s'%(url) gd = gd + 1 elif len(getKY) >= 3: GG='GOOD!!! %s'%(url) elif len(getKY) > 0: GG='simple' elif len(getKY) == 0: GG='-----' print(numb,keywd,len(getKY),GG), f=f+1 numb=numb+1 m = m + 1 p.close() p.join() print('------------------------------------') print('Vulnerabilities Detect Completed') fintime = time.clock() #计时器结束 print('Program running time %fs'%(fintime-strtime)) print('We detected %u targets through %u websites, %u of them is valuable'%(numb,m,gd))
{ "repo_name": "cyankw/Python-Crawler-Practice-01-Vulnerabilities-Detector", "path": "Vulnerabilities Detector.py", "copies": "1", "size": "3089", "license": "apache-2.0", "hash": 4426454802812552000, "line_mean": 27.8988764045, "line_max": 128, "alpha_frac": 0.5302745393, "autogenerated": false, "ratio": 2.3468667255075024, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.3377141264807503, "avg_score": null, "num_lines": null }
__author__ = 'cyhis' from suds.client import Client from suds import WebFault from model.project import Project class SoapHelper: def __init__(self, app): self.app = app self.username = self.app.config['webadmin']['username'] self.password = self.app.config['webadmin']['password'] def can_login(self, username, password): client = Client("http://localhost/mantisbt-1.2.19/api/soap/mantisconnect.php?wsdl") try: client.service.mc_login(username, password) return True except WebFault: return False def create(self, project): client = Client("http://localhost/mantisbt-1.2.19/api/soap/mantisconnect.php?wsdl") try: pData = client.factory.create('ProjectData') pData.name = project.name client.service.mc_project_add(self.username, self.password, pData) return True except WebFault: return False def get_projects(self): client = Client("http://localhost/mantisbt-1.2.19/api/soap/mantisconnect.php?wsdl") projects = [] try: projectsData = client.service.mc_projects_get_user_accessible(self.username, self.password) for projectData in projectsData: project = Project(name=projectData['name']) projects.append(project) return projects except WebFault: return False def delete(self, project): client = Client("http://localhost/mantisbt-1.2.19/api/soap/mantisconnect.php?wsdl") try: projectsData = client.service.mc_projects_get_user_accessible(self.username, self.password) for projectData in projectsData: if projectData['name'] == project.name: client.service.mc_project_delete(self.username, self.password, projectData['id']) return True except WebFault: return False
{ "repo_name": "galaktika81/python_training_mantis", "path": "fixture/soap.py", "copies": "1", "size": "1991", "license": "apache-2.0", "hash": -2530711195864215000, "line_mean": 34.5714285714, "line_max": 103, "alpha_frac": 0.611752888, "autogenerated": false, "ratio": 4.08829568788501, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.520004857588501, "avg_score": null, "num_lines": null }
__author__ = 'cyrbuzz' from base import (ScrollArea, QLabel, QFrame, QVBoxLayout, QPushButton, QHBoxLayout, QTableWidget, QAbstractItemView) class DownloadFrame(ScrollArea): def __init__(self, parent=None): super().__init__() self.parent = parent self.setObjectName('downloadMusic') with open('QSS/downloadFrame.qss', 'r', encoding="utf-8") as f: self.setStyleSheet(f.read()) self.mainLayout = QVBoxLayout(self) self.setHeader() self.setMusicTable() def setHeader(self): # self.titleLabel = QLabel("我的下载") self.spaceLine = QFrame(self) self.spaceLine.setObjectName("spaceLine") self.spaceLine.setFrameShape(QFrame.HLine) self.spaceLine.setFrameShadow(QFrame.Plain) self.spaceLine.setLineWidth(2) self.currentStorageFolderLabel = QLabel("当前存储目录: ") self.currentStorageFolder = QLabel() self.selectButton = QPushButton("选择目录") self.selectButton.setObjectName('selectButton') self.topShowLayout = QHBoxLayout() self.topShowLayout.addSpacing(20) # self.topShowLayout.addWidget(self.titleLabel) self.topShowLayout.addWidget(self.currentStorageFolderLabel) self.topShowLayout.addWidget(self.currentStorageFolder) self.topShowLayout.addWidget(self.selectButton) self.topShowLayout.addStretch(1) self.mainLayout.addLayout(self.topShowLayout) self.mainLayout.addWidget(self.spaceLine) def setMusicTable(self): self.singsTable = QTableWidget() self.singsTable.setObjectName('singsTable') self.singsTable.setMinimumWidth(self.width()) self.singsTable.setColumnCount(3) self.singsTable.setHorizontalHeaderLabels(['音乐标题', '歌手', '时长']) self.singsTable.setColumnWidth(0, self.width()/3*1.25) self.singsTable.setColumnWidth(1, self.width()/3*1.25) self.singsTable.setColumnWidth(2, self.width()/3*0.5) self.singsTable.horizontalHeader().setStretchLastSection(True) self.singsTable.verticalHeader().setVisible(False) self.singsTable.setShowGrid(False) self.singsTable.setAlternatingRowColors(True) self.singsTable.setEditTriggers(QAbstractItemView.NoEditTriggers) self.singsTable.setSelectionBehavior(QAbstractItemView.SelectRows) self.mainLayout.addWidget(self.singsTable)
{ "repo_name": "HuberTRoy/MusicPlayer", "path": "MusicPlayer/widgets/downloadFrame.py", "copies": "1", "size": "2503", "license": "mit", "hash": 7190364056942079000, "line_mean": 36.2727272727, "line_max": 98, "alpha_frac": 0.6823912159, "autogenerated": false, "ratio": 3.548340548340548, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9711616367852425, "avg_score": 0.0038230792776247323, "num_lines": 66 }
__author__ = 'cyrbuzz' import addition from base import QObject, QTableWidgetItem transTime = addition.itv2time class ConfigRecommendFrame(QObject): def __init__(self, parent): super().__init__() self.recommendFrame = parent self.musicList = [] self.bindConnect() def setSongs(self, musicInfo): self.recommendFrame.singsTable.setRowCount(len(musicInfo)) for index, data in enumerate(musicInfo): self.musicList.append(data._asdict()) self.recommendFrame.singsTable.setItem(index, 0, QTableWidgetItem(data.name)) self.recommendFrame.singsTable.setItem(index, 1, QTableWidgetItem(data.author)) self.recommendFrame.singsTable.setItem(index, 2, QTableWidgetItem(transTime(data.time/1000))) def bindConnect(self): self.recommendFrame.singsTable.itemDoubleClicked.connect(self.itemDoubleClickedEvent) def itemDoubleClickedEvent(self): currentRow = self.recommendFrame.singsTable.currentRow() data = self.musicList[currentRow] self.recommendFrame.parent.playWidgets.setPlayerAndPlayList(data)
{ "repo_name": "HuberTRoy/MusicPlayer", "path": "MusicPlayer/features/configRecommendFrameFeatures.py", "copies": "1", "size": "1143", "license": "mit", "hash": 2782863357357006300, "line_mean": 29.8918918919, "line_max": 105, "alpha_frac": 0.7034120735, "autogenerated": false, "ratio": 3.7475409836065574, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.49509530571065574, "avg_score": null, "num_lines": null }
__author__ = 'cyrbuzz' import os import glob import pickle import os.path try: import eyed3 except ImportError: print('eyed3没有成功加载或安装,请不要使用本地音乐功能!') from base import QFileDialog, QObject, QTableWidgetItem, checkFolder from addition import itv2time def getAllFolder(topFolder): result = [] def findFolder(topFolder): folders = [os.path.join(topFolder, i) for i in os.listdir(topFolder) if not os.path.isfile(os.path.join(topFolder, i))] if not folders: return else: result.extend(folders) for i in folders: findFolder(i) findFolder(topFolder) return result class ConfigNative(QObject): loadLocalFolder = 'cookies/native/local.cks' allCookiesFolder = [loadLocalFolder] def __init__(self, native): super(ConfigNative, self).__init__() self.native = native self.musicList = [] self.folder = [] self.bindConnect() self.loadCookies() def bindConnect(self): self.native.selectButton.clicked.connect(self.selectFolder) self.native.singsTable.itemDoubleClicked.connect(self.itemDoubleClickedEvent) def selectFolder(self): folder = QFileDialog() selectFolder = folder.getExistingDirectory() if not selectFolder: pass else: self.folder.append(selectFolder) self.loadMusic() def loadMusic(self): for folder in self.folder: mediaFiles = glob.glob(folder+'/*.mp3') allFolder = getAllFolder(folder) for i in allFolder: mediaFiles.extend(glob.glob(i+'/*.mp3')) length = len(mediaFiles) self.native.singsTable.clearContents() self.native.singsTable.setRowCount(length) self.musicList = [] for i in enumerate(mediaFiles): music = eyed3.load(i[1]) if not music: self.singsTable.removeRow(i[0]) continue try: name = music.tag.title author = music.tag.artist if not name: filePath = i[1].replace(folder, '') name = filePath[1:][:-4] if not author: author = '' except: try: # TODO # if more folders exist. filePath = i[1].replace(folder, '') name = filePath[1:][:-4] except Exception as e: name = i[1] author = '' try: time = itv2time(music.info.time_secs) except: time = '00:00' self.musicList.append({'name': name, 'author': author, 'time': time, 'url': i[1], 'music_img': 'None'}) self.native.singsTable.setItem(i[0], 0, QTableWidgetItem(name)) self.native.singsTable.setItem(i[0], 1, QTableWidgetItem(author)) self.native.singsTable.setItem(i[0], 2, QTableWidgetItem(time)) # 事件。 def itemDoubleClickedEvent(self): currentRow = self.native.singsTable.currentRow() data = self.musicList[currentRow] self.native.parent.playWidgets.setPlayerAndPlayList(data) @checkFolder(allCookiesFolder) def saveCookies(self): with open(self.loadLocalFolder, 'wb') as f: pickle.dump(self.folder, f) @checkFolder(allCookiesFolder) def loadCookies(self): with open(self.loadLocalFolder, 'rb') as f: self.folder = pickle.load(f) self.loadMusic()
{ "repo_name": "HuberTRoy/MusicPlayer", "path": "MusicPlayer/features/configNativeFeatures.py", "copies": "1", "size": "3832", "license": "mit", "hash": 8029725500049695000, "line_mean": 29.0158730159, "line_max": 127, "alpha_frac": 0.5446853517, "autogenerated": false, "ratio": 4.002116402116402, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5046801753816402, "avg_score": null, "num_lines": null }
__author__ = 'cyrbuzz' import os import re import glob import pickle import os.path import logging try: import eyed3 except ImportError: print('eyed3没有成功加载或安装,当再次打开时下载的音乐会加载不到!') from apiRequestsBase import HttpRequest from asyncBase import aAsync, toTask from base import QFileDialog, QObject, QTableWidgetItem, checkFolder from addition import itv2time import logger logger = logging.getLogger(__name__) myRequests = HttpRequest() def getAllFolder(topFolder): result = [] def findFolder(topFolder): folders = [os.path.join(topFolder, i) for i in os.listdir(topFolder) if not os.path.isfile(os.path.join(topFolder, i))] if not folders: return else: result.extend(folders) for i in folders: findFolder(i) findFolder(topFolder) return result def replace_forbidden_sym(string): return re.sub(r'[\\/:*?"<>|]{1}', ' ', string) class ConfigDownloadFrame(QObject): myDownloadFrameCookiesFolder = 'cookies/downloadInfo/downloadFolder.cks' allCookiesFolder = [myDownloadFrameCookiesFolder] def __init__(self, downloadFrame): super(ConfigDownloadFrame, self).__init__() self.downloadFrame = downloadFrame self.showTable = self.downloadFrame.singsTable self.musicList = [] self.folder = [] self.myDownloadFolder = os.path.join(os.getcwd(), 'downloads') self._setDownloadFolder(self.myDownloadFolder) self.bindConnect() self.loadCookies() def bindConnect(self): self.downloadFrame.selectButton.clicked.connect(self.selectFolder) self.downloadFrame.singsTable.itemDoubleClicked.connect(self.itemDoubleClickedEvent) def getDownloadSignal(self): # window = self.downloadFrame.parent try: window.searchArea.config.download.connect(self.downloadSong) window.detailSings.config.download.connect(self.downloadSong) except Exception as e: logger.error("下载时遇到未知错误", exc_info=True) def _setDownloadFolder(self, folderName): logger.info("下载目标变更{}".format(folderName)) self.fromPathLoadSong(folderName) self.myDownloadFolder = folderName self.downloadFrame.currentStorageFolder.setText(folderName) @toTask def downloadSong(self, musicInfo): logger.info("正在下载的音乐的信息: {}".format(musicInfo)) url = musicInfo.get('url') allMusicName = re.search(r'.*\.[a-zA-Z0-9]+', url[url.rfind('/')+1:]).group(0) if allMusicName: musicSuffix = allMusicName[allMusicName.rfind('.')+1:] musicName = '{name}.{suf}'.format(name=musicInfo.get('name') + ' - ' + musicInfo.get('author'), suf=musicSuffix) else: # TODO MD5。 musicName = "random_name.mp3" musicName = replace_forbidden_sym(musicName) self.downloadFrame.parent.systemTray.showMessage("~~~", '{musicName} 加入下载队列'.format(musicName=musicName)) # TODO # Streaming. future = aAsync(myRequests.httpRequest, url, 'GET') data = yield from future localPath = '{myDownloadFolder}/{musicName}'.format(myDownloadFolder=self.myDownloadFolder, musicName=musicName) with open(localPath, 'wb') as f: f.write(data.content) musicInfo['url'] = localPath # 从托盘栏给出提示。 self.downloadFrame.parent.systemTray.showMessage("~~~", '{musicName} 下载完成'.format(musicName=musicName)) # 将音乐信息加到musicList中。 self.musicList.append(musicInfo) self.updateDownloadShowTable(musicInfo) def updateDownloadShowTable(self, musicInfo): showInfo = [musicInfo.get("name"), musicInfo.get("author"), musicInfo.get("time")] # 这里写"我的下载"的实例对象。 # 首先获取出当前总共多少行。 rowCount = self.showTable.rowCount() self.showTable.setRowCount(rowCount+1) # 然后直接添加过去就好啦。 for i in range(3): self.showTable.setItem(rowCount, i, QTableWidgetItem(showInfo[i])) def fromPathLoadSong(self, selectFolder): if not os.path.isdir(selectFolder): os.mkdir(selectFolder) return mediaFiles = glob.glob(selectFolder+'/*.mp3') allFolder = getAllFolder(selectFolder) for i in allFolder: mediaFiles.extend(glob.glob(i+'/*.mp3')) length = len(mediaFiles) self.downloadFrame.singsTable.clearContents() self.downloadFrame.singsTable.setRowCount(length) self.musicList = [] for i in enumerate(mediaFiles): music = eyed3.load(i[1]) if not music: self.singsTable.removeRow(i[0]) continue try: name = music.tag.title author = music.tag.artist if not name: filePath = i[1].replace(selectFolder, '') name = filePath[1:][:-4] if not author: author = '' except: try: # TODO # if more folders exist. filePath = i[1].replace(selectFolder, '') name = filePath[1:][:-4] except Exception as e: name = i[1] author = '' try: time = itv2time(music.info.time_secs) except: time = '00:00' self.musicList.append({'name': name, 'author': author, 'time': time, 'url': i[1], 'music_img': 'None'}) self.downloadFrame.singsTable.setItem(i[0], 0, QTableWidgetItem(name)) self.downloadFrame.singsTable.setItem(i[0], 1, QTableWidgetItem(author)) self.downloadFrame.singsTable.setItem(i[0], 2, QTableWidgetItem(time)) def selectFolder(self): folder = QFileDialog() selectFolder = folder.getExistingDirectory() if not selectFolder: pass else: self.folder.append(selectFolder) self._setDownloadFolder(selectFolder) self.fromPathLoadSong(selectFolder) @checkFolder(allCookiesFolder) def saveCookies(self): with open(self.myDownloadFrameCookiesFolder, 'wb') as f: pickle.dump(self.myDownloadFolder, f) @checkFolder(allCookiesFolder) def loadCookies(self): with open(self.myDownloadFrameCookiesFolder, 'rb') as f: self.myDownloadFolder = pickle.load(f) self._setDownloadFolder(self.myDownloadFolder) # 事件。 def itemDoubleClickedEvent(self): currentRow = self.downloadFrame.singsTable.currentRow() data = self.musicList[currentRow] self.downloadFrame.parent.playWidgets.setPlayerAndPlayList(data)
{ "repo_name": "HuberTRoy/MusicPlayer", "path": "MusicPlayer/features/configDownloadFrameFeatures.py", "copies": "1", "size": "7055", "license": "mit", "hash": -2431072981149056500, "line_mean": 31.75, "line_max": 127, "alpha_frac": 0.6148876817, "autogenerated": false, "ratio": 3.5979926043317487, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9698294776333933, "avg_score": 0.0029171019395631613, "num_lines": 208 }
__author__ = 'cyrbuzz' """ 提供用于将函数包装成异步的基本功能。 """ import asyncio def aAsync(func, *args, **kwargs): """ future 用于将任意函数包装成一个awaitable对象。 例: future = aAsync(requests.get, 'http://www.xxx.com', headers=headers) data = yield from future """ # run_in_evecutor不支持**kwargs. # args是个元组,kwargs是个字典。 # 再用到时 run_in_evecutor(None, makeUp, args, kwargs) # 注意这边不要带*,带*为解包,不要解包。 def makeUp(args, kwargs): return func(*args, **kwargs) eventLoop = asyncio.get_event_loop() future = eventLoop.run_in_executor(None, makeUp, args, kwargs) return future def toTask(func): """ 一个将普通函数包装成异步函数的装饰器。 例: @toTask def test(x): future = aAsync(requests.get, 'http://www.xxx.com', headers=headers) print(x) data = yield from future print(data) 在运行test时(test())就会变成一个异步函数, 里面I/O部分就会由asyncio提供的事件循环处理。 >>> for i in range(5): test(x) ``` 0 1 2 3 4 <Response [200]> <Response [200]> <Response [200]> <Response [200]> <Response [200]> ``` """ def makeUp(*args, **kwargs): eventLoop = asyncio.get_event_loop() future = eventLoop.create_task(func(*args, **kwargs)) return future return makeUp def toTaskWCb(func): """ 一个将普通函数包装成异步函数并添加回调的装饰器。 """ def makeUp(callback): def makeUps(*args, **kwargs): eventLoop = asyncio.get_event_loop() future = eventLoop.create_task(func(*args, **kwargs)) future.add_done_callback(callback) return future return makeUps return makeUp if __name__ == '__main__': help(aAsync) print('\n') help(toTask)
{ "repo_name": "HuberTRoy/MusicPlayer", "path": "MusicPlayer/features/asyncBase.py", "copies": "1", "size": "2150", "license": "mit", "hash": 4915872713588908000, "line_mean": 19.6333333333, "line_max": 80, "alpha_frac": 0.5350215517, "autogenerated": false, "ratio": 2.662840746054519, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.3697862297754519, "avg_score": null, "num_lines": null }
__author__ = 'cyrbuzz' """制作登陆区盒子。""" from base import QDialog, QFrame, HBoxLayout, HStretchBox, QLabel, QLineEdit, QPushButton, Qt, VBoxLayout, RequestThread class InputLine(QLineEdit): def __init__(self, parent=None, width=0, height=0, placeholderText=None): super(QLineEdit, self).__init__() self.parent = parent if width: self.setMaximumWidth(width) self.setMinimumWidth(width) if height: self.setMaximumHeight(height) self.setMinimumHeight(height) if placeholderText: self.setPlaceholderText(placeholderText) class Header(QFrame): myStyle = """ QFrame {background: #2D2D2D;} QLabel { margin-left: 8px; color: white; font-weight: bold; font-size: 15px; } QPushButton { border: none; font: bold; font-size: 13px; color: #7C7C7C; margin-right: 8px; } QPushButton:hover{ color: #DCDDE4; } """ def __init__(self, title:str, parent=None): super(Header, self).__init__() self.parent = None self.setStyleSheet(self.myStyle) self.mainLayout = HBoxLayout(self) self.title = QLabel(title) self.mainLayout.addWidget(self.title) self.mainLayout.addStretch(1) self.closeButton = QPushButton('×') self.mainLayout.addWidget(self.closeButton) def connectCloseButton(self, functionName): self.closeButton.clicked.connect(functionName) class LoginBox(QDialog): def __init__(self, parent=None): super(LoginBox, self).__init__() self.parent = parent self.setWindowFlags(Qt.FramelessWindowHint) self.setWindowTitle('登陆') self.setObjectName('LoginBox') self.resize(520, 300) # 可能会有多个渠道登陆的后续扩展。 self.currentFrame = 0 self.mainLayout = VBoxLayout(self) self.phoneAndEMailFrame = PhoneAndEMailFrame(self) self.mainLayout.addWidget(self.phoneAndEMailFrame) def setWarningAndShowIt(self, warningStr): if not self.currentFrame: self.phoneAndEMailFrame.setWarningAndShowIt(warningStr) def connectLogin(self, functionName): if not self.currentFrame: self.phoneAndEMailFrame.connectLogin(functionName) def checkAndGetLoginInformation(self): if not self.currentFrame: return self.phoneAndEMailFrame.checkAndGetLoginInformation() class PhoneAndEMailFrame(QFrame): def __init__(self, parent=None): super(PhoneAndEMailFrame, self).__init__() self.parent = parent self.resize(520, 300) with open('QSS/phoneAndEMailFrame.qss', 'r') as f: self.setStyleSheet(f.read()) self.mainLayout = VBoxLayout(self) self.header = Header("用户名", self) self.header.setMinimumHeight(40) self.header.connectCloseButton(self.parent.accept) self.mainLayout.addWidget(self.header) self.mainLayout.addStretch(1) self.usernameLine = InputLine(self, 220, 32, '请输入用户名') self.usernameLine.setObjectName('usernameLine') self.usernameCenterBox = HStretchBox(self.mainLayout, self.usernameLine) self.mainLayout.addSpacing(10) self.passwordLine = InputLine(self, 220, 32, '请输入密码') self.passwordLine.setObjectName('passwordLine') self.passwordCenterBox = HStretchBox(self.mainLayout, self.passwordLine) self.passwordLine.setEchoMode(QLineEdit.Password) self.warningIconLabel = QLabel() self.warningIconLabel.setObjectName('warningIconLabel') self.warningIconLabel.setMaximumSize(14, 14) self.warningIconLabel.setMinimumSize(14, 14) self.warningIconLabel.hide() self.warningLabel = QLabel("请输入用户名") self.warningLabel.hide() self.warningLabel.setObjectName('warningLabel') self.warningCenterBox = HStretchBox(self.mainLayout, self.warningIconLabel, self.warningLabel, behindStretch=2) self.mainLayout.addSpacing(30) self.enterLoginButton = QPushButton("登 录") self.enterLoginButton.setObjectName("enterButton") self.enterLoginButton.setMaximumSize(217, 27) self.enterLoginButton.setMinimumSize(217, 27) self.enterLoginCenterBox = HStretchBox(self.mainLayout, self.enterLoginButton) self.mainLayout.addSpacing(30) self.mainLayout.addStretch(1) def checkAndGetLoginInformation(self): username = self.usernameLine.text() password = self.passwordLine.text() if not username or not password: self.warningIconLabel.show() if not username: self.warningLabel.setText('请输入用户名') self.warningLabel.show() return False if not password: self.warningLabel.setText('请输入密码') self.warningLabel.show() return False self.warningIconLabel.hide() self.warningLabel.hide() return username, password def setWarningAndShowIt(self, warningStr): self.warningLabel.setText(warningStr) self.warningLabel.show() self.warningIconLabel.show() def connectLogin(self, functionName): self.enterLoginButton.clicked.connect(functionName) if __name__ == '__main__': import os os.chdir('..') app = QApplication([]) main = LoginBox() main.show() app.exec_()
{ "repo_name": "HuberTRoy/MusicPlayer", "path": "MusicPlayer/widgets/loginFrames.py", "copies": "1", "size": "5711", "license": "mit", "hash": 4464577960630639600, "line_mean": 27.1055276382, "line_max": 120, "alpha_frac": 0.6348354793, "autogenerated": false, "ratio": 3.840659340659341, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9937374588101966, "avg_score": 0.007624046371475085, "num_lines": 199 }
__author__ = 'cyrbuzz' # 方便扩展,抽象成基类。 from base import (QCursor, QFrame, Qt, QTabWidget, QTextEdit, QLabel, QIcon, QPushButton, QHBoxLayout, QVBoxLayout, QGridLayout, QTableWidgetItem, PicLabel, ScrollArea, TableWidget, VBoxLayout, HBoxLayout, pyqtSignal) import addition class SingsFrameBase(ScrollArea): """全部歌单。""" def __init__(self, parent=None): super(SingsFrameBase, self).__init__() self.parent = parent self.transTime = addition.itv2time self.setObjectName("allSingsArea") # 为什么有的需要加utf-8呢,因为有中文。 with open('QSS/singsFrameBase.qss', 'r', encoding='utf-8') as f: self.setStyleSheet(f.read()) # 主布局。 self.mainLayout = QGridLayout(self.frame) self.mainLayout.setSpacing(0) self.mainLayout.setHorizontalSpacing(10) self.mainLayout.setContentsMargins(0, 0, 0, 0) class SingsSearchResultFrameBase(QFrame): def __init__(self, parent): super(SingsSearchResultFrameBase, self).__init__() self.parent = parent self.singsFrameLayout = VBoxLayout(self) self.noSingsContentsLabel = QLabel(self) self.noSingsContentsLabel.setMaximumHeight(60) self.noSingsContentsLabel.setObjectName("noSingsLable") self.noSingsContentsLabel.hide() self.singsResultTable = TableWidget(3, ['音乐标题', '歌手', '时长']) self.singsResultTable.setObjectName('singsTable') self.singsResultTable.setMinimumWidth(self.parent.width()) self.singsResultTable.setColumnWidths({i:j for i,j in zip(range(3), [self.parent.width()/3*1.25,self.parent.width()/3*1.25,self.parent.width()/3*0.5])}) self.singsFrameLayout.addWidget(self.singsResultTable, Qt.AlignTop|Qt.AlignCenter) self.centerLabelLayout = HBoxLayout() self.centerLabelLayout.addStretch(1) self.centerLabelLayout.addWidget(self.noSingsContentsLabel) self.centerLabelLayout.addStretch(1) self.singsFrameLayout.addLayout(self.centerLabelLayout) # 歌单详情页。 class DetailSings(ScrollArea): def __init__(self, parent=None): super(DetailSings, self).__init__(self) # self.hide() self.parent = parent self.setObjectName('detailSings') with open('QSS/detailSings.qss', 'r', encoding='utf-8') as f: self.setStyleSheet(f.read()) self.setLabels() self.setButtons() self.setTabs() self.setLayouts() # 布局。 def setLabels(self): self.picLabel = PicLabel(width=200, height=200) self.picLabel.setObjectName('picLabel') self.titleLabel = QLabel(self.frame) self.titleLabel.setObjectName('titleLabel') self.titleLabel.setWordWrap(True) self.titleLabel.setMaximumHeight(40) self.authorPic = QLabel(self.frame) self.authorName = QLabel(self.frame) self.authorName.setObjectName('authorName') self.authorName.setMaximumHeight(28) self.descriptionText = QTextEdit(self.frame) self.descriptionText.setReadOnly(True) self.descriptionText.setObjectName('descriptionText') self.descriptionText.setMaximumWidth(450) self.descriptionText.setMaximumHeight(100) self.descriptionText.setMinimumHeight(100) def setButtons(self): self.showButton = QPushButton("歌单") self.showButton.setObjectName('showButton') self.showButton.setMaximumSize(36, 20) self.descriptionButton = QPushButton(" 简介 :") self.descriptionButton.setObjectName('descriptionButton') self.descriptionButton.setMaximumSize(36, 36) self.playAllButton = QPushButton("全部播放") self.playAllButton.setIcon(QIcon('resource/playAll.png')) self.playAllButton.setObjectName('playAllButton') self.playAllButton.setMaximumSize(90, 24) def setTabs(self): self.contentsTab = QTabWidget(self.frame) self.singsTable = TableWidget(3, ['音乐标题', '歌手', '时长']) self.singsTable.setObjectName('singsTable') self.singsTable.setMinimumWidth(self.width()) self.singsTable.setColumnWidths({i:j for i,j in zip(range(3), [self.width()/3*1.25,self.width()/3*1.25,self.width()/3*0.5])}) self.contentsTab.addTab(self.singsTable, "歌曲列表") def setLayouts(self): self.mainLayout = VBoxLayout() self.topLayout = HBoxLayout() self.descriptionLayout = VBoxLayout() self.titleLayout = HBoxLayout() self.titleLayout.addWidget(self.showButton) self.titleLayout.addSpacing(5) self.titleLayout.addWidget(self.titleLabel) self.authorLayout = HBoxLayout() self.authorLayout.addWidget(self.authorPic) self.authorLayout.addWidget(self.authorName) self.authorLayout.addStretch(1) self.descriptLayout = HBoxLayout() self.descriptLayout.addWidget(self.descriptionButton) self.descriptLayout.addWidget(self.descriptionText) self.descriptionLayout.addSpacing(5) self.descriptionLayout.addLayout(self.titleLayout) self.descriptionLayout.addLayout(self.authorLayout) self.descriptionLayout.addSpacing(5) self.descriptionLayout.addWidget(self.playAllButton) self.descriptionLayout.addSpacing(10) self.descriptionLayout.addLayout(self.descriptLayout) self.topLayout.addWidget(self.picLabel) self.topLayout.addSpacing(18) self.topLayout.addLayout(self.descriptionLayout) self.mainLayout.addLayout(self.topLayout) self.mainLayout.addWidget(self.contentsTab) self.frame.setLayout(self.mainLayout) class OneSing(QFrame): # 大量创建,这样可以省内存。 __solts__ = ('parent', 'ggparent', 'detailFrame', 'row', 'column', 'ids', 'picName', 'picLabel', 'nameLabel', 'mainLayout', 'mousePos', 'result','catch', 'singsIds', 'singsUrls') clicked = pyqtSignal(str, str) def __init__(self, row, column, ids=None, parent=None, picName=None): super(OneSing, self).__init__() self.setObjectName('oneSing') # 自己的位置信息。 self.row = row self.column = column # 歌单号。 self.ids = str(ids) # 大图的缓存名。 self.picName = picName self.setMinimumSize(180, 235) self.picLabel = QLabel() self.picLabel.setObjectName('picLabel') self.picLabel.setMinimumSize(180, 180) self.picLabel.setMaximumSize(180, 180) self.nameLabel = QLabel() self.nameLabel.setMaximumWidth(180) self.nameLabel.setWordWrap(True) self.mainLayout = QVBoxLayout(self) self.mainLayout.addWidget(self.picLabel) self.mainLayout.addWidget(self.nameLabel) # 功能。 def setStyleSheets(self, styleSheet=None): if styleSheet: self.setStyleSheet(styleSheet) # 事件。 def mousePressEvent(self, event): # 记录下当前鼠标的位置。 self.mousePos = QCursor.pos() def mouseReleaseEvent(self, event): # 先进行判断,防止误点将鼠标移开后还是会判断为已经点击的尴尬。 if QCursor.pos() != self.mousePos: return else: self.clicked.emit(self.ids, self.picName) class PlaylistButton(QPushButton): """ 提供一个简单点击的歌单自动切换点击状态,点击时会发出hasClicked信号。 Args: parent 父类。 ids 歌单的ids。 coverImgUrl 这个歌单应该包含的图片地址 -> None, url. *args其他作用于原生Button的参数。 singsIds 与 singsUrls暂时无用,不提供接受接口。 """ __solts__ = ('parent', 'grandparent', 'ids', 'coverImgUrl', 'catch', 'detailFrame', 'result', 'singsIds', 'singsUrls' ) hasClicked = pyqtSignal(int, str) def __init__(self, parent, ids, coverImgUrl, *args): super(PlaylistButton, self).__init__(*args) self.parent = parent self.grandparent = self.parent.parent self.setCheckable(True) self.setAutoExclusive(True) self.ids = ids self.coverImgUrl = coverImgUrl self.catch = None self.result = None self.singsIds = None self.singsUrls = None self.clicked.connect(self.clickedEvent) def clickedEvent(self): self.hasClicked.emit(self.ids, self.coverImgUrl)
{ "repo_name": "HuberTRoy/MusicPlayer", "path": "MusicPlayer/widgets/singsFrameBase.py", "copies": "1", "size": "8770", "license": "mit", "hash": 4592259841820868600, "line_mean": 31.6117647059, "line_max": 138, "alpha_frac": 0.6465848966, "autogenerated": false, "ratio": 3.2895569620253164, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4436141858625316, "avg_score": null, "num_lines": null }
__author__ = 'cyrfer' # epic pyparsing help provided by Paul McGuire # http://pyparsing.wikispaces.com/share/view/68890534 from ogre_parse.basereader import * from ogre_parse.submodel import * class ReadTextureUnit(ReadBase): def __init__(self): # TODO: need to separate the required member options # --------------------- cubicResourceDecl = Keyword('cubic_texture')('resource_type') cubicIdentifier = identspec ^ Group(identspec + identspec + identspec + identspec + identspec + identspec) # cubicPropList = Group(\ # cubicIdentifier('name') + \ # oneOf('combinedUVW separateUV')('batch_mode') \ # ) cubicPropList = propList cubicResource = Group( cubicResourceDecl + cubicPropList('cubic_properties') )('required') # --------------------- animResourceDecl = Keyword('anim_texture')('resource_type') animResource = Group( animResourceDecl + propList('keyframes') )('required') # --------------------- textureResourceDecl = Keyword('texture')('resource_type') texPropList = Group(identspec('name') + \ Optional( oneOf('1d 2d 3d cubic') )('type') + \ Optional(integer)('numMipMaps') + \ Optional(Literal('alpha'))('alpha') + \ Optional(oneOf(imageFormats)('format')) + \ Optional(Literal('gamma'))) textureResource = Group( textureResourceDecl + texPropList('resource_properties') )('required') addr_mode_val = oneOf('wrap clamp mirror border') addr_mode_spec = addr_mode_val + Optional(addr_mode_val + Optional(addr_mode_val)) c_op_src_spec = oneOf('src_current src_texture src_diffuse src_specular src_manual') c_op_spec = oneOf('source1 source2 modulate modulate_x2 modulate_x4 add add_signed add_smooth subtract blend_diffuse_alpha blend_texture_alpha blend_current_alpha blend_manual dotproduct blend_diffuse_colour') colour_op_ex_spec = Group(c_op_spec('operation') + c_op_src_spec('source1') + c_op_src_spec('source2') + Optional(real('manual_factor')) + Optional(coloraction('manual_colour1')) + Optional(coloraction('manual_colour2'))) fallback_spec = Group(scene_blend_long_spec('src_factor') + scene_blend_long_spec('dest_factor')) # define the optional members alias = Group(Keyword('texture_alias').suppress() + identspec)('texture_alias') coord_set = Group(Keyword('tex_coord_set').suppress() + integer)('tex_coord_set') address_mode = Group(Keyword('tex_address_mode').suppress() + addr_mode_spec)('tex_address_mode') border_colour = Group(Keyword('tex_border_colour').suppress() + coloraction)('tex_border_colour') filtering = Group(Keyword('filtering').suppress() + propList)('filtering') scale = Group(Keyword('scale').suppress() + (real('x') + real('y')))('scale') colour_op = Group(Keyword('colour_op').suppress() + oneOf('replace add modulate alpha_blend'))('colour_op') binding_type = Group(Keyword('binding_type').suppress() + oneOf('vertex fragment'))('binding_type') colour_op_ex = Group(Keyword('colour_op_ex').suppress() + colour_op_ex_spec)('colour_op_ex') colour_op_multipass_fallback = Group(Keyword('colour_op_multipass_fallback').suppress() + fallback_spec)('colour_op_multipass_fallback') env_map = Group(Keyword('env_map').suppress() + oneOf('off spherical planar cubic_reflection cubic_normal'))('env_map') content_type = Group(Keyword('content_type').suppress() + oneOf('named shadow compositor') + Optional(identspec('compositor') + identspec('texture') + Optional(integer('MRT'))))('content_type') # --- define the parser textureDecl = Keyword('texture_unit').suppress() + Optional(ident)('name') + \ lbrace + \ ( \ Optional(textureResource | animResource | cubicResource) & \ Optional(alias) & \ Optional(coord_set) & \ Optional(address_mode) & \ Optional(border_colour) & \ Optional(filtering) & \ Optional(scale) & \ Optional(colour_op) & \ Optional(binding_type) & \ Optional(colour_op_ex) & \ Optional(colour_op_multipass_fallback) & \ Optional(env_map) & \ Optional(content_type) \ ) + \ rbrace texture_ = Group(textureDecl).setParseAction(MTextureUnit) super(ReadTextureUnit, self).__init__(texture_('texture_unit')) class ReadShaderReference(ReadBase): def __init__(self): # --- define the shader_ref parser # shaderRefPropName = oneOf('param_indexed param_indexed_auto param_named param_named_auto shared_params_ref') param_named_auto_spec = Keyword('param_named_auto').suppress() + ident param_named_spec = Keyword('param_named').suppress() + ident shaderRefSpec = oneOf('vertex_program_ref fragment_program_ref') shaderRefDecl = shaderRefSpec('stage') + ident('resource_name') + \ lbrace + \ ( \ dictOf( param_named_auto_spec, propList )('param_named_auto') & \ dictOf( param_named_spec, propList )('param_named') \ ) + \ rbrace # (dictOf(param_named_auto, propList('system_params'))('param_named_auto')) + \ shader_ref_ = Group(shaderRefDecl) shader_ref_.setParseAction(MShaderRef) super(ReadShaderReference, self).__init__(shader_ref_('shader_ref')) # successful parsing produces a subreader.MPass in parsed.mpass # format documented here: # http://www.ogre3d.org/docs/manual/manual_16.html#Passes class ReadPass(ReadBase): def __init__(self): # define named parsers color_ambient = Group(Keyword('ambient').suppress() + colorspec)('ambient') color_diffuse = Group(Keyword('diffuse').suppress() + colorspec)('diffuse') color_emissive = Group(Keyword('emissive').suppress() + colorspec)('emissive') color_specular = Group(Keyword('specular').suppress() + specular_spec)('specular') # scene_blend # TODO: add action to turn short format into long format scene_blend_short = oneOf('add modulate colour_blend alpha_blend') scene_blend_long = scene_blend_long_spec + scene_blend_long_spec scene_blend = Group(Keyword('scene_blend').suppress() + (scene_blend_short | scene_blend_long))('scene_blend') # TODO: add action to turn short format into long format separate_blend_short = scene_blend_short + scene_blend_short separate_blend_long = scene_blend_long_spec + scene_blend_long_spec + scene_blend_long_spec + scene_blend_long_spec separate_scene_blend = Group(Keyword('separate_scene_blend').suppress() + (separate_blend_short | separate_blend_long))('separate_scene_blend') scene_blend_op_spec = oneOf('add subtract reverse_subtract min max') scene_blend_op = Group(Keyword('scene_blend_op').suppress() + scene_blend_op_spec)('scene_blend_op') separate_scene_blend_op = Group(Keyword('separate_scene_blend_op').suppress() + (scene_blend_op_spec+scene_blend_op_spec))('separate_scene_blend_op') # depth stuff depth_check = Group(Keyword('depth_check').suppress() + onoff_val_spec)('depth_check') depth_write = Group(Keyword('depth_write').suppress() + onoff_val_spec)('depth_write') depth_func_val_spec = oneOf('always_fail always_pass less less_equal equal not_equal greater_equal greater') depth_func = Group(Keyword('depth_func').suppress() + depth_func_val_spec)('depth_func') depth_bias = Group(Keyword('depth_bias').suppress() + real('constant') + Optional(real('slopescale')))('depth_bias') iter_depth_bias = Group(Keyword('iteration_depth_bias').suppress() + real('bias'))('iteration_depth_bias') # alpha stuff alpha_rejection_func = depth_func_val_spec('function') alpha_rejection = Group(Keyword('alpha_rejection').suppress() + alpha_rejection_func + real('threshold'))('alpha_rejection') alpha_to_coverage = Group(Keyword('alpha_to_coverage').suppress() + onoff_val_spec)('alpha_to_coverage') # light_scissor light_scissor = Group(Keyword('light_scissor').suppress() + onoff_val_spec)('light_scissor') light_clip_planes = Group(Keyword('light_clip_planes').suppress() + onoff_val_spec)('light_clip_planes') # other illum_stage_val = oneOf('ambient per_light decal') illumination_stage = Group(Keyword('illumination_stage').suppress() + illum_stage_val)('illumination_stage') onoffforce_val_spec = oneOf('on off force') transparent_sorting = Group(Keyword('transparent_sorting').suppress() + onoffforce_val_spec)('transparent_sorting') normalise_normals = Group(Keyword('normalise_normals').suppress() + onoff_val_spec)('normalise_normals') # cull cull_hardware = Group(Keyword('cull_hardware').suppress() + oneOf('clockwise anticlockwise none'))('cull_hardware') cull_software = Group(Keyword('cull_software').suppress() + oneOf('back front none'))('cull_software') # other lighting = Group(Keyword('lighting').suppress() + onoff_val_spec)('lighting') shading = Group(Keyword('shading').suppress() + oneOf('flat gouraud phong'))('shading') polygon_mode = Group(Keyword('polygon_mode').suppress() + oneOf('solid wireframe points'))('polygon_mode') polygon_mode_overrideable = Group(Keyword('polygon_mode_overrideable').suppress() + truefalse_spec)('polygon_mode_overrideable') fog_override_type = oneOf('none linear exp exp2') # must define number spec here because # something is wrong with the specs for number types at this point (real, integer) # and they are returning a number and not a string. fog_int = Word(nums) fog_real = Regex(r"\d+\.\d*") fog_num = fog_int ^ fog_real fog_override_colour = fog_num + fog_num + fog_num fog_override_args = fog_override_type('type') + fog_override_colour('colour') + fog_num('density') + fog_num('start') + fog_num('end') fog_override = Group(Keyword('fog_override').suppress() + truefalse_spec('enabled') + Optional(fog_override_args))('fog_override') colour_write = Group(Keyword('colour_write').suppress() + onoff_val_spec)('colour_write') start_light = Group(Keyword('start_light').suppress() + integer)('start_light') max_lights = Group(Keyword('max_lights').suppress() + integer)('max_lights') light_type_spec = oneOf('point directional spot') iteration_format1 = oneOf('once once_per_light') + Optional(light_type_spec) iteration_format2 = integerspec + Optional(Keyword('per_light') + light_type_spec) iteration_format3 = integerspec + Optional(Keyword('per_n_lights') + integerspec + Optional(light_type_spec)) iteration_args = (iteration_format1 ^ iteration_format2 ^ iteration_format3) iteration = Group(Keyword('iteration').suppress() + iteration_args)('iteration') point_size = Group(Keyword('point_size').suppress() + real)('point_size') point_sprites = Group(Keyword('point_sprites').suppress() + onoff_val_spec)('point_sprites') attenuation_spec = oneOf('constant linear quadratic') point_size_attenuation = Group(Keyword('point_size_attenuation').suppress() + onoff_val_spec('enabled') + Optional(attenuation_spec)('model'))('point_size_attenuation') point_size_min = Group(Keyword('point_size_min').suppress() + real)('point_size_min') point_size_max = Group(Keyword('point_size_max').suppress() + real)('point_size_max') tu = ReadTextureUnit() shader = ReadShaderReference() passBody = ( \ # color Optional(color_ambient) & \ Optional(color_diffuse) & \ Optional(color_emissive) & \ Optional(color_specular) & \ # blend Optional(scene_blend) & \ Optional(separate_scene_blend) & \ Optional(scene_blend_op) & \ Optional(separate_scene_blend_op) & \ # depth Optional(depth_check) & \ Optional(depth_write) & \ Optional(depth_func) & \ Optional(depth_bias) & \ Optional(iter_depth_bias) & \ # alpha Optional(alpha_rejection) & \ Optional(alpha_to_coverage) & \ # light scissor Optional(light_scissor) & \ Optional(light_clip_planes) & \ # other Optional(illumination_stage) & \ Optional(transparent_sorting) & \ Optional(normalise_normals) & \ # cull Optional(cull_hardware) & \ Optional(cull_software) & \ # lighting Optional(lighting) & \ Optional(shading) & \ # polygon Optional(polygon_mode) & \ Optional(polygon_mode_overrideable) & \ # other Optional(fog_override) & \ Optional(colour_write) & \ Optional(start_light) & \ Optional(max_lights) & \ Optional(iteration) & \ # point Optional(point_size) & \ Optional(point_sprites) & \ Optional(point_size_attenuation) & \ Optional(point_size_min) & \ Optional(point_size_max) & \ # texture ZeroOrMore(tu.getGrammar())('texture_units') & \ # shaders ZeroOrMore(shader.getGrammar())('shaders') \ ) # total parser parser = Group( Keyword('pass').suppress() + Optional(identspec('name')) + LBRACE + passBody + RBRACE)('mpass') parser.setParseAction(MPass) super(ReadPass, self).__init__(parser) # -------------- END NEW STUFF SINCE PAUL HELPED ------------ # class ReadTechnique(ReadBase): def __init__(self): pass_ = ReadPass() # --- define the technique parser scheme = Group(Keyword('scheme').suppress() + identspec)('scheme') lod_index = Group(Keyword('lod_index').suppress() + integer)('lod_index') shadow_caster_material = Group(Keyword('shadow_caster_material').suppress() + identspec)('shadow_caster_material') shadow_receiver_material = Group(Keyword('shadow_receiver_material').suppress() + identspec)('shadow_receiver_material') inex = oneOf('include exclude') gpu_vendor_rule = Group(Keyword('gpu_vendor_rule').suppress() + inex + identspec)('gpu_vendor_rule') gpu_device_rule = Group(Keyword('gpu_device_rule').suppress() + inex + propList)('gpu_device_rule') techDecl = Keyword('technique').suppress() + Optional(ident)('name') + \ lbrace + \ Optional( scheme ) + \ Optional( lod_index ) + \ Optional( shadow_caster_material ) + \ Optional( shadow_receiver_material ) + \ ZeroOrMore( gpu_vendor_rule )('gpu_vendor_rules') + \ ZeroOrMore( gpu_device_rule )('gpu_device_rules') + \ OneOrMore( pass_.getGrammar() )('passes') + \ rbrace technique_ = Group(techDecl) technique_.setParseAction(MTechnique) super(ReadTechnique, self).__init__(technique_('technique'))
{ "repo_name": "cyrfer/ogre_parse", "path": "ogre_parse/subreader.py", "copies": "1", "size": "16636", "license": "mit", "hash": 5522130532067399000, "line_mean": 52.8381877023, "line_max": 229, "alpha_frac": 0.5770016831, "autogenerated": false, "ratio": 4.018357487922706, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.009911498537122367, "num_lines": 309 }
__author__ = 'cyrfer' import array import math import copy def float_eq(a, b, epsilon=1e-7): # print('float_eq: type(a)=%s, type(b)=%s' % (type(a), type(b))) # print('float_eq(%s, %s)' % (a,b)) diff_val = math.fabs(a - b) # print('float_eq: diff_val = %s' % diff_val) return diff_val < epsilon # should be hooked up to a 'basereader.colorspec' instance class Color(object): def __init__(self, tokens=None, vals=None): self.vector = array.array('f', [0, 0, 0, 1]) if vals: for i in range(min(4,len(vals))): self.vector[i] = vals[i] if tokens and len(tokens) > 0: for index in range(min(len(self.vector), len(tokens[0]))): self.vector[index] = tokens[0][index] # print('self.vector = %s' % self.vector) def __str__(self): fmt = '{0:.6f}' rep = fmt.format(self.vector[0]).rstrip('0').rstrip('.') rep += ' ' + fmt.format(self.vector[1]).rstrip('0').rstrip('.') rep += ' ' + fmt.format(self.vector[2]).rstrip('0').rstrip('.') rep += ' ' + fmt.format(self.vector[3]).rstrip('0').rstrip('.') return rep __repr__ = __str__ def __getitem__(self, item): return self.vector.__getitem__(item) def __setitem__(self, key, value): return self.vector.__setitem__(key, value) def __eq__(self, other): if not isinstance(other, Color): return False if not other: # print('ogre_parse.basemodel.Color: why are we comparing with _%s_?' % other) return False if len(other)<4: # print('ogre_parse.basemodel.Color: len(other)=%s' % len(other)) return False res = self.vector[0] == other[0] \ and self.vector[1] == other[1] \ and self.vector[2] == other[2] \ and self.vector[3] == other[3] return res def __len__(self): return 4 # support RHS multiplication def __mul__(self, scalar): if not isinstance(scalar, float) and not isinstance(scalar, int): raise ValueError('argument to multiply with Color should be a scalar.') prod = copy.deepcopy(self) prod[0] = scalar * prod[0] prod[1] = scalar * prod[1] prod[2] = scalar * prod[2] prod[3] = scalar * prod[3] return prod # support LHS multiplication __rmul__ = __mul__
{ "repo_name": "cyrfer/ogre_parse", "path": "ogre_parse/basemodel.py", "copies": "1", "size": "2448", "license": "mit", "hash": -4869784341842968000, "line_mean": 28.1428571429, "line_max": 90, "alpha_frac": 0.5310457516, "autogenerated": false, "ratio": 3.4094707520891365, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9374034898460286, "avg_score": 0.013296321045770054, "num_lines": 84 }
__author__ = 'cyrfer' # from pyparsing import Optional, Word, Literal, Keyword, Forward, alphas, nums, alphanums, \ # Group, ZeroOrMore, OneOrMore, oneOf, delimitedList, cStyleComment, restOfLine, LineEnd, \ # cppStyleComment, Combine, Dict, dictOf, Regex, Suppress from pyparsing import * # import ogre_parse.model import ogre_parse.basemodel def printAll(s, l, toks): print('-----------') print(toks) print('-----------') # convenient definitions # TODO: find a way that does not pollute the global namespace EOL = LineEnd().suppress() ident = Word( alphanums+"_", alphanums+"_-/$@#.()" ) identspec = Word( alphanums+"_", alphanums+"_-$@#." ) lbrace = Literal("{").suppress() rbrace = Literal("}").suppress() integerspec = Word(nums) integer = Word(nums) integer.setParseAction(lambda t: int(t[0])) # propList.setParseAction(printAll) # reusable definitions LBRACE, RBRACE = map(Suppress,'{}') EOL = LineEnd().suppress() # another option for floating point parsing: # http://pyparsing.wikispaces.com/share/view/33656348 # Regex(r'\d+(\.\d*)?([eE]\d+)?') # realspec = Combine(Optional('-') + Regex(r"\d+(\.\d*)?")) realspec_frac_only = Literal('.') + Word(nums) #Regex(r".\d") realspec_whole_only = Word(nums) #Regex(r"\d") realspace_whole_and_frac = Regex(r"\d+(\.\d*)?") realspec = Combine(Optional('-') + (realspec_frac_only ^ realspec_whole_only ^ realspace_whole_and_frac)) int_or_real_spec = integerspec ^ realspec real = (int_or_real_spec).setParseAction(lambda t: float(t[0])) propVal = realspec | integerspec | ident propList = Group(OneOrMore(~EOL + propVal)) # colorspec = Group(~EOL + OneOrMore(realspec))('vector').setParseAction(Color) color3spec = Group(real('r') + real('g') + real('b')).setParseAction(ogre_parse.basemodel.Color) color4spec = Group(real('r') + real('g') + real('b') + real('a')).setParseAction(ogre_parse.basemodel.Color) coloraction = (color3spec ^ color4spec) colorspec = ( color3spec ^ color4spec )('args') specular_spec = Group( (Group(color3spec)('color') + Group(real)('shininess')) ^ (Group(color4spec)('color') + Group(real)('shininess')) ) scene_blend_long_spec = oneOf('one zero dest_colour src_colour one_minus_dest_colour one_minus_src_colour dest_alpha src_alpha one_minus_dest_alpha one_minus_src_alpha') truefalse_spec = oneOf('true false') onoff_val_spec = oneOf('on off') # on 8/8/2014, taken from: # http://www.ogre3d.org/docs/manual/manual_17.html#texture imageFormats = ''' PF_L8 PF_L16 PF_A8 PF_A4L4 PF_BYTE_LA PF_R5G6B5 PF_B5G6R5 PF_R3G3B2 PF_A4R4G4B4 PF_A1R5G5B5 PF_R8G8B8 PF_B8G8R8 PF_A8R8G8B8 PF_A8B8G8R8 PF_B8G8R8A8 PF_R8G8B8A8 PF_X8R8G8B8 PF_X8B8G8R8 PF_A2R10G10B10 PF_A2B10G10R10 PF_DXT1 PF_DXT2 PF_DXT3 PF_DXT4 PF_DXT5 PF_FLOAT16_R PF_FLOAT16_RGB PF_FLOAT16_RGBA PF_FLOAT32_R PF_FLOAT32_RGB PF_FLOAT32_RGBA PF_SHORT_RGBA PF_FLOAT16_GR PF_FLOAT32_GR PF_DEPTH PF_SHORT_GR PF_SHORT_RGB PF_PVRTC_RGB2 PF_PVRTC_RGBA2 PF_PVRTC_RGB4 PF_PVRTC_RGBA4 PF_R8 PF_RG8 ''' # base class for all parsers class ReadBase(object): # all derived classes will provide the grammar object def __init__(self, grammar): self.grammar_ = grammar self.grammar_.ignore( cppStyleComment ) self.debug_flag_ = False self.grammar_.setDebug(False) def getGrammar(self): return self.grammar_ def parseString(self, txt): if self.debug_flag_: print('parsing: [[\n' + txt + '\n]]\n') result = self.grammar_.parseString(txt) if self.debug_flag_: print('result = %s' % result ) return result
{ "repo_name": "cyrfer/ogre_parse", "path": "ogre_parse/basereader.py", "copies": "1", "size": "3633", "license": "mit", "hash": 4363126430530298000, "line_mean": 32.6388888889, "line_max": 169, "alpha_frac": 0.6710707404, "autogenerated": false, "ratio": 2.8316445830085737, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.40027153234085733, "avg_score": null, "num_lines": null }
__author__ = 'cyrfer' # This file defines 'readers' for the major classes of OGRE objects, including: # - materials # - shader declarations # - compositors # OGRE permits putting any of the major classes into any of the OGRE resource file types, including: # - .material # - .program # - .compositor # Because any script may define any of the definitions, a catch-all "script reader" is defined here. from ogre_parse.basereader import * import ogre_parse.subreader from ogre_parse.model import * # grammar to parse materials class ReadMaterial(ReadBase): def __init__(self): technique_ = ogre_parse.subreader.ReadTechnique() # --- define the material parser lod_strategy = Group(Keyword('lod_strategy').suppress() + ident)('lod_strategy') lod_values = Group(Keyword('lod_values').suppress() + OneOrMore(integer))('lod_values') receive_shadows = Group(Keyword('receive_shadows').suppress() + onoff_val_spec)('receive_shadows') transparency_casts_shadows = Group(Keyword('transparency_casts_shadows').suppress() + onoff_val_spec)('transparency_casts_shadows') set_texture_alias_key = Keyword('set_texture_alias').suppress() + identspec set_texture_alias_val = identspec matDecl = Keyword('material').suppress() + ident('name') + \ lbrace + \ Optional(lod_strategy) + \ Optional(lod_values) + \ Optional(receive_shadows) + \ Optional(transparency_casts_shadows) + \ dictOf(set_texture_alias_key, set_texture_alias_val)('texture_alias') + \ OneOrMore( technique_.getGrammar() )('techniques') + \ rbrace material_ = Group(matDecl)('material') material_.setParseAction(Material) super(ReadMaterial, self).__init__(material_) # grammar to parse shader declarations class ReadShaderDeclaration(ReadBase): def __init__(self): param_named_auto_spec = Keyword('param_named_auto').suppress() + ident param_named_spec = Keyword('param_named').suppress() + ident default_params = Group( Keyword('default_params').suppress() + \ lbrace + \ dictOf( param_named_auto_spec, propList )('param_named_auto') + \ dictOf( param_named_spec, propList )('param_named') + \ rbrace \ )('default_params') # shaderDeclPropName = oneOf('source entry_point target delegate') # shaderDeclProp = Group(shaderDeclPropName + propList) shaderStage = oneOf('vertex_program geometry_program fragment_program')('stage') shaderName = ident('name') shaderLang = oneOf('glsl hlsl cg asm')('language') source = Group(Keyword('source').suppress() + identspec)('source') entry_point = Group(Keyword('entry_point').suppress() + identspec)('entry_point') target = Group(Keyword('target').suppress() + identspec)('target') hlsl = ( \ source & \ entry_point & \ target \ ) glsl = ( \ source \ ) shaderDeclDecl = shaderStage + shaderName + shaderLang + \ lbrace + \ (\ (hlsl | glsl) & \ Optional(default_params) \ ) + \ rbrace # (shaderDeclProps_unified | shaderDeclProps_hlsl | shaderDeclProps_glsl | shaderDeclProps_cg | shaderDeclProps_asm) shader_declaration = Group(shaderDeclDecl)('shader') shader_declaration.setParseAction(ShaderDeclaration) super(ReadShaderDeclaration, self).__init__(shader_declaration) # grammar to parse compositors class ReadCompositor(ReadBase): def __init__(self): compTexture = Group(Keyword('texture') + OneOrMore(ident)) compTargetPropName = oneOf('input only_initial visibility_mask load_bias material_scheme shadows pass') compTargetProp = Group(compTargetPropName) compTarget = Group(Keyword('target') + ident + lbrace + ZeroOrMore(compTargetProp) + rbrace) compOutput = Group(Keyword('target_output') + lbrace + ZeroOrMore(compTargetProp) + rbrace) compTech = Group(Keyword('technique').suppress() + \ lbrace + \ ZeroOrMore(compTexture) + \ ZeroOrMore(compTarget) + \ compOutput + \ rbrace) compDecl = Keyword('compositor').suppress() + ident + \ lbrace + \ OneOrMore(compTech) + \ rbrace compositor = Group(compDecl) super(ReadCompositor, self).__init__(compositor) # grammar to parse anything in an ogre script (.material, .compositor, .program) class ReadScript(ReadBase): def __init__(self): material_ = ReadMaterial().getGrammar() shader_declaration_ = ReadShaderDeclaration().getGrammar() compositor_ = ReadCompositor().getGrammar() resourceType = material_ | compositor_ | shader_declaration_ scriptDecl = ZeroOrMore(resourceType) + StringEnd() scriptDecl.setParseAction(Script) super(ReadScript, self).__init__(scriptDecl('script')) def getGrammar(self): return self.grammar_
{ "repo_name": "cyrfer/ogre_parse", "path": "ogre_parse/reader.py", "copies": "1", "size": "5640", "license": "mit", "hash": -7213373833509549000, "line_mean": 41.0895522388, "line_max": 148, "alpha_frac": 0.5764184397, "autogenerated": false, "ratio": 4.28246013667426, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.535887857637426, "avg_score": null, "num_lines": null }
__author__ = 'Cyril de Vogelaere : 2814-11-00 & Thuin florian : 0656-11-00' import time import sys from copy import deepcopy from os import listdir, system from search import * # LEFT RIGHT UP DOWN directions = [[0, -1], [0, 1], [-1, 0], [1, 0]] global listOfGoalPos ############### # My function # ############### def readGridFromFile(Texte): with open(Texte, "r") as file: data_read = file.read() grid = data_read.split("\n") #Remove empty line grid.pop() #Remove last line grid.pop() #Remove first line grid.pop(0) #Truncate first and last collumn for i in range(0, len(grid)): grid[i] = grid[i][1 : len(grid[i])-1] return grid # Read goal file and return a grid containing the data def readStateFromGoal(goal): grid = readGridFromFile(goal) listOfGoalPos = [] #Orginal position of boxes i = 0 for line in grid: for j in range(0, len(line)): if line[j] == ".": #Avatar listOfGoalPos.append((i, j)) i+=1 return listOfGoalPos # Read init file and return a state containing the data def readStateFromInit(init): with open(init, "r") as file: # Lecture du fichier grid = readGridFromFile(init) # Creation d'un tableau equivalent au probleme avatarPos = (0,0) #Original position of the avatar listOfBoxesPos = [] #Orginal position of boxes #Read grid for important elem i = 0 for line in grid: for j in range(0, len(line)): if line[j] == "@": #Avatar avatarPos = (i, j) elif line[j] == "$": #Box listOfBoxesPos.append((i, j)) i+=1 return State(grid, listOfBoxesPos, avatarPos) #Add a line of wall def addALineOfWall(string, length): for i in range (0, length+2): string += "#" string += "\n" return string #Check if position is inbound def inBounds(grid, pos): return 0 <= pos[0] and pos[0] < len(grid) and 0 <= pos[1] and pos[1] < len(grid[0]) #Check if the state is a KO state def isKOState(state, box): #Check direction in which i can push if box in listOfGoalPos : # If box on goal state, it's never a KO state return False #Test LEFT AND RIGHT freedom = 0 for x in range(0, 2): i = box[0] + directions[x][0] j = box[1] + directions[x][1] if inBounds(state.grid, (i, j)) and (state.grid[i][j] == " " or state.grid[i][j] == "@"): freedom += 1; if freedom == 2: return False #Test LEFT AND RIGHT freedom = 0 for x in range(2, 4): i = box[0] + directions[x][0] j = box[1] + directions[x][1] if inBounds(state.grid, (i, j)) and (state.grid[i][j] == " " or state.grid[i][j] == "@"): freedom += 1; if freedom == 2: return False return False # Check if pushing box will lead to a KO state # Pre : box is pushable def isPushingOK(state, dir, x, y): result = False state.grid[x] = state.grid[x][:y] + " " + state.grid[x][y+1:] newBoxX = x + dir[0] newBoxY = y + dir[1] state.grid[newBoxX] = state.grid[newBoxX][:newBoxY] + "$" + state.grid[newBoxX][newBoxY+1:] result = not isKOState(state, (newBoxX, newBoxY)) state.grid[newBoxX] = state.grid[newBoxX][:newBoxY] + " " + state.grid[newBoxX][newBoxY+1:] state.grid[x] = state.grid[x][:y] + "$" + state.grid[x][y+1:] return result #Check if two position are adjacent def arePosAdjacent(posA, posB): distI = abs(posA[0] - posB[0]) distJ = abs(posA[1] - posB[1]) return (distI + distJ) < 2 #Check if char can push the box from this position def canPushBox(grid, char, box): if arePosAdjacent(char, box): i = 2*box[0] - char[0] j = 2*box[1] - char[1] if inBounds(grid, (i, j)) and grid[i][j] == " ": return True return False #Generate successor from state #Pre : Successor can be generated => if box it can be pushed def generateSuccessor(state, dir): newState = deepcopy(state) #Calculate new avatar pos newState.avatarPos = (state.avatarPos[0] + dir[0], state.avatarPos[1] + dir[1]) #Clear old pos in grid, update avatar pos in state newState.grid[state.avatarPos[0]] = newState.grid[state.avatarPos[0]][:state.avatarPos[1]] + " " + newState.grid[state.avatarPos[0]][state.avatarPos[1]+1:] if(newState.grid[newState.avatarPos[0]][newState.avatarPos[1]] == "$"): #Move box before updating avatar in grid for index in range(0, len(newState.listOfBoxesPos)): if (newState.avatarPos[0], newState.avatarPos[1]) == newState.listOfBoxesPos[index]: #Calculate new coordinate newX = dir[0] + newState.avatarPos[0] newY = dir[1] + newState.avatarPos[1] newState.listOfBoxesPos[index] = (newX, newY) newState.grid[newX] = newState.grid[newX][:newY] + "$" + newState.grid[newX][newY+1:] #Update avatar in grid newState.grid[newState.avatarPos[0]] = newState.grid[newState.avatarPos[0]][:newState.avatarPos[1]] + "@" + newState.grid[newState.avatarPos[0]][newState.avatarPos[1]+1:] return newState #Calculate the minimum position from the avatar to a box def calculateDistFromBoxes(state): best = len(state.grid) + len(state.grid[0]) for box in state.listOfBoxesPos: best = min(best, (abs(box[0] - state.avatarPos[0]) + abs(box[1] - state.avatarPos[1]))) return best # Return the minimum hamilton distance to reach a goal def minDistOfBoxToGoal(state, box): best = len(state.grid) + len(state.grid[0]) for goal in listOfGoalPos: best = min(best, (abs(goal[0] - box[0]) + abs(goal[1] - box[1]))) return best # Heuristic function # Minimal value will be explored first !!! def heuristicFunction(node): score = 0 for box in node.state.listOfBoxesPos: score += minDistOfBoxToGoal(node.state, box) * len(node.state.grid) # Passes everything score += calculateDistFromBoxes(node.state) return score ################# # My classes # ################# class Sokoban(Problem): def __init__(self, init): # Extract state from file global listOfGoalPos listOfGoalPos = readStateFromGoal(init + ".goal") initState = readStateFromInit(init + ".init") # Extend super init super().__init__(initState) def goal_test(self, state): for elem in listOfGoalPos: if not elem in state.listOfBoxesPos: return False return True def successor(self, state): #print(state) for i in range(0, len(directions)): x = state.avatarPos[0] + directions[i][0] y = state.avatarPos[1] + directions[i][1] if inBounds(state.grid, (x, y)) and (state.grid[x][y] == ' ' or (state.grid[x][y] == '$' and canPushBox(state.grid, state.avatarPos, (x,y)) and isPushingOK(state, directions[i], x, y))): #Yield result yield (i, generateSuccessor(state, directions[i])) class State: def __init__(self, gridInit, listOfBoxesPos, avatarPos): # Save state variable self.listOfBoxesPos = listOfBoxesPos self.avatarPos = avatarPos self.grid = gridInit def __str__(self): # Jolie representation string = "" string = addALineOfWall(string, len(self.grid[0])) for a in range(0, len(self.grid)): string += "#" for b in range(0, len(self.grid[a])): string += self.grid[a][b] string += "#" string += "\n" string = addALineOfWall(string, len(self.grid[0])) return string def __repr__(self): # Full representation return str((self.avatarPos, self.listOfBoxesPos, self.grid)) def __eq__(self, other): return (other.grid == self.grid) def __hash__(self): return self.__str__().__hash__() ##################### # Launch the search # ##################### # Init now = time.time() problem = Sokoban(sys.argv[1]) # Solve using bfs search #node = depth_first_graph_search(problem) node = astar_graph_search(problem, heuristicFunction) # Print path = node.path() path.reverse() #print(len(path)) for n in path: print(n.state) # assuming that the __str__ function of states output the correct format #Calculate time elapsed later = time.time() print(later - now)
{ "repo_name": "fthuin/artificial-intelligence", "path": "assignment2/Code/Sokoban.py", "copies": "1", "size": "8588", "license": "mit", "hash": 7897304180310724000, "line_mean": 32.8149606299, "line_max": 198, "alpha_frac": 0.5895435491, "autogenerated": false, "ratio": 3.286643704554152, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4376187253654152, "avg_score": null, "num_lines": null }
__author__ = 'czervenka' """ Simple database app settings BASIC USE --------- from models.config import settings version = settings['version'] # read settings settings['version'] = '1' # stores settings INITIALIZATION -------------- from models.config import settings settings['version'] = '1' # use Datastore Viewer to change your own settings """ from google.appengine.ext import ndb from google.appengine.api import namespace_manager class AppSettings(ndb.Model): version = ndb.StringProperty(default='0.1', indexed=False) class LazyAppSettings(object): """ Settings row property descriptor used in SettingsDict. """ def __get__(self, instance, owner=None, domain=None): if not hasattr(instance, '_cached_settings'): key = ndb.Key(AppSettings, 'app_settings') row = key.get() if row is None: row = AppSettings(key=key) instance._cached_settings = row return instance._cached_settings class SettingsDict(object): """ Dictionary which reads settings from database. To list all settings simply calll settings.items() """ _row = LazyAppSettings() def __getitem__(self, key): if hasattr(self._row, key): return getattr(self._row, key) else: raise KeyError('Key %s not found.' % key) def __setitem__(self, key, value): setattr(self._row, key, value) self._row.put() def __contains__(self, key): return hasattr(self._row, key) def keys(self): return self._row._properties.keys() def items(self): return [(key, self[key]) for key in self.keys()] def values(self): return [self[key] for key in self.keys()] def __iter__(self): for key in self.keys(): yield key class SettingsCache(object): def __init__(self): self.storage = {} def __getitem__(self, item): return self.get()[item] def __setitem__(self, key, value): self.get()[key] = value def get(self): return self.__for_namespace(namespace_manager.get_namespace()) def __for_namespace(self, namespace): return self.storage.setdefault(namespace, SettingsDict()) settings = SettingsCache()
{ "repo_name": "xaralis/gap-sandbox", "path": "src/app/settings.py", "copies": "1", "size": "2302", "license": "apache-2.0", "hash": 2730642240078861000, "line_mean": 24.0217391304, "line_max": 70, "alpha_frac": 0.6068635969, "autogenerated": false, "ratio": 4.074336283185841, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0, "num_lines": 92 }
__author__ = 'czhuang' import os import cPickle as pickle import numpy as np def is_valid_latin_square(grid): n_row, n_col = grid.shape assert n_row % n_col == 0 quotient = n_row / n_col sum_of_elements = np.sum(np.arange(n_col)) sum_of_col_wanted = quotient * sum_of_elements sum_of_cols = np.sum(grid, axis=0) # print sum_of_cols, sum_of_col_wanted if np.allclose(sum_of_cols, sum_of_col_wanted): return True else: return False def gen_latin_squares(num_conditions, num_participants): num_tries = 100 grid = np.zeros((num_participants, num_conditions)) for j in range(num_tries): # print j for i in range(num_participants): grid[i, :] = np.random.permutation(num_conditions) if is_valid_latin_square(grid): return grid return None def stack_condition_ordering(num_conditions, num_participants): quotient = num_participants / num_conditions ordering = None for i in range(quotient): local_ordering = gen_latin_squares(3, 3) if ordering is None: ordering = local_ordering else: ordering = np.vstack((ordering, local_ordering)) duplicated_ordering = np.zeros_like(ordering) duplicated_ordering = np.vstack((duplicated_ordering, duplicated_ordering)) for i in range(ordering.shape[0]): duplicated_ordering[i*2, :] = ordering[i, :] duplicated_ordering[i*2+1, :] = ordering[i, :] print duplicated_ordering assert np.allclose(np.sum(duplicated_ordering, axis=0), quotient * 2 * np.sum(np.arange(num_conditions))) fpath = os.path.join('pkls', 'condition_ordering.pkl') print 'fpath', fpath with open(fpath, 'wb') as p: pickle.dump(duplicated_ordering, p) def get_condition_ordering(): fpath = os.path.join('pkls', 'condition_ordering.pkl') with open(fpath, 'rb') as p: grid = pickle.load(p) return grid if __name__ == '__main__': print os.getcwd() condition_orderings = stack_condition_ordering(3, 3*10) condition_orderings = get_condition_ordering() print condition_orderings print condition_orderings.shape
{ "repo_name": "czhuang/ChordRipple", "path": "app/latin_squares_experiment_tools.py", "copies": "1", "size": "2219", "license": "mit", "hash": -5078612588845626000, "line_mean": 27.0886075949, "line_max": 79, "alpha_frac": 0.6349707075, "autogenerated": false, "ratio": 3.5110759493670884, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.46460466568670883, "avg_score": null, "num_lines": null }
__author__ = 'cz' import os import math from vmb_db.conn import get_one from flask import request, redirect, url_for, flash, current_app, abort, render_template from flask.ext.login import login_required, current_user from werkzeug import secure_filename from vmb_db.contact_info import get_contact, get_contact_by_casillero, set_contact_by_casillero, iter_pages from vmb_db.invoice_info import get_invoice_list, set_inv_panama_by_guia, INVOICES_PER_PAGE from vmb_db.upLoadZoom import accounts from vmb_db.conf import getModule configVMB = getModule('config') def allowed_file(filename): return '.' in filename and \ filename.rsplit('.', 1)[1] in configVMB.ALLOWED_EXTENSIONS @login_required def viewInvoice(): page = int(request.args.get('page', '1')) if request.method.lower() == 'post': file = request.files['file'] if file and allowed_file(file.filename): filename = secure_filename(file.filename) file.save(os.path.join(configVMB.UPLOAD_FOLDER, filename)) fileFullName = '%s/%s' % (configVMB.UPLOAD_FOLDER, filename) uploaded = accounts(fileFullName=fileFullName) if uploaded > 0: uploadedMes = 'The file was uploaded. %s rows were processed' % (uploaded) flash(uploadedMes) else: flash('There was a problem with uploading the file') client = {'casillero' : 0} sort = 'fecha_proceso DESC, hora_proceso DESC' invoices = get_invoice_list(where=None, sort=sort, limit=INVOICES_PER_PAGE, skip=(page - 1) * INVOICES_PER_PAGE) n = get_one(query='SELECT COUNT(*) AS count FROM VMB.INVOICES') count = int(n['count']) numpages = int(math.ceil(count / float(INVOICES_PER_PAGE))) return render_template('invoices/view.html', client=client, ilist=invoices, tcount=count, tnumpages=numpages, tpage=page, pagination=iter_pages) @login_required def packageArrived(): print 'packageArrived' if request.method.lower() == 'post': guia = request.form.get('update', '').strip() set_inv_panama_by_guia(guia) return redirect(url_for('viewInvoice'))
{ "repo_name": "cindy-zimmerman/vmb-mrest", "path": "flask_mrest/actions/invoice.py", "copies": "1", "size": "2217", "license": "mit", "hash": 6363787196036341000, "line_mean": 40.0555555556, "line_max": 116, "alpha_frac": 0.6580965268, "autogenerated": false, "ratio": 3.4803767660910516, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.46384732928910516, "avg_score": null, "num_lines": null }
__author__ = 'cz' from flask import request, redirect, url_for, flash, current_app, abort, render_template from flask.ext.login import login_required, current_user from vmb_db.contact_info import get_contact, get_contact_by_casillero, set_contact_by_casillero, iter_pages from vmb_db.invoice_info import get_invoice_list_by_casillero, set_inv_by_guia, INVOICES_PER_PAGE from vmb_db.accounts import insert_account @login_required def viewClient(cid): if request.method.lower() == 'post': guia = int(request.form.get('update', '').strip()) amt = request.form.get('amt', '').strip() subtotal = request.form.get('subtotal', '').strip() try: amt = float(amt) subtotal = float(subtotal) paid = 0 if amt >= subtotal: paid = 1 set_inv_by_guia(guia, paid, amt) except: flash('El numero de paga que ha introducido no es valido') client = get_contact_by_casillero(casillero=cid) invoices = get_invoice_list_by_casillero(casillero=cid) return render_template('clients/view.html', client=client, ilist=invoices) @login_required def editClient(cid): if cid == '0': client = {'contacto_nombre_1': '', 'casillero': '0', 'contacto_nombre_2': '', 'direccion_area': '', 'direccion_torre': '', 'direccion_calle': '', 'telefonocel': '', 'correo': '', 'ciudad': 'Panama', 'contacto_apellido_2': '', 'VMB_ACCOUNTS_id': 0, 'contacto_apellido_1': '', 'telefonofij': '', 'tarifa': 0, 'direccion_apt': ''} else: client = get_contact_by_casillero(casillero=cid) actItems = {'FT': True, 'telefonofij': False, 'telefonocel': False, 'tarifa': False} if not client: flash('User %s not found' % cid, 'error') return redirect(url_for('/')) if request.method.lower() == 'post': actItems['FT'] = False client['contacto_nombre_1'] = request.form.get('contacto_nombre_1', '').strip() client['contacto_nombre_2'] = request.form.get('contacto_nombre_2', '').strip() client['contacto_apellido_1'] = request.form.get('contacto_apellido_1', '').strip() client['contacto_apellido_2'] = request.form.get('contacto_apellido_2', '').strip() client['correo'] = request.form.get('correo', '').strip() client['direccion_calle'] = request.form.get('direccion_calle', '').strip() client['direccion_torre'] = request.form.get('direccion_torre', '').strip() client['direccion_apt'] = request.form.get('direccion_apt', '').strip() client['direccion_area'] = request.form.get('direccion_area', '').strip() client['ciudad'] = request.form.get('ciudad', '').strip() client['telefonofij'] = request.form.get('telefonofij', '').strip() client['telefonocel'] = request.form.get('telefonocel', '').strip() if request.form.get('tarifa', None): try: client['tarifa'] = float(request.form.get('tarifa', 0)) except: flash('El numero de tarifa que ha introducido no es valido') actItems['tarifa'] = True actItems['FT'] = True try: if client['telefonofij'][:3] not in ('507', '506'): client['telefonofij'] = '507%s' % (client['telefonofij']) if client['telefonocel'][:3] not in ('507', '506'): client['telefonocel'] = '507%s' % (client['telefonocel']) client['telefonofij'] = int(client['telefonofij']) client['telefonocel'] = int(client['telefonocel']) except: flash('El numero de telefono que ha introducido no es valido') actItems['telefonofij'] = True actItems['FT'] = True try: client['telefonocel'] = int(client['telefonocel']) except: flash('El numero de telefono que ha introducido no es valido') actItems['telefonocel'] = True actItems['FT'] = True update = request.form.get('update', '').strip() if int(update) > 0: if cid == '0': ncid = insert_account(contacto_nombre_1=client['contacto_nombre_1'], contacto_nombre_2=client['contacto_nombre_2'], contacto_apellido_1=client['contacto_apellido_1'], contacto_apellido_2=client['contacto_apellido_2'], telefonofij=client['telefonofij'], telefonocel=client['telefonocel'], correo=client['correo'], direccion_calle=client['direccion_calle'], direccion_torre=client['direccion_torre'], direccion_apt=client['direccion_apt'], direccion_area=client['direccion_area'], ciudad=client['ciudad'], tarifa=client['tarifa']) where = 'VMB_ACCOUNTS_id = %s' % (ncid) client = get_contact(where=where) cid = client['casillero'] else: set_contact_by_casillero(newClient=client, casillero=cid, updatedUser=current_user.username) return redirect(url_for('viewClient', cid=cid)) return render_template('clients/edit.html', client=client, actItems=actItems)
{ "repo_name": "cindy-zimmerman/vmb-mrest", "path": "flask_mrest/actions/client.py", "copies": "1", "size": "5538", "license": "mit", "hash": -1712777589072555300, "line_mean": 45.15, "line_max": 112, "alpha_frac": 0.5545323221, "autogenerated": false, "ratio": 3.496212121212121, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9520321808392898, "avg_score": 0.0060845269838449186, "num_lines": 120 }
from ps2_1 import evaluate_poly from ps2_2 import compute_deriv def compute_root(poly, x_0, epsilon): """ Uses Newton's Method to find and return a root of polynomial function. Returns a tuple containing the root and the number of iterations required to get the root. Example: >>> poly = (-13.39, 0.0, 17.5, 3.0, 1.0) x^4 + 3.0x^3 + 17.5x^2 - 13.39 >>> x_0 = 0.1 >>> epsilon = 0.0001 >>> print compute_root(poly, x_0, epsilon) (0.80679075379635201, 8) poly: tuple of numbers, length > 1. Represent a polynomial function containing at least one real root. The derivative of this polynomial function at x_0 is not 0. x_0: float epsilon: float > 0 returns: tuple (float, int) """ count = 0 while True: count += 1 result = evaluate_poly(poly, x_0) print 'x_0 =', x_0 if abs(result) <= epsilon: break else: x_0 = x_0 - (evaluate_poly(poly, x_0))/(evaluate_poly((compute_deriv(poly)), x_0)) return (x_0, count) ## testing.. if __name__ == '__main__': poly = (-13.39, 0.0, 17.5, 3.0, 1.0) x_0 = 0.1 epsilon = 0.0001 print 'The root and number of itearaion is\n', compute_root(poly, x_0, epsilon)
{ "repo_name": "MarcusHolloway/6.00SC", "path": "ps2_3.py", "copies": "1", "size": "1241", "license": "mit", "hash": -3367153847151883000, "line_mean": 30.025, "line_max": 85, "alpha_frac": 0.6518936342, "autogenerated": false, "ratio": 2.7334801762114536, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.8680802342936749, "avg_score": 0.04091429349494095, "num_lines": 40 }
import random import string WORD_LIST = "words.txt" def load_words(): """ Returns a list of valid words. Words are string of lowercase letters. Depending on the size of the word list, this function may take a while to finish. """ print "Loading word list from file..." fin = open(WORD_LIST, 'r', 0) line = fin.readline() wordlist = string.split(line) ##print wordlist print ' ', len(wordlist), 'words loaded.' return wordlist def choose_word(wordlist): """ wordlist: list of words Returns a word from wordlist at random. """ return random.choice(wordlist) wordlist = load_words() def part_of_word(mysterious_word, guessed_letters): result = '' for letter in mysterious_word: if letter in guessed_letters: result += letter else: result += '_' return result def hangman(): attemps = 8 isGuessed = False guessed_letters = '' mysterious_word = choose_word(wordlist) available_letters = ['a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z'] print 'Welcome to the game, Hangman!' print 'I am thinking of a word that is %d letters long.' %len(mysterious_word) while attemps > 0 and not isGuessed: print '--------------' print 'You have %d guesses left.' %attemps print 'Available letters: %s' %''.join(available_letters) guessed_char = raw_input('Please guess a letter:') if guessed_char not in available_letters: print 'Opps! You already enter that letter: ' + part_of_word(mysterious_word, guessed_letters) elif guessed_char not in mysterious_word: attemps -= 1 available_letters.remove(guessed_char) print 'Oops! That letter is not in my word: ' + part_of_word(mysterious_word, guessed_letters) else: available_letters.remove(guessed_char) guessed_letters += guessed_char print 'Good guess: ' + part_of_word(mysterious_word, guessed_letters) if mysterious_word == part_of_word(mysterious_word, guessed_letters): isGuessed = True if isGuessed: print 'Congratulations, you won!' else: print 'Oops! You lost!'
{ "repo_name": "MarcusHolloway/6.00SC", "path": "ps2_4.py", "copies": "1", "size": "2507", "license": "mit", "hash": 7463818048138955000, "line_mean": 33.3098591549, "line_max": 106, "alpha_frac": 0.5791783008, "autogenerated": false, "ratio": 3.5916905444126073, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9623952940217677, "avg_score": 0.009383180998986215, "num_lines": 71 }
def cal_credit_balance(): outstanding_balance = float(raw_input('Enter the outstanding balance on your credit card:')) annual_interest_rate = float(raw_input('Enter the annual credit card interest rate as decimal:')) min_monthly_payment_rate = float(raw_input('Enter the minimum monthly payment rate as decimal:')) monthly_interest_rate = annual_interest_rate / 12.0 months = 0 total_amount = 0 balance = outstanding_balance while(months < 12): min_monthly_payment = min_monthly_payment_rate * balance interest_paid = monthly_interest_rate * balance principal_paid = min_monthly_payment - interest_paid remaining_balance = balance - principal_paid months += 1 print 'Month:', months print 'Minimum monthly payment: $' + str(round(min_monthly_payment, 2)) print 'Principal paid: $' + str(round(principal_paid, 2)) print 'Remaining balance: $' + str(round(remaining_balance, 2)) total_amount += min_monthly_payment balance = remaining_balance print 'RESULT' print 'Total amount paid: $' + str(round(total_amount, 2)) print 'Remaining balance: $' + str(round(remaining_balance, 2)) ## testing... if __name__ == '__main__': cal_credit_balance() input('Press Enter to Exit!!')
{ "repo_name": "MarcusHolloway/6.00SC", "path": "ps_1a.py", "copies": "1", "size": "1386", "license": "mit", "hash": 2886143291776234500, "line_mean": 38.7647058824, "line_max": 101, "alpha_frac": 0.63997114, "autogenerated": false, "ratio": 3.860724233983287, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5000695373983287, "avg_score": null, "num_lines": null }
def cal_min_fixed_monthly_payment(): outstanding_balance = float(raw_input('Enter the outstanding balance on your credit card:')) annual_interest_rate = float(raw_input('Enter the annual credit card interest rate as decimal:')) monthly_interest_rate = annual_interest_rate / 12.0 updated_balance = outstanding_balance min_month_payment = 0.0 while updated_balance > 0.0: months = 0 min_month_payment += 10.0 previous_balance = outstanding_balance while months < 12 and updated_balance > 0.0: months += 1 updated_balance = previous_balance *(1 + monthly_interest_rate) - min_month_payment previous_balance = updated_balance print 'RESULT' print 'Monthly payment to pay off debt in 1 year:$' + str(round(min_month_payment, 2)) print 'Number of months needed:' + str(months) print 'Balance: $' + str(round(updated_balance, 2)) ## testing.. if __name__ == '__main__': cal_min_fixed_monthly_payment()
{ "repo_name": "MarcusHolloway/6.00SC", "path": "ps_1b.py", "copies": "1", "size": "1089", "license": "mit", "hash": 35078077625912436, "line_mean": 35.5517241379, "line_max": 101, "alpha_frac": 0.6271808999, "autogenerated": false, "ratio": 3.7294520547945207, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.4856632954694521, "avg_score": null, "num_lines": null }
def cal_min_payment(): outstanding_balance = float(raw_input('Enter the outstanding balance on your credit card:')) annual_interest_rate = float(raw_input('Enter the annual credit card interest rate as decimal:')) monthly_interest_rate = annual_interest_rate / 12.0 # New variables epsilon = 0.001 lower_payment_bound = outstanding_balance / 12.0 upper_payment_bound = (outstanding_balance * (1 + monthly_interest_rate)**12) / 12.0 min_monthly_payment = (upper_payment_bound + lower_payment_bound) / 2.0 month = 0 while True: previous_balance = outstanding_balance min_monthly_payment = (upper_payment_bound + lower_payment_bound) / 2.0 for month in range(1, 13): updated_balance = previous_balance * (1 + monthly_interest_rate) - min_monthly_payment previous_balance = updated_balance if updated_balance < 0: break if updated_balance == 0 or(upper_payment_bound - lower_payment_bound) / 2.0 < epsilon: break elif updated_balance < 0: upper_payment_bound = min_monthly_payment else: lower_payment_bound = min_monthly_payment print 'RESULT' print 'Monthly payment to pay off debt in 1 year:$' + str(round(min_monthly_payment, 2)) print 'Number of months needed:' + str(month) print 'Balance: $' + str(round(updated_balance, 2)) ## testing.. if __name__ == '__main__': cal_min_payment()
{ "repo_name": "MarcusHolloway/6.00SC", "path": "ps_1c.py", "copies": "1", "size": "1650", "license": "mit", "hash": 9053954728276165000, "line_mean": 37.2857142857, "line_max": 101, "alpha_frac": 0.6212121212, "autogenerated": false, "ratio": 3.793103448275862, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.9819221466411576, "avg_score": 0.019018820612857108, "num_lines": 42 }
__author__ = 'Daan Debie' from abc import ABCMeta from abc import abstractmethod from datetime import datetime def datetime_from_http_datestring(datestring): """ Converts a datestring as present in HTTP headers to a python datetime object """ return datetime.strptime(datestring, '%a, %d %b %Y %H:%M:%S %Z') def http_datestring_from_datetime(dt): # HTTP datetimes are always in TZ GMT return dt.strftime('%a, %d %b %Y %H:%M:%S GMT') class PlaylistCache(object): """ Abstract Base Class representing the cache for PlaylistItems This must be subclassed to implement a caching method. The subclass is responisble for serializing and deserializing PlaylistItem objects. """ __metaclass__ = ABCMeta @abstractmethod def get(self, key): """ Get a PlaylistItem from the cache """ pass @abstractmethod def put(self, key, value): """ Put a PlaylistItem in the cache """ pass @abstractmethod def remove(self, key): """ Remove a PlaylistItem from cache """ pass class PlaylistItem: def __init__(self, name, uri, last_modified, expires): self.name = name self.uri = uri self.last_modified = last_modified self.expires = expires def is_expired(self): return datetime.utcnow() > self.expires def __str__(self): return self.name class MemPlaylistCache(PlaylistCache): """ Naive caching implementation for in-memory caching, using a simple dictionary for storing PlaylistItems """ def __init__(self): self.cache = {} def get(self, key): if key in self.cache: return self.cache[key] else: return None def put(self, key, value): self.cache[key] = value def remove(self, key): del self.cache[key] def __str__(self): output = "" for key in self.cache.iterkeys(): output += "* {} / {} \n".format(key, self.cache[key].uri) return output
{ "repo_name": "davidsogn/PyPlaylist", "path": "playlist/cache.py", "copies": "1", "size": "2047", "license": "apache-2.0", "hash": 2729936508778319000, "line_mean": 23.6746987952, "line_max": 107, "alpha_frac": 0.6145578896, "autogenerated": false, "ratio": 4.0216110019646365, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5136168891564636, "avg_score": null, "num_lines": null }
__author__ = 'Daan Debie' from threading import Thread from Queue import Queue from generator import PlaylistGenerator class PLGeneratorThread(Thread): """ Class that provides threaded playlist generation. It takes one sentence off a queue and processes them. Results are stored in the Thread object itself. One of these Threads should be created for each message to be processed. Because of the average long running time of querying the API, race conditions or deadlocks should be rare to non-existant. Should one of the threads fail in some way, however, not all messages will be processed. """ def __init__(self, queue, generator): Thread.__init__(self) self.queue = queue self.generator = generator self.payload = None self.incomplete = True # If the Thread fails somehow, it should be considered incomplete self.playlist = None self.position = None def run(self): self.payload = self.queue.get() self.position = self.payload[1] print "running thread {}".format(str(self.position)) # We're processing multiple sentences, almost guaranteeing multiple playlist entries, # se we can use max_chunk_length self.playlist, self.incomplete = self.generator.generate_playlist(self.payload[0], True) self.queue.task_done() def generate_multiple_playlists_threaded(list_of_messages, cache): queue = Queue() generator = PlaylistGenerator(cache) # spawn a pool of threads, and pass them queue instance threads = [PLGeneratorThread(queue, generator) for message in list_of_messages] for thread in threads: thread.setDaemon(True) thread.start() # populate queue with data, also providing a position so we can sort them back later for x in range(len(list_of_messages)): queue.put((list_of_messages[x], x)) # wait on the queue until everything has been processed # NOTE: if a thread fails somehow (as in: doesn't execute at all), deadlock could occur, because part of the queue # won't be processed. queue.join() return [(thread.playlist, thread.incomplete) for thread in sorted(threads, key=lambda thread: thread.position)] def generate_multiple_playlists_naive(list_of_messages, cache): generator = PlaylistGenerator(cache) results = [generator.generate_playlist(message) for message in list_of_messages] return results
{ "repo_name": "davidsogn/PyPlaylist", "path": "playlist/plthreading.py", "copies": "1", "size": "2454", "license": "apache-2.0", "hash": 4595702250171069400, "line_mean": 39.2295081967, "line_max": 118, "alpha_frac": 0.7033414833, "autogenerated": false, "ratio": 4.312829525483304, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 0.5516171008783304, "avg_score": null, "num_lines": null }
__author__ = 'Daan Wierstra and Tom Schaul' from itertools import chain from scipy import zeros from pybrain.structure.networks.feedforward import FeedForwardNetwork from pybrain.structure.networks.recurrent import RecurrentNetwork from pybrain.structure.modules.neuronlayer import NeuronLayer from pybrain.structure.connections import FullConnection # CHECKME: allow modules that do not inherit from NeuronLayer? and treat them as single neurons? class NeuronDecomposableNetwork(object): """ A Network, that allows accessing parameters decomposed by their corresponding individual neuron. """ # ESP style treatment: espStyleDecomposition = True def addModule(self, m): assert isinstance(m, NeuronLayer) super(NeuronDecomposableNetwork, self).addModule(m) def sortModules(self): super(NeuronDecomposableNetwork, self).sortModules() self._constructParameterInfo() # contains a list of lists of indices self.decompositionIndices = {} for neuron in self._neuronIterator(): self.decompositionIndices[neuron] = [] for w in range(self.paramdim): inneuron, outneuron = self.paramInfo[w] if self.espStyleDecomposition and outneuron[0] in self.outmodules: self.decompositionIndices[inneuron].append(w) else: self.decompositionIndices[outneuron].append(w) def _neuronIterator(self): for m in self.modules: for n in range(m.dim): yield (m, n) def _constructParameterInfo(self): """ construct a dictionnary with information about each parameter: The key is the index in self.params, and the value is a tuple containing (inneuron, outneuron), where a neuron is a tuple of it's module and an index. """ self.paramInfo = {} index = 0 for x in self._containerIterator(): if isinstance(x, FullConnection): for w in range(x.paramdim): inbuf, outbuf = x.whichBuffers(w) self.paramInfo[index + w] = ((x.inmod, x.inmod.whichNeuron(outputIndex=inbuf)), (x.outmod, x.outmod.whichNeuron(inputIndex=outbuf))) elif isinstance(x, NeuronLayer): for n in range(x.paramdim): self.paramInfo[index + n] = ((x, n), (x, n)) else: raise index += x.paramdim def getDecomposition(self): """ return a list of arrays, each corresponding to one neuron's relevant parameters """ res = [] for neuron in self._neuronIterator(): nIndices = self.decompositionIndices[neuron] if len(nIndices) > 0: tmp = zeros(len(nIndices)) for i, ni in enumerate(nIndices): tmp[i] = self.params[ni] res.append(tmp) return res def setDecomposition(self, decomposedParams): """ set parameters by neuron decomposition, each corresponding to one neuron's relevant parameters """ nindex = 0 for neuron in self._neuronIterator(): nIndices = self.decompositionIndices[neuron] if len(nIndices) > 0: for i, ni in enumerate(nIndices): self.params[ni] = decomposedParams[nindex][i] nindex += 1 @staticmethod def convertNormalNetwork(n): """ convert a normal network into a decomposable one """ if isinstance(n, RecurrentNetwork): res = RecurrentDecomposableNetwork() for c in n.recurrentConns: res.addRecurrentConnection(c) else: res = FeedForwardDecomposableNetwork() for m in n.inmodules: res.addInputModule(m) for m in n.outmodules: res.addOutputModule(m) for m in n.modules: res.addModule(m) for c in chain(*list(n.connections.values())): res.addConnection(c) res.name = n.name res.sortModules() return res class FeedForwardDecomposableNetwork(NeuronDecomposableNetwork, FeedForwardNetwork): pass class RecurrentDecomposableNetwork(NeuronDecomposableNetwork, RecurrentNetwork): pass
{ "repo_name": "jackru/pybrain", "path": "pybrain/structure/networks/neurondecomposable.py", "copies": "25", "size": "4341", "license": "bsd-3-clause", "hash": -1961687884108220700, "line_mean": 36.4224137931, "line_max": 99, "alpha_frac": 0.6178299931, "autogenerated": false, "ratio": 4.2558823529411764, "config_test": false, "has_no_keywords": false, "few_assignments": false, "quality_score": 1, "avg_score": 0.0008518096198359535, "num_lines": 116 }