blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
2
616
content_id
stringlengths
40
40
detected_licenses
listlengths
0
69
license_type
stringclasses
2 values
repo_name
stringlengths
5
118
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringlengths
4
63
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
2.91k
686M
star_events_count
int64
0
209k
fork_events_count
int64
0
110k
gha_license_id
stringclasses
23 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
213 values
src_encoding
stringclasses
30 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
2
10.3M
extension
stringclasses
246 values
content
stringlengths
2
10.3M
authors
listlengths
1
1
author_id
stringlengths
0
212
9e08ed3d8b5fc2a5093ed686b9c5014984327021
9652657ce62bdbca2b0578d451ba5cf853a42400
/form/redisDB.py
0decff328a39fc332904a9ee2de4d16fef7c626e
[]
no_license
astange/ExpoProject
775ca1785032cfbc858c90dfddde5a1af22daf06
fc233802a078728b69e70ced112a94a4163053f4
refs/heads/master
2016-09-05T11:04:12.349586
2014-07-29T05:28:06
2014-07-29T05:28:06
null
0
0
null
null
null
null
UTF-8
Python
false
false
13
py
../redisDB.py
[ "anthony@Anthonys-Laptop.local" ]
anthony@Anthonys-Laptop.local
97450e3407268358d4f64aefe3120b8487b3401e
425db5a849281d333e68c26a26678e7c8ce11b66
/maths/fast_pow_and_matrix_multi.py
987f29bb269b191cf1b8759d9bc80770e1b3e800
[ "MIT" ]
permissive
lih627/python-algorithm-templates
e8092b327a02506086414df41bbfb2af5d6b06dc
a61fd583e33a769b44ab758990625d3381793768
refs/heads/master
2021-07-23T17:10:43.814639
2021-01-21T17:14:55
2021-01-21T17:14:55
238,456,498
29
8
null
null
null
null
UTF-8
Python
false
false
2,500
py
import random def fpowx(x, n): """ quick pow: x ** n """ res = 1 while n: if n & 1: res = res * x # compute x^2 x^4 x^8 x *= x n >>= 1 return res def fmulti(m, n, mod=10 ** 9 + 7): """ 并没有提速的效果 只是对于其他语言 如c 防止溢出 对 python 没有任何帮助 """ res = 0 while n: if n & 1: res += m m = (m + m) % mod res %= mod n >>= 1 return res def matrix_multiply(matrix_a, matrix_b): # 模 MOD 乘法/加法 MOD = 10 ** 9 + 7 n_row = len(matrix_a) n_col = len(matrix_b[0]) n_tmp = len(matrix_a[0]) matrix_c = [[0 for _ in range(n_col)] for _ in range(n_row)] for i in range(n_row): for j in range(n_col): for k in range(n_tmp): matrix_c[i][j] += matrix_a[i][k] * matrix_b[k][j] % MOD matrix_c[i][j] %= MOD return matrix_c def get_unit_matrix(n): # matrix I unit_matrix = [[0 for _ in range(n)] for _ in range(n)] for _ in range(n): unit_matrix[_][_] = 1 return unit_matrix def quick_matrix_pow(matrix_a, n): # A ^ n l = len(matrix_a) res = get_unit_matrix(l) while n: if n & 1: res = matrix_multiply(res, matrix_a) a = matrix_multiply(a, a) n >>= 1 return res def test_fmulti(): m = random.randint(10 ** 9, 10 ** 15) n = random.randint(10 ** 9, 10 ** 15) res = fmulti(m, n) return res def multi(m, n, mod=10 ** 9 + 7): return m * n % mod def test_multi(): m = random.randint(10 ** 9, 10 ** 15) n = random.randint(10 ** 9, 10 ** 15) res = multi(m, n) return res if __name__ == '__main__': print('fast pow: 2 ** 11: {}'.format(fpowx(2, 11))) print(fmulti(987654, 987654321)) print(987654 * 987654321 % (10 ** 9 + 7)) # test the speed of fast(?)-multi import timeit T_fmulti = timeit.Timer('test_fmulti()', 'from __main__ import test_fmulti') print('f_multi: {:.6f}s'.format(T_fmulti.timeit(number=1000))) T_multi = timeit.Timer('test_multi()', 'from __main__ import test_multi') print('s_multi: {:.6f}s'.format(T_multi.timeit(number=1000))) # test matrix multiply a = [[1, 2, 3], [4, 5, 6]] b = [[1, 2], [3, 4], [5, 6]] c = matrix_multiply(a, b) print("a = {}\nb = {}\nc = {}".format(a, b, c))
[ "lih627@outlook.com" ]
lih627@outlook.com
bf052740a6a34a83103709163327ad7948d74ae6
d9072f607b96701eb50cde1c184e29c83a27f597
/code/python/playground.py
bb1df391f8dc5604f024d7708ef0ccfa26e9d6ae
[]
no_license
glaukon-ariston/2007-TIP-HistogramSegmentation
f35111a87447241a8c67b77aac83485d80270f31
b8d250bbe80ee17fe33d56c6092bfc1049f45234
refs/heads/master
2022-12-28T17:50:11.269621
2020-10-09T08:42:23
2020-10-09T08:42:23
295,377,236
0
0
null
2020-09-14T10:08:29
2020-09-14T10:08:28
null
UTF-8
Python
false
false
5,001
py
#!/usr/bin/python # -*- coding: utf-8 -*- # # Author: Glaukon Ariston # Date: 26.09.2020 # Abstract: # Debugging harness for histogramSegmentation.fine_to_coarse_histogram_segmentation(H, e=0) # # (findpeaks in Python get different results than in MatLab) # print('idx = [%s];' % (' '.join(['%d' % (v+1,) for v in idxs]),)) # idx = [1 48 49 50 67 68 69 71 72 73 74 76 77 79 81 82 83 84 85 92 93 100 102 103 110 112 113 114 116 119 120 127 128 130 136 137 138 145 147 148 150 155 169 170 171 173 176 178 186 188 189 196 197 199 200 202 203 206 207 209 211 212 221 222 234 235 237 239 241 243 244 245 256]; # # print('val[%d] = [%s]' % (len(maxEntropies), ' '.join(['%f' % (v,) for v in maxEntropies]))) # val[70] = [-7.044258 -5.163717 469.061042 -5.331204 -2.602968 -1.984528 -2.600699 -2.296912 -2.674793 -1.536530 -2.815287 -2.116291 -2.991723 -2.703314 -1.531241 -0.981140 -3.991059 7.867301 -4.869646 69.468887 -2.245196 -4.134025 41.480205 -4.046769 -2.674656 -2.402819 -2.774352 0.178781 -4.328724 47.723254 -4.186474 -3.127704 19.860393 -3.645382 -3.957270 17.571696 -1.331898 -2.549457 -3.221580 12.974305 453.162900 -4.956052 -2.065192 -0.415594 -1.134470 -4.444906 107.802875 -3.603364 -3.993935 20.479040 -2.063956 -2.681444 -2.386676 -1.899041 -2.988679 -0.823756 -3.226691 -0.842533 -0.517108 -3.725309 223.395585 -5.520483 170.863611 -4.879836 0.966133 -3.032572 -2.957388 -2.141432 -2.179776 -4.337235]; # val[70] = [-7.044258 -5.163717 469.061042 -5.331204 -2.602968 -1.984528 -2.600699 -2.296912 -2.674793 -1.536530 -2.815287 -2.116291 -2.991723 -2.703314 -1.531241 -0.981140 -3.991059 7.867301 -4.869646 69.468887 -2.245196 -4.134025 41.480205 -4.046769 -2.674656 -2.402819 -2.774352 0.178781 -4.328724 47.723254 -4.186474 -3.127704 19.860393 -3.645382 -3.957270 17.571696 -1.331898 -2.549457 -3.221580 12.974305 453.162900 -4.956052 -2.065192 -0.415594 -1.134470 -4.444906 107.802875 -3.603364 -3.993935 20.479040 -2.063956 -2.681444 -2.386676 -1.899041 -2.988679 -0.823756 -3.226691 -0.842533 -0.517108 -3.725309 223.395585 -5.520483 170.863611 -4.879836 0.966133 -3.032572 -2.957388 -2.141432 -2.179776 -4.337235]; # val[68] = [-7.044258 -5.331204 -2.602968 -1.984528 -2.600699 -2.296912 -2.674793 -1.536530 -2.815287 -2.116291 -2.991723 -2.703314 -1.531241 -0.981140 -3.991059 7.867301 -4.869646 69.468887 -2.245196 -4.134025 41.480205 -4.046769 -2.674656 -2.402819 -2.774352 0.178781 -4.328724 47.723254 -4.186474 -3.127704 19.860393 -3.645382 -3.957270 17.571696 -1.331898 -2.549457 -3.221580 12.974305 453.162900 -4.956052 -2.065192 -0.415594 -1.134470 -4.444906 107.802875 -3.603364 -3.993935 20.479040 -2.063956 -2.681444 -2.386676 -1.899041 -2.988679 -0.823756 -3.226691 -0.842533 -0.517108 -3.725309 223.395585 -5.520483 170.863611 -4.879836 0.966133 -3.032572 -2.957388 -2.141432 -2.179776 -4.337235]; # # GNU Octave # logId = fopen('val_matlab.m', 'w'); # fprintf(logId, "val[%d] = [%s]\n", length(val), sprintf("%f ", val)) # fprintf(logId, 'j %d inc %d val(j) %f\n', j, inc, val(j)) # val[68] = [-7.044258 -5.331204 -2.602968 -1.984528 -2.600699 -2.296912 -2.674793 -1.536530 -2.815 287 -2.116291 -2.991723 -2.703314 -1.531241 -0.981140 -3.991059 7.867301 -4.869646 69.468887 -2.2 45196 -4.134025 41.480205 -4.046769 -2.674656 -2.402819 -2.774352 0.178781 -4.328724 47.723254 -4 .186474 -3.127704 19.860393 -3.645382 -3.957270 17.571696 -1.331898 -2.549457 -3.221580 12.974305 453.162900 -4.956052 -2.065192 -0.415594 -1.134470 -4.444906 107.802875 -3.603364 -3.993935 20.4 79040 -2.063956 -2.681444 -2.386676 -1.899041 -2.988679 -0.823756 -3.226691 -0.842533 -0.517108 - 3.725309 223.395585 -5.520483 170.863611 -4.879836 0.966133 -3.032572 -2.957388 -2.141432 -2.1797 76 -4.337235 ] import imageio from PIL import Image import numpy as np import matplotlib.pyplot as plt from skimage import filters from scipy import signal import histogramSegmentation import math import os.path import os import glob def playgroundHistogramSegmentation(imagePath): img = np.array(Image.open(imagePath).convert('L')) #histogram, bins = np.histogram(img.ravel(), 256, [0,256]) histogram = np.bincount(img.ravel(), minlength=256) plt.plot(histogram) H = histogram N = histogram.size #idxs_max, _ = signal.find_peaks(H) #idxs_min, _ = signal.find_peaks(-H) #idxs = np.sort(np.concatenate([np.array([0]), idxs_min, idxs_max, np.array([N-1])])) #for i in range(idxs_max.size): # plt.axvline(x=idxs_max[i], color='cyan') #for i in range(idxs_min.size): # plt.axvline(x=idxs_min[i], color='magenta') indices = histogramSegmentation.fine_to_coarse_histogram_segmentation(H, e=0) for i in range(indices.size): plt.axvline(x=indices[i], color='cyan') #_ = plt.hist(img.ravel(), 256, [0,256]) #plt.imshow(img, cmap='gray', interpolation='none') plt.title("Histogram") plt.show() def main(): playgroundHistogramSegmentation('../../images/lena.png') if __name__ == "__main__": main()
[ "glaukon.ariston@gmail.com" ]
glaukon.ariston@gmail.com
15efb02d15bd410d19b8018e6c307a75b9f04eb4
5fdcb39eaa9d1f44e2ba0130bc0d6ece3f5ff354
/code/cheshire3/record.py
728ddee847b4f75341864dad4eff2244263dd3c4
[]
no_license
Cheshire-Grampa/cheshire3
0a653d6372497290d938e098b6acf8366348133f
616ab36cd8442cd5f4712a9fccf65ca7ae9f692c
refs/heads/master
2020-12-25T07:26:16.366754
2012-06-06T09:52:53
2012-06-06T10:32:17
null
0
0
null
null
null
null
UTF-8
Python
false
false
41,269
py
from cheshire3.baseObjects import Record from cheshire3.exceptions import C3Exception from cheshire3.utils import flattenTexts, elementType from cheshire3.marc_utils import MARC, MARC8_to_Unicode import unicodedata import types, utils, os, re, sys from cStringIO import StringIO from xml.sax.saxutils import escape from xml.sax import ContentHandler # 1 <name> <attrHash> parent predicate end # Element # 4 <as 1> # Namespaced Element # 2 <name> <startLine> # End Element # 5 <as 2> # End Namespaced # 3 <text> # Characters # 9 <element hash> # pickled hash of locations # Split to separate object to allow for DOM->SAX direct conversion # by throwing events from DOM tree to handler. class SaxContentHandler(ContentHandler): currentText = [] currentPath = [] pathLines = [] currentLine = -1 recordWordCount = 0 elementHash = {} namespaces = [] hashAttributesNames = {} hashAttributes = [] stripWS = 0 saveElementIndexes = 1 def __init__(self): self.saveElementIndexes = 1 self.hashAttributesNames = {} self.hashAttributes = [] self.stripWS = 0 self.reinit() def reinit(self): self.currentText = [] self.currentPath = [] self.pathLines = [] self.currentLine = -1 self.recordWordCount = 0 self.elementHash = {} self.elementIndexes = [] self.namespaces = [] def startPrefixMapping(self, pfx, uri): self.currentLine += 1 if (pfx == None): pfx = '' self.currentText.append("6 %r, %r" % (pfx, uri)) # We want to fwd elems to NS elem handlers with default NS? def startElement(self, name, attrs): self.currentLine += 1 self.pathLines.append(self.currentLine) try: parent = self.pathLines[-2] except IndexError: parent = -1 attrHash = {} if (attrs): for k in attrs.keys(): attrHash[k] = escape(attrs[k]) try: npred = self.elementIndexes[-1][name] + 1 self.elementIndexes[-1][name] += 1 except IndexError: # Empty npred = 1 self.elementIndexes = [{name: npred}] except KeyError: # First occurence of Element npred = 1 self.elementIndexes[-1][name] = 1 except: raise self.elementIndexes.append({}) self.currentText.append("1 %s %s %d %d" % (name, repr(attrHash), parent, npred)) saveAttrs = [] try: hashAttrList = self.hashAttributesNames[name] for a in hashAttrList: try: saveAttrs.append("%s[@%s='%s']" % (name, a, attrHash[a])) except: pass except: pass try: starAttrList = self.hashAttributesNames['*'] for a in starAttrList: try: saveAttrs.append("*[@%s='%s']" % (a, attrHash[a])) except: pass except: pass if saveAttrs: self.hashAttributes.append((self.currentLine, saveAttrs)) def endElement(self, name): self.currentLine += 1 start = self.pathLines.pop() self.currentText.append("2 %s %d" % (name, start)) self.currentText[start] = "%s %d" % (self.currentText[start], self.currentLine) self.elementIndexes.pop() try: self.elementHash[name].append([start, self.currentLine]) except: self.elementHash[name] = [[start, self.currentLine]] if self.hashAttributes and self.hashAttributes[-1][0] == start: attrs = self.hashAttributes.pop()[1] for sa in attrs: try: self.elementHash[sa].append([start, self.currentLine]) except: self.elementHash[sa] = [[start, self.currentLine]] def startElementNS(self, name, qname, attrs): self.currentLine += 1 self.pathLines.append(self.currentLine) try: parent = self.pathLines[-2] except: parent = -1 attrHash = {} # Convert from weird sax thing if (attrs): for k in attrs.keys(): attrHash[k] = attrs[k] simpleName = name[1] try: npred = self.elementIndexes[-1][simpleName] + 1 self.elementIndexes[-1][simpleName] += 1 except IndexError: # Empty npred = 1 self.elementIndexes = [{simpleName: npred}] except KeyError: # First occurence of Element npred = 1 self.elementIndexes[-1][simpleName] = 1 self.elementIndexes.append({}) self.currentText.append("4 %r, %r, %r, %r %d %d" % (name[0], simpleName, qname, attrHash, parent, npred)) saveAttrs = [] try: hashAttrList = self.hashAttributesNames[simpleName] for a in hashAttrList: try: saveAttrs.append("%s[@%s='%s']" % (simpleName, a, attrHash[a])) except: pass except: pass try: starAttrList = self.hashAttributesNames['*'] for a in starAttrList: try: saveAttrs.append("*[@%s='%s']" % (a, attrHash[a])) except: pass except: pass if saveAttrs: self.hashAttributes.append((self.currentLine, saveAttrs)) def endElementNS(self, name, qname): self.currentLine += 1 start = self.pathLines.pop() self.currentText.append("5 %r, %r, %r %d" % (name[0], name[1], qname, start)) self.currentText[start] ="%s %d" % (self.currentText[start], self.currentLine) self.elementIndexes.pop() try: self.elementHash[name[1]].append([start, self.currentLine]) except: self.elementHash[name[1]] = [[start, self.currentLine]] if self.hashAttributes and self.hashAttributes[-1][0] == start: attrs = self.hashAttributes.pop()[1] for sa in attrs: try: self.elementHash[sa].append([start, self.currentLine]) except: self.elementHash[sa] = [[start, self.currentLine]] def characters(self, text, start=0, length=-1): # if text.isspace(): # text = " " prev = self.currentText[-1] if self.stripWS and text.isspace(): return self.currentLine += 1 if (len(text) != 1 and len(prev) != 3 and prev[0] == "3" and not prev[-1] in [' ', '-']): # Adjacent lines of text, ensure spaces text = ' ' + text self.currentText.append("3 %s" % (text)) self.recordWordCount += len(text.split()) def ignorableWhitespace(self, ws): # ... ignore! :D pass def processingInstruction(self, target, data): pass def skippedEntity(self, name): pass class SaxToDomHandler: nodeStack = [] document = None currText = "" def initState(self): self.nodeStack = [] self.document=None self.top = None def startElement(self, name, attribs={}): if (not self.document): self.document = implementation.createDocument(None, name, None) elem = self.document.childNodes[0] else: elem = self.document.createElementNS(None,name) for a in attribs: elem.setAttributeNS(None,a,attribs[a]) if (self.nodeStack): self.nodeStack[-1].appendChild(elem) else: self.document.appendChild(elem) self.nodeStack.append(elem) def endElement(self, foo): self.nodeStack.pop() def characters(self, text, zero=0, length=0): if (self.nodeStack): if (text.isspace()): text = " " # Is this escape necessary? text = escape(text) d = self.document.createTextNode(text) self.nodeStack[-1].appendChild(d) def startElementNS(self, name, qname, attribs): if (not self.document): self.document = implementation.createDocument(name[0], name[1], None) elem = self.document.childNodes[0] else: elem = self.document.createElementNS(name[0],name[1]) for a in attribs: elem.setAttributeNS(a[0],a[1],attribs[a]) if (self.nodeStack): self.nodeStack[-1].appendChild(elem) else: self.document.appendChild(elem) self.nodeStack.append(elem) def endElementNS(self, name,qname): self.nodeStack.pop() def startPrefixMapping(self, pref, uri): pass def getRootNode(self): return self.document s2dhandler = SaxToDomHandler() class SaxToXmlHandler: xml = [] currNs = 0 newNamespaces = {} def initState(self): self.xml = [] self.namespaces = {} self.currNs = 0 self.newNamespaces = {} def startPrefixMapping(self, pref, uri): self.namespaces[uri] = pref self.newNamespaces[pref] = uri def startElement(self, name, attribs={}): attrs = [] for a in attribs: attrs.append('%s="%s"' % (a, attribs[a])) attribtxt = ' '.join(attrs) if (attribtxt): attribtxt = " " + attribtxt self.xml.append("<%s%s>" % (name, attribtxt)) def endElement(self, name): self.xml.append("</%s>" % (name)) def _getPrefix(self, ns): if (not ns): return "" pref = self.namespaces.get(ns, None) if (pref == None): self.currNs += 1 pref = "ns%d" % (self.currNs) self.namespaces[ns] = pref self.newNamespaces[pref] = ns return pref def startElementNS(self, n, qn=None, attrs={}): pref = self._getPrefix(n[0]) if (pref): name = "%s:%s" % (pref, n[1]) else: name = n[1] attrlist = [] for ns,aname in attrs: p2 = self._getPrefix(ns) if (p2): nsaname = "%s:%s" % (p2, aname) else: nsaname = aname attrlist.append('%s="%s"' % (nsaname, attrs[(ns,aname)])) for x in self.newNamespaces.iteritems(): if (x[0]): attrlist.append('xmlns:%s="%s"' % (x[0], x[1])) else: attrlist.append('xmlns="%s"' % (x[1])) self.newNamespaces = {} attribtxt = ' '.join(attrlist) if (attribtxt): attribtxt = " " + attribtxt self.xml.append("<%s%s>" % (name,attribtxt)) def endElementNS(self, n, qn=None): pref = self._getPrefix(n[0]) if (pref): name = "%s:%s" % (pref, n[1]) else: name = n[1] self.xml.append("</%s>" % (name)) def characters(self, text, zero=0, length=0): text = escape(text) self.xml.append(text) def get_xmlString(self): return ''.join(self.xml) s2xhandler = SaxToXmlHandler() class NumericPredicateException(C3Exception): pass class DomRecord(Record): context = None size = 0 def __init__(self, data, xml="", docId=None, wordCount=0, byteCount=0): self.dom = data self.xml = xml self.id = docId self.parent = ('','',-1) self.context = None self.metadata = {} if wordCount: self.wordCount = wordCount else: try: # Sometimes this blows up self.wordCount = len(flattenTexts(data).split()) except: self.wordCount = 0 self.byteCount = byteCount def _walk(self, node): pass def get_sax(self, session): if (not self.sax): self.handler = SaxContentHandler() for c in self.dom.childNodes: self._walkTop(c) self.sax = self.handler.currentText self.sax.append("9 %r" % self.handler.elementHash) self.handler = None return self.sax def get_dom(self, session): return self.dom def fetch_vector(self, session, index, summary=False): return index.indexStore.fetch_vector(session, index, self, summary) def fetch_proxVector(self, session, index, elem=-1): return index.indexStore.fetch_proxVector(session, index, self, elem) class MinidomRecord(DomRecord): useNamespace = 1 def get_xml(self, session): if (self.xml): return self.xml else: self.xml = self.dom.toxml() return self.xml def _walkTop(self, node): # top level node if node.nodeType == utils.elementType: self.namespaces = node.namespaceURI != None self._walk(node) def _walk(self, node): if (node.nodeType == utils.elementType): name = node.localName ns = node.namespaceURI attrHash = {} for ai in range(node.attributes.length): attr = node.attributes.item(ai) if self.namespaces: if attr.namespaceURI == 'http://www.w3.org/2000/xmlns/': self.handler.startPrefixMapping(attr.localName, attr.value) else: attrHash[(attr.namespaceURI, attr.localName)] = attr.value else: attrHash[attr.localName] = attr.value if self.namespaces: self.handler.startElementNS((node.namespaceURI, node.localName), None, attrHash) else: self.handler.startElement(node.localName, attrHash) for c in node.childNodes: self._walk(c) if self.namespaces: self.handler.endElementNS((node.namespaceURI, node.localName), None) else: self.handler.endElement(node.localName) elif node.nodeType == utils.textType: self.handler.characters(node.data) def process_xpath(self, session, xpath, maps={}): raise NotImplementedError try: from lxml import etree, sax class LxmlRecord(DomRecord): def process_xpath(self, session, xpath, maps={}): global prefixRe if (isinstance(xpath, list)): xpath = repr(xpath[0]) if xpath[0] != "/" and xpath[-1] != ')': xpath = "//" + xpath if maps: retval = self.dom.xpath(xpath, namespaces=maps) else: retval = self.dom.xpath(xpath) if isinstance(retval, list): return retval else: return [retval] def get_xml(self, session): return etree.tostring(self.dom) def get_sax(self, session): if (not self.sax): handler = SaxContentHandler() sax.saxify(self.dom, handler) self.sax = handler.currentText self.sax.append("9 %r" % handler.elementHash) return self.sax def get_dom(self, session): try: return self.dom.getroot() except AttributeError: return self.dom except: class LxmlRecord(DomRecord): pass try: from xpath import ParsedRelativeLocationPath, ParsedAbsoluteLocationPath, \ ParsedStep, ParsedNodeTest, ParsedExpr, Compile, \ ParsedAbbreviatedAbsoluteLocationPath, ParsedAbbreviatedRelativeLocationPath, \ ParsedNodeTest except: # This means we can't do xpaths on SaxRecords... # making them a bit pointless, but not fatal as we likely don't need them pass def traversePath(node): if (isinstance(node, ParsedRelativeLocationPath.ParsedRelativeLocationPath)): left = traversePath(node._left) right = traversePath(node._right) if (left == []): # self::node() return [right] elif (type(left[0]) in types.StringTypes): return [left, right] else: left.append(right) return left elif (isinstance(node, ParsedAbsoluteLocationPath.ParsedAbsoluteLocationPath)): left = ['/'] if (node._child): right = traversePath(node._child) else: return left if (type(right[0]) == types.StringType): return [left, right] else: left.extend(right) return left elif (isinstance(node, ParsedAbbreviatedRelativeLocationPath.ParsedAbbreviatedRelativeLocationPath)): left = traversePath(node._left) right = traversePath(node._right) right[0] = 'descendant' if (left == []): # self::node() return [right] elif (type(left[0]) in types.StringTypes): return [left, right] else: left.append(right) return left elif (isinstance(node, ParsedStep.ParsedStep)): # TODO: Check that axis is something we can parse a = node._axis._axis if (a == 'self'): return [] n = node._nodeTest local = ParsedNodeTest.NodeNameTest nameattr = "_nodeName" if (isinstance(n, local)): n = getattr(n, nameattr) elif (isinstance(n, ParsedNodeTest.TextNodeTest)): n = "__text()" elif (isinstance(n, ParsedNodeTest.QualifiedNameTest)): n = n._prefix + ":" + n._localName elif (isinstance(n, ParsedNodeTest.PrincipalTypeTest)): n = "*" else: raise(NotImplementedError) preds = node._predicates pp = [] if (preds): for pred in preds: pp.append(traversePath(pred)) return [a, n, pp] elif (isinstance(node, ParsedExpr.ParsedEqualityExpr) or isinstance(node, ParsedExpr.ParsedRelationalExpr)): # @id="fish" op = node._op # Override check for common: [position()=int] if (op == '=' and isinstance(node._left, ParsedExpr.FunctionCall) and node._left._name == 'position' and isinstance(node._right, ParsedExpr.ParsedNLiteralExpr)): return node._right._literal left = traversePath(node._left) if (type(left) == types.ListType and left[0] == "attribute"): left = left[1] right = traversePath(node._right) if not op in ('=', '!='): op = ['<', '<=', '>', '>='][op] return [left, op, right] elif (isinstance(node, ParsedExpr.ParsedNLiteralExpr) or isinstance(node, ParsedExpr.ParsedLiteralExpr)): # 7 or "fish" return node._literal elif (isinstance(node, ParsedExpr.FunctionCall)): if (node._name == 'last'): # Override for last using Pythonic expr return -1 elif node._name == 'name': return ['FUNCTION', '__name()'] elif node._name == 'starts-with': # only for foo[starts-with(@bar, 'baz')] return ['FUNCTION', 'starts-with', traversePath(node._arg0)[1], node._arg1._literal] elif node._name == 'regexp': return ['FUNCTION', 'regexp', traversePath(node._arg0)[1], re.compile(node._arg1._literal)] elif node._name == 'count': return ['FUNCTION', 'count', traversePath(node._arg0)] else: raise(NotImplementedError) elif (isinstance(node, ParsedExpr.ParsedAndExpr)): return [traversePath(node._left), 'and', traversePath(node._right)] elif (isinstance(node, ParsedExpr.ParsedOrExpr)): return [traversePath(node._left), 'or', traversePath(node._right)] else: # We'll need to do full XPath vs DOM raise(NotImplementedError) def parseOldXPath(p): xpObj = Compile(p) t = traversePath(xpObj) if (t[0] <> '/' and type(t[0]) in types.StringTypes): t= [t] return [xpObj, t] class SaxRecord(Record): def __init__(self, data, xml="", docId=None, wordCount=0, byteCount=0): self.sax = data self.id = docId self.xml = xml self.history = [] self.rights = [] self.elementHash = {} self.wordCount = wordCount self.byteCount = byteCount self.parent = ('','',-1) self.attrRe = re.compile("u['\"](.+?)['\"]: u['\"](.*?)['\"](, |})") #self.attrRe = re.compile("u(?P<quote>['\"])(.+?)(?P=quote): u(?P<quoteb>['\"])(.*?)(?P=quoteb)(, |})") self.recordStore = "" def process_xpath(self, session, xpath, maps={}): if (not isinstance(xpath, list)): # Raw XPath xpath = parseOldXPath(xpath) xp = xpath[1] try: flatten = 0 if xp[0][0] == "FUNCTION" and xp[0][1] == 'count': # process xpath and return number of matches if isinstance(xp[0][2][0], str) and xp[0][2][0] != '/': data = self.process_xpath(session, [None, [xp[0][2]]], maps) else: data = self.process_xpath(session, [None, xp[0][2]], maps) return len(data) if (xp[-1][0] == 'child' and xp[-1][1] == "__text()"): flatten = 1 xp = xp[:-1] if (xp[-1][0] == 'attribute'): return self._handleAttribute(xp, maps) elif (xp[-1][0] == "/"): # Return top level element for x in xrange(len(self.sax)): if self.sax[x][0] in ['1', '4']: return self.sax[x:] elif(xp[-1][0] in ['child', 'descendant']): data = [] # Extracting element elemName = xp[-1][1] nselem = elemName.split(":") if (len(nselem) == 2): # Namespaced. nsUri = maps[nselem[0]] elemName = nselem[1] else: nsUri = "" attr = xp[-1][2] elemLines = [] if elemName == '*' and attr: for p in attr: if p[0] == 'FUNCTION' and p[2] == '__name()': names = self.elementHash.keys() if p[1] == 'starts-with' and p[2] == '__name()': for x in names: if x.find(p[3]) == 0: elemLines.extend(self.elementHash[x]) elif p[1] == 'regexp' and p[2] == '__name()': for x in names: if p[3].search(x): elemLines.extend(self.elementHash[x]) elif (not elemName in self.elementHash): return [] if (len(attr) == 1 and type(attr[0]) == types.ListType and attr[0][1] == "="): n = u"%s[@%s='%s']" % (elemName, attr[0][0], attr[0][2]) elemLines = self.elementHash.get(n, []) if elemLines == []: try: elemLines = self.elementHash[elemName] except: # might really be empty pass for e in elemLines: if (not nsUri or self.sax[e[0]][4:4+len(nsUri)] == nsUri): match = self._checkSaxXPathLine(xp, e[0]) if (match): # Return event chunk l = self.sax[e[0]] end = int(l[l.rfind(' ')+1:]) data.append(self.sax[e[0]:end+1]) else: # Unsupported final axis raise(NotImplementedError) if flatten and data: # Flatten to text nodes ndata = [] for match in data: txt = [] for ev in match: if ev[0] == '3': txt.append(ev[2:]) ndata.append(''.join(txt)) return ndata else: return data except NotImplementedError: # Convert to DOM (slow) and reapply (slower still) dom = self.get_dom(session) xp = xpTuple[0] try: return utils.evaluateXPath(xp, dom) except: self.log_critical("Buggy Xpath: %r" % xp) return [] # Otherwise just fall over as we've hit a real bug def _handleAttribute(self, xp, maps={}): attrName = xp[-1][1] nselem = attrName.split(":") if (len(nselem) == 2): # Namespaced attribute nsUri = maps[nselem[0]] attrName = nselem[1] else: nsUri = None data = [] if (len(xp) == 1): # Extracting all occs of attribute anywhere!? # Check predicates... (only support one numeric predicate) if (len(xp[0][2]) == 1 and type(xp[0][2][0]) == types.FloatType): nth = int(xp[0][2][0]) elif (len(xp[0][2])): # Non index or multiple predicates?? raise(NotImplementedError) else: nth = 0 currn = 0 for l in self.sax: if (l[0] == "1"): (name, attrs) = self._convert_elem(l) if (attrName in attrs): currn += 1 content = attrs[attrName] if (currn == nth): data.append(content) break elif (not nth): data.append(content) else: elemName = xp[-2][1] flatten = 0 if (elemName == "*"): # Let DOM code handle this monstrosity :P raise(NotImplementedError) nselem = elemName.split(":") if (len(nselem) == 2): # Namespaced. elemNsUri = maps[nselem[0]] elemName = nselem[1] else: elemNsUri = "" if (elemName in self.elementHash): elemLines = self.elementHash[elemName] for e in elemLines: if (not elemNsUri or self.sax[e[0]][4:4+len(elemNsUri)] == elemNsUri): line = self.sax[e[0]] (name, attrs) = self._convert_elem(line) if (attrName == '*'): # All attributes' values match = self._checkSaxXPathLine(xp[:-1], e[0]) if (match): for k in attrs.keys(): data.append(attrs[k]) else: if (not attrName in attrs): attrName = (nsUri, attrName) if (not attrName in attrs and not nsUri): # step through and take first content = None for key in attrs: if key[1] == attrName[1]: content = attrs[key] else: content = attrs.get(attrName, None) if (content): # Now check rest of path match = self._checkSaxXPathLine(xp[:-1], e[0]) if (match): data.append(content) return data def _checkSaxXPathLine(self, xp, line): # Check that event at line in record matches xpath up tree # Pass by reference, need a copy to pop! Looks like a hack... xpath = xp[:] climb = False while (xpath): posn = len(xpath) node = xpath.pop() if (line == -1): if node != "/" and node != ['/']: return 0 else: elem = self.sax[line] (name, attrs) = self._convert_elem(elem) match = self._checkSaxXPathNode(node, name, attrs, line, posn) if not match: if not climb: return 0 else: # Previous was a descendant, keep looking while not match: start = elem.rfind("}") + 2 end = elem.find(" ", start) line = int(elem[start:end]) if line != -1: elem = self.sax[line] (name, attrs) = self._convert_elem(elem) match = self._checkSaxXPathNode(node, name, attrs, line, posn) else: return 0 if xpath: start = elem.rfind("}") + 2 end = elem.find(" ", start) line = int(elem[start:end]) climb = (node and node[0] == "descendant") return 1 def _checkSaxXPathNode(self, step, name, attrs, line, posn): # name already checked, strip if step in ['/', ['/']] and name: return 0 if (step[1] != name and step[1] != '*' and step[1][step[1].find(":")+1:] != name): return 0 elif (not step[0] in ['child', 'descendant']): # Unsupported axis raise(NotImplementedError) elif (step[2]): # Check predicates predPosn = 0 for pred in (step[2]): predPosn += 1 m = self._checkSaxXPathPredicate(pred, name, attrs, line, posn, predPosn) if (not m): return 0 return 1 def _checkSaxXPathPredicate(self, pred, name, attrs, line, posn, predPosn): if (type(pred) != types.ListType): # Numeric Predicate. (eg /foo/bar[1]) if (predPosn != 1): # Can't do numeric predicate on already predicated nodeset # eg: text[@type='main'][2] raise(NotImplementedError) if (posn == 1): # First position in relative path. # Check against position in elementHash if (name in self.elementHash): all = self.elementHash[name] p = int(pred) if (len(all) < p): return 0 return all[int(pred)-1][0] == line return 0 else: # Not first position, so it applies to parent elem # Which we record during parsing elem = self.sax[line] end = elem.rfind("}") + 2 start = elem.find(' ', end) + 1 end = elem.find(' ', start) npred = float(elem[start:end]) return npred == pred elif (pred[1] in ['=', '!=', '<', '>', '<=', '>=']): # Single attribute return self._checkSaxXPathAttr(pred, attrs) elif (pred[1] in ['and', 'or']): # Attribute combinations left = self._checkSaxXPathPredicate(pred[0], name, attrs, line, posn, predPosn) right = self._checkSaxXPathPredicate(pred[2], name, attrs, line, posn, predPosn) if (pred[1] == 'and' and left and right): return 1 elif (pred[1] == 'or' and (left or right)): return 1 return 0 elif (pred[0] == 'attribute'): # Attribute exists test return pred[1] in attrs elif (pred[0] == 'FUNCTION'): if pred[2] == "__name()": return True if pred[1] == 'starts-with': if pred[2] in attrs: val = attrs[pred[2]] return not val.find(pred[3]) else: return False elif pred[1] == 'regexp': if pred[2] in attrs: return pred[3].search(attrs[pred[2]]) != None else: return False raise NotImplementedError else: # No idea!! raise(NotImplementedError) return 1 def _checkSaxXPathAttr(self, pred, attrs): # Namespacey if (not pred[0] in attrs): if ((None, pred[0]) in attrs): pred[0] = (None, pred[0]) else: return 0 rel = pred[1] # -Much- faster than eval if (type(pred[2]) == types.FloatType): attrValue = float(attrs[pred[0]]) else: attrValue = attrs[pred[0]] comp = cmp(attrValue, pred[2]) if rel == "=": return comp == 0 elif rel == ">": return comp == 1 elif rel == "<": return comp == -1 elif rel == "<=": return comp in (-1, 0) elif rel == ">=": return comp in (1, 0) elif rel == "!=": return comp in (1, -1) else: raise(NotImplementedError) def _convert_elem(self, line): # Currently: 1 name {attrs} parent npred end if (line[0] == '1'): start = line.find("{") name = line[2:start-1] if line[start+1] == '}': attrs = {} else: attrList = self.attrRe.findall(line) attrs = {} for m in attrList: attrs[unicode(m[0])] = unicode(m[1]) return [name, attrs] elif (line[0] == '4'): end = line.rfind("}") stuff = eval(line[2:end+1]) return [stuff[1], stuff[3]] else: raise ValueError("Called convert on non element.") def saxify(self, session, handler=None, sax=[]): if handler == None: handler = self if not sax: sax = self.get_sax(session) for l in sax: line = l # line = l.strip() if line[0] == "1": # String manipulation method (name, attrs) = self._convert_elem(line) handler.startElement(name, attrs) elif line[0] == "3": handler.characters(line[2:], 0, len(line)-2) elif line[0] == "2": end = line.rfind(' ') handler.endElement(line[2:end]) elif line[0] == "9": pass elif line[0] == '4': # 4 ns,name,qname, {} idx = line.rfind(' ') idx = line[:idx].rfind(' ') idx = line[:idx].rfind(' ') line = line[:idx] (ns, name, qname, attrs) = eval(line[2:]) handler.startElementNS((ns,name), qname, attrs) elif line[0] == '5': # 5 ns,name,qname parent pred end idx = line.rfind(' ') line = line[:idx] (ns, name, qname) = eval(line[2:]) handler.endElementNS((ns,name),qname) elif line[0] == '6': # 6 pref, uri pref, uri = eval(line[2:]) handler.startPrefixMapping(pref, uri) else: # Unknown type raise ValueError(line) def get_dom(self, session): if (self.dom): return self.dom else: # Turn SAX into DOM and cache s2dhandler.initState() self.saxify(session, s2dhandler); self.dom = s2dhandler.getRootNode() return self.dom def get_xml(self, session, events=[]): if (not events and self.xml): return self.xml else: # Turn SAX into XML and cache if not events: process = self.sax else: process = events s2xhandler.initState() self.saxify(session, s2xhandler, process) if not events: self.xml = s2xhandler.get_xmlString() return self.xml else: return s2xhandler.get_xmlString() def get_sax(self, session): return self.sax def fetch_vector(self, session, index, summary=False): return index.indexStore.fetch_vector(session, index, self, summary) class MarcRecord(Record): """For dealing with Library MARC Records.""" def __init__(self, data, xml="", docId=0, wordCount=0, byteCount=0): txt = doc.get_raw(session) self.marc = MARC(txt) self.id = docId # Estimate number of words... display = str(self.marc) if wordCount: self.wordCount=wordCount else: self.wordCount = len(display.split()) - ( len(display.split('\n')) * 2) if byteCount: self.byteCount = byteCount else: self.byteCount = len(display) self.decoder = MARC8_to_Unicode() self.asciiRe = re.compile('([\x0e-\x1f]|[\x7b-\xff])') def process_xpath(self, session, xpath, maps={}): if (not isinstance(xpath, list)): # Raw XPath # c = utils.verifyXPaths([xpath]) if (not c or not c[0][1]): return [] else: xpath = c[0] xp = xpath[1] # format: fldNNN/a try: fld = int(xp[0][1][3:]) except ValueError: # not a NNN not an int return [] if fld in self.marc.fields: data = self.marc.fields[fld] else: return [] if len(xp) > 1: subfield = xp[1][1] else: subfield = "" vals = [] if fld in [0,1]: vals = data else: for d in data: if not subfield: vals.append(' '.join([x[1] for x in d[2]])) elif subfield == 'ind1': vals.append(d[0]) elif subfield == 'ind2': vals.append(d[1]) elif fld == 8: if not subfield: vals.append(d) elif subfield == 'lang': vals.append(d[35:38]) elif subfield == 'date': vals.append(d[:6]) elif subfield == 'pubStatus': vals.append(d[6]) elif subfield == 'date1': vals.append(d[7:11]) elif subfield == 'date2': vals.append(d[11:15]) elif subfield == 'pubPlace': vals.append(d[15:18]) else: for x in d[2]: try: if x[0] == subfield: vals.append(x[1]) except: # broken pass nvals = [] for v in vals: try: nvals.append(v.decode('utf-8')) except: try: convtd = self.decoder.translate(v) nvals.append(unicodedata.normalize('NFC', convtd)) except: # strip out any totally @^%(ed characters v = self.asciiRe.sub('?', v) nvals.append(v) return nvals def get_dom(self, session): raise(NotImplementedError) def get_sax(self, session): raise(NotImplementedError) def get_xml(self, session): return self.marc.toMARCXML() def fetch_vector(self, session, index, summary=False): return index.indexStore.fetch_vector(session, index, self, summary)
[ "info@cheshire3.org" ]
info@cheshire3.org
22f32bf0fd55088a1f7d6a101468f6c446d7e723
244e9b0fab8f641e21566987b8577f48012a08f7
/fibonacci.py
22d1dae859177a0ab61fa0cfb249c2b2de0e92ef
[]
no_license
iopkelvin/Leetcode-problems
55253a7e845c11ded792f8f3ea1ad1c0c37f1956
6fe8dde4b9d76fc8b797de2c71f7e0041d725267
refs/heads/main
2023-03-05T01:14:24.395087
2021-02-08T18:24:09
2021-02-08T18:24:09
335,342,086
0
0
null
null
null
null
UTF-8
Python
false
false
311
py
import sys def fib(n): # Recursive # if n == 0: # return 0 # if n == 1: # return 1 # while n >= 0: # return fib(n - 1) + fib(n - 2) a, b = 0, 1 for _ in range(n): a, b = b, a + b return a if __name__ == '__main__': print(fib(int(sys.argv[1])))
[ "36288868+iopkelvin@users.noreply.github.com" ]
36288868+iopkelvin@users.noreply.github.com
5ee331590a0342ddcead80522daf670bc75601fb
f27413138fa72a18308d2732112319e6de408876
/CacheDecorator.py
d092877b0eacc298416cf2f0111f6b2bb8b65ec7
[]
no_license
noamalka228/PythonExercise
18cb90fb0b021c41429bd6575ed0e3bd4bbca049
0f306a019b9629123c6095918a9ad389931307db
refs/heads/main
2023-04-15T18:22:56.956642
2021-04-29T12:07:51
2021-04-29T12:07:51
362,802,792
0
0
null
null
null
null
UTF-8
Python
false
false
1,113
py
class CacheDecorator(dict): # constructor def __init__(self, func_name): self.func_name = func_name # if the function is already cached, return the cached value def __call__(self, *args): return self[args] # if the result is not cached, return the function call and cache the result for future calls def __missing__(self, key): result = self[key] = self.func_name(*key) return result # calling the class of the CacheDecorator @CacheDecorator # this function gets a number as parameter and recursively sums all numbers from 1 to num def sum_all(num): if type(num) is int: if num == 0 or num == 1: return 1 return num + sum_all(num - 1) def main(): print("Sum: ", sum_all(5)) # the call caches all sums from 1-5 print("Cached sums: ", sum_all) # therefore, calling the func without parameter prints sum(5) print("Sum: ", sum_all(10)) # the call caches all sums from 5-10 print("Cached sums: ", sum_all) # therefore, calling the func without parameter prints sum(10) if __name__ == "__main__": main()
[ "noreply@github.com" ]
noamalka228.noreply@github.com
b72ed4fab9d517cf24f9b0a77dbf51f8da8c46a9
f8965d7b16e3cf70370b3bd181ef855a2ab89768
/ebook/Chapter08/format_invoice.py
406defa9a09b82cef2a7b0a8fa3f64401411bdf3
[ "Apache-2.0" ]
permissive
tuannguyendang/montypython
59cae6fc6069cf5356670132470cdd52bad00d67
c0b8ff7a8130e811ba16bfab8d5e013eac37f432
refs/heads/main
2023-05-12T02:24:50.693432
2021-05-31T02:14:16
2021-05-31T02:14:16
325,188,168
0
0
null
null
null
null
UTF-8
Python
false
false
698
py
# This won't render right subtotal = 12.32 tax = subtotal * 0.07 total = subtotal + tax print(f"Sub: ${subtotal} Tax: ${tax} Total: ${total}") print(f"Sub: ${subtotal:0.2f} Tax: ${tax:0.2f} Total: ${total:0.2f}") orders = [("burger", 2, 5), ("fries", 3.5, 1), ("cola", 1.75, 3)] print("PRODUCT QUANTITY PRICE SUBTOTAL") for product, price, quantity in orders: subtotal = price * quantity print( f"{product:10s}{quantity: ^9d} " f"${price: <8.2f}${subtotal: >7.2f}" ) print("PRODUCT QUANTITY PRICE SUBTOTAL") for p, pr, q in orders: subt = pr * q print( f"{p:20s}{q:^10d}" f"${pr:<9.3f}${subt: > 9.3f}" )
[ "tuan193@gmail.com" ]
tuan193@gmail.com
babd1c65ab124645809ec009676a41dc323194be
935a4ebc2cb54553c9d51e1fcd885d5d5a993218
/fulfillment/fulfillment.py
1359858cc480454ef0e9f5eea67947f74b7cce36
[ "MIT" ]
permissive
ahmetyazar/adj-demo
c03a6a481470c0548294e448a158b370e79cd6a0
9f7c48faa65d951c040dcfed9904e2186415221a
refs/heads/master
2021-01-06T20:34:25.652447
2017-08-11T02:45:21
2017-08-11T02:45:21
99,522,948
0
0
null
null
null
null
UTF-8
Python
false
false
2,462
py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Jul 30 15:28:00 2017 @author: yazar """ import sys import logging import rds_config import pymysql import datetime import json import requests # rds settings rds_host = "creditadjudication.cz3lpdttkjpo.us-east-1.rds.amazonaws.com" name = rds_config.db_username password = rds_config.db_password db_name = rds_config.db_name logger = logging.getLogger() logger.setLevel(logging.INFO) try: conn = pymysql.connect(rds_host, user=name, passwd=password, db=db_name, connect_timeout=5, cursorclass=pymysql.cursors.DictCursor) except: logger.error("ERROR: Unexpected error: Could not connect to MySql instance.") sys.exit() logger.info("SUCCESS: Connection to RDS mysql instance succeeded") def datetime_handler(x): if isinstance(x, datetime.datetime): return x.isoformat() raise TypeError("Unknown type") json.JSONEncoder.default = datetime_handler def addBlockedCustomers(event, context): """ This function fetches content from mysql RDS instance """ logger.info('#################') logger.info(event) logger.info('#################') message = event['Records'][0]['Sns']['Message'] logger.info('From SNS: ' + message) message = json.loads(message) sql = "insert into FulfillmentDb.BlockedCustomers (partyID, effectiveDate) " sql += "values ({0}, NOW())".format(message['partyID']) logger.info(sql) with conn.cursor() as cur: # create Customers db and load some sample records cur.execute(sql) conn.commit() cur.execute("select * from FulfillmentDb.BlockedCustomers") item_count = 0 for row in cur: item_count += 1 logger.info(row) logger.info("Added %d items from RDS MySQL table".format(item_count)) conn.commit() return { 'statusCode': 200, 'headers': {'Content-Type': 'application/json'}, 'body': json.dumps(event) } def checkPotentialFraud(event, context): logger.info('#################') logger.info(event) logger.info('#################') # check if there is a fraud sql = "select * from FulfillmentDb.PotentialFraud where partyID = {}".format(event['partyID']) def processCreditLimit(event, context): logger.info('#################') logger.info(event) logger.info('#################')
[ "ahmetyazar@yahoo.com" ]
ahmetyazar@yahoo.com
fa56238a5d5fa13f842dcbd60a00c830bce51d0a
dcdb7a05d52cd1f9d558a70570b3ecbd85cefbe6
/dj_blog/settings.py
afa71345b7ec029e8bcda1221ebdb1f056053ba3
[]
no_license
GDCenter/blog_django
ea7a9a556292b212a6d5a2de3d02f7b1e9002871
5cb12f630618bb49bd955bcc9072339ff3a01387
refs/heads/master
2020-09-09T15:51:49.918515
2018-05-11T09:21:05
2018-05-11T09:21:05
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,064
py
""" Django settings for dj_blog project. Generated by 'django-admin startproject' using Django 1.8.2. For more information on this file, see https://docs.djangoproject.com/en/1.8/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.8/ref/settings/ """ # Build paths inside the project like this: os.path.join(BASE_DIR, ...) import os BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/1.8/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = '%7&gfs!u_s4ci@5o2*j6tv^$*ju_2ds7-83o#$s#423(4+y4ds' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = [] # Application definition INSTALLED_APPS = ( 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', # 添加apps 'apps.blog_article', 'apps.blog_console', 'apps.blog_index', 'apps.blog_sign', ) MIDDLEWARE_CLASSES = ( 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.auth.middleware.SessionAuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', 'django.middleware.security.SecurityMiddleware', ) ROOT_URLCONF = 'dj_blog.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [os.path.join(BASE_DIR, 'template')], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', # 使用media文件夹 'django.template.context_processors.media', ], }, }, ] WSGI_APPLICATION = 'dj_blog.wsgi.application' # Database # https://docs.djangoproject.com/en/1.8/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'blog_django', 'HOST': '106.14.1.9', 'PORT': 3306, 'USER': 'liduo', 'PASSWORD': '624695549' } } # Internationalization # https://docs.djangoproject.com/en/1.8/topics/i18n/ LANGUAGE_CODE = 'zh-hans' # 使用中文 TIME_ZONE = 'Asia/Shanghai' # 设定时区 USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.8/howto/static-files/ STATIC_URL = '/static/' # 添加静态文件路径 STATICFILES_DIRS = [os.path.join(BASE_DIR, 'static')] STATIC_ROOT = os.path.join(BASE_DIR, 'statics') # 设置media文件夹url MEDIA_URL = '/media/' # 添加media文件夹 MEDIA_DIR = os.path.join(BASE_DIR, 'media') MEDIA_ROOT = MEDIA_DIR #################################################### # 自定义User模型类设置 # 继承重写User模型类时,指定自定义的模型类,'app名字.类名' AUTH_USER_MODEL = 'blog_sign.User' #################################################### # login_required装饰器设置 # 配置登录界面,供登录验证装饰器login_required使用 LOGIN_URL = '/sign/login' #################################################### # 邮件设置 # 发送邮件配置 EMAIL_BACKEND = 'django.core.mail.backends.smtp.EmailBackend' # 使用ssl加密方式 EMAIL_USE_SSL = True # smpt服务地址 EMAIL_HOST = 'smtp.163.com' # 使用ssl的smtp端口465,非ssl端口25 EMAIL_PORT = 465 # 发送邮件的邮箱 EMAIL_HOST_USER = 'liduo945@163.com' # 在邮箱中设置的客户端授权密码 EMAIL_HOST_PASSWORD = 'liduo945163' # 收件人看到的发件人 EMAIL_FROM = 'LeeBlog<liduo945@163.com>' #################################################### # Redis缓存设置 # Django的缓存配置 CACHES = { "default": { "BACKEND": "django_redis.cache.RedisCache", "LOCATION": "redis://:624695549@106.14.1.9:6379/9", "OPTIONS": { "CLIENT_CLASS": "django_redis.client.DefaultClient", } } } # 配置session存储 SESSION_ENGINE = "django.contrib.sessions.backends.cache" SESSION_CACHE_ALIAS = "default" #################################################### # FastDFS分布式文件存储系统设置 # 设置Django的文件存储类 DEFAULT_FILE_STORAGE = 'utils.fdfs.storage.FDFSStorage' # 设置fdfs使用的client.conf文件路径 FDFS_CLIENT_CONF = os.path.join(BASE_DIR, 'dj_blog/client.conf') # 设置fdfs存储服务器上nginx的IP和端口号 FDFS_URL = 'http://106.14.1.9/' ####################################################
[ "liduo945@163.com" ]
liduo945@163.com
bb5d93cb6f7562c1613483c8b5d2df9e1d874fac
46192947e8a87bdb5ddf4de92f832f8c2d53bccb
/REST API/api_project/testpython.py
ca9eb6794f25f1aa41ba61c45ca544892381354a
[]
no_license
mkansari31/Rest_API
216a060330923c75e003a2b9d75071bc717a60be
b0325c907e3d8dd93f5711c274f57a0592efb912
refs/heads/master
2023-01-31T00:10:11.129491
2020-12-14T09:47:39
2020-12-14T09:47:39
321,298,669
0
0
null
null
null
null
UTF-8
Python
false
false
790
py
from flask import Flask, render_template, request from flask_mysqldb import MySQL app = Flask(__name__) app.config['MYSQL_HOST'] = 'localhost' app.config['MYSQL_USER'] = 'root' app.config['MYSQL_PASSWORD'] = '' app.config['MYSQL_DB'] = 'MyDB' app.config['DEBUG'] = True mysql = MySQL(app) @app.route('/', methods=['GET', 'POST']) def index(): if request.method == "POST": details = request.form firstName = details['fname'] lastName = details['lname'] cur = mysql.connection.cursor() cur.execute("INSERT INTO MyUsers(firstName, lastName) VALUES (%s, %s)", (firstName, lastName)) mysql.connection.commit() cur.close() return 'success' return render_template('index.html') if __name__ == '__main__': app.run()
[ "mkansari515@gmail.com" ]
mkansari515@gmail.com
f4506a41f21652bd250f6896810cd6fbdec72bfb
ca7aa979e7059467e158830b76673f5b77a0f5a3
/Python_codes/p03042/s013075072.py
044f87c3be49952ef7be8bf867e28108c9b4cd05
[]
no_license
Aasthaengg/IBMdataset
7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901
f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8
refs/heads/main
2023-04-22T10:22:44.763102
2021-05-13T17:27:22
2021-05-13T17:27:22
367,112,348
0
0
null
null
null
null
UTF-8
Python
false
false
186
py
s=int(input()) a=s//100 b=s%100 if a>0 and a<=12: if b>0 and b<=12: print("AMBIGUOUS") else: print("MMYY") else: if b>0 and b<=12: print("YYMM") else: print("NA")
[ "66529651+Aastha2104@users.noreply.github.com" ]
66529651+Aastha2104@users.noreply.github.com
2fae047ea5b7af3cba687716d80fa7aab18a4d0a
4d259f441632f5c45b94e8d816fc31a4f022af3c
/date/tt.py
9f4bc74af2f7f817b5cc2a96f52b570bd76401f0
[]
no_license
xiaoruiguo/lab
c37224fd4eb604aa2b39fe18ba64e93b7159a1eb
ec99f51b498244c414b025d7dae91fdad2f8ef46
refs/heads/master
2020-05-25T01:37:42.070770
2016-05-16T23:24:26
2016-05-16T23:24:26
null
0
0
null
null
null
null
UTF-8
Python
false
false
40
py
a=['1','2','3'] s = ['sss'*3]+a print s
[ "junmein@junmeinde-macbook-pro-3.local" ]
junmein@junmeinde-macbook-pro-3.local
74c6a926ed7a08e4d04f9fed809bdbf9a1a4c9b6
f0f9e59b3b62a89b1e9cf3f02627d80be3359711
/webpersonal/core/views.py
8cc05b98cc068c425cdf03633c4553a6812fef81
[]
no_license
gemamoreira/web-personal-django3
372d500bc9dd16768ef0223dcfb68d4eb1e6498c
3177db3bdaa7478f979e1f1d071904faef693035
refs/heads/main
2023-01-01T03:52:34.047986
2020-10-13T05:27:27
2020-10-13T05:27:27
291,835,070
0
0
null
null
null
null
UTF-8
Python
false
false
554
py
from django.shortcuts import render, HttpResponse html_base = """ <h1>Mi web Personal</h1> <ul> <li><a href="/">Portada</a></li> <li><a href="/about/">Acerca de...</a></li> <li><a href="/portfolio/">Portafolio</a></li> <li><a href="/contact/">Contacto</a></li> </ul> """ # Create your views here. def home(request): return render(request, "core/home.html") def about (request): return render(request, "core/about.html") def contact (request): return render(request, "core/contact.html")
[ "glmoremu20@gmail.com" ]
glmoremu20@gmail.com
62b6273166486acf1ece5437a98e41a0350b1124
9743d5fd24822f79c156ad112229e25adb9ed6f6
/xai/brain/wordbase/verbs/_celebrating.py
305a78d8f0d008577d0f029e5a82a8910f663133
[ "MIT" ]
permissive
cash2one/xai
de7adad1758f50dd6786bf0111e71a903f039b64
e76f12c9f4dcf3ac1c7c08b0cc8844c0b0a104b6
refs/heads/master
2021-01-19T12:33:54.964379
2017-01-28T02:00:50
2017-01-28T02:00:50
null
0
0
null
null
null
null
UTF-8
Python
false
false
261
py
from xai.brain.wordbase.verbs._celebrate import _CELEBRATE #calss header class _CELEBRATING(_CELEBRATE, ): def __init__(self,): _CELEBRATE.__init__(self) self.name = "CELEBRATING" self.specie = 'verbs' self.basic = "celebrate" self.jsondata = {}
[ "xingwang1991@gmail.com" ]
xingwang1991@gmail.com
2339baadffae6ec1d43540f2b5a2f88e5c5dddd0
4d74a14506b95289084379d85d09e0da020ba951
/condet/apps.py
de69d6066b4c6e98a8927d9908d1d593112724ce
[]
no_license
eduardozamudio/unam-tourism-research
6213e2763010cee57a85c2aa755f6e978cc56efe
3a307877b610cb250ba142d34933261d657d837a
refs/heads/master
2022-12-11T16:31:30.005504
2017-10-26T11:39:02
2017-10-26T11:39:02
102,898,354
0
0
null
2022-12-08T00:36:55
2017-09-08T19:42:37
Jupyter Notebook
UTF-8
Python
false
false
87
py
from django.apps import AppConfig class CondetConfig(AppConfig): name = 'condet'
[ "eduardozamudio@gmail.com" ]
eduardozamudio@gmail.com
89dd04e8b979a718563ec8de70bc2e2eff652161
6380adb492ee95afbbad559e5017af32830512e7
/lab-2/PCY.py
96abcdc03451b0a23e397d4ad754ecb7a20ec094
[]
no_license
square-brackets/avsp-2018
743cf258037cb9ab8f2cfa13edd6ae37c36f4540
7076f086a5198d04d4367ff4d7396e1c3eb4ac05
refs/heads/master
2020-04-03T06:38:42.252574
2018-10-28T14:38:20
2018-10-28T14:38:20
155,080,142
0
0
null
null
null
null
UTF-8
Python
false
false
2,059
py
import sys import time from itertools import combinations def count_items(bucket_count): # start_time = time.time() item_count = {} bucket_items = [] for bucket in range(bucket_count): bucket_items_line = sys.stdin.readline() bucket_items.append([]) for item in bucket_items_line.split(' '): item_int = int(item) item_count[item_int] = item_count.get(item_int, 0) + 1 bucket_items[bucket].append(item_int) # print('[COUNT_ITEMS]: ', time.time() - start_time) return item_count, bucket_items def create_itemsets(buckets_items, itemset_count, item_count, threshold): # start_time = time.time() itemsets = [0] * itemset_count; for bucket_items in buckets_items: for itemA, itemB in combinations(bucket_items, 2): if item_count[itemA] >= threshold and item_count[itemB] >= threshold: k = (itemA * len(item_count) + itemB) % itemset_count itemsets[k] += 1; # print('[CREATE_ITEMSETS]: ', time.time() - start_time) return itemsets if __name__ == '__main__': bucket_count = int(sys.stdin.readline()) s = float(sys.stdin.readline()) itemset_count = int(sys.stdin.readline()) threshold = int(s * bucket_count) item_count, buckets_items = count_items(bucket_count) itemsets = create_itemsets(buckets_items, itemset_count, item_count, threshold) # start_time = time.time() pairs = {} for bucket_items in buckets_items: for itemA, itemB in combinations(bucket_items, 2): if item_count[itemA] >= threshold and item_count[itemB] >= threshold: k = (itemA * len(item_count) + itemB) % itemset_count if (itemsets[k] >= threshold): pairs[(itemA, itemB)] = pairs.get((itemA, itemB), 0) + 1 # print('[COUNT_PAIRS]: ', time.time() - start_time) print(len(list(filter(lambda x: x > 0, itemsets)))) print(len(pairs)) for s in sorted(pairs, key=pairs.get, reverse=True): print(pairs[s])
[ "stjepan.petrusa@gmail.com" ]
stjepan.petrusa@gmail.com
226b1a82100d5a5ec91accefe4709d3526693183
2fe868ab7e641629013445af85f412cfd0fc323d
/04_문자열자료형/문자열생성.py
11a405fd099d52523f193b556ee27d3ae638f5d2
[]
no_license
lgy94/pythonOjt
9e0988a8e9ee005688f1f7e841077d34f403b13c
8d1eb5ed52527153153d3a2d47f346d0810e2d1a
refs/heads/master
2021-03-28T21:36:39.095570
2020-03-20T07:40:08
2020-03-20T07:40:08
247,898,759
0
0
null
null
null
null
UTF-8
Python
false
false
450
py
s = 'Python is great!' print (s) s = "Python is great!" print (s) s = '''Python is great!''' print (s) s = """Python is great!""" print (s) sentence = 'Python is the\ most popular programming\ language in these days.' print (sentence) a = 'say "hello" to mom' b = "say 'hello' to mom" c = '''say 'hello' to "mom"''' print (a) print (b) print (c) #letter to Alice print ('''Dear Alice, How are you? Say hello to your parents. Sincerely, Bob''')
[ "dudrk94@naver.com" ]
dudrk94@naver.com
a81b4c8c52d69ca1c7a214f2eef794aea07a5f65
8610b91f0f36e0df7f343c55929e5861bf0eb144
/Smart Reply_02_Apr_2018.py
9afb0d7cd1992c0664fe01816ecabf33d4ff609d
[]
no_license
abhijitdalavi/SmartReply
1236aa3e85cee2aeefa1362d7c54f5a9009109fc
a0f2c384550b579fc56b86a28412720793210f8d
refs/heads/master
2020-11-30T01:06:13.369949
2018-04-09T17:28:05
2018-04-09T17:28:05
null
0
0
null
null
null
null
UTF-8
Python
false
false
18,750
py
# coding: utf-8 # In[1]: from keras.models import Model from keras.layers.recurrent import LSTM from keras.layers import Dense, Input, Embedding from keras.preprocessing.sequence import pad_sequences from keras.callbacks import ModelCheckpoint from keras.utils.vis_utils import plot_model from keras.preprocessing.text import Tokenizer from nltk.corpus import stopwords from nltk.tokenize import word_tokenize from nltk.tokenize import TweetTokenizer from nltk.stem.wordnet import WordNetLemmatizer #from wordcloud import WordCloud import gensim from gensim.models import word2vec import logging import tensorflow as tf from collections import Counter import nltk import numpy as np import pandas as pd from sklearn.model_selection import train_test_split import urllib.request import os import sys import zipfile import logging import pydot import graphviz import re #os.environ['CUDA_VISIBLE_DEVICES'] = '-1' #import logging #logger = logging.getLogger("Testing_Model") #logger.setLevel(logging.INFO) #https://github.com/chen0040/keras-chatbot-web-api/blob/master/chatbot_train/cornell_word_seq2seq_glove_train.py # In[11]: # ********************************************************************** # Reading a pre-trained word embedding and addapting to our vocabulary: # ********************************************************************** def load_glove(): embeddings_index = {} #f = open(os.path.join(GLOVE_DIR, 'glove.6B.100d.txt')) f = open('glove.6B.100d.txt', encoding = 'utf8') for line in f: values = line.split() word = values[0] coefs = np.asarray(values[1:], dtype='float32') embeddings_index[word] = coefs f.close() return embeddings_index def init_stopwords(): lem = WordNetLemmatizer() #https://drive.google.com/file/d/0B1yuv8YaUVlZZ1RzMFJmc1ZsQmM/view # Aphost lookup dict APPO = { "aren't" : "are not", "can't" : "cannot", "couldn't" : "could not", "didn't" : "did not", "doesn't" : "does not", "don't" : "do not", "hadn't" : "had not", "hasn't" : "has not", "haven't" : "have not", "he'd" : "he would", "he'll" : "he will", "he's" : "he is", "i'd" : "I would", "i'd" : "I had", "i'll" : "I will", "i'm" : "I am", "isn't" : "is not", "it's" : "it is", "it'll":"it will", "i've" : "I have", "let's" : "let us", "mightn't" : "might not", "mustn't" : "must not", "shan't" : "shall not", "she'd" : "she would", "she'll" : "she will", "she's" : "she is", "shouldn't" : "should not", "that's" : "that is", "there's" : "there is", "they'd" : "they would", "they'll" : "they will", "they're" : "they are", "they've" : "they have", "we'd" : "we would", "we're" : "we are", "weren't" : "were not", "we've" : "we have", "what'll" : "what will", "what're" : "what are", "what's" : "what is", "what've" : "what have", "where's" : "where is", "who'd" : "who would", "who'll" : "who will", "who're" : "who are", "who's" : "who is", "who've" : "who have", "won't" : "will not", "wouldn't" : "would not", "you'd" : "you would", "you'll" : "you will", "you're" : "you are", "you've" : "you have", "'re": " are", "wasn't": "was not", "we'll":" will", "didn't": "did not", "tryin'":"trying" } eng_stopwords = set(stopwords.words("english")) user_stop_words = [] print("Eng stopwords before user stopwords:: ", len(eng_stopwords)) for w in user_stop_words: if w not in eng_stopwords: eng_stopwords.add(w) print("Eng stopwords after user stopwords:: ",len(eng_stopwords)) return lem, APPO, eng_stopwords def pre_process(text, lem, APPO, eng_stopwords): text=text.lower() text = re.sub("was good.", " ", text) text = re.sub("was bad.", " ", text) words = word_tokenize(text) # (')aphostophe replacement (ie) you're --> you are words=[APPO[word] if word in APPO else word for word in words] words=[lem.lemmatize(word, "v") for word in words] # remove punctuations words = [w.lower() for w in words if w.isalpha()] words = [w for w in words if not w in eng_stopwords] clean_sent = " ".join(words) return clean_sent # ********************************************************************** # Developing our vocabulary from the dataset: # ********************************************************************** def load_vocab(self): #print(type(inp_clean_corpus)) #sentences = inp_clean_corpus.tolist() sentences = self.input_texts print(sentences[1:5]) print(type(sentences)) model = word2vec.Word2Vec(sentences, iter=5, min_count=5, size=100, workers=4) vocab_size = len(model.wv.vocab) print(vocab_size) # get the most common words print(model.wv.index2word[0:10]) # get the least common words print(model.wv.index2word[vocab_size - 1:vocab_size-10]) # convert the wv word vectors into a numpy matrix # embedding_matrix = np.zeros((len(model.wv.vocab), 10)) # for i in range(len(model.wv.vocab)): # embedding_vector = model.wv[model.wv.index2word[i]] # if embedding_vector is not None: # embedding_matrix[i] = embedding_vector # print(embedding_matrix[0:5]) # convert the wv word vectors into a numpy matrix embedding_matrix = {} for i, word in enumerate(model.wv.vocab): embedding_vector = model.wv[model.wv.index2word[i]] if embedding_vector is not None: embedding_matrix[word] = embedding_vector embeddings_index = {} #f = open(os.path.join(GLOVE_DIR, 'glove.6B.100d.txt')) f = open('glove.6B.100d.txt', encoding = 'utf8') for line in f: values = line.split() word = values[0] if word in ('start','stop'): coefs = np.asarray(values[1:], dtype='float32') embeddings_index[word] = coefs f.close() embedding_matrix['start']=embeddings_index['start'] embedding_matrix['stop']=embeddings_index['stop'] return embedding_matrix # ********************************************************************** # Reading input text and the replies # ********************************************************************** def read_input(): df = pd.read_csv(DATA, encoding = 'latin-1') input1 = df['SentimentText'].fillna("") output1 = df['ResponseText'].fillna("") print(len(df['SentimentText'])) # pre-processing only the input data lem, APPO, eng_stopwords = init_stopwords() inp_clean_corpus = input1.apply(lambda x: pre_process(str(x), lem, APPO, eng_stopwords)) print(type(inp_clean_corpus)) # vocab = load_vocab(inp_clean_corpus) # some stats about input data num_words = 5000 tokenize = Tokenizer(num_words=num_words) tokenize.fit_on_texts(inp_clean_corpus) tok_inp = tokenize.texts_to_sequences(inp_clean_corpus) inp_len = [len(words) for words in tok_inp] mod_tok_inp = int(np.mean(inp_len) + 2 * np.std(inp_len)) print("\nmean:: ", np.mean(inp_len)) print("max:: ", np.max(inp_len)) print("inp token size - average + 2*sd --> :: ", (mod_tok_inp)) print("inp - total inp tokens:: ", len(inp_len)) final_len_inp = [(tok) for tok in inp_len if tok < mod_tok_inp] final_len_inp = list(final_len_inp) print("inp - no of tokens < mod_tok :: ", len(final_len_inp)) input1 = inp_clean_corpus.tolist() output1 = output1.tolist() target_counter = Counter() input_texts = [] target_texts = [] print(type(input1)) print("input:: \n", input1[1:5]) print(type(output1)) print("output:: \n", output1[1:5]) for line in input1: inp_words = [w.lower() for w in nltk.word_tokenize(line)] if len(inp_words) > MAX_TARGET_SEQ_LENGTH: inp_words = inp_words[0:MAX_TARGET_SEQ_LENGTH] input_texts.append(inp_words) for line1 in output1: out_words = [w.lower() for w in nltk.word_tokenize(line1)] if len(out_words) > MAX_TARGET_SEQ_LENGTH: out_words = out_words[0:MAX_TARGET_SEQ_LENGTH] tar_words = out_words[:] tar_words.insert(0, 'start') tar_words.append('end') for w in tar_words: target_counter[w] += 1 target_texts.append(tar_words) print("\n Input texts :: \n\n ", input_texts[1:5]) print("\n Target texts :: \n\n ",target_texts[1:5]) return input_texts, target_texts, target_counter def get_target(self): target_word2idx = dict() for idx, word in enumerate(self.target_counter.most_common(MAX_VOCAB_SIZE)): target_word2idx[word[0]] = idx + 1 if 'UNK' not in target_word2idx: target_word2idx['UNK'] = 0 target_idx2word = dict([(idx, word) for word, idx in target_word2idx.items()]) num_decoder_tokens = len(target_idx2word)+1 input_texts_word2em = [] encoder_max_seq_length = 0 decoder_max_seq_length = 0 for input_words, target_words in zip(self.input_texts, self.target_texts): encoder_input_wids = [] for w in input_words: emb = np.zeros(shape=GLOVE_EMBEDDING_SIZE) if w in self.word2em: emb = self.word2em[w] encoder_input_wids.append(emb) input_texts_word2em.append(encoder_input_wids) encoder_max_seq_length = max(len(encoder_input_wids), encoder_max_seq_length) decoder_max_seq_length = max(len(target_words), decoder_max_seq_length) #print("input_texts_word2em for first 2 sentenses:: \n", input_texts_word2em[1:3]) context = dict() context['num_decoder_tokens'] = num_decoder_tokens context['encoder_max_seq_length'] = encoder_max_seq_length context['decoder_max_seq_length'] = decoder_max_seq_length return target_word2idx, target_idx2word, context, input_texts_word2em def generate_batch(input_word2em_data, output_text_data, self): num_batches = len(input_word2em_data) // BATCH_SIZE print("context:: \n", self.context) print("len of input data :: ", len(input_word2em_data)) print("num of batches :: ", num_batches) while True: for batchIdx in range(0, num_batches): start = batchIdx * BATCH_SIZE end = (batchIdx + 1) * BATCH_SIZE encoder_input_data_batch = pad_sequences(input_word2em_data[start:end], self.context['encoder_max_seq_length']) decoder_target_data_batch = np.zeros(shape=(BATCH_SIZE, self.context['decoder_max_seq_length'], self.num_decoder_tokens)) decoder_input_data_batch = np.zeros(shape=(BATCH_SIZE, self.context['decoder_max_seq_length'], GLOVE_EMBEDDING_SIZE)) for lineIdx, target_words in enumerate(output_text_data[start:end]): for idx, w in enumerate(target_words): w2idx = self.target_word2idx['UNK'] # default UNK if w in self.target_word2idx: w2idx = self.target_word2idx[w] if w in self.word2em: decoder_input_data_batch[lineIdx, idx, :] = self.word2em[w] if idx > 0: decoder_target_data_batch[lineIdx, idx - 1, w2idx] = 1 yield [encoder_input_data_batch, decoder_input_data_batch], decoder_target_data_batch class CornellWordGloveChatBot(object): model = None encoder_model = None decoder_model = None target_counter = None target_word2idx = None target_idx2word = None max_decoder_seq_length = None max_encoder_seq_length = None num_decoder_tokens = None word2em = None context = None input_texts = None target_texts = None def __init__(self): #self.word2em = load_glove() self.input_texts, self.target_texts, self.target_counter = read_input() print("input texts:: \n", self.input_texts[0:5]) self.word2em = load_vocab(self) print("Length of word2em :: ", len(self.word2em)) for idx, (input_words, target_words) in enumerate(zip(self.input_texts, self.target_texts)): if idx > 10: break print([input_words, target_words]) self.target_word2idx, self.target_idx2word , self.context, input_texts_word2em = get_target(self) self.max_encoder_seq_length = self.context['encoder_max_seq_length'] self.max_decoder_seq_length = self.context['decoder_max_seq_length'] self.num_decoder_tokens = self.context['num_decoder_tokens'] print("context: ",self.context) encoder_inputs = Input(shape=(None, GLOVE_EMBEDDING_SIZE), name='encoder_inputs') encoder_lstm1 = LSTM(units=HIDDEN_UNITS, return_state=True, name="encoder_lstm1" , dropout=0.2) #logger.info("Added LSTM Layer") #encoder_lstm2 = LSTM(units=HIDDEN_UNITS, return_state=True, name="encoder_lstm2", dropout=0.2) #encoder_lstm3 = LSTM(units=HIDDEN_UNITS, return_state=True, name="encoder_lstm3") #x = encoder_lstm1(encoder_inputs) encoder_outputs, encoder_state_h, encoder_state_c = encoder_lstm1(encoder_inputs) encoder_states = [encoder_state_h, encoder_state_c] decoder_inputs = Input(shape=(None, GLOVE_EMBEDDING_SIZE), name='decoder_inputs') decoder_lstm = LSTM(units=HIDDEN_UNITS, return_sequences=True, return_state=True, name='decoder_lstm', dropout=0.2) decoder_outputs, _, _ = decoder_lstm(decoder_inputs, initial_state=encoder_states) decoder_dense = Dense(self.num_decoder_tokens, activation='softmax', name='decoder_dense') decoder_outputs = decoder_dense(decoder_outputs) self.model = Model([encoder_inputs, decoder_inputs], decoder_outputs) print(self.model.summary()) #plot_model(self.model, to_file='RNN_model.png', show_shapes=True) #self.model.load_weights('../chatbot_train/models/' + DATA_SET_NAME + '/word-glove-weights.h5') self.model.compile(optimizer='rmsprop', loss='categorical_crossentropy') Xtrain, Xtest, Ytrain, Ytest = train_test_split(input_texts_word2em, self.target_texts, test_size=0.2, random_state=42) print("Length of train data:: ", len(Xtrain)) print("Length of test data:: ", len(Xtest)) train_gen = generate_batch(Xtrain, Ytrain, self) test_gen = generate_batch(Xtest, Ytest, self) train_num_batches = len(Xtrain) // BATCH_SIZE test_num_batches = len(Xtest) // BATCH_SIZE #checkpoint = ModelCheckpoint(filepath=WEIGHT_FILE_PATH, save_best_only=True) self.model.fit_generator(generator=train_gen, steps_per_epoch=train_num_batches, epochs=NUM_EPOCHS, verbose=1, validation_data=test_gen, validation_steps=test_num_batches ) #, callbacks=[checkpoint]) self.model.save_weights(WEIGHT_FILE_PATH) self.encoder_model = Model(encoder_inputs, encoder_states) #plot_model(self.encoder_model, to_file='RNN_encoder_model.png', show_shapes=True) decoder_state_inputs = [Input(shape=(HIDDEN_UNITS,)), Input(shape=(HIDDEN_UNITS,))] decoder_outputs, state_h, state_c = decoder_lstm(decoder_inputs, initial_state=decoder_state_inputs) decoder_states = [state_h, state_c] decoder_outputs = decoder_dense(decoder_outputs) self.decoder_model = Model([decoder_inputs] + decoder_state_inputs, [decoder_outputs] + decoder_states) #plot_model(self.decoder_model, to_file='RNN_decoder_model.png', show_shapes=True) def reply(self, input_text): input_seq = [] input_emb = [] print("input text:: \n\n ", input_text) # pre-processing only the input data lem, APPO, eng_stopwords = init_stopwords() clean_input_text = pre_process(str(input_text), lem, APPO, eng_stopwords) for word in nltk.word_tokenize(clean_input_text.lower()): emb = np.zeros(shape=GLOVE_EMBEDDING_SIZE) if word in self.word2em: emb = self.word2em[word] input_emb.append(emb) input_seq.append(input_emb) input_seq = pad_sequences(input_seq, self.max_encoder_seq_length) states_value = self.encoder_model.predict(input_seq) target_seq = np.zeros((1, 1, GLOVE_EMBEDDING_SIZE)) target_seq[0, 0, :] = self.word2em['start'] target_text = '' target_text_len = 0 terminated = False while not terminated: output_tokens, h, c = self.decoder_model.predict([target_seq] + states_value) #print("output tokens shape :: \n\n ", output_tokens.shape) sample_token_idx = np.argmax(output_tokens[0, -1, :]) sample_word = self.target_idx2word[sample_token_idx] target_text_len += 1 if sample_word != 'start' and sample_word != 'end': #print("sample word :: ", sample_word) target_text += ' ' + sample_word if sample_word == 'end' or target_text_len >= self.max_decoder_seq_length: terminated = True target_seq = np.zeros((1, 1, GLOVE_EMBEDDING_SIZE)) if sample_word in self.word2em: target_seq[0, 0, :] = self.word2em[sample_word] states_value = [h, c] return target_text.strip() def test_run(self): print(self.reply('Not so good experience. Washroom was not cleaned properly and room service was not quick to resond.')) print(self.reply('Hotel was ok. Food was good and staff was very cooperative in providing services.')) print(self.reply('I loved the environment of the hotel !!!. It was great living there ')) def main(): np.random.seed(42) model = CornellWordGloveChatBot() model.test_run() if __name__ == '__main__': MAX_VOCAB_SIZE = 10000 BATCH_SIZE = 32 NUM_EPOCHS = 100 GLOVE_EMBEDDING_SIZE = 100 HIDDEN_UNITS = 32 MAX_INPUT_SEQ_LENGTH = 150 MAX_TARGET_SEQ_LENGTH = 150 DATA_SET_NAME = 'cornell' DATA = 'D:/CBA/PositiveOnly.csv' DATA_PATH = 'movie_lines_cleaned_10k.txt' WHITELIST = 'abcdefghijklmnopqrstuvwxyz1234567890?.,' WEIGHT_FILE_PATH = 'D:/CBA/word-glove-weights.h5' main()
[ "asksonu.sunil@gmail.com" ]
asksonu.sunil@gmail.com
3982714fe0bab687a1e4f4ee2cab7d5183777153
d846460c9fcb294fb88a68c14f081013466f47d5
/matplotlib.py
b2af1a9ac17bdec4142f724cb00879ebde715844
[]
no_license
divyajaincs/Python-Practice-
52d5dc0a7d998d7476fe332e723dee7b89b182cb
679df72102555a7f1afe055c1257aca2cfada0f3
refs/heads/master
2021-04-12T10:07:28.868872
2018-03-25T19:32:14
2018-03-25T19:32:14
126,732,330
0
0
null
null
null
null
UTF-8
Python
false
false
120
py
import numpy as np from matplotlib import pyplot as plt x=np.arange(0,3*np.pi,0.1) y=np.tan(x) plt.plot(x,y) plt.show()
[ "divyajnvh@gmail.com" ]
divyajnvh@gmail.com
c5020aa411c33ba9eb808cd247fe814f9c0ece17
8f5f92beeaefcd9effc93da87b26acb5ea159274
/xtorch/modules/seq2seq_encoders/seq2seq_encoder.py
edcdada140696dba36c224bbb20440c20a1c8b5f
[ "MIT" ]
permissive
altescy/xtorch
15f984bf08654dc00fc1be603cca696676428cc1
bcbbbe645f4d62c211af5b3555c526cc60792c32
refs/heads/main
2023-04-12T15:45:52.192602
2021-04-25T11:35:45
2021-04-25T11:35:45
361,373,990
0
0
null
null
null
null
UTF-8
Python
false
false
805
py
from typing import Optional import torch class Seq2seqEncoder(torch.nn.Module): def forward( self, inputs: torch.Tensor, mask: Optional[torch.BoolTensor] = None, ) -> torch.Tensor: """ Parameters ========== inputs: `torch.Tensor` Tensor of shape (batch_size, sequence_length, embedding_size). mask: `torch.BoolTensor`, optional (default = None) BoolTensor of shape (batch_size, sequence_length). Return ====== output: Tensor of shape (batch_size, sequence_length, encoding_size). """ raise NotImplementedError def get_input_dim(self) -> int: raise NotImplementedError def get_output_dim(self) -> int: raise NotImplementedError
[ "altescy@fastmail.com" ]
altescy@fastmail.com
08ffa38d648d852c717fc3bd315fe33ff3f51547
2a8b5252ab21612e692999a4ca2d7d2e3a1d5143
/table.py
ae467dd8c4cbaf6b7693b5d680cdf9434c5cc66b
[]
no_license
faunic/ideal-enigma
777835d77d0148d2fda9a2070127d14a1a870b3e
bf55c84f7879fc1c98a68cb37170536788ecdb56
refs/heads/main
2023-05-07T15:28:48.476205
2021-05-25T12:54:34
2021-05-25T12:54:34
347,642,502
0
0
null
null
null
null
UTF-8
Python
false
false
733
py
table = { 'ATA':'I', 'ATC':'I', 'ATT':'I', 'ATG':'M', 'ACA':'T', 'ACC':'T', 'ACG':'T', 'ACT':'T', 'AAC':'N', 'AAT':'N', 'AAA':'K', 'AAG':'K', 'AGC':'S', 'AGT':'S', 'AGA':'R', 'AGG':'R', 'CTA':'L', 'CTC':'L', 'CTG':'L', 'CTT':'L', 'CCA':'P', 'CCC':'P', 'CCG':'P', 'CCT':'P', 'CAC':'H', 'CAT':'H', 'CAA':'Q', 'CAG':'Q', 'CGA':'R', 'CGC':'R', 'CGG':'R', 'CGT':'R', 'GTA':'V', 'GTC':'V', 'GTG':'V', 'GTT':'V', 'GCA':'A', 'GCC':'A', 'GCG':'A', 'GCT':'A', 'GAC':'D', 'GAT':'D', 'GAA':'E', 'GAG':'E', 'GGA':'G', 'GGC':'G', 'GGG':'G', 'GGT':'G', 'TCA':'S', 'TCC':'S', 'TCG':'S', 'TCT':'S', 'TTC':'F', 'TTT':'F', 'TTA':'L', 'TTG':'L', 'TAC':'Y', 'TAT':'Y', 'TAA':'_', 'TAG':'_', 'TGC':'C', 'TGT':'C', 'TGA':'_', 'TGG':'W', }
[ "noreply@github.com" ]
faunic.noreply@github.com
44b5d7f99598cb32ba93499d8f2d394a484573cf
49edf974d6502d339095601b101e705911426c07
/project_files/noah_analysis/ordered_analysis/process_mle_slope_int_soft.py
9a78dc727874ee47ab3e7cc339e6e3007ae25bdc
[]
no_license
minghao2016/protein_design_and_site_variability
1ca7b7486464b109cb1b054e8d20fe15af41976a
605ad641c0061234f841e3ceed9d885dabc0d2ce
refs/heads/master
2021-05-30T05:32:37.344639
2015-06-27T04:50:25
2015-06-27T04:50:25
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,174
py
import os, re import subprocess from subprocess import Popen from numpy import * from scipy.stats import pearsonr as pearson #list of files that contain L vs RSA data data = [] #search string to use searchStr = "^align_data_array_ordered" + "[a-zA-Z0-9_\.\-]*" + "_soft.dat" x = [0.0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9] fpO = open("raw_mle_lines_ordered_soft_noah.csv","w") #find all csv files that match the search string for path, names, filename in os.walk('.',False): for file in filename: #print "Testing file: %s" %file if(re.search(searchStr, file)!=None): print "Found file: %s" % file output = subprocess.Popen(["/usr/bin/Rscript /Users/Eleisha/Documents/Wilke_Lab/Project_1/project/r_scripts/MLE_calc.R " + file], shell=True, stdout=subprocess.PIPE).communicate() print output result = re.findall("\-?[0-9]+\.?[0-9]*", re.split("\n",output[0])[-2]) print result slop = result[0] int = result[1] print file + " y(x) = " + str(int) + " + x" + str(slop) fpO.write(file+","+str(int)+","+str(slop)+"\n") data.append([int, slop]) fpO.close()
[ "eleishaj@utexas.edu" ]
eleishaj@utexas.edu
2f0cb96aaa337f7309712bd930d65de11673c433
55c250525bd7198ac905b1f2f86d16a44f73e03a
/Python/Pytest/pytest-django/pytest_django/plugin.py
cbfe15f79cb04f0e152ebe02bc8b4d3886108f5f
[ "BSD-3-Clause" ]
permissive
NateWeiler/Resources
213d18ba86f7cc9d845741b8571b9e2c2c6be916
bd4a8a82a3e83a381c97d19e5df42cbababfc66c
refs/heads/master
2023-09-03T17:50:31.937137
2023-08-28T23:50:57
2023-08-28T23:50:57
267,368,545
2
1
null
2022-09-08T15:20:18
2020-05-27T16:18:17
null
UTF-8
Python
false
false
130
py
version https://git-lfs.github.com/spec/v1 oid sha256:4b9c174912c01ae59fb496601d8c4ecf26765ee33134d079295304c25873875a size 26008
[ "nateweiler84@gmail.com" ]
nateweiler84@gmail.com
d0e4e18eaed6527f644f4b7cd84bcb12327c970d
1ae45340a0c8d64a481b1fb96d835fcd537539bd
/test/unit/controllers/job_list_api_test.py
2f405dc01d04f4614d82c1733a55da864307a63f
[ "MIT" ]
permissive
hazmat345/brew-view
6bc7fcad16bf73505f46e2426aabbff4e4e75c65
effd67819f7e995595471e0dc1c4e03a63942b96
refs/heads/master
2020-04-14T18:02:09.907639
2018-10-09T19:15:37
2018-10-09T19:15:37
164,003,765
0
0
MIT
2019-01-03T17:35:35
2019-01-03T17:35:35
null
UTF-8
Python
false
false
3,218
py
# -*- coding: utf-8 -*- import copy import json from datetime import datetime from mock import patch from bg_utils.models import Job, DateTrigger, RequestTemplate from . import TestHandlerBase class JobListAPITest(TestHandlerBase): def setUp(self): self.ts_epoch = 1451606400000 self.ts_dt = datetime(2016, 1, 1) self.job_dict = { 'name': 'job_name', 'trigger_type': 'date', 'trigger': { 'run_date': self.ts_epoch, 'timezone': 'utc', }, 'request_template': { 'system': 'system', 'system_version': '1.0.0', 'instance_name': 'default', 'command': 'speak', 'parameters': {'message': 'hey!'}, 'comment': 'hi!', 'metadata': {'request': 'stuff'}, }, 'misfire_grace_time': 3, 'coalesce': True, 'next_run_time': self.ts_epoch, 'success_count': 0, 'error_count': 0, 'status': 'RUNNING', } db_dict = copy.deepcopy(self.job_dict) db_dict['request_template'] = RequestTemplate(**db_dict['request_template']) db_dict['trigger']['run_date'] = self.ts_dt db_dict['trigger'] = DateTrigger(**db_dict['trigger']) db_dict['next_run_time'] = self.ts_dt self.job = Job(**db_dict) super(JobListAPITest, self).setUp() def tearDown(self): Job.objects.delete() def test_get(self): self.job.save() self.job_dict['id'] = str(self.job.id) response = self.fetch('/api/v1/jobs') self.assertEqual(200, response.code) self.assertEqual(json.loads(response.body.decode('utf-8')), [self.job_dict]) def test_get_with_filter_param(self): self.job.save() self.job_dict['id'] = str(self.job.id) response = self.fetch('/api/v1/jobs?name=DOES_NOT_EXIST') self.assertEqual(200, response.code) self.assertEqual(json.loads(response.body.decode('utf-8')), []) response = self.fetch('/api/v1/jobs?name=job_name') self.assertEqual(200, response.code) self.assertEqual( json.loads(response.body.decode('utf-8')), [self.job_dict] ) @patch('brew_view.request_scheduler') def test_post(self, scheduler_mock): body = json.dumps(self.job_dict) self.job_dict['id'] = None response = self.fetch('/api/v1/jobs', method='POST', body=body) self.assertEqual(response.code, 201) data_without_id = json.loads(response.body.decode('utf-8')) data_without_id['id'] = None self.assertEqual( data_without_id, self.job_dict ) self.assertEqual(scheduler_mock.add_job.call_count, 1) @patch('brew_view.request_scheduler') def test_post_error_delete(self, scheduler_mock): body = json.dumps(self.job_dict) self.job_dict['id'] = None scheduler_mock.add_job.side_effect = ValueError response = self.fetch('/api/v1/jobs', method='POST', body=body) self.assertGreaterEqual(response.code, 500)
[ "loganasherjones@gmail.com" ]
loganasherjones@gmail.com
1731a6bc44fffbafb6437d4bb39a9bb76acfeb29
45c170fb0673deece06f3055979ece25c3210380
/toontown/coghq/BossbotCountryClubMazeRoom_Battle00.py
218b80966c9553066709cc1c2f781554cc97b785
[]
no_license
MTTPAM/PublicRelease
5a479f5f696cfe9f2d9dcd96f378b5ce160ec93f
825f562d5021c65d40115d64523bb850feff6a98
refs/heads/master
2021-07-24T09:48:32.607518
2018-11-13T03:17:53
2018-11-13T03:17:53
119,129,731
2
6
null
2018-11-07T22:10:10
2018-01-27T03:43:39
Python
UTF-8
Python
false
false
2,389
py
#Embedded file name: toontown.coghq.BossbotCountryClubMazeRoom_Battle00 from toontown.coghq.SpecImports import * GlobalEntities = {1000: {'type': 'levelMgr', 'name': 'LevelMgr', 'comment': '', 'parentEntId': 0, 'cogLevel': 0, 'farPlaneDistance': 1500, 'modelFilename': 'phase_12/models/bossbotHQ/BossbotMazex1_C', 'wantDoors': 1}, 1001: {'type': 'editMgr', 'name': 'EditMgr', 'parentEntId': 0, 'insertEntity': None, 'removeEntity': None, 'requestNewEntity': None, 'requestSave': None}, 0: {'type': 'zone', 'name': 'UberZone', 'comment': '', 'parentEntId': 0, 'scale': 1, 'description': '', 'visibility': []}, 110000: {'type': 'battleBlocker', 'name': '<unnamed>', 'comment': '', 'parentEntId': 0, 'pos': Point3(-131.21, 84.92, 0), 'hpr': Point3(270, 0, 0), 'scale': Vec3(1, 1, 1), 'cellId': 0, 'radius': 10}, 110202: {'type': 'door', 'name': '<unnamed>', 'comment': '', 'parentEntId': 110001, 'pos': Point3(0, 0, 0), 'hpr': Vec3(0, 0, 0), 'scale': 1, 'color': Vec4(1, 1, 1, 1), 'isLock0Unlocked': 1, 'isLock1Unlocked': 0, 'isLock2Unlocked': 1, 'isLock3Unlocked': 1, 'isOpen': 0, 'isOpenEvent': 0, 'isVisBlocker': 0, 'secondsOpen': 1, 'unlock0Event': 0, 'unlock1Event': 110000, 'unlock2Event': 0, 'unlock3Event': 0}, 110002: {'type': 'maze', 'name': '<unnamed>', 'comment': '', 'parentEntId': 0, 'pos': Point3(-141.563, -78.8353, 0), 'hpr': Vec3(0, 0, 0), 'scale': Vec3(1, 1, 1), 'numSections': 1}, 10002: {'type': 'nodepath', 'name': 'props', 'comment': '', 'parentEntId': 0, 'pos': Point3(0, 0, 0), 'hpr': Vec3(0, 0, 0), 'scale': 1}, 110001: {'type': 'nodepath', 'name': '<unnamed>', 'comment': '', 'parentEntId': 0, 'pos': Point3(-106.91, 82.6953, 0), 'hpr': Point3(270, 0, 0), 'scale': Vec3(1, 1, 1)}} Scenario0 = {} levelSpec = {'globalEntities': GlobalEntities, 'scenarios': [Scenario0]}
[ "linktlh@gmail.com" ]
linktlh@gmail.com
3f0b424620cadbd7007d11df02a06feeb6089c28
24c2132b45590c3e1af9b8383fc3d3b4d85afb1f
/20_DNN/scratch/nn/basic/nn_mnist_batch.py
9a93d20ec7bc9520e769dd3c536b337fc4d8ca68
[]
no_license
harperfu6/ML_Tips
cbee8029ec8b5ef1a03d0b0e3bb9818e73e56442
e5776d17102113fc5e3187b1cdfb4d4bafe891f4
refs/heads/master
2022-02-27T09:24:42.946713
2019-11-28T13:15:48
2019-11-28T13:15:48
183,880,194
0
0
null
null
null
null
UTF-8
Python
false
false
1,374
py
# coding: utf-8 # データセットに対し,まとめて予測するだけ import sys, os sys.path.append(os.pardir) # 親ディレクトリのファイルをインポート import numpy as np import pickle from dataset.mnist import load_mnist from common.functions import sigmoid, softmax def get_data(): (x_train, t_train), (x_test, t_test) = load_mnist(normalize=True, flatten=True, one_hot_label=False) return x_test, t_test def init_network(): with open('sample_weight.pkl', 'rb') as f: network = pickle.load(f) return network def predict(network, x): W1, W2, W3 = network['W1'], network['W2'], network['W3'] b1, b2, b3 = network['b1'], network['b2'], network['b3'] a1 = np.dot(x, W1) + b1 z1 = sigmoid(a1) a2 = np.dot(z1, W2) + b2 z2 = sigmoid(a2) a3 = np.dot(z2, W3) + b3 y = sigmoid(a3) return y def main(): x, t = get_data() network = init_network() batch_size = 100 # バッチ数 accuracy_cnt = 0 for i in range(0, len(x), batch_size): x_batch = x[i:i+batch_size] y_batch = predict(network, x_batch) p = np.argmax(y_batch, axis=1) # 最も確率の高い要素のインデックスを取得 accuracy_cnt += np.sum(p == t[i:i+batch_size]) print('Accuracy:' + str(float(accuracy_cnt) / len(x))) if __name__ == '__main__': main()
[ "beharp8a8@gmail.com" ]
beharp8a8@gmail.com
07737492b88c075fe7d11b3f01a276520d1b854b
f7e459e0a9bc5bfa7c635abe6ed9c922bae27339
/dfvfs/analyzer/luksde_analyzer_helper.py
465bd7ffdc366edbcf727ae526ca4f4605627966
[ "Apache-2.0" ]
permissive
sanjaymsh/dfvfs
bcf5153a1743cb4bdc1d9e5bd45e383a2e6e675d
049c71df15f46ac0ef552f0c6f71f7c61797af87
refs/heads/master
2022-12-23T23:25:14.035173
2020-09-27T06:06:33
2020-09-27T06:06:33
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,012
py
# -*- coding: utf-8 -*- """The LUKSDE format analyzer helper implementation.""" from __future__ import unicode_literals from dfvfs.analyzer import analyzer from dfvfs.analyzer import analyzer_helper from dfvfs.analyzer import specification from dfvfs.lib import definitions class LUKSDEAnalyzerHelper(analyzer_helper.AnalyzerHelper): """LUKSDE analyzer helper.""" FORMAT_CATEGORIES = frozenset([ definitions.FORMAT_CATEGORY_VOLUME_SYSTEM]) TYPE_INDICATOR = definitions.TYPE_INDICATOR_LUKSDE def GetFormatSpecification(self): """Retrieves the format specification. Returns: FormatSpecification: format specification or None if the format cannot be defined by a specification object. """ format_specification = specification.FormatSpecification( self.type_indicator) # LUKSDE signature. format_specification.AddNewSignature(b'LUKS\xba\xbe', offset=0) return format_specification analyzer.Analyzer.RegisterHelper(LUKSDEAnalyzerHelper())
[ "noreply@github.com" ]
sanjaymsh.noreply@github.com
4c6318615664d63994ac62eccd6557cda896e2a9
2476d1dbac904aed784cc51550f74febe3b783c4
/farmer/config/__init__.py
f5de67a172870543bf9b9c7b151f084a878d6972
[]
no_license
9231058/farmer
ccdbae0cc168adfa84bda016d87809ee5797c554
83904d86e30ab31622eaa0d4534041a4972bf842
refs/heads/master
2022-11-08T04:57:54.166269
2020-06-18T21:21:39
2020-06-18T21:21:39
266,701,420
2
1
null
null
null
null
UTF-8
Python
false
false
58
py
from .config import Config, RequiredSeed, NonrequiredSeed
[ "parham.alvani@gmail.com" ]
parham.alvani@gmail.com
01cc3441b94a034279d96bc8ad271736ee058bcd
8a2a76d66a92c91ee07c44e533f8bbc81778ff28
/rund.py
76776386f4c4858a96a99032c7460c3953455839
[]
no_license
Shashant-R/GSCEventMOD
9941935caf966ca601b1e3c59b05eefd7dc18c67
25e85af84b12f725a188f6b872d44010f2f385bc
refs/heads/main
2023-08-14T00:01:00.469037
2021-09-23T12:55:32
2021-09-23T12:55:32
409,591,561
0
1
null
2021-10-01T18:17:30
2021-09-23T12:54:45
Python
UTF-8
Python
false
false
5,774
py
import os import matplotlib.pyplot as plt import random import h5py import numpy as np import warnings from sklearn.neighbors import kneighbors_graph from sklearn.cluster import SpectralClustering from sklearn.cluster import DBSCAN from sklearn.cluster import MeanShift from sklearn.metrics import silhouette_score from sklearn.metrics import calinski_harabasz_score from sklearn.ensemble import IsolationForest warnings.filterwarnings('ignore', '.*Graph is not fully connected*') print('reading Cars_sequence...') file_name = "Object Motion Data (mat files)/Cars_sequence.mat" f = h5py.File(file_name, "r") davis = f['davis'] dvs = davis['dvs'] pol = dvs['p'][0] #ts = dvs['t'][0] ts = np.load("street_ts.npy") ts = ts*0.000001 #x = dvs['x'][0] #y = dvs['y'][0] aps_ts = np.load("street_img_ts.npy") #dvs_ts = np.load("cars_all_ts.npy") print(len(ts), len(aps_ts)) ''' # events frequency distribution y_eve = [] i = 0 ctr = 0 j = 0 while i<len(ts): if ts[i] < aps_ts[j]: ctr += 1 else: y_eve.append(ctr) ctr = 1 j += 1 if j==len(aps_ts): break i += 1 np.save("event_dist_street.npy", np.asarray(y_eve)) ''' # plot frequency distribution y_eve = np.load("event_dist_street.npy") print(y_eve) print(len(y_eve)) fig = plt.figure() plt.bar(range(200), y_eve, color='r') plt.xlabel("Segments") plt.ylabel("No. of events") plt.title("Frequency of events in different segments") plt.show() print(sum(y_eve)) ''' #without cleaning n = len(dvs_ts) last = 0 ALL = len(pol) NEIGHBORS = 100 ctr = -1 for idx in dvs_ts: ctr+=1 xx = '0000000000' yy = str(ctr) file_name = xx[:len(xx) - len(yy)] + yy print(last) selected_events = [] for i in range(0, ALL)[last:idx]: selected_events.append([y[i], x[i], ts[i] * 0.0001, pol[i] * 0]) if len(selected_events)==6000: break last = idx selected_events = np.asarray(selected_events) cleaned_events = IsolationForest(random_state=0, n_jobs=-1, contamination=0.05).fit(selected_events) unwanted_events = cleaned_events.predict(selected_events) selected_events = selected_events[np.where(unwanted_events == 1, True, False)] adMat = kneighbors_graph(selected_events, n_neighbors=NEIGHBORS) max_score = -20 opt_clusters = 2 scores = [] print('predicting number of clusters...') for CLUSTERS in range(2, 10): clustering = SpectralClustering(n_clusters=CLUSTERS, random_state=0, affinity='precomputed_nearest_neighbors', n_neighbors=NEIGHBORS, assign_labels='kmeans', n_jobs=-1).fit_predict(adMat) curr_score = silhouette_score(selected_events, clustering) scores.append(curr_score) if curr_score > max_score: max_score = curr_score opt_clusters = CLUSTERS np.save(os.path.join('results/656/predict_k', file_name + '.npy'), np.asarray(scores)) clustering = SpectralClustering(n_clusters=opt_clusters, random_state=0, affinity='precomputed_nearest_neighbors', n_neighbors=NEIGHBORS, assign_labels='kmeans', n_jobs=-1).fit_predict(adMat) np.save(os.path.join('results/656/selected_events', file_name + '.npy'), selected_events) np.save(os.path.join('results/656/clusters', file_name + '.npy'), clustering) print('done') ''' ''' # indices of nearest timestamps event_idx = [] for t in aps_ts: idx_t = (np.abs(ts - t)).argmin() print(t) event_idx.append(idx_t) event_idx = np.asarray(event_idx) np.save("cars_all_ts.npy", event_idx) print(len(event_idx)) ''' ''' #with cleaning and cluster prediction ALL = len(pol) NEIGHBORS = 30 print(str(ALL)+' events in dataset...') seg = 64 while seg >= 64: print('dividing the sequence into '+str(seg)+' segments...') X = ALL//seg print('each segment has '+str(X)+' events, out of which '+str(X//4)+' events will be selected...') for sl_no in range(seg): print('segment no: '+str(sl_no+1)) selected_events = [] for i in range(0,ALL)[sl_no*X:sl_no*X+X:4]: selected_events.append([y[i], x[i], ts[i]*0.0001, pol[i]*0]) selected_events = np.asarray(selected_events) cleaned_events = IsolationForest(random_state=0, n_jobs=-1, contamination=0.1).fit(selected_events) unwanted_events = cleaned_events.predict(selected_events) selected_events_cleaned = selected_events[np.where(unwanted_events == 1, True, False)] adMat_cleaned = kneighbors_graph(selected_events_cleaned, n_neighbors=NEIGHBORS) print('clustering...') clustering_cleaned = SpectralClustering(n_clusters=2, random_state=0, affinity='precomputed_nearest_neighbors', n_neighbors=NEIGHBORS, assign_labels='kmeans', n_jobs=-1).fit_predict(adMat_cleaned) xx = '0000000000' yy = str(sl_no) file_name = xx[:len(xx) - len(yy)] + yy np.save(os.path.join('results/clean/64/selected_events', file_name+'.npy'), selected_events_cleaned) np.save(os.path.join('results/clean/64/clusters', file_name + '.npy'), clustering_cleaned) seg = seg // 2 break print('done') '''
[ "noreply@github.com" ]
Shashant-R.noreply@github.com
ec67fcd528e10f1d86e9a1a620b455fc3c212ba6
10f57edc51f50742625b405f2f7c37cdd4734700
/app/recipe/urls.py
034085c433bce22da70cdc48e8a58fb0632aba75
[ "MIT" ]
permissive
smkempin/recipe-app-api
5eea18b812bf822c292db8b66499ac1020dd886c
453223e68616ff092964c8414ede5cccc2e38351
refs/heads/master
2021-01-26T05:17:09.742683
2020-03-30T13:13:26
2020-03-30T13:13:26
243,324,015
0
0
null
null
null
null
UTF-8
Python
false
false
359
py
from django.urls import path, include from rest_framework.routers import DefaultRouter from recipe import views router = DefaultRouter() router.register('tags', views.TagViewSet) router.register('ingredients', views.IngredientViewSet) router.register('recipes', views.RecipeViewSet) app_name = 'recipe' urlpatterns=[ path('', include(router.urls)) ]
[ "scott@precisionwre.com" ]
scott@precisionwre.com
e32d9ecd5addc70ef1833cfb869c834a230a4f2c
7f97814acd76ca96aee877fd70d401380f848fae
/7_training/re_start_end.py
e5842c00b391813441ccd2346854697e29805bbb
[]
no_license
tberhanu/all_trainings
80cc4948868928af3da16cc3c5b8a9ab18377d08
e4e83d7c71a72e64c6e55096a609cec9091b78fa
refs/heads/master
2020-04-13T12:12:21.272316
2019-03-16T04:22:20
2019-03-16T04:22:20
163,195,802
0
0
null
null
null
null
UTF-8
Python
false
false
485
py
""" https://www.hackerrank.com/challenges/re-start-re-end/problem?h_r=next-challenge&h_v=zen """ # Enter your code here. Read input from STDIN. Print output to STDOUT import re s, k = input(), input() i = 0 found = False while i < len(s): string = s[i:] match = re.match(r'{}'.format(k), string) if match == None: i = i + 1 else: found = True print((match.start() + i, match.end() + i - 1)) i = i + 1 if not found: print('(-1, -1')
[ "tberhanu@berkeley.edu" ]
tberhanu@berkeley.edu
16f02a9531c8dbb7c2e6d252e5094a83efbd7217
a89bcfe5a2fff6727a39a64e36e92a5f5a72644f
/929_unique_email_addresses/solution.py
fa0a932ff6febfdfcffa431123abb50352a9fc0e
[]
no_license
vanshaw2017/leetcode_vanshaw
6fdde2c012f53470efa9f4b13b0d123f3fef0e89
12393cfaf4b1b758e8a0407787a2a8150285678d
refs/heads/master
2020-04-06T13:59:18.119910
2019-02-25T02:10:28
2019-02-25T02:10:28
157,522,844
1
0
null
null
null
null
UTF-8
Python
false
false
473
py
class Solution: def numUniqueEmails(self, emails: 'List[str]') -> 'int': result = [] for i in emails: local = i.split("@")[0] far = i.split("@")[1] if '+' in local: local = local.split("+")[0] if '.' in local: local = local.replace('.', '') email = local + far if email not in result: result.append(email) return len(result)
[ "614664248@qq.com" ]
614664248@qq.com
5a21f18a39ff0c0dda0b64a7e34fc7fc468fd600
330146ad205bb1c21b63c2eeaf11c8a2996e2a4f
/mylab/coursera/timeofday.py
21979edc23726092bf550c7561ed29d32721bbd8
[]
no_license
bkrishna2006/system
f3c63c03cbdf817f102a7b84075018a21e9e3c54
c478c7a001722af2b300fdaca9e10fdbec4d6a04
refs/heads/master
2020-04-10T20:01:49.806318
2018-01-22T16:34:36
2018-01-22T16:34:36
null
0
0
null
null
null
null
UTF-8
Python
false
false
541
py
filename = raw_input("Enter your file name: ") try: fd = open(filename) except: print "File not found: %s" % filename quit() theDict = dict() for line in fd: words = line.split() if len(words) < 1 or words[0] != "From" or words[0][-1] == ":" or len(words) < 6: continue time = words[5] time_split = time.split(":") if len(time_split) < 1: continue hour = time_split[0] theDict[hour] = theDict.get(hour, 0) + 1 result = theDict.items() result.sort() for hour, count in result: print hour, count
[ "luke.nothingness@gmail.com" ]
luke.nothingness@gmail.com
e5051b8cb2577762cfa4eefebf5dafbcea28c428
56b69a58a8844d09e213dc38aab9aa62422dd58e
/128.py
9f90303f0eefb77e7f24e3e9b2c106715dc321dc
[]
no_license
dilkas/project-euler
d697daf4087a0b436e2dc7b2840d5e53c2ff07b8
e637fc34d406de7b05755d9c85b370aef1beb2a7
refs/heads/master
2021-05-31T15:46:45.342257
2016-05-15T10:29:45
2016-05-15T10:29:45
null
0
0
null
null
null
null
UTF-8
Python
false
false
517
py
import math target = 2000 def prime(n): for k in range(2, math.floor(n ** 0.5) + 1): if n % k == 0: return False return True n = 1 counter = 0 while True: if prime(6 * n - 1) and prime(6 * n + 1) and prime(12 * n + 5): counter += 1 if counter == target: print(3 * n * n - 3 * n + 2) break if prime(6 * n - 1) and prime(6 * n + 5) and prime(12 * n - 7): counter += 1 if counter == target: print(3 * n * n + 3 * n + 1) break n += 1
[ "paulius.dilkas@gmail.com" ]
paulius.dilkas@gmail.com
eb78517e4f6aebbbb42fb241163e359f5183bcfc
9feb3d3c59113506cbd5e86aef84f42fb257f165
/Session04/cross_check.py
79654247efdbdc93feafa2e1ae9dc9a5d95a1dc6
[]
no_license
reanimation47/ledoananhquan-fundametal-c4e13
112a67bbb3093e62f9c53d9f0e7fc57870384fdd
6e1e576702c4d68e0aa33a3b21a193b112fa70b6
refs/heads/master
2021-09-03T11:03:52.598663
2018-01-08T14:33:34
2018-01-08T14:33:34
108,993,661
2
0
null
null
null
null
UTF-8
Python
false
false
112
py
# l = [0, 1, 2, 3, 4] n = 4 #0,1,2,3 for i in range(n-1): for j in range(i + 1, n): print(i,"vs",j)
[ "reanimation47@gmail.com" ]
reanimation47@gmail.com
16a9c914cb9d2272c13f251203a93fa646574f5b
3e2f7ff88aabbf17ee93a30176a40396adfbc7ec
/core/migrations/0003_auto_20171118_0536.py
fdb20ee9327f420442d4608bc28d9fa64520b0bd
[]
no_license
mahima-c/uhvpe
04cbc519904261b45193dea4450f646b83a4184d
9dc47d4fae17b6e405d5d9500c75585c45afbf20
refs/heads/master
2021-05-25T19:08:21.414764
2019-10-15T07:19:13
2019-10-15T07:19:13
null
0
0
null
null
null
null
UTF-8
Python
false
false
896
py
# -*- coding: utf-8 -*- # Generated by Django 1.11.7 on 2017-11-18 05:36 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('core', '0002_auto_20171117_1726'), ] operations = [ migrations.CreateModel( name='Presentation', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=50, unique=True)), ('language', models.CharField(choices=[('HI', 'Hindi'), ('EN', 'English')], max_length=10)), ('file', models.FileField(unique=True, upload_to='presentation/')), ], ), migrations.RemoveField( model_name='workshop', name='start_date', ), ]
[ "apoorvapandey365@gmail.com" ]
apoorvapandey365@gmail.com
be2ea70bbdeaae35443f15931b34646e0979c465
489ffb5efea81b6f374037d1e4b5856041f6a5a1
/main.py
c555233cd622eca6a3b73e8d8fa869339f3abc8f
[]
no_license
petar-tomov/RentACar
c183223e5b8b28dca6fe384904d00d99dbec0de0
cc9d3d36b72f99f8f4641f9f3d9bba027d1ea394
refs/heads/main
2023-03-18T21:06:09.081059
2021-03-12T08:31:04
2021-03-12T08:31:04
345,719,621
0
0
null
null
null
null
UTF-8
Python
false
false
638
py
from Customer import Customer pesho = Customer("Pesho") pesho.check_catalogue() pesho.rentacar_hours("PB5079TT", 5) pesho.rentacar_day("CB4078BM") pesho.rentacar_week("PB9189PC") pesho.rentacar_hours("PB9189PC", 10) # You can't rent a car which is already rented pesho.rentacar_day("PB9999AH") # You also can't rent a car which isn't in the catalogue, of course pesho.rentacar_hours("PB0168MX", 3) # Pesho's fourth car pesho.checkout() drago = Customer("Drago") drago.check_catalogue() # The cars rented by Pesho are no longer in the catalogue drago.rentacar_hours("PA5460AB", 4) drago.rentacar_hours("EB6633AH", 4) drago.checkout()
[ "pepyy.tommyy@gmail.com" ]
pepyy.tommyy@gmail.com
f7e150fe2c2755c2ee7c29b21367a36323defc16
e8977e740aa31a501e0be51ab9df159e1e9834d7
/Main.py
da7f5e22c36be88b80fa54d361a9836d5f4eb343
[]
no_license
bharatkumar7/stuff
67a06e552aa6839f596e508cf26973df36574ad2
99cb02e6926d6465e4773df4c178f278d9bc7fa6
refs/heads/master
2021-01-10T14:09:25.814402
2016-01-13T11:03:36
2016-01-13T11:03:36
49,219,047
1
0
null
null
null
null
UTF-8
Python
false
false
16,119
py
'''# -*- coding: utf-8 -*- If you are just trying to use UTF-8 characters or don't care if they are in your code, add this line''' import numpy as np import time,os,talib import matplotlib.pyplot as plt from matplotlib.finance import candlestick_ohlc, candlestick2_ohlc from matplotlib import gridspec,colors from math import pi import funct import mibian import pandas as pd from pandas import * stock_data=1 # 1 - Streamline data, 2 - OHLC data tickskip=1 #1 denotes every tick colm=0 # This must be dump 123456789 for k in range(0,1): if k==0:fname="code07.01.16.txt" #if k==0:fname="1sec_close.txt" status = 0 pruy=0 AccountSize=1000000 maxtick=25 target_profit=1.02 #Risk to Reward ratio 3:1 stop_loss=0.97 #stop loss when entering a position tpf=target_profit tsl=stop_loss graph=1 #1 - Yes, 0 - None tick_analysis_display=1 #1 - Yes , 0 - No exchange_charges=8 # 16.69 for equities. 90 for option. 8 for futures file_format=2 # 1- OHLC format, 2 - Live ticker format(Time OHLC), 3 - Live ticker format(Range OHLC) tf_sec='30s' initcap=AccountSize #----------------------- File reader ------------------------------------------------------------------------------------------------------ if file_format==1: tedhi,price_open,price_high,price_low,price_close, vol=(np.loadtxt(fname, dtype=float, delimiter=',', usecols=(1,3,4,5,6,7), skiprows=1,unpack=True)) numb=len(price_open) xoi=np.min(price_low) yoi=np.max(price_high) xoi=xoi-(xoi*0.05) yoi=yoi+(yoi*0.05) elif file_format==2: tedhi,dummy=(np.loadtxt(fname, dtype=str, delimiter=',', usecols=(0,1),skiprows=1,unpack=True)) cl_price, vol=(np.loadtxt(fname, dtype=float, delimiter=',', usecols=(15,2), skiprows=1,unpack=True)) d={'Datetime': Series(to_datetime(tedhi.astype(int)*1e9)), 'price': Series((cl_price)), 'volume':Series((vol))} df=DataFrame(d) df.set_index('Datetime', inplace=True) vol_sum = (df['volume'].resample(tf_sec, how='sum',fill_method='backfill',limit=0)).dropna(how='any', axis=0) price_ohlc = (df['price'].resample(tf_sec, how='ohlc',fill_method='backfill',limit=0)).dropna(how='any', axis=0) numb=len(price_ohlc) xoi=np.min(price_ohlc.low) yoi=np.max(price_ohlc.high) xoi=xoi-(xoi*0.001) yoi=yoi+(yoi*0.001) elif file_format==3: cl_price, vol=(np.loadtxt(fname, dtype=float, delimiter=',', usecols=(11,16), skiprows=1,unpack=True)) if cl_price[0]<=25:rangeper=3.5/100 #constant range percent "0.01" represents 1% elif cl_price[0]<=50:rangeper=2.50/100 elif cl_price[0]<=100:rangeper=2.0/100 elif cl_price[0]<=250:rangeper=2.5/100 else: rangeper=0.001/100 price_open,price_high,price_low,price_close=funct.range_bar(cl_price,vol,rangeper) numb=len(price_close) xoi=np.min(price_low) yoi=np.max(price_high) xoi=xoi-(xoi*0.001) yoi=yoi+(yoi*0.001) #----------------------- END OF FILE READER ------------------------------------------------------------------------------------------------------ tdata_ltp=np.zeros(maxtick,'f') tdata_vol=np.zeros(maxtick,'f') tdata_op=np.zeros(maxtick,'f') tdata_hi=np.zeros(maxtick,'f') tdata_lo=np.zeros(maxtick,'f') buy=np.full(numb-maxtick,-10) #full fills all the array with the defned number "-10" this case bstime=np.arange(0,numb-maxtick) #arange puts all real numbers in order 1,2,3 sell=np.full(numb-maxtick,-10) short=np.full(numb-maxtick,-10) cover=np.full(numb-maxtick,-10) alltrades=np.zeros(numb-maxtick) PLP=np.zeros(numb-maxtick) PL=np.zeros(numb-maxtick) AS=np.zeros(numb-maxtick) minute=np.arange(0,numb) trade=0 fp=0.0 #final percentage pot=AccountSize broker=0.0 AS[0]=AccountSize kfp=0.0 ktrades=0 kpot=0.0 tech_one=np.zeros(numb-maxtick,'f') tech_two=np.zeros(numb-maxtick,'f') tech_three=np.zeros(numb-maxtick,'f') flatornot=np.zeros(numb-maxtick,'f') #flatornot_ema=np.zeros(numb-maxtick,'f') # ---------------------- Analysis ----------------------------------- for i in range(0,numb-maxtick): bstime[i]=i-1+maxtick if file_format==1 or file_format==3: for j in range(i,maxtick+i): tdata_ltp[j-i]=price_close[j] tdata_vol[j-i]=vol[j] tdata_op[j-i]=price_open[j] tdata_hi[j-i]=price_high[j] tdata_lo[j-i]=price_low[j] #time.sleep(2) if file_format==2: for j in range(i,maxtick+i): tdata_ltp[j-i]=price_ohlc.close[j] tdata_vol[j-i]=vol_sum[j] tdata_op[j-i]=price_ohlc.open[j] tdata_hi[j-i]=price_ohlc.high[j] tdata_lo[j-i]=price_ohlc.low[j] float_data = [float(x) for x in tdata_ltp] tdata_ltp = np.array(float_data) #float_data = [float(x) for x in tdata_ltp_nifty] #tdata_ltp_nifty = np.array(float_data) float_data = [float(x) for x in tdata_vol] tdata_vol = np.array(float_data) float_data = [float(x) for x in tdata_op] tdata_op = np.array(float_data) float_data = [float(x) for x in tdata_hi] tdata_hi = np.array(float_data) float_data = [float(x) for x in tdata_lo] tdata_lo = np.array(float_data) upordown_ltp = talib.EMA(tdata_ltp,5) upordown_ltpl = talib.EMA(tdata_ltp,15) upordown_ltplong = talib.EMA(upordown_ltpl,10) kmav = talib.KAMA(tdata_ltp,10) upordown_kmav = talib.EMA(kmav,10) atrv = talib.ATR(tdata_hi, tdata_lo, tdata_ltp,timeperiod=5) #upordown_ltplong = talib.EMA(tdata_ltp,40) '''macd, macdsignal, macdhist = talib.MACD(tdata_ltp, fastperiod=6, slowperiod=13, signalperiod=4) upordown_vol = talib.EMA(tdata_vol,5) rocv = talib.ROC(tdata_ltp,5) tanv = talib.TAN(tdata_ltp) rsiv=talib.RSI(tdata_ltp, 14) tanv = talib.TAN(tdata_ltp)''' #bbuv, bbmv, bblv = talib.BBANDS(tdata_ltp, timeperiod=5, nbdevup=1, nbdevdn=1, matype=0) #adxv=talib.ADX(tdata_hi, tdata_lo, tdata_ltp, timeperiod=14) '''cciv=talib.CCI(tdata_hi, tdata_lo, tdata_ltp, timeperiod=5) ultoscv=talib.ULTOSC(tdata_hi, tdata_lo, tdata_ltp, timeperiod1=5, timeperiod2=10, timeperiod3=15) willrv=talib.WILLR(tdata_hi, tdata_lo, tdata_ltp, timeperiod=5) midpointv = talib.MIDPOINT(tdata_ltp, timeperiod=5) momv=talib.MOM(tdata_ltp, timeperiod=10) stfastkv, stfastdv = talib.STOCHF(tdata_hi, tdata_lo, tdata_ltp, fastk_period=5, fastd_period=3, fastd_matype=0) strsifastkv, strsifastdv = talib.STOCHRSI(tdata_ltp, timeperiod=14, fastk_period=5, fastd_period=3, fastd_matype=0) ''' hhv=np.max(tdata_hi[maxtick-14:maxtick-1]) llv=np.min(tdata_lo[maxtick-14:maxtick-1]) latest=maxtick-1 if (kmav[latest]>1.001*kmav[latest-1]>1.001*kmav[latest-2]>1.001*kmav[latest-3]) or \ (kmav[latest]<0.999*kmav[latest-1]<0.999*kmav[latest-2]<0.999*kmav[latest-3]):flatornot[latest]=kmav[latest] else: flatornot[latest]=0 Uvolt=(hhv-(2.0*atrv[latest])) Dvolt=(llv+(2.0*atrv[latest])) #---- FOR PLOT ----------- if graph==1: tech_one[i]= Uvolt #hhv-(3*atrv[latest])#upordown_ltplong[latest] tech_two[i]=Dvolt #atrv[latest-1] tech_three[i]=flatornot[latest]#adxv[latest] #atrv[latest-1] #print tech_three[i] #d = mibian.GK([8395, 8700, 6, 0, 11], volatility=15.7) #print d.callPrice # ------------------------ ENTRY -------------------------------------- #-------------------Uptrend ------------------------------ '''if (status == 0 and tdata_hi[latest]<tdata_hi[latest-1]) and \ (status == 0 and tdata_hi[latest]<tdata_hi[latest-2]) and \ (status == 0 and tdata_hi[latest]<tdata_hi[latest-3]) and \ (status == 0 and upordown_ltplong[latest]>upordown_ltplong[latest-1]) :''' #if (status == 0 and upordown_ltpl[latest]>upordown_ltpl[latest-1]): if (status == 0 and tdata_ltp[latest]>tdata_op[latest] and tdata_ltp[latest-1]>tdata_op[latest-1]) and \ (status == 0 and tdata_ltp[latest]>Dvolt): trend=1 if flatornot[latest]>0: trend=1 if(trend>=1): status=1 lot_size,StopPrice=funct.PositionSizing_Method(status,AccountSize,stop_loss,tdata_ltp[latest],atrv[latest],Uvolt,Dvolt,alltrades,PL,i) leftamt=AccountSize-(tdata_ltp[latest]*lot_size) if (tick_analysis_display==1):print " BUY - %.2f, Investment - %.2f, Lots - %.2f, Accountsize - %.2f, StopPrice - %.2f"%(tdata_ltp[latest],(lot_size*tdata_ltp[latest]),lot_size,AccountSize,StopPrice) bprice=tdata_ltp[latest] if graph==1:buy[i]=bprice tprice=bprice #-------------------Down trend (sell buy strategy)------------------------------ '''if (status == 0 and tdata_hi[latest]>tdata_hi[latest-1]) and \ (status == 0 and tdata_hi[latest]>tdata_hi[latest-2]) and \ (status == 0 and tdata_hi[latest]>tdata_hi[latest-3]) and \ (status == 0 and upordown_ltplong[latest]<upordown_ltplong[latest-1]) :''' #if (status == 0 and upordown_ltpl[latest]<upordown_ltpl[latest-1]): if (status == 0 and tdata_ltp[latest]<tdata_op[latest] and tdata_ltp[latest-1]<tdata_op[latest-1]) and \ (status == 0 and tdata_ltp[latest]<Uvolt): trend=1 if flatornot[latest]>0: trend=1 if(trend>=1): status=2 lot_size,StopPrice=funct.PositionSizing_Method(status,AccountSize,stop_loss,tdata_ltp[latest],atrv[latest],Uvolt,Dvolt,alltrades,PL,i) leftamt=AccountSize-(tdata_ltp[latest]*lot_size) if (tick_analysis_display==1):print " Short - %.2f, Investment - %.2f, Lots - %.2f, Accountsize - %.2f, StopPrice - %.2f"%(tdata_ltp[latest],(lot_size*tdata_ltp[latest]),lot_size,AccountSize,StopPrice) stprice=tdata_ltp[latest] if graph==1:short[i]=stprice tprice=stprice # ------------------------ EXIT STRATEGY ----------------------------------------- #------------------------ Trailing Stop Loss --------------------------- # SELL trailing for uptrend buy strategy '''if (status == 1 and ((tdata_ltp[latest] >= bprice*target_profit))): # if only bought bprice=tdata_ltp[latest] target_profit=1.0025 stop_loss=0.9975''' #if (status == 1 and tick_analysis_display==1): print "%d, LTP - %.2f, StopPrice - %.2f, Uvolt- %.2f"%(i+maxtick,tdata_ltp[latest],StopPrice,Uvolt) # --------------------------SELL STRATEGY -------------------------------------------------------- '''if (status == 1 and tdata_ltp[latest]<Uvolt) or \ (status == 1 and i==(numb-maxtick-1)) or \ (status == 1 and StopPrice > tdata_ltp[latest]) or \ (status == 1 and stop_loss*tprice > tdata_ltp[latest]) ''' if (status == 1 and tdata_ltp[latest]<tdata_op[latest] and tdata_ltp[latest-1]<tdata_op[latest-1]): #(status == 1 and tdata_ltp[latest] >= tprice*target_profit) : bk_amt=((tprice+tdata_ltp[latest])*lot_size) broker=(bk_amt*exchange_charges/100000.0) # 16.69 for equities. 90 for option. 8 for futures PL[i]=(lot_size*(tdata_ltp[latest]-tprice))-broker PLP[i]=(PL[i]*100/(tprice*lot_size)) fp=fp+PLP[i] pot=pot+PL[i] trade=trade+1 if PLP[i]>0.0: alltrades[i]=1 elif PLP[i]<0.0:alltrades[i]=-1 else:pass if graph==1:sell[i]=tdata_ltp[latest] if (tick_analysis_display==1):print " SELL - %.2f, Percentage %.2f"%(tdata_ltp[latest],PLP[i]) if (tick_analysis_display==1):print "-------------------------------" stop_loss=tsl target_profit=tpf AccountSize=leftamt+PL[i]+(tprice*lot_size) AS[i]=AccountSize status=0 #---------------------------------------------- # Cover trailing for downtrend '''if (status == 2 and ((tdata_ltp[latest]*target_profit <= stprice))): # if only bought stprice=tdata_ltp[latest] target_profit=1.0025 stop_loss=0.9975''' #if (status == 2 and tick_analysis_display==1): print "%d, LTP - %.2f, StopPrice - %.2f,Dvolt- %.2f"%(i+maxtick,tdata_ltp[latest],StopPrice,Dvolt) # -----------------COVER strategy --------------------------------------- '''if (status == 2 and tdata_ltp[latest]>Dvolt) or \ (status == 2 and i==(numb-maxtick-1)) or \ (status == 2 and StopPrice < tdata_ltp[latest]) or \ (status == 2 and tprice < stop_loss*tdata_ltp[latest]) or \ ''' if (status == 2 and tdata_ltp[latest]>tdata_op[latest] and tdata_ltp[latest-1]>tdata_op[latest-1]): #(status == 2 and tdata_ltp[latest]*target_profit <= tprice): bk_amt=((tprice+tdata_ltp[latest])*lot_size) broker=(bk_amt*exchange_charges/100000.0) # 16.69 for equities. 90 for option. 8 for futures PL[i]=-((lot_size*(tdata_ltp[latest]-tprice))+broker) PLP[i]=(PL[i]*100/(tprice*lot_size)) fp=fp+PLP[i] pot=pot+PL[i] trade=trade+1 if PLP[i]>0.0: alltrades[i]=1 elif PLP[i]<0.0:alltrades[i]=-1 else:pass if graph==1:cover[i]=tdata_ltp[latest] if (tick_analysis_display==1):print " Cover - %.2f, Percentage %.2f"%(tdata_ltp[latest],PLP[i]) if (tick_analysis_display==1):print "-------------------------------" stop_loss=tsl target_profit=tpf AccountSize=leftamt+PL[i]+(tprice*lot_size) AS[i]=AccountSize status=0 if AS[i]==0:AS[i]=AccountSize #------------------------------------------------------------------- #if trade>=20 and (pot-broker>=3 or pot-broker<=-15):break #if fp<=0 and trade>5: break #print "-----------------------------------------------------------------" #print " %d. Final Per = %.2f,Final Amt = %.2f, Trades = %d"%(k+1,fp, pot,trade) #print "-----------------------------------------------------------------" kfp=kfp+fp kpot=kpot+pot ktrades=ktrades+trade #print "-------------------------------" #print " %d. File= %s, Final Per = %.2f,Final Amt = %.2f, Trades = %d"%(k+1, fname,kfp,kpot,ktrades) print " %d. Final Amt = %.2f, Trades = %d"%(k+1, kpot,ktrades) #---------------Expectancy --------------------------- funct.Expectancy(alltrades,PLP,PL,numb-maxtick) #------------- PLOTS ---------------------------------------- if graph==1: #---------------- CANDLE STICK PLOTS ---------------------------------------------------------------------------- quotes=np.zeros((numb-1,5)) for i in range(0,numb-1): if file_format==1 or file_format==3:quotes[i]=(minute[i],price_open[i],price_high[i],price_low[i],price_close[i]) if file_format==2:quotes[i]=(minute[i],price_ohlc.open[i],price_ohlc.high[i],price_ohlc.low[i],price_ohlc.close[i]) #axes = plt.gca() #axes.set_xlim([0,numb]) #axes.set_ylim([xoi,yoi]) fig, ax1 = plt.subplots() ax2 = ax1.twinx() ax1.set_xlim([0,numb]) ax2.set_xlim([0,numb]) ax1.set_ylim([xoi,yoi]) ax2.set_ylim([xoi,yoi]) candlestick_ohlc(ax1,quotes,width=0.6, colorup=u'g', colordown=u'r', alpha=1.0) #ax2.plot(minute,price_ohlc.close,'gray',bstime,buy,'ko', marker=r'$\downarrow$', markersize=20,bstime,sell,'ro',bstime,short,'r*',bstime,cover,'g*') ax2.plot(bstime,buy-1,'go', marker=r'$\Uparrow$', markersize=8) ax2.plot(bstime,sell+1,'ro', marker=r'$\Downarrow$', markersize=8) ax2.plot(bstime,short+1,'ro', marker=r'$\blacktriangledown$', markersize=8) ax2.plot(bstime,cover-1,'go', marker=r'$\blacktriangle$', markersize=8) ax2.plot(bstime,tech_one,'blue',bstime,tech_two,'orange') plt.fill_between(bstime,tech_three,facecolor='seagreen', alpha=0.5, interpolate=True) plt.grid(b=True, which='major', color='grey', linestyle='--') plt.show() #---------------------------------------- NORMAL LINE PLOTS ---------------------------------------------------- ''' plt.figure(1) gs = gridspec.GridSpec(2, 1, height_ratios=[3, 1]) ax1=plt.subplot(gs[0]) axes = plt.gca() axes.set_xlim([0,numb]) axes.set_ylim([xoi,yoi]) #figure,ax1 = plt.subplots() ax2 = ax1.twinx() if file_format==1 or file_format==3:ax1.plot(minute,price_close,'black',bstime,buy,'go',bstime,sell,'ro',bstime,short,'bo',bstime,cover,'ro',bstime,tech_one,'seagreen',bstime,tech_two,'lightcoral')#,bstime,tech_three,'bo') if file_format==2:ax1.plot(minute,price_ohlc.close,'black',bstime,buy,'go',bstime,sell,'ro',bstime,short,'bo',bstime,cover,'ro',bstime,tech_one,'seagreen',bstime,tech_two,'lightcoral')#,bstime,tech_three,'bo') #ax2.plot(bstime,tech_three,'b-') plt.fill_between(bstime,tech_three,facecolor='seagreen', alpha=0.5, interpolate=True) plt.grid(b=True, which='major', color='grey', linestyle='-') plt.subplot(gs[1]) axes = plt.gca() axes.set_xlim([0,numb]) plt.plot(bstime,AS,'r-') plt.grid(b=True, which='major', color='grey', linestyle='-') plt.show() '''
[ "bharatkumar7@gmail.com" ]
bharatkumar7@gmail.com
c463bc62085e0e254656df795762807e09e2e229
949d9ed95d94c2cbce94e76120009c9d6b370fb1
/app/core/routers/__init__.py
eb3c2c8a4e245e044d5d6ff836de1452a79d69ab
[]
no_license
dexer13/guane-intern-fastapi
d8bc5fb808570c5c584f8fb0f685ed9f47b6a497
40f4ee5facf523ec93d974ba6613a959ebadae7c
refs/heads/main
2023-08-24T18:45:28.938807
2021-10-30T15:58:22
2021-10-30T15:58:22
420,717,715
0
0
null
null
null
null
UTF-8
Python
false
false
85
py
from . import animals from . import users from . import security from . import files
[ "disidro@campus.udes.edu.co" ]
disidro@campus.udes.edu.co
f3bf9cff8bd771da7387a3f9128836a954960113
661437f8881d9eb5b1d1fd0e28591c27e074326e
/Python/interesting/NOT_SUPPORT/parallel/Process_condition.py
97a9623051b5c7c0966f9736904a7dccaf913910
[]
no_license
gnosiop/fromPhone
c16c03bac3a3920a4daaa1f8b41dacfede257de0
ac427cf125573319c2adcc437b6286ea1da372f8
refs/heads/master
2023-08-14T23:46:14.919026
2021-09-25T08:11:38
2021-09-25T08:11:38
404,894,533
0
0
null
null
null
null
UTF-8
Python
false
false
575
py
import multiprocessing cv = multiprocessing.Condition() def produce(): with cv: for i in range(1, 6): print(i) cv.notify() # как только один поток выполнил операцию # он уведомил другой и "разбудил" его def consume(): with cv: cv.wait(timeout=2) # время на "пробуждение" print("второй поток") t2 = multiprocessing.Process(target=consume) t1 = multiprocessing.Process(target=produce) t1.start() t2.start()
[ "idoodi@ya.ru" ]
idoodi@ya.ru
4c7a6ca0278e20ba7b9ba747006fe1ff9e5d0326
e6c58d75f3cea45639b6dd0f8fe1d1ec6a00bae5
/weather/views.py
1a56ce2cef21e16d096be18a5f9308784b69156d
[ "MIT" ]
permissive
cindyjialiu/WeatherApp
4d2d0ae092b410ad6f35008a00ce938426d81a6c
c91f5928708d4cd79286bcd51a3934cb9d2e3a92
refs/heads/master
2020-04-11T19:00:55.616764
2018-12-21T15:54:44
2018-12-21T15:54:44
162,019,211
0
0
null
null
null
null
UTF-8
Python
false
false
1,968
py
import os import datetime import requests from django.shortcuts import render def index(request): api_key=os.environ['WEATHER_API_KEY'] today = datetime.datetime.today() response = requests.get(f'http://api.openweathermap.org/data/2.5/weather?q=London,uk&units=metric&appid={api_key}') # TODO: handle extra errors weather_data = response.json() response2 = requests.get(f'http://api.openweathermap.org/data/2.5/forecast?q=London,uk&units=metric&appid={api_key}') # TODO: handle extra errors weather_forecast_data = response2.json() weather_summary = get_weather_summary(weather_data) weather_forecast_summary = get_temps_for_tomorrow(get_weather_forecast_temp_and_dt(weather_forecast_data['list']), today) weather_forecast_tomorrow = get_temps_for_tomorrow_without_date(weather_forecast_summary) return render(request, 'index.html', { 'weather_forecast_tomorrow': weather_forecast_tomorrow, 'weather_summary': weather_summary }) def get_weather_summary(weather_data): return { 'temp': weather_data['main']['temp'], 'min': weather_data['main']['temp_min'], 'max': weather_data['main']['temp_max'], 'humidity':weather_data['main']['humidity'] } def get_weather_forecast_temp_and_dt(weather_forecast_data): return list(map(lambda x: { 'y': x['main']['temp'], 'x': x['dt_txt'] }, weather_forecast_data)) def get_temps_for_tomorrow(filtered_forecast_data, today): tomorrow = str(today + datetime.timedelta(days = 1)).split(' ')[0] return list(filter(lambda x: x['x'].split(' ')[0] == tomorrow, filtered_forecast_data )) def get_temps_for_tomorrow_without_date(tomorrow_temps_data): return list(map(lambda x: { 'x': dt_txt_formatter(x['x']), 'y': x['y']}, tomorrow_temps_data)) def dt_txt_formatter(dateTime): return dateTime.split(' ')[1][:-3]
[ "jl7e12@gmail.com" ]
jl7e12@gmail.com
cbefe3934df661d760f6292ce580d43fddcb9dac
00d809abff2460c051cf3aeaf0f98005bd5f0397
/API/api_open_subtitles.py
748d2f77e2667ea0a234b8b7701134410a2497c0
[ "MIT" ]
permissive
andreibastos/movie-info
e57d821becb8aea820e7858459d71ac34a537a1c
57da167b4d0ed602b5a9dce8c3f2544f7dfdacd9
refs/heads/master
2016-09-01T18:36:47.182014
2015-08-15T01:09:20
2015-08-15T01:09:20
40,301,196
0
0
null
null
null
null
UTF-8
Python
false
false
4,679
py
# coding: utf-8 import urllib from lxml import html import lxml.html as html from lxml import etree from lxml.html import fromstring, tostring import json url_base = 'http://www.opensubtitles.org' url_search = 'http://www.opensubtitles.org/pb/search/sublanguageid-pob/imdbid-' def get_legend(imdbID): legend = {} imdbID = imdbID.strip('tt') print (imdbID) link = url_search + imdbID print (link) # link = '/home/andrei/scripts/torrent/imdb/legend.html' # link = 'http://www.opensubtitles.org/pb/search/sublanguageid-all/subtrusted-on/hearingimpaired-on/hd-on/autotranslation-on' page = None tree = None response = False try: page = urllib.urlopen(link) tree = etree.HTML(page.read()) except Exception as inst: return json.dumps({'error':{'code':1,'mensage':str(inst)},'response':response}) try: for div in tree.xpath("//div[@style='text-align:center']"): for a in div.xpath('h1/a'): legend['link_download'] = a.get('href') legend['name_movie'] = a.xpath('span')[0].text legend['downloads'] = 0 for span in div.xpath("//div[@itemtype='http://schema.org/Movie']/h2"): legend['name_movie_file'] = span.text legends = [] legends.append(legend) response = True return json.dumps({'response':response,'legends':legends}) except Exception as msg: legends = [] try: for tr in tree.xpath("//table[@id='search_results']//tr"): # for columns of tr index_column = 0 # to legend name_movie = None quality_legend = [] name_movie_file = None language = None cd = None date = None fps = 0.0 downloads = None link_download = None legend_type = None punctuation = None comments = 0 imdbVote = 0 autor_name = None autor_rank = None len_columns = len(tr) legend = None tmp_class = tr.get('class') if len_columns == 9 and not 'head' in tmp_class: legend = {} for td in tr.xpath('td'): if index_column == 0: for a in td.xpath('strong/a'): name_movie = ((a.text.replace('\n','').replace('\t',''))) for img in td.xpath('img'): src = img.get('src') if src: if '/' in src: tmp_src = src.split('/') src = tmp_src[len(tmp_src)-1] if '.' in src: src = src.split('.')[0] quality_legend.append(src) for br in td.xpath("//br"): if br.tail: name_movie_file = br.tail for span in td.xpath('span'): name_movie_file = span.get('title') if index_column == 1: a = td.xpath('a') if a: language = a[0].get('title') if index_column == 2: cd = td.text.replace('\n','').replace('\t','') if index_column == 3: for time in td.xpath('time'): date = time.text for span in td.xpath('span'): fps = float(span.text) if index_column == 4: for a in td.xpath('a'): link_download = url_base + a.get('href') downloads = int (a.text.replace('x','').replace('\n','')) for span in td.xpath('span'): legend_type = span.text if index_column == 5: punctuation = td.text for img in td.xpath('img'): punctuation = (img.get('src').split('/')[len(img)-1]) if index_column == 6: comments = td.text if index_column == 7: imdbVote = td.xpath('a')[0].text if index_column == 8: if len(td.xpath('a'))>0: autor_name = td.xpath('a')[0].text if len(td.xpath('a/img'))>0: autor_rank = td.xpath('a/img')[0].get('alt') index_column +=1; legend['name_movie'] = name_movie legend['name_movie_file'] = name_movie_file legend['quality_legend'] = quality_legend legend['language'] = language legend['cd'] = cd legend['date'] = date legend['fps'] = fps legend['downloads'] = downloads legend['link_download'] = link_download legend['punctuation'] = punctuation legend['comments'] = comments legend['imdbVote'] = imdbVote legend['autor_name'] = autor_name legend['autor_rank'] = autor_rank legends.append(legend) response = True return json.dumps({'response':response,'legends':legends}, indent=4, sort_keys=True) except Exception as inst: print 'error 2 ' return json.dumps({'error':{'code':2,'mensage':str(inst)},'response':response}) # imdbID = "tt3235888" # f = open(imdbID +".json",'w') # legendas = get_legend(imdbID) # print (legendas) # f.write(legendas) # f.close
[ "andreibastos@outlook.com" ]
andreibastos@outlook.com
80ade65f1adcd28e82e85178c589f646d0eb1e6b
d5f518c8f23705396fd8da3317520cff0ab543f7
/algorithms/team 3/spikefinder/__init__.py
4413e5d15c8306e460e19e9e225a26b4cc57d8eb
[]
no_license
j-friedrich/spikefinder_analysis
78322f8ead7579b1b8bfb50769d6467cee66930e
def1f8c2c5268eb71e83f57c265d4b3c102fb5f8
refs/heads/master
2021-10-27T17:14:49.134302
2018-02-19T11:42:18
2018-02-19T11:42:18
93,436,281
1
0
null
2017-06-05T18:51:23
2017-06-05T18:51:23
null
UTF-8
Python
false
false
50
py
from .main import load, score __version__='1.0.0'
[ "noreply@github.com" ]
j-friedrich.noreply@github.com
dd55eae4011f0cb80d47c940385e7a3ff85cd7a3
602fa0e4ce194d3073d78230c61f7053281f9f9b
/code/python/src/categories/catutil.py
df03a0027b66f8d76d4265de7c7074d56b487bab
[]
no_license
ziqizhang/wop
111cfdda1686a874ff1fc11a453a23fb52d43af1
ea0c37f444de9f2d5303f74b989f6d1a09feb61d
refs/heads/master
2022-09-14T20:14:11.575021
2021-12-10T21:23:24
2021-12-10T21:23:24
166,239,995
2
1
null
2022-09-01T23:11:13
2019-01-17T14:33:51
Python
UTF-8
Python
false
false
2,128
py
import pandas as pd from nltk import PorterStemmer, WordNetLemmatizer import numpy from categories import cleanCategories as cc stemmer = PorterStemmer() lemmatizer = WordNetLemmatizer() #0=stem; 1=lem; else=nothing def normalise_categories(in_file_name, col, stem_or_lem): df = pd.read_csv(in_file_name, header=0, delimiter=";", quoting=0, encoding="utf-8", ).as_matrix() norm_cats=set() max_toks=0 for r in df: c = r[col] if type(c) is not str and numpy.isnan(c): c="NONE" toks = len(c.split(" ")) if toks>max_toks: max_toks=toks if stem_or_lem==0: c=stemmer.stem(c).strip() if len(c)>2: norm_cats.add(c) elif stem_or_lem==1: c=lemmatizer.lemmatize(c).strip() if len(c)>2: norm_cats.add(c) else: norm_cats.add(c) norm_cats_list=list(norm_cats) norm_cats_list=sorted(norm_cats_list) print(len(norm_cats_list)) print(max_toks) for nc in norm_cats_list: print(nc) def get_parent_category_level(in_file_name, col): df = pd.read_csv(in_file_name, header=0, delimiter=";", quoting=0, encoding="utf-8", ).as_matrix() norm_cats = set() norm_cats_list=[] for r in df: c = r[col] if type(c) is not str and numpy.isnan(c): continue c= cc.normaliseCategories(c) try: trim = c.index(">") except ValueError: continue c=c[0:trim].strip() norm_cats.add(c) norm_cats_list.append(c) norm_cats_unique_list=sorted(list(norm_cats)) norm_cats=sorted(norm_cats) for nc in norm_cats: print(nc) print("\n\n>>>>>>>>>\n\n") for nc in norm_cats_unique_list: print(nc) if __name__ == "__main__": # normalise_categories("/home/zz/Work/data/wop_data/goldstandard_eng_v1_cleanedCategories.csv", # 13,0) get_parent_category_level("/home/zz/Work/data/wop_data/goldstandard_eng_v1_utf8.csv", 8)
[ "ziqizhang.email@gmail.com" ]
ziqizhang.email@gmail.com
aa122ff357ac2276d0569e9bae610a18e81b9c11
45be54f14406418be8bf9c1c9a695e77f2c79d1e
/workflow/rules/quality_control.smk
bcb80c300cab633d395dd1523e96c80286362b8c
[]
no_license
G-Molano-LA/circrna-workflow
41d48c097a7909c8e843beab5e59685950daddf0
2be5e61e0eea5395758819c41f39ff9cac279fa0
refs/heads/main
2023-06-11T00:49:36.596498
2021-07-02T12:32:38
2021-07-02T12:32:38
348,354,579
0
0
null
null
null
null
UTF-8
Python
false
false
2,490
smk
#!/usr/bin/python3 __author__ = "G. Molano, LA (gonmola@hotmail.es)" __state__ = "ALMOST FINISHED" # requires execution to finish it ################################################################################ # Snakefile to realize a quality control of RNA-seq reads. #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # Author: G. Molano, LA (gonmola@hotmail.es) #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # Date : # Last modification : 01-06-2021 ################################################################################ RAW_READS = expand("{path}/{sample}{ext}", path = config["quality_control"]["reads"], sample = SAMPLES, ext = [config["quality_control"]["suffix"][1],config["quality_control"]["suffix"][2]] ) # TARGET RULE rule quality_control_results: input: html = f'{OUTDIR}/quality_control/raw_data/summary/multiqc_report.html' #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~FASTQC~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ rule fastqc1: input: RAW_READS output: html = expand("{outdir}/quality_control/raw_data/{sample}_{replicate}_fastqc.html", outdir = OUTDIR, sample = SAMPLES, replicate = [1,2]), zip = expand("{outdir}/quality_control/raw_data/{sample}_{replicate}_fastqc.zip", outdir = OUTDIR, sample = SAMPLES, replicate = [1,2]) params: outdir = f'{OUTDIR}/quality_control/raw_data/' threads: config["trimming"]["threads"] conda: config["envs"]["quality_control"] # message: # "Starting quality analysis control with FASTQC programm on the " # "following files {input.reads}. Number of threads used are {threads}." priority: 1 shell: "fastqc -t {threads} {input} --outdir={params.outdir}" #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~MULTIQC~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ rule multiqc1: input: zip = expand("{outdir}/quality_control/raw_data/{sample}_{replicate}_fastqc.zip", outdir = OUTDIR, sample = SAMPLES, replicate = [1,2]) output: html = f'{OUTDIR}/quality_control/raw_data/summary/multiqc_report.html', params: replace_old = "--force", # revisar que no remplaze al anterior outdir = f'{OUTDIR}/quality_control/raw_data/summary/' conda: config["envs"]["quality_control"] priority: 2 shell: "multiqc --interactive {params.replace_old} {input.zip} --outdir {params.outdir}"
[ "gonmola@hotmail.es" ]
gonmola@hotmail.es
6753822442fee034044704f8fce55be9c1448475
4128e5c41fabbe2289ea7d7faae3d970d0244514
/jeffy/sdk/kinesis.py
c8e3b18282e2c439b629b5950a6a35cb64a9d3c2
[ "MIT", "LicenseRef-scancode-unknown-license-reference" ]
permissive
sinofseven/jeffy-my-extended
d772e6a95e9551de637d5ec70dca1d65f27e60d5
036ef14e8be5c93f19af4fd0012cc482a77717bb
refs/heads/master
2021-01-15T06:07:33.383973
2020-02-25T03:37:42
2020-02-25T03:37:42
242,897,880
1
0
null
null
null
null
UTF-8
Python
false
false
1,065
py
import boto3 import json from typing import Any class Kinesis(): """ Kinesis Client. """ _resource = None @classmethod def get_resource(cls) -> boto3.client: """ Get boto3 client for Kinesis. Usage:: >>> from jeffy.sdk.kinesis import Kinesis >>> Kinesis.get_resource().put_record(...) """ if Kinesis._resource is None: Kinesis._resource = boto3.client('kinesis') return Kinesis._resource @classmethod def put_record(cls, stream_name: str, data: Any, partition_key: str, correlation_id: str = ''): """ Put recourd to Kinesis Stream with correlation_id. Usage:: >>> from jeffy.sdk.kinesis import Kinesis >>> Sqs.put_record(...) """ return cls.get_resource().put_record( StreamName=stream_name, Data=json.dumps({ 'correlation_id': correlation_id, 'item': data }), PartitionKey=partition_key, )
[ "info@serverless-operations.com" ]
info@serverless-operations.com
c38344d14b0ab7fc67752b5c676ed3fb625393cf
414c0bd290c264a55dc03d0657c5f812914bf050
/ApnaBazaar/urls.py
7ae85e3480a521c384db656e40d4b499b0ea12c2
[]
no_license
sakshamsin09/ApnaBazaar
b8af9aea2538d69f27ef80caeb05df3d10d08993
02c2a69e6950d8e81ec0bc51e7247f96c2f3f77a
refs/heads/main
2023-07-09T14:20:40.971602
2021-08-24T10:57:16
2021-08-24T10:57:16
383,352,840
0
0
null
2021-08-22T13:30:28
2021-07-06T05:42:14
HTML
UTF-8
Python
false
false
824
py
"""ApnaBazaar URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/3.2/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.urls import include, path 2. Add a URL to urlpatterns: path('blog/', include('blog.urls')) """ from django.contrib import admin from django.urls import path from django.urls.conf import include urlpatterns = [ path('admin/', admin.site.urls), path('', include('app.urls')), ]
[ "sakshamsinghal09@gmail.com" ]
sakshamsinghal09@gmail.com
edcbbc430b0d1a558d19be8a4a2625b7c762eb20
5add80be09ee754fced03e512a9acc214971cddf
/python-code/openvx-learning/helloworld.py
61352b55542a81f5e56cc66c6767ea1beb6c1d65
[ "Apache-2.0" ]
permissive
juxiangwu/image-processing
f774a9164de9c57e88742e6185ac3b28320eae69
c644ef3386973b2b983c6b6b08f15dc8d52cd39f
refs/heads/master
2021-06-24T15:13:08.900960
2019-04-03T10:28:44
2019-04-03T10:28:44
134,564,878
15
5
null
null
null
null
UTF-8
Python
false
false
935
py
from pyvx import vx context = vx.CreateContext() images = [ vx.CreateImage(context, 640, 480, vx.DF_IMAGE_UYVY), vx.CreateImage(context, 640, 480, vx.DF_IMAGE_S16), vx.CreateImage(context, 640, 480, vx.DF_IMAGE_U8), ] graph = vx.CreateGraph(context) virts = [ vx.CreateVirtualImage(graph, 0, 0, vx.DF_IMAGE_VIRT), vx.CreateVirtualImage(graph, 0, 0, vx.DF_IMAGE_VIRT), vx.CreateVirtualImage(graph, 0, 0, vx.DF_IMAGE_VIRT), vx.CreateVirtualImage(graph, 0, 0, vx.DF_IMAGE_VIRT), ] vx.ChannelExtractNode(graph, images[0], vx.CHANNEL_Y, virts[0]) vx.Gaussian3x3Node(graph, virts[0], virts[1]) vx.Sobel3x3Node(graph, virts[1], virts[2], virts[3]) vx.MagnitudeNode(graph, virts[2], virts[3], images[1]) vx.PhaseNode(graph, virts[2], virts[3], images[2]) status = vx.VerifyGraph(graph) if status == vx.SUCCESS: status = vx.ProcessGraph(graph) else: print("Verification failed.") vx.ReleaseContext(context)
[ "kkoolerter@gmail.com" ]
kkoolerter@gmail.com
6b21ff7967cabb367eae29730fcc4d5cd9aee141
78adcbb441d703c64553c09f6e08ae01a6d95ad0
/main.py
c59c9f7a49d137b8ca303ed85c45e90d9e858529
[]
no_license
victor369basu/MongoDBFlask
8dfa6fb576a23e6d47f7211fbb2949bbdbfcddbf
c7de5027e655b8768d11f0c4184aeb5bfa9ccd17
refs/heads/master
2023-03-04T04:18:31.649237
2023-02-21T12:08:39
2023-02-21T12:08:39
303,291,017
1
0
null
null
null
null
UTF-8
Python
false
false
2,647
py
from typing import Optional import uvicorn from fastapi import FastAPI, Request from MongoAPI import MongoAPI import json app = FastAPI() @app.get("/") async def base(): return {'response':{"Status": "Health Check!"}, 'status':200, 'mimetype':'application/json' } @app.get('/mongodb') async def mongo_read(info : Request): ''' Reading the data when a request is sent using the GET HTTP Method ''' data = await info.json() if data is None or data == {}: return {'response': {"Error": "Please provide connection information"}, 'status':400, 'mimetype':'application/json'} obj = MongoAPI(data) response = obj.read() return {'response':response, 'status':200, 'mimetype':'application/json'} @app.post('/mongodb') async def mongo_write(info : Request): ''' Writing the data when a request is sent using the POST HTTP Method. ''' data = await info.json() print(data['Document']) if data is None or data == {} or 'Document' not in data: return {'response': {"Error": "Please provide connection information"}, 'status':400, 'mimetype':'application/json'} obj = MongoAPI(data) response = obj.write(data) return {'response':response, 'status':200, 'mimetype':'application/json'} @app.put('/mongodb') async def mongo_update(info : Request): ''' Updating the data when a request is sent using the PUT HTTP Method. ''' data = await info.json() if data is None or data == {} or 'DataToBeUpdated' not in data: return {'response': {"Error": "Please provide connection information"}, 'status':400, 'mimetype':'application/json'} obj = MongoAPI(data) response = obj.update() return {'response':response, 'status':200, 'mimetype':'application/json'} @app.delete('/mongodb') async def mongo_delete(info : Request): ''' Deleting the data when a request is sent using the DELETE HTTP Method. ''' data = await info.json() if data is None or data == {} or 'Filter' not in data: return {'response': {"Error": "Please provide connection information"}, 'status':400, 'mimetype':'application/json'} obj = MongoAPI(data) response = obj.delete(data) return {'response':response, 'status':200, 'mimetype':'application/json'} # if __name__ == '__main__': # uvicorn.run("main:app", host="0.0.0.0", port=8000, reload=True, access_log=False)
[ "victor.basu@lumiq.ai" ]
victor.basu@lumiq.ai
d92df5cd630581d42b06e50bdc1070c5d414a17c
9647524c0f4d93fb1c8a992c20fe9f9d2710cde3
/2-content/Python/intro_programming-master/scripts/remove_input_references.py
2ab8878b1a362f079adf49a971ef71aa7677a4ea
[ "MIT" ]
permissive
bgoonz/web-dev-notes-resource-site
16161aa68e8eecafeaba4dc7abeb957aaee864c5
e7dc9c30393597cb39830c49c3f51c1486b97584
refs/heads/master
2023-09-01T14:04:20.867818
2021-06-17T07:56:20
2021-06-17T07:56:20
329,194,347
7
5
MIT
2021-07-05T06:36:49
2021-01-13T04:34:20
JavaScript
UTF-8
Python
false
false
1,306
py
# This script removes the input reference numbers from html pages. # They play a useful role in scientific notebooks, but they are really # just visual clutter in this project. # Could be an nbconvert setting, but it's an easy enough scripting job. import os import sys print("\nStripping input reference numbers from code cells...") # Find all files to work with. path_to_notebooks = '/srv/projects/intro_programming/intro_programming/notebooks/' filenames = [] for filename in os.listdir(path_to_notebooks): if '.html' in filename and filename != 'index.html': filenames.append(filename) # one file for testing: #filenames = ['hello_world.html'] for filename in filenames: f = open(path_to_notebooks + filename, 'r') lines = f.readlines() f.close() f = open(path_to_notebooks + filename, 'wb') for line in lines: # Unwanted lines have opening and closing div on same line, # with input reference number between them. if ('<div class="prompt input_prompt">' in line and '</div>' in line): # Don't write this line. continue else: # Regular line, write it. f.write(line.encode('utf-8')) f.close() print(" Stripped input reference numbers.\n")
[ "bryan.guner@gmail.com" ]
bryan.guner@gmail.com
eab0841d48237a5fda2b01c56721707bb39a40fb
156cac8bf5192a3d0b93a105539f4e6d5108fa1c
/ciyunapi/ciyunapi/asgi.py
b9e43c8bfd8ab576339ef2ba6040afa333af0c34
[]
no_license
yutu-75/ciyun
220de2c4e4c6365145c16d3bf634838ecb9921ca
c25b2555ac527097d4d27380ae75a47670f42687
refs/heads/main
2023-05-07T08:46:18.285524
2021-05-31T13:22:49
2021-05-31T13:22:49
365,187,271
0
0
null
null
null
null
UTF-8
Python
false
false
413
py
""" ASGI config for ciyunapi project. It exposes the ASGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.2/howto/deployment/asgi/ """ import os from django.core.asgi import get_asgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'ciyunapi.settings.dev') application = get_asgi_application()
[ "xiao3952@foxmail.com" ]
xiao3952@foxmail.com
93013a6c44645ef61cb45e633030c20663c3fde6
8ef8e6818c977c26d937d09b46be0d748022ea09
/cv/classification/torchvision/pytorch/train.py
1c16c81bc51ace035a2653350c088a3888b0904f
[ "Apache-2.0" ]
permissive
Deep-Spark/DeepSparkHub
eb5996607e63ccd2c706789f64b3cc0070e7f8ef
9d643e88946fc4a24f2d4d073c08b05ea693f4c5
refs/heads/master
2023-09-01T11:26:49.648759
2023-08-25T01:50:18
2023-08-25T01:50:18
534,133,249
7
6
Apache-2.0
2023-03-28T02:54:59
2022-09-08T09:07:01
Python
UTF-8
Python
false
false
15,577
py
# Copyright (c) 2022 Iluvatar CoreX. All rights reserved. # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import warnings warnings.filterwarnings('ignore') import datetime import os import logging import time import torch import torch.utils.data try: from apex import amp as apex_amp except: apex_amp = None try: from torch.cuda.amp import autocast, GradScaler scaler = GradScaler() except: autocast = None scaler = None from torch import nn import torch.distributed as dist import torchvision import utils from utils import (MetricLogger, SmoothedValue, accuracy, mkdir,\ init_distributed_mode, manual_seed,\ is_main_process, save_on_master, write_on_master) from dataloader.classification import get_datasets, create_dataloader def compute_loss(model, image, target, criterion): output = model(image) if not isinstance(output, (tuple, list)): output = [output] losses = [] for out in output: losses.append(criterion(out, target)) loss = sum(losses) return loss, output[0] def train_one_epoch(model, criterion, optimizer, data_loader, device, epoch, print_freq, use_amp=False, use_dali=False): model.train() metric_logger = MetricLogger(delimiter=" ") metric_logger.add_meter('lr', SmoothedValue(window_size=1, fmt='{value}')) metric_logger.add_meter('img/s', SmoothedValue(window_size=10, fmt='{value}')) header = 'Epoch: [{}]'.format(epoch) all_fps = [] for data in metric_logger.log_every(data_loader, print_freq, header): if use_dali: image, target = data[0]["data"], data[0]["label"][:, 0].long() else: image, target = data start_time = time.time() image, target = image.to(device, non_blocking=True), target.to(device, non_blocking=True) loss, output = compute_loss(model, image, target, criterion) if use_amp: with apex_amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() else: loss.backward() optimizer.step() optimizer.zero_grad() end_time = time.time() acc1, acc5 = accuracy(output, target, topk=(1, 5)) batch_size = image.shape[0] metric_logger.update(loss=loss.item(), lr=optimizer.param_groups[0]["lr"]) metric_logger.meters['acc1'].update(acc1.item(), n=batch_size) metric_logger.meters['acc5'].update(acc5.item(), n=batch_size) fps = batch_size / (end_time - start_time) * utils.get_world_size() metric_logger.meters['img/s'].update(fps) all_fps.append(fps) fps = round(sum(all_fps) / len(all_fps), 2) print(header, 'Avg img/s:', fps) return fps def evaluate(model, criterion, data_loader, device, print_freq=100, use_dali=False): model.eval() metric_logger = MetricLogger(delimiter=" ") header = 'Test:' with torch.no_grad(): for data in metric_logger.log_every(data_loader, print_freq, header): if use_dali: image, target = data[0]["data"], data[0]["label"][:, 0].long() else: image, target = data image = image.to(device, non_blocking=True) target = target.to(device, non_blocking=True) output = model(image) loss = criterion(output, target) acc1, acc5 = accuracy(output, target, topk=(1, 5)) # FIXME need to take into account that the datasets # could have been padded in distributed setup batch_size = image.shape[0] metric_logger.update(loss=loss.item()) metric_logger.meters['acc1'].update(acc1.item(), n=batch_size) metric_logger.meters['acc5'].update(acc5.item(), n=batch_size) # gather the stats from all processes metric_logger.synchronize_between_processes() print(' * Acc@1 {top1.global_avg:.3f} Acc@5 {top5.global_avg:.3f}' .format(top1=metric_logger.acc1, top5=metric_logger.acc5)) return round(metric_logger.acc1.global_avg, 2) def _get_cache_path(filepath): import hashlib h = hashlib.sha1(filepath.encode()).hexdigest() cache_path = os.path.join("~", ".torch", "vision", "datasets", "imagefolder", h[:10] + ".pt") cache_path = os.path.expanduser(cache_path) return cache_path def create_optimzier(params, args): opt_name = args.opt.lower() if opt_name == 'sgd': optimizer = torch.optim.SGD(params, lr=args.lr, momentum=args.momentum, weight_decay=args.weight_decay) elif opt_name == 'rmsprop': optimizer = torch.optim.RMSprop(params, lr=args.lr, momentum=args.momentum, weight_decay=args.weight_decay, eps=0.0316, alpha=0.9) elif opt_name == "fused_sgd": from apex.optimizers import FusedSGD optimizer = FusedSGD(params, lr=args.lr, momentum=args.momentum, weight_decay=args.weight_decay) else: raise RuntimeError("Invalid optimizer {}. Only SGD and RMSprop are supported.".format(args.opt)) return optimizer def main(args): init_distributed_mode(args) print(args) device = torch.device(args.device) manual_seed(args.seed, deterministic=args.deterministic) # WARN: if dist.is_initialized(): num_gpu = dist.get_world_size() else: num_gpu = 1 global_batch_size = num_gpu * args.batch_size train_dir = os.path.join(args.data_path, 'train') val_dir = os.path.join(args.data_path, 'val') num_classes = len(os.listdir(train_dir)) if 0 < num_classes < 13: if global_batch_size > 512: if is_main_process(): print("WARN: Updating global batch size to 512, avoid non-convergence when training small dataset.") args.batch_size = 512 // num_gpu if args.pretrained: num_classes = 1000 args.num_classes = num_classes print("Creating model") if hasattr(args, "model_cls"): model = args.model_cls(args) else: model = torchvision.models.__dict__[args.model](pretrained=args.pretrained, num_classes=num_classes) if args.padding_channel: print("WARN: Cannot convert first conv to N4HW.") data_loader, data_loader_test = create_dataloader(train_dir, val_dir, args) if args.padding_channel and isinstance(data_loader, torch.utils.data.DataLoader): data_loader.collate_fn = utils.nhwc_collect_fn(data_loader.collate_fn, fp16=args.amp, padding=args.padding_channel) data_loader_test.collate_fn = utils.nhwc_collect_fn(data_loader_test.collate_fn, fp16=args.amp, padding=args.padding_channel) model.to(device) if args.distributed and args.sync_bn: model = torch.nn.SyncBatchNorm.convert_sync_batchnorm(model) criterion = nn.CrossEntropyLoss() if args.nhwc: model = model.cuda().to(memory_format=torch.channels_last) optimizer = create_optimzier(model.parameters(), args) if args.amp: model, optimizer = apex_amp.initialize(model, optimizer, opt_level="O2", loss_scale="dynamic", master_weights=True) lr_scheduler = torch.optim.lr_scheduler.StepLR(optimizer, step_size=args.lr_step_size, gamma=args.lr_gamma) model_without_ddp = model if args.distributed: model = torch.nn.parallel.DistributedDataParallel(model, device_ids=[args.gpu]) model_without_ddp = model.module if args.resume: checkpoint = torch.load(args.resume, map_location='cpu') model_without_ddp.load_state_dict(checkpoint['model']) optimizer.load_state_dict(checkpoint['optimizer']) lr_scheduler.load_state_dict(checkpoint['lr_scheduler']) args.start_epoch = checkpoint['epoch'] + 1 if args.test_only: evaluate(model, criterion, data_loader_test, device=device) return print("Start training") start_time = time.time() best_acc1 = 0 best_epoch = 0 for epoch in range(args.start_epoch, args.epochs): epoch_start_time = time.time() if args.distributed and not args.dali: data_loader.sampler.set_epoch(epoch) fps = train_one_epoch(model, criterion, optimizer, data_loader, device, epoch, args.print_freq, args.amp, use_dali=args.dali) lr_scheduler.step() acc1 = evaluate(model, criterion, data_loader_test, device=device, use_dali=args.dali) if acc1 > best_acc1: best_acc1 = acc1 best_epoch = epoch if args.output_dir is not None: checkpoint = { 'model': model_without_ddp.state_dict(), 'optimizer': optimizer.state_dict(), 'lr_scheduler': lr_scheduler.state_dict(), 'epoch': epoch, 'args': args} save_on_master( checkpoint, os.path.join(args.output_dir, 'best.pth'.format(epoch))) save_on_master( checkpoint, os.path.join(args.output_dir, 'latest.pth')) epoch_total_time = time.time() - epoch_start_time epoch_total_time_str = str(datetime.timedelta(seconds=int(epoch_total_time))) print('epoch time {}'.format(epoch_total_time_str)) if args.dali: data_loader.reset() data_loader_test.reset() total_time = time.time() - start_time total_time_str = str(datetime.timedelta(seconds=int(total_time))) print('* Acc@1: {} at epoch {}'.format(round(best_acc1, 2), best_epoch)) print('Training time {}'.format(total_time_str)) if args.output_dir: write_on_master({"Name":os.path.basename(args.output_dir), "Model": args.model, "Dataset": os.path.basename(args.data_path), "AMP":args.amp, "Acc@1":best_acc1, "FPS":fps, "Time": total_time_str}, os.path.join(args.output_dir, 'result.json')) def get_args_parser(add_help=True): import argparse parser = argparse.ArgumentParser(description='PyTorch Classification Training', add_help=add_help) parser.add_argument('--data-path', default='/datasets01/imagenet_full_size/061417/', help='dataset') parser.add_argument('--model', default='resnet18', help='model') parser.add_argument('--device', default='cuda', help='device') parser.add_argument('-b', '--batch-size', default=32, type=int) parser.add_argument('--epochs', default=90, type=int, metavar='N', help='number of total epochs to run') parser.add_argument('-j', '--workers', default=4, type=int, metavar='N', help='number of data loading workers (default: 4)') parser.add_argument('--opt', default='sgd', type=str, help='optimizer') parser.add_argument('--lr', default=0.128, type=float, help='initial learning rate') parser.add_argument('--momentum', default=0.9, type=float, metavar='M', help='momentum') parser.add_argument('--wd', '--weight-decay', default=1e-4, type=float, metavar='W', help='weight decay (default: 1e-4)', dest='weight_decay') parser.add_argument('--lr-step-size', default=30, type=int, help='decrease lr every step-size epochs') parser.add_argument('--lr-gamma', default=0.1, type=float, help='decrease lr by a factor of lr-gamma') parser.add_argument('--print-freq', default=10, type=int, help='print frequency') parser.add_argument('--output-dir', default=None, help='path where to save') parser.add_argument('--resume', default='', help='resume from checkpoint') parser.add_argument('--start-epoch', default=0, type=int, metavar='N', help='start epoch') parser.add_argument( "--cache-dataset", dest="cache_dataset", help="Cache the datasets for quicker initialization. It also serializes the transforms", action="store_true", ) parser.add_argument( "--sync-bn", dest="sync_bn", help="Use sync batch norm", action="store_true", ) parser.add_argument( "--deterministic", help="Do not benchmark conv algo", action="store_true", ) parser.add_argument( "--test-only", dest="test_only", help="Only test the model", action="store_true", ) parser.add_argument( "--pretrained", dest="pretrained", help="Use pre-trained models from the modelzoo", action="store_true", ) parser.add_argument('--auto-augment', default=None, help='auto augment policy (default: None)') parser.add_argument('--random-erase', default=0.0, type=float, help='random erasing probability (default: 0.0)') parser.add_argument( "--dali", help="Use dali as dataloader", default=False, action="store_true", ) # distributed training parameters parser.add_argument('--local_rank', default=-1, type=int, help='Local rank') parser.add_argument('--world-size', default=1, type=int, help='number of distributed processes') parser.add_argument('--dist-url', default='env://', help='url used to set up distributed training') # other parser.add_argument('--amp', action='store_true', help='Automatic Mixed Precision training') parser.add_argument('--nhwc', action='store_true', help='Use NHWC') parser.add_argument('--padding-channel', action='store_true', help='Padding the channels of image to 4') parser.add_argument('--dali-cpu', action='store_true') parser.add_argument('--seed', default=42, type=int, help='Random seed') parser.add_argument('--crop-size', default=224, type=int) parser.add_argument('--base-size', default=256, type=int) return parser def check_agrs(args): if args.nhwc: args.amp = True if args.output_dir: prefix=args.output_dir names = [args.model, os.path.basename(args.data_path)] if args.amp: names.append("amp") if torch.cuda.device_count(): names.append(f"dist_{utils.get_world_size()}x{torch.cuda.device_count()}") exp_dir = "_".join(map(str, names)) args.output_dir = os.path.join(prefix, exp_dir) if not os.path.exists(args.output_dir): os.makedirs(args.output_dir, exist_ok=True) if args.amp: if apex_amp is None: raise RuntimeError("Not found apex in installed packages, cannot enable amp.") def train_model(model_cls=None): args = get_args_parser().parse_args() check_agrs(args) if utils.is_main_process(): setup_logging(args.output_dir) if hasattr(torch, "corex") and args.dali: args.dali_cpu = True if model_cls is not None: args.model_cls = model_cls main(args) def setup_logging(prefix): if prefix: handlers=[ logging.FileHandler(os.path.join(prefix, "train.log"), mode='w'), logging.StreamHandler(), ] else: handlers = None logging.basicConfig( level=logging.DEBUG, format="%(asctime)s [%(levelname)s] %(message)s", handlers=handlers ) if __name__ == "__main__": args = get_args_parser().parse_args() check_agrs(args) if utils.is_main_process(): setup_logging(args.output_dir) try: main(args) except Exception as e: logging.exception(e)
[ "jia.guo@iluvatar.ai" ]
jia.guo@iluvatar.ai
d0a3f8fea955cd6b7239c30eb4bde72572683e27
f2f88a578165a764d2ebb4a022d19e2ea4cc9946
/pyvisdk/do/guest_authentication.py
f16ac39d82372db0665b605fca27476d5d281d82
[ "MIT" ]
permissive
pombredanne/pyvisdk
1ecc68a1bf264095f72f274c776e5868fb302673
de24eb4426eb76233dc2e57640d3274ffd304eb3
refs/heads/master
2021-01-21T16:18:39.233611
2014-07-28T19:50:38
2014-07-28T19:50:38
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,039
py
import logging from pyvisdk.exceptions import InvalidArgumentError ######################################## # Automatically generated, do not edit. ######################################## log = logging.getLogger(__name__) def GuestAuthentication(vim, *args, **kwargs): '''GuestAuthentication is an abstract base class for authentication in the guest.''' obj = vim.client.factory.create('ns0:GuestAuthentication') # do some validation checking... if (len(args) + len(kwargs)) < 1: raise IndexError('Expected at least 2 arguments got: %d' % len(args)) required = [ 'interactiveSession' ] optional = [ 'dynamicProperty', 'dynamicType' ] for name, arg in zip(required+optional, args): setattr(obj, name, arg) for name, value in kwargs.items(): if name in required + optional: setattr(obj, name, value) else: raise InvalidArgumentError("Invalid argument: %s. Expected one of %s" % (name, ", ".join(required + optional))) return obj
[ "guy@rzn.co.il" ]
guy@rzn.co.il
2033b58ac5c7b829f095eafefba8e88a252ab286
5cb6907c93b4d8d4efdc318ccb44bc7dd2f7789f
/factors.py
3eddb30138797bc77b43b0c51a24523dd64128b6
[]
no_license
malempati0/malempati00
d4ab09cad24ee9cfe5b4b8c20c7ba8fc414bd31c
b951bd215cb79a04305bdc5039d6b7597725d3ee
refs/heads/master
2020-04-07T07:11:13.201614
2018-11-26T04:45:58
2018-11-26T04:45:58
158,167,178
0
0
null
null
null
null
UTF-8
Python
false
false
120
py
def print_factors(x): for i in range(1, x + 1): if x % i == 0: print(i) num = 6 print_factors(num)
[ "noreply@github.com" ]
malempati0.noreply@github.com
8a438d371dcd47d1c7a958b870491293517d1a86
cf39aeabaae2fc0a16ddf4b458308d5ebde10a33
/modules/grindhold_plainhtml/__init__.py
a969ecaf93e1343f741158687ed33350effc5806
[]
no_license
joker234/skarphed
03dbb774a7605d6523926d082009a24f29455fd7
a7dc2bd758bf24cf8819d36e67633e68e87cf008
refs/heads/master
2021-01-18T06:36:20.996757
2013-09-06T02:20:25
2013-09-06T02:20:25
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,682
py
#!/usr/bin/python #-*- coding: utf-8 -*- ########################################################### # © 2011 Daniel 'grindhold' Brendle and Team # # This file is part of Skarphed. # # Skarphed is free software: you can redistribute it and/or # modify it under the terms of the GNU Affero General Public License # as published by the Free Software Foundation, either # version 3 of the License, or (at your option) any later # version. # # Skarphed is distributed in the hope that it will be # useful, but WITHOUT ANY WARRANTY; without even the implied # warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR # PURPOSE. See the GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public # License along with Skarphed. # If not, see http://www.gnu.org/licenses/. ########################################################### import os from StringIO import StringIO from module import AbstractModule class ModuleException(Exception): ERRORS = { 0:"""This instance does not have a WidgetId. Therefore, Widget-bound methods cannot be used""" } @classmethod def get_msg(cls,nr, info=""): return "DB_"+str(nr)+": "+cls.ERRORS[nr]+" "+info class Module(AbstractModule): def __init__(self, core): AbstractModule.__init__(self,core) self._path = os.path.dirname(__file__) self._load_manifest() """ BEGIN IMPLEMENTING YOUR MODULE HERE """ def render_pure_html(self,widget_id,args={}): content = self.get_content(widget_id) return"<h2>%s</h2>%s"%(content['title'],content['html']) def render_html(self,widget_id,args={}): return self.render_pure_html() def render_javascript(self,widget_id,args={}): return "" def set_content(self, widget_id, content="", title=""): title = str(title) content = StringIO(str(content)) db = self._core.get_db() stmnt = "UPDATE OR INSERT INTO ${html} (MOD_INSTANCE_ID, HTM_TITLE, HTM_HTML) \ VALUES (?,?,?) MATCHING (MOD_INSTANCE_ID) ;" db.query(self, stmnt, (widget_id, title, content), commit=True) def get_content(self, widget_id): db = self._core.get_db() stmnt = "SELECT HTM_TITLE, HTM_HTML FROM ${html} WHERE MOD_INSTANCE_ID = ? ;" cur = db.query(self, stmnt, (widget_id,)) row = cur.fetchonemap() if row is not None: return {'title':row["HTM_TITLE"], 'html':row["HTM_HTML"]} else: return {"title":"Widget not found", "html":"<p>This widget does apparently not exist</p>"}
[ "grindhold@gmx.net" ]
grindhold@gmx.net
d384f24b5c0b0b257f66b1db1a63854c59b95395
3e4c69317323bca865b025503b60bf83d3ae65f8
/tests/server/blueprints/variants/test_variant_views_variant.py
c1fd7fe078f8967099df90b24cb215c5a79a60ac
[ "BSD-3-Clause" ]
permissive
tapaswenipathak/scout
f59beaa997a45487ac96c3b3e560b5e5aa9b30ae
c9b3ec14f5105abe6066337110145a263320b4c5
refs/heads/master
2020-05-30T11:13:25.662300
2019-05-28T09:26:25
2019-05-28T09:26:25
189,694,812
1
0
BSD-3-Clause
2019-06-01T05:36:35
2019-06-01T05:36:34
null
UTF-8
Python
false
false
1,207
py
# -*- coding: utf-8 -*- import logging from flask import url_for log = logging.getLogger(__name__) def test_server_variant(app, real_adapter): # GIVEN an initialized app # GIVEN a valid user, institute, case and variant adapter = real_adapter variant_obj = adapter.variant_collection.find_one() assert variant_obj with app.test_client() as client: # GIVEN that the user could be logged in resp = client.get(url_for('auto_login')) assert resp.status_code == 200 internal_case_id = variant_obj['case_id'] case = adapter.case(internal_case_id) case_name = case['display_name'] owner = case['owner'] # NOTE needs the actual document_id, not the variant_id variant_id = variant_obj['_id'] log.debug('Inst {} case {} variant {}'.format(owner,case_name, variant_id)) # WHEN accessing the variant page resp = client.get(url_for('variants.variant', institute_id=owner, case_name=case_name, variant_id=variant_id)) log.debug("{}",resp.data) # THEN it should return a page assert resp.status_code == 200
[ "rasi.chiara@gmail.com" ]
rasi.chiara@gmail.com
dd42b52d712e69767f647a33a975f897d68b913f
5a52ccea88f90dd4f1acc2819997fce0dd5ffb7d
/alipay/aop/api/domain/OssDirectoryDetail.py
7b7aed746981c86b4885e7159246c6f7d6a7017c
[ "Apache-2.0" ]
permissive
alipay/alipay-sdk-python-all
8bd20882852ffeb70a6e929038bf88ff1d1eff1c
1fad300587c9e7e099747305ba9077d4cd7afde9
refs/heads/master
2023-08-27T21:35:01.778771
2023-08-23T07:12:26
2023-08-23T07:12:26
133,338,689
247
70
Apache-2.0
2023-04-25T04:54:02
2018-05-14T09:40:54
Python
UTF-8
Python
false
false
2,270
py
#!/usr/bin/env python # -*- coding: utf-8 -*- import json from alipay.aop.api.constant.ParamConstants import * class OssDirectoryDetail(object): def __init__(self): self._acl = None self._file_id = None self._file_name = None self._last_modified = None @property def acl(self): return self._acl @acl.setter def acl(self, value): self._acl = value @property def file_id(self): return self._file_id @file_id.setter def file_id(self, value): self._file_id = value @property def file_name(self): return self._file_name @file_name.setter def file_name(self, value): self._file_name = value @property def last_modified(self): return self._last_modified @last_modified.setter def last_modified(self, value): self._last_modified = value def to_alipay_dict(self): params = dict() if self.acl: if hasattr(self.acl, 'to_alipay_dict'): params['acl'] = self.acl.to_alipay_dict() else: params['acl'] = self.acl if self.file_id: if hasattr(self.file_id, 'to_alipay_dict'): params['file_id'] = self.file_id.to_alipay_dict() else: params['file_id'] = self.file_id if self.file_name: if hasattr(self.file_name, 'to_alipay_dict'): params['file_name'] = self.file_name.to_alipay_dict() else: params['file_name'] = self.file_name if self.last_modified: if hasattr(self.last_modified, 'to_alipay_dict'): params['last_modified'] = self.last_modified.to_alipay_dict() else: params['last_modified'] = self.last_modified return params @staticmethod def from_alipay_dict(d): if not d: return None o = OssDirectoryDetail() if 'acl' in d: o.acl = d['acl'] if 'file_id' in d: o.file_id = d['file_id'] if 'file_name' in d: o.file_name = d['file_name'] if 'last_modified' in d: o.last_modified = d['last_modified'] return o
[ "jishupei.jsp@alibaba-inc.com" ]
jishupei.jsp@alibaba-inc.com
068ea44d67a7cb5384e48561163e2762d5fce31c
8565e0f12bb11e14096964eeef3e34535c513d7f
/LunarisBot.py
029724af17fd46a914baa8e334a908ee3611a044
[]
no_license
No17Namsan/NightSkyK
2be02740cd3520bf8216abe6069721a39ee0ad4b
73aa3890780814726f5dc23bfe4416f75df200cb
refs/heads/main
2023-07-18T04:41:12.381368
2021-09-06T11:21:00
2021-09-06T11:21:00
403,484,701
0
0
null
null
null
null
UTF-8
Python
false
false
9,855
py
import asyncio import discord import random import re import urllib.request import urllib.parse from urllib.parse import quote import json from datetime import datetime, timedelta client = discord.Client() # 봇이 구동되었을 때 동작되는 코드입니다. @client.event async def on_ready(): print("Logged in as ") # 화면에 봇의 아이디, 닉네임이 출력됩니다. print(client.user.name) print(client.user.id) print(datetime.now()) print("===========") # 봇이 새로운 메시지를 수신했을때 동작되는 코드입니다. @client.event async def on_message(message): if message.author.bot: # 만약 메시지를 보낸사람이 봇일 경우에는 return None # 동작하지 않고 무시합니다. dice = ['<:d1:683941256711241744>', '<:d2:683941257000910869>', '<:d3:683941256606253068>', '<:d4:683941256937734203>', '<:d5:683941256862105620>', '<:d6:683941256891858964>'] food = ['워..... 워.... 네 체중을 생각해!', '너구리 순한맛! (그후 봇을 볼 수 없었다고 한다)', '굶어ㅋㅋ', '피자사주세요!', '오늘은 치느님을 영접할 시간입니다!', '갓스터치가 있는데 버거킹이 없을리가 없잖아! 버거킹?', '밥버거', '집밥이 최고!', '빵야! 빵야! 빵!', '루나랑 라면먹고 가실래요?', '수르스트뢰밍 https://namu.wiki/w/%EC%88%98%EB%A5%B4%EC%8A%A4%ED%8A%B8%EB%A2%B0%EB%B0%8D', '진리 콩 까네~ 진리 콩 까네~ 칠리 콘 카르네~', '김밥말아서 소풍가요!', '스시....? 라고 불리는 초밥!', '이봐, 친구! 그거 알아? 레몬 한 개엔 자그마치 레몬 한 개 분량의 비타민 C가 있다는 놀라운 사실을!', '일단 3D안경을 쓰고..:popcorn:', '부어라! 마셔라! 죽어라! :beer:', '도넛!', '커피는 좋은 도핑제입니다.', '넌... 아직... 준비가... 안되었다..!:baby_bottle: ', '후헤헤헷 숙제(일)을 머거랑 헤헷', '까까먹어', '빵과 계란후라이!', ':pig2: 족발! 돼지고기! 보쌈!', ':fish: 회?', '술에 부대찌개먹고싶... 아니 부대찌개에 술마시고싶다.', '어제도 오늘도 내일도 마라탕. 당연한거잖아요?', '떡뽀끼? 떡뽁이? 알게뭐람, 떡볶이 주세요!', '버거킹이 있는데 갓스터치가 없을리가 없잖아! 갓스터치!', '워워... 진정해..! 빡친 당신을 위한 엽떡을 가져왔다구!', '발은 역시 닭발이지!:chicken:', '말해 뭐해 곱창이 최고 아니야?', '삶은 감자.... Life is Egg... Egg...?', '아야어여오요**우유** :milk:', '쌀국수 뚝배기!', '아... 시리얼에 우유부어먹고싶다... ... ...?', '풀리와 웰시가 맛나게 먹는 밀웜 한번 먹어보실?', '민트초코가 치약맛일까 치약이 민트초코맛일까?'] do = ['잠만보처럼 잠만 자던가! :zzz:', '톰 아저씨의 무지개 여섯 공성할래?', '데스티니 가디언즈는 죽었지만 우리 케이드는 마음 속에 살아있어!', '생존마 낚으러 희생제갈까나~ 살인마 괴롭히러 희생제갈까나~', 'WINNER WINNER CHICKEN DINNER!', '느껴지지않아..? 우리의 심장에 뛰고있는 이 뜨거운 :cy:가!', '역시 힐링은 마인크래프트', '나만 없어 모동숲...ㅠ', '오늘도 싱글벙글 롤 협곡생활!', '우리집에서 넷플릭스보고갈래?(으브븝)', '밥머겅 많이머겅', '저 오늘떠나요~ 공항으로~ :airplane:', 'TFT 모바일 ㄷㄷㄷㅈ, ㅇㅍㄷㄷ', '타르코프에서 도망쳐! 도망치라구!'] fates = [':spades:', ':clubs:', ':diamonds:', ':hearts:'] fatecall = ['!합기', '!gkqrl', '!GKQRL'] num = [':regional_indicator_a:', ':two:', ':three:', ':four:', ':five:', ':six:', ':seven:', ':eight:', ':nine:', ':keycap_ten:', ':regional_indicator_j:', ':regional_indicator_q:', ':regional_indicator_k:'] result = [] eat = ['!뭐먹지', '!뭐먹지?', '!머먹지?', '!머먹지', '!멀먹징?', '!멀먹징', '!뭐먹징?', '!뭐먹징', '!뭐먹제?', '!뭐먹지?', '!뭐먹'] doing = ['!뭐하지?', '!뭐하지', '!뭐할까?', '!뭐할까'] up = ['!업', '!djq', '!DJQ', '!up', '!UP'] meow = ['애옹', '야옹', '먀옹', 'meow', 'moew', '냐오', '냐옹', '냥', '미야옹', '마오', '앩옹', '이얏호응', '애-옹', '야-옹'] meowe = ['<:meow1:682071155943014415>', '<:meow2:682071408540647563>', '<:meow3:684983336178810888>', '<:meow4:684983336824733697>', '<:meow5:684984172963692545>'] blackcow = ['음머', '살고시퍼여ㅠㅠㅠ', '음머어어어어어엉'] guild = message.author.guild # id라는 변수에는 메시지를 보낸사람의 ID를 담습니다. textchannel = message.channel # textchannel이라는 변수에는 메시지를 받은 채널의 ID를 담습니다. member = message.author now = datetime.now() if message.content == ('!패치노트'): print(member, guild, now, '!패치노트') print('==========') await textchannel.send(embed=discord.Embed(title='Ver.1.0.3a', description='0.재구동 시작했습니다.\n', colour=0xe3da13)) if message.content == ("!도움말"): print(member, now, guild, '!도움말') print('==========') await textchannel.send(embed=discord.Embed(title='도움말', description='1. !루나리스: 봇이 인사를 합니다.\n2. !뭐먹지?, !머먹지?, !멀먹징?, !뭐먹징?, !뭐먹제?: 봇이 음식을 추천합니다.\n3. 야옹, 애옹, 냥 등등: 고양이 이모지를 가져옵니다!\n' '4. !n(dDㅇ)N: N면체 주사위를 n개 던집니다. (N=1~6,n=1~9)', colour=0xe3da13)) if message.content.startswith('!루나리스'): # 인사 print(member, guild, now, '!루나리스') print('==========') await textchannel.send('안녕하세요. 여러분!') return None if message.content in doing: print(member, guild, now, '!뭐하지?') print('==========') await textchannel.send(do[random.randint(0, len(do) - 1)]) if message.content in eat: print(member, guild, now, '!뭐먹지?') print('==========') await textchannel.send(food[random.randint(0, len(food) - 1)]) if message.content in meow: print(member, guild, now, '야옹') print('==========') await textchannel.send(meowe[random.randint(0, len(meowe) - 1)]) if message.content.startswith("!갈고리"): print(member, guild, now, '!갈고리') print('==========') await textchannel.send( '<:QuestionSpam:767992761491259443><:QuestionSpam:767992761491259443><:QuestionSpam:767992761491259443><:QuestionSpam:767992761491259443><:QuestionSpam:767992761491259443>') if message.content == ("!나스"): print(member, guild, now, '!나스') print('==========') await textchannel.send("<:NBSB:766596746762649621> sp? 잠깐만요. 아니, 잠깐만 sp?") if message.content == ("!나스바보"): print(member, guild, now, '!나스바보') print('==========') await textchannel.send( "<:NBSB:766596746762649621> ㄴ <:UnIm:684328036065476613> ㄱ <:NBSB:766596746762649621>") if message.content == ("!나바스보"): print(member, guild, now, '!나바스보') print('==========') await textchannel.send( "<:NBSB:766596746762649621><:NBSB:766596746762649621>\n<:NBSB:766596746762649621><:NBSB:766596746762649621>") if message.content.startswith("!멜라"): print(member, guild, now, '!멜라') print('==========') await textchannel.send('<:D2Ghost:685817640366768174> **현실**을 사세요, 수호자!!') if message.content.startswith("!흑우"): print(member, guild, now, '!흑우') print('==========') await textchannel.send(blackcow[random.randint(0, len(blackcow) - 1)]) if message.content.startswith("!힐카"): print(member, guild, now, '!힐카') print('==========') await textchannel.send('힐카는 힝해') dice_set = re.findall('^!([0-9]+)[dDㅇ]([1-6])$', message.content) if len(dice_set) != 0: dice_set = dice_set[0] for _ in range(int(dice_set[0])): result.append(dice[random.randint(0, int(dice_set[1]) - 1)]) print(member, now, guild, '!주사위', result) print('==========') await textchannel.send(result) # if message.content in fatecall: # result = (fates[random.randint(0, 3)]), (num[random.randint(0, 12)]) # print(member, guild, now, '!합기', result) # print('==========') # await textchannel.send(result) # if message.content in up: # result = (dice[random.randint(0, 5)]) # print(member, guild, now, '!업', result) # print('==========') # await textchannel.send(result) if message.content.startswith("!이스포츠"): print(member, guild, now, '!이스포츠') print('==========') await textchannel.send(embed=discord.Embed(title='이스포츠 일정 정리 21년 07월 4주차' , colour=0xe3da13)) client.run("Discord_API")
[ "noreply@github.com" ]
No17Namsan.noreply@github.com
dfc0cc855a774de8fa89bf5d0af2e7761c1399da
cf0ab8503d4d704045070deea1e2125375711e86
/apps/apikeys/v1/urls.py
1a8b15c264dc105260d2432da2775b98a3fb3a99
[]
no_license
faierbol/syncano-platform
c3c6468600115752fd9fa5e46a0ad59f75f6bc9c
879111874d1ef70418b4890cf970720b0a2be4d8
refs/heads/master
2023-07-20T10:13:40.066127
2021-02-08T15:01:13
2021-02-08T15:01:13
null
0
0
null
null
null
null
UTF-8
Python
false
false
198
py
# coding=UTF8 from rest_framework.routers import SimpleRouter from apps.apikeys.v1 import views router = SimpleRouter() router.register('api_keys', views.ApiKeyViewSet) urlpatterns = router.urls
[ "rk@23doors.com" ]
rk@23doors.com
bd8a2b6e104c59c4b57a5cac5db23d29db1db3ec
908e60d308ca9458b89980be1095f58a07fce0bb
/playNFS.py
7b26a28595bb3a090c047db2100d5e011b07db5b
[]
no_license
KartikPatelOfficial/Ai-Nfs
f6220d2f1cf479ce613aa206a9150c4dd88602a5
00e39a9d3ab4bec6d26987c6849799587f3d02cc
refs/heads/master
2021-07-19T14:57:30.108037
2017-10-28T16:07:32
2017-10-28T16:07:32
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,365
py
import numpy as np from PIL import ImageGrab import cv2 import pyautogui import time from directKeys import ReleaseKey, PressKey, W, A, S, D def drawLines(img, lines): try: for line in lines: coords = line[0] cv2.line(img, (coords[0],coords[1]), (coords[2],coords[3]), [255,255,255], 3) except: pass def roi(img, vertices): mask = np.zeros_like(img) cv2.fillPoly(mask,vertices,255) masked = cv2.bitwise_and(img, mask) return masked def processImg(originalImage): processedImg = cv2.cvtColor(originalImage,cv2.COLOR_BGR2GRAY) processedImg = cv2.Canny(processedImg,threshold1=300,threshold2=400) processedImg = cv2.GaussianBlur(processedImg, (5,5),0) verticies = np.array([[10,500],[10,300],[300,200],[500,200],[800,300],[800,400]]) processedImg = roi(processedImg,[verticies]) lines = cv2.HoughLinesP(processedImg, 1, np.pi/180, 180, 20, 15) drawLines(processedImg,lines) return processedImg def main(): lastTime = time.time() while(True): screen = np.array(ImageGrab.grab(bbox=(0,40,800,640))) newScreen = processImg(screen) print('Loop took {} seconds'.format(time.time()-lastTime)) lastTime = time.time() cv2.imshow('window',newScreen) # cv2.imshow('window',cv2.cvtColor(screen,cv2.COLOR_BGR2RGB)) if cv2.waitKey(25) & 0xFF == ord('q'): cv2.destroyAllWindows() break main()
[ "patelkartik1910@gmail.com" ]
patelkartik1910@gmail.com
42bdb6a885ac58d51bad36beea8877307f7902a5
eda9187adfd53c03f55207ad05d09d2d118baa4f
/algo/Transfer_Learning/Transfer_learning.py
725a6e82bceb8aa1d09e9cb263fc2fdf9da6aea1
[]
no_license
HuiZhaozh/python_tutorials
168761c9d21ad127a604512d7c6c6b38b4faa3c7
bde4245741081656875bcba2e4e4fcb6b711a3d9
refs/heads/master
2023-07-07T20:36:20.137647
2020-04-24T07:18:25
2020-04-24T07:18:25
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,586
py
# -*- coding:utf-8 -*- # /usr/bin/python ''' ------------------------------------------------- File Name : Transfer_learning Description : 迁移学习 Envs : pytorch Author : yanerrol Date : 2020/2/17 09:58 ------------------------------------------------- Change Activity: 2020/2/17 : new ------------------------------------------------- ''' __author__ = 'yanerrol' import torch import time import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from torchvision import datasets from torchvision import transforms from torch.utils.data import DataLoader ####################################### ### PRE-TRAINED MODELS AVAILABLE HERE ## https://pytorch.org/docs/stable/torchvision/models.html from torchvision import models ####################################### if torch.cuda.is_available(): torch.backends.cudnn.deterministic = True ########################## ### SETTINGS ########################## # Device DEVICE = torch.device("cuda:0" if torch.cuda.is_available() else "cpu") print('Device:', DEVICE) NUM_CLASSES = 10 # Hyperparameters random_seed = 1 learning_rate = 0.0001 num_epochs = 10 batch_size = 128 ########################## ### MNIST DATASET ########################## custom_transform = transforms.Compose([ transforms.Resize((224, 224)), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) ]) ## Note that this particular normalization scheme is ## necessary since it was used for pre-training ## the network on ImageNet. ## These are the channel-means and standard deviations ## for z-score normalization. train_dataset = datasets.CIFAR10(root='data', train=True, transform=custom_transform, download=True) test_dataset = datasets.CIFAR10(root='data', train=False, transform=custom_transform) train_loader = DataLoader(dataset=train_dataset, batch_size=batch_size, num_workers=8, shuffle=True) test_loader = DataLoader(dataset=test_dataset, batch_size=batch_size, num_workers=8, shuffle=False) # Checking the dataset for images, labels in train_loader: print('Image batch dimensions:', images.shape) print('Image label dimensions:', labels.shape) break ########################## ### Loading Pre-Trained Model ########################## model = models.vgg16(pretrained=True) ########################## ### Freezing Model ########################## for param in model.parameters(): param.requires_grad = False model.classifier[3].requires_grad = True model.classifier[6] = nn.Sequential( nn.Linear(4096, 512), nn.ReLU(), nn.Dropout(0.5), nn.Linear(512, NUM_CLASSES)) ########################## ### Training as usual ########################## model = model.to(DEVICE) optimizer = torch.optim.Adam(model.parameters()) def compute_accuracy(model, data_loader): model.eval() correct_pred, num_examples = 0, 0 for i, (features, targets) in enumerate(data_loader): features = features.to(DEVICE) targets = targets.to(DEVICE) logits = model(features) _, predicted_labels = torch.max(logits, 1) num_examples += targets.size(0) correct_pred += (predicted_labels == targets).sum() return correct_pred.float() / num_examples * 100 def compute_epoch_loss(model, data_loader): model.eval() curr_loss, num_examples = 0., 0 with torch.no_grad(): for features, targets in data_loader: features = features.to(DEVICE) targets = targets.to(DEVICE) logits = model(features) loss = F.cross_entropy(logits, targets, reduction='sum') num_examples += targets.size(0) curr_loss += loss curr_loss = curr_loss / num_examples return curr_loss start_time = time.time() for epoch in range(num_epochs): model.train() for batch_idx, (features, targets) in enumerate(train_loader): features = features.to(DEVICE) targets = targets.to(DEVICE) ### FORWARD AND BACK PROP logits = model(features) cost = F.cross_entropy(logits, targets) optimizer.zero_grad() cost.backward() ### UPDATE MODEL PARAMETERS optimizer.step() ### LOGGING if not batch_idx % 50: print('Epoch: %03d/%03d | Batch %04d/%04d | Cost: %.4f' % (epoch + 1, num_epochs, batch_idx, len(train_loader), cost)) model.eval() with torch.set_grad_enabled(False): # save memory during inference print('Epoch: %03d/%03d | Train: %.3f%% | Loss: %.3f' % ( epoch + 1, num_epochs, compute_accuracy(model, train_loader), compute_epoch_loss(model, train_loader))) print('Time elapsed: %.2f min' % ((time.time() - start_time) / 60)) print('Total Training Time: %.2f min' % ((time.time() - start_time) / 60)) with torch.set_grad_enabled(False): # save memory during inference print('Test accuracy: %.2f%%' % (compute_accuracy(model, test_loader)) ########################## ### Training as usual ########################## import matplotlib.pyplot as plt classes = ('plane', 'car', 'bird', 'cat', 'deer', 'dog', 'frog', 'horse', 'ship', 'truck') for batch_idx, (features, targets) in enumerate(test_loader): features = features targets = targets break logits = model(features.to(DEVICE)) _, predicted_labels = torch.max(logits, 1) def unnormalize(tensor, mean, std): for t, m, s in zip(tensor, mean, std): t.mul_(s).add_(m) return tensor n_images = 10 fig, axes = plt.subplots(nrows=1, ncols=n_images, sharex=True, sharey=True, figsize=(20, 2.5)) orig_images = features[:n_images] for i in range(n_images): curr_img = orig_images[i].detach().to(torch.device('cpu')) curr_img = unnormalize(curr_img, torch.tensor([0.485, 0.456, 0.406]), torch.tensor([0.229, 0.224, 0.225])) curr_img = curr_img.permute((1, 2, 0)) axes[i].imshow(curr_img) axes[i].set_title(classes[predicted_labels[i]])
[ "2681506@gmail.com" ]
2681506@gmail.com
cab5f90d462f2fc8397fd3e49936dc077c525358
b37601e91fc9e6d4ad7dab19880747bca82f2e50
/General_Practice/listdictionary_practice.py
3d78becc649383215c0f2fd6cabe3711c437197e
[]
no_license
sarahannali/pythoncourse
d26389cd3fda830af3778051d7205021d5e150b3
b9ab543c19511e00a7b3e174b817c5eb0706a49a
refs/heads/master
2020-08-29T05:53:21.388641
2019-12-05T22:47:56
2019-12-05T22:47:56
217,947,564
0
1
null
null
null
null
UTF-8
Python
false
false
1,226
py
# #List Comprehension Practice a = input("List 5 numbers less than 10, separated by a space: ") b = input("List 5 more numbers less than 10: ") a = a.replace(" ", "") answer = [item for item in a and b if item in a and b] #Shared in both inputOne = input("What is your name? ") inputTwo = input("What is your pet's name? ") inputThree = input("What is your favorite color? ") listNames = [inputOne, inputTwo, inputThree] answer2 = [name[::-1].lower() for name in listNames] #Reverse a string print(f"\nYour number inputs shared the following numbers: {answer}") print(f"\nThose words backward are {answer2}") #Dictionary Practice import random options = ["pizza", "bread", "cookies", "cakes", "coffee"] print(f"\nWELCOME! We sell {options}") choice = input("What would you like to buy? ") stock = { "pizza" : random.randint(1,10), "bread": random.randint(1,20), "cookies": random.randint(1,100), "cakes": random.randint(1,5), "coffee": random.randint(1,100) } amount = stock.get(choice) if choice in stock: if amount > 1: choiceFix = choice + "s" else: choiceFix = choice print(f"\nWe have {amount} {choiceFix} left") else: print("\nSorry, we don't make that!")
[ "asarahali00@gmail.com" ]
asarahali00@gmail.com
8d70adcda10da24ef6cac3a8733ad6a6cd20df78
d53e9e90cb085046a3419be05966d3f0eef5b8e9
/v2/ovn_ovs.py
1e5a00c5a57e5b064d6629554e7a3c337619e2ff
[]
no_license
e2e-win/k8s-ovn-ovs
99ba544485e726919b2aa1cfe83865c5152cd0a8
ed21beeb2ddf30fc94555ba99745a2f8c9963de2
refs/heads/master
2020-03-28T23:56:20.237212
2020-01-03T10:19:59
2020-01-03T10:19:59
149,316,665
0
10
null
2020-01-03T10:20:01
2018-09-18T16:08:41
Python
UTF-8
Python
false
false
16,023
py
import ci import configargparse import openstack_wrap as openstack import log import utils import os import time import shutil import constants import yaml p = configargparse.get_argument_parser() p.add("--linuxVMs", action="append", help="Name for linux VMS. List.") p.add("--linuxUserData", help="Linux VMS user-data.") p.add("--linuxFlavor", help="Linux VM flavor.") p.add("--linuxImageID", help="ImageID for linux VMs.") p.add("--windowsVMs", action="append", help="Name for Windows VMs. List.") p.add("--windowsUserData", help="Windows VMS user-data.") p.add("--windowsFlavor", help="Windows VM flavor.") p.add("--windowsImageID", help="ImageID for windows VMs.") p.add("--keyName", help="Openstack SSH key name") p.add("--keyFile", help="Openstack SSH private key") p.add("--internalNet", help="Internal Network for VMs") p.add("--externalNet", help="External Network for floating ips") p.add("--ansibleRepo", default="http://github.com/openvswitch/ovn-kubernetes", help="Ansible Repository for ovn-ovs playbooks.") p.add("--ansibleBranch", default="master", help="Ansible Repository branch for ovn-ovs playbooks.") class OVN_OVS_CI(ci.CI): DEFAULT_ANSIBLE_PATH="/tmp/ovn-kubernetes" ANSIBLE_PLAYBOOK="ovn-kubernetes-cluster.yml" ANSIBLE_PLAYBOOK_ROOT="%s/contrib" % DEFAULT_ANSIBLE_PATH ANSIBLE_HOSTS_TEMPLATE=("[kube-master]\nKUBE_MASTER_PLACEHOLDER\n\n[kube-minions-linux]\nKUBE_MINIONS_LINUX_PLACEHOLDER\n\n" "[kube-minions-windows]\nKUBE_MINIONS_WINDOWS_PLACEHOLDER\n") ANSIBLE_HOSTS_PATH="%s/contrib/inventory/hosts" % DEFAULT_ANSIBLE_PATH DEFAULT_ANSIBLE_WINDOWS_ADMIN="Admin" DEFAULT_ANSIBLE_HOST_VAR_WINDOWS_TEMPLATE="ansible_user: USERNAME_PLACEHOLDER\nansible_password: PASS_PLACEHOLDER\n" DEFAULT_ANSIBLE_HOST_VAR_DIR="%s/contrib/inventory/host_vars" % DEFAULT_ANSIBLE_PATH HOSTS_FILE="/etc/hosts" ANSIBLE_CONFIG_FILE="%s/contrib/ansible.cfg" % DEFAULT_ANSIBLE_PATH KUBE_CONFIG_PATH="/root/.kube/config" KUBE_TLS_SRC_PATH="/etc/kubernetes/tls/" def __init__(self): self.opts = p.parse_known_args()[0] self.cluster = {} self.default_ansible_path = OVN_OVS_CI.DEFAULT_ANSIBLE_PATH self.ansible_playbook = OVN_OVS_CI.ANSIBLE_PLAYBOOK self.ansible_playbook_root = OVN_OVS_CI.ANSIBLE_PLAYBOOK_ROOT self.ansible_hosts_template = OVN_OVS_CI.ANSIBLE_HOSTS_TEMPLATE self.ansible_hosts_path = OVN_OVS_CI.ANSIBLE_HOSTS_PATH self.ansible_windows_admin = OVN_OVS_CI.DEFAULT_ANSIBLE_WINDOWS_ADMIN self.ansible_host_var_windows_template = OVN_OVS_CI.DEFAULT_ANSIBLE_HOST_VAR_WINDOWS_TEMPLATE self.ansible_host_var_dir = OVN_OVS_CI.DEFAULT_ANSIBLE_HOST_VAR_DIR self.ansible_config_file = OVN_OVS_CI.ANSIBLE_CONFIG_FILE self.logging = log.getLogger(__name__) self.post_deploy_reboot_required = True def _add_linux_vm(self, vm_obj): if self.cluster.get("linuxVMs") == None: self.cluster["linuxVMs"] = [] self.cluster["linuxVMs"].append(vm_obj) def _add_windows_vm(self, vm_obj): if self.cluster.get("windowsVMs") == None: self.cluster["windowsVMs"] = [] self.cluster["windowsVMs"].append(vm_obj) def _get_windows_vms(self): return self.cluster.get("windowsVMs") def _get_linux_vms(self): return self.cluster.get("linuxVMs") def _get_all_vms(self): return self._get_linux_vms() + self._get_windows_vms() def _get_vm_fip(self, vm_obj): return vm_obj.get("FloatingIP") def _set_vm_fip(self, vm_obj, ip): vm_obj["FloatingIP"] = ip def _create_vms(self): self.logging.info("Creating Openstack VMs") vmPrefix = self.opts.cluster_name for vm in self.opts.linuxVMs: openstack_vm = openstack.server_create("%s-%s" % (vmPrefix, vm), self.opts.linuxFlavor, self.opts.linuxImageID, self.opts.internalNet, self.opts.keyName, self.opts.linuxUserData) fip = openstack.get_floating_ip(openstack.floating_ip_list()[0]) openstack.server_add_floating_ip(openstack_vm['name'], fip) self._set_vm_fip(openstack_vm, fip) self._add_linux_vm(openstack_vm) for vm in self.opts.windowsVMs: openstack_vm = openstack.server_create("%s-%s" % (vmPrefix, vm), self.opts.windowsFlavor, self.opts.windowsImageID, self.opts.internalNet, self.opts.keyName, self.opts.windowsUserData) fip = openstack.get_floating_ip(openstack.floating_ip_list()[0]) openstack.server_add_floating_ip(openstack_vm['name'], fip) self._set_vm_fip(openstack_vm, fip) self._add_windows_vm(openstack_vm) self.logging.info("Succesfuly created VMs %s" % [ vm.get("name") for vm in self._get_all_vms()]) def _wait_for_windows_machines(self): self.logging.info("Waiting for Windows VMs to obtain Admin password.") for vm in self._get_windows_vms(): openstack.server_get_password(vm['name'], self.opts.keyFile) self.logging.info("Windows VM: %s succesfully obtained password." % vm.get("name")) def _prepare_env(self): self._create_vms() self._wait_for_windows_machines() def _destroy_cluster(self): vmPrefix = self.opts.cluster_name for vm in self.opts.linuxVMs: openstack.server_delete("%s-%s" % (vmPrefix, vm)) for vm in self.opts.windowsVMs: openstack.server_delete("%s-%s" % (vmPrefix, vm)) def _prepare_ansible(self): utils.clone_repo(self.opts.ansibleRepo, self.opts.ansibleBranch, self.default_ansible_path) # Creating ansible hosts file linux_master = self._get_linux_vms()[0].get("name") linux_minions = [vm.get("name") for vm in self._get_linux_vms()[1:]] windows_minions = [vm.get("name") for vm in self._get_windows_vms()] hosts_file_content = self.ansible_hosts_template.replace("KUBE_MASTER_PLACEHOLDER", linux_master) hosts_file_content = hosts_file_content.replace("KUBE_MINIONS_LINUX_PLACEHOLDER", "\n".join(linux_minions)) hosts_file_content = hosts_file_content.replace("KUBE_MINIONS_WINDOWS_PLACEHOLDER","\n".join(windows_minions)) self.logging.info("Writing hosts file for ansible inventory.") with open(self.ansible_hosts_path, "w") as f: f.write(hosts_file_content) # Creating hosts_vars for hosts for vm in self._get_windows_vms(): vm_name = vm.get("name") vm_username = self.ansible_windows_admin # TO DO: Have this configurable trough opts vm_pass = openstack.server_get_password(vm_name, self.opts.keyFile) hosts_var_content = self.ansible_host_var_windows_template.replace("USERNAME_PLACEHOLDER", vm_username).replace("PASS_PLACEHOLDER", vm_pass) filepath = os.path.join(self.ansible_host_var_dir, vm_name) with open(filepath, "w") as f: f.write(hosts_var_content) # Populate hosts file with open(OVN_OVS_CI.HOSTS_FILE,"a") as f: for vm in self._get_all_vms(): vm_name = vm.get("name") if vm_name.find("master") > 0: vm_name = vm_name + " kubernetes" hosts_entry=("%s %s\n" % (self._get_vm_fip(vm), vm_name)) self.logging.info("Adding entry %s to hosts file." % hosts_entry) f.write(hosts_entry) # Enable ansible log and set ssh options with open(self.ansible_config_file, "a") as f: log_file = os.path.join(self.opts.log_path, "ansible-deploy.log") log_config = "log_path=%s\n" % log_file # This probably goes better in /etc/ansible.cfg (set in dockerfile ) ansible_config="\n\n[ssh_connection]\nssh_args=-o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null\n" f.write(log_config) f.write(ansible_config) full_ansible_tmp_path = os.path.join(self.ansible_playbook_root, "tmp") utils.mkdir_p(full_ansible_tmp_path) # Copy kubernetes prebuilt binaries for file in ["kubelet","kubectl","kube-apiserver","kube-controller-manager","kube-scheduler","kube-proxy"]: full_file_path = os.path.join(utils.get_k8s_folder(), constants.KUBERNETES_LINUX_BINS_LOCATION, file) self.logging.info("Copying %s to %s." % (full_file_path, full_ansible_tmp_path)) shutil.copy(full_file_path, full_ansible_tmp_path) for file in ["kubelet.exe", "kubectl.exe", "kube-proxy.exe"]: full_file_path = os.path.join(utils.get_k8s_folder(), constants.KUBERNETES_WINDOWS_BINS_LOCATION, file) self.logging.info("Copying %s to %s." % (full_file_path, full_ansible_tmp_path)) shutil.copy(full_file_path, full_ansible_tmp_path) def _deploy_ansible(self): self.logging.info("Starting Ansible deployment.") cmd = "ansible-playbook %s -v" % self.ansible_playbook cmd = cmd.split() cmd.append("--key-file=%s" % self.opts.keyFile) out, _ ,ret = utils.run_cmd(cmd, stdout=True, cwd=self.ansible_playbook_root) if ret != 0: self.logging.error("Failed to deploy ansible-playbook with error: %s" % out) raise Exception("Failed to deploy ansible-playbook with error: %s" % out) self.logging.info("Succesfully deployed ansible-playbook.") def _waitForConnection(self, machine, windows): self.logging.info("Waiting for connection to machine %s." % machine) cmd = ["ansible"] cmd.append(machine) if not windows: cmd.append("--key-file=%s" % self.opts.keyFile) cmd.append("-m") cmd.append("wait_for_connection") cmd.append("-a") cmd.append("'connect_timeout=5 sleep=5 timeout=600'") out, _, ret = utils.run_cmd(cmd, stdout=True, cwd=self.ansible_playbook_root, shell=True) return ret, out def _copyTo(self, src, dest, machine, windows=False, root=False): self.logging.info("Copying file %s to %s:%s." % (src, machine, dest)) cmd = ["ansible"] if root: cmd.append("--become") if not windows: cmd.append("--key-file=%s" % self.opts.keyFile) cmd.append(machine) cmd.append("-m") module = "win_copy" if windows else "copy" cmd.append(module) cmd.append("-a") cmd.append("'src=%(src)s dest=%(dest)s flat=yes'" % {"src": src, "dest": dest}) ret, _ = self._waitForConnection(machine, windows=windows) if ret != 0: self.logging.error("No connection to machine: %s", machine) raise Exception("No connection to machine: %s", machine) # Ansible logs everything to stdout out, _, ret = utils.run_cmd(cmd, stdout=True, cwd=self.ansible_playbook_root, shell=True) if ret != 0: self.logging.error("Ansible failed to copy file to %s with error: %s" % (machine, out)) raise Exception("Ansible failed to copy file to %s with error: %s" % (machine, out)) def _copyFrom(self, src, dest, machine, windows=False, root=False): self.logging.info("Copying file %s:%s to %s." % (machine, src, dest)) cmd = ["ansible"] if root: cmd.append("--become") if not windows: cmd.append("--key-file=%s" % self.opts.keyFile) cmd.append(machine) cmd.append("-m") cmd.append("fetch") cmd.append("-a") cmd.append("'src=%(src)s dest=%(dest)s flat=yes'" % {"src": src, "dest": dest}) # TO DO: (atuvenie) This could really be a decorator ret, _ = self._waitForConnection(machine, windows=windows) if ret != 0: self.logging.error("No connection to machine: %s", machine) raise Exception("No connection to machine: %s", machine) out, _, ret = utils.run_cmd(cmd, stdout=True, cwd=self.ansible_playbook_root, shell=True) if ret != 0: self.logging.error("Ansible failed to fetch file from %s with error: %s" % (machine, out)) raise Exception("Ansible failed to fetch file from %s with error: %s" % (machine, out)) def _runRemoteCmd(self, command, machine, windows=False, root=False): self.logging.info("Running cmd on remote machine %s." % (machine)) cmd=["ansible"] if root: cmd.append("--become") if windows: task = "win_shell" else: task = "shell" cmd.append("--key-file=%s" % self.opts.keyFile) cmd.append(machine) cmd.append("-m") cmd.append(task) cmd.append("-a") cmd.append("'%s'" % command) ret, _ = self._waitForConnection(machine, windows=windows) if ret != 0: self.logging.error("No connection to machine: %s", machine) raise Exception("No connection to machine: %s", machine) out, _, ret = utils.run_cmd(cmd, stdout=True, cwd=self.ansible_playbook_root, shell=True) if ret != 0: self.logging.error("Ansible failed to run command %s on machine %s with error: %s" % (cmd, machine, out)) raise Exception("Ansible failed to run command %s on machine %s with error: %s" % (cmd, machine, out)) def _prepullImages(self): # TO DO: This path should be passed as param prepull_script="/tmp/k8s-ovn-ovs/v2/prepull.ps1" for vm in self._get_windows_vms(): self.logging.info("Copying prepull script to node %s" % vm["name"]) self._copyTo(prepull_script, "c:\\", vm["name"], windows=True) self._runRemoteCmd("c:\\prepull.ps1", vm["name"], windows=True) def _prepareTestEnv(self): # For OVN-OVS CI: copy config file from .kube folder of the master node # Replace Server in config with dns-name for the machine # Export appropriate env vars linux_master = self._get_linux_vms()[0].get("name") self.logging.info("Copying kubeconfig from master") self._copyFrom("/root/.kube/config","/tmp/kubeconfig", linux_master, root=True) self._copyFrom("/etc/kubernetes/tls/ca.pem","/etc/kubernetes/tls/ca.pem", linux_master, root=True) self._copyFrom("/etc/kubernetes/tls/admin.pem","/etc/kubernetes/tls/admin.pem", linux_master, root=True) self._copyFrom("/etc/kubernetes/tls/admin-key.pem","/etc/kubernetes/tls/admin-key.pem", linux_master, root=True) with open("/tmp/kubeconfig") as f: content = yaml.load(f) for cluster in content["clusters"]: cluster["cluster"]["server"] = "https://kubernetes" with open("/tmp/kubeconfig", "w") as f: yaml.dump(content, f) os.environ["KUBE_MASTER"] = "local" os.environ["KUBE_MASTER_IP"] = "kubernetes" os.environ["KUBE_MASTER_URL"] = "https://kubernetes" os.environ["KUBECONFIG"] = "/tmp/kubeconfig" try: if self.post_deploy_reboot_required: for vm in self._get_windows_vms(): openstack.reboot_server(vm["name"]) self._prepullImages() except: pass def up(self): self.logging.info("Bringing cluster up.") try: self._prepare_env() self._prepare_ansible() self._deploy_ansible() except Exception as e: raise e def build(self): self.logging.info("Building k8s binaries.") utils.get_k8s(repo=self.opts.k8s_repo, branch=self.opts.k8s_branch) utils.build_k8s_binaries() def down(self): self.logging.info("Destroying cluster.") try: self._destroy_cluster() except Exception as e: raise e
[ "atuvenie@cloudbasesolutions.com" ]
atuvenie@cloudbasesolutions.com
30fd9ea784dffc56bf761d1938ac9c00617ccab4
8467f6026afd620aa7efc6bf8d5db7970a25460e
/calc.py
ad775e41cdf051c726c1954980c2cd0e20f73396
[]
no_license
alisebruevich/Graphing-Calculator
848563a89d1e28c7dcf760a35034bcd2c5a24bad
06c3ae99f46408d475bd31f715cf4bc448922af5
refs/heads/master
2020-08-19T14:08:55.178307
2019-10-18T02:40:19
2019-10-18T02:40:19
215,926,899
0
0
null
null
null
null
UTF-8
Python
false
false
13,595
py
#newest, latest, freshest #fix derivatives with masking and unmasking arrays for x and y: Bonnie #presentation (what we did, learned; 3-5 slides): Alise #code block diagram: Bonnie #test to see if exponential functions will work: Alise -> IT DOESN'T CRY #make labels for what different colors/symbols mean on the graph: Alise -> DONE #double check that log stuff works: Alise -> IT DOESN'T CRY ## #f is original, g is first derivative, h is second derivative import math as m from tkinter import * from sympy import * def graph(input,a,b): import numpy.ma as M import numpy as np from numpy import linspace from sympy.parsing.sympy_parser import parse_expr import matplotlib.pyplot as mpl from sympy.parsing.sympy_parser import standard_transformations,\ implicit_multiplication_application transformations = (standard_transformations +(implicit_multiplication_application,)) mpl.axhline(color="black") mpl.axvline(color="black") x = symbols('x') listofx=[] x_vals = np.arange(-50,50,0.01) for i in range(0, len(x_vals)): x_vals[i] = round(x_vals[i],4) specialx_vals=x_vals f=parse_expr(input, transformations=transformations)#parsing function fy_vals=[] zoo = parse_expr("1/0", transformations = transformations) nan = parse_expr("0/0", transformations = transformations) for i in x_vals: fy_vals.append(f.subs(x,i)) for i in specialx_vals: if f.subs(x,i).is_real==False: x_vals=x_vals[1:len(x_vals)] fy_vals=fy_vals[1:len(fy_vals)] specialx_vals=x_vals maskedy_vals=fy_vals vals=[] for i in range(0, len(fy_vals)): #Case 1: 1/0 , or zoo, which is an Asymptote if (fy_vals[i] == zoo): vals.append(x_vals[i]) #Case 2: 0/0, or nan, which is a Hole elif (fy_vals[i] == nan): vals.append(x_vals[i]) #graph derivative gy_vals=differentiate(x_vals,fy_vals) maskedgy_vals=gy_vals x_vals=x_vals[0:len(x_vals)-1] for i in range(0, len(vals)): maskedgy_vals = M.masked_where(x_vals == vals[i], maskedgy_vals) maskedgy_vals = M.masked_where(x_vals == x_vals[0], maskedgy_vals) mpl.plot(x_vals, maskedgy_vals,color="orange")#orange if isnumeric(a) and isnumeric(b): a=float(a) b=float(b) integrate(f,a,b) #graph 2nd derivative hy_vals=differentiate(x_vals,gy_vals) maskedhy_vals=hy_vals x_vals=x_vals[0:len(x_vals)-1] for i in range(0, len(vals)): maskedhy_vals = M.masked_where(x_vals == vals[i], maskedhy_vals) maskedhy_vals= M.masked_where(x_vals==x_vals[0],maskedhy_vals) maskedhy_vals= M.masked_where(x_vals==x_vals[1],maskedhy_vals) mpl.plot(x_vals, maskedhy_vals, color="green")#green #graph discontinuities lastFunction = simplify(f) print(lastFunction) for i in range(0, len(fy_vals)): #Case 1: 1/0 , or zoo, which is an Asymptote if (fy_vals[i] == zoo): print("it's asymptote!") maskedy_vals = M.masked_where(specialx_vals == specialx_vals[i], maskedy_vals) mpl.axvline(x=specialx_vals[i], color='r') #Case 2: 0/0, or nan, which is a Hole elif (fy_vals[i] == nan): print("it's a hole!") maskedy_vals = M.masked_where(specialx_vals == specialx_vals[i], maskedy_vals) mpl.plot(specialx_vals[i],lastFunction.subs(x, specialx_vals[i]),color="black",marker="p") mpl.plot(specialx_vals, maskedy_vals,color="blue")#plots#blue #find extrema #extrema are graphed as pink pentagons basically_zero = 1*10**-4 #print(x_vals) for i in range(0,len(gy_vals)-1): if (not(gy_vals[i]== nan or gy_vals[i]==zoo)) and abs(gy_vals[i])<basically_zero: if gy_vals[i-1]>basically_zero and gy_vals[i+1]<(basically_zero*-1): mpl.plot(x_vals[i],maskedy_vals[i],color="#FF00FC",marker="p") print("extrema:") print(x_vals[i]) print(maskedy_vals[i]) if gy_vals[i-1]<(-1*basically_zero) and gy_vals[i+1]>basically_zero: mpl.plot(x_vals[i],maskedy_vals[i],color="#FF00FC",marker="p") print("extrema:") print(x_vals[i]) print(maskedy_vals[i]) elif (not(gy_vals[i]== nan or gy_vals[i]==zoo)) and (not(gy_vals[i+1]== nan or gy_vals[i+1]==zoo)) and gy_vals[i]>0 and gy_vals[i+1]<0: xbetween=np.arange(x_vals[i],x_vals[i+1],0.00001) betweenvals=[] fbetween=[] for i in xbetween: fbetween.append(f.subs(x,i)) maskedfbetween=fbetween for i in range(0,len(maskedfbetween)): if maskedfbetween[i] == zoo or maskedfbetween == nan: betweenvals.append(xbetween[i]) for i in range(0, len(betweenvals)): maskedfbetween = M.masked_where(xbetween == betweenvals[i], maskedfbetween) gbetween=differentiate(xbetween,fbetween) for i in range(0,len(gbetween)-1): if abs(gbetween[i])<basically_zero: if gbetween[i-1]>0 and gbetween[i+1]<0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="p") print("extrema:") print(xbetween[i]) print(maskedfbetween[i]) if gbetween[i-1]<0 and gbetween[i+1]>0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="p") print("extrema:") print(xbetween[i]) print(maskedfbetween[i]) elif (not(gy_vals[i]== nan or gy_vals[i]==zoo)) and (not(gy_vals[i+1]== nan or gy_vals[i+1]==zoo)) and gy_vals[i]<0 and gy_vals[i+1]>0: xbetween=np.arange(x_vals[i],x_vals[i+1],0.00001) betweenvals=[] fbetween=[] for i in xbetween: fbetween.append(f.subs(x,i)) maskedfbetween=fbetween for i in range(0,len(maskedfbetween)): if maskedfbetween[i] == zoo or maskedfbetween == nan: betweenvals.append(xbetween[i]) for i in range(0, len(betweenvals)): maskedfbetween = M.masked_where(xbetween == betweenvals[i], maskedfbetween) gbetween=differentiate(xbetween,fbetween) for i in range(0,len(gbetween)-1): if abs(gbetween[i])<basically_zero: if gbetween[i-1]>0 and gbetween[i+1]<0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="p") print("extrema:") print(xbetween[i]) print(maskedfbetween[i]) if gbetween[i-1]<0 and gbetween[i+1]>0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="p") print("extrema:") print(xbetween[i]) print(maskedfbetween[i]) #find inflection points #inflection points are graphed as green stars for i in range(0,len(hy_vals)-1): if (not(hy_vals[i]== nan or hy_vals[i]==zoo)) and abs(hy_vals[i])<basically_zero: if not(hy_vals[i-1]== nan or hy_vals[i+1]==zoo): if hy_vals[i-1]>basically_zero and hy_vals[i+1]<(basically_zero*-1): mpl.plot(x_vals[i],maskedy_vals[i],color="#FF00FC",marker="*") print("inflection:") print(x_vals[i]) print(maskedy_vals[i]) if hy_vals[i-1]<(-1*basically_zero) and hy_vals[i+1]>basically_zero: mpl.plot(x_vals[i],maskedy_vals[i],color="#FF00FC",marker="*") print("inflection:") print(x_vals[i]) print(maskedy_vals[i]) elif (not(hy_vals[i]== nan or hy_vals[i]==zoo)) and (not(hy_vals[i+1]== nan or hy_vals[i+1]==zoo)) and hy_vals[i]>0 and hy_vals[i+1]<0: xbetween=np.arange(x_vals[i],x_vals[i+1],0.00001) betweenvals=[] fbetween=[] for i in xbetween: fbetween.append(f.subs(x,i)) maskedfbetween=fbetween for i in range(0,len(maskedfbetween)): if maskedfbetween[i] == zoo or maskedfbetween == nan: betweenvals.append(xbetween[i]) gbetween=differentiate(xbetween,fbetween) xbetween=xbetween[0:len(xbetween)-1] maskedgbetween=gbetween for i in range(0, len(betweenvals)): maskedgbetween = M.masked_where(xbetween == betweenvals[i], maskedgbetween) hbetween=differentiate(xbetween,gbetween) for i in range(0,len(hbetween)-1): if abs(hbetween[i])<basically_zero: if hbetween[i-1]>0 and hbetween[i+1]<0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="*") print("inflection:") print(xbetween[i]) print(maskedfbetween[i]) if hbetween[i-1]<0 and hbetween[i+1]>0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="*") print("inflection:") print(xbetween[i]) print(maskedfbetween[i]) elif (not(hy_vals[i]== nan or hy_vals[i]==zoo)) and (not(hy_vals[i+1]== nan or hy_vals[i+1]==zoo)) and hy_vals[i]<0 and hy_vals[i+1]>0: xbetween=np.arange(x_vals[i],x_vals[i+1],0.00001) betweenvals=[] fbetween=[] for i in xbetween: fbetween.append(f.subs(x,i)) maskedfbetween=fbetween for i in range(0,len(maskedfbetween)): if maskedfbetween[i] == zoo or maskedfbetween == nan: betweenvals.append(xbetween[i]) gbetween=differentiate(xbetween,fbetween) xbetween=xbetween[0:len(xbetween)-1] maskedgbetween=gbetween for i in range(0, len(betweenvals)): maskedgbetween = M.masked_where(xbetween == betweenvals[i], maskedgbetween) hbetween=differentiate(xbetween,gbetween) for i in range(0,len(hbetween)-1): if abs(hbetween[i])<basically_zero: if hbetween[i-1]>0 and hbetween[i+1]<0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="*") print("inflection:") print(xbetween[i]) print(maskedfbetween[i]) if hbetween[i-1]<0 and hbetween[i+1]>0: mpl.plot(xbetween[i],maskedfbetween[i],color="#FF00FC",marker="*") print("inflection:") print(xbetween[i]) print(maskedfbetween[i]) mpl.axhline(color="black") mpl.axvline(color="black") mpl.xlim(-5,5) mpl.ylim(-10, 10) mpl.grid(b=True)#sets grid mpl.show() def isnumeric(s): try: float(s) return True except ValueError: return False def differentiate(x,y): #takes in arguments: list of x values and list of y values gy_vals=[] #makes list of derivative values for n in range(1,len(x)-1): #loops through x values hi=(y[n+1]-y[n-1])/(x[n+1]-x[n-1]) #finds the slope using the values around it gy_vals.append(hi) #adds it to the list gy_vals.insert(0,gy_vals[0]) return gy_vals #returns derivative values ##############integrate function def integrate(f,a,b): #takes in arguments: function f, lower bound a, and upper bound b import numpy.ma as M import numpy as np from numpy import linspace from sympy.parsing.sympy_parser import parse_expr import matplotlib.pyplot as mpl x = symbols('x') #generates list of x values in small increments in between a and b if a<b: xvals=np.arange(a,b,0.001) if a>b: xvals=np.arange(b,a,0.001) if a==b: return fyvals=[] for i in xvals: fyvals.append(f.subs(x,i)) #creates list of y vals that correspont to x vals gyvals=differentiate(xvals,fyvals) #gets list of derivative values xvals=xvals[0:len(xvals)-1] sum=0 for n in range(0,len(xvals)-1): #finds areas of trapezoids and then adds them to sum trapezoid=0.5*(xvals[n+1]-xvals[n])*(gyvals[n]+gyvals[n+1]) sum+=trapezoid ftc=f.subs(x,b)-f.subs(x,a) #calculates f(b)-f(b) which should be equal to the sum thing="%s = %s" %(sum,ftc) #puts the equality on the graph mpl.text(0,0,thing) def evaluate(event): input= entry1.get() a=entrya.get() b=entryb.get() graph(input,a,b) w = Tk() w.title("Graphing Calculator") Label(w, text="Your Expression:").pack() entry1 = Entry(w) entry1.bind("<Return>", evaluate) entry1.pack() Label(w, text="a:").pack() entrya = Entry(w) entrya.bind("<Return>", evaluate) entrya.pack() Label(w, text="b:").pack() entryb = Entry(w) entryb.bind("<Return>", evaluate) entryb.pack() Label(w, text="MARKERS:").pack() Label(w, text="Extrema are graphed as pink pentagons").pack() Label(w, text="Inflection points are graphed as pink stars").pack() Label(w, text="Asymptotes graphed as a red line").pack() Label(w, text="Holes graphed as black pentagons").pack() res = Label(w) res.pack() w.mainloop()
[ "noreply@github.com" ]
alisebruevich.noreply@github.com
40ccefb509a8910430c7b9e7396fe596022f0fdf
3de6b4bbaf86afe1ff87b0d829bcba014a8a9696
/.history/home/models_20200512135208.py
77a219ab1076ce6b855a9fb57b4f9eeb54707ffb
[]
no_license
codermythbuster/bs
f0b758ab396e50f744aa29c7ecd58354b7df06df
3687302da8f9fe5a8f75d52ba429f14e2f09c67e
refs/heads/master
2022-08-04T22:27:59.987215
2020-05-26T11:24:01
2020-05-26T11:24:01
263,302,705
0
0
null
2020-05-26T09:58:03
2020-05-12T10:16:30
HTML
UTF-8
Python
false
false
1,225
py
from django.db import models from django.contrib.auth.models import User # book keeping record model class Book(models.Model): book_id = models.IntegerField(primary_key=True,verbose_name="BOOK ID") book_name = models.TextField(verbose_name="BOOK NAME") book_category = models.CharField(max_length=100,verbose_name="BOOK CATEGORY") book_description = models.TextField(verbose_name="BOOK DESCRIPTION") book_author_name = models.CharField(max_length=256, verbose_name="AUTHOR Name") book_price_to_sell = models.FloatField(verbose_name="Selling Price") username = models.ForeignKey(User,verbose_name="USERNAME") book_status = models.BooleanField(default=False, verbose_name="AVAILABILITY") book_image = models.ImageField(verbose_name="IMAGE URL") def __str__(self): return " {} {} {} {} {} {} ".format(self.book_id,self.book_name,self.book_author_name,self.book_category,self.book_status,self.book_image) class Books_purchased (models.Model): trans_id = models.AutoField(verbose_name="Transaction ID") user1 = models.ForeignKey(to_field=User.USERNAME_FIELD) user2 = models.ForeignKey(to_field=User.USERNAME_FIELD) book_id = models.ForeignKey(Book)
[ "namdev373@gmail.com" ]
namdev373@gmail.com
c207cd4f3194bfde257b57a18093edeb474a8c31
9a5fa4f1fcd2d335347dda8b6672ff0392003823
/backend/exprgram/migrations/0006_auto_20180724_0331.py
6f9ef516992a157f06519ae48738df51e0240ff7
[]
no_license
kyungjejo/exprgram-evaluation
ea9db3defc215f0f7ff4fde878267dcb7618e6c7
d0a549fbe3072bfc5aa20d645e3e4cfb0ba60f06
refs/heads/master
2023-01-13T03:09:26.063495
2018-09-10T11:16:08
2018-09-10T11:16:08
146,848,149
0
0
null
null
null
null
UTF-8
Python
false
false
874
py
# Generated by Django 2.0.6 on 2018-07-24 03:31 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('exprgram', '0005_auto_20180719_1659'), ] operations = [ migrations.AlterField( model_name='emotionlables', name='count', field=models.IntegerField(default=1), ), migrations.AlterField( model_name='intentionlables', name='count', field=models.IntegerField(default=1), ), migrations.AlterField( model_name='locationlables', name='count', field=models.IntegerField(default=1), ), migrations.AlterField( model_name='relationshiplables', name='count', field=models.IntegerField(default=1), ), ]
[ "kyungjejo@gmail.com" ]
kyungjejo@gmail.com
13a0070714639e0e12858309ad4019cba7a00079
b0d91025a0c188b8ecc4c328a38e3d0c158309a3
/mysite/settings.py
96b107a7f1405ad0f5f26aafebb1e9461262af79
[]
no_license
sdumi/dj_tut
26e9c742e2d72f5ccb0ae08e4a65bbacabf25192
fb7c0c979372da9de0ed30068fc12fe891ab6d12
refs/heads/master
2020-12-24T16:06:15.022099
2010-11-23T12:04:14
2010-11-23T12:04:14
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,241
py
# Django settings for mysite project. import os.path DEBUG = True #DEBUG = False TEMPLATE_DEBUG = DEBUG PROJECT_DIR = os.path.dirname(__file__) ADMINS = ( ('Dumi', 'dumitru.sipos@gmail.com'), ('Dumitru Sipos', 'dumitru.sipos@alcatel-lucent.com') ) EMAIL_HOST='smtp.tm.alcatel.ro' EMAIL_PORT=25 MANAGERS = ADMINS # shortcut for os.path.join # called: pj("templates") ==> os.path.join(PROJECT_DIR, "templates") pj = lambda filename: os.path.join(PROJECT_DIR, filename) # just an alias for os.path.joi j = os.path.join # cannot use pj here: it takes only the new path to be added to PROJECT_DIR # and I do not want to pass "../databases"... not sure how well that works on Windows... dbname = j(j(j(PROJECT_DIR, ".."), "databases"), "mysite.db") DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', # Add 'postgresql_psycopg2', 'postgresql', 'mysql', 'sqlite3' or 'oracle'. # 'NAME': '/home/dsipos/prg/web/dj_tut/databases/mysite.db', # Or path to database file if using sqlite3. 'NAME': dbname, 'USER': '', # Not used with sqlite3. 'PASSWORD': '', # Not used with sqlite3. 'HOST': '', # Set to empty string for localhost. Not used with sqlite3. 'PORT': '', # Set to empty string for default. Not used with sqlite3. } } # Local time zone for this installation. Choices can be found here: # http://en.wikipedia.org/wiki/List_of_tz_zones_by_name # although not all choices may be available on all operating systems. # On Unix systems, a value of None will cause Django to use the same # timezone as the operating system. # If running in a Windows environment this must be set to the same as your # system time zone. TIME_ZONE = 'Europe/Bucharest' # Language code for this installation. All choices can be found here: # http://www.i18nguy.com/unicode/language-identifiers.html LANGUAGE_CODE = 'en-us' SITE_ID = 1 # If you set this to False, Django will make some optimizations so as not # to load the internationalization machinery. USE_I18N = True # If you set this to False, Django will not format dates, numbers and # calendars according to the current locale USE_L10N = True # Absolute path to the directory that holds media. # Example: "/home/media/media.lawrence.com/" MEDIA_ROOT = j(j(PROJECT_DIR, ".."), "media") # URL that handles the media served from MEDIA_ROOT. Make sure to use a # trailing slash if there is a path component (optional in other cases). # Examples: "http://media.lawrence.com", "http://example.com/media/" MEDIA_URL = 'http://127.0.0.1:8080/media/' # URL prefix for admin media -- CSS, JavaScript and images. Make sure to use a # trailing slash. # Examples: "http://foo.com/media/", "/media/". ADMIN_MEDIA_PREFIX = '/media/' # Make this unique, and don't share it with anybody. SECRET_KEY = '%3g5=u^-vb-+xb&k$)zm(6z&n^^79k8m8f&5489z37hm9x3#w%' # List of callables that know how to import templates from various sources. TEMPLATE_LOADERS = ( 'django.template.loaders.filesystem.Loader', 'django.template.loaders.app_directories.Loader', # 'django.template.loaders.eggs.Loader', ) MIDDLEWARE_CLASSES = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', ) ROOT_URLCONF = 'mysite.urls' #TEMPLATE_DIRS = ( # Put strings here, like "/home/html/django_templates" or "C:/www/django/templates". # Always use forward slashes, even on Windows. # Don't forget to use absolute paths, not relative paths. # "/home/dsipos/prg/web/dj_tut/templates", # TEMPLATE_DIRS = ( j(j(PROJECT_DIR, ".."), "templates"), ) INSTALLED_APPS = ( 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.messages', # Uncomment the next line to enable the admin: 'django.contrib.admin', 'mysite.polls', 'mysite.books', 'mysite.aprozar', 'mysite.notes', 'mysite.xo' )
[ "dumitru.sipos@gmail.com" ]
dumitru.sipos@gmail.com
f281fed287dbd357fea0ab3bb3bd35efc0794cf4
51d65cbed3df1e9e3a0d51f79590ee12f88291d1
/object_detection/inference_over_image.py
0bbbdb9954ca69ffd0cf92de7a7cbb7577cf8043
[ "MIT" ]
permissive
apacha/Mensural-Detector
f9332c23854263c6a3f89e8b92f3f666f8377ed8
05c91204cf268feaae84cd079dbe7a1852fba216
refs/heads/master
2022-09-23T21:20:53.376367
2022-08-31T08:36:35
2022-08-31T08:36:35
137,372,669
12
6
null
null
null
null
UTF-8
Python
false
false
6,444
py
import numpy as np import tensorflow as tf import argparse from PIL import Image from object_detection.utils import ops as utils_ops, label_map_util, visualization_utils as vis_util if tf.__version__ < '1.4.0': raise ImportError('Please upgrade your tensorflow installation to v1.4.* or later!') def load_image_into_numpy_array(image): (im_width, im_height) = image.size return np.array(image.getdata()).reshape( (im_height, im_width, 3)).astype(np.uint8) def run_inference_for_single_image(image, graph): with graph.as_default(): with tf.Session() as sess: # Get handles to input and output tensors ops = tf.get_default_graph().get_operations() all_tensor_names = {output.name for op in ops for output in op.outputs} tensor_dict = {} for key in [ 'num_detections', 'detection_boxes', 'detection_scores', 'detection_classes', 'detection_masks' ]: tensor_name = key + ':0' if tensor_name in all_tensor_names: tensor_dict[key] = tf.get_default_graph().get_tensor_by_name(tensor_name) if 'detection_masks' in tensor_dict: # The following processing is only for single image detection_boxes = tf.squeeze(tensor_dict['detection_boxes'], [0]) detection_masks = tf.squeeze(tensor_dict['detection_masks'], [0]) # Reframe is required to translate mask from box coordinates to image coordinates and fit the image size. real_num_detection = tf.cast(tensor_dict['num_detections'][0], tf.int32) detection_boxes = tf.slice(detection_boxes, [0, 0], [real_num_detection, -1]) detection_masks = tf.slice(detection_masks, [0, 0, 0], [real_num_detection, -1, -1]) detection_masks_reframed = utils_ops.reframe_box_masks_to_image_masks( detection_masks, detection_boxes, image.shape[0], image.shape[1]) detection_masks_reframed = tf.cast(tf.greater(detection_masks_reframed, 0.5), tf.uint8) # Follow the convention by adding back the batch dimension tensor_dict['detection_masks'] = tf.expand_dims(detection_masks_reframed, 0) image_tensor = tf.get_default_graph().get_tensor_by_name('image_tensor:0') # Run inference output_dict = sess.run(tensor_dict, feed_dict={image_tensor: np.expand_dims(image, 0)}) # all outputs are float32 numpy arrays, so convert types as appropriate output_dict['num_detections'] = int(output_dict['num_detections'][0]) output_dict['detection_classes'] = output_dict['detection_classes'][0].astype(np.uint8) output_dict['detection_boxes'] = output_dict['detection_boxes'][0] output_dict['detection_scores'] = output_dict['detection_scores'][0] if 'detection_masks' in output_dict: output_dict['detection_masks'] = output_dict['detection_masks'][0] return output_dict def load_detection_graph(path_to_checkpoint): detection_graph = tf.Graph() with detection_graph.as_default(): od_graph_def = tf.GraphDef() with tf.gfile.GFile(path_to_checkpoint, 'rb') as fid: serialized_graph = fid.read() od_graph_def.ParseFromString(serialized_graph) tf.import_graph_def(od_graph_def, name='') return detection_graph def load_category_index(path_to_labels, number_of_classes): # Load label map label_map = label_map_util.load_labelmap(path_to_labels) categories = label_map_util.convert_label_map_to_categories(label_map, max_num_classes=number_of_classes, use_display_name=True) category_index = label_map_util.create_category_index(categories) return category_index if __name__ == "__main__": parser = argparse.ArgumentParser(description='Performs detection over input image given a trained detector.') parser.add_argument('--inference_graph', dest='inference_graph', type=str, required=True, help='Path to the frozen inference graph.') parser.add_argument('--label_map', dest='label_map', type=str, required=True, help='Path to the label map, which is json-file that maps each category name to a unique number.', default="mapping.txt") parser.add_argument('--number_of_classes', dest='number_of_classes', type=int, default=32, help='Number of classes.') parser.add_argument('--input_image', dest='input_image', type=str, required=True, help='Path to the input image.') parser.add_argument('--output_image', dest='output_image', type=str, default='detection.jpg', help='Path to the output image.') args = parser.parse_args() # Path to frozen detection graph. This is the actual model that is used for the object detection. # PATH_TO_CKPT = '/home/jcalvo/Escritorio/Current/Mensural Detector/mensural-detector/output_inference_graph.pb/frozen_inference_graph.pb' path_to_frozen_inference_graph = args.inference_graph path_to_labels = args.label_map number_of_classes = args.number_of_classes input_image = args.input_image output_image = args.output_image # Read frozen graph detection_graph = load_detection_graph(path_to_frozen_inference_graph) category_index = load_category_index(path_to_labels, number_of_classes) image = Image.open(input_image) # the array based representation of the image will be used later in order to prepare the # result image with boxes and labels on it. image_np = load_image_into_numpy_array(image) # Actual detection. output_dict = run_inference_for_single_image(image_np, detection_graph) # Visualization of the results of a detection. vis_util.visualize_boxes_and_labels_on_image_array( image_np, output_dict['detection_boxes'], output_dict['detection_classes'], output_dict['detection_scores'], category_index, instance_masks=output_dict.get('detection_masks'), use_normalized_coordinates=True, line_thickness=2) Image.fromarray(image_np).save(output_image)
[ "alexander.pacha@gmail.com" ]
alexander.pacha@gmail.com
90eb5fd3f16495104732f25945188ffbca0336ac
95d7484f512f2ef0b62a0da1feb900e436214c8b
/models/contact_template.py
1bcc94bd095d43236881fd97e45402ddc616740d
[]
no_license
Sundaya-Indo/product_test
d3b090384284b3efe449952b63f7face735a0ccf
e1d7a6dd40b224024df9955e67e50c0a09cf7b8b
refs/heads/master
2021-07-31T14:00:13.001861
2021-07-22T08:42:57
2021-07-22T08:42:57
248,193,227
0
0
null
null
null
null
UTF-8
Python
false
false
663
py
from odoo import models, fields, api, _ from odoo.exceptions import UserError class Partners(models.Model): _inherit = 'res.partner' file_datasheet_partner = fields.Many2many( comodel_name="ir.attachment", relation="m2m_ir_file_datasheet_partner", column1="m2m_id", column2="attachment_id", string="Datasheet File") weblinks_partner = fields.Many2many(comodel_name="ir.attachment", relation="m2m_ir_file_weblinks_partner", column1="m2m_id", column2="attachment_id", string="Weblinks") is_employee = fields.Boolean('Is an Employee', default=False)
[ "agis@sundaya.com" ]
agis@sundaya.com
3e993a5460ef71cfd268ce842d87012816743b71
f2656962b7bb2b0bb120718feb4b9f124b7ecc6c
/AssociationRuleMining/utils.py
27d5abd683c4cf625a574825a63f41d1380be05a
[]
no_license
Fasgort/AIA-Recomendacion
24c42634953efaef2b650320f83a7a17c8a5bc06
2340b6196552083c8b9ef75c2be245881e1bd9d1
refs/heads/master
2021-01-17T23:21:09.207808
2017-03-15T11:27:51
2017-03-15T11:27:51
84,218,017
0
0
null
null
null
null
UTF-8
Python
false
false
1,138
py
# -*- coding: utf-8 -*- # !/usr/bin/env python3 import logging from tabulate import tabulate from association_rule_tools import get_conviction, get_lift, get_confidence, get_support def load_transaction_db(path, limit=0): logging.debug("Start loading transactions database") transaction_db = list() loaded = 0 with open(path) as fd: for line in fd: transaction = frozenset([int(i_char) for i_char in line.split()]) transaction_db.append(transaction) loaded += 1 if 0 < limit <= loaded: break return transaction_db def association_rules_report(rules, transaction_db): res = list() for r in rules: s = get_support((r[0]+r[1]), transaction_db) c = get_confidence(r, transaction_db) l = get_lift(r, transaction_db) conv = get_conviction(r, transaction_db) res.append(((sorted(r[0]),r[1]), s, c, l, conv)) res = sorted(res, key=lambda x: (len(x[0][0]) * 100000000) + (x[0][0][0]), reverse=False) return tabulate(res, headers=['Rule X=>Y', 'Support', 'Confidence', 'Lift', 'Conviction'])
[ "valentin.sallop@gmail.com" ]
valentin.sallop@gmail.com
57bfefceefd25252047dcd608dff497f0c347b82
988dd821269be12c2f56f62b0c35546fd3050537
/python/quaternions/rotations.py
852c8839c1435519fcbc0675bd055c4d8af732b7
[]
no_license
gdiazh/adcs_models
fb19f541eeb9b01ae49ec98719c508d084e4fd7a
51d0829cc777d2e345e4fabe406ec7f54e661117
refs/heads/master
2020-03-28T13:04:56.174852
2018-09-28T22:08:25
2018-09-28T22:08:25
148,364,081
0
0
null
null
null
null
UTF-8
Python
false
false
3,050
py
#!/usr/bin/python __author__ = 'gdiaz' import matplotlib as mpl from plotVectors import PlotVectors import numpy as np class Rotation(object): def __init__(self): self.vectors = PlotVectors() self.a = [0, 0, 0] def rotate_z(self, a, yaw): Az = np.matrix([[np.cos(yaw), -np.sin(yaw), 0], [np.sin(yaw), np.cos(yaw), 0], [0, 0, 1]]) a_ = np.matrix([[a[0]], [a[1]], [a[2]]]) u = Az*a_ return [u.item(0), u.item(1), u.item(2)] def rotate_frame_z(self, I, J, K, yaw): Az = np.matrix([[np.cos(yaw), np.sin(yaw), 0], [-np.sin(yaw), np.cos(yaw), 0], [0, 0, 1]]) I_ = np.matrix([I[0], I[1], I[2]]) J_ = np.matrix([J[0], J[1], J[2]]) K_ = np.matrix([K[0], K[1], K[2]]) i_ = I_*Az j_ = J_*Az k_ = K_*Az i = [i_.item(0), i_.item(1), i_.item(2)] j = [j_.item(0), j_.item(1), j_.item(2)] k = [k_.item(0), k_.item(1), k_.item(2)] return [i, j, k] def vectorRotationTest(self): # Calcs p1 = [2, 0, 0] yaw = 90*np.pi/180 p1_rot = self.rotate_z(p1, yaw) print p1_rot # Plot self.vectors.plotAxes() self.vectors.config() self.vectors.plot(p1) self.vectors.plot(p1_rot) self.vectors.show() def frameRotationTest(self): # Calcs I = [1, 0, 0] J = [0, 1, 0] K = [0, 0, 1] yaw = 45*np.pi/180 ijk = self.rotate_frame_z(I, J, K, yaw) print ijk # Plot self.vectors.plotAxes() self.vectors.config() self.vectors.plot(ijk[0]) self.vectors.plot(ijk[1]) self.vectors.plot(ijk[2]) self.vectors.show() def get_qT(self, yawT): #Return quaternion target given yaw target AT = np.matrix([[np.cos(yawT), np.sin(yawT), 0], [-np.sin(yawT), np.cos(yawT), 0], [0, 0, 1]]) q4 = 0.5*np.sqrt(1+AT[0,0]+AT[1,1]+AT[2,2]) q1 = 0.25*(AT[1,2]-AT[2,1])/q4 q2 = 0.25*(AT[2,0]-AT[0,2])/q4 q3 = 0.25*(AT[0,1]-AT[1,0])/q4 return [q4, q1, q2, q3] def get_qE_(self, qT, qS): qT_ = np.matrix([[qT[0], qT[3], -qT[2], qT[1]], [-qT[3], qT[0], qT[1], qT[2]], [qT[2], -qT[1], qT[0], qT[3]], [-qT[1], -qT[2], -qT[3], qT[0]]]) qS_ = np.matrix([[-qS[1]], [-qS[2]], [-qS[3]], [qS[0]]]) qE = qT_*qS_ return [qE.item(0), qE.item(1), qE.item(2), qE.item(3)] def get_qE(self, yawT, qS): qT = self.get_qT(yawT) qE = self.get_qE_(qT, qS) return qE if __name__ == '__main__': rotation = Rotation() # Test Example # rotation.vectorRotationTest() rotation.frameRotationTest()
[ "g.hernan.diaz@gmail.com" ]
g.hernan.diaz@gmail.com
4314eb2e3669ee41547bd4d12cc4d8689c34d0aa
74b65dee638e73b07032b4d26a9e0ce7a50b7ccc
/neural_network/network.py
a8d781a35c4b16245ae5b9b31c4d9c88e34b61b6
[]
no_license
mpwillia/Tensorflow-Network-Experiments
6aec1d0a645d18536f0293185be553d67b584ad6
6d43f9a71c0b80a4d634de812e5141a8b295a4f8
refs/heads/master
2021-01-11T16:23:44.949529
2017-04-27T23:10:06
2017-04-27T23:10:06
80,074,966
0
0
null
null
null
null
UTF-8
Python
false
false
32,465
py
import tensorflow as tf import tensorflow.contrib as tfc import tensorflow.contrib.layers as tfcl print("Using Tensorflow Version: {}".format(tf.__version__)) import numpy as np import sys import math import random import os from functools import partial from network_util import match_tensor_shape, batch_dataset, get_num_batches, \ make_per_class_eval_tensor, print_eval_results, print_fit_results from summary import NetworkSummary from collections import namedtuple EvalResults = namedtuple('EvalResults', ['overall', 'per_class']) FitResults = namedtuple('FitResults', ['train', 'validation', 'test']) class Network(object): def __init__(self, input_shape, layers, pred_act_fn = None, logdir = None, network_name = 'network'): """ For |layers| see: https://www.tensorflow.org/api_docs/python/contrib.layers/higher_level_ops_for_building_neural_network_layers_ """ self.input_shape = input_shape self.layers = layers self.pred_act_fn = pred_act_fn self.network_name = network_name self.input_shape = input_shape self.sess = None self.saver = None self.train_step = None # setup global step counter self.global_step = tf.Variable(0, trainable = False, name = "net_global_step") # setup the network's summaries self.logdir = logdir self.network_summary = NetworkSummary(logdir, max_queue = 3, flush_secs = 60) # setup the network's input shape and input placeholder if type(input_shape) is int: self.net_input_shape = [None, input_shape] else: self.net_input_shape = (None,) + tuple(input_shape) with tf.name_scope('net_input'): self.net_input = tf.placeholder(tf.float32, shape = self.net_input_shape, name = "network_input_tensor") print("\nConstructing {} Layer Network".format(len(layers))) print(" {:35s} : {}".format("Input Shape", self.net_input.get_shape())) self.using_dropout = False self.keep_prob = tf.placeholder(tf.float32, shape=[], name = "dropout_keep_prob") # layer states are only applicable for recurrent layers self.layer_states = [] made_kernel_images = False prev_layer_output = self.net_input for layer_num, layer in enumerate(layers): layer_type = layer.func.__name__ layer_name = "layer_{:d}_{}".format(layer_num, layer_type) layer_kwargs = {'inputs' : prev_layer_output, 'scope' : layer_name} # handle dropout layers if 'dropout' in layer_type: self.using_dropout = True layer_kwargs['keep_prob'] = self.keep_prob with tf.name_scope(layer_name) as layer_scope: layer_output = layer(**layer_kwargs) try: # check if the layer is recurrent, if so extract the state if len(layer_output) == 2: prev_layer_output, state = layer_output else: prev_layer_output = layer_output[0] state = None except: prev_layer_output = layer_output state = None self.layer_states.append(state) self.network_summary.add_layer_summary(layer_name, prev_layer_output, layer_scope) layer_msg = "Layer {:d} ({}) Shape".format(layer_num, layer_type) print(" {:35s} : {}".format(layer_msg, prev_layer_output.get_shape())) print("") with tf.name_scope('net_output') as output_scope: self.net_output = prev_layer_output self.network_summary.add_output_summary(self.net_output, scope = output_scope) if self.pred_act_fn is not None: self.pred_net_output = self.pred_act_fn(prev_layer_output) else: self.pred_net_output = prev_layer_output self.network_summary.add_output_summary(self.pred_net_output, scope = output_scope) self.exp_output = tf.placeholder(tf.float32, self.net_output.get_shape(), name = "loss_expected_output") self.eval_net_output = tf.placeholder(tf.float32, self.net_output.get_shape(), name = "eval_net_output") # Various Network Getters ------------------------------------------------- def get_global_step(self): """ Returns the current global step if the network has an active session, otherwise returns None """ if self.sess is not None: return self.sess.run(self.global_step) def _get_weight_variables(self): vars = tf.trainable_variables() return [v for v in vars if 'weight' in v.name] # Session Handling -------------------------------------------------------- def init_session(self): """ Initializes the network's tensorflow session along with initializing all tensorflow variables. Will also create a new tensorflow Saver instance for the network if needed. If a session has already been created when this method is called, for example through loading a saved network, then all uninitialized variables will be initialized. If a session has not yet been created then: - A new Saver instance will be created - A new session will be created - All variables will be initialized If a session already exists (through loading a saved network) then: - All uninitialized variables will be initialized """ if self.sess is None: sess_config = tf.ConfigProto( log_device_placement = False, allow_soft_placement = False) self.saver = tf.train.Saver() self.sess = tf.Session(config = sess_config) self.sess.run(tf.global_variables_initializer()) else: list_of_variables = tf.global_variables() uninitialized_variables = list(tf.get_variable(name) for name in self.sess.run(tf.report_uninitialized_variables(list_of_variables))) self.sess.run(tf.initialize_variables(uninitialized_variables)) def close(self): """ Closes and deletes this Network's session. """ if self.sess is not None: self.sess.close() self.sess = None # Network Prediction ------------------------------------------------------ def predict(self, input_data, chunk_size = 500): """ Makes predictions on the given input data. Returns the index of the output with the highest value for each item in the input data. Arguments: |input_data| the input data to make predictions on. Optional: |chunk_size| the maximum number of items to process with one run of the network. All of the input data will be processed but it will be broken into the smaller chunks for better memory usage. If the chunk size is None then all of the input data will be processed in one run of the network. Returns the index of the output with the highest value for each item in the input. """ with self.sess.as_default(): feed_dict = dict() if self.using_dropout: feed_dict[self.keep_prob] = 1.0 results = [] for chunk_x in batch_dataset(input_data, chunk_size, has_outputs = False): feed_dict[self.net_input] = chunk_x results.extend(self.pred_net_output.eval(feed_dict=feed_dict)) return np.argmax(results, 1) def sample(self, input_data, temperature = 1.0, filter_zero = True, chunk_size = 500): """ Samples the networks response to the given input data. Returns a randomly selected output index based on the network's predicted probability of each possible output for each item in the input data. In otherwords the probability of selecting a given output index is given by the network's predicted probabilities. Arguments: |input_data| the input data to make predictions on. Optional: |temperature| the temperature value changes the distribution of the network's predicted probabilities for each output. Accepts any nonzero float value. Defaults to 1.0. A higher temperature value makes the resulting probability distribution more evenly spread while a lower temperature value makes the resulting probability distribution less evenly spread. There are four distinct effects that can be achieved with the temperature value. Temperature Effects (from low to high): temp=0.0 - Has the same effect as calling predict() where the output with the highest value will always be selected. temp<1.0 - Higher probabilities are increased further while lower probabilities are decreased further. Results in less randomness in the output. temp=1.0 - Has no effect on the network's predicted probabilities. temp>1.0 - Higher probabilities are made smaller while lower probabilities are made larger. Results in more randomness in the output. |filter_zero| if True then when applying temperature to the predicted probabilities, probabilities of zero will be filtered out. |chunk_size| the maximum number of items to process with one run of the network. All of the input data will be processed but it will be broken into the smaller chunks for better memory usage. If the chunk size is None then all of the input data will be processed in one run of the network. """ with self.sess.as_default(): feed_dict = dict() if self.using_dropout: feed_dict[self.keep_prob] = 1.0 results = [] for chunk_x in batch_dataset(input_data, chunk_size, has_outputs = False): feed_dict[self.net_input] = chunk_x results.extend(self.pred_net_output.eval(feed_dict=feed_dict)) results = np.asarray(results) if temperature <= 0.0: return np.argmax(results, 1) num_choices = results.shape[1] # (batch, outputs) if filter_zero: choices = np.arange(num_choices) def apply_temperature(results_1d): non_zero = np.nonzero(results_1d) nz_results = results_1d[non_zero] nz_choices = choices[non_zero] probs = np.exp(np.log(nz_results) / temperature) probs /= np.sum(probs) return np.random.choice(nz_choices, p = probs) return np.apply_along_axis(apply_temperature, 1, results) else: probs = np.exp(np.log(results) / temperature) probs /= np.sum(probs, 1) f = lambda p: np.random.choice(num_choices, p=p) return np.apply_along_axis(f, 1, probs) # Network Training -------------------------------------------------------- def fit(self, train_data, optimizer, loss, epochs, mb_size = None, evaluation_freq = None, evaluation_func = None, evaluation_fmt = None, evaluation_target = None, max_step = None, per_class_evaluation = False, validation_data = None, test_data = None, shuffle_freq = None, l1_reg_strength = 0.0, l2_reg_strength = 0.0, dropout_keep_prob = 1.0, summaries_per_epoch = None, save_checkpoints = False, checkpoint_freq = None, verbose = False): """ For |optimizer| see: https://www.tensorflow.org/api_docs/python/train/optimizers For |loss| see: https://www.tensorflow.org/api_docs/python/contrib.losses/other_functions_and_classes https://www.tensorflow.org/api_docs/python/nn/classification """ # reshape given data #train_data = self._reshape_dataset(train_data) #validation_data = self._reshape_dataset(validation_data) #test_data = self._reshape_dataset(test_data) train_feed_dict = dict() # handle dropout if self.using_dropout: train_feed_dict[self.keep_prob] = dropout_keep_prob if summaries_per_epoch <= 0: summaries_per_epoch = None self.network_summary.add_input_summary(self.net_input, mb_size) # setting up our loss tensor with tf.name_scope("loss") as loss_scope: grad_loss = loss(self.net_output, self.exp_output) # setup regularization if l1_reg_strength > 0.0 or l2_reg_strength > 0.0: l1_reg = None if l1_reg_strength > 0.0: l1_reg = tfcl.l1_regularizer(l1_reg_strength) l2_reg = None if l2_reg_strength > 0.0: l2_reg = tfcl.l2_regularizer(l2_reg_strength) l1_l2_reg = tfcl.sum_regularizer((l1_reg, l2_reg)) reg_penalty = tfcl.apply_regularization(l1_l2_reg, self._get_weight_variables()) loss_tensor = grad_loss + reg_penalty else: reg_penalty = None loss_tensor = grad_loss self.network_summary.add_loss_summary(loss_tensor, grad_loss, reg_penalty, loss_scope) # adds a summary for all trainable variables self.network_summary.add_variable_summary() # setting up our optimizer try: opt_name = optimizer.__class__.__name__ except: opt_name = 'optimizer' with tf.name_scope(opt_name): self.train_step = optimizer.minimize(loss_tensor, global_step = self.global_step) # setting up our evaluation function and summaries if evaluation_func is None: evaluation_func = loss with tf.name_scope('evaluation') as eval_scope: # overall eval tensor eval_tensor = evaluation_func(self.eval_net_output, self.exp_output) self.network_summary.add_eval_summary(eval_tensor, 'train', eval_scope) self.network_summary.add_eval_summary(eval_tensor, 'validation', eval_scope) self.network_summary.add_eval_summary(eval_tensor, 'test', eval_scope) # setting up our per class evaluation function and summaries with tf.name_scope('per_class_evaluation') as per_class_eval_scope: # per class eval tensor per_class_evaluation = False, per_class_eval_tensor = None if per_class_evaluation: per_class_eval_tensor = make_per_class_eval_tensor(evaluation_func, self.eval_net_output, self.exp_output, scope = per_class_eval_scope) def add_per_class_summary(name): self.network_summary.add_per_class_eval_summary(per_class_eval_tensor, max_val = 1.0, name = name, scope = per_class_eval_scope) add_per_class_summary('train') add_per_class_summary('validation') add_per_class_summary('test') # setting up the formating for printing the evaluation results if evaluation_fmt is None: evaluation_fmt = ".5f" # initialize our session self.init_session() # add a graph summary self.network_summary.add_graph(self.sess.graph) epoch_eval_results = [] initial_step = self.get_global_step() for epoch in range(epochs): # execute our training epoch epoch_msg = "Training Epoch {:4d} / {:4d}".format(epoch, epochs) self._run_training_epoch(train_data, mb_size, feed_dict_kwargs = train_feed_dict, summaries_per_epoch = summaries_per_epoch, verbose = True, verbose_prefix = epoch_msg) # check for mid-train evaluations if evaluation_freq is not None and epoch % evaluation_freq == 0: # evaluate on the training dataset if verbose > 1: print("\nMid-Train Evaluation") train_eval = self._evaluate(train_data, eval_tensor, per_class_eval_tensor, name = 'train') # evaluate on the validation dataset if validation_data is not None: validation_eval = self._evaluate(validation_data, eval_tensor, per_class_eval_tensor, name = 'validation') else: validation_eval = None # check if we've met our early stopping evaluation target if evaluation_target: if validation_eval is not None: met_target = validation_eval.overall >= evaluation_target else: met_target = train_eval.overall >= evaluation_target else: met_target = None # add the mid train evaluation results to our list epoch_fit_results = FitResults(train = train_eval, validation = validation_eval, test = None) epoch_eval_results.append(epoch_fit_results) # print the fit results if verbose > 1: print_fit_results(epoch_fit_results, evaluation_fmt) # break early if we've met our evaluation target if met_target is not None and met_target: print("\n\nReached Evaluation Target of {}".format(evaluation_target)) break # break early if we've met our step target if max_step is not None and self.get_global_step() >= max_step: print("\n\nReached Max Step Target of {}".format(max_step)) break if verbose > 1: print("") # save a checkpoint if needed if save_checkpoints and checkpoint_freq is not None and epoch % checkpoint_freq == 0: if verbose > 1: print("Saving Mid-Train Checkpoint") self._save_checkpoint() # shuffle the dataset if needed if shuffle_freq is not None and epoch % shuffle_freq == 0: train_data = self._shuffle_dataset(train_data) self.network_summary.flush() # report the number of training steps taken final_step = self.get_global_step() total_steps = final_step - initial_step if verbose > 0: print("\nTrained for {:d} Steps".format(total_steps)) # save the final checkpoint if save_checkpoints: if verbose > 1: print("Saving Final Checkpoint") self._save_checkpoint() if verbose == 1: print("") # Perform final fit result evaluations # final training evaluation if verbose > 1: print("Final Evaluation") train_eval = self._evaluate(train_data, eval_tensor, per_class_eval_tensor, name = 'train') # final validation evaluation if validation_data is not None: validation_eval = self._evaluate(validation_data, eval_tensor, per_class_eval_tensor, name = 'validation') else: validation_eval = None # final test evaluation if test_data is not None: test_eval = self._evaluate(test_data, eval_tensor, per_class_eval_tensor, name = 'test') else: test_eval = None # print and return the final fit results fit_results = FitResults(train = train_eval, validation = validation_eval, test = test_eval) if verbose > 1: print_fit_results(fit_results, evaluation_fmt) self.network_summary.flush() return fit_results # Single Network Training Epoch ------------------------------------------- def _run_training_epoch(self, train_data, mb_size = None, feed_dict_kwargs = dict(), summaries_per_epoch = None, verbose = False, verbose_prefix = None): """ Runs a single training epoch. Arguments: |train_data| the data to train on Optional: |mb_size| the size of the minibatches, if None then no minibatching will be done. |feed_dict_kwargs| any extra kwargs to be passed to the network. |summaries_per_epoch| how many summaries to produce for this epoch. The summaries will be evenly distributed across the minibatches. If None then no summaries will be made. |verbose| if True the progress information will be printed |verbose_prefix| extra information to append to the progress messages printed when the |verbose| argument is True. """ train_x, train_y = train_data mb_total = get_num_batches(len(train_x), mb_size) # Compute when to generate summaries if summaries_per_epoch is None: summary_every = None elif summaries_per_epoch >= mb_total: summary_every = 1 elif summaries_per_epoch == 1: summary_every = mb_total else: summary_every = int(math.ceil(mb_total / float(summaries_per_epoch))) with self.sess.as_default(): # Iterate over the batches for mb_x, mb_y, mb_num, mb_total in self._batch_for_train(train_data, mb_size, True): if verbose: # print progress message if verbose prefix = '' if verbose_prefix is not None: prefix = verbose_prefix + " " mb_msg = "Mini-Batch {:5d} / {:5d}".format(mb_num, mb_total) sys.stdout.write("{}{} \r".format(prefix, mb_msg)) sys.stdout.flush() feed_dict_kwargs[self.net_input] = mb_x feed_dict_kwargs[self.exp_output] = mb_y fetches = [self.train_step] # if this is a summary epoch then add those if (summary_every is not None) and (mb_num >= mb_total-1 or (mb_num+1) % summary_every == 0): train_summary = self.network_summary.get_training_summary() if train_summary is not None: fetches.extend([train_summary, self.global_step]) run_results = self.sess.run(fetches, feed_dict = feed_dict_kwargs) self._process_run_results(run_results) def _batch_for_train(self, dataset, batch_size, include_progress = False): """used to define batching specific to the training epochs""" return batch_dataset(dataset, batch_size, include_progress, True) # Network Performance Evaluation ------------------------------------------ def _evaluate(self, dataset, eval_tensor, per_class_eval_tensor = None, chunk_size = 2000, name = 'eval'): """ Evaluates the network's performance on the given dataset. Arguments: |dataset| the dataset to evaluate the network's performance on |eval_tensor| the tensor to use for evaluation. This tensor should accept the results of the network's predictions on the dataset and the expected outputs to produce a metric for how good the network's predictions are. For example, it could compute the network's accuracy. Optional: |per_class_eval_tensor| if the network is performing classification then this tensor can be used to evaluate the network's performance for each class individually. This tensor accepts the same inputs as the |eval_tensor| but is expected to produce a vector of metrics where each element is the metric for each class. |chunk_size| the maximum number of items to process with one run of the network. All of the input data will be processed but it will be broken into the smaller chunks for better memory usage. If the chunk size is None then all of the input data will be processed in one run of the network. |name| gives a name for the evaluation being performed. Used for grouping like summaries together. For example, can be used to group evaluation of validation data seperately from the evaluation of the testing data. Returns the evaluation results as an EvalResults tuple. """ eval_x, eval_y = dataset with self.sess.as_default(): feed_dict = dict() if self.using_dropout: feed_dict[self.keep_prob] = 1.0 results = [] for chunk_x, chunk_y, in self._batch_for_eval(dataset, chunk_size): feed_dict[self.net_input] = chunk_x results.extend(self.net_output.eval(feed_dict=feed_dict)) feed_dict = {self.eval_net_output : results, self.exp_output : eval_y} fetches = [eval_tensor] if per_class_eval_tensor is not None: fetches.append(per_class_eval_tensor) non_summary_size = len(fetches) eval_summary = self.network_summary.get_evaluation_summary(name) if eval_summary is not None: fetches.extend([eval_summary, self.global_step]) eval_results = self.sess.run(fetches, feed_dict = feed_dict) return self._process_eval_results(eval_results, non_summary_size) def _batch_for_eval(self, dataset, batch_size, include_progress = False): """used to define batching specific to the evaluation epochs""" return batch_dataset(dataset, batch_size, include_progress, True) # Result Handling (Training and Evaluation) ------------------------------- def _process_eval_results(self, eval_results, non_summary_size = 1): """ Takes the raw, unprocessed evaluation results and extracts out the relevant information as an EvalResults tuple. Arguments: |eval_results| the raw, unprocessed evaluation results Optional: |non_summary_size| the expected number of items not used for summaries Returns the processed EvalResults tuple """ eval_results = self._process_run_results(eval_results, non_summary_size) if len(eval_results) == 1: return EvalResults(overall = eval_results[0]) elif len(eval_results) == 2: return EvalResults(overall = eval_results[0], per_class = eval_results[1]) else: raise ValueError("Don't know how to process eval_results with length {:d}!".format(len(eval_results))) def _process_run_results(self, run_results, non_summary_size = 1): """ Takes the raw, unprocessed run results and extracts out the nonsummary information as a tuple. Arguments: |run_results| the raw, unprocessed run results Optional: |non_summary_size| the expected number of items not used for summaries Returns a tuple of the non-summary items from the run """ if len(run_results) == non_summary_size + 2: summary, step = run_results[-2:] self.network_summary.write(summary, step) elif len(run_results) != non_summary_size: raise ValueError("Don't know how to process run_results with length {:d}!".format(len(run_results))) return tuple(run_results[:non_summary_size]) # Dataset Utilities ------------------------------------------------------- def _reshape_dataset(self, dataset): if dataset is None: return None x,y = dataset return match_tensor_shape(x, self.net_input), \ match_tensor_shape(y, self.net_output) def _shuffle_dataset(self, dataset): zipped_dataset = zip(*dataset) random.shuffle(zipped_dataset) return list(zip(*zipped_dataset)) # Network Pickling and Variable I/O --------------------------------------- def __getstate__(self): odict = self.__dict__.copy() # Strip Tensorflow Content del odict['sess'] del odict['saver'] del odict['global_step'] del odict['train_step'] del odict['network_summary'] del odict['exp_output'] del odict['eval_net_output'] del odict['net_input'] del odict['net_output'] del odict['pred_net_output'] del odict['net_input_shape'] del odict['layer_states'] del odict['using_dropout'] del odict['keep_prob'] return odict def __setstate__(self, state): self.__init__(**state) def save_variables(self, path): """ Saves the networks tensorflow variable states to the given filepath Arguments: |path| the filepath to save the tensorflow variable states to """ if self.saver is None or self.sess is None: raise Exception("Cannot save variables without a session and saver") self.saver.save(self.sess, path) def load_variables(self, path): """ Loads the networks tensorflow variable states from the given filepath Arguments: |path| the filepath to load the tensorflow variable states from """ self.init_session() self.saver.restore(self.sess, path) def _save_checkpoint(self): if self.sess is None: raise Exception("Cannot save checkpoint without an active session!") if self.saver is None: raise Exception("Cannot save checkpoint without a tf.train.Saver instance!") if self.logdir is not None: save_path = os.path.join(self.logdir, self.network_name) else: save_path = os.path.join('./', self.network_name) self.saver.save(self.sess, save_path, self.global_step)
[ "mike@clwill.com" ]
mike@clwill.com
120ee9bc839ff0a3903105aacf109d63e4c539be
7774f3549007ea06046ff06abe85efb6433062b9
/textapp/forms.py
fe60215b1bf50eafe9943fed60819425e3743837
[]
no_license
rikuriku1999/textbook
92be520f064c74986c62437fe585db0d39216a61
49814eac2ed318c13882966a35bec60760084a2b
refs/heads/master
2021-01-01T12:39:22.769329
2020-03-15T13:18:42
2020-03-15T13:18:42
239,282,516
0
0
null
null
null
null
UTF-8
Python
false
false
6,025
py
from django import forms from . import models from django.contrib.auth import get_user_model from django.contrib.auth.forms import ( AuthenticationForm, UserCreationForm ) User=get_user_model() COLLEGE_CHOICES = ( ('慶應義塾大学','慶應義塾大学'), ('早稲田大学','早稲田大学'), ('青山学院大学','青山学院大学')) GENDER_CHOICES = ( ('男性','男性'), ('女性','女性')) STATUS_CHOICES = ( ('e','全て'), (False,'販売中のみ') ) QUALITY_CHOICES = ( ('','状態を選択'), ('きれい','きれい'), ('少し書き込みあり','少し書き込みあり'), ('かなり書き込みあり','かなり書き込みあり'), ('あまりきれいでない','あまりきれいでない') ) class CommentForm(forms.ModelForm): class Meta: model = models.Commentmodel fields = ('text',) text = forms.CharField( widget=forms.Textarea(), required=False, max_length=30, initial='' ) class UserForm(forms.Form): class Meta: model = models.Usermodel fields = ('username','college','gender','intro') username = forms.CharField( widget=forms.TextInput( attrs={'placeholder':'ユーザー名を入力(変更不可)'} ), required=True, max_length=20, ) college = forms.CharField( widget=forms.TextInput( attrs={'placeholder':'大学名、学部学科を入力'} ), #choices=COLLEGE_CHOICES, required=True, max_length=20, ) intro = forms.CharField( required=False, widget=forms.Textarea( attrs={'placeholder':'自己紹介文を入力'} ), max_length=1000 ) gender = forms.ChoiceField( widget=forms.Select, choices=GENDER_CHOICES ) class UserForm2(forms.Form): class Meta: model = models.Usermodel fields = ('college','intro') college = forms.CharField( widget=forms.TextInput(), #choices=COLLEGE_CHOICES, required=True, max_length=20, ) intro = forms.CharField( required=False, widget=forms.Textarea( attrs={'placeholder':'自己紹介文を入力'} ), max_length=1000 ) class DetailForm(forms.Form): class Meta: model = models.Textbookmodel fields = ('title','content','images','collegecategory','status','price','campus',) images = forms.ImageField( required=True, ) title = forms.CharField( required=True, widget=forms.TextInput( attrs={'placeholder':'例 経済学入門'} ), max_length=30, ) content = forms.CharField( max_length=200, required=True, widget=forms.Textarea( attrs={'placeholder':'例 経済学部の必修科目の教科書です。定価5000円です。テストに出そうな部分も書き込んであります。よろしくお願いします。'} ) ) collegecategory = forms.CharField( max_length=20, required=True, widget=forms.TextInput(), ) status = forms.ChoiceField( widget=forms.Select, choices=QUALITY_CHOICES, ) price = forms.IntegerField( required=True, widget=forms.NumberInput( attrs={'placeholder':'例 2000'} ) ) campus = forms.CharField( max_length=30, required=False, widget=forms.TextInput( attrs={'placeholder':'例 渋谷キャンパス'} ) ) class ChatForm(forms.ModelForm): class Meta: model = models.Chatmodel fields = ('text',) text = forms.CharField( initial='', max_length=32, required = True, widget = forms.Textarea( attrs={'placeholder':'メッセージを入力'} ) ) class SearchForm(forms.Form): search = forms.CharField( initial='', label='search', required = False, # 必須ではない widget=forms.TextInput( attrs={'placeholder':'キーワードを入力'} )) class SqueezeForm(forms.Form): title = forms.CharField( initial='', label='title', required = False, widget=forms.TextInput( attrs={'placeholder':'タイトルを入力'} ) ) college = forms.CharField( initial='', label='college', required = False, widget = forms.TextInput( attrs={'placeholder':'大学名等を入力'} ) ) price = forms.IntegerField( initial='', label='price', required = False, widget=forms.NumberInput( attrs={'placeholder':'数字のみ ~円以下'} ) ) sellstatus = forms.ChoiceField( label='sellstatus', widget = forms.Select, choices = STATUS_CHOICES ) class LoginForm(AuthenticationForm): """ログインフォーム""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) for field in self.fields.values(): field.widget.attrs['class'] = 'form-control' field.widget.attrs['placeholder'] = field.label # placeholderにフィールドのラベルを入れる class UserCreateForm(UserCreationForm): """ユーザー登録用フォーム""" class Meta: model = User fields = ('email',) def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) for field in self.fields.values(): field.widget.attrs['placeholder'] = field.label # placeholderにフィールドのラベルを入れる field.widget.attrs['class'] = 'form-control' def clean_email(self): email = self.cleaned_data['email'] User.objects.filter(email=email, is_active=False).delete() return email
[ "kalashnikova1120@gmail.com" ]
kalashnikova1120@gmail.com
f0f4aaf831a274f5022dbc1d1fa68fe08e28c63e
a2fd9491d11d9982d1ce82765b6dbed7653a954d
/Adpy/Lesson 1.5/Lesson 1.5.py
cefec4862090dd0924ccf7bd7ccd0636ff58ea19
[]
no_license
nikolaydmukha/Netology
bc3dc541c1de9329671daeacca385de3b55d9c48
b3577a4d414e76876412de01c2120e81ba82c697
refs/heads/master
2022-12-11T04:34:27.270476
2019-06-30T09:04:02
2019-06-30T09:04:02
167,036,184
1
0
null
2022-11-22T03:15:02
2019-01-22T17:23:40
Python
UTF-8
Python
false
false
1,943
py
import email import smtplib import imaplib from email.mime.text import MIMEText from email.mime.multipart import MIMEMultipart class Email: def __init__(self, sender, password, subject, recipients, message, header=None): self.sender = sender self.password = password self.subject = subject self.recipients = recipients self.message = message self.header = header self.gmail_smtp = "smtp.gmail.com" self.gmail_imap = "imap.gmail.com" # send message def send_message(self): msg = MIMEMultipart() msg['From'] = self.sender msg['To'] = ', '.join(self.recipients) msg['Subject'] = self.subject msg.attach(MIMEText(self.message)) ms = smtplib.SMTP(self.gmail_smtp, 587) # identify ourselves to smtp gmail client ms.ehlo() # secure our email with tls encryption ms.starttls() # re-identify ourselves as an encrypted connection ms.ehlo() ms.login(self.sender, self.password) ms.sendmail(self.sender, ms, msg.as_string()) ms.quit() # receive messages def receive(self): mail = imaplib.IMAP4_SSL(self.gmail_imap) mail.login(self.sender, self.password) mail.list() mail.select("inbox") criterion = '(HEADER Subject "%s")' % self.header if self.header else 'ALL' result, data = mail.uid('search', None, criterion) assert data[0], 'There are no letters with current header' latest_email_uid = data[0].split()[-1] result, data = mail.uid('fetch', latest_email_uid, '(RFC822)') raw_email = data[0][1] email_message = email.message_from_string(raw_email) mail.logout() if __name__ == '__main__': send_mail = Email('login@gmail.com', 'qwerty', 'Subject', ['vasya@email.com', 'petya@email.com'], 'Message', header='Refactoring')
[ "45710335+nikolaydmukha@users.noreply.github.com" ]
45710335+nikolaydmukha@users.noreply.github.com
480449f9654885b91bebc78cabcb012d20f26abb
6d0d27377707b4ac36d5d7858a590869e4eeff7c
/src/backend/utils/upload.py
ba670386065fdb7720aee57d3337ffee6efe620b
[]
no_license
Hiraishi-Ryota/mediado-hack-2021-a
6ce03f63a3b254c80e615248e85834df2af8718a
1c570a2d807c4531949639846220cbde15beada9
refs/heads/main
2023-08-06T21:10:57.354095
2021-09-15T05:50:54
2021-09-15T05:50:54
404,998,912
0
0
null
2021-09-15T05:50:55
2021-09-10T07:45:52
Python
UTF-8
Python
false
false
559
py
import os from pathlib import Path import shutil import sys from tempfile import NamedTemporaryFile, SpooledTemporaryFile BASE_DIR = os.getcwd() def upload(filename: str, file: SpooledTemporaryFile, dir: str): """e_pub_fileをstaticに保存し、そのBASEDIR下のパスを返す""" try: with NamedTemporaryFile(delete=False, suffix=Path(filename).suffix, dir=dir) as tmp: shutil.copyfileobj(file, tmp) e_pub_path = Path(tmp.name) finally: file.close() return str(e_pub_path.relative_to(BASE_DIR))
[ "leonard.t1028@gmail.com" ]
leonard.t1028@gmail.com
a717e726db9b4dae2b92ba552163a7eb6f742c6f
711ebd8c54be73934154d3f3b325bbc2a13a5fde
/weather app.py
6dc4cd8ffc05b6af1499c0d481df5b6bdd07ec51
[]
no_license
Balajiigor/python
3b5ed72c57002eec7d2eebc8a1f1f84277eb5da8
3e6817b2447c5b0265147870758d00e30dbbc238
refs/heads/using-json-in-python
2023-07-08T19:08:40.490641
2021-08-16T02:32:03
2021-08-16T02:32:03
382,550,766
0
0
null
2021-08-16T02:34:18
2021-07-03T07:11:36
Python
UTF-8
Python
false
false
3,192
py
import requests from bs4 import BeautifulSoup from tkinter import Tk from tkinter import Label from PIL import ImageTk, Image Tamil_nadu = "https://weather.com/en-IN/weather/today/l/4a5f6abb61cf684f3b18578ada1c5647346a0c273b8d5cd86c1eb48842d572e5" Seattle = "https://weather.com/en-IN/weather/today/l/ced0de18c1d771856e6012f3abf0a952cfe22952e72e516e6e098d54ca737114" Dubai = "https://weather.com/en-IN/weather/today/l/af60f113ba123ce93774fed531be2e1e51a1666be5d6012f129cfa27bae1ee6c" Paris = "https://weather.com/en-IN/weather/today/l/501361e097b79e8221d5c0b1447e80a0bf1c48b8fee1e4d98d4dad397ba2f204" Norway = "https://weather.com/en-IN/weather/today/l/cb003c6f366a3ae14b6a78ac5f2cfd18285fa02d15f892112dd00961afcb043b" Los_Angeles = "https://weather.com/en-IN/weather/today/l/0f4e045fdd139c3280846cf4eaae5b3f1c6ca58d13169016e6209f7b86872fc1" Arab = "https://weather.com/en-IN/weather/today/l/9eb72583100b2852c7d0da1a9f6d6d523dc38cfeb848d2ba82517e7f8bb44626" Moscow = "https://weather.com/en-IN/weather/today/l/34f2aafc84cff75ae0b014754856ea5e7f8ddf618cf9735549dfb5e016c28e10" master = Tk() master.title("Weather app") master.config(background = "black") img = Image.open("/home/balaji/Pictures/weather.png") img = img.resize((150, 150)) img = ImageTk.PhotoImage(img) def getWeather(): page = requests.get(Tamil_nadu) soup = BeautifulSoup(page.content, "html.parser") location =soup.find("h1", class_="CurrentConditions--location--2_osB").text time = soup.find("div",class_="CurrentConditions--timestamp--3_-CV" ).text temperature = soup.find("span",class_="CurrentConditions--tempValue--1RYJJ" ).text weatherPrediction = soup.find("div",class_="CurrentConditions--phraseValue--17s79" ).text alert = soup.find("div", class_="CurrentConditions--precipValue--1RgXi").text print(location) print(time) print(temperature) print(weatherPrediction) print(alert) locationLabel.config(text = location) timeLabel.config(text = time) temperaturLabel.config(text = temperature) weatherPredictionLabel.config(text = weatherPrediction) alertLabel.config(text = alert) timeLabel.after(60000, getWeather) temperaturLabel.after(60000, getWeather) weatherPredictionLabel.after(60000, getWeather) alertLabel.after(60000, getWeather) locationLabel = Label(master, font = ("calibri bold", 30), background = "black", foreground = "white") locationLabel.grid(row = 0, sticky="W", padx=40) timeLabel = Label(master, font =("calibri bold", 20), background = "black", foreground = "white") timeLabel.grid(row = 1, sticky = "W", padx = 40) temperaturLabel = Label(master, font=("calibri bold", 70), background="black", foreground = "white") temperaturLabel.grid(row =2, sticky = "W", padx = 40) Label(master, image = img, background = "black").grid(row = 2, sticky = "E") weatherPredictionLabel = Label(master, font= ("calibri bold", 40), background = "black", foreground ="white") weatherPredictionLabel.grid(row = 3, sticky = "W", padx=40) alertLabel = Label(master, font = ("calibri bold", 15), background = "black", foreground = "white") alertLabel.grid(row= 4, sticky ="W", padx = 40) getWeather() master.mainloop()
[ "noreply@github.com" ]
Balajiigor.noreply@github.com
e173a1acb6f004419a36a21a69349963e12720f5
1f1de940fd030db12ece5a4037fc1b9291f884cf
/src/main/python/config.py
d14218a38f34ce6ddfe81273e18ef2c5b271faf0
[]
no_license
psy2013GitHub/sklearn-utils
1ea747827f7a36332a049388b4abc03c43501d94
acea4bf3423883dd8e6782741234c6493648c820
refs/heads/master
2020-06-12T10:52:47.160022
2016-12-05T06:42:12
2016-12-05T06:42:12
75,586,943
0
0
null
null
null
null
UTF-8
Python
false
false
24
py
__author__ = 'flappy'
[ "deng.zhou@immomo.com" ]
deng.zhou@immomo.com
01e5eebcd277d1c73463b5afec695c16d394fa57
b5aea34ce585e4462775a838c27135f92c5a852d
/portfolio/migrations/0001_initial.py
152480a5681f5374338900b09a932854832e0688
[]
no_license
theethaj/portfolio-project
80851eaf4ed61eec8861a98398d0bbcad5223eb7
a954d3b1d7f1ec17eefb8c8e67e7ce801dc5f201
refs/heads/main
2023-02-25T12:16:45.619964
2021-01-31T10:59:04
2021-01-31T10:59:04
null
0
0
null
null
null
null
UTF-8
Python
false
false
600
py
# Generated by Django 3.1.5 on 2021-01-30 13:15 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Contact', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('email', models.CharField(max_length=200)), ('subject', models.CharField(max_length=200)), ('message', models.TextField()), ], ), ]
[ "ding533@hotmail.com" ]
ding533@hotmail.com
5f064814d535825806233dee2469180a93dec6bb
7df7eb32424b40fa98378af298716759ae39d198
/Laboratorios/Lab8/Lab8Ejercicio1.py
3a0679cfc04af8dd79cbb5820216ec63516cbcbc
[]
no_license
tratohecho3/Algoritmos-1
289b2ee990da943d7a1622af04d815839882937a
b009a780de7e56dccf4d7832010b06d0254a35c1
refs/heads/master
2021-01-12T00:48:36.848983
2017-01-07T20:19:24
2017-01-07T20:19:24
78,298,851
0
0
null
null
null
null
UTF-8
Python
false
false
5,170
py
"""Autores: Cesar Colina 13-10299 Francisco Marquez 12-11163 Lab8Ejercicio.py Descripcion: Programa que lee una matriz en un archivo .txt y devuelve en otro archivo .txt una serie de datos de interes. """ #CALCULOS import pygame,sys, os.path from pygame.locals import * pygame.init() while True: tiempo = pygame.time.get_ticks() / 1000 if tiempo > 360: sys.exit() pygame.quit() x = int(input("Introduzca el valor de la fila donde se encuentra su numero: ")) y = int(input("introduzca el valor de la columna donde se encuentra su numero: ")) with open("matrix-entry.txt") as f: lineas = f.readlines() f.closed matriz = [] for i in range(len(lineas)): matriz.append(lineas[i].split()) for i in range(len(matriz)): for j in range(len(matriz[i])): matriz[i][j] = int(matriz[i][j]) Diag_ppal = [] for i in range(len(lineas)): for j in range(len(matriz[i])): if i == j: if matriz[i][j] > 0 and matriz[i][j]% 2 == 0: Diag_ppal.append(matriz[i][j]) Diag_sec = [] for i in range(len(lineas)): for j in range(len(matriz[i])): if i + j == 5: if matriz[i][j] < 0 and matriz[i][j]% 2 != 0: Diag_sec.append(matriz[i][j]) minimos = [] for i in range(len(lineas)): for j in matriz[i]: if j % 2 != 0: minimos.append(j) valor = 1000 for i in minimos: if i % 2 != 0: if i < valor: valor = i for i in range(len(lineas)): for j in range(len(matriz[i])): if matriz[i][j] == valor: #RESULTADO3 posicion_impar_menor = [i, j] maximos = [] for i in range(len(lineas)): for j in matriz[i]: if j % 2 == 0: maximos.append(j) valor2 = max(maximos) for i in range(len(lineas)): for j in range(len(matriz[i])): if matriz[i][j] == valor2: #RESULTADO3 posicion_par_mayor = [i, j] columna0 = [] columna1 = [] columna2 = [] columna3 = [] columna4 = [] columna5 = [] for i in range(len(matriz)): columna0.append(matriz[i][0]) for i in range(len(matriz)): columna1.append(matriz[i][1]) for i in range(len(matriz)): columna2.append(matriz[i][2]) for i in range(len(matriz)): columna3.append(matriz[i][3]) for i in range(len(matriz)): columna4.append(matriz[i][4]) for i in range(len(matriz)): columna5.append(matriz[i][5]) area1 = [] for i in range(3): for j in range(3): area1.append(matriz[i][j]) area2 = [] for i in range(3): for j in range(3, 6): area2.append(matriz[i][j]) area3 = [] for i in range(3, 6): for j in range(3): area3.append(matriz[i][j]) area4 = [] for i in range(3, 6): for j in range(3, 6): area4.append(matriz[i][j]) total = (sum(area1) + sum(area2) + sum(area3) + sum(area4))/ 4 promedio = sum(Diag_ppal) / len(Diag_ppal) sumatoria = sum(Diag_sec) with open("report-output.txt", "w") as f: f.write("El promedio de los valores pares positivos de la diagonal principal es\n") f.write(str(promedio) + "\n") f.write("La suma de los valores impares negativos de la diagonal secundaria es \n") f.write(str(sumatoria) + "\n") f.write("El valor del impar menor de la matriz es" + "\n") f.write(str(valor) + "\n") f.write("La posicion, o posiciones, del impar menor es, o son" + "\n") for i in range(len(lineas)): for j in range(len(matriz[i])): if matriz[i][j] == valor: #RESULTADO3 f.write("La fila" + "\n") f.write(str(i) + "\n") f.write("La columna" + "\n") f.write(str(j) + "\n") f.write("El valor del par mayor de la matriz es" + "\n") f.write(str(valor2) + "\n") f.write("La posicion, o posiciones, del par mayor es, o son" + "\n") for i in range(len(lineas)): for j in range(len(matriz[i])): if matriz[i][j] == valor2: #RESULTADO3 f.write("La fila" + "\n") f.write(str(i) + "\n") f.write("La columna" + "\n") f.write(str(j) + "\n") f.write("El promedio de los valores de cada fila es" + "\n") for i in range(len(matriz)): #RESULTADO4 f.write("El promedio de la fila " + str(i) + " es" + "\n") f.write(str(sum(matriz[i]) / len(matriz[i])) + "\n") f.write("El promedio de los valores de cada columna es" + "\n") f.write("Para la columna 0" + "\n") f.write(str(sum(columna0) / len(columna0)) + "\n") f.write("Para la columna 1" + "\n") f.write(str(sum(columna1) / len(columna1)) + "\n") f.write("Para la columna 2" + "\n") f.write(str(sum(columna2) / len(columna2)) + "\n") f.write("Para la columna 3" + "\n") f.write(str(sum(columna3) / len(columna3)) + "\n") f.write("Para la columna 4" + "\n") f.write(str(sum(columna4) / len(columna4)) + "\n") f.write("Para la columna 5" + "\n") f.write(str(sum(columna5) / len(columna5)) + "\n") f.write("La suma de los valores de cada una de las subregiones de la matriz es" + "\n") f.write("Para la subregion A" + "\n") f.write(str(sum(area1)) + "\n") f.write("Para la subregion B" + "\n") f.write(str(sum(area2)) + "\n") f.write("Para la subregion C" + "\n") f.write(str(sum(area3)) + "\n") f.write("Para la subregion D" + "\n") f.write(str(sum(area4)) + "\n") f.write("El promedio de la suma total de las subregiones es" + "\n") f.write(str(total)) f.closed print(matriz[x][y])
[ "noreply@github.com" ]
tratohecho3.noreply@github.com
1466dcdda5a60e77319ce48fd9043140a50877ca
1c6d58e5b2bbce4a457350302fa9845f43f076a2
/Python Programming/EXC 030 - 7-0-0.py
28a0e6b50ec8b2ba0c20dba9c68e67ff8b2174f4
[]
no_license
Daviswww/Toys
fa8b481bf5106a0f984c6bfd5260f3ec55ccee1d
680c260ebb8d385a3dbcdd985a447fd5d2b74f3b
refs/heads/master
2022-07-21T03:35:41.590248
2020-01-11T11:04:02
2020-01-11T11:04:02
144,127,014
0
0
null
2022-06-22T00:05:19
2018-08-09T08:57:15
Python
UTF-8
Python
false
false
1,693
py
A = [89, 56, 92 ,79, 51] B = [70, 86, 77, 83, '缺考'] C = [0, 0, 0, 0, 0] D = [0, 0, 0, 0, 0] print('期中考缺考: ') for i in range(5): if(A[i] =='缺考'): print(i,'號缺考') print('\n期末考缺考: ') for i in range(5): if(B[i] =='缺考'): print(i,'號缺考') print('\n平均分數: ') for i in range(5): if(A[i] == '缺考' or B[i] == '缺考'): if(A[i] == '缺考'): C[i] = B[i] / 2 elif(B[i] == '缺考'): C[i] = A[i] / 2 else: C[i] = 0 print('第',i,'號\t期中考:', A[i], '\t期末考:',B[i], '\t平均:', C[i]) else: C[i] = (A[i] + B[i]) / 2 if(C[i] > 100): C[i] = 100 print('第',i,'號\t期中考:', A[i], '\t期末考:',B[i], '\t平均:', C[i]) print(C.index(max(C[0:5])),'號學期成績最高分',max(C[0:5])) print(C.index(min(C[0:5])),'號學期成績最低分',min(C[0:5])) print('\n條分後平均分數: ') for i in range(5): if(A[i] == '缺考' or B[i] == '缺考'): if(A[i] == '缺考'): D[i] = ((B[i]*1.5)) / (1 + 1.5) elif(B[i] == '缺考'): D[i] = ((A[i])) / (1 + 1.5) else: D[i] = 0 print('第',i,'號\t期中考:', A[i], '\t期末考:',B[i], '\t平均:', D[i]) else: D[i] = (A[i] + (B[i]*1.5)) / (1 + 1.5) if(D[i] > 100): D[i] = 100 print('第',i,'號\t期中考:', A[i], '\t期末考:',B[i], '\t平均:', D[i]) print(D.index(max(D[0:5])),'號學期成績最高分',max(D[0:5])) print(D.index(min(D[0:5])),'號學期成績最低分',min(D[0:5]))
[ "noreply@github.com" ]
Daviswww.noreply@github.com
524db47926d6c1b18a65735cec61aad5f9e91b97
d2c163f246d28b8519f8c89de23556e43be91684
/www/ad_board/urls.py
9309b9dfb201f43c13a2ec3d393148de00aea612
[]
no_license
boogiiieee/Iskcon
d7a2b8bdc3002ef3306fc5e7ddc577504d8533c9
b672dbafee06af3ee6d646c75f442d97133f5ec9
refs/heads/master
2021-09-04T03:11:06.770094
2018-01-15T04:21:36
2018-01-15T04:21:36
null
0
0
null
null
null
null
UTF-8
Python
false
false
388
py
# -*- coding: utf-8 -*- from django.conf.urls.defaults import patterns, include, url urlpatterns = patterns('ad_board.views', url(r'^$', 'full', name='ad_board_url'), url(r'^category/(?P<id>[0-9]+)/$', 'category', name='category_ad_board_url'), url(r'^(?P<id>[0-9]+)/$', 'item', name='ad_board_item_url'), url(r'^category/(?P<id>[0-9]+)/add/$', 'add', name='add_ad_board_url'), )
[ "shalyapinalexander@gmail.com" ]
shalyapinalexander@gmail.com
35b57a408d049fe970e3d7bb1fcf28c9e89d7f4c
faa3c49ce63590c298ffcd5ecc4c4b1808efb5db
/docker-images/docker-madminer-all/code/configurate.py
d540ebcfa111bf9b88f45e96d77090405b44974c
[ "MIT" ]
permissive
johannbrehmer/workflow-madminer
bf4063e3793db2f0340ed4c0fe4e13052d7f6d06
bb648503bc5b6df301dea7708cc05fb567a4be57
refs/heads/master
2020-04-28T12:11:37.970977
2020-04-09T19:13:04
2020-04-09T19:13:04
175,268,087
0
0
MIT
2019-03-12T17:52:33
2019-03-12T17:52:33
null
UTF-8
Python
false
false
2,821
py
from __future__ import absolute_import, division, print_function, unicode_literals import numpy as np #import matplotlib #from matplotlib import pyplot as plt #%matplotlib inline import sys import yaml import inspect from madminer.core import MadMiner from madminer.plotting import plot_2d_morphing_basis from madminer.sampling import combine_and_shuffle from madminer.sampling import SampleAugmenter from madminer.sampling import benchmark, benchmarks from madminer.sampling import morphing_point, morphing_points, random_morphing_points mg_dir = '/home/software/MG5_aMC_v2_6_2' miner = MadMiner()#(debug=False) input_file = str(sys.argv[1]) print('inputfile: ',input_file) ########### ADD parameters and benchmarks from input file with open(input_file) as f: # use safe_load instead load dict_all = yaml.safe_load(f) #get default values of miner.add_parameters() default_arr = inspect.getargspec(miner.add_parameter) default = dict(zip(reversed(default_arr.args), reversed(default_arr.defaults))) #ADD PARAMETERS for parameter in dict_all['parameters']: #format range_input to tuple range_input = parameter['parameter_range'] range_tuple = map(float, range_input.replace('(','').replace(')','').split(',')) miner.add_parameter( lha_block=parameter['lha_block'], #required lha_id=parameter['lha_id'], #required parameter_name=parameter.get('parameter_name', default['parameter_name']), #optional morphing_max_power=int( parameter.get('morphing_max_power', default['morphing_max_power']) ), #optional param_card_transform=parameter.get('param_card_transform',default['param_card_transform']), #optional parameter_range=range_tuple #optional ) n_parameters = len(dict_all['parameters']) #ADD BENCHMARKS for benchmark in dict_all['benchmarks']: dict_of_parameters_this_benchmark = dict() for i in range(1, n_parameters+1): try: #add to the dictionary: key is parameter name, value is value dict_of_parameters_this_benchmark[ benchmark['parameter_name_'+str(i)] ] = float(benchmark['value_'+str(i)]) except KeyError as e: print('Number of benchmark parameters does not match number of global parameters in input file') raise e #add miner.add_benchmark( dict_of_parameters_this_benchmark, benchmark['name'] ) ########### #SET morphing settings = dict_all['set_morphing'] miner.set_morphing( include_existing_benchmarks=True, max_overall_power=int(settings['max_overall_power']) ) #fig = plot_2d_morphing_basis( # miner.morpher, # xlabel=r'$c_{W} v^2 / \Lambda^2$', # ylabel=r'$c_{\tilde{W}} v^2 / \Lambda^2$', # xrange=(-10.,10.), # yrange=(-10.,10.) #) miner.save('/home/data/madminer_example.h5')
[ "iem244@nyu.edu" ]
iem244@nyu.edu
1a94d4955bc1347ae86d5992a523abcfbfb17267
5da2c116d3d0dc4f3811cec144c9f8b5a74afede
/lncrawl/assets/user_agents.py
fbec17aabe02c7b79f52106cf5ee397fca225e17
[ "Apache-2.0" ]
permissive
NNTin/lightnovel-crawler
a08bd252f2e72f41f931f0b2165f906b64d33692
451e816ab03c8466be90f6f0b3eaa52d799140ce
refs/heads/master
2021-06-23T12:07:43.668329
2021-04-25T01:51:26
2021-04-25T01:51:26
361,695,538
2
0
Apache-2.0
2021-04-26T16:48:21
2021-04-26T09:40:46
null
UTF-8
Python
false
false
6,302
py
# -*- coding: utf-8 -*- user_agents = [ # "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:66.0) Gecko/20100101 Firefox/66.0", # "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:15.0) Gecko/20100101 Firefox/15.0.1", # "Mozilla/5.0 (X11; CrOS x86_64 8172.45.0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/51.0.2704.64 Safari/537.36", # "Mozilla/5.0 (Linux; Android 8.0.0; SM-G960F Build/R16NW) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/62.0.3202.84 Mobile Safari/537.36", # "Mozilla/5.0 (Linux; Android 6.0.1; Nexus 6P Build/MMB29P) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.83 Mobile Safari/537.36", # "Mozilla/5.0 (Linux; Android 6.0; HTC One M9 Build/MRA58K) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.98 Mobile Safari/537.3", # "Mozilla/5.0 (Linux; Android 7.0; Pixel C Build/NRD90M; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/52.0.2743.98 Safari/537.36", # "Mozilla/5.0 (Linux; Android 6.0.1; SHIELD Tablet K1 Build/MRA58K; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/55.0.2883.91 Safari/537.36", # "Mozilla/5.0 (iPhone; CPU iPhone OS 12_0 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/12.0 Mobile/15E148 Safari/604.1", # "Mozilla/5.0 (iPhone; CPU iPhone OS 12_0 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) CriOS/69.0.3497.105 Mobile/15E148 Safari/605.1", # "Mozilla/5.0 (iPhone; CPU iPhone OS 12_0 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/13.2b11866 Mobile/16A366 Safari/605.1.15", # "Mozilla/5.0 (iPhone; CPU iPhone OS 11_0 like Mac OS X) AppleWebKit/604.1.38 (KHTML, like Gecko) Version/11.0 Mobile/15A372 Safari/604.1", # "Mozilla/5.0 (iPhone; CPU iPhone OS 11_0 like Mac OS X) AppleWebKit/604.1.34 (KHTML, like Gecko) Version/11.0 Mobile/15A5341f Safari/604.1", # "Mozilla/5.0 (iPhone; CPU iPhone OS 11_0 like Mac OS X) AppleWebKit/604.1.38 (KHTML, like Gecko) Version/11.0 Mobile/15A5370a Safari/604.1", # "Mozilla/5.0 (iPhone9,3; U; CPU iPhone OS 10_0_1 like Mac OS X) AppleWebKit/602.1.50 (KHTML, like Gecko) Version/10.0 Mobile/14A403 Safari/602.1", # "Mozilla/5.0 (Windows Phone 10.0; Android 6.0.1; Microsoft; RM-1152) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Mobile Safari/537.36 Edge/15.15254", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/79.0.3945.117 Safari/537.36", "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.103 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.78 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.79 Safari/537.36" ]
[ "dipu.sudipta@gmail.com" ]
dipu.sudipta@gmail.com
da41b6d51ae8d2de2dd7a45e0120555d35750c8d
f0187406babf1be73626fa2a4fbeb790e177dd7a
/assignment2/assignment2_2015004120.py
708e78808230d9427a30fe1383f4dfb056f5b6ec
[]
no_license
CameliaOvO/CSE4007
bc150084f00f3ab484a8194002022cc96d169414
89929cb6f5c61b9f89de06f14a2a03f2a43e5378
refs/heads/master
2020-03-19T03:32:53.572535
2018-05-29T04:38:28
2018-05-29T04:38:28
135,737,859
1
0
null
null
null
null
UTF-8
Python
false
false
6,084
py
from bisect import bisect_left from collections import Counter from math import log2 def complete_link_clustering(sim_name): sim = cosine_similarity if sim_name == 'c' else euclidean_distance most = min if sim == euclidean_distance else max levels, clusters = [], [[x] for x in range(num_of_words)] prox_mat = [[sim(vectors[i], vectors[j]) for j in range(num_of_words) if i > j] for i in range(num_of_words)][1:] while len(clusters) > 1: most_sim = most(enumerate([(i.index(most(i)), most(i)) for i in prox_mat]), key=lambda x: x[1][1]) r, s = clusters[most_sim[0] + 1], clusters[most_sim[1][0]] levels.append([find_least_sim(r, s, sim), r + s]) clusters.remove(r), clusters.remove(s), clusters.append(r + s) del prox_mat[most_sim[0]] if most_sim[1][0] > 0: del prox_mat[most_sim[1][0] - 1] it = 0 while it < len(prox_mat): if len(prox_mat[it]) > most_sim[0] + 1: del prox_mat[it][most_sim[0] + 1] if len(prox_mat[it]) > most_sim[1][0]: del prox_mat[it][most_sim[1][0]] if len(prox_mat[it]) == 0: prox_mat.remove(prox_mat[it]) else: it += 1 prox_mat.append([find_least_sim(r + s, clusters[t], sim) for t in range(len(clusters) - 1)]) return levels def cosine_similarity(a, b): return sum([x * y for x, y in zip(a, b)]) / ((sum([x ** 2 for x in a]) ** 0.5) * (sum([x ** 2 for x in b]) ** 0.5)) def euclidean_distance(x, y): return sum([(xk - yk) ** 2 for xk, yk in zip(x, y)]) ** 0.5 def find_least_sim(c1, c2, sim): least = max if sim == euclidean_distance else min return least([sim(vectors[p1], vectors[p2]) for p1 in c1 for p2 in c2]) def normalize(level): max_val = max(level, key=lambda x: x[0])[0] min_val = min(level, key=lambda x: x[0])[0] for l in level: l[0] = 1 - ((l[0] - min_val) / (max_val - min_val)) return level def get_words_vectors(): word_list, vector_list = [], [] with open("WordEmbedding.txt", 'r') as f: for word, vector in zip(*[f] * 2): word_list.append(word.strip()) vector_list.append(list(map(float, vector.split(",")))) return word_list, vector_list, len(vector_list) def divide_cluster(levels, threshold): cluster_idx, cluster_num = 0, [0 for _ in range(num_of_words)] limit = bisect_left([x[0] for x in levels], threshold) levels = levels[limit:] for level in levels: cluster_idx += 1 flag = False for c in level[1]: if cluster_num[c] == 0: flag = True cluster_num[c] = cluster_idx if not flag: cluster_idx -= 1 for i in range(len(cluster_num)): if cluster_num[i] == 0: cluster_idx += 1 cluster_num[i] = cluster_idx return cluster_idx, cluster_num def write_on_file(result): write_word, write_vector = [], [] with open("WordEmbedding.txt", 'r') as rf: for word, vector in zip(*[rf] * 2): write_word.append(word.strip()) write_vector.append(vector.strip()) with open("WordClustering.txt", 'w') as wf: for word, vector, cluster in zip(write_word, write_vector, result): wf.write(word + "\n" + vector + "\n" + str(cluster) + "\n") def get_word_class(): with open("WordTopic.txt", 'r') as f: whole = [x.strip().lower() for x in f.readlines()] word_topic, topic, word_cls = [], [], [] for word in whole: if (not word.isalnum()) and topic != []: word_topic.append(topic) topic = [] if word.isalnum(): topic.append(word) word_topic.append(topic) for word in words: for cls in word_topic: if word in cls: word_cls.append(word_topic.index(cls)) break return word_cls def entropy_measure(n_clusters, c_list, word_cls): clustered = [[] for _ in range(n_clusters)] for i in range(len(c_list)): clustered[c_list[i] - 1].append(word_cls[i]) counter_clustered = [[x[1] for x in Counter(clusters).items()] for clusters in clustered] cluster_entropy = [sum([-(x / sum(lis)) * log2(x / sum(lis)) for x in lis]) for lis in counter_clustered] cluster_size = [len(cluster) / num_of_words for cluster in clustered] weighted_sum = sum([x * y for x, y in zip(cluster_size, cluster_entropy)]) return weighted_sum def silhouette_measure(n_clusters, c_list): silhouette_list = [] dist_mat = [[euclidean_distance(x, y) for x in vectors] for y in vectors] clustered_idx = [[] for _ in range(n_clusters)] for i in range(len(c_list)): clustered_idx[c_list[i] - 1].append(i) for i in range(num_of_words): inner_cluster_idx = c_list[i] - 1 if len(clustered_idx[inner_cluster_idx]) == 1: silhouette_list.append(0) else: c_i = clustered_idx[inner_cluster_idx] a_i = sum([dist_mat[i][cx] for cx in c_i if cx != i]) / (len(c_i) - 1) b_i = min([sum([dist_mat[i][cx] for cx in c]) / len(c) for c in clustered_idx if c != c_i]) silhouette_coef = (b_i - a_i) / max([a_i, b_i]) silhouette_list.append(silhouette_coef) sil_measure = sum(silhouette_list) / len(silhouette_list) return sil_measure argument = ['e', 0.6] words, vectors, num_of_words = get_words_vectors() word_class = get_word_class() level_cluster = complete_link_clustering(argument[0])[::-1] if argument[0] != 'c': level_cluster = normalize(level_cluster) num_of_clusters, clustered_list = divide_cluster(level_cluster, argument[1]) write_on_file(clustered_list) print('cosine similarity' if argument[0] == 'c' else 'euclidean distance') print("divided into", num_of_clusters, "clusters with threshold", argument[1]) print("entropy : \t", entropy_measure(num_of_clusters, clustered_list, word_class)) print("silhouette : \t", silhouette_measure(num_of_clusters, clustered_list))
[ "camelia0858@gmail.com" ]
camelia0858@gmail.com
57cbb17eae32ce8daed7bf554a568c0f8d9328db
36e13e0219419b6a0c9d913b99b9330c7894f32a
/LifelongMixture_64_Dirichlet.py
5c20b33432794fcd54c1b387c2aa5778c3182873
[]
no_license
WN1695173791/LifelongMixtureVAEs
4ef6f5c62f3a9480bd010fabce249020cca71b5b
b1f858cae35f8f0b91981f398ec431d9a8afb061
refs/heads/main
2023-06-15T10:09:31.087605
2021-07-09T15:11:53
2021-07-09T15:11:53
null
0
0
null
null
null
null
UTF-8
Python
false
false
23,976
py
import tensorflow as tf import mnist_data import tensorflow.contrib.slim as slim import time import seaborn as sns from Assign_Dataset import * from tensorflow.examples.tutorials.mnist import input_data from keras.datasets import mnist from Support import * from Mnist_DataHandle import * from HSICSupport import * from scipy.misc import imsave as ims from utils import * from glob import glob import keras from keras.datasets import reuters from keras.models import Sequential from keras.layers import Dense, Dropout, Activation from keras.preprocessing.text import Tokenizer import os, gzip from data_hand import * os.environ['CUDA_VISIBLE_DEVICES']='7' distributions = tf.distributions from Mixture_Models import * import keras.datasets.cifar10 as cifar10 def file_name2_(file_dir): t1 = [] for root, dirs, files in os.walk(file_dir): for a1 in dirs: b1 = "C:/CommonData//rendered_chairs/" + a1 + "/renders/*.png" img_path = glob(b1) t1.append(img_path) cc = [] for i in range(len(t1)): a1 = t1[i] for p1 in a1: cc.append(p1) return cc def file_name_(file_dir): t1 = [] file_dir = "E:/LifelongMixtureModel/data/images_background/" for root, dirs, files in os.walk(file_dir): for a1 in dirs: b1 = "E:/LifelongMixtureModel/data/images_background/" + a1 + "/renders/*.png" b1 = "E:/LifelongMixtureModel/data/images_background/" + a1 for root2, dirs2, files2 in os.walk(b1): for c1 in dirs2: b2 = b1 + "/" + c1 + "/*.png" img_path = glob(b2) t1.append(img_path) cc = [] for i in range(len(t1)): a1 = t1[i] for p1 in a1: cc.append(p1) return cc def file_name(file_dir): t1 = [] file_dir = "../images_background/" for root, dirs, files in os.walk(file_dir): for a1 in dirs: b1 = "../images_background/" + a1 + "/renders/*.png" b1 = "../images_background/" + a1 for root2, dirs2, files2 in os.walk(b1): for c1 in dirs2: b2 = b1 + "/" + c1 + "/*.png" img_path = glob(b2) t1.append(img_path) cc = [] for i in range(len(t1)): a1 = t1[i] for p1 in a1: cc.append(p1) return cc def file_name2(file_dir): t1 = [] for root, dirs, files in os.walk(file_dir): for a1 in dirs: b1 = "../rendered_chairs/" + a1 + "/renders/*.png" img_path = glob(b1) t1.append(img_path) print('root_dir:', root) # 当前目录路径 print('sub_dirs:', dirs) # 当前路径下所有子目录 print('files:', files) # 当前路径下所有非目录子文件 cc = [] for i in range(len(t1)): a1 = t1[i] for p1 in a1: cc.append(p1) return cc # Gateway def autoencoder(x_hat, x, dim_img, dim_z, n_hidden, keep_prob, task_state, disentangledCount): # encoding mu1, sigma1 = Encoder_64(x_hat, "encoder1") mu2, sigma2 = Encoder_64(x_hat, "encoder2") mu3, sigma3 = Encoder_64(x_hat, "encoder3") mu4, sigma4 = Encoder_64(x_hat, "encoder4") z1 = mu1 + sigma1 * tf.random_normal(tf.shape(mu1), 0, 1, dtype=tf.float32) z2 = mu2 + sigma2 * tf.random_normal(tf.shape(mu2), 0, 1, dtype=tf.float32) z3 = mu3 + sigma3 * tf.random_normal(tf.shape(mu3), 0, 1, dtype=tf.float32) z4 = mu4 + sigma4 * tf.random_normal(tf.shape(mu4), 0, 1, dtype=tf.float32) s1 = Generator_64(z1, "decoder1") s2 = Generator_64(z2, "decoder2") s3 = Generator_64(z3, "decoder3") s4 = Generator_64(z4, "decoder4") imageSize = 64 s1_1 = tf.reshape(s1,(-1,imageSize*imageSize*3))*task_state[:, 0:1] s2_1 = tf.reshape(s2,(-1,imageSize*imageSize*3))*task_state[:, 1:2] s3_1 = tf.reshape(s3,(-1,imageSize*imageSize*3))*task_state[:, 2:3] s4_1 = tf.reshape(s4,(-1,imageSize*imageSize*3))*task_state[:, 3:4] reco = s1_1 + s2_1 + s3_1 + s4_1 reco = reco / (task_state[0, 0] + task_state[0, 1] + task_state[0, 2] + task_state[0, 3]) reco = tf.reshape(reco,(-1,imageSize,imageSize,3)) #Calculate task relationship # Select tasks reco1 = tf.reduce_mean(tf.reduce_sum(tf.square(s1 - x_hat), [1, 2, 3])) reco2 = tf.reduce_mean(tf.reduce_sum(tf.square(s2 - x_hat), [1, 2, 3])) reco3 = tf.reduce_mean(tf.reduce_sum(tf.square(s3 - x_hat), [1, 2, 3])) reco4 = tf.reduce_mean(tf.reduce_sum(tf.square(s4 - x_hat), [1, 2, 3])) reco1_ = reco1 + (1 - task_state[0, 0]) * 1000000 reco2_ = reco2 + (1 - task_state[0, 1]) * 1000000 reco3_ = reco3 + (1 - task_state[0, 2]) * 1000000 reco4_ = reco4 + (1 - task_state[0, 3]) * 1000000 totalScore = tf.stack((reco1_, reco2_, reco3_, reco4_), axis=0) mixParameter = task_state[0] sum = mixParameter[0] + mixParameter[1] + mixParameter[2] + mixParameter[3] mixParameter = mixParameter / sum dist = tf.distributions.Dirichlet(mixParameter) mix_samples = dist.sample() b1 = mix_samples[0] * task_state[0, 0] b2 = mix_samples[1] * task_state[0, 1] b3 = mix_samples[2] * task_state[0, 2] b4 = mix_samples[3] * task_state[0, 3] mix_samples2 = tf.stack((b1,b2,b3,b4),axis=0) # loss reco1_loss = reco1 * mix_samples2[0] reco2_loss = reco2 * mix_samples2[1] reco3_loss = reco3 * mix_samples2[2] reco4_loss = reco4 * mix_samples2[3] # loss marginal_likelihood = (reco1_loss + reco2_loss + reco3_loss + reco4_loss) k1 = 0.5 * tf.reduce_sum( tf.square(mu1) + tf.square(sigma1) - tf.log(1e-8 + tf.square(sigma1)) - 1, 1) k2 = 0.5 * tf.reduce_sum( tf.square(mu2) + tf.square(sigma2) - tf.log(1e-8 + tf.square(sigma2)) - 1, 1) k3 = 0.5 * tf.reduce_sum( tf.square(mu3) + tf.square(sigma3) - tf.log(1e-8 + tf.square(sigma3)) - 1, 1) k4 = 0.5 * tf.reduce_sum( tf.square(mu4) + tf.square(sigma4) - tf.log(1e-8 + tf.square(sigma4)) - 1, 1) k1 = tf.reduce_mean(k1) k2 = tf.reduce_mean(k2) k3 = tf.reduce_mean(k3) k4 = tf.reduce_mean(k4) KL_divergence = k1 * mix_samples2[0] + k2 * mix_samples2[1] + k3 * mix_samples2[2] + k4 * mix_samples2[3] KL_divergence = KL_divergence p2 = 1 gamma = 4 loss = marginal_likelihood + gamma * tf.abs(KL_divergence - disentangledCount) z = z1 y = reco return y, z, loss, -marginal_likelihood, KL_divergence,totalScore def decoder(z, dim_img, n_hidden): y = bernoulli_MLP_decoder(z, n_hidden, dim_img, 1.0, reuse=True) return y n_hidden = 500 IMAGE_SIZE_MNIST = 28 dim_img = IMAGE_SIZE_MNIST ** 2 # number of pixels for a MNIST image dim_z = 256 # train n_epochs = 100 batch_size = 64 learn_rate = 0.001 train_total_data, train_size, _, _, test_data, test_labels = mnist_data.prepare_MNIST_data() n_samples = train_size # input placeholders # In denoising-autoencoder, x_hat == x + noise, otherwise x_hat == x x_hat = tf.placeholder(tf.float32, shape=[64, 64, 64, 3], name='input_img') x = tf.placeholder(tf.float32, shape=[64, 64, 64, 3], name='target_img') # dropout keep_prob = tf.placeholder(tf.float32, name='keep_prob') # input for PMLR z_in = tf.placeholder(tf.float32, shape=[None, dim_z], name='latent_variable') task_state = tf.placeholder(tf.float32, shape=[64, 4]) disentangledCount = tf.placeholder(tf.float32) # network architecture y, z, loss, neg_marginal_likelihood, KL_divergence,totalScore = autoencoder(x_hat, x, dim_img, dim_z, n_hidden, keep_prob, task_state, disentangledCount) # optimization t_vars = tf.trainable_variables() train_op = tf.train.AdamOptimizer(learn_rate).minimize(loss, var_list=t_vars) # train total_batch = int(n_samples / batch_size) min_tot_loss = 1e99 ADD_NOISE = False train_data2_ = train_total_data[:, :-mnist_data.NUM_LABELS] train_y = train_total_data[:, 784:784 + mnist_data.NUM_LABELS] # MNIST dataset load datasets img_path = glob('../img_celeba2/*.jpg') # 获取新文件夹下所有图片 data_files = img_path data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches celebaFiles = data_files # load 3D chairs img_path = glob('../CACD2000/CACD2000/*.jpg') # 获取新文件夹下所有图片 data_files = img_path data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches cacdFiles = data_files file_dir = "../rendered_chairs/" files = file_name2(file_dir) data_files = files data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches chairFiles = data_files files = file_name(1) data_files = files data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches zimuFiles = data_files saver = tf.train.Saver() isWeight = False currentTask = 4 def max_list(lt): temp = 0 for i in lt: if lt.count(i) > temp: max_str = i temp = lt.count(i) return max_str isWeight = False with tf.Session() as sess: sess.run(tf.global_variables_initializer(), feed_dict={keep_prob: 0.9}) if isWeight: saver.restore(sess, 'models/LifelongMixture_64_Dirichlet') img_path = glob('C:/CommonData/img_celeba2/*.jpg') # 获取新文件夹下所有图片 data_files = img_path data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches myIndex = 10 celebaFiles = data_files[myIndex * batch_size:(myIndex + 2) * batch_size] # load 3D chairs img_path = glob('C:/CommonData/CACD2000/CACD2000/*.jpg') # 获取新文件夹下所有图片 data_files = img_path data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches cacdFiles = data_files[myIndex * batch_size:(myIndex + 2) * batch_size] file_dir = "C:/CommonData/rendered_chairs/" files = file_name2_(file_dir) data_files = files data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches chairFiles = data_files[myIndex * batch_size:(myIndex + 2) * batch_size] files = file_name_(1) data_files = files data_files = sorted(data_files) data_files = np.array(data_files) # for tl.iterate.minibatches zimuFiles = data_files[myIndex * batch_size:(myIndex + 2) * batch_size] dataArray = [] for taskIndex in range(4): taskIndex = 2 if taskIndex == 0: x_train = celebaFiles x_fixed = x_train[0:batch_size] x_fixed2 = x_train[batch_size:batch_size * 2] elif taskIndex == 1: x_train = cacdFiles x_fixed = x_train[0:batch_size] x_fixed2 = x_train[batch_size:batch_size * 2] elif taskIndex == 2: x_train = chairFiles x_fixed = x_train[0:batch_size] x_fixed2 = x_train[batch_size:batch_size * 2] elif taskIndex == 3: x_train = zimuFiles x_fixed = x_train[0:batch_size] x_fixed2 = x_train[batch_size:batch_size * 2] batchFiles = x_fixed batchFiles2 = x_fixed2 if taskIndex == 0: batch = [get_image( sample_file, input_height=128, input_width=128, resize_height=64, resize_width=64, crop=True) for sample_file in batchFiles] batch2 = [get_image( sample_file, input_height=128, input_width=128, resize_height=64, resize_width=64, crop=True) for sample_file in batchFiles2] elif taskIndex == 1: batch = [get_image( sample_file, input_height=250, input_width=250, resize_height=64, resize_width=64, crop=True) for sample_file in batchFiles] batch2 = [get_image( sample_file, input_height=250, input_width=250, resize_height=64, resize_width=64, crop=True) for sample_file in batchFiles2] elif taskIndex == 2: image_size = 64 batch = [get_image2(batch_file, 300, is_crop=True, resize_w=image_size, is_grayscale=0) \ for batch_file in batchFiles] batch2 = [get_image2(batch_file, 300, is_crop=True, resize_w=image_size, is_grayscale=0) \ for batch_file in batchFiles2] elif taskIndex == 3: batch = [get_image(batch_file, 105, 105, resize_height=64, resize_width=64, crop=False, grayscale=False) \ for batch_file in batchFiles] batch = np.array(batch) batch = np.reshape(batch, (64, 64, 64, 1)) batch = np.concatenate((batch, batch, batch), axis=-1) batch2 = [get_image(batch_file, 105, 105, resize_height=64, resize_width=64, crop=False, grayscale=False) \ for batch_file in batchFiles2] batch2 = np.array(batch2) batch2 = np.reshape(batch2, (64, 64, 64, 1)) batch2 = np.concatenate((batch2, batch2, batch2), axis=-1) dataArray.append(batch) x_fixed = batch x_fixed = np.array(x_fixed) x_fixed2 = batch2 x_fixed2 = np.array(x_fixed2) # select the most relevant component stateState = np.zeros((batch_size, 4)) stateState[:, 0] = 1 stateState[:, 1] = 1 stateState[:, 2] = 1 stateState[:, 3] = 1 score = sess.run(totalScore, feed_dict={x_hat: x_fixed, keep_prob: 1, task_state: stateState}) a = np.argmin(score, axis=0) index = a z = 0 generator_outputs = 0 if index == 0: mu1, sigma1 = Encoder_64(x_hat, "encoder1", reuse=True) z1 = mu1 + sigma1 * tf.random_normal(tf.shape(mu1), 0, 1, dtype=tf.float32) Reco = Generator_64(z1, "decoder1", reuse=True) generator_outputs = Generator_64(z_in, "decoder1", reuse=True) z = z1 elif index == 1: mu2, sigma2 = Encoder_64(x_hat, "encoder2", reuse=True) z2 = mu2 + sigma2 * tf.random_normal(tf.shape(mu2), 0, 1, dtype=tf.float32) Reco = Generator_64(z2, "decoder2", reuse=True) generator_outputs = Generator_64(z_in, "decoder2", reuse=True) z = z2 elif index == 2: mu3, sigma3 = Encoder_64(x_hat, "encoder3", reuse=True) z3 = mu3 + sigma3 * tf.random_normal(tf.shape(mu3), 0, 1, dtype=tf.float32) Reco = Generator_64(z3, "decoder3", reuse=True) generator_outputs = Generator_64(z_in, "decoder3", reuse=True) z = z3 elif index == 3: mu4, sigma4 = Encoder_64(x_hat, "encoder4", reuse=True) z4 = mu4 + sigma4 * tf.random_normal(tf.shape(mu4), 0, 1, dtype=tf.float32) Reco = Generator_64(z4, "decoder4", reuse=True) generator_outputs = Generator_64(z_in, "decoder4", reuse=True) z = z4 code1 = sess.run(z, feed_dict={x_hat: x_fixed, keep_prob: 1, task_state: stateState}) code2 = sess.run(z, feed_dict={x_hat: x_fixed2, keep_prob: 1, task_state: stateState}) recoArr = [] minV = -3 maxV = 3 tv = 6.0 / 12.0 ''' for j in range(256): code1 = sess.run(z, feed_dict={x_hat: x_fixed, keep_prob: 1, task_state: stateState}) recoArr = [] myIndex = 0 for i in range(12): code1[:, j] = minV + tv * i myReco = sess.run(generator_outputs, feed_dict={z_in: code1, keep_prob: 1, task_state: stateState}) recoArr.append(myReco[myIndex]) recoArr = np.array(recoArr) ims("results/" + "inter" + str(j) + ".png", merge2(recoArr, [1, 12])) bc = 2 BC =0 ''' for t1 in range(64): for j in range(256): code1 = sess.run(z, feed_dict={x_hat: x_fixed, keep_prob: 1, task_state: stateState}) recoArr = [] j = 224 myIndex = t1 for i in range(12): code1[:,j] = minV + tv * i myReco = sess.run(generator_outputs, feed_dict={z_in: code1, keep_prob: 1, task_state: stateState}) recoArr.append(myReco[myIndex]) recoArr = np.array(recoArr) ims("results/" + "inter" + str(t1) + ".png", merge2(recoArr, [1, 12])) bc = 2 break c=0 for t in range(2): if t ==1 : t = t+20 recoArr.append(x_fixed2[t]) for i in range(10): newCode = code2 + distance*i myReco = sess.run(generator_outputs, feed_dict={z_in: newCode, keep_prob: 1, task_state: stateState}) recoArr.append(myReco[t]) recoArr.append(x_fixed[t]) recoArr = np.array(recoArr) ims("results/" + "inter" + str(taskIndex) + ".png", merge2(recoArr, [2, 12])) myReco = sess.run(Reco, feed_dict={x_hat: x_fixed, keep_prob: 1, task_state: stateState}) ims("results/" + "Dataset" + str(taskIndex) + "_mini.png", merge2(x_fixed[:16], [2, 8])) ims("results/" + "Reco" + str(taskIndex) + "_H_mini.png", merge2(myReco[:16], [2, 8])) bc = 0 bc = 0 # training n_epochs = 20 stateState = np.zeros((batch_size, 4)) stateState[:, 0] = 1 stateState[:, 1] = 1 stateState[:, 2] = 1 stateState[:, 3] = 1 disentangledScore = 0.5 vChange = 25.0 / n_epochs for taskIndex in range(currentTask): taskIndex = 1 if taskIndex == 0: x_train = celebaFiles x_fixed = x_train[0:batch_size] elif taskIndex == 1: x_train = cacdFiles x_fixed = x_train[0:batch_size] elif taskIndex == 2: x_train = chairFiles x_fixed = x_train[0:batch_size] elif taskIndex == 3: x_train = zimuFiles x_fixed = x_train[0:batch_size] disentangledScore = disentangledScore + vChange n_samples = np.shape(np.array(x_train))[0] total_batch = int(n_samples / batch_size) for epoch in range(n_epochs): # Random shuffling index = [i for i in range(np.shape(x_train)[0])] random.shuffle(index) x_train = x_train[index] image_size = 64 # Loop over all batches for i in range(total_batch): batchFiles = x_train[i * batch_size:i * batch_size + batch_size] if taskIndex == 0: batch = [get_image( sample_file, input_height=128, input_width=128, resize_height=64, resize_width=64, crop=True) for sample_file in batchFiles] elif taskIndex == 1: batch = [get_image( sample_file, input_height=250, input_width=250, resize_height=64, resize_width=64, crop=True) for sample_file in batchFiles] elif taskIndex == 2: batch = [get_image2(batch_file, 300, is_crop=True, resize_w=image_size, is_grayscale=0) \ for batch_file in batchFiles] elif taskIndex == 3: batch = [get_image(batch_file, 105, 105, resize_height=64, resize_width=64, crop=False, grayscale=False) \ for batch_file in batchFiles] batch = np.array(batch) batch = np.reshape(batch, (64, 64, 64, 1)) batch = np.concatenate((batch, batch, batch), axis=-1) # Compute the offset of the current minibatch in the data. batch_xs_target = batch x_fixed = batch batch_xs_input = batch if ADD_NOISE: batch_xs_input = batch_xs_input * np.random.randint(2, size=batch_xs_input.shape) batch_xs_input += np.random.randint(2, size=batch_xs_input.shape) _, tot_loss, loss_likelihood, loss_divergence = sess.run( (train_op, loss, neg_marginal_likelihood, KL_divergence), feed_dict={x_hat: batch_xs_input, x: batch_xs_target, keep_prob: 1.0, task_state: stateState,disentangledCount:disentangledScore}) print("epoch %f: L_tot %03.2f L_likelihood %03.2f L_divergence %03.2f" % ( epoch, tot_loss, loss_likelihood, loss_divergence)) y_PRR = sess.run(y, feed_dict={x_hat: x_fixed, keep_prob: 1,task_state:stateState,disentangledCount:disentangledScore}) y_RPR = np.reshape(y_PRR, (-1, 64, 64,3)) ims("results/" + "VAE" + str(epoch) + ".jpg", merge2(y_RPR[:64], [8, 8])) if epoch > 0: x_fixed_image = np.reshape(x_fixed, (-1, 64, 64,3)) ims("results/" + "Real" + str(epoch) + ".jpg", merge2(x_fixed_image[:64], [8, 8])) #select the most relevant component score = sess.run(totalScore, feed_dict={x_hat: x_fixed, keep_prob: 1, task_state: stateState}) a = np.argmin(score, axis=0) index = a if index == 0: stateState[:, 0:1] = 0 elif index == 1: stateState[:, 1:2] = 0 elif index == 2: stateState[:, 2:3] = 0 elif index == 3: stateState[:, 3:4] = 0 saver.save(sess, 'models/LifelongMixture_64_Dirichlet')
[ "noreply@github.com" ]
WN1695173791.noreply@github.com
de57cedbc86dec255b93ebc77daf153a873f5256
1422a57e98aba02321b772d72f8f0ada6d8b8cba
/friday/friday-vendor/vendor-scripts/test-resources/scripts/pylib/hue_turn_on_light.py
152b15f1a6ee7c7306946bab089ea4f1578d9421
[ "MIT" ]
permissive
JonasRSV/Friday
e1908a411aa133bc5bd2f383b0a995f7e028092d
f959eff95ba7b11525f97099c8f5ea0e325face7
refs/heads/main
2023-05-15T03:33:21.542621
2021-06-12T10:34:50
2021-06-12T10:34:50
315,309,991
7
2
null
null
null
null
UTF-8
Python
false
false
196
py
import phue import sys if __name__ == "__main__": b = phue.Bridge(config_file_path="credentials.json") b.set_light(int(sys.argv[1]), parameter={"on": True, "bri": 200}, transitiontime=5)
[ "jonas@valfridsson.net" ]
jonas@valfridsson.net
198442838c9414d3f62f9b0af071a325589a66ae
8840b69e4341f4ed030c8b33151db205b8db3640
/flask_minijax.py
a5036e1c916ae910ed2af7e28ecdc01b86534110
[ "MIT" ]
permissive
FidgetYou/proj3-anagrams
b5fe7ccc333bca0895c12590142b9f0e30f10b83
86923a696794b7098940023d57aaef679a52b3ac
refs/heads/master
2021-01-11T01:03:32.507679
2016-10-18T01:58:25
2016-10-18T01:58:25
70,846,302
0
0
null
2016-10-13T20:39:51
2016-10-13T20:39:50
null
UTF-8
Python
false
false
1,317
py
""" Tiny demo of Ajax interaction """ import flask from flask import request # Data from a submitted form from flask import url_for from flask import jsonify # For AJAX transactions import json import logging import argparse # For the vocabulary list import sys ### # Globals ### app = flask.Flask(__name__) import CONFIG app.secret_key = CONFIG.secret_key # Should allow using session variables ### # Pages ### @app.route("/") def index(): return flask.render_template('minijax.html') ############### # AJAX request handlers # These return JSON to the JavaScript function on # an existing page, rather than rendering a new page. ############### @app.route("/_countem") def countem(): text = request.args.get("text", type=str) length = len(text) rslt = { "long_enough": length >= 5 } return jsonify(result=rslt) ############# # Run locally if __name__ == "__main__": # Standalone. app.debug = True app.logger.setLevel(logging.DEBUG) print("Opening for global access on port {}".format(CONFIG.PORT)) app.run(port=CONFIG.PORT, host="0.0.0.0") # If we run 'python3 flask_minijax.py, we get the above 'main'. # If we run 'gunicorn flask_minijax:app', we instead get a # 'main' inside gunicorn, which loads this file as a module # and accesses the Flask 'app' object. #
[ "michal.young@gmail.com" ]
michal.young@gmail.com
24f57a7e04375267f91a3c4ad24656b5ee311e60
ae0feb6bf0c3851ea1fcb57476a4012cdd09f0c9
/listMap.py
2fa46e48fd3ee50e5766208659d71aa385432771
[]
no_license
beyondzhou/algorithms
f79e4832c53c528d972475aba5ab48bd33feefbb
a0f30d9a8efb51af7f58ad75b8a9f1ebb084d99f
refs/heads/master
2021-01-10T14:12:12.322067
2015-11-08T22:40:44
2015-11-08T22:40:44
44,276,856
0
0
null
null
null
null
UTF-8
Python
false
false
2,372
py
class MyMap: # init def __init__(self): self._entryList = list() # length def __len__(self): return len(self._entryList) # Contain def __contains__(self, key): ndx = self._findPosition(key) return ndx is not None # Add def add(self, key, value): ndx = self._findPosition(key) if ndx is not None: self._entryList[ndx].value = value return False else: entry = _MapEntry(key, value) self._entryList.append(entry) return True # Remove def remove(self, key): ndx = self._findPosition(key) assert ndx is not None, "key is not in the map." self._entryList.pop(ndx) # valueOf def valueOf(self, key): ndx = self._findPosition(key) assert ndx is not None, "key is not in the map." return self._entryList[ndx].value # find position def _findPosition(self, key): for i in range(len(self)): if self._entryList[i].key == key: return i return None # iter def __iter__(self): return _MapIterator(self._entryList) # Map storage class _MapEntry: # init def __init__(self, key, value): self.key = key self.value = value # Map Iterator class _MapIterator: # init def __init__(self, entryList): self._entryList = entryList self._ndx = 0 def __iter__(self): return self def next(self): if self._ndx < len(self._entryList): entry = self._entryList[self._ndx].key self._ndx += 1 return entry else: raise StopIteration # Test Map Function def testMyMap(): # init mapEntryList = MyMap() # Add some key,value pair mapEntryList.add('tim', 100) mapEntryList.add('dog', 10) mapEntryList.add('cat', 1) # Print the length print len(mapEntryList) # Print all the item for i in mapEntryList: print i, print '' # In check print 'dog' in mapEntryList print 'cat1' in mapEntryList # Remove mapEntryList.remove('dog') # Print all the item for i in mapEntryList: print i, print '' # Print the value print mapEntryList.valueOf('cat') if __name__ == "__main__": testMyMap()
[ "guaguastd@gmail.com" ]
guaguastd@gmail.com
ec02d70df62b2b0336e9a3155848509a1793fc6c
d31432d77d775bde32fe51e7584d68cfc465808a
/Tema_6_3_Patrones_organizacion_datos/jerarquico.py
e8eec174864fe48b3867697cc2ba65fad1b241e7
[]
no_license
surtich/TFG_Manuel_R
ac8d4685afeb7180d8beb3169372a1e27e484eea
ec1fda88695dd95d1add70d4869ea1e6e623313e
refs/heads/main
2023-07-02T06:45:19.506802
2021-08-03T16:55:32
2021-08-03T16:55:32
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,676
py
#!/usr/bin/env python #Usamos el archivo foros.csv from mrjob.job import MRJob from mrjob.protocol import RawValueProtocol import xmlify class jerarquico(MRJob): OUTPUT_PROTOCOL = RawValueProtocol def mapper(self,_, line): linea=line.split(";") mensaje=linea[4] # Recogemos el mensaje de la posición 4 de la línea tipoMensaje=linea[5] #Recogemos de la posición 5, si es una pregunta o respuesta if tipoMensaje=="question": idMensaje=linea[0] #Almacenamos el id único del mensaje yield idMensaje,(tipoMensaje,mensaje) else: idMensaje=linea[7] #Almacenamos el identificador del mensaje idMensaje yield idMensaje,(tipoMensaje,mensaje) def reducer(self, key, values): listaValores=[] listaPrincipal=[] listaAuxiliar=[] for v in values: #Metemos los valores que vienen en un matriz listaValores.append(v) #Matriz que contiene el tipo de mensaje y el mensaje asociado for valor in listaValores: if valor[0]=="question":#Si es una pregunta la metemos en la lista principal listaPrincipal.append(valor[1]) else: listaAuxiliar.append(valor[1]) # Si son respuestas, las vamos agregando a una lista listaPrincipal.append(listaAuxiliar) #agregamos la lista de respuestas a la lista principal #Conversion a XML indicando en el raiz el id del mensaje yield "Creada linea XML: " ,xmlify.dumps(listaPrincipal,root = key) if __name__ == '__main__': jerarquico.run()
[ "mrodrigue212@alumno.uned.es" ]
mrodrigue212@alumno.uned.es
6d346848a2eed9d5be67fdb017a17285227f874a
bd5a3b59a5ca9f0c0394c8bf90e818c3967778d9
/vre/apps/xauth/urls.py
2ba5dfc62bf27aafa163e3cf36365c4b0ea01be0
[]
no_license
BlickLabs/vre
85f377c04406c163464f7ddade7eafb579f1dfb1
6f3644fb9295f6355057cfa64a1156a329b4b4b8
refs/heads/develop
2020-05-22T04:28:31.913667
2018-07-06T21:12:14
2018-07-06T21:12:14
62,763,239
0
0
null
null
null
null
UTF-8
Python
false
false
297
py
#!/usr/bin/env python # -*- coding: utf-8 -*- from django.conf.urls import url from . import views urlpatterns = [ url(regex=r'^login/$', view=views.LoginView.as_view(), name='login'), url(regex=r'^logout/$', view=views.logout_view, name='logout'), ]
[ "mauriciodinki@gmail.com" ]
mauriciodinki@gmail.com
4679f028fe213a090bbd604db9707043887751db
8459dc3a3edebdd27b5589e9aa7215a55a42055b
/report/Lib/site-packages/wx/lib/mixins/listctrl.py
95ac8fbe9a65036e7334cd84fbfa901f88e58e2e
[]
no_license
lyj21803/ReportsTest
adb5ef9c057d0bd0669ed9807eccc7edb77655d7
37df534d61f7d1f781dd91299ffa99fba4ba0e49
refs/heads/master
2022-12-02T08:59:27.587331
2020-07-24T07:29:50
2020-07-24T07:29:50
278,009,685
1
0
null
null
null
null
UTF-8
Python
false
false
31,250
py
#---------------------------------------------------------------------------- # Name: wx.lib.mixins.listctrl # Purpose: Helpful mix-in classes for wxListCtrl # # Author: Robin Dunn # # Created: 15-May-2001 # Copyright: (c) 2001-2020 by Total Control Software # Licence: wxWindows license # Tags: phoenix-port, py3-port #---------------------------------------------------------------------------- # 12/14/2003 - Jeff Grimmett (grimmtooth@softhome.net) # # o 2.5 compatibility update. # o ListCtrlSelectionManagerMix untested. # # 12/21/2003 - Jeff Grimmett (grimmtooth@softhome.net) # # o wxColumnSorterMixin -> ColumnSorterMixin # o wxListCtrlAutoWidthMixin -> ListCtrlAutoWidthMixin # ... # 13/10/2004 - Pim Van Heuven (pim@think-wize.com) # o wxTextEditMixin: Support Horizontal scrolling when TAB is pressed on long # ListCtrls, support for WXK_DOWN, WXK_UP, performance improvements on # very long ListCtrls, Support for virtual ListCtrls # # 15-Oct-2004 - Robin Dunn # o wxTextEditMixin: Added Shift-TAB support # # 2008-11-19 - raf <raf@raf.org> # o ColumnSorterMixin: Added GetSortState() # import locale import wx import six if six.PY3: # python 3 lacks cmp: def cmp(a, b): return (a > b) - (a < b) #---------------------------------------------------------------------------- class ColumnSorterMixin: """ A mixin class that handles sorting of a wx.ListCtrl in REPORT mode when the column header is clicked on. There are a few requirments needed in order for this to work genericly: 1. The combined class must have a GetListCtrl method that returns the wx.ListCtrl to be sorted, and the list control must exist at the time the wx.ColumnSorterMixin.__init__ method is called because it uses GetListCtrl. 2. Items in the list control must have a unique data value set with list.SetItemData. 3. The combined class must have an attribute named itemDataMap that is a dictionary mapping the data values to a sequence of objects representing the values in each column. These values are compared in the column sorter to determine sort order. Interesting methods to override are GetColumnSorter, GetSecondarySortValues, and GetSortImages. See below for details. """ def __init__(self, numColumns): self.SetColumnCount(numColumns) list = self.GetListCtrl() if not list: raise ValueError("No wx.ListCtrl available") list.Bind(wx.EVT_LIST_COL_CLICK, self.__OnColClick, list) def SetColumnCount(self, newNumColumns): self._colSortFlag = [0] * newNumColumns self._col = -1 def SortListItems(self, col=-1, ascending=1): """Sort the list on demand. Can also be used to set the sort column and order.""" oldCol = self._col if col != -1: self._col = col self._colSortFlag[col] = ascending self.GetListCtrl().SortItems(self.GetColumnSorter()) self.__updateImages(oldCol) def GetColumnWidths(self): """ Returns a list of column widths. Can be used to help restore the current view later. """ list = self.GetListCtrl() rv = [] for x in range(len(self._colSortFlag)): rv.append(list.GetColumnWidth(x)) return rv def GetSortImages(self): """ Returns a tuple of image list indexesthe indexes in the image list for an image to be put on the column header when sorting in descending order. """ return (-1, -1) # (decending, ascending) image IDs def GetColumnSorter(self): """Returns a callable object to be used for comparing column values when sorting.""" return self.__ColumnSorter def GetSecondarySortValues(self, col, key1, key2): """Returns a tuple of 2 values to use for secondary sort values when the items in the selected column match equal. The default just returns the item data values.""" return (key1, key2) def __OnColClick(self, evt): oldCol = self._col self._col = col = evt.GetColumn() self._colSortFlag[col] = int(not self._colSortFlag[col]) self.GetListCtrl().SortItems(self.GetColumnSorter()) if wx.Platform != "__WXMAC__" or wx.SystemOptions.GetOptionInt("mac.listctrl.always_use_generic") == 1: self.__updateImages(oldCol) evt.Skip() self.OnSortOrderChanged() def OnSortOrderChanged(self): """ Callback called after sort order has changed (whenever user clicked column header). """ pass def GetSortState(self): """ Return a tuple containing the index of the column that was last sorted and the sort direction of that column. Usage: col, ascending = self.GetSortState() # Make changes to list items... then resort self.SortListItems(col, ascending) """ return (self._col, self._colSortFlag[self._col]) def __ColumnSorter(self, key1, key2): col = self._col ascending = self._colSortFlag[col] item1 = self.itemDataMap[key1][col] item2 = self.itemDataMap[key2][col] #--- Internationalization of string sorting with locale module if isinstance(item1, six.text_type) and isinstance(item2, six.text_type): # both are unicode (py2) or str (py3) cmpVal = locale.strcoll(item1, item2) elif isinstance(item1, six.binary_type) or isinstance(item2, six.binary_type): # at least one is a str (py2) or byte (py3) cmpVal = locale.strcoll(str(item1), str(item2)) else: cmpVal = cmp(item1, item2) #--- # If the items are equal then pick something else to make the sort value unique if cmpVal == 0: cmpVal = cmp(*self.GetSecondarySortValues(col, key1, key2)) if ascending: return cmpVal else: return -cmpVal def __updateImages(self, oldCol): sortImages = self.GetSortImages() if self._col != -1 and sortImages[0] != -1: img = sortImages[self._colSortFlag[self._col]] list = self.GetListCtrl() if oldCol != -1: list.ClearColumnImage(oldCol) list.SetColumnImage(self._col, img) #---------------------------------------------------------------------------- #---------------------------------------------------------------------------- class ListCtrlAutoWidthMixin: """ A mix-in class that automatically resizes the last column to take up the remaining width of the wx.ListCtrl. This causes the wx.ListCtrl to automatically take up the full width of the list, without either a horizontal scroll bar (unless absolutely necessary) or empty space to the right of the last column. NOTE: This only works for report-style lists. WARNING: If you override the EVT_SIZE event in your wx.ListCtrl, make sure you call event.Skip() to ensure that the mixin's _OnResize method is called. This mix-in class was written by Erik Westra <ewestra@wave.co.nz> """ def __init__(self): """ Standard initialiser. """ self._resizeColMinWidth = None self._resizeColStyle = "LAST" self._resizeCol = 0 self.Bind(wx.EVT_SIZE, self._onResize) self.Bind(wx.EVT_LIST_COL_END_DRAG, self._onResize, self) def setResizeColumn(self, col): """ Specify which column that should be autosized. Pass either 'LAST' or the column number. Default is 'LAST'. """ if col == "LAST": self._resizeColStyle = "LAST" else: self._resizeColStyle = "COL" self._resizeCol = col def resizeLastColumn(self, minWidth): """ Resize the last column appropriately. If the list's columns are too wide to fit within the window, we use a horizontal scrollbar. Otherwise, we expand the right-most column to take up the remaining free space in the list. This method is called automatically when the wx.ListCtrl is resized; you can also call it yourself whenever you want the last column to be resized appropriately (eg, when adding, removing or resizing columns). 'minWidth' is the preferred minimum width for the last column. """ self.resizeColumn(minWidth) def resizeColumn(self, minWidth): self._resizeColMinWidth = minWidth self._doResize() # ===================== # == Private Methods == # ===================== def _onResize(self, event): """ Respond to the wx.ListCtrl being resized. We automatically resize the last column in the list. """ if 'gtk2' in wx.PlatformInfo or 'gtk3' in wx.PlatformInfo: self._doResize() else: wx.CallAfter(self._doResize) event.Skip() def _doResize(self): """ Resize the last column as appropriate. If the list's columns are too wide to fit within the window, we use a horizontal scrollbar. Otherwise, we expand the right-most column to take up the remaining free space in the list. We remember the current size of the last column, before resizing, as the preferred minimum width if we haven't previously been given or calculated a minimum width. This ensure that repeated calls to _doResize() don't cause the last column to size itself too large. """ if not self: # avoid a PyDeadObject error return if self.GetSize().height < 32: return # avoid an endless update bug when the height is small. numCols = self.GetColumnCount() if numCols == 0: return # Nothing to resize. if(self._resizeColStyle == "LAST"): resizeCol = self.GetColumnCount() else: resizeCol = self._resizeCol resizeCol = max(1, resizeCol) if self._resizeColMinWidth is None: self._resizeColMinWidth = self.GetColumnWidth(resizeCol - 1) # Get total width listWidth = self.GetClientSize().width totColWidth = 0 # Width of all columns except last one. for col in range(numCols): if col != (resizeCol-1): totColWidth = totColWidth + self.GetColumnWidth(col) resizeColWidth = self.GetColumnWidth(resizeCol - 1) if totColWidth + self._resizeColMinWidth > listWidth: # We haven't got the width to show the last column at its minimum # width -> set it to its minimum width and allow the horizontal # scrollbar to show. self.SetColumnWidth(resizeCol-1, self._resizeColMinWidth) return # Resize the last column to take up the remaining available space. self.SetColumnWidth(resizeCol-1, listWidth - totColWidth) #---------------------------------------------------------------------------- #---------------------------------------------------------------------------- SEL_FOC = wx.LIST_STATE_SELECTED | wx.LIST_STATE_FOCUSED def selectBeforePopup(event): """Ensures the item the mouse is pointing at is selected before a popup. Works with both single-select and multi-select lists.""" ctrl = event.GetEventObject() if isinstance(ctrl, wx.ListCtrl): n, flags = ctrl.HitTest(event.GetPosition()) if n >= 0: if not ctrl.GetItemState(n, wx.LIST_STATE_SELECTED): for i in range(ctrl.GetItemCount()): ctrl.SetItemState(i, 0, SEL_FOC) #for i in getListCtrlSelection(ctrl, SEL_FOC): # ctrl.SetItemState(i, 0, SEL_FOC) ctrl.SetItemState(n, SEL_FOC, SEL_FOC) def getListCtrlSelection(listctrl, state=wx.LIST_STATE_SELECTED): """ Returns list of item indexes of given state (selected by defaults) """ res = [] idx = -1 while 1: idx = listctrl.GetNextItem(idx, wx.LIST_NEXT_ALL, state) if idx == -1: break res.append(idx) return res wxEVT_DOPOPUPMENU = wx.NewEventType() EVT_DOPOPUPMENU = wx.PyEventBinder(wxEVT_DOPOPUPMENU, 0) class ListCtrlSelectionManagerMix: """Mixin that defines a platform independent selection policy As selection single and multi-select list return the item index or a list of item indexes respectively. """ _menu = None def __init__(self): self.Bind(wx.EVT_RIGHT_DOWN, self.OnLCSMRightDown) self.Bind(EVT_DOPOPUPMENU, self.OnLCSMDoPopup) # self.Connect(-1, -1, self.wxEVT_DOPOPUPMENU, self.OnLCSMDoPopup) def getPopupMenu(self): """ Override to implement dynamic menus (create) """ return self._menu def setPopupMenu(self, menu): """ Must be set for default behaviour """ self._menu = menu def afterPopupMenu(self, menu): """ Override to implement dynamic menus (destroy) """ pass def getSelection(self): res = getListCtrlSelection(self) if self.GetWindowStyleFlag() & wx.LC_SINGLE_SEL: if res: return res[0] else: return -1 else: return res def OnLCSMRightDown(self, event): selectBeforePopup(event) event.Skip() menu = self.getPopupMenu() if menu: evt = wx.PyEvent() evt.SetEventType(wxEVT_DOPOPUPMENU) evt.menu = menu evt.pos = event.GetPosition() wx.PostEvent(self, evt) def OnLCSMDoPopup(self, event): self.PopupMenu(event.menu, event.pos) self.afterPopupMenu(event.menu) #---------------------------------------------------------------------------- #---------------------------------------------------------------------------- from bisect import bisect class TextEditMixin: """ A mixin class that enables any text in any column of a multi-column listctrl to be edited by clicking on the given row and column. You close the text editor by hitting the ENTER key or clicking somewhere else on the listctrl. You switch to the next column by hiting TAB. To use the mixin you have to include it in the class definition and call the __init__ function:: class TestListCtrl(wx.ListCtrl, TextEditMixin): def __init__(self, parent, ID, pos=wx.DefaultPosition, size=wx.DefaultSize, style=0): wx.ListCtrl.__init__(self, parent, ID, pos, size, style) TextEditMixin.__init__(self) Authors: Steve Zatz, Pim Van Heuven (pim@think-wize.com) """ editorBgColour = wx.Colour(255,255,175) # Yellow editorFgColour = wx.Colour(0,0,0) # black def __init__(self): #editor = wx.TextCtrl(self, -1, pos=(-1,-1), size=(-1,-1), # style=wx.TE_PROCESS_ENTER|wx.TE_PROCESS_TAB \ # |wx.TE_RICH2) self.make_editor() self.Bind(wx.EVT_TEXT_ENTER, self.CloseEditor) self.Bind(wx.EVT_LEFT_DOWN, self.OnLeftDown) self.Bind(wx.EVT_LEFT_DCLICK, self.OnLeftDown) self.Bind(wx.EVT_LIST_ITEM_SELECTED, self.OnItemSelected, self) def make_editor(self, col_style=wx.LIST_FORMAT_LEFT): style =wx.TE_PROCESS_ENTER|wx.TE_PROCESS_TAB|wx.TE_RICH2 style |= {wx.LIST_FORMAT_LEFT: wx.TE_LEFT, wx.LIST_FORMAT_RIGHT: wx.TE_RIGHT, wx.LIST_FORMAT_CENTRE : wx.TE_CENTRE }[col_style] editor = wx.TextCtrl(self, -1, style=style) editor.SetBackgroundColour(self.editorBgColour) editor.SetForegroundColour(self.editorFgColour) font = self.GetFont() editor.SetFont(font) self.curRow = 0 self.curCol = 0 editor.Hide() if hasattr(self, 'editor'): self.editor.Destroy() self.editor = editor self.col_style = col_style self.editor.Bind(wx.EVT_CHAR, self.OnChar) self.editor.Bind(wx.EVT_KILL_FOCUS, self.CloseEditor) def OnItemSelected(self, evt): self.curRow = evt.GetIndex() evt.Skip() def OnChar(self, event): ''' Catch the TAB, Shift-TAB, cursor DOWN/UP key code so we can open the editor at the next column (if any).''' keycode = event.GetKeyCode() if keycode == wx.WXK_TAB and event.ShiftDown(): self.CloseEditor() if self.curCol-1 >= 0: self.OpenEditor(self.curCol-1, self.curRow) elif keycode == wx.WXK_TAB: self.CloseEditor() if self.curCol+1 < self.GetColumnCount(): self.OpenEditor(self.curCol+1, self.curRow) elif keycode == wx.WXK_ESCAPE: self.CloseEditor() elif keycode == wx.WXK_DOWN: self.CloseEditor() if self.curRow+1 < self.GetItemCount(): self._SelectIndex(self.curRow+1) self.OpenEditor(self.curCol, self.curRow) elif keycode == wx.WXK_UP: self.CloseEditor() if self.curRow > 0: self._SelectIndex(self.curRow-1) self.OpenEditor(self.curCol, self.curRow) else: event.Skip() def OnLeftDown(self, evt=None): ''' Examine the click and double click events to see if a row has been click on twice. If so, determine the current row and columnn and open the editor.''' if self.editor.IsShown(): self.CloseEditor() x,y = evt.GetPosition() row,flags = self.HitTest((x,y)) if row != self.curRow: # self.curRow keeps track of the current row evt.Skip() return # the following should really be done in the mixin's init but # the wx.ListCtrl demo creates the columns after creating the # ListCtrl (generally not a good idea) on the other hand, # doing this here handles adjustable column widths self.col_locs = [0] loc = 0 for n in range(self.GetColumnCount()): loc = loc + self.GetColumnWidth(n) self.col_locs.append(loc) col = bisect(self.col_locs, x+self.GetScrollPos(wx.HORIZONTAL)) - 1 self.OpenEditor(col, row) def OpenEditor(self, col, row): ''' Opens an editor at the current position. ''' # give the derived class a chance to Allow/Veto this edit. evt = wx.ListEvent(wx.wxEVT_COMMAND_LIST_BEGIN_LABEL_EDIT, self.GetId()) evt.Index = row evt.Column = col item = self.GetItem(row, col) evt.Item.SetId(item.GetId()) evt.Item.SetColumn(item.GetColumn()) evt.Item.SetData(item.GetData()) evt.Item.SetText(item.GetText()) ret = self.GetEventHandler().ProcessEvent(evt) if ret and not evt.IsAllowed(): return # user code doesn't allow the edit. if self.GetColumn(col).Align != self.col_style: self.make_editor(self.GetColumn(col).Align) x0 = self.col_locs[col] x1 = self.col_locs[col+1] - x0 scrolloffset = self.GetScrollPos(wx.HORIZONTAL) # scroll forward if x0+x1-scrolloffset > self.GetSize()[0]: if wx.Platform == "__WXMSW__": # don't start scrolling unless we really need to offset = x0+x1-self.GetSize()[0]-scrolloffset # scroll a bit more than what is minimum required # so we don't have to scroll everytime the user presses TAB # which is very tireing to the eye addoffset = self.GetSize()[0]/4 # but be careful at the end of the list if addoffset + scrolloffset < self.GetSize()[0]: offset += addoffset self.ScrollList(offset, 0) scrolloffset = self.GetScrollPos(wx.HORIZONTAL) else: # Since we can not programmatically scroll the ListCtrl # close the editor so the user can scroll and open the editor # again self.editor.SetValue(self.GetItem(row, col).GetText()) self.curRow = row self.curCol = col self.CloseEditor() return y0 = self.GetItemRect(row)[1] def _activate_editor(editor): editor.SetSize(x0-scrolloffset,y0, x1,-1, wx.SIZE_USE_EXISTING) editor.SetValue(self.GetItem(row, col).GetText()) editor.Show() editor.Raise() editor.SetSelection(-1,-1) editor.SetFocus() wx.CallAfter(_activate_editor, self.editor) self.curRow = row self.curCol = col # FIXME: this function is usually called twice - second time because # it is binded to wx.EVT_KILL_FOCUS. Can it be avoided? (MW) def CloseEditor(self, evt=None): ''' Close the editor and save the new value to the ListCtrl. ''' if not self.editor.IsShown(): return text = self.editor.GetValue() self.editor.Hide() self.SetFocus() # post wxEVT_COMMAND_LIST_END_LABEL_EDIT # Event can be vetoed. It doesn't has SetEditCanceled(), what would # require passing extra argument to CloseEditor() evt = wx.ListEvent(wx.wxEVT_COMMAND_LIST_END_LABEL_EDIT, self.GetId()) evt.Index = self.curRow evt.Column = self.curCol item = wx.ListItem(self.GetItem(self.curRow, self.curCol)) item.SetText(text) evt.SetItem(item) ret = self.GetEventHandler().ProcessEvent(evt) if not ret or evt.IsAllowed(): if self.IsVirtual(): # replace by whather you use to populate the virtual ListCtrl # data source self.SetVirtualData(self.curRow, self.curCol, text) else: self.SetItem(self.curRow, self.curCol, text) self.RefreshItem(self.curRow) def _SelectIndex(self, row): listlen = self.GetItemCount() if row < 0 and not listlen: return if row > (listlen-1): row = listlen -1 self.SetItemState(self.curRow, ~wx.LIST_STATE_SELECTED, wx.LIST_STATE_SELECTED) self.EnsureVisible(row) self.SetItemState(row, wx.LIST_STATE_SELECTED, wx.LIST_STATE_SELECTED) #---------------------------------------------------------------------------- #---------------------------------------------------------------------------- """ FILENAME: CheckListCtrlMixin.py AUTHOR: Bruce Who (bruce.who.hk at gmail.com) DATE: 2006-02-09 DESCRIPTION: This script provide a mixin for ListCtrl which add a checkbox in the first column of each row. It is inspired by limodou's CheckList.py(which can be got from his NewEdit) and improved: - You can just use InsertStringItem() to insert new items; - Once a checkbox is checked/unchecked, the corresponding item is not selected; - You can use SetItemData() and GetItemData(); - Interfaces are changed to OnCheckItem(), IsChecked(), CheckItem(). You should not set a imagelist for the ListCtrl once this mixin is used. HISTORY: 1.3 - You can check/uncheck a group of sequential items by <Shift-click>: First click(or <Shift-Click>) item1 to check/uncheck it, then Shift-click item2 to check/uncheck it, and you'll find that all items between item1 and item2 are check/unchecked! 1.2 - Add ToggleItem() 1.1 - Initial version """ class CheckListCtrlMixin(object): """ This is a mixin for ListCtrl which add a checkbox in the first column of each row. It is inspired by limodou's CheckList.py(which can be got from his NewEdit) and improved: - You can just use InsertStringItem() to insert new items; - Once a checkbox is checked/unchecked, the corresponding item is not selected; - You can use SetItemData() and GetItemData(); - Interfaces are changed to OnCheckItem(), IsChecked(), CheckItem(). You should not set a imagelist for the ListCtrl once this mixin is used. """ def __init__(self, check_image=None, uncheck_image=None, imgsz=(16,16)): if check_image is not None: imgsz = check_image.GetSize() elif uncheck_image is not None: imgsz = check_image.GetSize() self.__imagelist_ = wx.ImageList(*imgsz) # Create default checkbox images if none were specified if check_image is None: check_image = self.__CreateBitmap(wx.CONTROL_CHECKED, imgsz) if uncheck_image is None: uncheck_image = self.__CreateBitmap(0, imgsz) self.uncheck_image = self.__imagelist_.Add(uncheck_image) self.check_image = self.__imagelist_.Add(check_image) self.AssignImageList(self.__imagelist_, wx.IMAGE_LIST_SMALL) self.__last_check_ = None self.Bind(wx.EVT_LEFT_DOWN, self.__OnLeftDown_) # Monkey-patch in a new InsertItem so we can also set the image ID for the item self._origInsertItem = self.InsertItem self.InsertItem = self.__InsertItem_ def __InsertItem_(self, *args, **kw): index = self._origInsertItem(*args, **kw) self.SetItemImage(index, self.uncheck_image) return index def __CreateBitmap(self, flag=0, size=(16, 16)): """Create a bitmap of the platforms native checkbox. The flag is used to determine the checkboxes state (see wx.CONTROL_*) """ bmp = wx.Bitmap(*size) dc = wx.MemoryDC(bmp) dc.SetBackground(wx.WHITE_BRUSH) dc.Clear() wx.RendererNative.Get().DrawCheckBox(self, dc, (0, 0, size[0], size[1]), flag) dc.SelectObject(wx.NullBitmap) return bmp def __OnLeftDown_(self, evt): (index, flags) = self.HitTest(evt.GetPosition()) if flags == wx.LIST_HITTEST_ONITEMICON: img_idx = self.GetItem(index).GetImage() flag_check = img_idx == 0 begin_index = index end_index = index if self.__last_check_ is not None \ and wx.GetKeyState(wx.WXK_SHIFT): last_index, last_flag_check = self.__last_check_ if last_flag_check == flag_check: # XXX what if the previous item is deleted or new items # are inserted? item_count = self.GetItemCount() if last_index < item_count: if last_index < index: begin_index = last_index end_index = index elif last_index > index: begin_index = index end_index = last_index else: assert False while begin_index <= end_index: self.CheckItem(begin_index, flag_check) begin_index += 1 self.__last_check_ = (index, flag_check) else: evt.Skip() def OnCheckItem(self, index, flag): pass def IsChecked(self, index): return self.GetItem(index).GetImage() == 1 def CheckItem(self, index, check=True): img_idx = self.GetItem(index).GetImage() if img_idx == 0 and check: self.SetItemImage(index, 1) self.OnCheckItem(index, True) elif img_idx == 1 and not check: self.SetItemImage(index, 0) self.OnCheckItem(index, False) def ToggleItem(self, index): self.CheckItem(index, not self.IsChecked(index)) #---------------------------------------------------------------------------- #---------------------------------------------------------------------------- # Mode Flags HIGHLIGHT_ODD = 1 # Highlight the Odd rows HIGHLIGHT_EVEN = 2 # Highlight the Even rows class ListRowHighlighter: """Editra Control Library: ListRowHighlighter Mixin class that handles automatic background highlighting of alternate rows in the a ListCtrl. The background of the rows are highlighted automatically as items are added or inserted in the control based on the mixins Mode and set Color. By default the Even rows will be highlighted with the systems highlight color. """ def __init__(self, color=None, mode=HIGHLIGHT_EVEN): """Initialize the highlighter mixin @keyword color: Set a custom highlight color (default uses system color) @keyword mode: HIGHLIGHT_EVEN (default) or HIGHLIGHT_ODD """ # Attributes self._color = color self._defaultb = wx.SystemSettings.GetColour(wx.SYS_COLOUR_LISTBOX) self._mode = mode # Event Handlers self.Bind(wx.EVT_LIST_INSERT_ITEM, lambda evt: self.RefreshRows()) self.Bind(wx.EVT_LIST_DELETE_ITEM, lambda evt: self.RefreshRows()) def RefreshRows(self): """Re-color all the rows""" if self._color is None: if wx.Platform in ('__WXGTK__', '__WXMSW__'): color = wx.SystemSettings.GetColour(wx.SYS_COLOUR_3DLIGHT) else: color = wx.Colour(237, 243, 254) else: color = self._color local_defaultb = self._defaultb local_mode = self._mode for row in range(self.GetItemCount()): if local_mode & HIGHLIGHT_EVEN: dohlight = not row % 2 else: dohlight = row % 2 if dohlight: self.SetItemBackgroundColour(row, color) elif local_defaultb: self.SetItemBackgroundColour(row, local_defaultb) else: # This part of the loop should only happen once if self._defaultb is None. local_defaultb = self._defaultb = self.GetItemBackgroundColour(row) self.SetItemBackgroundColour(row, local_defaultb) def SetHighlightColor(self, color): """Set the color used to highlight the rows. Call :meth:`RefreshRows` after this if you wish to update all the rows highlight colors. @param color: wx.Color or None to set default """ self._color = color def SetHighlightMode(self, mode): """Set the highlighting mode to either HIGHLIGHT_EVEN or to HIGHLIGHT_ODD. Call :meth:`RefreshRows` afterwards to update the list state. @param mode: HIGHLIGHT_* mode value """ self._mode = mode #----------------------------------------------------------------------------
[ "lyj218@qq.com" ]
lyj218@qq.com
3ea0abb3bd098265da02a905ab08ccb1c2b9a663
aa8ca649dfe718398bc57ec00133e67df71ea407
/cbz_notes/wsgi.py
438f30169f7ec9e8bbed857bf5de8f73f7d5f0c9
[]
no_license
habelash/cbz-notes
516a148f030b99a33445a06b1933b518a81d3667
d1f6ab75dde56f28fd421e4dcf77d69e7d491e54
refs/heads/master
2022-11-09T08:49:32.274230
2020-06-17T18:55:16
2020-06-17T18:55:16
271,054,937
0
0
null
null
null
null
UTF-8
Python
false
false
395
py
""" WSGI config for cbz_notes project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.0/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'cbz_notes.settings') application = get_wsgi_application()
[ "habelash@gmail.com" ]
habelash@gmail.com
a79f0d2f37b1ef05ad04d86eb3a3f170aa616237
482ee9f8972bb01b0de68c921cddb27aa9470a8c
/raspi_server/servomot.py
f31c565c0db7dbe0c901a91e0bb113b35791e157
[]
no_license
bhavika022/Multi-Terrain-Robot
1a793ad5ac7f6f0ddd20cbca4b9a035efe2f8ed9
d2aa39f4b471d126e928d4bc51bb88096f4c2cc3
refs/heads/master
2023-08-18T22:07:18.188112
2021-10-15T16:20:32
2021-10-15T16:20:32
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,138
py
import RPi.GPIO as GPIO import time def setup(): GPIO.setmode(GPIO.BOARD) GPIO.setup(11, GPIO.OUT) global p p=GPIO.PWM(11, 50) p.start(0) p.ChangeDutyCycle(0) def setangle_up(): global duty duty=(90/18)+2 GPIO.output(11, True) p.ChangeDutyCycle(duty) time.sleep(0.5) GPIO.output(11, False) p.ChangeDutyCycle(0) print('Degrees the Servo was rotated by: ') print('90') def setangle_down(): global duty duty=(180/18)+2 GPIO.output(11, True) p.ChangeDutyCycle(duty) time.sleep(0.5) GPIO.output(11, False) p.ChangeDutyCycle(0) print('Degrees the Servo was rotated by: ') print('180') def rotmul(): global i i=0 for i in range (5): global duty1, duty2 duty1=(90/18)+2 duty2=(180/18)+2 GPIO.output(11, True) p.ChangeDutyCycle(duty1) time.sleep(0.5) GPIO.output(11, False) p.ChangeDutyCycle(0) GPIO.output(11, True) p.ChangeDutyCycle(duty2) time.sleep(0.5) GPIO.output(11, False) p.ChangeDutyCycle(0) def close(): p.stop() #GPIO.cleanup()
[ "ameya.k.kale@gmail.com" ]
ameya.k.kale@gmail.com
c9800b7561104d8aa6fcc841bb12aac744f3d879
4ee74237ad3230147674546223a0ff9644adf944
/quickstart/migrations/0024_poresizedistribution.py
8f479c54adbb664e1cb109894b3673f45dc70ec9
[]
no_license
PMEAL/porespy-backend
d5641c8b1ae1930b5dd9185c43f74036d5e95f94
bb76cae9e752a95e428ec417bf1524f90b110790
refs/heads/master
2023-03-31T05:35:48.043698
2021-04-11T01:21:34
2021-04-11T01:21:34
323,444,231
0
0
null
null
null
null
UTF-8
Python
false
false
529
py
# Generated by Django 3.1.3 on 2021-03-16 20:26 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('quickstart', '0023_auto_20210314_2350'), ] operations = [ migrations.CreateModel( name='PoreSizeDistribution', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('psd_im', models.TextField(default='')), ], ), ]
[ "espitiaandres123@gmail.com" ]
espitiaandres123@gmail.com
f3e768e706777c6ba9fb873bf9632cbe6fddb951
ebb4fcf4b95e9143136f78aa3cba426829b1b2ff
/urls.py
be6411b662095da3d336edc4302660a233c3f043
[]
no_license
brasky/scheduler
26881e1056753d26f2deecbb5920d40f3642b628
4adbd99a629d4b9eab0d56fc8b8e5ef14b366354
refs/heads/master
2021-01-20T17:33:49.134819
2013-11-13T18:35:05
2013-11-13T18:35:05
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,029
py
from django.conf.urls import patterns, include, url urlpatterns = patterns('', url(r'^$','scheduler.views.home_view'), url(r'^accounts/login/$', 'django.contrib.auth.views.login', {'template_name': 'scheduler/login.html'}), url(r'^accounts/logout/$','scheduler.views.logout_view'), url(r'^accounts/is_loggedin/$','scheduler.views.is_loggedin_view'), url(r'^accounts/register/$','scheduler.views.register_view'), url(r'^accounts/profile/$','scheduler.views.account_view'), url(r'^createevent/(?P<scheduleid>\d+)$','scheduler.views.create_event_view'), url(r'^schedule/(?P<scheduleid>\d+)$','scheduler.views.schedule_view'), url(r'^accounts/createschedule/$','scheduler.views.create_schedule_view'), url(r'^friends/$','scheduler.views.friends_view'), url(r'^friends/accept/(?P<friendid>\d+)$','scheduler.views.friends_accept_view'), url(r'^friends/add/$','scheduler.views.friends_add_view'), )
[ "redx47@gmail.com" ]
redx47@gmail.com
ebb1c1e58b95234c46d4cd1a13b51cec064b0756
620a79597511cecd55cef6cdc9ac09062c1fe12b
/valid/valid/urls.py
85db8c10cd4cafc533bdf30a4ba59925b253b97f
[]
no_license
ashfan6339/myfristproject
6e4a0d76a3232022e94c8706342f97e9ae6c65ff
535864d1179521144c2b518feddbbc9a43836f50
refs/heads/master
2021-04-20T22:29:03.952907
2020-05-21T10:51:57
2020-05-21T10:51:57
249,721,885
0
0
null
null
null
null
UTF-8
Python
false
false
839
py
"""valid URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/3.0/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.urls import include, path 2. Add a URL to urlpatterns: path('blog/', include('blog.urls')) """ from django.contrib import admin from django.urls import path,include from registor import views as v urlpatterns = [ path('admin/', admin.site.urls), path("registor", v.registor, name="registor"), ]
[ "shaikashfan3@gmail.com" ]
shaikashfan3@gmail.com
753b5b2ec561ad28d7410e49144c4be4fac47627
28f52b0e9c8f7fe15a008127f2a76c8854efaa7e
/pkgs/clean-pkg/src/genie/libs/clean/stages/nxos/aci/image_handler.py
2c5fe6e13db35982ecd8d592e92175249ab32a43
[ "Apache-2.0" ]
permissive
dthangap/genielibs
cb5098e675c51f2c2c46a929faf630cbad6aa7b3
778edb3b310bac960f507dae55e82ac027d8c6c8
refs/heads/master
2023-02-01T05:18:04.575913
2020-12-16T12:55:54
2020-12-16T12:55:54
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,399
py
""" NXOS ACI: Image Handler Class """ import yaml from genie.libs.clean.stages.image_handler import BaseImageHandler from pyats.utils.schemaengine import Schema, ListOf, Optional class ImageLoader(object): EXPECTED_IMAGE_STRUCTURE_MSG = """\ Expected one of the following structures for 'images' in the clean yaml Structure #1 ------------ images: - /path/to/controller_image.bin - /path/to/switch_image.bin Structure #2 ------------ images: controller: - /path/to/controller_image.bin switch: - /path/to/switch_image.bin Structure #3 ------------ images: controller: file: - /path/to/controller_image.bin switch: file: - /path/to/switch_image.bin But got the following structure ------------------------------- {}""" def load(self, images): if (not self.valid_structure_1(images) and not self.valid_structure_2(images) and not self.valid_structure_3(images)): raise Exception(self.EXPECTED_IMAGE_STRUCTURE_MSG.format( yaml.dump({'images': images}))) def valid_structure_1(self, images): schema = ListOf(str) try: Schema(schema).validate(images) except Exception: return False if len(images) == 1: # This is not a bug. It is optional to clean only switches or only # controllers but we do not know what type of image the user # provided if they only provide 1. setattr(self, 'controller', images) setattr(self, 'switch', images) return True if len(images) == 2: setattr(self, 'controller', images[:1]) setattr(self, 'switch', images[1:]) return True else: return False def valid_structure_2(self, images): schema = { Optional('controller'): ListOf(str), Optional('switch'): ListOf(str) } try: Schema(schema).validate(images) except Exception: return False if ('controller' in images and 'switch' in images and len(images['controller']) == 1 and len(images['switch']) == 1): setattr(self, 'controller', images['controller']) setattr(self, 'switch', images['switch']) return True elif ('controller' in images and len(images['controller']) == 1): setattr(self, 'controller', images['controller']) return True elif ('switch' in images and len(images['switch']) == 1): setattr(self, 'switch', images['switch']) return True else: return False def valid_structure_3(self, images): schema = { Optional('controller'): { 'file': ListOf(str) }, Optional('switch'): { 'file': ListOf(str) } } try: Schema(schema).validate(images) except Exception: return False if ('controller' in images and 'switch' in images and len(images['controller']['file']) == 1 and len(images['switch']['file']) == 1): setattr(self, 'controller', images['controller']['file']) setattr(self, 'switch', images['switch']['file']) return True elif ('controller' in images and len(images['controller']['file']) == 1): setattr(self, 'controller', images['controller']['file']) return True elif ('switch' in images and len(images['switch']['file']) == 1): setattr(self, 'switch', images['switch']['file']) return True else: return False class ImageHandler(BaseImageHandler, ImageLoader): def __init__(self, device, images, *args, **kwargs): # Set defaults self.controller = [] self.switch = [] # Check if images is one of the valid structures and # load into a consolidated structure ImageLoader.load(self, images) # Temp workaround for XPRESSO if self.controller: self.controller = [self.controller[0].replace('file://', '')] if self.switch: self.switch = [self.switch[0].replace('file://', '')] super().__init__(device, images, *args, **kwargs) def update_image_references(self, section): if 'image_mapping' in section.parameters: for index, image in enumerate(self.controller): # change the saved image to the new image name/path self.controller[index] = section.parameters['image_mapping'].get( image, self.controller[index]) for index, image in enumerate(self.switch): # change the saved image to the new image name/path self.switch[index] = section.parameters['image_mapping'].get( image, self.switch[index]) def update_copy_to_linux(self): '''Update clean section 'copy_to_linux' with image information''' # Init 'copy_to_linux' defaults origin = self.device.clean.setdefault('copy_to_linux', {}).\ setdefault('origin', {}) origin.update({'files': self.controller + self.switch}) def update_copy_to_device(self): '''Update clean stage 'copy_to_device' with image information''' origin = self.device.clean.setdefault('copy_to_device', {}).\ setdefault('origin', {}) origin.update({'files': self.controller + self.switch}) def update_fabric_upgrade(self): '''Update clean stage 'fabric_upgrade' with image information''' fabric_upgrade = self.device.clean.setdefault('fabric_upgrade', {}) fabric_upgrade.update({'controller_image': self.controller}) fabric_upgrade.update({'switch_image': self.switch}) def update_fabric_clean(self): '''Update clean stage 'fabric_clean' with image information ''' fabric_clean = self.device.clean.setdefault('fabric_clean', {}) if fabric_clean.get('copy_boot_image', {}).get('origin', {}): fabric_clean['copy_boot_image']['origin'].update({'files': self.switch})
[ "tahigash@cisco.com" ]
tahigash@cisco.com
c43501f1134f44d9e0c3c38a8ce719ea17e5bbcb
3253da5603971958d69df0ed442e3341a8d3bff4
/1-Iniciante/1914.py
67fa34c039b20ad33bd528808a4ce2d4016000af
[]
no_license
CleitonSilvaT/URI_Python
1c73ec0852ae87c6138baa148ad8c2cb56bb723e
a8510bab2fa8f680b54058fafebff3a2727617d9
refs/heads/master
2021-06-20T08:18:50.104839
2021-05-20T08:59:19
2021-05-20T08:59:19
213,665,657
0
0
null
null
null
null
UTF-8
Python
false
false
959
py
# -*- coding: utf-8 -*- if __name__ == '__main__': # Entrada casos_teste = int(input()) while(casos_teste > 0): # Entrada dados = input() escolha = dados.split(' ') # nomepessoa1 - escolha[0] # escolhapessoa1 - escolha[1] # nomepessoa2 - escolha[2] # escolhapessoa2 - escolha[3] # Entrada valores = input() numeros = valores.split(' ') # Calculando soma dos valores total = int(numeros[0]) + int(numeros[1]) # Identificando se a soma eh PAR ou IMPAR if((total % 2) == 0): # Imprimindo o vencedor if(escolha[1] == 'PAR'): print(escolha[0]) else: print(escolha[2]) else: # Imprimindo o vencedor if(escolha[1] == 'IMPAR'): print(escolha[0]) else: print(escolha[2]) casos_teste -= 1
[ "cleitonsilvatavares@gmail.com" ]
cleitonsilvatavares@gmail.com
b672c87e3458490ceb0e8b3852355a8c15a2c399
d1fadc514274711a7986a6b3caaaee7e8d48b4a6
/plot_scripts/scratch29.py
9b454212d7485e7e1237f495490e6b1a3e2c0169
[ "MIT" ]
permissive
lbaiao/sys-simulator-2
24d940db6423070818c23b6ffefbc5da4a1030a0
94f00d43309fe7b56dac5099bd4024695ba317b6
refs/heads/master
2021-08-20T08:30:06.864473
2021-06-30T10:37:26
2021-06-30T10:37:26
230,333,523
1
0
null
2021-06-30T10:37:27
2019-12-26T22:02:59
Jupyter Notebook
UTF-8
Python
false
false
1,688
py
import pickle import matplotlib.pyplot as plt import numpy as np filepath = 'D:/Dev/sys-simulator-2/data/scratch29.pickle' file = open(filepath, 'rb') data = pickle.load(file) aux_range = [10,15,20] action_counts_total = data['action_counts_total'] d2d_spectral_effs = data['d2d_speffs_avg_total'] mue_success_rate = data['mue_success_rate'] equals_counts_total = data['equals_counts_total'] d2d_speffs_avg = list() for i, d in enumerate(d2d_spectral_effs): d2d_speffs_avg.append(np.average(d)) fig2, ax1 = plt.subplots() ax1.set_xlabel('Number of D2D pairs in the RB') ax1.set_ylabel('D2D Average Spectral Efficiency [bps/Hz]', color='tab:blue') ax1.plot(d2d_speffs_avg, '.', color='tab:blue') ax2 = ax1.twinx() ax2.set_ylabel('MUE Success Rate', color='tab:red') ax2.plot(mue_success_rate, '.', color='tab:red') fig2.tight_layout() xi = list(range(len(aux_range))) ax = [0,1,2,3,4] axi = list(range(len(ax))) for i, c in enumerate(action_counts_total): if i in aux_range: plt.figure() plt.plot(np.mean(c, axis=0)/i*100, '*',label='mean') plt.plot(np.std(c, axis=0)/i*100, 'x', label='std') plt.legend() plt.title(f'N={i}') plt.xlabel('Action Index') plt.ylabel('Average Action Ocurrency [%]') plt.xticks(axi, ax) mean_equals = np.array([np.mean(c) for c in equals_counts_total]) std_equals = np.array([np.std(c) for c in equals_counts_total]) plt.figure() plt.plot(mean_equals[aux_range]*100, '*',label='mean') plt.plot(std_equals[aux_range]*100, 'x', label='std') plt.legend() plt.xlabel('Amount of D2D Devices') plt.ylabel('Average Equal Actions Ocurrency [%]') plt.xticks(xi, aux_range) plt.show()
[ "lucasbaiao@gmail.com" ]
lucasbaiao@gmail.com
36cfad3cd196894fc61a19c79f88a05598a87dd2
207d7f6d16a19bc78e27881841d7088b3eabc3c2
/day5/homework-ATM/modules/repay.py
ab498edd2effcf53edd276c143455289193ea672
[]
no_license
pangguoping/python-study
b0c00f73177ec86148d06f780556a4340c45e1a8
769e828d41403b89d101c2ff915699bba91390cd
refs/heads/master
2021-01-20T19:33:52.601637
2017-03-07T23:34:04
2017-03-07T23:34:04
63,377,955
0
0
null
null
null
null
UTF-8
Python
false
false
890
py
#!/usr/bin/env python # -*- coding:utf-8 -*- # Auther: pangguoping import json import os from conf import setting from modules.write_log import write_record #还款函数 def repay(card_num,**userinfo): qiankuan = userinfo['credit'] - userinfo['balance'] print('现在还欠:',qiankuan) fee = int(input("请输入还款金额:")) if fee <= qiankuan: userinfo['balance'] += fee balance = userinfo['balance'] #log(card_num,'信用卡还款',+fee,balance,**userinfo) json.dump(userinfo, open(os.path.join(setting.USER_DIR_FOLDER, card_num, "basic_info.json"), 'w', encoding='utf-8')) write_record('%s - 信用卡账户还款:%f;本月额度:%f' % ("还款", fee, balance), card_num) print('你成功还款%d,当前可用额度%s' %(fee,balance)) else: print('输入还款金额错误')
[ "work2312@163.com" ]
work2312@163.com