text
string
size
int64
token_count
int64
import sys print("Hello world, I am",sys.executable, sys.version) x=input("Dame un numero mayor que cero: ") x = int(x) if x < 0: print('Negative changed to zero') x = 0 print(f"El valor final de x es: {x}")
218
85
def convertToHEXForChar(charList): convertedCharList = [] for message in charList: convertedCharList.append(ord(message)) return convertedCharList def displayChar(line, *args): concatedList = [] for argItem in args: concatedList.extend(argItem) print(len(concatedList)) for message in concatedList: print(message) def main(): displayChar(0, [0x00], convertToHEXForChar("! Rasbperry Pi")) main()
436
146
import os as _os from enum import Enum as _Enum from datetime import datetime as _datetime import time as _time from Acquire.Service import call_function as _call_function from Acquire.Service import Service as _Service from Acquire.ObjectStore import bytes_to_string as _bytes_to_string from Acquire.ObjectStore import string_to_bytes as _string_to_bytes from Acquire.Crypto import PrivateKey as _PrivateKey from Acquire.Crypto import PublicKey as _PublicKey from ._qrcode import create_qrcode as _create_qrcode from ._qrcode import has_qrcode as _has_qrcode from ._errors import UserError, LoginError # If we can, import socket to get the hostname and IP address try: import socket as _socket _has_socket = True except: _has_socket = False __all__ = ["User", "username_to_uid", "uid_to_username", "get_session_keys"] class _LoginStatus(_Enum): EMPTY = 0 LOGGING_IN = 1 LOGGED_IN = 2 LOGGED_OUT = 3 ERROR = 4 def _get_identity_url(): """Function to discover and return the default identity url""" return "http://130.61.60.88:8080/t/identity" def _get_identity_service(identity_url=None): """Function to return the identity service for the system""" if identity_url is None: identity_url = _get_identity_url() privkey = _PrivateKey() response = _call_function(identity_url, response_key=privkey) try: service = _Service.from_data(response["service_info"]) except: raise LoginError("Have not received the identity service info from " "the identity service at '%s' - got '%s'" % (identity_url, response)) if not service.is_identity_service(): raise LoginError( "You can only use a valid identity service to log in! " "The service at '%s' is a '%s'" % (identity_url, service.service_type())) if identity_url != service.service_url(): service.update_service_url(identity_url) return service def uid_to_username(user_uid, identity_url=None): """Function to return the username for the passed uid""" if identity_url is None: identity_url = _get_identity_url() response = _call_function(identity_url, "whois", user_uid=str(user_uid)) return response["username"] def username_to_uid(username, identity_url=None): """Function to return the uid for the passed username""" if identity_url is None: identity_url = _get_identity_url() response = _call_function(identity_url, "whois", username=str(username)) return response["user_uid"] def get_session_keys(username=None, user_uid=None, session_uid=None, identity_url=None): """Function to return the session keys for the specified user""" if username is None and user_uid is None: raise ValueError("You must supply either the username or user_uid!") if session_uid is None: raise ValueError("You must supply a valid UID for a login session") if identity_url is None: identity_url = _get_identity_url() response = _call_function(identity_url, "whois", username=username, user_uid=user_uid, session_uid=session_uid) try: response["public_key"] = _PublicKey.from_data(response["public_key"]) except: pass try: response["public_cert"] = _PublicKey.from_data(response["public_cert"]) except: pass return response class User: """This class holds all functionality that would be used by a user to authenticate with and access the service. This represents a single client login, and is the user-facing part of Acquire """ def __init__(self, username, identity_url=None): """Construct a null user""" self._username = username self._status = _LoginStatus.EMPTY self._identity_service = None if identity_url: self._identity_url = identity_url self._user_uid = None def __str__(self): return "User(name='%s', status=%s)" % (self.username(), self.status()) def __enter__(self): """Enter function used by 'with' statements'""" pass def __exit__(self, exception_type, exception_value, traceback): """Ensure that we logout""" self.logout() def __del__(self): """Make sure that we log out before deleting this object""" self.logout() def _set_status(self, status): """Internal function used to set the status from the string obtained from the LoginSession""" if status == "approved": self._status = _LoginStatus.LOGGED_IN elif status == "denied": self._set_error_state("Permission to log in was denied!") elif status == "logged_out": self._status = _LoginStatus.LOGGED_OUT def username(self): """Return the username of the user""" return self._username def uid(self): """Return the UID of this user. This uniquely identifies the user across all systems """ if self._user_uid is None: self._user_uid = username_to_uid(self.username(), self.identity_service_url()) return self._user_uid def status(self): """Return the current status of this user""" return self._status def _check_for_error(self): """Call to ensure that this object is not in an error state. If it is in an error state then raise an exception""" if self._status == _LoginStatus.ERROR: raise LoginError(self._error_string) def _set_error_state(self, message): """Put this object into an error state, displaying the passed message if anyone tries to use this object""" self._status = _LoginStatus.ERROR self._error_string = message def session_key(self): """Return the session key for the current login session""" self._check_for_error() try: return self._session_key except: return None def signing_key(self): """Return the signing key used for the current login session""" self._check_for_error() try: return self._signing_key except: return None def identity_service(self): """Return the identity service info object for the identity service used to validate the identity of this user """ if self._identity_service: return self._identity_service self._identity_service = _get_identity_service( self.identity_service_url()) return self._identity_service def identity_service_url(self): """Return the URL to the identity service. This is the full URL to the service, minus the actual function to be called, e.g. https://function_service.com/t/identity """ self._check_for_error() try: return self._identity_url except: # return the default URL - this should be discovered... return _get_identity_url() def login_url(self): """Return the URL that the user must connect to to authenticate this login session""" self._check_for_error() try: return self._login_url except: return None def login_qr_code(self): """Return a QR code of the login URL that the user must connect to to authenticate this login session""" self._check_for_error() try: return self._login_qrcode except: return None def session_uid(self): """Return the UID of the current login session. Returns None if there is no valid login session""" self._check_for_error() try: return self._session_uid except: return None def is_empty(self): """Return whether or not this is an empty login (so has not been used for anything yet...""" return self._status == _LoginStatus.EMPTY def is_logged_in(self): """Return whether or not the user has successfully logged in""" return self._status == _LoginStatus.LOGGED_IN def is_logging_in(self): """Return whether or not the user is in the process of loggin in""" return self._status == _LoginStatus.LOGGING_IN def logout(self): """Log out from the current session""" if self.is_logged_in() or self.is_logging_in(): identity_url = self.identity_service_url() if identity_url is None: return # create a permission message that can be signed # and then validated by the user permission = "Log out request for %s" % self._session_uid signature = self.signing_key().sign(permission) print("Logging out %s from session %s" % (self._username, self._session_uid)) result = _call_function( identity_url, "logout", args_key=self.identity_service().public_key(), username=self._username, session_uid=self._session_uid, permission=permission, signature=_bytes_to_string(signature)) print(result) self._status = _LoginStatus.LOGGED_OUT return result def register(self, password, identity_url=None): """Request to register this user with the identity service running at 'identity_url', using the supplied 'password'. This will return a QR code that you must use immediately to add this user on the identity service to a QR code generator""" if self._username is None: return None if identity_url is None: identity_url = _get_identity_url() privkey = _PrivateKey() result = _call_function( identity_url, "register", args_key=self.identity_service().public_key(), response_key=privkey, public_cert=self.identity_service().public_certificate(), username=self._username, password=password) try: provisioning_uri = result["provisioning_uri"] except: raise UserError( "Cannot register the user '%s' on " "the identity service at '%s'!" % (self._username, identity_url)) # return a QR code for the provisioning URI return (provisioning_uri, _create_qrcode(provisioning_uri)) def request_login(self, login_message=None): """Request to authenticate as this user. This returns a login URL that you must connect to to supply your login credentials If 'login_message' is supplied, then this is passed to the identity service so that it can be displayed when the user accesses the login page. This helps the user validate that they have accessed the correct login page. Note that if the message is None, then a random message will be generated. """ self._check_for_error() if not self.is_empty(): raise LoginError("You cannot try to log in twice using the same " "User object. Create another object if you want " "to try to log in again.") # first, create a private key that will be used # to sign all requests and identify this login session_key = _PrivateKey() signing_key = _PrivateKey() args = {"username": self._username, "public_key": session_key.public_key().to_data(), "public_certificate": signing_key.public_key().to_data(), "ipaddr": None} # get information from the local machine to help # the user validate that the login details are correct if _has_socket: hostname = _socket.gethostname() ipaddr = _socket.gethostbyname(hostname) args["ipaddr"] = ipaddr args["hostname"] = hostname if login_message is None: login_message = "User '%s' in process '%s' wants to log in..." % \ (_os.getlogin(), _os.getpid()) args["message"] = login_message result = _call_function( self.identity_service_url(), "request_login", args_key=self.identity_service().public_key(), response_key=session_key, public_cert=self.identity_service().public_certificate(), username=self._username, public_key=session_key.public_key().to_data(), public_certificate=signing_key.public_key().to_data(), ipaddr=None, message=login_message) # look for status = 0 try: status = int(result["status"]) except: status = -1 try: message = result["message"] except: message = str(result) try: prune_message = result["prune_message"] print("Pruning old sessions...\n%s" % "\n".join(prune_message)) except: pass if status != 0: error = "Failed to login. Error = %d. Message = %s" % \ (status, message) self._set_error_state(error) raise LoginError(error) try: login_url = result["login_url"] except: login_url = None if login_url is None: error = "Failed to login. Could not extract the login URL! " \ "Result is %s" % (str(result)) self._set_error_state(error) raise LoginError(error) try: session_uid = result["session_uid"] except: session_uid = None if session_uid is None: error = "Failed to login. Could not extract the login " \ "session UID! Result is %s" % (str(result)) self._set_error_state(error) raise LoginError(error) # now save all of the needed data self._login_url = result["login_url"] self._session_key = session_key self._signing_key = signing_key self._session_uid = session_uid self._status = _LoginStatus.LOGGING_IN self._user_uid = result["user_uid"] qrcode = None if _has_qrcode(): try: self._login_qrcode = _create_qrcode(self._login_url) qrcode = self._login_qrcode except: pass return (self._login_url, qrcode) def _poll_session_status(self): """Function used to query the identity service for this session to poll for the session status""" identity_url = self.identity_service_url() if identity_url is None: return result = _call_function(identity_url, "get_status", username=self._username, session_uid=self._session_uid) # look for status = 0 try: status = int(result["status"]) except: status = -1 try: message = result["message"] except: message = str(result) if status != 0: error = "Failed to query identity service. Error = %d. " \ "Message = %s" % (status, message) self._set_error_state(error) raise LoginError(error) # now update the status... status = result["session_status"] self._set_status(status) def wait_for_login(self, timeout=None, polling_delta=5): """Block until the user has logged in. If 'timeout' is set then we will wait for a maximum of that number of seconds This will check whether we have logged in by polling the identity service every 'polling_delta' seconds. """ self._check_for_error() if not self.is_logging_in(): return self.is_logged_in() polling_delta = int(polling_delta) if polling_delta > 60: polling_delta = 60 elif polling_delta < 1: polling_delta = 1 if timeout is None: # block forever.... while True: self._poll_session_status() if self.is_logged_in(): return True elif not self.is_logging_in(): return False _time.sleep(polling_delta) else: # only block until the timeout has been reached timeout = int(timeout) if timeout < 1: timeout = 1 start_time = _datetime.now() while (_datetime.now() - start_time).seconds < timeout: self._poll_session_status() if self.is_logged_in(): return True elif not self.is_logging_in(): return False _time.sleep(polling_delta) return False
17,670
4,658
#! /usr/bin/python2.7 # -*- coding: utf-8 -*- """ Created on 2017-4-06 @module: MyMAIL @used: send mail """ import smtplib import mimetypes from email.mime.text import MIMEText from email.mime.multipart import MIMEMultipart from email.mime.image import MIMEImage from MyLOG import MyLog from botasky.utils.MyFILE import project_abdir, recursiveSearchFile logConfig = recursiveSearchFile(project_abdir, '*logConfig.ini')[0] mylog = MyLog(logConfig, 'MyMAIL.py') logger = mylog.outputLog() __all__ = ['MyMail'] __author__ = 'zhihao' mail_info = {'mail_host': 'smtp.163.com', 'mail_user': '15895890858', 'mail_pass': 'zhi@hao@111', 'mail_postfix': '163.com'} class MyMail(): ''' used : send mail ''' def __init__(self, mail_info): ''' used : init mail :param mail_info: smtp server config ''' self.mail_info = mail_info def send_mail(self, to_list, mail_type, subject, content, attachment_list, img_list): ''' used : send mail :param to_list: target mail adresses :param mail_type: plain or html :param subject: title :param content: main body :param attachment_list: attachment :param img_list: picture :return: ''' my_adress = "0905zhihao" + "<" + self.mail_info['mail_user'] + "@" + self.mail_info['mail_postfix'] + ">" msg = MIMEMultipart() msg['Subject'] = subject msg['From'] = my_adress msg['To'] = ";".join(to_list) #main text if mail_type == 'plain' or mail_type == 'html': try: body_msg = MIMEText(content, _subtype=mail_type, _charset='gb2312') msg.attach(body_msg) exec_info = "[action]:init msg" \ "[status]:OK" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(Subject=msg['Subject'], From=msg['From'], To=msg['To']) logger.info(exec_info) except Exception, e: print Exception, ":", e error_msg = "[action]:init msg" \ "[status]:FAIL" \ "[Errorcode]:{e}" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(Subject=msg['Subject'], From=msg['From'], To=msg['To'], e=e) logger.error(error_msg) else: error_msg = "[action]:send mail_type" \ "[status]:FAIL" \ "[Errorcode]mail_type is not format" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(Subject=msg['Subject'], From=msg['From'], To=msg['To']) print error_msg logger.info(error_msg) #attachment if attachment_list == '' or len(attachment_list) == 0: pass else: for attachment in attachment_list: try: att = MIMEText(open(attachment, 'rb').read(), 'base64', 'gb2312') att["Content-Type"] = 'application/octet-stream' #display name att["Content-Disposition"] = 'attachment; filename="'+attachment+'\"\'' msg.attach(att) exec_info = "[action]:add attachment" \ "[status]:OK" \ "[attachment]:{attachment}" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(attachment=attachment, Subject=msg['Subject'], From=msg['From'], To=msg['To']) logger.info(exec_info) except Exception, e: print Exception, ":", e error_msg = "[action]:add attachment" \ "[status]:FAIL" \ "[Errorcode]:{e}" \ "[attachment]={attachment}" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(Subject=msg['Subject'], From=msg['From'], attachment=attachment, To=msg['To'], e=e) logger.error(error_msg) #img if img_list == '' or len(img_list) == 0: pass else: for image_adress in img_list: try: image = MIMEImage(open(image_adress, 'rb').read()) image.add_header('Content-ID', '<image1>') msg.attach(image) exec_info = "[action]:add image" \ "[status]:OK" \ "[image]:{image}" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(image=image_adress, Subject=msg['Subject'], From=msg['From'], To=msg['To']) logger.info(exec_info) except Exception, e: print Exception, ":", e error_msg = "[action]:add image" \ "[status]:FAIL" \ "[Errorcode]:{e}" \ "[image]:{image}" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(Subject=msg['Subject'], From=msg['From'], image=image_adress, To=msg['To'], e=e) logger.error(error_msg) #send mail try: server = smtplib.SMTP() server.connect(self.mail_info['mail_host']) server.login(self.mail_info['mail_user'], self.mail_info['mail_pass']) server.sendmail(msg['from'], msg['to'], msg.as_string()) server.quit() exec_info = "[action]:send mail" \ "[status]:OK" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(Subject=msg['Subject'], From=msg['From'],To=msg['To']) logger.info(exec_info) except Exception, e: print Exception, ":", e error_msg = "[action]:send mail" \ "[status]:FAIL" \ "[Errorcode]:{e}" \ "[Subject]:{Subject}" \ "[From]:{From}" \ "[To]:{To}".format(Subject=msg['Subject'], From=msg['From'], To=msg['To'], e=e) logger.error(error_msg) if __name__ == '__main__': ''' mail_info = {'mail_host': 'smtp.163.com', 'mail_user': '15002283621', 'mail_pass': 'zhihao1206', 'mail_postfix': '163.com'} #to_list = ['15002283621@163.com'] to_list = ['1204207658@qq.com'] subject = 'xxxxxxxxxxxxx' content = 'xxxxxxxxxxxxx' #attachment_list = ['F:\img\img.rar', 'F:\img\img2.rar'] attachment_list = [] #img_list = ['F:\img\\1025.jpg', 'F:\img\\1041.jpg'] img_list = [] mail = MyMail(mail_info) mail.send_mail(to_list, 'plain', subject, content, attachment_list, img_list) ''' import MyMAIL help(MyMAIL)
8,007
2,413
import json import traceback from db_adapter.exceptions import DatabaseAdapterError from db_adapter.logger import logger """ Source JSON Object would looks like this e.g.: { 'model' : 'wrfSE', 'version' : 'v3', 'parameters': { } } { 'model' : 'OBS_WATER_LEVEL', 'version' : '', 'parameters': { "CHANNEL_CELL_MAP" : { "594" : "Wellawatta", "1547": "Ingurukade", "3255": "Yakbedda", "3730": "Wellampitiya", "7033": "Janakala Kendraya" }, "FLOOD_PLAIN_CELL_MAP": { } } } """ def get_source_by_id(pool, id_): """ Retrieve source by id :param pool: database connection pool :param id_: source id :return: Source if source exists in the database, else None """ connection = pool.connection() try: with connection.cursor() as cursor: sql_statement = "SELECT * FROM `source` WHERE `id`=%s" row_count = cursor.execute(sql_statement, id_) if row_count > 0: return cursor.fetchone() else: return None except Exception as exception: error_message = "Retrieving source with source_id {} failed".format(id_) logger.error(error_message) traceback.print_exc() raise exception finally: if connection is not None: connection.close() def get_source_id(pool, model, version) -> str: """ Retrieve Source id :param pool: database connection pool :param model: :param version: :return: str: source id if source exists in the database, else None """ connection = pool.connection() try: with connection.cursor() as cursor: sql_statement = "SELECT `id` FROM `source` WHERE `model`=%s and `version`=%s" row_count = cursor.execute(sql_statement, (model, version)) if row_count > 0: return cursor.fetchone()['id'] else: return None except Exception as exception: error_message = "Retrieving source id: model={} and version={} failed.".format(model, version) logger.error(error_message) traceback.print_exc() raise exception finally: if connection is not None: connection.close() def add_source(pool, model, version, parameters=None): """ Insert sources into the database :param pool: database connection pool :param model: string :param version: string :param parameters: JSON :return: True if the source has been added to the "Source' table of the database, else False """ connection = pool.connection() try: if get_source_id(pool=pool, model=model, version=version) is None: with connection.cursor() as cursor: sql_statement = "INSERT INTO `source` (`model`, `version`, `parameters`) VALUES ( %s, %s, %s)" row_count = cursor.execute(sql_statement, (model, version, json.dumps(parameters))) connection.commit() return True if row_count > 0 else False else: logger.info("Source with model={} and version={} already exists in the database".format(model, version)) return False except Exception as exception: connection.rollback() error_message = "Insertion of source: model={}, version={} and parameters={} failed".format(model, version, parameters) logger.error(error_message) traceback.print_exc() raise exception finally: if connection is not None: connection.close() def add_sources(sources, pool): """ Add sources into Source table :param sources: list of json objects that define source attributes e.g.: { 'model' : 'wrfSE', 'version' : 'v3', 'parameters': { } } { 'model' : 'OBS_WATER_LEVEL', 'version' : '', 'parameters': { "CHANNEL_CELL_MAP" : { "594" : "Wellawatta", "1547": "Ingurukade", "3255": "Yakbedda", "3730": "Wellampitiya", "7033": "Janakala Kendraya" }, "FLOOD_PLAIN_CELL_MAP": { } } } :return: """ for source in sources: print(add_source(pool=pool, model=source.get('model'), version=source.get('version'), parameters=source.get('parameters'))) print(source.get('model')) def delete_source(pool, model, version): """ Delete source from Source table, given model and version :param pool: database connection pool :param model: str :param version: str :return: True if the deletion was successful """ connection = pool.connection() try: with connection.cursor() as cursor: sql_statement = "DELETE FROM `source` WHERE `model`=%s and `version`=%s" row_count = cursor.execute(sql_statement, (model, version)) connection.commit() if row_count > 0: return True else: logger.info("There's no record of source in the database with model={} and version={}".format(model, version)) return False except Exception as exception: connection.rollback() error_message = "Deleting source with model={} and version={} failed.".format(model, version) logger.error(error_message) traceback.print_exc() raise exception finally: if connection is not None: connection.close() def delete_source_by_id(pool, id_): """ Delete source from Source table by id :param pool: database connection pool :param id_: :return: True if the deletion was successful, else False """ connection = pool.connection() try: with connection.cursor() as cursor: sql_statement = "DELETE FROM `source` WHERE `id`=%s" row_count = cursor.execute(sql_statement, id_) connection.commit() if row_count > 0 : return True else: logger.info("There's no record of source in the database with the source id {}".format(id_)) return False except Exception as exception: connection.rollback() error_message = "Deleting source with id {} failed.".format(id_) logger.error(error_message) traceback.print_exc() raise exception finally: if connection is not None: connection.close() def get_source_parameters(pool, model, version): """ Retrieve Source parameters :param pool: database connection pool :param model: :param version: :return: str: json object parameters if source exists in the database, else None """ connection = pool.connection() try: with connection.cursor() as cursor: sql_statement = "SELECT `parameters` FROM `source` WHERE `model`=%s and `version`=%s" row_count = cursor.execute(sql_statement, (model, version)) if row_count > 0: return cursor.fetchone()['parameters'] else: return None except Exception as exception: error_message = "Retrieving source parameters: model={} and version={} failed.".format(model, version) logger.error(error_message) traceback.print_exc() raise exception finally: if connection is not None: connection.close()
7,694
2,111
import unittest from katas.kyu_7.binary_addition import add_binary class AddBinaryTestCase(unittest.TestCase): def test_equals(self): self.assertEqual(add_binary(1, 1), '10') def test_equals_2(self): self.assertEqual(add_binary(0, 1), '1') def test_equals_3(self): self.assertEqual(add_binary(1, 0), '1') def test_equals_4(self): self.assertEqual(add_binary(2, 2), '100') def test_equals_5(self): self.assertEqual(add_binary(51, 12), '111111')
511
201
# Copyright 2013-2020 Lawrence Livermore National Security, LLC and other # Spack Project Developers. See the top-level COPYRIGHT file for details. # # SPDX-License-Identifier: (Apache-2.0 OR MIT) from spack import * class Netdata(AutotoolsPackage): """Real-time performance monitoring, done right!""" homepage = "https://www.netdata.cloud/" url = "https://github.com/netdata/netdata/archive/v1.22.1.tar.gz" version('1.22.1', sha256='6efd785eab82f98892b4b4017cadfa4ce1688985915499bc75f2f888765a3446') depends_on('m4', type='build') depends_on('autoconf', type='build') depends_on('automake', type='build') depends_on('libtool', type='build') depends_on('libuv') depends_on('uuid')
739
301
import flask import flask.ext.sqlalchemy import flask.ext.restless # Create the Flask application and the Flask-SQLAlchemy object. app = flask.Flask(__name__) app.config['DEBUG'] = True app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:////tmp/test.db' db = flask.ext.sqlalchemy.SQLAlchemy(app) # Create your Flask-SQLALchemy models as usual but with the following two # (reasonable) restrictions: # 1. They must have a primary key column of type sqlalchemy.Integer or # type sqlalchemy.Unicode. # 2. They must have an __init__ method which accepts keyword arguments for # all columns (the constructor in flask.ext.sqlalchemy.SQLAlchemy.Model # supplies such a method, so you don't need to declare a new one). class Person(db.Model): id = db.Column(db.Integer, primary_key=True) name = db.Column(db.Unicode, unique=True) birth_date = db.Column(db.Date) class Computer(db.Model): id = db.Column(db.Integer, primary_key=True) name = db.Column(db.Unicode, unique=True) vendor = db.Column(db.Unicode) purchase_time = db.Column(db.DateTime) owner_id = db.Column(db.Integer, db.ForeignKey('person.id')) owner = db.relationship('Person', backref=db.backref('computers', lazy='dynamic')) # Create the database tables. db.create_all() # Create the Flask-Restless API manager. manager = flask.ext.restless.APIManager(app, flask_sqlalchemy_db=db) # Create API endpoints, which will be available at /api/<tablename> by # default. Allowed HTTP methods can be specified as well. manager.create_api(Person, methods=['GET', 'POST', 'DELETE']) manager.create_api(Computer, methods=['GET']) # start the flask loop app.run()
1,727
555
""" This is a script that takes a calculates the value of a cryptocurrency portfolio It uses JSON in the with quantities of different cryptocurrencies in the form { "ticker" : volume, "ticker" : volume } gets the live price from an API endpoint and returns the price of each item in the portfolio and the total It also writes these into a sqlite3 database for future reference with a timestamp """ import os, logging, argparse, json import sqlite3 import requests import datetime import time """ TODO: Error handling & logging """ # Need API from https://min-api.cryptocompare.com/ API_KEY = os.getenv("CRYPTO_API_KEY") HEADER = {"authorization": f"Apikey {API_KEY}"} # Taken from https://docs.python.org/3/library/sqlite3.html#registering-an-adapter-callable def adapt_datetime(ts): return time.mktime(ts.timetuple()) sqlite3.register_adapter(datetime.datetime, adapt_datetime) def setup_db(db_path): """ Initialises a local sqlite3 database and create the table required to hold data. Parameters ------------- db_path string : A filepath to a target sqlite database Returns ------------- con: Connection : Returns a connection to that database """ con = sqlite3.connect(db_path) # Create table with con: con.execute( """CREATE TABLE IF NOT EXISTS CRYPTO_PRICE (DATE timestamp, TICKER text, QTY real, PRICE real, VALUE real )""" ) logging.info("Database and table created") return con def insert_into_db(connection, ticker, price, dict): """ Writes crypto price data to specified sqlite3 database Parameters ------------- connection string : Connection to sqlite3 database output of setup_db() fn ticker string : String of the Ticker for a cryptocurrency e.g. BTC price float : Price of a cryptocurrency dict Dictionary : Dictionary loaded from portfolio JSON. output of parse_json() fn """ now = datetime.datetime.now() with connection as con: if ticker != "SUM": con.execute( """insert into CRYPTO_PRICE values (?,?,?,?,?)""", (now, ticker, dict[ticker], price, price * dict[ticker]), ) else: con.execute( """insert into CRYPTO_PRICE values (?,?,?,?,?)""", (now, ticker, 0, price, price), ) logging.info(f"Inserted {ticker} values into database") def parse_json(json_path): """ Loads portfolio in JSON into a python dictionary. Parameters ------------- json_path string : Path to portfolio JSON described in header documentation Returns ------------- crypto_dict Dictionary : Dictionary loaded from portfolio json. output of parse_json() fn """ with open(json_path) as j: crypto_dict = json.load(j) return crypto_dict def get_price(ticker): """ Returns the live price of a unit a cryptocurrency in GBP. Parameters ------------- ticker string : String of the Ticker for a cryptocurrency e.g. BTC Returns ------------- price float : Price of a cryptocurrency """ API_ENDPOINT = f"https://min-api.cryptocompare.com/data/price?fsym={ticker}&tsyms=GBP" response = requests.get(API_ENDPOINT, headers=HEADER) price = response.json()["GBP"] return price def main(json_path, connection): crypto_dict = parse_json(json_path) wallet_dict = dict() for key_ in crypto_dict.keys(): price = get_price(key_) print(f"{key_}: £{round(price*crypto_dict[key_],2)}") wallet_dict[key_] = price * crypto_dict[key_] insert_into_db(connection, key_, price, crypto_dict) insert_into_db(connection, "SUM", sum(wallet_dict.values()), crypto_dict) print(f"Total: £{sum(wallet_dict.values())}") return sum(wallet_dict.values()) if __name__ == "__main__": logging.basicConfig( level=logging.INFO, format="[%(levelname)s: %(asctime)s] %(filename)s, %(funcName)s, line %(lineno)d : %(message)s" ) parser = argparse.ArgumentParser() parser.add_argument( "--filepath_in", required=False, type=str, default=os.getcwd(), help="Filepath to json holding volumes of crypto" ) parser.add_argument( "--db_path", required=False, type=str, default=f"{os.getcwd()}/crypto.db", help="Filepath to sqlite database" ) args = parser.parse_args() FILEPATH_IN = args.filepath_in con = setup_db(args.db_path) main(FILEPATH_IN, con) con.close()
4,687
1,450
#np43.py #43. 名詞を含む文節が動詞を含む文節に係るものを抽出 """名詞を含む文節が,動詞を含む文節に係るとき, これらをタブ区切り形式で抽出せよ.ただし,句読点などの記号は出力しないようにせよ. """ import re # 区切り文字 separator = re.compile('\t|,') # かかりうけ kakari = re.compile(r'''(?:\*\s\d+\s) # キャプチャ対象外 (-?\d+) # 数字(係り先) ''', re.VERBOSE) class Morph: def __init__(self, morph): #タブとカンマで分割 tab = separator.split(morph) self.surface = tab[0] # 表層形(surface) self.base = tab[7] # 基本形(base) self.pos = tab[1] # 品詞(pos) self.pos1 = tab[2] # 品詞細分類1(pos1) class Chunk: def __init__(self, morphs, dst): self.morphs = morphs self.srcs = [] # 係り元文節インデックス番号のリスト self.dst = dst # 係り先文節インデックス番号 self.phrase = ''.join([morph.surface for morph in morphs if morph.pos!='記号']) # 係り元を代入,Chunkリストを文のリストを追加 def append_sentence(chunks, sentences): # 係り元を代入 for i, chunk in enumerate(chunks): if chunk.dst != -1: chunks[chunk.dst].srcs.append(i) sentences.append(chunks) return sentences import np41sss sentences = np41sss.Ai_morphs() sentence = sentences[1] for chunk in sentence: if int(chunk.dst) != -1: modifier = ''.join([morph.surface if morph.pos != '記号' else '' for morph in chunk.morphs]) modifier_pos = [morph.pos for morph in chunk.morphs]#記号等々の排除chunk,morphs内 modifiee = ''.join([morph.surface if morph.pos != '記号' else '' for morph in sentence[int(chunk.dst)].morphs]) modifiee_pos = [morph.pos for morph in sentence[int(chunk.dst)].morphs] if '名詞' in modifier_pos and '動詞' in modifiee_pos:#動詞と名詞を含むものを取り出す print(modifier, modifiee, sep='\t')#タブ区切り
1,685
865
from presentation.models import Author, Follower, Post, Comment from django.shortcuts import get_object_or_404 from presentation.Serializers.comment_serializer import CommentSerializer from rest_framework import viewsets, status from django.http import JsonResponse from rest_framework.response import Response import uuid from urllib.parse import urlparse from . import urlutil ''' URL: ://service/author/{author_id}/posts/{post_id}/comments access GET get comments of the post POST if you post an object of “type”:”comment”, it will add your comment to the post ''' def getAuthorIDFromRequestURL(request, id): host = urlutil.getSafeURL(request.build_absolute_uri()) author_id = f"{host}/author/{id}" return author_id def getPostIDFromRequestURL(request, id): post_id = f"/posts/{id}" return post_id def getCommentIDFromRequestURL(request, id): comment_id = f"/comments/{id}" return comment_id class CommentViewSet(viewsets.ModelViewSet): serializer_class = CommentSerializer queryset = Comment.objects.all() # GET a list of comments of the post def list(self, request, *args, **kwargs): author_id = getAuthorIDFromRequestURL( request, self.kwargs['author_id']) post_id = getPostIDFromRequestURL( request, self.kwargs['post_id']) post_id = author_id + post_id queryset = Comment.objects.filter(post=post_id) if queryset.exists(): comments = list(queryset.values()) # May have mistakes here, do we need to change comment model? return JsonResponse(comments, safe=False) else: return JsonResponse([], safe=False) # GET a single comment using comment_id def retrieve(self, request, *args, **kwargs): comment_id = request.build_absolute_uri()[:-1] queryset = Comment.objects.get(id=comment_id) if queryset.exists(): serializer = CommentSerializer(queryset) return Response(serializer.data, 200) else: return Response({"msg": "No comment for given id"}, 404) # POST a new comment under a post def create(self, request, *args, **kwargs): request_data = request.data.copy() # assume the id of the commmenter is part of the data # CHANGE THIS LATER! commenter_id = request_data.get('author', None) author_id = getAuthorIDFromRequestURL( request, self.kwargs['author_id']) post_id = getPostIDFromRequestURL( request, self.kwargs['post_id']) post_id = author_id + post_id comment = request_data.get('comment', None) content_type = request_data.get('contentType', None) # create comment id cuuid = str(uuid.uuid4().hex) comment_id = f"{post_id}/comments/{cuuid}" comment_data = {'type': 'comment', 'author': commenter_id, 'comment': comment, 'contentType': content_type, 'post': post_id, 'id': comment_id} serializer = self.serializer_class(data=comment_data) if serializer.is_valid(): serializer.save() return Response(serializer.data, 200) else: return Response(serializer.errors, status=400) def delete(self, request, *args, **kwargs): author_id = getAuthorIDFromRequestURL( request, self.kwargs['author_id']) post_id = getPostIDFromRequestURL( request, self.kwargs['post_id']) post_id = author_id + post_id comments = get_object_or_404(Comment, post=post_id) comment_id = getCommentIDFromRequestURL( request, self.kwargs['comment_id']) comment_id = post_id + comment_id comment = get_object_or_404(Comment, id=comment_id) # Possible mistake? try: comment.delete() except ValueError: return Response("No such a comment. Deletion fails.", 500) return Response("Delete successful")
4,026
1,142
# encoding=utf-8 import vim import re import sys import os.path import argparse import shlex from subprocess import Popen, PIPE from itertools import chain from vim_pandoc.utils import plugin_enabled_modules, ensure_string from vim_pandoc.bib.vim_completer import find_bibfiles from vim_pandoc.helpparser import PandocInfo class PandocCommand(object): def __init__(self): self.pandoc_info = PandocInfo(vim.vars["pandoc#command#path"]) self.formats_table = {} self.build_formats_table() self._output_file_path = None self._run_command = None self._out = None def build_formats_table(self): for i in self.pandoc_info.output_formats: if i in ("asciidoc", "plain"): self.formats_table[i] = "txt" elif i in ("beamer", "pdf"): self.formats_table[i] = "pdf" elif i in ("dzslides", "html", "html5", "mediawiki", "revealjs", "s5", "slideous", "slidy"): self.formats_table[i] = "html" elif i in ("markdown", "gfm", "markdown_github", "markdown_mmd", "markdown_phpextra", "markdown_strict"): self.formats_table[i] = "md" elif i in ("odt", "opendocument"): self.formats_table[i] = "odt" elif i == "native": self.formats_table[i] = "hs" elif i == "texinfo": self.formats_table[i] = "info" elif i == "latex": self.formats_table[i] = "tex" else: self.formats_table[i] = i if "latex" in self.formats_table or "beamer" in self.formats_table and "pdf" not in self.formats_table: self.formats_table["pdf"] = "pdf" def __call__(self, args, should_open): largs = shlex.split(args) if largs == []: largs = ['html'] # make sure we pass an output format p = self.pandoc_info.build_argument_parser() c_vars = vars(p.parse_args(largs)) # Infer arguments from vim environment # a) bibliographies if 'bibliographies' in plugin_enabled_modules(): local_bibs = vim.eval('b:pandoc_biblio_bibs') found_bibs = find_bibfiles() if local_bibs or found_bibs and not c_vars['bibliography']: c_vars['bibliography'] = [] if local_bibs: c_vars['bibliography'].extend(local_bibs) if found_bibs: c_vars['bibliography'].extend(found_bibs) # Now, we must determine what are our input and output files # a) First, let's see what is the desired output format... output_format = c_vars['output_format'] \ if self.pandoc_info.is_valid_output_format(c_vars['output_format']) \ or c_vars['output_format'] == 'pdf' \ else "html" # overwrite --to with this value # 'pdf' is not a valid output format, we pass it to -o instead) if output_format != 'pdf': c_vars['to'] = output_format if output_format == 'pdf': # pdf engine if self.pandoc_info.version >= '2': engine_option = 'pdf_engine' else: engine_option = 'latex_engine' if not c_vars[engine_option]: try: # try a buffer local engine engine_var = ensure_string(vim.current.buffer.vars['pandoc_command_latex_engine']) except: # use te global value engine_var = ensure_string(vim.vars['pandoc#command#latex_engine']) c_vars[engine_option] = str(engine_var) if not c_vars['output']: self._output_file_path = vim.eval('expand("%:r")') + '.' \ + self.formats_table[re.split("[-+]", output_format)[0]] c_vars['output'] = self._output_file_path else: self._output_file_path = os.path.expanduser(c_vars['output'][0]) input_arg = '"' + vim.eval('expand("%")') + '"' # Now, we reconstruct the pandoc call arglist = [] arglist.append(ensure_string(vim.vars['pandoc#compiler#command'])) arglist.append(ensure_string(vim.vars['pandoc#compiler#arguments'])) # Only consider enabled flags and arguments with values extra_arg_vars_keys = [k for k in c_vars.keys() if c_vars[k] and k != 'output_format'] for var in extra_arg_vars_keys: real_var = var.replace("_", "-") val = c_vars[var] if type(val) == list and len(val) > 1: # multiple values, repeat keys for vv in val: if type(vv) == list and type(vv[0]) == list: vv = vv[0][0] elif type(vv) == list: vv = vv[0] elif type(val) == bool: vv = None if vv: vv = os.path.expanduser(vv) arglist.append("--" + real_var + '="' + str(vv) + '"') else: arglist.append("--" + real_var) else: if type(val) == list and type(val[0]) == list: val = val[0][0] elif type(val) == list: val = val[0] elif type(val) == bool: val = None if val: val = os.path.expanduser(val) arglist.append('--' + real_var + '="' + str(val) + '"') else: arglist.append('--' + real_var) arglist.append(input_arg) self._run_command = " ".join(arglist) # execute self.execute(should_open) def execute(self, should_open): with open("pandoc.out", 'w') as tmp: # for nvim if vim.eval("has('nvim')") == '1': try: should_open_s = str(int(should_open)) except: should_open_s = '0' vim.command("call jobstart(" + \ "['"+ "','".join(shlex.split(self._run_command)) + "'], " + \ " extend({'should_open': '" + should_open_s + "'}," +\ " {'on_exit': 'pandoc#command#JobHandler'," + \ "'on_stdout': 'pandoc#command#JobHandler'," + \ "'on_stderr': 'pandoc#command#JobHandler'}))") # for vim versions with clientserver support elif vim.eval("has('clientserver')") == '1' and \ vim.eval("v:servername") != "" and \ vim.eval("executable('python')") == '1': async_runner = '"' + os.path.join(os.path.dirname(__file__), "async.py") + '"' servername_arg = "--servername=" + vim.eval("v:servername") open_arg = "--open" if should_open else "--noopen" async_command = " ".join(["python", async_runner, servername_arg, open_arg, self._run_command]) try: Popen(shlex.split(async_command), stdout=tmp, stderr=tmp) except: vim.command('echoe "vim-pandoc: could not execute pandoc asynchronously"') else: try: # fallback to synchronous execution com = Popen(shlex.split(self._run_command), stdout=tmp, stderr=tmp) com.wait() except: vim.command('echoe "vim-pandoc: could not execute pandoc"') return self.on_done(should_open, com.returncode) def on_done(self, should_open, returncode): if self._run_command and self._output_file_path: vim.command("echohl Statement") vim.command("echom 'vim-pandoc:ran " + self._run_command + "'") vim.command("echohl None") if vim.eval("g:pandoc#command#use_message_buffers") == '1' \ and returncode not in ('0', 0): vim.command("let split = &splitbelow") vim.command("set splitbelow") vim.command("5new pandoc\ output") vim.command("let &splitbelow = split") vim.command("setlocal wrap") vim.command("setlocal linebreak") vim.current.buffer[0] = "# Press q to close this" vim.current.buffer.append("> " + self._run_command) vim.command("normal! G") if vim.eval('filereadable("pandoc.out")') == '1': vim.command("silent r pandoc.out") vim.command("setlocal buftype=nofile") vim.command("setlocal nobuflisted") # pressing q on the buffer will delete it vim.command("map <buffer> q :bd<cr>") # we will highlight some elements in the buffer vim.command("syn match PandocOutputMarks /^>>/") vim.command("syn match PandocCommand /^>.*$/hs=s+1") vim.command("syn match PandocInstructions /^#.*$/") vim.command("hi! link PandocOutputMarks Operator") vim.command("hi! link PandocCommand Debug") vim.command("hi! link PandocInstructions Comment") # under windows, pandoc.out is not closed by async.py in time sometimes, # so we wait a bit if sys.platform.startswith("win"): from time import sleep sleep(1) if os.path.exists("pandoc.out"): os.remove("pandoc.out") # open file if needed # nvim's python host doesn't change the directory the same way vim does if vim.eval('has("nvim")') == '1': os.chdir(vim.eval('expand("%:p:h")')) if os.path.exists(os.path.abspath(self._output_file_path)) and should_open: # if g:pandoc#command#custom_open is defined and is a valid funcref if vim.eval("g:pandoc#command#custom_open") != "" \ and vim.eval("exists('*"+vim.eval("g:pandoc#command#custom_open")+"')") == '1': custom_command = vim.eval(vim.eval("g:pandoc#command#custom_open") \ + "('"+self._output_file_path+"')") Popen(shlex.split(custom_command)) # otherwise use platform defaults: else: if sys.platform == "darwin" or sys.platform.startswith("linux"): if sys.platform == "darwin": open_command = "open" #OSX elif sys.platform.startswith("linux"): open_command = "xdg-open" # freedesktop/linux with open(os.devnull, 'wb') as fnull: Popen([open_command, self._output_file_path], stderr=fnull) elif sys.platform.startswith("win"): Popen('cmd /c "start ' + self._output_file_path + '"') # we reset this self._output_file_path = None self._run_command = None vim.command("redraw") if returncode in ("0", 0): vim.command("echohl Statement") vim.command("echom 'vim-pandoc:ran successfully.'") vim.command("echohl None") pandoc = PandocCommand()
11,520
3,403
# # Copyright (C) 2019 Databricks, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ String functions on Koalas Series """ from typing import TYPE_CHECKING import numpy as np from pyspark.sql.types import StringType, BinaryType, BooleanType from databricks.koalas.base import _wrap_accessor_pandas if TYPE_CHECKING: import databricks.koalas as ks class StringMethods(object): """String methods for Koalas Series""" def __init__(self, series: 'ks.Series'): if not isinstance(series.spark_type, (StringType, BinaryType)): raise ValueError( "Cannot call StringMethods on type {}" .format(series.spark_type)) self._data = series self.name = self._data.name # Methods def capitalize(self) -> 'ks.Series': """ Convert Strings in the series to be capitalized. """ return _wrap_accessor_pandas( self, lambda x: x.str.capitalize(), StringType() ).alias(self.name) def lower(self) -> 'ks.Series': """ Convert strings in the Series/Index to all lowercase. """ return _wrap_accessor_pandas( self, lambda x: x.str.lower(), StringType() ).alias(self.name) def upper(self) -> 'ks.Series': """ Convert strings in the Series/Index to all uppercase. """ return _wrap_accessor_pandas( self, lambda x: x.str.upper(), StringType() ).alias(self.name) def swapcase(self) -> 'ks.Series': """ Convert strings in the Series/Index to be swapcased. """ return _wrap_accessor_pandas( self, lambda x: x.str.swapcase(), StringType() ).alias(self.name) def startswith(self, pattern, na=np.NaN) -> 'ks.Series': """ Test if the start of each string element matches a pattern. Equivalent to :func:`str.startswith`. Parameters ---------- pattern : str Character sequence. Regular expressions are not accepted. na : object, defulat NaN Object shown if element is not a string. Returns ------- Series of bool Koalas Series of booleans indicating whether the given pattern matches the start of each string element. """ return _wrap_accessor_pandas( self, lambda x: x.str.startswith(pattern, na), BooleanType() ).alias(self.name) def endswith(self, pattern, na=np.NaN) -> 'ks.Series': """ Test if the end of each string element matches a pattern. Equivalent to :func:`str.endswith`. Parameters ---------- pattern : str Character sequence. Regular expressions are not accepted. na : object, defulat NaN Object shown if element is not a string. Returns ------- Series of bool Koalas Series of booleans indicating whether the given pattern matches the end of each string element. """ return _wrap_accessor_pandas( self, lambda x: x.str.endswith(pattern, na), BooleanType() ).alias(self.name) def strip(self, to_strip=None) -> 'ks.Series': """ Remove leading and trailing characters. Strip whitespaces (including newlines) or a set of specified characters from each string in the Series/Index from left and right sides. Equivalent to :func:`str.strip`. Parameters ---------- to_strip : str Specifying the set of characters to be removed. All combinations of this set of characters will be stripped. If None then whitespaces are removed. Returns ------- Series of str """ return _wrap_accessor_pandas( self, lambda x: x.str.strip(to_strip), StringType() ).alias(self.name) def lstrip(self, to_strip=None) -> 'ks.Series': """ Remove leading characters. Strip whitespaces (including newlines) or a set of specified characters from each string in the Series/Index from left side. Equivalent to :func:`str.lstrip`. Parameters ---------- to_strip : str Specifying the set of characters to be removed. All combinations of this set of characters will be stripped. If None then whitespaces are removed. Returns ------- Series of str """ return _wrap_accessor_pandas( self, lambda x: x.str.lstrip(to_strip), StringType() ).alias(self.name) def rstrip(self, to_strip=None) -> 'ks.Series': """ Remove trailing characters. Strip whitespaces (including newlines) or a set of specified characters from each string in the Series/Index from right side. Equivalent to :func:`str.rstrip`. Parameters ---------- to_strip : str Specifying the set of characters to be removed. All combinations of this set of characters will be stripped. If None then whitespaces are removed. Returns ------- Series of str """ return _wrap_accessor_pandas( self, lambda x: x.str.rstrip(to_strip), StringType() ).alias(self.name) def get(self, i) -> 'ks.Series': """ Extract element from each string in the Series/Index at the specified position. Parameters ---------- i : int Position of element to extract. Returns ------- Series of objects """ return _wrap_accessor_pandas( self, lambda x: x.str.get(i), StringType() ).alias(self.name) def isalnum(self) -> 'ks.Series': """ Check whether all characters in each string are alphanumeric. This is equivalent to running the Python string method :func:`str.isalnum` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.isalnum(), BooleanType() ).alias(self.name) def isalpha(self) -> 'ks.Series': """ Check whether all characters in each string are alphabetic. This is equivalent to running the Python string method :func:`str.isalpha` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.isalpha(), BooleanType() ).alias(self.name) def isdigit(self) -> 'ks.Series': """ Check whether all characters in each string are digits. This is equivalent to running the Python string method :func:`str.isdigit` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.isdigit(), BooleanType() ).alias(self.name) def isspace(self) -> 'ks.Series': """ Check whether all characters in each string are whitespaces. This is equivalent to running the Python string method :func:`str.isspace` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.isspace(), BooleanType() ).alias(self.name) def islower(self) -> 'ks.Series': """ Check whether all characters in each string are lowercase. This is equivalent to running the Python string method :func:`str.islower` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.islower(), BooleanType() ).alias(self.name) def isupper(self) -> 'ks.Series': """ Check whether all characters in each string are uppercase. This is equivalent to running the Python string method :func:`str.isupper` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.isupper(), BooleanType() ).alias(self.name) def istitle(self) -> 'ks.Series': """ Check whether all characters in each string are titlecase. This is equivalent to running the Python string method :func:`str.istitle` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.istitle(), BooleanType() ).alias(self.name) def isnumeric(self) -> 'ks.Series': """ Check whether all characters in each string are numeric. This is equivalent to running the Python string method :func:`str.isnumeric` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.isnumeric(), BooleanType() ).alias(self.name) def isdecimal(self) -> 'ks.Series': """ Check whether all characters in each string are decimals. This is equivalent to running the Python string method :func:`str.isdecimal` for each element of the Series/Index. If a string has zero characters, False is returned for that check. """ return _wrap_accessor_pandas( self, lambda x: x.str.isdecimal(), BooleanType() ).alias(self.name)
10,955
2,850
import pytest import six from turbodbc import connect from query_fixture import query_fixture from helpers import for_one_database, open_cursor @for_one_database def test_many_batches_with_async_io(dsn, configuration): with open_cursor(configuration, use_async_io=True) as cursor: with query_fixture(cursor, configuration, 'INSERT INTEGER') as table_name: # insert 2^16 rows cursor.execute("INSERT INTO {} VALUES (1)".format(table_name)) for _ in six.moves.range(16): cursor.execute("INSERT INTO {} SELECT * FROM {}".format(table_name, table_name)) cursor.execute("SELECT * FROM {}".format(table_name)) assert sum(1 for _ in cursor) == 2**16
807
225
""" URL's for the LList Game Board app. """ from django.urls import path from llist_gameboard.api import llist_api from . import views urlpatterns = [ # Views path('', views.llist_game_board, name='llist-game-board'), #Game Play API Calls For Linked List path('llist_api', llist_api.api_overview, name='llist-game-board-api_overview'), path('llist_api/start_game/<str:difficulty>/<str:player_ids>/<str:data_structures>', llist_api.start_game, name='llist-game-board-start_game'), path('llist_api/board/<str:game_id>', llist_api.board, name='llist-game-board-game_status'), path('llist_api/dig_tunnel/<str:game_id>/<str:origin>/<str:destination>', llist_api.dig_tunnel, name='llist-game-board-dig_tunnel'), path('llist_api/dig_chamber/<str:game_id>/<str:origin>/<str:move_ant>/<str:ant>', llist_api.dig_chamber, name='llist-game-board-dig_chamber'), path('llist_api/fill_chamber/<str:game_id>/<str:to_fill>', llist_api.fill_chamber, name='llist-game-board-fill_chamber'), path('llist_api/spawn_ant/<str:game_id>', llist_api.spawn_ant, name='llist-game-board-spawn_ant'), path('llist_api/forage/<str:game_id>/<str:difficulty>/<str:dest>', llist_api.forage, name='llist-game-board-forage'), path('llist_api/move_food/<str:game_id>/<str:start>/<str:dest>', llist_api.move_food, name='llist-game-board-move_food'), path('llist_api/move_ant/<str:game_id>/<str:start>/<str:dest>', llist_api.move_ant, name='llist-game-board-move_ant'), ]
1,491
611
from src import most_invoices EXPECTED_RESULT = (14, 'Berlin') def test_most_invoices() -> None: tested_result = most_invoices.get_city_with_most_invoices() assert tested_result == EXPECTED_RESULT
208
81
import random import numpy as np import operator from scipy import optimize from matplotlib.backends.backend_qt4agg import FigureCanvasQTAgg from matplotlib.figure import Figure as MatplotlibFigure from mpl_toolkits.mplot3d import Axes3D from matplotlib import cm as color_map from matplotlib.ticker import LinearLocator, FormatStrFormatter import interface.auxiliary_functions as auxi import dictionaries.constants as cs ################################################################################# class ResultsCanvas(FigureCanvasQTAgg): def __init__(self, canvas_ref, vip): self._Figure = MatplotlibFigure(figsize = cs.FIG_SIZE, dpi = cs.DPI)#, tight_layout=True, frameon=True) super(ResultsCanvas, self).__init__(self._Figure) self.update_figure(canvas_ref, vip) def _from_options(self, canvas_ref, vip): self.Axes.set_position(self._get_axes_position(vip)) labels_x = self.Axes.xaxis.get_ticklabels() labels_y = self.Axes.yaxis.get_ticklabels() fontsize = vip.get('Options', 'R_axes_font_size') angle = vip.get('Options', 'R_x_plot_label_rotation') for label in labels_x+labels_y: label.set_fontsize(fontsize) if vip.get(canvas_ref, 'F_plot_function') == 'density': for label in labels_x: label.set_rotation(angle) def _get_axes_position(self, vip): session_keys = ['R_x_plot_position', 'R_y_plot_position', 'R_x_plot_size', 'R_y_plot_size'] f = lambda k: float(vip.get('Options', k)) return map(f, session_keys) ################################################################################# class Canvas2dData(ResultsCanvas): def __init__(self, canvas_ref, vip): super(Canvas2dData, self).__init__(canvas_ref, vip) def update_figure(self, canvas_ref, vip): self._Figure.clear() #from numpy.random import rand #x, y, c, s = rand(4, 100) #def onpick3(event): # ind = event.ind # print 'onpick3 scatter:', ind, np.take(x_axis, ind), np.take(y_axis, ind) #self._Figure.canvas.mpl_connect('pick_event', onpick3) try: data_set = vip.get(canvas_ref, 'F_data_set') plot_data2D = vip.plot_data[data_set]['2d_data'] ########## Axes self.Axes = self._Figure.add_axes(cs.AXES_POSITION_INIT) x_axis = plot_data2D['axis_1'] y_axis = plot_data2D['axis_r'] self.Axes.plot(x_axis, y_axis, auxi.colour(cs.PLOT_COLOR_RANGE)) #self.Axes.set_xlim([x_axis[0], x_axis[-1]]) self.Axes.set_xlim(sorted([x_axis[0], x_axis[-1]])) self._from_options(canvas_ref, vip) self.Axes.set_xlabel(plot_data2D['label_1']) self.Axes.set_ylabel(plot_data2D['label_r']) #self.Axes.hold(False) ########## Extrema #max_index, max_y = max(enumerate(y_axis), key=operator.itemgetter(1)) #vip.maximal_x = x_axis[max_index] min_index, min_y = min(enumerate(y_axis), key=operator.itemgetter(1)) vip.minimal_x = x_axis[min_index] print "* GLOBAL MINIMUM:\n{0}".format(vip.minimal_x) if canvas_ref in ['Plot_column_1']: ########## Savitzky Golay Filter ws = len(y_axis)/cs.SAVITZKY_GOLAY_FILTER_RANGE_DENOMINATOR ws = ws if (ws % 2 == 1) else (ws + 1) try: y_axis_sg = auxi.savitzky_golay_filter(y_axis, window_size=ws, order=cs.SAVITZKY_GOLAY_FILTER_ORDER) self.Axes.plot(x_axis, y_axis_sg, cs.FILTER_CURVE_STYLE, linewidth=cs.FILTER_LINEWIDTH) except TypeError as exception: print "! (update_figure) couldn't compute 'savitzky_golay_filter':" print exception ########## Fit try: def lorenzian_fit(x, A, k, ke): """Take min_x of this session and define a fit function""" def h(ke_): return (k / 2 - ke_)**2 + (x - vip.minimal_x)**2 r = A * h(ke) / h(0) return auxi.to_dB(r) parameters, covariance = optimize.curve_fit(lorenzian_fit, x_axis, y_axis_sg) LINE = 40 * "." + "\n" print LINE print "LORENZIAN FIT AT FILTER CUVE MINIMUM:\n" print "* PARAMETERS:\n\n [A, kappa, kappa_e]\n= {0}\n".format(parameters) print "* PARAMETERS:\n\n kappa_e / kappa\n= {0}\n" .format(parameters[1] / parameters[0]) print "* COVARIANCE:\n\n Matrix\n= {0}\n" .format(covariance) print "* MINIMUM: \n\n (x,y)\n= ({0}, {1})\n" .format(x_axis[min_index], y_axis[min_index]) print LINE fit_function = lambda x: lorenzian_fit(x, *parameters) y_axis_fit = map(fit_function, x_axis) self.Axes.plot(x_axis, y_axis_fit, cs.FITTING_CURVE_STYLE, linewidth=cs.FITTING_LINEWIDTH, linestyle=cs.FITTING_LINESTYLE) except: print "! (update_figure) couldn't fit to lorenzian_fit." else: pass try: self.draw() except ValueError: message = "! (update_figure, ValueError) at vip.draw." vip.GUI_feedback(message) except KeyError: message = "! (update_figure) The specified dataset might not exist." vip.GUI_feedback(message) ################################################################################# class Canvas3dData(ResultsCanvas): def __init__(self, canvas_ref, vip): super(Canvas3dData, self).__init__(canvas_ref, vip) def update_figure(self, canvas_ref, vip): self._Figure.clear() try: data_set = vip.get(canvas_ref, 'F_data_set') plot_data3D = vip.plot_data[data_set]['3d_data'] ########## Axes X, Y = np.meshgrid(plot_data3D['axis_1'], plot_data3D['axis_2']) Z = np.array(plot_data3D['axis_r']) if vip.get(canvas_ref, 'F_plot_function') == 'density': self.Axes = self._Figure.add_axes(cs.AXES_POSITION_INIT) self.Axes.pcolormesh(X, Y, Z, cmap = color_map.coolwarm) elif vip.get(canvas_ref, 'F_plot_function') == 'surface': self.Axes = Axes3D(self._Figure) surf = self.Axes.plot_surface(X, Y, Z, cmap = color_map.coolwarm, rstride = 1, cstride = 1, linewidth = 0.15, antialiased = False) self.Axes.zaxis.set_major_locator(LinearLocator(10)) self.Axes.zaxis.set_major_formatter(FormatStrFormatter('%.02f')) #self.Axes.set_zlim(-1.01, 1.01) position_color_bar = [0.015, 0.17, 0.015, 0.75] Axes_color_bar = self._Figure.add_axes(position_color_bar) self._Figure.colorbar(surf, cax = Axes_color_bar) self._from_options(canvas_ref, vip) #self.Axes.hold(False) self.Axes.set_xlabel(plot_data3D['label_1']) self.Axes.set_ylabel(plot_data3D['label_2']) ########## / Axes try: self.draw() except ValueError: message = "(update_figure, vip.draw, ValueError)" vip.GUI_feedback(message) except KeyError: message = "The specified dataset might not exist" vip.GUI_feedback(message)
8,223
2,788
#!/usr/bin/env python # Given weighted graph, perform kruskal-based clustering from common import * from cluster_common import * import argparse import csv import pickle as pickle from collections import defaultdict class unionfind: mp = {} blacklisted_edges = set() # blacklisted_e_nodes = set() # blacklist_edges_adj = defaultdict(set) def get_id(self, a): if a not in self.mp: self.mp[a] = a return a if self.mp[a] == a: return a else: self.mp[a] = self.get_id(self.mp[a]) return self.mp[a] def mergeset(self, a, b): self.mp[self.get_id(b)] = self.get_id(a) def mergeall(self, a): d = self.get_id(a[0]) for b in a[1:]: if not self.check_for_blacklist(b, d): self.mp[self.get_id(b)] = d def disallow(self, v1, v2): if v2 > v1: v1, v2 = v2, v1 self.blacklisted_edges.add((v1, v2)) # self.blacklisted_e_nodes.add(v1) # self.blacklisted_e_nodes.add(v2) # self.blacklist_edges[v1].add(v2) # self.blacklist_edges[v2].add(v1) def check_for_blacklist(self, v1, v2): v1, v2 = self.get_id(v1), self.get_id(v2) if v2 > v1: v1, v2 = v2, v1 for e1, e2 in self.blacklisted_edges: c1, c2 = self.get_id(e1), self.get_id(e2) if c2 > c1: c1, c2 = c2, c1 if c1 == v1 and c2 == v2: return True return False def trymerge(self, v1, v2): if self.get_id(v1) != self.get_id(v2) and not self.check_for_blacklist(v1, v2): self.mergeset(v1, v2) def main(): parser = argparse.ArgumentParser() parser.add_argument('edgelist') parser.add_argument('outfile', nargs='?') parser.add_argument('-t', '--interconnectivity', default=0.82, type=float) parser.add_argument('-A', '--with-analysis', action='store_true') parser.add_argument('-a', '--authorprefeat', default='generated/Author_prefeat.pickle') parser.add_argument('-s', '--seedset', nargs='*', default=['data/goldstd_clusters.csv', 'data/seedset_clusters.csv']) parser.add_argument('-S', '--seededges', nargs='*', default=['data/train.csv']) parser.add_argument('-b', '--blacklist', nargs='*', default=['data/blacklist_edges.csv', 'data/train.csv', 'data/train_extra.csv']) args = parser.parse_args() if args.outfile == None: args.outfile = args.edgelist.replace('.prob','') + '.clusters' threshold_interconnectivity = args.interconnectivity print_err("Loading graph") reader = csv.reader(enforce_min(skip_comments(open(args.edgelist, 'rb')), threshold_interconnectivity)) edges = [] for i, line in enumerate(reader): line[0:2] = list(map(int, line[0:2])) line[2] = float(line[2]) edges.append((line[2], line[0], line[1])) if (i+1) % 10000 == 0: print_err(i+1, "edges done") print_err("Sorting edges by weight") edges = sorted(edges, reverse=True) uf = unionfind() if args.blacklist: for filename in args.blacklist: with open(filename, 'rb') as f: reader = csv.reader(skip_comments(f)) for line in reader: line[0:3] = list(map(int, line[0:3])) if len(line) > 2: if line[0] != 0: continue line = line[1:] uf.disallow(line[0], line[1]) if args.seedset: print_err("Loading seedset(s)") for filename in args.seedset: for cl in loadClusters(filename): if len(cl) < 2: continue uf.mergeall(cl) if args.seededges: for filename in args.blacklist: with open(filename, 'rb') as f: reader = csv.reader(skip_comments(f)) for line in reader: line[0:3] = list(map(int, line[0:3])) if line[0] != 1: continue line = line[1:] uf.trymerge(line[0], line[1]) print_err("Clustering") for i, (w, v1, v2) in enumerate(edges): uf.trymerge(v1, v2) if (i+1) % 10000 == 0: print_err(i+1, "edges done") clusters = defaultdict(list) for v in uf.mp: clusters[uf.get_id(v)].append(v) clusters = [v for v in list(clusters.values()) if len(v) > 1] clusters = sorted(clusters, key=len, reverse=True) print_err("Writing clusters") f_out = open(args.outfile, 'wb') if not args.with_analysis: for cl in clusters: f_out.write(','.join(map(str, sorted(cl))) + '\n') if args.with_analysis: print_err("Loading pickled author pre-features") authors = pickle.load(open(args.authorprefeat, 'rb')) import networkx as nx G_sim = nx.read_weighted_edgelist(skip_comments(open(args.edgelist, 'rb')), nodetype=int, delimiter=',') outputClusters(clusters, f_out, G_sim, authors) if __name__ == "__main__": main()
4,538
2,125
# Copyright 2011 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Helper routines to facilitate use of oauth2_client in gsutil.""" import sys import time import webbrowser import oauth2_client GSUTIL_CLIENT_ID = '909320924072.apps.googleusercontent.com' # Google OAuth2 clients always have a secret, even if the client is an installed # application/utility such as gsutil. Of course, in such cases the "secret" is # actually publicly known; security depends entirly on the secrecy of refresh # tokens, which effectively become bearer tokens. GSUTIL_CLIENT_NOTSOSECRET = 'p3RlpR10xMFh9ZXBS/ZNLYUu' GOOGLE_OAUTH2_PROVIDER_LABEL = 'Google' GOOGLE_OAUTH2_PROVIDER_AUTHORIZATION_URI = ( 'https://accounts.google.com/o/oauth2/auth') GOOGLE_OAUTH2_PROVIDER_TOKEN_URI = ( 'https://accounts.google.com/o/oauth2/token') OOB_REDIRECT_URI = 'urn:ietf:wg:oauth:2.0:oob' def OAuth2ClientFromBotoConfig(config): token_cache = None token_cache_type = config.get('OAuth2', 'token_cache', 'file_system') if token_cache_type == 'file_system': if config.has_option('OAuth2', 'token_cache_path_pattern'): token_cache = oauth2_client.FileSystemTokenCache( path_pattern=config.get('OAuth2', 'token_cache_path_pattern')) else: token_cache = oauth2_client.FileSystemTokenCache() elif token_cache_type == 'in_memory': token_cache = oauth2_client.InMemoryTokenCache() else: raise Exception( "Invalid value for config option OAuth2/token_cache: %s" % token_cache_type) proxy = None if (config.has_option('Boto', 'proxy') and config.has_option('Boto', 'proxy_port')): proxy = "%s:%s" % (config.get('Boto', 'proxy'), config.get('Boto', 'proxy_port')) provider_label = config.get( 'OAuth2', 'provider_label', GOOGLE_OAUTH2_PROVIDER_LABEL) provider_authorization_uri = config.get( 'OAuth2', 'provider_authorization_uri', GOOGLE_OAUTH2_PROVIDER_AUTHORIZATION_URI) provider_token_uri = config.get( 'OAuth2', 'provider_token_uri', GOOGLE_OAUTH2_PROVIDER_TOKEN_URI) client_id = config.get('OAuth2', 'client_id', GSUTIL_CLIENT_ID) client_secret = config.get( 'OAuth2', 'client_secret', GSUTIL_CLIENT_NOTSOSECRET) return oauth2_client.OAuth2Client( oauth2_client.OAuth2Provider( provider_label, provider_authorization_uri, provider_token_uri), client_id, client_secret, proxy=proxy, access_token_cache=token_cache) def OAuth2ApprovalFlow(oauth2_client, scopes, launch_browser=False): approval_url = oauth2_client.GetAuthorizationUri(OOB_REDIRECT_URI, scopes) if launch_browser: sys.stdout.write( 'Attempting to launch a browser with the OAuth2 approval dialog at ' 'URL: %s\n\n' '[Note: due to a Python bug, you may see a spurious error message "object is not\n' 'callable [...] in [...] Popen.__del__" which can be ignored.]\n\n' % approval_url) else: sys.stdout.write( 'Please navigate your browser to the following URL:\n%s\n\n' % approval_url) if (launch_browser and not webbrowser.open(approval_url, new=1, autoraise=True)): sys.stdout.write( 'Launching browser appears to have failed; please navigate a browser ' 'to the following URL:\n%s\n' % approval_url) code = raw_input('Enter the authorization code: ') refresh_token, access_token = oauth2_client.ExchangeAuthorizationCode( code, OOB_REDIRECT_URI, scopes) return refresh_token
4,019
1,394
# -*- coding: utf-8 -*- # Python script to manage automatically generated dependents # Author: Rintze M. Zelle # Version: 2014-04-17 # * Requires lxml library (http://lxml.de/) import os, glob, re, inspect, shutil from lxml import etree # http://stackoverflow.com/questions/50499 folderPath = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))) parentFolderPath = os.path.dirname (folderPath) path = os.path.join(parentFolderPath, 'styles') pathGeneratedStyles = os.path.join(parentFolderPath, 'utilities', 'generate_dependent_styles', 'generated_styles', 'aacr') pathRemovedStyles = os.path.join(parentFolderPath, 'removed-styles') dependentStyles = [] commentMatchingStyles = [] parentMatchingStyles = [] timestampMatchingStyles = [] generatedStyles = [] for stylepath in glob.glob( os.path.join(path, 'dependent', '*.csl') ): dependentStyles.append(os.path.join(stylepath)) for stylepath in glob.glob( os.path.join(pathGeneratedStyles, '*.csl') ): generatedStyles.append(os.path.basename(stylepath)) #Filter dependent styles by their parent (set A), print number #Of set A, print style ID if XML comment doesn't match that of dependent style template #Of set A, print style ID if timestamp doesn't match that of dependent style template #Have a toggle to move remaining styles out of root folder #(it would be better to filter by the XML comment on the first pass, since styles from #a set may have different parents, but XML comments aren't currently unique to a set) for style in dependentStyles: parser = etree.XMLParser(remove_blank_text=True) parsedStyle = etree.parse(style, parser) styleElement = parsedStyle.getroot() parentLink = styleElement.find(".//{http://purl.org/net/xbiblio/csl}link[@rel='independent-parent']") if(parentLink.attrib.get("href") == "http://www.zotero.org/styles/american-association-for-cancer-research"): parentMatchingStyles.append(os.path.basename(style)) comments = styleElement.xpath("//comment()", namespaces={"cs": "http://purl.org/net/xbiblio/csl"}) for comment in comments: if(comment.text == " Generated with https://github.com/citation-style-language/utilities/tree/master/generate_dependent_styles/data/aacr "): commentMatchingStyles.append(os.path.basename(style)) timestamp = styleElement.find(".//{http://purl.org/net/xbiblio/csl}updated") if(timestamp.text == "2014-04-23T12:00:00+00:00"): timestampMatchingStyles.append(os.path.basename(style)) print("Number of dependent styles with selected parent: " + str(len(parentMatchingStyles))) print("Number of generated styles: " + str(len(generatedStyles))) for style in parentMatchingStyles: badStyle = False if not (style in commentMatchingStyles): print "bad comment!: " + style badStyle = True if not (style in timestampMatchingStyles): print "bad timestamp!: " + style badStyle = True if not (style in generatedStyles): print "not generated!: " + style badStyle = True if badStyle: parentMatchingStyles.remove(style) print("Number of consistent styles: " + str(len(parentMatchingStyles))) moveStyles = False if moveStyles == True: #move styles out of "styles/dependent" folder if not os.path.exists(pathRemovedStyles): os.makedirs(pathRemovedStyles) for style in parentMatchingStyles: shutil.move(os.path.join(path, 'dependent', style), os.path.join(pathRemovedStyles, style)) # counter = [] # for infoNodeIndex, infoNode in enumerate(csInfo): # # check if node is an element # if isinstance(infoNode.tag, basestring): # # get rid of namespace # infoElement = infoNode.tag.replace("{http://purl.org/net/xbiblio/csl}","") # if(infoElement == "link"): # infoElement += "[@" + infoNode.get("rel") + "]" # if((infoElement == "category") & (infoNode.get("citation-format") is not None)): # infoElement += "[@citation-format]" # if((infoElement == "category") & (infoNode.get("field") is not None)): # infoElement += "[@field]" # # check if node is a comment # elif (etree.tostring(infoNode, encoding='UTF-8', xml_declaration=False) == ("<!--" + infoNode.text.encode("utf-8") + "-->")): # # keep comments that precede any element at the top # if(sum(counter) == 0): # counter.append(desiredOrder.index("preceding-comment")) # # keep a comment at the end at the end # elif(len(counter) == (len(csInfo) - 1)): # counter.append(desiredOrder.index("end-comment")) # # keep other comments with preceding element # else: # counter.append(counter[-1]) # # # Possible improvements: # # * exceptions for recognizable comments (issn, category) # else: # print(infoNode) # # # Reorder attributes on cs:link # try: # links = styleElement.findall(".//{http://purl.org/net/xbiblio/csl}link") # for link in links: # rel = link.get("rel") # del link.attrib["rel"] # link.set("rel",rel) # except: # pass
5,311
1,618
# ------------------------------------------------------- class: Constants ------------------------------------------------------- # class Constants: # --------------------------------------------------- Public properties -------------------------------------------------- # USER_AGENT_FILE_NAME = 'ua.txt' GENERAL_COOKIES_FOLDER_NAME = 'cookies' GENERAL_PROFILE_FOLDER_NAME = 'selenium-python-profile' DEFAULT_PROFILE_ID = 'test' DEFAULT_FIND_FUNC_TIMEOUT = 2.5 # -------------------------------------------------------------------------------------------------------------------------------- #
642
150
class Game: def __init__(self, *args, **kwargs): self.player_0 = 'white' self.player_1 = 'black' self._new_pieces_game() self.current_player_turn = self.player_0 def _new_pieces_game(self): self.board = { 0: { 0: Piece(self.player_0, "rook", 0, (0, 0)) 1: Piece(self.player_0, "knight", 0, (0, 1)) 2: Piece(self.player_0, "bishop", 0, (0, 2)) 3: Piece(self.player_0, "king", 0, (0, 3)) 4: Piece(self.player_0, "queen", 0, (0, 4)) 5: Piece(self.player_0, "bishop", 1, (0, 5)) 6: Piece(self.player_0, "knight", 1, (0, 6)) 7: Piece(self.player_0, "rook", 1, (0, 7)) }, 1: { {ii: Piece(self.player_0, "pawn", ii, (1, ii)) }, 7: { 0: Piece(self.player_1, "rook", 0, (7, 0)) 1: Piece(self.player_1, "knight", 0, (7, 1)) 2: Piece(self.player_1, "bishop", 0, (7, 2)) 3: Piece(self.player_1, "king", 0, (7, 3)) 4: Piece(self.player_1, "queen", 0, (7, 4)) 5: Piece(self.player_1, "bishop", 1, (7, 5)) 6: Piece(self.player_1, "knight", 1, (7, 6)) 7: Piece(self.player_1, "rook", 1, (7, 7)) }, 6: { {ii: Piece(self.player_1, "pawn", ii, (6, ii)) } } pieces = { self.player_0: { 'rook': {0: self.board[0][0], 1: self.board[0][7]}, 'knight': {0: self.board[0][1], 1: self.board[0][6]}, 'bishop': {0: self.board[0][2], 1: self.board[0][5]}, 'king': {0: self.board[0][3]}, 'queen': {0: self.board[0][4]}, 'pawn': {} }, self.player_1: { 'rook': {0: self.board[7][0], 1: self.board[7][7]}, 'knight': {0: self.board[7][1], 1: self.board[7][6]}, 'bishop': {0: self.board[7][2], 1: self.board[7][5]}, 'king': {0: self.board[7][3]}, 'queen': {0: self.board[7][4]}, 'pawn': {} } } for ii in range(0, 8): pieces[self.player_0]["pawn"][ii] = self.board[1][ii] pieces[self.player_1]["pawn"][ii] = [6][ii] self.pieces = pieces def find_piece(self, x_coordinate, y_coordinate): piece = self.board.get(x_coordinate, {}).get(y_coordinate) return piece def to_dict(self): return {'current_player_turn': self.current_player_turn, 'pieces': self.pieces} class Piece: def __init__(self, player, kind, ii, coordinate): self.player = player self.kind = kind self.ii = ii self.coordinate = coordinate
2,893
1,080
from __future__ import annotations import csv import pathlib import textwrap from operator import itemgetter from typing import Any from typing import Callable from typing import Dict from typing import IO from typing import Iterable from typing import Iterator from typing import List from typing import NamedTuple from typing import Optional from typing import Set from typing import Tuple from typing import TypeVar from typing import Union from typing import cast import openpyxl import xlsxwriter from lark import ParseError from spinta import commands from spinta import spyna from spinta.backends import Backend from spinta.backends.components import BackendOrigin from spinta.components import Context from spinta.datasets.components import Resource from spinta.dimensions.comments.components import Comment from spinta.dimensions.enum.components import EnumItem from spinta.components import Model from spinta.components import Namespace from spinta.components import Property from spinta.core.enums import Access from spinta.core.ufuncs import unparse from spinta.datasets.components import Dataset from spinta.dimensions.enum.components import Enums from spinta.dimensions.lang.components import LangData from spinta.dimensions.prefix.components import UriPrefix from spinta.exceptions import MultipleErrors from spinta.exceptions import PropertyNotFound from spinta.manifests.components import Manifest from spinta.manifests.helpers import load_manifest_nodes from spinta.manifests.tabular.components import ACCESS from spinta.manifests.tabular.components import BackendRow from spinta.manifests.tabular.components import BaseRow from spinta.manifests.tabular.components import CommentData from spinta.manifests.tabular.components import DESCRIPTION from spinta.manifests.tabular.components import DatasetRow from spinta.manifests.tabular.components import ParamRow from spinta.manifests.tabular.components import EnumRow from spinta.manifests.tabular.components import ID from spinta.manifests.tabular.components import MANIFEST_COLUMNS from spinta.manifests.tabular.components import ManifestColumn from spinta.manifests.tabular.components import ManifestRow from spinta.manifests.tabular.components import ManifestTableRow from spinta.manifests.tabular.components import ModelRow from spinta.manifests.tabular.components import PREPARE from spinta.manifests.tabular.components import PROPERTY from spinta.manifests.tabular.components import PrefixRow from spinta.manifests.tabular.components import PropertyRow from spinta.manifests.tabular.components import REF from spinta.manifests.tabular.components import ResourceRow from spinta.manifests.tabular.components import SOURCE from spinta.manifests.tabular.components import TITLE from spinta.manifests.tabular.components import TabularFormat from spinta.manifests.tabular.constants import DATASET from spinta.manifests.tabular.formats.gsheets import read_gsheets_manifest from spinta.spyna import SpynaAST from spinta.types.datatype import Ref from spinta.utils.data import take from spinta.utils.schema import NA from spinta.utils.schema import NotAvailable ParsedRow = Tuple[int, Dict[str, Any]] MAIN_DIMENSIONS = [ 'dataset', 'resource', 'base', 'model', 'property', ] EXTRA_DIMENSIONS = [ '', 'prefix', 'enum', 'param', 'comment', 'ns', 'lang', ] class TabularManifestError(Exception): pass def _detect_header( path: Optional[str], line: int, # Line number row: Iterable[str], ) -> List[str]: header = [h.strip().lower() for h in row] unknown_columns = set(header[:len(DATASET)]) - set(DATASET) if unknown_columns: unknown_columns = ', '.join(sorted(unknown_columns, key=header.index)) raise TabularManifestError( f"{path}:{line}: Unknown columns: {unknown_columns}." ) return header def _detect_dimension( path: Optional[pathlib.Path], line: str, # Line number with a prefix (depends on manifest format) row: Dict[str, str], ) -> Optional[str]: dimensions = [k for k in MAIN_DIMENSIONS if row[k]] if len(dimensions) == 1: return dimensions[0] if len(dimensions) > 1: dimensions = ', '.join(dimensions) raise TabularManifestError( f"{path}:{line}: In one row only single dimension can be used, " f"but found more than one: {dimensions}" ) if row['type']: if row['type'] not in EXTRA_DIMENSIONS: raise TabularManifestError( f"{path}:{line}:type: Unknown additional dimension name " f"{row['type']}." ) return row['type'] return '' def _parse_spyna( reader: TabularReader, formula: str, ) -> Union[SpynaAST, NotAvailable, None]: if formula: try: return spyna.parse(formula) except ParseError as e: reader.error(f"Error while parsing formula {formula!r}:\n{e}") return NA class TabularReader: state: State path: str line: str type: str name: str data: ManifestRow # Used when `appendable` is False rows: List[Dict[str, Any]] # Used when `appendable` is True appendable: bool = False # Tells if reader is appendable. def __init__( self, state: State, path: str, line: str, ): self.state = state self.path = path self.line = line self.data = {} self.rows = [] def __str__(self): return f"<{type(self).__name__} name={self.name!r}>" def read(self, row: Dict[str, str]) -> None: raise NotImplementedError def append(self, row: Dict[str, str]) -> None: if any(row.values()): self.error( f"Updates are not supported in context of {self.type!r}." ) def release(self, reader: TabularReader = None) -> bool: raise NotImplementedError def items(self) -> Iterator[ParsedRow]: if self.appendable: for data in self.rows: yield self.line, data else: yield self.line, self.data def enter(self) -> None: raise NotImplementedError def leave(self) -> None: raise NotImplementedError def error(self, message: str) -> None: raise TabularManifestError(f"{self.path}:{self.line}: {message}") class ManifestReader(TabularReader): type: str = 'manifest' datasets: Set[str] namespaces: Set[str] data: ManifestTableRow def read(self, row: ManifestRow) -> None: self.name = str(self.path) self.data = { 'type': 'manifest', } def release(self, reader: TabularReader = None) -> bool: return reader is None def enter(self) -> None: self.datasets = set() self.namespaces = set() self.state.manifest = self def leave(self) -> None: self.state.manifest = None class DatasetReader(TabularReader): type: str = 'dataset' data: DatasetRow def read(self, row: Dict[str, str]) -> None: self.name = row['dataset'] if row['dataset'] in self.state.manifest.datasets: self.error("Dataset already defined.") self.data = { 'type': 'dataset', 'id': row['id'], 'name': row['dataset'], 'level': row['level'], 'access': row['access'], 'title': row['title'], 'description': row['description'], 'resources': {}, } def release(self, reader: TabularReader = None) -> bool: return reader is None or isinstance(reader, ( ManifestReader, DatasetReader, )) def enter(self) -> None: self.state.dataset = self def leave(self) -> None: self.state.dataset = None class ResourceReader(TabularReader): type: str = 'resource' data: Union[BackendRow, ResourceRow] def read(self, row: Dict[str, str]) -> None: self.name = row['resource'] if self.state.dataset is None: self.read_backend(row) else: self.read_resource(row) def read_backend(self, row: Dict[str, str]) -> None: # Backends will be loaded using # `spinta.manifests.helpers._load_manifest_backends`. if 'backends' not in self.state.manifest.data: self.state.manifest.data['backends'] = {} backends = self.state.manifest.data['backends'] if self.name in backends: self.error( f"Backend {self.name!r} with the same name already defined." ) self.data = { 'type': row['type'], 'name': self.name, 'dsn': row['source'], 'title': row['title'], 'description': row['description'], } backends[self.name] = self.data def read_resource(self, row: Dict[str, str]) -> None: dataset = self.state.dataset.data if self.name in dataset['resources']: self.error("Resource with the same name already defined in ") self.data = { 'type': row['type'], 'backend': row['ref'], 'external': row['source'], 'prepare': _parse_spyna(self, row[PREPARE]), 'level': row['level'], 'access': row['access'], 'title': row['title'], 'description': row['description'], } dataset['resources'][self.name] = self.data def release(self, reader: TabularReader = None) -> bool: return reader is None or isinstance(reader, ( ManifestReader, DatasetReader, ResourceReader, )) def enter(self) -> None: self.state.resource = self def leave(self) -> None: self.state.resource = None class BaseReader(TabularReader): type: str = 'base' data: BaseRow def read(self, row: Dict[str, str]) -> None: self.name = row['base'] dataset = self.state.dataset.data if self.state.dataset else None self.data = { 'model': get_relative_model_name(dataset, row['base']), 'pk': row['ref'], } def release(self, reader: TabularReader = None) -> bool: return reader is None or isinstance(reader, ( ManifestReader, DatasetReader, ResourceReader, BaseReader, )) def enter(self) -> None: self.state.base = self def leave(self) -> None: self.state.base = None class ModelReader(TabularReader): type: str = 'model' data: ModelRow def read(self, row: Dict[str, str]) -> None: dataset = self.state.dataset resource = self.state.resource base = self.state.base name = get_relative_model_name( dataset.data if dataset else None, row['model'], ) if self.state.rename_duplicates: dup = 1 _name = name while _name in self.state.models: _name = f'{name}_{dup}' dup += 1 name = _name elif name in self.state.models: self.error(f"Model {name!r} with the same name is already defined.") self.name = name self.data = { 'type': 'model', 'id': row['id'], 'name': name, 'base': base.name if base else None, 'level': row['level'], 'access': row['access'], 'title': row['title'], 'description': row['description'], 'properties': {}, 'external': { 'dataset': dataset.name if dataset else '', 'resource': resource.name if dataset and resource else '', 'pk': ( [x.strip() for x in row['ref'].split(',')] if row['ref'] else [] ), 'name': row['source'], 'prepare': _parse_spyna(self, row[PREPARE]), }, } if resource and not dataset: self.data['backend'] = resource.name def release(self, reader: TabularReader = None) -> bool: return reader is None or isinstance(reader, ( ManifestReader, DatasetReader, ResourceReader, BaseReader, ModelReader, )) def enter(self) -> None: self.state.model = self self.state.models.add(self.name) def leave(self) -> None: self.state.model = None def _parse_property_ref(ref: str) -> Tuple[str, List[str]]: if '[' in ref: ref = ref.rstrip(']') ref_model, ref_props = ref.split('[', 1) ref_props = [p.strip() for p in ref_props.split(',')] else: ref_model = ref ref_props = [] return ref_model, ref_props class PropertyReader(TabularReader): type: str = 'property' data: PropertyRow enums: Set[str] def read(self, row: Dict[str, str]) -> None: self.name = row['property'] if self.state.model is None: context = self.state.stack[-1] self.error( f"Property {self.name!r} must be defined in a model context. " f"Now it is defined in {context.name!r} {context.type} context." ) if row['property'] in self.state.model.data['properties']: self.error( f"Property {self.name!r} with the same name is already " f"defined for this {self.state.model.name!r} model." ) self.data = { 'type': row['type'], 'prepare': _parse_spyna(self, row[PREPARE]), 'level': row['level'], 'access': row['access'], 'uri': row['uri'], 'title': row['title'], 'description': row['description'], } dataset = self.state.dataset.data if self.state.dataset else None if row['ref']: if row['type'] in ('ref', 'backref', 'generic'): ref_model, ref_props = _parse_property_ref(row['ref']) self.data['model'] = get_relative_model_name(dataset, ref_model) self.data['refprops'] = ref_props else: # TODO: Detect if ref is a unit or an enum. self.data['enum'] = row['ref'] if dataset or row['source']: self.data['external'] = { 'name': row['source'], 'prepare': self.data.pop('prepare'), } self.state.model.data['properties'][row['property']] = self.data def release(self, reader: TabularReader = None) -> bool: return reader is None or isinstance(reader, ( ManifestReader, DatasetReader, ResourceReader, BaseReader, ModelReader, PropertyReader, )) def enter(self) -> None: self.state.prop = self def leave(self) -> None: self.state.prop = None class AppendReader(TabularReader): type: str = 'append' data: ManifestRow def read(self, row: ManifestRow) -> None: self.name = row[REF] self.data = row def release(self, reader: TabularReader = None) -> bool: return True def enter(self) -> None: pass def leave(self) -> None: self.state.stack[-1].append(self.data) class PrefixReader(TabularReader): type: str = 'prefix' data: PrefixRow def read(self, row: Dict[str, str]) -> None: if not row['ref']: # `ref` is a required parameter. return self.name = row['ref'] node = ( self.state.prop or self.state.model or self.state.base or self.state.resource or self.state.dataset or self.state.manifest ) if 'prefixes' not in node.data: node.data['prefixes'] = {} prefixes = node.data['prefixes'] if self.name in prefixes: self.error( f"Prefix {self.name!r} with the same name is already " f"defined for this {node.name!r} {node.type}." ) self.data = { 'id': row['id'], 'eid': f'{self.path}:{self.line}', 'type': self.type, 'name': self.name, 'uri': row['uri'], 'title': row['title'], 'description': row['description'], } prefixes[self.name] = self.data def append(self, row: Dict[str, str]) -> None: self.read(row) def release(self, reader: TabularReader = None) -> bool: return not isinstance(reader, AppendReader) def enter(self) -> None: pass def leave(self) -> None: pass class NamespaceReader(TabularReader): type: str = 'ns' appendable: bool = True def read(self, row: Dict[str, str]) -> None: if not row['ref']: # `ref` is a required parameter. return self.name = row['ref'] manifest = self.state.manifest if self.name in manifest.namespaces: self.error( f"Namespace {self.name!r} with the same name is already " f"defined." ) manifest.namespaces.add(self.name) self.rows.append({ 'id': row['id'], 'type': self.type, 'name': self.name, 'title': row['title'], 'description': row['description'], }) def append(self, row: Dict[str, str]) -> None: self.read(row) def release(self, reader: TabularReader = None) -> bool: return not isinstance(reader, AppendReader) def enter(self) -> None: pass def leave(self) -> None: pass class ParamReader(TabularReader): type: str = 'param' data: ParamRow name: str = None def _get_node(self) -> TabularReader: return ( self.state.prop or self.state.model or self.state.base or self.state.resource or self.state.dataset or self.state.manifest ) def _get_data(self, name: str, row: ManifestRow): return { 'name': name, 'source': [row[SOURCE]], 'prepare': [_parse_spyna(self, row[PREPARE])], 'title': row[TITLE], 'description': row[DESCRIPTION], } def _ensure_params_list(self, node: TabularReader, name: str) -> None: if 'params' not in node.data: node.data['params'] = {} if name not in node.data['params']: node.data['params'][name] = [] def _check_param_name(self, node: TabularReader, name: str) -> None: if 'params' in node.data and name in node.data['params']: self.error( f"Parameter {name!r} with the same name already defined!" ) def read(self, row: ManifestRow) -> None: node = self._get_node() self.name = row[REF] if not self.name: self.error("Parameter must have a name.") self._check_param_name(node, self.name) self._ensure_params_list(node, self.name) self.data = self._get_data(self.name, row) node.data['params'][self.name].append(self.data) def append(self, row: ManifestRow) -> None: node = self._get_node() if row[REF]: self.name = row[REF] self._check_param_name(node, self.name) self._ensure_params_list(node, self.name) self.data = self._get_data(self.name, row) node.data['params'][self.name].append(self.data) def release(self, reader: TabularReader = None) -> bool: return not isinstance(reader, (AppendReader, LangReader)) def enter(self) -> None: pass def leave(self) -> None: pass class EnumReader(TabularReader): type: str = 'enum' data: EnumRow name: str = None def read(self, row: ManifestRow) -> None: if row[REF]: self.name = row[REF] else: self.name = self.name or '' if not any([ row[SOURCE], row[PREPARE], row[ACCESS], row[TITLE], row[DESCRIPTION], ]): return # source = row[SOURCE] if row[SOURCE] is not None else row[PREPARE] source = str(row[SOURCE]) or row[PREPARE] if not source: self.error( "At least source or prepare must be specified for an enum." ) self.data = { 'name': self.name, 'source': row[SOURCE], 'prepare': _parse_spyna(self, row[PREPARE]), 'access': row[ACCESS], 'title': row[TITLE], 'description': row[DESCRIPTION], } node = ( self.state.prop or self.state.model or self.state.base or self.state.resource or self.state.dataset or self.state.manifest ) if 'enums' not in node.data: node.data['enums'] = {} if self.name not in node.data['enums']: node.data['enums'][self.name] = {} enum = node.data['enums'][self.name] if source in enum: self.error( f"Enum {self.name!r} item {source!r} with the same value is " f"already defined." ) enum[source] = self.data def append(self, row: ManifestRow) -> None: self.read(row) def release(self, reader: TabularReader = None) -> bool: return not isinstance(reader, (AppendReader, LangReader)) def enter(self) -> None: pass def leave(self) -> None: pass class LangReader(TabularReader): type: str = 'lang' def read(self, row: ManifestRow) -> None: reader = self.state.stack[-1] if not isinstance(reader, ( DatasetReader, ResourceReader, BaseReader, ModelReader, PropertyReader, EnumReader, )): self.error(f'Language metadata is not supported on {reader.type}.') return if 'lang' not in reader.data: reader.data['lang'] = {} lang = reader.data['lang'] self.name = row[REF] if self.name in lang: self.error( f"Language {self.name!r} with the same name is already " f"defined for this {reader.name!r} {reader.type}." ) lang[self.name] = { 'id': row[ID], 'eid': f'{self.path}:{self.line}', 'type': self.type, 'ref': self.name, 'title': row[TITLE], 'description': row[DESCRIPTION], } def append(self, row: ManifestRow) -> None: self.read(row) def release(self, reader: TabularReader = None) -> bool: return not isinstance(reader, AppendReader) def enter(self) -> None: pass def leave(self) -> None: pass class CommentReader(TabularReader): type: str = 'comment' data: CommentData def read(self, row: ManifestRow) -> None: reader = self.state.stack[-1] if 'comments' not in reader.data: reader.data['comments'] = [] comments = reader.data['comments'] comments.append({ 'id': row[ID], 'parent': row[REF], 'author': row[SOURCE], 'access': row[ACCESS], # TODO: parse datetime 'created': row[TITLE], 'comment': row[DESCRIPTION], }) def append(self, row: ManifestRow) -> None: self.read(row) def release(self, reader: TabularReader = None) -> bool: return not isinstance(reader, AppendReader) def enter(self) -> None: pass def leave(self) -> None: pass READERS = { # Main dimensions 'dataset': DatasetReader, 'resource': ResourceReader, 'base': BaseReader, 'model': ModelReader, 'property': PropertyReader, # Extra dimensions '': AppendReader, 'prefix': PrefixReader, 'ns': NamespaceReader, 'param': ParamReader, 'enum': EnumReader, 'lang': LangReader, 'comment': CommentReader, } class State: stack: List[TabularReader] backends: Dict[str, Dict[str, str]] = None models: Set[str] manifest: ManifestReader = None dataset: DatasetReader = None resource: ResourceReader = None base: BaseReader = None model: ModelReader = None prop: PropertyReader = None rename_duplicates: bool = False def __init__(self): self.stack = [] self.models = set() def release(self, reader: TabularReader = None) -> Iterator[ParsedRow]: for parent in list(reversed(self.stack)): if parent.release(reader): if isinstance(parent, ( ManifestReader, NamespaceReader, DatasetReader, ModelReader, )): yield from parent.items() self.stack.pop() parent.leave() else: break if reader: reader.enter() self.stack.append(reader) def _read_tabular_manifest_rows( path: Optional[str], rows: Iterator[Tuple[str, List[str]]], *, rename_duplicates: bool = True, ) -> Iterator[ParsedRow]: _, header = next(rows, (None, None)) if header is None: # Looks like an empty file. return header = _detect_header(path, 1, header) defaults = {k: '' for k in MANIFEST_COLUMNS} state = State() state.rename_duplicates = rename_duplicates reader = ManifestReader(state, path, '1') reader.read({}) yield from state.release(reader) for line, row in rows: row = dict(zip(header, row)) row = {**defaults, **row} dimension = _detect_dimension(path, line, row) Reader = READERS[dimension] reader = Reader(state, path, line) reader.read(row) yield from state.release(reader) yield from state.release() def read_tabular_manifest( format_: TabularFormat = None, *, path: str = None, file: IO = None, rename_duplicates: bool = False, ) -> Iterator[ParsedRow]: if format_ == TabularFormat.GSHEETS: rows = read_gsheets_manifest(path) elif format_ == TabularFormat.CSV: rows = _read_csv_manifest(path, file) elif format_ == TabularFormat.ASCII: rows = _read_txt_manifest(path, file) elif format_ == TabularFormat.XLSX: rows = _read_xlsx_manifest(path) else: raise ValueError(f"Unknown tabular manifest format {format_!r}.") yield from _read_tabular_manifest_rows( path, rows, rename_duplicates=rename_duplicates, ) def _read_txt_manifest( path: str, file: IO[str] = None, ) -> Iterator[Tuple[str, List[str]]]: if file: yield from _read_ascii_tabular_manifest(file) else: with pathlib.Path(path).open(encoding='utf-8-sig') as f: yield from _read_ascii_tabular_manifest(f) def _read_csv_manifest( path: str, file: IO[str] = None, ) -> Iterator[Tuple[str, List[str]]]: if file: rows = csv.reader(file) for i, row in enumerate(rows, 1): yield str(i), row else: with pathlib.Path(path).open(encoding='utf-8-sig') as f: rows = csv.reader(f) for i, row in enumerate(rows, 1): yield str(i), row def _read_xlsx_manifest(path: str) -> Iterator[Tuple[str, List[str]]]: wb = openpyxl.load_workbook(path) yield '1', DATASET for sheet in wb: rows = sheet.iter_rows(values_only=True) cols = next(rows, None) if cols is None: continue cols = normalizes_columns(cols) cols = [cols.index(c) if c in cols else None for c in DATASET] for i, row in enumerate(rows, 2): row = [row[c] if c is not None else None for c in cols] yield f'{sheet.title}:{i}', row def striptable(table): return textwrap.dedent(table).strip() def _join_escapes(row: List[str]) -> List[str]: res = [] for v in row: if res and res[-1] and res[-1].endswith('\\'): res[-1] = res[-1][:-1] + '|' + v else: res.append(v) return res def _read_ascii_tabular_manifest( lines: Iterable[str], *, check_column_names: bool = True, ) -> Iterator[Tuple[str, List[str]]]: lines = (line.strip() for line in lines) lines = filter(None, lines) # Read header header = next(lines, None) if header is None: return header = normalizes_columns( header.split('|'), check_column_names=check_column_names, ) yield '1', header # Find index where dimension columns end. dim = sum(1 for h in header if h in DATASET[:6]) for i, line in enumerate(lines, 2): row = _join_escapes(line.split('|')) row = [x.strip() for x in row] row = row[:len(header)] rem = len(header) - len(row) row = row[:dim - rem] + [''] * rem + row[dim - rem:] assert len(header) == len(row), line yield str(i), row def read_ascii_tabular_rows( manifest: str, *, strip: bool = False, check_column_names: bool = True, ) -> Iterator[List[str]]: if strip: manifest = striptable(manifest) rows = _read_ascii_tabular_manifest( manifest.splitlines(), check_column_names=check_column_names, ) for line, row in rows: yield row def read_ascii_tabular_manifest( manifest: str, *, strip: bool = False, rename_duplicates: bool = False, ) -> Iterator[ParsedRow]: if strip: manifest = striptable(manifest) rows = _read_ascii_tabular_manifest(manifest.splitlines()) yield from _read_tabular_manifest_rows( None, rows, rename_duplicates=rename_duplicates, ) def load_ascii_tabular_manifest( context: Context, manifest: Manifest, manifest_ascii_table: str, *, strip: bool = False, ) -> None: schemas = read_ascii_tabular_manifest(manifest_ascii_table, strip=strip) load_manifest_nodes(context, manifest, schemas) commands.link(context, manifest) def get_relative_model_name(dataset: dict, name: str) -> str: if name.startswith('/'): return name[1:] elif dataset is None: return name else: return '/'.join([ dataset['name'], name, ]) def to_relative_model_name(model: Model, dataset: Dataset = None) -> str: """Convert absolute model `name` to relative.""" if dataset is None: return model.name if model.name.startswith(dataset.name): prefix = dataset.name return model.name[len(prefix) + 1:] else: return '/' + model.name def tabular_eid(model: Model): if isinstance(model.eid, int): return model.eid else: return 0 class OrderBy(NamedTuple): func: Callable[[Union[Dataset, Model, Property, EnumItem]], Any] reverse: bool = False def _order_datasets_by_access(dataset: Dataset): return dataset.access or Access.private def _order_datasets_by_name(dataset: Dataset): return dataset.name DATASETS_ORDER_BY = { 'access': OrderBy(_order_datasets_by_access, reverse=True), 'default': OrderBy(_order_datasets_by_name), } def _order_models_by_access(model: Model): return model.access or Access.private MODELS_ORDER_BY = { 'access': OrderBy(_order_models_by_access, reverse=True), 'default': OrderBy(tabular_eid), } def _order_properties_by_access(prop: Property): return prop.access or Access.private PROPERTIES_ORDER_BY = { 'access': OrderBy(_order_properties_by_access, reverse=True), } T = TypeVar('T', Dataset, Model, Property, EnumItem) def sort( ordering: Dict[str, OrderBy], items: Iterable[T], order_by: Optional[str], ) -> Iterable[T]: order: Optional[OrderBy] = None if order_by: order = ordering[order_by] elif 'default' in ordering: order = ordering['default'] if order: return sorted(items, key=order.func, reverse=order.reverse) else: return items def _prefixes_to_tabular( prefixes: Dict[str, UriPrefix], *, separator: bool = False, ) -> Iterator[ManifestRow]: first = True for name, prefix in prefixes.items(): yield torow(DATASET, { 'id': prefix.id, 'type': prefix.type if first else '', 'ref': name, 'uri': prefix.uri, 'title': prefix.title, 'description': prefix.description, }) first = False if separator and prefixes: yield torow(DATASET, {}) def _backends_to_tabular( backends: Dict[str, Backend], *, separator: bool = False, ) -> Iterator[ManifestRow]: for name, backend in backends.items(): yield torow(DATASET, { 'type': backend.type, 'resource': name, 'source': backend.config.get('dsn'), }) if separator and backends: yield torow(DATASET, {}) def _namespaces_to_tabular( namespaces: Dict[str, Namespace], *, separator: bool = False, ) -> Iterator[ManifestRow]: namespaces = { k: ns for k, ns in namespaces.items() if not ns.generated } first = True for name, ns in namespaces.items(): yield torow(DATASET, { 'type': ns.type if first else '', 'ref': name, 'title': ns.title, 'description': ns.description, }) first = False if separator and namespaces: yield torow(DATASET, {}) def _order_enums_by_access(item: EnumItem): return item.access or Access.private ENUMS_ORDER_BY = { 'access': OrderBy(_order_enums_by_access, reverse=True), } def _enums_to_tabular( enums: Optional[Enums], *, external: bool = True, access: Access = Access.private, order_by: ManifestColumn = None, separator: bool = False, ) -> Iterator[ManifestRow]: if enums is None: return for name, enum in enums.items(): first = True items = sort(ENUMS_ORDER_BY, enum.values(), order_by) for item in items: if item.access is not None and item.access < access: continue yield torow(DATASET, { 'type': 'enum' if first else '', 'ref': name if first else '', 'source': item.source if external else '', 'prepare': unparse(item.prepare), 'access': item.given.access, 'title': item.title, 'description': item.description, }) if lang := list(_lang_to_tabular(item.lang)): first = True yield from lang else: first = False if separator and enums: yield torow(DATASET, {}) def _lang_to_tabular( lang: Optional[LangData], ) -> Iterator[ManifestRow]: if lang is None: return first = True for name, data in sorted(lang.items(), key=itemgetter(0)): yield torow(DATASET, { 'type': 'lang' if first else '', 'ref': name if first else '', 'title': data['title'], 'description': data['description'], }) first = False def _comments_to_tabular( comments: Optional[List[Comment]], *, access: Access = Access.private, ) -> Iterator[ManifestRow]: if comments is None: return first = True for comment in comments: if comment.access < access: return yield torow(DATASET, { 'id': comment.id, 'type': 'comment' if first else '', 'ref': comment.parent, 'source': comment.author, 'access': comment.given.access, 'title': comment.created, 'description': comment.comment, }) first = False def _dataset_to_tabular( dataset: Dataset, *, external: bool = True, access: Access = Access.private, order_by: ManifestColumn = None, ) -> Iterator[ManifestRow]: yield torow(DATASET, { 'id': dataset.id, 'dataset': dataset.name, 'level': dataset.level, 'access': dataset.given.access, 'title': dataset.title, 'description': dataset.description, }) yield from _lang_to_tabular(dataset.lang) yield from _prefixes_to_tabular(dataset.prefixes, separator=True) yield from _enums_to_tabular( dataset.ns.enums, external=external, access=access, order_by=order_by, ) def _resource_to_tabular( resource: Resource, *, external: bool = True, ) -> Iterator[ManifestRow]: backend = resource.backend yield torow(DATASET, { 'resource': resource.name, 'source': resource.external if external else '', 'prepare': unparse(resource.prepare or NA) if external else '', 'type': resource.type, 'ref': ( backend.name if ( external and backend and backend.origin != BackendOrigin.resource ) else '' ), 'level': resource.level, 'access': resource.given.access, 'title': resource.title, 'description': resource.description, }) yield from _lang_to_tabular(resource.lang) def _property_to_tabular( prop: Property, *, external: bool = True, access: Access = Access.private, order_by: ManifestColumn = None, ) -> Iterator[ManifestRow]: if prop.name.startswith('_'): return if prop.access < access: return data = { 'property': prop.place, 'type': prop.dtype.name, 'level': prop.level, 'access': prop.given.access, 'uri': prop.uri, 'title': prop.title, 'description': prop.description, } if external and prop.external: if isinstance(prop.external, list): # data['source'] = ', '.join(x.name for x in prop.external) # data['prepare'] = ', '.join( # unparse(x.prepare or NA) # for x in prop.external if x.prepare # ) raise DeprecationWarning( "Source can't be a list, use prepare instead." ) elif prop.external: data['source'] = prop.external.name data['prepare'] = unparse(prop.external.prepare or NA) if isinstance(prop.dtype, Ref): model = prop.model if model.external and model.external.dataset: data['ref'] = to_relative_model_name( prop.dtype.model, model.external.dataset, ) pkeys = prop.dtype.model.external.pkeys rkeys = prop.dtype.refprops if rkeys and pkeys != rkeys: rkeys = ', '.join([p.place for p in rkeys]) data['ref'] += f'[{rkeys}]' else: data['ref'] = prop.dtype.model.name elif prop.enum is not None: data['ref'] = prop.given.enum elif prop.unit is not None: data['ref'] = prop.given.unit yield torow(DATASET, data) yield from _comments_to_tabular(prop.comments, access=access) yield from _lang_to_tabular(prop.lang) yield from _enums_to_tabular( prop.enums, external=external, access=access, order_by=order_by, ) def _model_to_tabular( model: Model, *, external: bool = True, access: Access = Access.private, order_by: ManifestColumn = None, ) -> Iterator[ManifestRow]: data = { 'id': model.id, 'model': model.name, 'level': model.level, 'access': model.given.access, 'title': model.title, 'description': model.description, } if model.external and model.external.dataset: data['model'] = to_relative_model_name( model, model.external.dataset, ) if external and model.external: data.update({ 'source': model.external.name, 'prepare': unparse(model.external.prepare or NA), }) if ( not model.external.unknown_primary_key and all(p.access >= access for p in model.external.pkeys) ): # Add `ref` only if all properties are available in the # resulting manifest. data['ref'] = ', '.join([ p.name for p in model.external.pkeys ]) yield torow(DATASET, data) yield from _comments_to_tabular(model.comments, access=access) yield from _lang_to_tabular(model.lang) props = sort(PROPERTIES_ORDER_BY, model.properties.values(), order_by) for prop in props: yield from _property_to_tabular( prop, external=external, access=access, order_by=order_by, ) def datasets_to_tabular( manifest: Manifest, *, external: bool = True, # clean content of source and prepare access: Access = Access.private, internal: bool = False, # internal models with _ prefix like _txn order_by: ManifestColumn = None, ) -> Iterator[ManifestRow]: yield from _prefixes_to_tabular(manifest.prefixes, separator=True) yield from _backends_to_tabular(manifest.backends, separator=True) yield from _namespaces_to_tabular(manifest.namespaces, separator=True) yield from _enums_to_tabular( manifest.enums, external=external, access=access, order_by=order_by, separator=True, ) seen_datasets = set() dataset = None resource = None models = manifest.models if internal else take(manifest.models) models = sort(MODELS_ORDER_BY, models.values(), order_by) separator = False for model in models: if model.access < access: continue if model.external: if dataset is None or dataset.name != model.external.dataset.name: dataset = model.external.dataset if dataset: seen_datasets.add(dataset.name) resource = None separator = True yield from _dataset_to_tabular( dataset, external=external, access=access, order_by=order_by, ) if external and model.external and model.external.resource and ( resource is None or resource.name != model.external.resource.name ): resource = model.external.resource if resource: separator = True yield from _resource_to_tabular(resource, external=external) if separator: yield torow(DATASET, {}) else: separator = False yield from _model_to_tabular( model, external=external, access=access, order_by=order_by, ) datasets = sort(DATASETS_ORDER_BY, manifest.datasets.values(), order_by) for dataset in datasets: if dataset.name in seen_datasets: continue yield from _dataset_to_tabular( dataset, external=external, access=access, order_by=order_by, ) for resource in dataset.resources.values(): yield from _resource_to_tabular(resource) def torow(keys, values) -> ManifestRow: return {k: values.get(k) for k in keys} def render_tabular_manifest( manifest: Manifest, cols: List[ManifestColumn] = None, *, sizes: Dict[ManifestColumn, int] = None, ) -> str: rows = datasets_to_tabular(manifest) return render_tabular_manifest_rows(rows, cols, sizes=sizes) def render_tabular_manifest_rows( rows: Iterable[ManifestRow], cols: List[ManifestColumn] = None, *, sizes: Dict[ManifestColumn, int] = None, ) -> str: cols = cols or MANIFEST_COLUMNS hs = 1 if ID in cols else 0 # hierarchical cols start he = cols.index(PROPERTY) # hierarchical cols end hsize = 1 # hierarchical column size bsize = 3 # border size if sizes is None: sizes = dict( [(c, len(c)) for c in cols[:hs]] + [(c, 1) for c in cols[hs:he]] + [(c, len(c)) for c in cols[he:]] ) rows = list(rows) for row in rows: for i, col in enumerate(cols): val = '' if row[col] is None else str(row[col]) if col == ID: sizes[col] = 2 elif i < he: size = (hsize + bsize) * (he - hs - i) + sizes[PROPERTY] if size < len(val): sizes[PROPERTY] += len(val) - size elif sizes[col] < len(val): sizes[col] = len(val) line = [] for col in cols: size = sizes[col] line.append(col[:size].ljust(size)) lines = [line] for row in rows: if ID in cols: line = [row[ID][:2] if row[ID] else ' '] else: line = [] for i, col in enumerate(cols[hs:he + 1]): val = row[col] or '' if val: depth = i break else: val = '' depth = 0 line += [' ' * hsize] * depth size = (hsize + bsize) * (he - hs - depth) + sizes[PROPERTY] line += [val.ljust(size)] for col in cols[he + 1:]: val = '' if row[col] is None else str(row[col]) val = val.replace('|', '\\|') size = sizes[col] line.append(val.ljust(size)) lines.append(line) lines = [' | '.join(line) for line in lines] lines = [l.rstrip() for l in lines] return '\n'.join(lines) SHORT_NAMES = { 'd': 'dataset', 'r': 'resource', 'b': 'base', 'm': 'model', 'p': 'property', 't': 'type', } def normalizes_columns( cols: List[str], *, check_column_names: bool = True, ) -> List[ManifestColumn]: result: List[ManifestColumn] = [] unknown: List[str] = [] invalid: List[str] = [] for col in cols: col = col or '' col = col.strip().lower() col = SHORT_NAMES.get(col, col) col = cast(ManifestColumn, col) if col not in MANIFEST_COLUMNS: unknown.append(col) else: if unknown: result += unknown invalid += unknown unknown = [] result.append(col) if check_column_names and invalid: if len(invalid) == 1: raise PropertyNotFound(property=invalid[0]) else: raise MultipleErrors( PropertyNotFound(property=col) for col in invalid ) return result def write_tabular_manifest( path: str, rows: Union[ Manifest, Iterable[ManifestRow], None, ] = None, cols: List[ManifestColumn] = None, ) -> None: cols = cols or DATASET if rows is None: rows = [] elif isinstance(rows, Manifest): rows = datasets_to_tabular(rows) rows = ({c: row[c] for c in cols} for row in rows) if path.endswith('.csv'): _write_csv(pathlib.Path(path), rows, cols) elif path.endswith('.xlsx'): _write_xlsx(pathlib.Path(path), rows, cols) else: raise ValueError(f"Unknown tabular manifest format {path!r}.") def _write_csv( path: pathlib.Path, rows: Iterator[ManifestRow], cols: List[ManifestColumn], ) -> None: with path.open('w') as f: writer = csv.DictWriter(f, fieldnames=cols) writer.writeheader() writer.writerows(rows) def _write_xlsx( path: pathlib.Path, rows: Iterator[ManifestRow], cols: List[ManifestColumn], ) -> None: workbook = xlsxwriter.Workbook(path, { 'strings_to_formulas': False, 'strings_to_urls': False, }) bold = workbook.add_format({'bold': True}) formats = { 'id': workbook.add_format({ 'align': 'right', 'valign': 'top', }), 'dataset': workbook.add_format({ 'bold': True, 'valign': 'top', 'font_color': '#127622', }), 'resource': workbook.add_format({ 'valign': 'top', }), 'base': workbook.add_format({ 'valign': 'top', }), 'model': workbook.add_format({ 'bold': True, 'valign': 'top', 'font_color': '#127622', }), 'property': workbook.add_format({ 'valign': 'top', 'font_color': '#127622', }), 'type': workbook.add_format({ 'valign': 'top', }), 'ref': workbook.add_format({ 'valign': 'top', 'font_color': '#127622', }), 'source': workbook.add_format({ 'valign': 'top', 'font_color': '#c9211e', }), 'prepare': workbook.add_format({ 'valign': 'top', 'font_color': '#c9211e', }), 'level': workbook.add_format({ 'valign': 'top', }), 'access': workbook.add_format({ 'valign': 'top', }), 'uri': workbook.add_format({ 'valign': 'top', 'font_color': '#284f80', }), 'title': workbook.add_format({ 'valign': 'top', 'text_wrap': True, }), 'description': workbook.add_format({ 'valign': 'top', 'text_wrap': True, }), } sheet = workbook.add_worksheet() sheet.freeze_panes(1, 0) # Freeze the first row. sheet.set_column('A:E', 2) # id, d, r, b, m sheet.set_column('F:F', 20) # property sheet.set_column('I:J', 20) # source, prepare sheet.set_column('N:N', 20) # title sheet.set_column('O:O', 30) # description for j, col in enumerate(cols): sheet.write(0, j, col, bold) for i, row in enumerate(rows, 1): for j, col in enumerate(cols): val = row[col] fmt = formats.get(col) sheet.write(i, j, val, fmt) workbook.close()
50,976
15,256
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # coding: utf-8 """Context management API of mxnet.""" from __future__ import absolute_import import threading import warnings from .base import classproperty, with_metaclass, _MXClassPropertyMetaClass class Context(with_metaclass(_MXClassPropertyMetaClass, object)): """Constructs a context. MXNet can run operations on CPU and different GPUs. A context describes the device type and ID on which computation should be carried on. One can use mx.cpu and mx.gpu for short. See also ---------- `How to run MXNet on multiple CPU/GPUs <http://mxnet.io/faq/multi_devices.html>` for more details. Parameters ---------- device_type : {'cpu', 'gpu'} or Context. String representing the device type. device_id : int (default=0) The device id of the device, needed for GPU. Note ---- Context can also be used as a way to change the default context. Examples -------- >>> # array on cpu >>> cpu_array = mx.nd.ones((2, 3)) >>> # switch default context to GPU(2) >>> with mx.Context(mx.gpu(2)): ... gpu_array = mx.nd.ones((2, 3)) >>> gpu_array.context gpu(2) One can also explicitly specify the context when creating an array. >>> gpu_array = mx.nd.ones((2, 3), mx.gpu(1)) >>> gpu_array.context gpu(1) """ # static class variable _default_ctx = threading.local() devtype2str = {1: 'cpu', 2: 'gpu', 3: 'cpu_pinned', 5: 'cpu_shared'} devstr2type = {'cpu': 1, 'gpu': 2, 'cpu_pinned': 3, 'cpu_shared': 5} def __init__(self, device_type, device_id=0): if isinstance(device_type, Context): self.device_typeid = device_type.device_typeid self.device_id = device_type.device_id else: self.device_typeid = Context.devstr2type[device_type] self.device_id = device_id self._old_ctx = None @property def device_type(self): """Returns the device type of current context. Examples ------- >>> mx.context.current_context().device_type 'cpu' >>> mx.current_context().device_type 'cpu' Returns ------- device_type : str """ return Context.devtype2str[self.device_typeid] def __hash__(self): """Compute hash value of context for dictionary lookup""" return hash((self.device_typeid, self.device_id)) def __eq__(self, other): """Compares two contexts. Two contexts are equal if they have the same device type and device id. """ return isinstance(other, Context) and \ self.device_typeid == other.device_typeid and \ self.device_id == other.device_id def __str__(self): return '%s(%d)' % (self.device_type, self.device_id) def __repr__(self): return self.__str__() def __enter__(self): if not hasattr(Context._default_ctx, "value"): Context._default_ctx.value = Context('cpu', 0) self._old_ctx = Context._default_ctx.value Context._default_ctx.value = self return self def __exit__(self, ptype, value, trace): Context._default_ctx.value = self._old_ctx #pylint: disable=no-self-argument @classproperty def default_ctx(cls): warnings.warn("Context.default_ctx has been deprecated. " "Please use Context.current_context() instead. " "Please use test_utils.set_default_context to set a default context", DeprecationWarning) if not hasattr(Context._default_ctx, "value"): cls._default_ctx.value = Context('cpu', 0) return cls._default_ctx.value @default_ctx.setter def default_ctx(cls, val): warnings.warn("Context.default_ctx has been deprecated. " "Please use Context.current_context() instead. " "Please use test_utils.set_default_context to set a default context", DeprecationWarning) cls._default_ctx.value = val #pylint: enable=no-self-argument # initialize the default context in Context Context._default_ctx.value = Context('cpu', 0) def cpu(device_id=0): """Returns a CPU context. This function is a short cut for ``Context('cpu', device_id)``. For most operations, when no context is specified, the default context is `cpu()`. Examples ---------- >>> with mx.cpu(): ... cpu_array = mx.nd.ones((2, 3)) >>> cpu_array.context cpu(0) >>> cpu_array = mx.nd.ones((2, 3), ctx=mx.cpu()) >>> cpu_array.context cpu(0) Parameters ---------- device_id : int, optional The device id of the device. `device_id` is not needed for CPU. This is included to make interface compatible with GPU. Returns ------- context : Context The corresponding CPU context. """ return Context('cpu', device_id) def cpu_pinned(device_id=0): """Returns a CPU pinned memory context. Copying from CPU pinned memory to GPU is faster than from normal CPU memory. This function is a short cut for ``Context('cpu_pinned', device_id)``. Examples ---------- >>> with mx.cpu_pinned(): ... cpu_array = mx.nd.ones((2, 3)) >>> cpu_array.context cpu_pinned(0) >>> cpu_array = mx.nd.ones((2, 3), ctx=mx.cpu_pinned()) >>> cpu_array.context cpu_pinned(0) Parameters ---------- device_id : int, optional The device id of the device. `device_id` is not needed for CPU. This is included to make interface compatible with GPU. Returns ------- context : Context The corresponding CPU pinned memory context. """ return Context('cpu_pinned', device_id) def gpu(device_id=0): """Returns a GPU context. This function is a short cut for Context('gpu', device_id). The K GPUs on a node are typically numbered as 0,...,K-1. Examples ---------- >>> cpu_array = mx.nd.ones((2, 3)) >>> cpu_array.context cpu(0) >>> with mx.gpu(1): ... gpu_array = mx.nd.ones((2, 3)) >>> gpu_array.context gpu(1) >>> gpu_array = mx.nd.ones((2, 3), ctx=mx.gpu(1)) >>> gpu_array.context gpu(1) Parameters ---------- device_id : int, optional The device id of the device, needed for GPU. Returns ------- context : Context The corresponding GPU context. """ return Context('gpu', device_id) def current_context(): """Returns the current context. By default, `mx.cpu()` is used for all the computations and it can be overridden by using `with mx.Context(x)` statement where x can be cpu(device_id) or gpu(device_id). Examples ------- >>> mx.current_context() cpu(0) >>> with mx.Context('gpu', 1): # Context changed in `with` block. ... mx.current_context() # Computation done here will be on gpu(1). ... gpu(1) >>> mx.current_context() # Back to default context. cpu(0) Returns ------- default_ctx : Context """ if not hasattr(Context._default_ctx, "value"): Context._default_ctx.value = Context('cpu', 0) return Context._default_ctx.value
8,080
2,509
# theory MPD client # Copyright (C) 2008 Ryan Roemmich <ralfonso@gmail.com> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import ConfigParser import pickle from pylons import config from pylons import app_globals as g class ConfigFileError(Exception): pass class TConfig: """ handles the global configuration. loaded into app globals at application startup also handles committing the configuration to disk to maintain across app restarts """ def __init__(self): """ try to read the configuration from disk """ self.server = None self.port = None self.password = None self.webpassword = '' self.timeout = False self.awskey = None self.aws_secret = None self.streams = [] self.default_search = 'Any' conf = ConfigParser.ConfigParser() conf.read(config['localconf']) try: self.server = conf.get('mpd','server') self.port = conf.get('mpd','port') self.password = conf.get('mpd','password') self.awskey = conf.get('services','awskey') self.aws_secret = conf.get('services','aws_secret') self.webpassword = conf.get('main','webpassword') self.timeout = conf.getboolean('main','timeout') self.default_search = conf.get('main','default_search') conf_stream = conf.get('ext','streams') except (ConfigParser.NoSectionError,ConfigParser.NoOptionError): pass try: self.streams = pickle.loads(eval(conf_stream)) except: # we don't really care what happened, the user must have messed with the magic pickled string :) pass def commit_config(self): """ commit the configuration to disk """ conf = ConfigParser.ConfigParser() conf.add_section("mpd") conf.set("mpd", "server",self.server) conf.set("mpd", "port",self.port) conf.set("mpd", "password",self.password) conf.add_section("services") conf.set('services','awskey',self.awskey) conf.set('services','aws_secret',self.aws_secret) conf.add_section('main') conf.set('main','webpassword',self.webpassword) conf.set('main','timeout',self.timeout) conf.set('main','default_search',self.default_search) conf.add_section('ext') conf.set('ext','streams',repr(pickle.dumps(self.streams))) try: conffile = open(config['localconf'],"w") conf.write(conffile) except IOError,e: raise ConfigFileErro def get_stream_name(self,url): """ search the list of streams for a particular name """ for s in self.streams: if s[1] == url: return s[0]
3,422
972
"""Test `iotanbo_py_utils.error.py`.""" from iotanbo_py_utils.error import Error from iotanbo_py_utils.error import ErrorKind def test_error_one_of_arithmetic_errors() -> None: errs = ( Error(ErrorKind.ArithmeticError), Error(ErrorKind.FloatingPointError), Error(ErrorKind.OverflowError), Error(ErrorKind.ZeroDivisionError), ) for err in errs: assert err.one_of_arithmetic_errors() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_arithmetic_errors() def test_error_one_of_import_errors() -> None: errs = ( Error(ErrorKind.ImportError), Error(ErrorKind.ModuleNotFoundError), ) for err in errs: assert err.one_of_import_errors() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_import_errors() def test_error_one_of_lookup_errors() -> None: errs = ( Error(ErrorKind.LookupError), Error(ErrorKind.IndexError), Error(ErrorKind.KeyError), ) for err in errs: assert err.one_of_lookup_errors() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_lookup_errors() def test_error_one_of_name_errors() -> None: errs = ( Error(ErrorKind.NameError), Error(ErrorKind.UnboundLocalError), ) for err in errs: assert err.one_of_name_errors() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_name_errors() def test_error_one_of_os_errors() -> None: errs = ( Error(ErrorKind.OSError), Error(ErrorKind.BlockingIOError), Error(ErrorKind.ChildProcessError), Error(ErrorKind.ConnectionError), Error(ErrorKind.BrokenPipeError), Error(ErrorKind.ConnectionAbortedError), Error(ErrorKind.ConnectionRefusedError), Error(ErrorKind.ConnectionResetError), Error(ErrorKind.FileExistsError), Error(ErrorKind.FileNotFoundError), Error(ErrorKind.InterruptedError), Error(ErrorKind.IsADirectoryError), Error(ErrorKind.NotADirectoryError), Error(ErrorKind.PermissionError), Error(ErrorKind.ProcessLookupError), Error(ErrorKind.TimeoutError), ) for err in errs: assert err.one_of_os_errors() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_os_errors() def test_error_one_of_runtime_errors() -> None: errs = ( Error(ErrorKind.RuntimeError), Error(ErrorKind.NotImplementedError), Error(ErrorKind.RecursionError), ) for err in errs: assert err.one_of_runtime_errors() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_runtime_errors() def test_error_one_of_syntax_errors() -> None: errs = ( Error(ErrorKind.SyntaxError), Error(ErrorKind.IndentationError), Error(ErrorKind.TabError), ) for err in errs: assert err.one_of_syntax_errors() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_syntax_errors() def test_error_one_of_value_errors() -> None: errs = ( Error(ErrorKind.ValueError), Error(ErrorKind.UnicodeError), Error(ErrorKind.UnicodeDecodeError), Error(ErrorKind.UnicodeEncodeError), Error(ErrorKind.UnicodeTranslateError), ) for err in errs: assert err.one_of_value_errors() # == negative path == err = Error(ErrorKind.SyntaxError) assert not err.one_of_value_errors() def test_error_one_of_warnings() -> None: errs = ( Error(ErrorKind.Warning), Error(ErrorKind.DeprecationWarning), Error(ErrorKind.PendingDeprecationWarning), Error(ErrorKind.RuntimeWarning), Error(ErrorKind.SyntaxWarning), Error(ErrorKind.UserWarning), Error(ErrorKind.FutureWarning), Error(ErrorKind.ImportWarning), Error(ErrorKind.UnicodeWarning), Error(ErrorKind.BytesWarning), Error(ErrorKind.ResourceWarning), ) for err in errs: assert err.one_of_warnings() # == negative path == err = Error(ErrorKind.ValueError) assert not err.one_of_warnings() class _CustomException(Exception): ... def test_error_from_exception() -> None: # error from exception, preserve kind e = Error.from_exception(ValueError("test")) assert e.kind == ErrorKind.ValueError assert not e.cause assert e.msg == "test" # new kind replaces exception's kind e = Error.from_exception(ValueError("test"), new_kind=ErrorKind.Warning) assert e.kind == ErrorKind.Warning assert e.cause == ErrorKind.ValueError assert e.msg == "test" # error from custom exception, preserve kind try: raise _CustomException() except _CustomException as ex: e = Error.from_exception(ex) assert e.kind == "_CustomException" assert not e.cause assert e.msg == ""
5,055
1,559
from setuptools import setup, find_packages from os import path from io import open here = path.abspath(path.dirname(__file__)) # Get the long description from the README file with open(path.join(here, 'README.md'), encoding='utf-8') as f: long_description = f.read() setup( name='orvibo', version='1.5.0', description='Python module to controll Orvibo devices, such as s20 wifi sockets and AllOne IR blasters', long_description=long_description, long_description_content_type='text/markdown', url='https://github.com/cherezov/orvibo', packages=find_packages(exclude=['contrib', 'docs', 'tests']), # Required )
651
218
import glob, os import subprocess from difflib import context_diff class TestRunner(object): def __init__(self, context): self.context = context self.error_count = 0 self.test_count = 0 self.success_count = 0 def run(self): os.getcwd() os.chdir(self.context.repo_dir) search = os.path.join(self.context.testcasedir, '*.in') problem_files = glob.glob(search) if len(problem_files) == 0: self.warn("No problem files found. Does directory '" + self.context.testcasedir + "' exist?") return file = self.context.runscript if not os.path.isfile(file): self.error("Could not find file '{}' to run the program.".format(file)) return for problem_file in problem_files: self.test_count += 1 output = self.get_output(problem_file) expected_output = self.get_expected_output(problem_file) if self.compare(problem_file, output, expected_output): self.success_count += 1 else: self.error_count += 1 def compare(self, problem_file, output, expected_output): diff_iterator = context_diff(output, expected_output, fromfile='program output', tofile='expected') diff = "" for char in diff_iterator: diff += char if len(diff) == 0: self.info("Testing '" + problem_file + "'. Result: output CORRECT") return True else: self.error("Testing '" + problem_file + "'. Result: output DIFFERENT") self.error(" Expected:") for line in expected_output.split('\n'): self.error(" " + line) self.error(" Actual:") for line in output.split('\n'): self.error(" " + line) return False def info(self, message): self.context.logger.info(message) def warn(self, message): self.context.logger.warn(message) def error(self, message): self.context.logger.error(message) def get_expected_output(self, problem_file): expected_output_file = problem_file[:-2] + 'out' with open(expected_output_file, 'r') as file: return file.read() def get_output(self, problem_file): runscript = os.path.join(self.context.repo_dir, self.context.runscript) out = '' err = '' try: with open(problem_file, 'r') as input: p = subprocess.Popen(runscript, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin = input) out, err = p.communicate() p.wait() if len(err) > 0: self.warn('Stderr is outputting text:') for line in err.split('\n'): self.warn(line) except Exception as e: self.error('Caught unexpected error: ' + str(e)) return out
3,053
864
from amazon.model import db def __search_by_admin_name(username): query={'username': username} matching_user = db['users'].find(query) if matching_user.count() > 0: return matching_user.next() else: return None
245
76
expected = [ { "abstract_type": None, "content": "RET can be activated in cis or trans by its co-receptors and ligands in vitro, but the physiological roles of trans signaling are unclear. Rapidly adapting (RA) mechanoreceptors in dorsal root ganglia (DRGs) express Ret and the co-receptor Gfr\u03b12 and depend on Ret for survival and central projection growth. Here, we show that Ret and Gfr\u03b12 null mice display comparable early central projection deficits, but Gfr\u03b12 null RA mechanoreceptors recover later. Loss of Gfr\u03b11, the co-receptor implicated in activating RET in trans, causes no significant central projection or cell survival deficit, but Gfr\u03b11;Gfr\u03b12 double nulls phenocopy Ret nulls. Finally, we demonstrate that GFR\u03b11 produced by neighboring DRG neurons activates RET in RA mechanoreceptors. Taken together, our results suggest that trans and cis RET signaling could function in the same developmental process and that the availability of both forms of activation likely enhances but not diversifies outcomes of RET signaling.", "full_content": "<p>RET can be activated in <italic>cis</italic> or <italic>trans</italic> by its co-receptors and ligands <italic>in vitro</italic>, but the physiological roles of <italic>trans</italic> signaling are unclear. Rapidly adapting (RA) mechanoreceptors in dorsal root ganglia (DRGs) express <italic>Ret</italic> and the co-receptor <italic>Gfr\u03b12</italic> and depend on <italic>Ret</italic> for survival and central projection growth. Here, we show that <italic>Ret</italic> and <italic>Gfr\u03b12</italic> null mice display comparable early central projection deficits, but <italic>Gfr\u03b12</italic> null RA mechanoreceptors recover later. Loss of <italic>Gfr\u03b11</italic>, the co-receptor implicated in activating RET <italic>in trans</italic>, causes no significant central projection or cell survival deficit, but <italic>Gfr\u03b11;Gfr\u03b12</italic> double nulls phenocopy <italic>Ret</italic> nulls. Finally, we demonstrate that GFR\u03b11 produced by neighboring DRG neurons activates RET in RA mechanoreceptors. Taken together, our results suggest that <italic>trans</italic> and <italic>cis</italic> RET signaling could function in the same developmental process and that the availability of both forms of activation likely enhances but not diversifies outcomes of RET signaling.</p>", }, ]
2,429
739
# coding=utf-8 import optparse import torch import time import pickle from torch.autograd import Variable from loader import * from utils import * # python -m visdom.server optparser = optparse.OptionParser() optparser.add_option( "-t", "--test", default="data/eng.testb", help="Test set location" ) optparser.add_option( '--score', default='evaluation/temp/score.txt', help='score file location' ) optparser.add_option( "-g", '--use_gpu', default='1', type='int', help='whether or not to ues gpu' ) optparser.add_option( '--loss', default='loss.txt', help='loss file location' ) optparser.add_option( '--model_path', default='models/test', help='model path' ) optparser.add_option( '--map_path', default='models/mapping.pkl', help='model path' ) optparser.add_option( '--char_mode', choices=['CNN', 'LSTM'], default='CNN', help='char_CNN or char_LSTM' ) opts = optparser.parse_args()[0] mapping_file = opts.map_path with open(mapping_file, 'rb') as f: mappings = pickle.load(f) word_to_id = mappings['word_to_id'] tag_to_id = mappings['tag_to_id'] id_to_tag = {k[1]: k[0] for k in tag_to_id.items()} char_to_id = mappings['char_to_id'] parameters = mappings['parameters'] word_embeds = mappings['word_embeds'] use_gpu = opts.use_gpu == 1 and torch.cuda.is_available() assert os.path.isfile(opts.test) assert parameters['tag_scheme'] in ['iob', 'iobes'] if not os.path.isfile(eval_script): raise Exception('CoNLL evaluation script not found at "%s"' % eval_script) if not os.path.exists(eval_temp): os.makedirs(eval_temp) lower = parameters['lower'] zeros = parameters['zeros'] tag_scheme = parameters['tag_scheme'] test_sentences = load_sentences(opts.test, lower, zeros) update_tag_scheme(test_sentences, tag_scheme) test_data = prepare_dataset( test_sentences, word_to_id, char_to_id, tag_to_id, lower ) model = torch.load(opts.model_path) model_name = opts.model_path.split('/')[-1].split('.')[0] if use_gpu: model.cuda() model.eval() def eval(model, datas): prediction = [] confusion_matrix = torch.zeros((len(tag_to_id) - 2, len(tag_to_id) - 2)) for data in datas: ground_truth_id = data['tags'] words = data['str_words'] chars2 = data['chars'] caps = data['caps'] if parameters['char_mode'] == 'LSTM': chars2_sorted = sorted(chars2, key=lambda p: len(p), reverse=True) d = {} for i, ci in enumerate(chars2): for j, cj in enumerate(chars2_sorted): if ci == cj and not j in d and not i in d.values(): d[j] = i continue chars2_length = [len(c) for c in chars2_sorted] char_maxl = max(chars2_length) chars2_mask = np.zeros((len(chars2_sorted), char_maxl), dtype='int') for i, c in enumerate(chars2_sorted): chars2_mask[i, :chars2_length[i]] = c chars2_mask = Variable(torch.LongTensor(chars2_mask)) if parameters['char_mode'] == 'CNN': d = {} chars2_length = [len(c) for c in chars2] char_maxl = max(chars2_length) chars2_mask = np.zeros((len(chars2_length), char_maxl), dtype='int') for i, c in enumerate(chars2): chars2_mask[i, :chars2_length[i]] = c chars2_mask = Variable(torch.LongTensor(chars2_mask)) dwords = Variable(torch.LongTensor(data['words'])) dcaps = Variable(torch.LongTensor(caps)) if use_gpu: val, out = model(dwords.cuda(), chars2_mask.cuda(), dcaps.cuda(),chars2_length, d) else: val, out = model(dwords, chars2_mask, dcaps, chars2_length, d) predicted_id = out for (word, true_id, pred_id) in zip(words, ground_truth_id, predicted_id): line = ' '.join([word, id_to_tag[true_id], id_to_tag[pred_id]]) prediction.append(line) confusion_matrix[true_id, pred_id] += 1 prediction.append('') predf = eval_temp + '/pred.' + model_name scoref = eval_temp + '/score.' + model_name with open(predf, 'w') as f: f.write('\n'.join(prediction)) os.system('%s < %s > %s' % (eval_script, predf, scoref)) with open(scoref, 'r') as f: for l in f.readlines(): print(l.strip()) print(("{: >2}{: >7}{: >7}%s{: >9}" % ("{: >7}" * confusion_matrix.size(0))).format( "ID", "NE", "Total", *([id_to_tag[i] for i in range(confusion_matrix.size(0))] + ["Percent"]) )) for i in range(confusion_matrix.size(0)): print(("{: >2}{: >7}{: >7}%s{: >9}" % ("{: >7}" * confusion_matrix.size(0))).format( str(i), id_to_tag[i], str(confusion_matrix[i].sum().item()), *([confusion_matrix[i][j] for j in range(confusion_matrix.size(0))] + ["%.3f" % (confusion_matrix[i][i] * 100. / max(1, confusion_matrix[i].sum()))]) )) t = time.time() eval(model, test_data) print(time.time() - t)
5,059
1,825
from openbiolink.graph_creation.metadata_db_file.edge.dbMetadataEdge import DbMetadataEdge from openbiolink.graph_creation.types.dbType import DbType class DbMetaEdgeSiderInd(DbMetadataEdge): NAME = 'Edge - Sider - Indications' URL = "http://sideeffects.embl.de/media/download/meddra_all_indications.tsv.gz" OFILE_NAME = "SIDER_dis_drug.tsv.gz" COLS = ['stichID', 'umlsID', 'method', 'umlsName', 'medDRAumlsType', 'medDRAumlsID', 'medDRAumlsName'] FILTER_COLS = ['umlsID', 'stichID', 'method'] HEADER = 0 DB_TYPE = DbType.DB_EDGE_SIDER_IND def __init__(self): super().__init__(url= DbMetaEdgeSiderInd.URL, ofile_name= DbMetaEdgeSiderInd.OFILE_NAME, dbType= DbMetaEdgeSiderInd.DB_TYPE)
789
292
import mahotas as mh import glob import os # rename files from 0 to number of files. files = glob.glob("/path/*") files.sort() if not os.path.isdir("/path/renamed/"): os.mkdir("/path/renamed/") for i in range(len(files)): print("Processing " + files[i]) renamedFile = mh.imread(files[i]) mh.imsave("/path/renamed/" + str(i).zfill(5) + ".tif", renamedFile)
377
144
import os import re from base64 import b64decode from pathlib import Path import requests username = os.environ["GITHUB_USERNAME"] password = os.environ["GITHUB_PERSONAL_ACCESS_TOKEN"] auth = requests.auth.HTTPBasicAuth(username, password) directory = Path(__file__).resolve().parent.parent / "github" directory.mkdir(exist_ok=True) start_url = "https://api.github.com/search/code?q=view+language:lookml" next_url = None page = 1 with requests.Session() as session: session.auth = auth while True: response = session.get(next_url or start_url) response.raise_for_status() links = response.headers["Link"] finds = re.findall( r"<(https://api.github.com/search/code\?" r'q=view\+language%3Alookml&page=\d+)>; rel="next"', links, ) if finds: next_url = finds[0] else: next_url = None print(next_url) urls = [item["url"] for item in response.json()["items"]] print(f"Downloading all content from page {page}") for url in urls: response = session.get(url) response.raise_for_status() response_json = response.json() name = response_json["name"] encoded = response_json["content"] content = b64decode(encoded).decode("utf-8") if ( name.endswith(".lookml") or content.startswith("-") or "- view" in content ): continue file_path = directory / name with file_path.open("w+") as file: file.write(content) if next_url is None: break else: page += 1
1,749
512
import os import sys sys.path.insert(0, os.path.abspath('..')) def get_version(filename): from re import findall with open(filename) as f: metadata = dict(findall(r"__([a-z]+)__ = '([^']+)'", f.read())) return metadata['version'] project = 'carreralib' copyright = '2015-2017 Thomas Kemmer' version = get_version(b'../carreralib/__init__.py') release = version extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.coverage', 'sphinx.ext.doctest', 'sphinx.ext.todo' ] exclude_patterns = ['_build'] master_doc = 'index' html_theme = 'default'
579
230
def lcs(x, y): """ Longest Common Subsequence """ n = len(x) + 1 m = len(y) + 1 table = [ [0]*m for i in range(n) ] for i in range(n): for j in range(m): # If either string is empty, then lcs = 0 if i == 0 or j == 0: table[i][j] = 0 elif x[i - 1] == y[j - 1]: table[i][j] = 1 + table[i-1][j-1] else: table[i][j] = max(table[i-1][j], table[i][j-1]) return table[len(x)][len(y)] if __name__ == '__main__': x = "AGGTAB" y = "GXTXAYB" print lcs(x, y)
503
276
from models.methods import method from typing import Any class SetDeviceInfoMethod(method.Method): def __init__(self, params: Any): super().__init__("set_device_info", params) def set_request_time_milis(self, t: float): self.requestTimeMils = t def set_terminal_uuid(self, uuid: str): self.terminalUUID = uuid
350
112
#!/usr/bin/env python # encoding: utf-8 """ @author: sherlock @contact: sherlockliao01@gmail.com """ import logging import os import sys sys.path.append('.') from fastreid.config import get_cfg from projects.InterpretationReID.interpretationreid.engine import DefaultTrainer, default_argument_parser, default_setup, launch from fastreid.utils.checkpoint import Checkpointer from projects.InterpretationReID.interpretationreid.evaluation import ReidEvaluator import projects.InterpretationReID.interpretationreid as PII from fastreid.utils.logger import setup_logger class Trainer(DefaultTrainer): @classmethod def build_train_loader(cls, cfg): """ Returns: iterable It now calls :func:`fastreid.data.build_detection_train_loader`. Overwrite it if you'd like a different data loader. """ logger = logging.getLogger(__name__) logger.info("Prepare training set") return PII.add_build_reid_train_loader(cfg) @classmethod def build_test_loader(cls, cfg, dataset_name): """ Returns: iterable It now calls :func:`fastreid.data.build_detection_test_loader`. Overwrite it if you'd like a different data loader. """ return PII.add_build_reid_test_loader(cfg, dataset_name) @classmethod def build_evaluator(cls, cfg, num_query, output_folder=None): if output_folder is None: output_folder = os.path.join(cfg.OUTPUT_DIR, "inference") return ReidEvaluator(cfg, num_query) def setup(args): """ Create configs_old and perform basic setups. """ cfg = get_cfg() PII.add_interpretation_config(cfg) cfg.merge_from_file(args.config_file) cfg.merge_from_list(args.opts) cfg.freeze() default_setup(cfg, args) return cfg def main(args): cfg = setup(args) if args.eval_only: cfg.defrost() cfg.MODEL.BACKBONE.PRETRAIN = False model = Trainer.build_model(cfg) Checkpointer(model).load(cfg.MODEL.WEIGHTS) # load trained model res = Trainer.test(cfg, model) return res trainer = Trainer(cfg) trainer.resume_or_load(resume=args.resume) if cfg.INTERPRETATION.PRETRAIN_MODEL: #print("trainer.load_n_or_not()") trainer.load_n_or_not() #print("load success") #print(trainer.model) #for p in trainer.model.backbone_1.parameters(): #p.requires_grad=False #for p in trainer.model.backbone_2.parameters(): #p.requires_grad=False #for p in trainer.model.heads.parameters(): #.module #p.requires_grad=False #print("trainer.train()") #print(cfg) #print(trainer._hooks) #setup_logger() return trainer.train() if __name__ == "__main__": args = default_argument_parser().parse_args() print("Command Line Args:", args) launch( main, args.num_gpus, num_machines=args.num_machines, machine_rank=args.machine_rank, dist_url=args.dist_url, args=(args,), )
3,077
1,039
from collections import deque from asynciomeasures.events import Event class Collector(deque): """Caped list of metrics """ def flush(self, rate=None, formatter=None): while True: try: metric = self.popleft() if isinstance(metric, Event): yield formatter(metric) continue if metric.value is None: continue if rate and metric.rate and rate < metric.rate: continue if formatter: try: yield formatter(metric) except ValueError: continue else: yield metric except IndexError: return
826
178
# Copyright 2021, Peter Birch, mailto:peter@lightlogic.co.uk # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from cocotb_bus.drivers import Driver from cocotb.triggers import RisingEdge from ..driver_common import BaseDriver class StateInitiator(BaseDriver): """ Drivers signal state updates as an initiator """ async def _driver_send(self, transaction, sync=True, **kwargs): """ Send queued transactions onto the interface. Args: transaction: Transaction to send sync : Align to the rising clock edge before sending **kwargs : Any other arguments """ # Synchronise to the rising edge if sync: await RisingEdge(self.clock) # Wait for reset to clear while self.reset == 1: await RisingEdge(self.clock) # Drive the request self.intf.index <= transaction.index self.intf.is_seq <= transaction.sequential self.intf.value <= transaction.state self.intf.update <= 1 await RisingEdge(self.clock) self.intf.update <= 0
1,576
447
# Prim's Algorithm in Python INF = 9999999 # number of vertices in graph V = 5 # create a 2d array of size 5x5 # for adjacency matrix to represent graph G = [ [0, 9, 75, 0, 0], [9, 0, 95, 19, 42], [75, 95, 0, 51, 66], [0, 19, 51, 0, 31], [0, 42, 66, 31, 0], ] # create a array to track selected vertex # selected will become true otherwise false selected = [0, 0, 0, 0, 0] # set number of edge to 0 no_edge = 0 # the number of egde in minimum spanning tree will be # always less than(V - 1), where V is number of vertices in # graph # choose 0th vertex and make it true selected[0] = True # print for edge and weight print("Edge : Weight\n") while no_edge < V - 1: # For every vertex in the set S, find the all adjacent vertices # , calculate the distance from the vertex selected at step 1. # if the vertex is already in the set S, discard it otherwise # choose another vertex nearest to selected vertex at step 1. minimum = INF x = 0 y = 0 for i in range(V): if selected[i]: for j in range(V): if (not selected[j]) and G[i][j]: # not in selected and there is an edge if minimum > G[i][j]: minimum = G[i][j] x = i y = j print(str(x) + "-" + str(y) + ":" + str(G[x][y])) selected[y] = True no_edge += 1
1,463
553
#!/usr/bin/env python '''Tests for the likelihood.py module''' from time import perf_counter_ns import pytest import numpy as np from numpy.testing import assert_array_equal, assert_almost_equal from scipy.stats import gamma import likelihood SMALL_FIT_PARAMS = { 'baseline_intensities': np.asarray([1, 2, np.nan, np.nan]), 'r_h': 1.5, 'r_c': 0.5 } SIMPLE_DIST_PARAMS = { 'self_excitation_shape': 2, 'self_excitation_scale': 1, 'discharge_excitation_shape': 3, 'discharge_excitation_scale': 2 } SMALL_CASES_FILE = 'tests/fixtures/small.csv' SMALL_COVARIATES_FILE = 'tests/fixtures/small_covariates.csv' LARGE_FIT_PARAMS = { 'baseline_intensities': np.asarray([0.3, 0.4, 0.6, 0.9]), 'r_h': 1.5, 'r_c': 0.5 } FULL_DIST_PARAMS = { 'self_excitation_shape': 2.6, 'self_excitation_scale': 2.5, 'discharge_excitation_shape': 2.6, 'discharge_excitation_scale': 2.5 } def test_gamma_pdf(): x = np.linspace(0, 10, 100) shape = FULL_DIST_PARAMS['self_excitation_shape'] scale = FULL_DIST_PARAMS['self_excitation_scale'] assert_almost_equal( gamma.pdf(x, a=shape, scale=scale), likelihood.gamma_pdf(x, shape, scale) ) @pytest.mark.parametrize( "test_element,result_dtype", [(123_456_789, np.uint32), (65_535, np.uint16), (255, np.uint8)] ) def test_compactify(test_element, result_dtype): '''Test that arrays compactify correctly, and to the correct data types''' array = np.asarray([[1, 2], [3, test_element]], dtype=np.uint32) result = likelihood.compactify(array) assert result.dtype == result_dtype assert_array_equal(array, result) def test_read_and_tidy_data(): '''Test that a CSV file with care home IDs as a header row is read, sorted, and split correctly.''' ids, values = likelihood.read_and_tidy_data(SMALL_CASES_FILE) assert_array_equal(ids, [14, 16, 35]) assert_array_equal( values, [[4, 1, 6], [4, 0, 3], [6, 66, 2]] ) @pytest.fixture def small_cases(): '''Get a small data file that could be cases or discharges.''' return likelihood.read_and_tidy_data(SMALL_CASES_FILE) @pytest.fixture def small_covariates(): '''Get a small data file containing covariates.''' return likelihood.read_and_tidy_data(SMALL_COVARIATES_FILE) def test_carehome_intensity_null(small_cases, small_covariates): '''Test that calculating the null-case intensity (based on mapping banded carehome size to a base intensity) gives the correct result''' _, cases = small_cases _, covariates = small_covariates intensity = likelihood.carehome_intensity_null( covariates=covariates, cases=cases, fit_params=SMALL_FIT_PARAMS ) assert_array_equal(intensity, [[1, 2, 2], [1, 2, 2], [1, 2, 2]]) def test_single_excitation(small_cases): '''Test that excitation terms of the form e_i(t) = \\sum_{s<t} f(t - s) triggers_i(s) are correctly calculated''' _, cases = small_cases excitation = likelihood.single_excitation(cases, 2, 1) assert_almost_equal( excitation, [[0, 0, 0], [1.472, 0.368, 2.207], [2.554, 0.271, 2.728]], decimal=3 ) def test_cached_single_excitation(small_cases): ''' Test that the caching of the single_excitation function works correctly. ''' _, cases = small_cases cases.flags.writeable = False shape = SIMPLE_DIST_PARAMS['self_excitation_shape'] scale = SIMPLE_DIST_PARAMS['self_excitation_scale'] uncached_start = perf_counter_ns() uncached_excitation = likelihood.single_excitation(cases, shape, scale) uncached_end = perf_counter_ns() first_excitation = likelihood.cached_single_excitation( cases, shape, scale ) assert_array_equal(uncached_excitation, first_excitation) cached_start = perf_counter_ns() cached_excitation = likelihood.cached_single_excitation( cases, shape, scale ) cached_end = perf_counter_ns() assert_array_equal(uncached_excitation, cached_excitation) # Cached version should be quicker assert (cached_end - cached_start) < (uncached_end - uncached_start) def test_carehome_intensity_no_discharges(small_cases, small_covariates): '''Test that the behaviour of carehome_intensity in the case where discharges are not considered.''' _, cases = small_cases _, covariates = small_covariates fit_params_no_rh = {**SMALL_FIT_PARAMS, 'r_h': None} intensity = likelihood.carehome_intensity( covariates=covariates, cases=cases, fit_params=fit_params_no_rh, dist_params=SIMPLE_DIST_PARAMS ) assert_almost_equal( intensity, [[1, 2, 2], [1.736, 2.184, 3.104], [2.277, 2.135, 3.364]], decimal=3 ) def test_carehome_intensity_with_discharges(small_cases, small_covariates): '''Test that the behaviour of carehome_intensity is correct in the case where discharges are considered.''' _, cases = small_cases _, covariates = small_covariates discharges = cases[::-1] intensity = likelihood.carehome_intensity( covariates=covariates, cases=cases, fit_params=SMALL_FIT_PARAMS, dist_params=SIMPLE_DIST_PARAMS, discharges=discharges ) assert_almost_equal( intensity, [[1, 2, 2], [2.077, 5.937, 3.217], [3.332, 11.240, 3.810]], decimal=3 ) @pytest.mark.parametrize("mean, cv, expected_shape, expected_scale", [(1, 1, 1, 1), (6.5, 0.62, 2.601, 2.499)]) def test_calculate_gamma_parameters(mean, cv, expected_shape, expected_scale): '''Test that calculation of Scipy-style gamma parameters from "descriptive" gamma parameters is correct.''' shape, scale = likelihood.calculate_gamma_parameters(mean, cv) assert_almost_equal([shape, scale], [expected_shape, expected_scale], decimal=3) def test_likelihood(): '''Test that the likelihood calculation is correct''' cases = np.asarray([[3, 1, 0, 1], [1, 0, 2, 1], [0, 0, 0, 1]]) intensity = np.asarray( [[1, 3, 1.5, 6], [4.2, 3.1, 7, 1.4], [2, 5.1, 4.2, 8.9]] ) result = likelihood.likelihood(intensity, cases) assert_almost_equal(result, -39.145, decimal=3) def test_calculate_likelihood_from_files_no_discharges(): '''Test that likelihood is correctly calculated from input files when discharges are not considered.''' fit_params_no_rh = {**SMALL_FIT_PARAMS, 'r_h': None} result = likelihood.calculate_likelihood_from_files( SMALL_CASES_FILE, SMALL_COVARIATES_FILE, fit_params=fit_params_no_rh, dist_params=SIMPLE_DIST_PARAMS ) assert_almost_equal(result, -187.443, decimal=3) def test_calculate_likelihood_from_files_no_cases(): '''Test that likelihood is correctly calculated from input files when cases are not considered.''' fit_params_no_rh = {**SMALL_FIT_PARAMS, 'r_c': 0} result = likelihood.calculate_likelihood_from_files( SMALL_CASES_FILE, SMALL_COVARIATES_FILE, discharges_file=SMALL_CASES_FILE, fit_params=fit_params_no_rh, dist_params=SIMPLE_DIST_PARAMS ) assert_almost_equal(result, -189.046, decimal=3) def test_calculate_likelihood_from_files_no_discharges_or_cases(): '''Test that likelihood is correctly calculated from input files when neither cases nor discharges are considered.''' fit_params_no_rh = {**SMALL_FIT_PARAMS, 'r_h': None, 'r_c': 0} result = likelihood.calculate_likelihood_from_files( SMALL_CASES_FILE, SMALL_COVARIATES_FILE, fit_params=fit_params_no_rh, dist_params=SIMPLE_DIST_PARAMS ) assert_almost_equal(result, -196.466, decimal=3) def test_calculate_likelihood_from_files_with_discharges(): '''Test that likelihood is correctly calculated from input files when discharges are considered.''' result = likelihood.calculate_likelihood_from_files( SMALL_CASES_FILE, SMALL_COVARIATES_FILE, discharges_file=SMALL_CASES_FILE, fit_params=SMALL_FIT_PARAMS, dist_params=SIMPLE_DIST_PARAMS ) assert_almost_equal(result, -182.761, decimal=3) def test_calculate_likelihood_from_files_missing_discharges(): '''Test that an error is generated when r_h is provided but discharge data are not''' with pytest.raises(AssertionError): likelihood.calculate_likelihood_from_files( SMALL_CASES_FILE, SMALL_COVARIATES_FILE, fit_params=SMALL_FIT_PARAMS, dist_params=SIMPLE_DIST_PARAMS ) @pytest.mark.parametrize( 'r_c, r_h, expect', [(0, 0, 196.466), (0.5, 1.5, 182.761), (0.5, 0, 187.443), (0, 1.5, 189.046)] ) def test_fittable_likelihood(r_c, r_h, expect): '''Test that the closure to give a version of intensity and likelihood that can be fitted by scipy works correctly.''' fittable_likelihood = likelihood.get_fittable_likelihood( SMALL_CASES_FILE, SMALL_COVARIATES_FILE, SMALL_CASES_FILE ) fit_params = np.asarray( [r_c, r_h, *SMALL_FIT_PARAMS['baseline_intensities']] ) assert_almost_equal( fittable_likelihood( fit_params, *map( SIMPLE_DIST_PARAMS.get, (('self_excitation_shape', 'self_excitation_scale', 'discharge_excitation_shape', 'discharge_excitation_scale')) ) ), expect, decimal=3 ) @pytest.fixture def large_test_data(): '''Generate test data of the size expected from SAIL.''' max_categories = 4 num_care_homes = 1000 num_cases = 2000 num_case_homes = 330 num_discharges = 3000 num_discharge_homes = 500 num_days = 181 num_covariates = 1 max_carehome_id = 32767 cases = np.zeros((num_days, num_care_homes), dtype=np.int8) discharges = np.zeros((num_days, num_care_homes), dtype=np.int8) covariates = np.zeros((num_covariates, num_care_homes), dtype=np.int8) # For runs with the same version of numpy, we should get the same # test data each time. Not guaranteed to work between versions # because default_rng can change. rng = np.random.default_rng(seed=0) care_home_ids = rng.choice( max_carehome_id, size=num_care_homes, replace=False ) for sample_array, num_instances, num_places in ( (cases, num_cases, num_case_homes), (discharges, num_discharges, num_discharge_homes) ): for _ in range(num_instances): sample_array[rng.integers(num_days), rng.integers(num_places)] += 1 covariates[0] = rng.choice(max_categories, size=num_care_homes) for array in care_home_ids, cases, covariates, discharges: array.flags.writeable = False return care_home_ids, cases, covariates, discharges def test_intensity_performance_base(large_test_data, benchmark): ''' Test the performance of the intensity function for the base case ''' _, cases, covariates, _ = large_test_data kwargs = { 'fit_params': {**LARGE_FIT_PARAMS, 'r_h': None, 'r_c': None}, 'covariates': covariates, 'cases': cases } # Ensure that numba can jit the function before timing it likelihood.carehome_intensity_null(**kwargs) benchmark(likelihood.carehome_intensity_null, **kwargs) @pytest.mark.parametrize("use_cache", [True, False]) def test_intensity_performance_self(large_test_data, benchmark, use_cache): ''' Test the performance of the intensity function with self-excitation ''' _, cases, covariates, _ = large_test_data if not use_cache: # Writeable arrays are not cached cases.flags.writeable = True covariates.flags.writeable = True kwargs = { 'fit_params': {**LARGE_FIT_PARAMS, 'r_h': None}, 'covariates': covariates, 'cases': cases, 'dist_params': FULL_DIST_PARAMS } # Ensure that numba can jit the function before timing it likelihood.carehome_intensity(**kwargs) benchmark(likelihood.carehome_intensity, **kwargs) @pytest.mark.parametrize("use_cache", [True, False]) def test_intensity_performance_hospitals( large_test_data, benchmark, use_cache ): ''' Test the performance of the intensity function with self- and discharge excitations.''' _, cases, covariates, discharges = large_test_data if not use_cache: # Writeable arrays are not cached cases.flags.writeable = True covariates.flags.writeable = True discharges.flags.writeable = True kwargs = { 'fit_params': LARGE_FIT_PARAMS, 'covariates': covariates, 'cases': cases, 'discharges': discharges, 'dist_params': FULL_DIST_PARAMS } # Ensure that numba can jit the function before timing it likelihood.carehome_intensity(**kwargs) benchmark(likelihood.carehome_intensity, **kwargs) def test_likelihood_performance(large_test_data, benchmark): ''' Test the performance of the calculation of likelihood from the intensity and case distribution.''' _, cases, covariates, discharges = large_test_data intensity = likelihood.carehome_intensity( fit_params=LARGE_FIT_PARAMS, covariates=covariates, cases=cases, discharges=discharges, dist_params=FULL_DIST_PARAMS ) benchmark(likelihood.likelihood, intensity, cases)
13,449
5,009
from django.db import models class Url(models.Model): long_url = models.CharField(max_length=200) short_url = models.CharField(max_length=100) visit_time = models.DateTimeField(auto_now_add=True) def __unicode__(self): return self.long_url
266
91
import unittest class Queue: """Simple Queue implementation - First in First Out""" def __init__(self): self.__data = [] def enqueue(self, text): """Add new element to queue Arguments: text {string} -- An element which needs to be added to an end of a queue """ self.__data.append(text) def dequeue(self): """Gets a first element in a front of a queue Returns: string -- A first element in a front of a queue """ if len(self.__data) == 0: return None taken = self.__data[0] new_queue = [] for index in range(1, len(self.__data)): new_queue.append(self.__data[index]) self.__data = new_queue return taken def front(self): """Checks a first element in a front of a queue Returns: string -- A first element in a front of a queue """ if len(self.__data) == 0: return None return self.__data[0] def rear(self): """Checks a last element in a queue Returns: string -- A last element in a queue """ if len(self.__data) == 0: return None return self.__data[-1] class QueueTest(unittest.TestCase): def test_empty_queue(self): queue = Queue() self.assertIsNone(queue.front()) self.assertIsNone(queue.rear()) self.assertIsNone(queue.dequeue()) def test_add_one(self): queue = Queue() queue.enqueue("one") self.assertEqual(queue.front(), "one", "Should be 'one'") self.assertEqual(queue.rear(), "one", "Should be 'one'") def test_add_three(self): queue = Queue() queue.enqueue("one") queue.enqueue("two") queue.enqueue("three") self.assertEqual(queue.front(), "one", "Should be 'one'") self.assertEqual(queue.rear(), "three", "Should be 'three'") def test_add_three_get_one(self): queue = Queue() queue.enqueue("one") queue.enqueue("two") queue.enqueue("three") taken = queue.dequeue() self.assertEqual(queue.front(), "two", "Should be 'two'") self.assertEqual(queue.rear(), "three", "Should be 'three'") self.assertEqual(taken, "one", "Should be 'one'") if __name__ == "__main__": unittest.main()
2,411
724
import logging from iemit_plugin import IEmitter from plugins.emitters.base_http_emitter import BaseHttpEmitter logger = logging.getLogger('crawlutils') class HttpEmitter(BaseHttpEmitter, IEmitter): def get_emitter_protocol(self): return 'http'
262
86
import torch import torch.nn as nn import numpy as np import torch.nn.functional as F import pywt class DWTForward(nn.Module): def __init__(self, wave_name="haar"): super(DWTForward, self).__init__() wavelet = pywt.Wavelet(wave_name) ll = np.outer(wavelet.dec_lo, wavelet.dec_lo) lh = np.outer(wavelet.dec_hi, wavelet.dec_lo) hl = np.outer(wavelet.dec_lo, wavelet.dec_hi) hh = np.outer(wavelet.dec_hi, wavelet.dec_hi) filts = np.stack([ll[None, :, :], lh[None, :, :], hl[None, :, :], hh[None, :, :]], axis=0) self.weight = nn.Parameter( torch.tensor(filts).to(torch.get_default_dtype()), requires_grad=False) def forward(self, x): channel = x.shape[1] filters = torch.cat([self.weight, ] * channel, dim=0) # in tf2 self.strides = [1, 1, 2, 2, 1] # x = tf.nn.conv3d(x, self.filter, padding='VALID', strides=self.strides) y = F.conv2d(x, filters, groups=channel, stride=2) return y class DWTInverse(nn.Module): def __init__(self, wave_name="haar"): super(DWTInverse, self).__init__() wavelet = pywt.Wavelet(wave_name) ll = np.outer(wavelet.dec_lo, wavelet.dec_lo) lh = np.outer(wavelet.dec_hi, wavelet.dec_lo) hl = np.outer(wavelet.dec_lo, wavelet.dec_hi) hh = np.outer(wavelet.dec_hi, wavelet.dec_hi) filts = np.stack([ll[None, :, :], lh[None, :, :], hl[None, :, :], hh[None, :, :]], axis=0) self.weight = nn.Parameter( torch.tensor(filts).to(torch.get_default_dtype()), requires_grad=False) def forward(self, x): channel = int(x.shape[1] / 4) filters = torch.cat([self.weight, ] * channel, dim=0) y = F.conv_transpose2d(x, filters, groups=channel, stride=2) return y
1,946
730
import sys import numpy as np import json import matplotlib.pyplot as plt from matplotlib.dates import DateFormatter from matplotlib.backends.backend_pdf import PdfPages from datetime import datetime, timedelta import matplotlib.dates as mdates import emoji class FacebookMessengerConversation(): """Module for getting stats of a Facebook Messenger conversation. Attributes: data (dict): The conversation of interest. title (str) : Title of the conversation. p (list): List of conversation participants. """ def __init__(self, conversation): """Prepares `conversation` and fetches its participants. Args: conversation (json): Conversation downloaded from Facebook (see https://www.facebook.com/help/ 212802592074644?helpref=uf_permalink) """ self.data = json.load(open(conversation)) self.title = self.data['title'] # Convert unicode characters for p in self.data['participants']: p['name'] = p['name'].encode('raw_unicode_escape').decode('utf-8') for message in self.data['messages']: message['sender_name'] = message['sender_name'].encode( 'raw_unicode_escape').decode('utf-8') if 'content' in message: message['content'] = message['content'].encode( 'raw_unicode_escape').decode('utf-8') # Set names of conversation participants nbr_participants = len(self.data['participants']) self.p = nbr_participants * [None] for i in range(nbr_participants): self.p[i] = self.data['participants'][i]['name'] def get_participants(self): """Returns the names of the conversation participants. Returns: list: Contains the conversation participants """ return self.p def get_time_interval(self, type): """Returns the start and end time of the conversation. Args: type (str): Decides what type should be returned. Either 'datetime' or 'str'. Returns: tuple: (start, end). Either as datetimes or strings. Raises: ValueError: If a not supported `type` was entered. """ start = datetime.fromtimestamp( self.data['messages'][-1]['timestamp_ms']/1000) end = datetime.fromtimestamp( self.data['messages'][0]['timestamp_ms']/1000) if type == 'datetime': return start, end elif type == 'str': return start.strftime('%Y-%m-%d %H:%M:%S'), \ end.strftime('%Y-%m-%d %H:%M:%S') else: raise ValueError('Type not supported. Must be '\ 'either datetime or str.') def get_nbr_days(self): """Returns the number days between the first and last message. Returns: int: Days between start and end. """ start, end = self.get_time_interval('datetime') return (end - start).days + 1 def get_nbr_msg(self): """Returns the total number of messages. Returns: int: Number of messages. """ return len(self.data['messages']) def get_nbr_words(self): """Returns the total number of words. Returns: int: Number of words. """ nbr_words = 0 for message in self.data['messages']: if 'content' in message: nbr_words += len(message['content'].split()) return nbr_words def get_avg_len_msg(self): """Returns the average length of a message. Returns: float: Average length of message. """ return round(self.get_nbr_words()/self.get_nbr_msg(), 1) def get_avg_msg_day(self): """Returns the average number of messages sent each day. Returns: float: Average number of messages sent per day. """ return round(self.get_nbr_msg()/self.get_nbr_days(), 1) def activity(self): """Activity by each conversation participant. Returns: dict: Contains a list (value) with the number of messages sent and the percentage it corresponds to per participant (key). """ nbr_msg = self.get_nbr_msg() act = {p: 0 for p in self.p} for message in self.data['messages']: try: act[message['sender_name']] += 1 except KeyError: pass for key in act: nbr_msg_p = act[key] act[key] = [nbr_msg_p, 100*round(nbr_msg_p/nbr_msg, 2)] return act def timeline(self): """Fetches data when messages are sent. Returns: tuple: Containing which days messages were sent and also how many were sent per day, weekday and hour. """ nbr_days = self.get_nbr_days() timeline = [None] * nbr_days hour = list(range(24)) weekday_arr = [0, 1, 2, 3, 4, 5, 6] nbr_times_hour = [0] * 24 nbr_times_weekday = [0] * 7 nbr_times_day = [0] * nbr_days _, end = self.get_time_interval('datetime') current_day = end.date() index = len(timeline) - 1 timeline[index] = current_day nbr_times_day[index] = 1 for message in self.data['messages']: current = datetime.fromtimestamp( message['timestamp_ms']/1000) h = int(round(current.hour + current.minute/60. +\ current.second/3600)) if h == 24: h = 0 nbr_times_hour[h] = nbr_times_hour[h] + 1 wd = current.weekday() nbr_times_weekday[wd] = nbr_times_weekday[wd] + 1 current = current.date() if current == current_day: nbr_times_day[index] = nbr_times_day[index] + 1 elif current < current_day: diff = (current_day - current).days index = index - diff current_day = current timeline[index] = current_day nbr_times_day[index] = 1 dates = [None] * len(timeline) for i in range(0, len(timeline)): if timeline[i] == None: timeline[i] = timeline[i - 1] + timedelta(days=1) dates[i] = timeline[i].strftime('%Y-%m-%d') return timeline, nbr_times_day, nbr_times_weekday, nbr_times_hour def top_emojis(self, nbr): """Returns the top `nbr` emojis used and who sent them. Args: nbr (int): The number of emojis to include in top list. Returns: tuple: List of top emojis and dict showing how many of these were sent by each participant. """ emojis = {e: 0 for e in iter(emoji.UNICODE_EMOJI['en'].values())} emojis_p = {p: 0 for p in self.p} for p in emojis_p: emojis_p[p] = {e: 0 for e in iter(emoji.UNICODE_EMOJI['en'].values())} for message in self.data['messages']: if 'content' in message: msg = message['content'] sender = message['sender_name'] for c in msg: emoji_str = emoji.demojize(c) if emoji_str in emojis and sender in emojis_p: emojis_p[sender][emoji_str] += 1 emojis[emoji_str] += 1 top_emojis = [emoji_key for emoji_key, count in sorted(emojis.items(), key=lambda kv: (-kv[1], kv[0]))[:nbr]] emojis_count_p = {p: {} for p in self.p} for p in self.p: emojis_count_p[p] = [emojis_p[p][e] for e in top_emojis] top_emojis = [emoji.emojize(top_emoji) for top_emoji in top_emojis] return top_emojis, emojis_count_p
7,973
2,411
""" Setup file """ import os import sys from setuptools import setup, find_packages HERE = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(HERE, 'README.rst')).read() CHANGES = open(os.path.join(HERE, 'CHANGES.rst')).read() REQUIREMENTS = [ 'pyramid', 'six', ] TEST_REQUIREMENTS = [ 'mock', ] if sys.version_info[:2] < (2, 7): TEST_REQUIREMENTS.extend(['unittest2']) if __name__ == "__main__": setup( name='pyramid_duh', version='0.1.2', description='Useful utilities for every pyramid app', long_description=README + '\n\n' + CHANGES, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Pyramid', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Topic :: Internet :: WWW/HTTP', 'Topic :: Utilities', ], author='Steven Arcangeli', author_email='arcangeli07@gmail.com', url='http://pyramid-duh.readthedocs.org/', keywords='pyramid util utility', license='MIT', zip_safe=False, include_package_data=True, packages=find_packages(exclude=('tests',)), install_requires=REQUIREMENTS, tests_require=REQUIREMENTS + TEST_REQUIREMENTS, test_suite='tests', )
1,758
572
import requests, os from dotenv import load_dotenv load_dotenv() API_TOKEN = os.getenv('API_TOKEN') def sendMessage(user_id: str, text: str, max_retries: int = 1): url = f'https://api.telegram.org/bot{API_TOKEN}/sendMessage' payload = { "chat_id": user_id, "text": text } for i in range(max_retries): r = requests.get(url, params=payload) isOk = False try: isOk = r.json()["ok"] == True except: pass if isOk == True: return isOk return isOk
555
192
from unittest import TestCase from mock import patch, MagicMock from mangrove.datastore.database import DatabaseManager from datawinners.questionnaire.library import QuestionnaireLibrary class TestQuestionnaireTemplate(TestCase): def test_get_category_to_doc_mappings(self): with patch('datawinners.questionnaire.library.get_db_manager') as get_db_manager: mock_dbm = MagicMock(spec=DatabaseManager) get_db_manager.return_value = mock_dbm mock_dbm.load_all_rows_in_view.return_value = [ {'key': 'Health', 'value': {'name': 'one', 'id': 'health1'}}, {'key': 'Health', 'value': {'name': 'two', 'id': 'health2'}}, {'key': 'Agriculture', 'value': {'name': 'three', 'id': 'agri1'}} ] library = QuestionnaireLibrary() result = library.get_template_groupings('en') expected = [ {'category': 'Agriculture', 'templates': [{'id': 'agri1', 'name': 'three'}]}, {'category': 'Health', 'templates': [{'id': 'health1', 'name': 'one'}, {'id': 'health2', 'name': 'two'}]}] self.assertDictEqual(expected[0], result[0]) self.assertDictEqual(expected[1], result[1]) mock_dbm.load_all_rows_in_view.assert_called_with('by_template_category_en') def test_template_details_with_french_loaded_when_language_is_french(self): with patch('datawinners.questionnaire.library.get_db_manager') as get_db_manager: mock_dbm = MagicMock(spec=DatabaseManager) get_db_manager.return_value = mock_dbm mock_dbm.load_all_rows_in_view.return_value = [] library = QuestionnaireLibrary() library.get_template_groupings('fr') mock_dbm.load_all_rows_in_view.assert_called_with('by_template_category_fr')
1,880
572
import chainer from chainer import functions from chainer import links import chainer_chemistry from chainer_chemistry.links.connection.graph_linear import GraphLinear from chainer_chemistry.utils import is_sparse class GGNNUpdate(chainer.Chain): """GGNN submodule for update part. Args: in_channels (int or None): input dim of feature vector for each node hidden_channels (int): dimension of feature vector for each node out_channels (int or None): output dime of feature vector for each node When `None`, `hidden_channels` is used. n_edge_types (int): number of types of edge """ def __init__(self, in_channels=None, hidden_channels=16, out_channels=None, n_edge_types=4, **kwargs): if out_channels is None: out_channels = hidden_channels super(GGNNUpdate, self).__init__() if in_channels is None: gru_in_channels = None else: gru_in_channels = in_channels + hidden_channels with self.init_scope(): self.graph_linear = GraphLinear( in_channels, n_edge_types * hidden_channels) self.update_layer = links.GRU(gru_in_channels, out_channels) self.n_edge_types = n_edge_types self.in_channels = in_channels self.hidden_channels = hidden_channels self.out_channels = out_channels def __call__(self, h, adj, **kwargs): hidden_ch = self.hidden_channels # --- Message part --- mb, atom, in_ch = h.shape m = functions.reshape(self.graph_linear(h), (mb, atom, hidden_ch, self.n_edge_types)) # m: (minibatch, atom, ch, edge_type) # Transpose m = functions.transpose(m, (0, 3, 1, 2)) # m: (minibatch, edge_type, atom, ch) # (minibatch * edge_type, atom, out_ch) m = functions.reshape(m, (mb * self.n_edge_types, atom, hidden_ch)) if is_sparse(adj): m = functions.sparse_matmul(adj, m) else: adj = functions.reshape(adj, (mb * self.n_edge_types, atom, atom)) m = chainer_chemistry.functions.matmul(adj, m) # (minibatch * edge_type, atom, out_ch) m = functions.reshape(m, (mb, self.n_edge_types, atom, hidden_ch)) m = functions.sum(m, axis=1) # (minibatch, atom, out_ch) # --- Update part --- # Contraction h = functions.reshape(h, (mb * atom, in_ch)) # Contraction m = functions.reshape(m, (mb * atom, hidden_ch)) out_h = self.update_layer(functions.concat((h, m), axis=1)) # Expansion out_h = functions.reshape(out_h, (mb, atom, self.out_channels)) return out_h def reset_state(self): self.update_layer.reset_state()
2,826
906
def func(a, b): return a + b def func2(a): print(a) print("Hello")
78
36
from numpy.testing import assert_array_almost_equal as array_assert from badboids.boids import SimulationParameters def test_simulation_parameters_init(): """Tests Simulation Parameters constructor""" # Arrange formation_flying_distance = 800 formation_flying_strength = 0.10 alert_distance = 8 move_to_middle_strength = 0.2 delta_t = 1.5 # Act sut = SimulationParameters(formation_flying_distance, formation_flying_strength, alert_distance, move_to_middle_strength, delta_t) # Assert array_assert(sut.formation_flying_distance, formation_flying_distance) array_assert(sut.formation_flying_strength, formation_flying_strength) array_assert(sut.alert_distance, alert_distance) array_assert(sut.move_to_middle_strength, move_to_middle_strength) array_assert(sut.delta_t, delta_t) def test_get_defaults(): """Tests Simulation Parameters get defaults method""" # Arrange expected_formation_flying_distance = 10000 expected_formation_flying_strength = 0.125 expected_alert_distance = 100 expected_move_to_middle_strength = 0.01 expected_delta_t = 1.0 # Act parameters = SimulationParameters.get_defaults() # Assert assert parameters.formation_flying_distance == expected_formation_flying_distance assert parameters.formation_flying_strength == expected_formation_flying_strength assert parameters.alert_distance == expected_alert_distance assert parameters.move_to_middle_strength == expected_move_to_middle_strength assert parameters.delta_t == expected_delta_t
1,621
511
#-*- coding: utf-8 -*- from django_town.core.settings import OAUTH2_SETTINGS try: if not OAUTH2_SETTINGS.ACCESS_TOKEN_SECRET_KEY: raise ImportError except KeyError: # import traceback # traceback.print_exc() raise ImportError from django.db import models from django.conf import settings from django.contrib import admin from django_town.cache.model import CachingModel from django_town.core.fields import JSONField from django_town.utils import generate_random_from_vschar_set class Service(models.Model): name = models.CharField(max_length=200) def __unicode__(self): return self.name # class ServiceSecretKey(CachingModel): # cache_key_format = "_ut_o2ss:%(service__pk)d" # # service = models.ForeignKey(Service, unique=True) # secret_key = models.CharField(max_length=OAUTH2_SETTINGS.SERVICE_SECRET_KEY_LENGTH, # default=lambda: generate_random_from_vschar_set( # OAUTH2_SETTINGS.SERVICE_SECRET_KEY_LENGTH)) def _generate_random_from_vschar_set_for_client_id(): return generate_random_from_vschar_set(OAUTH2_SETTINGS.CLIENT_ID_LENGTH) def _generate_random_from_vschar_set_for_client_secret(): return generate_random_from_vschar_set(OAUTH2_SETTINGS.CLIENT_ID_LENGTH) class Client(CachingModel): IOS_CLIENT = 1 CLIENT_TYPE = ( (0, "Web"), (1, "iOS"), (2, "Android"), (3, "Win"), ) cache_key_format = "_ut_o2c:%(client_id)s" name = models.CharField(max_length=200) service = models.ForeignKey(Service) client_id = models.CharField(max_length=OAUTH2_SETTINGS.CLIENT_ID_LENGTH, unique=True, default=_generate_random_from_vschar_set_for_client_id) client_secret = models.CharField(max_length=OAUTH2_SETTINGS.CLIENT_SECRET_LENGTH, default=_generate_random_from_vschar_set_for_client_secret) redirect_uris = JSONField(blank=True) default_redirect_uri = models.URLField() available_scope = JSONField(blank=True) client_type = models.IntegerField(default=IOS_CLIENT, choices=CLIENT_TYPE) client_min_version = models.CharField(max_length=20, default="") client_cur_version = models.CharField(max_length=20, default="") client_store_id = models.CharField(max_length=30, default="") def __unicode__(self): return self.name def _generate_random_from_vschar_set_for_secret_key(): return generate_random_from_vschar_set(OAUTH2_SETTINGS.USER_SECRET_KEY_LENGTH) class UserClientSecretKey(CachingModel): cache_key_format = "_ut_o2u:%(user_id)d:%(client__pk)d" user_id = models.IntegerField() client = models.ForeignKey(Client) secret_key = models.CharField(max_length=OAUTH2_SETTINGS.USER_SECRET_KEY_LENGTH, default=_generate_random_from_vschar_set_for_secret_key) unique_together = (("user_id", "client"),) class Scope(models.Model): name = models.CharField(max_length=30, unique=True) class ClientAdmin(admin.ModelAdmin): readonly_fields = ['client_id', 'client_secret'] admin.site.register(Client, admin.ModelAdmin) admin.site.register(Service, admin.ModelAdmin)
3,247
1,117
# List files in a directory. Useful for testing the path from local_tools import * from image_renamer import ImageRenamer if confirm_config('path'): img_path = get_config('path') fl = ImageRenamer(img_path) for ff in fl.image_files(): print(ff)
265
83
import os def deepmedic_config(config_files_path, niftis_path, test_flair_file_name, test_t1c_file_name, mask, prediction_file_name, output_path): with open(os.path.join(config_files_path, 'model', 'modelConfig.cfg'), 'r') as f: lines = f.readlines() lines[8] = 'folderForOutput = "%s"\n' % output_path with open(os.path.join(config_files_path, 'model', 'modelConfig.cfg'), 'w') as f: f.writelines(lines) with open(os.path.join(config_files_path, 'test', 'testConfig.cfg'), 'r') as f: lines = f.readlines() lines[8] = 'folderForOutput = "%s"\n' % output_path with open(os.path.join(config_files_path, 'test', 'testConfig.cfg'), 'w') as f: f.writelines(lines) with open(os.path.join(config_files_path, 'test', 'testChannels_flair.cfg'), 'w') as f: f.write(os.path.join(niftis_path, test_flair_file_name)) with open (os.path.join(config_files_path, 'test', 'testChannels_t1c.cfg'), 'w') as f: f.write(os.path.join(niftis_path, test_t1c_file_name)) with open(os.path.join(config_files_path, 'test', 'testRoiMasks.cfg'), 'w') as f: f.write(os.path.join(niftis_path, mask)) with open(os.path.join(config_files_path, 'test' 'testNamesOfPredictions.cfg'), 'w') as f: f.write(prediction_file_name)
1,394
536
import os from .default import DefaultModelConfig class ModelConfig(DefaultModelConfig): def __init__(self): super().__init__() self.MODEL_NAME = 'AOTT'
174
52
from __future__ import print_function import time import tensorflow as tf from data import DataGenerator def define_net(input_batch): return input_batch + 20 # simplest network I could think of. def main(): batch_size = 1 coord = tf.train.Coordinator() with tf.name_scope('create_inputs'): reader = DataGenerator(coord) input_batch = reader.dequeue(batch_size) sess = tf.Session(config=tf.ConfigProto(log_device_placement=False)) init = tf.global_variables_initializer() sess.run(init) threads = reader.start_threads(sess) net = define_net(input_batch) queue_size = reader.queue_size for step in range(10000): print('size queue =', queue_size.eval(session=sess)) print(sess.run(net)) # Make this thread slow. You can comment this line. If you do so, you will dequeue # faster than you enqueue, so expect the queue not to reach its maximum (32 by default) time.sleep(1) coord.request_stop() print("stop requested.") for thread in threads: thread.join() if __name__ == '__main__': main()
1,125
354
# Dimmer Switch class class DimmerSwitch(): def __init__(self, label): self.label = label self.isOn = False self.brightness = 0 def turnOn(self): self.isOn = True # turn the light on at self.brightness def turnOff(self): self.isOn = False # turn the light off def raiseLevel(self): if self.brightness < 10: self.brightness = self.brightness + 1 def lowerLevel(self): if self.brightness > 0: self.brightness = self.brightness - 1 # Extra method for debugging def show(self): print('Label:', self.label) print('Light is on?', self.isOn) print('Brightness is:', self.brightness) print() # Main code # Create first DimmerSwitch, turn on and raise level twice oDimmer1 = DimmerSwitch('Dimmer1') oDimmer1.turnOn() oDimmer1.raiseLevel() oDimmer1.raiseLevel() # Create second DimmerSwitch, turn on and raise level 3 times oDimmer2 = DimmerSwitch('Dimmer2') oDimmer2.turnOn() oDimmer2.raiseLevel() oDimmer2.raiseLevel() oDimmer2.raiseLevel() # Create third DimmerSwitch, using the default settings oDimmer3 = DimmerSwitch('Dimmer3') # Ask each switch to show itself oDimmer1.show() oDimmer2.show() oDimmer3.show()
1,285
430
#!/usr/bin/env python3 import logging import argparse from pathlib import Path from vosk import list_models, list_languages from vosk.transcriber.transcriber import Transcriber parser = argparse.ArgumentParser( description = 'Transcribe audio file and save result in selected format') parser.add_argument( '--model', '-m', type=str, help='model path') parser.add_argument( '--list-models', default=False, action='store_true', help='list available models') parser.add_argument( '--list-languages', default=False, action='store_true', help='list available languages') parser.add_argument( '--model-name', '-n', type=str, help='select model by name') parser.add_argument( '--lang', '-l', default='en-us', type=str, help='select model by language') parser.add_argument( '--input', '-i', type=str, help='audiofile') parser.add_argument( '--output', '-o', default='', type=str, help='optional output filename path') parser.add_argument( '--output-type', '-t', default='txt', type=str, help='optional arg output data type') parser.add_argument( '--log-level', default='INFO', help='logging level') def main(): args = parser.parse_args() log_level = args.log_level.upper() logging.getLogger().setLevel(log_level) if args.list_models == True: list_models() return if args.list_languages == True: list_languages() return if not args.input: logging.info('Please specify input file or directory') exit(1) if not Path(args.input).exists(): logging.info("File/folder '%s' does not exist, please specify an existing file/directory" % (args.input)) exit(1) transcriber = Transcriber(args) if Path(args.input).is_dir(): transcriber.process_dir(args) return elif Path(args.input).is_file(): transcriber.process_file(args) else: logging.info('Wrong arguments') exit(1) if __name__ == "__main__": main()
2,101
632
import numpy as np import cv2 from imutils.object_detection import non_max_suppression import pytesseract from matplotlib import pyplot as plt def ocr(images): results = [] for image in images: args = {"image": image, "east": "frozen_east_text_detection.pb", "min_confidence": 0.5, "width": 320, "height": 320} args['image'] = image image = cv2.imread(args['image']) orig = image.copy() (origH, origW) = image.shape[:2] (newW, newH) = (args["width"], args["height"]) rW = origW / float(newW) rH = origH / float(newH) image = cv2.resize(image, (newW, newH)) (H, W) = image.shape[:2] blob = cv2.dnn.blobFromImage(image, 1.0, (W, H), (123.68, 116.78, 103.94), swapRB=True, crop=False) net = cv2.dnn.readNet(args["east"]) layerNames = [ "feature_fusion/Conv_7/Sigmoid", "feature_fusion/concat_3"] net.setInput(blob) (scores, geometry) = net.forward(layerNames) def predictions(prob_score, geo): (numR, numC) = prob_score.shape[2:4] boxes = [] confidence_val = [] for y in range(0, numR): scoresData = prob_score[0, 0, y] x0 = geo[0, 0, y] x1 = geo[0, 1, y] x2 = geo[0, 2, y] x3 = geo[0, 3, y] anglesData = geo[0, 4, y] for i in range(0, numC): if scoresData[i] < args["min_confidence"]: continue (offX, offY) = (i * 4.0, y * 4.0) angle = anglesData[i] cos = np.cos(angle) sin = np.sin(angle) h = x0[i] + x2[i] w = x1[i] + x3[i] endX = int(offX + (cos * x1[i]) + (sin * x2[i])) endY = int(offY - (sin * x1[i]) + (cos * x2[i])) startX = int(endX - w) startY = int(endY - h) boxes.append((startX, startY, endX, endY)) confidence_val.append(scoresData[i]) return (boxes, confidence_val) (boxes, confidence_val) = predictions(scores, geometry) boxes = non_max_suppression(np.array(boxes), probs=confidence_val) result = [] for (startX, startY, endX, endY) in boxes: startX = int(startX * rW) startY = int(startY * rH) endX = int(endX * rW) endY = int(endY * rH) r = orig[startY:endY, startX:endX] configuration = ("-l eng --oem 1 --psm 8") text = pytesseract.image_to_string(r, config=configuration) result.append(text) results.append(result) return results print(ocr(["./images/car_wash.png"]))
2,923
1,021
# Generated by Django 2.0 on 2018-01-28 21:44 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('bug', '0002_auto_20180110_1107'), ] operations = [ migrations.AlterField( model_name='bug', name='buglevel', field=models.CharField(choices=[('1', '1'), ('2', '2'), ('3', '3')], default='3', max_length=200, null=True, verbose_name='严重程度'), ), migrations.AlterField( model_name='bug', name='bugstatus', field=models.CharField(choices=[('激活', '激活'), ('已解决', '已解决'), ('已关闭', '已关闭')], default='激活', max_length=200, null=True, verbose_name='解决状态'), ), ]
727
291
""" Tests for validators """ from decimal import Decimal from unittest.mock import Mock import random import string from styler_validation import validators as va from styler_validation import messages as msg class MyModel: pass class TestIsRequired: def test_is_required(self): val = va.is_required() model = MyModel() model.prop = None valid, error = val(model, 'prop') assert not valid assert error == (msg.REQUIRED_VALUE,) def test_valid(self): val = va.is_required() model = MyModel() model.prop = 'something' valid, error = val(model, 'prop') assert valid assert error is None def test_accepts(self): val = va.is_required(accepts={0}) model = MyModel() model.prop = 0 valid, error = val(model, 'prop') assert valid assert error is None class TestIsInteger: def test_is_integer(self): val = va.is_integer() model = MyModel() model.prop = '123' valid, error = val(model, 'prop') assert valid assert error is None def test_is_none(self): val = va.is_integer() model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None def test_type_mismatch(self): val = va.is_integer() model = MyModel() model.prop = {'123'} valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_value_mismatch(self): val = va.is_integer() model = MyModel() model.prop = 'abc123' valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) class TestIsBetween: def test_is_between(self): val = va.is_between(min_=0, max_=10) model = MyModel() model.prop = 2 valid, error = val(model, 'prop') assert valid assert error is None def test_outside_interval(self): val = va.is_between(min_=0, max_=10) model = MyModel() model.prop = 20 model.prop2 = -1 valid, error = val(model, 'prop') valid2, error2 = val(model, 'prop2') assert not valid assert error == (msg.LESS_OR_EQUAL_THAN, 10) assert not valid2 assert error2 == (msg.GREATER_OR_EQUAL_THAN, 0) def test_no_interval_set(self): val = va.is_between() model = MyModel() model.prop = 20 model.prop2 = -1 valid, error = val(model, 'prop') valid2, error2 = val(model, 'prop2') assert valid assert error is None assert valid2 assert error2 is None def test_one_sided_interval(self): val = va.is_between(min_=0) val2 = va.is_between(max_=10) model = MyModel() model.prop = 20 model.prop2 = -1 valid, error = val(model, 'prop') valid2, error2 = val(model, 'prop2') assert valid assert error is None assert not valid2 assert error2 == (msg.GREATER_OR_EQUAL_THAN, 0) valid, error = val2(model, 'prop') valid2, error2 = val2(model, 'prop2') assert not valid assert error == (msg.LESS_OR_EQUAL_THAN, 10) assert valid2 assert error2 is None def test_none(self): val = va.is_between(min_=0, max_=10) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None def test_type_mismatch(self): val = va.is_between(min_=0, max_=10) model = MyModel() model.prop = {'123'} valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) class TestIsInside: def test_is_inside(self): accepted_values = {'a', 'b'} val = va.is_inside(accepted=accepted_values) model = MyModel() model.prop = 'b' valid, error = val(model, 'prop') assert valid assert error is None def test_not_inside(self): accepted_values = {'a', 'b'} val = va.is_inside(accepted=accepted_values) model = MyModel() model.prop = 'c' valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_none(self): accepted_values = {'a', 'b'} val = va.is_inside(accepted=accepted_values) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None class TestIsOfType: def test_is_of_type(self): val = va.is_of_type(Decimal) model = MyModel() model.prop = Decimal('12.33') valid, error = val(model, 'prop') assert valid assert error is None def test_not_type(self): val = va.is_of_type(Decimal) model = MyModel() model.prop = '12.33' valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_none(self): val = va.is_of_type(Decimal) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None class TestIsMoney: def test_is_money(self): val = va.is_money() model = MyModel() model.prop = Decimal('12.33') valid, error = val(model, 'prop') assert valid assert error is None def test_not_allow_zero(self): val = va.is_money(allow_zero=False) model = MyModel() model.prop = Decimal('0.0') valid, error = val(model, 'prop') assert not valid assert error == (msg.NOT_ZERO,) def test_negative(self): val = va.is_money() model = MyModel() model.prop = Decimal('-12.33') valid, error = val(model, 'prop') assert not valid assert error == (msg.POSITIVE_VALUE,) def test_none(self): val = va.is_money() model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None def test_type_mismatch(self): val = va.is_money() model = MyModel() model.prop = {'sdfads'} valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_value_mismatch(self): val = va.is_money() model = MyModel() model.prop = 'sdfads' valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) class TestIsValidTime: def test_is_valid_time(self): val = va.is_valid_time() model = MyModel() model.prop = '12:33' valid, error = val(model, 'prop') assert valid assert error is None def test_not_valid(self): val = va.is_valid_time() model = MyModel() model.prop = '12:73' valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_TIME,) def test_none(self): val = va.is_valid_time() model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None class TestIsGreaterThanField: def test_is_greater_than_field(self): val = va.is_greater_than_field('prop2') model = MyModel() model.prop = 333 model.prop2 = 222 valid, error = val(model, 'prop') assert valid assert error is None def test_not_valid(self): val = va.is_greater_than_field('prop2') model = MyModel() model.prop = 11 model.prop2 = 12 valid, error = val(model, 'prop') assert not valid assert error == (msg.GREATER_THAN, 'mymodel.prop2') def test_none(self): val = va.is_greater_than_field('prop2') model = MyModel() model.prop = 1 model.prop2 = None valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_default(self): val = va.is_greater_than_field('prop2', default=True) model = MyModel() model.prop = 1 model.prop2 = None valid, error = val(model, 'prop') assert valid assert error is None def test_allow_equal(self): val = va.is_greater_than_field('prop2', allow_equal=True) model = MyModel() model.prop = 1 model.prop2 = 1 valid, error = val(model, 'prop') assert valid assert error is None class TestIsLessThanField: def test_is_less_than_field(self): val = va.is_less_than_field('prop2') model = MyModel() model.prop = 111 model.prop2 = 222 valid, error = val(model, 'prop') assert valid assert error is None def test_not_valid(self): val = va.is_less_than_field('prop2') model = MyModel() model.prop = 13 model.prop2 = 12 valid, error = val(model, 'prop') assert not valid assert error == (msg.LESS_THAN, 'mymodel.prop2') def test_none(self): val = va.is_less_than_field('prop2') model = MyModel() model.prop = 1 model.prop2 = None valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_default(self): val = va.is_less_than_field('prop2', default=True) model = MyModel() model.prop = 1 model.prop2 = None valid, error = val(model, 'prop') assert valid assert error is None def test_allow_equal(self): val = va.is_less_than_field('prop2', allow_equal=True) model = MyModel() model.prop = 1 model.prop2 = 1 valid, error = val(model, 'prop') assert valid assert error is None class TestIsGreaterThanNumber: def test_is_greater_than_number(self): val = va.is_greater_than_number(10) model = MyModel() model.prop = 111 valid, error = val(model, 'prop') assert valid assert error is None def test_not_valid(self): val = va.is_greater_than_number(10) model = MyModel() model.prop = 1 valid, error = val(model, 'prop') assert not valid assert error == (msg.GREATER_THAN, 10) def test_none(self): val = va.is_greater_than_number(10) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_default(self): val = va.is_greater_than_number(10, default=True) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None def test_allow_equal(self): val = va.is_greater_than_number(10, allow_equal=True) model = MyModel() model.prop = 10 valid, error = val(model, 'prop') assert valid assert error is None class TestIsLessThanNumber: def test_is_less_than_number(self): val = va.is_less_than_number(10) model = MyModel() model.prop = 1 valid, error = val(model, 'prop') assert valid assert error is None def test_not_valid(self): val = va.is_less_than_number(10) model = MyModel() model.prop = 11 valid, error = val(model, 'prop') assert not valid assert error == (msg.LESS_THAN, 10) def test_none(self): val = va.is_less_than_number(10) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_default(self): val = va.is_less_than_number(10, default=True) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None def test_allow_equal(self): val = va.is_less_than_number(10, allow_equal=True) model = MyModel() model.prop = 10 valid, error = val(model, 'prop') assert valid assert error is None class TestIsNotEmpty: def test_is_not_empty(self): val = va.is_not_empty() model = MyModel() model.prop = 'something' valid, error = val(model, 'prop') assert valid assert error is None def test_invalid(self): val = va.is_not_empty() model = MyModel() model.prop = ' ' valid, error = val(model, 'prop') assert not valid assert error == (msg.NOT_EMPTY,) def test_none(self): val = va.is_not_empty() model = MyModel() model.prop = 0 valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_default(self): val = va.is_not_empty(default=True) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None class TestObjectValidator: def test_child_object(self): child = Mock() child.is_valid.return_value = (True, {}) model = MyModel() model.prop = child val = va.object_validator() valid, error = val(model, 'prop') assert valid assert error == {} child.is_valid.assert_called_once() def test_multiple_child_objects(self): child1 = Mock() child1.is_valid.return_value = (False, {'error1': 'error'}) child2 = Mock() child2.is_valid.return_value = (True, {}) child3 = Mock() child3.is_valid.return_value = (False, {'error3': 'error'}) model = MyModel() model.prop = [child1, child2, child3] val = va.object_validator() valid, error = val(model, 'prop') assert not valid assert error == { 'error1': 'error', 'error3': 'error' } child1.is_valid.assert_called_once() child2.is_valid.assert_called_once() child3.is_valid.assert_called_once() def test_none(self): model = MyModel() model.prop = None val = va.object_validator() valid, error = val(model, 'prop') assert valid assert error == {} class TestIsUuid: def test_is_uuid(self): val = va.is_uuid() model = MyModel() model.prop = '42fb4cf1-bd85-469c-8266-9dfcd54796a4' valid, error = val(model, 'prop') assert valid assert error is None def test_invalid(self): val = va.is_uuid() model = MyModel() model.prop = 'anything' valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) def test_none(self): val = va.is_uuid() model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None def test_type_mismatch(self): val = va.is_uuid() model = MyModel() model.prop = 1234 valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,) class TestIf_: def test_if_true(self): validation = Mock(return_value=(True, None)) val = va.if_(lambda x: True, validation) model = MyModel() model.prop = '123' valid, error = val(model, 'prop') assert valid assert error is None validation.assert_called_once() def test_if_false(self): validation = Mock(return_value=(True, None)) val = va.if_(lambda x: False, validation) model = MyModel() model.prop = '123' valid, error = val(model, 'prop') assert valid assert error is None validation.assert_not_called() class TestMaxLength: def test_valid_max_length(self): val = va.max_length(255) model = MyModel() model.prop = 'string_with_length_under_255' valid, error = val(model, 'prop') assert valid assert error is None def test_invalid_max_length(self): length = 255 val = va.max_length(length) model = MyModel() model.prop = ''.join(random.choices(string.ascii_uppercase + string.digits, k=256)) valid, error = val(model, 'prop') assert not valid assert error == (msg.STRING_TOO_LONG, length) def test_none(self): val = va.max_length(255) model = MyModel() model.prop = None valid, error = val(model, 'prop') assert valid assert error is None def test_invalid_type(self): val = va.max_length(255) model = MyModel() model.prop = 1 valid, error = val(model, 'prop') assert not valid assert error == (msg.INVALID_VALUE,)
17,439
5,626
from django.core.files.storage import FileSystemStorage from django.db import models # Create your models here. from datetime import date from django.urls import reverse #Used to generate URLs by reversing the URL patterns from django.contrib.auth.models import User #Blog author or commenter def attachment_path(instance, filename): return "media/animal/" + str(instance.animal.id) + "/attachments/" + filename """ Metoda vrací cestu k uploadovanému plakátu. """ def poster_path(instance, filename): return "animals/" + str(instance.name) +"/foto/"+ filename class Type(models.Model): name = models.CharField(max_length=50, unique=True, verbose_name="Type of animal", help_text='Enter a type of animal (e.g. Savec)') class Meta: ordering = ["name"] def __str__(self): return self.name class Animal(models.Model): name = models.CharField(max_length=50, unique=True, verbose_name="Name of animal") poster = models.ImageField(upload_to=poster_path, verbose_name="Poster") latin = models.CharField(max_length=50, unique=True, verbose_name="Latin name for animal") description = models.CharField(max_length=5000, verbose_name="Description of animal") # Metadata class Meta: ordering = ["name"] # Methods def __str__(self): """Součástí textové reprezentace filmu bude jeho název, rok uvedení a hodnocení""" return f"{self.name}, {str(self.poster)}, {str(self.latin)}" def get_absolute_url(self): """Metoda vrací URL stránky, na které se vypisují podrobné informace o filmu""" return reverse('animal-detail', args=[str(self.id)]) class Attachment(models.Model): # Fields # Povinný titulek přílohy - text do délky 200 znaků title = models.CharField(max_length=200, verbose_name="Title") # Časový údaj o poslední aktualizaci přílohy - automaticky se ukládá aktuální čas last_update = models.DateTimeField(auto_now=True) # Pole pro upload souboru # Parametr upload_to zajistí uložení souboru do složky specifikované v návratové hodnotě metody attachment_path file = models.FileField(upload_to=attachment_path, null=True, verbose_name="File") # Konstanta, v níž jsou ve formě n-tic (tuples) předdefinovány různé typy příloh TYPE_OF_ATTACHMENT = ( ('audio', 'Audio'), ('image', 'Image'), ('text', 'Text'), ('video', 'Video'), ('other', 'Other'), ) # Pole s definovanými předvolbami pro uložení typu přílohy #type = models.CharField(max_length=5, choices=TYPE_OF_ATTACHMENT, blank=True, default='image', # help_text='Select allowed attachment type', verbose_name="Attachment type") # Cizí klíč, který zajišťuje propojení přílohy s daným filmem (vztah N:1) # Parametr on_delete slouží k zajištění tzv. referenční integrity - v případě odstranění filmu # budou odstraněny i všechny jeho přílohy (models.CASCADE) film = models.ForeignKey(Animal, on_delete=models.CASCADE) # Metadata class Meta: # Primární seřazeno podle poslední aktualizace souborů, sekundárně podle typu přílohy ordering = ["-last_update"] # Methods def __str__(self): """ Textová reprezentace objektu """ return f"{self.title})"
3,305
1,093
print(list(range(10, 0, -2))) # if start > end and step > 0: # a list generated from start to no more than end with step as constant increment # if start > end and step < 0: # an empty list generated # if start < end and step > 0: # an empty list generated # if start < end and step < 0 # a list generated from start to no more than end with step as constant decrement
369
111
""" github : https://github.com/amingolnari/Deep-Learning-Course Author : Amin Golnari Keras Version : 2.2.4 Date : 4/12/2018 Keras CNN Classification on MNIST Data Code 301 """ ## If your GPU is AMD , you can use PlaidML Backend # import os # os.environ["KERAS_BACKEND"] = "plaidml.keras.backend" from keras.models import Sequential from keras.layers import Dense, Flatten, Dropout, Conv2D, MaxPool2D from keras.optimizers import SGD from keras.datasets import mnist from keras.utils.np_utils import to_categorical import matplotlib.pyplot as plt # Load MNIST Data (Download for First) def LoadData(): (Xtrain, Ytrain), (Xtest, Ytest) = mnist.load_data() Xtrain = Xtrain.reshape(60000, 28, 28, 1).astype('float32') Xtrain = Xtrain / 255 # Normalize to 0-1 Xtest = Xtest.reshape(10000, 28, 28, 1).astype('float32') Xtest = Xtest / 255 Ytrain = to_categorical(Ytrain, 10) # for exam Label 2 : [0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0] Ytest = to_categorical(Ytest, 10) return (Xtrain, Xtest), (Ytrain, Ytest) def BuildModel(): model = Sequential() model.add(Conv2D(filters = 128, kernel_size = (5, 5), activation = 'relu', padding = 'same', input_shape = (28, 28, 1))) # Just First (Input) Layer Need Init input_shape model.add(Conv2D(filters = 64, kernel_size = (3, 3), padding = 'same', activation = 'relu')) model.add(MaxPool2D(pool_size = (2, 2), padding = 'same')) model.add(Conv2D(filters = 64, kernel_size = (3, 3), activation = 'relu', padding = 'same')) model.add(MaxPool2D(pool_size = (2, 2))) model.add(Flatten()) model.add(Dropout(0.25)) model.add(Dense(300, activation = 'tanh')) # Hidden Layer #1 model.add(Dense(200, activation = 'sigmoid')) # Hidden Layer #1 model.add(Dense(10, activation = 'softmax'))# Output Layer return model def PlotHistory(history): plt.title('Keras Model loss/accuracy') plt.ylabel('loss/accuracy') plt.xlabel('epochs') # Accuracy plt.plot(history.history['acc'], '.-') plt.plot(history.history['val_acc'], '-.') # Loss plt.plot(history.history['loss'], '-*') plt.plot(history.history['val_loss'], '*-') plt.legend(['Train loss', 'Validation loss', 'Train acc', 'Validation acc'], loc='upper right') plt.grid(True, linestyle = '-.') plt.tick_params(labelcolor = 'b', labelsize = 'medium', width = 3) fig = plt.gcf() fig.savefig('images/model loss.jpg') plt.show() return def main(): (Xtrain, Xtest), (Ytrain, Ytest) = LoadData() model = BuildModel() model.summary() model.compile(loss = 'categorical_crossentropy', optimizer = SGD(lr = 0.1), metrics = ['accuracy']) History = model.fit(Xtrain, Ytrain, batch_size = 256, epochs = 100, validation_split = .3) PlotHistory(History) score, acc = model.evaluate(Xtest, Ytest) print('Test Accuracy : ', acc) if __name__ == "__main__": main()
3,141
1,185
start = 104200 end = 702648265 for arm1 in range(start, end + 1): exp = len(str(arm1)) num_sum = 0 c = arm1 while c > 0: num = c % 10 num_sum += num ** exp c //= 10 if arm1 != num_sum: continue else: if arm1 == num_sum: print("The first Armstrong number encountered is:", arm1) break #####OUTPUT##### ## The first Armstrong number encountered is: 548834 ## Process finished with exit code 0
526
214
import requests import json def Neji_version(): # import neji # print(neji.Neji_version()) # output: Hydrogen-0.0.0.1 return "Hydrogen-0.0.0.1" def Neji_upload(put_file_directory_in_string_format): # import neji # print(neji.Neji_upload(r"C:\...\Pictures\trial\black1 - Copy (5).png")) # output: https://file.io/CfJlLI r = requests.post('https://file.io', files={'file': open(str(put_file_directory_in_string_format), 'rb')}) al1 = json.loads(r.text) if al1['link'][0] == "h": return al1['link'] else: r2 = requests.post('https://api.anonymousfiles.io', files={'file': open(str(put_file_directory_in_string_format), 'rb')}) al2 = json.loads(r2.text) if al2["url"][0] == "h": return al2["url"] else: return "inform us"
787
359
#!/bin/env python3 """Handling events as tickets The goal here is, provided a maintenance event, create an event if not a duplicate. To determine if not duplicate, use some combination of values to form a key. Methods to delete, update, and otherwise transform the ticket should be available A base class, Ticket, is provided to do some boiler plate things and enforce a consistent interface. """ from textwrap import dedent from jira import JIRA class Ticket(object): """Base class for a ticket Purpose of this is to provide standard methods for retrieving duplicates, creating event, and deleting. Implementation details should be self-contained to each subclass but not really different from the interface perspective. Attributes: event (XMaintNoteEvent) acconut (str) impact (str) maintenance_id (str) object_id (str) provider (str) key (str): String that can try to be used to be unique among maintenances title (str): Generated title that may be used as a ticket title body (str): Generated body thath may be used as a ticket description ticket: Optional to add by subclass, instance of ticket in the ticket system """ def __init__(self, event, **kwargs): """Initializes and runs _post_init() Event is the only required input with any kwargs being accepted and forworded to ``self._post_init``. Purpose of the ``_post_init`` method is to facilitate each type of ticketing system to mutate the event data in however it needs without overloading ``__init__`` itself. A key is created using the provider, account, and maintenance-id keys of the event. How this is implemented by a ticketing system to take advantage of is up to the subclass. Args: event (XMaintNoteEvent): Maintenance Event """ self.event = event self.account = event['X-MAINTNOTE-ACCOUNT'] self.impact = event['X-MAINTNOTE-IMPACT'] self.maintenance_id = event['X-MAINTNOTE-MAINTENANCE-ID'] self.object_id = event['X-MAINTNOTE-OBJECT-ID'] self.provider = event['X-MAINTNOTE-PROVIDER'] self.ticket = None factors = [ self.provider, self.account, self.maintenance_id, ] self.key = '{}:{}:{}'.format(*factors) self.title = '{provider} {impact} Maintenance for {account}'.format( provider=self.provider, impact=self.impact, account=self.account, ) body = ''' {provider} is having a maintenance of {impact}. Affected account number is {account}. Start time: {start_time} End time: {end_time} Impact: {impact} Account: {account} '''.format( provider=self.provider, impact=self.impact, account=self.account, start_time=str(event['DTSTART'].dt), end_time=str(event['DTEND'].dt), ) self.body = dedent(body) self._post_init(**kwargs) def _post_init(self, **kwargs): pass def create(self): """Overload to create a ticket in the system""" raise NotImplemented('Subclass must overload this method') def close(self): """Overload to close a ticket in the system""" raise NotImplemented('Subclass must overload this method') def exists(self): """Overload to determine if this event exists in ticket form already""" raise NotImplemented('Subclass must overload this method') class JiraTicket(Ticket): """Ticket driver for JIRA Supports adding list of watchers to maintenance issues created, custom finishing transition for when calling close, and custom issue types. Priorities will be mapped according to the impact status of the maintenance. A preferred mapping can be provided otherwise it defaults to using the Vanilla JIRA install names, eg: >>> { 'NO-IMPACT': {'name': 'Low'}, 'REDUCED-REDUNDANCY': {'name': 'Medium'}, 'DEGRADED': {'name': 'High'}, 'OUTAGE': {'name': 'Highest'}, } Example: >>> type(event) xmaintnote.event.XMaintNoteEvent >>> tkt = JiraTicket( event, url='http://localhost', username='admin', password='admin', watchers='noc', ) >>> tkt.exists() False >>> tkt.create() True >>> tkt.exists() True >>> tkt.ticket <JIRA Issue: key=u'MAINT-14', id=u'10013'> >>> tkt.impact vText('NO-IMPACT') >>> tkt.ticket.fields.priority <JIRA Priority: name=u'Low', id=u'4'> >>> tkt.ticket.fields.labels [u'example.com:137.035999173:WorkOrder-31415'] """ def _post_init( self, url='http://localhost:8080', username=None, password=None, project='MAINT', issuetype='Task', finished_transition='Done', watchers=None, pri_mapping=None, ): """Setup to initialize Jira client and any required settings If username or password aren't provided, will attempt to do actions as anonymous Args: url (str): URL to jira server. MUST have the URL scheme (http://) username (str): Username (if applicable) password (str): Password (if applicable) project (str): JIRA project handle issuetype (str): Issue type to file these issues as watchers (list): List of usernames to add as watchers to the maints finished_transition (str): Transition to move the issue into when calling the ``.close`` method. Default: Done pri_mapping (str): Map of maintenance impact name to JIRA priority dict. eg, {'NO-IMPACT': {'name': 'Low'}} """ # If either part of the credential tuple is unprovided, default to # anonymous credentials = (username, password) if not all(credentials): basic_auth = None else: basic_auth = credentials if not watchers: watchers = [] if not pri_mapping: pri_mapping = { 'NO-IMPACT': {'name': 'Low'}, 'REDUCED-REDUNDANCY': {'name': 'Medium'}, 'DEGRADED': {'name': 'High'}, 'OUTAGE': {'name': 'Highest'}, } self.jira = JIRA(url, basic_auth=basic_auth) self.project = project self.issuetype = issuetype self.finished_transition = finished_transition self.watchers = watchers self.pri_mapping = pri_mapping def exists(self): """Return bool for whether maintenance issue exists for this event Improvements: Currently not handling the case where multiple issues are returned which may hint that the key used isn't unique enough or people have manually added the same label to other things. Also no exception handling mostly because the exception return by JIRA is pretty descriptive Returns: exists (bool) """ existing = self.jira.search_issues('labels = {}'.format(self.key)) if existing: self.ticket = existing[0] return True if existing else False def create(self): """Create issue for event Pre-check factors such as chehcking if this is a duplicate. If so, stop further actions. Returns: success (bool) """ jira = self.jira # If issue doesn't exist, create it. Else return False for inability # Add watchers to the new ticket if not self.exists(): options = { 'project': self.project, 'summary': self.title, 'labels': [self.key], 'description': self.body, 'issuetype': {'name': self.issuetype}, 'priority': self.pri_mapping[self.impact], } new_issue = jira.create_issue(fields=options) self.ticket = new_issue [self._add_watcher(new_issue, w) for w in self.watchers] return True else: return False def close(self): """Return bool representing success or failure for closing issue If issue doesn't exist, will return False because it can't close. Returns: success (bool) """ jira = self.jira finished_transition = self.finished_transition if self.exists(): # Fetch the transitions that we can put the current issue into. # Search through these for the provided ``finished_transition`` # from init. If not found, raise error. tkt = self.ticket transitions = jira.transitions(tkt) transition_ids = [ t['id'] for t in transitions if t['name'] == self.finished_transition ] if not transition_ids: raise ValueError( 'Transition "{}" not found'.format(finished_transition) ) t = transition_ids[0] jira.transition_issue(tkt, t) else: return False def _add_watcher(self, issue, watcher): """Add watcher to issue""" self.jira.add_watcher(issue, watcher)
9,731
2,611
def _do_set(env, name, value): if env.contains(name): env.set(name, value) elif env.parent is not None: _do_set(env.parent, name, value) else: raise Exception( "Attempted to set name '%s' but it does not exist." % name ) def set_(env, symbol_name, value): if symbol_name[0] != "string": raise Exception( "set() takes a string as its first argument, but was: %s" % str(symbol_name) ) _do_set(env, symbol_name[1], value) return value
554
177
from cicdctl.commands.types.flag import FlagParamType def test_flag(): flag = '-no-color' parsed = FlagParamType().convert(value=flag, param=None, context={}) assert flag == parsed
195
60
class PostgresQueryPart: """ Object representing Postgres query part """ def get_query(self) -> str: """ Get query Returns: str """ pass
194
51
#!/usr/bin/env python # Copyright 2010 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ Merge translations from a set of .po or XMB files into a set of .po files. Usage: ../tools/merge_messages <source-dir> <template-file> ../tools/merge_messages <source-dir> <template-file> <target-dir> ../tools/merge_messages <source-po-file> <template-file> <target-po-file> <source-dir> should be a directory containing a subdirectories named with locale codes (e.g. pt_BR). For each locale, this script looks for the first .po or .xml file it finds anywhere under <source-dir>/<locale-code>/ and adds all its messages and translations to the corresponding django.po file in the target directory, at <target-dir>/<locale-code>/LC_MESSAGES/django.po. <template-file> is the output file from running: 'find_missing_translations --format=po' With the name that corresponds to the --format=xmb output. Make sure to run this in a tree that corresponds to the version used for generating the xmb file or the resulting merge will be wrong. See validate_merge for directions on verifying the merge was correct. If <target-dir> is unspecified, it defaults to the app/locale directory of the current app. Alternatively, you can specify a single source file and a single target file to update. When merging messages from a source file into a target file: - Empty messages and messages marked "fuzzy" in the source file are ignored. - Translations in the source file will replace any existing translations for the same messages in the target file. - Other translations in the source file will be added to the target file. - If the target file doesn't exist, it will be created. - To minimize unnecessary changes from version to version, the target file has no "#: filename:line" comments and the messages are sorted by msgid. """ import babel.messages from babel.messages import pofile import codecs import os import sys import xml.sax class XmbCatalogReader(xml.sax.handler.ContentHandler): """A SAX handler that populates a babel.messages.Catalog with messages read from an XMB file.""" def __init__(self, template): """template should be a Catalog containing the untranslated messages in the same order as the corresponding messages in the XMB file.""" self.tags = [] self.catalog = babel.messages.Catalog() self.iter = iter(template) assert self.iter.next().id == '' # skip the blank metadata message def startElement(self, tag, attrs): self.tags.append(tag) if tag == 'msg': self.string = '' self.message = babel.messages.Message(self.iter.next().id) if tag == 'ph': self.string += '%(' + attrs['name'] + ')s' self.message.flags.add('python-format') def endElement(self, tag): assert self.tags.pop() == tag if tag == 'msg': self.message.string = self.string self.catalog[self.message.id] = self.message def characters(self, content): if self.tags[-1] == 'msg': self.string += content def log(text): """Prints out Unicode text.""" print text.encode('utf-8') def log_change(old_message, new_message): """Describes an update to a message.""" if not old_message: if new_message.id: log('+ msgid "%s"' % str(new_message.id)) else: print >>sys.stderr, 'no message id: %s' % new_message log('+ msgstr "%s"' % str(new_message.string.encode('ascii', 'ignore'))) if new_message.flags: log('+ #, %s' % ', '.join(sorted(new_message.flags))) else: if (new_message.string != old_message.string or new_message.flags != old_message.flags): log(' msgid "%s"' % old_message.id) log('- msgstr "%s"' % old_message.string) if old_message.flags: log('- #, %s' % ', '.join(sorted(old_message.flags))) log('+ msgstr "%s"' % new_message.string) if new_message.flags: log('+ #, %s' % ', '.join(sorted(new_message.flags))) def create_file(filename): """Opens a file for writing, creating any necessary parent directories.""" if not os.path.exists(os.path.dirname(filename)): os.makedirs(os.path.dirname(filename)) return open(filename, 'w') def merge(source, target_filename): """Merges the messages from the source Catalog into a .po file at target_filename. Creates the target file if it doesn't exist.""" if os.path.exists(target_filename): target = pofile.read_po(open(target_filename)) for message in source: if message.id and message.string and not message.fuzzy: log_change(message.id in target and target[message.id], message) # This doesn't actually replace the message! It just updates # the fields other than the string. See Catalog.__setitem__. target[message.id] = message # We have to mutate the message to update the string and flags. target[message.id].string = message.string target[message.id].flags = message.flags else: for message in source: log_change(None, message) target = source target_file = create_file(target_filename) pofile.write_po(target_file, target, no_location=True, sort_output=True, ignore_obsolete=True) target_file.close() def merge_file(source_filename, target_filename, template_filename): if source_filename.endswith('.po'): merge(pofile.read_po(open(source_filename)), target_filename) elif source_filename.endswith('.xml'): handler = XmbCatalogReader(pofile.read_po(open(template_filename))) xml.sax.parse(open(source_filename), handler) merge(handler.catalog, target_filename) if __name__ == '__main__': args = sys.argv[1:] if len(args) not in [1, 2, 3]: print __doc__ sys.exit(1) args = (args + [None, None])[:3] source_path = args[0] template_path = args[1] target_path = args[2] or os.path.join(os.environ['APP_DIR'], 'locale') # If a single file is specified, merge it. if ((source_path.endswith('.po') or source_path.endswith('.xml')) and target_path.endswith('.po')): print target_path merge_file(source_path, target_path, template_path) sys.exit(0) # Otherwise, we expect two directories. if not os.path.isdir(source_path) or not os.path.isdir(target_path): print __doc__ sys.exit(1) # Find all the source files. source_filenames = {} # {locale: po_filename} def find_po_file(key, dir, filenames): """Looks for a .po file and records it in source_filenames.""" for filename in filenames: if filename.endswith('.po') or filename.endswith('.xml'): source_filenames[key] = os.path.join(dir, filename) for locale in os.listdir(source_path): os.path.walk(os.path.join(source_path, locale), find_po_file, locale.replace('-', '_')) # Merge them into the target files. for locale in sorted(source_filenames.keys()): target = os.path.join(target_path, locale, 'LC_MESSAGES', 'django.po') print target merge_file(source_filenames[locale], target, template_path)
7,957
2,345
# derwin.py - testing a window within a window import curses def main(stdscr): # Create container window from stdscr sh, sw = stdscr.getmaxyx() container_win = curses.newwin(sh-1, sw-1, 1, 1) # Create inner window from container win bh, bw = container_win.getmaxyx() box_win = container_win.derwin(bh-2, bw-2, 1, 1) # Add size of inner win box_win.addstr(1, 1, f"{bh}x{bw}") # Draw borders container_win.box() box_win.box() # Render and wait for char container_win.refresh() container_win.getch() # main if __name__ == "__main__": curses.wrapper(main)
621
240
import sys import logging from unicorn import * from unicorn.arm_const import * from androidemu.emulator import Emulator from UnicornTraceDebugger import udbg logging.basicConfig(stream=sys.stdout, level=logging.DEBUG, format="%(asctime)s %(levelname)7s %(name)34s | %(message)s") logger = logging.getLogger(__name__) emulator = Emulator() libc = emulator.load_library('jnilibs/libc.so', do_init=False) libso = emulator.load_library('jnilibs/libnative-lib.so', do_init=False) # data segment data_base = 0xa00000 data_size = 0x10000 * 3 emulator.mu.mem_map(data_base, data_size) emulator.mu.mem_write(data_base, b'123') emulator.mu.reg_write(UC_ARM_REG_R0, data_base) try: dbg = udbg.UnicornDebugger(emulator.mu) addr_start = 0xcbc66000 + 0x9B68 + 1 addr_end = 0xcbc66000 + 0x9C2C emulator.mu.emu_start(addr_start, addr_end) r2 = emulator.mu.reg_read(UC_ARM_REG_R2) result = emulator.mu.mem_read(r2, 16) print(result.hex()) except UcError as e: list_tracks = dbg.get_tracks() for addr in list_tracks[-100:-1]: print(hex(addr - 0xcbc66000)) print (e)
1,110
479
#!/usr/bin/python # -*- coding: utf-8 -*- """ This package implements several community detection. Originally based on community aka python-louvain library from Thomas Aynaud (https://github.com/taynaud/python-louvain) """ from .community_ext import ( partition_at_level, modularity, best_partition, generate_dendrogram, induced_graph, load_binary, estimate_gamma, estimate_mu, ilfr_mu_loglikelihood, compare_partitions, model_log_likelihood ) __author__ = """Aleksey Tikhonov (altsoph@gmail.com)""" __author__ = """Liudmila Ostroumova Prokhorenkova (ostroumova-la@yandex-team.ru)""" # Copyright (C) 2018 by # Aleksey Tikhonov (altsoph@gmail.com> # Liudmila Ostroumova Prokhorenkova (ostroumova-la@yandex-team.ru) # All rights reserved. # BSD license.
816
305
#!/usr/bin/env python import rospy from std_msgs.msg import String from nav_msgs.msg import Odometry from trajectory_msgs.msg import JointTrajectory from control_msgs.msg import JointTrajectoryControllerState def callback_odom(data): print("odom\n" + str(data)) def callback_JointTrajectory(data): print("gripper_controller/command\n" + str(data)) def callback_gripper_JointTrajectory(data): print("gripper_controller\n" + str(data)) def listener(): rospy.init_node("listener12", anonymous=True) # rospy.Subscriber("/odom", Odometry, callback_odom) # rospy.Subscriber("/arm_1/gripper_controller/command", JointTrajectory, callback_JointTrajectory) rospy.Subscriber( "/arm_1/gripper_controller/state", JointTrajectoryControllerState, callback_gripper_JointTrajectory, ) rospy.spin() if __name__ == "__main__": listener()
895
323
#Botpic:https://upload.wikimedia.org/wikipedia/commons/thumb/b/b8/Red_Rose_Photography.jpg/800px-Red_Rose_Photography.jpg #Botpic:https://commons.wikimedia.org/wiki/File:Red_Rose_Photography.jpg #reference:https://www.youtube.com/watch?v=SPTfmiYiuok import discord import os import requests import json import math, random from replit import db from keep_alive import keep_alive import asyncpraw, asyncprawcore #import commands import time, asyncio, datetime from discord.ext import tasks from discord import Member from discord.ext.commands import has_permissions, MissingPermissions from prawcore import NotFound import ffmpeg from discord import FFmpegPCMAudio from dotenv import load_dotenv from youtube_search import YoutubeSearch load_dotenv() client = discord.Client() # To cache the every user For on_remove_reaction to be usable # Also enable members intent from https://discord.com/developers/ in bot secition intents = discord.Intents.default() intents.members = True global playing, stream global currently_playing_message def say_hello(): print(time.ctime()) #await message.channel.send("hello :-)" + str(joke)) #t1 = threading.Timer(10, say_hello) #t1.start() #---------- To keep the bot alive -------------------------- #1. keeping the bot alive ''' #------------------- adding a background task ----------------- status = cycle(['with Python','JetHub']) @bot.event async def on_ready(): change_status.start() print("Your bot is ready") @tasks.loop(seconds=10) async def change_status(): await bot.change_presence(activity=discord.Game(next(status))) #-------------------------------------------------------------- 3. Setup the Uptime Robot : create an account on uptime robot. After creating an account, go to the dashboard and click on Add new monitor (preview) select monitor type Http(s) (preview) then go to to ur project on repl.it and copy the url from the top of the console and paste it in url section of the monitor (preview) now set the monitoring interval to every 5 mins (so that it will ping the bot every 5 mins) and click on create monitor twice (preview) That’s it…Now go to ur project on repl.it and hit the Run button ''' class MySchedule: async def schedule_message(sth, author='anonymous', message='please provide a message', id=863298114949218324, seconds=0): print('received:') print(author, message, id, seconds) #await ctx.message.delete() if author == 'anonymous': #author = 'anonymous' description = 'command: .anon your_message' else: author = author + ' <scheduled_message>' description = "command: .schedule time_in_seconds your_message" time.sleep(seconds) print('sleep 10 seconds') print('author : ', author) #channel = bot.get_channel(id=ctx.channel.id) #print('sending {}'.format(message)) #retStr = str("""```css\nThis is some colored Text```""") #embed = discord.Embed(title="Random test") #embed.add_field(name="Name field can't be colored as it seems",value=retStr) #await ctx.send(embed=embed) #message = str(ctx.message.author).split('#')[0] + ' : ' + message embed = discord.Embed(title=author, colour=discord.Color.blue()) embed.add_field( name=message, value=description, ) channel = bot.get_channel(id=id) await channel.send(embed=embed) reddit = asyncpraw.Reddit( client_id="nnhGBCiBxSJysTobl6SLPQ", client_secret=os.environ['rd_client_secret'], password=os.environ['rd_pass'], user_agent="praw_test", username="Alternative-Ad-8849", ) async def sub_exists(subreddit_name): exists = True if subreddit_name.startswith(('/r/', 'r/')): subreddit_name = subreddit_name.split('r/')[-1] # -1 gets the last element in the list try: subreddit = await reddit.subreddit(subreddit_name, fetch=True) # by default Async PRAW doesn't make network requests when subreddit is called # do something with subreddit except asyncprawcore.Redirect: exists=False return(exists) # Reddit will redirect to reddit.com/search if the subreddit doesn't exist #await ctx.send(f"Subreddit {subreddit_name} doesn't exist.") def get_nude(): memes_submissions = reddit.subreddit('BustyPetite').hot() print('got memes') post_to_pick = random.randint(1, 15) print('choosen random') for i in range(0, post_to_pick): print('for loop:{}'.format(i)) submission = next(x for x in memes_submissions if not x.stickied) return (submission.url) def get_crazy(sub_reddit_name='memes'): memes_submissions = reddit.subreddit(sub_reddit_name).hot() #print('got memes') #post_to_pick = random.randint(1, 15) #print('choosen random') start = random.randint(100, 1000) end = random.randint(start, start + 100) print('start:{} end:{}'.format(start, end)) for i in range(start, end): #print('for loop:{}'.format(i)) submission = next(x for x in memes_submissions if not x.stickied) yield (submission.url) def get_memes_crazy(): memes_submissions = reddit.subreddit('memes').hot() print('got memes') #post_to_pick = random.randint(1, 50) print('choosen random') for i in range(0, 50): #post_to_pick): print('for loop:{}'.format(i)) submission = next(x for x in memes_submissions if not x.stickied) yield (submission.url) #return submission async def get_one(sub_reddit='memes'): #Working #submission = list(reddit.subreddit(sub_reddit_name).random()#.hot(limit=None)) #submissions = list(reddit.subreddit('redditdev').hot(limit=None)) '''urls=[] submissions = await list(reddit.subreddit('redditdev').hot(limit=None)) print(await submissions)''' #submissions = await reddit.subreddit("memes").hot(limit=random.randint(1,150)) #for submission in submissions: # pass subreddit = await reddit.subreddit(sub_reddit) async for submission in subreddit.random_rising( limit=random.randint(1, 150)): pass #print(submission.title) #urls.append([submission.title,submission.url]) #yield(submission.title, submission.url) #print(submission.title)''' #submissionn = random.choice(submissions) #submission = reddit.subreddit("AskReddit").random() #submissions = reddit.subreddit('redditdev').hot(limit=None)) #submission = random.choice(submissions) #print('got memes') #post_to_pick = random.randint(1, 50) #print('choosen random') '''for i in range(0, 50):#post_to_pick): print('for loop:{}'.format(i)) submission = next(x for x in memes_submissions if not x.stickied)''' #submission = await random.choice(memes_submissions) #return(submission.url) #print(submissionn.url) #print(submission.title) #return('hi') embed = discord.Embed(title=submission.title, url=submission.url, description=submission.selftext, colour=discord.Color.red()) embed.set_image(url=submission.url) #await channel.send(embed=embed) return (embed) from discord.ext import commands bot = commands.Bot(command_prefix='.', help_command=None, intents=intents) ''' class MyHelpCommand(commands.MinimalHelpCommand): async def send_pages(self): destination = self.get_destination() e = discord.Embed(colour=discord.Color.blurple(), description='') for page in self.paginator.pages: e.description += page await destination.send(embed=e) bot.help_command = MyHelpCommand()''' # My sample help command: @bot.command(name='help', brief='`.help` for help', help='Plesae enter `.help` for help') async def help(ctx, args=None): """ Check which mods are online on current guild """ help_embed = discord.Embed( title="Encouragement Bot Help!", #url="https:ioee.herokuapp.com/", description= "Type `.help <command name>` for more details about each command. e.g. `.help joke`", ) command_names_list = [x.name for x in bot.commands] # If there are no arguments, just list the commands: if not args: help_embed.add_field( name="List of supported commands:", value='value', #value="\n".join([str(i+1)+". "+x.name for i,x in enumerate(bot.commands)]), inline=False, #colour=discord.Color.blue() ) #bot.get_command(x.name).help for i in bot.commands: help_embed.add_field( name='***{}***'.format(i.name), #value='value' value='> {}\n\n\n'.format(bot.get_command(i.name).brief), inline=False, #colour=discord.Color.blue() ) #print(i.name) #print(i) #print(bot.get_command(i.name).help) '''for i,command in enumerate(bot.commands): help_embed.add_field( name = command, value = bot.get_command(command), inline=True )''' help_embed.add_field( name="Details", value= "Type `.help <command name>` for more details about each command.", inline=False) # If the argument is a command, get the help text from that command: elif args in command_names_list: help_embed.add_field(name=args, value=str(bot.get_command(args).brief) + ' \n' + str(bot.get_command(args).help)) # If someone is just trolling: else: help_embed.add_field(name="Nope.", value="Don't think I got that command, boss!") await ctx.send(embed=help_embed) # My sample help command: @bot.command(name='share_info', brief='`.share_info` for share_info', help='Plesae enter `.share_info` for mero_share_info') async def info(ctx, args=None): response = requests.get('http://ioee.herokuapp.com/meroshare/') response = response.text.strip() print(response) try: previous_messages = await ctx.channel.history(limit=1).flatten() prev_message = previous_messages[0].content print('previous_message:') print(prev_message) except: pass if (str(prev_message).strip() != response): print('not same messages:prev_message and rseponse') await ctx.send(response) else: print('same message as previous message, so not sending') pass @bot.command(name='ping', brief=" short_help:to test if bot responding ", help='long_help: e.g. .ping') async def ping(ctx, subreddit='jokes', no_of_posts=1, user='.'): #channel = bot.get_channel(id=int(channel_id)) '''for n, submission in enumerate(reddit.subreddit('memes').top('day',limit=int(no_of_posts/3))): print('Unleash for loop:{}'.format(n)) title = submission.title body = submission.selftext embed = discord.Embed( title=submission.title, url=submission.url, description=body, colour=discord.Color.green()) embed.set_image(url=submission.url) await ctx.send(embed=embed)''' await ctx.send('pong ') print('Ping-Pong is invoked: ', user, ctx) @bot.command(name='embed', help='e.g.`.embed`', brief='embedding help') async def embed(ctx): embed = discord.Embed(title="Text Formatting", url="https://realdrewdata.medium.com/", description="Here are some ways to format text", colour=discord.Color.blue()) embed.set_author( name="RealDrewData", url="https://twitter.com/RealDrewData", icon_url= "https://cdn-images-1.medium.com/fit/c/32/32/1*QVYjh50XJuOLQBeH_RZoGw.jpeg" ) #embed.set_author(name=ctx.author.display_name, url="https://twitter.com/RealDrewData", icon_url=ctx.author.avatar_url) embed.set_thumbnail(url="https://i.imgur.com/axLm3p6.jpeg") embed.add_field(name="*Italics*", value="Surround your text in asterisks (\*)", inline=False) embed.add_field(name="**Bold**", value="Surround your text in double asterisks (\*\*)", inline=False) embed.add_field(name="__Underline__", value="Surround your text in double underscores (\_\_)", inline=False) embed.add_field(name="~~Strikethrough~~", value="Surround your text in double tildes (\~\~)", inline=False) embed.add_field(name="`Code Chunks`", value="Surround your text in backticks (\`)", inline=False) embed.add_field(name="Blockquotes", value="> Start your text with a greater than symbol (\>)", inline=False) embed.add_field(name="Secrets", value="||Surround your text with double pipes (\|\|)||", inline=False) embed.set_footer(text="Learn more here: realdrewdata.medium.com") await ctx.send(embed=embed) @bot.command(name='schedule', brief='to schedule message to be sent in any group.', help='e.g. `.schedule 10 scheduled for ten seconds.') async def schedule(ctx, seconds: int = 3, *, message='Hello There'): #print(ctx.channel.id) print('Seconds: ', seconds) msg = str(message) #print(msg) await ctx.message.delete() id = ctx.channel.id author = str(ctx.message.author).split('#')[0] #print(author) #print(type(id)) sch = { '1': '', '2': '', '3': '', '4': '', '5': '', } r = random.randint(1, 5) sch[str(r)] = MySchedule() await sch[str(r)].schedule_message(author=author, message=msg, id=id, seconds=int(seconds)) #await schedule_message(author=author, message=msg, id=id, seconds=3) #print(id) #print(channel) #await channel.send('hi') '''@bot.command() async def schedule(ctx, message='Hello There', seconds = 3): #print(ctx.channel.id) m=str(message) id = ctx.message.id print('\n\n\n{}\n\n'.format(m)) author = str(ctx.message.author).split('#')[0] await ctx.message.delete() #id=ctx.channel.id channel = bot.get_channel(id=id) print(id) print(channel) #await channel.send('hi') #await schedule_message(author, m, id, seconds = seconds) #print(ctx.message) #await ctx.message.delete(ctx.message) #await channel.send('hi') #await ctx.send('pong') #print('Im invoked')''' @bot.command(name='anon', brief='to send message anonymously', help='e.g. `.anon Guess who!`') async def anon(ctx, *, message='please provide a message'): msg = str(message) #print(msg) await ctx.message.delete() id = ctx.channel.id a = {'anon': ''} a['anon'] = MySchedule() await a['anon'].schedule_message('anonymous', msg, id) print('send') print(msg, id) #await schedule_message(author='', message=msg, id=id) @bot.command(name="echo", pass_context=True, brief='ehhoes/repeat the message deleting the user\'s message', help='e.g. `.echo I am echoed`') async def echo(ctx, *, message='please provide a message'): msg = message #print(ctx.message) try: await ctx.message.delete() except: pass #id=ctx.channel.id await ctx.send(msg) @echo.error async def echo_error(ctx, error): if isinstance(error, MissingPermissions): text = "Sorry {}, you do not have permissions to do that!".format( ctx.message.author) await bot.send_message(ctx.message.channel, text) @bot.command(name='unleash', brief='unleahes the subreddit to c channel', help='e.g.To unleash r/jokes `.unleash jokes`') async def unleash(ctx, subreddit='none'): if subreddit == 'none': await ctx.send('Please enter the subreddit to be unleashed') else: print(ctx.channel.id) #if "unleash" not in db.keys():db['unleash']={} if await sub_exists(subreddit): if str(ctx.channel.id) not in db['unleash']: #i.e. channel doesn't exists in database db['unleash'][str(ctx.channel.id)] = [] #db['unleash'][str(ctx.channel.id)].append(str(subreddit)) else: #i.e. channel doesn't exists in database if str(subreddit) not in db['unleash'][str(ctx.channel.id)]: db['unleash'][str(ctx.channel.id)].append(str(subreddit)) await ctx.send('unleashing r/{} to {}'.format(subreddit, ctx.channel)) else: await ctx.send('r/{} already unleashed to {}'.format(subreddit, ctx.channel)) else: await ctx.send('Sorry! subreddit r/{} doesnot exists.'.format( subreddit, ctx.channel)) @bot.command(name='contain', brief='to contain/stop unleashed subreddit message', help='e.g. `.contain jokes`') async def contain(ctx, subreddit='none'): if subreddit == 'none': await ctx.send('Please enter the subreddit to be unleashed') else: print(ctx.channel.id) if str(ctx.channel.id) in db['unleash'] and str( subreddit) in db['unleash'][str(ctx.channel.id)]: db['unleash'][str(ctx.channel.id)].remove(str(subreddit)) await ctx.send( 'successfully contained subreddit r/{} from {}'.format( subreddit, ctx.channel)) else: await ctx.send('Subreddit r/{} not unleashed in .'.format( subreddit, ctx.channel)) #print(ctx.channel.id) #await ctx.send(ctx.channel.id) @bot.command( name='go', brief='to see memes from r/memes or nude from r/\'BustyPetite\'', help='e.g. `.go meme`, `.go meme crazy`, `.go nude`, `.go nude crazy`') async def go(ctx, what='', what2=''): print('wHat:{} what2:{}'.format(what, what2)) if what == 'nude': if what2 == 'crazy': print('1') urls = get_crazy('BustyPetite') for url in urls: print('sending nude') await ctx.send(url) else: print('11') urls = get_one('BustyPetite') await ctx.send(urls) #print(urls) else: if what == 'meme' and what2 == 'crazy': urls = get_crazy() #print(submission.url) for url in urls: await ctx.send(url) else: urls = get_one() await ctx.send(urls) #name='', brief='', help='e.g. `.`' '''@bot.command(name='', brief='', help='e.g. `.`') async def h(ctx, what='general'): #await ctx.send('pong') if str(what).lower()=='general': for command in commands: await ctx.send(command) elif str(what).lower() == 'fuse': for command in fuse_help_commands: await ctx.send(command)''' @bot.command( name='add_user', brief='to activate fuse user\'s auto attendance', help='get code from https://ioee.herokuapp.com e.g. `.add_user *code*`') async def add_user(ctx, pseudo_id): add_fuse_user(pseudo_id) await ctx.send("User {} has been activated successfully.".format(pseudo_id) ) @bot.command(name='check', brief='checks if live class has been started', help='e.g. `.check`') async def check(ctx, pseudo_id): add_fuse_user(pseudo_id) await ctx.send("Sorry mannual checking is unavailable for a while") #await test.start(True) @bot.command( name='remove_user', brief='to deactivate fuse auto attadance of specific', help='get code from https://ioee.herokuapp.com e.g. `.remove_user *code*`') async def remove_user(ctx, pseudo_id): remove_fuse_user(pseudo_id) #users = [] if "users" in db.keys(): #pseudocode = pseudo_id.split(".remove_user",1)[1].strip() remove_fuse_user(pseudo_id) await ctx.send("User {} has been removed successfully.".format(pseudo_id)) #------------------------------------------ @bot.command(name='joke', brief='to get jokes', help='e.g. `.joke`, `.joke 10`') async def joke(ctx, n=1): if n == 1: joke = get_joke() await ctx.send("\n\n\nJoke:" + str(joke)) else: jokes = list(get_jokes(n)) for joke in jokes: ctx.send("\n\n\nJoke:" + str(joke)) @bot.command(name='jokes', brief='to get jokes', help='e.g. `.jokes`, `.jokes 10`') async def jokes(ctx, n=5, subreddit='jokes'): await unleash_reddit(subreddit, str(ctx.channel.id), n) @bot.command(name='riddle', brief='to get a riddle', help='e.g. `.riddle`') async def riddle(ctx): riddle = get_riddles() await ctx.send(riddle) @bot.command(name='quote', brief='to get an inspiring quote', help='e.g. `.`') async def quote(ctx): quote = get_quote() await ctx.send(quote) @bot.command(name='inspire', brief='To get inspirational message', help='e.g. `.inspire`') async def inspire(ctx): inspiration = get_quote() await ctx.send(inspiration) @bot.command(name='puns', brief='To get puns', help='e.g. `.puns`') async def puns(ctx): puns = get_puns() await ctx.send(puns) @bot.command(name='one_liners', brief='to get one liner jokes', help='e.g. `.one_liners`') async def one_liners(ctx): #add_fuse_user() await ctx.send("Sorry mannual checking is unavailable for a while") @bot.command(name='meme', brief='to display meme from r/memes', help='e.g. `.meme`') async def meme(ctx, what='memes'): embed = await get_one(what) await ctx.send(embed=embed) #for i,j in get_one(): # await ctx.send(i) # await ctx.send(j) @bot.command(name='memes', brief='to display memes from r/memes', help='e.g. `.memes 3`') async def memes(ctx, n=5, subreddit='memes'): await unleash_reddit(subreddit, str(ctx.channel.id), n) #embed = await get_one(what) #await ctx.send(embed=embed) @bot.command(name='reddit', brief='to display subreddit from r/subreddit', help='e.g. `.reddit motivation 3`') async def memes(ctx, subreddit='motivation', n=3): await unleash_reddit(subreddit, str(ctx.channel.id), n) @bot.command(name='deactivate', brief='to deactivate the bot', help='e.g. `.deactivate`') async def deactivate(ctx): db["responding"] = False await ctx.send( "Encouragement bot is deactivated.\nPlease enter: .activate to activate." ) @bot.command(name='activate', brief='to activate bot', help='e.g. `.activate`') async def activate(ctx): db["responding"] = True #await ctx.send("YaY.. I'm turned on baby...")" await ctx.send("Encouragement bot is enabled.. sorry for being rude.") @bot.command( name='mute', brief= 'to mute sucesful/unsuccessful attendance attempts of fuse auto attend.', help='e.g. `.mute successful`, `.mute unsuccessful`') async def mute(ctx, what): if what == 'unsuccessful': db["unsuccessful_logs"] = False await ctx.send("unsuccessful attending_logs are muted.") elif what == 'successful': db["successful_logs"] = False await ctx.send( "successful attending_log are muted. to unmute please enter: .unmute successful" ) @bot.command( name='unmute', brief= 'to unmute sucesful/unsuccessful attendance attempts of fuse auto attend.', help='e.g. `.unmute successful`, `.unmute unsuccessful`') async def unmute(ctx, what): if what == 'unsuccessful': db["unsuccessful_logs"] = True await ctx.send("successful attending_logs are unmuted.") elif what == 'successful': db["successful_logs"] = True await ctx.send( "usuccessful attending_log are unmuted. to mute please enter: .mute unsuccessful" ) #print('\n\nwhat==\'\'', end=' ') #print(what=='') # db["responding"] = False # await message.channel.send("Encouragement bot is deactivated.\nPlease Enter: .activate to activate") @bot.command(name='list', brief='to list the current encouraging messages', help='e.g. `.list`') async def list(ctx, what='encouragements'): if what == 'users': users = [] if "users" in db.keys(): users = list(db["users"]) await ctx.send('Users: ' + str(users)) else: encouragements = [] if "encouragements" in db.keys(): encouragements = list(db["encouragements"]) await ctx.send(encouragements) @bot.command(name='delete', brief='To delete encouragement message', help='e.g. `.delete Every passsing second is makingyou better`') async def delete(ctx, index): encouragements = [] if "encouragements" in db.keys(): index = index.split(".del", 1)[1].strip() delete_encouragment(index) encouragements = list(db["encouragements"]) await ctx.send(encouragements) @bot.command(name='new', brief='To add new encouraging message to database', help='e.g. `.new Every passsing second is makingyou better`') async def new(ctx, msg): encouraging_message = msg.strip() update_encouragements(encouraging_message) await ctx.send("New encouraging message added.") @bot.command(name='avatar', brief='To see avatar of specific member in the group', help='e.g. `.avatar @Encouragement Bot`') async def avatar(ctx, *, avamember: discord.Member = None): userAvatarUrl = avamember.avatar_url await ctx.send(userAvatarUrl) # _______________________________________________________________________ # ---------------------------- For Music Bot : https://medium.com/pythonland/build-a-discord-bot-in-python-that-plays-music-and-send-gifs-856385e605a1 # _______________________________________________________________________ import os, youtube_dl import ffmpeg @bot.command( name='join', help='Tells the bot to join the voice channel before playing music ') async def join(ctx): if not ctx.message.author.voice: await ctx.send("{} is not connected to a voice channel".format( ctx.message.author.name)) return else: channel = ctx.message.author.voice.channel await channel.connect() @bot.command(name='leave', help='To make the bot leave the voice channel') async def leave(ctx): voice_client = ctx.message.guild.voice_client if voice_client.is_connected(): await voice_client.disconnect() else: await ctx.send("The bot is not connected to a voice channel.") class YTDLSource(discord.PCMVolumeTransformer): def __init__(self, source, *, data, volume=0.5): super().__init__(source, volume) self.data = data self.title = data.get('title') self.url = "" @classmethod async def from_url(cls, url, *, loop=None, stream=False, download=False): SAVE_PATH = os.path.join(os.getcwd(), 'downloads') ydl_opts = { 'format': 'bestaudio/best', 'restrictfilenames': True, 'noplaylist': True, 'nocheckcertificate': True, 'ignoreerrors': False, 'logtostderr': False, 'quiet': True, 'no_warnings': True, 'default_search': 'auto', 'source_address': '0.0.0.0', # bind to ipv4 since ipv6 addresses cause issues sometimes 'preferredcodec': [{ 'key': 'FFmpegExtractAudio', 'preferredcodec': 'webm', 'preferredquality': '192', }], 'outtmpl':SAVE_PATH + '/%(title)s.%(ext)s', } #results = YoutubeSearch(url, max_results=3).to_dict() #vid_url = 'https://www.youtube.com' + results[0]['url_suffix'] #thumbnails = results[0]['thumbnails'] #title = results[0]['title'] #print('vid_url:{}, thumbnails:{}, title:{}, download:{},url:{}'.format(vid_url, thumbnails, title, download, url)) with youtube_dl.YoutubeDL(ydl_opts) as ydl: data = ydl.extract_info(f"ytsearch:{url}", download=download)['entries'][0] URL = data['url'] thumbnails = data['thumbnails'] title = data['title'] vid_url = data['webpage_url'] print(URL) #Renaming files if downloaded if download==True: files = os.listdir(os.path.join(os.getcwd(), 'downloads')) for file_name in files: if not file_name.endswith('.part'): # To download files as .mp3 #mp3_format = os.path.join(os.getcwd(), 'downloads', file_name.replace(file_name.split('.')[-1], 'mp3')) file_name = os.path.join(os.getcwd(), 'downloads', file_name) os.rename(file_name, title + '.mp3') return(URL,thumbnails, title, vid_url) @bot.command(name='p', brief='To play song note: Please enter: `.join` first', help="example: `.play gangnam style`") async def play(ctx, *, url): global playing playing = url if not ctx.message.author.voice: await ctx.send("{} is not connected to a voice channel".format( ctx.message.author.name)) return else: channel = ctx.message.author.voice.channel try: global player player = await channel.connect() except: pass #joined the channel try: server = ctx.message.guild voice_channel = server.voice_client #print('voice_channel : ' + str(voice_channel)) async with ctx.typing(): URL, thumbnails, title, vid_url = await YTDLSource.from_url(url, loop=bot.loop) #to stop playing if already playing another player.stop() player.play(discord.FFmpegPCMAudio(URL)) print('vid_url:{}, thumbnails:{}, title:{}, URL:{},url:{}'.format(vid_url, thumbnails, title, URL, url)) embed=discord.Embed(title=title, #description=stream['longDesc'], color=0x00FFFF, url=vid_url) embed.set_author(name=ctx.message.author) embed.set_thumbnail(url=thumbnails[0]['url']) embed.timestamp = datetime.datetime.utcnow() embed.set_footer(text=f'Added by {ctx.author}') message = await ctx.send(embed=embed) emos=['⏸️','⏹️', '⬇️']#['⏮️', '⏸️', '⏹️', '⏭️', '⬇️'] for emoji in emos: await message.add_reaction(emoji) except Exception as e: print(e) await ctx.send("The bot is not connected to a voice channel.") #Downloads videb name/url and returns full filename async def download_from_youtube(url): SAVE_PATH = os.path.join(os.getcwd(), 'downloads') ydl_opts = { 'format': 'bestaudio/best', 'preferredcodec': [{ 'key': 'FFmpegExtractAudio', 'preferredcodec': 'webm', 'preferredquality': '192', }],'outtmpl':SAVE_PATH + '/%(title)s.%(ext)s', } print(' downloading!!! ') with youtube_dl.YoutubeDL(ydl_opts) as ydl: try: ydl.download([url]) except: video = ydl.extract_info(f"ytsearch:{url}", download=True)['entries'][0] else: video = ydl.extract_info(url, download=False) #return video #print('type_of'+str(type(video))) # Didnot work for filename we extracted did not match with actual file_name '''file_name=str(video['title'] + '-' +video['id'] + '.' +video['formats'][3]['ext']) file_name = file_name.replace('/','_') ''' files = os.listdir(os.path.join(os.getcwd(), 'downloads')) for file_name in files: if not file_name.endswith('.part'): # To download files as .mp3 mp3_format = os.path.join(os.getcwd(), 'downloads', file_name.replace(file_name.split('.')[-1], 'mp3')) file_name = os.path.join(os.getcwd(), 'downloads', file_name) os.rename(file_name, mp3_format) print('file_name: {}'.format(file_name)) print('mp3_format: {}'.format(mp3_format)) return(mp3_format) @bot.command(name='d', brief='To download song note: Please enter: `.d song name` ', help="example: `.d gangnam style`") async def d(ctx, *, url:str): if not 'downloads' in os.listdir(): os.mkdir('downloads') print('Try download') async with ctx.typing(): URL, thumbnails, title, vid_url = await YTDLSource.from_url(url, loop=bot.loop, download=True) full_downloaded_file_name = title + '.mp3' await ctx.send(file=discord.File(full_downloaded_file_name)) os.remove(full_downloaded_file_name) print(' downloaded!!! ') @bot.command(name='pause', help='This command pauses the song') async def pause(ctx): voice_client = ctx.message.guild.voice_client if voice_client.is_playing(): await voice_client.pause() else: await ctx.send("The bot is not playing anything at the moment.") @bot.command(name='resume', help='Resumes the song') async def resume(ctx): voice_client = ctx.message.guild.voice_client if voice_client.is_paused(): await voice_client.resume() else: await ctx.send( "The bot was not playing anything before this. Use play_song command" ) @bot.command(name='stop', help='Stops the song') async def stop(ctx): await ctx.message.add_reaction('🛑') voice_client = ctx.message.guild.voice_client if voice_client.is_playing(): await voice_client.stop() voice_client #os.remove( else: await ctx.send("The bot is not playing anything at the moment.") #To make leave voice channel if bot is alone in voice channel @bot.event async def on_voice_state_update(member, before, after): print('\n\n Fired on_voice_state_update function \n\n') voice_state = member.guild.voice_client if voice_state is None: # Exiting if the bot it's not connected to a voice channel return if len(voice_state.channel.members) == 1: await voice_state.disconnect() @bot.command(aliases=['donation', 'support']) async def donate(ctx, url: str = 'http://stream.radioparadise.com/rock-128'): embed=discord.Embed(title='Support:', description='''Thank you :-) \nesewa/khalti id:\n 9840445934 \n\n Paytreon:\nhttps://www.patreon.com/join/7095305? \n\n Coinbase:\n https://commerce.coinbase.com/checkout/63a4b635-8510-459f-b091-a4f0697993e6 \n\n And please vote for me here: https://top.gg/bot/862191340355715093/vote ''', color=0x00FFFF, #url=stream['url'] ) embed.set_author( name=ctx.message.author, ) #embed.set_thumbnail(url=stream['image']) #embed.pfp = author.avatar_url embed.timestamp = datetime.datetime.utcnow() embed.set_footer(text=f'Added by {ctx.author}') message = await ctx.send(embed=embed) #_______________________________________________________________________ # ----------------------------- --------------------------------------- # _______________________________________________________________________ # ----------------------------- FM Player ----------------------------- from discord import FFmpegPCMAudio from discord.ext.commands import Bot from dotenv import load_dotenv load_dotenv() #To be implemented global streams streams = None def start_load_streams(): global streams try: streams[0] except: with open('test_fm_list.json','r') as F: streams = json.load(F) #To get current, next, previous streams def get_stream(which=None, current=None): global streams try: streams[0] print('Streams already defined') except: with open('test_fm_list.json','r') as F: streams = json.load(F) streams = streams['stream_links'] print(streams) #global streams_url #streams=streams['stream_links'] #streams_url = [i['url'] for i in streams] finally: if current==None: current={ "name": "Radio Nepal", "city" : "kathmandu", "url": "https://radionepal.news/live/audio/mp3", "image": "https://radionepal.gov.np/wp-content/themes/rdnp/images/logo-en.png", "desc": "am/sw/fm radio", "longDesc": "Radio Nepal, oldest radio of nepal." } if which=='next': nxt = streams.index(current) + 1 # Triggred to get next station at the end of stations list if nxt >= len(streams): nxt -= len(streams) current = streams[nxt] print(nxt) elif which=='prev': prev = streams.index(current) - 1 print(prev) # Triggred to get previous station at the beginning of stations list if prev < 0: prev += len(streams) print('current:{}, prev:{}'.format(streams.index(current),prev)) current = streams[prev] return(current) @bot.command(aliases=['fm', 'radio']) async def playfm(ctx, url: str = 'http://stream.radioparadise.com/rock-128'): global playing playing = "fm" global currently_playing_message global stream stream = get_stream() #url = "https://radio-streaming-serv-1.hamropatro.com/radio/8050/radio.mp3" #url = 'https://radionepal.news/live/audio/mp3' #global channel channel = ctx.message.author.voice.channel global player try: player = await channel.connect() except: pass player.play(FFmpegPCMAudio(stream['url'])) #global message embed=discord.Embed(title=stream['name'], description=stream['longDesc'], color=0x00FFFF, url=stream['url']) embed.set_author( name=ctx.message.author, ) #icon_url=ctx.message.author.avatar_url) embed.set_thumbnail(url=stream['image']) #embed.pfp = author.avatar_url embed.timestamp = datetime.datetime.utcnow() embed.set_footer(text=f'Added by {ctx.author}') currently_playing_message = await ctx.send(embed=embed) #emojis = [':track_previous:', ':pause_button:', ':stop_button:', ':track_next:', ':record_button:', ':arrow_down:'] emos=['⏮️', '⏸️', '⏹️', '⏭️']#, '⏺️', '⬇️'] for emoji in emos: await currently_playing_message.add_reaction(emoji) def get_embed(reaction, user, stream): embed=discord.Embed(title=stream['name'], #description=stream['longDesc'], color=0x00FFFF, url=stream['url']) embed.set_author( name=user, ) #icon_url=ctx.message.author.avatar_url) embed.set_thumbnail(url=stream['image']) #embed.pfp = author.avatar_url embed.timestamp = datetime.datetime.utcnow() embed.set_footer(text=f'Added by {user}') return embed @bot.event async def on_reaction_add(reaction, user,a=''): #embed = reaction.embeds[0] #emoji = reaction.emoji #print('hii') #await reaction.message.add_reaction('♥️') global stream if not user.bot: # stop emoji if str(reaction.emoji) == "⏹️": player.stop() # pause emoji elif str(reaction.emoji) == "⏸️": if player.is_playing(): player.pause() print('paused') else: player.resume() print('resume') # next emoji elif str(reaction.emoji) == "⏭️": if playing=='fm': print('Playing next, current:{}'.format(stream)) stream = get_stream('next',stream) player.stop() player.play(FFmpegPCMAudio(stream['url'])) embed=get_embed(reaction, user, stream) await currently_playing_message.edit(embed=embed) #message.send('Hello World') #play_next() # previous emoji elif str(reaction.emoji) == "⏮️": if playing=='fm': print('Playing next, current:{}'.format(stream)) stream = get_stream('prev', stream) player.stop() player.play(FFmpegPCMAudio(stream['url'])) embed=get_embed(reaction, user, stream) await currently_playing_message.edit(embed=embed) print('Playing next') # download emoji elif str(reaction.emoji) == "⬇️": if playing!='fm': if not 'downloads' in os.listdir(): os.mkdir('downloads') print('Try download') async with reaction.message.channel.typing(): URL, thumbnails, title, vid_url = await YTDLSource.from_url(playing, loop=bot.loop, download=True) full_downloaded_file_name = title + '.mp3' await reaction.message.channel.send(file=discord.File(full_downloaded_file_name)) os.remove(full_downloaded_file_name) print(' downloaded!!! ') else: await reaction.message.add_reaction(reaction) #print('hii') #print(reaction) #print(reaction.message) #print(user) #if user.bot: # return #else: # previous_messages = await channel.history(limit=1).flatten() # prev_message.add_reaction('♥️') '''if emoji == "emoji 1": fixed_channel = bot.get_channel(channel_id) await fixed_channel.send(embed=embed) elif emoji == "emoji 2": #do stuff elif emoji == "emoji 3": #do stuff else: return''' @bot.event async def on_reaction_remove(reaction, user): print('\nremoved reaction\n') global stream if not user.bot: # stop emoji if str(reaction.emoji) == "⏹️": player.stop() # pause emoji elif str(reaction.emoji) == "⏸️": if player.is_playing(): player.pause() print('paused') else: player.resume() print('resume') # next emoji elif str(reaction.emoji) == "⏭️": if playing=='fm': print('Playing next, current:{}'.format(stream)) stream = get_stream('next',stream) player.stop() player.play(FFmpegPCMAudio(stream['url'])) embed=get_embed(reaction, user, stream) await currently_playing_message.edit(embed=embed) #message.send('Hello World') #play_next() # previous emoji elif str(reaction.emoji) == "⏮️": if playing=='fm': print('Playing next, current:{}'.format(stream)) stream = get_stream('prev', stream) player.stop() player.play(FFmpegPCMAudio(stream['url'])) embed=get_embed(reaction, user, stream) await currently_playing_message.edit(embed=embed) print('Playing next') # download emoji elif str(reaction.emoji) == "⬇️": if playing=='fm': if not 'downloads' in os.listdir(): os.mkdir('downloads') print('Try download') async with reaction.message.channel.typing(): full_downloaded_file_name = await download_from_youtube(playing) await reaction.message.channel.send(file=discord.File(full_downloaded_file_name)) os.remove(full_downloaded_file_name) print(' downloaded!!! ') else: await reaction.message.add_reaction(reaction) # _____________________________________________________ # ///////////////////// FM Player ///////////////////// # _____________________________________________________ @bot.command(aliases=['s', 'sto']) async def stopfm(ctx): player.stop() @bot.command( name='disable_unleashing', brief='To disable/stop add unleashing all reddit posts to the server', help='e.g. `.disable_unleashing`') async def disable_unleashing(ctx): try: unleashing.stop() await ctx.send('unleashing disabled successfully.') except: await ctx.send('already disabled.') @bot.command( name='enable_unleashing', brief= 'To enable/start unleashing previously stopped reddit posts to the server', help='e.g. `.enable_unleashing`') async def enable_unleashing(ctx): try: unleashing.start() await ctx.send('unleashing enabled successfully.') except: await ctx.send('already enabled.') @bot.command(name='disable_autoattend', brief='To start autoattending in fuse classroom.', help='e.g. `.start_unleashing`') async def disable_autoattend(ctx): try: auto_attend.stop() await ctx.send('fuse auto-attend disabled successfully.') except: await ctx.send('already disabled.') @bot.command(name='enable_autoattend', brief='To enable/stopsrt autoattending in fuse classroom.', help='e.g. `.enable_unleashing`') async def enable_autoattend(ctx): try: auto_attend.start() await ctx.send('fuse auto-attend enabled successfully.') except: await ctx.send('already enabled.') @bot.command(name='video_embed_test', brief='', help='e') async def video(ctx): embed = discord.Embed( title='title', url='https://thumbs2.redgifs.com/WelcomeSweetTadpole-mobile.mp4', description='body', colour=discord.Color.red()) embed.set_image( url="https://thumbs2.redgifs.com/WelcomeSweetTadpole-mobile.mp4") embed.set_video( url="https://www.redgifs.com/watch/blissfulimperfectyardant") await ctx.send(embed=embed) ''' async def my_background_task(): await client.wait_until_ready() counter = 0 channel = client.get_channel(id=123456789) # replace with channel_id while not client.is_closed(): counter += 1 await channel.send(counter) print(counter) await asyncio.sleep(60) # task runs every 60 seconds''' '' sad_words = [ "sad", "depressed", "unhappy", "angry", "miserable", "depressing", "hurt", "pain" ] starter_encouragements = [ "Cheer up!", "You are a great person / bot!", ] commandss = [ '\".h fuse\" or \".help fuse\" -> for fuse_auto_attend help', 'fuse auto-attend registration at: https://ioee.herokuapp.com/', '\".inspire\" or \".quote\" -> to display quote ', '\".joke\" -> to display joke', '\".meme\" -> displays best random meme', '\".riddle\" -> displays best random riddle', '\".puns\" -> displays best random puns', '\".knock knock\" -> displays knock knock joke', '\".deactivate\" -> deactivates the bot .activate -> activates the bot', '\".new inspirational_message\" -> Adds new inspirationsl message to db', '\".del inspirational_message\" -> deletes inspirational message from db', '\".list\" -> lists the current inspirational messages', ] fuse_help_commands = [ '\".h\" or \".help\" - for general help', '----------- ------------------------- -----------', 'fuse auto-attend registration at: https://ioee.herokuapp.com/', '---------------------------------', '\".add_user user_token\" -> to add user for auto-fuse attandance', '.remove_user user_token -> to remove user', '\".list_user\" -> to list available users', '\".check class\" or \".snoop class\" -> checks if live class started.', '\".mute unsuccessful\" -> to mute unsuccessful attending_logs. ie. hide \"Live Class not started\" messages', '\".mute successful\" -> to mute successful attending_logs ie. hide messages when attended successfully', '\".unmute unsuccessful\" -> to unmute unsuccessful attending_logs ie. show \"Live Class not started\" messages', '\".umute successful\" -> to unmute successful attending_logs ie. show messages when attended successfully', ] #from discord.ext import commands #bot = commands.Bot(command_prefix='.') #@bot.command() #async def test(ctx): # await ctx.send('I heard you! {0}'.format(ctx.author)) '''print('--------------Test Mode--------------------------------') print(client.servers) print('-------------------------------------------------------')''' if "responding" not in db.keys(): db["responding"] = True if "unsuccessful_logs" not in db.keys(): db["unsuccessful_logs"] = False if "successful_logs" not in db.keys(): db["successful_logs"] = True def get_quote(): response = requests.get("https://zenquotes.io/api/random") json_data = json.loads(response.text) quote = json_data[0]['q'] + " -" + json_data[0]['a'] return (quote) def get_joke(): response = requests.get("https://imao.herokuapp.com/jokes/api/random/") json_data = response.json() joke = str(json_data['title']) + ' : ' + str( json_data['body']) + ' - ' + str(json_data['author']) return (joke) def get_jokes(no_of_jokes): response = requests.get("https://imao.herokuapp.com/jokes/api/{}/".format( int(no_of_jokes))) jokes = [] for joke in response.json()['jokes']: jokes.append( str(joke['title']) + ' : ' + str(joke['body']) + ' - ' + str(joke['author'])) return (jokes) def get_puns(): return ('Puns are comming very very soon!') def get_riddles(): return ('Riddles are comming very very soon!') def add_fuse_user(pseudoid): if "users" in db.keys(): users = db["users"] if pseudoid not in users: users.append(pseudoid) db["users"] = users else: db["users"] = [pseudoid] def remove_fuse_user(pseudoid): users = list(db["users"]) #if len(encouragements) > index: if pseudoid in users: #del encouragements[index] users.remove(pseudoid) db["users"] = users def update_encouragements(encouraging_message): if "encouragements" in db.keys(): encouragements = db["encouragements"] encouragements.append(encouraging_message) db["encouragements"] = encouragements else: db["encouragements"] = [encouraging_message] def delete_encouragment(index): encouragements = list(db["encouragements"]) #if len(encouragements) > index: if index in encouragements: #del encouragements[index] encouragements.remove(index) db["encouragements"] = encouragements def sanitize_db(): users = list(set(list(db["users"]))) users_sanitized = [] for user in users: users_sanitized.append( user.replace('\'', '').replace('\"', '').strip()) db["users"] = users_sanitized print('Users sanitized. \n Users:') print(list(db["users"])) def attend_each(usr): custom_url = 'https:ioee.herokuapp.com/attend/{}/'.format(usr) response = requests.get(custom_url) return (response.text) #---------------Working------------------------ # For scrapping quotes every 1 min. @tasks.loop(minutes=1) async def start_scrapping(): with open('quotes.json','r') as f: saved_quotes = json.load(f) # Got saved quotes saved_quotes = saved_quotes['quotes'] new_quotes=requests.get('https://zenquotes.io/api/quotes').json() # To combine new and old quotes n=0 for quote in new_quotes: if quote not in saved_quotes: saved_quotes.append(quote) n+=1 total_quotes = len(saved_quotes) with open('quotes.json','w') as file: json.dump({'quotes' : saved_quotes}, file, indent = 4) print('Saved {} quotes, total:{}'.format(n,total_quotes)) @tasks.loop(minutes=30) async def auto_attend(mannual_attempt=False): intents = discord.Intents.default() intents.members = True #user= await client.get_user("487904509670337509") #await client.send_message(user, "Your message goes here") #client.get_user(487904509670337509).send('hi') #sanitize_db() print("Users: ") users = list(db['users']) print(users) # To limit attend time from 9 am to 5:45 # i.e. (03:15 to 11:45) UTC now = datetime.datetime.now() morning = now.replace(hour=3, minute=5, second=0, microsecond=0) evening = now.replace(hour=11, minute=45, second=0, microsecond=0) #print('hello fella') if (now.strftime("%A") != "Saturday") and (now >= morning and now <= evening): channel = bot.get_channel(id=862205194283253763) #user = client.get_user("487904509670337509") #await user.send_message('hi') #username = bot.get_user('861131196779331624') #print('bot:') #print(username) users = [] if "users" in db.keys(): users = db["users"] for user in users: #response = str(attend_each(user)) custom_url = 'https://ioee.herokuapp.com/attend/{}/'.format(user) print(custom_url) response = requests.get(str(custom_url)) response = response.text.strip() print(response) print(response) if response == "Live Class not started" and db[ "unsuccessful_logs"] == False and mannual_attempt == False: continue elif db["successful_logs"] == False and mannual_attempt == False: await channel.send( "Successful attending attempt_logs are muted. to unmute please enter: .unmute unsuccessful" ) else: try: previous_messages = await channel.history( limit=1).flatten() prev_message = previous_messages[0].content print('previous_message:') print(prev_message) except: prev_message = "Your Attendance is done.Discord prevened previous message view" if (str(prev_message).strip() != response): #print("mannual_attempt:{} db[\"successful_logs\"]:{} db[\"unsuccessful_logs\"]:{} response=={}, response:{}".format(mannual_attempt, db["successful_logs"],db["unsuccessful_logs"], response == "Live Class not started", response ) ) print('not same messages:prev_message and rseponse') await channel.send(response) #await channel.send('user:'+str(user)) else: print('same message as previous message, so not sending') #print(prev_message) #print(response) #print(str(prev_message)==str(response)) #print(type(response)) pass #----------- To list discord servers --------- @tasks.loop(hours=25) async def share_info(): intents = discord.Intents.default() intents.members = True channel = bot.get_channel(id=882664470692909056) response = requests.get('http://ioee.herokuapp.com/meroshare/') response = response.text.strip() print(response) try: previous_messages = await channel.history(limit=1).flatten() prev_message = previous_messages[0].content print('previous_message:') print(prev_message) except: pass if (str(prev_message).strip() != response): print('not same messages:prev_message and rseponse') await channel.send(response) else: print('same message as previous message, so not sending') pass @tasks.loop(hours=25) async def meroshare(mannual_attempt=False): intents = discord.Intents.default() intents.members = True #user= await client.get_user("487904509670337509") #await client.send_message(user, "Your message goes here") #client.get_user(487904509670337509).send('hi') #sanitize_db() #print("Users: ") #users=list(db['users']) #print(users) # To limit attend time from 9 am to 5:45 # i.e. (03:15 to 11:45) UTC now = datetime.datetime.now() morning = now.replace(hour=1, minute=00, second=0, microsecond=0) evening = now.replace(hour=10, minute=15, second=0, microsecond=0) #print('hello fella') if (now.strftime("%A") != "Saturday") and (now >= morning and now <= evening): channel = bot.get_channel(id=882655060050444288) #user = client.get_user("487904509670337509") #await user.send_message('hi') #username = bot.get_user('861131196779331624') #print('bot:') #print(username) users = [] if "users" in db.keys(): users = db["users"] for user in users: #response = str(attend_each(user)) custom_url = 'https://ioee.herokuapp.com/attend/{}/'.format(user) print(custom_url) response = requests.get(str(custom_url)) response = response.text.strip() print(response) print(response) if response == "Live Class not started" and db[ "unsuccessful_logs"] == False and mannual_attempt == False: continue elif db["successful_logs"] == False and mannual_attempt == False: await channel.send( "Successful attending attempt_logs are muted. to unmute please enter: .unmute unsuccessful" ) else: try: previous_messages = await channel.history(limit=1 ).flatten() prev_message = previous_messages[0].content print('previous_message:') print(prev_message) except: prev_message = "Your Attendance is done.Discord prevened previous message view" if (str(prev_message).strip() != response): #print("mannual_attempt:{} db[\"successful_logs\"]:{} db[\"unsuccessful_logs\"]:{} response=={}, response:{}".format(mannual_attempt, db["successful_logs"],db["unsuccessful_logs"], response == "Live Class not started", response ) ) print('not same messages:prev_message and rseponse') await channel.send(response) #await channel.send('user:'+str(user)) else: print('same message as previous message, so not sending') #print(prev_message) #print(response) #print(str(prev_message)==str(response)) #print(type(response)) pass #----------- To list discord servers --------- class OwnerCommands(commands.Cog): def __init__(self, client): self.client = client @commands.Cog.listener() async def on_ready(self): print("OwnerCommands Is Ready") @bot.command() async def servers(ctx): discord.Intents.members = True activeservers = bot.guilds embed = discord.Embed( title='Servers and members', description='', colour=discord.Color.green(), ) '''for guild in activeservers: print('guild.channels') print(guild.channels) embed.add_field( name = str(guild.name) + ' ({}) own({})'.format(guild.member_count, guild.owner), value=str([i.name for i in guild.members]), ) ''' print('members:') for i in bot.guilds[1:]: a = i.fetch_members(limit=None) aa = [] async for ii in a: aa.append(ii.name) #print(i, ii) embed.add_field(name=str(i) + ' ({}) own({})'.format(i.member_count, i.owner), value=str(aa)) await ctx.send(embed=embed) #print(a) # print(channel) #await ctx.send(guild.name) #print(guild.name) def setup(client): bot.add_cog(OwnerCommands(bot)) #------------------------- async def unleash_reddit_jokes(subreddit, channel_id, no_of_posts=7): channel = bot.get_channel(id=int(channel_id)) for n, submission in enumerate( reddit.subreddit('jokes').top('day', limit=int(no_of_posts / 2))): print('Unleash for loop:{}'.format(n)) title = submission.title body = submission.selftext embed = discord.Embed(title=title, url=submission.url, description=body, colour=discord.Color.blue()) embed.set_image(url=submission.url) await channel.send(embed=embed) for n, submission in enumerate( reddit.subreddit('jokes').hot(limit=no_of_posts - int(no_of_posts / 4))): print('Unleash for loop:{}'.format(n)) title = str(submission.title)[:256] body = submission.selftext embed = discord.Embed(title=title, url=submission.url, description=body, colour=discord.Color.blue()) await channel.send(embed=embed) for n, submission in enumerate( reddit.subreddit('jokes').new(limit=no_of_posts - math.ceil(no_of_posts / 4))): print('Unleash for loop:{}'.format(n)) title = submission.title body = submission.selftext embed = discord.Embed(title=title, url=submission.url, description=body, colour=discord.Color.blue()) await channel.send(embed=embed) async def unleash_reddit(subreddit, channel_id, no_of_posts=5): channel = bot.get_channel(id=int(channel_id)) submissions = await reddit.subreddit(subreddit) donot_proceed = 0 #To display hot post if only one is to be fetched if no_of_posts == 1: donot_proceed = 1 no_of_posts = 2 async for submission in submissions.hot(limit=int(no_of_posts / 4)): print('Unleash for loop:{}'.format(0)) title = submission.title body = submission.selftext embed = discord.Embed(title=title, url=submission.url, description=body, colour=discord.Color.red()) embed.set_image(url=submission.url) print('Submission_url: ', submission.url) try: #To filter lenthy messages > 2500 letters if len(str(body)) < 2500: image_formats = ['jpg', 'jpeg', 'png'] #checks if image_format in submission.url if sum([(i in str(submission.url)) for i in image_formats]): await channel.send(embed=embed) else: await channel.send(submission.url) except: pass if donot_proceed != 1: async for submission in submissions.top('day', limit=int(no_of_posts / 2)): print('Unleash for loop:{}'.format('n')) title = submission.title body = submission.selftext embed = discord.Embed(title=title, url=submission.url, description=body, colour=discord.Color.red()) embed.set_image(url=submission.url) print('Submission_url: \"', submission.url, '\"') if submission.url == '': print('Guess What') try: if len(str(body)) < 2500: image_formats = ['jpg', 'jpeg', 'png'] #checks if image_format in submission.url if sum([(i in str(submission.url)) for i in image_formats]): await channel.send(embed=embed) else: await channel.send(submission.url) except: pass async for submission in submissions.new(limit=no_of_posts - math.ceil(no_of_posts / 4)): print('Unleash for loop:{}'.format(0)) title = submission.title body = submission.selftext embed = discord.Embed(title=title, url=submission.url, description=body, colour=discord.Color.red()) embed.set_image(url=submission.url) print('Submission_url: ', submission.url) try: if len(str(body)) < 2500: image_formats = ['jpg', 'jpeg', 'png'] #checks if image_format in submission.url if sum([(i in str(submission.url)) for i in image_formats]): await channel.send(embed=embed) else: await channel.send(submission.url) except: pass ''' async def unleash_reddit(subreddit, channel_id, no_of_posts=5): channel = bot.get_channel(id=int(channel_id)) submissions_top = await reddit.subreddit(subreddit) submissions_hot = await reddit.subreddit(subreddit) submissions_new = await reddit.subreddit(subreddit) #30% top, 40%hot, 30%new for i in range(0, no_of_posts): print('Unleash for loop:{}'.format(i)) if i < int(no_of_posts/3): submission=random.choice([x async for x in submissions_top.top(limit=25)]) print(a) ''async for x in submissions_top.top(limit=15): if not x.stickied: submission = x #submission = next(x async for x in submissions_top.top('all') if not x.stickied)'' elif i < int(no_of_posts/7): #submission = next(x async for x in submissions_hot.hot('all') if not x.stickied) submission=random.choice([x async for x in submissions_top.hot(limit=35)]) else: #submission = next(x async for x in submissions_new.new('all') if not x.stickied) submission=random.choice([x async for x in #submissions_top.new(limit=15)]) embed=discord.Embed( title=submission.title, description=submission.selftext, #description=submission.title, colour=discord.Color.green()) embed.set_image(url=submission.url) await channel.send(embed=embed)''' @tasks.loop(hours=6) async def unleashing(): print('\nstart Unleashing') intents = discord.Intents.default() #discord.Intents.members = True intents.members = True intents.all() for channel_id in dict(db['unleash']).keys(): for each_subreddit in db['unleash'][str(channel_id)]: await unleash_reddit(each_subreddit, str(channel_id), 10) print('Unleashed') @bot.event async def on_ready(): print('We have logged in as \"{0.user.name}\"'.format(bot)) print(bot.user.id) #For fuse attendance trying #auto_attend.start() #For viewing share_info #share_info.start() #for unleashing from reddit unleashing.start() start_scrapping.start() game = discord.Game("Chilling out.") streaming = discord.Streaming(name='pubg lite', url="https://www.twitch.tv/caterpileer") #movie.url="https://thumbs2.redgifs.com/WelcomeSweetTadpole-mobile.mp4" await bot.change_presence(status=discord.Status.online, activity=streaming) #await bot.process_commands(message) @bot.event async def on_message(message): # we do not want the bot to reply to itself if message.author == bot.user: return if message.content.startswith('.guess'): await message.channel.send('Guess a number between 1 to 10') def guess_check(m): return m.content.isdigit() guess = await bot.wait_for('message', timeout=5.0, check=guess_check) answer = random.randint(1, 10) if guess is None: fmt = 'Sorry, you took too long. It was {}.' await message.channel.send(fmt.format(answer)) return if int(guess.content) == answer: await message.channel.send('You are right!') else: await message.channel.send( 'Sorry. It is actually {}.'.format(answer)) else: if message.guild is None and message.author != bot.user: #await channel.send(str(message.author) + str(message.content)) embed = discord.Embed(title=message.author, description=message.content) channel = bot.get_channel(id=873477235477184522) await channel.send(embed=embed) print(str(message.content)) if any(word in str(message) for word in sad_words): options = starter_encouragements if "encouragements" in db.keys(): #print(list(db["encouragements"])) options = options #+ list(db["encouragements"]) await message.channel.send(random.choice(options)) await bot.process_commands(message) #await message.channel.send('hello') #if message!='fuck': # await message.add_reaction('♥️') # return '''print('Author_id:') print(message.author.id) #message.author.send_message('hi') print('Hello') msg = message.content.strip() if msg==".help" or msg == '.h': for command in commands: await message.channel.send(command) elif msg==".help fuse" or msg == '.h fuse': for command in fuse_help_commands: await message.channel.send(command) elif msg.startswith(".add_user"): pseudocode = msg.split(".add_user",1)[1].replace('\"','').replace('\'','').strip() add_fuse_user(pseudocode) await message.channel.send("User {} has been added/activated successfully.".format(pseudocode)) elif msg.startswith(".remove_user"): users = [] if "users" in db.keys(): pseudocode = msg.split(".remove_user",1)[1].strip() remove_fuse_user(pseudocode) await message.channel.send("User {} has been removed successfully.".format(pseudocode)) elif msg==".snoop class" or msg==".check class" or message == ".check live-class" or message == ".check live_class" or message == ".check live class": await message.channel.send("Sorry mannual checking is unavailable for a while"); await test.start(True) \'''if db["successful_logs"] == False: await message.channel.send("Successful attending attempt_logs are muted. to unmute please enter: .unmute unsuccessful") await message.channel.send("Checking Live Classes: ") m = test_mannual() await message.channel.send(m)\''' elif msg.startswith(".deactivate"): #switch = msg.split(".deac ",1)[1].lower() db["responding"] = False await message.channel.send("Encouragement bot is deactivated.\nPlease Enter: .activate to activate") elif msg.startswith(".activate"): db["responding"] = True await message.channel.send("Encouragement bot is activated.\nPlease enter: .deactivate to deactivate.") elif msg == (".mute unsuccessful") or msg == (".mute unsuccessful logs"): db["unsuccessful_logs"] = False await message.channel.send("unsuccessful attending_logs are muted.") elif msg == (".unmute unsuccessful") or msg == (".unmute unsuccessful logs"): db["unsuccessful_logs"] = True await message.channel.send("unsuccessful attending_logs are unmuted.") elif msg == (".mute successful") or msg == (".mute successful logs"): db["successful_logs"] = False await message.channel.send("successful attending_log are muted. to unmute please enter: .unmute successful") elif msg == (".unmute successful") or msg == (".unmute successful logs"): db["successful_logs"] = True await message.channel.send("successful attending_logs are unmuted.") if db["responding"]: if msg == ".list_users" or msg == ".list_user": users = [] if "users" in db.keys(): users = list(db["users"]) await message.channel.send('Users: '+str(users)) elif msg.startswith('.inspire') or msg.startswith('.quote'): quote = get_quote() await message.channel.send(quote) elif msg.startswith('.joke'): if msg == '.joke': joke = get_joke() await message.channel.send("\n\n\nJoke:" + str(joke)) else: try: n = int(msg.split(' ')[1].strip()) except: n = int(msg.split(' ')[2].strip()) jokes = list(get_jokes(n)) for joke in jokes: await message.channel.send("\n\n\nJoke:" + str(joke)) elif msg.startswith('.riddle'): riddle = get_riddles() await message.channel.send(riddle) elif msg.startswith('.puns'): puns = get_puns() await message.channel.send(puns) elif msg.startswith('.memes') or msg.startswith('.knock knock'): await message.channel.send('Sorry! ' + str(msg) + ' are comming very very soon!' ) elif msg.startswith(".new"): encouraging_message = msg.split(".new ",1)[1].strip() update_encouragements(encouraging_message) await message.channel.send("New encouraging message added.") elif msg.startswith(".del"): encouragements = [] if "encouragements" in db.keys(): index = msg.split(".del",1)[1].strip() delete_encouragment(index) encouragements = list(db["encouragements"]) await message.channel.send(encouragements) elif msg.startswith(".list"): encouragements = [] if "encouragements" in db.keys(): encouragements = list(db["encouragements"]) await message.channel.send(encouragements)''' keep_alive() bot.run(os.environ['TOKEN']) #client.loop.create_task(my_background_task()) bot.run('token') #
75,699
23,773
# Code Taken from https://github.com/LYH-YF/MWPToolkit # -*- encoding: utf-8 -*- # @Author: Yihuai Lan # @Time: 2021/08/29 22:05:03 # @File: transformer_layer.py import torch import math from torch import nn from torch.nn import functional as F from transformers.activations import gelu_new as gelu_bert from module.Attention.multi_head_attention import MultiHeadAttention from module.Attention.multi_head_attention import EPTMultiHeadAttention from module.Attention.group_attention import GroupAttention from utils.utils import clones class TransformerLayer(nn.Module): r"""Transformer Layer, including a multi-head self-attention, a external multi-head self-attention layer (only for conditional decoder) and a point-wise feed-forward layer. Args: self_padding_mask (torch.bool): the padding mask for the multi head attention sublayer. self_attn_mask (torch.bool): the attention mask for the multi head attention sublayer. external_states (torch.Tensor): the external context for decoder, e.g., hidden states from encoder. external_padding_mask (torch.bool): the padding mask for the external states. Returns: feedforward_output (torch.Tensor): the output of the point-wise feed-forward sublayer, is the output of the transformer layer """ def __init__(self, embedding_size, ffn_size, num_heads, attn_dropout_ratio=0.0, attn_weight_dropout_ratio=0.0, ffn_dropout_ratio=0.0, with_external=False): super(TransformerLayer, self).__init__() self.multi_head_attention = MultiHeadAttention(embedding_size, num_heads, attn_weight_dropout_ratio) self.feed_forward_1 = nn.Linear(embedding_size, ffn_size) self.feed_forward_2 = nn.Linear(ffn_size, embedding_size) self.attn_layer_norm = nn.LayerNorm(embedding_size, eps=1e-6) self.ffn_layer_norm = nn.LayerNorm(embedding_size, eps=1e-6) self.attn_dropout = nn.Dropout(attn_dropout_ratio) self.ffn_dropout = nn.Dropout(ffn_dropout_ratio) self.with_external = with_external if self.with_external: self.external_multi_head_attention = MultiHeadAttention(embedding_size, num_heads, attn_weight_dropout_ratio) self.external_layer_norm = nn.LayerNorm(embedding_size) self.reset_parameters() def reset_parameters(self): nn.init.normal_(self.feed_forward_1.weight, std=0.02) nn.init.normal_(self.feed_forward_2.weight, std=0.02) nn.init.constant_(self.feed_forward_1.bias, 0.) nn.init.constant_(self.feed_forward_2.bias, 0.) def gelu(self, x): return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) def forward(self, x, kv=None, self_padding_mask=None, self_attn_mask=None, external_states=None, external_padding_mask=None): residual = x if kv is None: x, self_attn_weights = self.multi_head_attention(query=x, key=x, value=x, key_padding_mask=self_padding_mask, attn_mask=self_attn_mask) else: x, self_attn_weights = self.multi_head_attention(query=x, key=kv, value=kv, key_padding_mask=self_padding_mask, attn_mask=self_attn_mask) x = self.attn_dropout(x) x = self.attn_layer_norm(residual + x) if self.with_external: residual = x x, external_attn_weights = self.external_multi_head_attention(query=x, key=external_states, value=external_states, key_padding_mask=external_padding_mask) x = self.attn_dropout(x) x = self.external_layer_norm(residual + x) else: external_attn_weights = None residual = x x = self.feed_forward_2(self.gelu(self.feed_forward_1(x))) x = self.ffn_dropout(x) x = self.ffn_layer_norm(residual + x) return x, self_attn_weights, external_attn_weights class GAEncoderLayer(nn.Module): """Group attentional encoder layer, encoder is made up of self-attn and feed forward. """ def __init__(self, size, self_attn, feed_forward, dropout): super(GAEncoderLayer, self).__init__() self.self_attn = self_attn self.feed_forward = feed_forward self.sublayer = clones(SublayerConnection(size, dropout), 2) self.size = size def forward(self, x, mask): """Follow Figure 1 (left) for connections.""" x = self.sublayer[0](x, lambda x: self.self_attn(x, x, x, mask)) return self.sublayer[1](x, self.feed_forward) class SublayerConnection(nn.Module): """ A residual connection followed by a layer norm. Note for code simplicity the norm is first as opposed to last. """ def __init__(self, size, dropout): super(SublayerConnection, self).__init__() #self.norm = LayerNorm(size) self.norm = nn.LayerNorm(size, eps=1e-6) self.dropout = nn.Dropout(dropout) def forward(self, x, sublayer): """Apply residual connection to any sublayer with the same size.""" return x + self.dropout(sublayer(self.norm(x))) class LayerNorm(nn.Module): """Construct a layernorm module (See citation for details).""" def __init__(self, features, eps=1e-6): super(LayerNorm, self).__init__() self.a_2 = nn.Parameter(torch.ones(features)) self.b_2 = nn.Parameter(torch.zeros(features)) self.eps = eps def forward(self, x): mean = x.mean(-1, keepdim=True) std = x.std(-1, keepdim=True) return self.a_2 * (x - mean) / (std + self.eps) + self.b_2 class PositionwiseFeedForward(nn.Module): """Implements FFN equation.""" def __init__(self, d_model, d_ff, dropout=0.1): super(PositionwiseFeedForward, self).__init__() self.w_1 = nn.Linear(d_model, d_ff) self.w_2 = nn.Linear(d_ff, d_model) self.dropout = nn.Dropout(dropout) def forward(self, x): return self.w_2(self.dropout(F.relu(self.w_1(x)))) class EPTTransformerLayer(nn.Module): """ Class for Transformer Encoder/Decoder layer (follows the paper, 'Attention is all you need') """ def __init__(self, hidden_dim = None, num_decoder_heads = None, layernorm_eps = None,intermediate_dim= None): """ Initialize TransformerLayer class :param ModelConfig config: Configuration of this Encoder/Decoder layer """ super().__init__() # Self-attention layer self.attn = EPTMultiHeadAttention(hidden_dim=hidden_dim, num_heads=num_decoder_heads, layernorm_eps=layernorm_eps, dropout=0.0) # Source-Target attention layer self.mem = EPTMultiHeadAttention(hidden_dim=hidden_dim, num_heads=num_decoder_heads, layernorm_eps=layernorm_eps, dropout=0.0) # Dropout for self-attention self.dropout_attn = nn.Dropout(0.0) # Dropout for source-target attention self.dropout_mem = nn.Dropout(0.0) # Dropout for expansion before outputting self.dropout_expand = nn.Dropout(0.0) # Dropout for outputting self.dropout_out = nn.Dropout(0.0) # Linear transformation layer for expansion (H -> I) where I = vector dimension of intermediate state self.lin_expand = nn.Linear(hidden_dim, intermediate_dim) # Linear transformation layer for output (I -> H) self.lin_collapse = nn.Linear(intermediate_dim, hidden_dim) # Post Layer Normalization for self-attention self.norm_attn = nn.LayerNorm(hidden_dim, eps=layernorm_eps) # Post Layer Normalization for source-target attention self.norm_mem = nn.LayerNorm(hidden_dim, eps=layernorm_eps) # Post Layer Normalization for outputting self.norm_out = nn.LayerNorm(hidden_dim, eps=layernorm_eps) def forward(self, target, target_ignorance_mask=None, target_attention_mask=None, memory=None, memory_ignorance_mask=None): """ Forward-computation of Transformer Encoder/Decoder layers Args: target (torch.Tensor): FloatTensor indicating Sequence of target vectors. Shape [batch_size, target_length, hidden_size]. target_ignorance_mask (torch.Tensor): BoolTensor indicating Mask for target tokens that should be ignored. Shape [batch_size, target_length]. target_attention_mask (torch.Tensor) : BoolTensor indicating Target-to-target Attention mask for target tokens. Shape [target_length, target_length]. memory (torch.Tensor): FloatTensor indicating Sequence of source vectors. Shape [batch_size, sequence_length, hidden_size]. This can be None when you want to use this layer as an encoder layer. memory_ignorance_mask (torch.Tensor): BoolTensor indicating Mask for source tokens that should be ignored. Shape [batch_size, sequence_length]. Returns: torch.FloatTensor: Decoder hidden states per each target token, shape [batch_size, sequence_length, hidden_size]. """ # Compute self-attention attented = self.attn(query=target, attention_mask=target_attention_mask, key_ignorance_mask=target_ignorance_mask) target = target + self.dropout_attn(attented) target = self.norm_attn(target) # Compute attention over targets with source as queries. if memory is not None: attented = self.mem(query=target, key_value=memory, key_ignorance_mask=memory_ignorance_mask) target = target + self.dropout_mem(attented) target = self.norm_mem(target) # Pass linear transformations output = self.lin_collapse(self.dropout_expand(gelu_bert(self.lin_expand(target)))) target = target + self.dropout_out(output) target = self.norm_out(target) return target
9,862
3,196
import math from seal import * from seal_helper import * def example_ckks_basics(): print_example_banner("Example: CKKS Basics") parms = EncryptionParameters(scheme_type.CKKS) poly_modulus_degree = 8192 parms.set_poly_modulus_degree(poly_modulus_degree) parms.set_coeff_modulus(CoeffModulus.Create( poly_modulus_degree, [60, 40, 40, 60])) scale = pow(2.0, 40) context = SEALContext.Create(parms) print_parameters(context) keygen = KeyGenerator(context) public_key = keygen.public_key() secret_key = keygen.secret_key() relin_keys = keygen.relin_keys() encryptor = Encryptor(context, public_key) evaluator = Evaluator(context) decryptor = Decryptor(context, secret_key) encoder = CKKSEncoder(context) slot_count = encoder.slot_count() print("Number of slots: " + str(slot_count)) inputs = DoubleVector() curr_point = 0.0 step_size = 1.0 / (slot_count - 1) for i in range(slot_count): inputs.append(curr_point) curr_point += step_size print("Input vector: ") print_vector(inputs, 3, 7) print("Evaluating polynomial PI*x^3 + 0.4x + 1 ...") ''' We create plaintexts for PI, 0.4, and 1 using an overload of CKKSEncoder::encode that encodes the given floating-point value to every slot in the vector. ''' plain_coeff3 = Plaintext() plain_coeff1 = Plaintext() plain_coeff0 = Plaintext() encoder.encode(3.14159265, scale, plain_coeff3) encoder.encode(0.4, scale, plain_coeff1) encoder.encode(1.0, scale, plain_coeff0) x_plain = Plaintext() print("-" * 50) print("Encode input vectors.") encoder.encode(inputs, scale, x_plain) x1_encrypted = Ciphertext() encryptor.encrypt(x_plain, x1_encrypted) x3_encrypted = Ciphertext() print("-" * 50) print("Compute x^2 and relinearize:") evaluator.square(x1_encrypted, x3_encrypted) evaluator.relinearize_inplace(x3_encrypted, relin_keys) print(" + Scale of x^2 before rescale: " + "%.0f" % math.log(x3_encrypted.scale(), 2) + " bits") print("-" * 50) print("Rescale x^2.") evaluator.rescale_to_next_inplace(x3_encrypted) print(" + Scale of x^2 after rescale: " + "%.0f" % math.log(x3_encrypted.scale(), 2) + " bits") print("-" * 50) print("Compute and rescale PI*x.") x1_encrypted_coeff3 = Ciphertext() evaluator.multiply_plain(x1_encrypted, plain_coeff3, x1_encrypted_coeff3) print(" + Scale of PI*x before rescale: " + "%.0f" % math.log(x1_encrypted_coeff3.scale(), 2) + " bits") evaluator.rescale_to_next_inplace(x1_encrypted_coeff3) print(" + Scale of PI*x after rescale: " + "%.0f" % math.log(x1_encrypted_coeff3.scale(), 2) + " bits") print("-" * 50) print("Compute, relinearize, and rescale (PI*x)*x^2.") evaluator.multiply_inplace(x3_encrypted, x1_encrypted_coeff3) evaluator.relinearize_inplace(x3_encrypted, relin_keys) print(" + Scale of PI*x^3 before rescale: " + "%.0f" % math.log(x3_encrypted.scale(), 2) + " bits") evaluator.rescale_to_next_inplace(x3_encrypted) print(" + Scale of PI*x^3 after rescale: " + "%.0f" % math.log(x3_encrypted.scale(), 2) + " bits") print("-" * 50) print("Compute and rescale 0.4*x.") evaluator.multiply_plain_inplace(x1_encrypted, plain_coeff1) print(" + Scale of 0.4*x before rescale: " + "%.0f" % math.log(x1_encrypted.scale(), 2) + " bits") evaluator.rescale_to_next_inplace(x1_encrypted) print(" + Scale of 0.4*x after rescale: " + "%.0f" % math.log(x1_encrypted.scale(), 2) + " bits") print() print("-" * 50) print("Parameters used by all three terms are different.") print(" + Modulus chain index for x3_encrypted: " + str(context.get_context_data(x3_encrypted.parms_id()).chain_index())) print(" + Modulus chain index for x1_encrypted: " + str(context.get_context_data(x1_encrypted.parms_id()).chain_index())) print(" + Modulus chain index for x1_encrypted: " + str(context.get_context_data(plain_coeff0.parms_id()).chain_index())) print() print("-" * 50) print("The exact scales of all three terms are different:") print(" + Exact scale in PI*x^3: " + "%.10f" % x3_encrypted.scale()) print(" + Exact scale in 0.4*x: " + "%.10f" % x1_encrypted.scale()) print(" + Exact scale in 1: " + "%.10f" % plain_coeff0.scale()) print("-" * 50) print("Normalize scales to 2^40.") # set_scale() this function should be add to seal/ciphertext.h line 632 x3_encrypted.set_scale(pow(2.0, 40)) x1_encrypted.set_scale(pow(2.0, 40)) ''' We still have a problem with mismatching encryption parameters. This is easy to fix by using traditional modulus switching (no rescaling). CKKS supports modulus switching just like the BFV scheme, allowing us to switch away parts of the coefficient modulus when it is simply not needed. ''' print("-" * 50) print("Normalize encryption parameters to the lowest level.") last_parms_id = x3_encrypted.parms_id() evaluator.mod_switch_to_inplace(x1_encrypted, last_parms_id) evaluator.mod_switch_to_inplace(plain_coeff0, last_parms_id) ''' All three ciphertexts are now compatible and can be added. ''' print("-" * 50) print("Compute PI*x^3 + 0.4*x + 1.") encrypted_result = Ciphertext() evaluator.add(x3_encrypted, x1_encrypted, encrypted_result) evaluator.add_plain_inplace(encrypted_result, plain_coeff0) ''' First print the true result. ''' plain_result = Plaintext() print("-" * 50) print("Decrypt and decode PI*x^3 + 0.4x + 1.") print(" + Expected result:") true_result = [] for x in inputs: true_result.append((3.14159265 * x * x + 0.4) * x + 1) print_vector(true_result, 3, 7) ''' Decrypt, decode, and print the result. ''' decryptor.decrypt(encrypted_result, plain_result) result = DoubleVector() encoder.decode(plain_result, result) print(" + Computed result ...... Correct.") print_vector(result, 3, 7) ''' While we did not show any computations on complex numbers in these examples, the CKKSEncoder would allow us to have done that just as easily. Additions and multiplications of complex numbers behave just as one would expect. ''' if __name__ == '__main__': example_ckks_basics()
6,564
2,471
from correlate import * import matplotlib import matplotlib.pyplot as plt import seaborn as sns from matplotlib import rc, rcParams rc('axes', linewidth=1) rc('font', weight='bold', size=10) def barplots(prefixes, regions, stains, corre, error, name, folder, ylim): for stain in stains: barplot(prefixes, regions, stain, corre, error, name, folder, ylim) def barplot(prefixes, regions, stain, corre, error, name, folder, ylim): colors = ['#bfe2e3', '#69869c', '#36896e', '#c22e00', '#c6d645', '#ffd3b6', '#b2b2b2', '#4724a9', '#9bc84d', '#7141ae', '#d2a782', '#933b61', '#435299', '#d88770', '#765aa8', '#719795'] Val, Std = [], [] for i, prefix in enumerate(prefixes): val, std = corre[prefix][stain], error[prefix][stain] Val.append(val) Std.append(std) fig, ax = plt.subplots(dpi=300, figsize=(6, 3)) index = [i for i in range(len(regions))] ax.bar(index, Val, yerr=Std, capsize=2, color=colors[:len(prefixes)]) ax.set_ylabel('Spearman\'s rank correlation', fontweight='bold') ax.set_ylim(ylim[0]-0.05, ylim[1]+0.05) ax.set_xticks(index) ax.set_xticklabels(regions, rotation=45, ha="right") ax.grid(which='major', axis='both', linestyle='--') plt.savefig(folder + 'bar_{}_{}.png'.format(stain, name), bbox_inches='tight') plt.close() if __name__ == "__main__": years = 2 layername = 'block2BN' time_threshold, type = 365*years, 'COG' folder = type + '_correlation_{}_years/'.format(years) if not os.path.exists(folder): os.mkdir(folder) interval = file_interval_info(type) y_lim = [0, 0] corre = collections.defaultdict(dict) error = collections.defaultdict(dict) pool = [[0, prefixes[i], regions[i]] for i in range(len(regions))] for i, region in enumerate(prefixes): for stain in stains: corr, std = get_correlation(region + '_' + stain, prefix_idx[region], time_threshold, interval, folder, type, layername, missing=0) corre[region][stain] = corr error[region][stain] = 0 y_lim[1] = max(y_lim[1], corr) y_lim[0] = min(y_lim[0], corr) pool[i][0] -= corr pool.sort() prefixes = [p[1] for p in pool] regions = [p[2] for p in pool] barplots(prefixes, regions, stains, corre, error, '{}days_{}shap_{}'.format(time_threshold, type, layername), folder, y_lim)
2,422
935
""" An example of running autofaiss by pyspark to produce N indices. You need to install pyspark before using the following example. """ from typing import Dict import faiss import numpy as np from autofaiss import build_index # You'd better create a spark session before calling build_index, # otherwise, a spark session would be created by autofaiss with the least configuration. _, index_path2_metric_infos = build_index( embeddings="hdfs://root/path/to/your/embeddings/folder", distributed="pyspark", file_format="parquet", temporary_indices_folder="hdfs://root/tmp/distributed_autofaiss_indices", current_memory_available="10G", max_index_memory_usage="100G", nb_indices_to_keep=10, ) index_paths = sorted(index_path2_metric_infos.keys()) ########################################### # Use case 1: merging 10 indices into one # ########################################### merged = faiss.read_index(index_paths[0]) for rest_index_file in index_paths[1:]: index = faiss.read_index(rest_index_file) faiss.merge_into(merged, index, shift_ids=False) with open("merged-knn.index", "wb") as f: faiss.write_index(merged, faiss.PyCallbackIOWriter(f.write)) ######################################## # Use case 2: searching from N indices # ######################################## K, DIM, all_distances, all_ids, NB_QUERIES = 5, 512, [], [], 2 queries = faiss.rand((NB_QUERIES, DIM)) for rest_index_file in index_paths: index = faiss.read_index(rest_index_file) distances, ids = index.search(queries, k=K) all_distances.append(distances) all_ids.append(ids) dists_arr = np.stack(all_distances, axis=1).reshape(NB_QUERIES, -1) knn_ids_arr = np.stack(all_ids, axis=1).reshape(NB_QUERIES, -1) sorted_k_indices = np.argsort(-dists_arr)[:, :K] sorted_k_dists = np.take_along_axis(dists_arr, sorted_k_indices, axis=1) sorted_k_ids = np.take_along_axis(knn_ids_arr, sorted_k_indices, axis=1) print(f"{K} nearest distances: {sorted_k_dists}") print(f"{K} nearest ids: {sorted_k_ids}") ############################################ # Use case 3: on disk merging of N indices # ############################################ # using faiss.merge_ondisk (https://github.com/facebookresearch/faiss/blob/30abcd6a865afef7cf86df7e8b839a41b5161505/contrib/ondisk.py ) # https://github.com/facebookresearch/faiss/blob/151e3d7be54aec844b6328dc3e7dd0b83fcfa5bc/demos/demo_ondisk_ivf.py # to merge indices on disk without using memory # this is useful in particular to use a very large index with almost no memory usage. from faiss.contrib.ondisk import merge_ondisk import faiss block_fnames = index_paths empty_index = faiss.read_index(block_fnames[0], faiss.IO_FLAG_MMAP) empty_index.ntotal = 0 merge_ondisk(empty_index, block_fnames, "merged_index.ivfdata") faiss.write_index(empty_index, "populated.index") pop = faiss.read_index("populated.index", faiss.IO_FLAG_ONDISK_SAME_DIR) ######################################################## # Use case 4: use N indices using HStackInvertedLists # ######################################################## # This allows using N indices as a single combined index # without changing anything on disk or loading anything to memory # it works well but it's slower than first using merge_ondisk # because it requires explore N pieces of inverted list for each # list to explore import os class CombinedIndex: """ combines a set of inverted lists into a hstack adds these inverted lists to an empty index that contains the info on how to perform searches """ def __init__(self, invlist_fnames): ilv = faiss.InvertedListsPtrVector() for fname in invlist_fnames: if os.path.exists(fname): index = faiss.read_index(fname, faiss.IO_FLAG_MMAP) index_ivf = faiss.extract_index_ivf(index) il = index_ivf.invlists index_ivf.own_invlists = False else: raise FileNotFoundError ilv.push_back(il) self.big_il = faiss.HStackInvertedLists(ilv.size(), ilv.data()) ntotal = self.big_il.compute_ntotal() self.index = faiss.read_index(invlist_fnames[0], faiss.IO_FLAG_MMAP) index_ivf = faiss.extract_index_ivf(self.index) index_ivf.replace_invlists(self.big_il, True) index_ivf.ntotal = self.index.ntotal = ntotal def search(self, x, k): D, I = self.index.search(x, k) return D, I index = CombinedIndex(index_paths) index.search(queries, K)
4,549
1,546
"""Callback module for rbkcli.""" import json from rbkcli.core.handlers.inputs import InputHandler from rbkcli.base.essentials import DotDict, RbkcliException from rbkcli.core.handlers import ApiTargetTools from rbkcli.core.handlers.outputs import OutputHandler class CallBack(ApiTargetTools): """Class to provide rbkcli internal api calls.""" def __init__(self, operations, base_kit): """Initialize callback class.""" ApiTargetTools.__init__(self, base_kit) self.operations = operations self.base_kit = base_kit self.validator = InputHandler(self.base_kit, self.operations) self.formatter = OutputHandler(base_kit, self.operations) def parseit(self, args): """Parse arguments provided.""" self.args = args if not isinstance(self.args, list): self.args = self.args.replace('rbkcli ', '') new_args = [] self.args = self.args.split('"') if len(self.args) % 2 != 0: for arg in enumerate(self.args): newarg = arg[1] if arg[0] % 2 == 0: newarg = arg[1].split() if isinstance(newarg, list): new_args = new_args + newarg else: new_args.append(newarg) else: print('Error ## Danger, danger, high voltage...') self.args = new_args self.request = self.base_kit.parser.parse_args(self.args) self.request = vars(self.request) self.request = self.base_kit.parser.un_list(self.request) return self.request, self.args def structreit(self, args, request): """Reestructure arguments provided.""" self.args = args self.request = {} requet_1 = {} for key in request.keys(): self.request[key] = request[key] requet_1[key] = request[key] # Structure the request to the needed format. stct_request = self.base_kit.parser.create_request_structure(requet_1, self.args) # Pass data to dot dictionary self.stct_request = DotDict() for key in stct_request.keys(): self.stct_request[key] = stct_request[key] # Normalizing request dictionary self.stct_request.endpoint = ' '.join(self.stct_request.api_endpoint) self.stct_request.formatt = 'raw' self.stct_request.param = self.stct_request.query self.stct_request.data = self.stct_request.parameter self.stct_request.structured = True return self.stct_request def callit(self, stct_request, args=None): """Call endpoint provided with arguments.""" if 'structured' not in stct_request.keys(): if args is None: args = [] stct_request = self.structreit(args, stct_request) self.stct_request = stct_request self.req = self.validator.validate(self.stct_request) api_result = self.operations.execute(self.req) if '{' in api_result.text or '[' in api_result.text: try: self.call_result = json.loads(api_result.text) except: self.call_result = { 'result_text': api_result.text} else: self.call_result = { 'result_text': api_result.text} return self.call_result def call_back(self, args): """Perform same level of parsing (even CLI) as any other request.""" self.request, self.args = self.parseit(args) self.stct_request = self.structreit(self.args, self.request) result = DotDict() result.text = self.callit(self.stct_request) result.status_code = 200 result.text = json.dumps(result.text, indent=2) return self.formatter.outputfy(self.req, result) def call_back_text(self, args): """Returns dict directly instead of API result.""" result = self.call_back(args) return json.loads(result.text)
4,170
1,291
# Generated by Django 2.2.1 on 2019-05-16 23:28 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operations = [ migrations.CreateModel( name='SocialProfile', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('provider', models.SmallIntegerField(choices=[('1', 'Facebook')])), ('social_id', models.CharField(max_length=255, unique=True)), ('photo', models.TextField(blank=True)), ('extra_data', models.TextField(blank=True)), ('user', models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, related_name='social_profile', to=settings.AUTH_USER_MODEL)), ], ), ]
1,001
303
from rest_framework import serializers, exceptions from greenbudget.lib.rest_framework_utils.serializers import ( EnhancedModelSerializer) from greenbudget.app.account.models import BudgetAccount, TemplateAccount from greenbudget.app.tagging.serializers import ColorField from greenbudget.app.subaccount.models import ( BudgetSubAccount, TemplateSubAccount) from .models import ( Group, BudgetAccountGroup, TemplateAccountGroup, BudgetSubAccountGroup, TemplateSubAccountGroup ) class GroupSerializer(EnhancedModelSerializer): id = serializers.IntegerField(read_only=True) name = serializers.CharField( required=False, allow_blank=False, allow_null=False ) created_by = serializers.PrimaryKeyRelatedField(read_only=True) updated_by = serializers.PrimaryKeyRelatedField(read_only=True) created_at = serializers.DateTimeField(read_only=True) updated_at = serializers.DateTimeField(read_only=True) estimated = serializers.FloatField(read_only=True) color = ColorField(content_type_model=Group, required=False) class Meta: model = Group fields = ( 'id', 'name', 'created_by', 'created_at', 'updated_by', 'updated_at', 'color', 'estimated') def validate_name(self, value): parent = self.context.get('parent') if parent is None: parent = self.instance.parent validator = serializers.UniqueTogetherValidator( queryset=parent.groups.all(), fields=('name', ), ) validator({'name': value}, self) return value def create(self, *args, **kwargs): """ Overridden to perform cleanup of empty :obj:`Group` instances. When a :obj:`Group` is created, it can be created with children that already belong to another :obj:`Group`. When this happens, the children are removed from the other :obj:`Group` and included in the new :obj:`Group`. We want to perform cleanup, and remove empty :obj:`Group` instances that have no children. This is partially accomplished with :obj:`track_model`. However, when the above situation occurs, the :obj:`track_model` will not work, because both of these implementations will fail: (1) :obj:`track_model` on :obj:`Group` @track_model(on_field_change_hooks={'parent': remove_empty_groups}) class BudgetAccountGroup(Group): ... parent = models.ForeignKey( to='budget.Budget', ... ) The mechanics of :obj:`track_model` responsible for removing the :obj:`Group` if it has no more children will not be triggered because :obj:`track_model` will only trigger when the :obj:`Group` with no more children itself is updated. In this situation, we are not updating the :obj:`Group` that has no more children, we are updating the new :obj:`Group` that will have the children that previously existed on the old :obj:`Group`. (2) :obj:`track_model` on :obj:`Account` or :obj:`SubAccount` class BudgetAccountGroup(Group): ... parent = models.ForeignKey( to='budget.Budget', ... ) @track_model(on_field_change_hooks={'group': remove_empty_groups}) class BudgetAccount(Account): group = models.ForeignKey( to='group.BudgetAccountGroup', related_name='children' ) Here, we cannot use the :obj:`track_model` on the :obj:`Account` or :obj:`SubAccount` models to remove empty groups after the group assigned to those models changes because for DRF we are updating the models via the `children` attribute, which is the reverse FK accessor, and apparently that does not trigger the post_save hooks on the primary model: PATCH /v1/groups/<pk>/ { children: [...] } -> Updating the `children` on the :obj:`BudgetAccountGroup` (i.e. updating a reverse FK accessor) will not trigger the `post_save` on :obj:`BudgetAccount`. For the above reason, we need to address this problem without the :obj:`track_model` behavior. (3) `post_save` signals directly on :obj:`Group` We cannot accomplish this at the database level (or model level) via post_save signals. Consider we try to accomplish this with the following signal: @dispatch.receiver(post_save, sender=BudgetSubAccountGroup) def remove_empty_groups(instance, **kwargs): for sibling_group in instance.parent.groups.all(): if sibling_group.children.count() == 0: sibling_group.delete() If we were to do this, we would run into issues creating instances of :obj:`Group`. Since the `children` field is a reverse FK accessor, the :obj:`Group` has to be created before an entity can be assigned a group. That means a :obj:`Group` will at times have no children just before children are assigned - and we cannot have those groups incident- ally deleted before children are assigned. For this reason, we need to accomplish this behavior at the request/ response interface - which is why we override this method here. TODO: ---- We should investigate whether or not there is a better way around this problem. At the very least, we should develop CRON tasks that should remove remnant empty groups. """ instance = super().create(*args, **kwargs) for sibling_group in instance.parent.groups.all(): if sibling_group != instance \ and sibling_group.children.count() == 0: sibling_group.delete() return instance def update(self, *args, **kwargs): """ Overridden to perform cleanup of empty :obj:`Group` instances. See docstring in `.create()` method for a more detailed explanation. """ instance = super().update(*args, **kwargs) siblings = [ sib for sib in instance.parent.groups.all() if sib != instance ] for sibling_group in siblings: if sibling_group.children.count() == 0: sibling_group.delete() return instance class AbstractAccountGroupSerializer(GroupSerializer): class Meta: abstract = True def validate_children(self, value): parent = self.context.get('parent') if parent is None: parent = self.instance.parent for child in value: if child.budget != parent: raise exceptions.ValidationError( "The %s %s does not belong to the same %s " "that the Group does (%s)." % ( type(child).__name__, child.pk, type(parent).__name__, parent.pk) ) return value class BudgetAccountGroupSerializer(AbstractAccountGroupSerializer): actual = serializers.FloatField(read_only=True) variance = serializers.FloatField(read_only=True) children = serializers.PrimaryKeyRelatedField( many=True, required=True, queryset=BudgetAccount.objects.active() ) class Meta(GroupSerializer.Meta): model = BudgetAccountGroup fields = GroupSerializer.Meta.fields + ( 'children', 'actual', 'variance') class TemplateAccountGroupSerializer(AbstractAccountGroupSerializer): children = serializers.PrimaryKeyRelatedField( many=True, required=True, queryset=TemplateAccount.objects.active() ) class Meta(GroupSerializer.Meta): model = TemplateAccountGroup fields = GroupSerializer.Meta.fields + ('children', ) class AbstractSubAccountGroupSerializer(GroupSerializer): class Meta: abstract = True def validate_children(self, value): parent = self.context.get('parent') if parent is None: parent = self.instance.parent for child in value: if child.parent != parent: raise exceptions.ValidationError( "The %s %s does not belong to the same %s " "that the Group does (%s)." % ( type(child).__name__, child.pk, type(parent).__name__, parent.pk) ) # Is this check necessary? Would this otherwise be constrained # by model restrictions? elif child.budget != parent.budget: raise exceptions.ValidationError( "The %s %s does not belong to the same %s " "that the Group does (%s)." % ( type(child).__name__, child.pk, type(child.budget).__name__, parent.pk) ) return value class BudgetSubAccountGroupSerializer(AbstractSubAccountGroupSerializer): actual = serializers.FloatField(read_only=True) variance = serializers.FloatField(read_only=True) children = serializers.PrimaryKeyRelatedField( many=True, required=True, queryset=BudgetSubAccount.objects.active() ) class Meta(GroupSerializer.Meta): model = BudgetSubAccountGroup fields = GroupSerializer.Meta.fields + ( 'children', 'actual', 'variance') class TemplateSubAccountGroupSerializer(AbstractSubAccountGroupSerializer): children = serializers.PrimaryKeyRelatedField( many=True, required=True, queryset=TemplateSubAccount.objects.active() ) class Meta(GroupSerializer.Meta): model = TemplateSubAccountGroup fields = GroupSerializer.Meta.fields + ('children', )
9,950
2,553
def email_subject(center_name): return f"[Urgent Reminder] Vaccine slot is now available at {center_name}" def email_body(email_data): return f"Hi, \n" \ f"Vaccine slot is available for below centers \n " \ f"Center name and available data \n {email_data} \n" \ f"Please register at cowin website https://cowin.gov.in \n" \ f"Have a lovely day ahead! :) \n" \ f"Thanks, \n" \ f"Shishir Bondre \n" \ f"To unsubscribe click here http://3.6.160.147:8000/vaccine_tracker/unsubscribe/"
570
198
# -*- coding: utf-8 -*- """ INTRO @author: Yi Zhang. Created on Tue May 21 11:57:52 2019 Department of Aerodynamics Faculty of Aerospace Engineering TU Delft, Delft, the Netherlands """ import inspect from screws.freeze.main import FrozenOnly from typing import Dict, Union import numpy as np from screws.decorators.classproperty.main import classproperty class DomainInputBase(FrozenOnly): def __init__(self, domain_name='domain without name'): self.domain_name = domain_name self._ndim_ = 2 self._region_corner_coordinates_ = None self._region_edge_types_ = None self._boundary_region_edges_ = None self._region_interpolators_ = None self._boundary_names_ = None self._periodic_boundary_pairs_ = dict() self._periodic_boundaries_ = set() self._region_sequence_ = None self._region_type_wr2_metric_ = None self._internal_parameters_ = list() INSP = inspect.getfullargspec(self.__init__) self.__arg_names___ = INSP[0][1:] assert INSP[1] is None and INSP[2] is None, "A domain input class can not have *args and **kwargs." assert len(INSP[3]) == len(self.__arg_names___), "A domain input class can only have kwargs." self._freeze_self_() @property def internal_parameters(self): """Internal parameters only affect internal metric, does not affect the domain shape.""" return self._internal_parameters_ @internal_parameters.setter def internal_parameters(self, internal_parameters): if isinstance(internal_parameters, list): pass elif isinstance(internal_parameters, str): internal_parameters = [internal_parameters,] elif isinstance(internal_parameters, (tuple, set)): internal_parameters = list(internal_parameters) else: raise NotImplementedError(f"internal_parameters = {internal_parameters} not acceptable.") assert isinstance(internal_parameters, list), \ f"please put internal_parameters({internal_parameters}) in list." if len(internal_parameters) > 0: assert all([ip in self.__arg_names___ for ip in internal_parameters]) self._internal_parameters_ = internal_parameters @property def domain_name(self): """ Mesh name. """ return self._domain_name_ @domain_name.setter def domain_name(self, dn): assert isinstance(dn, str), " <DomainInput> : domain name needs to be str." self._domain_name_ = dn @property def ndim(self): """ dimensions n. """ return self._ndim_ @property def region_interpolators(self): return self._region_interpolators_ @region_interpolators.setter def region_interpolators(self, region_interpolators): self._region_interpolators_ = region_interpolators def ___PRIVATE_region_name_requirement_checker___(self, regionDict): """ Requirements: 1). must be str 2). != domain name. 3). length > 2 4). Starts with 'R:' 5). can only have letters and _ """ for R in regionDict: assert isinstance(R, str), f"region name={R} wrong, need be str!" assert R != self.domain_name, f"region name == domain.name! wrong!" assert len(R) > 2, f"region name = {R} too short, must > 2." assert R[0:2] == 'R:', f"regions name = {R} does not start with 'R:'" R2 = R[2:].replace('_', '') assert R2.isalpha(),f"region_name = {R} wrong, can only have letter and _ (at >2)." @property def region_corner_coordinates(self): """ Store the regions 4 corners' coordinates. Returns ------- region_coordinates : dict A dict whose keys represent the regions names, and values represent the coordinates of regions corner points. In 2D: (UL, DL, UR, DR). L: Left, R: Right, U: Upper, D: Down """ return self._region_corner_coordinates_ @region_corner_coordinates.setter def region_corner_coordinates(self, _dict_): assert isinstance(_dict_, dict), " <DomainInput> : region_coordinates needs to be a dict." self.___PRIVATE_region_name_requirement_checker___(_dict_) for R in _dict_: assert np.shape(_dict_[R])[0] == 4, \ " <DomainInput> : region_coordinates[{}]={} is wrong.".format(R, _dict_[R]) self._region_corner_coordinates_ = _dict_ @property def region_edge_types(self): """ Store the regions' boundaries' types. Returns ------- region_boundary_type : dict A dict that contains the region boundary info. The keys indicate the region boundary, the value indicate the info. value[0] indicate the type, value[1:] indicate the rest info which will be parsed into full information. The not mentioned regions boundaries will be set into default type: ('plane',) Notice that the value will be sent to edge_geometry. And if this info (value[1:]) to be parsed, it will be done there in edge_geometry. And the value is stored in the `edge_geometry.edge_types`. """ return self._region_edge_types_ @region_edge_types.setter def region_edge_types(self, _dict_): assert self.region_corner_coordinates is not None, " <DomainInput> : please first set region_coordinates." assert isinstance(_dict_, dict), " <DomainInput> : region_boundary_type needs to be a dict." for item in _dict_: R, S = item.split('-') assert R in self.region_corner_coordinates and S in ('U', 'D', 'L', 'R'), \ " <DomainInput> : region_edge_type key {} is wrong.".format(item) self._region_edge_types_ = _dict_ def ___PRIVATE_boundary_name_requirement_checker___(self, boundaryRegionSidesDict): """ Requirements: 1). != domain name. 2). Length > 2 3). Can not start with 'R:' (So it must be different from regions names). 4). Only have letters """ for boundary_name in boundaryRegionSidesDict.keys(): assert boundary_name != self.domain_name assert len(boundary_name) > 2, f"boundary_name = {boundary_name} is too short (>2 must)." assert boundary_name[0:2] != 'R:', f"boundary_name = {boundary_name} wrong." assert boundary_name.isalpha(), f"boundary_name = {boundary_name} wrong, boundary_name can only contain letters." @property def boundary_region_edges(self): """ Store the domain boundary information. Returns ------- domain_boundary : dict For example: {'Down': ("Body_center-D", 'Body_back-D', ...), 'West': ("Body_center-R", 'Body_back-R', ...), ......} This means we have domain boundaries: South, West and so on. """ return self._boundary_region_edges_ @boundary_region_edges.setter def boundary_region_edges(self, _dict_): assert self.region_corner_coordinates is not None, " <DomainInput> : please first set region_coordinates." assert isinstance(_dict_, dict), " <DomainInput> : domain_boundary needs to be a dict." self.___PRIVATE_boundary_name_requirement_checker___(_dict_) for boundary_names in _dict_.keys(): assert isinstance(boundary_names, str) and boundary_names != '' and '-' not in boundary_names, \ " <DomainInput> : boundary_names = {} is wrong.".format(boundary_names) assert boundary_names not in self.region_corner_coordinates.keys(), \ " <DomainInput>: boundary_names={} is taken by one of the regions.".format(boundary_names) for item in _dict_: if isinstance(_dict_[item], str): _dict_[item] = (_dict_[item],) if isinstance(_dict_[item], list) or isinstance(_dict_[item], tuple): for item_i in _dict_[item]: R, S = item_i.split('-') assert R in self.region_corner_coordinates and S in ('U', 'D', 'L', 'R'), \ " <DomainInput> : domain_boundary[{}]={} is wrong.".format(item, _dict_[item]) else: raise Exception(" <DomainInput> : boundary_region_edges input value accepts only str, tuple of list.") self._boundary_region_edges_ = _dict_ self._boundary_names_ = list(_dict_.keys()) def ___PRIVATE_periodic_boundary_requirement_checker___(self, pBd): """ Here we only do a simple check. We make sure that the keys are in format of: 0). boundary_name_1=boundary_name_2. 1). A boundary name at most appear in one pair. """ assert isinstance(pBd, dict) bnPOOL = set() for pair in pBd: assert '=' in pair bn1, bn2 = pair.split('=') lengthPOOL = len(bnPOOL) assert bn1 in self._boundary_names_ and bn2 in self._boundary_names_ bnPOOL.add(bn1) bnPOOL.add(bn2) newLengthPOOL = len(bnPOOL) assert newLengthPOOL == lengthPOOL + 2, "Boundary(s) used for multiple periodic pairs!" self._periodic_boundaries_ = bnPOOL @property def periodic_boundary_pairs(self): return self._periodic_boundary_pairs_ @periodic_boundary_pairs.setter def periodic_boundary_pairs(self, pBd): """ """ self.___PRIVATE_periodic_boundary_requirement_checker___(pBd) self._periodic_boundary_pairs_ = pBd @property def periodic_boundaries(self): """(set) Return a set of all boundary names those involved in the periodic boundary setting.""" return self._periodic_boundaries_ @property def periodic_boundaries_involved_regions(self): """The regions that involve periodic boundaries.""" regions = set() for pb in self.periodic_boundaries: region_sides = self.boundary_region_edges[pb] for rs in region_sides: rn = rs.split('-')[0] if rn not in regions: regions.add(rn) return regions @property def region_sequence(self): """ This will fix the sequence of regions by fix their names in property region_names or regions.names. This is very important for numbering. Sometimes, a bad regions sequence can make the numbering wrong. """ return self._region_sequence_ @region_sequence.setter def region_sequence(self, rS: tuple): assert len(rS) == len(self.region_corner_coordinates.keys()) assert all([rSi in self.region_corner_coordinates for rSi in rS]) & \ all([rSi in rS for rSi in self.region_corner_coordinates.keys()]), \ f"region_sequence={rS} has invalid regions name(s)." self._region_sequence_ = rS @property def region_type_wr2_metric(self): return self._region_type_wr2_metric_ @region_type_wr2_metric.setter def region_type_wr2_metric(self, rTwr2M: Union[str, Dict[str, str]]): if isinstance(rTwr2M, str): _D_ = dict() for region_name in self.region_corner_coordinates: _D_[region_name] = rTwr2M rTwr2M = _D_ assert isinstance(rTwr2M, dict), "region_type_wr2_metric needs to be a dictionary." for key in rTwr2M: assert key in self.region_corner_coordinates, f"Region name={key} not valid." self._region_type_wr2_metric_ = rTwr2M # class properties ------------------------- @classproperty def statistic(cls): raise NotImplementedError() @classproperty def random_parameters(cls): raise NotImplementedError()
12,103
3,563
""" Vectors over the symbolic ring. Implements vectors over the symbolic ring. AUTHORS: - Robert Bradshaw (2011-05-25): Added more element-wise simplification methods - Joris Vankerschaver (2011-05-15): Initial version EXAMPLES:: sage: x, y = var('x, y') sage: u = vector([sin(x)^2 + cos(x)^2, log(2*y) + log(3*y)]); u (cos(x)^2 + sin(x)^2, log(3*y) + log(2*y)) sage: type(u) <class 'sage.modules.free_module.FreeModule_ambient_field_with_category.element_class'> sage: u.simplify_full() (1, log(3*y) + log(2*y)) TESTS: Check that the outcome of arithmetic with symbolic vectors is again a symbolic vector (:trac:`11549`):: sage: v = vector(SR, [1, 2]) sage: w = vector(SR, [sin(x), 0]) sage: type(v) <class 'sage.modules.free_module.FreeModule_ambient_field_with_category.element_class'> sage: type(w) <class 'sage.modules.free_module.FreeModule_ambient_field_with_category.element_class'> sage: type(v + w) <class 'sage.modules.free_module.FreeModule_ambient_field_with_category.element_class'> sage: type(-v) <class 'sage.modules.free_module.FreeModule_ambient_field_with_category.element_class'> sage: type(5*w) <class 'sage.modules.free_module.FreeModule_ambient_field_with_category.element_class'> Test pickling/unpickling:: sage: u = vector(SR, [sin(x^2)]) sage: loads(dumps(u)) == u True """ from __future__ import absolute_import #***************************************************************************** # Copyright (C) 2011 Joris Vankerschaver (jv@caltech.edu) # # Distributed under the terms of the GNU General Public License (GPL) # as published by the Free Software Foundation; either version 2 of # the License, or (at your option) any later version. # http://www.gnu.org/licenses/ #***************************************************************************** from . import free_module_element from sage.symbolic.all import Expression def apply_map(phi): """ Returns a function that applies phi to its argument. EXAMPLES:: sage: from sage.modules.vector_symbolic_dense import apply_map sage: v = vector([1,2,3]) sage: f = apply_map(lambda x: x+1) sage: f(v) (2, 3, 4) """ def apply(self, *args, **kwds): """ Generic function used to implement common symbolic operations elementwise as methods of a vector. EXAMPLES:: sage: var('x,y') (x, y) sage: v = vector([sin(x)^2 + cos(x)^2, log(x*y), sin(x/(x^2 + x)), factorial(x+1)/factorial(x)]) sage: v.simplify_trig() (1, log(x*y), sin(1/(x + 1)), factorial(x + 1)/factorial(x)) sage: v.canonicalize_radical() (cos(x)^2 + sin(x)^2, log(x) + log(y), sin(1/(x + 1)), factorial(x + 1)/factorial(x)) sage: v.simplify_rational() (cos(x)^2 + sin(x)^2, log(x*y), sin(1/(x + 1)), factorial(x + 1)/factorial(x)) sage: v.simplify_factorial() (cos(x)^2 + sin(x)^2, log(x*y), sin(x/(x^2 + x)), x + 1) sage: v.simplify_full() (1, log(x*y), sin(1/(x + 1)), x + 1) sage: v = vector([sin(2*x), sin(3*x)]) sage: v.simplify_trig() (2*cos(x)*sin(x), (4*cos(x)^2 - 1)*sin(x)) sage: v.simplify_trig(False) (sin(2*x), sin(3*x)) sage: v.simplify_trig(expand=False) (sin(2*x), sin(3*x)) """ return self.apply_map(lambda x: phi(x, *args, **kwds)) apply.__doc__ += "\nSee Expression." + phi.__name__ + "() for optional arguments." return apply class Vector_symbolic_dense(free_module_element.FreeModuleElement_generic_dense): pass # Add elementwise methods. for method in ['simplify', 'simplify_factorial', 'simplify_log', 'simplify_rational', 'simplify_trig', 'simplify_full', 'trig_expand', 'canonicalize_radical', 'trig_reduce']: setattr(Vector_symbolic_dense, method, apply_map(getattr(Expression, method)))
4,098
1,510
from app import db from app.models.serializer import Serializer class Weather(db.Model, Serializer): id = db.Column(db.Integer, primary_key=True) ip = db.Column(db.String(15), index=True, unique=True) country = db.Column(db.String(80)) flag = db.Column(db.String(512)) town = db.Column(db.String(80)) tendency = db.Column(db.String(80)) wind_speed = db.Column(db.String(20)) temperature_min = db.Column(db.String(20)) temperature_max = db.Column(db.String(20)) temperature = db.Column(db.String(20)) humidity = db.Column(db.String(40)) clouds = db.Column(db.String(80)) def __repr__(self): return '<Weather {} : {}>'.format(self.town, self.temperature) def serialize(self): return Serializer.serialize(self)
784
283
# -*- coding: utf-8 -*- # # Copyright Contributors to the Conu project. # SPDX-License-Identifier: MIT # # TODO: move this line to some generic constants, instead of same in # docker and nspawn CONU_ARTIFACT_TAG = 'CONU.' CONU_IMAGES_STORE = "/opt/conu-nspawn-images/" CONU_NSPAWN_BASEPACKAGES = [ "dnf", "iproute", "dhcp-client", "initscripts", "passwd", "systemd", "rpm", "bash", "shadow-utils", "sssd-client", "util-linux", "libcrypt", "sssd-client", "coreutils", "glibc-all-langpacks", "vim-minimal"] BOOTSTRAP_IMAGE_SIZE_IN_MB = 5000 BOOTSTRAP_FS_UTIL = "mkfs.ext4" BOOTSTRAP_PACKAGER = [ "dnf", "-y", "install", "--nogpgcheck", "--setopt=install_weak_deps=False", "--allowerasing"] DEFAULT_RETRYTIMEOUT = 30 DEFAULT_SLEEP = 1
894
356
import itertools from ..quant_utils import QuantizedValue, QuantizedValueType, attribute_to_kwarg, quantize_nparray from .base_operator import QuantOperatorBase class QDQOperatorBase: def __init__(self, onnx_quantizer, onnx_node): self.quantizer = onnx_quantizer self.node = onnx_node self.disable_qdq_for_node_output = ( True if onnx_node.op_type in onnx_quantizer.op_types_to_exclude_output_quantization else False ) def quantize(self): node = self.node if self.disable_qdq_for_node_output: tensors_to_quantize = node.input else: tensors_to_quantize = itertools.chain(node.input, node.output) for tensor_name in tensors_to_quantize: self.quantizer.quantize_tensor(tensor_name)
805
256
from django.db import models # Create your models here. class Blag(models.Model): name = models.CharField(max_length=128) class BlagEntry(models.Model): blag = models.ForeignKey(Blag) title = models.CharField(max_length=255) body = models.TextField() class Meta: verbose_name_plural = "Entries" class PhotoBlagEntry(models.Model): blag = models.ForeignKey(Blag) title = models.CharField(max_length=255) # and a file field for a photo but whatever
495
166
import bottleneck as bn import numpy as np from scipy import sparse """ All the data should be in the shape of (n_users, n_items) All the latent factors should in the shape of (n_users/n_items, n_components) 1. train_data refers to the data that was used to train the model 2. heldout_data refers to the data that was used for evaluation (could be test set or validation set) 3. vad_data refers to the data that should be excluded as validation set, which should only be used when calculating test scores """ def prec_at_k(train_data, heldout_data, U, V, batch_users=5000, k=20, vad_data=None): n_users = train_data.shape[0] res = list() for user_idx in user_idx_generator(n_users, batch_users): res.append(precision_at_k_batch(train_data, heldout_data, U, V.T, user_idx, k=k, vad_data=vad_data)) mn_prec = np.hstack(res) return np.nanmean(mn_prec) def recall_at_k(train_data, heldout_data, U, V, batch_users=5000, k=20, vad_data=None): n_users = train_data.shape[0] res = list() for user_idx in user_idx_generator(n_users, batch_users): res.append(recall_at_k_batch(train_data, heldout_data, U, V.T, user_idx, k=k, vad_data=vad_data)) mrecall = np.hstack(res) return np.nanmean(mrecall) def ric_rank_at_k(train_data, heldout_data, U, V, batch_users=5000, k=5, vad_data=None): n_users = train_data.shape[0] res = list() for user_idx in user_idx_generator(n_users, batch_users): res.append(mean_rrank_at_k_batch(train_data, heldout_data, U, V.T, user_idx, k=k, vad_data=vad_data)) mrrank = np.hstack(res) #return mrrank[mrrank > 0].mean(), np.median(mrrank[mrrank > 0]) return mrrank[mrrank > 0].mean() def mean_perc_rank(train_data, heldout_data, U, V, batch_users=5000, vad_data=None): n_users = train_data.shape[0] mpr = 0 for user_idx in user_idx_generator(n_users, batch_users): mpr += mean_perc_rank_batch(train_data, heldout_data, U, V.T, user_idx, vad_data=vad_data) mpr /= heldout_data.sum() return mpr def normalized_dcg(train_data, heldout_data, U, V, batch_users=5000, vad_data=None): n_users = train_data.shape[0] res = list() for user_idx in user_idx_generator(n_users, batch_users): res.append(NDCG_binary_batch(train_data, heldout_data, U, V.T, user_idx, vad_data=vad_data)) ndcg = np.hstack(res) return np.nanmean(ndcg) # helper functions # def user_idx_generator(n_users, batch_users): ''' helper function to generate the user index to loop through the dataset ''' for start in xrange(0, n_users, batch_users): end = min(n_users, start + batch_users) yield slice(start, end) def _make_prediction(train_data, Et, Eb, user_idx, batch_users, vad_data=None): n_songs = train_data.shape[1] # exclude examples from training and validation (if any) item_idx = np.zeros((batch_users, n_songs), dtype=bool) item_idx[train_data[user_idx].nonzero()] = True if vad_data is not None: item_idx[vad_data[user_idx].nonzero()] = True X_pred = Et[user_idx].dot(Eb) X_pred[item_idx] = -np.inf return X_pred def precision_at_k_batch(train_data, heldout_data, Et, Eb, user_idx, k=20, normalize=True, vad_data=None): batch_users = user_idx.stop - user_idx.start X_pred = _make_prediction(train_data, Et, Eb, user_idx, batch_users, vad_data=vad_data) idx = bn.argpartsort(-X_pred, k, axis=1) X_pred_binary = np.zeros_like(X_pred, dtype=bool) X_pred_binary[np.tile(np.arange(batch_users), (k, 1)).T, idx[:, :k]] = True X_true_binary = (heldout_data[user_idx] > 0).toarray() tmp = (np.logical_and(X_true_binary, X_pred_binary).sum(axis=1)).astype( np.float32) if normalize: precision = tmp / np.minimum(k, X_true_binary.sum(axis=1)) else: precision = tmp / k return precision def recall_at_k_batch(train_data, heldout_data, Et, Eb, user_idx, k=20, vad_data=None): batch_users = user_idx.stop - user_idx.start X_pred = _make_prediction(train_data, Et, Eb, user_idx, batch_users, vad_data=vad_data) idx = bn.argpartsort(-X_pred, k, axis=1) X_pred_binary = np.zeros_like(X_pred, dtype=bool) X_pred_binary[np.tile(np.arange(batch_users), (k, 1)).T, idx[:, :k]] = True X_true_binary = (heldout_data[user_idx] > 0).toarray() tmp = (np.logical_and(X_true_binary, X_pred_binary).sum(axis=1)).astype( np.float32) recall = tmp / X_true_binary.sum(axis=1) return recall def mean_rrank_at_k_batch(train_data, heldout_data, Et, Eb, user_idx, k=5, vad_data=None): ''' mean reciprocal rank@k: For each user, make predictions and rank for all the items. Then calculate the mean reciprocal rank for the top K that are in the held-out set. ''' batch_users = user_idx.stop - user_idx.start X_pred = _make_prediction(train_data, Et, Eb, user_idx, batch_users, vad_data=vad_data) all_rrank = 1. / (np.argsort(np.argsort(-X_pred, axis=1), axis=1) + 1) X_true_binary = (heldout_data[user_idx] > 0).toarray() heldout_rrank = X_true_binary * all_rrank top_k = bn.partsort(-heldout_rrank, k, axis=1) return -top_k[:, :k].mean(axis=1) def NDCG_binary_batch(train_data, heldout_data, Et, Eb, user_idx, vad_data=None): ''' normalized discounted cumulative gain for binary relevance ''' batch_users = user_idx.stop - user_idx.start n_items = train_data.shape[1] X_pred = _make_prediction(train_data, Et, Eb, user_idx, batch_users, vad_data=vad_data) all_rank = np.argsort(np.argsort(-X_pred, axis=1), axis=1) # build the discount template tp = np.hstack((1, 1. / np.log2(np.arange(2, n_items + 1)))) all_disc = tp[all_rank] X_true_binary = (heldout_data[user_idx] > 0).tocoo() disc = sparse.csr_matrix((all_disc[X_true_binary.row, X_true_binary.col], (X_true_binary.row, X_true_binary.col)), shape=all_disc.shape) DCG = np.array(disc.sum(axis=1)).ravel() IDCG = np.array([tp[:n].sum() for n in heldout_data[user_idx].getnnz(axis=1)]) return DCG / IDCG def mean_perc_rank_batch(train_data, heldout_data, Et, Eb, user_idx, vad_data=None): ''' mean percentile rank for a batch of users MPR of the full set is the sum of batch MPR's divided by the sum of all the feedbacks. (Eq. 8 in Hu et al.) This metric not necessarily constrains the data to be binary ''' batch_users = user_idx.stop - user_idx.start X_pred = _make_prediction(train_data, Et, Eb, user_idx, batch_users, vad_data=vad_data) all_perc = np.argsort(np.argsort(-X_pred, axis=1), axis=1) / \ np.isfinite(X_pred).sum(axis=1, keepdims=True).astype(np.float32) perc_batch = (all_perc[heldout_data[user_idx].nonzero()] * heldout_data[user_idx].data).sum() return perc_batch
7,598
2,788
from pgdrive.component.blocks.curve import Curve from pgdrive.component.blocks.first_block import FirstPGBlock from pgdrive.component.blocks.std_t_intersection import StdTInterSection from pgdrive.component.blocks.straight import Straight from pgdrive.component.road.road_network import RoadNetwork from pgdrive.tests.vis_block.vis_block_base import TestBlock if __name__ == "__main__": test = TestBlock(True) from pgdrive.engine.asset_loader import initialize_asset_loader initialize_asset_loader(test) global_network = RoadNetwork() first = FirstPGBlock(global_network, 3.0, 2, test.render, test.world, 1) curve = Curve(1, first.get_socket(0), global_network, 1) curve.construct_block(test.render, test.world) straight = Straight(2, curve.get_socket(0), global_network, 1) straight.construct_block(test.render, test.world) intersection = StdTInterSection(3, straight.get_socket(0), global_network, 1) print(intersection.construct_block(test.render, test.world)) id = 4 for socket_idx in range(intersection.SOCKET_NUM): block = Curve(id, intersection.get_socket(socket_idx), global_network, id + 1) block.construct_block(test.render, test.world) id += 1 test.show_bounding_box(global_network) test.run()
1,295
420
from selenium import webdriver BASE_URL = 'http://news.163.com/' # 首页ajax内容,无人机ajax,其实后面?callback=data_callback'可以省略 # 如何快速找到json链接?现在只能手工完成..添加每一个json文档 JSON_INDEX_URLS = [ 'http://temp.163.com/special/00804KVA/cm_yaowen.js?callback=data_callback', 'http://house.163.com/special/00078GU7/guangzhou_xw_news_v1.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_shehui.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_guonei.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_guoji.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_dujia.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_war.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_money.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_tech.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_sports.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_ent.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_lady.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_auto.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_houseguangzhou.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_hangkong.js?callback=data_callback', 'http://temp.163.com/special/00804KVA/cm_jiankang.js?callback=data_callback', # 无人机标签 'http://news.163.com/uav/special/000189N0/uav_index.js?callback=data_callback', # 体育 'http://sports.163.com/special/000587PR/newsdata_n_index.js?callback=data_callback', 'http://sports.163.com/special/000587PR/newsdata_n_world.js?callback=data_callback', 'http://sports.163.com/special/000587PR/newsdata_n_china.js?callback=data_callback', 'http://sports.163.com/special/000587PR/newsdata_n_cba.js?callback=data_callback', 'http://sports.163.com/special/000587PR/newsdata_n_allsports.js?callback=data_callback', # NBA 'http://sports.163.com/special/000587PK/newsdata_nba_hj.js?callback=data_callback', 'http://sports.163.com/special/000587PK/newsdata_nba_qsh.js?callback=data_callback', 'http://sports.163.com/special/000587PK/newsdata_nba_ysh.js?callback=data_callback', 'http://sports.163.com/special/000587PK/newsdata_nba_ketr.js?callback=data_callback', 'http://sports.163.com/special/000587PK/newsdata_nba_okc.js?callback=data_callback', 'http://sports.163.com/special/000587PK/newsdata_nba_huren.js?callback=data_callback', 'http://sports.163.com/special/000587PK/newsdata_nba_mc.js?callback=data_callback', # 娱乐 'http://ent.163.com/special/000380VU/newsdata_index.js?callback=data_callback', 'http://ent.163.com/special/000380VU/newsdata_music.js?callback=data_callback', 'http://ent.163.com/special/000380VU/newsdata_star.js?callback=data_callback', 'http://ent.163.com/special/000380VU/newsdata_movie.js?callback=data_callback', 'http://ent.163.com/special/000380VU/newsdata_tv.js?callback=data_callback', 'http://ent.163.com/special/000380VU/newsdata_show.js?callback=data_callback', # 财经 'http://money.163.com/special/002557S5/newsdata_idx_index.js?callback=data_callback', 'http://money.163.com/special/002557S5/newsdata_idx_stock.js?callback=data_callback', 'http://money.163.com/special/002557S5/newsdata_idx_finance.js?callback=data_callback', 'http://money.163.com/special/002557S5/newsdata_idx_fund.js?callback=data_callback', 'http://money.163.com/special/002557S5/newsdata_idx_licai.js?callback=data_callback', 'http://money.163.com/special/002557S5/newsdata_idx_biz.js?callback=data_callback', 'http://money.163.com/special/002557S5/newsdata_idx_bitcoin.js?callback=data_callback', # 股票 'http://money.163.com/special/002557S6/newsdata_gp_index.js?callback=data_callback', 'http://money.163.com/special/002557S6/newsdata_gp_hkstock.js?callback=data_callback', 'http://money.163.com/special/002557S6/newsdata_gp_usstock.js?callback=data_callback', 'http://money.163.com/special/002557S6/newsdata_gp_ipo.js?callback=data_callback', 'http://money.163.com/special/002557S6/newsdata_gp_bitcoin.js?callback=data_callback', 'http://money.163.com/special/002557S6/newsdata_gp_dy.js?callback=data_callback', # 科技 'http://tech.163.com/special/00097UHL/tech_datalist.js?callback=data_callback', # 全国,找不到规律,不想copy了 烦 'http://bendi.news.163.com/beijing/special/04388GGG/bjxinxiliu.js', 'http://bendi.news.163.com/shanghai/special/04188GP4/shxinxiliu.js', 'http://tj.news.163.com/special/04208F5D/tjxxl.js', 'http://bendi.news.163.com/jiangsu/special/04248H8U/njxxl.js', 'http://bendi.news.163.com/zhejiang/special/04098FBT/xinxiliu.js', 'http://sc.news.163.com/special/04268EVT/xinxiliu.js', 'http://bendi.news.163.com/heilongjiang/special/04238DR5/haerbin.js', 'http://bendi.news.163.com/jilin/special/04118E6D/center_news_cc.js', 'http://bendi.news.163.com/liaoning/special/04228EED/xinxiliu.js', 'http://bendi.news.163.com/neimengu/special/04138EHT/nmgxxl.js' ] URLs = ['http://news.163.com/', 'http://news.163.com/rank/', 'http://news.163.com/photo/#Current', 'http://news.163.com/domestic/', 'http://news.163.com/world/', 'http://news.163.com/shehui/', 'http://data.163.com/special/datablog/', 'http://war.163.com/', 'http://news.163.com/air/', 'http://news.163.com/uav/', 'http://news.163.com/college', 'http://gov.163.com/', 'http://gongyi.163.com/', 'http://media.163.com/'] # config mongoDB MONGODB_HOST = 'localhost' MONGODB_PORT = 27017 MONGODB_DBNAME = 'neteasenews' # 存储热更新的内容, 即更新频率快的 MONGODB_TABLE_1 = 'article' # 存储新闻排行榜内容, 即更新频率中等 MONGODB_TABLE_2 = 'newsrank' # 存储冷更新的内容,即更新时间长的 MONGODB_TABLE_3 = 'coldpage' # 存储图片数据 MONGODB_TABLE_4 = 'picture' # config chromedriver: prefs = { 'profile.default_content_setting_values': { 'images': 2, # 'javascript': 2 # 'User-Agent': ua } } options = webdriver.ChromeOptions() options.add_experimental_option('prefs', prefs) # 开启无界面模式 options.add_argument('--headless')
6,241
2,992
#!/usr/bin/env python3 """ Author : Brian Scott <brianscott@email.arizona.edu> Date : 2021-11-09 Purpose: FInd the similarities between sequences. """ import argparse # -------------------------------------------------- def get_args(): """Get command-line arguments""" parser = argparse.ArgumentParser( description='Rock the Casbah', formatter_class=argparse.ArgumentDefaultsHelpFormatter) parser.add_argument('file', help='Input file', metavar='FILE', type=argparse.FileType('rt')) return parser.parse_args() # -------------------------------------------------- def main(): """Makes it work""" args = get_args() sequences = [] for line in args.file: temp_list = [] for char in line.strip(): temp_list.append(char) sequences.append(temp_list) out_seq = "" for base in range(len(sequences[0])): temp_base = [] # for line in range(len(sequences)): # had to use enumerate to get pylint to shut up. # also had to then use the variable so pylint would shut up # because If i use the range(len()) method pylint doesn't like it # even though it makes more sense than enumerate becuase # it doesn't create unnecessary variables# for line, value in enumerate(sequences): temp_base.append(sequences[line][base]) # print(temp_base) all_same(value) # only here to get enumerate to shup up if all_same(temp_base): out_seq = out_seq + "|" else: out_seq = out_seq + "X" for line in sequences: temp_line = "" for char in line: temp_line = temp_line + char print(temp_line) print(out_seq) def all_same(list1): """checks if all items in list are equal""" return all(x == list1[0] for x in list1) # -------------------------------------------------- if __name__ == '__main__': main()
2,044
597
#!/usr/bin/env python # -*- coding: utf8 -*- # Loading a few python packages import os import glob import warnings from astropy import log from astropy.io import fits as pyfits import json # Loading iraf packages from pyraf import iraf from pyraf.iraf import onedspec from pyraf.iraf import twodspec, apextract class SpecRed(): """ This class runs a few methods over a data set defined in an python dictionary. """ def __init__(self, raw_path, red_path, obs_log_file, flat_correction=True, fit_continuum=False): project_dir = os.path.abspath(os.path.dirname(os.path.dirname((os.path.dirname(__file__))))) os.system('export PYTHONPATH={}'.format(project_dir)) self.data_dir = '{}/ldss_spec/data/'.format(project_dir) # Get code path # Getting raw and red paths self.raw_path = str(os.path.join(raw_path[0], '')) self.red_path = str(os.path.join(red_path[0], '')) self.obs_log_file = obs_log_file self.fit_continuum = fit_continuum # Getting observation log dictionary + sky definition self.obs_dict = self.get_obs_log(self.obs_log_file[0]) self.database_dir = str(os.path.join(self.red_path+'database', '')) if fit_continuum: self.fit_continuum = True if flat_correction: self.flat_correction = True ## Getting disp solution dictionary. # Keys are the names of the files inside tye dispsol folder # Values are the names that files will have inside the database folder self.disp_solution_files = {'vph_all_open': 'id_vph_all_open.0001', 'vph_blue_open': 'id_vph_blue_open.0001', 'vph_red_og590': 'id_vph_red_og590.0001'} # Sci extraction parameters self.spec_ext_flags = {'format': 'onedspec', 'interactive': 'yes', 'find': 'yes', 'nfind': '1', 'resize': 'no', 'recenter': 'yes', 'edit': 'yes', 'nsum': '10', 'width': '10', 'lower': '-10', 'upper': '10', 'b_naverage': '-100', 'extract': 'yes', 'extras': 'no', 'trace': 'yes', 'fittrace': 'yes', 't_function': 'chebyshev', 't_order': '3', 't_nsum': '15', 't_step': '15', 'background': 'fit', 'b_function': 'chebyshev', 'b_order': '1', 'clean': 'yes', 'readnoise': 'enoise', 'gain': 'egain', 'weights': 'variance'} # Arc extraction parameters self.arc_ext_flags = {'format': 'onedspec', 'interactive': 'no', 'find': 'no', 'resize': 'no', 'recenter': 'no', 'edit': 'no', 'lower': '-10.0', 'upper': '10.0', 'extract': 'yes', 'extras': 'no', 'trace': 'no', 'fittrace': 'no', 'background': 'none', 'review': 'yes'} # Identify parameters self.identify_flags = {'ftype': 'emission', 'fwidth': '40', 'cradius': '30', 'thresho': '10.', 'minsep': '10.'} self.wrange = { 'vph_all': [3900.0, 10300.0], 'vph_red': [6050.0, 10300.0], 'vph_blue': [3850.0, 6580.0] } # About warnings warnings.filterwarnings('ignore') log.propagate = False def run(self, *args, **kwargs): # cleaning up the reduction dir self.clean_path(path=self.red_path, ext_list=['fits', 'log', 'txt', 'dat']) # Loading dispsol content dictionary self.disp_sol_content_dictionary = self.get_dict_of_dispsol_content(self.disp_solution_files, dispsol_dirname='dispsol/') # Changing to red_path dir self.change_dir(path=self.red_path) # Creating directory database if does not exist if not os.path.exists(self.database_dir): os.makedirs(str(self.database_dir)) # Extracting spectra and corresponding arcs self.extract_spec(path=self.raw_path, obs_dict=self.obs_dict, out_prefix='e_') # Obtaining and applying dispersion function self.wcalibrate_spec(obs_dict=self.obs_dict, disp_sol_content_dict=self.disp_sol_content_dictionary, prefix='e_', out_prefix='w') @staticmethod def get_obs_log(filename): """ It reads a json file and return it as a python dictionary. Args: filename (str): The filename of a json file with the obslog definitions. Returns: obs_log_dict (dict): A python dictionary containing the obs log definitions. """ filename = str(filename) with open(filename, 'r') as f: obs_log_dict = json.load(f) #except FileNotFoundError as file_not_found_message: # print(file_not_found_message) return obs_log_dict def get_spec_setup(self, file): hdu = pyfits.open(file) header = hdu[0].header hdu.close() # Getting spec setup grism = header['GRISM'].replace('-', '_').lower() filter = header['FILTER'].lower() slit = header['APERTURE'].lower() return slit, grism, filter def check_spec_setup(self, sci_file, arc_file): sci_slit, sci_grism, sci_filter = self.get_spec_setup(sci_file) arc_slit, arc_grism, arc_filter = self.get_spec_setup(arc_file) # Check Slits message = 'SLIT {} from SCI file {} does not match SLIT {} from ARC file {}'.format(sci_slit.upper(), sci_file, arc_slit.upper(), arc_file) assert sci_slit == arc_slit, message # Check Grisms message = 'GRISM {} from SCI file {} does not match SLIT {} from ARC file {}'.format(sci_grism.upper(), sci_file, arc_grism.upper(), arc_file) assert sci_grism == arc_grism, message # Check Filter if sci_grism != 'vph_all': message = 'FILTER {} from SCI file {} does not match FILTER {} from ARC file {}'.format(sci_filter.upper(), sci_file, arc_filter.upper(), arc_file) assert sci_filter == arc_filter, message @staticmethod def get_dispsol_content(fname): # Getting disp sol content from input disp file with open(fname, 'r') as f: content = f.read() f.close() return content def get_dict_of_dispsol_content(self, disp_solutions_files_dictionary, dispsol_dirname='dispsol'): data_dir = '{}{}'.format(self.data_dir, dispsol_dirname) keys = disp_solutions_files_dictionary.keys() values = disp_solutions_files_dictionary.values() content_dict = {} for i, key in enumerate(keys): content = self.get_dispsol_content(fname=data_dir + values[i]) content_dict[key] = content return content_dict def write_modified_disp_sol_to_database(self, arc_file, dispsol_content, database_dir): # Getting arc_file setup slit, grism, filter = self.get_spec_setup(arc_file) # This is the file name format as given in the dispsol directory, which is based on the # grism - filter combination inp_disp_file = 'id_{}_{}.0001'.format(grism, filter) # Output file name with the dispersion solution (as required by IRAF) output_disp_file_name = 'id{}'.format(arc_file.replace('.fits', '')) # Correspoing txt inside the dispersion solution file txt_to_be_replaced = str(inp_disp_file) txt_new = output_disp_file_name.replace('id', '') modified_disp_content = dispsol_content.replace(txt_to_be_replaced, txt_new) # Writting new content to out_disp_file in the self.database_dir with open(database_dir + output_disp_file_name, 'w') as f: print >> f, modified_disp_content f.close() @staticmethod def change_dir(path): """ Change to directory defined in the path variable. Args: path (str): absolute or relative path. """ # Changing system to path if os.getcwd() is not path: os.chdir(path) @staticmethod def clean_path(path, ext_list): """ Clean up files from path. If the path does not exist, it will be created. It's not recursive, so it will not clean up a sub-folder within the path. Args: path (string): path to be cleaned. ext_list (list): list of strings containing the extension of files to be deleted. """ path = str(os.path.join(path, '')) if not os.path.exists(path): log.info('Provided path does not exist, but it was created!') os.makedirs(path) elif os.path.exists(path): log.info('Cleaning up ' + path + '.') iraf.imdelete('*tmp*', verify='no') for ext in ext_list: for _file in glob.glob(os.path.join(path, '*.' + str(ext))): os.remove(_file) def extract_spec(self, path, obs_dict, out_prefix='e_'): """ This method runs *apall* over the data set inside the path folder as defined by the obs_dict dictionary. Args: path (string): path that contaings the data to be extracted with apall. obs_dict (dict): dictionary containing the filename of the spectra to be extracted, the corresponding arc file and the sky background definition. The dictionary should be something like this: self.obs_dict = {'LTTT1788_1': {'spec': 'ccd0093c1', 'arc': 'ccd0095c1', 'sky': '-35:-20,20:35'}, 'LTTT1788_2': {'spec': 'ccd0094c1', 'arc': 'ccd0095c1', 'sky': '-35:-20,20:35'}, 'LTTT1788_3': {'spec': 'ccd0096c1', 'arc': 'ccd0098c1', 'sky': '-35:-20,20:35'}, 'LTTT1788_4': {'spec': 'ccd0097c1', 'arc': 'ccd0098c1', 'sky': '-35:-20,20:35'}} out_prefix (string): letter that will be added to the filename of the data extracted. Returns: It returns 1d extracted spectra. """ path = str(os.path.join(path, '')) # Set task parameters. twodspec.apextract.apall.unlearn() # Make sure that in "apextract" the dispaxis parameter is set to "2". If the commando below does not work, # do it manually. twodspec.apextract(dispaxis=2) for target, p in obs_dict.iteritems(): spec_in = p['spec'] + '.fits' arc_in = p['arc'] + '.fits' if len(obs_dict) > 0: print('\n') log.info('Extracting star spectrum ' + p['spec'] + '...') apextract.apall(input=path + spec_in, output=out_prefix + spec_in, b_sample=p['sky'], **self.spec_ext_flags) print('\n') log.info('Extracting arc spectrum ' + p['arc'] + '...') apextract.apall(input=path + arc_in, output=out_prefix + arc_in, reference=path + spec_in, **self.arc_ext_flags) def wcalibrate_spec(self, obs_dict, disp_sol_content_dict, prefix='e_', out_prefix='w'): """ This method runs a couple of IRAF routines to obtain the dispersion function for a arc files and then apply it to the corresponding spectra. Args: obs_dict (dict): prefix (str): out_prefix (str): Returns: """ onedspec.identify.unlearn() onedspec.refspec.unlearn() onedspec.dispcor.unlearn() for target, p in obs_dict.iteritems(): spec_in = prefix + p['spec'] + '.0001.fits' arc_ref = prefix + p['arc'] + '.0001.fits' # Checking sepctral setup self.check_spec_setup(spec_in, arc_ref) ##### Copying disp solution to 'database' dir # 1. Getting content dictionary with disp solutions of the corresponding arc slit, grism, filter = self.get_spec_setup(arc_ref) w1, w2 = self.wrange[grism][0], self.wrange[grism][1] # Getting specific disp sol content of the corresponding arc file key = '{}_{}'.format(grism,filter) content_dict = disp_sol_content_dict[key] # 2. Writting solution to database dir self.write_modified_disp_sol_to_database(arc_ref, content_dict, database_dir=self.database_dir) ##### Running iraf to get updated disp sol print('\n') log.info('Finding wavelength solution to reference arc ' + arc_ref + '...') onedspec.identify(arc_ref, **self.identify_flags) print('\n') log.info('Associating the obtained wavelength solution with the spectrum of the star:') log.info(spec_in + ' -----> REFSPEC = ' + arc_ref + '.') onedspec.refspec(spec_in, reference=arc_ref, sort='', group='') print('\n') log.info('Applying wavelength calibration to ' + spec_in + '.') onedspec.dispcor(spec_in, out=out_prefix + spec_in, w1=w1, w2=w2) if self.fit_continuum: onedspec.continuum.unlearn() print('\n') log.info('Fitting continuum to ' + out_prefix + spec_in + '.') input = out_prefix + spec_in output = 'cont_' + out_prefix + spec_in onedspec.continuum(input=input, output=output, type='fit', function='legendre', order=15, niterate=10, low_reject=2.0, high_reject=0.0)
13,999
4,379
from ..models import Job from engine.scripts.mirmachine_args import run_mirmachine from .socket_helper import announce_status_change, announce_queue_position, announce_initiation, announce_completed from .maintainer import clean_up_temporary_files from django.utils import timezone from MirMachineWebapp import user_config as config def schedule_job(stop): ongoing = Job.objects.filter(status='ongoing') # check if already job running if ongoing.exists(): return queued = Job.objects.filter(status='queued').order_by('initiated') # check if queue is empty if not queued.exists(): if config.AUTO_CLEANUP_TEMP_FILES: clean_up_temporary_files() return next_in_line = queued[0] next_in_line.status = 'ongoing' next_in_line.initiated = timezone.now() next_in_line.save() announce_status_change(next_in_line) announce_initiation(next_in_line) for i in range(len(queued)): announce_queue_position(queued[i], i+1) try: process, job_object = run_mirmachine(next_in_line, stop) handle_job_end(process, job_object) except OSError: next_in_line.status = 'halted' next_in_line.save() announce_status_change(next_in_line) except RuntimeError: print('Interrupted, exiting thread') return schedule_job(stop) def handle_job_end(process, job_object): if process.returncode != 0: job_object.status = 'halted' else: job_object.status = 'completed' job_object.completed = timezone.now() job_object.save() announce_completed(job_object) announce_status_change(job_object)
1,667
541
""" 33 - Faca um programa que leia um vetor de 15 posicoes e o compacte, ou seja, elimine as posicoes com valor zero. Para isso, todos os elementos a frente do valor zero, devem ser movidos uma posicao para tras no vetor. """ """ vetor = [] count = 0 for x in range(1, 16): vetor.append(int(input(f'Digite o {x}/15: '))) n = len(vetor) for i in range(n): if vetor[i] != 0: vetor[count] = vetor[i] count += 1 while n > count: vetor[count] = 0 count += 1 print(vetor) # [5, 6, 9, 8, 10, 15, 33, 22, 66, 99, 10, 100, 0, 0, 0] Este os zeros vao para tras """ from itertools import compress, repeat, chain vetor = [] for x in range(1, 16): vetor.append(int(input(f'Digite o {x}/15: '))) # usando list.count e itertools.compress y = [0] * vetor.count(0) y.extend(compress(vetor, vetor)) print(y)
836
384
import pygame import my_colors as color pygame.init() screen_width = 800 screen_height = 600 screen_size = (screen_width, screen_height) screen = pygame.display.set_mode(screen_size) pygame.display.set_caption("Search the green pixel at the coordinates (x=10, y=100)") running = True while running: screen.set_at((1, 1), color.white) screen.set_at((10, 100), color.green) pygame.display.update() event = pygame.event.wait() if event.type == pygame.QUIT: running = False pygame.quit() print("Goodbye!")
533
189