commit
stringlengths 40
40
| old_file
stringlengths 4
236
| new_file
stringlengths 4
236
| old_contents
stringlengths 1
3.26k
| new_contents
stringlengths 16
4.43k
| subject
stringlengths 16
624
| message
stringlengths 17
3.29k
| lang
stringclasses 5
values | license
stringclasses 13
values | repos
stringlengths 5
91.5k
|
|---|---|---|---|---|---|---|---|---|---|
77f8e99ca67489caa75aceb76f79fd5a5d32ded8
|
setup.py
|
setup.py
|
from distutils.core import setup
import re
def get_version():
init_py = open('pykka/__init__.py').read()
metadata = dict(re.findall("__([a-z]+)__ = '([^']+)'", init_py))
return metadata['version']
setup(
name='Pykka',
version=get_version(),
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://pykka.readthedocs.org/',
license='Apache License, Version 2.0',
description='Pykka is a Python implementation of the actor model',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries',
],
)
|
from distutils.core import setup
import re
def get_version():
init_py = open('pykka/__init__.py').read()
metadata = dict(re.findall("__([a-z]+)__ = '([^']+)'", init_py))
return metadata['version']
setup(
name='Pykka',
version=get_version(),
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://pykka.readthedocs.org/',
license='Apache License, Version 2.0',
description='Pykka is a Python implementation of the actor model',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries',
],
)
|
Add more Python version/implementation classifiers
|
pypi: Add more Python version/implementation classifiers
|
Python
|
apache-2.0
|
jodal/pykka,tamland/pykka,tempbottle/pykka
|
8e3abcd310b7e932d769f05fa0a7135cc1a53b76
|
setup.py
|
setup.py
|
from cx_Freeze import setup, Executable
# Dependencies are automatically detected, but it might need
# fine tuning.
build_exe_options = {
"excludes": [
"numpy"
],
"bin_includes": [
"libcrypto.so.1.0.0",
"libssl.so.1.0.0"
],
"packages": [
"_cffi_backend",
"appdirs",
"asyncio",
"bcrypt",
"encodings",
"idna",
"motor",
"packaging",
"raven",
"uvloop"
]
}
options = {
"build_exe": build_exe_options
}
executables = [
Executable('run.py', base="Console")
]
setup(name='virtool', executables=executables, options=options)
|
from cx_Freeze import setup, Executable
# Dependencies are automatically detected, but it might need
# fine tuning.
build_exe_options = {
"bin_includes": [
"libcrypto.so.1.0.0",
"libssl.so.1.0.0"
],
"includes": [
"numpy",
"numpy.core._methods",
"numpy.lib",
"numpy.lib.format"
],
"packages": [
"_cffi_backend",
"appdirs",
"asyncio",
"bcrypt",
"encodings",
"idna",
"motor",
"packaging",
"raven",
"uvloop"
]
}
options = {
"build_exe": build_exe_options
}
executables = [
Executable('run.py', base="Console")
]
setup(name='virtool', executables=executables, options=options)
|
Include missing numpy modules in build
|
Include missing numpy modules in build
|
Python
|
mit
|
igboyes/virtool,virtool/virtool,igboyes/virtool,virtool/virtool
|
4c90264d744b177aabcaa1cecba4fe17e30cf308
|
corehq/apps/accounting/migrations/0026_auto_20180508_1956.py
|
corehq/apps/accounting/migrations/0026_auto_20180508_1956.py
|
# -*- coding: utf-8 -*-
# Generated by Django 1.11.13 on 2018-05-08 19:56
from __future__ import unicode_literals
from __future__ import absolute_import
from django.db import migrations
from corehq.sql_db.operations import HqRunPython
def _convert_emailed_to_array_field(apps, schema_editor):
BillingRecord = apps.get_model('accounting', 'BillingRecord')
for record in BillingRecord.objects.all():
if record.emailed_to != '':
record.emailed_to_list = record.emailed_to.split(',')
WireBillingRecord = apps.get_model('accounting', 'WireBillingRecord')
for wirerecord in WireBillingRecord.objects.all():
if wirerecord.emailed_to != '':
wirerecord.emailed_to_list = wirerecord.emailed_to.split(',')
class Migration(migrations.Migration):
dependencies = [
('accounting', '0025_auto_20180508_1952'),
]
operations = [
HqRunPython(_convert_emailed_to_array_field)
]
|
# -*- coding: utf-8 -*-
# Generated by Django 1.11.13 on 2018-05-08 19:56
from __future__ import unicode_literals
from __future__ import absolute_import
from django.db import migrations
from corehq.sql_db.operations import HqRunPython
def noop(*args, **kwargs):
pass
def _convert_emailed_to_array_field(apps, schema_editor):
BillingRecord = apps.get_model('accounting', 'BillingRecord')
for record in BillingRecord.objects.all():
if record.emailed_to != '':
record.emailed_to_list = record.emailed_to.split(',')
WireBillingRecord = apps.get_model('accounting', 'WireBillingRecord')
for wirerecord in WireBillingRecord.objects.all():
if wirerecord.emailed_to != '':
wirerecord.emailed_to_list = wirerecord.emailed_to.split(',')
class Migration(migrations.Migration):
dependencies = [
('accounting', '0025_auto_20180508_1952'),
]
operations = [
HqRunPython(_convert_emailed_to_array_field, reverse_code=noop)
]
|
Add noop to migration file
|
Add noop to migration file
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
49f1715067df0208c79a1af2e73d6aa314b96bef
|
django_su/utils.py
|
django_su/utils.py
|
# -*- coding: utf-8 -*-
import warnings
import collections
from django.conf import settings
from django.utils.module_loading import import_string
def su_login_callback(user):
if hasattr(settings, 'SU_LOGIN'):
warnings.warn(
"SU_LOGIN is deprecated, use SU_LOGIN_CALLBACK",
DeprecationWarning,
)
func = getattr(settings, 'SU_LOGIN_CALLBACK', None)
if func is not None:
if not isinstance(func, collections.Callable):
func = import_string(func)
return func(user)
return user.has_perm('auth.change_user')
def custom_login_action(request, user):
func = getattr(settings, 'SU_CUSTOM_LOGIN_ACTION', None)
if func is None:
return False
if not isinstance(func, collections.Callable):
func = import_string(func)
func(request, user)
return True
|
# -*- coding: utf-8 -*-
import warnings
from collections.abc import Callable
from django.conf import settings
from django.utils.module_loading import import_string
def su_login_callback(user):
if hasattr(settings, 'SU_LOGIN'):
warnings.warn(
"SU_LOGIN is deprecated, use SU_LOGIN_CALLBACK",
DeprecationWarning,
)
func = getattr(settings, 'SU_LOGIN_CALLBACK', None)
if func is not None:
if not isinstance(func, Callable):
func = import_string(func)
return func(user)
return user.has_perm('auth.change_user')
def custom_login_action(request, user):
func = getattr(settings, 'SU_CUSTOM_LOGIN_ACTION', None)
if func is None:
return False
if not isinstance(func, Callable):
func = import_string(func)
func(request, user)
return True
|
Update collections.Callable typecheck to collections.abc.Callable
|
Update collections.Callable typecheck to collections.abc.Callable
|
Python
|
mit
|
adamcharnock/django-su,PetrDlouhy/django-su,PetrDlouhy/django-su,adamcharnock/django-su
|
a45c79b10ef5ca6eb4b4e792f2229b2f9b0a7bbf
|
thinglang/foundation/definitions.py
|
thinglang/foundation/definitions.py
|
import itertools
from thinglang.lexer.values.identifier import Identifier
"""
The internal ordering of core types used by the compiler and runtime
"""
INTERNAL_TYPE_COUNTER = itertools.count(1)
# TODO: map dynamically at runtime
INTERNAL_TYPE_ORDERING = {
Identifier("text"): next(INTERNAL_TYPE_COUNTER),
Identifier("number"): next(INTERNAL_TYPE_COUNTER),
Identifier("bool"): next(INTERNAL_TYPE_COUNTER),
Identifier("list"): next(INTERNAL_TYPE_COUNTER),
Identifier("map"): next(INTERNAL_TYPE_COUNTER),
Identifier("iterator"): next(INTERNAL_TYPE_COUNTER),
Identifier("Console"): next(INTERNAL_TYPE_COUNTER),
Identifier("File"): next(INTERNAL_TYPE_COUNTER),
Identifier("Directory"): next(INTERNAL_TYPE_COUNTER),
Identifier("Time"): next(INTERNAL_TYPE_COUNTER),
Identifier("Exception"): next(INTERNAL_TYPE_COUNTER)
}
|
import glob
import os
from thinglang.lexer.values.identifier import Identifier
"""
The internal ordering of core types used by the compiler and runtime
"""
CURRENT_PATH = os.path.dirname(os.path.abspath(__file__))
SOURCE_PATTERN = os.path.join(CURRENT_PATH, 'source/**/*.thing')
def list_types():
for path in glob.glob(SOURCE_PATTERN, recursive=True):
name = os.path.basename(path).replace('.thing', '')
yield name, path
PRIMITIVE_TYPES = [
'text',
'number'
]
INTERNAL_SOURCES = {Identifier(name): path for name, path in list_types()}
|
Remove manual INTERNAL_TYPE_ORDERING map in favor of explicit import tables
|
Remove manual INTERNAL_TYPE_ORDERING map in favor of explicit import tables
|
Python
|
mit
|
ytanay/thinglang,ytanay/thinglang,ytanay/thinglang,ytanay/thinglang
|
3b2fae7875d89adb8537b75c7e9b48a8663a9d4f
|
src/rnaseq_lib/web/synapse.py
|
src/rnaseq_lib/web/synapse.py
|
import os
from synapseclient import Synapse, File
expression = 'syn11311347'
metadata = 'syn11311931'
def upload_file(file_path, login, parent, description=None):
"""
Uploads file to Synapse. Password must be stored in environment variable SYNAPSE_PASS
:param str file_path: Path to file
:param str login: Login (usually an email address)
:param str parent: Parent Synapse ID (example: syn12312415) where file will be placed
:param str description: Optional description to add
"""
description = '' if None else description
f = File(file_path, description=description, parent=parent)
assert 'SYNAPSE_PASS' in os.environ, 'SYNAPSE_PASS must be set as an environment variable'
syn = Synapse()
syn.login(login, os.environ['SYNAPSE_PASS'])
syn.store(f)
|
import os
from synapseclient import Synapse, File
expression = 'syn11311347'
metadata = 'syn11311931'
def upload_file(file_path, login, parent, description=None):
"""
Uploads file to Synapse. Password must be stored in environment variable SYNAPSE_PASS
:param str file_path: Path to file
:param str login: Login (usually an email address)
:param str parent: Parent Synapse ID (example: syn12312415) where file will be placed
:param str description: Optional description to add
"""
description = '' if None else description
f = File(file_path, description=description, parent=parent)
syn = _syn_login(login)
syn.store(f)
def download_file(synid, login, download_location='.'):
"""
Synapse ID of file to download
:param str synid: Synapse ID
:param str login: Synapse ID
:param str download_location: Download location for file
"""
syn = _syn_login(login)
syn.get(synid, downloadLocation=download_location)
def _syn_login(login):
"""
Login to synapse. Set environment variable SYNAPSE_PASS to the password for `login`
:param str login:
:return: Synapse instance
:rtype: instance
"""
assert 'SYNAPSE_PASS' in os.environ, 'SYNAPSE_PASS must be set as an environment variable'
syn = Synapse()
syn.login(login, os.environ['SYNAPSE_PASS'])
return syn
|
Add download and login functions
|
Add download and login functions
|
Python
|
mit
|
jvivian/rnaseq-lib,jvivian/rnaseq-lib
|
b7335f5c011d9fad3570a097fb1165cc6fbd3cef
|
src/python/grpcio_tests/tests/unit/_logging_test.py
|
src/python/grpcio_tests/tests/unit/_logging_test.py
|
# Copyright 2018 gRPC authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Test of gRPC Python's interaction with the python logging module"""
import unittest
import six
import grpc
import logging
class LoggingTest(unittest.TestCase):
def test_logger_not_occupied(self):
self.assertEqual(0, len(logging.getLogger().handlers))
if __name__ == '__main__':
unittest.main(verbosity=2)
|
# Copyright 2018 gRPC authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Test of gRPC Python's interaction with the python logging module"""
import unittest
import six
from six.moves import reload_module
import logging
import grpc
import functools
import sys
class LoggingTest(unittest.TestCase):
def test_logger_not_occupied(self):
self.assertEqual(0, len(logging.getLogger().handlers))
def test_handler_found(self):
old_stderr = sys.stderr
sys.stderr = six.StringIO()
try:
reload_module(logging)
logging.basicConfig()
reload_module(grpc)
self.assertFalse("No handlers could be found" in sys.stderr.getvalue())
finally:
sys.stderr = old_stderr
reload_module(logging)
if __name__ == '__main__':
unittest.main(verbosity=2)
|
Add test for 'No handlers could be found' problem
|
Add test for 'No handlers could be found' problem
|
Python
|
apache-2.0
|
mehrdada/grpc,sreecha/grpc,stanley-cheung/grpc,vjpai/grpc,mehrdada/grpc,muxi/grpc,pszemus/grpc,stanley-cheung/grpc,jtattermusch/grpc,donnadionne/grpc,grpc/grpc,mehrdada/grpc,pszemus/grpc,ctiller/grpc,nicolasnoble/grpc,firebase/grpc,donnadionne/grpc,ctiller/grpc,jtattermusch/grpc,donnadionne/grpc,vjpai/grpc,muxi/grpc,donnadionne/grpc,grpc/grpc,vjpai/grpc,carl-mastrangelo/grpc,carl-mastrangelo/grpc,stanley-cheung/grpc,ctiller/grpc,jboeuf/grpc,donnadionne/grpc,ejona86/grpc,jboeuf/grpc,carl-mastrangelo/grpc,carl-mastrangelo/grpc,ctiller/grpc,nicolasnoble/grpc,grpc/grpc,pszemus/grpc,stanley-cheung/grpc,sreecha/grpc,jtattermusch/grpc,stanley-cheung/grpc,ctiller/grpc,nicolasnoble/grpc,nicolasnoble/grpc,nicolasnoble/grpc,grpc/grpc,pszemus/grpc,carl-mastrangelo/grpc,mehrdada/grpc,nicolasnoble/grpc,grpc/grpc,jtattermusch/grpc,pszemus/grpc,muxi/grpc,carl-mastrangelo/grpc,ctiller/grpc,vjpai/grpc,grpc/grpc,ctiller/grpc,jtattermusch/grpc,sreecha/grpc,vjpai/grpc,firebase/grpc,donnadionne/grpc,sreecha/grpc,donnadionne/grpc,muxi/grpc,grpc/grpc,muxi/grpc,sreecha/grpc,pszemus/grpc,vjpai/grpc,firebase/grpc,grpc/grpc,jboeuf/grpc,jboeuf/grpc,carl-mastrangelo/grpc,firebase/grpc,ejona86/grpc,pszemus/grpc,ejona86/grpc,stanley-cheung/grpc,stanley-cheung/grpc,ejona86/grpc,vjpai/grpc,ejona86/grpc,vjpai/grpc,vjpai/grpc,mehrdada/grpc,pszemus/grpc,muxi/grpc,jtattermusch/grpc,jtattermusch/grpc,stanley-cheung/grpc,ctiller/grpc,mehrdada/grpc,ctiller/grpc,grpc/grpc,ejona86/grpc,pszemus/grpc,jtattermusch/grpc,firebase/grpc,ejona86/grpc,firebase/grpc,nicolasnoble/grpc,firebase/grpc,ejona86/grpc,nicolasnoble/grpc,mehrdada/grpc,firebase/grpc,donnadionne/grpc,stanley-cheung/grpc,pszemus/grpc,jboeuf/grpc,donnadionne/grpc,vjpai/grpc,donnadionne/grpc,mehrdada/grpc,ctiller/grpc,muxi/grpc,vjpai/grpc,pszemus/grpc,stanley-cheung/grpc,jboeuf/grpc,mehrdada/grpc,carl-mastrangelo/grpc,jtattermusch/grpc,carl-mastrangelo/grpc,mehrdada/grpc,muxi/grpc,jboeuf/grpc,ctiller/grpc,mehrdada/grpc,nicolasnoble/grpc,carl-mastrangelo/grpc,ejona86/grpc,ejona86/grpc,mehrdada/grpc,muxi/grpc,muxi/grpc,pszemus/grpc,donnadionne/grpc,nicolasnoble/grpc,sreecha/grpc,jboeuf/grpc,sreecha/grpc,carl-mastrangelo/grpc,jtattermusch/grpc,donnadionne/grpc,ctiller/grpc,firebase/grpc,vjpai/grpc,carl-mastrangelo/grpc,jboeuf/grpc,firebase/grpc,jtattermusch/grpc,jtattermusch/grpc,muxi/grpc,grpc/grpc,sreecha/grpc,sreecha/grpc,ejona86/grpc,grpc/grpc,sreecha/grpc,stanley-cheung/grpc,firebase/grpc,muxi/grpc,stanley-cheung/grpc,jboeuf/grpc,jboeuf/grpc,sreecha/grpc,nicolasnoble/grpc,grpc/grpc,firebase/grpc,sreecha/grpc,ejona86/grpc,nicolasnoble/grpc,jboeuf/grpc
|
62d9fdfe0ad3fc37286aa19a87e2890aaf90f639
|
tasks/check_rd2_enablement.py
|
tasks/check_rd2_enablement.py
|
import simple_salesforce
from cumulusci.tasks.salesforce import BaseSalesforceApiTask
class is_rd2_enabled(BaseSalesforceApiTask):
def _run_task(self):
try:
settings = self.sf.query(
"SELECT IsRecurringDonations2Enabled__c "
"FROM npe03__Recurring_Donations_Settings__c "
"WHERE SetupOwnerId IN (SELECT Id FROM Organization)"
)
except simple_salesforce.exceptions.SalesforceMalformedRequest:
# The field does not exist in the target org, meaning it's
# pre-RD2
self.return_values = False
return
if settings.get("records"):
if settings["records"][0]["IsRecurringDonations2Enabled__c"]:
self.return_values = True
self.return_values = False
|
import simple_salesforce
from cumulusci.tasks.salesforce import BaseSalesforceApiTask
class is_rd2_enabled(BaseSalesforceApiTask):
def _run_task(self):
try:
settings = self.sf.query(
"SELECT IsRecurringDonations2Enabled__c "
"FROM npe03__Recurring_Donations_Settings__c "
"WHERE SetupOwnerId IN (SELECT Id FROM Organization)"
)
except simple_salesforce.exceptions.SalesforceMalformedRequest:
# The field does not exist in the target org, meaning it's
# pre-RD2
self.return_values = False
return
if settings.get("records"):
if settings["records"][0]["IsRecurringDonations2Enabled__c"]:
self.return_values = True
return
self.return_values = False
|
Correct bug in preflight check
|
Correct bug in preflight check
|
Python
|
bsd-3-clause
|
SalesforceFoundation/Cumulus,SalesforceFoundation/Cumulus,SalesforceFoundation/Cumulus,SalesforceFoundation/Cumulus
|
7cbd21a050a9e94d0f8f1f5c3ce4f81c812e279c
|
trump/templating/tests/test_templates.py
|
trump/templating/tests/test_templates.py
|
from ..templates import QuandlFT
class TestTemplates(object):
def test_quandl_ft(self):
ftemp = QuandlFT("xxx", trim_start="yyyy-mm-dd", authtoken="yyy")
assert ftemp.sourcing == {'authtoken': 'yyy',
'trim_start': 'yyyy-mm-dd',
'dataset': 'xxx'}
|
from ..templates import QuandlFT, QuandlSecureFT, GoogleFinanceFT
class TestTemplates(object):
def test_quandl_ft(self):
ftemp = QuandlFT("xxx", trim_start="yyyy-mm-dd", authtoken="yyy")
assert ftemp.sourcing == {'authtoken': 'yyy',
'trim_start': 'yyyy-mm-dd',
'dataset': 'xxx'}
def test_quandl_secure_ft(self):
ftemp = QuandlSecureFT("xxx", trim_start="yyyy-mm-dd")
assert ftemp.sourcing == {'trim_start': 'yyyy-mm-dd',
'dataset': 'xxx'}
assert ftemp.meta == {'sourcing_key' : 'userone',
'stype' : 'Quandl'}
def test_google_finance_ft(self):
ftemp = GoogleFinanceFT("xxx")
assert ftemp.sourcing == {'name': 'xxx',
'start': '2000-01-01,
'end': 'now',
'data_source' : 'google',
'data_column' : 'Close'}
assert ftemp.meta == {'stype' : 'PyDataDataReaderST'}
|
Add two tests for templates
|
Add two tests for templates
|
Python
|
bsd-3-clause
|
Equitable/trump,Asiant/trump,jnmclarty/trump
|
ea17a76c4ada65dac9e909b930c938a24ddb99b2
|
tests/formatter/test_csver.py
|
tests/formatter/test_csver.py
|
import unittest, argparse
from echolalia.formatter.csver import Formatter
class CsverTestCase(unittest.TestCase):
def setUp(self):
self.parser = argparse.ArgumentParser()
self.data = [{'char': chr(i), 'order': i - 96} for i in xrange(97, 100)]
self.formatter = Formatter()
def test_add_args(self):
new_parser = self.formatter.add_args(self.parser)
self.assertEqual(new_parser, self.parser)
args = new_parser.parse_args(['--with_header'])
self.assertTrue(args.with_header)
args = new_parser.parse_args([])
self.assertFalse(args.with_header)
def test_marshall_no_header(self):
new_parser = self.formatter.add_args(self.parser)
args = new_parser.parse_args([])
result = self.formatter.marshall(args, self.data)
expect = "a,1\r\nb,2\r\nc,3\r\n"
def test_marshall_with_header(self):
new_parser = self.formatter.add_args(self.parser)
args = new_parser.parse_args(['--with_header'])
result = self.formatter.marshall(args, self.data)
expect = "char,order\r\na,1\r\nb,2\r\nc,3\r\n"
self.assertEqual(result, expect)
|
import unittest, argparse
from echolalia.formatter.csver import Formatter
class CsverTestCase(unittest.TestCase):
def setUp(self):
self.parser = argparse.ArgumentParser()
self.data = [{'char': chr(i), 'order': i - 96} for i in xrange(97, 100)]
self.formatter = Formatter()
def test_add_args(self):
new_parser = self.formatter.add_args(self.parser)
self.assertEqual(new_parser, self.parser)
args = new_parser.parse_args(['--with_header'])
self.assertTrue(args.with_header)
args = new_parser.parse_args([])
self.assertFalse(args.with_header)
def test_marshall_no_header(self):
new_parser = self.formatter.add_args(self.parser)
args = new_parser.parse_args([])
result = self.formatter.marshall(args, self.data)
expect = "a,1\r\nb,2\r\nc,3\r\n"
self.assertEqual(result, expect)
def test_marshall_with_header(self):
new_parser = self.formatter.add_args(self.parser)
args = new_parser.parse_args(['--with_header'])
result = self.formatter.marshall(args, self.data)
expect = "char,order\r\na,1\r\nb,2\r\nc,3\r\n"
self.assertEqual(result, expect)
|
Fix no header test for csv formatter
|
Fix no header test for csv formatter
|
Python
|
mit
|
eiri/echolalia-prototype
|
3a5432e14c18852758afdf92b913c93906808e3e
|
cinder/db/sqlalchemy/migrate_repo/versions/115_add_shared_targets_to_volumes.py
|
cinder/db/sqlalchemy/migrate_repo/versions/115_add_shared_targets_to_volumes.py
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from sqlalchemy import Boolean, Column, MetaData, Table
def upgrade(migrate_engine):
"""Add shared_targets column to Volumes."""
meta = MetaData()
meta.bind = migrate_engine
volumes = Table('volumes', meta, autoload=True)
# NOTE(jdg): We use a default of True because it's harmless for a device
# that does NOT use shared_targets to be treated as if it does
shared_targets = Column('shared_targets',
Boolean,
default=True)
volumes.create_column(shared_targets)
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from sqlalchemy import Boolean, Column, MetaData, Table
def upgrade(migrate_engine):
"""Add shared_targets column to Volumes."""
meta = MetaData()
meta.bind = migrate_engine
volumes = Table('volumes', meta, autoload=True)
# NOTE(jdg): We use a default of True because it's harmless for a device
# that does NOT use shared_targets to be treated as if it does
if not hasattr(volumes.c, 'shared_targets'):
volumes.create_column(Column('shared_targets', Boolean, default=True))
|
Add 'shared_targets' only when it doesn't exist
|
Add 'shared_targets' only when it doesn't exist
Add existence check before actually create it.
Change-Id: I96946f736d7263f80f7ad24f8cbbc9a09eb3cc63
|
Python
|
apache-2.0
|
phenoxim/cinder,Datera/cinder,mahak/cinder,openstack/cinder,j-griffith/cinder,openstack/cinder,mahak/cinder,j-griffith/cinder,Datera/cinder,phenoxim/cinder
|
052042e2f48b7936a6057c18a128f497d5e5b1a4
|
folium/__init__.py
|
folium/__init__.py
|
# -*- coding: utf-8 -*-
from __future__ import absolute_import
__version__ = '0.2.0.dev'
from folium.folium import Map, initialize_notebook
|
# -*- coding: utf-8 -*-
from __future__ import absolute_import
__version__ = '0.2.0.dev'
from folium.folium import Map, initialize_notebook, CircleMarker
from folium.map import FeatureGroup, FitBounds,Icon, LayerControl, Marker, Popup, TileLayer
from folium.features import (ClickForMarker, ColorScale, CustomIcon, DivIcon, GeoJson, GeoJsonStyle,
ImageOverlay, LatLngPopup, MarkerCluster, MultiPolyLine, PolyLine,
RegularPolygonMarker, TopoJson, Vega, WmsTileLayer)
|
Make features accessible from root
|
Make features accessible from root
|
Python
|
mit
|
QuLogic/folium,talespaiva/folium,andrewgiessel/folium,themiurgo/folium,shankari/folium,python-visualization/folium,talespaiva/folium,QuLogic/folium,BibMartin/folium,ocefpaf/folium,themiurgo/folium,talespaiva/folium,andrewgiessel/folium,BibMartin/folium,ocefpaf/folium,python-visualization/folium,shankari/folium,shankari/folium,themiurgo/folium,BibMartin/folium,andrewgiessel/folium,talespaiva/folium,QuLogic/folium
|
bb6f4302937e477f23c4de0d6a265d1d6f8985a0
|
geometry_export.py
|
geometry_export.py
|
print "Loading ", __name__
import geometry, from_poser, to_lux
reload(geometry)
reload(from_poser)
reload(to_lux)
import from_poser, to_lux
class GeometryExporter(object):
def __init__(self, subject, convert_material = None,
write_mesh_parameters = None, options = {}):
geom = from_poser.get(subject)
if geom is None or geom.is_empty:
print "Mesh is empty."
self.write = lambda file: None
else:
print "Mesh has", geom.number_of_polygons, "polygons and",
print geom.number_of_points, "vertices"
mats = geom.materials
key = geom.material_key
if convert_material:
materials = [convert_material(mat, key) for mat in mats]
else:
materials = [' NamedMaterial "%s/%s"' % (key, mat.Name())
for mat in mats]
if options.get('compute_normals', True) in [True, 1, '1', 'true']:
geom.compute_normals()
for i in xrange(int(options.get('subdivisionlevel', 0))):
print " subdividing: pass", (i+1)
geom.subdivide()
to_lux.preprocess(geom)
self.write = lambda file: to_lux.write(file, geom, materials,
write_mesh_parameters)
|
print "Loading ", __name__
import geometry, from_poser, to_lux
reload(geometry)
reload(from_poser)
reload(to_lux)
import from_poser, to_lux
def get_materials(geometry, convert = None):
f = convert or (lambda mat, k: ' NamedMaterial "%s/%s"' % (k, mat.Name()))
return [f(mat, geometry.material_key) for mat in geometry.materials]
def preprocess(geometry, options = {}):
if options.get('compute_normals', True) in [True, 1, '1', 'true']:
geometry.compute_normals()
for i in xrange(int(options.get('subdivisionlevel', 0))):
print " subdividing: pass", (i+1)
geometry.subdivide()
class GeometryExporter(object):
def __init__(self, subject, convert_material = None,
write_mesh_parameters = None, options = {}):
geom = from_poser.get(subject)
if geom is None or geom.is_empty:
print "Mesh is empty."
self.write = lambda file: None
else:
print "Mesh has", geom.number_of_polygons, "polygons and",
print geom.number_of_points, "vertices"
materials = get_materials(geom, convert_material)
preprocess(geom, options)
to_lux.preprocess(geom)
self.write = lambda file: to_lux.write(file, geom, materials,
write_mesh_parameters)
|
Split off two functions from GeometryExporter.__init__
|
Split off two functions from GeometryExporter.__init__
|
Python
|
mit
|
odf/pydough
|
77b1f64633d2b70e4e4fc490916e2a9ccae7228f
|
gignore/__init__.py
|
gignore/__init__.py
|
__version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
|
__version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
valid = True
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
def is_valid(self):
"""
:rtype: bool
"""
return self.valid
def set_valid(self, valid):
"""
:type valid: bool
"""
self.valid = valid
|
Add valid attribute with setter/getter
|
Add valid attribute with setter/getter
|
Python
|
bsd-3-clause
|
Alir3z4/python-gignore
|
0986bbba02a4bb4d2c13835dd91281cce3bb5f10
|
alembic/versions/174eb928136a_gdpr_restrict_processing.py
|
alembic/versions/174eb928136a_gdpr_restrict_processing.py
|
"""GDPR restrict processing
Revision ID: 174eb928136a
Revises: d5b07c8f0893
Create Date: 2018-05-14 11:21:55.138387
"""
# revision identifiers, used by Alembic.
revision = '174eb928136a'
down_revision = 'd5b07c8f0893'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column('user', sa.Column('restrict', sa.Boolean))
def downgrade():
op.drop_column('user', 'restrict')
|
"""GDPR restrict processing
Revision ID: 174eb928136a
Revises: d5b07c8f0893
Create Date: 2018-05-14 11:21:55.138387
"""
# revision identifiers, used by Alembic.
revision = '174eb928136a'
down_revision = 'd5b07c8f0893'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column('user', sa.Column('restrict', sa.Boolean, default=False))
sql = 'update "user" set restrict=false'
op.execute(sql)
def downgrade():
op.drop_column('user', 'restrict')
|
Set default to False, and update existing users.
|
Set default to False, and update existing users.
|
Python
|
agpl-3.0
|
Scifabric/pybossa,PyBossa/pybossa,Scifabric/pybossa,PyBossa/pybossa
|
75d435e55e42fefe1c28095dadb9abb56284c1fb
|
marked/__init__.py
|
marked/__init__.py
|
import markgen
from bs4 import BeautifulSoup
TAGS = {
'p': 'paragraph',
'div': 'paragraph',
'a': 'link',
'strong': 'emphasis',
'em': 'emphasis',
'b': 'emphasis',
'i': 'emphasis',
'u': 'emphasis',
'img': 'image',
'image': 'image',
'blockquote': 'quote',
'pre': 'pre',
'code': 'pre',
'h1': 'header',
'h2': 'header',
'h3': 'header',
'h4': 'header',
'h5': 'header',
'h6': 'header',
'ul': 'ulist',
'ol': 'olist'
}
def markup_to_markdown(content):
soup = BeautifulSoup(content)
# Account for HTML snippets and full documents alike
contents = soup.body.contents if soup.body is not None else soup.contents
return _iterate_over_contents(contents)
def _iterate_over_contents(contents):
out = u''
for c in contents:
if hasattr(c, 'contents'):
c = _iterate_over_contents(c.contents)
if c.name in TAGS:
wrap = getattr(markgen, TAGS[c.name])
c = wrap(c)
out += u"\n{0}".format(c)
return out
|
import markgen
from bs4 import BeautifulSoup
TAGS = {
'p': 'paragraph',
'div': 'paragraph',
'a': 'link',
'strong': 'emphasis',
'em': 'emphasis',
'b': 'emphasis',
'i': 'emphasis',
'u': 'emphasis',
'img': 'image',
'image': 'image',
'blockquote': 'quote',
'pre': 'pre',
'code': 'pre',
'h1': 'header',
'h2': 'header',
'h3': 'header',
'h4': 'header',
'h5': 'header',
'h6': 'header',
'ul': 'ulist',
'ol': 'olist'
}
def markup_to_markdown(content):
soup = BeautifulSoup(content)
# Account for HTML snippets and full documents alike
contents = soup.body.contents if soup.body is not None else soup.contents
return _iterate_over_contents(contents)
def _iterate_over_contents(contents):
out = u''
for c in contents:
if hasattr(c, 'contents'):
c.string = _iterate_over_contents(c.contents)
if c.name in TAGS:
wrap = getattr(markgen, TAGS[c.name])
c = wrap(c.string)
out += u"\n{0}".format(c)
return out
|
Use .string so we keep within BS parse tree
|
Use .string so we keep within BS parse tree
|
Python
|
bsd-3-clause
|
1stvamp/marked
|
6d964e5ce83b8f07de64ef8ed5b531271725d9c4
|
peering/management/commands/deploy_configurations.py
|
peering/management/commands/deploy_configurations.py
|
from __future__ import unicode_literals
import logging
from django.core.management.base import BaseCommand
from peering.models import InternetExchange
class Command(BaseCommand):
help = ('Deploy configurations each IX having a router and a configuration'
' template attached.')
logger = logging.getLogger('peering.manager.peering')
def handle(self, *args, **options):
self.logger.info('Deploying configurations...')
for ix in InternetExchange.objects.all():
if ix.configuration_template and ix.router:
self.logger.info(
'Deploying configuration on {}'.format(ix.name))
ix.router.set_napalm_configuration(ix.generate_configuration(),
commit=True)
else:
self.logger.info(
'No configuration to deploy on {}'.format(ix.name))
self.logger.info('Configurations deployed')
|
from __future__ import unicode_literals
import logging
from django.core.management.base import BaseCommand
from peering.models import InternetExchange
class Command(BaseCommand):
help = ('Deploy configurations each IX having a router and a configuration'
' template attached.')
logger = logging.getLogger('peering.manager.peering')
def handle(self, *args, **options):
self.logger.info('Deploying configurations...')
for ix in InternetExchange.objects.all():
# Only deploy config if there are at least a configuration
# template, a router and a platform for the router
if ix.configuration_template and ix.router and ix.router.platform:
self.logger.info(
'Deploying configuration on {}'.format(ix.name))
ix.router.set_napalm_configuration(ix.generate_configuration(),
commit=True)
else:
self.logger.info(
'No configuration to deploy on {}'.format(ix.name))
self.logger.info('Configurations deployed')
|
Check for router platform in auto-deploy script.
|
Check for router platform in auto-deploy script.
|
Python
|
apache-2.0
|
respawner/peering-manager,respawner/peering-manager,respawner/peering-manager,respawner/peering-manager
|
8ef3e88c99602dbdac8fca1b223c7bab8308d820
|
backend/backend/serializers.py
|
backend/backend/serializers.py
|
from rest_framework import serializers
from .models import Animal
class AnimalSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Animal
fields = ('id', 'name', 'dob', 'active', 'own')
|
from rest_framework import serializers
from .models import Animal
class AnimalSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Animal
fields = ('id', 'name', 'dob', 'gender', 'active', 'own', 'father', 'mother')
|
Add parents and gender to the list of values in serializer
|
Add parents and gender to the list of values in serializer
|
Python
|
apache-2.0
|
mmlado/animal_pairing,mmlado/animal_pairing
|
4d7f94e7ee5b2ffdfe58b353688ae5bfc280332c
|
boris/reporting/management.py
|
boris/reporting/management.py
|
'''
Created on 3.12.2011
@author: xaralis
'''
from os.path import dirname, join
from django.db import models, connection
from boris import reporting
from boris.reporting import models as reporting_app
def install_views(app, created_models, verbosity, **kwargs):
if verbosity >= 1:
print "Installing reporting views ..."
cursor = connection.cursor()
sql_file = open(join(dirname(reporting.__file__), 'sql', 'reporting-views.mysql.sql'), 'r')
cursor.execute(sql_file.read())
sql_file.close()
models.signals.post_syncdb.connect(install_views, sender=reporting_app)
|
from os.path import dirname, join
from django.db import connection
from south.signals import post_migrate
from boris import reporting
from boris.reporting import models as reporting_app
def install_views(app, **kwargs):
print "Installing reporting views ..."
cursor = connection.cursor()
sql_file = open(join(dirname(reporting.__file__), 'sql', 'reporting-views.mysql.sql'), 'r')
try:
cursor.execute(sql_file.read())
finally:
sql_file.close()
post_migrate.connect(install_views, sender=reporting_app)
|
Install views on post_migrate rather than post_syncdb.
|
Install views on post_migrate rather than post_syncdb.
|
Python
|
mit
|
fragaria/BorIS,fragaria/BorIS,fragaria/BorIS
|
c6ba057d2e8a1b75edb49ce3c007676f4fe46a16
|
tv-script-generation/helper.py
|
tv-script-generation/helper.py
|
import os
import pickle
def load_data(path):
"""
Load Dataset from File
"""
input_file = os.path.join(path)
with open(input_file, "r") as f:
data = f.read()
return data
def preprocess_and_save_data(dataset_path, token_lookup, create_lookup_tables):
"""
Preprocess Text Data
"""
text = load_data(dataset_path)
token_dict = token_lookup()
for key, token in token_dict.items():
text = text.replace(key, ' {} '.format(token))
text = text.lower()
text = text.split()
vocab_to_int, int_to_vocab = create_lookup_tables(text)
int_text = [vocab_to_int[word] for word in text]
pickle.dump((int_text, vocab_to_int, int_to_vocab, token_dict), open('preprocess.p', 'wb'))
def load_preprocess():
"""
Load the Preprocessed Training data and return them in batches of <batch_size> or less
"""
return pickle.load(open('preprocess.p', mode='rb'))
def save_params(params):
"""
Save parameters to file
"""
pickle.dump(params, open('params.p', 'wb'))
def load_params():
"""
Load parameters from file
"""
return pickle.load(open('params.p', mode='rb'))
|
import os
import pickle
def load_data(path):
"""
Load Dataset from File
"""
input_file = os.path.join(path)
with open(input_file, "r") as f:
data = f.read()
return data
def preprocess_and_save_data(dataset_path, token_lookup, create_lookup_tables):
"""
Preprocess Text Data
"""
text = load_data(dataset_path)
# Ignore notice, since we don't use it for analysing the data
text = text[81:]
token_dict = token_lookup()
for key, token in token_dict.items():
text = text.replace(key, ' {} '.format(token))
text = text.lower()
text = text.split()
vocab_to_int, int_to_vocab = create_lookup_tables(text)
int_text = [vocab_to_int[word] for word in text]
pickle.dump((int_text, vocab_to_int, int_to_vocab, token_dict), open('preprocess.p', 'wb'))
def load_preprocess():
"""
Load the Preprocessed Training data and return them in batches of <batch_size> or less
"""
return pickle.load(open('preprocess.p', mode='rb'))
def save_params(params):
"""
Save parameters to file
"""
pickle.dump(params, open('params.p', 'wb'))
def load_params():
"""
Load parameters from file
"""
return pickle.load(open('params.p', mode='rb'))
|
Remove copyright notice during preprocessing
|
Remove copyright notice during preprocessing
|
Python
|
mit
|
spencer2211/deep-learning
|
0983361e6fba5812416d8fb5b695f6b3034bc927
|
registration/management/commands/cleanupregistration.py
|
registration/management/commands/cleanupregistration.py
|
"""
A management command which deletes expired accounts (e.g.,
accounts which signed up but never activated) from the database.
Calls ``RegistrationProfile.objects.delete_expired_users()``, which
contains the actual logic for determining which accounts are deleted.
"""
from django.core.management.base import NoArgsCommand
from ...models import RegistrationProfile
class Command(NoArgsCommand):
help = "Delete expired user registrations from the database"
def handle_noargs(self, **options):
RegistrationProfile.objects.delete_expired_users()
|
"""
A management command which deletes expired accounts (e.g.,
accounts which signed up but never activated) from the database.
Calls ``RegistrationProfile.objects.delete_expired_users()``, which
contains the actual logic for determining which accounts are deleted.
"""
from django.core.management.base import BaseCommand
from ...models import RegistrationProfile
class Command(BaseCommand):
help = "Delete expired user registrations from the database"
def handle(self, *args, **options):
RegistrationProfile.objects.delete_expired_users()
|
Fix deprecated class NoArgsCommand class.
|
Fix deprecated class NoArgsCommand class.
Solve the warning RemovedInDjango110Warning: NoArgsCommand class is deprecated and will be removed in Django 1.10. Use BaseCommand instead, which takes no arguments by default.
|
Python
|
bsd-3-clause
|
sergafts/django-registration,timgraham/django-registration,sergafts/django-registration,pando85/django-registration,pando85/django-registration,allo-/django-registration,allo-/django-registration,timgraham/django-registration
|
da66b82b4a5d5c0b0bb716b05a8bfd2dae5e2f4c
|
ookoobah/glutil.py
|
ookoobah/glutil.py
|
from contextlib import contextmanager
from pyglet.gl import *
def ptr(*args):
return (GLfloat * len(args))(*args)
@contextmanager
def gl_disable(*bits):
glPushAttrib(GL_ENABLE_BIT)
map(glDisable, bits)
yield
glPopAttrib(GL_ENABLE_BIT)
@contextmanager
def gl_ortho(window):
# clobbers current modelview matrix
glMatrixMode(GL_PROJECTION)
glPushMatrix()
glLoadIdentity()
gluOrtho2D(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
yield
glMatrixMode(GL_PROJECTION)
glPopMatrix()
glMatrixMode(GL_MODELVIEW)
|
from contextlib import contextmanager
from pyglet.gl import *
__all__ = [
'ptr',
'gl_disable',
'gl_ortho',
]
def ptr(*args):
return (GLfloat * len(args))(*args)
@contextmanager
def gl_disable(*bits):
glPushAttrib(GL_ENABLE_BIT)
map(glDisable, bits)
yield
glPopAttrib(GL_ENABLE_BIT)
@contextmanager
def gl_ortho(window):
# clobbers current modelview matrix
glMatrixMode(GL_PROJECTION)
glPushMatrix()
glLoadIdentity()
gluOrtho2D(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
yield
glMatrixMode(GL_PROJECTION)
glPopMatrix()
glMatrixMode(GL_MODELVIEW)
|
Fix pyglet breackage by controlling exports.
|
Fix pyglet breackage by controlling exports.
|
Python
|
mit
|
vickenty/ookoobah,vickenty/ookoobah,vickenty/ookoobah,vickenty/ookoobah,vickenty/ookoobah
|
04c8a36c5713e4279f8bf52fa45cdb03de721dbb
|
example/deploy.py
|
example/deploy.py
|
from pyinfra import inventory, state
from pyinfra_docker import deploy_docker
from pyinfra_etcd import deploy_etcd
from pyinfra_kubernetes import deploy_kubernetes_master, deploy_kubernetes_node
SUDO = True
FAIL_PERCENT = 0
def get_etcd_nodes():
return [
'http://{0}:2379'.format(
etcd_node.fact.network_devices[etcd_node.data.etcd_interface]
['ipv4']['address'],
)
for etcd_node in inventory.get_group('etcd_nodes')
]
# Install/configure etcd cluster
with state.limit('etcd_nodes'):
deploy_etcd()
# Install/configure the masters (apiserver, controller, scheduler)
with state.limit('kubernetes_masters'):
deploy_kubernetes_master(etcd_nodes=get_etcd_nodes())
# Install/configure the nodes
with state.limit('kubernetes_nodes'):
# Install Docker
deploy_docker()
# Install Kubernetes node components (kubelet, kube-proxy)
first_master = inventory.get_group('kubernetes_masters')[0]
deploy_kubernetes_node(
master_address='http://{0}'.format((
first_master
.fact.network_devices[first_master.data.network_interface]
['ipv4']['address']
)),
)
# deploy_docker(config={
# # Make Docker use the Vagrant provided interface which has it's own /24
# 'bip': '{{ host.fact.network_devices[host.data.network_interface].ipv4.address }}',
# })
|
from pyinfra import inventory, state
from pyinfra_docker import deploy_docker
from pyinfra_etcd import deploy_etcd
from pyinfra_kubernetes import deploy_kubernetes_master, deploy_kubernetes_node
SUDO = True
FAIL_PERCENT = 0
def get_etcd_nodes():
return [
'http://{0}:2379'.format(
etcd_node.fact.network_devices[etcd_node.data.etcd_interface]
['ipv4']['address'],
)
for etcd_node in inventory.get_group('etcd_nodes')
]
# Install/configure etcd cluster
with state.limit('etcd_nodes'):
deploy_etcd()
# Install/configure the masters (apiserver, controller, scheduler)
with state.limit('kubernetes_masters'):
deploy_kubernetes_master(etcd_nodes=get_etcd_nodes())
# Install/configure the nodes
with state.limit('kubernetes_nodes'):
# Install Docker
deploy_docker(config={
# Make Docker use the Vagrant provided interface which has it's own /24
'bip': '{{ host.fact.network_devices[host.data.network_interface].ipv4.address }}',
})
# Install Kubernetes node components (kubelet, kube-proxy)
first_master = inventory.get_group('kubernetes_masters')[0]
deploy_kubernetes_node(
master_address='http://{0}'.format((
first_master
.fact.network_devices[first_master.data.network_interface]
['ipv4']['address']
)),
)
|
Use Docker config pointing at the correct interface/subnect for networking.
|
Use Docker config pointing at the correct interface/subnect for networking.
|
Python
|
mit
|
EDITD/pyinfra-kubernetes,EDITD/pyinfra-kubernetes
|
4714f803b22eda26eb2fc867c1d9e2c7230bdd11
|
pythonforandroid/recipes/pysdl2/__init__.py
|
pythonforandroid/recipes/pysdl2/__init__.py
|
from pythonforandroid.recipe import PythonRecipe
class PySDL2Recipe(PythonRecipe):
version = '0.9.3'
url = 'https://bitbucket.org/marcusva/py-sdl2/downloads/PySDL2-{version}.tar.gz'
depends = ['sdl2']
recipe = PySDL2Recipe()
|
from pythonforandroid.recipe import PythonRecipe
class PySDL2Recipe(PythonRecipe):
version = '0.9.6'
url = 'https://files.pythonhosted.org/packages/source/P/PySDL2/PySDL2-{version}.tar.gz'
depends = ['sdl2']
recipe = PySDL2Recipe()
|
Fix outdated PySDL2 version and non-PyPI install source
|
Fix outdated PySDL2 version and non-PyPI install source
|
Python
|
mit
|
kronenpj/python-for-android,rnixx/python-for-android,germn/python-for-android,PKRoma/python-for-android,germn/python-for-android,kivy/python-for-android,rnixx/python-for-android,rnixx/python-for-android,rnixx/python-for-android,PKRoma/python-for-android,kivy/python-for-android,germn/python-for-android,kronenpj/python-for-android,rnixx/python-for-android,kronenpj/python-for-android,kivy/python-for-android,kivy/python-for-android,kronenpj/python-for-android,germn/python-for-android,PKRoma/python-for-android,rnixx/python-for-android,germn/python-for-android,germn/python-for-android,kronenpj/python-for-android,kivy/python-for-android,PKRoma/python-for-android,PKRoma/python-for-android
|
32cc988e81bbbecf09f7e7a801e92c6cfc281e75
|
docs/autogen_config.py
|
docs/autogen_config.py
|
#!/usr/bin/env python
from os.path import join, dirname, abspath
from IPython.terminal.ipapp import TerminalIPythonApp
from ipykernel.kernelapp import IPKernelApp
here = abspath(dirname(__file__))
options = join(here, 'source', 'config', 'options')
generated = join(options, 'generated.rst')
def write_doc(name, title, app, preamble=None):
filename = '%s.rst' % name
with open(join(options, filename), 'w') as f:
f.write(title + '\n')
f.write(('=' * len(title)) + '\n')
f.write('\n')
if preamble is not None:
f.write(preamble + '\n\n')
f.write(app.document_config_options())
with open(generated, 'a') as f:
f.write(filename + '\n')
if __name__ == '__main__':
# create empty file
with open(generated, 'w'):
pass
write_doc('terminal', 'Terminal IPython options', TerminalIPythonApp())
write_doc('kernel', 'IPython kernel options', IPKernelApp(),
preamble=("These options can be used in :file:`ipython_kernel_config.py`. "
"The kernel also respects any options in `ipython_config.py`"),
)
|
#!/usr/bin/env python
from os.path import join, dirname, abspath
from IPython.terminal.ipapp import TerminalIPythonApp
from ipykernel.kernelapp import IPKernelApp
here = abspath(dirname(__file__))
options = join(here, 'source', 'config', 'options')
def write_doc(name, title, app, preamble=None):
filename = '%s.rst' % name
with open(join(options, filename), 'w') as f:
f.write(title + '\n')
f.write(('=' * len(title)) + '\n')
f.write('\n')
if preamble is not None:
f.write(preamble + '\n\n')
f.write(app.document_config_options())
if __name__ == '__main__':
write_doc('terminal', 'Terminal IPython options', TerminalIPythonApp())
write_doc('kernel', 'IPython kernel options', IPKernelApp(),
preamble=("These options can be used in :file:`ipython_kernel_config.py`. "
"The kernel also respects any options in `ipython_config.py`"),
)
|
Remove generation of unnecessary generated.rst file
|
Remove generation of unnecessary generated.rst file
|
Python
|
bsd-3-clause
|
ipython/ipython,ipython/ipython
|
b43b555a7803c6afd50fe5992f455cc5d1ad5d86
|
stonemason/service/tileserver/health/views.py
|
stonemason/service/tileserver/health/views.py
|
# -*- encoding: utf-8 -*-
__author__ = 'ray'
__date__ = '3/2/15'
from flask import make_response
def health_check():
"""Return a dummy response"""
response = make_response()
response.headers['Content-Type'] = 'text/plain'
response.headers['Cache-Control'] = 'public, max-age=0'
return response
|
# -*- encoding: utf-8 -*-
__author__ = 'ray'
__date__ = '3/2/15'
from flask import make_response
import stonemason
import sys
import platform
VERSION_STRING = '''stonemason:%s
Python: %s
Platform: %s''' % (stonemason.__version__,
sys.version,
platform.version())
del stonemason, sys, platform
def health_check():
"""Return a dummy response"""
response = make_response(VERSION_STRING)
response.headers['Content-Type'] = 'text/plain'
response.headers['Cache-Control'] = 'public, max-age=0'
return response
|
Return sys/platform version in tileserver health check
|
FEATURE: Return sys/platform version in tileserver health check
|
Python
|
mit
|
Kotaimen/stonemason,Kotaimen/stonemason
|
de6ac0596b58fac2efc547fe6f81a48f4a06f527
|
tests/grammar_creation_test/TerminalAdding.py
|
tests/grammar_creation_test/TerminalAdding.py
|
#!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import TestCase, main
from grammpy import *
class TerminalAddingTest(TestCase):
pass
if __name__ == '__main__':
main()
|
#!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import TestCase, main
from grammpy import *
class TerminalAddingTest(TestCase):
def test_shouldAddOneTerminal(self):
g = Grammar(terminals=['asdf'])
self.assertTrue(g.have_term('asdf'))
self.assertFalse(g.have_term('a'))
def test_shouldAddMoreTerminals(self):
g = Grammar(terminals=[0, 1, 2])
self.assertTrue(g.have_term([0, 1, 2]))
self.assertFalse(g.have_term('a'))
self.assertFalse(g.have_term('asdf'))
self.assertFalse(g.have_term(3))
if __name__ == '__main__':
main()
|
Add tests of terminal adding when grammar is create
|
Add tests of terminal adding when grammar is create
|
Python
|
mit
|
PatrikValkovic/grammpy
|
3081fcd1e37520f504804a3efae62c33d3371a21
|
temba/msgs/migrations/0034_move_recording_domains.py
|
temba/msgs/migrations/0034_move_recording_domains.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
('msgs', '0033_exportmessagestask_uuid'),
]
def move_recording_domains(apps, schema_editor):
Msg = apps.get_model('msgs', 'Msg')
# this is our new bucket name
bucket_name = settings.AWS_STORAGE_BUCKET_NAME
# our old bucket name had periods instead of dashes
old_bucket_domain = 'http://' + bucket_name.replace('-', '.')
# our new domain is more specific
new_bucket_domain = 'https://' + settings.AWS_BUCKET_DOMAIN
for msg in Msg.objects.filter(msg_type='V').exclude(recording_url=None):
# if our recording URL is on our old bucket
if msg.recording_url.find(old_bucket_domain) >= 0:
# rename it to our new bucket
old_recording_url = msg.recording_url
msg.recording_url = msg.recording_url.replace(old_bucket_domain,
new_bucket_domain)
print "[%d] %s to %s" % (msg.id, old_recording_url, msg.recording_url)
msg.save(update_fields=['recording_url'])
operations = [
migrations.RunPython(move_recording_domains)
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
('msgs', '0033_exportmessagestask_uuid'),
]
def move_recording_domains(apps, schema_editor):
Msg = apps.get_model('msgs', 'Msg')
# this is our new bucket name
bucket_name = settings.AWS_STORAGE_BUCKET_NAME
# our old bucket name had periods instead of dashes
old_bucket_domain = 'http://' + bucket_name.replace('-', '.')
# our new domain is more specific
new_bucket_domain = 'https://' + settings.AWS_BUCKET_DOMAIN
for msg in Msg.objects.filter(direction='I', msg_type='V').exclude(recording_url=None):
# if our recording URL is on our old bucket
if msg.recording_url.find(old_bucket_domain) >= 0:
# rename it to our new bucket
old_recording_url = msg.recording_url
msg.recording_url = msg.recording_url.replace(old_bucket_domain,
new_bucket_domain)
print "[%d] %s to %s" % (msg.id, old_recording_url, msg.recording_url)
msg.save(update_fields=['recording_url'])
operations = [
migrations.RunPython(move_recording_domains)
]
|
Tweak to migration so it is a bit faster for future migraters
|
Tweak to migration so it is a bit faster for future migraters
|
Python
|
agpl-3.0
|
tsotetsi/textily-web,tsotetsi/textily-web,pulilab/rapidpro,praekelt/rapidpro,ewheeler/rapidpro,reyrodrigues/EU-SMS,reyrodrigues/EU-SMS,ewheeler/rapidpro,reyrodrigues/EU-SMS,tsotetsi/textily-web,tsotetsi/textily-web,pulilab/rapidpro,praekelt/rapidpro,tsotetsi/textily-web,praekelt/rapidpro,pulilab/rapidpro,ewheeler/rapidpro,pulilab/rapidpro,praekelt/rapidpro,ewheeler/rapidpro,pulilab/rapidpro
|
7755dda1449f6264d7d7fe57dc776c731ab22d84
|
src/satosa/micro_services/processors/scope_processor.py
|
src/satosa/micro_services/processors/scope_processor.py
|
from ..attribute_processor import AttributeProcessorError
from .base_processor import BaseProcessor
CONFIG_KEY_SCOPE = 'scope'
CONFIG_DEFAULT_SCOPE = ''
class ScopeProcessor(BaseProcessor):
def process(self, internal_data, attribute, **kwargs):
scope = kwargs.get(CONFIG_KEY_SCOPE, CONFIG_DEFAULT_SCOPE)
if scope is None or scope == '':
raise AttributeProcessorError("No scope set.")
attributes = internal_data.attributes
value = attributes.get(attribute, [None])[0]
attributes[attribute][0] = value + '@' + scope
|
from ..attribute_processor import AttributeProcessorError
from .base_processor import BaseProcessor
CONFIG_KEY_SCOPE = 'scope'
CONFIG_DEFAULT_SCOPE = ''
class ScopeProcessor(BaseProcessor):
def process(self, internal_data, attribute, **kwargs):
scope = kwargs.get(CONFIG_KEY_SCOPE, CONFIG_DEFAULT_SCOPE)
if scope is None or scope == '':
raise AttributeProcessorError("No scope set.")
attributes = internal_data.attributes
values = attributes.get(attribute, [])
if not isinstance(values, list):
values = [values]
if values:
new_values=[]
for value in values:
new_values.append(value + '@' + scope)
attributes[attribute] = new_values
|
Allow scope processor to handle multivalued attributes
|
Allow scope processor to handle multivalued attributes
|
Python
|
apache-2.0
|
its-dirg/SATOSA,irtnog/SATOSA,SUNET/SATOSA,SUNET/SATOSA,irtnog/SATOSA
|
adf3a500e8ab8115520daa16bc008faeec7cfca9
|
gitfs/views/view.py
|
gitfs/views/view.py
|
import os
from abc import ABCMeta, abstractmethod
from gitfs import FuseMethodNotImplemented
from gitfs.filesystems.passthrough import PassthroughFuse
class View(PassthroughFuse):
__metaclass__ = ABCMeta
def __init__(self, *args, **kwargs):
self.args = args
for attr in kwargs:
setattr(self, attr, kwargs[attr])
def getxattr(self, path, name, position=0):
"""Get extended attributes"""
raise FuseMethodNotImplemented
|
import os
from abc import ABCMeta, abstractmethod
from gitfs import FuseMethodNotImplemented
class View(object):
__metaclass__ = ABCMeta
def __init__(self, *args, **kwargs):
self.args = args
for attr in kwargs:
setattr(self, attr, kwargs[attr])
def getxattr(self, path, name, position=0):
"""Get extended attributes"""
raise FuseMethodNotImplemented
|
Make View inherit from objects instead of PassthroughFuse
|
Make View inherit from objects instead of PassthroughFuse
|
Python
|
apache-2.0
|
PressLabs/gitfs,PressLabs/gitfs,rowhit/gitfs,bussiere/gitfs,ksmaheshkumar/gitfs
|
ee28fdc66fbb0f91821ff18ff219791bf5de8f4d
|
corehq/apps/fixtures/tasks.py
|
corehq/apps/fixtures/tasks.py
|
from __future__ import absolute_import
from __future__ import unicode_literals
from corehq.apps.fixtures.upload import upload_fixture_file
from soil import DownloadBase
from celery.task import task
@task(serializer='pickle')
def fixture_upload_async(domain, download_id, replace):
task = fixture_upload_async
DownloadBase.set_progress(task, 0, 100)
download_ref = DownloadBase.get(download_id)
result = upload_fixture_file(domain, download_ref.get_filename(), replace, task)
DownloadBase.set_progress(task, 100, 100)
return {
'messages': {
'success': result.success,
'messages': result.messages,
'errors': result.errors,
'number_of_fixtures': result.number_of_fixtures,
},
}
@task(serializer='pickle')
def fixture_download_async(prepare_download, *args, **kw):
task = fixture_download_async
DownloadBase.set_progress(task, 0, 100)
prepare_download(task=task, *args, **kw)
DownloadBase.set_progress(task, 100, 100)
|
from __future__ import absolute_import, unicode_literals
from celery.task import task
from soil import DownloadBase
from corehq.apps.fixtures.upload import upload_fixture_file
@task
def fixture_upload_async(domain, download_id, replace):
task = fixture_upload_async
DownloadBase.set_progress(task, 0, 100)
download_ref = DownloadBase.get(download_id)
result = upload_fixture_file(domain, download_ref.get_filename(), replace, task)
DownloadBase.set_progress(task, 100, 100)
return {
'messages': {
'success': result.success,
'messages': result.messages,
'errors': result.errors,
'number_of_fixtures': result.number_of_fixtures,
},
}
@task(serializer='pickle')
def fixture_download_async(prepare_download, *args, **kw):
task = fixture_download_async
DownloadBase.set_progress(task, 0, 100)
prepare_download(task=task, *args, **kw)
DownloadBase.set_progress(task, 100, 100)
|
Change fixture upload task to json serializer
|
Change fixture upload task to json serializer
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
57024104a5951d62ff8a87a281a6d232583dabed
|
python/new_year_chaos.py
|
python/new_year_chaos.py
|
#!/bin/python3
import math
import os
import random
import re
import sys
# Complete the minimumBribes function below.
def minimumBribes(finalLine):
if invalid(finalLine):
return "Too chaotic"
return bubbleSort(finalLine)
def invalid(finalLine):
return any(didBribeMoreThanTwoPeople(person, index) for index, person in enumerate(finalLine))
def didBribeMoreThanTwoPeople(person, index):
return index + 2 < person - 1
def bubbleSort(line):
swaps = 0
numberOfPeople = len(line)
for person in range(numberOfPeople):
for i in range(0, numberOfPeople - person - 1):
if line[i] > line[i + 1]:
line[i], line[i + 1] = line[i + 1], line[i]
swaps += 1
return swaps
if __name__ == '__main__':
t = int(input())
for t_itr in range(t):
n = int(input())
q = list(map(int, input().rstrip().split()))
print(minimumBribes(q))
|
#!/bin/python3
import math
import os
import random
import re
import sys
# Complete the minimumBribes function below.
def minimumBribes(finalLine):
if invalid(finalLine):
return "Too chaotic"
return bubbleSort(finalLine)
def invalid(finalLine):
return any(didBribeMoreThanTwoPeople(person, index) for index, person in enumerate(finalLine))
def didBribeMoreThanTwoPeople(person, index):
return index + 2 < person - 1
def bubbleSort(line):
swaps = 0
swappedInCurrentPass = False
for person in range(len(line)):
for i in range(0, len(line) - 1):
if line[i] > line[i + 1]:
line[i], line[i + 1] = line[i + 1], line[i]
swaps += 1
swappedInCurrentPass = True
if swappedInCurrentPass:
swappedInCurrentPass = False
else:
break
return swaps
if __name__ == '__main__':
t = int(input())
for t_itr in range(t):
n = int(input())
q = list(map(int, input().rstrip().split()))
print(minimumBribes(q))
|
Improve efficiency of new year chaos
|
Improve efficiency of new year chaos
|
Python
|
mit
|
rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank
|
d90edf3b4d8fa714e7e24acbc22fb35bc828911d
|
services/controllers/interpolator.py
|
services/controllers/interpolator.py
|
class Interpolator:
def __init__(self):
self.data = []
def addIndexValue(self, index, value):
self.data.append((index, value))
def valueAtIndex(self, target_index):
if target_index < self.data[0][0]:
return None
elif self.data[-1][0] < target_index:
return None
else:
start = None
end = None
for (index, value) in self.data:
if index == target_index:
return value
else:
if index <= target_index:
start = (index, value)
elif target_index < index:
end = (index, value)
break
index_delta = end[0] - start[0]
percent = (target_index - start[0]) / index_delta
value_delta = end[1] - start[1]
return start[1] + value_delta * percent
if __name__ == "__main__":
pass
|
class Interpolator:
def __init__(self):
self.data = []
def addIndexValue(self, index, value):
self.data.append((index, value))
def valueAtIndex(self, target_index):
if target_index < self.data[0][0]:
return None
elif self.data[-1][0] < target_index:
return None
else:
start = None
end = None
for (index, value) in self.data:
if index == target_index:
return value
else:
if index <= target_index:
start = (index, value)
elif target_index < index:
end = (index, value)
break
index_delta = end[0] - start[0]
percent = (target_index - start[0]) / index_delta
value_delta = end[1] - start[1]
return start[1] + value_delta * percent
def to_array(self):
result = []
for (index, value) in self.data:
result.append(index)
result.append(value)
return result
def from_array(self, array):
self.data = []
for i in range(0, len(array), 2):
self.addIndexValue(array[i], array[i + 1])
if __name__ == "__main__":
pass
|
Add ability to convert to/from an array
|
Add ability to convert to/from an array
This is needed as an easy way to serialize an interpolator for sending/receiving over HTTP
|
Python
|
bsd-3-clause
|
gizmo-cda/g2x-submarine-v2,gizmo-cda/g2x-submarine-v2,gizmo-cda/g2x-submarine-v2,gizmo-cda/g2x-submarine-v2
|
3deffc39e1a489255272c35f7171b7e85942b108
|
shipyard/shipyard/host/node/build.py
|
shipyard/shipyard/host/node/build.py
|
"""Host-only environment for Node.js."""
from pathlib import Path
from foreman import define_parameter, decorate_rule
from shipyard import install_packages
(define_parameter('npm_prefix')
.with_doc("""Location host-only npm.""")
.with_type(Path)
.with_derive(lambda ps: ps['//base:build'] / 'host/npm-host')
)
@decorate_rule('//base:build')
def install(parameters):
"""Set up host-only environment for Node.js."""
if not Path('/usr/bin/nodejs').exists():
install_packages(['nodejs', 'npm'])
contents = 'prefix = %s\n' % parameters['npm_prefix'].absolute()
(Path.home() / '.npmrc').write_text(contents)
|
"""Host-only environment for Node.js."""
from pathlib import Path
from foreman import define_parameter, decorate_rule
from shipyard import (
ensure_file,
execute,
install_packages,
)
(define_parameter('npm_prefix')
.with_doc("""Location host-only npm.""")
.with_type(Path)
.with_derive(lambda ps: ps['//base:build'] / 'host/npm-host')
)
@decorate_rule('//base:build')
def install(parameters):
"""Set up host-only environment for Node.js."""
if not Path('/usr/bin/node').exists():
install_packages(['nodejs', 'npm'])
contents = 'prefix = %s\n' % parameters['npm_prefix'].absolute()
(Path.home() / '.npmrc').write_text(contents)
# Ubuntu systems use `nodejs` rather than `node` :(
if not Path('/usr/bin/node').exists():
ensure_file('/usr/bin/nodejs')
execute('sudo ln --symbolic nodejs node'.split(), cwd='/usr/bin')
|
Fix node/nodejs name conflict on Ubuntu systems
|
Fix node/nodejs name conflict on Ubuntu systems
|
Python
|
mit
|
clchiou/garage,clchiou/garage,clchiou/garage,clchiou/garage
|
df8618c185108aa71e42da7d9569e16fb350b4c0
|
hackeriet/doorcontrold/__init__.py
|
hackeriet/doorcontrold/__init__.py
|
#!/usr/bin/env python
from hackeriet.mqtt import MQTT
from hackeriet.door import Doors
import threading, os, logging
logging.basicConfig(level=logging.INFO, format='%(asctime)-15s %(message)s')
piface = False
# Determine if piface is used on the Pi
if "PIFACE" in os.environ:
piface = True
logging.info('Using piface configuration')
# Be backwards compatible with old env variable name
gpio_pin = int(os.getenv("DOOR_GPIO_PIN", os.getenv("DOOR_PIN", 0)))
# How many seconds should the door lock remain open
timeout = int(os.getenv("DOOR_TIMEOUT", 2))
door = Doors(piface=piface,pin=gpio_pin,timeout=timeout)
mqtt = MQTT()
door_name = os.getenv("DOOR_NAME", 'hackeriet')
door_topic = "hackeriet/door/%s/open" % door_name
mqtt.subscribe(door_topic, 0)
def on_message(mosq, obj, msg):
door.open()
logging('Door opened: %s' % msg.payload.decode())
mqtt.on_message = on_message
# Block forever
def main():
for t in threading.enumerate():
if t us threading.currentThread():
continue
t.join()
if __name__ == "__main__":
main()
|
#!/usr/bin/env python
from hackeriet.mqtt import MQTT
from hackeriet.door import Doors
import threading, os, logging
logging.basicConfig(level=logging.INFO, format='%(asctime)-15s %(message)s')
piface = False
# Determine if piface is used on the Pi
if "PIFACE" in os.environ:
piface = True
logging.info('Using piface configuration')
# Be backwards compatible with old env variable name
gpio_pin = int(os.getenv("DOOR_GPIO_PIN", os.getenv("DOOR_PIN", 0)))
# How many seconds should the door lock remain open
timeout = int(os.getenv("DOOR_TIMEOUT", 2))
door = Doors(piface=piface,pin=gpio_pin,timeout=timeout)
def on_message(mosq, obj, msg):
door.open()
logging.info('Door opened: %s' % msg.payload
door_name = os.getenv("DOOR_NAME", 'hackeriet')
door_topic = "hackeriet/door/%s/open" % door_name
mqtt = MQTT(on_message)
mqtt.subscribe(door_topic, 0)
# Block forever
def main():
for t in threading.enumerate():
if t us threading.currentThread():
continue
t.join()
if __name__ == "__main__":
main()
|
Fix incompatibilities with latest paho lib
|
Fix incompatibilities with latest paho lib
|
Python
|
apache-2.0
|
hackeriet/pyhackeriet,hackeriet/pyhackeriet,hackeriet/nfcd,hackeriet/nfcd,hackeriet/pyhackeriet,hackeriet/nfcd
|
edc5564d4c3677dc8b545e9c9a6a51b481247eab
|
contentcuration/contentcuration/tests/test_makemessages.py
|
contentcuration/contentcuration/tests/test_makemessages.py
|
import os
import subprocess
import pathlib
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
# this test can make changes to committed files, so only run it
# on the CI server
if 'CI' not in os.environ or not os.environ['CI']:
return
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
)
|
import os
import subprocess
import pathlib
import pytest
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
# this test can make changes to committed files, so only run it
# on the CI server
@pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server")
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
)
|
Use pytest.skip so we can check the test wasn't skipped on the CI.
|
Use pytest.skip so we can check the test wasn't skipped on the CI.
|
Python
|
mit
|
DXCanas/content-curation,DXCanas/content-curation,DXCanas/content-curation,DXCanas/content-curation
|
c3f8069435f0f1c09c00ed6dba2e4f3bdb7ab91b
|
grow/testing/testdata/pod/extensions/preprocessors.py
|
grow/testing/testdata/pod/extensions/preprocessors.py
|
from grow import Preprocessor
from protorpc import messages
class CustomPreprocessor(Preprocessor):
KIND = 'custom_preprocessor'
class Config(messages.Message):
value = messages.StringField(1)
def run(self):
# To allow the test to check the result
self.pod._custom_preprocessor_value = self.config.value
|
from grow import Preprocessor
from protorpc import messages
class CustomPreprocessor(Preprocessor):
KIND = 'custom_preprocessor'
class Config(messages.Message):
value = messages.StringField(1)
def run(self, **kwargs):
# To allow the test to check the result
self.pod._custom_preprocessor_value = self.config.value
|
Update extension testdata to take **kwargs.
|
Update extension testdata to take **kwargs.
|
Python
|
mit
|
grow/grow,grow/pygrow,denmojo/pygrow,grow/pygrow,grow/grow,denmojo/pygrow,denmojo/pygrow,grow/pygrow,denmojo/pygrow,grow/grow,grow/grow
|
e29b1f6243fb7f9d2322b80573617ff9a0582d01
|
pinax/blog/parsers/markdown_parser.py
|
pinax/blog/parsers/markdown_parser.py
|
from markdown import Markdown
from markdown.inlinepatterns import ImagePattern, IMAGE_LINK_RE
from ..models import Image
class ImageLookupImagePattern(ImagePattern):
def sanitize_url(self, url):
if url.startswith("http"):
return url
else:
try:
image = Image.objects.get(pk=int(url))
return image.image_path.url
except Image.DoesNotExist:
pass
except ValueError:
return url
return ""
def parse(text):
md = Markdown(extensions=["codehilite"])
md.inlinePatterns["image_link"] = ImageLookupImagePattern(IMAGE_LINK_RE, md)
html = md.convert(text)
return html
|
from markdown import Markdown
from markdown.inlinepatterns import ImagePattern, IMAGE_LINK_RE
from ..models import Image
class ImageLookupImagePattern(ImagePattern):
def sanitize_url(self, url):
if url.startswith("http"):
return url
else:
try:
image = Image.objects.get(pk=int(url))
return image.image_path.url
except Image.DoesNotExist:
pass
except ValueError:
return url
return ""
def parse(text):
md = Markdown(extensions=["codehilite", "tables", "smarty", "admonition", "toc"])
md.inlinePatterns["image_link"] = ImageLookupImagePattern(IMAGE_LINK_RE, md)
html = md.convert(text)
return html
|
Add some extensions to the markdown parser
|
Add some extensions to the markdown parser
Ultimately we should make this a setting or hookset so it could be overridden at the site level.
|
Python
|
mit
|
swilcox/pinax-blog,pinax/pinax-blog,miurahr/pinax-blog,miurahr/pinax-blog,swilcox/pinax-blog,easton402/pinax-blog,pinax/pinax-blog,pinax/pinax-blog,easton402/pinax-blog
|
044e55544529aa8eb3a755428d990f0400403687
|
xunit-autolabeler-v2/ast_parser/core/test_data/parser/exclude_tags/exclude_tags_main.py
|
xunit-autolabeler-v2/ast_parser/core/test_data/parser/exclude_tags/exclude_tags_main.py
|
# Copyright 2020 Google LLC.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# [START main_method]
def included():
return 'included method one'
# [START_EXCLUDE]
def also_included():
return 'also included method'
# [END_EXCLUDE]
# [END main_method]
|
# Copyright 2020 Google LLC.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# [START included]
def included():
return 'included method one'
# [START_EXCLUDE]
def also_included():
return 'also included method'
# [END_EXCLUDE]
# [END included]
|
Fix stepping on other tests >:(
|
Fix stepping on other tests >:(
|
Python
|
apache-2.0
|
GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground,GoogleCloudPlatform/repo-automation-playground
|
606b2b6c84e9f9f67606a4d7e521cf4805855a98
|
migrations/versions/0311_populate_returned_letters.py
|
migrations/versions/0311_populate_returned_letters.py
|
"""
Revision ID: 0311_populate_returned_letters
Revises: 0310_returned_letters_table
Create Date: 2019-12-09 12:13:49.432993
"""
from alembic import op
from app.dao.returned_letters_dao import insert_or_update_returned_letters
revision = '0311_populate_returned_letters'
down_revision = '0310_returned_letters_table'
def upgrade():
conn = op.get_bind()
sql = """
select id, service_id, reference
from notification_history
where notification_type = 'letter'
and notification_status = 'returned-letter'"""
results = conn.execute(sql)
returned_letters = results.fetchall()
references = [x.reference for x in returned_letters]
insert_or_update_returned_letters(references)
def downgrade():
pass
|
"""
Revision ID: 0311_populate_returned_letters
Revises: 0310_returned_letters_table
Create Date: 2019-12-09 12:13:49.432993
"""
from alembic import op
revision = '0311_populate_returned_letters'
down_revision = '0310_returned_letters_table'
def upgrade():
conn = op.get_bind()
sql = """
select id, service_id, reference, updated_at
from notification_history
where notification_type = 'letter'
and notification_status = 'returned-letter'"""
insert_sql = """
insert into returned_letters(id, reported_at, service_id, notification_id, created_at, updated_at)
values(uuid_in(md5(random()::text)::cstring), '{}', '{}', '{}', now(), null)
"""
results = conn.execute(sql)
returned_letters = results.fetchall()
for x in returned_letters:
f = insert_sql.format(x.updated_at.date(), x.service_id, x.id)
conn.execute(f)
def downgrade():
pass
|
Change the insert to use updated_at as the reported_at date
|
Change the insert to use updated_at as the reported_at date
|
Python
|
mit
|
alphagov/notifications-api,alphagov/notifications-api
|
853d2907432a8d7fbedbed12ff28efbe520d4c80
|
project_euler/library/number_theory/continued_fractions.py
|
project_euler/library/number_theory/continued_fractions.py
|
from fractions import Fraction
from math import sqrt
from itertools import chain, cycle
from typing import Generator, Iterable, List, Tuple
def convergent_sequence(generator: Iterable[int]) -> \
Generator[Fraction, None, None]:
h = (0, 1)
k = (1, 0)
for a in generator:
h = h[1], a * h[1] + h[0]
k = k[1], a * k[1] + k[0]
yield Fraction(h[-1], k[-1])
def continued_fraction_sqrt(n: int) -> Tuple[List[int], List[int]]:
remainders = []
continued_fraction = []
remainder = (Fraction(1), Fraction(0)) # remainder is sqrt(n) + 0.
sqrt_n = sqrt(n)
while remainder not in remainders:
remainders.append(remainder)
a = int(remainder[0] * sqrt_n + remainder[1])
continued_fraction.append(a)
norm = (remainder[1] - a) ** 2 - remainder[0] ** 2 * n
remainder = (-remainder[0] / norm, (remainder[1] - a) / norm)
index = remainders.index(remainder)
return continued_fraction[:index], continued_fraction[index:]
def convergents_sqrt(n: int) -> Generator[Fraction, None, None]:
initial, repeat = continued_fraction_sqrt(n)
convergents = convergent_sequence(chain(initial, cycle(repeat)))
yield from convergents
|
from fractions import Fraction
from math import sqrt
from itertools import chain, cycle
from typing import Generator, Iterable, List, Tuple
from .gcd import gcd
from ..sqrt import fsqrt
def convergent_sequence(generator: Iterable[int]) -> \
Generator[Fraction, None, None]:
h = (0, 1)
k = (1, 0)
for a in generator:
h = h[1], a * h[1] + h[0]
k = k[1], a * k[1] + k[0]
yield Fraction(h[-1], k[-1])
def continued_fraction_sqrt(n: int) -> Tuple[List[int], List[int]]:
sqrt_n = sqrt(n)
remainders = []
remainder = (0, 1)
# remainder is an + (sqrt(n) - p) / q and these are initial.
continued_fraction = []
while remainder not in remainders:
remainders.append(remainder)
p, q = remainder
q = (n - (p * p)) // q
a = int((sqrt_n + p) / q)
p = a * q - p
continued_fraction.append(a)
remainder = (p, q)
index = remainders.index(remainder)
return continued_fraction[1:index], continued_fraction[index:]
def convergents_sqrt(n: int) -> Generator[Fraction, None, None]:
initial, repeat = continued_fraction_sqrt(n)
convergents = convergent_sequence(chain(initial, cycle(repeat)))
yield from convergents
|
Make continued fractions sqrt much faster
|
Make continued fractions sqrt much faster
|
Python
|
mit
|
cryvate/project-euler,cryvate/project-euler
|
36df41cf3f5345ab599b5a748562aec2af414239
|
python/crypto-square/crypto_square.py
|
python/crypto-square/crypto_square.py
|
import string
import math
import itertools
class CryptoSquare:
@classmethod
def encode(cls, msg):
if len(cls.normalize(msg)) == 0:
return ''
return ' '.join(cls.transpose_square(cls.squarify(cls.normalize(msg))))
@classmethod
def squarify(cls, msg):
return [msg[i:i + cls.square_size(len(msg))]
for i in range(0, len(msg), cls.square_size(len(msg)))]
@classmethod
def transpose_square(cls, square):
matrix = [list(row) for row in square]
transposed_matrix = cls.filter_out_none(cls.transpose_uneven_matrix(matrix))
return [''.join(row) for row in transposed_matrix]
@staticmethod
def normalize(msg):
return ''.join(ch.lower() for ch in msg if ch not in
set(string.punctuation + ' '))
@staticmethod
def square_size(msg_length):
return int(math.ceil(msg_length ** 0.5))
# https://stackoverflow.com/a/4938130/2813210
@staticmethod
def transpose_uneven_matrix(matrix):
return list(itertools.zip_longest(*matrix))
@staticmethod
def filter_out_none(matrix):
return [[val for val in row if val is not None] for row in matrix]
def encode(msg):
return CryptoSquare.encode(msg)
|
import string
import math
import itertools
class CryptoSquare:
@classmethod
def encode(cls, msg):
if len(cls.normalize(msg)) == 0:
return ''
return ' '.join(cls.transpose_square(cls.squarify(cls.normalize(msg))))
@classmethod
def squarify(cls, msg):
return [msg[i:i + cls.square_size(len(msg))]
for i in range(0, len(msg), cls.square_size(len(msg)))]
@classmethod
def transpose_square(cls, square):
matrix = [list(row) for row in square]
transposed_matrix = cls.transpose_uneven_matrix(matrix)
return [''.join(row) for row in transposed_matrix]
@staticmethod
def normalize(msg):
return ''.join(ch.lower() for ch in msg if ch not in
set(string.punctuation + ' '))
@staticmethod
def square_size(msg_length):
return int(math.ceil(msg_length ** 0.5))
# https://stackoverflow.com/a/4938130/2813210
@staticmethod
def transpose_uneven_matrix(matrix):
transposed_matrix = list(itertools.zip_longest(*matrix))
return [[val for val in row if val is not None] for row in transposed_matrix] # Remove None's
def encode(msg):
return CryptoSquare.encode(msg)
|
Clean up transpose helper method
|
Clean up transpose helper method
|
Python
|
mit
|
rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism
|
c8301f1e3165a5e5eaac46de9bdf97c4c1109718
|
dht.py
|
dht.py
|
#!/usr/bin/env python
import time
import thread
import Adafruit_DHT as dht
import config
h = 0.0
t = 0.0
def get_ht_thread():
while True:
ht = dht.read_retry(dht.DHT22, config.DHT22_GPIO_NUM)
h = '{0:0.1f}'.format(ht[0])
t = '{0:0.1f}'.format(ht[1])
time.sleep(2)
def get_ht():
return (h, t)
thread.start_new_thread(get_ht_thread, ())
if __name__ == '__main__':
ht = get_ht()
print('The humidity and temperature:')
print(ht)
|
#!/usr/bin/env python
import time
import thread
import Adafruit_DHT as dht
import config
h = 0.0
t = 0.0
def get_ht_thread():
global h
global t
while True:
ht = dht.read_retry(dht.DHT22, config.DHT22_GPIO_NUM)
h = '{0:0.1f}'.format(ht[0])
t = '{0:0.1f}'.format(ht[1])
time.sleep(2)
def get_ht():
return (h, t)
thread.start_new_thread(get_ht_thread, ())
if __name__ == '__main__':
ht = get_ht()
print('The humidity and temperature:')
print(ht)
|
Fix a DHT reading error
|
Fix a DHT reading error
|
Python
|
mit
|
yunbademo/yunba-smarthome,yunbademo/yunba-smarthome
|
b86d23b0302bb4d0efa2aa203883a78d3dcbf26e
|
scipy/integrate/_ivp/tests/test_rk.py
|
scipy/integrate/_ivp/tests/test_rk.py
|
import pytest
from numpy.testing import assert_allclose
import numpy as np
from scipy.integrate import RK23, RK45, DOP853
from scipy.integrate._ivp import dop853_coefficients
@pytest.mark.parametrize("solver", [RK23, RK45, DOP853])
def test_coefficient_properties(solver):
assert_allclose(np.sum(solver.B), 1, rtol=1e-15)
assert_allclose(np.sum(solver.A, axis=1), solver.C, rtol=1e-14)
def test_coefficient_properties_dop853():
assert_allclose(np.sum(dop853_coefficients.B), 1, rtol=1e-15)
assert_allclose(np.sum(dop853_coefficients.A, axis=1),
dop853_coefficients.C,
rtol=1e-14)
|
import pytest
from numpy.testing import assert_allclose, assert_
import numpy as np
from scipy.integrate import RK23, RK45, DOP853
from scipy.integrate._ivp import dop853_coefficients
@pytest.mark.parametrize("solver", [RK23, RK45, DOP853])
def test_coefficient_properties(solver):
assert_allclose(np.sum(solver.B), 1, rtol=1e-15)
assert_allclose(np.sum(solver.A, axis=1), solver.C, rtol=1e-14)
def test_coefficient_properties_dop853():
assert_allclose(np.sum(dop853_coefficients.B), 1, rtol=1e-15)
assert_allclose(np.sum(dop853_coefficients.A, axis=1),
dop853_coefficients.C,
rtol=1e-14)
@pytest.mark.parametrize("solver_class", [RK23, RK45, DOP853])
def test_error_estimation(solver_class):
step = 0.2
solver = solver_class(lambda t, y: y, 0, [1], 1, first_step=step)
solver.step()
error_estimate = solver._estimate_errors(solver.K, step)
error = solver.y - np.exp([step])
assert_(np.abs(error) < np.abs(error_estimate))
|
Test of error estimation of Runge-Kutta methods
|
TST: Test of error estimation of Runge-Kutta methods
|
Python
|
bsd-3-clause
|
jor-/scipy,zerothi/scipy,mdhaber/scipy,anntzer/scipy,ilayn/scipy,Eric89GXL/scipy,mdhaber/scipy,matthew-brett/scipy,endolith/scipy,jor-/scipy,anntzer/scipy,grlee77/scipy,vigna/scipy,mdhaber/scipy,andyfaff/scipy,aarchiba/scipy,aeklant/scipy,tylerjereddy/scipy,aeklant/scipy,andyfaff/scipy,perimosocordiae/scipy,tylerjereddy/scipy,aeklant/scipy,person142/scipy,jamestwebber/scipy,e-q/scipy,rgommers/scipy,endolith/scipy,perimosocordiae/scipy,jor-/scipy,Stefan-Endres/scipy,pizzathief/scipy,mdhaber/scipy,rgommers/scipy,perimosocordiae/scipy,lhilt/scipy,Eric89GXL/scipy,grlee77/scipy,person142/scipy,matthew-brett/scipy,aarchiba/scipy,gertingold/scipy,mdhaber/scipy,person142/scipy,pizzathief/scipy,matthew-brett/scipy,jamestwebber/scipy,scipy/scipy,perimosocordiae/scipy,aarchiba/scipy,endolith/scipy,Eric89GXL/scipy,Stefan-Endres/scipy,tylerjereddy/scipy,perimosocordiae/scipy,e-q/scipy,nmayorov/scipy,pizzathief/scipy,jor-/scipy,arokem/scipy,vigna/scipy,gertingold/scipy,andyfaff/scipy,scipy/scipy,pizzathief/scipy,matthew-brett/scipy,aeklant/scipy,scipy/scipy,rgommers/scipy,jamestwebber/scipy,Stefan-Endres/scipy,lhilt/scipy,nmayorov/scipy,jor-/scipy,mdhaber/scipy,vigna/scipy,grlee77/scipy,Eric89GXL/scipy,Eric89GXL/scipy,andyfaff/scipy,ilayn/scipy,jamestwebber/scipy,person142/scipy,gertingold/scipy,nmayorov/scipy,nmayorov/scipy,nmayorov/scipy,perimosocordiae/scipy,ilayn/scipy,WarrenWeckesser/scipy,arokem/scipy,lhilt/scipy,WarrenWeckesser/scipy,Stefan-Endres/scipy,person142/scipy,jamestwebber/scipy,scipy/scipy,zerothi/scipy,anntzer/scipy,Stefan-Endres/scipy,scipy/scipy,scipy/scipy,grlee77/scipy,zerothi/scipy,ilayn/scipy,endolith/scipy,zerothi/scipy,rgommers/scipy,zerothi/scipy,endolith/scipy,tylerjereddy/scipy,WarrenWeckesser/scipy,andyfaff/scipy,lhilt/scipy,rgommers/scipy,gertingold/scipy,e-q/scipy,pizzathief/scipy,vigna/scipy,ilayn/scipy,aarchiba/scipy,e-q/scipy,WarrenWeckesser/scipy,matthew-brett/scipy,lhilt/scipy,arokem/scipy,zerothi/scipy,WarrenWeckesser/scipy,aeklant/scipy,ilayn/scipy,grlee77/scipy,tylerjereddy/scipy,Stefan-Endres/scipy,e-q/scipy,Eric89GXL/scipy,andyfaff/scipy,anntzer/scipy,endolith/scipy,vigna/scipy,anntzer/scipy,aarchiba/scipy,arokem/scipy,WarrenWeckesser/scipy,anntzer/scipy,arokem/scipy,gertingold/scipy
|
81dfb5cb952fbca90882bd39e76887f0fa6479eb
|
msmexplorer/tests/test_msm_plot.py
|
msmexplorer/tests/test_msm_plot.py
|
import numpy as np
from msmbuilder.msm import MarkovStateModel, BayesianMarkovStateModel
from matplotlib.axes import SubplotBase
from seaborn.apionly import JointGrid
from ..plots import plot_pop_resids, plot_msm_network, plot_timescales
rs = np.random.RandomState(42)
data = rs.randint(low=0, high=10, size=100000)
msm = MarkovStateModel()
msm.fit(data)
bmsm = BayesianMarkovStateModel()
bmsm.fit(data)
def test_plot_pop_resids():
ax = plot_pop_resids(msm)
assert isinstance(ax, JointGrid)
def test_plot_msm_network():
ax = plot_msm_network(msm)
assert isinstance(ax, SubplotBase)
def test_plot_timescales_msm():
ax = plot_timescales(msm, n_timescales=3, xlabel='x', ylabel='y')
assert isinstance(ax, SubplotBase)
def test_plot_timescales_bmsm():
ax = plot_timescales(bmsm)
assert isinstance(ax, SubplotBase)
|
import numpy as np
from msmbuilder.msm import MarkovStateModel, BayesianMarkovStateModel
from matplotlib.axes import SubplotBase
from seaborn.apionly import JointGrid
from ..plots import plot_pop_resids, plot_msm_network, plot_timescales, plot_implied_timescales
rs = np.random.RandomState(42)
data = rs.randint(low=0, high=10, size=100000)
msm = MarkovStateModel()
msm.fit(data)
bmsm = BayesianMarkovStateModel()
bmsm.fit(data)
def test_plot_pop_resids():
ax = plot_pop_resids(msm)
assert isinstance(ax, JointGrid)
def test_plot_msm_network():
ax = plot_msm_network(msm)
assert isinstance(ax, SubplotBase)
def test_plot_timescales_msm():
ax = plot_timescales(msm, n_timescales=3, xlabel='x', ylabel='y')
assert isinstance(ax, SubplotBase)
def test_plot_timescales_bmsm():
ax = plot_timescales(bmsm)
assert isinstance(ax, SubplotBase)
def test_plot_implied_timescales():
lag_times = [1, 10, 50, 100, 200, 250, 500]
msm_objs = []
for lag in lag_times:
# Construct MSM
msm = MarkovStateModel(lag_time=lag, n_timescales=5)
msm.fit(clustered_trajs)
msm_objs.append(msm)
ax = plot_implied_timescales(msm_objs)
assert isinstance(ax, SubplotBase)
|
Add test for implied timescales plot
|
Add test for implied timescales plot
|
Python
|
mit
|
msmexplorer/msmexplorer,msmexplorer/msmexplorer
|
5f39fd311c735593ac41ba17a060f9cadbe80e18
|
nlpipe/scripts/amcat_background.py
|
nlpipe/scripts/amcat_background.py
|
"""
Assign articles from AmCAT sets for background processing in nlpipe
"""
import sys, argparse
from nlpipe import tasks
from nlpipe.pipeline import parse_background
from nlpipe.backend import get_input_ids
from nlpipe.celery import app
modules = {n.split(".")[-1]: t for (n,t) in app.tasks.iteritems() if n.startswith("nlpipe")}
parser = argparse.ArgumentParser(description=__doc__)
parser.add_argument('module', help='nlpipe module (task) name ({})'.format(", ".join(sorted(modules))),
choices=modules, metavar="module")
parser.add_argument('sets', type=int, nargs='+', help='Article set id(s)')
parser.add_argument('--max', type=int, help='maximum number of articles to assign')
parser.add_argument('--queue', default='background', help='Celery queue to put the articles on')
args = parser.parse_args()
task = modules[args.module]
body = {u'filter': {'terms': {u'sets': args.sets}}}
print("Assigning {max} articles from set(s) {args.sets} for processing by {task.name}"
.format(max=("up to {}".format(args.max) if args.max is not None else "all"), **locals()))
ids = list(get_input_ids(body))
parse_background(ids, task, max=args.max, queue=args.queue)
|
"""
Assign articles from AmCAT sets for background processing in nlpipe
"""
import sys, argparse
from nlpipe import tasks
from nlpipe.pipeline import parse_background
from nlpipe.backend import get_input_ids
from nlpipe.celery import app
import logging
FORMAT = '[%(asctime)-15s] %(message)s'
logging.basicConfig(format=FORMAT, level=logging.INFO)
modules = {n.split(".")[-1]: t for (n,t) in app.tasks.iteritems() if n.startswith("nlpipe")}
parser = argparse.ArgumentParser(description=__doc__)
parser.add_argument('module', help='nlpipe module (task) name ({})'.format(", ".join(sorted(modules))),
choices=modules, metavar="module")
parser.add_argument('sets', type=int, nargs='+', help='Article set id(s)')
parser.add_argument('--max', type=int, help='maximum number of articles to assign')
parser.add_argument('--queue', default='background', help='Celery queue to put the articles on')
args = parser.parse_args()
task = modules[args.module]
body = {u'filter': {'terms': {u'sets': args.sets}}}
logging.info("Assigning {max} articles from set(s) {args.sets} for processing by {task.name}"
.format(max=("up to {}".format(args.max) if args.max is not None else "all"), **locals()))
ids = list(get_input_ids(body))
logging.info("... Found {} articles".format(len(ids)))
parse_background(ids, task, max=args.max, queue=args.queue)
|
Add logging to background assign
|
Add logging to background assign
|
Python
|
mit
|
amcat/nlpipe
|
8c11b2db7f09844aa860bfe7f1c3ff23c0d30f94
|
sentry/migrations/0062_correct_del_index_sentry_groupedmessage_logger__view__checksum.py
|
sentry/migrations/0062_correct_del_index_sentry_groupedmessage_logger__view__checksum.py
|
# -*- coding: utf-8 -*-
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
# FIXES 0015
try:
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
except:
# TODO: we should probaby remove this since it seems to only affect previous installs that
# hadnt applied migrations, and MySQL doesnt like try/excepts (aka South)
pass
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['checksum', 'logger', 'view']
#FIXES 0015
db.create_unique('sentry_groupedmessage', ['checksum', 'logger', 'view'])
complete_apps = ['sentry']
|
# -*- coding: utf-8 -*-
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
pass
def backwards(self, orm):
pass
complete_apps = ['sentry']
|
Remove bad delete_unique call as it was already applied in migration 0015
|
Remove bad delete_unique call as it was already applied in migration 0015
|
Python
|
bsd-3-clause
|
camilonova/sentry,1tush/sentry,vperron/sentry,drcapulet/sentry,fuziontech/sentry,boneyao/sentry,mvaled/sentry,ifduyue/sentry,pauloschilling/sentry,boneyao/sentry,beni55/sentry,Kryz/sentry,beeftornado/sentry,jean/sentry,gg7/sentry,JamesMura/sentry,rdio/sentry,wong2/sentry,songyi199111/sentry,daevaorn/sentry,looker/sentry,kevinastone/sentry,beni55/sentry,gencer/sentry,jokey2k/sentry,BayanGroup/sentry,fuziontech/sentry,JamesMura/sentry,gencer/sentry,TedaLIEz/sentry,zenefits/sentry,BayanGroup/sentry,jean/sentry,NickPresta/sentry,imankulov/sentry,mvaled/sentry,gencer/sentry,JackDanger/sentry,looker/sentry,nicholasserra/sentry,JTCunning/sentry,camilonova/sentry,mvaled/sentry,Natim/sentry,alexm92/sentry,llonchj/sentry,zenefits/sentry,drcapulet/sentry,hongliang5623/sentry,BuildingLink/sentry,wujuguang/sentry,songyi199111/sentry,pauloschilling/sentry,kevinlondon/sentry,vperron/sentry,ngonzalvez/sentry,SilentCircle/sentry,JTCunning/sentry,jean/sentry,TedaLIEz/sentry,gg7/sentry,Natim/sentry,camilonova/sentry,fotinakis/sentry,argonemyth/sentry,ewdurbin/sentry,jean/sentry,imankulov/sentry,ifduyue/sentry,1tush/sentry,BuildingLink/sentry,fotinakis/sentry,SilentCircle/sentry,jean/sentry,NickPresta/sentry,NickPresta/sentry,gg7/sentry,wujuguang/sentry,BuildingLink/sentry,looker/sentry,fotinakis/sentry,daevaorn/sentry,pauloschilling/sentry,BuildingLink/sentry,rdio/sentry,daevaorn/sentry,1tush/sentry,nicholasserra/sentry,fotinakis/sentry,llonchj/sentry,rdio/sentry,kevinlondon/sentry,zenefits/sentry,looker/sentry,ifduyue/sentry,hongliang5623/sentry,BayanGroup/sentry,SilentCircle/sentry,chayapan/django-sentry,kevinastone/sentry,felixbuenemann/sentry,JamesMura/sentry,boneyao/sentry,Natim/sentry,mitsuhiko/sentry,hongliang5623/sentry,Kryz/sentry,korealerts1/sentry,imankulov/sentry,alexm92/sentry,wujuguang/sentry,vperron/sentry,JTCunning/sentry,beeftornado/sentry,JamesMura/sentry,ifduyue/sentry,ewdurbin/sentry,felixbuenemann/sentry,mitsuhiko/sentry,wong2/sentry,JackDanger/sentry,chayapan/django-sentry,nicholasserra/sentry,SilentCircle/sentry,beeftornado/sentry,drcapulet/sentry,TedaLIEz/sentry,gencer/sentry,ewdurbin/sentry,gencer/sentry,BuildingLink/sentry,korealerts1/sentry,felixbuenemann/sentry,Kryz/sentry,jokey2k/sentry,chayapan/django-sentry,kevinastone/sentry,JackDanger/sentry,songyi199111/sentry,daevaorn/sentry,ngonzalvez/sentry,mvaled/sentry,looker/sentry,ifduyue/sentry,fuziontech/sentry,JamesMura/sentry,argonemyth/sentry,zenefits/sentry,mvaled/sentry,zenefits/sentry,ngonzalvez/sentry,mvaled/sentry,korealerts1/sentry,rdio/sentry,jokey2k/sentry,kevinlondon/sentry,alexm92/sentry,argonemyth/sentry,wong2/sentry,llonchj/sentry,NickPresta/sentry,beni55/sentry
|
457f2d1d51b2bf008f837bf3ce8ee3cb47d5ba6b
|
var/spack/packages/libpng/package.py
|
var/spack/packages/libpng/package.py
|
from spack import *
class Libpng(Package):
"""libpng graphics file format"""
homepage = "http://www.libpng.org/pub/png/libpng.html"
url = "http://sourceforge.net/projects/libpng/files/libpng16/1.6.14/libpng-1.6.14.tar.gz/download"
version('1.6.14', '2101b3de1d5f348925990f9aa8405660')
def install(self, spec, prefix):
configure("--prefix=%s" % prefix)
make()
make("install")
|
from spack import *
class Libpng(Package):
"""libpng graphics file format"""
homepage = "http://www.libpng.org/pub/png/libpng.html"
url = "http://download.sourceforge.net/libpng/libpng-1.6.16.tar.gz"
version('1.6.14', '2101b3de1d5f348925990f9aa8405660')
version('1.6.15', '829a256f3de9307731d4f52dc071916d')
version('1.6.16', '1a4ad377919ab15b54f6cb6a3ae2622d')
def install(self, spec, prefix):
configure("--prefix=%s" % prefix)
make()
make("install")
|
Fix libpng to use a better URL
|
Fix libpng to use a better URL
Sourceforge URLs like this eventually die when the libpng version is bumped:
http://sourceforge.net/projects/libpng/files/libpng16/1.6.14/libpng-1.6.14.tar.gz/download
But ones like this give you a "permanently moved", which curl -L will follow:
http://download.sourceforge.net/libpng/libpng-1.6.16.tar.gz
|
Python
|
lgpl-2.1
|
mfherbst/spack,tmerrick1/spack,iulian787/spack,TheTimmy/spack,tmerrick1/spack,krafczyk/spack,EmreAtes/spack,matthiasdiener/spack,TheTimmy/spack,lgarren/spack,EmreAtes/spack,lgarren/spack,krafczyk/spack,EmreAtes/spack,mfherbst/spack,LLNL/spack,lgarren/spack,krafczyk/spack,krafczyk/spack,skosukhin/spack,TheTimmy/spack,mfherbst/spack,TheTimmy/spack,skosukhin/spack,tmerrick1/spack,matthiasdiener/spack,EmreAtes/spack,krafczyk/spack,tmerrick1/spack,iulian787/spack,lgarren/spack,mfherbst/spack,matthiasdiener/spack,LLNL/spack,LLNL/spack,matthiasdiener/spack,LLNL/spack,mfherbst/spack,iulian787/spack,lgarren/spack,matthiasdiener/spack,TheTimmy/spack,EmreAtes/spack,iulian787/spack,iulian787/spack,skosukhin/spack,LLNL/spack,skosukhin/spack,tmerrick1/spack,skosukhin/spack
|
f4429e49c8b493fa285d169a41b82cb761716705
|
tests/explorers_tests/test_additive_ou.py
|
tests/explorers_tests/test_additive_ou.py
|
from __future__ import unicode_literals
from __future__ import print_function
from __future__ import division
from __future__ import absolute_import
from future import standard_library
standard_library.install_aliases()
import unittest
import numpy as np
from chainerrl.explorers.additive_ou import AdditiveOU
class TestAdditiveOU(unittest.TestCase):
def test(self):
action_size = 3
dt = 0.5
sigma = 0.001
theta = 0.3
def greedy_action_func():
return np.asarray([0] * action_size, dtype=np.float32)
explorer = AdditiveOU(action_size, dt=dt, theta=theta, sigma=sigma)
for t in range(10000):
a = explorer.select_action(t, greedy_action_func)
print(a)
|
from __future__ import unicode_literals
from __future__ import print_function
from __future__ import division
from __future__ import absolute_import
from future import standard_library
standard_library.install_aliases()
import unittest
import numpy as np
from chainerrl.explorers.additive_ou import AdditiveOU
class TestAdditiveOU(unittest.TestCase):
def test(self):
action_size = 3
def greedy_action_func():
return np.asarray([0] * action_size, dtype=np.float32)
explorer = AdditiveOU()
for t in range(100):
a = explorer.select_action(t, greedy_action_func)
print(t, a)
|
Fix a test for AdditiveOU
|
Fix a test for AdditiveOU
|
Python
|
mit
|
toslunar/chainerrl,toslunar/chainerrl
|
bea258e2affc165f610de83248d9f958eec1ef4e
|
cmsplugin_markdown/models.py
|
cmsplugin_markdown/models.py
|
from django.db import models
from cms.models import CMSPlugin
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
|
from django.db import models
from cms.models import CMSPlugin
from cms.utils.compat.dj import python_2_unicode_compatible
@python_2_unicode_compatible
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
def __str__(self):
text = self.markdown_text
return (text[:50] + '...') if len(text) > 53 else text
|
Add __str__ method for better representation in frontend
|
Add __str__ method for better representation in frontend
|
Python
|
mit
|
bitmazk/cmsplugin-markdown,bitmazk/cmsplugin-markdown,bitmazk/cmsplugin-markdown
|
9828e5125cdbc01a773c60b1e211d0e434a2c5aa
|
tests/test_modules/test_pmac/test_pmacstatuspart.py
|
tests/test_modules/test_pmac/test_pmacstatuspart.py
|
from malcolm.core import Process
from malcolm.modules.builtin.controllers import ManagerController
from malcolm.modules.pmac.blocks import pmac_status_block
from malcolm.modules.pmac.parts import PmacStatusPart
from malcolm.testutil import ChildTestCase
class TestPmacStatusPart(ChildTestCase):
def setUp(self):
self.process = Process("Process")
child = self.create_child_block(
pmac_status_block, self.process, mri="my_mri", pv_prefix="PV:PRE"
)
self.set_attributes(child, i10=1705244)
c = ManagerController("PMAC", "/tmp", use_git=False)
self.o = PmacStatusPart(name="part", mri="my_mri", initial_visibility=True)
c.add_part(self.o)
self.process.add_controller(c)
self.process.start()
self.b = c.block_view()
def tearDown(self):
self.process.stop(timeout=1)
def test_servo_freq(self):
freq = self.b.servoFrequency()
assert freq == 4919.300698316487
|
from malcolm.core import Process
from malcolm.modules.builtin.controllers import ManagerController
from malcolm.modules.pmac.blocks import pmac_status_block
from malcolm.modules.pmac.parts import PmacStatusPart
from malcolm.testutil import ChildTestCase
class TestPmacStatusPart(ChildTestCase):
def setUp(self):
self.process = Process("Process")
child = self.create_child_block(
pmac_status_block, self.process, mri="my_mri", pv_prefix="PV:PRE"
)
self.set_attributes(child, servoFreq=2500.04)
c = ManagerController("PMAC", "/tmp", use_git=False)
self.o = PmacStatusPart(name="part", mri="my_mri", initial_visibility=True)
c.add_part(self.o)
self.process.add_controller(c)
self.process.start()
self.b = c.block_view()
def tearDown(self):
self.process.stop(timeout=1)
def test_servo_freq(self):
freq = self.b.servoFrequency()
assert freq == 2500.04
|
Change TestPmacStatusPart to not use i10
|
Change TestPmacStatusPart to not use i10
|
Python
|
apache-2.0
|
dls-controls/pymalcolm,dls-controls/pymalcolm,dls-controls/pymalcolm
|
58dbfa0b449b8e4171c5f9cef1c15db39b52c1f0
|
tests/run_tests.py
|
tests/run_tests.py
|
#!/usr/bin/env python
import os.path
import sys
import subprocess
import unittest
tests_dir = os.path.dirname(__file__)
sys.path.insert(0, os.path.dirname(tests_dir))
import secretstorage
if __name__ == '__main__':
major, minor, patch = sys.version_info[:3]
print('Running with Python %d.%d.%d (SecretStorage from %s)' %
(major, minor, patch, os.path.dirname(secretstorage.__file__)))
mock = None
if len(sys.argv) > 1 and os.path.isfile(sys.argv[1]):
mock = subprocess.Popen(('/usr/bin/python3', sys.argv[1],),
stdout=subprocess.PIPE,
universal_newlines=True)
bus_name = mock.stdout.readline().rstrip()
secretstorage.util.BUS_NAME = bus_name
print('Bus name set to %r' % secretstorage.util.BUS_NAME)
loader = unittest.TestLoader()
runner = unittest.TextTestRunner(verbosity=2)
result = runner.run(loader.discover(tests_dir))
if mock is not None:
mock.terminate()
sys.exit(not result.wasSuccessful())
|
#!/usr/bin/env python
import os.path
import sys
import subprocess
import unittest
tests_dir = os.path.dirname(__file__)
sys.path.insert(0, os.path.dirname(tests_dir))
import secretstorage
if __name__ == '__main__':
major, minor, patch = sys.version_info[:3]
print('Running with Python %d.%d.%d (SecretStorage from %s)' %
(major, minor, patch, os.path.dirname(secretstorage.__file__)))
mock = None
if len(sys.argv) > 1 and os.path.isfile(sys.argv[1]):
mock = subprocess.Popen(('/usr/bin/python3', sys.argv[1],),
stdout=subprocess.PIPE,
universal_newlines=True)
assert mock.stdout is not None # for mypy
bus_name = mock.stdout.readline().rstrip()
secretstorage.util.BUS_NAME = bus_name
print('Bus name set to %r' % secretstorage.util.BUS_NAME)
loader = unittest.TestLoader()
runner = unittest.TextTestRunner(verbosity=2)
result = runner.run(loader.discover(tests_dir))
if mock is not None:
mock.terminate()
sys.exit(not result.wasSuccessful())
|
Add an assert to make mypy check pass again
|
Add an assert to make mypy check pass again
|
Python
|
bsd-3-clause
|
mitya57/secretstorage
|
99496d97f3e00284840d2127556bba0e21d1a99e
|
frappe/tests/test_commands.py
|
frappe/tests/test_commands.py
|
# Copyright (c) 2020, Frappe Technologies Pvt. Ltd. and Contributors
from __future__ import unicode_literals
import shlex
import subprocess
import unittest
import frappe
def clean(value):
if isinstance(value, (bytes, str)):
value = value.decode().strip()
return value
class BaseTestCommands:
def execute(self, command):
command = command.format(**{"site": frappe.local.site})
command = shlex.split(command)
self._proc = subprocess.run(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
self.stdout = clean(self._proc.stdout)
self.stderr = clean(self._proc.stderr)
self.returncode = clean(self._proc.returncode)
|
# Copyright (c) 2020, Frappe Technologies Pvt. Ltd. and Contributors
from __future__ import unicode_literals
import shlex
import subprocess
import unittest
import frappe
def clean(value):
if isinstance(value, (bytes, str)):
value = value.decode().strip()
return value
class BaseTestCommands:
def execute(self, command):
command = command.format(**{"site": frappe.local.site})
command = shlex.split(command)
self._proc = subprocess.run(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
self.stdout = clean(self._proc.stdout)
self.stderr = clean(self._proc.stderr)
self.returncode = clean(self._proc.returncode)
class TestCommands(BaseTestCommands, unittest.TestCase):
def test_execute(self):
# execute a command expecting a numeric output
self.execute("bench --site {site} execute frappe.db.get_database_size")
self.assertEquals(self.returncode, 0)
self.assertIsInstance(float(self.stdout), float)
# execute a command expecting an errored output as local won't exist
self.execute("bench --site {site} execute frappe.local.site")
self.assertEquals(self.returncode, 1)
self.assertIsNotNone(self.stderr)
# execute a command with kwargs
self.execute("""bench --site {site} execute frappe.bold --kwargs '{{"text": "DocType"}}'""")
self.assertEquals(self.returncode, 0)
self.assertEquals(self.stdout[1:-1], frappe.bold(text='DocType'))
|
Add tests for bench execute
|
test: Add tests for bench execute
|
Python
|
mit
|
saurabh6790/frappe,StrellaGroup/frappe,adityahase/frappe,mhbu50/frappe,adityahase/frappe,yashodhank/frappe,mhbu50/frappe,yashodhank/frappe,mhbu50/frappe,mhbu50/frappe,StrellaGroup/frappe,saurabh6790/frappe,yashodhank/frappe,frappe/frappe,almeidapaulopt/frappe,almeidapaulopt/frappe,StrellaGroup/frappe,adityahase/frappe,frappe/frappe,saurabh6790/frappe,almeidapaulopt/frappe,yashodhank/frappe,adityahase/frappe,saurabh6790/frappe,almeidapaulopt/frappe,frappe/frappe
|
fac280a022c8728f14bbe1194cf74af761b7ec3f
|
vfp2py/__main__.py
|
vfp2py/__main__.py
|
import argparse
import vfp2py
def parse_args(argv=None):
parser = argparse.ArgumentParser(description='Tool for rewriting Foxpro code in Python')
parser.add_argument("infile", help="file to convert", type=str)
parser.add_argument("outfile", help="file to output to", type=str)
parser.add_argument("search", help="directories to search for included files", type=str, nargs='*')
return parser.parse_args(argv)
def main(argv=None):
args = parse_args(argv)
global SEARCH_PATH
SEARCH_PATH = args.search
vfp2py.convert_file(args.infile, args.outfile)
if __name__ == '__main__':
try:
main()
except KeyboardInterrupt:
pass
|
import argparse
import vfp2py
def parse_args(argv=None):
parser = argparse.ArgumentParser(description='Tool for rewriting Foxpro code in Python')
parser.add_argument("infile", help="file to convert", type=str)
parser.add_argument("outfile", help="file to output to", type=str)
parser.add_argument("search", help="directories to search for included files", type=str, nargs='*')
return parser.parse_args(argv)
def main(argv=None):
args = parse_args(argv)
vfp2py.SEARCH_PATH += args.search
vfp2py.convert_file(args.infile, args.outfile)
if __name__ == '__main__':
try:
main()
except KeyboardInterrupt:
pass
|
Fix search paths not being added from arguments.
|
Fix search paths not being added from arguments.
|
Python
|
mit
|
mwisslead/vfp2py,mwisslead/vfp2py
|
2088b3df274fd31c28baa6193c937046c04b98a6
|
scripts/generate_wiki_languages.py
|
scripts/generate_wiki_languages.py
|
from urllib2 import urlopen
import csv
import lxml.builder as lb
from lxml import etree
# Returns CSV of all wikipedias, ordered by number of 'good' articles
URL = "https://wikistats.wmflabs.org/api.php?action=dump&table=wikipedias&format=csv&s=good"
data = csv.reader(urlopen(URL))
# Column 2 is the language code
lang_keys = [row[2] for row in data]
del lang_keys[0] # Get rid of the headers
# Generate the XML
x = lb.E
keys = [x.item(k) for k in lang_keys]
resources = x.resources(
getattr(x, 'string-array')(*keys, name="preference_language_keys"),
)
open("languages_list.xml", "w").write(
etree.tostring(resources, pretty_print=True, encoding="utf-8", xml_declaration=True)
)
|
from urllib2 import urlopen
import csv
import json
import lxml.builder as lb
from lxml import etree
# Returns CSV of all wikipedias, ordered by number of 'good' articles
URL = "https://wikistats.wmflabs.org/api.php?action=dump&table=wikipedias&format=csv&s=good"
data = csv.reader(urlopen(URL))
lang_keys = []
lang_local_names = []
lang_eng_names = []
for row in data:
lang_keys.append(row[2])
lang_local_names.append(row[10])
lang_eng_names.append(row[1])
# Generate the XML, for Android
x = lb.E
keys = [x.item(k) for k in lang_keys]
# Skip the headers!
del keys[0]
resources = x.resources(
getattr(x, 'string-array')(*keys, name="preference_language_keys"),
)
open("languages_list.xml", "w").write(
etree.tostring(resources, pretty_print=True, encoding="utf-8", xml_declaration=True)
)
# Generate the JSON, for iOS
langs_json = []
# Start from 1, to skip the headers
for i in xrange(1, len(lang_keys)):
langs_json.append({
"code": lang_keys[i],
"name": lang_local_names[i],
"canonical_name": lang_eng_names[i]
})
open("languages_list.json", "w").write(json.dumps(langs_json, indent=4))
|
Modify language generation script to make JSON for iOS
|
Modify language generation script to make JSON for iOS
Change-Id: Ib5aec2f6cfcb5bd1187cf8863ecd50f1b1a2d20c
|
Python
|
apache-2.0
|
Wikinaut/wikipedia-app,carloshwa/apps-android-wikipedia,dbrant/apps-android-wikipedia,creaITve/apps-android-tbrc-works,reproio/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,reproio/apps-android-wikipedia,wikimedia/apps-android-wikipedia,BrunoMRodrigues/apps-android-tbrc-work,BrunoMRodrigues/apps-android-tbrc-work,carloshwa/apps-android-wikipedia,creaITve/apps-android-tbrc-works,BrunoMRodrigues/apps-android-tbrc-work,Wikinaut/wikipedia-app,Wikinaut/wikipedia-app,BrunoMRodrigues/apps-android-tbrc-work,wikimedia/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,carloshwa/apps-android-wikipedia,wikimedia/apps-android-wikipedia,Wikinaut/wikipedia-app,parvez3019/apps-android-wikipedia,carloshwa/apps-android-wikipedia,Duct-and-rice/KrswtkhrWiki4Android,anirudh24seven/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,parvez3019/apps-android-wikipedia,dbrant/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,dbrant/apps-android-wikipedia,reproio/apps-android-wikipedia,reproio/apps-android-wikipedia,creaITve/apps-android-tbrc-works,anirudh24seven/apps-android-wikipedia,dbrant/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,Duct-and-rice/KrswtkhrWiki4Android,parvez3019/apps-android-wikipedia,parvez3019/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,creaITve/apps-android-tbrc-works,wikimedia/apps-android-wikipedia,Duct-and-rice/KrswtkhrWiki4Android,SAGROUP2/apps-android-wikipedia,parvez3019/apps-android-wikipedia,carloshwa/apps-android-wikipedia,dbrant/apps-android-wikipedia,reproio/apps-android-wikipedia,Duct-and-rice/KrswtkhrWiki4Android,Duct-and-rice/KrswtkhrWiki4Android,Wikinaut/wikipedia-app
|
f340c674737431c15875007f92de4dbe558ba377
|
molo/yourwords/templatetags/competition_tag.py
|
molo/yourwords/templatetags/competition_tag.py
|
from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.live().child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
competitions = []
context.update({
'competitions': [
a.get_translation_for(locale_code) or a for a in competitions]
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.live().child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return [a.get_translation_for(locale) or a for a in qs]
else:
return []
|
from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
from molo.core.core_tags import get_pages
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
competitions = []
context.update({
'competitions': get_pages(context, competitions, locale_code)
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return get_pages(context, qs, locale)
else:
return []
|
Add support for hiding untranslated content
|
Add support for hiding untranslated content
|
Python
|
bsd-2-clause
|
praekelt/molo.yourwords,praekelt/molo.yourwords
|
abdd6d6e75fb7c6f9cff4b42f6b12a2cfb7a342a
|
fpsd/test/test_sketchy_sites.py
|
fpsd/test/test_sketchy_sites.py
|
#!/usr/bin/env python3.5
# This test crawls some sets that have triggered http.client.RemoteDisconnected
# exceptions
import unittest
from crawler import Crawler
class CrawlBadSitesTest(unittest.TestCase):
bad_sites = ["http://jlve2diknf45qwjv.onion/",
"http://money2mxtcfcauot.onion",
"http://22222222aziwzse2.onion"]
def test_crawl_of_bad_sites(self):
with Crawler() as crawler:
crawler.collect_set_of_traces(self.bad_sites, shuffle=False)
if __name__ == "__main__":
unittest.main()
|
#!/usr/bin/env python3.5
# This test crawls some sets that have triggered http.client.RemoteDisconnected
# exceptions
import unittest
from crawler import Crawler
class CrawlBadSitesTest(unittest.TestCase):
bad_sites = ["http://jlve2diknf45qwjv.onion/",
"http://money2mxtcfcauot.onion",
"http://22222222aziwzse2.onion"]
def test_crawl_of_bad_sites(self):
with Crawler(restart_on_sketchy_exception=True) as crawler:
crawler.collect_set_of_traces(self.bad_sites)
if __name__ == "__main__":
unittest.main()
|
Use known-to-trigger-exceptions sites to test crawler restart method
|
Use known-to-trigger-exceptions sites to test crawler restart method
|
Python
|
agpl-3.0
|
freedomofpress/fingerprint-securedrop,freedomofpress/FingerprintSecureDrop,freedomofpress/fingerprint-securedrop,freedomofpress/fingerprint-securedrop,freedomofpress/FingerprintSecureDrop
|
053147c19acbf467bb0e044f2fb58304b759b72d
|
frameworks/Python/pyramid/create_database.py
|
frameworks/Python/pyramid/create_database.py
|
import codecs
from frameworkbenchmarks.models import DBSession
if __name__ == "__main__":
"""
Initialize database
"""
with codecs.open('../config/create-postgres.sql', 'r', encoding='utf-8') as fp:
sql = fp.read()
DBSession.execute(sql)
DBSession.commit()
|
import codecs
from frameworkbenchmarks.models import DBSession
if __name__ == "__main__":
"""
Initialize database
"""
with codecs.open('../../../config/create-postgres.sql',
'r',
encoding='utf-8') as fp:
sql = fp.read()
DBSession.execute(sql)
DBSession.commit()
|
Fix the path to create-postgres.sql
|
Fix the path to create-postgres.sql
|
Python
|
bsd-3-clause
|
k-r-g/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,sxend/FrameworkBenchmarks,doom369/FrameworkBenchmarks,herloct/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,actframework/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,zloster/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,khellang/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Verber/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,methane/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,denkab/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,Verber/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,testn/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,sxend/FrameworkBenchmarks,grob/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,doom369/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,denkab/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zloster/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,khellang/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,joshk/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,zapov/FrameworkBenchmarks,zloster/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,doom369/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,valyala/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,methane/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zloster/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,grob/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,testn/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,joshk/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,grob/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zloster/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,Verber/FrameworkBenchmarks,khellang/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,zapov/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,actframework/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,joshk/FrameworkBenchmarks,denkab/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,valyala/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,joshk/FrameworkBenchmarks,doom369/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,doom369/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jamming/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Verber/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,grob/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,grob/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,actframework/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,doom369/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,denkab/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,sxend/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,grob/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,grob/FrameworkBenchmarks,methane/FrameworkBenchmarks,joshk/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,valyala/FrameworkBenchmarks,jamming/FrameworkBenchmarks,sxend/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,methane/FrameworkBenchmarks,sxend/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,zloster/FrameworkBenchmarks,testn/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,grob/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,jamming/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,jamming/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,valyala/FrameworkBenchmarks,herloct/FrameworkBenchmarks,khellang/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,valyala/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,methane/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,denkab/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,denkab/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,methane/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,zapov/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Verber/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,testn/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zloster/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,jamming/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,valyala/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,valyala/FrameworkBenchmarks,doom369/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,herloct/FrameworkBenchmarks,valyala/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,denkab/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,methane/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sgml/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zloster/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,sgml/FrameworkBenchmarks,testn/FrameworkBenchmarks,khellang/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Verber/FrameworkBenchmarks,denkab/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,methane/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Verber/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,zloster/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,testn/FrameworkBenchmarks,testn/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,valyala/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,sgml/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,khellang/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,methane/FrameworkBenchmarks,Verber/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,joshk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,joshk/FrameworkBenchmarks,denkab/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Verber/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,sgml/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,khellang/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,denkab/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,testn/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zapov/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,valyala/FrameworkBenchmarks,jamming/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,joshk/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,actframework/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,sxend/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,jamming/FrameworkBenchmarks,joshk/FrameworkBenchmarks,zapov/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,sgml/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zapov/FrameworkBenchmarks,zloster/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,denkab/FrameworkBenchmarks,actframework/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,jamming/FrameworkBenchmarks,denkab/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,doom369/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,methane/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,actframework/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,testn/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,joshk/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,herloct/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zloster/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,methane/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,sgml/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,sgml/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,khellang/FrameworkBenchmarks,sxend/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,grob/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,grob/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,sgml/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,jamming/FrameworkBenchmarks,sxend/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,zapov/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Verber/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,herloct/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,grob/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,joshk/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,sxend/FrameworkBenchmarks,zloster/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,sgml/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,sgml/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,zapov/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,herloct/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,denkab/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,khellang/FrameworkBenchmarks,actframework/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,actframework/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,sxend/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,actframework/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,actframework/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,jamming/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,jamming/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sxend/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zapov/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,grob/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Verber/FrameworkBenchmarks,actframework/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,joshk/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jamming/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jamming/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,khellang/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,actframework/FrameworkBenchmarks,zloster/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sgml/FrameworkBenchmarks,doom369/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,testn/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,testn/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,grob/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,herloct/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,sxend/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,sxend/FrameworkBenchmarks,doom369/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,methane/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,zloster/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,actframework/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Verber/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zapov/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,actframework/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,zapov/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,zapov/FrameworkBenchmarks,herloct/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,khellang/FrameworkBenchmarks,methane/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,grob/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zapov/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,zloster/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,doom369/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,sxend/FrameworkBenchmarks,sgml/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,methane/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,testn/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,herloct/FrameworkBenchmarks,testn/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zapov/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,valyala/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,khellang/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Verber/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,sgml/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,Verber/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,denkab/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,doom369/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,zloster/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,sxend/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sgml/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,zapov/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,actframework/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,testn/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks
|
310553e1282231c35093ff355c61129e9f073a0a
|
src/lib/verify_email_google.py
|
src/lib/verify_email_google.py
|
import DNS
from validate_email import validate_email
from DNS.Lib import PackError
def is_google_apps_email(email):
hostname = email[email.find('@')+1:]
try:
mx_hosts = DNS.mxlookup(hostname)
except DNS.ServerError as e:
return False
except PackError as e:
return False
for mx in mx_hosts:
if len(mx) == 2:
priority, host_server = mx
else:
host_server = mx
if 'google' in str(host_server).lower() and 'aspmx' in str(host_server).lower():
return True
return False
|
import DNS
import re
from validate_email import validate_email
from DNS.Lib import PackError
EMAIL_RE = re.compile('^[a-zA-Z0-9\.\@]+$')
def is_valid_email(email):
if email.count('@') != 1:
return False
return bool(EMAIL_RE.match(email))
def is_google_apps_email(email):
if not is_valid_email(email):
return False
hostname = email[email.find('@')+1:]
try:
mx_hosts = DNS.mxlookup(hostname)
except DNS.ServerError as e:
return False
except PackError as e:
return False
for mx in mx_hosts:
if len(mx) == 2:
priority, host_server = mx
else:
host_server = mx
if 'google' in str(host_server).lower() and 'aspmx' in str(host_server).lower():
return True
return False
|
Add Google Apps email address validation
|
Add Google Apps email address validation
|
Python
|
agpl-3.0
|
juposocial/jupo,juposocial/jupo,juposocial/jupo,juposocial/jupo
|
0dc1412ad6e7cbe47eda1e476ce16603b7f6a030
|
raspigibbon_bringup/scripts/raspigibbon_joint_subscriber.py
|
raspigibbon_bringup/scripts/raspigibbon_joint_subscriber.py
|
#!/usr/bin/env python
# coding: utf-8
from futaba_serial_servo import RS30X
import rospy
from sensor_msgs.msg import JointState
class Slave:
def __init__(self):
self.rs = RS30X.RS304MD()
self.sub = rospy.Subscriber("/raspigibbon/master_joint_state", JointState, self.joint_callback, queue_size=10)
for i in range(1,6):
self.rs.setTorque(i, True)
rospy.sleep(0.01)
rospy.loginfo("servo initialized")
def joint_callback(self, msg):
for i in range(1, 6):
self.rs.setAngle(i, msg.position[i-1])
rospy.sleep(0.01)
if __name__ == "__main__":
try:
while not rospy.is_shutdown():
rospy.init_node("slave_joint_state")
slave = Slave()
rospy.spin()
except rospy.ROSInterruptException:
pass
|
#!/usr/bin/env python
# coding: utf-8
from futaba_serial_servo import RS30X
import rospy
from sensor_msgs.msg import JointState
class Slave:
def __init__(self):
self.rs = RS30X.RS304MD()
self.sub = rospy.Subscriber("/raspigibbon/master_joint_state", JointState, self.joint_callback, queue_size=10)
for i in range(1,6):
self.rs.setTorque(i, True)
rospy.sleep(0.01)
rospy.loginfo("servo initialized")
def joint_callback(self, msg):
for i in range(1, 6):
self.rs.setAngle(i, msg.position[i-1])
rospy.sleep(0.01)
def shutdown(self):
for i in range(1,6):
self.rs.setTorque(i, False)
rospy.sleep(0.01)
rospy.loginfo("set all servo torque_off")
if __name__ == "__main__":
try:
while not rospy.is_shutdown():
rospy.init_node("slave_joint_state")
rospy.on_shutdown(self.shutdown)
slave = Slave()
rospy.spin()
except rospy.ROSInterruptException:
pass
|
Add shutdown scripts to turn_off servo after subscribing
|
Add shutdown scripts to turn_off servo after subscribing
|
Python
|
mit
|
raspberrypigibbon/raspigibbon_ros
|
cf58ebf492cd0dfaf640d2fd8d3cf4e5b2706424
|
alembic/versions/47dd43c1491_create_category_tabl.py
|
alembic/versions/47dd43c1491_create_category_tabl.py
|
"""create category table
Revision ID: 47dd43c1491
Revises: 27bf0aefa49d
Create Date: 2013-05-21 10:41:43.548449
"""
# revision identifiers, used by Alembic.
revision = '47dd43c1491'
down_revision = '27bf0aefa49d'
from alembic import op
import sqlalchemy as sa
import datetime
def make_timestamp():
now = datetime.datetime.utcnow()
return now.isoformat()
def upgrade():
op.create_table(
'category',
sa.Column('id', sa.Integer, primary_key=True),
sa.Column('name', sa.Text, nullable=False, unique=True),
sa.Column('short_name', sa.Text, nullable=False, unique=True),
sa.Column('created', sa.Text, default=make_timestamp),
)
def downgrade():
op.drop_table('category')
|
"""create category table
Revision ID: 47dd43c1491
Revises: 27bf0aefa49d
Create Date: 2013-05-21 10:41:43.548449
"""
# revision identifiers, used by Alembic.
revision = '47dd43c1491'
down_revision = '27bf0aefa49d'
from alembic import op
import sqlalchemy as sa
import datetime
def make_timestamp():
now = datetime.datetime.utcnow()
return now.isoformat()
def upgrade():
op.create_table(
'category',
sa.Column('id', sa.Integer, primary_key=True),
sa.Column('name', sa.Text, nullable=False, unique=True),
sa.Column('short_name', sa.Text, nullable=False, unique=True),
sa.Column('description', sa.Text, nullable=False),
sa.Column('created', sa.Text, default=make_timestamp),
)
# Add two categories
query = 'INSERT INTO category (name, short_name, description) VALUES (\'Thinking\', \'thinking\', \'Applications where you can help using your skills\')'
op.execute(query)
query = 'INSERT INTO category (name, short_name, description) VALUES (\'Sensing\', \'sensing\', \'Applications where you can help gathering data\')'
op.execute(query)
def downgrade():
op.drop_table('category')
|
Add description to the table and populate it with two categories
|
Add description to the table and populate it with two categories
|
Python
|
agpl-3.0
|
geotagx/geotagx-pybossa-archive,OpenNewsLabs/pybossa,PyBossa/pybossa,proyectos-analizo-info/pybossa-analizo-info,Scifabric/pybossa,CulturePlex/pybossa,geotagx/pybossa,proyectos-analizo-info/pybossa-analizo-info,CulturePlex/pybossa,OpenNewsLabs/pybossa,geotagx/geotagx-pybossa-archive,harihpr/tweetclickers,geotagx/geotagx-pybossa-archive,PyBossa/pybossa,geotagx/pybossa,harihpr/tweetclickers,stefanhahmann/pybossa,stefanhahmann/pybossa,inteligencia-coletiva-lsd/pybossa,proyectos-analizo-info/pybossa-analizo-info,Scifabric/pybossa,jean/pybossa,inteligencia-coletiva-lsd/pybossa,jean/pybossa,CulturePlex/pybossa,geotagx/geotagx-pybossa-archive,geotagx/geotagx-pybossa-archive
|
8b7ab303340ba65aa219103c568ce9d88ea39689
|
airmozilla/main/context_processors.py
|
airmozilla/main/context_processors.py
|
from django.conf import settings
from airmozilla.main.models import Event
def sidebar(request):
featured = Event.objects.approved().filter(public=True, featured=True)
upcoming = Event.objects.upcoming().order_by('start_time')
if not request.user.is_active:
featured = featured.filter(public=True)
upcoming = upcoming.filter(public=True)
upcoming = upcoming[:settings.UPCOMING_SIDEBAR_COUNT]
return {
'upcoming': upcoming,
'featured': featured
}
|
from django.conf import settings
from airmozilla.main.models import Event
def sidebar(request):
featured = Event.objects.approved().filter(featured=True)
upcoming = Event.objects.upcoming().order_by('start_time')
if not request.user.is_active:
featured = featured.filter(public=True)
upcoming = upcoming.filter(public=True)
upcoming = upcoming[:settings.UPCOMING_SIDEBAR_COUNT]
return {
'upcoming': upcoming,
'featured': featured
}
|
Fix context processor to correctly display internal featured videos.
|
Fix context processor to correctly display internal featured videos.
|
Python
|
bsd-3-clause
|
EricSekyere/airmozilla,lcamacho/airmozilla,kenrick95/airmozilla,tannishk/airmozilla,tannishk/airmozilla,a-buck/airmozilla,bugzPDX/airmozilla,ehsan/airmozilla,mythmon/airmozilla,Nolski/airmozilla,blossomica/airmozilla,EricSekyere/airmozilla,blossomica/airmozilla,zofuthan/airmozilla,bugzPDX/airmozilla,EricSekyere/airmozilla,bugzPDX/airmozilla,chirilo/airmozilla,lcamacho/airmozilla,mythmon/airmozilla,kenrick95/airmozilla,Nolski/airmozilla,lcamacho/airmozilla,chirilo/airmozilla,anjalymehla/airmozilla,tannishk/airmozilla,ehsan/airmozilla,ehsan/airmozilla,anjalymehla/airmozilla,peterbe/airmozilla,chirilo/airmozilla,mozilla/airmozilla,EricSekyere/airmozilla,peterbe/airmozilla,EricSekyere/airmozilla,zofuthan/airmozilla,chirilo/airmozilla,anu7495/airmozilla,Nolski/airmozilla,chirilo/airmozilla,anu7495/airmozilla,mozilla/airmozilla,anjalymehla/airmozilla,anjalymehla/airmozilla,mozilla/airmozilla,mythmon/airmozilla,lcamacho/airmozilla,ehsan/airmozilla,kenrick95/airmozilla,lcamacho/airmozilla,tannishk/airmozilla,zofuthan/airmozilla,a-buck/airmozilla,mythmon/airmozilla,zofuthan/airmozilla,anjalymehla/airmozilla,peterbe/airmozilla,kenrick95/airmozilla,blossomica/airmozilla,Nolski/airmozilla,kenrick95/airmozilla,a-buck/airmozilla,anu7495/airmozilla,ehsan/airmozilla,a-buck/airmozilla,bugzPDX/airmozilla,Nolski/airmozilla,zofuthan/airmozilla,tannishk/airmozilla,blossomica/airmozilla,mythmon/airmozilla,mozilla/airmozilla,anu7495/airmozilla,anu7495/airmozilla
|
ee55ce9cc95e0e058cac77f45fac0f899398061e
|
api/preprint_providers/serializers.py
|
api/preprint_providers/serializers.py
|
from rest_framework import serializers as ser
from api.base.utils import absolute_reverse
from api.base.serializers import JSONAPISerializer, LinksField
class PreprintProviderSerializer(JSONAPISerializer):
filterable_fields = frozenset([
'name',
'description',
'id'
])
name = ser.CharField(required=True)
description = ser.CharField(required=False)
id = ser.CharField(max_length=200, source='_id')
logo_path = ser.CharField(read_only=True)
banner_path = ser.CharField(read_only=True)
links = LinksField({
'self': 'get_absolute_url',
'preprints': 'get_preprints_url'
})
class Meta:
type_ = 'preprint_providers'
def get_absolute_url(self, obj):
return obj.absolute_api_v2_url
def get_preprints_url(self, obj):
return absolute_reverse('preprint_providers:preprints-list', kwargs={'provider_id': obj._id})
|
from rest_framework import serializers as ser
from api.base.utils import absolute_reverse
from api.base.serializers import JSONAPISerializer, LinksField
class PreprintProviderSerializer(JSONAPISerializer):
filterable_fields = frozenset([
'name',
'description',
'id'
])
name = ser.CharField(required=True)
description = ser.CharField(required=False)
id = ser.CharField(max_length=200, source='_id')
logo_path = ser.CharField(read_only=True)
banner_path = ser.CharField(read_only=True)
links = LinksField({
'self': 'get_absolute_url',
'preprints': 'get_preprints_url',
'external_url': 'get_external_url'
})
class Meta:
type_ = 'preprint_providers'
def get_absolute_url(self, obj):
return obj.absolute_api_v2_url
def get_preprints_url(self, obj):
return absolute_reverse('preprint_providers:preprints-list', kwargs={'provider_id': obj._id})
def get_external_url(self, obj):
return obj.external_url
|
Add external url to preprint provider serializer
|
Add external url to preprint provider serializer
|
Python
|
apache-2.0
|
chrisseto/osf.io,adlius/osf.io,samchrisinger/osf.io,laurenrevere/osf.io,cslzchen/osf.io,mluo613/osf.io,binoculars/osf.io,adlius/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,felliott/osf.io,CenterForOpenScience/osf.io,caneruguz/osf.io,binoculars/osf.io,Nesiehr/osf.io,alexschiller/osf.io,cwisecarver/osf.io,HalcyonChimera/osf.io,mluo613/osf.io,TomBaxter/osf.io,mattclark/osf.io,hmoco/osf.io,brianjgeiger/osf.io,CenterForOpenScience/osf.io,leb2dg/osf.io,Nesiehr/osf.io,chrisseto/osf.io,cwisecarver/osf.io,mfraezz/osf.io,icereval/osf.io,aaxelb/osf.io,alexschiller/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,samchrisinger/osf.io,erinspace/osf.io,cslzchen/osf.io,leb2dg/osf.io,mluo613/osf.io,alexschiller/osf.io,erinspace/osf.io,emetsger/osf.io,acshi/osf.io,caneruguz/osf.io,TomBaxter/osf.io,pattisdr/osf.io,caseyrollins/osf.io,mluo613/osf.io,rdhyee/osf.io,emetsger/osf.io,sloria/osf.io,laurenrevere/osf.io,alexschiller/osf.io,aaxelb/osf.io,aaxelb/osf.io,chrisseto/osf.io,crcresearch/osf.io,TomBaxter/osf.io,Johnetordoff/osf.io,crcresearch/osf.io,hmoco/osf.io,emetsger/osf.io,emetsger/osf.io,monikagrabowska/osf.io,alexschiller/osf.io,caseyrollins/osf.io,sloria/osf.io,icereval/osf.io,CenterForOpenScience/osf.io,mfraezz/osf.io,chennan47/osf.io,pattisdr/osf.io,chennan47/osf.io,adlius/osf.io,samchrisinger/osf.io,monikagrabowska/osf.io,saradbowman/osf.io,icereval/osf.io,cslzchen/osf.io,HalcyonChimera/osf.io,acshi/osf.io,samchrisinger/osf.io,HalcyonChimera/osf.io,leb2dg/osf.io,chrisseto/osf.io,baylee-d/osf.io,rdhyee/osf.io,CenterForOpenScience/osf.io,baylee-d/osf.io,acshi/osf.io,mattclark/osf.io,Nesiehr/osf.io,crcresearch/osf.io,mluo613/osf.io,caneruguz/osf.io,rdhyee/osf.io,baylee-d/osf.io,cwisecarver/osf.io,aaxelb/osf.io,saradbowman/osf.io,Johnetordoff/osf.io,rdhyee/osf.io,mattclark/osf.io,hmoco/osf.io,acshi/osf.io,felliott/osf.io,leb2dg/osf.io,hmoco/osf.io,chennan47/osf.io,felliott/osf.io,mfraezz/osf.io,felliott/osf.io,brianjgeiger/osf.io,HalcyonChimera/osf.io,sloria/osf.io,acshi/osf.io,erinspace/osf.io,caseyrollins/osf.io,Nesiehr/osf.io,Johnetordoff/osf.io,brianjgeiger/osf.io,monikagrabowska/osf.io,cwisecarver/osf.io,adlius/osf.io,caneruguz/osf.io,mfraezz/osf.io,pattisdr/osf.io,monikagrabowska/osf.io,laurenrevere/osf.io,binoculars/osf.io
|
ac44332d53736f1ac3e067eecf1064bcef038b3a
|
core/platform/transactions/django_transaction_services.py
|
core/platform/transactions/django_transaction_services.py
|
# coding: utf-8
#
# Copyright 2013 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Provides a seam for transaction services."""
__author__ = 'Sean Lip'
def run_in_transaction(fn, *args, **kwargs):
"""Run a function in a transaction."""
# TODO(sll): Actually run the function in a transaction.
return fn(*args, **kwargs)
|
# coding: utf-8
#
# Copyright 2013 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Provides a seam for transaction services."""
__author__ = 'Sean Lip'
from django.db import transaction
def run_in_transaction(fn, *args, **kwargs):
"""Run a function in a transaction."""
with transaction.commit_on_success():
return fn(*args, **kwargs)
|
Add transaction support for django models.
|
Add transaction support for django models.
|
Python
|
apache-2.0
|
oulan/oppia,directorlive/oppia,google-code-export/oppia,oulan/oppia,michaelWagner/oppia,edallison/oppia,terrameijar/oppia,Dev4X/oppia,amitdeutsch/oppia,zgchizi/oppia-uc,virajprabhu/oppia,won0089/oppia,sunu/oppia,mit0110/oppia,sanyaade-teachings/oppia,kennho/oppia,BenHenning/oppia,CMDann/oppia,whygee/oppia,gale320/oppia,kevinlee12/oppia,won0089/oppia,bjvoth/oppia,kaffeel/oppia,won0089/oppia,cleophasmashiri/oppia,danieljjh/oppia,openhatch/oh-missions-oppia-beta,nagyistoce/oppia,kaffeel/oppia,mit0110/oppia,kevinlee12/oppia,kennho/oppia,rackstar17/oppia,toooooper/oppia,won0089/oppia,jestapinski/oppia,Dev4X/oppia,mit0110/oppia,BenHenning/oppia,sdulal/oppia,sanyaade-teachings/oppia,himanshu-dixit/oppia,leandrotoledo/oppia,kevinlee12/oppia,edallison/oppia,toooooper/oppia,souravbadami/oppia,openhatch/oh-missions-oppia-beta,google-code-export/oppia,aldeka/oppia,MaximLich/oppia,kingctan/oppia,oppia/oppia,google-code-export/oppia,AllanYangZhou/oppia,sunu/oppia,sdulal/oppia,sbhowmik89/oppia,sunu/oppia,danieljjh/oppia,Cgruppo/oppia,nagyistoce/oppia,terrameijar/oppia,toooooper/oppia,souravbadami/oppia,mindpin/mindpin_oppia,fernandopinhati/oppia,bjvoth/oppia,danieljjh/oppia,felipecocco/oppia,Atlas-Sailed-Co/oppia,miyucy/oppia,kaffeel/oppia,Atlas-Sailed-Co/oppia,nagyistoce/oppia,aldeka/oppia,dippatel1994/oppia,Cgruppo/oppia,leandrotoledo/oppia,leandrotoledo/oppia,mindpin/mindpin_oppia,raju249/oppia,BenHenning/oppia,kennho/oppia,raju249/oppia,dippatel1994/oppia,VictoriaRoux/oppia,oppia/oppia,mindpin/mindpin_oppia,wangsai/oppia,virajprabhu/oppia,miyucy/oppia,VictoriaRoux/oppia,fernandopinhati/oppia,infinyte/oppia,infinyte/oppia,anthkris/oppia,Dev4X/oppia,MaximLich/oppia,CMDann/oppia,brianrodri/oppia,jestapinski/oppia,kingctan/oppia,amitdeutsch/oppia,brylie/oppia,brianrodri/oppia,MAKOSCAFEE/oppia,nagyistoce/oppia,fernandopinhati/oppia,mindpin/mindpin_oppia,whygee/oppia,Atlas-Sailed-Co/oppia,amitdeutsch/oppia,MAKOSCAFEE/oppia,whygee/oppia,infinyte/oppia,danieljjh/oppia,Cgruppo/oppia,directorlive/oppia,CMDann/oppia,asandyz/oppia,gale320/oppia,souravbadami/oppia,cleophasmashiri/oppia,virajprabhu/oppia,brylie/oppia,cleophasmashiri/oppia,amgowano/oppia,sarahfo/oppia,bjvoth/oppia,sunu/oh-missions-oppia-beta,dippatel1994/oppia,prasanna08/oppia,brylie/oppia,himanshu-dixit/oppia,edallison/oppia,mit0110/oppia,zgchizi/oppia-uc,DewarM/oppia,anthkris/oppia,infinyte/oppia,DewarM/oppia,edallison/oppia,cleophasmashiri/oppia,himanshu-dixit/oppia,sanyaade-teachings/oppia,dippatel1994/oppia,sarahfo/oppia,michaelWagner/oppia,amgowano/oppia,bjvoth/oppia,sanyaade-teachings/oppia,amgowano/oppia,kevinlee12/oppia,shaz13/oppia,oulan/oppia,sbhowmik89/oppia,kevinlee12/oppia,zgchizi/oppia-uc,felipecocco/oppia,wangsai/oppia,openhatch/oh-missions-oppia-beta,BenHenning/oppia,MAKOSCAFEE/oppia,rackstar17/oppia,michaelWagner/oppia,sdulal/oppia,leandrotoledo/oppia,oppia/oppia,gale320/oppia,shaz13/oppia,sanyaade-teachings/oppia,virajprabhu/oppia,prasanna08/oppia,amitdeutsch/oppia,kingctan/oppia,himanshu-dixit/oppia,rackstar17/oppia,sunu/oppia,MAKOSCAFEE/oppia,oppia/oppia,felipecocco/oppia,hazmatzo/oppia,sunu/oppia,VictoriaRoux/oppia,aldeka/oppia,directorlive/oppia,sdulal/oppia,shaz13/oppia,wangsai/oppia,toooooper/oppia,oulan/oppia,bjvoth/oppia,aldeka/oppia,zgchizi/oppia-uc,jestapinski/oppia,danieljjh/oppia,Dev4X/oppia,anthkris/oppia,Atlas-Sailed-Co/oppia,kaffeel/oppia,oppia/oppia,BenHenning/oppia,asandyz/oppia,DewarM/oppia,CMDann/oppia,won0089/oppia,VictoriaRoux/oppia,miyucy/oppia,sunu/oh-missions-oppia-beta,kennho/oppia,wangsai/oppia,fernandopinhati/oppia,oulan/oppia,terrameijar/oppia,Cgruppo/oppia,shaz13/oppia,brylie/oppia,anthkris/oppia,miyucy/oppia,souravbadami/oppia,mit0110/oppia,sarahfo/oppia,kingctan/oppia,felipecocco/oppia,openhatch/oh-missions-oppia-beta,hazmatzo/oppia,anggorodewanto/oppia,amitdeutsch/oppia,felipecocco/oppia,prasanna08/oppia,dippatel1994/oppia,sarahfo/oppia,kennho/oppia,CMDann/oppia,terrameijar/oppia,fernandopinhati/oppia,prasanna08/oppia,google-code-export/oppia,gale320/oppia,hazmatzo/oppia,leandrotoledo/oppia,sdulal/oppia,brylie/oppia,toooooper/oppia,hazmatzo/oppia,anggorodewanto/oppia,Cgruppo/oppia,gale320/oppia,sunu/oh-missions-oppia-beta,kingctan/oppia,anggorodewanto/oppia,brianrodri/oppia,MaximLich/oppia,AllanYangZhou/oppia,raju249/oppia,anggorodewanto/oppia,sbhowmik89/oppia,asandyz/oppia,sunu/oh-missions-oppia-beta,brianrodri/oppia,DewarM/oppia,hazmatzo/oppia,sbhowmik89/oppia,asandyz/oppia,asandyz/oppia,sbhowmik89/oppia,AllanYangZhou/oppia,directorlive/oppia,DewarM/oppia,whygee/oppia,Atlas-Sailed-Co/oppia,jestapinski/oppia,wangsai/oppia,amgowano/oppia,infinyte/oppia,prasanna08/oppia,brianrodri/oppia,google-code-export/oppia,raju249/oppia,VictoriaRoux/oppia,sarahfo/oppia,virajprabhu/oppia,whygee/oppia,michaelWagner/oppia,MaximLich/oppia,souravbadami/oppia,kaffeel/oppia,Dev4X/oppia,michaelWagner/oppia,rackstar17/oppia,directorlive/oppia,AllanYangZhou/oppia,cleophasmashiri/oppia,nagyistoce/oppia
|
e5bd4884fc7ea4389315d0d2b8ff248bbda9a905
|
custom/enikshay/integrations/utils.py
|
custom/enikshay/integrations/utils.py
|
from corehq.apps.locations.models import SQLLocation
from dimagi.utils.logging import notify_exception
def is_submission_from_test_location(person_case):
try:
phi_location = SQLLocation.objects.get(location_id=person_case.owner_id)
except SQLLocation.DoesNotExist:
message = ("Location with id {location_id} not found. This is the owner for person with id: {person_id}"
.format(location_id=person_case.owner_id, person_id=person_case.case_id))
notify_exception(None, message="[ENIKSHAY] {}".format(message))
return True
return phi_location.metadata.get('is_test', "yes") == "yes"
|
from corehq.apps.locations.models import SQLLocation
from custom.enikshay.exceptions import NikshayLocationNotFound
def is_submission_from_test_location(person_case):
try:
phi_location = SQLLocation.objects.get(location_id=person_case.owner_id)
except SQLLocation.DoesNotExist:
raise NikshayLocationNotFound(
"Location with id {location_id} not found. This is the owner for person with id: {person_id}"
.format(location_id=person_case.owner_id, person_id=person_case.case_id)
)
return phi_location.metadata.get('is_test', "yes") == "yes"
|
Revert "Fallback is test location"
|
Revert "Fallback is test location"
This reverts commit 2ba9865fa0f05e9ae244b2513e046c961540fca1.
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
78136c619ebafb54e4bd65af3cfd85a8ff67766b
|
osfclient/tests/test_cloning.py
|
osfclient/tests/test_cloning.py
|
"""Test `osf clone` command."""
import os
from mock import patch, mock_open, call
from osfclient import OSF
from osfclient.cli import clone
from osfclient.tests.mocks import MockProject
from osfclient.tests.mocks import MockArgs
@patch.object(OSF, 'project', return_value=MockProject('1234'))
def test_clone_project(OSF_project):
# check that `osf clone` opens files with the right names and modes
args = MockArgs(project='1234')
mock_open_func = mock_open()
with patch('osfclient.cli.open', mock_open_func):
with patch('osfclient.cli.os.makedirs'):
clone(args)
OSF_project.assert_called_once_with('1234')
# check that the project and the files have been accessed
for store in OSF_project.return_value.storages:
assert store._name_mock.called
for f in store.files:
assert f._path_mock.called
fname = f._path_mock.return_value
if fname.startswith('/'):
fname = fname[1:]
full_path = os.path.join('1234',
store._name_mock.return_value,
fname)
assert call(full_path, 'wb') in mock_open_func.mock_calls
|
"""Test `osf clone` command."""
import os
from mock import patch, mock_open, call
from osfclient import OSF
from osfclient.cli import clone
from osfclient.tests.mocks import MockProject
from osfclient.tests.mocks import MockArgs
@patch.object(OSF, 'project', return_value=MockProject('1234'))
def test_clone_project(OSF_project):
# check that `osf clone` opens files with the right names and modes
args = MockArgs(project='1234')
mock_open_func = mock_open()
with patch('osfclient.cli.open', mock_open_func):
with patch('osfclient.cli.os.makedirs'):
with patch('osfclient.cli.os.getenv', side_effect='SECRET'):
clone(args)
OSF_project.assert_called_once_with('1234')
# check that the project and the files have been accessed
for store in OSF_project.return_value.storages:
assert store._name_mock.called
for f in store.files:
assert f._path_mock.called
fname = f._path_mock.return_value
if fname.startswith('/'):
fname = fname[1:]
full_path = os.path.join('1234',
store._name_mock.return_value,
fname)
assert call(full_path, 'wb') in mock_open_func.mock_calls
|
Fix osf clone test that was asking for a password
|
Fix osf clone test that was asking for a password
|
Python
|
bsd-3-clause
|
betatim/osf-cli,betatim/osf-cli
|
f17baf70d08f47dc4ebb8e0142ce0a3566aa1e9a
|
tests/window/WINDOW_CAPTION.py
|
tests/window/WINDOW_CAPTION.py
|
#!/usr/bin/env python
'''Test that the window caption can be set.
Expected behaviour:
Two windows will be opened, one with the caption "Window caption 1"
counting up every second; the other with a Unicode string including
some non-ASCII characters.
Press escape or close either window to finished the test.
'''
__docformat__ = 'restructuredtext'
__version__ = '$Id: $'
import time
import unittest
from pyglet import window
class WINDOW_CAPTION(unittest.TestCase):
def test_caption(self):
w1 = window.Window(200, 200)
w2 = window.Window(200, 200)
count = 1
w1.set_caption('Window caption %d' % count)
w2.set_caption(u'\u00bfHabla espa\u00f1ol?')
last_time = time.time()
while not (w1.has_exit or w2.has_exit):
if time.time() - last_time > 1:
count += 1
w1.set_caption('Window caption %d' % count)
last_time = time.time()
w1.dispatch_events()
w2.dispatch_events()
w1.close()
w2.close()
if __name__ == '__main__':
unittest.main()
|
#!/usr/bin/env python
'''Test that the window caption can be set.
Expected behaviour:
Two windows will be opened, one with the caption "Window caption 1"
counting up every second; the other with a Unicode string including
some non-ASCII characters.
Press escape or close either window to finished the test.
'''
__docformat__ = 'restructuredtext'
__version__ = '$Id: $'
import time
import unittest
from pyglet import window
class WINDOW_CAPTION(unittest.TestCase):
def test_caption(self):
w1 = window.Window(400, 200, resizable=True)
w2 = window.Window(400, 200, resizable=True)
count = 1
w1.set_caption('Window caption %d' % count)
w2.set_caption(u'\u00bfHabla espa\u00f1ol?')
last_time = time.time()
while not (w1.has_exit or w2.has_exit):
if time.time() - last_time > 1:
count += 1
w1.set_caption('Window caption %d' % count)
last_time = time.time()
w1.dispatch_events()
w2.dispatch_events()
w1.close()
w2.close()
if __name__ == '__main__':
unittest.main()
|
Make windows bigger in this test so the captions can be read.
|
Make windows bigger in this test so the captions can be read.
Index: tests/window/WINDOW_CAPTION.py
===================================================================
--- tests/window/WINDOW_CAPTION.py (revision 777)
+++ tests/window/WINDOW_CAPTION.py (working copy)
@@ -19,8 +19,8 @@
class WINDOW_CAPTION(unittest.TestCase):
def test_caption(self):
- w1 = window.Window(200, 200)
- w2 = window.Window(200, 200)
+ w1 = window.Window(400, 200, resizable=True)
+ w2 = window.Window(400, 200, resizable=True)
count = 1
w1.set_caption('Window caption %d' % count)
w2.set_caption(u'\u00bfHabla espa\u00f1ol?')
git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@781 14d46d22-621c-0410-bb3d-6f67920f7d95
|
Python
|
bsd-3-clause
|
regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations
|
eca659b789cc80c7d99bc38e551def972af11607
|
cs251tk/student/markdownify/check_submit_date.py
|
cs251tk/student/markdownify/check_submit_date.py
|
import os
from dateutil.parser import parse
from ...common import run, chdir
def check_dates(spec_id, username, spec, basedir):
""" Port of the CheckDates program from C++
Finds the first submission date for an assignment
by comparing first commits for all files in the spec
and returning the earliest """
basedir = os.path.join(basedir, 'students', username, spec_id)
dates = []
with chdir(basedir):
for file in spec['files']:
# Run a git log on each file with earliest commits listed first
status, res, _ = run(['git', 'log', '--reverse', '--pretty=format:%ad', '--date=iso8601',
os.path.join(basedir, file['filename'])])
# If we didn't get an error, add date to array
if status == 'success':
# Parse the first line
dates.append(parse(res.splitlines()[0]))
# Return earliest date as a string with the format mm/dd/yyyy hh:mm:ss
return min(dates).strftime("%x %X")
|
import os
from dateutil.parser import parse
from ...common import run, chdir
def check_dates(spec_id, username, spec, basedir):
""" Port of the CheckDates program from C++
Finds the first submission date for an assignment
by comparing first commits for all files in the spec
and returning the earliest """
basedir = os.path.join(basedir, 'students', username, spec_id)
dates = []
with chdir(basedir):
for file in spec['files']:
# Run a git log on each file with earliest commits listed first
status, res, _ = run(['git', 'log', '--reverse', '--pretty=format:%ad', '--date=iso8601',
os.path.join(basedir, file['filename'])])
# If we didn't get an error, add date to array
if status == 'success':
# Parse the first line
dates.append(parse(res.splitlines()[0]))
# Return earliest date as a string with the format mm/dd/yyyy hh:mm:ss
if not dates:
return "ERROR"
return min(dates).strftime("%x %X")
|
Add check for unsuccessful date checks
|
Add check for unsuccessful date checks
|
Python
|
mit
|
StoDevX/cs251-toolkit,StoDevX/cs251-toolkit,StoDevX/cs251-toolkit,StoDevX/cs251-toolkit
|
9c7ff0d98d324e3a52664f9dcd6fe64334778e00
|
web/dbconfig/dbconfigbock7k.py
|
web/dbconfig/dbconfigbock7k.py
|
#
# Configuration for the will database
#
import dbconfig
class dbConfigBock7k ( dbconfig.dbConfig ):
# cubedim is a dictionary so it can vary
# size of the cube at resolution
cubedim = { 0: [128, 128, 16] }
#information about the image stack
slicerange = [0,61]
tilesz = [ 256,256 ]
#resolution information -- lowest resolution and list of resolution
resolutions = [ 0 ]
imagesz = { 0: [ 7198, 7352 ] }
# Resize factor to eliminate distortion
zscale = { 0: 1.0 }
|
#
# Configuration for the will database
#
import dbconfig
class dbConfigBock7k ( dbconfig.dbConfig ):
# cubedim is a dictionary so it can vary
# size of the cube at resolution
cubedim = { 0: [128, 128, 16],
1: [128, 128, 16],
2: [128, 128, 16],
3: [128, 128, 16] }
#information about the image stack
slicerange = [0,61]
tilesz = [ 256,256 ]
#resolution information -- lowest resolution and list of resolution
resolutions = [ 0, 1, 2, 3 ]
imagesz = { 0: [ 7198, 7352 ],
1: [ 3599, 3676 ],
2: [ 1800, 1838 ],
3: [ 900, 919 ] }
# Resize factor to eliminate distortion
zscale = { 0: 10.0,
1: 5.0,
2: 2.5,
3: 1.25 }
|
Expand bock7k to be a multi-resolution project.
|
Expand bock7k to be a multi-resolution project.
|
Python
|
apache-2.0
|
neurodata/ndstore,openconnectome/open-connectome,openconnectome/open-connectome,neurodata/ndstore,neurodata/ndstore,openconnectome/open-connectome,openconnectome/open-connectome,neurodata/ndstore,openconnectome/open-connectome,openconnectome/open-connectome
|
d82111c5415176ea07674723151f14445e4b52ab
|
fire_rs/firemodel/test_propagation.py
|
fire_rs/firemodel/test_propagation.py
|
import unittest
import fire_rs.firemodel.propagation as propagation
class TestPropagation(unittest.TestCase):
def test_propagate(self):
env = propagation.Environment([[475060.0, 477060.0], [6200074.0, 6202074.0]], wind_speed=4.11, wind_dir=0)
prop = propagation.propagate(env, 10, 20)
# prop.plot(blocking=True)
|
import unittest
import fire_rs.firemodel.propagation as propagation
class TestPropagation(unittest.TestCase):
def test_propagate(self):
env = propagation.Environment([[480060.0, 490060.0], [6210074.0, 6220074.0]], wind_speed=4.11, wind_dir=0)
prop = propagation.propagate(env, 10, 20, horizon=3*3600)
# prop.plot(blocking=True)
|
Set test area to a burnable one.
|
[fire-models] Set test area to a burnable one.
|
Python
|
bsd-2-clause
|
fire-rs-laas/fire-rs-saop,fire-rs-laas/fire-rs-saop,fire-rs-laas/fire-rs-saop,fire-rs-laas/fire-rs-saop
|
d919c1e29645a52e795e85686de6de8f1e57196e
|
glue/plugins/ginga_viewer/__init__.py
|
glue/plugins/ginga_viewer/__init__.py
|
try:
from .client import *
from .qt_widget import *
except ImportError:
import warnings
warnings.warn("Could not import ginga plugin, since ginga is required")
# Register qt client
from ...config import qt_client
qt_client.add(GingaWidget)
|
try:
from .client import *
from .qt_widget import *
except ImportError:
import warnings
warnings.warn("Could not import ginga plugin, since ginga is required")
else:
# Register qt client
from ...config import qt_client
qt_client.add(GingaWidget)
|
Fix if ginga is not installed
|
Fix if ginga is not installed
|
Python
|
bsd-3-clause
|
JudoWill/glue,stscieisenhamer/glue,saimn/glue,JudoWill/glue,saimn/glue,stscieisenhamer/glue
|
ee425b43502054895986c447e4cdae2c7e6c9278
|
Lib/fontTools/misc/timeTools.py
|
Lib/fontTools/misc/timeTools.py
|
"""fontTools.misc.timeTools.py -- miscellaneous routines."""
from __future__ import print_function, division, absolute_import
from fontTools.misc.py23 import *
import time
import calendar
# OpenType timestamp handling
epoch_diff = calendar.timegm((1904, 1, 1, 0, 0, 0, 0, 0, 0))
def timestampToString(value):
try:
value = time.asctime(time.gmtime(max(0, value + epoch_diff)))
except ValueError:
value = time.asctime(time.gmtime(0))
def timestampFromString(value):
return calendar.timegm(time.strptime(value)) - epoch_diff
def timestampNow():
return int(time.time() - epoch_diff)
|
"""fontTools.misc.timeTools.py -- miscellaneous routines."""
from __future__ import print_function, division, absolute_import
from fontTools.misc.py23 import *
import time
import calendar
# OpenType timestamp handling
epoch_diff = calendar.timegm((1904, 1, 1, 0, 0, 0, 0, 0, 0))
def timestampToString(value):
# https://github.com/behdad/fonttools/issues/99#issuecomment-66776810
try:
value = time.asctime(time.gmtime(max(0, value + epoch_diff)))
except (OverflowError, ValueError):
value = time.asctime(time.gmtime(0))
def timestampFromString(value):
return calendar.timegm(time.strptime(value)) - epoch_diff
def timestampNow():
return int(time.time() - epoch_diff)
|
Adjust for Python 3.3 change in gmtime() exception type
|
Adjust for Python 3.3 change in gmtime() exception type
https://github.com/behdad/fonttools/issues/99#issuecomment-66776810
Fixes https://github.com/behdad/fonttools/issues/99
|
Python
|
mit
|
googlefonts/fonttools,fonttools/fonttools
|
fbdc69e218a71e984982a39fc36de19b7cf56f90
|
Publishers/SamplePachube.py
|
Publishers/SamplePachube.py
|
import clr
from System import *
from System.Net import WebClient
from System.Xml import XmlDocument
from System.Diagnostics import Trace
url = "http://pachube.com/api/"
apiKey = "40ab667a92d6f892fef6099f38ad5eb31e619dffd793ff8842ae3b00eaf7d7cb"
environmentId = 2065
def Publish(topic, data):
ms = MemoryStream()
Trace.WriteLine("Pachube Sample")
client = WebClient()
client.Headers.Add('X-PachubeApiKey', apiKey)
watts, temp = 25, 44
resp = client.UploadString(CreateFullUrl(), "PUT", str(watts) + "," + str(temp))
client.Dispose();
return 1
def CreateFullUrl():
return url + str(environmentId) + '.csv'
def Shutdown():
return 1
def GetTopics():
return ["PowerMeter/CC128/Mark"]
|
import clr
from System import *
from System.Net import WebClient
from System.Xml import XmlDocument
from System.Diagnostics import Trace
url = "http://pachube.com/api/"
apiKey = "<Your-Pachube-Api-Key-Here>"
environmentId = -1
def Publish(topic, data):
ms = MemoryStream()
Trace.WriteLine("Pachube Sample")
client = WebClient()
client.Headers.Add('X-PachubeApiKey', apiKey)
watts, temp = 25, 44
resp = client.UploadString(CreateFullUrl(), "PUT", str(watts) + "," + str(temp))
client.Dispose();
return 1
def CreateFullUrl():
return url + str(environmentId) + '.csv'
def Shutdown():
return 1
def GetTopics():
return ["PowerMeter/CC128/Mark"]
|
Change to sample pachube script
|
Change to sample pachube script
|
Python
|
mit
|
markallanson/sspe,markallanson/sspe
|
7f6c151d8d5c18fb78a5603792ee19738d625aab
|
python_scripts/extractor_python_readability_server.py
|
python_scripts/extractor_python_readability_server.py
|
#!/usr/bin/python
import sys
import glob
sys.path.append("python_scripts/gen-py")
sys.path.append("gen-py/thrift_solr/")
from thrift.transport import TSocket
from thrift.server import TServer
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
server = TServer.TThreadPoolServer(processor, listening_socket)
print ("[Server] Started")
server.serve()
|
#!/usr/bin/python
import sys
import os
import glob
#sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py"))
sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/"))
sys.path.append(os.path.dirname(__file__) )
from thrift.transport import TSocket
from thrift.server import TServer
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
#print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
#print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
server = TServer.TThreadPoolServer(processor, listening_socket)
print ("[Server] Started")
server.serve()
|
Fix include path and ascii / utf8 errors.
|
Fix include path and ascii / utf8 errors.
|
Python
|
agpl-3.0
|
AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud
|
2a8a564fbd48fba25c4876ff3d4317152a1d647c
|
tests/basics/builtin_range.py
|
tests/basics/builtin_range.py
|
# test builtin range type
# print
print(range(4))
# bool
print(bool(range(0)))
print(bool(range(10)))
# len
print(len(range(0)))
print(len(range(4)))
print(len(range(1, 4)))
print(len(range(1, 4, 2)))
print(len(range(1, 4, -1)))
print(len(range(4, 1, -1)))
print(len(range(4, 1, -2)))
# subscr
print(range(4)[0])
print(range(4)[1])
print(range(4)[-1])
# slice
print(range(4)[0:])
print(range(4)[1:])
print(range(4)[1:2])
print(range(4)[1:3])
print(range(4)[1::2])
print(range(4)[1:-2:2])
# attrs
print(range(1, 2, 3).start)
print(range(1, 2, 3).stop)
print(range(1, 2, 3).step)
# bad unary op
try:
-range(1)
except TypeError:
print("TypeError")
# bad subscription (can't store)
try:
range(1)[0] = 1
except TypeError:
print("TypeError")
|
# test builtin range type
# print
print(range(4))
# bool
print(bool(range(0)))
print(bool(range(10)))
# len
print(len(range(0)))
print(len(range(4)))
print(len(range(1, 4)))
print(len(range(1, 4, 2)))
print(len(range(1, 4, -1)))
print(len(range(4, 1, -1)))
print(len(range(4, 1, -2)))
# subscr
print(range(4)[0])
print(range(4)[1])
print(range(4)[-1])
# slice
print(range(4)[0:])
print(range(4)[1:])
print(range(4)[1:2])
print(range(4)[1:3])
print(range(4)[1::2])
print(range(4)[1:-2:2])
print(range(1,4)[:])
print(range(1,4)[0:])
print(range(1,4)[1:])
print(range(1,4)[:-1])
print(range(7,-2,-4)[:])
# attrs
print(range(1, 2, 3).start)
print(range(1, 2, 3).stop)
print(range(1, 2, 3).step)
# bad unary op
try:
-range(1)
except TypeError:
print("TypeError")
# bad subscription (can't store)
try:
range(1)[0] = 1
except TypeError:
print("TypeError")
|
Test slicing a range that does not start at zero.
|
tests: Test slicing a range that does not start at zero.
|
Python
|
mit
|
torwag/micropython,TDAbboud/micropython,dinau/micropython,dmazzella/micropython,pramasoul/micropython,adafruit/micropython,danicampora/micropython,misterdanb/micropython,trezor/micropython,misterdanb/micropython,redbear/micropython,noahwilliamsson/micropython,adafruit/circuitpython,alex-robbins/micropython,torwag/micropython,alex-robbins/micropython,mpalomer/micropython,ernesto-g/micropython,drrk/micropython,trezor/micropython,dinau/micropython,adamkh/micropython,pfalcon/micropython,matthewelse/micropython,xhat/micropython,cloudformdesign/micropython,chrisdearman/micropython,puuu/micropython,ernesto-g/micropython,turbinenreiter/micropython,bvernoux/micropython,alex-robbins/micropython,dxxb/micropython,alex-march/micropython,deshipu/micropython,MrSurly/micropython,adafruit/micropython,hiway/micropython,SHA2017-badge/micropython-esp32,ryannathans/micropython,ernesto-g/micropython,matthewelse/micropython,HenrikSolver/micropython,jmarcelino/pycom-micropython,MrSurly/micropython-esp32,pramasoul/micropython,adafruit/micropython,ganshun666/micropython,puuu/micropython,selste/micropython,martinribelotta/micropython,ruffy91/micropython,omtinez/micropython,Peetz0r/micropython-esp32,hosaka/micropython,lowRISC/micropython,kerneltask/micropython,selste/micropython,micropython/micropython-esp32,misterdanb/micropython,cloudformdesign/micropython,deshipu/micropython,supergis/micropython,tuc-osg/micropython,mianos/micropython,alex-robbins/micropython,xuxiaoxin/micropython,mpalomer/micropython,ChuckM/micropython,blmorris/micropython,turbinenreiter/micropython,PappaPeppar/micropython,Timmenem/micropython,MrSurly/micropython-esp32,micropython/micropython-esp32,pramasoul/micropython,redbear/micropython,xhat/micropython,dhylands/micropython,ganshun666/micropython,supergis/micropython,infinnovation/micropython,praemdonck/micropython,cloudformdesign/micropython,mhoffma/micropython,AriZuu/micropython,bvernoux/micropython,turbinenreiter/micropython,bvernoux/micropython,mhoffma/micropython,selste/micropython,xuxiaoxin/micropython,infinnovation/micropython,bvernoux/micropython,dmazzella/micropython,Timmenem/micropython,tralamazza/micropython,dxxb/micropython,adafruit/circuitpython,SHA2017-badge/micropython-esp32,kerneltask/micropython,infinnovation/micropython,vitiral/micropython,misterdanb/micropython,lowRISC/micropython,tobbad/micropython,adamkh/micropython,jmarcelino/pycom-micropython,pfalcon/micropython,tuc-osg/micropython,pramasoul/micropython,omtinez/micropython,blmorris/micropython,puuu/micropython,swegener/micropython,Peetz0r/micropython-esp32,emfcamp/micropython,hosaka/micropython,cloudformdesign/micropython,mianos/micropython,tuc-osg/micropython,xuxiaoxin/micropython,xuxiaoxin/micropython,pozetroninc/micropython,MrSurly/micropython-esp32,misterdanb/micropython,mianos/micropython,swegener/micropython,noahwilliamsson/micropython,pozetroninc/micropython,cwyark/micropython,chrisdearman/micropython,hosaka/micropython,deshipu/micropython,toolmacher/micropython,dhylands/micropython,SHA2017-badge/micropython-esp32,henriknelson/micropython,cwyark/micropython,drrk/micropython,MrSurly/micropython,redbear/micropython,henriknelson/micropython,matthewelse/micropython,mhoffma/micropython,dinau/micropython,trezor/micropython,kerneltask/micropython,feilongfl/micropython,galenhz/micropython,ChuckM/micropython,toolmacher/micropython,alex-march/micropython,dhylands/micropython,torwag/micropython,selste/micropython,EcmaXp/micropython,noahwilliamsson/micropython,galenhz/micropython,micropython/micropython-esp32,feilongfl/micropython,ruffy91/micropython,micropython/micropython-esp32,SHA2017-badge/micropython-esp32,kerneltask/micropython,ryannathans/micropython,Timmenem/micropython,vitiral/micropython,chrisdearman/micropython,supergis/micropython,adamkh/micropython,lowRISC/micropython,MrSurly/micropython,dhylands/micropython,blazewicz/micropython,adamkh/micropython,toolmacher/micropython,omtinez/micropython,cwyark/micropython,tuc-osg/micropython,swegener/micropython,danicampora/micropython,Peetz0r/micropython-esp32,dxxb/micropython,xhat/micropython,toolmacher/micropython,ruffy91/micropython,henriknelson/micropython,mpalomer/micropython,oopy/micropython,danicampora/micropython,ChuckM/micropython,deshipu/micropython,EcmaXp/micropython,PappaPeppar/micropython,feilongfl/micropython,ganshun666/micropython,alex-march/micropython,hiway/micropython,trezor/micropython,mhoffma/micropython,dhylands/micropython,AriZuu/micropython,chrisdearman/micropython,puuu/micropython,swegener/micropython,drrk/micropython,mhoffma/micropython,HenrikSolver/micropython,emfcamp/micropython,pfalcon/micropython,turbinenreiter/micropython,AriZuu/micropython,blazewicz/micropython,hiway/micropython,hosaka/micropython,neilh10/micropython,adafruit/circuitpython,EcmaXp/micropython,hosaka/micropython,martinribelotta/micropython,hiway/micropython,feilongfl/micropython,xhat/micropython,pozetroninc/micropython,pfalcon/micropython,supergis/micropython,emfcamp/micropython,HenrikSolver/micropython,praemdonck/micropython,hiway/micropython,vitiral/micropython,oopy/micropython,chrisdearman/micropython,pozetroninc/micropython,noahwilliamsson/micropython,PappaPeppar/micropython,pramasoul/micropython,ryannathans/micropython,xuxiaoxin/micropython,blmorris/micropython,adafruit/micropython,blmorris/micropython,tobbad/micropython,ChuckM/micropython,selste/micropython,matthewelse/micropython,Timmenem/micropython,jmarcelino/pycom-micropython,PappaPeppar/micropython,lowRISC/micropython,EcmaXp/micropython,galenhz/micropython,MrSurly/micropython-esp32,deshipu/micropython,TDAbboud/micropython,matthewelse/micropython,HenrikSolver/micropython,AriZuu/micropython,pozetroninc/micropython,feilongfl/micropython,blazewicz/micropython,henriknelson/micropython,vitiral/micropython,SHA2017-badge/micropython-esp32,omtinez/micropython,bvernoux/micropython,tralamazza/micropython,adafruit/circuitpython,galenhz/micropython,jmarcelino/pycom-micropython,turbinenreiter/micropython,AriZuu/micropython,xhat/micropython,tobbad/micropython,TDAbboud/micropython,danicampora/micropython,ChuckM/micropython,infinnovation/micropython,tralamazza/micropython,martinribelotta/micropython,blmorris/micropython,MrSurly/micropython,noahwilliamsson/micropython,henriknelson/micropython,martinribelotta/micropython,redbear/micropython,puuu/micropython,ryannathans/micropython,ernesto-g/micropython,omtinez/micropython,adamkh/micropython,dxxb/micropython,EcmaXp/micropython,dmazzella/micropython,mianos/micropython,torwag/micropython,cwyark/micropython,drrk/micropython,danicampora/micropython,ruffy91/micropython,ernesto-g/micropython,TDAbboud/micropython,MrSurly/micropython-esp32,drrk/micropython,ganshun666/micropython,neilh10/micropython,MrSurly/micropython,alex-march/micropython,neilh10/micropython,adafruit/circuitpython,TDAbboud/micropython,jmarcelino/pycom-micropython,cwyark/micropython,ruffy91/micropython,praemdonck/micropython,swegener/micropython,cloudformdesign/micropython,Peetz0r/micropython-esp32,pfalcon/micropython,infinnovation/micropython,supergis/micropython,Peetz0r/micropython-esp32,blazewicz/micropython,dxxb/micropython,Timmenem/micropython,matthewelse/micropython,oopy/micropython,PappaPeppar/micropython,tuc-osg/micropython,trezor/micropython,praemdonck/micropython,ganshun666/micropython,lowRISC/micropython,oopy/micropython,tobbad/micropython,dinau/micropython,mianos/micropython,tralamazza/micropython,mpalomer/micropython,tobbad/micropython,dmazzella/micropython,praemdonck/micropython,redbear/micropython,alex-march/micropython,neilh10/micropython,galenhz/micropython,vitiral/micropython,toolmacher/micropython,neilh10/micropython,oopy/micropython,blazewicz/micropython,micropython/micropython-esp32,kerneltask/micropython,adafruit/micropython,mpalomer/micropython,emfcamp/micropython,alex-robbins/micropython,martinribelotta/micropython,emfcamp/micropython,adafruit/circuitpython,ryannathans/micropython,torwag/micropython,HenrikSolver/micropython,dinau/micropython
|
73cb3c6883940e96e656b9b7dd6033ed2e41cb33
|
custom/intrahealth/reports/recap_passage_report_v2.py
|
custom/intrahealth/reports/recap_passage_report_v2.py
|
from __future__ import absolute_import
from __future__ import unicode_literals
from memoized import memoized
from custom.intrahealth.filters import RecapPassageLocationFilter2, FRMonthFilter, FRYearFilter
from custom.intrahealth.sqldata import RecapPassageData2, DateSource2
from custom.intrahealth.reports.tableu_de_board_report_v2 import MultiReport
class RecapPassageReport2(MultiReport):
title = "Recap Passage NEW"
name = "Recap Passage NEW"
slug = 'recap_passage2'
report_title = "Recap Passage"
exportable = True
default_rows = 10
fields = [FRMonthFilter, FRYearFilter, RecapPassageLocationFilter2]
def config_update(self, config):
if self.location and self.location.location_type_name.lower() == 'pps':
config['location_id'] = self.location.location_id
@property
@memoized
def data_providers(self):
dates = DateSource2(config=self.report_config).rows
data_providers = []
for date in dates:
config = self.report_config
config.update(dict(startdate=date, enddate=date))
data_providers.append(RecapPassageData2(config=config))
if not data_providers:
data_providers.append(RecapPassageData2(config=self.report_config))
return data_providers
|
from __future__ import absolute_import
from __future__ import unicode_literals
from memoized import memoized
from corehq.apps.reports.standard import MonthYearMixin
from custom.intrahealth.filters import RecapPassageLocationFilter2, FRMonthFilter, FRYearFilter
from custom.intrahealth.sqldata import RecapPassageData2, DateSource2
from custom.intrahealth.reports.tableu_de_board_report_v2 import MultiReport
class RecapPassageReport2(MonthYearMixin, MultiReport):
title = "Recap Passage NEW"
name = "Recap Passage NEW"
slug = 'recap_passage2'
report_title = "Recap Passage"
exportable = True
default_rows = 10
fields = [FRMonthFilter, FRYearFilter, RecapPassageLocationFilter2]
def config_update(self, config):
if self.location and self.location.location_type_name.lower() == 'pps':
config['location_id'] = self.location.location_id
@property
@memoized
def data_providers(self):
dates = DateSource2(config=self.report_config).rows
data_providers = []
for date in dates:
config = self.report_config
config.update(dict(startdate=date, enddate=date))
data_providers.append(RecapPassageData2(config=config))
if not data_providers:
data_providers.append(RecapPassageData2(config=self.report_config))
return data_providers
|
Fix month filter for recap passage report
|
Fix month filter for recap passage report
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
23a88191e5d827dea84ad533853657110c94c840
|
app/public/views.py
|
app/public/views.py
|
from flask import Blueprint, render_template, redirect, session, url_for
from app.decorators import login_required
blueprint = Blueprint('public', __name__)
@blueprint.route('/')
def home():
"""Return Home Page"""
return render_template('public/index.html')
@blueprint.route('/login', methods=['GET', 'POST'])
def login():
"""Return Login page"""
session['logged_in'] = True
return render_template('public/login.html')
@blueprint.route('/logout')
def logout():
session.pop('logged_in', None)
return redirect(url_for('home'))
|
import os
from flask import Blueprint, redirect, render_template, request, session, url_for
from app.decorators import login_required
ADMIN_USERNAME = os.environ['CUSTOMER_INFO_ADMIN_USERNAME']
ADMIN_PASSWORD_HASH = os.environ['CUSTOMER_INFO_ADMIN_PASSWORD_HASH']
blueprint = Blueprint('public', __name__)
@blueprint.route('/')
def home():
"""Return Home Page"""
return render_template('public/index.html')
def _validate_credentials(username, password):
return (username == ADMIN_USERNAME and
check_password_hash(ADMIN_PASSWORD_HASH, password))
@blueprint.route('/login', methods=['GET', 'POST'])
def login():
"""Return Login page"""
error = None
if request.method == 'POST':
username = request.form['username']
password = request.form['password']
if _validate_credentials(username, password):
session['logged_in'] = True
return redirect(url_for('customer/index.html'))
else:
error = 'Invalid username or password'
return render_template('public/login.html', error=error)
@blueprint.route('/logout')
def logout():
session.pop('logged_in', None)
return redirect(url_for('home'))
|
Add logic to verify and login admin
|
Add logic to verify and login admin
|
Python
|
apache-2.0
|
ueg1990/customer-info,ueg1990/customer-info
|
9c9fff8617a048a32cbff3fb72b3b3ba23476996
|
thinc/neural/_classes/softmax.py
|
thinc/neural/_classes/softmax.py
|
from .affine import Affine
from ... import describe
from ...describe import Dimension, Synapses, Biases
@describe.attributes(
W=Synapses("Weights matrix",
lambda obj: (obj.nO, obj.nI),
lambda W, ops: None)
)
class Softmax(Affine):
def predict(self, input__BI):
output__BO = self.ops.affine(self.W, self.b, input__BI)
self.ops.softmax(output__BO, inplace=True)
return output__BO
def begin_update(self, input__BI, drop=0.):
output__BO = self.predict(input__BI)
def finish_update(grad__BO, sgd=None):
self.d_W += self.ops.batch_outer(grad__BO, input__BI)
self.d_b += grad__BO.sum(axis=0)
if sgd is not None:
sgd(self._mem.weights, self._mem.gradient, key=self.id)
return self.ops.batch_dot(grad__BO, self.W.T)
return output__BO, finish_update
|
from .affine import Affine
from ... import describe
from ...describe import Dimension, Synapses, Biases
@describe.attributes(
W=Synapses("Weights matrix",
lambda obj: (obj.nO, obj.nI),
lambda W, ops: None)
)
class Softmax(Affine):
name = 'softmax'
def predict(self, input__BI):
output__BO = self.ops.affine(self.W, self.b, input__BI)
self.ops.softmax(output__BO, inplace=True)
return output__BO
def begin_update(self, input__BI, drop=0.):
output__BO = self.predict(input__BI)
def finish_update(grad__BO, sgd=None):
self.d_W += self.ops.batch_outer(grad__BO, input__BI)
self.d_b += grad__BO.sum(axis=0)
if sgd is not None:
sgd(self._mem.weights, self._mem.gradient,
key=id(self._mem))
return self.ops.batch_dot(grad__BO, self.W.T)
return output__BO, finish_update
|
Fix passing of params to optimizer in Softmax
|
Fix passing of params to optimizer in Softmax
|
Python
|
mit
|
spacy-io/thinc,explosion/thinc,spacy-io/thinc,explosion/thinc,explosion/thinc,explosion/thinc,spacy-io/thinc
|
0c6dfa4ad297562ec263a8e98bb75d836d2ab054
|
src/python/expedient/ui/html/forms.py
|
src/python/expedient/ui/html/forms.py
|
'''
Created on Jun 20, 2010
@author: jnaous
'''
from django import forms
from expedient.ui.html.models import SliceFlowSpace
class FlowSpaceForm(forms.ModelForm):
"""
Form to edit flowspace.
"""
class Meta:
model = SliceFlowSpace
exclude = ["slice"]
|
'''
Created on Jun 20, 2010
@author: jnaous
'''
from django import forms
from openflow.plugin.models import FlowSpaceRule
class FlowSpaceForm(forms.ModelForm):
"""
Form to edit flowspace.
"""
class Meta:
model = FlowSpaceRule
def __init__(self, sliver_qs, *args, **kwargs):
super(FlowSpaceForm, self).__init__(*args, **kwargs)
self.fields["slivers"].queryset = sliver_qs
|
Modify FlowSpaceForm to use actual stored rules
|
Modify FlowSpaceForm to use actual stored rules
|
Python
|
bsd-3-clause
|
avlach/univbris-ocf,avlach/univbris-ocf,avlach/univbris-ocf,avlach/univbris-ocf
|
cf1da65820085a84eee51884431b0020d3018f23
|
bot/project_info.py
|
bot/project_info.py
|
# Shared project info
name = 'telegram-bot-framework'
description = 'Python Telegram bot API framework'
url = 'https://github.com/alvarogzp/telegram-bot-framework'
author_name = 'Alvaro Gutierrez Perez'
author_email = 'alvarogzp@gmail.com'
authors_credits = (
("@AlvaroGP", "main developer"),
("@KouteiCheke", "i18n & UI/UX support")
)
is_open_source = True
license_name = 'GNU AGPL 3.0+'
license_url = 'https://www.gnu.org/licenses/agpl-3.0.en.html'
donation_addresses = ()
|
# Shared project info
name = 'telegram-bot-framework'
description = 'Python Telegram bot API framework'
url = 'https://github.com/alvarogzp/telegram-bot-framework'
author_name = 'Alvaro Gutierrez Perez'
author_email = 'alvarogzp@gmail.com'
authors_credits = (
("@AlvaroGP", "main developer"),
("@KouteiCheke", "i18n & UI/UX support")
)
is_open_source = True
license_name = 'GNU AGPL 3.0+'
license_url = 'https://www.gnu.org/licenses/agpl-3.0.en.html'
donation_addresses = (
("Bitcoin", "36rwcSgcU1H9fuMvZoebZD3auus6h9wVXk"),
("Bitcoin (bech32 format)", "bc1q4943c5p5dl0hujmmcg2g0568hetynajd3qqtv0")
)
|
Add bitcoin address to donation addresses
|
Add bitcoin address to donation addresses
|
Python
|
agpl-3.0
|
alvarogzp/telegram-bot,alvarogzp/telegram-bot
|
2adf8e8bbf1d0f623e14b8490d511ac45cbb7430
|
djangochurch_data/management/commands/djangochurchimages.py
|
djangochurch_data/management/commands/djangochurchimages.py
|
import os.path
from blanc_basic_assets.models import Image
from django.apps import apps
from django.core.files import File
from django.core.management.base import BaseCommand
IMAGE_LIST = [
(1, 'remember.jpg'),
(2, 'sample-image-1.jpg'),
(3, 'sample-image-2.jpg'),
(4, 'sample-image-3.jpg'),
(5, 'sample-image-4.jpg'),
]
class Command(BaseCommand):
help = 'Load Django Church images'
def handle(self, directory=None, *args, **options):
image_dir = os.path.join(apps.get_app_path('djangochurch_data'), 'images')
for image_id, image_name in IMAGE_LIST:
self.stdout.write('Importing: %s' % (image_name,))
image = Image.objects.get(id=image_id)
image_file = os.path.join(image_dir, image_name)
with open(image_file, 'rb') as f:
image.file.save(image_name, File(f))
|
import os.path
from blanc_basic_assets.models import Image
from django.apps import apps
from django.core.files import File
from django.core.management.base import BaseCommand
IMAGE_LIST = [
(1, 'remember.jpg'),
(2, 'sample-image-1.jpg'),
(3, 'sample-image-2.jpg'),
(4, 'sample-image-3.jpg'),
(5, 'sample-image-4.jpg'),
]
class Command(BaseCommand):
help = 'Load Django Church images'
def handle(self, directory=None, *args, **options):
data_app = apps.get_app_config('djangochurch_data')
image_dir = os.path.join(data_app.path, 'images')
for image_id, image_name in IMAGE_LIST:
self.stdout.write('Importing: %s' % (image_name,))
image = Image.objects.get(id=image_id)
image_file = os.path.join(image_dir, image_name)
with open(image_file, 'rb') as f:
image.file.save(image_name, File(f))
|
Use updated app config for getting the path
|
Use updated app config for getting the path
Prevent warning with Django 1.8, fixes #3
|
Python
|
bsd-3-clause
|
djangochurch/djangochurch-data
|
43e3df5a07caa1370e71858f593c9c8bd73d1e2f
|
cloudly/rqworker.py
|
cloudly/rqworker.py
|
from rq import Worker, Queue, Connection
from rq.job import Job
from cloudly.cache import redis
from cloudly.memoized import Memoized
def enqueue(function, *args):
return _get_queue().enqueue(function, *args)
def fetch_job(job_id):
return Job.fetch(job_id, redis)
@Memoized
def _get_queue():
return Queue(connection=redis)
def work(setup_fct=None):
if setup_fct:
setup_fct()
listen = ['high', 'default', 'low']
with Connection(redis):
worker = Worker(map(Queue, listen))
worker.work()
if __name__ == '__main__':
work()
|
from rq import Worker, Queue, Connection
from rq.job import Job
from cloudly.cache import redis
from cloudly.memoized import Memoized
def enqueue(function, *args, **kwargs):
return _get_queue().enqueue(function, *args, **kwargs)
def fetch_job(job_id):
return Job.fetch(job_id, redis)
@Memoized
def _get_queue():
return Queue(connection=redis)
def work(setup_fct=None):
if setup_fct:
setup_fct()
listen = ['high', 'default', 'low']
with Connection(redis):
worker = Worker(map(Queue, listen))
worker.work()
if __name__ == '__main__':
work()
|
Fix missing `kwargs` argument to enqueue.
|
Fix missing `kwargs` argument to enqueue.
|
Python
|
mit
|
ooda/cloudly,ooda/cloudly
|
0c0e81798b078547bc5931c26dd2b0ab6507db94
|
devilry/project/common/devilry_test_runner.py
|
devilry/project/common/devilry_test_runner.py
|
import warnings
from django.test.runner import DiscoverRunner
from django.utils.deprecation import RemovedInDjango20Warning, RemovedInDjango110Warning
class DevilryTestRunner(DiscoverRunner):
def setup_test_environment(self, **kwargs):
# warnings.filterwarnings('ignore', category=RemovedInDjango)
super(DevilryTestRunner, self).setup_test_environment(**kwargs)
warnings.filterwarnings('ignore', category=RemovedInDjango20Warning)
warnings.filterwarnings('ignore', category=DeprecationWarning)
warnings.filterwarnings('ignore', category=RemovedInDjango110Warning)
|
import warnings
from django.test.runner import DiscoverRunner
from django.utils.deprecation import RemovedInDjango20Warning
class DevilryTestRunner(DiscoverRunner):
def setup_test_environment(self, **kwargs):
# warnings.filterwarnings('ignore', category=RemovedInDjango)
super(DevilryTestRunner, self).setup_test_environment(**kwargs)
warnings.filterwarnings('ignore', category=RemovedInDjango20Warning)
warnings.filterwarnings('ignore', category=DeprecationWarning)
|
Update warning ignores for Django 1.10.
|
project...DevilryTestRunner: Update warning ignores for Django 1.10.
|
Python
|
bsd-3-clause
|
devilry/devilry-django,devilry/devilry-django,devilry/devilry-django,devilry/devilry-django
|
979d84f965b0118f86a8df7aa0311f65f8e36170
|
indra/tools/reading/readers/trips/__init__.py
|
indra/tools/reading/readers/trips/__init__.py
|
from indra.tools.reading.readers.core import EmptyReader
from indra.sources import trips
class TripsReader(EmptyReader):
"""A stand-in for TRIPS reading.
Currently, we do not run TRIPS (more specifically DRUM) regularly at large
scales, however on occasion we have outputs from TRIPS that were generated
a while ago.
"""
name = 'TRIPS'
def __init__(self, *args, **kwargs):
self.version = self.get_version()
return
def _read(self, *args, **kwargs):
return []
@classmethod
def get_version(cls):
return 'STATIC'
@staticmethod
def get_processor(content):
return trips.process_xml(content)
|
import os
import subprocess as sp
from indra.tools.reading.readers.core import Reader
from indra.sources.trips import client, process_xml
from indra_db import formats
class TripsReader(Reader):
"""A stand-in for TRIPS reading.
Currently, we do not run TRIPS (more specifically DRUM) regularly at large
scales, however on occasion we have outputs from TRIPS that were generated
a while ago.
"""
name = 'TRIPS'
result_format = formats.XML
def __init__(self, *args, **kwargs):
self.version = self.get_version()
return
def _read(self, content_iter, verbose=False, log=False, n_per_proc=None):
# Start trips running
if os.environ.get("IN_TRIPS_DOCKER", 'false') != 'true':
return []
p = sp.Popen('/sw/drum/bin/startup.sh', stdout=sp.PIPE,
stderr=sp.STDOUT)
service_endpoint = 'http://localhost:80/cgi/'
service_host = 'drum'
# Process all the content.
for content in content_iter:
html = client.send_query(content.get_text(),
service_endpoint=service_endpoint,
service_host=service_host)
xml = client.get_xml(html)
self.add_result(content.get_id(), xml)
return self.results
@classmethod
def get_version(cls):
return 'STATIC'
@staticmethod
def get_processor(reading_content):
return process_xml(reading_content)
|
Implement the basics of the TRIPS reader.
|
Implement the basics of the TRIPS reader.
|
Python
|
bsd-2-clause
|
sorgerlab/indra,johnbachman/belpy,sorgerlab/belpy,sorgerlab/belpy,sorgerlab/indra,johnbachman/belpy,johnbachman/indra,johnbachman/belpy,bgyori/indra,sorgerlab/indra,johnbachman/indra,sorgerlab/belpy,johnbachman/indra,bgyori/indra,bgyori/indra
|
493ce497e5d84d8db9c37816aefea9099df42e90
|
pywatson/answer/synonym.py
|
pywatson/answer/synonym.py
|
class Synonym(object):
def __init__(self):
pass
|
from pywatson.util.map_initializable import MapInitializable
class SynSetSynonym(MapInitializable):
def __init__(self, is_chosen, value, weight):
self.is_chosen = is_chosen
self.value = value
self.weight = weight
@classmethod
def from_mapping(cls, syn_mapping):
return cls(is_chosen=syn_mapping['isChosen'],
value=syn_mapping['value'],
weight=syn_mapping['weight'])
class SynSet(MapInitializable):
def __init__(self, name, synonyms=()):
self.name = name
self.synonyms = tuple(synonyms)
@classmethod
def from_mapping(cls, synset_mapping):
return cls(name=synset_mapping[0]['name'],
synonyms=(SynSetSynonym.from_mapping(s) for s in synset_mapping[0]['synonym']))
class Synonym(MapInitializable):
def __init__(self, part_of_speech, lemma, value, syn_set):
self.part_of_speech = part_of_speech
self.lemma = lemma
self.value = value
self.syn_set = syn_set
@classmethod
def from_mapping(cls, synonym_mapping):
return cls(part_of_speech=synonym_mapping['partOfSpeech'],
lemma=synonym_mapping['lemma'],
value=synonym_mapping['value'],
syn_set=SynSet.from_mapping(synonym_mapping['synSet']))
|
Add Synonym and related classes
|
Add Synonym and related classes
|
Python
|
mit
|
sherlocke/pywatson
|
10426b049baeceb8dda1390650503e1d75ff8b64
|
us_ignite/common/management/commands/common_load_fixtures.py
|
us_ignite/common/management/commands/common_load_fixtures.py
|
import urlparse
from django.conf import settings
from django.core.management.base import BaseCommand
from django.contrib.sites.models import Site
from us_ignite.profiles.models import Interest
INTEREST_LIST = (
('SDN', 'sdn'),
('OpenFlow', 'openflow'),
('Ultra fast', 'ultra-fast'),
('Advanced wireless', 'advanced-wireless'),
('Low-latency', 'low-latency'),
('Local cloud/edge computing', 'local-cloud-edge-computing'),
('Healthcare', 'healthcare'),
('Education & Workforce development', 'education-workforce-development'),
('Energy', 'energy'),
('Transportation', 'transportation'),
('Advanced Manufacturing', 'advanced-manufacturing'),
('Public Safety', 'public-safety'),
)
class Command(BaseCommand):
def handle(self, *args, **options):
parsed_url = urlparse.urlparse(settings.SITE_URL)
Site.objects.all().update(domain=parsed_url.netloc,
name=parsed_url.netloc)
print "Updated site URL."
for name, slug in INTEREST_LIST:
interest, is_new = (Interest.objects
.get_or_create(name=name, slug=slug))
if is_new:
print u'Imported interest: %s' % interest
|
import urlparse
from django.conf import settings
from django.core.management.base import BaseCommand
from django.contrib.sites.models import Site
from us_ignite.profiles.models import Category, Interest
INTEREST_LIST = (
('SDN', 'sdn'),
('OpenFlow', 'openflow'),
('Ultra fast', 'ultra-fast'),
('Advanced wireless', 'advanced-wireless'),
('Low-latency', 'low-latency'),
('Local cloud/edge computing', 'local-cloud-edge-computing'),
('Healthcare', 'healthcare'),
('Education & Workforce development', 'education-workforce-development'),
('Energy', 'energy'),
('Transportation', 'transportation'),
('Advanced Manufacturing', 'advanced-manufacturing'),
('Public Safety', 'public-safety'),
)
CATEGORY_LIST = [
'Developer',
'Community leader',
'Subject matter expert',
'Designer',
'Project manager',
'Network engineer',
'Funder',
'Press/media',
'Interested party',
]
class Command(BaseCommand):
def handle(self, *args, **options):
parsed_url = urlparse.urlparse(settings.SITE_URL)
Site.objects.all().update(domain=parsed_url.netloc,
name=parsed_url.netloc)
print u'Updated site URL.'
for name, slug in INTEREST_LIST:
interest, is_new = (Interest.objects
.get_or_create(name=name, slug=slug))
if is_new:
print u'Imported interest: %s.' % interest
for name in CATEGORY_LIST:
category, is_new = Category.objects.get_or_create(name=name)
if is_new:
print u'Imported category: %s.' % category
|
Add initial fixtures for the categories.
|
Add initial fixtures for the categories.
|
Python
|
bsd-3-clause
|
us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite
|
fb53f2ed0e6337d6f5766f47cb67c204c89c0568
|
src/oauth2client/__init__.py
|
src/oauth2client/__init__.py
|
# Copyright 2015 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Client library for using OAuth2, especially with Google APIs."""
__version__ = '4.1.3'
GOOGLE_AUTH_URI = 'https://accounts.google.com/o/oauth2/v2/auth'
GOOGLE_DEVICE_URI = 'https://oauth2.googleapis.com/device/code'
GOOGLE_REVOKE_URI = 'https://oauth2.googleapis.com/revoke'
GOOGLE_TOKEN_URI = 'https://oauth2.googleapis.com/token'
GOOGLE_TOKEN_INFO_URI = 'https://oauth2.googleapis.com/tokeninfo'
|
# Copyright 2015 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Client library for using OAuth2, especially with Google APIs."""
__version__ = '4.1.3'
GOOGLE_AUTH_URI = 'https://accounts.google.com/o/oauth2/v2/auth'
GOOGLE_DEVICE_URI = 'https://oauth2.googleapis.com/device/code'
GOOGLE_REVOKE_URI = 'https://accounts.google.com/o/oauth2/revoke'
GOOGLE_TOKEN_URI = 'https://oauth2.googleapis.com/token'
GOOGLE_TOKEN_INFO_URI = 'https://oauth2.googleapis.com/tokeninfo'
|
Fix oauth2 revoke URI, new URL doesn't seem to work
|
Fix oauth2 revoke URI, new URL doesn't seem to work
|
Python
|
apache-2.0
|
GAM-team/GAM,GAM-team/GAM
|
83e820209f9980e6c9103908b14ff07fee23dc41
|
getCheckedOut.py
|
getCheckedOut.py
|
import requests
from bs4 import BeautifulSoup
import json
from dotenv import load_dotenv
import os
load_dotenv(".env")
s = requests.Session()
r = s.get("https://kcls.bibliocommons.com/user/login", verify=False)
payload = {
"name": os.environ.get("USER"),
"user_pin": os.environ.get("PIN")
}
s.post("https://kcls.bibliocommons.com/user/login", data=payload)
r = s.get("https://kcls.bibliocommons.com/checkedout?display_quantity=50&page=1&view=medium")
soup = BeautifulSoup(r.text, "html.parser")
checkedOutList = soup.find("div", { "id": "bibList" })
checkedOutItems = []
for title in checkedOutList.find_all("div", { "class": "listItem" }):
title_name = title.find("span", { "class": "title" })
due_date = title.find("span", { "class": "item_due_date" })
checkedOutItems.append({ "title": title_name.text.strip(), "due": due_date.text.strip() })
with open("checkedout.json", "w") as f:
print "%d title(s) checked out" % len(checkedOutItems)
f.write(json.dumps(checkedOutItems))
|
import requests
from bs4 import BeautifulSoup
import json
from dotenv import load_dotenv
import os
load_dotenv(".env")
s = requests.Session()
r = s.get("https://kcls.bibliocommons.com/user/login", verify=False)
payload = {
"name": os.environ.get("KCLS_USER"),
"user_pin": os.environ.get("PIN")
}
p = s.post("https://kcls.bibliocommons.com/user/login", data=payload)
r = s.get("https://kcls.bibliocommons.com/checkedout?display_quantity=50&page=1&view=medium")
soup = BeautifulSoup(r.text, "html.parser")
checkedOutList = soup.find("div", { "id": "bibList" })
checkedOutItems = []
for title in checkedOutList.find_all("div", { "class": "listItem" }):
title_name = title.find("span", { "class": "title" })
due_date = title.find("span", { "class": "item_due_date" })
checkedOutItems.append({ "title": title_name.text.strip(), "due": due_date.text.strip() })
with open("checkedout.json", "w") as f:
print "%d title(s) checked out" % len(checkedOutItems)
f.write(json.dumps(checkedOutItems))
|
Change .env variable to KCLS_USER
|
Change .env variable to KCLS_USER
|
Python
|
apache-2.0
|
mphuie/kcls-myaccount
|
f0246b9897d89c1ec6f2361bbb488c4e162e5c5e
|
reddit_liveupdate/utils.py
|
reddit_liveupdate/utils.py
|
import itertools
import pytz
from babel.dates import format_time
from pylons import c
def pairwise(iterable):
a, b = itertools.tee(iterable)
next(b, None)
return itertools.izip(a, b)
def pretty_time(dt):
display_tz = pytz.timezone(c.liveupdate_event.timezone)
return format_time(
time=dt,
tzinfo=display_tz,
format="HH:mm z",
locale=c.locale,
)
|
import datetime
import itertools
import pytz
from babel.dates import format_time, format_datetime
from pylons import c
def pairwise(iterable):
a, b = itertools.tee(iterable)
next(b, None)
return itertools.izip(a, b)
def pretty_time(dt):
display_tz = pytz.timezone(c.liveupdate_event.timezone)
today = datetime.datetime.now(display_tz).date()
date = dt.astimezone(display_tz).date()
if date == today:
return format_time(
time=dt,
tzinfo=display_tz,
format="HH:mm z",
locale=c.locale,
)
elif today - date < datetime.timedelta(days=365):
return format_datetime(
datetime=dt,
tzinfo=display_tz,
format="dd MMM HH:mm z",
locale=c.locale,
)
else:
return format_datetime(
datetime=dt,
tzinfo=display_tz,
format="dd MMM YYYY HH:mm z",
locale=c.locale,
)
|
Make timestamps more specific as temporal context fades.
|
Make timestamps more specific as temporal context fades.
Fixes #6.
|
Python
|
bsd-3-clause
|
madbook/reddit-plugin-liveupdate,sim642/reddit-plugin-liveupdate,florenceyeun/reddit-plugin-liveupdate,sim642/reddit-plugin-liveupdate,florenceyeun/reddit-plugin-liveupdate,madbook/reddit-plugin-liveupdate,sim642/reddit-plugin-liveupdate,madbook/reddit-plugin-liveupdate,florenceyeun/reddit-plugin-liveupdate
|
540c5f2969e75a0f461e9d46090cfe8d92c53b00
|
Simulator/plot.py
|
Simulator/plot.py
|
from Simulator import *
import XMLParser
import textToXML
def getHistoryFileName(xmlFileName):
y = xmlFileName[:-3]
return 'history_' + y + 'txt'
def plotFromXML(fileName,simulationTime,chemicalList):
historyFile = getHistoryFileName(fileName)
sim = XMLParser.getSimulator(fileName)
sim.simulate(int(simulationTime),historyFile)
sim.plot(chemicalList)
def plotFromTxt(fileName,simulationTime,chemicalList):
xmlFile = textToXML.getXMLFromTxt(fileName)
plotFromXML(xmlFile,simulationTime,chemicalList)
|
from Simulator import *
import XMLParser
import textToXML
def getHistoryFileName(xmlFileName):
y = xmlFileName[:-3]
y = y + 'txt'
i = len(y) - 1
while i>=0 :
if y[i]=='\\' or y[i]=='/' :
break
i-=1
if i>=0 :
return y[:i+1] + 'history_' + y[i+1:]
else:
return 'history_' + y
def plotFromXML(fileName,simulationTime,chemicalList):
historyFile = getHistoryFileName(fileName)
sim = XMLParser.getSimulator(fileName)
sim.simulate(int(simulationTime),historyFile)
sim.plot(chemicalList)
def plotFromTxt(fileName,simulationTime,chemicalList):
xmlFile = textToXML.getXMLFromTxt(fileName)
plotFromXML(xmlFile,simulationTime,chemicalList)
|
Remove history name error for absolute paths
|
Remove history name error for absolute paths
|
Python
|
mit
|
aayushkapadia/chemical_reaction_simulator
|
ffab98b03588cef69ab11a10a440d02952661edf
|
cyder/cydns/soa/forms.py
|
cyder/cydns/soa/forms.py
|
from django.forms import ModelForm
from cyder.base.mixins import UsabilityFormMixin
from cyder.base.eav.forms import get_eav_form
from cyder.cydns.soa.models import SOA, SOAAV
class SOAForm(ModelForm, UsabilityFormMixin):
class Meta:
model = SOA
fields = ('root_domain', 'primary', 'contact', 'expire',
'retry', 'refresh', 'minimum', 'ttl', 'description',
'is_signed', 'dns_enabled')
exclude = ('serial', 'dirty',)
SOAAVForm = get_eav_form(SOAAV, SOA)
|
from django.forms import ModelForm
from cyder.base.mixins import UsabilityFormMixin
from cyder.base.eav.forms import get_eav_form
from cyder.cydns.soa.models import SOA, SOAAV
class SOAForm(ModelForm, UsabilityFormMixin):
class Meta:
model = SOA
fields = ('root_domain', 'primary', 'contact', 'expire',
'retry', 'refresh', 'minimum', 'ttl', 'description',
'is_signed', 'dns_enabled')
exclude = ('serial', 'dirty',)
def clean(self, *args, **kwargs):
contact = self.cleaned_data['contact']
self.cleaned_data['contact'] = contact.replace('@', '.')
return super(SOAForm, self).clean(*args, **kwargs)
SOAAVForm = get_eav_form(SOAAV, SOA)
|
Replace @ with . in soa form clean
|
Replace @ with . in soa form clean
|
Python
|
bsd-3-clause
|
OSU-Net/cyder,OSU-Net/cyder,akeym/cyder,drkitty/cyder,murrown/cyder,OSU-Net/cyder,drkitty/cyder,akeym/cyder,murrown/cyder,drkitty/cyder,akeym/cyder,murrown/cyder,akeym/cyder,drkitty/cyder,murrown/cyder,OSU-Net/cyder
|
26f984a7732491e87e4eb756caf0056a7ac71484
|
contract_invoice_merge_by_partner/models/account_analytic_analysis.py
|
contract_invoice_merge_by_partner/models/account_analytic_analysis.py
|
# -*- coding: utf-8 -*-
# © 2016 Carlos Dauden <carlos.dauden@tecnativa.com>
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from openerp import api, models
class PurchaseOrderLine(models.Model):
_inherit = 'account.analytic.account'
@api.multi
def _recurring_create_invoice(self, automatic=False):
invoice_obj = self.env['account.invoice']
invoices = invoice_obj.browse(
super(PurchaseOrderLine, self)._recurring_create_invoice(automatic))
res = []
unlink_list = []
for partner in invoices.mapped('partner_id'):
inv_to_merge = invoices.filtered(
lambda x: x.partner_id.id == partner)
if partner.contract_invoice_merge:
invoices_merged = inv_to_merge.do_merge()
res.extend(invoices_merged)
unlink_list.extend(inv_to_merge)
else:
res.extend(inv_to_merge)
if unlink_list:
invoice_obj.unlink([x.id for x in unlink_list])
return res
|
# -*- coding: utf-8 -*-
# © 2016 Carlos Dauden <carlos.dauden@tecnativa.com>
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from openerp import api, models
class PurchaseOrderLine(models.Model):
_inherit = 'account.analytic.account'
@api.multi
def _recurring_create_invoice(self, automatic=False):
invoice_obj = self.env['account.invoice']
invoices = invoice_obj.browse(
super(PurchaseOrderLine, self)._recurring_create_invoice(
automatic))
res = []
unlink_list = []
for partner in invoices.mapped('partner_id'):
inv_to_merge = invoices.filtered(
lambda x: x.partner_id.id == partner)
if partner.contract_invoice_merge and len(inv_to_merge) > 1:
invoices_merged = inv_to_merge.do_merge()
res.extend(invoices_merged)
unlink_list.extend(inv_to_merge)
else:
res.extend(inv_to_merge)
if unlink_list:
invoice_obj.browse(unlink_list).unlink()
return res
|
Fix unlink, >1 filter and lines too long
|
Fix unlink, >1 filter and lines too long
|
Python
|
agpl-3.0
|
bullet92/contract,open-synergy/contract
|
7ad47fad53be18a07aede85c02e41176a96c5de2
|
learnwithpeople/__init__.py
|
learnwithpeople/__init__.py
|
# This will make sure the app is always imported when
# Django starts so that shared_task will use this app.
from .celery import app as celery_app
__version__ = "dev"
GIT_REVISION = "dev"
|
# This will make sure the app is always imported when
# Django starts so that shared_task will use this app.
from .celery import app as celery_app
__all__ = ('celery_app',)
__version__ = "dev"
GIT_REVISION = "dev"
|
Update celery setup according to docs
|
Update celery setup according to docs
|
Python
|
mit
|
p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles
|
e67c57128f88b61eac08e488e54343d48f1454c7
|
ddcz/forms/authentication.py
|
ddcz/forms/authentication.py
|
import logging
from django import forms
from django.contrib.auth import forms as authforms
from ..models import UserProfile
logger = logging.getLogger(__name__)
class LoginForm(forms.Form):
nick = forms.CharField(label="Nick", max_length=20)
password = forms.CharField(label="Heslo", max_length=50, widget=forms.PasswordInput)
class PasswordResetForm(authforms.PasswordResetForm):
def get_users(self, email):
"""Given an email, return matching user(s) who should receive a reset.
This is overridem from original form to use UserProfile instead of standard
user model since that is normative for email storage.
"""
user_profiles = UserProfile.objects.filter(email_uzivatele__iexact=email)
users = tuple(
list(
up.user
for up in user_profiles
if up.user.has_usable_password() and up.user.is_active
)
)
logger.info(
"Selected users for password reset: %s"
% ", ".join([str(u.pk) for u in users])
)
return users
|
import logging
from django import forms
from django.contrib.auth import forms as authforms
from ..models import UserProfile
logger = logging.getLogger(__name__)
class LoginForm(forms.Form):
nick = forms.CharField(label="Nick", max_length=25)
password = forms.CharField(
label="Heslo", max_length=100, widget=forms.PasswordInput
)
class PasswordResetForm(authforms.PasswordResetForm):
def get_users(self, email):
"""Given an email, return matching user(s) who should receive a reset.
This is overridem from original form to use UserProfile instead of standard
user model since that is normative for email storage.
"""
user_profiles = UserProfile.objects.filter(email_uzivatele__iexact=email)
users = tuple(
list(
up.user
for up in user_profiles
if up.user.has_usable_password() and up.user.is_active
)
)
logger.info(
"Selected users for password reset: %s"
% ", ".join([str(u.pk) for u in users])
)
return users
|
Update LoginForm to match reality
|
Update LoginForm to match reality
|
Python
|
mit
|
dracidoupe/graveyard,dracidoupe/graveyard,dracidoupe/graveyard,dracidoupe/graveyard
|
14d6955118893c532c1d9f8f6037d1da1b18dbbb
|
analysis/plot-skeleton.py
|
analysis/plot-skeleton.py
|
#!/usr/bin/env python
import climate
import database
import plots
@climate.annotate(
root='plot data rooted at this path',
pattern=('plot data from files matching this pattern', 'option'),
)
def main(root, pattern='*/*block02/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
with plots.space() as ax:
plots.skeleton(ax, trial, 100)
if __name__ == '__main__':
climate.call(main)
|
#!/usr/bin/env python
import climate
import pandas as pd
import database
import plots
@climate.annotate(
root='plot data rooted at this path',
pattern=('plot data from files matching this pattern', 'option'),
)
def main(root, pattern='*/*block03/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
with plots.space() as ax:
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3)
#trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df)))
trial.make_body_relative()
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3)
if __name__ == '__main__':
climate.call(main)
|
Add multiple skeletons for the moment.
|
Add multiple skeletons for the moment.
|
Python
|
mit
|
lmjohns3/cube-experiment,lmjohns3/cube-experiment,lmjohns3/cube-experiment
|
bfd75a927da2b46cb8630fab0cd3828ba71bf4ee
|
dependencies.py
|
dependencies.py
|
#! /usr/bin/env python3
from setuptools.command import easy_install
requires = ["dnslib", "dkimpy>=0.7.1", "pyyaml", "ddt", "authheaders"]
for module in requires:
easy_install.main( ["-U",module] )
|
#! /usr/bin/env python3
import subprocess
import sys
requires = ["dnslib", "dkimpy>=0.7.1", "pyyaml", "ddt", "authheaders"]
def install(package):
subprocess.call([sys.executable, "-m", "pip", "install", package])
for module in requires:
install(module)
|
Use pip instead of easy_install
|
Use pip instead of easy_install
|
Python
|
mit
|
ValiMail/arc_test_suite
|
3171e7e355536f41a6c517ca7128a152c2577829
|
anndata/tests/test_uns.py
|
anndata/tests/test_uns.py
|
import numpy as np
import pandas as pd
from anndata import AnnData
def test_uns_color_subset():
# Tests for https://github.com/theislab/anndata/issues/257
obs = pd.DataFrame(index=[f"cell{i}" for i in range(5)])
obs["cat1"] = pd.Series(list("aabcd"), index=obs.index, dtype="category")
obs["cat2"] = pd.Series(list("aabbb"), index=obs.index, dtype="category")
uns = dict(
cat1_colors=["red", "green", "blue"], cat2_colors=["red", "green", "blue"],
)
adata = AnnData(np.ones((5, 5)), obs=obs, uns=uns)
# If number of categories does not match number of colors,
# they should be reset
v = adata[:, [0, 1]]
assert "cat1_colors" not in v.uns
assert "cat2_colors" not in v.uns
# Otherwise the colors should still match after reseting
adata.uns["cat1_colors"] = ["red", "green", "blue", "yellow"]
v = adata[[0, 1], :]
assert len(v.uns["cat1_colors"]) == 1
assert v.uns["cat1_colors"][0] == "red"
|
import numpy as np
import pandas as pd
from anndata import AnnData
def test_uns_color_subset():
# Tests for https://github.com/theislab/anndata/issues/257
obs = pd.DataFrame(index=[f"cell{i}" for i in range(5)])
obs["cat1"] = pd.Series(list("aabcd"), index=obs.index, dtype="category")
obs["cat2"] = pd.Series(list("aabbb"), index=obs.index, dtype="category")
uns = dict(
cat1_colors=["red", "green", "blue"], cat2_colors=["red", "green", "blue"],
)
adata = AnnData(np.ones((5, 5)), obs=obs, uns=uns)
# If number of categories does not match number of colors,
# they should be reset
v = adata[:, [0, 1]]
assert "cat1_colors" not in v.uns
assert "cat2_colors" not in v.uns
# Otherwise the colors should still match after reseting
cat1_colors = ["red", "green", "blue", "yellow"]
adata.uns["cat1_colors"] = cat1_colors.copy()
v = adata[[0, 1], :]
assert len(v.uns["cat1_colors"]) == 1
assert v.uns["cat1_colors"][0] == "red"
# But original object should not change
assert list(adata.uns["cat1_colors"]) == cat1_colors
|
Add test for categorical colors staying around after subsetting
|
Add test for categorical colors staying around after subsetting
|
Python
|
bsd-3-clause
|
theislab/anndata
|
2dece45476170e24e14903f19f9bf400c10ebf42
|
djangocms_wow/cms_plugins.py
|
djangocms_wow/cms_plugins.py
|
# -*- coding: utf-8 -*-
from django.utils.translation import ugettext_lazy as _
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from . import models
class AnimationPlugin(CMSPluginBase):
model = models.Animation
name = _('Animation')
render_template = 'djangocms_wow/animation.html'
allow_children = True
cache = True
def render(self, context, instance, placeholder):
context.update({'instance': instance})
return context
plugin_pool.register_plugin(AnimationPlugin)
class WOWAnimationPlugin(CMSPluginBase):
model = models.WOWAnimation
name = _("Wow Animation")
render_template = 'djangocms_wow/wow_animation.html'
allow_children = True
cache = True
def render(self, context, instance, placeholder):
context.update({'instance': instance})
return context
plugin_pool.register_plugin(WOWAnimationPlugin)
|
# -*- coding: utf-8 -*-
from django.utils.translation import ugettext_lazy as _
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from . import models
class AnimationPlugin(CMSPluginBase):
model = models.Animation
name = _('Animation')
render_template = 'djangocms_wow/animation.html'
allow_children = True
text_enabled = True
cache = True
def render(self, context, instance, placeholder):
context.update({'instance': instance})
return context
plugin_pool.register_plugin(AnimationPlugin)
class WOWAnimationPlugin(CMSPluginBase):
model = models.WOWAnimation
name = _("Wow Animation")
render_template = 'djangocms_wow/wow_animation.html'
allow_children = True
text_enabled = True
cache = True
def render(self, context, instance, placeholder):
context.update({'instance': instance})
return context
plugin_pool.register_plugin(WOWAnimationPlugin)
|
Allow WOW animations to be used in text plugin.
|
Allow WOW animations to be used in text plugin.
|
Python
|
bsd-3-clause
|
narayanaditya95/djangocms-wow,narayanaditya95/djangocms-wow,narayanaditya95/djangocms-wow
|
c81b07f93253acc49cbc5028ec83e5334fb47ed9
|
flask_admin/model/typefmt.py
|
flask_admin/model/typefmt.py
|
from jinja2 import Markup
from flask_admin._compat import text_type
def null_formatter(view, value):
"""
Return `NULL` as the string for `None` value
:param value:
Value to check
"""
return Markup('<i>NULL</i>')
def empty_formatter(view, value):
"""
Return empty string for `None` value
:param value:
Value to check
"""
return ''
def bool_formatter(view, value):
"""
Return check icon if value is `True` or empty string otherwise.
:param value:
Value to check
"""
glyph = 'ok-circle' if value else 'minus-sign'
fa = 'check-circle' if value else 'minus-circle'
return Markup('<span class="fa fa-%s glyphicon glyphicon-%s icon-%s"></span>' % (fa, glyph, glyph))
def list_formatter(view, values):
"""
Return string with comma separated values
:param values:
Value to check
"""
return u', '.join(text_type(v) for v in values)
BASE_FORMATTERS = {
type(None): empty_formatter,
bool: bool_formatter,
list: list_formatter,
}
EXPORT_FORMATTERS = {
type(None): empty_formatter,
list: list_formatter,
}
|
from jinja2 import Markup
from flask_admin._compat import text_type
try:
from enum import Enum
except ImportError:
Enum = None
def null_formatter(view, value):
"""
Return `NULL` as the string for `None` value
:param value:
Value to check
"""
return Markup('<i>NULL</i>')
def empty_formatter(view, value):
"""
Return empty string for `None` value
:param value:
Value to check
"""
return ''
def bool_formatter(view, value):
"""
Return check icon if value is `True` or empty string otherwise.
:param value:
Value to check
"""
glyph = 'ok-circle' if value else 'minus-sign'
fa = 'check-circle' if value else 'minus-circle'
return Markup('<span class="fa fa-%s glyphicon glyphicon-%s icon-%s"></span>' % (fa, glyph, glyph))
def list_formatter(view, values):
"""
Return string with comma separated values
:param values:
Value to check
"""
return u', '.join(text_type(v) for v in values)
def enum_formatter(view, value):
"""
Return the name of the enumerated member.
:param value:
Value to check
"""
return value.name
BASE_FORMATTERS = {
type(None): empty_formatter,
bool: bool_formatter,
list: list_formatter,
}
EXPORT_FORMATTERS = {
type(None): empty_formatter,
list: list_formatter,
}
if Enum is not None:
BASE_FORMATTERS[Enum] = enum_formatter
EXPORT_FORMATTERS[Enum] = enum_formatter
|
Add default type formatters for Enum
|
Add default type formatters for Enum
|
Python
|
bsd-3-clause
|
jschneier/flask-admin,jschneier/flask-admin,jschneier/flask-admin,jmagnusson/flask-admin,likaiguo/flask-admin,quokkaproject/flask-admin,flask-admin/flask-admin,lifei/flask-admin,likaiguo/flask-admin,ArtemSerga/flask-admin,iurisilvio/flask-admin,flask-admin/flask-admin,flask-admin/flask-admin,jschneier/flask-admin,jmagnusson/flask-admin,betterlife/flask-admin,closeio/flask-admin,closeio/flask-admin,lifei/flask-admin,quokkaproject/flask-admin,betterlife/flask-admin,quokkaproject/flask-admin,betterlife/flask-admin,lifei/flask-admin,quokkaproject/flask-admin,lifei/flask-admin,iurisilvio/flask-admin,likaiguo/flask-admin,iurisilvio/flask-admin,ArtemSerga/flask-admin,closeio/flask-admin,ArtemSerga/flask-admin,likaiguo/flask-admin,closeio/flask-admin,rochacbruno/flask-admin,jmagnusson/flask-admin,flask-admin/flask-admin,ArtemSerga/flask-admin,rochacbruno/flask-admin,jmagnusson/flask-admin,rochacbruno/flask-admin,iurisilvio/flask-admin,betterlife/flask-admin,rochacbruno/flask-admin
|
a2fd2436cb1c0285dfdd18fad43e505d7c246535
|
modules/module_spotify.py
|
modules/module_spotify.py
|
import re
import urllib
def handle_url(bot, user, channel, url, msg):
"""Handle IMDB urls"""
m = re.match("(http:\/\/open.spotify.com\/|spotify:)(album|artist|track)([:\/])([a-zA-Z0-9]+)\/?", url)
if not m: return
dataurl = "http://spotify.url.fi/%s/%s?txt" % (m.group(2), m.group(4))
f = urllib.urlopen(dataurl)
songinfo = f.read()
f.close()
artist, album, song = songinfo.split("/", 2)
bot.say(channel, "[Spotify] %s - %s (%s)" % (artist.strip(), song.strip(), album.strip()))
|
import re
import urllib
def do_spotify(bot, user, channel, dataurl):
f = urllib.urlopen(dataurl)
songinfo = f.read()
f.close()
artist, album, song = songinfo.split("/", 2)
bot.say(channel, "[Spotify] %s - %s (%s)" % (artist.strip(), song.strip(), album.strip()))
def handle_privmsg(bot, user, reply, msg):
"""Grab Spotify URLs from the messages and handle them"""
m = re.match("(http:\/\/open.spotify.com\/|spotify:)(album|artist|track)([:\/])([a-zA-Z0-9]+)\/?", msg)
if not m: return
dataurl = "http://spotify.url.fi/%s/%s?txt" % (m.group(2), m.group(4))
do_spotify(bot, user, reply, dataurl)
|
Handle spotify: -type urls Cleanup
|
Handle spotify: -type urls
Cleanup
git-svn-id: 056f9092885898c4775d98c479d2d33d00273e45@144 dda364a1-ef19-0410-af65-756c83048fb2
|
Python
|
bsd-3-clause
|
rnyberg/pyfibot,huqa/pyfibot,lepinkainen/pyfibot,EArmour/pyfibot,nigeljonez/newpyfibot,EArmour/pyfibot,huqa/pyfibot,lepinkainen/pyfibot,rnyberg/pyfibot,aapa/pyfibot,aapa/pyfibot
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.