commit
stringlengths
40
40
old_file
stringlengths
4
106
new_file
stringlengths
4
106
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
2.95k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.31k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
diff
stringlengths
49
3.61k
6ccc85832aeff2ca9800cd9e2af8461515ff680d
cartography/midi_utils.py
cartography/midi_utils.py
import mido def open_output(): return open_steinberg_output() def get_steinberg_device_name(): output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()] if len(output_names) != 1: raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one") return output_names[0] def open_steinberg_output(): return mido.open_output(get_steinberg_device_name()) def open_steinberg_input(): return mido.open_input(get_steinberg_device_name())
import mido def open_output(): return open_steinberg_output() def get_steinberg_device_name(): output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()] if len(output_names) != 1: raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one") return output_names[0] def open_steinberg_output(): return mido.open_output(get_steinberg_device_name(), autoreset=True) def open_steinberg_input(): return mido.open_input(get_steinberg_device_name())
Add dump presets and utils
Add dump presets and utils
Python
mit
tingled/synthetic-cartography,tingled/synthetic-cartography
import mido def open_output(): return open_steinberg_output() def get_steinberg_device_name(): output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()] if len(output_names) != 1: raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one") return output_names[0] def open_steinberg_output(): - return mido.open_output(get_steinberg_device_name()) + return mido.open_output(get_steinberg_device_name(), autoreset=True) def open_steinberg_input(): return mido.open_input(get_steinberg_device_name())
Add dump presets and utils
## Code Before: import mido def open_output(): return open_steinberg_output() def get_steinberg_device_name(): output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()] if len(output_names) != 1: raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one") return output_names[0] def open_steinberg_output(): return mido.open_output(get_steinberg_device_name()) def open_steinberg_input(): return mido.open_input(get_steinberg_device_name()) ## Instruction: Add dump presets and utils ## Code After: import mido def open_output(): return open_steinberg_output() def get_steinberg_device_name(): output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()] if len(output_names) != 1: raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one") return output_names[0] def open_steinberg_output(): return mido.open_output(get_steinberg_device_name(), autoreset=True) def open_steinberg_input(): return mido.open_input(get_steinberg_device_name())
import mido def open_output(): return open_steinberg_output() def get_steinberg_device_name(): output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()] if len(output_names) != 1: raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one") return output_names[0] def open_steinberg_output(): - return mido.open_output(get_steinberg_device_name()) + return mido.open_output(get_steinberg_device_name(), autoreset=True) ? ++++++++++++++++ def open_steinberg_input(): return mido.open_input(get_steinberg_device_name())
04ff248c628e5523bc22c532cf4518f210376307
setup.py
setup.py
import os import sys sys.path.insert(0, os.path.abspath('lib')) from ansible import __version__, __author__ from distutils.core import setup setup(name='ansible', version=__version__, description='Minimal SSH command and control', author=__author__, author_email='michael.dehaan@gmail.com', url='http://ansible.github.com/', license='GPLv3', install_requires=['paramiko', 'jinja2', "PyYAML"], package_dir={ 'ansible': 'lib/ansible' }, packages=[ 'ansible', 'ansible.inventory', 'ansible.playbook', 'ansible.runner', 'ansible.runner.connection', ], scripts=[ 'bin/ansible', 'bin/ansible-playbook' ] )
import os import sys sys.path.insert(0, os.path.abspath('lib')) from ansible import __version__, __author__ from distutils.core import setup setup(name='ansible', version=__version__, description='Minimal SSH command and control', author=__author__, author_email='michael.dehaan@gmail.com', url='http://ansible.github.com/', license='GPLv3', install_requires=['paramiko', 'jinja2', "PyYAML"], package_dir={ 'ansible': 'lib/ansible' }, packages=[ 'ansible', 'ansible.inventory', 'ansible.playbook', 'ansible.runner', 'ansible.runner.connection', ], scripts=[ 'bin/ansible', 'bin/ansible-playbook', 'bin/ansible-pull' ] )
Include bin/ansible-pull as part of the sdist in distutils.
Include bin/ansible-pull as part of the sdist in distutils.
Python
mit
thaim/ansible,thaim/ansible
import os import sys sys.path.insert(0, os.path.abspath('lib')) from ansible import __version__, __author__ from distutils.core import setup setup(name='ansible', version=__version__, description='Minimal SSH command and control', author=__author__, author_email='michael.dehaan@gmail.com', url='http://ansible.github.com/', license='GPLv3', install_requires=['paramiko', 'jinja2', "PyYAML"], package_dir={ 'ansible': 'lib/ansible' }, packages=[ 'ansible', 'ansible.inventory', 'ansible.playbook', 'ansible.runner', 'ansible.runner.connection', ], scripts=[ 'bin/ansible', - 'bin/ansible-playbook' + 'bin/ansible-playbook', + 'bin/ansible-pull' ] )
Include bin/ansible-pull as part of the sdist in distutils.
## Code Before: import os import sys sys.path.insert(0, os.path.abspath('lib')) from ansible import __version__, __author__ from distutils.core import setup setup(name='ansible', version=__version__, description='Minimal SSH command and control', author=__author__, author_email='michael.dehaan@gmail.com', url='http://ansible.github.com/', license='GPLv3', install_requires=['paramiko', 'jinja2', "PyYAML"], package_dir={ 'ansible': 'lib/ansible' }, packages=[ 'ansible', 'ansible.inventory', 'ansible.playbook', 'ansible.runner', 'ansible.runner.connection', ], scripts=[ 'bin/ansible', 'bin/ansible-playbook' ] ) ## Instruction: Include bin/ansible-pull as part of the sdist in distutils. ## Code After: import os import sys sys.path.insert(0, os.path.abspath('lib')) from ansible import __version__, __author__ from distutils.core import setup setup(name='ansible', version=__version__, description='Minimal SSH command and control', author=__author__, author_email='michael.dehaan@gmail.com', url='http://ansible.github.com/', license='GPLv3', install_requires=['paramiko', 'jinja2', "PyYAML"], package_dir={ 'ansible': 'lib/ansible' }, packages=[ 'ansible', 'ansible.inventory', 'ansible.playbook', 'ansible.runner', 'ansible.runner.connection', ], scripts=[ 'bin/ansible', 'bin/ansible-playbook', 'bin/ansible-pull' ] )
import os import sys sys.path.insert(0, os.path.abspath('lib')) from ansible import __version__, __author__ from distutils.core import setup setup(name='ansible', version=__version__, description='Minimal SSH command and control', author=__author__, author_email='michael.dehaan@gmail.com', url='http://ansible.github.com/', license='GPLv3', install_requires=['paramiko', 'jinja2', "PyYAML"], package_dir={ 'ansible': 'lib/ansible' }, packages=[ 'ansible', 'ansible.inventory', 'ansible.playbook', 'ansible.runner', 'ansible.runner.connection', ], scripts=[ 'bin/ansible', - 'bin/ansible-playbook' + 'bin/ansible-playbook', ? + + 'bin/ansible-pull' ] )
a4d1659197c0c3da706065d5362fd3b060223c87
newaccount/views.py
newaccount/views.py
from django.shortcuts import render from django.http import JsonResponse import common.render from common.settings import get_page_config def form(request): ''' The signup form webpage ''' context = get_page_config(title='New User Sign Up') context['form'] = [ {'label': 'User Name', 'name': 'username'}, {'label': 'Email Address', 'name': 'email'}, {'label': 'Password', 'type': 'password', 'name':'password'}, {'label': 'Re-enter Password', 'type': 'password', 'id':'repass'} ] context['validators'] = [ 'signup/newaccount_validator.js' ] return common.render.singleform(request, context) def validate(request): ''' Signup form validation handler ''' return JsonResponse({"status": "error", "message": "Test Case"}) def submit(request): ''' Signup form submission handler ''' return HttpResponse('')
from django.http import JsonResponse from django.contrib.auth.models import User from django.core.validators import validate_email from django.core.exceptions import ValidationError from django.shortcuts import render import urllib import common.render from common.settings import get_page_config def form(request): ''' The signup form webpage ''' context = get_page_config(title='New User Sign Up') context['form'] = [ {'label': 'User Name', 'name': 'username'}, {'label': 'Email Address', 'name': 'email'}, {'label': 'Password', 'type': 'password', 'name':'password'}, {'label': 'Re-enter Password', 'type': 'password', 'id':'repass'} ] context['validators'] = [ 'signup/newaccount_validator.js' ] return common.render.singleform(request, context) def validate(request): ''' Signup form validation handler ''' msg = '' if not('username' in request.GET): msg = 'Name not given!' elif not('email' in request.GET): msg = 'Email not given!' name = urllib.unquote(request.GET['username']) email = urllib.unquote(request.GET['email']) if not msg: msg = __validate_name__(name) if not msg: msg = __validate_email__(email) status = 'error' if msg else 'success' return JsonResponse({"status": status, "message": msg}) def submit(request): ''' Signup form submission handler ''' return HttpResponse('') def __validate_name__(name): ''' Internal validation function for username ''' lname = len(name) if lname < 5: return 'User name must be at least 5 characters long' if lname > 64: return 'User name must not be longer than 64 characters' if len(User.objects.filter(username=name)): return 'User name already in use' return '' def __validate_email__(email): ''' Internal validation function for email ''' try: validate_email(email) except ValidationError: return 'Invalid email address: '+email return ''
Implement backend newaccount form verification
Implement backend newaccount form verification
Python
mit
NicolasKiely/Robit-Tracker,NicolasKiely/Robit-Tracker,NicolasKiely/Robit-Tracker
+ from django.http import JsonResponse + from django.contrib.auth.models import User + from django.core.validators import validate_email + from django.core.exceptions import ValidationError from django.shortcuts import render - from django.http import JsonResponse + import urllib import common.render from common.settings import get_page_config def form(request): ''' The signup form webpage ''' context = get_page_config(title='New User Sign Up') context['form'] = [ {'label': 'User Name', 'name': 'username'}, {'label': 'Email Address', 'name': 'email'}, {'label': 'Password', 'type': 'password', 'name':'password'}, {'label': 'Re-enter Password', 'type': 'password', 'id':'repass'} ] context['validators'] = [ 'signup/newaccount_validator.js' ] return common.render.singleform(request, context) def validate(request): ''' Signup form validation handler ''' + msg = '' + if not('username' in request.GET): + msg = 'Name not given!' + elif not('email' in request.GET): + msg = 'Email not given!' + name = urllib.unquote(request.GET['username']) + email = urllib.unquote(request.GET['email']) + if not msg: msg = __validate_name__(name) + if not msg: msg = __validate_email__(email) + + status = 'error' if msg else 'success' - return JsonResponse({"status": "error", "message": "Test Case"}) + return JsonResponse({"status": status, "message": msg}) def submit(request): ''' Signup form submission handler ''' return HttpResponse('') + + def __validate_name__(name): + ''' Internal validation function for username ''' + lname = len(name) + if lname < 5: + return 'User name must be at least 5 characters long' + if lname > 64: + return 'User name must not be longer than 64 characters' + if len(User.objects.filter(username=name)): + return 'User name already in use' + return '' + + def __validate_email__(email): + ''' Internal validation function for email ''' + try: + validate_email(email) + except ValidationError: + return 'Invalid email address: '+email + return '' +
Implement backend newaccount form verification
## Code Before: from django.shortcuts import render from django.http import JsonResponse import common.render from common.settings import get_page_config def form(request): ''' The signup form webpage ''' context = get_page_config(title='New User Sign Up') context['form'] = [ {'label': 'User Name', 'name': 'username'}, {'label': 'Email Address', 'name': 'email'}, {'label': 'Password', 'type': 'password', 'name':'password'}, {'label': 'Re-enter Password', 'type': 'password', 'id':'repass'} ] context['validators'] = [ 'signup/newaccount_validator.js' ] return common.render.singleform(request, context) def validate(request): ''' Signup form validation handler ''' return JsonResponse({"status": "error", "message": "Test Case"}) def submit(request): ''' Signup form submission handler ''' return HttpResponse('') ## Instruction: Implement backend newaccount form verification ## Code After: from django.http import JsonResponse from django.contrib.auth.models import User from django.core.validators import validate_email from django.core.exceptions import ValidationError from django.shortcuts import render import urllib import common.render from common.settings import get_page_config def form(request): ''' The signup form webpage ''' context = get_page_config(title='New User Sign Up') context['form'] = [ {'label': 'User Name', 'name': 'username'}, {'label': 'Email Address', 'name': 'email'}, {'label': 'Password', 'type': 'password', 'name':'password'}, {'label': 'Re-enter Password', 'type': 'password', 'id':'repass'} ] context['validators'] = [ 'signup/newaccount_validator.js' ] return common.render.singleform(request, context) def validate(request): ''' Signup form validation handler ''' msg = '' if not('username' in request.GET): msg = 'Name not given!' elif not('email' in request.GET): msg = 'Email not given!' name = urllib.unquote(request.GET['username']) email = urllib.unquote(request.GET['email']) if not msg: msg = __validate_name__(name) if not msg: msg = __validate_email__(email) status = 'error' if msg else 'success' return JsonResponse({"status": status, "message": msg}) def submit(request): ''' Signup form submission handler ''' return HttpResponse('') def __validate_name__(name): ''' Internal validation function for username ''' lname = len(name) if lname < 5: return 'User name must be at least 5 characters long' if lname > 64: return 'User name must not be longer than 64 characters' if len(User.objects.filter(username=name)): return 'User name already in use' return '' def __validate_email__(email): ''' Internal validation function for email ''' try: validate_email(email) except ValidationError: return 'Invalid email address: '+email return ''
+ from django.http import JsonResponse + from django.contrib.auth.models import User + from django.core.validators import validate_email + from django.core.exceptions import ValidationError from django.shortcuts import render - from django.http import JsonResponse + import urllib import common.render from common.settings import get_page_config def form(request): ''' The signup form webpage ''' context = get_page_config(title='New User Sign Up') context['form'] = [ {'label': 'User Name', 'name': 'username'}, {'label': 'Email Address', 'name': 'email'}, {'label': 'Password', 'type': 'password', 'name':'password'}, {'label': 'Re-enter Password', 'type': 'password', 'id':'repass'} ] context['validators'] = [ 'signup/newaccount_validator.js' ] return common.render.singleform(request, context) def validate(request): ''' Signup form validation handler ''' + msg = '' + if not('username' in request.GET): + msg = 'Name not given!' + elif not('email' in request.GET): + msg = 'Email not given!' + name = urllib.unquote(request.GET['username']) + email = urllib.unquote(request.GET['email']) + if not msg: msg = __validate_name__(name) + if not msg: msg = __validate_email__(email) + + status = 'error' if msg else 'success' - return JsonResponse({"status": "error", "message": "Test Case"}) ? ^^^^^^^ ^^^ ^^^^^^^ + return JsonResponse({"status": status, "message": msg}) ? ^^^^^^ ^ ^ def submit(request): ''' Signup form submission handler ''' return HttpResponse('') + + + def __validate_name__(name): + ''' Internal validation function for username ''' + lname = len(name) + if lname < 5: + return 'User name must be at least 5 characters long' + if lname > 64: + return 'User name must not be longer than 64 characters' + if len(User.objects.filter(username=name)): + return 'User name already in use' + return '' + + def __validate_email__(email): + ''' Internal validation function for email ''' + try: + validate_email(email) + except ValidationError: + return 'Invalid email address: '+email + return ''
0d2079b1dcb97708dc55c32d9e2c1a0f12595875
salt/runners/launchd.py
salt/runners/launchd.py
''' Manage launchd plist files ''' # Import python libs import os import sys def write_launchd_plist(program): ''' Write a launchd plist for managing salt-master or salt-minion CLI Example: .. code-block:: bash salt-run launchd.write_launchd_plist salt-master ''' plist_sample_text = """ <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>Label</key> <string>org.saltstack.{program}</string> <key>ProgramArguments</key> <array> <string>{python}</string> <string>{script}</string> </array> <key>RunAtLoad</key> <true/> </dict> </plist> """.strip() supported_programs = ['salt-master', 'salt-minion'] if program not in supported_programs: sys.stderr.write("Supported programs: %r\n" % supported_programs) sys.exit(-1) sys.stdout.write( plist_sample_text.format( program=program, python=sys.executable, script=os.path.join(os.path.dirname(sys.executable), program) ) )
''' Manage launchd plist files ''' # Import python libs import os import sys def write_launchd_plist(program): ''' Write a launchd plist for managing salt-master or salt-minion CLI Example: .. code-block:: bash salt-run launchd.write_launchd_plist salt-master ''' plist_sample_text = ''' <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>Label</key> <string>org.saltstack.{program}</string> <key>ProgramArguments</key> <array> <string>{python}</string> <string>{script}</string> </array> <key>RunAtLoad</key> <true/> </dict> </plist> '''.strip() supported_programs = ['salt-master', 'salt-minion'] if program not in supported_programs: sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs)) sys.exit(-1) sys.stdout.write( plist_sample_text.format( program=program, python=sys.executable, script=os.path.join(os.path.dirname(sys.executable), program) ) )
Replace string substitution with string formatting
Replace string substitution with string formatting
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
''' Manage launchd plist files ''' # Import python libs import os import sys def write_launchd_plist(program): ''' Write a launchd plist for managing salt-master or salt-minion CLI Example: .. code-block:: bash salt-run launchd.write_launchd_plist salt-master ''' - plist_sample_text = """ + plist_sample_text = ''' <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>Label</key> <string>org.saltstack.{program}</string> <key>ProgramArguments</key> <array> <string>{python}</string> <string>{script}</string> </array> <key>RunAtLoad</key> <true/> </dict> </plist> - """.strip() + '''.strip() supported_programs = ['salt-master', 'salt-minion'] if program not in supported_programs: - sys.stderr.write("Supported programs: %r\n" % supported_programs) + sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs)) sys.exit(-1) sys.stdout.write( plist_sample_text.format( program=program, python=sys.executable, script=os.path.join(os.path.dirname(sys.executable), program) ) )
Replace string substitution with string formatting
## Code Before: ''' Manage launchd plist files ''' # Import python libs import os import sys def write_launchd_plist(program): ''' Write a launchd plist for managing salt-master or salt-minion CLI Example: .. code-block:: bash salt-run launchd.write_launchd_plist salt-master ''' plist_sample_text = """ <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>Label</key> <string>org.saltstack.{program}</string> <key>ProgramArguments</key> <array> <string>{python}</string> <string>{script}</string> </array> <key>RunAtLoad</key> <true/> </dict> </plist> """.strip() supported_programs = ['salt-master', 'salt-minion'] if program not in supported_programs: sys.stderr.write("Supported programs: %r\n" % supported_programs) sys.exit(-1) sys.stdout.write( plist_sample_text.format( program=program, python=sys.executable, script=os.path.join(os.path.dirname(sys.executable), program) ) ) ## Instruction: Replace string substitution with string formatting ## Code After: ''' Manage launchd plist files ''' # Import python libs import os import sys def write_launchd_plist(program): ''' Write a launchd plist for managing salt-master or salt-minion CLI Example: .. code-block:: bash salt-run launchd.write_launchd_plist salt-master ''' plist_sample_text = ''' <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>Label</key> <string>org.saltstack.{program}</string> <key>ProgramArguments</key> <array> <string>{python}</string> <string>{script}</string> </array> <key>RunAtLoad</key> <true/> </dict> </plist> '''.strip() supported_programs = ['salt-master', 'salt-minion'] if program not in supported_programs: sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs)) sys.exit(-1) sys.stdout.write( plist_sample_text.format( program=program, python=sys.executable, script=os.path.join(os.path.dirname(sys.executable), program) ) )
''' Manage launchd plist files ''' # Import python libs import os import sys def write_launchd_plist(program): ''' Write a launchd plist for managing salt-master or salt-minion CLI Example: .. code-block:: bash salt-run launchd.write_launchd_plist salt-master ''' - plist_sample_text = """ ? ^^^ + plist_sample_text = ''' ? ^^^ <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>Label</key> <string>org.saltstack.{program}</string> <key>ProgramArguments</key> <array> <string>{python}</string> <string>{script}</string> </array> <key>RunAtLoad</key> <true/> </dict> </plist> - """.strip() ? ^^^ + '''.strip() ? ^^^ supported_programs = ['salt-master', 'salt-minion'] if program not in supported_programs: - sys.stderr.write("Supported programs: %r\n" % supported_programs) ? ^ ^ ^^^^ + sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs)) ? ^ ^^^ + ^^^^^^^^^ + sys.exit(-1) sys.stdout.write( plist_sample_text.format( program=program, python=sys.executable, script=os.path.join(os.path.dirname(sys.executable), program) ) )
ad415f26eec5c6a20c26123ccb6ce3e320ea9a69
zou/app/blueprints/crud/asset_instance.py
zou/app/blueprints/crud/asset_instance.py
from zou.app.models.asset_instance import AssetInstance from zou.app.services import assets_service, user_service from zou.app.utils import permissions from .base import BaseModelResource, BaseModelsResource class AssetInstancesResource(BaseModelsResource): def __init__(self): BaseModelsResource.__init__(self, AssetInstance) class AssetInstanceResource(BaseModelResource): def __init__(self): BaseModelResource.__init__(self, AssetInstance) def check_read_permissions(self, instance): if permissions.has_manager_permissions(): return True else: asset_instance = self.get_model_or_404(instance["id"]) asset = assets_service.get_asset(asset_instance.asset_id) return user_service.check_has_task_related(asset["project_id"]) def check_update_permissions(self, asset_instance, data): if permissions.has_manager_permissions(): return True else: return user_service.check_working_on_entity( asset_instance["entity_id"] )
from zou.app.models.asset_instance import AssetInstance from zou.app.services import assets_service, user_service from zou.app.utils import permissions from .base import BaseModelResource, BaseModelsResource class AssetInstancesResource(BaseModelsResource): def __init__(self): BaseModelsResource.__init__(self, AssetInstance) class AssetInstanceResource(BaseModelResource): def __init__(self): BaseModelResource.__init__(self, AssetInstance) self.protected_fields.append(["number"]) def check_read_permissions(self, instance): if permissions.has_manager_permissions(): return True else: asset_instance = self.get_model_or_404(instance["id"]) asset = assets_service.get_asset(asset_instance.asset_id) return user_service.check_has_task_related(asset["project_id"]) def check_update_permissions(self, asset_instance, data): if permissions.has_manager_permissions(): return True else: asset = assets_service.get_asset(asset_instance["asset_id"]) return user_service.check_has_task_related(asset["project_id"])
Change asset instance update permissions
Change asset instance update permissions * Do not allow to change instance number * Allow to change instance name by a CG artist
Python
agpl-3.0
cgwire/zou
from zou.app.models.asset_instance import AssetInstance from zou.app.services import assets_service, user_service from zou.app.utils import permissions from .base import BaseModelResource, BaseModelsResource class AssetInstancesResource(BaseModelsResource): def __init__(self): BaseModelsResource.__init__(self, AssetInstance) class AssetInstanceResource(BaseModelResource): def __init__(self): BaseModelResource.__init__(self, AssetInstance) + self.protected_fields.append(["number"]) def check_read_permissions(self, instance): if permissions.has_manager_permissions(): return True else: asset_instance = self.get_model_or_404(instance["id"]) asset = assets_service.get_asset(asset_instance.asset_id) return user_service.check_has_task_related(asset["project_id"]) def check_update_permissions(self, asset_instance, data): if permissions.has_manager_permissions(): return True else: + asset = assets_service.get_asset(asset_instance["asset_id"]) + return user_service.check_has_task_related(asset["project_id"]) - return user_service.check_working_on_entity( - asset_instance["entity_id"] - )
Change asset instance update permissions
## Code Before: from zou.app.models.asset_instance import AssetInstance from zou.app.services import assets_service, user_service from zou.app.utils import permissions from .base import BaseModelResource, BaseModelsResource class AssetInstancesResource(BaseModelsResource): def __init__(self): BaseModelsResource.__init__(self, AssetInstance) class AssetInstanceResource(BaseModelResource): def __init__(self): BaseModelResource.__init__(self, AssetInstance) def check_read_permissions(self, instance): if permissions.has_manager_permissions(): return True else: asset_instance = self.get_model_or_404(instance["id"]) asset = assets_service.get_asset(asset_instance.asset_id) return user_service.check_has_task_related(asset["project_id"]) def check_update_permissions(self, asset_instance, data): if permissions.has_manager_permissions(): return True else: return user_service.check_working_on_entity( asset_instance["entity_id"] ) ## Instruction: Change asset instance update permissions ## Code After: from zou.app.models.asset_instance import AssetInstance from zou.app.services import assets_service, user_service from zou.app.utils import permissions from .base import BaseModelResource, BaseModelsResource class AssetInstancesResource(BaseModelsResource): def __init__(self): BaseModelsResource.__init__(self, AssetInstance) class AssetInstanceResource(BaseModelResource): def __init__(self): BaseModelResource.__init__(self, AssetInstance) self.protected_fields.append(["number"]) def check_read_permissions(self, instance): if permissions.has_manager_permissions(): return True else: asset_instance = self.get_model_or_404(instance["id"]) asset = assets_service.get_asset(asset_instance.asset_id) return user_service.check_has_task_related(asset["project_id"]) def check_update_permissions(self, asset_instance, data): if permissions.has_manager_permissions(): return True else: asset = assets_service.get_asset(asset_instance["asset_id"]) return user_service.check_has_task_related(asset["project_id"])
from zou.app.models.asset_instance import AssetInstance from zou.app.services import assets_service, user_service from zou.app.utils import permissions from .base import BaseModelResource, BaseModelsResource class AssetInstancesResource(BaseModelsResource): def __init__(self): BaseModelsResource.__init__(self, AssetInstance) class AssetInstanceResource(BaseModelResource): def __init__(self): BaseModelResource.__init__(self, AssetInstance) + self.protected_fields.append(["number"]) def check_read_permissions(self, instance): if permissions.has_manager_permissions(): return True else: asset_instance = self.get_model_or_404(instance["id"]) asset = assets_service.get_asset(asset_instance.asset_id) return user_service.check_has_task_related(asset["project_id"]) def check_update_permissions(self, asset_instance, data): if permissions.has_manager_permissions(): return True else: + asset = assets_service.get_asset(asset_instance["asset_id"]) + return user_service.check_has_task_related(asset["project_id"]) - return user_service.check_working_on_entity( - asset_instance["entity_id"] - )
8ef4ca2166167f6370dd8c2f724e752210adf067
sirius/SI_V07/__init__.py
sirius/SI_V07/__init__.py
from . import lattice as _lattice from . import accelerator as _accelerator from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = _accelerator.create_accelerator # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
Fix bug when family_data.py was deleted
Fix bug when family_data.py was deleted
Python
mit
lnls-fac/sirius
from . import lattice as _lattice from . import accelerator as _accelerator - from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator - get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
Fix bug when family_data.py was deleted
## Code Before: from . import lattice as _lattice from . import accelerator as _accelerator from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar] ## Instruction: Fix bug when family_data.py was deleted ## Code After: from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = _accelerator.create_accelerator # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
from . import lattice as _lattice from . import accelerator as _accelerator - from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator - get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
baaeb4fe0998bac8e0cb853d8124aa6134f55996
poradnia/letters/admin.py
poradnia/letters/admin.py
from django.contrib import admin from .models import Attachment, Letter class AttachmentInline(admin.StackedInline): ''' Stacked Inline View for Attachment ''' model = Attachment class LetterAdmin(admin.ModelAdmin): ''' Admin View for Letter ''' inlines = [ AttachmentInline, ] admin.site.register(Letter, LetterAdmin)
from django.contrib import admin from .models import Attachment, Letter class AttachmentInline(admin.StackedInline): ''' Stacked Inline View for Attachment ''' model = Attachment @admin.register(Letter) class LetterAdmin(admin.ModelAdmin): ''' Admin View for Letter ''' inlines = [ AttachmentInline, ]
Rewrite DjangoAdmin in letters for decorators
Rewrite DjangoAdmin in letters for decorators
Python
mit
rwakulszowa/poradnia,rwakulszowa/poradnia,watchdogpolska/poradnia.siecobywatelska.pl,watchdogpolska/poradnia.siecobywatelska.pl,watchdogpolska/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia.siecobywatelska.pl,rwakulszowa/poradnia,rwakulszowa/poradnia
from django.contrib import admin from .models import Attachment, Letter class AttachmentInline(admin.StackedInline): ''' Stacked Inline View for Attachment ''' model = Attachment + @admin.register(Letter) class LetterAdmin(admin.ModelAdmin): ''' Admin View for Letter ''' inlines = [ AttachmentInline, ] - admin.site.register(Letter, LetterAdmin) -
Rewrite DjangoAdmin in letters for decorators
## Code Before: from django.contrib import admin from .models import Attachment, Letter class AttachmentInline(admin.StackedInline): ''' Stacked Inline View for Attachment ''' model = Attachment class LetterAdmin(admin.ModelAdmin): ''' Admin View for Letter ''' inlines = [ AttachmentInline, ] admin.site.register(Letter, LetterAdmin) ## Instruction: Rewrite DjangoAdmin in letters for decorators ## Code After: from django.contrib import admin from .models import Attachment, Letter class AttachmentInline(admin.StackedInline): ''' Stacked Inline View for Attachment ''' model = Attachment @admin.register(Letter) class LetterAdmin(admin.ModelAdmin): ''' Admin View for Letter ''' inlines = [ AttachmentInline, ]
from django.contrib import admin from .models import Attachment, Letter class AttachmentInline(admin.StackedInline): ''' Stacked Inline View for Attachment ''' model = Attachment + @admin.register(Letter) class LetterAdmin(admin.ModelAdmin): ''' Admin View for Letter ''' inlines = [ AttachmentInline, ] - - admin.site.register(Letter, LetterAdmin)
aacb30d2d87cf83c3f09225532c06d81e399cae2
spanky/commands/cmd_roster.py
spanky/commands/cmd_roster.py
import json import click from spanky.cli import pass_context from spanky.lib.enroll import roster @click.command('roster', short_help='Enroll / leave') @click.argument('name') @pass_context def cli(ctx, name): config = ctx.config.load('enroll.yml') click.echo(json.dumps(list(roster(config, name))))
import json import click from spanky.cli import pass_context from spanky.lib.enroll import roster @click.command('roster', short_help='Enroll / leave') @click.argument('name') @pass_context def cli(ctx, name): config = ctx.config.load('enroll.yml') members = list(roster(config, name)) click.echo(json.dumps({'memebers': members}))
Return members in roster cmd out.
Return members in roster cmd out.
Python
bsd-3-clause
pglbutt/spanky,pglbutt/spanky,pglbutt/spanky
import json import click from spanky.cli import pass_context from spanky.lib.enroll import roster @click.command('roster', short_help='Enroll / leave') @click.argument('name') @pass_context def cli(ctx, name): config = ctx.config.load('enroll.yml') - click.echo(json.dumps(list(roster(config, name)))) + members = list(roster(config, name)) + click.echo(json.dumps({'memebers': members}))
Return members in roster cmd out.
## Code Before: import json import click from spanky.cli import pass_context from spanky.lib.enroll import roster @click.command('roster', short_help='Enroll / leave') @click.argument('name') @pass_context def cli(ctx, name): config = ctx.config.load('enroll.yml') click.echo(json.dumps(list(roster(config, name)))) ## Instruction: Return members in roster cmd out. ## Code After: import json import click from spanky.cli import pass_context from spanky.lib.enroll import roster @click.command('roster', short_help='Enroll / leave') @click.argument('name') @pass_context def cli(ctx, name): config = ctx.config.load('enroll.yml') members = list(roster(config, name)) click.echo(json.dumps({'memebers': members}))
import json import click from spanky.cli import pass_context from spanky.lib.enroll import roster @click.command('roster', short_help='Enroll / leave') @click.argument('name') @pass_context def cli(ctx, name): config = ctx.config.load('enroll.yml') - click.echo(json.dumps(list(roster(config, name)))) + members = list(roster(config, name)) + click.echo(json.dumps({'memebers': members}))
0177066012b3373753cba8baf86f00a365d7147b
findaconf/tests/config.py
findaconf/tests/config.py
from decouple import config from findaconf.tests.fake_data import fake_conference, seed def set_app(app, db=False): unset_app(db) app.config['TESTING'] = True app.config['WTF_CSRF_ENABLED'] = False if db: app.config['SQLALCHEMY_DATABASE_URI'] = config( 'DATABASE_URL_TEST', default='sqlite:///' + app.config['BASEDIR'].child('findaconf', 'tests', 'tests.db') ) test_app = app.test_client() if db: db.create_all() seed(app, db) [db.session.add(fake_conference(db)) for i in range(1, 43)] db.session.commit() return test_app def unset_app(db=False): if db: db.session.remove() db.drop_all()
from decouple import config from findaconf.tests.fake_data import fake_conference, seed def set_app(app, db=False): # set test vars app.config['TESTING'] = True app.config['WTF_CSRF_ENABLED'] = False # set test db if db: app.config['SQLALCHEMY_DATABASE_URI'] = config( 'DATABASE_URL_TEST', default='sqlite:///' + app.config['BASEDIR'].child('findaconf', 'tests', 'tests.db') ) # create test app test_app = app.test_client() # create and feed db tables if db: # start from a clean db db.session.remove() db.drop_all() # create tables and feed them db.create_all() seed(app, db) [db.session.add(fake_conference(db)) for i in range(1, 43)] db.session.commit() # return test app return test_app def unset_app(db=False): if db: db.session.remove() db.drop_all()
Fix bug that used dev db instead of test db
Fix bug that used dev db instead of test db
Python
mit
cuducos/findaconf,koorukuroo/findaconf,cuducos/findaconf,koorukuroo/findaconf,koorukuroo/findaconf,cuducos/findaconf
from decouple import config from findaconf.tests.fake_data import fake_conference, seed def set_app(app, db=False): - unset_app(db) + + # set test vars app.config['TESTING'] = True app.config['WTF_CSRF_ENABLED'] = False + + # set test db if db: app.config['SQLALCHEMY_DATABASE_URI'] = config( 'DATABASE_URL_TEST', default='sqlite:///' + app.config['BASEDIR'].child('findaconf', 'tests', 'tests.db') ) + + # create test app test_app = app.test_client() + + # create and feed db tables if db: + + # start from a clean db + db.session.remove() + db.drop_all() + + # create tables and feed them db.create_all() seed(app, db) [db.session.add(fake_conference(db)) for i in range(1, 43)] db.session.commit() + + # return test app return test_app def unset_app(db=False): if db: db.session.remove() db.drop_all()
Fix bug that used dev db instead of test db
## Code Before: from decouple import config from findaconf.tests.fake_data import fake_conference, seed def set_app(app, db=False): unset_app(db) app.config['TESTING'] = True app.config['WTF_CSRF_ENABLED'] = False if db: app.config['SQLALCHEMY_DATABASE_URI'] = config( 'DATABASE_URL_TEST', default='sqlite:///' + app.config['BASEDIR'].child('findaconf', 'tests', 'tests.db') ) test_app = app.test_client() if db: db.create_all() seed(app, db) [db.session.add(fake_conference(db)) for i in range(1, 43)] db.session.commit() return test_app def unset_app(db=False): if db: db.session.remove() db.drop_all() ## Instruction: Fix bug that used dev db instead of test db ## Code After: from decouple import config from findaconf.tests.fake_data import fake_conference, seed def set_app(app, db=False): # set test vars app.config['TESTING'] = True app.config['WTF_CSRF_ENABLED'] = False # set test db if db: app.config['SQLALCHEMY_DATABASE_URI'] = config( 'DATABASE_URL_TEST', default='sqlite:///' + app.config['BASEDIR'].child('findaconf', 'tests', 'tests.db') ) # create test app test_app = app.test_client() # create and feed db tables if db: # start from a clean db db.session.remove() db.drop_all() # create tables and feed them db.create_all() seed(app, db) [db.session.add(fake_conference(db)) for i in range(1, 43)] db.session.commit() # return test app return test_app def unset_app(db=False): if db: db.session.remove() db.drop_all()
from decouple import config from findaconf.tests.fake_data import fake_conference, seed def set_app(app, db=False): - unset_app(db) + + # set test vars app.config['TESTING'] = True app.config['WTF_CSRF_ENABLED'] = False + + # set test db if db: app.config['SQLALCHEMY_DATABASE_URI'] = config( 'DATABASE_URL_TEST', default='sqlite:///' + app.config['BASEDIR'].child('findaconf', 'tests', 'tests.db') ) + + # create test app test_app = app.test_client() + + # create and feed db tables if db: + + # start from a clean db + db.session.remove() + db.drop_all() + + # create tables and feed them db.create_all() seed(app, db) [db.session.add(fake_conference(db)) for i in range(1, 43)] db.session.commit() + + # return test app return test_app def unset_app(db=False): if db: db.session.remove() db.drop_all()
1726a73b81c8a7dfc3610690fe9272776e930f0f
aero/adapters/bower.py
aero/adapters/bower.py
__author__ = 'oliveiraev' __all__ = ['Bower'] from re import sub from re import split from aero.__version__ import enc from .base import BaseAdapter class Bower(BaseAdapter): """ Twitter Bower - Browser package manager - Adapter """ def search(self, query): return {} response = self.command('search', query, ['--no-color'])[0].decode(*enc) lst = dict([(self.package_name(k), v) for k, v in [ line.lstrip(' -').split(' ') for line in response.splitlines() if line.startswith(' - ')] ]) if lst: return lst def install(self, query): return self.shell('install', query) def info(self, query): response = self.command('info', query, ['--no-color'])[0].decode(*enc) return response or ['Aborted: No info available']
__author__ = 'oliveiraev' __all__ = ['Bower'] from re import sub from re import split from aero.__version__ import enc from .base import BaseAdapter class Bower(BaseAdapter): """ Twitter Bower - Browser package manager - Adapter """ def search(self, query): response = self.command('search', query, ['--no-color'])[0].decode(*enc) lst = dict([(self.package_name(k), v) for k, v in [ line.lstrip(' -').split(' ') for line in response.splitlines() if line.startswith(' - ')] ]) return lst or {} def install(self, query): return self.shell('install', query) def info(self, query): response = self.command('info', query, ['--no-color'])[0].decode(*enc) return response or ['Aborted: No info available']
Simplify return while we're at it
Simplify return while we're at it
Python
bsd-3-clause
Aeronautics/aero
__author__ = 'oliveiraev' __all__ = ['Bower'] from re import sub from re import split from aero.__version__ import enc from .base import BaseAdapter class Bower(BaseAdapter): """ Twitter Bower - Browser package manager - Adapter """ def search(self, query): - return {} - response = self.command('search', query, ['--no-color'])[0].decode(*enc) lst = dict([(self.package_name(k), v) for k, v in [ line.lstrip(' -').split(' ') for line in response.splitlines() if line.startswith(' - ')] ]) - if lst: - return lst + return lst or {} def install(self, query): return self.shell('install', query) def info(self, query): response = self.command('info', query, ['--no-color'])[0].decode(*enc) return response or ['Aborted: No info available']
Simplify return while we're at it
## Code Before: __author__ = 'oliveiraev' __all__ = ['Bower'] from re import sub from re import split from aero.__version__ import enc from .base import BaseAdapter class Bower(BaseAdapter): """ Twitter Bower - Browser package manager - Adapter """ def search(self, query): return {} response = self.command('search', query, ['--no-color'])[0].decode(*enc) lst = dict([(self.package_name(k), v) for k, v in [ line.lstrip(' -').split(' ') for line in response.splitlines() if line.startswith(' - ')] ]) if lst: return lst def install(self, query): return self.shell('install', query) def info(self, query): response = self.command('info', query, ['--no-color'])[0].decode(*enc) return response or ['Aborted: No info available'] ## Instruction: Simplify return while we're at it ## Code After: __author__ = 'oliveiraev' __all__ = ['Bower'] from re import sub from re import split from aero.__version__ import enc from .base import BaseAdapter class Bower(BaseAdapter): """ Twitter Bower - Browser package manager - Adapter """ def search(self, query): response = self.command('search', query, ['--no-color'])[0].decode(*enc) lst = dict([(self.package_name(k), v) for k, v in [ line.lstrip(' -').split(' ') for line in response.splitlines() if line.startswith(' - ')] ]) return lst or {} def install(self, query): return self.shell('install', query) def info(self, query): response = self.command('info', query, ['--no-color'])[0].decode(*enc) return response or ['Aborted: No info available']
__author__ = 'oliveiraev' __all__ = ['Bower'] from re import sub from re import split from aero.__version__ import enc from .base import BaseAdapter class Bower(BaseAdapter): """ Twitter Bower - Browser package manager - Adapter """ def search(self, query): - return {} - response = self.command('search', query, ['--no-color'])[0].decode(*enc) lst = dict([(self.package_name(k), v) for k, v in [ line.lstrip(' -').split(' ') for line in response.splitlines() if line.startswith(' - ')] ]) - if lst: - return lst ? ---- + return lst or {} ? ++++++ def install(self, query): return self.shell('install', query) def info(self, query): response = self.command('info', query, ['--no-color'])[0].decode(*enc) return response or ['Aborted: No info available']
2158edb92cba6c19fa258f19445191d0308c4153
utils/async_tasks.py
utils/async_tasks.py
from utils.redis_store import store def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60): # Get task results previously stored in store output, elapsed_time = store.get(store_key, include_elapsed_time=True) # If there are no previously stored results (elapsed_time will be a magically big number) or # if the previously stored results are older than refresh_time, then we trigger recompute of the # task so that results are ready for next load. if elapsed_time > refresh_time: task_func.delay(store_key, *task_args, **task_kwargs) return output
from utils.redis_store import store from celery.signals import task_postrun, task_prerun def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True): # Get task results previously stored in store output, elapsed_time = store.get(store_key, include_elapsed_time=True) # If there are no previously stored results (elapsed_time will be a magically big number) or # if the previously stored results are older than refresh_time, then we trigger recompute of the # task so that results are ready for next load. # If run_once=True, we only trigger the recompute if the task is not already running if elapsed_time > refresh_time: if run_once: # Check that it is not already running computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__) if store.get(computing_store_key): # Task is already running, don't trigger running again print('Skip computing data for {0}, already running'.format(store_key)) return output task_func.delay(store_key, *task_args, **task_kwargs) return output @task_prerun.connect() def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None): # Set computing key computing_store_key = 'computing-{0}'.format(task.name) store.set(computing_store_key, {'running': True}) @task_postrun.connect() def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None): # Delete computing key (if present) computing_store_key = 'computing-{0}'.format(task.name) store.delete(computing_store_key)
Add option to run async tasks only on at a time
Add option to run async tasks only on at a time This is implemented with a simple lock like mechanism using redis.
Python
agpl-3.0
MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets
from utils.redis_store import store + from celery.signals import task_postrun, task_prerun - def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60): + def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True): + # Get task results previously stored in store output, elapsed_time = store.get(store_key, include_elapsed_time=True) # If there are no previously stored results (elapsed_time will be a magically big number) or # if the previously stored results are older than refresh_time, then we trigger recompute of the # task so that results are ready for next load. + # If run_once=True, we only trigger the recompute if the task is not already running if elapsed_time > refresh_time: + if run_once: + # Check that it is not already running + computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__) + if store.get(computing_store_key): + # Task is already running, don't trigger running again + print('Skip computing data for {0}, already running'.format(store_key)) + return output - task_func.delay(store_key, *task_args, **task_kwargs) + task_func.delay(store_key, *task_args, **task_kwargs) return output + + @task_prerun.connect() + def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None): + # Set computing key + computing_store_key = 'computing-{0}'.format(task.name) + store.set(computing_store_key, {'running': True}) + + + @task_postrun.connect() + def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None): + # Delete computing key (if present) + computing_store_key = 'computing-{0}'.format(task.name) + store.delete(computing_store_key) +
Add option to run async tasks only on at a time
## Code Before: from utils.redis_store import store def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60): # Get task results previously stored in store output, elapsed_time = store.get(store_key, include_elapsed_time=True) # If there are no previously stored results (elapsed_time will be a magically big number) or # if the previously stored results are older than refresh_time, then we trigger recompute of the # task so that results are ready for next load. if elapsed_time > refresh_time: task_func.delay(store_key, *task_args, **task_kwargs) return output ## Instruction: Add option to run async tasks only on at a time ## Code After: from utils.redis_store import store from celery.signals import task_postrun, task_prerun def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True): # Get task results previously stored in store output, elapsed_time = store.get(store_key, include_elapsed_time=True) # If there are no previously stored results (elapsed_time will be a magically big number) or # if the previously stored results are older than refresh_time, then we trigger recompute of the # task so that results are ready for next load. # If run_once=True, we only trigger the recompute if the task is not already running if elapsed_time > refresh_time: if run_once: # Check that it is not already running computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__) if store.get(computing_store_key): # Task is already running, don't trigger running again print('Skip computing data for {0}, already running'.format(store_key)) return output task_func.delay(store_key, *task_args, **task_kwargs) return output @task_prerun.connect() def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None): # Set computing key computing_store_key = 'computing-{0}'.format(task.name) store.set(computing_store_key, {'running': True}) @task_postrun.connect() def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None): # Delete computing key (if present) computing_store_key = 'computing-{0}'.format(task.name) store.delete(computing_store_key)
from utils.redis_store import store + from celery.signals import task_postrun, task_prerun - def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60): + def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True): ? +++++++++++++++ + # Get task results previously stored in store output, elapsed_time = store.get(store_key, include_elapsed_time=True) # If there are no previously stored results (elapsed_time will be a magically big number) or # if the previously stored results are older than refresh_time, then we trigger recompute of the # task so that results are ready for next load. + # If run_once=True, we only trigger the recompute if the task is not already running if elapsed_time > refresh_time: + if run_once: + # Check that it is not already running + computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__) + if store.get(computing_store_key): + # Task is already running, don't trigger running again + print('Skip computing data for {0}, already running'.format(store_key)) + return output - task_func.delay(store_key, *task_args, **task_kwargs) ? - + task_func.delay(store_key, *task_args, **task_kwargs) return output + + + @task_prerun.connect() + def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None): + # Set computing key + computing_store_key = 'computing-{0}'.format(task.name) + store.set(computing_store_key, {'running': True}) + + + @task_postrun.connect() + def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None): + # Delete computing key (if present) + computing_store_key = 'computing-{0}'.format(task.name) + store.delete(computing_store_key)
b5fa8ff1d86485c7f00ddecaef040ca66a817dfc
setup.py
setup.py
from distutils.core import setup setup( name='freki', version='0.3.0-develop', description='PDF-Extraction helper for RiPLEs pipeline.', author='Michael Goodman, Ryan Georgi', author_email='goodmami@uw.edu, rgeorgi@uw.edu', url='https://github.com/xigt/freki', license='MIT', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Text Processing :: Linguistic', 'Topic :: Utilities' ], keywords='nlp pdf ie text', packages=['freki', 'freki.readers', 'freki.analyzers'], install_requires=[ 'numpy', 'matplotlib' ], entry_points={ 'console_scripts': [ 'freki=freki.main:main' ] }, )
from distutils.core import setup setup( name='freki', version='0.3.0-develop', description='PDF-Extraction helper for RiPLEs pipeline.', author='Michael Goodman, Ryan Georgi', author_email='goodmami@uw.edu, rgeorgi@uw.edu', url='https://github.com/xigt/freki', license='MIT', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Text Processing :: Linguistic', 'Topic :: Utilities' ], keywords='nlp pdf ie text', packages=['freki', 'freki.readers', 'freki.analyzers'], install_requires=[ 'numpy', 'matplotlib', 'chardet' ], entry_points={ 'console_scripts': [ 'freki=freki.main:main' ] }, )
Add Chardet as installation dependency
Add Chardet as installation dependency
Python
mit
xigt/freki,xigt/freki
from distutils.core import setup setup( name='freki', version='0.3.0-develop', description='PDF-Extraction helper for RiPLEs pipeline.', author='Michael Goodman, Ryan Georgi', author_email='goodmami@uw.edu, rgeorgi@uw.edu', url='https://github.com/xigt/freki', license='MIT', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Text Processing :: Linguistic', 'Topic :: Utilities' ], keywords='nlp pdf ie text', packages=['freki', 'freki.readers', 'freki.analyzers'], install_requires=[ 'numpy', - 'matplotlib' + 'matplotlib', + 'chardet' ], entry_points={ 'console_scripts': [ 'freki=freki.main:main' ] }, )
Add Chardet as installation dependency
## Code Before: from distutils.core import setup setup( name='freki', version='0.3.0-develop', description='PDF-Extraction helper for RiPLEs pipeline.', author='Michael Goodman, Ryan Georgi', author_email='goodmami@uw.edu, rgeorgi@uw.edu', url='https://github.com/xigt/freki', license='MIT', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Text Processing :: Linguistic', 'Topic :: Utilities' ], keywords='nlp pdf ie text', packages=['freki', 'freki.readers', 'freki.analyzers'], install_requires=[ 'numpy', 'matplotlib' ], entry_points={ 'console_scripts': [ 'freki=freki.main:main' ] }, ) ## Instruction: Add Chardet as installation dependency ## Code After: from distutils.core import setup setup( name='freki', version='0.3.0-develop', description='PDF-Extraction helper for RiPLEs pipeline.', author='Michael Goodman, Ryan Georgi', author_email='goodmami@uw.edu, rgeorgi@uw.edu', url='https://github.com/xigt/freki', license='MIT', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Text Processing :: Linguistic', 'Topic :: Utilities' ], keywords='nlp pdf ie text', packages=['freki', 'freki.readers', 'freki.analyzers'], install_requires=[ 'numpy', 'matplotlib', 'chardet' ], entry_points={ 'console_scripts': [ 'freki=freki.main:main' ] }, )
from distutils.core import setup setup( name='freki', version='0.3.0-develop', description='PDF-Extraction helper for RiPLEs pipeline.', author='Michael Goodman, Ryan Georgi', author_email='goodmami@uw.edu, rgeorgi@uw.edu', url='https://github.com/xigt/freki', license='MIT', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Text Processing :: Linguistic', 'Topic :: Utilities' ], keywords='nlp pdf ie text', packages=['freki', 'freki.readers', 'freki.analyzers'], install_requires=[ 'numpy', - 'matplotlib' + 'matplotlib', ? + + 'chardet' ], entry_points={ 'console_scripts': [ 'freki=freki.main:main' ] }, )
d6ff777c7fb3f645c021da1319bb5d78d13aa9db
meshnet/interface.py
meshnet/interface.py
import serial import struct from siphashc import siphash def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes): packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data) return struct.pack("Q", siphash(key, packed_data)) class SerialMessage(object): def __init__(self): pass def serialize(self): pass class Connection(object): def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200)
import serial import struct from siphashc import siphash def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes): packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data return struct.pack(">Q", siphash(key, packed_data)) class SerialMessage(object): def __init__(self): pass def serialize(self): pass class Connection(object): def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200)
Fix python siphashing to match c implementation
Fix python siphashing to match c implementation Signed-off-by: Jan Losinski <577c4104c61edf9f052c616c0c23e67bef4a9955@wh2.tu-dresden.de>
Python
bsd-3-clause
janLo/automation_mesh,janLo/automation_mesh,janLo/automation_mesh
import serial import struct from siphashc import siphash + - def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes): + def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes): - packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data) + packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data - return struct.pack("Q", siphash(key, packed_data)) + return struct.pack(">Q", siphash(key, packed_data)) class SerialMessage(object): def __init__(self): pass def serialize(self): pass - - class Connection(object): - def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200)
Fix python siphashing to match c implementation
## Code Before: import serial import struct from siphashc import siphash def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes): packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data) return struct.pack("Q", siphash(key, packed_data)) class SerialMessage(object): def __init__(self): pass def serialize(self): pass class Connection(object): def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200) ## Instruction: Fix python siphashing to match c implementation ## Code After: import serial import struct from siphashc import siphash def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes): packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data return struct.pack(">Q", siphash(key, packed_data)) class SerialMessage(object): def __init__(self): pass def serialize(self): pass class Connection(object): def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200)
import serial import struct from siphashc import siphash + - def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes): ? -- + def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes): ? ++++ - packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data) ? - - ^ - + packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data ? ^^^ - return struct.pack("Q", siphash(key, packed_data)) + return struct.pack(">Q", siphash(key, packed_data)) ? + class SerialMessage(object): def __init__(self): pass def serialize(self): pass - - class Connection(object): - def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200)
1a761c9360f185d6bd07be9f16ea2cfa239f4bdd
groupy/api/base.py
groupy/api/base.py
from groupy import utils class Manager: """Class for interacting with the endpoint for a resource. :param session: the requests session :type session: :class:`~groupy.session.Session` :param str path: path relative to the base URL """ #: the base URL base_url = 'https://api.groupme.com/v3/' def __init__(self, session, path=None): self.session = session self.url = utils.urljoin(self.base_url, path) class Resource: def __init__(self, **data): self.data = data def __getattr__(self, attr): if attr not in self.data: error_message = 'this {!s} resource does not have a {!r} field' raise AttributeError(error_message.format(self.__class__.__name__, attr)) return self.data[attr] class ManagedResource(Resource): """Class to represent an API object.""" def __init__(self, manager, **data): """Create an instance of the resource. :param manager: the resource's manager :type manager: :class:`~groupy.api.base.Manager` :param kwargs data: the resource data """ super().__init__(**data) self.manager = manager
from groupy import utils class Manager: """Class for interacting with the endpoint for a resource. :param session: the requests session :type session: :class:`~groupy.session.Session` :param str path: path relative to the base URL """ #: the base URL base_url = 'https://api.groupme.com/v3/' def __init__(self, session, path=None): self.session = session self.url = utils.urljoin(self.base_url, path) class Resource: def __init__(self, **data): self.data = data def __getattr__(self, attr): if attr not in self.data: error_message = 'this {!s} resource does not have a {!r} field' raise AttributeError(error_message.format(self.__class__.__name__, attr)) return self.data[attr] def __getstate__(self): return self.__dict__ def __setstate__(self, d): self.__dict__.update(d) class ManagedResource(Resource): """Class to represent an API object.""" def __init__(self, manager, **data): """Create an instance of the resource. :param manager: the resource's manager :type manager: :class:`~groupy.api.base.Manager` :param kwargs data: the resource data """ super().__init__(**data) self.manager = manager
Fix pickling/unpickling of Resource objects
Fix pickling/unpickling of Resource objects Add __getstate__ and __setstate__ methods to the Resource class to avoid hitting the recursion limit when trying to pickle/unpickle Resource objects. A similar issue/solution can be found here: https://stackoverflow.com/a/12102691
Python
apache-2.0
rhgrant10/Groupy
from groupy import utils class Manager: """Class for interacting with the endpoint for a resource. :param session: the requests session :type session: :class:`~groupy.session.Session` :param str path: path relative to the base URL """ #: the base URL base_url = 'https://api.groupme.com/v3/' def __init__(self, session, path=None): self.session = session self.url = utils.urljoin(self.base_url, path) class Resource: def __init__(self, **data): self.data = data def __getattr__(self, attr): if attr not in self.data: error_message = 'this {!s} resource does not have a {!r} field' raise AttributeError(error_message.format(self.__class__.__name__, attr)) return self.data[attr] + def __getstate__(self): + return self.__dict__ + + def __setstate__(self, d): + self.__dict__.update(d) + class ManagedResource(Resource): """Class to represent an API object.""" def __init__(self, manager, **data): """Create an instance of the resource. :param manager: the resource's manager :type manager: :class:`~groupy.api.base.Manager` :param kwargs data: the resource data """ super().__init__(**data) self.manager = manager
Fix pickling/unpickling of Resource objects
## Code Before: from groupy import utils class Manager: """Class for interacting with the endpoint for a resource. :param session: the requests session :type session: :class:`~groupy.session.Session` :param str path: path relative to the base URL """ #: the base URL base_url = 'https://api.groupme.com/v3/' def __init__(self, session, path=None): self.session = session self.url = utils.urljoin(self.base_url, path) class Resource: def __init__(self, **data): self.data = data def __getattr__(self, attr): if attr not in self.data: error_message = 'this {!s} resource does not have a {!r} field' raise AttributeError(error_message.format(self.__class__.__name__, attr)) return self.data[attr] class ManagedResource(Resource): """Class to represent an API object.""" def __init__(self, manager, **data): """Create an instance of the resource. :param manager: the resource's manager :type manager: :class:`~groupy.api.base.Manager` :param kwargs data: the resource data """ super().__init__(**data) self.manager = manager ## Instruction: Fix pickling/unpickling of Resource objects ## Code After: from groupy import utils class Manager: """Class for interacting with the endpoint for a resource. :param session: the requests session :type session: :class:`~groupy.session.Session` :param str path: path relative to the base URL """ #: the base URL base_url = 'https://api.groupme.com/v3/' def __init__(self, session, path=None): self.session = session self.url = utils.urljoin(self.base_url, path) class Resource: def __init__(self, **data): self.data = data def __getattr__(self, attr): if attr not in self.data: error_message = 'this {!s} resource does not have a {!r} field' raise AttributeError(error_message.format(self.__class__.__name__, attr)) return self.data[attr] def __getstate__(self): return self.__dict__ def __setstate__(self, d): self.__dict__.update(d) class ManagedResource(Resource): """Class to represent an API object.""" def __init__(self, manager, **data): """Create an instance of the resource. :param manager: the resource's manager :type manager: :class:`~groupy.api.base.Manager` :param kwargs data: the resource data """ super().__init__(**data) self.manager = manager
from groupy import utils class Manager: """Class for interacting with the endpoint for a resource. :param session: the requests session :type session: :class:`~groupy.session.Session` :param str path: path relative to the base URL """ #: the base URL base_url = 'https://api.groupme.com/v3/' def __init__(self, session, path=None): self.session = session self.url = utils.urljoin(self.base_url, path) class Resource: def __init__(self, **data): self.data = data def __getattr__(self, attr): if attr not in self.data: error_message = 'this {!s} resource does not have a {!r} field' raise AttributeError(error_message.format(self.__class__.__name__, attr)) return self.data[attr] + def __getstate__(self): + return self.__dict__ + + def __setstate__(self, d): + self.__dict__.update(d) + class ManagedResource(Resource): """Class to represent an API object.""" def __init__(self, manager, **data): """Create an instance of the resource. :param manager: the resource's manager :type manager: :class:`~groupy.api.base.Manager` :param kwargs data: the resource data """ super().__init__(**data) self.manager = manager
93870152b4afb04f1547378184e2cee0bd0dd45f
kobo/apps/languages/serializers/base.py
kobo/apps/languages/serializers/base.py
from collections import defaultdict, OrderedDict from django.db import models from rest_framework import serializers class BaseServiceSerializer(serializers.ModelSerializer): class Meta: fields = [ 'name', 'code', ] class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer): region = serializers.SerializerMethodField() service = serializers.SerializerMethodField() language = serializers.SerializerMethodField() class Meta: fields = '__all__' def get_list_serializer_class(self): pass def get_region(self, through_instance): if through_instance.region: return through_instance.region.code return None def get_service(self, through_instance): return through_instance.service.code def get_language(self, through_instance): return through_instance.language.code class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer): def to_representation(self, data): """ Override `ListSerializer` behaviour to display services as a dictionary instead of a list """ iterable = data.all() if isinstance(data, models.Manager) else data formatted_data = self._get_formatted_data(iterable) representation = [] for service_code, service in formatted_data.items(): representation.append({service_code: service}) return representation
from collections import defaultdict, OrderedDict from django.db import models from rest_framework import serializers class BaseServiceSerializer(serializers.ModelSerializer): class Meta: fields = [ 'name', 'code', ] class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer): region = serializers.SerializerMethodField() service = serializers.SerializerMethodField() language = serializers.SerializerMethodField() class Meta: fields = '__all__' def get_list_serializer_class(self): pass def get_region(self, through_instance): if through_instance.region: return through_instance.region.code return None def get_service(self, through_instance): return through_instance.service.code def get_language(self, through_instance): return through_instance.language.code class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer): # Force `ListSerializer` to return a dict, not a list @property def data(self): ret = serializers.BaseSerializer.data.fget(self) return serializers.ReturnDict(ret, serializer=self) def to_representation(self, data): """ Override `ListSerializer` behaviour to display services as a dictionary instead of a list """ iterable = data.all() if isinstance(data, models.Manager) else data formatted_data = self._get_formatted_data(iterable) return formatted_data
Return a dictionary for transcription/translation services (instead of list)
Return a dictionary for transcription/translation services (instead of list)
Python
agpl-3.0
kobotoolbox/kpi,kobotoolbox/kpi,kobotoolbox/kpi,kobotoolbox/kpi,kobotoolbox/kpi
from collections import defaultdict, OrderedDict from django.db import models from rest_framework import serializers class BaseServiceSerializer(serializers.ModelSerializer): class Meta: fields = [ 'name', 'code', ] class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer): region = serializers.SerializerMethodField() service = serializers.SerializerMethodField() language = serializers.SerializerMethodField() class Meta: fields = '__all__' def get_list_serializer_class(self): pass def get_region(self, through_instance): if through_instance.region: return through_instance.region.code return None def get_service(self, through_instance): return through_instance.service.code def get_language(self, through_instance): return through_instance.language.code class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer): + # Force `ListSerializer` to return a dict, not a list + @property + def data(self): + ret = serializers.BaseSerializer.data.fget(self) + return serializers.ReturnDict(ret, serializer=self) + def to_representation(self, data): """ Override `ListSerializer` behaviour to display services as a dictionary instead of a list """ iterable = data.all() if isinstance(data, models.Manager) else data formatted_data = self._get_formatted_data(iterable) + return formatted_data - representation = [] - for service_code, service in formatted_data.items(): - representation.append({service_code: service}) - return representation
Return a dictionary for transcription/translation services (instead of list)
## Code Before: from collections import defaultdict, OrderedDict from django.db import models from rest_framework import serializers class BaseServiceSerializer(serializers.ModelSerializer): class Meta: fields = [ 'name', 'code', ] class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer): region = serializers.SerializerMethodField() service = serializers.SerializerMethodField() language = serializers.SerializerMethodField() class Meta: fields = '__all__' def get_list_serializer_class(self): pass def get_region(self, through_instance): if through_instance.region: return through_instance.region.code return None def get_service(self, through_instance): return through_instance.service.code def get_language(self, through_instance): return through_instance.language.code class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer): def to_representation(self, data): """ Override `ListSerializer` behaviour to display services as a dictionary instead of a list """ iterable = data.all() if isinstance(data, models.Manager) else data formatted_data = self._get_formatted_data(iterable) representation = [] for service_code, service in formatted_data.items(): representation.append({service_code: service}) return representation ## Instruction: Return a dictionary for transcription/translation services (instead of list) ## Code After: from collections import defaultdict, OrderedDict from django.db import models from rest_framework import serializers class BaseServiceSerializer(serializers.ModelSerializer): class Meta: fields = [ 'name', 'code', ] class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer): region = serializers.SerializerMethodField() service = serializers.SerializerMethodField() language = serializers.SerializerMethodField() class Meta: fields = '__all__' def get_list_serializer_class(self): pass def get_region(self, through_instance): if through_instance.region: return through_instance.region.code return None def get_service(self, through_instance): return through_instance.service.code def get_language(self, through_instance): return through_instance.language.code class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer): # Force `ListSerializer` to return a dict, not a list @property def data(self): ret = serializers.BaseSerializer.data.fget(self) return serializers.ReturnDict(ret, serializer=self) def to_representation(self, data): """ Override `ListSerializer` behaviour to display services as a dictionary instead of a list """ iterable = data.all() if isinstance(data, models.Manager) else data formatted_data = self._get_formatted_data(iterable) return formatted_data
from collections import defaultdict, OrderedDict from django.db import models from rest_framework import serializers class BaseServiceSerializer(serializers.ModelSerializer): class Meta: fields = [ 'name', 'code', ] class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer): region = serializers.SerializerMethodField() service = serializers.SerializerMethodField() language = serializers.SerializerMethodField() class Meta: fields = '__all__' def get_list_serializer_class(self): pass def get_region(self, through_instance): if through_instance.region: return through_instance.region.code return None def get_service(self, through_instance): return through_instance.service.code def get_language(self, through_instance): return through_instance.language.code class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer): + # Force `ListSerializer` to return a dict, not a list + @property + def data(self): + ret = serializers.BaseSerializer.data.fget(self) + return serializers.ReturnDict(ret, serializer=self) + def to_representation(self, data): """ Override `ListSerializer` behaviour to display services as a dictionary instead of a list """ iterable = data.all() if isinstance(data, models.Manager) else data formatted_data = self._get_formatted_data(iterable) + return formatted_data - representation = [] - for service_code, service in formatted_data.items(): - representation.append({service_code: service}) - return representation
78f049ce9713dabd3eec544494dadcab7ff93d4c
sui_hei/templatetags/markdown.py
sui_hei/templatetags/markdown.py
import re from bs4 import BeautifulSoup from django import template from django.template.defaultfilters import stringfilter from markdown import markdown as md register = template.Library() @stringfilter @register.filter(is_safe=True) def text2md(value): ''' convert markdown-like text to html. strip header <p> and footer </p> if p is True. ''' returns = md(value, ['markdown.extensions.extra']) returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8") return returns.decode('utf8') @register.filter(is_safe=True) @stringfilter def line2md(value, p=True): returns = BeautifulSoup(value, 'html5lib').get_text() returns = re.sub("^([*+-]) ", r"\\\1 ", returns) returns = md(returns, ['markdown.extensions.extra']) if p: returns = returns[3:-4] return returns
import re from bs4 import BeautifulSoup from django import template from django.template.defaultfilters import stringfilter from markdown import markdown as md from markdown.extensions.headerid import HeaderIdExtension register = template.Library() @stringfilter @register.filter(is_safe=True) def text2md(value): ''' convert markdown-like text to html. strip header <p> and footer </p> if p is True. ''' returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)]) returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8") return returns.decode('utf8') @register.filter(is_safe=True) @stringfilter def line2md(value, p=True): returns = BeautifulSoup(value, 'html5lib').get_text() returns = re.sub("^([*+-]) ", r"\\\1 ", returns) returns = md(returns, ['markdown.extensions.extra']) if p: returns = returns[3:-4] return returns
Add header id extension for github preferences
Add header id extension for github preferences
Python
mit
heyrict/cindy,heyrict/cindy,heyrict/cindy
import re from bs4 import BeautifulSoup from django import template from django.template.defaultfilters import stringfilter from markdown import markdown as md + from markdown.extensions.headerid import HeaderIdExtension register = template.Library() @stringfilter @register.filter(is_safe=True) def text2md(value): ''' convert markdown-like text to html. strip header <p> and footer </p> if p is True. ''' - returns = md(value, ['markdown.extensions.extra']) + returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)]) returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8") return returns.decode('utf8') @register.filter(is_safe=True) @stringfilter def line2md(value, p=True): returns = BeautifulSoup(value, 'html5lib').get_text() returns = re.sub("^([*+-]) ", r"\\\1 ", returns) returns = md(returns, ['markdown.extensions.extra']) if p: returns = returns[3:-4] return returns
Add header id extension for github preferences
## Code Before: import re from bs4 import BeautifulSoup from django import template from django.template.defaultfilters import stringfilter from markdown import markdown as md register = template.Library() @stringfilter @register.filter(is_safe=True) def text2md(value): ''' convert markdown-like text to html. strip header <p> and footer </p> if p is True. ''' returns = md(value, ['markdown.extensions.extra']) returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8") return returns.decode('utf8') @register.filter(is_safe=True) @stringfilter def line2md(value, p=True): returns = BeautifulSoup(value, 'html5lib').get_text() returns = re.sub("^([*+-]) ", r"\\\1 ", returns) returns = md(returns, ['markdown.extensions.extra']) if p: returns = returns[3:-4] return returns ## Instruction: Add header id extension for github preferences ## Code After: import re from bs4 import BeautifulSoup from django import template from django.template.defaultfilters import stringfilter from markdown import markdown as md from markdown.extensions.headerid import HeaderIdExtension register = template.Library() @stringfilter @register.filter(is_safe=True) def text2md(value): ''' convert markdown-like text to html. strip header <p> and footer </p> if p is True. ''' returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)]) returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8") return returns.decode('utf8') @register.filter(is_safe=True) @stringfilter def line2md(value, p=True): returns = BeautifulSoup(value, 'html5lib').get_text() returns = re.sub("^([*+-]) ", r"\\\1 ", returns) returns = md(returns, ['markdown.extensions.extra']) if p: returns = returns[3:-4] return returns
import re from bs4 import BeautifulSoup from django import template from django.template.defaultfilters import stringfilter from markdown import markdown as md + from markdown.extensions.headerid import HeaderIdExtension register = template.Library() @stringfilter @register.filter(is_safe=True) def text2md(value): ''' convert markdown-like text to html. strip header <p> and footer </p> if p is True. ''' - returns = md(value, ['markdown.extensions.extra']) + returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)]) ? ++++++++++++++++++++++++++++ returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8") return returns.decode('utf8') @register.filter(is_safe=True) @stringfilter def line2md(value, p=True): returns = BeautifulSoup(value, 'html5lib').get_text() returns = re.sub("^([*+-]) ", r"\\\1 ", returns) returns = md(returns, ['markdown.extensions.extra']) if p: returns = returns[3:-4] return returns
7aaa385da78bef57c8b6339f6db04044ace08807
api/taxonomies/serializers.py
api/taxonomies/serializers.py
from rest_framework import serializers as ser from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField class TaxonomyField(ser.Field): def to_representation(self, obj): if obj is not None: return {'id': obj._id, 'text': obj.text} return None def to_internal_value(self, data): return data class TaxonomySerializer(JSONAPISerializer): filterable_fields = frozenset([ 'text', 'parents', 'id' ]) id = ser.CharField(source='_id', required=True) text = ser.CharField(max_length=200) parents = JSONAPIListField(child=TaxonomyField()) links = LinksField({ 'parents': 'get_parent_urls', 'self': 'get_absolute_url', }) def get_parent_urls(self, obj): return [p.get_absolute_url() for p in obj.parents] def get_absolute_url(self, obj): return obj.get_absolute_url() class Meta: type_ = 'taxonomies'
from rest_framework import serializers as ser from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField class TaxonomyField(ser.Field): def to_representation(self, obj): if obj is not None: return {'id': obj._id, 'text': obj.text} return None def to_internal_value(self, data): return data class TaxonomySerializer(JSONAPISerializer): filterable_fields = frozenset([ 'text', 'parents', 'id' ]) id = ser.CharField(source='_id', required=True) text = ser.CharField(max_length=200) parents = JSONAPIListField(child=TaxonomyField()) child_count = ser.IntegerField() links = LinksField({ 'parents': 'get_parent_urls', 'self': 'get_absolute_url', }) def get_parent_urls(self, obj): return [p.get_absolute_url() for p in obj.parents] def get_absolute_url(self, obj): return obj.get_absolute_url() class Meta: type_ = 'taxonomies'
Add child_count taken from new Subject property
Add child_count taken from new Subject property
Python
apache-2.0
adlius/osf.io,brianjgeiger/osf.io,chrisseto/osf.io,rdhyee/osf.io,sloria/osf.io,brianjgeiger/osf.io,sloria/osf.io,binoculars/osf.io,mattclark/osf.io,saradbowman/osf.io,aaxelb/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,laurenrevere/osf.io,samchrisinger/osf.io,binoculars/osf.io,aaxelb/osf.io,rdhyee/osf.io,alexschiller/osf.io,HalcyonChimera/osf.io,mfraezz/osf.io,cwisecarver/osf.io,CenterForOpenScience/osf.io,acshi/osf.io,cslzchen/osf.io,mattclark/osf.io,crcresearch/osf.io,TomBaxter/osf.io,acshi/osf.io,samchrisinger/osf.io,monikagrabowska/osf.io,laurenrevere/osf.io,adlius/osf.io,acshi/osf.io,mluo613/osf.io,HalcyonChimera/osf.io,felliott/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,felliott/osf.io,mattclark/osf.io,saradbowman/osf.io,erinspace/osf.io,hmoco/osf.io,emetsger/osf.io,alexschiller/osf.io,icereval/osf.io,mfraezz/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,chrisseto/osf.io,erinspace/osf.io,monikagrabowska/osf.io,leb2dg/osf.io,pattisdr/osf.io,CenterForOpenScience/osf.io,monikagrabowska/osf.io,laurenrevere/osf.io,Nesiehr/osf.io,acshi/osf.io,cslzchen/osf.io,caseyrollins/osf.io,monikagrabowska/osf.io,emetsger/osf.io,adlius/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,crcresearch/osf.io,cwisecarver/osf.io,brianjgeiger/osf.io,mluo613/osf.io,Nesiehr/osf.io,sloria/osf.io,cslzchen/osf.io,icereval/osf.io,caseyrollins/osf.io,binoculars/osf.io,aaxelb/osf.io,caseyrollins/osf.io,cslzchen/osf.io,icereval/osf.io,CenterForOpenScience/osf.io,erinspace/osf.io,acshi/osf.io,brianjgeiger/osf.io,adlius/osf.io,caneruguz/osf.io,pattisdr/osf.io,cwisecarver/osf.io,mluo613/osf.io,chrisseto/osf.io,leb2dg/osf.io,mluo613/osf.io,alexschiller/osf.io,samchrisinger/osf.io,chennan47/osf.io,caneruguz/osf.io,Nesiehr/osf.io,CenterForOpenScience/osf.io,rdhyee/osf.io,chrisseto/osf.io,mluo613/osf.io,Johnetordoff/osf.io,rdhyee/osf.io,felliott/osf.io,mfraezz/osf.io,aaxelb/osf.io,chennan47/osf.io,crcresearch/osf.io,alexschiller/osf.io,pattisdr/osf.io,HalcyonChimera/osf.io,emetsger/osf.io,leb2dg/osf.io,emetsger/osf.io,cwisecarver/osf.io,alexschiller/osf.io,Nesiehr/osf.io,hmoco/osf.io,hmoco/osf.io,felliott/osf.io,chennan47/osf.io,samchrisinger/osf.io,mfraezz/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,TomBaxter/osf.io,baylee-d/osf.io
from rest_framework import serializers as ser from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField class TaxonomyField(ser.Field): def to_representation(self, obj): if obj is not None: return {'id': obj._id, 'text': obj.text} return None def to_internal_value(self, data): return data class TaxonomySerializer(JSONAPISerializer): filterable_fields = frozenset([ 'text', 'parents', 'id' ]) id = ser.CharField(source='_id', required=True) text = ser.CharField(max_length=200) parents = JSONAPIListField(child=TaxonomyField()) + child_count = ser.IntegerField() links = LinksField({ 'parents': 'get_parent_urls', 'self': 'get_absolute_url', }) def get_parent_urls(self, obj): return [p.get_absolute_url() for p in obj.parents] def get_absolute_url(self, obj): return obj.get_absolute_url() class Meta: type_ = 'taxonomies'
Add child_count taken from new Subject property
## Code Before: from rest_framework import serializers as ser from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField class TaxonomyField(ser.Field): def to_representation(self, obj): if obj is not None: return {'id': obj._id, 'text': obj.text} return None def to_internal_value(self, data): return data class TaxonomySerializer(JSONAPISerializer): filterable_fields = frozenset([ 'text', 'parents', 'id' ]) id = ser.CharField(source='_id', required=True) text = ser.CharField(max_length=200) parents = JSONAPIListField(child=TaxonomyField()) links = LinksField({ 'parents': 'get_parent_urls', 'self': 'get_absolute_url', }) def get_parent_urls(self, obj): return [p.get_absolute_url() for p in obj.parents] def get_absolute_url(self, obj): return obj.get_absolute_url() class Meta: type_ = 'taxonomies' ## Instruction: Add child_count taken from new Subject property ## Code After: from rest_framework import serializers as ser from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField class TaxonomyField(ser.Field): def to_representation(self, obj): if obj is not None: return {'id': obj._id, 'text': obj.text} return None def to_internal_value(self, data): return data class TaxonomySerializer(JSONAPISerializer): filterable_fields = frozenset([ 'text', 'parents', 'id' ]) id = ser.CharField(source='_id', required=True) text = ser.CharField(max_length=200) parents = JSONAPIListField(child=TaxonomyField()) child_count = ser.IntegerField() links = LinksField({ 'parents': 'get_parent_urls', 'self': 'get_absolute_url', }) def get_parent_urls(self, obj): return [p.get_absolute_url() for p in obj.parents] def get_absolute_url(self, obj): return obj.get_absolute_url() class Meta: type_ = 'taxonomies'
from rest_framework import serializers as ser from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField class TaxonomyField(ser.Field): def to_representation(self, obj): if obj is not None: return {'id': obj._id, 'text': obj.text} return None def to_internal_value(self, data): return data class TaxonomySerializer(JSONAPISerializer): filterable_fields = frozenset([ 'text', 'parents', 'id' ]) id = ser.CharField(source='_id', required=True) text = ser.CharField(max_length=200) parents = JSONAPIListField(child=TaxonomyField()) + child_count = ser.IntegerField() links = LinksField({ 'parents': 'get_parent_urls', 'self': 'get_absolute_url', }) def get_parent_urls(self, obj): return [p.get_absolute_url() for p in obj.parents] def get_absolute_url(self, obj): return obj.get_absolute_url() class Meta: type_ = 'taxonomies'
a222d268ec1c12466db48bbfcd58d8ecf2907805
echo_server.py
echo_server.py
import socket class EchoServer(object): """a simple EchoServer""" def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5): self.ip = ip self.port = port self.backlog = backlog self.socket = socket.socket( socket.AF_INET, socket.SOCK_STREAM, socket.IPPROTO_IP) self.socket.bind((self.ip, self.port)) self.socket.listen(self.backlog) def start_listening(self): while True: self.connection, self.addr = self.socket.accept() words = self.connection.recv(32) if words: self.connection.sendall(unicode(words)) self.connection.close() self.socket.close() break if __name__ == "__main__": server = EchoServer() server.start_listening()
import socket class EchoServer(object): """a simple EchoServer""" def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5): self.ip = ip self.port = port self.backlog = backlog self.socket = socket.socket( socket.AF_INET, socket.SOCK_STREAM, socket.IPPROTO_IP) self.socket.bind((self.ip, self.port)) self.socket.listen(self.backlog) def start_listening(self): while True: request = [] self.connection, self.addr = self.socket.accept() while True: buffer_ = self.connection.recv(32) if buffer_: request.append(buffer_) else: break self.connection.sendall(" ".join(request)) self.connection.close() if __name__ == "__main__": server = EchoServer() server.start_listening()
Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests
Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests
Python
mit
jefrailey/network_tools
import socket class EchoServer(object): """a simple EchoServer""" def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5): self.ip = ip self.port = port self.backlog = backlog self.socket = socket.socket( socket.AF_INET, socket.SOCK_STREAM, socket.IPPROTO_IP) self.socket.bind((self.ip, self.port)) self.socket.listen(self.backlog) def start_listening(self): while True: + request = [] self.connection, self.addr = self.socket.accept() + + while True: - words = self.connection.recv(32) + buffer_ = self.connection.recv(32) - if words: + if buffer_: + request.append(buffer_) + else: + break - self.connection.sendall(unicode(words)) + self.connection.sendall(" ".join(request)) self.connection.close() + - self.socket.close() - break if __name__ == "__main__": server = EchoServer() server.start_listening()
Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests
## Code Before: import socket class EchoServer(object): """a simple EchoServer""" def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5): self.ip = ip self.port = port self.backlog = backlog self.socket = socket.socket( socket.AF_INET, socket.SOCK_STREAM, socket.IPPROTO_IP) self.socket.bind((self.ip, self.port)) self.socket.listen(self.backlog) def start_listening(self): while True: self.connection, self.addr = self.socket.accept() words = self.connection.recv(32) if words: self.connection.sendall(unicode(words)) self.connection.close() self.socket.close() break if __name__ == "__main__": server = EchoServer() server.start_listening() ## Instruction: Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests ## Code After: import socket class EchoServer(object): """a simple EchoServer""" def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5): self.ip = ip self.port = port self.backlog = backlog self.socket = socket.socket( socket.AF_INET, socket.SOCK_STREAM, socket.IPPROTO_IP) self.socket.bind((self.ip, self.port)) self.socket.listen(self.backlog) def start_listening(self): while True: request = [] self.connection, self.addr = self.socket.accept() while True: buffer_ = self.connection.recv(32) if buffer_: request.append(buffer_) else: break self.connection.sendall(" ".join(request)) self.connection.close() if __name__ == "__main__": server = EchoServer() server.start_listening()
import socket class EchoServer(object): """a simple EchoServer""" def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5): self.ip = ip self.port = port self.backlog = backlog self.socket = socket.socket( socket.AF_INET, socket.SOCK_STREAM, socket.IPPROTO_IP) self.socket.bind((self.ip, self.port)) self.socket.listen(self.backlog) def start_listening(self): while True: + request = [] self.connection, self.addr = self.socket.accept() + + while True: - words = self.connection.recv(32) ? ^^ ^^ + buffer_ = self.connection.recv(32) ? ^^^^^^^^^ ^ - if words: + if buffer_: + request.append(buffer_) + else: + break - self.connection.sendall(unicode(words)) ? ---- ----- ----- + self.connection.sendall(" ".join(request)) ? ++++++++++++ + self.connection.close() + - self.socket.close() - break if __name__ == "__main__": server = EchoServer() server.start_listening()
c7a424cb3fb0a037cda04c30d44606515aed829d
chrome/test/functional/test_pyauto.py
chrome/test/functional/test_pyauto.py
import unittest import pyauto_functional # Must be imported before pyauto import pyauto class PyAutoTest(pyauto.PyUITest): """Test functionality of the PyAuto framework.""" _EXTRA_CHROME_FLAGS = [ '--scooby-doo=123', '--donald-duck=cool', '--super-mario', '--marvin-the-martian', ] def ExtraChromeFlags(self): """Ensures Chrome is launched with some custom flags. Overrides the default list of extra flags passed to Chrome. See ExtraChromeFlags() in pyauto.py. """ return self._EXTRA_CHROME_FLAGS def testSetCustomChromeFlags(self): """Ensures that Chrome can be launched with custom flags.""" self.NavigateToURL('about://version') for flag in self._EXTRA_CHROME_FLAGS: self.assertEqual(self.FindInPage(flag)['match_count'], 1, msg='Missing expected Chrome flag "%s"' % flag) if __name__ == '__main__': pyauto_functional.Main()
import unittest import pyauto_functional # Must be imported before pyauto import pyauto class PyAutoTest(pyauto.PyUITest): """Test functionality of the PyAuto framework.""" _EXTRA_CHROME_FLAGS = [ '--scooby-doo=123', '--donald-duck=cool', '--super-mario', '--marvin-the-martian', ] def ExtraChromeFlags(self): """Ensures Chrome is launched with some custom flags. Overrides the default list of extra flags passed to Chrome. See ExtraChromeFlags() in pyauto.py. """ return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS def testSetCustomChromeFlags(self): """Ensures that Chrome can be launched with custom flags.""" self.NavigateToURL('about://version') for flag in self._EXTRA_CHROME_FLAGS: self.assertEqual(self.FindInPage(flag)['match_count'], 1, msg='Missing expected Chrome flag "%s"' % flag) if __name__ == '__main__': pyauto_functional.Main()
Update testSetCustomChromeFlags to only append to chrome flags, not override
Update testSetCustomChromeFlags to only append to chrome flags, not override This is in keeping with the spirit of the ExtraChromeFlags() method. Whenever it's overridden, it should only ever append to the list of chrome flags, never completely override it. BUG=None TEST=None R=dennisjeffrey@chromium.org Review URL: https://chromiumcodereview.appspot.com/10041001 git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@131597 0039d316-1c4b-4281-b951-d872f2087c98
Python
bsd-3-clause
Jonekee/chromium.src,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,patrickm/chromium.src,littlstar/chromium.src,hujiajie/pa-chromium,jaruba/chromium.src,hgl888/chromium-crosswalk,hujiajie/pa-chromium,junmin-zhu/chromium-rivertrail,PeterWangIntel/chromium-crosswalk,keishi/chromium,jaruba/chromium.src,junmin-zhu/chromium-rivertrail,bright-sparks/chromium-spacewalk,jaruba/chromium.src,crosswalk-project/chromium-crosswalk-efl,littlstar/chromium.src,jaruba/chromium.src,anirudhSK/chromium,crosswalk-project/chromium-crosswalk-efl,jaruba/chromium.src,robclark/chromium,ChromiumWebApps/chromium,TheTypoMaster/chromium-crosswalk,timopulkkinen/BubbleFish,markYoungH/chromium.src,hgl888/chromium-crosswalk,ondra-novak/chromium.src,chuan9/chromium-crosswalk,robclark/chromium,hujiajie/pa-chromium,hgl888/chromium-crosswalk,markYoungH/chromium.src,chuan9/chromium-crosswalk,Jonekee/chromium.src,ltilve/chromium,hgl888/chromium-crosswalk-efl,ChromiumWebApps/chromium,PeterWangIntel/chromium-crosswalk,hgl888/chromium-crosswalk,keishi/chromium,Chilledheart/chromium,jaruba/chromium.src,zcbenz/cefode-chromium,axinging/chromium-crosswalk,fujunwei/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,pozdnyakov/chromium-crosswalk,hujiajie/pa-chromium,nacl-webkit/chrome_deps,hgl888/chromium-crosswalk-efl,PeterWangIntel/chromium-crosswalk,dushu1203/chromium.src,M4sse/chromium.src,PeterWangIntel/chromium-crosswalk,patrickm/chromium.src,mohamed--abdel-maksoud/chromium.src,krieger-od/nwjs_chromium.src,mohamed--abdel-maksoud/chromium.src,chuan9/chromium-crosswalk,junmin-zhu/chromium-rivertrail,PeterWangIntel/chromium-crosswalk,krieger-od/nwjs_chromium.src,anirudhSK/chromium,mohamed--abdel-maksoud/chromium.src,chuan9/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,axinging/chromium-crosswalk,axinging/chromium-crosswalk,timopulkkinen/BubbleFish,mohamed--abdel-maksoud/chromium.src,Fireblend/chromium-crosswalk,hujiajie/pa-chromium,anirudhSK/chromium,markYoungH/chromium.src,mogoweb/chromium-crosswalk,Chilledheart/chromium,nacl-webkit/chrome_deps,hgl888/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,Just-D/chromium-1,ChromiumWebApps/chromium,hujiajie/pa-chromium,axinging/chromium-crosswalk,ChromiumWebApps/chromium,markYoungH/chromium.src,ondra-novak/chromium.src,keishi/chromium,ondra-novak/chromium.src,pozdnyakov/chromium-crosswalk,bright-sparks/chromium-spacewalk,Pluto-tv/chromium-crosswalk,M4sse/chromium.src,timopulkkinen/BubbleFish,timopulkkinen/BubbleFish,Jonekee/chromium.src,Just-D/chromium-1,dushu1203/chromium.src,krieger-od/nwjs_chromium.src,keishi/chromium,junmin-zhu/chromium-rivertrail,hgl888/chromium-crosswalk-efl,Chilledheart/chromium,M4sse/chromium.src,chuan9/chromium-crosswalk,dednal/chromium.src,ChromiumWebApps/chromium,mohamed--abdel-maksoud/chromium.src,nacl-webkit/chrome_deps,ondra-novak/chromium.src,Pluto-tv/chromium-crosswalk,Just-D/chromium-1,PeterWangIntel/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,ondra-novak/chromium.src,markYoungH/chromium.src,littlstar/chromium.src,robclark/chromium,ltilve/chromium,ondra-novak/chromium.src,markYoungH/chromium.src,Chilledheart/chromium,robclark/chromium,hgl888/chromium-crosswalk-efl,Fireblend/chromium-crosswalk,axinging/chromium-crosswalk,mogoweb/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Jonekee/chromium.src,M4sse/chromium.src,pozdnyakov/chromium-crosswalk,hgl888/chromium-crosswalk,bright-sparks/chromium-spacewalk,dednal/chromium.src,zcbenz/cefode-chromium,Chilledheart/chromium,crosswalk-project/chromium-crosswalk-efl,mogoweb/chromium-crosswalk,Chilledheart/chromium,nacl-webkit/chrome_deps,dushu1203/chromium.src,zcbenz/cefode-chromium,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,junmin-zhu/chromium-rivertrail,Fireblend/chromium-crosswalk,patrickm/chromium.src,fujunwei/chromium-crosswalk,timopulkkinen/BubbleFish,PeterWangIntel/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,nacl-webkit/chrome_deps,M4sse/chromium.src,zcbenz/cefode-chromium,ondra-novak/chromium.src,dednal/chromium.src,bright-sparks/chromium-spacewalk,axinging/chromium-crosswalk,keishi/chromium,Fireblend/chromium-crosswalk,junmin-zhu/chromium-rivertrail,junmin-zhu/chromium-rivertrail,dushu1203/chromium.src,keishi/chromium,robclark/chromium,fujunwei/chromium-crosswalk,markYoungH/chromium.src,hgl888/chromium-crosswalk,ltilve/chromium,patrickm/chromium.src,Just-D/chromium-1,mogoweb/chromium-crosswalk,hujiajie/pa-chromium,PeterWangIntel/chromium-crosswalk,hujiajie/pa-chromium,chuan9/chromium-crosswalk,dednal/chromium.src,Chilledheart/chromium,crosswalk-project/chromium-crosswalk-efl,anirudhSK/chromium,mohamed--abdel-maksoud/chromium.src,junmin-zhu/chromium-rivertrail,littlstar/chromium.src,anirudhSK/chromium,zcbenz/cefode-chromium,krieger-od/nwjs_chromium.src,keishi/chromium,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,M4sse/chromium.src,dushu1203/chromium.src,ChromiumWebApps/chromium,jaruba/chromium.src,Pluto-tv/chromium-crosswalk,nacl-webkit/chrome_deps,fujunwei/chromium-crosswalk,zcbenz/cefode-chromium,M4sse/chromium.src,krieger-od/nwjs_chromium.src,chuan9/chromium-crosswalk,hgl888/chromium-crosswalk-efl,nacl-webkit/chrome_deps,patrickm/chromium.src,axinging/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,patrickm/chromium.src,Jonekee/chromium.src,anirudhSK/chromium,hgl888/chromium-crosswalk-efl,pozdnyakov/chromium-crosswalk,Fireblend/chromium-crosswalk,axinging/chromium-crosswalk,dushu1203/chromium.src,patrickm/chromium.src,Fireblend/chromium-crosswalk,bright-sparks/chromium-spacewalk,ChromiumWebApps/chromium,keishi/chromium,TheTypoMaster/chromium-crosswalk,jaruba/chromium.src,anirudhSK/chromium,hgl888/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,anirudhSK/chromium,Chilledheart/chromium,hujiajie/pa-chromium,keishi/chromium,robclark/chromium,robclark/chromium,Fireblend/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,keishi/chromium,robclark/chromium,hgl888/chromium-crosswalk-efl,Pluto-tv/chromium-crosswalk,dednal/chromium.src,mogoweb/chromium-crosswalk,hgl888/chromium-crosswalk-efl,fujunwei/chromium-crosswalk,Fireblend/chromium-crosswalk,ltilve/chromium,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,littlstar/chromium.src,Jonekee/chromium.src,Just-D/chromium-1,M4sse/chromium.src,Pluto-tv/chromium-crosswalk,mogoweb/chromium-crosswalk,nacl-webkit/chrome_deps,ChromiumWebApps/chromium,keishi/chromium,mogoweb/chromium-crosswalk,Jonekee/chromium.src,fujunwei/chromium-crosswalk,dushu1203/chromium.src,anirudhSK/chromium,zcbenz/cefode-chromium,mohamed--abdel-maksoud/chromium.src,littlstar/chromium.src,ltilve/chromium,dednal/chromium.src,axinging/chromium-crosswalk,jaruba/chromium.src,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,mogoweb/chromium-crosswalk,chuan9/chromium-crosswalk,pozdnyakov/chromium-crosswalk,Pluto-tv/chromium-crosswalk,ChromiumWebApps/chromium,bright-sparks/chromium-spacewalk,Jonekee/chromium.src,jaruba/chromium.src,Just-D/chromium-1,littlstar/chromium.src,mogoweb/chromium-crosswalk,markYoungH/chromium.src,nacl-webkit/chrome_deps,pozdnyakov/chromium-crosswalk,ChromiumWebApps/chromium,Just-D/chromium-1,mogoweb/chromium-crosswalk,dushu1203/chromium.src,Jonekee/chromium.src,bright-sparks/chromium-spacewalk,junmin-zhu/chromium-rivertrail,TheTypoMaster/chromium-crosswalk,dushu1203/chromium.src,ChromiumWebApps/chromium,junmin-zhu/chromium-rivertrail,dushu1203/chromium.src,zcbenz/cefode-chromium,hgl888/chromium-crosswalk-efl,littlstar/chromium.src,crosswalk-project/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,dednal/chromium.src,nacl-webkit/chrome_deps,anirudhSK/chromium,markYoungH/chromium.src,M4sse/chromium.src,TheTypoMaster/chromium-crosswalk,krieger-od/nwjs_chromium.src,M4sse/chromium.src,chuan9/chromium-crosswalk,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,zcbenz/cefode-chromium,Just-D/chromium-1,ltilve/chromium,ondra-novak/chromium.src,dednal/chromium.src,ltilve/chromium,M4sse/chromium.src,ondra-novak/chromium.src,jaruba/chromium.src,fujunwei/chromium-crosswalk,Jonekee/chromium.src,TheTypoMaster/chromium-crosswalk,markYoungH/chromium.src,Fireblend/chromium-crosswalk,ltilve/chromium,robclark/chromium,patrickm/chromium.src,dednal/chromium.src,anirudhSK/chromium,Jonekee/chromium.src,pozdnyakov/chromium-crosswalk,Chilledheart/chromium,ChromiumWebApps/chromium,zcbenz/cefode-chromium,timopulkkinen/BubbleFish,axinging/chromium-crosswalk,krieger-od/nwjs_chromium.src,anirudhSK/chromium,Pluto-tv/chromium-crosswalk,nacl-webkit/chrome_deps,ltilve/chromium,crosswalk-project/chromium-crosswalk-efl,fujunwei/chromium-crosswalk,timopulkkinen/BubbleFish,axinging/chromium-crosswalk,dednal/chromium.src,hujiajie/pa-chromium,timopulkkinen/BubbleFish,pozdnyakov/chromium-crosswalk,Pluto-tv/chromium-crosswalk,junmin-zhu/chromium-rivertrail,krieger-od/nwjs_chromium.src,zcbenz/cefode-chromium,robclark/chromium,markYoungH/chromium.src,hujiajie/pa-chromium,dushu1203/chromium.src,fujunwei/chromium-crosswalk,Pluto-tv/chromium-crosswalk,bright-sparks/chromium-spacewalk,patrickm/chromium.src,dednal/chromium.src,Just-D/chromium-1,bright-sparks/chromium-spacewalk
import unittest import pyauto_functional # Must be imported before pyauto import pyauto class PyAutoTest(pyauto.PyUITest): """Test functionality of the PyAuto framework.""" _EXTRA_CHROME_FLAGS = [ '--scooby-doo=123', '--donald-duck=cool', '--super-mario', '--marvin-the-martian', ] def ExtraChromeFlags(self): """Ensures Chrome is launched with some custom flags. Overrides the default list of extra flags passed to Chrome. See ExtraChromeFlags() in pyauto.py. """ - return self._EXTRA_CHROME_FLAGS + return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS def testSetCustomChromeFlags(self): """Ensures that Chrome can be launched with custom flags.""" self.NavigateToURL('about://version') for flag in self._EXTRA_CHROME_FLAGS: self.assertEqual(self.FindInPage(flag)['match_count'], 1, msg='Missing expected Chrome flag "%s"' % flag) if __name__ == '__main__': pyauto_functional.Main()
Update testSetCustomChromeFlags to only append to chrome flags, not override
## Code Before: import unittest import pyauto_functional # Must be imported before pyauto import pyauto class PyAutoTest(pyauto.PyUITest): """Test functionality of the PyAuto framework.""" _EXTRA_CHROME_FLAGS = [ '--scooby-doo=123', '--donald-duck=cool', '--super-mario', '--marvin-the-martian', ] def ExtraChromeFlags(self): """Ensures Chrome is launched with some custom flags. Overrides the default list of extra flags passed to Chrome. See ExtraChromeFlags() in pyauto.py. """ return self._EXTRA_CHROME_FLAGS def testSetCustomChromeFlags(self): """Ensures that Chrome can be launched with custom flags.""" self.NavigateToURL('about://version') for flag in self._EXTRA_CHROME_FLAGS: self.assertEqual(self.FindInPage(flag)['match_count'], 1, msg='Missing expected Chrome flag "%s"' % flag) if __name__ == '__main__': pyauto_functional.Main() ## Instruction: Update testSetCustomChromeFlags to only append to chrome flags, not override ## Code After: import unittest import pyauto_functional # Must be imported before pyauto import pyauto class PyAutoTest(pyauto.PyUITest): """Test functionality of the PyAuto framework.""" _EXTRA_CHROME_FLAGS = [ '--scooby-doo=123', '--donald-duck=cool', '--super-mario', '--marvin-the-martian', ] def ExtraChromeFlags(self): """Ensures Chrome is launched with some custom flags. Overrides the default list of extra flags passed to Chrome. See ExtraChromeFlags() in pyauto.py. """ return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS def testSetCustomChromeFlags(self): """Ensures that Chrome can be launched with custom flags.""" self.NavigateToURL('about://version') for flag in self._EXTRA_CHROME_FLAGS: self.assertEqual(self.FindInPage(flag)['match_count'], 1, msg='Missing expected Chrome flag "%s"' % flag) if __name__ == '__main__': pyauto_functional.Main()
import unittest import pyauto_functional # Must be imported before pyauto import pyauto class PyAutoTest(pyauto.PyUITest): """Test functionality of the PyAuto framework.""" _EXTRA_CHROME_FLAGS = [ '--scooby-doo=123', '--donald-duck=cool', '--super-mario', '--marvin-the-martian', ] def ExtraChromeFlags(self): """Ensures Chrome is launched with some custom flags. Overrides the default list of extra flags passed to Chrome. See ExtraChromeFlags() in pyauto.py. """ - return self._EXTRA_CHROME_FLAGS + return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS def testSetCustomChromeFlags(self): """Ensures that Chrome can be launched with custom flags.""" self.NavigateToURL('about://version') for flag in self._EXTRA_CHROME_FLAGS: self.assertEqual(self.FindInPage(flag)['match_count'], 1, msg='Missing expected Chrome flag "%s"' % flag) if __name__ == '__main__': pyauto_functional.Main()
f2139cad673ee50f027164bda80d86979d5ce7a0
passenger_wsgi.py
passenger_wsgi.py
import os import sys try: from flask import Flask, render_template, send_file, Response import requests except ImportError: INTERP = "venv/bin/python" if os.path.relpath(sys.executable, os.getcwd()) != INTERP: try: os.execl(INTERP, INTERP, *sys.argv) except OSError: sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`") else: sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`") application = Flask(__name__) @application.route("/") def index(): return "Hello, world!"
import os import sys try: from flask import Flask import flask_login from flask_restless import APIManager from flask_sqlalchemy import SQLAlchemy import requests except ImportError: INTERP = "venv/bin/python" if os.path.relpath(sys.executable, os.getcwd()) != INTERP: try: os.execl(INTERP, INTERP, *sys.argv) except OSError: sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`") else: sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`") application = Flask(__name__) @application.route("/") def index(): return "Hello, world!"
Add more imports for further functionality
Add more imports for further functionality `flask_login`, `flask_restless`, `flask_sqlalchemy`
Python
mit
GregBrimble/boilerplate-web-service,GregBrimble/boilerplate-web-service
import os import sys try: - from flask import Flask, render_template, send_file, Response + from flask import Flask + import flask_login + from flask_restless import APIManager + from flask_sqlalchemy import SQLAlchemy import requests except ImportError: INTERP = "venv/bin/python" if os.path.relpath(sys.executable, os.getcwd()) != INTERP: try: os.execl(INTERP, INTERP, *sys.argv) except OSError: sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`") else: sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`") application = Flask(__name__) @application.route("/") def index(): return "Hello, world!"
Add more imports for further functionality
## Code Before: import os import sys try: from flask import Flask, render_template, send_file, Response import requests except ImportError: INTERP = "venv/bin/python" if os.path.relpath(sys.executable, os.getcwd()) != INTERP: try: os.execl(INTERP, INTERP, *sys.argv) except OSError: sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`") else: sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`") application = Flask(__name__) @application.route("/") def index(): return "Hello, world!" ## Instruction: Add more imports for further functionality ## Code After: import os import sys try: from flask import Flask import flask_login from flask_restless import APIManager from flask_sqlalchemy import SQLAlchemy import requests except ImportError: INTERP = "venv/bin/python" if os.path.relpath(sys.executable, os.getcwd()) != INTERP: try: os.execl(INTERP, INTERP, *sys.argv) except OSError: sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`") else: sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`") application = Flask(__name__) @application.route("/") def index(): return "Hello, world!"
import os import sys try: - from flask import Flask, render_template, send_file, Response + from flask import Flask + import flask_login + from flask_restless import APIManager + from flask_sqlalchemy import SQLAlchemy import requests except ImportError: INTERP = "venv/bin/python" if os.path.relpath(sys.executable, os.getcwd()) != INTERP: try: os.execl(INTERP, INTERP, *sys.argv) except OSError: sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`") else: sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`") application = Flask(__name__) @application.route("/") def index(): return "Hello, world!"
b16c49cfd6a0ee659e4493ef959e0483e93d350a
os_client_config/defaults.py
os_client_config/defaults.py
_defaults = dict( auth_type='password', compute_api_version='2', floating_ip_source='neutron', identity_api_version='2', image_api_use_tasks=False, image_api_version='1', network_api_version='2', object_api_version='1', volume_api_version='1', )
_defaults = dict( auth_type='password', baremetal_api_version='1', compute_api_version='2', database_api_version='1.0', floating_ip_source='neutron', identity_api_version='2', image_api_use_tasks=False, image_api_version='1', network_api_version='2', object_api_version='1', volume_api_version='1', )
Add default versions for trove and ironic
Add default versions for trove and ironic Change-Id: Ib7af38664cfbe75c78c70693117f1193c4beb7e6
Python
apache-2.0
openstack/python-openstacksdk,stackforge/python-openstacksdk,redhat-openstack/os-client-config,dtroyer/python-openstacksdk,openstack/os-client-config,dtroyer/python-openstacksdk,switch-ch/os-client-config,stackforge/python-openstacksdk,dtroyer/os-client-config,openstack/python-openstacksdk
_defaults = dict( auth_type='password', + baremetal_api_version='1', compute_api_version='2', + database_api_version='1.0', floating_ip_source='neutron', identity_api_version='2', image_api_use_tasks=False, image_api_version='1', network_api_version='2', object_api_version='1', volume_api_version='1', )
Add default versions for trove and ironic
## Code Before: _defaults = dict( auth_type='password', compute_api_version='2', floating_ip_source='neutron', identity_api_version='2', image_api_use_tasks=False, image_api_version='1', network_api_version='2', object_api_version='1', volume_api_version='1', ) ## Instruction: Add default versions for trove and ironic ## Code After: _defaults = dict( auth_type='password', baremetal_api_version='1', compute_api_version='2', database_api_version='1.0', floating_ip_source='neutron', identity_api_version='2', image_api_use_tasks=False, image_api_version='1', network_api_version='2', object_api_version='1', volume_api_version='1', )
_defaults = dict( auth_type='password', + baremetal_api_version='1', compute_api_version='2', + database_api_version='1.0', floating_ip_source='neutron', identity_api_version='2', image_api_use_tasks=False, image_api_version='1', network_api_version='2', object_api_version='1', volume_api_version='1', )
216f0bb3680b86ac2dfc8c506b791db4e34eeee6
nextactions/board.py
nextactions/board.py
from nextactions.list import List class Board: def __init__(self, trello, json): self._trello = trello self.id = json['id'] self.name = json['name'] self.nextActionList = [] def getLists(self): json = self._trello.get( 'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)' ) return [List(j) for j in json] def getListByName(self, name): for l in self.getLists(): if l.name == name: return l return None
from nextactions.list import List class Board: def __init__(self, trello, json): self._trello = trello self.id = json['id'] self.name = json['name'] self.nextActionList = [] def getLists(self): json = self._trello.get( 'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)' ) return [List(j) for j in json] def getListByName(self, name): matches = [l for l in self.getLists() if l.name == name] return matches[0] if len(matches) else None
Tidy matching lists by name
Tidy matching lists by name
Python
mit
stevecshanks/trello-next-actions
from nextactions.list import List class Board: def __init__(self, trello, json): self._trello = trello self.id = json['id'] self.name = json['name'] self.nextActionList = [] def getLists(self): json = self._trello.get( 'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)' ) return [List(j) for j in json] def getListByName(self, name): + matches = [l for l in self.getLists() if l.name == name] + return matches[0] if len(matches) else None - for l in self.getLists(): - if l.name == name: - return l - return None
Tidy matching lists by name
## Code Before: from nextactions.list import List class Board: def __init__(self, trello, json): self._trello = trello self.id = json['id'] self.name = json['name'] self.nextActionList = [] def getLists(self): json = self._trello.get( 'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)' ) return [List(j) for j in json] def getListByName(self, name): for l in self.getLists(): if l.name == name: return l return None ## Instruction: Tidy matching lists by name ## Code After: from nextactions.list import List class Board: def __init__(self, trello, json): self._trello = trello self.id = json['id'] self.name = json['name'] self.nextActionList = [] def getLists(self): json = self._trello.get( 'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)' ) return [List(j) for j in json] def getListByName(self, name): matches = [l for l in self.getLists() if l.name == name] return matches[0] if len(matches) else None
from nextactions.list import List class Board: def __init__(self, trello, json): self._trello = trello self.id = json['id'] self.name = json['name'] self.nextActionList = [] def getLists(self): json = self._trello.get( 'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)' ) return [List(j) for j in json] def getListByName(self, name): + matches = [l for l in self.getLists() if l.name == name] + return matches[0] if len(matches) else None - for l in self.getLists(): - if l.name == name: - return l - return None
2facb0c8794c9529ccb17631a90b0ee181c4eb5b
xml_json_import/__init__.py
xml_json_import/__init__.py
from django.conf import settings from os import path class XmlJsonImportModuleException(Exception): pass if not hasattr(settings, 'XSLT_FILES_DIR'): raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter') if not path.exists(settings.XSLT_FILES_DIR): raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist')
from django.conf import settings from os import path, listdir from lxml import etree class XmlJsonImportModuleException(Exception): pass if not hasattr(settings, 'XSLT_FILES_DIR'): raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter') if not path.exists(settings.XSLT_FILES_DIR): raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist') for filename in listdir(settings.XSLT_FILES_DIR): filepath = path.join(settings.XSLT_FILES_DIR, filename) if path.isfile(filepath): try: xslt_etree = etree.parse(filepath) except etree.XMLSyntaxError as er: raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er)) try: transform = etree.XSLT(xslt_etree) except etree.XSLTParseError as er: raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er))
Add exception handling for invalid XSLT files
Add exception handling for invalid XSLT files
Python
mit
lev-veshnyakov/django-import-data,lev-veshnyakov/django-import-data
from django.conf import settings - from os import path + from os import path, listdir + from lxml import etree class XmlJsonImportModuleException(Exception): pass if not hasattr(settings, 'XSLT_FILES_DIR'): raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter') if not path.exists(settings.XSLT_FILES_DIR): raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist') + + for filename in listdir(settings.XSLT_FILES_DIR): + filepath = path.join(settings.XSLT_FILES_DIR, filename) + if path.isfile(filepath): + try: + xslt_etree = etree.parse(filepath) + except etree.XMLSyntaxError as er: + raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er)) + try: + transform = etree.XSLT(xslt_etree) + except etree.XSLTParseError as er: + raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er))
Add exception handling for invalid XSLT files
## Code Before: from django.conf import settings from os import path class XmlJsonImportModuleException(Exception): pass if not hasattr(settings, 'XSLT_FILES_DIR'): raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter') if not path.exists(settings.XSLT_FILES_DIR): raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist') ## Instruction: Add exception handling for invalid XSLT files ## Code After: from django.conf import settings from os import path, listdir from lxml import etree class XmlJsonImportModuleException(Exception): pass if not hasattr(settings, 'XSLT_FILES_DIR'): raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter') if not path.exists(settings.XSLT_FILES_DIR): raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist') for filename in listdir(settings.XSLT_FILES_DIR): filepath = path.join(settings.XSLT_FILES_DIR, filename) if path.isfile(filepath): try: xslt_etree = etree.parse(filepath) except etree.XMLSyntaxError as er: raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er)) try: transform = etree.XSLT(xslt_etree) except etree.XSLTParseError as er: raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er))
from django.conf import settings - from os import path + from os import path, listdir ? +++++++++ + from lxml import etree class XmlJsonImportModuleException(Exception): pass if not hasattr(settings, 'XSLT_FILES_DIR'): raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter') if not path.exists(settings.XSLT_FILES_DIR): raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist') + + for filename in listdir(settings.XSLT_FILES_DIR): + filepath = path.join(settings.XSLT_FILES_DIR, filename) + if path.isfile(filepath): + try: + xslt_etree = etree.parse(filepath) + except etree.XMLSyntaxError as er: + raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er)) + try: + transform = etree.XSLT(xslt_etree) + except etree.XSLTParseError as er: + raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er))
849fbdf724528df99f2ac53d389274f7c2631f11
invitation/admin.py
invitation/admin.py
from django.contrib import admin from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode class InvitationKeyAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired') class InvitationUserAdmin(admin.ModelAdmin): list_display = ('inviter', 'invitations_remaining') def invite_user(modeladmin, request, queryset): for invitation_request in queryset.all(): invitation = InvitationKey.objects.create_invitation(request.user) invitation.send_to(invitation_request.email) invitation_request.invited = True invitation_request.save() invite_user.short_description = "Invite this user" class InvitationRequestAdmin(admin.ModelAdmin): list_display = ('email', 'invited') actions = [invite_user] admin.site.register(InvitationKey, InvitationKeyAdmin) admin.site.register(InvitationUser, InvitationUserAdmin) admin.site.register(InvitationRequest, InvitationRequestAdmin) admin.site.register(InvitationCode)
from django.contrib import admin from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode class InvitationKeyAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired') class InvitationUserAdmin(admin.ModelAdmin): list_display = ('inviter', 'invitations_remaining') def invite_user(modeladmin, request, queryset): for invitation_request in queryset.all(): invitation = InvitationKey.objects.create_invitation(request.user) invitation.send_to(invitation_request.email) invitation_request.invited = True invitation_request.save() invite_user.short_description = "Invite selected invitation requests" class InvitationRequestAdmin(admin.ModelAdmin): list_display = ('email', 'invited') actions = [invite_user] admin.site.register(InvitationKey, InvitationKeyAdmin) admin.site.register(InvitationUser, InvitationUserAdmin) admin.site.register(InvitationRequest, InvitationRequestAdmin) admin.site.register(InvitationCode)
Improve the invite_user action name.
Improve the invite_user action name.
Python
bsd-3-clause
adieu/django-invitation
from django.contrib import admin from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode class InvitationKeyAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired') class InvitationUserAdmin(admin.ModelAdmin): list_display = ('inviter', 'invitations_remaining') def invite_user(modeladmin, request, queryset): for invitation_request in queryset.all(): invitation = InvitationKey.objects.create_invitation(request.user) invitation.send_to(invitation_request.email) invitation_request.invited = True invitation_request.save() - invite_user.short_description = "Invite this user" + invite_user.short_description = "Invite selected invitation requests" class InvitationRequestAdmin(admin.ModelAdmin): list_display = ('email', 'invited') actions = [invite_user] admin.site.register(InvitationKey, InvitationKeyAdmin) admin.site.register(InvitationUser, InvitationUserAdmin) admin.site.register(InvitationRequest, InvitationRequestAdmin) admin.site.register(InvitationCode)
Improve the invite_user action name.
## Code Before: from django.contrib import admin from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode class InvitationKeyAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired') class InvitationUserAdmin(admin.ModelAdmin): list_display = ('inviter', 'invitations_remaining') def invite_user(modeladmin, request, queryset): for invitation_request in queryset.all(): invitation = InvitationKey.objects.create_invitation(request.user) invitation.send_to(invitation_request.email) invitation_request.invited = True invitation_request.save() invite_user.short_description = "Invite this user" class InvitationRequestAdmin(admin.ModelAdmin): list_display = ('email', 'invited') actions = [invite_user] admin.site.register(InvitationKey, InvitationKeyAdmin) admin.site.register(InvitationUser, InvitationUserAdmin) admin.site.register(InvitationRequest, InvitationRequestAdmin) admin.site.register(InvitationCode) ## Instruction: Improve the invite_user action name. ## Code After: from django.contrib import admin from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode class InvitationKeyAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired') class InvitationUserAdmin(admin.ModelAdmin): list_display = ('inviter', 'invitations_remaining') def invite_user(modeladmin, request, queryset): for invitation_request in queryset.all(): invitation = InvitationKey.objects.create_invitation(request.user) invitation.send_to(invitation_request.email) invitation_request.invited = True invitation_request.save() invite_user.short_description = "Invite selected invitation requests" class InvitationRequestAdmin(admin.ModelAdmin): list_display = ('email', 'invited') actions = [invite_user] admin.site.register(InvitationKey, InvitationKeyAdmin) admin.site.register(InvitationUser, InvitationUserAdmin) admin.site.register(InvitationRequest, InvitationRequestAdmin) admin.site.register(InvitationCode)
from django.contrib import admin from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode class InvitationKeyAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired') class InvitationUserAdmin(admin.ModelAdmin): list_display = ('inviter', 'invitations_remaining') def invite_user(modeladmin, request, queryset): for invitation_request in queryset.all(): invitation = InvitationKey.objects.create_invitation(request.user) invitation.send_to(invitation_request.email) invitation_request.invited = True invitation_request.save() - invite_user.short_description = "Invite this user" + invite_user.short_description = "Invite selected invitation requests" class InvitationRequestAdmin(admin.ModelAdmin): list_display = ('email', 'invited') actions = [invite_user] admin.site.register(InvitationKey, InvitationKeyAdmin) admin.site.register(InvitationUser, InvitationUserAdmin) admin.site.register(InvitationRequest, InvitationRequestAdmin) admin.site.register(InvitationCode)
2261b3c6cb579ae65c1119db45f291e246f536c2
examples/main.py
examples/main.py
import asyncio import sys from contextlib import suppress sys.path.append("..") from asynccmd import Cmd class Commander(Cmd): def __init__(self, intro, prompt): if sys.platform == 'win32': super().__init__(mode="Run", run_loop=False) else: super().__init__(mode="Reader", run_loop=False) self.intro = intro self.prompt = prompt self.loop = None def do_tasks(self, arg): """ Fake command. Type "prodigy {arg}" :param arg: args occurred from cmd after command :return: """ print(print(asyncio.Task.all_tasks(loop=self.loop))) def start(self, loop=None): self.loop = loop super().cmdloop(loop) loop = asyncio.ProactorEventLoop() #loop = asyncio.get_event_loop() cmd = Commander(intro="This is example", prompt="example> ") cmd.start(loop) try: loop.run_forever() except KeyboardInterrupt: loop.stop() pending = asyncio.Task.all_tasks(loop=loop) for task in pending: task.cancel() with suppress(asyncio.CancelledError): loop.run_until_complete(task)
import asyncio import sys from contextlib import suppress sys.path.append("..") from asynccmd import Cmd class Commander(Cmd): def __init__(self, intro, prompt): if sys.platform == 'win32': super().__init__(mode="Run", run_loop=False) else: super().__init__(mode="Reader", run_loop=False) self.intro = intro self.prompt = prompt self.loop = None def do_tasks(self, arg): """ Fake command. Type "prodigy {arg}" :param arg: args occurred from cmd after command :return: """ print(print(asyncio.Task.all_tasks(loop=self.loop))) def start(self, loop=None): self.loop = loop super().cmdloop(loop) if sys.platform == 'win32': loop = asyncio.ProactorEventLoop() else: loop = asyncio.get_event_loop() cmd = Commander(intro="This is example", prompt="example> ") cmd.start(loop) try: loop.run_forever() except KeyboardInterrupt: loop.stop() pending = asyncio.Task.all_tasks(loop=loop) for task in pending: task.cancel() with suppress(asyncio.CancelledError): loop.run_until_complete(task)
FIX example for both Win and NIX
FIX example for both Win and NIX TODO: tasks wont work
Python
apache-2.0
valentinmk/asynccmd
import asyncio import sys from contextlib import suppress sys.path.append("..") from asynccmd import Cmd class Commander(Cmd): def __init__(self, intro, prompt): if sys.platform == 'win32': super().__init__(mode="Run", run_loop=False) else: super().__init__(mode="Reader", run_loop=False) self.intro = intro self.prompt = prompt self.loop = None def do_tasks(self, arg): """ Fake command. Type "prodigy {arg}" :param arg: args occurred from cmd after command :return: """ print(print(asyncio.Task.all_tasks(loop=self.loop))) def start(self, loop=None): self.loop = loop super().cmdloop(loop) + if sys.platform == 'win32': - loop = asyncio.ProactorEventLoop() + loop = asyncio.ProactorEventLoop() + else: - #loop = asyncio.get_event_loop() + loop = asyncio.get_event_loop() cmd = Commander(intro="This is example", prompt="example> ") cmd.start(loop) try: loop.run_forever() except KeyboardInterrupt: loop.stop() pending = asyncio.Task.all_tasks(loop=loop) for task in pending: task.cancel() with suppress(asyncio.CancelledError): loop.run_until_complete(task)
FIX example for both Win and NIX
## Code Before: import asyncio import sys from contextlib import suppress sys.path.append("..") from asynccmd import Cmd class Commander(Cmd): def __init__(self, intro, prompt): if sys.platform == 'win32': super().__init__(mode="Run", run_loop=False) else: super().__init__(mode="Reader", run_loop=False) self.intro = intro self.prompt = prompt self.loop = None def do_tasks(self, arg): """ Fake command. Type "prodigy {arg}" :param arg: args occurred from cmd after command :return: """ print(print(asyncio.Task.all_tasks(loop=self.loop))) def start(self, loop=None): self.loop = loop super().cmdloop(loop) loop = asyncio.ProactorEventLoop() #loop = asyncio.get_event_loop() cmd = Commander(intro="This is example", prompt="example> ") cmd.start(loop) try: loop.run_forever() except KeyboardInterrupt: loop.stop() pending = asyncio.Task.all_tasks(loop=loop) for task in pending: task.cancel() with suppress(asyncio.CancelledError): loop.run_until_complete(task) ## Instruction: FIX example for both Win and NIX ## Code After: import asyncio import sys from contextlib import suppress sys.path.append("..") from asynccmd import Cmd class Commander(Cmd): def __init__(self, intro, prompt): if sys.platform == 'win32': super().__init__(mode="Run", run_loop=False) else: super().__init__(mode="Reader", run_loop=False) self.intro = intro self.prompt = prompt self.loop = None def do_tasks(self, arg): """ Fake command. Type "prodigy {arg}" :param arg: args occurred from cmd after command :return: """ print(print(asyncio.Task.all_tasks(loop=self.loop))) def start(self, loop=None): self.loop = loop super().cmdloop(loop) if sys.platform == 'win32': loop = asyncio.ProactorEventLoop() else: loop = asyncio.get_event_loop() cmd = Commander(intro="This is example", prompt="example> ") cmd.start(loop) try: loop.run_forever() except KeyboardInterrupt: loop.stop() pending = asyncio.Task.all_tasks(loop=loop) for task in pending: task.cancel() with suppress(asyncio.CancelledError): loop.run_until_complete(task)
import asyncio import sys from contextlib import suppress sys.path.append("..") from asynccmd import Cmd class Commander(Cmd): def __init__(self, intro, prompt): if sys.platform == 'win32': super().__init__(mode="Run", run_loop=False) else: super().__init__(mode="Reader", run_loop=False) self.intro = intro self.prompt = prompt self.loop = None def do_tasks(self, arg): """ Fake command. Type "prodigy {arg}" :param arg: args occurred from cmd after command :return: """ print(print(asyncio.Task.all_tasks(loop=self.loop))) def start(self, loop=None): self.loop = loop super().cmdloop(loop) + if sys.platform == 'win32': - loop = asyncio.ProactorEventLoop() + loop = asyncio.ProactorEventLoop() ? +++ + else: - #loop = asyncio.get_event_loop() ? ^ + loop = asyncio.get_event_loop() ? ^^^ cmd = Commander(intro="This is example", prompt="example> ") cmd.start(loop) try: loop.run_forever() except KeyboardInterrupt: loop.stop() pending = asyncio.Task.all_tasks(loop=loop) for task in pending: task.cancel() with suppress(asyncio.CancelledError): loop.run_until_complete(task)
85f8d0662901047115f2d852489a3a5be1a01226
datafilters/views.py
datafilters/views.py
try: from django.views.generic.base import ContextMixin as mixin_base except ImportError: mixin_base = object __all__ = ('FilterFormMixin',) class FilterFormMixin(mixin_base): """ Mixin that adds filtering behaviour for Class Based Views. Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and get_context_data(self, **kwargs) method. """ filter_form_cls = None use_filter_chaining = False def get_filter(self): return self.filter_form_cls(self.request.GET, runtime_context=self.get_runtime_context(), use_filter_chaining=self.use_filter_chaining) def get_queryset(self): qs = super(FilterFormMixin, self).get_queryset() qs = self.get_filter().filter(qs).distinct() return qs def get_context_data(self, **kwargs): context = super(FilterFormMixin, self).get_context_data(**kwargs) context['filterform'] = self.get_filter() return context def get_runtime_context(self): return {'user': self.request.user}
from django.views.generic.list import MultipleObjectMixin __all__ = ('FilterFormMixin',) class FilterFormMixin(MultipleObjectMixin): """ Mixin that adds filtering behaviour for Class Based Views. Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and get_context_data(self, **kwargs) method. """ filter_form_cls = None use_filter_chaining = False def get_filter(self): return self.filter_form_cls(self.request.GET, runtime_context=self.get_runtime_context(), use_filter_chaining=self.use_filter_chaining) def get_queryset(self): qs = super(FilterFormMixin, self).get_queryset() qs = self.get_filter().filter(qs).distinct() return qs def get_context_data(self, **kwargs): context = super(FilterFormMixin, self).get_context_data(**kwargs) context['filterform'] = self.get_filter() return context def get_runtime_context(self): return {'user': self.request.user}
Set base class for view mixin to MultipleObjectMixin
Set base class for view mixin to MultipleObjectMixin
Python
mit
freevoid/django-datafilters,zorainc/django-datafilters,zorainc/django-datafilters
+ from django.views.generic.list import MultipleObjectMixin - try: - from django.views.generic.base import ContextMixin as mixin_base - except ImportError: - mixin_base = object __all__ = ('FilterFormMixin',) - class FilterFormMixin(mixin_base): + class FilterFormMixin(MultipleObjectMixin): """ Mixin that adds filtering behaviour for Class Based Views. Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and get_context_data(self, **kwargs) method. """ filter_form_cls = None use_filter_chaining = False def get_filter(self): return self.filter_form_cls(self.request.GET, runtime_context=self.get_runtime_context(), use_filter_chaining=self.use_filter_chaining) def get_queryset(self): qs = super(FilterFormMixin, self).get_queryset() qs = self.get_filter().filter(qs).distinct() return qs def get_context_data(self, **kwargs): context = super(FilterFormMixin, self).get_context_data(**kwargs) context['filterform'] = self.get_filter() return context def get_runtime_context(self): return {'user': self.request.user}
Set base class for view mixin to MultipleObjectMixin
## Code Before: try: from django.views.generic.base import ContextMixin as mixin_base except ImportError: mixin_base = object __all__ = ('FilterFormMixin',) class FilterFormMixin(mixin_base): """ Mixin that adds filtering behaviour for Class Based Views. Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and get_context_data(self, **kwargs) method. """ filter_form_cls = None use_filter_chaining = False def get_filter(self): return self.filter_form_cls(self.request.GET, runtime_context=self.get_runtime_context(), use_filter_chaining=self.use_filter_chaining) def get_queryset(self): qs = super(FilterFormMixin, self).get_queryset() qs = self.get_filter().filter(qs).distinct() return qs def get_context_data(self, **kwargs): context = super(FilterFormMixin, self).get_context_data(**kwargs) context['filterform'] = self.get_filter() return context def get_runtime_context(self): return {'user': self.request.user} ## Instruction: Set base class for view mixin to MultipleObjectMixin ## Code After: from django.views.generic.list import MultipleObjectMixin __all__ = ('FilterFormMixin',) class FilterFormMixin(MultipleObjectMixin): """ Mixin that adds filtering behaviour for Class Based Views. Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and get_context_data(self, **kwargs) method. """ filter_form_cls = None use_filter_chaining = False def get_filter(self): return self.filter_form_cls(self.request.GET, runtime_context=self.get_runtime_context(), use_filter_chaining=self.use_filter_chaining) def get_queryset(self): qs = super(FilterFormMixin, self).get_queryset() qs = self.get_filter().filter(qs).distinct() return qs def get_context_data(self, **kwargs): context = super(FilterFormMixin, self).get_context_data(**kwargs) context['filterform'] = self.get_filter() return context def get_runtime_context(self): return {'user': self.request.user}
+ from django.views.generic.list import MultipleObjectMixin - try: - from django.views.generic.base import ContextMixin as mixin_base - except ImportError: - mixin_base = object __all__ = ('FilterFormMixin',) - class FilterFormMixin(mixin_base): + class FilterFormMixin(MultipleObjectMixin): """ Mixin that adds filtering behaviour for Class Based Views. Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and get_context_data(self, **kwargs) method. """ filter_form_cls = None use_filter_chaining = False def get_filter(self): return self.filter_form_cls(self.request.GET, runtime_context=self.get_runtime_context(), use_filter_chaining=self.use_filter_chaining) def get_queryset(self): qs = super(FilterFormMixin, self).get_queryset() qs = self.get_filter().filter(qs).distinct() return qs def get_context_data(self, **kwargs): context = super(FilterFormMixin, self).get_context_data(**kwargs) context['filterform'] = self.get_filter() return context def get_runtime_context(self): return {'user': self.request.user}
937a5e32c77ca57917d60a891616fbcf19ab19f9
respite/utils.py
respite/utils.py
from django import forms def generate_form(model): """ Generate a form from a model. Arguments: model -- A Django model. """ _model = model class Form(forms.ModelForm): class Meta: model = _model return Form def parse_http_accept_header(header): """ Return a list of content types listed in the HTTP Accept header ordered by quality. Arguments: header -- A string describing the contents of the HTTP Accept header. """ components = header.split(',') l = [] for component in components: if ';' in component: subcomponents = component.split(';') l.append( ( subcomponents[0], # eg. 'text/html' subcomponents[1][2:] # eg. 'q=0.9' ) ) else: l.append((component, '1')) l.sort( key = lambda i: i[1], reverse = True ) content_types = [] for i in l: content_types.append(i[0]) return content_types
from django import forms def generate_form(model): """ Generate a form from a model. Arguments: model -- A Django model. """ _model = model class Form(forms.ModelForm): class Meta: model = _model return Form def parse_http_accept_header(header): """ Return a list of content types listed in the HTTP Accept header ordered by quality. Arguments: header -- A string describing the contents of the HTTP Accept header. """ components = [item.strip() for item in header.split(',')] l = [] for component in components: if ';' in component: subcomponents = [item.strip() for item in component.split(';')] l.append( ( subcomponents[0], # eg. 'text/html' subcomponents[1][2:] # eg. 'q=0.9' ) ) else: l.append((component, '1')) l.sort( key = lambda i: i[1], reverse = True ) content_types = [] for i in l: content_types.append(i[0]) return content_types
Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly
Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly
Python
mit
jgorset/django-respite,jgorset/django-respite,jgorset/django-respite
from django import forms def generate_form(model): """ Generate a form from a model. Arguments: model -- A Django model. """ _model = model class Form(forms.ModelForm): class Meta: model = _model return Form def parse_http_accept_header(header): """ Return a list of content types listed in the HTTP Accept header ordered by quality. Arguments: header -- A string describing the contents of the HTTP Accept header. """ - components = header.split(',') + components = [item.strip() for item in header.split(',')] l = [] for component in components: if ';' in component: - subcomponents = component.split(';') + subcomponents = [item.strip() for item in component.split(';')] l.append( ( subcomponents[0], # eg. 'text/html' subcomponents[1][2:] # eg. 'q=0.9' ) ) else: l.append((component, '1')) l.sort( key = lambda i: i[1], reverse = True ) content_types = [] for i in l: content_types.append(i[0]) return content_types
Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly
## Code Before: from django import forms def generate_form(model): """ Generate a form from a model. Arguments: model -- A Django model. """ _model = model class Form(forms.ModelForm): class Meta: model = _model return Form def parse_http_accept_header(header): """ Return a list of content types listed in the HTTP Accept header ordered by quality. Arguments: header -- A string describing the contents of the HTTP Accept header. """ components = header.split(',') l = [] for component in components: if ';' in component: subcomponents = component.split(';') l.append( ( subcomponents[0], # eg. 'text/html' subcomponents[1][2:] # eg. 'q=0.9' ) ) else: l.append((component, '1')) l.sort( key = lambda i: i[1], reverse = True ) content_types = [] for i in l: content_types.append(i[0]) return content_types ## Instruction: Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly ## Code After: from django import forms def generate_form(model): """ Generate a form from a model. Arguments: model -- A Django model. """ _model = model class Form(forms.ModelForm): class Meta: model = _model return Form def parse_http_accept_header(header): """ Return a list of content types listed in the HTTP Accept header ordered by quality. Arguments: header -- A string describing the contents of the HTTP Accept header. """ components = [item.strip() for item in header.split(',')] l = [] for component in components: if ';' in component: subcomponents = [item.strip() for item in component.split(';')] l.append( ( subcomponents[0], # eg. 'text/html' subcomponents[1][2:] # eg. 'q=0.9' ) ) else: l.append((component, '1')) l.sort( key = lambda i: i[1], reverse = True ) content_types = [] for i in l: content_types.append(i[0]) return content_types
from django import forms def generate_form(model): """ Generate a form from a model. Arguments: model -- A Django model. """ _model = model class Form(forms.ModelForm): class Meta: model = _model return Form def parse_http_accept_header(header): """ Return a list of content types listed in the HTTP Accept header ordered by quality. Arguments: header -- A string describing the contents of the HTTP Accept header. """ - components = header.split(',') + components = [item.strip() for item in header.split(',')] l = [] for component in components: if ';' in component: - subcomponents = component.split(';') + subcomponents = [item.strip() for item in component.split(';')] ? ++++++++++++++++++++++++++ + l.append( ( subcomponents[0], # eg. 'text/html' subcomponents[1][2:] # eg. 'q=0.9' ) ) else: l.append((component, '1')) l.sort( key = lambda i: i[1], reverse = True ) content_types = [] for i in l: content_types.append(i[0]) return content_types
edc5564d4c3677dc8b545e9c9a6a51b481247eab
contentcuration/contentcuration/tests/test_makemessages.py
contentcuration/contentcuration/tests/test_makemessages.py
import os import subprocess import pathlib from django.conf import settings from django.test import TestCase class MakeMessagesCommandRunTestCase(TestCase): """ Sanity check to make sure makemessages runs to completion. """ def test_command_succeeds_without_postgres(self): """ Test that we can run makemessages when postgres is not activated. """ # this test can make changes to committed files, so only run it # on the CI server if 'CI' not in os.environ or not os.environ['CI']: return repo_root = pathlib.Path(settings.BASE_DIR).parent cmd = ["make", "makemessages"] env = os.environ.copy() # We fake postgres not being available, by setting the wrong IP address. # hopefully postgres isn't running at 127.0.0.2! env.update({"DATA_DB_HOST": "127.0.0.2"}) subprocess.check_output( cmd, env=env, cwd=str(repo_root) )
import os import subprocess import pathlib import pytest from django.conf import settings from django.test import TestCase class MakeMessagesCommandRunTestCase(TestCase): """ Sanity check to make sure makemessages runs to completion. """ # this test can make changes to committed files, so only run it # on the CI server @pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server") def test_command_succeeds_without_postgres(self): """ Test that we can run makemessages when postgres is not activated. """ repo_root = pathlib.Path(settings.BASE_DIR).parent cmd = ["make", "makemessages"] env = os.environ.copy() # We fake postgres not being available, by setting the wrong IP address. # hopefully postgres isn't running at 127.0.0.2! env.update({"DATA_DB_HOST": "127.0.0.2"}) subprocess.check_output( cmd, env=env, cwd=str(repo_root) )
Use pytest.skip so we can check the test wasn't skipped on the CI.
Use pytest.skip so we can check the test wasn't skipped on the CI.
Python
mit
DXCanas/content-curation,DXCanas/content-curation,DXCanas/content-curation,DXCanas/content-curation
import os import subprocess import pathlib + import pytest from django.conf import settings from django.test import TestCase class MakeMessagesCommandRunTestCase(TestCase): """ Sanity check to make sure makemessages runs to completion. """ + # this test can make changes to committed files, so only run it + # on the CI server + @pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server") def test_command_succeeds_without_postgres(self): """ Test that we can run makemessages when postgres is not activated. """ - - # this test can make changes to committed files, so only run it - # on the CI server - if 'CI' not in os.environ or not os.environ['CI']: - return repo_root = pathlib.Path(settings.BASE_DIR).parent cmd = ["make", "makemessages"] env = os.environ.copy() # We fake postgres not being available, by setting the wrong IP address. # hopefully postgres isn't running at 127.0.0.2! env.update({"DATA_DB_HOST": "127.0.0.2"}) subprocess.check_output( cmd, env=env, cwd=str(repo_root) )
Use pytest.skip so we can check the test wasn't skipped on the CI.
## Code Before: import os import subprocess import pathlib from django.conf import settings from django.test import TestCase class MakeMessagesCommandRunTestCase(TestCase): """ Sanity check to make sure makemessages runs to completion. """ def test_command_succeeds_without_postgres(self): """ Test that we can run makemessages when postgres is not activated. """ # this test can make changes to committed files, so only run it # on the CI server if 'CI' not in os.environ or not os.environ['CI']: return repo_root = pathlib.Path(settings.BASE_DIR).parent cmd = ["make", "makemessages"] env = os.environ.copy() # We fake postgres not being available, by setting the wrong IP address. # hopefully postgres isn't running at 127.0.0.2! env.update({"DATA_DB_HOST": "127.0.0.2"}) subprocess.check_output( cmd, env=env, cwd=str(repo_root) ) ## Instruction: Use pytest.skip so we can check the test wasn't skipped on the CI. ## Code After: import os import subprocess import pathlib import pytest from django.conf import settings from django.test import TestCase class MakeMessagesCommandRunTestCase(TestCase): """ Sanity check to make sure makemessages runs to completion. """ # this test can make changes to committed files, so only run it # on the CI server @pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server") def test_command_succeeds_without_postgres(self): """ Test that we can run makemessages when postgres is not activated. """ repo_root = pathlib.Path(settings.BASE_DIR).parent cmd = ["make", "makemessages"] env = os.environ.copy() # We fake postgres not being available, by setting the wrong IP address. # hopefully postgres isn't running at 127.0.0.2! env.update({"DATA_DB_HOST": "127.0.0.2"}) subprocess.check_output( cmd, env=env, cwd=str(repo_root) )
import os import subprocess import pathlib + import pytest from django.conf import settings from django.test import TestCase class MakeMessagesCommandRunTestCase(TestCase): """ Sanity check to make sure makemessages runs to completion. """ + # this test can make changes to committed files, so only run it + # on the CI server + @pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server") def test_command_succeeds_without_postgres(self): """ Test that we can run makemessages when postgres is not activated. """ - - # this test can make changes to committed files, so only run it - # on the CI server - if 'CI' not in os.environ or not os.environ['CI']: - return repo_root = pathlib.Path(settings.BASE_DIR).parent cmd = ["make", "makemessages"] env = os.environ.copy() # We fake postgres not being available, by setting the wrong IP address. # hopefully postgres isn't running at 127.0.0.2! env.update({"DATA_DB_HOST": "127.0.0.2"}) subprocess.check_output( cmd, env=env, cwd=str(repo_root) )
dafde564f3ea18655b1e15f410df70d05b3eb8f5
beets/util/collections.py
beets/util/collections.py
class IdentityFallbackDict(dict): """A dictionary which is "transparent" (maps keys to themselves) for all keys not in it. """ def __getitem__(self, key): try: return dict.__getitem__(self, key) except KeyError: return key
from __future__ import division, absolute_import, print_function class IdentityFallbackDict(dict): """A dictionary which is "transparent" (maps keys to themselves) for all keys not in it. """ def __getitem__(self, key): try: return dict.__getitem__(self, key) except KeyError: return key
Add __future__ imports to a new module
Add __future__ imports to a new module
Python
mit
mosesfistos1/beetbox,ibmibmibm/beets,mosesfistos1/beetbox,MyTunesFreeMusic/privacy-policy,artemutin/beets,jackwilsdon/beets,sampsyo/beets,pkess/beets,xsteadfastx/beets,shamangeorge/beets,diego-plan9/beets,MyTunesFreeMusic/privacy-policy,jackwilsdon/beets,beetbox/beets,sampsyo/beets,beetbox/beets,madmouser1/beets,beetbox/beets,madmouser1/beets,pkess/beets,shamangeorge/beets,diego-plan9/beets,MyTunesFreeMusic/privacy-policy,xsteadfastx/beets,Kraymer/beets,MyTunesFreeMusic/privacy-policy,SusannaMaria/beets,jackwilsdon/beets,lengtche/beets,Kraymer/beets,lengtche/beets,jackwilsdon/beets,artemutin/beets,madmouser1/beets,sampsyo/beets,ibmibmibm/beets,pkess/beets,Kraymer/beets,mosesfistos1/beetbox,shamangeorge/beets,Kraymer/beets,lengtche/beets,diego-plan9/beets,ibmibmibm/beets,SusannaMaria/beets,beetbox/beets,xsteadfastx/beets,madmouser1/beets,ibmibmibm/beets,artemutin/beets,pkess/beets,xsteadfastx/beets,artemutin/beets,SusannaMaria/beets,SusannaMaria/beets,lengtche/beets,shamangeorge/beets,mosesfistos1/beetbox,sampsyo/beets,diego-plan9/beets
+ + from __future__ import division, absolute_import, print_function class IdentityFallbackDict(dict): """A dictionary which is "transparent" (maps keys to themselves) for all keys not in it. """ def __getitem__(self, key): try: return dict.__getitem__(self, key) except KeyError: return key
Add __future__ imports to a new module
## Code Before: class IdentityFallbackDict(dict): """A dictionary which is "transparent" (maps keys to themselves) for all keys not in it. """ def __getitem__(self, key): try: return dict.__getitem__(self, key) except KeyError: return key ## Instruction: Add __future__ imports to a new module ## Code After: from __future__ import division, absolute_import, print_function class IdentityFallbackDict(dict): """A dictionary which is "transparent" (maps keys to themselves) for all keys not in it. """ def __getitem__(self, key): try: return dict.__getitem__(self, key) except KeyError: return key
+ + from __future__ import division, absolute_import, print_function class IdentityFallbackDict(dict): """A dictionary which is "transparent" (maps keys to themselves) for all keys not in it. """ def __getitem__(self, key): try: return dict.__getitem__(self, key) except KeyError: return key
8a821cb62a35547417fcd56d02486e5cc2d8494f
xzarr.py
xzarr.py
from .base import DataSourceMixin class ZarrSource(DataSourceMixin): """Open a xarray dataset. Parameters ---------- urlpath: str Path to source. This can be a local directory or a remote data service (i.e., with a protocol specifier like ``'s3://``). storage_options: dict Parameters passed to the backend file-system kwargs: Further parameters are passed to xr.open_zarr """ name = 'zarr' def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs): super(ZarrSource, self).__init__(metadata=metadata) self.urlpath = urlpath self.storage_options = storage_options self.kwargs = kwargs self._ds = None def _open_dataset(self): import xarray as xr from dask.bytes.core import get_fs, infer_options, \ update_storage_options urlpath, protocol, options = infer_options(self.urlpath) update_storage_options(options, self.storage_options) self._fs, _ = get_fs(protocol, options) if protocol != 'file': self._mapper = get_mapper(protocol, self._fs, urlpath) self._ds = xr.open_zarr(self._mapper, **self.kwargs) else: self._ds = xr.open_zarr(self.urlpath, **self.kwargs) def close(self): super(ZarrSource, self).close() self._fs = None self._mapper = None def get_mapper(protocol, fs, path): if protocol == 's3': from s3fs.mapping import S3Map return S3Map(path, fs) elif protocol == 'gcs': from gcsfs.mapping import GCSMap return GCSMap(path, fs) else: raise NotImplementedError
from .base import DataSourceMixin class ZarrSource(DataSourceMixin): """Open a xarray dataset. Parameters ---------- urlpath: str Path to source. This can be a local directory or a remote data service (i.e., with a protocol specifier like ``'s3://``). storage_options: dict Parameters passed to the backend file-system kwargs: Further parameters are passed to xr.open_zarr """ name = 'zarr' def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs): super(ZarrSource, self).__init__(metadata=metadata) self.urlpath = urlpath self.storage_options = storage_options or {} self.kwargs = kwargs self._ds = None def _open_dataset(self): import xarray as xr from fsspec import get_mapper self._mapper = get_mapper(self.urlpath, **self.storage_options) self._ds = xr.open_zarr(self._mapper, **self.kwargs) def close(self): super(ZarrSource, self).close() self._fs = None self._mapper = None
Make work with any filesystem
Make work with any filesystem
Python
bsd-3-clause
ericdill/databroker,ericdill/databroker
from .base import DataSourceMixin class ZarrSource(DataSourceMixin): """Open a xarray dataset. Parameters ---------- urlpath: str Path to source. This can be a local directory or a remote data service (i.e., with a protocol specifier like ``'s3://``). storage_options: dict Parameters passed to the backend file-system kwargs: Further parameters are passed to xr.open_zarr """ name = 'zarr' def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs): super(ZarrSource, self).__init__(metadata=metadata) self.urlpath = urlpath - self.storage_options = storage_options + self.storage_options = storage_options or {} self.kwargs = kwargs self._ds = None def _open_dataset(self): import xarray as xr + from fsspec import get_mapper - from dask.bytes.core import get_fs, infer_options, \ - update_storage_options - urlpath, protocol, options = infer_options(self.urlpath) - update_storage_options(options, self.storage_options) + self._mapper = get_mapper(self.urlpath, **self.storage_options) - self._fs, _ = get_fs(protocol, options) - if protocol != 'file': - self._mapper = get_mapper(protocol, self._fs, urlpath) - self._ds = xr.open_zarr(self._mapper, **self.kwargs) + self._ds = xr.open_zarr(self._mapper, **self.kwargs) - else: - self._ds = xr.open_zarr(self.urlpath, **self.kwargs) def close(self): super(ZarrSource, self).close() self._fs = None self._mapper = None - - def get_mapper(protocol, fs, path): - if protocol == 's3': - from s3fs.mapping import S3Map - return S3Map(path, fs) - elif protocol == 'gcs': - from gcsfs.mapping import GCSMap - return GCSMap(path, fs) - else: - raise NotImplementedError -
Make work with any filesystem
## Code Before: from .base import DataSourceMixin class ZarrSource(DataSourceMixin): """Open a xarray dataset. Parameters ---------- urlpath: str Path to source. This can be a local directory or a remote data service (i.e., with a protocol specifier like ``'s3://``). storage_options: dict Parameters passed to the backend file-system kwargs: Further parameters are passed to xr.open_zarr """ name = 'zarr' def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs): super(ZarrSource, self).__init__(metadata=metadata) self.urlpath = urlpath self.storage_options = storage_options self.kwargs = kwargs self._ds = None def _open_dataset(self): import xarray as xr from dask.bytes.core import get_fs, infer_options, \ update_storage_options urlpath, protocol, options = infer_options(self.urlpath) update_storage_options(options, self.storage_options) self._fs, _ = get_fs(protocol, options) if protocol != 'file': self._mapper = get_mapper(protocol, self._fs, urlpath) self._ds = xr.open_zarr(self._mapper, **self.kwargs) else: self._ds = xr.open_zarr(self.urlpath, **self.kwargs) def close(self): super(ZarrSource, self).close() self._fs = None self._mapper = None def get_mapper(protocol, fs, path): if protocol == 's3': from s3fs.mapping import S3Map return S3Map(path, fs) elif protocol == 'gcs': from gcsfs.mapping import GCSMap return GCSMap(path, fs) else: raise NotImplementedError ## Instruction: Make work with any filesystem ## Code After: from .base import DataSourceMixin class ZarrSource(DataSourceMixin): """Open a xarray dataset. Parameters ---------- urlpath: str Path to source. This can be a local directory or a remote data service (i.e., with a protocol specifier like ``'s3://``). storage_options: dict Parameters passed to the backend file-system kwargs: Further parameters are passed to xr.open_zarr """ name = 'zarr' def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs): super(ZarrSource, self).__init__(metadata=metadata) self.urlpath = urlpath self.storage_options = storage_options or {} self.kwargs = kwargs self._ds = None def _open_dataset(self): import xarray as xr from fsspec import get_mapper self._mapper = get_mapper(self.urlpath, **self.storage_options) self._ds = xr.open_zarr(self._mapper, **self.kwargs) def close(self): super(ZarrSource, self).close() self._fs = None self._mapper = None
from .base import DataSourceMixin class ZarrSource(DataSourceMixin): """Open a xarray dataset. Parameters ---------- urlpath: str Path to source. This can be a local directory or a remote data service (i.e., with a protocol specifier like ``'s3://``). storage_options: dict Parameters passed to the backend file-system kwargs: Further parameters are passed to xr.open_zarr """ name = 'zarr' def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs): super(ZarrSource, self).__init__(metadata=metadata) self.urlpath = urlpath - self.storage_options = storage_options + self.storage_options = storage_options or {} ? ++++++ self.kwargs = kwargs self._ds = None def _open_dataset(self): import xarray as xr + from fsspec import get_mapper - from dask.bytes.core import get_fs, infer_options, \ - update_storage_options - urlpath, protocol, options = infer_options(self.urlpath) - update_storage_options(options, self.storage_options) + self._mapper = get_mapper(self.urlpath, **self.storage_options) - self._fs, _ = get_fs(protocol, options) - if protocol != 'file': - self._mapper = get_mapper(protocol, self._fs, urlpath) - self._ds = xr.open_zarr(self._mapper, **self.kwargs) ? ---- + self._ds = xr.open_zarr(self._mapper, **self.kwargs) - else: - self._ds = xr.open_zarr(self.urlpath, **self.kwargs) def close(self): super(ZarrSource, self).close() self._fs = None self._mapper = None - - - def get_mapper(protocol, fs, path): - if protocol == 's3': - from s3fs.mapping import S3Map - return S3Map(path, fs) - elif protocol == 'gcs': - from gcsfs.mapping import GCSMap - return GCSMap(path, fs) - else: - raise NotImplementedError
07fd8bf23917e18ba419859d788d9f51735f3b39
conda_gitenv/__init__.py
conda_gitenv/__init__.py
from __future__ import absolute_import, division, print_function, unicode_literals from distutils.version import StrictVersion from conda import __version__ as CONDA_VERSION from ._version import get_versions __version__ = get_versions()['version'] del get_versions _conda_base = StrictVersion('4.3.0') _conda_support = StrictVersion(CONDA_VERSION) >= _conda_base assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base) manifest_branch_prefix = 'manifest/'
from __future__ import absolute_import, division, print_function, unicode_literals from distutils.version import StrictVersion from conda import __version__ as CONDA_VERSION from ._version import get_versions __version__ = get_versions()['version'] del get_versions _conda_base = StrictVersion('4.3.0') _conda_version = StrictVersion(CONDA_VERSION) _conda_supported = _conda_version >= _conda_base assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version) manifest_branch_prefix = 'manifest/'
Update minimum conda version diagnostic
Update minimum conda version diagnostic
Python
bsd-3-clause
SciTools/conda-gitenv
from __future__ import absolute_import, division, print_function, unicode_literals from distutils.version import StrictVersion from conda import __version__ as CONDA_VERSION from ._version import get_versions __version__ = get_versions()['version'] del get_versions _conda_base = StrictVersion('4.3.0') - _conda_support = StrictVersion(CONDA_VERSION) >= _conda_base + _conda_version = StrictVersion(CONDA_VERSION) + _conda_supported = _conda_version >= _conda_base - assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base) + assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version) manifest_branch_prefix = 'manifest/'
Update minimum conda version diagnostic
## Code Before: from __future__ import absolute_import, division, print_function, unicode_literals from distutils.version import StrictVersion from conda import __version__ as CONDA_VERSION from ._version import get_versions __version__ = get_versions()['version'] del get_versions _conda_base = StrictVersion('4.3.0') _conda_support = StrictVersion(CONDA_VERSION) >= _conda_base assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base) manifest_branch_prefix = 'manifest/' ## Instruction: Update minimum conda version diagnostic ## Code After: from __future__ import absolute_import, division, print_function, unicode_literals from distutils.version import StrictVersion from conda import __version__ as CONDA_VERSION from ._version import get_versions __version__ = get_versions()['version'] del get_versions _conda_base = StrictVersion('4.3.0') _conda_version = StrictVersion(CONDA_VERSION) _conda_supported = _conda_version >= _conda_base assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version) manifest_branch_prefix = 'manifest/'
from __future__ import absolute_import, division, print_function, unicode_literals from distutils.version import StrictVersion from conda import __version__ as CONDA_VERSION from ._version import get_versions __version__ = get_versions()['version'] del get_versions _conda_base = StrictVersion('4.3.0') - _conda_support = StrictVersion(CONDA_VERSION) >= _conda_base ? ^^^ ^^ --------------- + _conda_version = StrictVersion(CONDA_VERSION) ? +++ ^ ^ + _conda_supported = _conda_version >= _conda_base - assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base) + assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version) ? ++++++++ ++++++++++++++++ manifest_branch_prefix = 'manifest/'
b5d812504924af2e2781f4be63a6191e5c47879d
test_project/urls.py
test_project/urls.py
from django.conf.urls import url, include from django.conf import settings from django.contrib import admin from django.views.generic import TemplateView TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html') urlpatterns = [ url('^accounts/', include('django.contrib.auth.urls')), url(r'^admin/', admin.site.urls), url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)), ]
from django.conf.urls import url, include from django.conf import settings from django.contrib import admin from django.views.generic import TemplateView urlpatterns = [ url('^accounts/', include('django.contrib.auth.urls')), url(r'^admin/', admin.site.urls), ] TEST_TEMPLATES = getattr( settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')]) for path, template in TEST_TEMPLATES: urlpatterns.append(url(path, TemplateView.as_view(template_name=template)))
Support multiple templates in TEST_TEMPLATES setting.
Support multiple templates in TEST_TEMPLATES setting. Unit tests need to be able to test redirects and other features involving multiple web pages. This commit changes the singleton TEST_TEMPLATE setting to TEST_TEMPLATES, which is a list of path, template tuples.
Python
bsd-3-clause
nimbis/django-selenium-testcase,nimbis/django-selenium-testcase
from django.conf.urls import url, include from django.conf import settings from django.contrib import admin from django.views.generic import TemplateView - TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html') - urlpatterns = [ url('^accounts/', include('django.contrib.auth.urls')), url(r'^admin/', admin.site.urls), - url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)), ] + TEST_TEMPLATES = getattr( + settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')]) + + for path, template in TEST_TEMPLATES: + urlpatterns.append(url(path, TemplateView.as_view(template_name=template))) +
Support multiple templates in TEST_TEMPLATES setting.
## Code Before: from django.conf.urls import url, include from django.conf import settings from django.contrib import admin from django.views.generic import TemplateView TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html') urlpatterns = [ url('^accounts/', include('django.contrib.auth.urls')), url(r'^admin/', admin.site.urls), url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)), ] ## Instruction: Support multiple templates in TEST_TEMPLATES setting. ## Code After: from django.conf.urls import url, include from django.conf import settings from django.contrib import admin from django.views.generic import TemplateView urlpatterns = [ url('^accounts/', include('django.contrib.auth.urls')), url(r'^admin/', admin.site.urls), ] TEST_TEMPLATES = getattr( settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')]) for path, template in TEST_TEMPLATES: urlpatterns.append(url(path, TemplateView.as_view(template_name=template)))
from django.conf.urls import url, include from django.conf import settings from django.contrib import admin from django.views.generic import TemplateView - TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html') - urlpatterns = [ url('^accounts/', include('django.contrib.auth.urls')), url(r'^admin/', admin.site.urls), - url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)), ] + + TEST_TEMPLATES = getattr( + settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')]) + + for path, template in TEST_TEMPLATES: + urlpatterns.append(url(path, TemplateView.as_view(template_name=template)))
836845abde53ee55bca93f098ece78880ab6b5c6
examples/events/create_massive_dummy_events.py
examples/events/create_massive_dummy_events.py
from pymisp import PyMISP from keys import misp_url, misp_key, misp_verifycert import argparse import tools def init(url, key): return PyMISP(url, key, misp_verifycert, 'json') if __name__ == '__main__': parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.') parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)") parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)") args = parser.parse_args() misp = init(misp_url, misp_key) if args.limit is None: args.limit = 1 if args.attribute is None: args.attribute = 3000 for i in range(args.limit): tools.create_massive_dummy_events(misp, args.attribute)
from pymisp import PyMISP from keys import url, key import argparse import tools if __name__ == '__main__': parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.') parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)") parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)") args = parser.parse_args() misp = PyMISP(url, key, True, 'json') if args.limit is None: args.limit = 1 if args.attribute is None: args.attribute = 3000 for i in range(args.limit): tools.create_massive_dummy_events(misp, args.attribute)
Use same variable names as testing environment
Use same variable names as testing environment
Python
bsd-2-clause
pombredanne/PyMISP,iglocska/PyMISP
from pymisp import PyMISP - from keys import misp_url, misp_key, misp_verifycert + from keys import url, key import argparse import tools - def init(url, key): - return PyMISP(url, key, misp_verifycert, 'json') if __name__ == '__main__': parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.') parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)") parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)") args = parser.parse_args() - misp = init(misp_url, misp_key) + misp = PyMISP(url, key, True, 'json') if args.limit is None: args.limit = 1 if args.attribute is None: args.attribute = 3000 for i in range(args.limit): tools.create_massive_dummy_events(misp, args.attribute)
Use same variable names as testing environment
## Code Before: from pymisp import PyMISP from keys import misp_url, misp_key, misp_verifycert import argparse import tools def init(url, key): return PyMISP(url, key, misp_verifycert, 'json') if __name__ == '__main__': parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.') parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)") parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)") args = parser.parse_args() misp = init(misp_url, misp_key) if args.limit is None: args.limit = 1 if args.attribute is None: args.attribute = 3000 for i in range(args.limit): tools.create_massive_dummy_events(misp, args.attribute) ## Instruction: Use same variable names as testing environment ## Code After: from pymisp import PyMISP from keys import url, key import argparse import tools if __name__ == '__main__': parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.') parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)") parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)") args = parser.parse_args() misp = PyMISP(url, key, True, 'json') if args.limit is None: args.limit = 1 if args.attribute is None: args.attribute = 3000 for i in range(args.limit): tools.create_massive_dummy_events(misp, args.attribute)
from pymisp import PyMISP - from keys import misp_url, misp_key, misp_verifycert + from keys import url, key import argparse import tools - def init(url, key): - return PyMISP(url, key, misp_verifycert, 'json') if __name__ == '__main__': parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.') parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)") parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)") args = parser.parse_args() - misp = init(misp_url, misp_key) + misp = PyMISP(url, key, True, 'json') if args.limit is None: args.limit = 1 if args.attribute is None: args.attribute = 3000 for i in range(args.limit): tools.create_massive_dummy_events(misp, args.attribute)
1cbab715a647689aeda4679d7dcf4e60ff9ab5b1
api/webview/models.py
api/webview/models.py
from django.db import models from django_pgjson.fields import JsonField class Document(models.Model): source = models.CharField(max_length=100) docID = models.CharField(max_length=100) providerUpdatedDateTime = models.DateTimeField(null=True) raw = JsonField() normalized = JsonField()
import json import six from requests.structures import CaseInsensitiveDict from django.db import models from django_pgjson.fields import JsonField class Document(models.Model): source = models.CharField(max_length=100) docID = models.CharField(max_length=100) providerUpdatedDateTime = models.DateTimeField(null=True) raw = JsonField() normalized = JsonField() class HarvesterResponse(models.Model): method = models.TextField(primary_key=True) url = models.TextField(primary_key=True, required=True) # Raw request data ok = models.BooleanField() content = models.BinaryField() encoding = models.TextField() headers_str = models.TextField() status_code = models.IntegerField() time_made = models.DateTimeField(auto_now=True) def json(self): return json.loads(self.content) @property def headers(self): return CaseInsensitiveDict(json.loads(self.headers_str)) @property def text(self): return six.u(self.content)
Add harvester response model in django ORM
Add harvester response model in django ORM
Python
apache-2.0
felliott/scrapi,fabianvf/scrapi,erinspace/scrapi,mehanig/scrapi,felliott/scrapi,erinspace/scrapi,CenterForOpenScience/scrapi,CenterForOpenScience/scrapi,mehanig/scrapi,fabianvf/scrapi
+ import json + + import six + from requests.structures import CaseInsensitiveDict + from django.db import models - from django_pgjson.fields import JsonField class Document(models.Model): source = models.CharField(max_length=100) docID = models.CharField(max_length=100) providerUpdatedDateTime = models.DateTimeField(null=True) raw = JsonField() normalized = JsonField() + + class HarvesterResponse(models.Model): + + method = models.TextField(primary_key=True) + url = models.TextField(primary_key=True, required=True) + + # Raw request data + ok = models.BooleanField() + content = models.BinaryField() + encoding = models.TextField() + headers_str = models.TextField() + status_code = models.IntegerField() + time_made = models.DateTimeField(auto_now=True) + + def json(self): + return json.loads(self.content) + + @property + def headers(self): + return CaseInsensitiveDict(json.loads(self.headers_str)) + + @property + def text(self): + return six.u(self.content) +
Add harvester response model in django ORM
## Code Before: from django.db import models from django_pgjson.fields import JsonField class Document(models.Model): source = models.CharField(max_length=100) docID = models.CharField(max_length=100) providerUpdatedDateTime = models.DateTimeField(null=True) raw = JsonField() normalized = JsonField() ## Instruction: Add harvester response model in django ORM ## Code After: import json import six from requests.structures import CaseInsensitiveDict from django.db import models from django_pgjson.fields import JsonField class Document(models.Model): source = models.CharField(max_length=100) docID = models.CharField(max_length=100) providerUpdatedDateTime = models.DateTimeField(null=True) raw = JsonField() normalized = JsonField() class HarvesterResponse(models.Model): method = models.TextField(primary_key=True) url = models.TextField(primary_key=True, required=True) # Raw request data ok = models.BooleanField() content = models.BinaryField() encoding = models.TextField() headers_str = models.TextField() status_code = models.IntegerField() time_made = models.DateTimeField(auto_now=True) def json(self): return json.loads(self.content) @property def headers(self): return CaseInsensitiveDict(json.loads(self.headers_str)) @property def text(self): return six.u(self.content)
+ import json + + import six + from requests.structures import CaseInsensitiveDict + from django.db import models - from django_pgjson.fields import JsonField class Document(models.Model): source = models.CharField(max_length=100) docID = models.CharField(max_length=100) providerUpdatedDateTime = models.DateTimeField(null=True) raw = JsonField() normalized = JsonField() + + + class HarvesterResponse(models.Model): + + method = models.TextField(primary_key=True) + url = models.TextField(primary_key=True, required=True) + + # Raw request data + ok = models.BooleanField() + content = models.BinaryField() + encoding = models.TextField() + headers_str = models.TextField() + status_code = models.IntegerField() + time_made = models.DateTimeField(auto_now=True) + + def json(self): + return json.loads(self.content) + + @property + def headers(self): + return CaseInsensitiveDict(json.loads(self.headers_str)) + + @property + def text(self): + return six.u(self.content)
11c30f5dd765475a9f5f0f847f31c47af8c40a39
user_agent/device.py
user_agent/device.py
import os.path import json PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__)) with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f: SMARTPHONE_DEV_IDS = json.load(open(f)) with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f: TABLET_DEV_IDS = json.load(open())
import os.path import json PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__)) with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f: SMARTPHONE_DEV_IDS = json.load(f) with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f: TABLET_DEV_IDS = json.load(f)
Fix uses of file objects
Fix uses of file objects
Python
mit
lorien/user_agent
import os.path import json PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__)) with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f: - SMARTPHONE_DEV_IDS = json.load(open(f)) + SMARTPHONE_DEV_IDS = json.load(f) with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f: - TABLET_DEV_IDS = json.load(open()) + TABLET_DEV_IDS = json.load(f)
Fix uses of file objects
## Code Before: import os.path import json PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__)) with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f: SMARTPHONE_DEV_IDS = json.load(open(f)) with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f: TABLET_DEV_IDS = json.load(open()) ## Instruction: Fix uses of file objects ## Code After: import os.path import json PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__)) with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f: SMARTPHONE_DEV_IDS = json.load(f) with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f: TABLET_DEV_IDS = json.load(f)
import os.path import json PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__)) with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f: - SMARTPHONE_DEV_IDS = json.load(open(f)) ? ----- - + SMARTPHONE_DEV_IDS = json.load(f) with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f: - TABLET_DEV_IDS = json.load(open()) ? ^^^^^ - + TABLET_DEV_IDS = json.load(f) ? ^
ea315b018fb3fab6925f1194fcd3e341166ab6fb
opt/resource/common.py
opt/resource/common.py
import json from urllib.parse import urljoin import sys def get_payload(): return json.load(sys.stdin) def get_index_url(payload): source = payload['source'] uri = source['uri'] index = source['index'] if not uri.endswith('/'): uri += '/' if not index.endswith('/'): index += '/' return urljoin(uri, index) def get_package_url(payload): package = payload['source']['package'] return get_index_url(payload) + package def get_auth(payload): source = payload['source'] return source['username'], source['password'] def get_version(payload): if 'version' in payload: version = payload['version']['version'] else: version = None return version
import json from urllib.parse import urljoin import sys def get_payload(): return json.load(sys.stdin) def get_index_url(payload): source = payload['source'] uri = source['uri'] index = source['index'] if not uri.endswith('/'): uri += '/' if not index.endswith('/'): index += '/' return urljoin(uri, index) def get_package_url(payload): package = payload['source']['package'] return get_index_url(payload) + package def get_auth(payload): source = payload['source'] return source['username'], source['password'] def get_version(payload): try: version = payload['version']['version'] except TypeError: version = None return version
Handle missing version in payload.
Handle missing version in payload.
Python
mit
mdomke/concourse-devpi-resource
import json from urllib.parse import urljoin import sys def get_payload(): return json.load(sys.stdin) def get_index_url(payload): source = payload['source'] uri = source['uri'] index = source['index'] if not uri.endswith('/'): uri += '/' if not index.endswith('/'): index += '/' return urljoin(uri, index) def get_package_url(payload): package = payload['source']['package'] return get_index_url(payload) + package def get_auth(payload): source = payload['source'] return source['username'], source['password'] def get_version(payload): - if 'version' in payload: + try: version = payload['version']['version'] - else: + except TypeError: version = None return version
Handle missing version in payload.
## Code Before: import json from urllib.parse import urljoin import sys def get_payload(): return json.load(sys.stdin) def get_index_url(payload): source = payload['source'] uri = source['uri'] index = source['index'] if not uri.endswith('/'): uri += '/' if not index.endswith('/'): index += '/' return urljoin(uri, index) def get_package_url(payload): package = payload['source']['package'] return get_index_url(payload) + package def get_auth(payload): source = payload['source'] return source['username'], source['password'] def get_version(payload): if 'version' in payload: version = payload['version']['version'] else: version = None return version ## Instruction: Handle missing version in payload. ## Code After: import json from urllib.parse import urljoin import sys def get_payload(): return json.load(sys.stdin) def get_index_url(payload): source = payload['source'] uri = source['uri'] index = source['index'] if not uri.endswith('/'): uri += '/' if not index.endswith('/'): index += '/' return urljoin(uri, index) def get_package_url(payload): package = payload['source']['package'] return get_index_url(payload) + package def get_auth(payload): source = payload['source'] return source['username'], source['password'] def get_version(payload): try: version = payload['version']['version'] except TypeError: version = None return version
import json from urllib.parse import urljoin import sys def get_payload(): return json.load(sys.stdin) def get_index_url(payload): source = payload['source'] uri = source['uri'] index = source['index'] if not uri.endswith('/'): uri += '/' if not index.endswith('/'): index += '/' return urljoin(uri, index) def get_package_url(payload): package = payload['source']['package'] return get_index_url(payload) + package def get_auth(payload): source = payload['source'] return source['username'], source['password'] def get_version(payload): - if 'version' in payload: + try: version = payload['version']['version'] - else: + except TypeError: version = None return version
0138eacf0d518b86e819a70000b7b527434a6b35
libretto/signals.py
libretto/signals.py
from __future__ import unicode_literals from celery_haystack.signals import CelerySignalProcessor from django.contrib.admin.models import LogEntry from reversion.models import Version, Revision from .tasks import auto_invalidate class CeleryAutoInvalidator(CelerySignalProcessor): def enqueue(self, action, instance, sender, **kwargs): if sender in (LogEntry, Revision, Version): return auto_invalidate.delay(action, instance)
from __future__ import unicode_literals from celery_haystack.signals import CelerySignalProcessor from django.contrib.admin.models import LogEntry from django.contrib.sessions.models import Session from reversion.models import Version, Revision from .tasks import auto_invalidate class CeleryAutoInvalidator(CelerySignalProcessor): def enqueue(self, action, instance, sender, **kwargs): if sender in (LogEntry, Session, Revision, Version): return auto_invalidate.delay(action, instance.__class__, instance.pk)
Change les arguments passés à celery pour gérer la sérialisation JSON.
Change les arguments passés à celery pour gérer la sérialisation JSON.
Python
bsd-3-clause
dezede/dezede,dezede/dezede,dezede/dezede,dezede/dezede
from __future__ import unicode_literals from celery_haystack.signals import CelerySignalProcessor from django.contrib.admin.models import LogEntry + from django.contrib.sessions.models import Session from reversion.models import Version, Revision from .tasks import auto_invalidate class CeleryAutoInvalidator(CelerySignalProcessor): def enqueue(self, action, instance, sender, **kwargs): - if sender in (LogEntry, Revision, Version): + if sender in (LogEntry, Session, Revision, Version): return - auto_invalidate.delay(action, instance) + auto_invalidate.delay(action, instance.__class__, instance.pk)
Change les arguments passés à celery pour gérer la sérialisation JSON.
## Code Before: from __future__ import unicode_literals from celery_haystack.signals import CelerySignalProcessor from django.contrib.admin.models import LogEntry from reversion.models import Version, Revision from .tasks import auto_invalidate class CeleryAutoInvalidator(CelerySignalProcessor): def enqueue(self, action, instance, sender, **kwargs): if sender in (LogEntry, Revision, Version): return auto_invalidate.delay(action, instance) ## Instruction: Change les arguments passés à celery pour gérer la sérialisation JSON. ## Code After: from __future__ import unicode_literals from celery_haystack.signals import CelerySignalProcessor from django.contrib.admin.models import LogEntry from django.contrib.sessions.models import Session from reversion.models import Version, Revision from .tasks import auto_invalidate class CeleryAutoInvalidator(CelerySignalProcessor): def enqueue(self, action, instance, sender, **kwargs): if sender in (LogEntry, Session, Revision, Version): return auto_invalidate.delay(action, instance.__class__, instance.pk)
from __future__ import unicode_literals from celery_haystack.signals import CelerySignalProcessor from django.contrib.admin.models import LogEntry + from django.contrib.sessions.models import Session from reversion.models import Version, Revision from .tasks import auto_invalidate class CeleryAutoInvalidator(CelerySignalProcessor): def enqueue(self, action, instance, sender, **kwargs): - if sender in (LogEntry, Revision, Version): + if sender in (LogEntry, Session, Revision, Version): ? +++++++++ return - auto_invalidate.delay(action, instance) + auto_invalidate.delay(action, instance.__class__, instance.pk) ? +++++++++++++++++++++++
396df5eac473fccc16e103d3d3316aefd653789a
changeling/models.py
changeling/models.py
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, 'tags': {'type': 'array'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None, tags=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description self.tags = tags or [] @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
Add tags - that was too easy
Add tags - that was too easy
Python
apache-2.0
bcwaldon/changeling,bcwaldon/changeling
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, + 'tags': {'type': 'array'}, }, 'additionalProperties': False, } - def __init__(self, id=None, name=None, description=None): + def __init__(self, id=None, name=None, description=None, tags=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description + self.tags = tags or [] @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
Add tags - that was too easy
## Code Before: import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True ## Instruction: Add tags - that was too easy ## Code After: import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, 'tags': {'type': 'array'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None, tags=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description self.tags = tags or [] @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, + 'tags': {'type': 'array'}, }, 'additionalProperties': False, } - def __init__(self, id=None, name=None, description=None): + def __init__(self, id=None, name=None, description=None, tags=None): ? +++++++++++ self.id = id or str(uuid.uuid4()) self.name = name self.description = description + self.tags = tags or [] @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
1cafb39b6204010d3e17b059254af6042f4a9efc
apts/__init__.py
apts/__init__.py
import os import shutil import configparser from .equipment import Equipment from .observations import Observation from .place import Place from .weather import Weather from .notify import Notify from .catalogs import Catalogs from .utils import Utils user_config = os.path.expanduser("~") + "/.config/apts/apts.ini" config = configparser.ConfigParser() config.read(user_config) # Load static fields from config Weather.API_KEY = config['weather']['api_key'] Weather.API_URL = config['weather']['api_url'] Notify.EMAIL_ADDRESS = config['notification']['email_address'] Notify.EMAIL_PASSWORD = config['notification']['email_password'] __version__ = '0.2.9'
import os import shutil import configparser from .equipment import Equipment from .observations import Observation from .place import Place from .weather import Weather from .notify import Notify from .catalogs import Catalogs from .utils import Utils # Default values for configuration values DEFAULTS = { 'weather': { 'api_key': 'unknown', 'api_url': 'unknown' }, 'notification': { 'email_address': 'unknown', 'email_password': 'unknown' } } # Init config with default values config = configparser.ConfigParser(DEFAULTS) # Read users configuration user_config = os.path.expanduser("~") + "/.config/apts/apts.ini" config.read(user_config) # Load static fields from config Weather.API_KEY = config['weather']['api_key'] Weather.API_URL = config['weather']['api_url'] Notify.EMAIL_ADDRESS = config['notification']['email_address'] Notify.EMAIL_PASSWORD = config['notification']['email_password'] __version__ = '0.2.9'
Add default values for config
Add default values for config
Python
apache-2.0
pozar87/apts
import os import shutil import configparser from .equipment import Equipment from .observations import Observation from .place import Place from .weather import Weather from .notify import Notify from .catalogs import Catalogs from .utils import Utils + # Default values for configuration values + DEFAULTS = { + 'weather': { + 'api_key': 'unknown', + 'api_url': 'unknown' + }, + 'notification': { + 'email_address': 'unknown', + 'email_password': 'unknown' + } + } + + # Init config with default values + config = configparser.ConfigParser(DEFAULTS) + + # Read users configuration user_config = os.path.expanduser("~") + "/.config/apts/apts.ini" - - config = configparser.ConfigParser() config.read(user_config) # Load static fields from config Weather.API_KEY = config['weather']['api_key'] Weather.API_URL = config['weather']['api_url'] Notify.EMAIL_ADDRESS = config['notification']['email_address'] Notify.EMAIL_PASSWORD = config['notification']['email_password'] __version__ = '0.2.9'
Add default values for config
## Code Before: import os import shutil import configparser from .equipment import Equipment from .observations import Observation from .place import Place from .weather import Weather from .notify import Notify from .catalogs import Catalogs from .utils import Utils user_config = os.path.expanduser("~") + "/.config/apts/apts.ini" config = configparser.ConfigParser() config.read(user_config) # Load static fields from config Weather.API_KEY = config['weather']['api_key'] Weather.API_URL = config['weather']['api_url'] Notify.EMAIL_ADDRESS = config['notification']['email_address'] Notify.EMAIL_PASSWORD = config['notification']['email_password'] __version__ = '0.2.9' ## Instruction: Add default values for config ## Code After: import os import shutil import configparser from .equipment import Equipment from .observations import Observation from .place import Place from .weather import Weather from .notify import Notify from .catalogs import Catalogs from .utils import Utils # Default values for configuration values DEFAULTS = { 'weather': { 'api_key': 'unknown', 'api_url': 'unknown' }, 'notification': { 'email_address': 'unknown', 'email_password': 'unknown' } } # Init config with default values config = configparser.ConfigParser(DEFAULTS) # Read users configuration user_config = os.path.expanduser("~") + "/.config/apts/apts.ini" config.read(user_config) # Load static fields from config Weather.API_KEY = config['weather']['api_key'] Weather.API_URL = config['weather']['api_url'] Notify.EMAIL_ADDRESS = config['notification']['email_address'] Notify.EMAIL_PASSWORD = config['notification']['email_password'] __version__ = '0.2.9'
import os import shutil import configparser from .equipment import Equipment from .observations import Observation from .place import Place from .weather import Weather from .notify import Notify from .catalogs import Catalogs from .utils import Utils + # Default values for configuration values + DEFAULTS = { + 'weather': { + 'api_key': 'unknown', + 'api_url': 'unknown' + }, + 'notification': { + 'email_address': 'unknown', + 'email_password': 'unknown' + } + } + + # Init config with default values + config = configparser.ConfigParser(DEFAULTS) + + # Read users configuration user_config = os.path.expanduser("~") + "/.config/apts/apts.ini" - - config = configparser.ConfigParser() config.read(user_config) # Load static fields from config Weather.API_KEY = config['weather']['api_key'] Weather.API_URL = config['weather']['api_url'] Notify.EMAIL_ADDRESS = config['notification']['email_address'] Notify.EMAIL_PASSWORD = config['notification']['email_password'] __version__ = '0.2.9'
8c3e3ec6076d8b9ee858fca00d92717d77c67ade
time_lapse.py
time_lapse.py
import sys import time import picamera import settings from settings import IMAGE, SNAP import uploader def main(): with picamera.PiCamera() as camera: camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y) time.sleep(2) output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg' capture = camera.capture_continuous(output_file, quality=IMAGE.quality) for i, _ in enumerate(capture): if i == SNAP.total - 1: break time.sleep(SNAP.interval) if __name__ == '__main__': while True: main()
import time import picamera from settings import Job, IMAGES_DIRECTORY def main(): job = Job() if job.exists(): resolution_x = job.image_settings.resolution_x resolution_y = job.image_settings.resolution_y image_quality = job.image_settings.quality snap_interval = job.snap_settings.interval snap_total = job.snap_settings.total with picamera.PiCamera() as camera: camera.resolution = (resolution_x, resolution_y) time.sleep(2) output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg' capture = camera.capture_continuous(output_file, quality=image_quality) for i, _ in enumerate(capture): if i == snap_total - 1: job.archive() break time.sleep(snap_interval) if __name__ == '__main__': while True: main()
Check for job in main loop
Check for job in main loop
Python
mit
projectweekend/Pi-Camera-Time-Lapse,projectweekend/Pi-Camera-Time-Lapse
- import sys import time import picamera + from settings import Job, IMAGES_DIRECTORY - import settings - from settings import IMAGE, SNAP - import uploader def main(): + job = Job() + if job.exists(): + resolution_x = job.image_settings.resolution_x + resolution_y = job.image_settings.resolution_y + image_quality = job.image_settings.quality + snap_interval = job.snap_settings.interval + snap_total = job.snap_settings.total - with picamera.PiCamera() as camera: + with picamera.PiCamera() as camera: - camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y) + camera.resolution = (resolution_x, resolution_y) - time.sleep(2) + time.sleep(2) - output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg' + output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg' - capture = camera.capture_continuous(output_file, quality=IMAGE.quality) + capture = camera.capture_continuous(output_file, quality=image_quality) - for i, _ in enumerate(capture): + for i, _ in enumerate(capture): - if i == SNAP.total - 1: + if i == snap_total - 1: + job.archive() - break + break - time.sleep(SNAP.interval) + time.sleep(snap_interval) if __name__ == '__main__': while True: main()
Check for job in main loop
## Code Before: import sys import time import picamera import settings from settings import IMAGE, SNAP import uploader def main(): with picamera.PiCamera() as camera: camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y) time.sleep(2) output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg' capture = camera.capture_continuous(output_file, quality=IMAGE.quality) for i, _ in enumerate(capture): if i == SNAP.total - 1: break time.sleep(SNAP.interval) if __name__ == '__main__': while True: main() ## Instruction: Check for job in main loop ## Code After: import time import picamera from settings import Job, IMAGES_DIRECTORY def main(): job = Job() if job.exists(): resolution_x = job.image_settings.resolution_x resolution_y = job.image_settings.resolution_y image_quality = job.image_settings.quality snap_interval = job.snap_settings.interval snap_total = job.snap_settings.total with picamera.PiCamera() as camera: camera.resolution = (resolution_x, resolution_y) time.sleep(2) output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg' capture = camera.capture_continuous(output_file, quality=image_quality) for i, _ in enumerate(capture): if i == snap_total - 1: job.archive() break time.sleep(snap_interval) if __name__ == '__main__': while True: main()
- import sys import time import picamera + from settings import Job, IMAGES_DIRECTORY - import settings - from settings import IMAGE, SNAP - import uploader def main(): + job = Job() + if job.exists(): + resolution_x = job.image_settings.resolution_x + resolution_y = job.image_settings.resolution_y + image_quality = job.image_settings.quality + snap_interval = job.snap_settings.interval + snap_total = job.snap_settings.total - with picamera.PiCamera() as camera: + with picamera.PiCamera() as camera: ? ++++ - camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y) ? ------ ------ + camera.resolution = (resolution_x, resolution_y) ? ++++ - time.sleep(2) + time.sleep(2) ? ++++ - output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg' ? --------- + output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg' ? ++++ - capture = camera.capture_continuous(output_file, quality=IMAGE.quality) ? ^^^^^^ + capture = camera.capture_continuous(output_file, quality=image_quality) ? ++++ ^^^^^^ - for i, _ in enumerate(capture): + for i, _ in enumerate(capture): ? ++++ - if i == SNAP.total - 1: ? ^^^^^ + if i == snap_total - 1: ? ++++ ^^^^^ + job.archive() - break + break ? ++++ - time.sleep(SNAP.interval) ? ^^^^^ + time.sleep(snap_interval) ? ++++ ^^^^^ if __name__ == '__main__': while True: main()
ebac72a3753205d3e45041c6db636a378187e3cf
pylua/tests/test_compiled.py
pylua/tests/test_compiled.py
import os import subprocess from pylua.tests.helpers import test_file class TestCompiled(object): """ Tests compiled binary """ def test_addition(self, capsys): f = test_file(src=""" -- short add x = 10 y = 5 z = y + y + x print(z) print(z+y) --a = 100+y lx = 1234567890 ly = 99999999 print(lx+ly) --print(lx+1234567890) """, suffix=".l" ) out = subprocess.check_output(['bin/pylua', f.name]) assert out == "20.000000\n25.000000\n1334567889.000000\n"
import os import subprocess from pylua.tests.helpers import test_file class TestCompiled(object): """ Tests compiled binary """ PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua')) def test_addition(self, capsys): f = test_file(src=""" -- short add x = 10 y = 5 z = y + y + x print(z) print(z+y) --a = 100+y lx = 1234567890 ly = 99999999 print(lx+ly) --print(lx+1234567890) """, suffix=".l" ) out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name]) assert out == "20.000000\n25.000000\n1334567889.000000\n"
Use absolute path for lua binary in tests
Use absolute path for lua binary in tests
Python
bsd-3-clause
fhahn/luna,fhahn/luna
import os import subprocess from pylua.tests.helpers import test_file class TestCompiled(object): """ Tests compiled binary """ + + PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua')) def test_addition(self, capsys): f = test_file(src=""" -- short add x = 10 y = 5 z = y + y + x print(z) print(z+y) --a = 100+y lx = 1234567890 ly = 99999999 print(lx+ly) --print(lx+1234567890) """, suffix=".l" ) - out = subprocess.check_output(['bin/pylua', f.name]) + out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name]) assert out == "20.000000\n25.000000\n1334567889.000000\n"
Use absolute path for lua binary in tests
## Code Before: import os import subprocess from pylua.tests.helpers import test_file class TestCompiled(object): """ Tests compiled binary """ def test_addition(self, capsys): f = test_file(src=""" -- short add x = 10 y = 5 z = y + y + x print(z) print(z+y) --a = 100+y lx = 1234567890 ly = 99999999 print(lx+ly) --print(lx+1234567890) """, suffix=".l" ) out = subprocess.check_output(['bin/pylua', f.name]) assert out == "20.000000\n25.000000\n1334567889.000000\n" ## Instruction: Use absolute path for lua binary in tests ## Code After: import os import subprocess from pylua.tests.helpers import test_file class TestCompiled(object): """ Tests compiled binary """ PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua')) def test_addition(self, capsys): f = test_file(src=""" -- short add x = 10 y = 5 z = y + y + x print(z) print(z+y) --a = 100+y lx = 1234567890 ly = 99999999 print(lx+ly) --print(lx+1234567890) """, suffix=".l" ) out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name]) assert out == "20.000000\n25.000000\n1334567889.000000\n"
import os import subprocess from pylua.tests.helpers import test_file class TestCompiled(object): """ Tests compiled binary """ + + PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua')) def test_addition(self, capsys): f = test_file(src=""" -- short add x = 10 y = 5 z = y + y + x print(z) print(z+y) --a = 100+y lx = 1234567890 ly = 99999999 print(lx+ly) --print(lx+1234567890) """, suffix=".l" ) - out = subprocess.check_output(['bin/pylua', f.name]) ? ^^ ---- ^^^ + out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name]) ? ^^^^^^^^ ^^^^^^^^^^^^ assert out == "20.000000\n25.000000\n1334567889.000000\n"
46816c4d8470192e76e730969ddcedeb8391fdcf
setup.py
setup.py
from distutils.core import setup setup(name="Neighborhoodize", version='0.9', description='Utility for translating lat, long coordinates into ' 'neighborhoods in various cities', author='Brian Lange', author_email='brian.lange@datascopeanalytics.com', url='https://github.com/bjlange/neighborhoodize', packages=['neighborhoodize', ], package_data={'neighborhoodize': ['data/*']}, install_requires=[ "Shapely >= 1.5.7", "beautifulsoup4 >= 4.3.2", "fastkml >= 0.9"], )
from distutils.core import setup setup(name="Neighborhoodize", version='0.9', description='Utility for translating lat, long coordinates into ' 'neighborhoods in various cities', author='Brian Lange', author_email='brian.lange@datascopeanalytics.com', url='https://github.com/bjlange/neighborhoodize', packages=['neighborhoodize', ], package_data={'neighborhoodize': ['data/*']}, download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9', install_requires=[ "Shapely >= 1.5.7", "beautifulsoup4 >= 4.3.2", "fastkml >= 0.9"], )
Add download url for pypi
Add download url for pypi
Python
mit
bjlange/neighborhoodize
from distutils.core import setup setup(name="Neighborhoodize", version='0.9', description='Utility for translating lat, long coordinates into ' 'neighborhoods in various cities', author='Brian Lange', author_email='brian.lange@datascopeanalytics.com', url='https://github.com/bjlange/neighborhoodize', packages=['neighborhoodize', ], package_data={'neighborhoodize': ['data/*']}, + download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9', install_requires=[ "Shapely >= 1.5.7", "beautifulsoup4 >= 4.3.2", "fastkml >= 0.9"], )
Add download url for pypi
## Code Before: from distutils.core import setup setup(name="Neighborhoodize", version='0.9', description='Utility for translating lat, long coordinates into ' 'neighborhoods in various cities', author='Brian Lange', author_email='brian.lange@datascopeanalytics.com', url='https://github.com/bjlange/neighborhoodize', packages=['neighborhoodize', ], package_data={'neighborhoodize': ['data/*']}, install_requires=[ "Shapely >= 1.5.7", "beautifulsoup4 >= 4.3.2", "fastkml >= 0.9"], ) ## Instruction: Add download url for pypi ## Code After: from distutils.core import setup setup(name="Neighborhoodize", version='0.9', description='Utility for translating lat, long coordinates into ' 'neighborhoods in various cities', author='Brian Lange', author_email='brian.lange@datascopeanalytics.com', url='https://github.com/bjlange/neighborhoodize', packages=['neighborhoodize', ], package_data={'neighborhoodize': ['data/*']}, download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9', install_requires=[ "Shapely >= 1.5.7", "beautifulsoup4 >= 4.3.2", "fastkml >= 0.9"], )
from distutils.core import setup setup(name="Neighborhoodize", version='0.9', description='Utility for translating lat, long coordinates into ' 'neighborhoods in various cities', author='Brian Lange', author_email='brian.lange@datascopeanalytics.com', url='https://github.com/bjlange/neighborhoodize', packages=['neighborhoodize', ], package_data={'neighborhoodize': ['data/*']}, + download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9', install_requires=[ "Shapely >= 1.5.7", "beautifulsoup4 >= 4.3.2", "fastkml >= 0.9"], )
011949b266ab33df8c0f9bec29ba693824e7d8ef
setup.py
setup.py
import sys from distutils.core import setup setup( name='pathlib', version=open('VERSION.txt').read().strip(), py_modules=['pathlib'], license='MIT License', description='Object-oriented filesystem paths', long_description=open('README.txt').read(), author='Antoine Pitrou', author_email='solipsis@pitrou.net', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Topic :: Software Development :: Libraries', 'Topic :: System :: Filesystems', ], download_url='https://pypi.python.org/pypi/pathlib/', url='http://readthedocs.org/docs/pathlib/', )
import sys from distutils.core import setup setup( name='pathlib', version=open('VERSION.txt').read().strip(), py_modules=['pathlib'], license='MIT License', description='Object-oriented filesystem paths', long_description=open('README.txt').read(), author='Antoine Pitrou', author_email='solipsis@pitrou.net', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Topic :: Software Development :: Libraries', 'Topic :: System :: Filesystems', ], download_url='https://pypi.python.org/pypi/pathlib/', url='http://readthedocs.org/docs/pathlib/', )
Add classifier for Python 3.3
Add classifier for Python 3.3
Python
mit
pombreda/pathlib
import sys from distutils.core import setup setup( name='pathlib', version=open('VERSION.txt').read().strip(), py_modules=['pathlib'], license='MIT License', description='Object-oriented filesystem paths', long_description=open('README.txt').read(), author='Antoine Pitrou', author_email='solipsis@pitrou.net', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', + 'Programming Language :: Python :: 3.3', 'Topic :: Software Development :: Libraries', 'Topic :: System :: Filesystems', ], download_url='https://pypi.python.org/pypi/pathlib/', url='http://readthedocs.org/docs/pathlib/', )
Add classifier for Python 3.3
## Code Before: import sys from distutils.core import setup setup( name='pathlib', version=open('VERSION.txt').read().strip(), py_modules=['pathlib'], license='MIT License', description='Object-oriented filesystem paths', long_description=open('README.txt').read(), author='Antoine Pitrou', author_email='solipsis@pitrou.net', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Topic :: Software Development :: Libraries', 'Topic :: System :: Filesystems', ], download_url='https://pypi.python.org/pypi/pathlib/', url='http://readthedocs.org/docs/pathlib/', ) ## Instruction: Add classifier for Python 3.3 ## Code After: import sys from distutils.core import setup setup( name='pathlib', version=open('VERSION.txt').read().strip(), py_modules=['pathlib'], license='MIT License', description='Object-oriented filesystem paths', long_description=open('README.txt').read(), author='Antoine Pitrou', author_email='solipsis@pitrou.net', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Topic :: Software Development :: Libraries', 'Topic :: System :: Filesystems', ], download_url='https://pypi.python.org/pypi/pathlib/', url='http://readthedocs.org/docs/pathlib/', )
import sys from distutils.core import setup setup( name='pathlib', version=open('VERSION.txt').read().strip(), py_modules=['pathlib'], license='MIT License', description='Object-oriented filesystem paths', long_description=open('README.txt').read(), author='Antoine Pitrou', author_email='solipsis@pitrou.net', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', + 'Programming Language :: Python :: 3.3', 'Topic :: Software Development :: Libraries', 'Topic :: System :: Filesystems', ], download_url='https://pypi.python.org/pypi/pathlib/', url='http://readthedocs.org/docs/pathlib/', )
d4412f8573dbfc1b06f2a298cc5c3042c6c468e6
tests/test_api.py
tests/test_api.py
from django.test import TestCase from django_snooze import apis class APITestCase(TestCase): def setUp(self): """Sets up an API object to play with. :returns: None """ self.api = apis.api self.api.discover_models() def test_apps(self): """Test if the right apps are present. :returns: None """ self.assertIn('tests', self.api._resources.keys()) self.assertIn('auth', self.api._resources.keys())
from django.test import TestCase from django_snooze import apis class APITestCase(TestCase): def setUp(self): """Sets up an API object to play with. :returns: None """ self.api = apis.api self.api.discover_models() def test_apps(self): """Test if the right apps are present. :returns: None """ self.assertIn('tests', self.api._resources.keys()) self.assertIn('auth', self.api._resources.keys()) tests_models = [x.model_name for x in self.api._resources['tests']] self.assertNotIn('abstract', tests_models)
Test to see if abstract classes sneak in.
Test to see if abstract classes sneak in. Now that get_models has been found to skip abstract classes, we want to test for this in case this behaviour ever changes.
Python
bsd-3-clause
ainmosni/django-snooze,ainmosni/django-snooze
from django.test import TestCase from django_snooze import apis class APITestCase(TestCase): def setUp(self): """Sets up an API object to play with. :returns: None """ self.api = apis.api self.api.discover_models() def test_apps(self): """Test if the right apps are present. :returns: None """ self.assertIn('tests', self.api._resources.keys()) self.assertIn('auth', self.api._resources.keys()) + tests_models = [x.model_name for x in self.api._resources['tests']] + self.assertNotIn('abstract', tests_models)
Test to see if abstract classes sneak in.
## Code Before: from django.test import TestCase from django_snooze import apis class APITestCase(TestCase): def setUp(self): """Sets up an API object to play with. :returns: None """ self.api = apis.api self.api.discover_models() def test_apps(self): """Test if the right apps are present. :returns: None """ self.assertIn('tests', self.api._resources.keys()) self.assertIn('auth', self.api._resources.keys()) ## Instruction: Test to see if abstract classes sneak in. ## Code After: from django.test import TestCase from django_snooze import apis class APITestCase(TestCase): def setUp(self): """Sets up an API object to play with. :returns: None """ self.api = apis.api self.api.discover_models() def test_apps(self): """Test if the right apps are present. :returns: None """ self.assertIn('tests', self.api._resources.keys()) self.assertIn('auth', self.api._resources.keys()) tests_models = [x.model_name for x in self.api._resources['tests']] self.assertNotIn('abstract', tests_models)
from django.test import TestCase from django_snooze import apis class APITestCase(TestCase): def setUp(self): """Sets up an API object to play with. :returns: None """ self.api = apis.api self.api.discover_models() def test_apps(self): """Test if the right apps are present. :returns: None """ self.assertIn('tests', self.api._resources.keys()) self.assertIn('auth', self.api._resources.keys()) + tests_models = [x.model_name for x in self.api._resources['tests']] + self.assertNotIn('abstract', tests_models)
63a7b11d3ae51a944bf2e70637dea503e455c2f5
fontdump/cli.py
fontdump/cli.py
from collections import OrderedDict import requests import cssutils USER_AGENTS = OrderedDict() USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 def main(): font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400' sheets={} for (format, ua) in USER_AGENTS.items(): headers = { 'User-Agent': ua, } r =requests.get(font_url, headers=headers) sheets[format] = cssutils.parseString(r.content) if __name__ == '__main__': main()
import requests import cssutils USER_AGENTS = { 'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome 'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 } def main(): font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400' sheets={} for (format, ua) in USER_AGENTS.items(): headers = { 'User-Agent': ua, } r =requests.get(font_url, headers=headers) sheets[format] = cssutils.parseString(r.content) if __name__ == '__main__': main()
Revert "The order of the formats matters. Use OrderedDict instead of dict"
Revert "The order of the formats matters. Use OrderedDict instead of dict" I can't rely on the order of dict. The control flow is more complex. This reverts commit 3389ed71971ddacd185bbbf8fe667a8651108c70.
Python
mit
glasslion/fontdump
- from collections import OrderedDict - import requests import cssutils - USER_AGENTS = OrderedDict() + USER_AGENTS = { - USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome + 'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome + 'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 - USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 + 'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 - USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 - USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 + 'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 - + } def main(): font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400' sheets={} for (format, ua) in USER_AGENTS.items(): headers = { 'User-Agent': ua, } r =requests.get(font_url, headers=headers) sheets[format] = cssutils.parseString(r.content) if __name__ == '__main__': main()
Revert "The order of the formats matters. Use OrderedDict instead of dict"
## Code Before: from collections import OrderedDict import requests import cssutils USER_AGENTS = OrderedDict() USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 def main(): font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400' sheets={} for (format, ua) in USER_AGENTS.items(): headers = { 'User-Agent': ua, } r =requests.get(font_url, headers=headers) sheets[format] = cssutils.parseString(r.content) if __name__ == '__main__': main() ## Instruction: Revert "The order of the formats matters. Use OrderedDict instead of dict" ## Code After: import requests import cssutils USER_AGENTS = { 'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome 'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 } def main(): font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400' sheets={} for (format, ua) in USER_AGENTS.items(): headers = { 'User-Agent': ua, } r =requests.get(font_url, headers=headers) sheets[format] = cssutils.parseString(r.content) if __name__ == '__main__': main()
- from collections import OrderedDict - import requests import cssutils - USER_AGENTS = OrderedDict() + USER_AGENTS = { - USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome ? ^^^^^^^^^^^^ ^^^ + 'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome ? ^^^^ ^ + 'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 - USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 ? ^^^^^^^^^^^^ ^^^ + 'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2 ? ^^^^ ^ - USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6 - USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 ? ^^^^^^^^^^^^ ^^^^ ^^^ + 'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2 ? ^^^^ ^^^ ^ - + } def main(): font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400' sheets={} for (format, ua) in USER_AGENTS.items(): headers = { 'User-Agent': ua, } r =requests.get(font_url, headers=headers) sheets[format] = cssutils.parseString(r.content) if __name__ == '__main__': main()
eb5294f0df32442dbd7431fd9200388ca4c63d62
tests/builtins/test_reversed.py
tests/builtins/test_reversed.py
from .. utils import TranspileTestCase, BuiltinFunctionTestCase class ReversedTests(TranspileTestCase): pass class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["reversed"] not_implemented = [ 'test_range', ]
from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase def _iterate_test(datatype): def test_func(self): code = '\n'.join([ '\nfor x in {value}:\n print(x)\n'.format(value=value) for value in SAMPLE_DATA[datatype] ]) self.assertCodeExecution(code) return test_func class ReversedTests(TranspileTestCase): # test_iterate_bytearray = _iterate_test('bytearray') test_iterate_bytes = _iterate_test('bytes') test_iterate_list = _iterate_test('list') test_iterate_range = _iterate_test('range') test_iterate_str = _iterate_test('str') test_iterate_tuple = _iterate_test('tuple') class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["reversed"] not_implemented = [ 'test_range', ]
Add iteration tests for reversed type
Add iteration tests for reversed type
Python
bsd-3-clause
cflee/voc,cflee/voc,freakboy3742/voc,freakboy3742/voc
- from .. utils import TranspileTestCase, BuiltinFunctionTestCase + from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase + + + def _iterate_test(datatype): + + def test_func(self): + code = '\n'.join([ + '\nfor x in {value}:\n print(x)\n'.format(value=value) + for value in SAMPLE_DATA[datatype] + ]) + self.assertCodeExecution(code) + + return test_func class ReversedTests(TranspileTestCase): - pass + # test_iterate_bytearray = _iterate_test('bytearray') + test_iterate_bytes = _iterate_test('bytes') + test_iterate_list = _iterate_test('list') + test_iterate_range = _iterate_test('range') + test_iterate_str = _iterate_test('str') + test_iterate_tuple = _iterate_test('tuple') class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["reversed"] not_implemented = [ 'test_range', ]
Add iteration tests for reversed type
## Code Before: from .. utils import TranspileTestCase, BuiltinFunctionTestCase class ReversedTests(TranspileTestCase): pass class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["reversed"] not_implemented = [ 'test_range', ] ## Instruction: Add iteration tests for reversed type ## Code After: from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase def _iterate_test(datatype): def test_func(self): code = '\n'.join([ '\nfor x in {value}:\n print(x)\n'.format(value=value) for value in SAMPLE_DATA[datatype] ]) self.assertCodeExecution(code) return test_func class ReversedTests(TranspileTestCase): # test_iterate_bytearray = _iterate_test('bytearray') test_iterate_bytes = _iterate_test('bytes') test_iterate_list = _iterate_test('list') test_iterate_range = _iterate_test('range') test_iterate_str = _iterate_test('str') test_iterate_tuple = _iterate_test('tuple') class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["reversed"] not_implemented = [ 'test_range', ]
- from .. utils import TranspileTestCase, BuiltinFunctionTestCase + from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase ? +++++++++++++ + + + def _iterate_test(datatype): + + def test_func(self): + code = '\n'.join([ + '\nfor x in {value}:\n print(x)\n'.format(value=value) + for value in SAMPLE_DATA[datatype] + ]) + self.assertCodeExecution(code) + + return test_func class ReversedTests(TranspileTestCase): - pass + # test_iterate_bytearray = _iterate_test('bytearray') + test_iterate_bytes = _iterate_test('bytes') + test_iterate_list = _iterate_test('list') + test_iterate_range = _iterate_test('range') + test_iterate_str = _iterate_test('str') + test_iterate_tuple = _iterate_test('tuple') class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["reversed"] not_implemented = [ 'test_range', ]
4c1237d2969d735cfcf9f3c10cf27cb801996e32
tests/test_integration.py
tests/test_integration.py
"""Unit test module for Selenium testing""" from selenium import webdriver from flask.ext.testing import LiveServerTestCase from tests import TestCase from pages import LoginPage class TestUI(TestCase, LiveServerTestCase): """Test class for UI integration/workflow testing""" def setUp(self): """Reset all tables before testing.""" super(TestUI, self).setUp() self.driver = webdriver.Firefox() self.driver.implicitly_wait(60) self.driver.root_uri = self.get_server_url() def tearDown(self): """Clean db session, drop all tables.""" self.driver.quit() super(TestUI, self).tearDown() def test_login_page(self): """Ensure login page loads successfully""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text) def test_login_form_facebook_exists(self): """Ensure Facebook button present on login form""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertIsNotNone(page.facebook_button)
"""Unit test module for Selenium testing""" import os from selenium import webdriver from flask.ext.testing import LiveServerTestCase from tests import TestCase from pages import LoginPage class TestUI(TestCase, LiveServerTestCase): """Test class for UI integration/workflow testing""" def setUp(self): """Reset all tables before testing.""" super(TestUI, self).setUp() if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ: capabilities = { "tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"], "build": os.environ["TRAVIS_BUILD_NUMBER"], "tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"], } url = "http://{username}:{access_key}@localhost:4445/wd/hub".format( username=os.environ["SAUCE_USERNAME"], access_key=os.environ["SAUCE_ACCESS_KEY"], ) self.driver = webdriver.Remote( desired_capabilities=capabilities, command_executor=url ) else: self.driver = webdriver.Firefox() self.driver.implicitly_wait(60) self.driver.root_uri = self.get_server_url() def tearDown(self): """Clean db session, drop all tables.""" self.driver.quit() super(TestUI, self).tearDown() def test_login_page(self): """Ensure login page loads successfully""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text) def test_login_form_facebook_exists(self): """Ensure Facebook button present on login form""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertIsNotNone(page.facebook_button)
Use Sauce Labs for selenium testing when available
Use Sauce Labs for selenium testing when available
Python
bsd-3-clause
uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal
"""Unit test module for Selenium testing""" + import os from selenium import webdriver from flask.ext.testing import LiveServerTestCase from tests import TestCase from pages import LoginPage class TestUI(TestCase, LiveServerTestCase): """Test class for UI integration/workflow testing""" def setUp(self): """Reset all tables before testing.""" super(TestUI, self).setUp() + if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ: + + capabilities = { + "tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"], + "build": os.environ["TRAVIS_BUILD_NUMBER"], + "tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"], + } + url = "http://{username}:{access_key}@localhost:4445/wd/hub".format( + username=os.environ["SAUCE_USERNAME"], + access_key=os.environ["SAUCE_ACCESS_KEY"], + ) + + self.driver = webdriver.Remote( + desired_capabilities=capabilities, + command_executor=url + ) + + else: - self.driver = webdriver.Firefox() + self.driver = webdriver.Firefox() + self.driver.implicitly_wait(60) self.driver.root_uri = self.get_server_url() def tearDown(self): """Clean db session, drop all tables.""" self.driver.quit() super(TestUI, self).tearDown() def test_login_page(self): """Ensure login page loads successfully""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text) def test_login_form_facebook_exists(self): """Ensure Facebook button present on login form""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertIsNotNone(page.facebook_button)
Use Sauce Labs for selenium testing when available
## Code Before: """Unit test module for Selenium testing""" from selenium import webdriver from flask.ext.testing import LiveServerTestCase from tests import TestCase from pages import LoginPage class TestUI(TestCase, LiveServerTestCase): """Test class for UI integration/workflow testing""" def setUp(self): """Reset all tables before testing.""" super(TestUI, self).setUp() self.driver = webdriver.Firefox() self.driver.implicitly_wait(60) self.driver.root_uri = self.get_server_url() def tearDown(self): """Clean db session, drop all tables.""" self.driver.quit() super(TestUI, self).tearDown() def test_login_page(self): """Ensure login page loads successfully""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text) def test_login_form_facebook_exists(self): """Ensure Facebook button present on login form""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertIsNotNone(page.facebook_button) ## Instruction: Use Sauce Labs for selenium testing when available ## Code After: """Unit test module for Selenium testing""" import os from selenium import webdriver from flask.ext.testing import LiveServerTestCase from tests import TestCase from pages import LoginPage class TestUI(TestCase, LiveServerTestCase): """Test class for UI integration/workflow testing""" def setUp(self): """Reset all tables before testing.""" super(TestUI, self).setUp() if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ: capabilities = { "tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"], "build": os.environ["TRAVIS_BUILD_NUMBER"], "tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"], } url = "http://{username}:{access_key}@localhost:4445/wd/hub".format( username=os.environ["SAUCE_USERNAME"], access_key=os.environ["SAUCE_ACCESS_KEY"], ) self.driver = webdriver.Remote( desired_capabilities=capabilities, command_executor=url ) else: self.driver = webdriver.Firefox() self.driver.implicitly_wait(60) self.driver.root_uri = self.get_server_url() def tearDown(self): """Clean db session, drop all tables.""" self.driver.quit() super(TestUI, self).tearDown() def test_login_page(self): """Ensure login page loads successfully""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text) def test_login_form_facebook_exists(self): """Ensure Facebook button present on login form""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertIsNotNone(page.facebook_button)
"""Unit test module for Selenium testing""" + import os from selenium import webdriver from flask.ext.testing import LiveServerTestCase from tests import TestCase from pages import LoginPage class TestUI(TestCase, LiveServerTestCase): """Test class for UI integration/workflow testing""" def setUp(self): """Reset all tables before testing.""" super(TestUI, self).setUp() + if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ: + + capabilities = { + "tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"], + "build": os.environ["TRAVIS_BUILD_NUMBER"], + "tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"], + } + url = "http://{username}:{access_key}@localhost:4445/wd/hub".format( + username=os.environ["SAUCE_USERNAME"], + access_key=os.environ["SAUCE_ACCESS_KEY"], + ) + + self.driver = webdriver.Remote( + desired_capabilities=capabilities, + command_executor=url + ) + + else: - self.driver = webdriver.Firefox() + self.driver = webdriver.Firefox() ? ++++ + self.driver.implicitly_wait(60) self.driver.root_uri = self.get_server_url() def tearDown(self): """Clean db session, drop all tables.""" self.driver.quit() super(TestUI, self).tearDown() def test_login_page(self): """Ensure login page loads successfully""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text) def test_login_form_facebook_exists(self): """Ensure Facebook button present on login form""" page = LoginPage(self.driver) page.get("/user/sign-in") self.assertIsNotNone(page.facebook_button)
9f484e6eb4fcf37f5515d62eb80be6a0f8d5a097
swteams/views.py
swteams/views.py
import iris.views from teams.models import Team def activity(request, slug, *args, **kw): template_name = 'teams/activity.html' team = get_object_or_404(Team, slug=slug) template_context = { 'group': team, } return render_to_response(template_name, template_context, RequestContext(request)) def topics_with_team_slug(*args, **kwargs): # Insert the team into the template context as 'group' so that # breadcrumbs, etc. render properly. slug = kwargs['slug'] extra_context = kwargs.setdefault('extra_context', {}) extra_context['group'] = Team.objects.get(slug=slug) return iris.views.topics(*args, **kwargs)
from django.shortcuts import get_object_or_404, render_to_response from django.template.context import RequestContext import iris.views from teams.models import Team def activity(request, slug, *args, **kw): template_name = 'teams/activity.html' team = get_object_or_404(Team, slug=slug) template_context = { 'group': team, } return render_to_response(template_name, template_context, RequestContext(request)) def topics_with_team_slug(*args, **kwargs): # Insert the team into the template context as 'group' so that # breadcrumbs, etc. render properly. slug = kwargs['slug'] extra_context = kwargs.setdefault('extra_context', {}) extra_context['group'] = Team.objects.get(slug=slug) return iris.views.topics(*args, **kwargs)
Fix team activity history view (was missing imports)
Fix team activity history view (was missing imports)
Python
apache-2.0
snswa/swsites,snswa/swsites,snswa/swsites
+ from django.shortcuts import get_object_or_404, render_to_response + from django.template.context import RequestContext + import iris.views from teams.models import Team def activity(request, slug, *args, **kw): template_name = 'teams/activity.html' team = get_object_or_404(Team, slug=slug) template_context = { 'group': team, } return render_to_response(template_name, template_context, RequestContext(request)) def topics_with_team_slug(*args, **kwargs): # Insert the team into the template context as 'group' so that # breadcrumbs, etc. render properly. slug = kwargs['slug'] extra_context = kwargs.setdefault('extra_context', {}) extra_context['group'] = Team.objects.get(slug=slug) return iris.views.topics(*args, **kwargs)
Fix team activity history view (was missing imports)
## Code Before: import iris.views from teams.models import Team def activity(request, slug, *args, **kw): template_name = 'teams/activity.html' team = get_object_or_404(Team, slug=slug) template_context = { 'group': team, } return render_to_response(template_name, template_context, RequestContext(request)) def topics_with_team_slug(*args, **kwargs): # Insert the team into the template context as 'group' so that # breadcrumbs, etc. render properly. slug = kwargs['slug'] extra_context = kwargs.setdefault('extra_context', {}) extra_context['group'] = Team.objects.get(slug=slug) return iris.views.topics(*args, **kwargs) ## Instruction: Fix team activity history view (was missing imports) ## Code After: from django.shortcuts import get_object_or_404, render_to_response from django.template.context import RequestContext import iris.views from teams.models import Team def activity(request, slug, *args, **kw): template_name = 'teams/activity.html' team = get_object_or_404(Team, slug=slug) template_context = { 'group': team, } return render_to_response(template_name, template_context, RequestContext(request)) def topics_with_team_slug(*args, **kwargs): # Insert the team into the template context as 'group' so that # breadcrumbs, etc. render properly. slug = kwargs['slug'] extra_context = kwargs.setdefault('extra_context', {}) extra_context['group'] = Team.objects.get(slug=slug) return iris.views.topics(*args, **kwargs)
+ from django.shortcuts import get_object_or_404, render_to_response + from django.template.context import RequestContext + import iris.views from teams.models import Team def activity(request, slug, *args, **kw): template_name = 'teams/activity.html' team = get_object_or_404(Team, slug=slug) template_context = { 'group': team, } return render_to_response(template_name, template_context, RequestContext(request)) def topics_with_team_slug(*args, **kwargs): # Insert the team into the template context as 'group' so that # breadcrumbs, etc. render properly. slug = kwargs['slug'] extra_context = kwargs.setdefault('extra_context', {}) extra_context['group'] = Team.objects.get(slug=slug) return iris.views.topics(*args, **kwargs)
f0871086222be67ba4add4574cf03d3cb9d39d63
tests/circular/template/test_interpolatedstr.py
tests/circular/template/test_interpolatedstr.py
from src.circular.template.interpolatedstr import InterpolatedStr from src.circular.template.context import Context from tests.utils import TObserver def test_string_interp(): ctx = Context() ctx.name = "James" s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.") s.bind_ctx(ctx) t = TObserver(s) assert s.value == "My name is , James ." ctx.surname = "Bond" data = t.events.pop().data assert s.value == "My name is Bond, James Bond."
from src.circular.template.interpolatedstr import InterpolatedStr from src.circular.template.context import Context from tests.utils import TObserver def test_string_interp(): ctx = Context() ctx.name = "James" s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.") s.bind_ctx(ctx) t = TObserver(s) assert s.value == "My name is , James ." ctx.surname = "Bond" data = t.events.pop().data assert s.value == "My name is Bond, James Bond." # Should correctly interpolate two immediately succeeding expressions ctx.sur="B" s = InterpolatedStr('{{name}}{{sur}}') s.bind_ctx(ctx) assert s.value == "JamesB"
Add test for run-together interpolated expressions.
Add test for run-together interpolated expressions.
Python
mit
jonathanverner/circular,jonathanverner/circular,jonathanverner/circular
from src.circular.template.interpolatedstr import InterpolatedStr from src.circular.template.context import Context from tests.utils import TObserver def test_string_interp(): ctx = Context() ctx.name = "James" s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.") s.bind_ctx(ctx) t = TObserver(s) assert s.value == "My name is , James ." ctx.surname = "Bond" data = t.events.pop().data assert s.value == "My name is Bond, James Bond." + # Should correctly interpolate two immediately succeeding expressions + ctx.sur="B" + s = InterpolatedStr('{{name}}{{sur}}') + s.bind_ctx(ctx) + assert s.value == "JamesB" +
Add test for run-together interpolated expressions.
## Code Before: from src.circular.template.interpolatedstr import InterpolatedStr from src.circular.template.context import Context from tests.utils import TObserver def test_string_interp(): ctx = Context() ctx.name = "James" s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.") s.bind_ctx(ctx) t = TObserver(s) assert s.value == "My name is , James ." ctx.surname = "Bond" data = t.events.pop().data assert s.value == "My name is Bond, James Bond." ## Instruction: Add test for run-together interpolated expressions. ## Code After: from src.circular.template.interpolatedstr import InterpolatedStr from src.circular.template.context import Context from tests.utils import TObserver def test_string_interp(): ctx = Context() ctx.name = "James" s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.") s.bind_ctx(ctx) t = TObserver(s) assert s.value == "My name is , James ." ctx.surname = "Bond" data = t.events.pop().data assert s.value == "My name is Bond, James Bond." # Should correctly interpolate two immediately succeeding expressions ctx.sur="B" s = InterpolatedStr('{{name}}{{sur}}') s.bind_ctx(ctx) assert s.value == "JamesB"
from src.circular.template.interpolatedstr import InterpolatedStr from src.circular.template.context import Context from tests.utils import TObserver def test_string_interp(): ctx = Context() ctx.name = "James" s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.") s.bind_ctx(ctx) t = TObserver(s) assert s.value == "My name is , James ." ctx.surname = "Bond" data = t.events.pop().data assert s.value == "My name is Bond, James Bond." + # Should correctly interpolate two immediately succeeding expressions + ctx.sur="B" + s = InterpolatedStr('{{name}}{{sur}}') + s.bind_ctx(ctx) + assert s.value == "JamesB" +
a287c1e7a6e96a2a2143e9270a5f9b2ec295022e
fireplace/cards/removed/all.py
fireplace/cards/removed/all.py
from ..utils import * # Adrenaline Rush class NEW1_006: action = drawCard combo = drawCards(2) # Bolstered (Bloodsail Corsair) class NEW1_025e: Health = 1
from ..utils import * # Dagger Mastery class CS2_083: def action(self): if self.hero.weapon: self.hero.weapon.buff("CS2_083e") else: self.hero.summon("CS2_082") class CS2_083e: Atk = 1 # Adrenaline Rush class NEW1_006: action = drawCard combo = drawCards(2) # Bolstered (Bloodsail Corsair) class NEW1_025e: Health = 1
Implement the old Dagger Mastery
Implement the old Dagger Mastery
Python
agpl-3.0
Ragowit/fireplace,beheh/fireplace,NightKev/fireplace,liujimj/fireplace,amw2104/fireplace,butozerca/fireplace,oftc-ftw/fireplace,jleclanche/fireplace,liujimj/fireplace,butozerca/fireplace,oftc-ftw/fireplace,smallnamespace/fireplace,Meerkov/fireplace,smallnamespace/fireplace,amw2104/fireplace,Ragowit/fireplace,Meerkov/fireplace
from ..utils import * + + + # Dagger Mastery + class CS2_083: + def action(self): + if self.hero.weapon: + self.hero.weapon.buff("CS2_083e") + else: + self.hero.summon("CS2_082") + + class CS2_083e: + Atk = 1 # Adrenaline Rush class NEW1_006: action = drawCard combo = drawCards(2) # Bolstered (Bloodsail Corsair) class NEW1_025e: Health = 1
Implement the old Dagger Mastery
## Code Before: from ..utils import * # Adrenaline Rush class NEW1_006: action = drawCard combo = drawCards(2) # Bolstered (Bloodsail Corsair) class NEW1_025e: Health = 1 ## Instruction: Implement the old Dagger Mastery ## Code After: from ..utils import * # Dagger Mastery class CS2_083: def action(self): if self.hero.weapon: self.hero.weapon.buff("CS2_083e") else: self.hero.summon("CS2_082") class CS2_083e: Atk = 1 # Adrenaline Rush class NEW1_006: action = drawCard combo = drawCards(2) # Bolstered (Bloodsail Corsair) class NEW1_025e: Health = 1
from ..utils import * + + + # Dagger Mastery + class CS2_083: + def action(self): + if self.hero.weapon: + self.hero.weapon.buff("CS2_083e") + else: + self.hero.summon("CS2_082") + + class CS2_083e: + Atk = 1 # Adrenaline Rush class NEW1_006: action = drawCard combo = drawCards(2) # Bolstered (Bloodsail Corsair) class NEW1_025e: Health = 1
07ccbc36fd5148db2efc5f676fd13d4b24aa004f
hackasmlexer/hacklexer.py
hackasmlexer/hacklexer.py
import re from pygments.lexer import RegexLexer, include from pygments.token import * class HackAsmLexer(RegexLexer): name = 'Hack Assembler' aliases = ['hack_asm'] filenames = ['*.asm'] identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*' flags = re.IGNORECASE | re.MULTILINE tokens = { 'root': [ include('whitespace'), (r'\(' + identifier + '\)', Name.Label), (r'[+-=;&|!]+', Operator), (r'\/\/.+$', Comment), (r'[\r\n]+', Text), (r'@[A-Za-z][A-Za-z0-9]+', Name.Variable), (r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword), (r'null', Keyword.Pseudo), (r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin), (r'@[0-9]+', Name.Constant) ], 'whitespace': [ (r'\n', Text), (r'\s+', Text), (r'#.*?\n', Comment) ] }
import re from pygments.lexer import RegexLexer, include from pygments.token import * class HackAsmLexer(RegexLexer): name = 'Hack Assembler' aliases = ['hack_asm'] filenames = ['*.asm'] identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*' flags = re.IGNORECASE | re.MULTILINE tokens = { 'root': [ include('whitespace'), (r'\(' + identifier + '\)', Name.Label), (r'[+-=;&|!]+', Operator), (r'\/\/.+$', Comment), (r'[\r\n]+', Text), (r'@[A-Za-z][A-Za-z0-9]+', Name.Variable), (r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword), (r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses (r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses (r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses (r'null', Keyword.Pseudo), (r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin), (r'@[0-9]+', Name.Constant) ], 'whitespace': [ (r'\n', Text), (r'\s+', Text), (r'#.*?\n', Comment) ] }
Add register and IO addresses
Add register and IO addresses
Python
mit
cprieto/pygments_hack_asm
import re from pygments.lexer import RegexLexer, include from pygments.token import * class HackAsmLexer(RegexLexer): name = 'Hack Assembler' aliases = ['hack_asm'] filenames = ['*.asm'] identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*' flags = re.IGNORECASE | re.MULTILINE tokens = { 'root': [ include('whitespace'), (r'\(' + identifier + '\)', Name.Label), (r'[+-=;&|!]+', Operator), (r'\/\/.+$', Comment), (r'[\r\n]+', Text), (r'@[A-Za-z][A-Za-z0-9]+', Name.Variable), (r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword), + (r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses + (r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses + (r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses (r'null', Keyword.Pseudo), (r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin), (r'@[0-9]+', Name.Constant) ], 'whitespace': [ (r'\n', Text), (r'\s+', Text), (r'#.*?\n', Comment) ] }
Add register and IO addresses
## Code Before: import re from pygments.lexer import RegexLexer, include from pygments.token import * class HackAsmLexer(RegexLexer): name = 'Hack Assembler' aliases = ['hack_asm'] filenames = ['*.asm'] identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*' flags = re.IGNORECASE | re.MULTILINE tokens = { 'root': [ include('whitespace'), (r'\(' + identifier + '\)', Name.Label), (r'[+-=;&|!]+', Operator), (r'\/\/.+$', Comment), (r'[\r\n]+', Text), (r'@[A-Za-z][A-Za-z0-9]+', Name.Variable), (r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword), (r'null', Keyword.Pseudo), (r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin), (r'@[0-9]+', Name.Constant) ], 'whitespace': [ (r'\n', Text), (r'\s+', Text), (r'#.*?\n', Comment) ] } ## Instruction: Add register and IO addresses ## Code After: import re from pygments.lexer import RegexLexer, include from pygments.token import * class HackAsmLexer(RegexLexer): name = 'Hack Assembler' aliases = ['hack_asm'] filenames = ['*.asm'] identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*' flags = re.IGNORECASE | re.MULTILINE tokens = { 'root': [ include('whitespace'), (r'\(' + identifier + '\)', Name.Label), (r'[+-=;&|!]+', Operator), (r'\/\/.+$', Comment), (r'[\r\n]+', Text), (r'@[A-Za-z][A-Za-z0-9]+', Name.Variable), (r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword), (r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses (r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses (r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses (r'null', Keyword.Pseudo), (r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin), (r'@[0-9]+', Name.Constant) ], 'whitespace': [ (r'\n', Text), (r'\s+', Text), (r'#.*?\n', Comment) ] }
import re from pygments.lexer import RegexLexer, include from pygments.token import * class HackAsmLexer(RegexLexer): name = 'Hack Assembler' aliases = ['hack_asm'] filenames = ['*.asm'] identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*' flags = re.IGNORECASE | re.MULTILINE tokens = { 'root': [ include('whitespace'), (r'\(' + identifier + '\)', Name.Label), (r'[+-=;&|!]+', Operator), (r'\/\/.+$', Comment), (r'[\r\n]+', Text), (r'@[A-Za-z][A-Za-z0-9]+', Name.Variable), (r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword), + (r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses + (r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses + (r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses (r'null', Keyword.Pseudo), (r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin), (r'@[0-9]+', Name.Constant) ], 'whitespace': [ (r'\n', Text), (r'\s+', Text), (r'#.*?\n', Comment) ] }
8570efd42f35b89d9a97d9aa5a5aa47765cd21f6
diary/logthread.py
diary/logthread.py
from threading import Thread try: from queue import Queue except ImportError: # python 2 from Queue import Queue class ElemThread(Thread): """A thread for logging as to not disrupt the logged application""" def __init__(self, elem, name="Elementary Logger"): """Construct a thread for logging :param elem: An Elementary instance to handle logging :param name: A string to represent this thread """ Thread.__init__(self, name=name) self.daemon = True # py2 constructor requires explicit self.elem = elem self.queue = Queue() self.start() def add(self, event): """Add a logged event to queue for logging""" self.queue.put(event) def run(self): """Main for thread to run""" while True: self.elem.write(self.queue.get())
from threading import Thread try: from queue import Queue except ImportError: # python 2 from Queue import Queue class DiaryThread(Thread): """A thread for logging as to not disrupt the logged application""" def __init__(self, diary, name="Diary Logger"): """Construct a thread for logging :param diary: An Diary instance to handle logging :param name: A string to represent this thread """ Thread.__init__(self, name=name) self.daemon = True # py2 constructor requires explicit self.diary = diary self.queue = Queue() self.start() def add(self, event): """Add a logged event to queue for logging""" self.queue.put(event) def run(self): """Main for thread to run""" while True: self.diary.write(self.queue.get())
Make last changes over to diary name
Make last changes over to diary name
Python
mit
GreenVars/diary
from threading import Thread try: from queue import Queue except ImportError: # python 2 from Queue import Queue - class ElemThread(Thread): + class DiaryThread(Thread): """A thread for logging as to not disrupt the logged application""" - def __init__(self, elem, name="Elementary Logger"): + def __init__(self, diary, name="Diary Logger"): """Construct a thread for logging - :param elem: An Elementary instance to handle logging + :param diary: An Diary instance to handle logging :param name: A string to represent this thread """ Thread.__init__(self, name=name) self.daemon = True # py2 constructor requires explicit - self.elem = elem + self.diary = diary self.queue = Queue() self.start() def add(self, event): """Add a logged event to queue for logging""" self.queue.put(event) def run(self): """Main for thread to run""" while True: - self.elem.write(self.queue.get()) + self.diary.write(self.queue.get())
Make last changes over to diary name
## Code Before: from threading import Thread try: from queue import Queue except ImportError: # python 2 from Queue import Queue class ElemThread(Thread): """A thread for logging as to not disrupt the logged application""" def __init__(self, elem, name="Elementary Logger"): """Construct a thread for logging :param elem: An Elementary instance to handle logging :param name: A string to represent this thread """ Thread.__init__(self, name=name) self.daemon = True # py2 constructor requires explicit self.elem = elem self.queue = Queue() self.start() def add(self, event): """Add a logged event to queue for logging""" self.queue.put(event) def run(self): """Main for thread to run""" while True: self.elem.write(self.queue.get()) ## Instruction: Make last changes over to diary name ## Code After: from threading import Thread try: from queue import Queue except ImportError: # python 2 from Queue import Queue class DiaryThread(Thread): """A thread for logging as to not disrupt the logged application""" def __init__(self, diary, name="Diary Logger"): """Construct a thread for logging :param diary: An Diary instance to handle logging :param name: A string to represent this thread """ Thread.__init__(self, name=name) self.daemon = True # py2 constructor requires explicit self.diary = diary self.queue = Queue() self.start() def add(self, event): """Add a logged event to queue for logging""" self.queue.put(event) def run(self): """Main for thread to run""" while True: self.diary.write(self.queue.get())
from threading import Thread try: from queue import Queue except ImportError: # python 2 from Queue import Queue - class ElemThread(Thread): ? ^^^^ + class DiaryThread(Thread): ? ^^^^^ """A thread for logging as to not disrupt the logged application""" - def __init__(self, elem, name="Elementary Logger"): ? ^^^^ ^^^^^^^ + def __init__(self, diary, name="Diary Logger"): ? ^^^^^ ^^ """Construct a thread for logging - :param elem: An Elementary instance to handle logging ? ^^^^ ^^^^^^^ + :param diary: An Diary instance to handle logging ? ^^^^^ ^^ :param name: A string to represent this thread """ Thread.__init__(self, name=name) self.daemon = True # py2 constructor requires explicit - self.elem = elem + self.diary = diary self.queue = Queue() self.start() def add(self, event): """Add a logged event to queue for logging""" self.queue.put(event) def run(self): """Main for thread to run""" while True: - self.elem.write(self.queue.get()) ? ^^^^ + self.diary.write(self.queue.get()) ? ^^^^^
59daf205869c42b3797aa9dbaaa97930cbca2417
nanshe_workflow/ipy.py
nanshe_workflow/ipy.py
__author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>" __date__ = "$Nov 10, 2015 17:09$" try: from IPython.utils.shimmodule import ShimWarning except ImportError: class ShimWarning(Warning): """Warning issued by IPython 4.x regarding deprecated API.""" pass import warnings with warnings.catch_warnings(): warnings.filterwarnings('error', '', ShimWarning) try: # IPython 3 from IPython.html.widgets import FloatProgress from IPython.parallel import Client except ShimWarning: # IPython 4 from ipywidgets import FloatProgress from ipyparallel import Client from IPython.display import display
__author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>" __date__ = "$Nov 10, 2015 17:09$" import json import re try: from IPython.utils.shimmodule import ShimWarning except ImportError: class ShimWarning(Warning): """Warning issued by IPython 4.x regarding deprecated API.""" pass import warnings with warnings.catch_warnings(): warnings.filterwarnings('error', '', ShimWarning) try: # IPython 3 from IPython.html.widgets import FloatProgress from IPython.parallel import Client except ShimWarning: # IPython 4 from ipywidgets import FloatProgress from ipyparallel import Client from IPython.display import display import ipykernel import notebook.notebookapp import requests def check_nbserverproxy(): """ Return the url of the current jupyter notebook server. """ kernel_id = re.search( "kernel-(.*).json", ipykernel.connect.get_connection_file() ).group(1) servers = notebook.notebookapp.list_running_servers() for s in servers: response = requests.get( requests.compat.urljoin(s["url"], "api/sessions"), params={"token": s.get("token", "")} ) for n in json.loads(response.text): if n["kernel"]["id"] == kernel_id: # Found server that is running this Jupyter Notebook. # Try to requests this servers port through nbserverproxy. url = requests.compat.urljoin( s["url"], "proxy/%i" % s["port"] ) # If the proxy is running, it will redirect. # If not, it will error out. try: requests.get(url).raise_for_status() except requests.HTTPError: return False else: return True
Add function to check if nbserverproxy is running
Add function to check if nbserverproxy is running Provides a simple check to see if the `nbserverproxy` is installed and running. As this is a Jupyter server extension and this code is run from the notebook, we can't simply import `nbserverproxy`. In fact that wouldn't even work when using the Python 2 kernel even though the proxy server could be running. Instead to solve this problem try to identify the Jupyter Notebook server we are running under. Once identified, attempt to query the proxy server with the port of the Jupyter Notebook server. If the proxy server is running, this will merely redirect to the Jupyter Notebook server and return an HTTP 200 status. However if the proxy server is not running, this will return a HTTP 404 error. There may be other errors that it could raise. In any event, if the proxy redirects us, we know it is working and if not we know it doesn't work.
Python
apache-2.0
nanshe-org/nanshe_workflow,DudLab/nanshe_workflow
__author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>" __date__ = "$Nov 10, 2015 17:09$" + + import json + import re try: from IPython.utils.shimmodule import ShimWarning except ImportError: class ShimWarning(Warning): """Warning issued by IPython 4.x regarding deprecated API.""" pass import warnings with warnings.catch_warnings(): warnings.filterwarnings('error', '', ShimWarning) try: # IPython 3 from IPython.html.widgets import FloatProgress from IPython.parallel import Client except ShimWarning: # IPython 4 from ipywidgets import FloatProgress from ipyparallel import Client from IPython.display import display + import ipykernel + import notebook.notebookapp + + import requests + + + def check_nbserverproxy(): + """ + Return the url of the current jupyter notebook server. + """ + kernel_id = re.search( + "kernel-(.*).json", + ipykernel.connect.get_connection_file() + ).group(1) + servers = notebook.notebookapp.list_running_servers() + for s in servers: + response = requests.get( + requests.compat.urljoin(s["url"], "api/sessions"), + params={"token": s.get("token", "")} + ) + for n in json.loads(response.text): + if n["kernel"]["id"] == kernel_id: + # Found server that is running this Jupyter Notebook. + # Try to requests this servers port through nbserverproxy. + url = requests.compat.urljoin( + s["url"], "proxy/%i" % s["port"] + ) + # If the proxy is running, it will redirect. + # If not, it will error out. + try: + requests.get(url).raise_for_status() + except requests.HTTPError: + return False + else: + return True +
Add function to check if nbserverproxy is running
## Code Before: __author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>" __date__ = "$Nov 10, 2015 17:09$" try: from IPython.utils.shimmodule import ShimWarning except ImportError: class ShimWarning(Warning): """Warning issued by IPython 4.x regarding deprecated API.""" pass import warnings with warnings.catch_warnings(): warnings.filterwarnings('error', '', ShimWarning) try: # IPython 3 from IPython.html.widgets import FloatProgress from IPython.parallel import Client except ShimWarning: # IPython 4 from ipywidgets import FloatProgress from ipyparallel import Client from IPython.display import display ## Instruction: Add function to check if nbserverproxy is running ## Code After: __author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>" __date__ = "$Nov 10, 2015 17:09$" import json import re try: from IPython.utils.shimmodule import ShimWarning except ImportError: class ShimWarning(Warning): """Warning issued by IPython 4.x regarding deprecated API.""" pass import warnings with warnings.catch_warnings(): warnings.filterwarnings('error', '', ShimWarning) try: # IPython 3 from IPython.html.widgets import FloatProgress from IPython.parallel import Client except ShimWarning: # IPython 4 from ipywidgets import FloatProgress from ipyparallel import Client from IPython.display import display import ipykernel import notebook.notebookapp import requests def check_nbserverproxy(): """ Return the url of the current jupyter notebook server. """ kernel_id = re.search( "kernel-(.*).json", ipykernel.connect.get_connection_file() ).group(1) servers = notebook.notebookapp.list_running_servers() for s in servers: response = requests.get( requests.compat.urljoin(s["url"], "api/sessions"), params={"token": s.get("token", "")} ) for n in json.loads(response.text): if n["kernel"]["id"] == kernel_id: # Found server that is running this Jupyter Notebook. # Try to requests this servers port through nbserverproxy. url = requests.compat.urljoin( s["url"], "proxy/%i" % s["port"] ) # If the proxy is running, it will redirect. # If not, it will error out. try: requests.get(url).raise_for_status() except requests.HTTPError: return False else: return True
__author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>" __date__ = "$Nov 10, 2015 17:09$" + + import json + import re try: from IPython.utils.shimmodule import ShimWarning except ImportError: class ShimWarning(Warning): """Warning issued by IPython 4.x regarding deprecated API.""" pass import warnings with warnings.catch_warnings(): warnings.filterwarnings('error', '', ShimWarning) try: # IPython 3 from IPython.html.widgets import FloatProgress from IPython.parallel import Client except ShimWarning: # IPython 4 from ipywidgets import FloatProgress from ipyparallel import Client from IPython.display import display + + import ipykernel + import notebook.notebookapp + + import requests + + + def check_nbserverproxy(): + """ + Return the url of the current jupyter notebook server. + """ + kernel_id = re.search( + "kernel-(.*).json", + ipykernel.connect.get_connection_file() + ).group(1) + servers = notebook.notebookapp.list_running_servers() + for s in servers: + response = requests.get( + requests.compat.urljoin(s["url"], "api/sessions"), + params={"token": s.get("token", "")} + ) + for n in json.loads(response.text): + if n["kernel"]["id"] == kernel_id: + # Found server that is running this Jupyter Notebook. + # Try to requests this servers port through nbserverproxy. + url = requests.compat.urljoin( + s["url"], "proxy/%i" % s["port"] + ) + # If the proxy is running, it will redirect. + # If not, it will error out. + try: + requests.get(url).raise_for_status() + except requests.HTTPError: + return False + else: + return True
eeac557b77a3a63a3497791a2716706801b20e37
kodos/main.py
kodos/main.py
def run(args=None): """Main entry point of the application.""" pass
import sys from PyQt4.QtGui import QApplication, QMainWindow from kodos.ui.ui_main import Ui_MainWindow class KodosMainWindow(QMainWindow, Ui_MainWindow): def __init__(self, parent=None): super(KodosMainWindow, self).__init__(parent) self.setupUi(self) self.connectActions() # Trigger the textChanged signal for widget in [self.regexText, self.searchText, self.replaceText]: widget.setPlainText('') def connectActions(self): # Connect input widgets to update the GUI when their text change for widget in [self.regexText, self.searchText, self.replaceText]: widget.textChanged.connect(self.on_compute_regex) def on_compute_regex(self): regex = self.regexText.toPlainText() search = self.searchText.toPlainText() replace = self.replaceText.toPlainText() if regex == "" or search == "": self.statusbar.showMessage( "Please enter a regex and a search to work on") else: self.statusbar.clearMessage() def run(args=None): """Main entry point of the application.""" app = QApplication(sys.argv) kodos = KodosMainWindow() kodos.show() app.exec_()
Connect the UI to the code and start to connect slots to actions.
Connect the UI to the code and start to connect slots to actions.
Python
bsd-2-clause
multani/kodos-qt4
+ import sys + from PyQt4.QtGui import QApplication, QMainWindow + + from kodos.ui.ui_main import Ui_MainWindow + + + class KodosMainWindow(QMainWindow, Ui_MainWindow): + def __init__(self, parent=None): + super(KodosMainWindow, self).__init__(parent) + self.setupUi(self) + self.connectActions() + + # Trigger the textChanged signal + for widget in [self.regexText, self.searchText, self.replaceText]: + widget.setPlainText('') + + def connectActions(self): + # Connect input widgets to update the GUI when their text change + for widget in [self.regexText, self.searchText, self.replaceText]: + widget.textChanged.connect(self.on_compute_regex) + + def on_compute_regex(self): + regex = self.regexText.toPlainText() + search = self.searchText.toPlainText() + replace = self.replaceText.toPlainText() + + if regex == "" or search == "": + self.statusbar.showMessage( + "Please enter a regex and a search to work on") + else: + self.statusbar.clearMessage() def run(args=None): """Main entry point of the application.""" - pass + app = QApplication(sys.argv) + kodos = KodosMainWindow() + kodos.show() + app.exec_() +
Connect the UI to the code and start to connect slots to actions.
## Code Before: def run(args=None): """Main entry point of the application.""" pass ## Instruction: Connect the UI to the code and start to connect slots to actions. ## Code After: import sys from PyQt4.QtGui import QApplication, QMainWindow from kodos.ui.ui_main import Ui_MainWindow class KodosMainWindow(QMainWindow, Ui_MainWindow): def __init__(self, parent=None): super(KodosMainWindow, self).__init__(parent) self.setupUi(self) self.connectActions() # Trigger the textChanged signal for widget in [self.regexText, self.searchText, self.replaceText]: widget.setPlainText('') def connectActions(self): # Connect input widgets to update the GUI when their text change for widget in [self.regexText, self.searchText, self.replaceText]: widget.textChanged.connect(self.on_compute_regex) def on_compute_regex(self): regex = self.regexText.toPlainText() search = self.searchText.toPlainText() replace = self.replaceText.toPlainText() if regex == "" or search == "": self.statusbar.showMessage( "Please enter a regex and a search to work on") else: self.statusbar.clearMessage() def run(args=None): """Main entry point of the application.""" app = QApplication(sys.argv) kodos = KodosMainWindow() kodos.show() app.exec_()
+ import sys + from PyQt4.QtGui import QApplication, QMainWindow + + from kodos.ui.ui_main import Ui_MainWindow + + + class KodosMainWindow(QMainWindow, Ui_MainWindow): + def __init__(self, parent=None): + super(KodosMainWindow, self).__init__(parent) + self.setupUi(self) + self.connectActions() + + # Trigger the textChanged signal + for widget in [self.regexText, self.searchText, self.replaceText]: + widget.setPlainText('') + + def connectActions(self): + # Connect input widgets to update the GUI when their text change + for widget in [self.regexText, self.searchText, self.replaceText]: + widget.textChanged.connect(self.on_compute_regex) + + def on_compute_regex(self): + regex = self.regexText.toPlainText() + search = self.searchText.toPlainText() + replace = self.replaceText.toPlainText() + + if regex == "" or search == "": + self.statusbar.showMessage( + "Please enter a regex and a search to work on") + else: + self.statusbar.clearMessage() def run(args=None): """Main entry point of the application.""" - pass + + app = QApplication(sys.argv) + kodos = KodosMainWindow() + kodos.show() + app.exec_()
7aa89902f8af2ca1f4b3c9e356a62062cc74696b
bot/anime_searcher.py
bot/anime_searcher.py
from itertools import chain from typing import Iterable from minoshiro import Medium, Minoshiro, Site from minoshiro.helpers import get_synonyms class AnimeSearcher(Minoshiro): async def get(self, query: str, medium: Medium, sites: Iterable[Site] = None, *, timeout=3): sites = sites if sites else list(Site) cached_data, cached_id = await self.__get_cached(query, medium) to_be_cached = {} names = [] return_val = {} for site in sites: res, id_ = await self.__get_result( cached_data, cached_id, query, names, site, medium, timeout ) if res: return_val[site] = res for title in get_synonyms(res, site): names.append(title) if id_: to_be_cached[site] = id_ return return_val, to_be_cached, names, medium async def cache(self, to_be_cached, names, medium): """ Cache search results into the db. :param to_be_cached: items to be cached. :param names: all names for the item. :param medium: the medium type. """ itere = set(chain(*names)) for site, id_ in to_be_cached.items(): await self.cache_one(site, id_, medium, itere) async def cache_one(self, site, id_, medium, iterator): """ Cache one id. :param site: the site. :param id_: the id. :param medium: the medium type. :param iterator: an iterator for all names. """ for name in iterator: if name: await self.db_controller.set_identifier( name, medium, site, id_ )
from typing import Iterable from minoshiro import Medium, Minoshiro, Site from minoshiro.helpers import get_synonyms class AnimeSearcher(Minoshiro): async def get(self, query: str, medium: Medium, sites: Iterable[Site] = None, *, timeout=3): sites = sites if sites else list(Site) cached_data, cached_id = await self._get_cached(query, medium) to_be_cached = {} names = [] return_val = {} for site in sites: res, id_ = await self._get_result( cached_data, cached_id, query, names, site, medium, timeout ) if res: return_val[site] = res for title in get_synonyms(res, site): names.append(title) if id_: to_be_cached[site] = id_ return return_val, to_be_cached, names, medium async def cache(self, to_be_cached, names, medium): await super()._cache(to_be_cached, names, medium)
Update anime searcher implementation to use super class methods
Update anime searcher implementation to use super class methods
Python
apache-2.0
MaT1g3R/YasenBaka
- from itertools import chain from typing import Iterable from minoshiro import Medium, Minoshiro, Site from minoshiro.helpers import get_synonyms class AnimeSearcher(Minoshiro): async def get(self, query: str, medium: Medium, sites: Iterable[Site] = None, *, timeout=3): sites = sites if sites else list(Site) - cached_data, cached_id = await self.__get_cached(query, medium) + cached_data, cached_id = await self._get_cached(query, medium) to_be_cached = {} names = [] return_val = {} for site in sites: - res, id_ = await self.__get_result( + res, id_ = await self._get_result( cached_data, cached_id, query, names, site, medium, timeout ) if res: return_val[site] = res for title in get_synonyms(res, site): names.append(title) if id_: to_be_cached[site] = id_ return return_val, to_be_cached, names, medium async def cache(self, to_be_cached, names, medium): + await super()._cache(to_be_cached, names, medium) - """ - Cache search results into the db. - :param to_be_cached: items to be cached. - :param names: all names for the item. - :param medium: the medium type. - """ - itere = set(chain(*names)) - for site, id_ in to_be_cached.items(): - await self.cache_one(site, id_, medium, itere) - async def cache_one(self, site, id_, medium, iterator): - """ - Cache one id. - :param site: the site. - :param id_: the id. - :param medium: the medium type. - :param iterator: an iterator for all names. - """ - for name in iterator: - if name: - await self.db_controller.set_identifier( - name, medium, site, id_ - ) -
Update anime searcher implementation to use super class methods
## Code Before: from itertools import chain from typing import Iterable from minoshiro import Medium, Minoshiro, Site from minoshiro.helpers import get_synonyms class AnimeSearcher(Minoshiro): async def get(self, query: str, medium: Medium, sites: Iterable[Site] = None, *, timeout=3): sites = sites if sites else list(Site) cached_data, cached_id = await self.__get_cached(query, medium) to_be_cached = {} names = [] return_val = {} for site in sites: res, id_ = await self.__get_result( cached_data, cached_id, query, names, site, medium, timeout ) if res: return_val[site] = res for title in get_synonyms(res, site): names.append(title) if id_: to_be_cached[site] = id_ return return_val, to_be_cached, names, medium async def cache(self, to_be_cached, names, medium): """ Cache search results into the db. :param to_be_cached: items to be cached. :param names: all names for the item. :param medium: the medium type. """ itere = set(chain(*names)) for site, id_ in to_be_cached.items(): await self.cache_one(site, id_, medium, itere) async def cache_one(self, site, id_, medium, iterator): """ Cache one id. :param site: the site. :param id_: the id. :param medium: the medium type. :param iterator: an iterator for all names. """ for name in iterator: if name: await self.db_controller.set_identifier( name, medium, site, id_ ) ## Instruction: Update anime searcher implementation to use super class methods ## Code After: from typing import Iterable from minoshiro import Medium, Minoshiro, Site from minoshiro.helpers import get_synonyms class AnimeSearcher(Minoshiro): async def get(self, query: str, medium: Medium, sites: Iterable[Site] = None, *, timeout=3): sites = sites if sites else list(Site) cached_data, cached_id = await self._get_cached(query, medium) to_be_cached = {} names = [] return_val = {} for site in sites: res, id_ = await self._get_result( cached_data, cached_id, query, names, site, medium, timeout ) if res: return_val[site] = res for title in get_synonyms(res, site): names.append(title) if id_: to_be_cached[site] = id_ return return_val, to_be_cached, names, medium async def cache(self, to_be_cached, names, medium): await super()._cache(to_be_cached, names, medium)
- from itertools import chain from typing import Iterable from minoshiro import Medium, Minoshiro, Site from minoshiro.helpers import get_synonyms class AnimeSearcher(Minoshiro): async def get(self, query: str, medium: Medium, sites: Iterable[Site] = None, *, timeout=3): sites = sites if sites else list(Site) - cached_data, cached_id = await self.__get_cached(query, medium) ? - + cached_data, cached_id = await self._get_cached(query, medium) to_be_cached = {} names = [] return_val = {} for site in sites: - res, id_ = await self.__get_result( ? - + res, id_ = await self._get_result( cached_data, cached_id, query, names, site, medium, timeout ) if res: return_val[site] = res for title in get_synonyms(res, site): names.append(title) if id_: to_be_cached[site] = id_ return return_val, to_be_cached, names, medium async def cache(self, to_be_cached, names, medium): + await super()._cache(to_be_cached, names, medium) - """ - Cache search results into the db. - :param to_be_cached: items to be cached. - :param names: all names for the item. - :param medium: the medium type. - """ - itere = set(chain(*names)) - for site, id_ in to_be_cached.items(): - await self.cache_one(site, id_, medium, itere) - - async def cache_one(self, site, id_, medium, iterator): - """ - Cache one id. - :param site: the site. - :param id_: the id. - :param medium: the medium type. - :param iterator: an iterator for all names. - """ - for name in iterator: - if name: - await self.db_controller.set_identifier( - name, medium, site, id_ - )
930ad35cee818e2d0b97f840ff0b3b772bd51af3
post_office/management/commands/send_queued_mail.py
post_office/management/commands/send_queued_mail.py
import tempfile import sys from optparse import make_option from django.core.management.base import BaseCommand from ...lockfile import FileLock, FileLocked from ...mail import send_queued from ...logutils import setup_loghandlers logger = setup_loghandlers() default_lockfile = tempfile.gettempdir() + "/post_office" class Command(BaseCommand): option_list = BaseCommand.option_list + ( make_option('-p', '--processes', type='int', help='Number of processes used to send emails', default=1), make_option('-L', '--lockfile', type='string', default=default_lockfile, help='Absolute path of lockfile to acquire'), make_option('-l', '--log-level', type='int', help='"0" to log nothing, "1" to only log errors'), ) def handle(self, *args, **options): logger.info('Acquiring lock for sending queued emails at %s.lock' % options['lockfile']) try: with FileLock(options['lockfile']): try: send_queued(options['processes'], options.get('log_level')) except Exception as e: logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500}) raise except FileLocked: logger.info('Failed to acquire lock, terminating now.')
import tempfile import sys from django.core.management.base import BaseCommand from ...lockfile import FileLock, FileLocked from ...mail import send_queued from ...logutils import setup_loghandlers logger = setup_loghandlers() default_lockfile = tempfile.gettempdir() + "/post_office" class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument('-p', '--processes', type='int', help='Number of processes used to send emails', default=1 ) parser.add_argument('-L', '--lockfile', type='string', default=default_lockfile, help='Absolute path of lockfile to acquire' ) parser.add_argument('-l', '--log-level', type='int', help='"0" to log nothing, "1" to only log errors' ) def handle(self, *args, **options): logger.info('Acquiring lock for sending queued emails at %s.lock' % options['lockfile']) try: with FileLock(options['lockfile']): try: send_queued(options['processes'], options.get('log_level')) except Exception as e: logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500}) raise except FileLocked: logger.info('Failed to acquire lock, terminating now.')
Switch to using the `add_arguments` method.
Switch to using the `add_arguments` method. This is an alternative to using the `option_list` and `optparse.make_option`. Django deprecated the use of `optparse` in management commands in Django 1.8 and removed it in Django 1.10.
Python
mit
jrief/django-post_office,ui/django-post_office,ui/django-post_office
import tempfile import sys - from optparse import make_option from django.core.management.base import BaseCommand from ...lockfile import FileLock, FileLocked from ...mail import send_queued from ...logutils import setup_loghandlers logger = setup_loghandlers() default_lockfile = tempfile.gettempdir() + "/post_office" class Command(BaseCommand): - - option_list = BaseCommand.option_list + ( - make_option('-p', '--processes', type='int', + def add_arguments(self, parser): + parser.add_argument('-p', '--processes', + type='int', - help='Number of processes used to send emails', default=1), + help='Number of processes used to send emails', - make_option('-L', '--lockfile', type='string', default=default_lockfile, + default=1 + ) + parser.add_argument('-L', '--lockfile', + type='string', + default=default_lockfile, - help='Absolute path of lockfile to acquire'), + help='Absolute path of lockfile to acquire' - make_option('-l', '--log-level', type='int', + ) + parser.add_argument('-l', '--log-level', + type='int', - help='"0" to log nothing, "1" to only log errors'), + help='"0" to log nothing, "1" to only log errors' - ) + ) def handle(self, *args, **options): logger.info('Acquiring lock for sending queued emails at %s.lock' % options['lockfile']) try: with FileLock(options['lockfile']): try: send_queued(options['processes'], options.get('log_level')) except Exception as e: logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500}) raise except FileLocked: logger.info('Failed to acquire lock, terminating now.')
Switch to using the `add_arguments` method.
## Code Before: import tempfile import sys from optparse import make_option from django.core.management.base import BaseCommand from ...lockfile import FileLock, FileLocked from ...mail import send_queued from ...logutils import setup_loghandlers logger = setup_loghandlers() default_lockfile = tempfile.gettempdir() + "/post_office" class Command(BaseCommand): option_list = BaseCommand.option_list + ( make_option('-p', '--processes', type='int', help='Number of processes used to send emails', default=1), make_option('-L', '--lockfile', type='string', default=default_lockfile, help='Absolute path of lockfile to acquire'), make_option('-l', '--log-level', type='int', help='"0" to log nothing, "1" to only log errors'), ) def handle(self, *args, **options): logger.info('Acquiring lock for sending queued emails at %s.lock' % options['lockfile']) try: with FileLock(options['lockfile']): try: send_queued(options['processes'], options.get('log_level')) except Exception as e: logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500}) raise except FileLocked: logger.info('Failed to acquire lock, terminating now.') ## Instruction: Switch to using the `add_arguments` method. ## Code After: import tempfile import sys from django.core.management.base import BaseCommand from ...lockfile import FileLock, FileLocked from ...mail import send_queued from ...logutils import setup_loghandlers logger = setup_loghandlers() default_lockfile = tempfile.gettempdir() + "/post_office" class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument('-p', '--processes', type='int', help='Number of processes used to send emails', default=1 ) parser.add_argument('-L', '--lockfile', type='string', default=default_lockfile, help='Absolute path of lockfile to acquire' ) parser.add_argument('-l', '--log-level', type='int', help='"0" to log nothing, "1" to only log errors' ) def handle(self, *args, **options): logger.info('Acquiring lock for sending queued emails at %s.lock' % options['lockfile']) try: with FileLock(options['lockfile']): try: send_queued(options['processes'], options.get('log_level')) except Exception as e: logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500}) raise except FileLocked: logger.info('Failed to acquire lock, terminating now.')
import tempfile import sys - from optparse import make_option from django.core.management.base import BaseCommand from ...lockfile import FileLock, FileLocked from ...mail import send_queued from ...logutils import setup_loghandlers logger = setup_loghandlers() default_lockfile = tempfile.gettempdir() + "/post_office" class Command(BaseCommand): - - option_list = BaseCommand.option_list + ( - make_option('-p', '--processes', type='int', + def add_arguments(self, parser): + parser.add_argument('-p', '--processes', + type='int', - help='Number of processes used to send emails', default=1), ? -------- ------------ + help='Number of processes used to send emails', - make_option('-L', '--lockfile', type='string', default=default_lockfile, + default=1 + ) + parser.add_argument('-L', '--lockfile', + type='string', + default=default_lockfile, - help='Absolute path of lockfile to acquire'), ? -------- -- + help='Absolute path of lockfile to acquire' - make_option('-l', '--log-level', type='int', + ) + parser.add_argument('-l', '--log-level', + type='int', - help='"0" to log nothing, "1" to only log errors'), ? -------- -- + help='"0" to log nothing, "1" to only log errors' - ) + ) def handle(self, *args, **options): logger.info('Acquiring lock for sending queued emails at %s.lock' % options['lockfile']) try: with FileLock(options['lockfile']): try: send_queued(options['processes'], options.get('log_level')) except Exception as e: logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500}) raise except FileLocked: logger.info('Failed to acquire lock, terminating now.')
e2be9eb27d6fc7cfa424cbf908347796ab595526
groundstation/broadcast_announcer.py
groundstation/broadcast_announcer.py
import socket import logger from groundstation.broadcast_socket import BroadcastSocket import logger log = logger.getLogger(__name__) class BroadcastAnnouncer(BroadcastSocket): def __init__(self, port): super(BroadcastAnnouncer, self).__init__() self._addr = '255.255.255.255', port self._name = None self.broadcast_payload = "PING None" @property def name(self): return self._name @name.setter def name(self, value): self._name = value self.broadcast_payload = "PING %s" % (self._name) def ping(self): log.info("ping payload: %s" % (self.broadcast_payload)) transmitted = self.socket.sendto(self.broadcast_payload, self._addr) if transmitted != len(self.broadcast_payload): log.warning("ping wasn't successfully broadcast")
import socket import logger from sockets.broadcast_socket import BroadcastSocket import logger log = logger.getLogger(__name__) class BroadcastAnnouncer(BroadcastSocket): def __init__(self, port): super(BroadcastAnnouncer, self).__init__() self._addr = '255.255.255.255', port self._name = None self.broadcast_payload = "PING None" @property def name(self): return self._name @name.setter def name(self, value): self._name = value self.broadcast_payload = "PING %s" % (self._name) def ping(self): log.info("ping payload: %s" % (self.broadcast_payload)) transmitted = self.socket.sendto(self.broadcast_payload, self._addr) if transmitted != len(self.broadcast_payload): log.warning("ping wasn't successfully broadcast")
Fix an import path bug masked by remaining .pyc files
Fix an import path bug masked by remaining .pyc files
Python
mit
richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation
import socket import logger - from groundstation.broadcast_socket import BroadcastSocket + from sockets.broadcast_socket import BroadcastSocket import logger log = logger.getLogger(__name__) class BroadcastAnnouncer(BroadcastSocket): def __init__(self, port): super(BroadcastAnnouncer, self).__init__() self._addr = '255.255.255.255', port self._name = None self.broadcast_payload = "PING None" @property def name(self): return self._name @name.setter def name(self, value): self._name = value self.broadcast_payload = "PING %s" % (self._name) def ping(self): log.info("ping payload: %s" % (self.broadcast_payload)) transmitted = self.socket.sendto(self.broadcast_payload, self._addr) if transmitted != len(self.broadcast_payload): log.warning("ping wasn't successfully broadcast")
Fix an import path bug masked by remaining .pyc files
## Code Before: import socket import logger from groundstation.broadcast_socket import BroadcastSocket import logger log = logger.getLogger(__name__) class BroadcastAnnouncer(BroadcastSocket): def __init__(self, port): super(BroadcastAnnouncer, self).__init__() self._addr = '255.255.255.255', port self._name = None self.broadcast_payload = "PING None" @property def name(self): return self._name @name.setter def name(self, value): self._name = value self.broadcast_payload = "PING %s" % (self._name) def ping(self): log.info("ping payload: %s" % (self.broadcast_payload)) transmitted = self.socket.sendto(self.broadcast_payload, self._addr) if transmitted != len(self.broadcast_payload): log.warning("ping wasn't successfully broadcast") ## Instruction: Fix an import path bug masked by remaining .pyc files ## Code After: import socket import logger from sockets.broadcast_socket import BroadcastSocket import logger log = logger.getLogger(__name__) class BroadcastAnnouncer(BroadcastSocket): def __init__(self, port): super(BroadcastAnnouncer, self).__init__() self._addr = '255.255.255.255', port self._name = None self.broadcast_payload = "PING None" @property def name(self): return self._name @name.setter def name(self, value): self._name = value self.broadcast_payload = "PING %s" % (self._name) def ping(self): log.info("ping payload: %s" % (self.broadcast_payload)) transmitted = self.socket.sendto(self.broadcast_payload, self._addr) if transmitted != len(self.broadcast_payload): log.warning("ping wasn't successfully broadcast")
import socket import logger - from groundstation.broadcast_socket import BroadcastSocket ? ^^ ^^^ ------ + from sockets.broadcast_socket import BroadcastSocket ? ^ ^^^^ import logger log = logger.getLogger(__name__) class BroadcastAnnouncer(BroadcastSocket): def __init__(self, port): super(BroadcastAnnouncer, self).__init__() self._addr = '255.255.255.255', port self._name = None self.broadcast_payload = "PING None" @property def name(self): return self._name @name.setter def name(self, value): self._name = value self.broadcast_payload = "PING %s" % (self._name) def ping(self): log.info("ping payload: %s" % (self.broadcast_payload)) transmitted = self.socket.sendto(self.broadcast_payload, self._addr) if transmitted != len(self.broadcast_payload): log.warning("ping wasn't successfully broadcast")
608325c33cb2d446b89c263ba0bb02ced5c4ffe8
portal/views.py
portal/views.py
import csv from django.shortcuts import render from django.http import HttpResponse from . import services def index(request): data = services.overview() return render(request, 'index.html', data) def meter_runs(request): """Render the table of exported MeterRun results in html""" data = services.meterruns_export() return render(request, 'project_run_table.html', data) def csv_export(request): """Return a dump of all the MeterRuns in CSV form""" response = HttpResponse(content_type="text/csv") response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"' writer = csv.writer(response) writer.writerow(['First row', 'Foo', 'Bar']) return response
import csv from django.shortcuts import render from django.http import HttpResponse from . import services def index(request): data = services.overview() return render(request, 'index.html', data) def meter_runs(request): """Render the table of exported MeterRun results in html""" data = services.meterruns_export() return render(request, 'project_run_table.html', data) def csv_export(request): """Return a dump of all the MeterRuns in CSV form""" response = HttpResponse(content_type="text/csv") response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"' data = services.meterruns_export() writer = csv.DictWriter(response, fieldnames=data['headers']) writer.writeheader() for meter_run in data['meter_runs']: writer.writerow(meter_run) return response
Use the meterrun_export service to power csv export
Use the meterrun_export service to power csv export
Python
mit
impactlab/oeem-energy-datastore,impactlab/oeem-energy-datastore,impactlab/oeem-energy-datastore
import csv from django.shortcuts import render from django.http import HttpResponse from . import services def index(request): data = services.overview() return render(request, 'index.html', data) def meter_runs(request): """Render the table of exported MeterRun results in html""" data = services.meterruns_export() return render(request, 'project_run_table.html', data) def csv_export(request): """Return a dump of all the MeterRuns in CSV form""" response = HttpResponse(content_type="text/csv") response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"' - writer = csv.writer(response) - writer.writerow(['First row', 'Foo', 'Bar']) + data = services.meterruns_export() + + writer = csv.DictWriter(response, fieldnames=data['headers']) + writer.writeheader() + for meter_run in data['meter_runs']: + writer.writerow(meter_run) return response
Use the meterrun_export service to power csv export
## Code Before: import csv from django.shortcuts import render from django.http import HttpResponse from . import services def index(request): data = services.overview() return render(request, 'index.html', data) def meter_runs(request): """Render the table of exported MeterRun results in html""" data = services.meterruns_export() return render(request, 'project_run_table.html', data) def csv_export(request): """Return a dump of all the MeterRuns in CSV form""" response = HttpResponse(content_type="text/csv") response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"' writer = csv.writer(response) writer.writerow(['First row', 'Foo', 'Bar']) return response ## Instruction: Use the meterrun_export service to power csv export ## Code After: import csv from django.shortcuts import render from django.http import HttpResponse from . import services def index(request): data = services.overview() return render(request, 'index.html', data) def meter_runs(request): """Render the table of exported MeterRun results in html""" data = services.meterruns_export() return render(request, 'project_run_table.html', data) def csv_export(request): """Return a dump of all the MeterRuns in CSV form""" response = HttpResponse(content_type="text/csv") response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"' data = services.meterruns_export() writer = csv.DictWriter(response, fieldnames=data['headers']) writer.writeheader() for meter_run in data['meter_runs']: writer.writerow(meter_run) return response
import csv from django.shortcuts import render from django.http import HttpResponse from . import services def index(request): data = services.overview() return render(request, 'index.html', data) def meter_runs(request): """Render the table of exported MeterRun results in html""" data = services.meterruns_export() return render(request, 'project_run_table.html', data) def csv_export(request): """Return a dump of all the MeterRuns in CSV form""" response = HttpResponse(content_type="text/csv") response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"' - writer = csv.writer(response) - writer.writerow(['First row', 'Foo', 'Bar']) + data = services.meterruns_export() + + writer = csv.DictWriter(response, fieldnames=data['headers']) + writer.writeheader() + for meter_run in data['meter_runs']: + writer.writerow(meter_run) return response
29110323469d20ff1e481ab2267812afd8e0a3a4
more/chameleon/main.py
more/chameleon/main.py
import morepath import chameleon class ChameleonApp(morepath.App): pass @ChameleonApp.setting_section(section='chameleon') def get_setting_section(): return {'auto_reload': False} @ChameleonApp.template_engine(extension='.pt') def get_chameleon_render(path, original_render, settings): config = settings.chameleon.__dict__ template = chameleon.PageTemplateFile(path, **config) def render(content, request): variables = {'request': request} variables.update(content) return original_render(template.render(**variables), request) return render
import os import morepath import chameleon class ChameleonApp(morepath.App): pass @ChameleonApp.setting_section(section='chameleon') def get_setting_section(): return {'auto_reload': False} @ChameleonApp.template_engine(extension='.pt') def get_chameleon_render(name, original_render, registry, search_path): config = registry.settings.chameleon.__dict__ template = chameleon.PageTemplateFile(os.path.join(search_path, name), **config) def render(content, request): variables = {'request': request} variables.update(content) return original_render(template.render(**variables), request) return render
Adjust to modifications in Morepath. But now to enable real explicit file support.
Adjust to modifications in Morepath. But now to enable real explicit file support.
Python
bsd-3-clause
morepath/more.chameleon
+ import os import morepath import chameleon class ChameleonApp(morepath.App): pass @ChameleonApp.setting_section(section='chameleon') def get_setting_section(): return {'auto_reload': False} @ChameleonApp.template_engine(extension='.pt') - def get_chameleon_render(path, original_render, settings): + def get_chameleon_render(name, original_render, registry, search_path): - config = settings.chameleon.__dict__ + config = registry.settings.chameleon.__dict__ - template = chameleon.PageTemplateFile(path, **config) + template = chameleon.PageTemplateFile(os.path.join(search_path, name), + **config) def render(content, request): variables = {'request': request} variables.update(content) return original_render(template.render(**variables), request) return render
Adjust to modifications in Morepath. But now to enable real explicit file support.
## Code Before: import morepath import chameleon class ChameleonApp(morepath.App): pass @ChameleonApp.setting_section(section='chameleon') def get_setting_section(): return {'auto_reload': False} @ChameleonApp.template_engine(extension='.pt') def get_chameleon_render(path, original_render, settings): config = settings.chameleon.__dict__ template = chameleon.PageTemplateFile(path, **config) def render(content, request): variables = {'request': request} variables.update(content) return original_render(template.render(**variables), request) return render ## Instruction: Adjust to modifications in Morepath. But now to enable real explicit file support. ## Code After: import os import morepath import chameleon class ChameleonApp(morepath.App): pass @ChameleonApp.setting_section(section='chameleon') def get_setting_section(): return {'auto_reload': False} @ChameleonApp.template_engine(extension='.pt') def get_chameleon_render(name, original_render, registry, search_path): config = registry.settings.chameleon.__dict__ template = chameleon.PageTemplateFile(os.path.join(search_path, name), **config) def render(content, request): variables = {'request': request} variables.update(content) return original_render(template.render(**variables), request) return render
+ import os import morepath import chameleon class ChameleonApp(morepath.App): pass @ChameleonApp.setting_section(section='chameleon') def get_setting_section(): return {'auto_reload': False} @ChameleonApp.template_engine(extension='.pt') - def get_chameleon_render(path, original_render, settings): ? ^ ^^ ^^^^^ + def get_chameleon_render(name, original_render, registry, search_path): ? ^ ^^ ++++++++++ +++++++ ^ - config = settings.chameleon.__dict__ + config = registry.settings.chameleon.__dict__ ? +++++++++ - template = chameleon.PageTemplateFile(path, **config) ? ---- ^^^ + template = chameleon.PageTemplateFile(os.path.join(search_path, name), ? ++++++++++++++++++++ ^^^ + + **config) def render(content, request): variables = {'request': request} variables.update(content) return original_render(template.render(**variables), request) return render
73c842af63a09add43c0e33336dd4eb21153fda1
bin/database.py
bin/database.py
import json from api import config CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new def confirm_schema_match(): """ Checks version of database schema Returns (0) if DB schema version matches requirements. Returns (42) if DB schema version does not match requirements and can be upgraded. Returns (43) if DB schema version does not match requirements and cannot be upgraded, perhaps because code is at lower version than the DB schema version. """ version = config.db.version.find_one({"_id": "version"}) if version is None or version.get('database', None) is None: return 42 # At version 0 db_version = version.get('database', 0) if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: return 43 elif db_version < CURRENT_DATABASE_VERSION: return 42 else: return 0 def upgrade_schema(): """ Upgrades db to the current schema version Returns (0) if upgrade is successful """ # In progress # db_version = version.get('database',0) # if db_version < 1: # # rename the metadata fields # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) return 0
import json from api import config CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made def confirm_schema_match(): """ Checks version of database schema Returns (0) if DB schema version matches requirements. Returns (42) if DB schema version does not match requirements and can be upgraded. Returns (43) if DB schema version does not match requirements and cannot be upgraded, perhaps because code is at lower version than the DB schema version. """ version = config.db.version.find_one({"_id": "version"}) if version is None or version.get('database', None) is None: return 42 # At version 0 db_version = version.get('database', 0) if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: return 43 elif db_version < CURRENT_DATABASE_VERSION: return 42 else: return 0 def upgrade_schema(): """ Upgrades db to the current schema version Returns (0) if upgrade is successful """ # In progress # db_version = version.get('database',0) # if db_version < 1: # # rename the metadata fields # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) return 0
Fix tab vs spaces issue
Fix tab vs spaces issue
Python
mit
scitran/api,scitran/api,scitran/core,scitran/core,scitran/core,scitran/core
import json from api import config - CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new + CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made def confirm_schema_match(): - """ + """ - Checks version of database schema + Checks version of database schema - Returns (0) if DB schema version matches requirements. + Returns (0) if DB schema version matches requirements. - Returns (42) if DB schema version does not match + Returns (42) if DB schema version does not match - requirements and can be upgraded. + requirements and can be upgraded. - Returns (43) if DB schema version does not match + Returns (43) if DB schema version does not match - requirements and cannot be upgraded, + requirements and cannot be upgraded, - perhaps because code is at lower version + perhaps because code is at lower version - than the DB schema version. + than the DB schema version. - """ + """ - version = config.db.version.find_one({"_id": "version"}) + version = config.db.version.find_one({"_id": "version"}) - if version is None or version.get('database', None) is None: + if version is None or version.get('database', None) is None: - return 42 # At version 0 + return 42 # At version 0 - db_version = version.get('database', 0) + db_version = version.get('database', 0) - if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: + if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: - return 43 + return 43 - elif db_version < CURRENT_DATABASE_VERSION: + elif db_version < CURRENT_DATABASE_VERSION: - return 42 - else: - return 0 + return 42 + else: + return 0 def upgrade_schema(): - """ + """ - Upgrades db to the current schema version + Upgrades db to the current schema version Returns (0) if upgrade is successful """ - # In progress + # In progress - # db_version = version.get('database',0) + # db_version = version.get('database',0) - + - # if db_version < 1: + # if db_version < 1: - # # rename the metadata fields + # # rename the metadata fields - # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) + # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) - # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) + # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) - return 0 + return 0
Fix tab vs spaces issue
## Code Before: import json from api import config CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new def confirm_schema_match(): """ Checks version of database schema Returns (0) if DB schema version matches requirements. Returns (42) if DB schema version does not match requirements and can be upgraded. Returns (43) if DB schema version does not match requirements and cannot be upgraded, perhaps because code is at lower version than the DB schema version. """ version = config.db.version.find_one({"_id": "version"}) if version is None or version.get('database', None) is None: return 42 # At version 0 db_version = version.get('database', 0) if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: return 43 elif db_version < CURRENT_DATABASE_VERSION: return 42 else: return 0 def upgrade_schema(): """ Upgrades db to the current schema version Returns (0) if upgrade is successful """ # In progress # db_version = version.get('database',0) # if db_version < 1: # # rename the metadata fields # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) return 0 ## Instruction: Fix tab vs spaces issue ## Code After: import json from api import config CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made def confirm_schema_match(): """ Checks version of database schema Returns (0) if DB schema version matches requirements. Returns (42) if DB schema version does not match requirements and can be upgraded. Returns (43) if DB schema version does not match requirements and cannot be upgraded, perhaps because code is at lower version than the DB schema version. """ version = config.db.version.find_one({"_id": "version"}) if version is None or version.get('database', None) is None: return 42 # At version 0 db_version = version.get('database', 0) if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: return 43 elif db_version < CURRENT_DATABASE_VERSION: return 42 else: return 0 def upgrade_schema(): """ Upgrades db to the current schema version Returns (0) if upgrade is successful """ # In progress # db_version = version.get('database',0) # if db_version < 1: # # rename the metadata fields # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) return 0
import json from api import config - CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new + CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made ? +++++++++++++++++++++ def confirm_schema_match(): - """ + """ - Checks version of database schema ? ^ + Checks version of database schema ? ^^^^ - Returns (0) if DB schema version matches requirements. ? ^ + Returns (0) if DB schema version matches requirements. ? ^^^^ - Returns (42) if DB schema version does not match ? ^ + Returns (42) if DB schema version does not match ? ^^^^ - requirements and can be upgraded. ? ^ + requirements and can be upgraded. ? ^^^^ - Returns (43) if DB schema version does not match ? ^ + Returns (43) if DB schema version does not match ? ^^^^ - requirements and cannot be upgraded, ? ^ + requirements and cannot be upgraded, ? ^^^^ - perhaps because code is at lower version ? ^ + perhaps because code is at lower version ? ^^^^ - than the DB schema version. ? ^ + than the DB schema version. ? ^^^^ - """ + """ - version = config.db.version.find_one({"_id": "version"}) ? ^ + version = config.db.version.find_one({"_id": "version"}) ? ^^^^ - if version is None or version.get('database', None) is None: ? ^ + if version is None or version.get('database', None) is None: ? ^^^^ - return 42 # At version 0 ? ^^ + return 42 # At version 0 ? ^^^^^^^^ - db_version = version.get('database', 0) ? ^ + db_version = version.get('database', 0) ? ^^^^ - if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: ? ^ + if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION: ? ^^^^ - return 43 + return 43 - elif db_version < CURRENT_DATABASE_VERSION: ? ^ + elif db_version < CURRENT_DATABASE_VERSION: ? ^^^^ - return 42 - else: - return 0 + return 42 + else: + return 0 def upgrade_schema(): - """ + """ - Upgrades db to the current schema version ? ^ + Upgrades db to the current schema version ? ^^^^ Returns (0) if upgrade is successful """ - # In progress ? ^ + # In progress ? ^^^^ - # db_version = version.get('database',0) ? ^ + # db_version = version.get('database',0) ? ^^^^ - + - # if db_version < 1: ? ^ + # if db_version < 1: ? ^^^^ - # # rename the metadata fields ? ^ ^ + # # rename the metadata fields ? ^^^^ ^^ - # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) ? ^ ^ + # config.db.container.update_many({}, {"$rename": {"metadata": "info"}}) ? ^^^^ ^^ - # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) ? ^ + # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}}) ? ^^^^ - return 0 ? ^ + return 0 ? ^^^^
592c6550255793772add694cb941a0db0883713b
kamboo/core.py
kamboo/core.py
import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ session = botocore.session.get_session() def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region)
import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.session = botocore.session.get_session() self.service = service_name self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) def __repr__(self): return "KambooConnection: [%s, %s, %s]" % (self.account_id, self.region, self.service)
Fix the issue: "session" shared in different connections
Fix the issue: "session" shared in different connections
Python
apache-2.0
henrysher/kamboo,henrysher/kamboo
import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ - session = botocore.session.get_session() def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): + self.session = botocore.session.get_session() + self.service = service_name self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) + def __repr__(self): + return "KambooConnection: [%s, %s, %s]" % (self.account_id, + self.region, + self.service) +
Fix the issue: "session" shared in different connections
## Code Before: import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ session = botocore.session.get_session() def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) ## Instruction: Fix the issue: "session" shared in different connections ## Code After: import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.session = botocore.session.get_session() self.service = service_name self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) def __repr__(self): return "KambooConnection: [%s, %s, %s]" % (self.account_id, self.region, self.service)
import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ - session = botocore.session.get_session() def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): + self.session = botocore.session.get_session() + self.service = service_name self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) + + def __repr__(self): + return "KambooConnection: [%s, %s, %s]" % (self.account_id, + self.region, + self.service)
017ba0d18acb83a5135dd7a23c085b3c93d539b3
linkatos/message.py
linkatos/message.py
import re link_re = re.compile("https?://\S+(\s|$)") def extract_url(message): """ Returns the first url in a message. If there aren't any returns None """ answer = link_re.search(message) if answer is not None: answer = answer.group(0).strip() return answer
import re link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)") def extract_url(message): """ Returns the first url in a message. If there aren't any returns None """ answer = link_re.search(message) if answer is not None: answer = answer.group(2).strip() return answer
Change regex to adapt to the <url> format
fix: Change regex to adapt to the <url> format
Python
mit
iwi/linkatos,iwi/linkatos
import re - link_re = re.compile("https?://\S+(\s|$)") + link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)") def extract_url(message): """ Returns the first url in a message. If there aren't any returns None """ answer = link_re.search(message) if answer is not None: - answer = answer.group(0).strip() + answer = answer.group(2).strip() return answer
Change regex to adapt to the <url> format
## Code Before: import re link_re = re.compile("https?://\S+(\s|$)") def extract_url(message): """ Returns the first url in a message. If there aren't any returns None """ answer = link_re.search(message) if answer is not None: answer = answer.group(0).strip() return answer ## Instruction: Change regex to adapt to the <url> format ## Code After: import re link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)") def extract_url(message): """ Returns the first url in a message. If there aren't any returns None """ answer = link_re.search(message) if answer is not None: answer = answer.group(2).strip() return answer
import re - link_re = re.compile("https?://\S+(\s|$)") ? ^ + link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)") ? ++++++++ + ^^^^ ++++ def extract_url(message): """ Returns the first url in a message. If there aren't any returns None """ answer = link_re.search(message) if answer is not None: - answer = answer.group(0).strip() ? ^ + answer = answer.group(2).strip() ? ^ return answer
84af44868ea742bb5f6d08991526a98c8c78a931
tellurium/teconverters/__init__.py
tellurium/teconverters/__init__.py
from __future__ import absolute_import # converts Antimony to/from SBML from .convert_antimony import antimonyConverter from .convert_omex import inlineOmexImporter, OmexFormatDetector try: from .convert_phrasedml import phrasedmlImporter except: pass from .antimony_sbo import SBOError from .inline_omex import inlineOmex, saveInlineOMEX
from __future__ import absolute_import # converts Antimony to/from SBML from .convert_antimony import antimonyConverter from .convert_omex import inlineOmexImporter, OmexFormatDetector try: from .convert_phrasedml import phrasedmlImporter from .inline_omex import inlineOmex, saveInlineOMEX except: pass from .antimony_sbo import SBOError
Drop inline omex if it fails.
Drop inline omex if it fails.
Python
apache-2.0
sys-bio/tellurium,sys-bio/tellurium
from __future__ import absolute_import # converts Antimony to/from SBML from .convert_antimony import antimonyConverter from .convert_omex import inlineOmexImporter, OmexFormatDetector try: from .convert_phrasedml import phrasedmlImporter + from .inline_omex import inlineOmex, saveInlineOMEX except: pass from .antimony_sbo import SBOError - from .inline_omex import inlineOmex, saveInlineOMEX
Drop inline omex if it fails.
## Code Before: from __future__ import absolute_import # converts Antimony to/from SBML from .convert_antimony import antimonyConverter from .convert_omex import inlineOmexImporter, OmexFormatDetector try: from .convert_phrasedml import phrasedmlImporter except: pass from .antimony_sbo import SBOError from .inline_omex import inlineOmex, saveInlineOMEX ## Instruction: Drop inline omex if it fails. ## Code After: from __future__ import absolute_import # converts Antimony to/from SBML from .convert_antimony import antimonyConverter from .convert_omex import inlineOmexImporter, OmexFormatDetector try: from .convert_phrasedml import phrasedmlImporter from .inline_omex import inlineOmex, saveInlineOMEX except: pass from .antimony_sbo import SBOError
from __future__ import absolute_import # converts Antimony to/from SBML from .convert_antimony import antimonyConverter from .convert_omex import inlineOmexImporter, OmexFormatDetector try: from .convert_phrasedml import phrasedmlImporter + from .inline_omex import inlineOmex, saveInlineOMEX except: pass from .antimony_sbo import SBOError - from .inline_omex import inlineOmex, saveInlineOMEX
ae5626eaf36c6be94860d2a9570a777ff7f4e148
apps/client_config.py
apps/client_config.py
import superdesk from flask import current_app as app from superdesk.utils import ListCursor class ClientConfigResource(superdesk.Resource): item_methods = [] public_methods = ['GET'] resource_methods = ['GET'] class ClientConfigService(superdesk.Service): def get(self, req, lookup): return ListCursor() def on_fetched(self, docs): docs['config'] = getattr(app, 'client_config', {}) def init_app(app): superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app) app.client_config.update({ 'schema': app.config.get('SCHEMA'), 'editor': app.config.get('EDITOR'), })
import superdesk from flask import current_app as app from superdesk.utils import ListCursor class ClientConfigResource(superdesk.Resource): item_methods = [] public_methods = ['GET'] resource_methods = ['GET'] class ClientConfigService(superdesk.Service): def get(self, req, lookup): return ListCursor() def on_fetched(self, docs): docs['config'] = getattr(app, 'client_config', {}) def init_app(app): superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app) app.client_config.update({ 'schema': app.config.get('SCHEMA'), 'editor': app.config.get('EDITOR'), 'feedback_url': app.config.get('FEEDBACK_URL'), })
Add feedback url to served client configuration
[SDESK-2128] Add feedback url to served client configuration
Python
agpl-3.0
superdesk/superdesk-core,petrjasek/superdesk-core,mugurrus/superdesk-core,mdhaman/superdesk-core,ioanpocol/superdesk-core,ioanpocol/superdesk-core,mugurrus/superdesk-core,ioanpocol/superdesk-core,superdesk/superdesk-core,mdhaman/superdesk-core,petrjasek/superdesk-core,petrjasek/superdesk-core,hlmnrmr/superdesk-core,mugurrus/superdesk-core,superdesk/superdesk-core,superdesk/superdesk-core,mdhaman/superdesk-core,hlmnrmr/superdesk-core,petrjasek/superdesk-core
import superdesk from flask import current_app as app from superdesk.utils import ListCursor class ClientConfigResource(superdesk.Resource): item_methods = [] public_methods = ['GET'] resource_methods = ['GET'] class ClientConfigService(superdesk.Service): def get(self, req, lookup): return ListCursor() def on_fetched(self, docs): docs['config'] = getattr(app, 'client_config', {}) def init_app(app): superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app) app.client_config.update({ 'schema': app.config.get('SCHEMA'), 'editor': app.config.get('EDITOR'), + 'feedback_url': app.config.get('FEEDBACK_URL'), })
Add feedback url to served client configuration
## Code Before: import superdesk from flask import current_app as app from superdesk.utils import ListCursor class ClientConfigResource(superdesk.Resource): item_methods = [] public_methods = ['GET'] resource_methods = ['GET'] class ClientConfigService(superdesk.Service): def get(self, req, lookup): return ListCursor() def on_fetched(self, docs): docs['config'] = getattr(app, 'client_config', {}) def init_app(app): superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app) app.client_config.update({ 'schema': app.config.get('SCHEMA'), 'editor': app.config.get('EDITOR'), }) ## Instruction: Add feedback url to served client configuration ## Code After: import superdesk from flask import current_app as app from superdesk.utils import ListCursor class ClientConfigResource(superdesk.Resource): item_methods = [] public_methods = ['GET'] resource_methods = ['GET'] class ClientConfigService(superdesk.Service): def get(self, req, lookup): return ListCursor() def on_fetched(self, docs): docs['config'] = getattr(app, 'client_config', {}) def init_app(app): superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app) app.client_config.update({ 'schema': app.config.get('SCHEMA'), 'editor': app.config.get('EDITOR'), 'feedback_url': app.config.get('FEEDBACK_URL'), })
import superdesk from flask import current_app as app from superdesk.utils import ListCursor class ClientConfigResource(superdesk.Resource): item_methods = [] public_methods = ['GET'] resource_methods = ['GET'] class ClientConfigService(superdesk.Service): def get(self, req, lookup): return ListCursor() def on_fetched(self, docs): docs['config'] = getattr(app, 'client_config', {}) def init_app(app): superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app) app.client_config.update({ 'schema': app.config.get('SCHEMA'), 'editor': app.config.get('EDITOR'), + 'feedback_url': app.config.get('FEEDBACK_URL'), })
67d4f376586c912f852b98c75f7de04aeb05979a
pag/words.py
pag/words.py
"""Get words from files in "src/dictionary/".""" import os def get_word_list(filepath): """ Get a list of words from a file. Input: file name Output: dict with formula {word: [synonym, synonym]}""" filepath = os.path.abspath(filepath) assert os.path.isfile(filepath), 'Must be a file' f = open(filepath, 'r') contents = f.read() txt = contents.strip().split('\n') if ':' in contents: ntxt = txt[:] for line in txt: if line[0] == '#': ntxt.remove(ntxt[ntxt.index(line)]) elif ':' not in line: ntxt[ntxt.index(line)] = line + ':' txt = ntxt words = {} for line in txt: index = line.split(':')[0] words[index] = line.split(':')[1].split(',') for syn in words[index]: if syn == '': words[index].remove(syn) else: words = [word.strip() for word in txt] f.close() return words verbs = get_word_list('dictionary/verbs.txt') nouns = get_word_list('dictionary/nouns.txt') extras = get_word_list('dictionary/extras.txt') directions = get_word_list('dictionary/directions.txt')
"""Get words from files in "src/dictionary/".""" import os def get_word_list(filepath): """ Get a list of words from a file. Input: file name Output: dict with formula {word: [synonym, synonym]}""" filepath = os.path.abspath(filepath) assert os.path.isfile(filepath), 'Must be a file' f = open(filepath, 'r') contents = f.read() txt = contents.strip().split('\n') ntxt = txt[:] for line in txt: if line[0] == '#': ntxt.remove(ntxt[ntxt.index(line)]) elif ':' not in line: ntxt[ntxt.index(line)] = line + ':' txt = ntxt words = {} for line in txt: index = line.split(':')[0] words[index] = line.split(':')[1].split(',') for syn in words[index]: if syn == '': words[index].remove(syn) f.close() return words verbs = get_word_list('dictionary/verbs.txt') nouns = get_word_list('dictionary/nouns.txt') extras = get_word_list('dictionary/extras.txt') directions = get_word_list('dictionary/directions.txt')
Remove useless and confusing code
Remove useless and confusing code
Python
mit
allanburleson/python-adventure-game,disorientedperson/python-adventure-game
"""Get words from files in "src/dictionary/".""" import os def get_word_list(filepath): """ Get a list of words from a file. Input: file name Output: dict with formula {word: [synonym, synonym]}""" filepath = os.path.abspath(filepath) assert os.path.isfile(filepath), 'Must be a file' f = open(filepath, 'r') contents = f.read() txt = contents.strip().split('\n') - if ':' in contents: - ntxt = txt[:] + ntxt = txt[:] - for line in txt: + for line in txt: - if line[0] == '#': + if line[0] == '#': - ntxt.remove(ntxt[ntxt.index(line)]) + ntxt.remove(ntxt[ntxt.index(line)]) - elif ':' not in line: + elif ':' not in line: - ntxt[ntxt.index(line)] = line + ':' + ntxt[ntxt.index(line)] = line + ':' - txt = ntxt + txt = ntxt - words = {} + words = {} - for line in txt: + for line in txt: - index = line.split(':')[0] + index = line.split(':')[0] - words[index] = line.split(':')[1].split(',') + words[index] = line.split(':')[1].split(',') - for syn in words[index]: + for syn in words[index]: - if syn == '': + if syn == '': - words[index].remove(syn) + words[index].remove(syn) - else: - words = [word.strip() for word in txt] f.close() return words verbs = get_word_list('dictionary/verbs.txt') nouns = get_word_list('dictionary/nouns.txt') extras = get_word_list('dictionary/extras.txt') directions = get_word_list('dictionary/directions.txt')
Remove useless and confusing code
## Code Before: """Get words from files in "src/dictionary/".""" import os def get_word_list(filepath): """ Get a list of words from a file. Input: file name Output: dict with formula {word: [synonym, synonym]}""" filepath = os.path.abspath(filepath) assert os.path.isfile(filepath), 'Must be a file' f = open(filepath, 'r') contents = f.read() txt = contents.strip().split('\n') if ':' in contents: ntxt = txt[:] for line in txt: if line[0] == '#': ntxt.remove(ntxt[ntxt.index(line)]) elif ':' not in line: ntxt[ntxt.index(line)] = line + ':' txt = ntxt words = {} for line in txt: index = line.split(':')[0] words[index] = line.split(':')[1].split(',') for syn in words[index]: if syn == '': words[index].remove(syn) else: words = [word.strip() for word in txt] f.close() return words verbs = get_word_list('dictionary/verbs.txt') nouns = get_word_list('dictionary/nouns.txt') extras = get_word_list('dictionary/extras.txt') directions = get_word_list('dictionary/directions.txt') ## Instruction: Remove useless and confusing code ## Code After: """Get words from files in "src/dictionary/".""" import os def get_word_list(filepath): """ Get a list of words from a file. Input: file name Output: dict with formula {word: [synonym, synonym]}""" filepath = os.path.abspath(filepath) assert os.path.isfile(filepath), 'Must be a file' f = open(filepath, 'r') contents = f.read() txt = contents.strip().split('\n') ntxt = txt[:] for line in txt: if line[0] == '#': ntxt.remove(ntxt[ntxt.index(line)]) elif ':' not in line: ntxt[ntxt.index(line)] = line + ':' txt = ntxt words = {} for line in txt: index = line.split(':')[0] words[index] = line.split(':')[1].split(',') for syn in words[index]: if syn == '': words[index].remove(syn) f.close() return words verbs = get_word_list('dictionary/verbs.txt') nouns = get_word_list('dictionary/nouns.txt') extras = get_word_list('dictionary/extras.txt') directions = get_word_list('dictionary/directions.txt')
"""Get words from files in "src/dictionary/".""" import os def get_word_list(filepath): """ Get a list of words from a file. Input: file name Output: dict with formula {word: [synonym, synonym]}""" filepath = os.path.abspath(filepath) assert os.path.isfile(filepath), 'Must be a file' f = open(filepath, 'r') contents = f.read() txt = contents.strip().split('\n') - if ':' in contents: - ntxt = txt[:] ? ---- + ntxt = txt[:] - for line in txt: ? ---- + for line in txt: - if line[0] == '#': ? ---- + if line[0] == '#': - ntxt.remove(ntxt[ntxt.index(line)]) ? ---- + ntxt.remove(ntxt[ntxt.index(line)]) - elif ':' not in line: ? ---- + elif ':' not in line: - ntxt[ntxt.index(line)] = line + ':' ? ---- + ntxt[ntxt.index(line)] = line + ':' - txt = ntxt ? ---- + txt = ntxt - words = {} ? ---- + words = {} - for line in txt: ? ---- + for line in txt: - index = line.split(':')[0] ? ---- + index = line.split(':')[0] - words[index] = line.split(':')[1].split(',') ? ---- + words[index] = line.split(':')[1].split(',') - for syn in words[index]: ? ---- + for syn in words[index]: - if syn == '': ? ---- + if syn == '': - words[index].remove(syn) ? ---- + words[index].remove(syn) - else: - words = [word.strip() for word in txt] f.close() return words verbs = get_word_list('dictionary/verbs.txt') nouns = get_word_list('dictionary/nouns.txt') extras = get_word_list('dictionary/extras.txt') directions = get_word_list('dictionary/directions.txt')
4848baf76e4972401530b624816ba48cb08d9398
appconf/utils.py
appconf/utils.py
import sys def import_attribute(import_path, exception_handler=None): from django.utils.importlib import import_module module_name, object_name = import_path.rsplit('.', 1) try: module = import_module(module_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise try: return getattr(module, object_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise
import sys def import_attribute(import_path, exception_handler=None): try: from importlib import import_module except ImportError: from django.utils.importlib import import_module module_name, object_name = import_path.rsplit('.', 1) try: module = import_module(module_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise try: return getattr(module, object_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise
Use import_module from standard library if exists
Use import_module from standard library if exists Django 1.8+ drops `django.utils.importlib`. I imagine because that is because an older version of Python (either 2.5 and/or 2.6) is being dropped. I haven't checked older versions but `importlib` exists in Python 2.7.
Python
bsd-3-clause
diox/django-appconf,carltongibson/django-appconf,django-compressor/django-appconf,jezdez/django-appconf,jessehon/django-appconf,treyhunner/django-appconf,jezdez-archive/django-appconf
import sys def import_attribute(import_path, exception_handler=None): + try: + from importlib import import_module + except ImportError: - from django.utils.importlib import import_module + from django.utils.importlib import import_module module_name, object_name = import_path.rsplit('.', 1) try: module = import_module(module_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise try: return getattr(module, object_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise
Use import_module from standard library if exists
## Code Before: import sys def import_attribute(import_path, exception_handler=None): from django.utils.importlib import import_module module_name, object_name = import_path.rsplit('.', 1) try: module = import_module(module_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise try: return getattr(module, object_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise ## Instruction: Use import_module from standard library if exists ## Code After: import sys def import_attribute(import_path, exception_handler=None): try: from importlib import import_module except ImportError: from django.utils.importlib import import_module module_name, object_name = import_path.rsplit('.', 1) try: module = import_module(module_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise try: return getattr(module, object_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise
import sys def import_attribute(import_path, exception_handler=None): + try: + from importlib import import_module + except ImportError: - from django.utils.importlib import import_module + from django.utils.importlib import import_module ? ++++ module_name, object_name = import_path.rsplit('.', 1) try: module = import_module(module_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise try: return getattr(module, object_name) except: # pragma: no cover if callable(exception_handler): exctype, excvalue, tb = sys.exc_info() return exception_handler(import_path, exctype, excvalue, tb) else: raise
9cd3bb79126fa2431ba4ae03811ac30fb77b9b46
netcat.py
netcat.py
import argparse import socket import sys parser = argparse.ArgumentParser(description='Simple netcat in pure python.') parser.add_argument('-z', '--scan', action='store_true') parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int) parser.add_argument('-v', '--verbose', action='store_true') parser.add_argument('host') parser.add_argument('port', type=int) args = parser.parse_args() if args.scan: try: connection = socket.create_connection((args.host, args.port), args.timeout) if args.verbose: print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port) sys.exit(0) except socket.error as msg: if args.verbose: print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg) sys.exit(1) else: print 'Not implemented'
import argparse import socket import sys parser = argparse.ArgumentParser(description='Simple netcat in pure python.') parser.add_argument('-s', '--source', metavar='ADDRESS') parser.add_argument('-v', '--verbose', action='store_true') parser.add_argument('-w', '--wait', metavar='SECONDS', type=int) parser.add_argument('-z', '--zero', action='store_true') parser.add_argument('host') parser.add_argument('port') args = parser.parse_args() # Set a souce address for socket connection source = ('', 0) if args.source: source = (args.source, 0) # port 0 specifies that the OS will choose a port # exit successfully if the connection succeeds if args.zero: try: connection = socket.create_connection((args.host, args.port), args.wait, source) if args.verbose: print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)) sys.exit(0) except socket.error as msg: if args.verbose: print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)) sys.exit(1) else: print('Not implemented')
Support python 2 and 3
Support python 2 and 3 Add source argument. Update arguments to use long names from GNU netcat.
Python
unlicense
benformosa/Toolbox,benformosa/Toolbox
import argparse import socket import sys parser = argparse.ArgumentParser(description='Simple netcat in pure python.') + parser.add_argument('-s', '--source', metavar='ADDRESS') - parser.add_argument('-z', '--scan', action='store_true') - parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int) parser.add_argument('-v', '--verbose', action='store_true') + parser.add_argument('-w', '--wait', metavar='SECONDS', type=int) + parser.add_argument('-z', '--zero', action='store_true') parser.add_argument('host') - parser.add_argument('port', type=int) + parser.add_argument('port') args = parser.parse_args() + + # Set a souce address for socket connection + source = ('', 0) - if args.scan: + if args.source: + source = (args.source, 0) # port 0 specifies that the OS will choose a port + + # exit successfully if the connection succeeds + if args.zero: try: - connection = socket.create_connection((args.host, args.port), args.timeout) + connection = socket.create_connection((args.host, args.port), args.wait, source) if args.verbose: - print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port) + print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)) sys.exit(0) except socket.error as msg: if args.verbose: - print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg) + print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)) sys.exit(1) else: - print 'Not implemented' + print('Not implemented')
Support python 2 and 3
## Code Before: import argparse import socket import sys parser = argparse.ArgumentParser(description='Simple netcat in pure python.') parser.add_argument('-z', '--scan', action='store_true') parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int) parser.add_argument('-v', '--verbose', action='store_true') parser.add_argument('host') parser.add_argument('port', type=int) args = parser.parse_args() if args.scan: try: connection = socket.create_connection((args.host, args.port), args.timeout) if args.verbose: print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port) sys.exit(0) except socket.error as msg: if args.verbose: print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg) sys.exit(1) else: print 'Not implemented' ## Instruction: Support python 2 and 3 ## Code After: import argparse import socket import sys parser = argparse.ArgumentParser(description='Simple netcat in pure python.') parser.add_argument('-s', '--source', metavar='ADDRESS') parser.add_argument('-v', '--verbose', action='store_true') parser.add_argument('-w', '--wait', metavar='SECONDS', type=int) parser.add_argument('-z', '--zero', action='store_true') parser.add_argument('host') parser.add_argument('port') args = parser.parse_args() # Set a souce address for socket connection source = ('', 0) if args.source: source = (args.source, 0) # port 0 specifies that the OS will choose a port # exit successfully if the connection succeeds if args.zero: try: connection = socket.create_connection((args.host, args.port), args.wait, source) if args.verbose: print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)) sys.exit(0) except socket.error as msg: if args.verbose: print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)) sys.exit(1) else: print('Not implemented')
import argparse import socket import sys parser = argparse.ArgumentParser(description='Simple netcat in pure python.') + parser.add_argument('-s', '--source', metavar='ADDRESS') - parser.add_argument('-z', '--scan', action='store_true') - parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int) parser.add_argument('-v', '--verbose', action='store_true') + parser.add_argument('-w', '--wait', metavar='SECONDS', type=int) + parser.add_argument('-z', '--zero', action='store_true') parser.add_argument('host') - parser.add_argument('port', type=int) ? ---------- + parser.add_argument('port') args = parser.parse_args() + + # Set a souce address for socket connection + source = ('', 0) - if args.scan: ? ^^ + if args.source: ? +++ ^ + source = (args.source, 0) # port 0 specifies that the OS will choose a port + + # exit successfully if the connection succeeds + if args.zero: try: - connection = socket.create_connection((args.host, args.port), args.timeout) ? ^^^ ^ + connection = socket.create_connection((args.host, args.port), args.wait, source) ? +++ ^^^ ^^^ if args.verbose: - print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port) ? ^ + print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)) ? ^ + sys.exit(0) except socket.error as msg: if args.verbose: - print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg) ? ^ + print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)) ? ^ + sys.exit(1) else: - print 'Not implemented' ? ^ + print('Not implemented') ? ^ +
7ace27a6a114e381a30ac9760880b68277a868fc
python_scripts/mc_config.py
python_scripts/mc_config.py
import yaml def read_config(): yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb') config_file = yaml.load( yml_file ) return config_file
import yaml import os.path _config_file_base_name = 'mediawords.yml' _config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml')) def read_config(): yml_file = open(_config_file_name, 'rb') config_file = yaml.load( yml_file ) return config_file
Use relative path location for mediawords.yml.
Use relative path location for mediawords.yml.
Python
agpl-3.0
berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud
import yaml + import os.path + + _config_file_base_name = 'mediawords.yml' + _config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml')) def read_config(): - yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb') + yml_file = open(_config_file_name, 'rb') config_file = yaml.load( yml_file ) return config_file
Use relative path location for mediawords.yml.
## Code Before: import yaml def read_config(): yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb') config_file = yaml.load( yml_file ) return config_file ## Instruction: Use relative path location for mediawords.yml. ## Code After: import yaml import os.path _config_file_base_name = 'mediawords.yml' _config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml')) def read_config(): yml_file = open(_config_file_name, 'rb') config_file = yaml.load( yml_file ) return config_file
import yaml + import os.path + + _config_file_base_name = 'mediawords.yml' + _config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml')) def read_config(): - yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb') + yml_file = open(_config_file_name, 'rb') config_file = yaml.load( yml_file ) return config_file
c1ed5befe3081f6812fc77fc694ea3e82d90f39c
telemetry/telemetry/core/backends/facebook_credentials_backend.py
telemetry/telemetry/core/backends/facebook_credentials_backend.py
from telemetry.core.backends import form_based_credentials_backend class FacebookCredentialsBackend( form_based_credentials_backend.FormBasedCredentialsBackend): @property def logged_in_javascript(self): """Evaluates to true iff already logged in.""" return ('document.getElementById("fbNotificationsList")!== null || ' 'document.getElementById("m_home_notice")!== null') @property def credentials_type(self): return 'facebook' @property def url(self): return 'http://www.facebook.com/' @property def login_form_id(self): return 'login_form' @property def login_input_id(self): return 'email' @property def password_input_id(self): return 'pass' class FacebookCredentialsBackend2(FacebookCredentialsBackend): @property def credentials_type(self): return 'facebook2'
from telemetry.core.backends import form_based_credentials_backend class FacebookCredentialsBackend( form_based_credentials_backend.FormBasedCredentialsBackend): @property def logged_in_javascript(self): """Evaluates to true iff already logged in.""" return ('document.getElementById("fbNotificationsList")!== null || ' 'document.getElementById("m_home_notice")!== null') @property def credentials_type(self): return 'facebook' @property def url(self): return 'http://www.facebook.com/' @property def login_form_id(self): return 'login_form' @property def login_input_id(self): return 'email' @property def password_input_id(self): return 'pass' class FacebookCredentialsBackend2(FacebookCredentialsBackend): """ Facebook credential backend for https client. """ @property def credentials_type(self): return 'facebook2' @property def url(self): return 'https://www.facebook.com/'
Set facebook_crendentials_backend_2's url to https
[Telemetry] Set facebook_crendentials_backend_2's url to https TBR=tonyg@chromium.org BUG=428098 Review URL: https://codereview.chromium.org/688113003 Cr-Commit-Position: 972c6d2dc6dd5efdad1377c0d224e03eb8f276f7@{#301945}
Python
bsd-3-clause
benschmaus/catapult,SummerLW/Perf-Insight-Report,sahiljain/catapult,SummerLW/Perf-Insight-Report,sahiljain/catapult,catapult-project/catapult,benschmaus/catapult,SummerLW/Perf-Insight-Report,catapult-project/catapult,sahiljain/catapult,catapult-project/catapult-csm,benschmaus/catapult,catapult-project/catapult-csm,sahiljain/catapult,catapult-project/catapult-csm,catapult-project/catapult-csm,SummerLW/Perf-Insight-Report,catapult-project/catapult,benschmaus/catapult,SummerLW/Perf-Insight-Report,catapult-project/catapult,SummerLW/Perf-Insight-Report,benschmaus/catapult,benschmaus/catapult,sahiljain/catapult,catapult-project/catapult-csm,catapult-project/catapult,catapult-project/catapult-csm,sahiljain/catapult,catapult-project/catapult-csm,catapult-project/catapult,benschmaus/catapult,catapult-project/catapult
from telemetry.core.backends import form_based_credentials_backend class FacebookCredentialsBackend( form_based_credentials_backend.FormBasedCredentialsBackend): @property def logged_in_javascript(self): """Evaluates to true iff already logged in.""" return ('document.getElementById("fbNotificationsList")!== null || ' 'document.getElementById("m_home_notice")!== null') @property def credentials_type(self): return 'facebook' @property def url(self): return 'http://www.facebook.com/' @property def login_form_id(self): return 'login_form' @property def login_input_id(self): return 'email' @property def password_input_id(self): return 'pass' class FacebookCredentialsBackend2(FacebookCredentialsBackend): + """ Facebook credential backend for https client. """ @property def credentials_type(self): return 'facebook2' + @property + def url(self): + return 'https://www.facebook.com/' +
Set facebook_crendentials_backend_2's url to https
## Code Before: from telemetry.core.backends import form_based_credentials_backend class FacebookCredentialsBackend( form_based_credentials_backend.FormBasedCredentialsBackend): @property def logged_in_javascript(self): """Evaluates to true iff already logged in.""" return ('document.getElementById("fbNotificationsList")!== null || ' 'document.getElementById("m_home_notice")!== null') @property def credentials_type(self): return 'facebook' @property def url(self): return 'http://www.facebook.com/' @property def login_form_id(self): return 'login_form' @property def login_input_id(self): return 'email' @property def password_input_id(self): return 'pass' class FacebookCredentialsBackend2(FacebookCredentialsBackend): @property def credentials_type(self): return 'facebook2' ## Instruction: Set facebook_crendentials_backend_2's url to https ## Code After: from telemetry.core.backends import form_based_credentials_backend class FacebookCredentialsBackend( form_based_credentials_backend.FormBasedCredentialsBackend): @property def logged_in_javascript(self): """Evaluates to true iff already logged in.""" return ('document.getElementById("fbNotificationsList")!== null || ' 'document.getElementById("m_home_notice")!== null') @property def credentials_type(self): return 'facebook' @property def url(self): return 'http://www.facebook.com/' @property def login_form_id(self): return 'login_form' @property def login_input_id(self): return 'email' @property def password_input_id(self): return 'pass' class FacebookCredentialsBackend2(FacebookCredentialsBackend): """ Facebook credential backend for https client. """ @property def credentials_type(self): return 'facebook2' @property def url(self): return 'https://www.facebook.com/'
from telemetry.core.backends import form_based_credentials_backend class FacebookCredentialsBackend( form_based_credentials_backend.FormBasedCredentialsBackend): @property def logged_in_javascript(self): """Evaluates to true iff already logged in.""" return ('document.getElementById("fbNotificationsList")!== null || ' 'document.getElementById("m_home_notice")!== null') @property def credentials_type(self): return 'facebook' @property def url(self): return 'http://www.facebook.com/' @property def login_form_id(self): return 'login_form' @property def login_input_id(self): return 'email' @property def password_input_id(self): return 'pass' class FacebookCredentialsBackend2(FacebookCredentialsBackend): + """ Facebook credential backend for https client. """ @property def credentials_type(self): return 'facebook2' + + @property + def url(self): + return 'https://www.facebook.com/'
18fec1124bb86f90183350e7b9c86eb946a01884
whatchanged/main.py
whatchanged/main.py
from __future__ import absolute_import, print_function # Standard library from os import walk from os.path import exists, isdir, join # Local library from .util import is_py_file from .diff import diff_files def main(): import sys if sys.argv < 3: print('Usage: %s <module1> <module2>' % sys.argv[0]) sys.exit(1) old, new = sys.argv[1:3] diff = set([]) if isdir(old): assert isdir(new) for dirpath, dirnames, filenames in walk(new): for file_ in filenames: if is_py_file(file_): new_file = join(dirpath, file_) old_file = new_file.replace(new, old) if exists(old_file): mdiff = diff_files(old_file, new_file) if mdiff is not None: diff.add(mdiff) else: diff.add(diff_files(old, new)) for module in diff: print(module) if __name__ == '__main__': main()
from __future__ import absolute_import, print_function # Standard library from os import walk from os.path import exists, isdir, join # Local library from .util import is_py_file from .diff import diff_files def main(): import sys if len(sys.argv) < 3: print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0]) sys.exit(1) old, new = sys.argv[1:3] diff = set([]) if isdir(old): assert isdir(new) for dirpath, dirnames, filenames in walk(new): for file_ in filenames: if is_py_file(file_): new_file = join(dirpath, file_) old_file = new_file.replace(new, old) if exists(old_file): mdiff = diff_files(old_file, new_file) if mdiff is not None: diff.add(mdiff) else: diff.add(diff_files(old, new)) for module in diff: print(module) if __name__ == '__main__': main()
Fix minor bug in length comparison.
Fix minor bug in length comparison.
Python
bsd-2-clause
punchagan/what-changed
from __future__ import absolute_import, print_function # Standard library from os import walk from os.path import exists, isdir, join # Local library from .util import is_py_file from .diff import diff_files def main(): import sys - if sys.argv < 3: + if len(sys.argv) < 3: - print('Usage: %s <module1> <module2>' % sys.argv[0]) + print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0]) sys.exit(1) old, new = sys.argv[1:3] diff = set([]) if isdir(old): assert isdir(new) for dirpath, dirnames, filenames in walk(new): for file_ in filenames: if is_py_file(file_): new_file = join(dirpath, file_) old_file = new_file.replace(new, old) if exists(old_file): mdiff = diff_files(old_file, new_file) if mdiff is not None: diff.add(mdiff) else: diff.add(diff_files(old, new)) for module in diff: print(module) if __name__ == '__main__': main()
Fix minor bug in length comparison.
## Code Before: from __future__ import absolute_import, print_function # Standard library from os import walk from os.path import exists, isdir, join # Local library from .util import is_py_file from .diff import diff_files def main(): import sys if sys.argv < 3: print('Usage: %s <module1> <module2>' % sys.argv[0]) sys.exit(1) old, new = sys.argv[1:3] diff = set([]) if isdir(old): assert isdir(new) for dirpath, dirnames, filenames in walk(new): for file_ in filenames: if is_py_file(file_): new_file = join(dirpath, file_) old_file = new_file.replace(new, old) if exists(old_file): mdiff = diff_files(old_file, new_file) if mdiff is not None: diff.add(mdiff) else: diff.add(diff_files(old, new)) for module in diff: print(module) if __name__ == '__main__': main() ## Instruction: Fix minor bug in length comparison. ## Code After: from __future__ import absolute_import, print_function # Standard library from os import walk from os.path import exists, isdir, join # Local library from .util import is_py_file from .diff import diff_files def main(): import sys if len(sys.argv) < 3: print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0]) sys.exit(1) old, new = sys.argv[1:3] diff = set([]) if isdir(old): assert isdir(new) for dirpath, dirnames, filenames in walk(new): for file_ in filenames: if is_py_file(file_): new_file = join(dirpath, file_) old_file = new_file.replace(new, old) if exists(old_file): mdiff = diff_files(old_file, new_file) if mdiff is not None: diff.add(mdiff) else: diff.add(diff_files(old, new)) for module in diff: print(module) if __name__ == '__main__': main()
from __future__ import absolute_import, print_function # Standard library from os import walk from os.path import exists, isdir, join # Local library from .util import is_py_file from .diff import diff_files def main(): import sys - if sys.argv < 3: + if len(sys.argv) < 3: ? ++++ + - print('Usage: %s <module1> <module2>' % sys.argv[0]) + print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0]) ? +++++++++ +++++++++ sys.exit(1) old, new = sys.argv[1:3] diff = set([]) if isdir(old): assert isdir(new) for dirpath, dirnames, filenames in walk(new): for file_ in filenames: if is_py_file(file_): new_file = join(dirpath, file_) old_file = new_file.replace(new, old) if exists(old_file): mdiff = diff_files(old_file, new_file) if mdiff is not None: diff.add(mdiff) else: diff.add(diff_files(old, new)) for module in diff: print(module) if __name__ == '__main__': main()
d7fdebdc4ce52e59c126a27ea06171994a6c846b
src/config/common/ssl_adapter.py
src/config/common/ssl_adapter.py
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # @author: Sanju Abraham, Juniper Networks, OpenContrail from requests.adapters import HTTPAdapter try: # This is required for RDO, which installs both python-requests # and python-urllib3, but symlinks python-request's internally packaged # urllib3 to the site installed one. from requests.packages.urllib3.poolmanager import PoolManager except ImportError: # Fallback to standard installation methods from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that can be configured with SSL/TLS version.''' def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version self.poolmanager = None super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # @author: Sanju Abraham, Juniper Networks, OpenContrail from requests.adapters import HTTPAdapter try: # This is required for RDO, which installs both python-requests # and python-urllib3, but symlinks python-request's internally packaged # urllib3 to the site installed one. from requests.packages.urllib3.poolmanager import PoolManager except ImportError: # Fallback to standard installation methods from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that can be configured with SSL/TLS version.''' HTTPAdapter.__attrs__.extend(['ssl_version']) def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version self.poolmanager = None super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module. This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__.
Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module. This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__. Courtesy: https://github.com/sigmavirus24/requests-toolbelt/commit/decadbd3512444889feb30cf1ff2f1448a3ecfca Closes-Bug:#1604247 Change-Id: Iee9e0348c005e88c535f4da33cf98149a8c1b19d
Python
apache-2.0
codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # @author: Sanju Abraham, Juniper Networks, OpenContrail from requests.adapters import HTTPAdapter try: # This is required for RDO, which installs both python-requests # and python-urllib3, but symlinks python-request's internally packaged # urllib3 to the site installed one. from requests.packages.urllib3.poolmanager import PoolManager except ImportError: # Fallback to standard installation methods from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that can be configured with SSL/TLS version.''' + HTTPAdapter.__attrs__.extend(['ssl_version']) def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version self.poolmanager = None super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module. This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__.
## Code Before: # -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # @author: Sanju Abraham, Juniper Networks, OpenContrail from requests.adapters import HTTPAdapter try: # This is required for RDO, which installs both python-requests # and python-urllib3, but symlinks python-request's internally packaged # urllib3 to the site installed one. from requests.packages.urllib3.poolmanager import PoolManager except ImportError: # Fallback to standard installation methods from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that can be configured with SSL/TLS version.''' def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version self.poolmanager = None super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version) ## Instruction: Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module. This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__. ## Code After: # -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # @author: Sanju Abraham, Juniper Networks, OpenContrail from requests.adapters import HTTPAdapter try: # This is required for RDO, which installs both python-requests # and python-urllib3, but symlinks python-request's internally packaged # urllib3 to the site installed one. from requests.packages.urllib3.poolmanager import PoolManager except ImportError: # Fallback to standard installation methods from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that can be configured with SSL/TLS version.''' HTTPAdapter.__attrs__.extend(['ssl_version']) def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version self.poolmanager = None super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # @author: Sanju Abraham, Juniper Networks, OpenContrail from requests.adapters import HTTPAdapter try: # This is required for RDO, which installs both python-requests # and python-urllib3, but symlinks python-request's internally packaged # urllib3 to the site installed one. from requests.packages.urllib3.poolmanager import PoolManager except ImportError: # Fallback to standard installation methods from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that can be configured with SSL/TLS version.''' + HTTPAdapter.__attrs__.extend(['ssl_version']) def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version self.poolmanager = None super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
a0e8c92a9d12846c8cfe6819ea26d1e08dd4098a
example/models.py
example/models.py
import i18n from i18n.models import TranslatableModel class Document(TranslatableModel): charfield = i18n.LocalizedCharField(max_length=50) textfield = i18n.LocalizedTextField(max_length=512) filefield = i18n.LocalizedFileField(null=True, upload_to='files') imagefield = i18n.LocalizedImageField(null=True, upload_to='images') booleanfield = i18n.LocalizedBooleanField() datefield = i18n.LocalizedDateField() fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True, related_name='+') urlfied = i18n.LocalizedURLField() decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2) integerfield = i18n.LocalizedIntegerField() def __str__(self): return '%d, %s' % (self.pk, self.charfield) class Meta: app_label = 'example'
from django.db import models import i18n from i18n.models import TranslatableModel class Document(TranslatableModel): untranslated_charfield = models.CharField(max_length=50, blank=True) charfield = i18n.LocalizedCharField(max_length=50) textfield = i18n.LocalizedTextField(max_length=500, blank=True) filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True) imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True) booleanfield = i18n.LocalizedBooleanField() datefield = i18n.LocalizedDateField(blank=True, null=True) fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True, related_name='+') urlfied = i18n.LocalizedURLField(null=True, blank=True) decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True, blank=True) integerfield = i18n.LocalizedIntegerField(null=True, blank=True) def __str__(self): return '%d, %s' % (self.pk, self.charfield) class Meta: app_label = 'example'
Make fields in example app non required
Make fields in example app non required
Python
bsd-3-clause
jonasundderwolf/django-localizedfields,jonasundderwolf/django-localizedfields
+ from django.db import models import i18n from i18n.models import TranslatableModel class Document(TranslatableModel): - + untranslated_charfield = models.CharField(max_length=50, blank=True) charfield = i18n.LocalizedCharField(max_length=50) - textfield = i18n.LocalizedTextField(max_length=512) + textfield = i18n.LocalizedTextField(max_length=500, blank=True) - filefield = i18n.LocalizedFileField(null=True, upload_to='files') + filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True) - imagefield = i18n.LocalizedImageField(null=True, upload_to='images') + imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True) booleanfield = i18n.LocalizedBooleanField() - datefield = i18n.LocalizedDateField() + datefield = i18n.LocalizedDateField(blank=True, null=True) fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True, related_name='+') - urlfied = i18n.LocalizedURLField() + urlfied = i18n.LocalizedURLField(null=True, blank=True) - decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2) + decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True, + blank=True) - integerfield = i18n.LocalizedIntegerField() + integerfield = i18n.LocalizedIntegerField(null=True, blank=True) def __str__(self): return '%d, %s' % (self.pk, self.charfield) class Meta: app_label = 'example'
Make fields in example app non required
## Code Before: import i18n from i18n.models import TranslatableModel class Document(TranslatableModel): charfield = i18n.LocalizedCharField(max_length=50) textfield = i18n.LocalizedTextField(max_length=512) filefield = i18n.LocalizedFileField(null=True, upload_to='files') imagefield = i18n.LocalizedImageField(null=True, upload_to='images') booleanfield = i18n.LocalizedBooleanField() datefield = i18n.LocalizedDateField() fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True, related_name='+') urlfied = i18n.LocalizedURLField() decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2) integerfield = i18n.LocalizedIntegerField() def __str__(self): return '%d, %s' % (self.pk, self.charfield) class Meta: app_label = 'example' ## Instruction: Make fields in example app non required ## Code After: from django.db import models import i18n from i18n.models import TranslatableModel class Document(TranslatableModel): untranslated_charfield = models.CharField(max_length=50, blank=True) charfield = i18n.LocalizedCharField(max_length=50) textfield = i18n.LocalizedTextField(max_length=500, blank=True) filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True) imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True) booleanfield = i18n.LocalizedBooleanField() datefield = i18n.LocalizedDateField(blank=True, null=True) fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True, related_name='+') urlfied = i18n.LocalizedURLField(null=True, blank=True) decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True, blank=True) integerfield = i18n.LocalizedIntegerField(null=True, blank=True) def __str__(self): return '%d, %s' % (self.pk, self.charfield) class Meta: app_label = 'example'
+ from django.db import models import i18n from i18n.models import TranslatableModel class Document(TranslatableModel): - + untranslated_charfield = models.CharField(max_length=50, blank=True) charfield = i18n.LocalizedCharField(max_length=50) - textfield = i18n.LocalizedTextField(max_length=512) ? ^^ + textfield = i18n.LocalizedTextField(max_length=500, blank=True) ? ^^^^^^^^^^^^^^ - filefield = i18n.LocalizedFileField(null=True, upload_to='files') + filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True) ? ++++++++++++ - imagefield = i18n.LocalizedImageField(null=True, upload_to='images') + imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True) ? ++++++++++++ booleanfield = i18n.LocalizedBooleanField() - datefield = i18n.LocalizedDateField() + datefield = i18n.LocalizedDateField(blank=True, null=True) ? +++++++++++++++++++++ fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True, related_name='+') - urlfied = i18n.LocalizedURLField() + urlfied = i18n.LocalizedURLField(null=True, blank=True) ? +++++++++++++++++++++ - decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2) ? ^ + decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True, ? ^^^^^^^^^^^^ + blank=True) - integerfield = i18n.LocalizedIntegerField() + integerfield = i18n.LocalizedIntegerField(null=True, blank=True) ? +++++++++++++++++++++ def __str__(self): return '%d, %s' % (self.pk, self.charfield) class Meta: app_label = 'example'
a31e62f2a981f7662aee8a35ad195252a542d08d
plugins/say.py
plugins/say.py
from motobot import command, action @command('say') def say_command(bot, message, database): masters = [ "Moto-chan", "Motoko11", "Akahige", "betholas", "Baradium", "Cold_slither", "Drahken" ] if message.nick.lower() not in [x.lower() for x in masters]: return "Check your privilege!" else: args = message.message.split(' ')[1:] if len(args) < 2: return "You must specify both a channel and a message" else: channel = args[0] message = ' '.join(args[1:]) if message.startswith('/me '): message = action(message[4:]) bot.send('PRIVMSG {} :{}'.format(channel, message))
from motobot import command, action @command('say') def say_command(bot, message, database): masters = [ "Moto-chan", "Motoko11", "MotoNyan", "Akahige", "betholas", "Baradium", "Cold_slither", "Drahken" ] if message.nick.lower() not in [x.lower() for x in masters]: return "Check your privilege!" else: args = message.message.split(' ')[1:] if len(args) < 2: return "You must specify both a channel and a message" else: channel = args[0] message = ' '.join(args[1:]) if message.startswith('/me '): message = action(message[4:]) bot.send('PRIVMSG {} :{}'.format(channel, message))
Add MotoNyan to mad hax
Add MotoNyan to mad hax
Python
mit
Motoko11/MotoBot
from motobot import command, action @command('say') def say_command(bot, message, database): masters = [ "Moto-chan", "Motoko11", + "MotoNyan", "Akahige", "betholas", "Baradium", "Cold_slither", "Drahken" ] if message.nick.lower() not in [x.lower() for x in masters]: return "Check your privilege!" else: args = message.message.split(' ')[1:] if len(args) < 2: return "You must specify both a channel and a message" else: channel = args[0] message = ' '.join(args[1:]) if message.startswith('/me '): message = action(message[4:]) bot.send('PRIVMSG {} :{}'.format(channel, message))
Add MotoNyan to mad hax
## Code Before: from motobot import command, action @command('say') def say_command(bot, message, database): masters = [ "Moto-chan", "Motoko11", "Akahige", "betholas", "Baradium", "Cold_slither", "Drahken" ] if message.nick.lower() not in [x.lower() for x in masters]: return "Check your privilege!" else: args = message.message.split(' ')[1:] if len(args) < 2: return "You must specify both a channel and a message" else: channel = args[0] message = ' '.join(args[1:]) if message.startswith('/me '): message = action(message[4:]) bot.send('PRIVMSG {} :{}'.format(channel, message)) ## Instruction: Add MotoNyan to mad hax ## Code After: from motobot import command, action @command('say') def say_command(bot, message, database): masters = [ "Moto-chan", "Motoko11", "MotoNyan", "Akahige", "betholas", "Baradium", "Cold_slither", "Drahken" ] if message.nick.lower() not in [x.lower() for x in masters]: return "Check your privilege!" else: args = message.message.split(' ')[1:] if len(args) < 2: return "You must specify both a channel and a message" else: channel = args[0] message = ' '.join(args[1:]) if message.startswith('/me '): message = action(message[4:]) bot.send('PRIVMSG {} :{}'.format(channel, message))
from motobot import command, action @command('say') def say_command(bot, message, database): masters = [ "Moto-chan", "Motoko11", + "MotoNyan", "Akahige", "betholas", "Baradium", "Cold_slither", "Drahken" ] if message.nick.lower() not in [x.lower() for x in masters]: return "Check your privilege!" else: args = message.message.split(' ')[1:] if len(args) < 2: return "You must specify both a channel and a message" else: channel = args[0] message = ' '.join(args[1:]) if message.startswith('/me '): message = action(message[4:]) bot.send('PRIVMSG {} :{}'.format(channel, message))
ef0a6968dedad74ddd40bd4ae81595be6092f24f
wrapper/__init__.py
wrapper/__init__.py
__version__ = '2.2.0' from libsbol import * import unit_tests
from __future__ import absolute_import __version__ = '2.2.0' from sbol.libsbol import * import sbol.unit_tests
Fix import issue with Python 3.6/Support future Python by forcing absolute import
Fix import issue with Python 3.6/Support future Python by forcing absolute import
Python
apache-2.0
SynBioDex/libSBOL,SynBioDex/libSBOL,SynBioDex/libSBOL,SynBioDex/libSBOL
+ from __future__ import absolute_import + __version__ = '2.2.0' - from libsbol import * + from sbol.libsbol import * - import unit_tests + import sbol.unit_tests
Fix import issue with Python 3.6/Support future Python by forcing absolute import
## Code Before: __version__ = '2.2.0' from libsbol import * import unit_tests ## Instruction: Fix import issue with Python 3.6/Support future Python by forcing absolute import ## Code After: from __future__ import absolute_import __version__ = '2.2.0' from sbol.libsbol import * import sbol.unit_tests
+ from __future__ import absolute_import + __version__ = '2.2.0' - from libsbol import * + from sbol.libsbol import * ? +++++ - import unit_tests + import sbol.unit_tests ? +++++
312c0d463940257cb1f777d3720778550b5bdb2d
bluebottle/organizations/serializers.py
bluebottle/organizations/serializers.py
from rest_framework import serializers from bluebottle.organizations.models import Organization from bluebottle.utils.serializers import URLField class OrganizationSerializer(serializers.ModelSerializer): class Meta: model = Organization fields = ('id', 'name', 'slug', 'address_line1', 'address_line2', 'city', 'state', 'country', 'postal_code', 'phone_number', 'website', 'email') class ManageOrganizationSerializer(serializers.ModelSerializer): slug = serializers.SlugField(required=False, allow_null=True) name = serializers.CharField(required=True) website = URLField(required=False, allow_blank=True) email = serializers.EmailField(required=False, allow_blank=True) class Meta: model = Organization fields = OrganizationSerializer.Meta.fields + ('partner_organizations', 'created', 'updated')
from rest_framework import serializers from bluebottle.organizations.models import Organization from bluebottle.utils.serializers import URLField class OrganizationSerializer(serializers.ModelSerializer): class Meta: model = Organization fields = ('id', 'name', 'slug', 'address_line1', 'address_line2', 'city', 'state', 'country', 'postal_code', 'phone_number', 'website', 'email') class ManageOrganizationSerializer(serializers.ModelSerializer): slug = serializers.SlugField(required=False, allow_null=True) name = serializers.CharField(required=True, allow_blank=True) website = URLField(required=False, allow_blank=True) email = serializers.EmailField(required=False, allow_blank=True) class Meta: model = Organization fields = OrganizationSerializer.Meta.fields + ('partner_organizations', 'created', 'updated')
Revert "Make the name of an organization required"
Revert "Make the name of an organization required" This reverts commit 02140561a29a2b7fe50f7bf2402da566e60be641.
Python
bsd-3-clause
jfterpstra/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle
from rest_framework import serializers from bluebottle.organizations.models import Organization from bluebottle.utils.serializers import URLField class OrganizationSerializer(serializers.ModelSerializer): class Meta: model = Organization fields = ('id', 'name', 'slug', 'address_line1', 'address_line2', 'city', 'state', 'country', 'postal_code', 'phone_number', 'website', 'email') class ManageOrganizationSerializer(serializers.ModelSerializer): slug = serializers.SlugField(required=False, allow_null=True) - name = serializers.CharField(required=True) + name = serializers.CharField(required=True, allow_blank=True) website = URLField(required=False, allow_blank=True) email = serializers.EmailField(required=False, allow_blank=True) class Meta: model = Organization fields = OrganizationSerializer.Meta.fields + ('partner_organizations', 'created', 'updated')
Revert "Make the name of an organization required"
## Code Before: from rest_framework import serializers from bluebottle.organizations.models import Organization from bluebottle.utils.serializers import URLField class OrganizationSerializer(serializers.ModelSerializer): class Meta: model = Organization fields = ('id', 'name', 'slug', 'address_line1', 'address_line2', 'city', 'state', 'country', 'postal_code', 'phone_number', 'website', 'email') class ManageOrganizationSerializer(serializers.ModelSerializer): slug = serializers.SlugField(required=False, allow_null=True) name = serializers.CharField(required=True) website = URLField(required=False, allow_blank=True) email = serializers.EmailField(required=False, allow_blank=True) class Meta: model = Organization fields = OrganizationSerializer.Meta.fields + ('partner_organizations', 'created', 'updated') ## Instruction: Revert "Make the name of an organization required" ## Code After: from rest_framework import serializers from bluebottle.organizations.models import Organization from bluebottle.utils.serializers import URLField class OrganizationSerializer(serializers.ModelSerializer): class Meta: model = Organization fields = ('id', 'name', 'slug', 'address_line1', 'address_line2', 'city', 'state', 'country', 'postal_code', 'phone_number', 'website', 'email') class ManageOrganizationSerializer(serializers.ModelSerializer): slug = serializers.SlugField(required=False, allow_null=True) name = serializers.CharField(required=True, allow_blank=True) website = URLField(required=False, allow_blank=True) email = serializers.EmailField(required=False, allow_blank=True) class Meta: model = Organization fields = OrganizationSerializer.Meta.fields + ('partner_organizations', 'created', 'updated')
from rest_framework import serializers from bluebottle.organizations.models import Organization from bluebottle.utils.serializers import URLField class OrganizationSerializer(serializers.ModelSerializer): class Meta: model = Organization fields = ('id', 'name', 'slug', 'address_line1', 'address_line2', 'city', 'state', 'country', 'postal_code', 'phone_number', 'website', 'email') class ManageOrganizationSerializer(serializers.ModelSerializer): slug = serializers.SlugField(required=False, allow_null=True) - name = serializers.CharField(required=True) + name = serializers.CharField(required=True, allow_blank=True) ? ++++++++++++++++++ website = URLField(required=False, allow_blank=True) email = serializers.EmailField(required=False, allow_blank=True) class Meta: model = Organization fields = OrganizationSerializer.Meta.fields + ('partner_organizations', 'created', 'updated')
5a785f725d68733561a7e5e82c57655e25439ec8
indra/tests/test_grounding_resources.py
indra/tests/test_grounding_resources.py
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map # Namespaces that are not currently handled but still appear in statements exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_exceptional_unhandled(): """Test that exceptional namespaces actually aren't handled. This will catch if we make an update that makes an exceptional namespace become a handled namespace. That way we can update the tests. """ actually_handled = [] for ns in exceptions: if validate_ns(ns): actually_handled.append(ns) assert not actually_handled, actually_handled
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries
Remove exceptional namespaces from test
Remove exceptional namespaces from test
Python
bsd-2-clause
johnbachman/indra,sorgerlab/indra,sorgerlab/belpy,sorgerlab/indra,bgyori/indra,sorgerlab/indra,bgyori/indra,bgyori/indra,johnbachman/indra,johnbachman/indra,sorgerlab/belpy,sorgerlab/belpy
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map - - # Namespaces that are not currently handled but still appear in statements - exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries - - def test_exceptional_unhandled(): - """Test that exceptional namespaces actually aren't handled. - - This will catch if we make an update that makes an exceptional namespace - become a handled namespace. That way we can update the tests. - """ - actually_handled = [] - for ns in exceptions: - if validate_ns(ns): - actually_handled.append(ns) - assert not actually_handled, actually_handled - - - - - - -
Remove exceptional namespaces from test
## Code Before: import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map # Namespaces that are not currently handled but still appear in statements exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_exceptional_unhandled(): """Test that exceptional namespaces actually aren't handled. This will catch if we make an update that makes an exceptional namespace become a handled namespace. That way we can update the tests. """ actually_handled = [] for ns in exceptions: if validate_ns(ns): actually_handled.append(ns) assert not actually_handled, actually_handled ## Instruction: Remove exceptional namespaces from test ## Code After: import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map - - # Namespaces that are not currently handled but still appear in statements - exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries - - - def test_exceptional_unhandled(): - """Test that exceptional namespaces actually aren't handled. - - This will catch if we make an update that makes an exceptional namespace - become a handled namespace. That way we can update the tests. - """ - actually_handled = [] - for ns in exceptions: - if validate_ns(ns): - actually_handled.append(ns) - assert not actually_handled, actually_handled - - - - - -
0c1caf49a18bcd862247cdca7a4efe2f6fc02d93
wafer/management/commands/wafer_talk_video_reviewers.py
wafer/management/commands/wafer_talk_video_reviewers.py
import sys import csv from django.core.management.base import BaseCommand from django.contrib.auth import get_user_model from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL class Command(BaseCommand): help = ("List talks and the associated video_reviewer emails." " Only reviewers for accepted talks are listed") def _video_reviewers(self, options): talks = Talk.objects.filter(status=ACCEPTED) csv_file = csv.writer(sys.stdout) for talk in talks: reviewer = talk.video_reviewer if not reviewer: reviewer = 'NO REVIEWER' row = [x.encode("utf-8") for x in ( talk.title, talk.get_authors_display_name(), reviewer, )] csv_file.writerow(row) def handle(self, *args, **options): self._video_reviewers(options)
import sys import csv from django.core.management.base import BaseCommand from django.contrib.auth import get_user_model from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL class Command(BaseCommand): help = ("List talks and the associated video_reviewer emails." " Only reviewers for accepted talks are listed") def _video_reviewers(self, options): talks = Talk.objects.filter(status=ACCEPTED) csv_file = csv.writer(sys.stdout) for talk in talks: reviewer = talk.video_reviewer if not reviewer: reviewer = 'NO REVIEWER' row = [talk.title, talk.get_authors_display_name(), reviewer, ] csv_file.writerow(row) def handle(self, *args, **options): self._video_reviewers(options)
Drop python2-era manual encode dance
Drop python2-era manual encode dance
Python
isc
CTPUG/wafer,CTPUG/wafer,CTPUG/wafer,CTPUG/wafer
import sys import csv from django.core.management.base import BaseCommand from django.contrib.auth import get_user_model from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL class Command(BaseCommand): help = ("List talks and the associated video_reviewer emails." " Only reviewers for accepted talks are listed") def _video_reviewers(self, options): talks = Talk.objects.filter(status=ACCEPTED) csv_file = csv.writer(sys.stdout) for talk in talks: reviewer = talk.video_reviewer if not reviewer: reviewer = 'NO REVIEWER' - row = [x.encode("utf-8") for x in ( - talk.title, + row = [talk.title, - talk.get_authors_display_name(), + talk.get_authors_display_name(), - reviewer, + reviewer, - )] + ] csv_file.writerow(row) def handle(self, *args, **options): self._video_reviewers(options)
Drop python2-era manual encode dance
## Code Before: import sys import csv from django.core.management.base import BaseCommand from django.contrib.auth import get_user_model from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL class Command(BaseCommand): help = ("List talks and the associated video_reviewer emails." " Only reviewers for accepted talks are listed") def _video_reviewers(self, options): talks = Talk.objects.filter(status=ACCEPTED) csv_file = csv.writer(sys.stdout) for talk in talks: reviewer = talk.video_reviewer if not reviewer: reviewer = 'NO REVIEWER' row = [x.encode("utf-8") for x in ( talk.title, talk.get_authors_display_name(), reviewer, )] csv_file.writerow(row) def handle(self, *args, **options): self._video_reviewers(options) ## Instruction: Drop python2-era manual encode dance ## Code After: import sys import csv from django.core.management.base import BaseCommand from django.contrib.auth import get_user_model from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL class Command(BaseCommand): help = ("List talks and the associated video_reviewer emails." " Only reviewers for accepted talks are listed") def _video_reviewers(self, options): talks = Talk.objects.filter(status=ACCEPTED) csv_file = csv.writer(sys.stdout) for talk in talks: reviewer = talk.video_reviewer if not reviewer: reviewer = 'NO REVIEWER' row = [talk.title, talk.get_authors_display_name(), reviewer, ] csv_file.writerow(row) def handle(self, *args, **options): self._video_reviewers(options)
import sys import csv from django.core.management.base import BaseCommand from django.contrib.auth import get_user_model from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL class Command(BaseCommand): help = ("List talks and the associated video_reviewer emails." " Only reviewers for accepted talks are listed") def _video_reviewers(self, options): talks = Talk.objects.filter(status=ACCEPTED) csv_file = csv.writer(sys.stdout) for talk in talks: reviewer = talk.video_reviewer if not reviewer: reviewer = 'NO REVIEWER' - row = [x.encode("utf-8") for x in ( - talk.title, ? ^^^^^^ + row = [talk.title, ? +++ + ^ - talk.get_authors_display_name(), ? - + talk.get_authors_display_name(), - reviewer, ? - + reviewer, - )] ? - + ] csv_file.writerow(row) def handle(self, *args, **options): self._video_reviewers(options)
fd054790ce32c3918f6edbe824540c09d7efce59
stagehand/providers/__init__.py
stagehand/providers/__init__.py
import asyncio from ..utils import load_plugins, invoke_plugins from .base import ProviderError plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage']) @asyncio.coroutine def start(manager): """ Called when the manager is starting. """ yield from invoke_plugins(plugins, 'start', manager) for name, error in broken_plugins.items(): log.warning('failed to load provider plugin %s: %s', name, error)
import asyncio from ..utils import load_plugins, invoke_plugins from .base import ProviderError plugins, broken_plugins = load_plugins('providers', ['thetvdb']) @asyncio.coroutine def start(manager): """ Called when the manager is starting. """ yield from invoke_plugins(plugins, 'start', manager) for name, error in broken_plugins.items(): log.warning('failed to load provider plugin %s: %s', name, error)
Remove tvrage from active providers as site is shut down
Remove tvrage from active providers as site is shut down
Python
mit
jtackaberry/stagehand,jtackaberry/stagehand
import asyncio from ..utils import load_plugins, invoke_plugins from .base import ProviderError - plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage']) + plugins, broken_plugins = load_plugins('providers', ['thetvdb']) @asyncio.coroutine def start(manager): """ Called when the manager is starting. """ yield from invoke_plugins(plugins, 'start', manager) for name, error in broken_plugins.items(): log.warning('failed to load provider plugin %s: %s', name, error)
Remove tvrage from active providers as site is shut down
## Code Before: import asyncio from ..utils import load_plugins, invoke_plugins from .base import ProviderError plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage']) @asyncio.coroutine def start(manager): """ Called when the manager is starting. """ yield from invoke_plugins(plugins, 'start', manager) for name, error in broken_plugins.items(): log.warning('failed to load provider plugin %s: %s', name, error) ## Instruction: Remove tvrage from active providers as site is shut down ## Code After: import asyncio from ..utils import load_plugins, invoke_plugins from .base import ProviderError plugins, broken_plugins = load_plugins('providers', ['thetvdb']) @asyncio.coroutine def start(manager): """ Called when the manager is starting. """ yield from invoke_plugins(plugins, 'start', manager) for name, error in broken_plugins.items(): log.warning('failed to load provider plugin %s: %s', name, error)
import asyncio from ..utils import load_plugins, invoke_plugins from .base import ProviderError - plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage']) ? ---------- + plugins, broken_plugins = load_plugins('providers', ['thetvdb']) @asyncio.coroutine def start(manager): """ Called when the manager is starting. """ yield from invoke_plugins(plugins, 'start', manager) for name, error in broken_plugins.items(): log.warning('failed to load provider plugin %s: %s', name, error)
64bc8ff452d03c7bb026be0b2edd9a047a88b386
foyer/forcefields/forcefields.py
foyer/forcefields/forcefields.py
import os import glob from pkg_resources import resource_filename from foyer import Forcefield def get_ff_path(): return [resource_filename('foyer', 'forcefields')] def get_forcefield_paths(forcefield_name=None): for dir_path in get_ff_path(): file_pattern = os.path.join(dir_path, 'xml/*.xml') file_paths = [file_path for file_path in glob.glob(file_pattern)] return file_paths def get_forcefield(name=None): if name is None: raise ValueError('Need a force field name') file_paths = get_forcefield_paths() try: ff_path = next(val for val in file_paths if name in val) except StopIteration: raise ValueError('Could not find force field with name {}' ' in path {}'.format(name, get_ff_path())) return Forcefield(ff_path) load_OPLSAA = get_forcefield(name='oplsaa') load_TRAPPE_UA = get_forcefield(name='trappe-ua')
import os import glob from pkg_resources import resource_filename from foyer import Forcefield def get_ff_path(): return [resource_filename('foyer', 'forcefields')] def get_forcefield_paths(forcefield_name=None): for dir_path in get_ff_path(): file_pattern = os.path.join(dir_path, 'xml/*.xml') file_paths = [file_path for file_path in glob.glob(file_pattern)] return file_paths def get_forcefield(name=None): if name is None: raise ValueError('Need a force field name') file_paths = get_forcefield_paths() try: ff_path = next(val for val in file_paths if name in val) except StopIteration: raise ValueError('Could not find force field with name {}' ' in path {}'.format(name, get_ff_path())) return Forcefield(ff_path) def load_OPLSAA(): return get_forcefield(name='oplsaa') def load_TRAPPE_UA(): return get_forcefield(name='trappe-ua') load_OPLSAA = load_OPLSAA load_TRAPPE_UA = load_TRAPPE_UA
Make discrete functions for each force field
Make discrete functions for each force field
Python
mit
mosdef-hub/foyer,mosdef-hub/foyer,iModels/foyer,iModels/foyer
import os import glob from pkg_resources import resource_filename from foyer import Forcefield def get_ff_path(): return [resource_filename('foyer', 'forcefields')] def get_forcefield_paths(forcefield_name=None): for dir_path in get_ff_path(): file_pattern = os.path.join(dir_path, 'xml/*.xml') file_paths = [file_path for file_path in glob.glob(file_pattern)] return file_paths def get_forcefield(name=None): if name is None: raise ValueError('Need a force field name') file_paths = get_forcefield_paths() try: ff_path = next(val for val in file_paths if name in val) except StopIteration: raise ValueError('Could not find force field with name {}' ' in path {}'.format(name, get_ff_path())) return Forcefield(ff_path) - load_OPLSAA = get_forcefield(name='oplsaa') - load_TRAPPE_UA = get_forcefield(name='trappe-ua') + def load_OPLSAA(): + return get_forcefield(name='oplsaa') + + + def load_TRAPPE_UA(): + return get_forcefield(name='trappe-ua') + + + load_OPLSAA = load_OPLSAA + load_TRAPPE_UA = load_TRAPPE_UA +
Make discrete functions for each force field
## Code Before: import os import glob from pkg_resources import resource_filename from foyer import Forcefield def get_ff_path(): return [resource_filename('foyer', 'forcefields')] def get_forcefield_paths(forcefield_name=None): for dir_path in get_ff_path(): file_pattern = os.path.join(dir_path, 'xml/*.xml') file_paths = [file_path for file_path in glob.glob(file_pattern)] return file_paths def get_forcefield(name=None): if name is None: raise ValueError('Need a force field name') file_paths = get_forcefield_paths() try: ff_path = next(val for val in file_paths if name in val) except StopIteration: raise ValueError('Could not find force field with name {}' ' in path {}'.format(name, get_ff_path())) return Forcefield(ff_path) load_OPLSAA = get_forcefield(name='oplsaa') load_TRAPPE_UA = get_forcefield(name='trappe-ua') ## Instruction: Make discrete functions for each force field ## Code After: import os import glob from pkg_resources import resource_filename from foyer import Forcefield def get_ff_path(): return [resource_filename('foyer', 'forcefields')] def get_forcefield_paths(forcefield_name=None): for dir_path in get_ff_path(): file_pattern = os.path.join(dir_path, 'xml/*.xml') file_paths = [file_path for file_path in glob.glob(file_pattern)] return file_paths def get_forcefield(name=None): if name is None: raise ValueError('Need a force field name') file_paths = get_forcefield_paths() try: ff_path = next(val for val in file_paths if name in val) except StopIteration: raise ValueError('Could not find force field with name {}' ' in path {}'.format(name, get_ff_path())) return Forcefield(ff_path) def load_OPLSAA(): return get_forcefield(name='oplsaa') def load_TRAPPE_UA(): return get_forcefield(name='trappe-ua') load_OPLSAA = load_OPLSAA load_TRAPPE_UA = load_TRAPPE_UA
import os import glob from pkg_resources import resource_filename from foyer import Forcefield def get_ff_path(): return [resource_filename('foyer', 'forcefields')] def get_forcefield_paths(forcefield_name=None): for dir_path in get_ff_path(): file_pattern = os.path.join(dir_path, 'xml/*.xml') file_paths = [file_path for file_path in glob.glob(file_pattern)] return file_paths def get_forcefield(name=None): if name is None: raise ValueError('Need a force field name') file_paths = get_forcefield_paths() try: ff_path = next(val for val in file_paths if name in val) except StopIteration: raise ValueError('Could not find force field with name {}' ' in path {}'.format(name, get_ff_path())) return Forcefield(ff_path) - load_OPLSAA = get_forcefield(name='oplsaa') - load_TRAPPE_UA = get_forcefield(name='trappe-ua') + + def load_OPLSAA(): + return get_forcefield(name='oplsaa') + + + def load_TRAPPE_UA(): + return get_forcefield(name='trappe-ua') + + + load_OPLSAA = load_OPLSAA + load_TRAPPE_UA = load_TRAPPE_UA
5dcec96b7af384f7f753cb2d67d7cbd0c361c504
tests/helpers.py
tests/helpers.py
from __future__ import unicode_literals import json from elasticsearch import ( Elasticsearch, TransportError ) ELASTICSEARCH_URL = "localhost" conn = Elasticsearch(ELASTICSEARCH_URL) def homogeneous(a, b): json.dumps(a).should.equal(json.dumps(b)) def heterogeneous(a, b): json.dumps(a).shouldnt.equal(json.dumps(b)) def add_document(index, document, **kwargs): document_type = "my_doc_type" conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs) def clean_elasticsearch(context): _delete_es_index("foo") def prepare_elasticsearch(context): clean_elasticsearch(context) _create_foo_index() conn.cluster.health(wait_for_status='yellow') def _create_foo_index(): conn.indices.create(index="foo", ignore=400) def _delete_es_index(index): conn.indices.delete(index=index, ignore=[400, 404]) prepare_data = [ prepare_elasticsearch ] cleanup_data = [ clean_elasticsearch ]
from __future__ import unicode_literals import json from elasticsearch import ( Elasticsearch, TransportError ) ELASTICSEARCH_URL = "localhost" conn = Elasticsearch(ELASTICSEARCH_URL) def homogeneous(a, b): json.dumps(a).should.equal(json.dumps(b)) def heterogeneous(a, b): json.dumps(a).shouldnt.equal(json.dumps(b)) def add_document(index, document, **kwargs): if "doc_type" not in kwargs: # Allow overriding doc type defaults doc_type = "my_doc_type" kwargs["doc_type"] = doc_type conn.create(index=index, body=document, refresh=True, **kwargs) def clean_elasticsearch(context): _delete_es_index("foo") def prepare_elasticsearch(context): clean_elasticsearch(context) _create_foo_index() conn.cluster.health(wait_for_status='yellow') def _create_foo_index(): conn.indices.create(index="foo", ignore=400) def _delete_es_index(index): conn.indices.delete(index=index, ignore=[400, 404]) prepare_data = [ prepare_elasticsearch ] cleanup_data = [ clean_elasticsearch ]
Allow overriding doc type defaults
Allow overriding doc type defaults
Python
mit
Yipit/pyeqs
from __future__ import unicode_literals import json from elasticsearch import ( Elasticsearch, TransportError ) ELASTICSEARCH_URL = "localhost" conn = Elasticsearch(ELASTICSEARCH_URL) def homogeneous(a, b): json.dumps(a).should.equal(json.dumps(b)) def heterogeneous(a, b): json.dumps(a).shouldnt.equal(json.dumps(b)) def add_document(index, document, **kwargs): + if "doc_type" not in kwargs: + # Allow overriding doc type defaults - document_type = "my_doc_type" + doc_type = "my_doc_type" + kwargs["doc_type"] = doc_type - conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs) + conn.create(index=index, body=document, refresh=True, **kwargs) def clean_elasticsearch(context): _delete_es_index("foo") def prepare_elasticsearch(context): clean_elasticsearch(context) _create_foo_index() conn.cluster.health(wait_for_status='yellow') def _create_foo_index(): conn.indices.create(index="foo", ignore=400) def _delete_es_index(index): conn.indices.delete(index=index, ignore=[400, 404]) prepare_data = [ prepare_elasticsearch ] cleanup_data = [ clean_elasticsearch ]
Allow overriding doc type defaults
## Code Before: from __future__ import unicode_literals import json from elasticsearch import ( Elasticsearch, TransportError ) ELASTICSEARCH_URL = "localhost" conn = Elasticsearch(ELASTICSEARCH_URL) def homogeneous(a, b): json.dumps(a).should.equal(json.dumps(b)) def heterogeneous(a, b): json.dumps(a).shouldnt.equal(json.dumps(b)) def add_document(index, document, **kwargs): document_type = "my_doc_type" conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs) def clean_elasticsearch(context): _delete_es_index("foo") def prepare_elasticsearch(context): clean_elasticsearch(context) _create_foo_index() conn.cluster.health(wait_for_status='yellow') def _create_foo_index(): conn.indices.create(index="foo", ignore=400) def _delete_es_index(index): conn.indices.delete(index=index, ignore=[400, 404]) prepare_data = [ prepare_elasticsearch ] cleanup_data = [ clean_elasticsearch ] ## Instruction: Allow overriding doc type defaults ## Code After: from __future__ import unicode_literals import json from elasticsearch import ( Elasticsearch, TransportError ) ELASTICSEARCH_URL = "localhost" conn = Elasticsearch(ELASTICSEARCH_URL) def homogeneous(a, b): json.dumps(a).should.equal(json.dumps(b)) def heterogeneous(a, b): json.dumps(a).shouldnt.equal(json.dumps(b)) def add_document(index, document, **kwargs): if "doc_type" not in kwargs: # Allow overriding doc type defaults doc_type = "my_doc_type" kwargs["doc_type"] = doc_type conn.create(index=index, body=document, refresh=True, **kwargs) def clean_elasticsearch(context): _delete_es_index("foo") def prepare_elasticsearch(context): clean_elasticsearch(context) _create_foo_index() conn.cluster.health(wait_for_status='yellow') def _create_foo_index(): conn.indices.create(index="foo", ignore=400) def _delete_es_index(index): conn.indices.delete(index=index, ignore=[400, 404]) prepare_data = [ prepare_elasticsearch ] cleanup_data = [ clean_elasticsearch ]
from __future__ import unicode_literals import json from elasticsearch import ( Elasticsearch, TransportError ) ELASTICSEARCH_URL = "localhost" conn = Elasticsearch(ELASTICSEARCH_URL) def homogeneous(a, b): json.dumps(a).should.equal(json.dumps(b)) def heterogeneous(a, b): json.dumps(a).shouldnt.equal(json.dumps(b)) def add_document(index, document, **kwargs): + if "doc_type" not in kwargs: + # Allow overriding doc type defaults - document_type = "my_doc_type" ? ----- + doc_type = "my_doc_type" ? ++++ + kwargs["doc_type"] = doc_type - conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs) ? ------------------------ + conn.create(index=index, body=document, refresh=True, **kwargs) def clean_elasticsearch(context): _delete_es_index("foo") def prepare_elasticsearch(context): clean_elasticsearch(context) _create_foo_index() conn.cluster.health(wait_for_status='yellow') def _create_foo_index(): conn.indices.create(index="foo", ignore=400) def _delete_es_index(index): conn.indices.delete(index=index, ignore=[400, 404]) prepare_data = [ prepare_elasticsearch ] cleanup_data = [ clean_elasticsearch ]
146e35f48774173c2000b8a9790cdbe6925ba94a
opps/contrib/multisite/admin.py
opps/contrib/multisite/admin.py
from django.contrib import admin from .models import SitePermission admin.site.register(SitePermission)
from django.contrib import admin from django.utils import timezone from .models import SitePermission class AdminViewPermission(admin.ModelAdmin): def queryset(self, request): queryset = super(AdminViewPermission, self).queryset(request) try: sitepermission = SitePermission.objects.get( user=request.user, date_available__lte=timezone.now(), published=True) return queryset.filter(site_iid=sitepermission.site_iid) except SitePermission.DoesNotExist: pass return queryset admin.site.register(SitePermission)
Create AdminViewPermission on contrib multisite
Create AdminViewPermission on contrib multisite
Python
mit
opps/opps,YACOWS/opps,williamroot/opps,jeanmask/opps,opps/opps,williamroot/opps,jeanmask/opps,YACOWS/opps,YACOWS/opps,YACOWS/opps,opps/opps,jeanmask/opps,williamroot/opps,opps/opps,jeanmask/opps,williamroot/opps
from django.contrib import admin + from django.utils import timezone + from .models import SitePermission + + + class AdminViewPermission(admin.ModelAdmin): + + def queryset(self, request): + queryset = super(AdminViewPermission, self).queryset(request) + try: + sitepermission = SitePermission.objects.get( + user=request.user, + date_available__lte=timezone.now(), + published=True) + return queryset.filter(site_iid=sitepermission.site_iid) + except SitePermission.DoesNotExist: + pass + return queryset admin.site.register(SitePermission)
Create AdminViewPermission on contrib multisite
## Code Before: from django.contrib import admin from .models import SitePermission admin.site.register(SitePermission) ## Instruction: Create AdminViewPermission on contrib multisite ## Code After: from django.contrib import admin from django.utils import timezone from .models import SitePermission class AdminViewPermission(admin.ModelAdmin): def queryset(self, request): queryset = super(AdminViewPermission, self).queryset(request) try: sitepermission = SitePermission.objects.get( user=request.user, date_available__lte=timezone.now(), published=True) return queryset.filter(site_iid=sitepermission.site_iid) except SitePermission.DoesNotExist: pass return queryset admin.site.register(SitePermission)
from django.contrib import admin + from django.utils import timezone + from .models import SitePermission + class AdminViewPermission(admin.ModelAdmin): + + def queryset(self, request): + queryset = super(AdminViewPermission, self).queryset(request) + try: + sitepermission = SitePermission.objects.get( + user=request.user, + date_available__lte=timezone.now(), + published=True) + return queryset.filter(site_iid=sitepermission.site_iid) + except SitePermission.DoesNotExist: + pass + return queryset + + admin.site.register(SitePermission)
38ce0d6b0433a68787c18691407c815d4eb1fdb2
txscrypt/__init__.py
txscrypt/__init__.py
from txscrypt.wrapper import computeKey, verifyPassword from txscrypt._version import __version__ __all__ = ["computeKey", "verifyPassword"]
from txscrypt.wrapper import checkPassword, computeKey from txscrypt._version import __version__ __all__ = ["verifyPassword", "computeKey"]
Make checkPassword the only public API, remove verifyPassword
Make checkPassword the only public API, remove verifyPassword
Python
isc
lvh/txscrypt
- from txscrypt.wrapper import computeKey, verifyPassword + from txscrypt.wrapper import checkPassword, computeKey from txscrypt._version import __version__ - __all__ = ["computeKey", "verifyPassword"] + __all__ = ["verifyPassword", "computeKey"]
Make checkPassword the only public API, remove verifyPassword
## Code Before: from txscrypt.wrapper import computeKey, verifyPassword from txscrypt._version import __version__ __all__ = ["computeKey", "verifyPassword"] ## Instruction: Make checkPassword the only public API, remove verifyPassword ## Code After: from txscrypt.wrapper import checkPassword, computeKey from txscrypt._version import __version__ __all__ = ["verifyPassword", "computeKey"]
- from txscrypt.wrapper import computeKey, verifyPassword + from txscrypt.wrapper import checkPassword, computeKey from txscrypt._version import __version__ - __all__ = ["computeKey", "verifyPassword"] + __all__ = ["verifyPassword", "computeKey"]
c39260e64c8820bad9243c35f10b352419425810
marble/tests/test_exposure.py
marble/tests/test_exposure.py
""" Tests for the exposure computation """ from nose.tools import * import marble as mb # Test maximum value of exposure # Test maximum value of isolation # Test minimum of exposure # Test minimum of isolation
""" Tests for the exposure computation """ from __future__ import division from nose.tools import * import itertools import marble as mb # # Synthetic data for tests # def segregated_city(): """ perfect segregation """ city = {"A":{1:7, 2:0, 3:0}, "B":{1:0, 2:0, 3:14}, "C":{1:0, 2:42, 3:0}} return city def two_way_city(): """ perfect two-way exposure for 1 and 2 """ city = {"A":{1:7, 2:13, 3:0}, "B":{1:7, 2:13, 3:0}, "C":{1:0, 2:0, 3:37}} return city def uniform_city(): """ Uniform representation """ city = {"A":{1:1, 2:10, 3:7}, "B":{1:2, 2:20, 3:14}, "C":{1:4, 2:40, 3:28}} return city # # Test # class TestExposure(object): def test_maximum_isolation(city): city = segregated_city() exp = mb.exposure(city) N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} N_tot = sum(N_cl.values()) for c in exp: assert_almost_equal(exp[c][c][0], N_tot/N_cl[c], places=3) def test_minimum_exposure(city): city = segregated_city() exp = mb.exposure(city) for c0,c1 in itertools.permutations([1,2,3], 2): assert_almost_equal(exp[c0][c1][0], 0.0) def test_maximum_exposure(city): city = two_way_city() exp = mb.exposure(city) N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} N_tot = sum(N_cl.values()) assert_almost_equal(exp[2][1][0], N_tot/(N_cl[1]+N_cl[2]), places=3) def test_minimum_isolation(city): city = uniform_city() exp = mb.exposure(city) for c in [1,2,3]: assert_almost_equal(exp[c][c][0], 1.0, places=3)
Write tests for the exposure
Write tests for the exposure
Python
bsd-3-clause
walkerke/marble,scities/marble
""" Tests for the exposure computation """ + from __future__ import division from nose.tools import * + import itertools import marble as mb - # Test maximum value of exposure - # Test maximum value of isolation - # Test minimum of exposure - # Test minimum of isolation + # + # Synthetic data for tests + # + def segregated_city(): + """ perfect segregation """ + city = {"A":{1:7, 2:0, 3:0}, + "B":{1:0, 2:0, 3:14}, + "C":{1:0, 2:42, 3:0}} + return city + + def two_way_city(): + """ perfect two-way exposure for 1 and 2 """ + city = {"A":{1:7, 2:13, 3:0}, + "B":{1:7, 2:13, 3:0}, + "C":{1:0, 2:0, 3:37}} + return city + + def uniform_city(): + """ Uniform representation """ + city = {"A":{1:1, 2:10, 3:7}, + "B":{1:2, 2:20, 3:14}, + "C":{1:4, 2:40, 3:28}} + return city + + + + # + # Test + # + class TestExposure(object): + + def test_maximum_isolation(city): + city = segregated_city() + exp = mb.exposure(city) + N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} + N_tot = sum(N_cl.values()) + for c in exp: + assert_almost_equal(exp[c][c][0], + N_tot/N_cl[c], + places=3) + + def test_minimum_exposure(city): + city = segregated_city() + exp = mb.exposure(city) + for c0,c1 in itertools.permutations([1,2,3], 2): + assert_almost_equal(exp[c0][c1][0], + 0.0) + + def test_maximum_exposure(city): + city = two_way_city() + exp = mb.exposure(city) + N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} + N_tot = sum(N_cl.values()) + assert_almost_equal(exp[2][1][0], + N_tot/(N_cl[1]+N_cl[2]), + places=3) + + def test_minimum_isolation(city): + city = uniform_city() + exp = mb.exposure(city) + for c in [1,2,3]: + assert_almost_equal(exp[c][c][0], + 1.0, + places=3) + +
Write tests for the exposure
## Code Before: """ Tests for the exposure computation """ from nose.tools import * import marble as mb # Test maximum value of exposure # Test maximum value of isolation # Test minimum of exposure # Test minimum of isolation ## Instruction: Write tests for the exposure ## Code After: """ Tests for the exposure computation """ from __future__ import division from nose.tools import * import itertools import marble as mb # # Synthetic data for tests # def segregated_city(): """ perfect segregation """ city = {"A":{1:7, 2:0, 3:0}, "B":{1:0, 2:0, 3:14}, "C":{1:0, 2:42, 3:0}} return city def two_way_city(): """ perfect two-way exposure for 1 and 2 """ city = {"A":{1:7, 2:13, 3:0}, "B":{1:7, 2:13, 3:0}, "C":{1:0, 2:0, 3:37}} return city def uniform_city(): """ Uniform representation """ city = {"A":{1:1, 2:10, 3:7}, "B":{1:2, 2:20, 3:14}, "C":{1:4, 2:40, 3:28}} return city # # Test # class TestExposure(object): def test_maximum_isolation(city): city = segregated_city() exp = mb.exposure(city) N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} N_tot = sum(N_cl.values()) for c in exp: assert_almost_equal(exp[c][c][0], N_tot/N_cl[c], places=3) def test_minimum_exposure(city): city = segregated_city() exp = mb.exposure(city) for c0,c1 in itertools.permutations([1,2,3], 2): assert_almost_equal(exp[c0][c1][0], 0.0) def test_maximum_exposure(city): city = two_way_city() exp = mb.exposure(city) N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} N_tot = sum(N_cl.values()) assert_almost_equal(exp[2][1][0], N_tot/(N_cl[1]+N_cl[2]), places=3) def test_minimum_isolation(city): city = uniform_city() exp = mb.exposure(city) for c in [1,2,3]: assert_almost_equal(exp[c][c][0], 1.0, places=3)
""" Tests for the exposure computation """ + from __future__ import division from nose.tools import * + import itertools import marble as mb - # Test maximum value of exposure - # Test maximum value of isolation - # Test minimum of exposure - # Test minimum of isolation + + # + # Synthetic data for tests + # + def segregated_city(): + """ perfect segregation """ + city = {"A":{1:7, 2:0, 3:0}, + "B":{1:0, 2:0, 3:14}, + "C":{1:0, 2:42, 3:0}} + return city + + def two_way_city(): + """ perfect two-way exposure for 1 and 2 """ + city = {"A":{1:7, 2:13, 3:0}, + "B":{1:7, 2:13, 3:0}, + "C":{1:0, 2:0, 3:37}} + return city + + def uniform_city(): + """ Uniform representation """ + city = {"A":{1:1, 2:10, 3:7}, + "B":{1:2, 2:20, 3:14}, + "C":{1:4, 2:40, 3:28}} + return city + + + + # + # Test + # + class TestExposure(object): + + def test_maximum_isolation(city): + city = segregated_city() + exp = mb.exposure(city) + N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} + N_tot = sum(N_cl.values()) + for c in exp: + assert_almost_equal(exp[c][c][0], + N_tot/N_cl[c], + places=3) + + def test_minimum_exposure(city): + city = segregated_city() + exp = mb.exposure(city) + for c0,c1 in itertools.permutations([1,2,3], 2): + assert_almost_equal(exp[c0][c1][0], + 0.0) + + def test_maximum_exposure(city): + city = two_way_city() + exp = mb.exposure(city) + N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]} + N_tot = sum(N_cl.values()) + assert_almost_equal(exp[2][1][0], + N_tot/(N_cl[1]+N_cl[2]), + places=3) + + def test_minimum_isolation(city): + city = uniform_city() + exp = mb.exposure(city) + for c in [1,2,3]: + assert_almost_equal(exp[c][c][0], + 1.0, + places=3) +
a9d4fab047249fbf5db26385779902d0f7483057
qsimcirq/__init__.py
qsimcirq/__init__.py
from .qsim_circuit import * from .qsim_simulator import * from .qsimh_simulator import *
from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator from .qsimh_simulator import QSimhSimulator
Replace star imports to fix mypy issue.
Replace star imports to fix mypy issue.
Python
apache-2.0
quantumlib/qsim,quantumlib/qsim,quantumlib/qsim,quantumlib/qsim
- from .qsim_circuit import * - from .qsim_simulator import * + from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit + from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator - from .qsimh_simulator import * + from .qsimh_simulator import QSimhSimulator - -
Replace star imports to fix mypy issue.
## Code Before: from .qsim_circuit import * from .qsim_simulator import * from .qsimh_simulator import * ## Instruction: Replace star imports to fix mypy issue. ## Code After: from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator from .qsimh_simulator import QSimhSimulator
- from .qsim_circuit import * - from .qsim_simulator import * + from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit + from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator - from .qsimh_simulator import * ? ^ + from .qsimh_simulator import QSimhSimulator ? ^^^^^^^^^^^^^^ - -
5b50b96b35c678ca17b069630875a9d86e2cbca3
scripts/i18n/commons.py
scripts/i18n/commons.py
msg = { 'en': { 'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].', 'commons-file-now-available' : u'File is now available on Wikimedia Commons.', 'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}', }, 'qqq': { 'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', 'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.', 'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.', }, }
msg = { 'en': { 'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].', 'commons-file-now-available' : u'File is now available on Wikimedia Commons.', }, 'qqq': { 'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', 'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.', }, }
Remove the template for now.
Remove the template for now. git-svn-id: 9a050473c2aca1e14f53d73349e19b938c2cf203@9344 6a7f98fc-eeb0-4dc1-a6e2-c2c589a08aa6
Python
mit
legoktm/pywikipedia-rewrite
msg = { 'en': { 'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].', 'commons-file-now-available' : u'File is now available on Wikimedia Commons.', - 'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}', }, 'qqq': { - 'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', + 'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', - 'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.', + 'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.', - 'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.', }, }
Remove the template for now.
## Code Before: msg = { 'en': { 'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].', 'commons-file-now-available' : u'File is now available on Wikimedia Commons.', 'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}', }, 'qqq': { 'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', 'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.', 'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.', }, } ## Instruction: Remove the template for now. ## Code After: msg = { 'en': { 'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].', 'commons-file-now-available' : u'File is now available on Wikimedia Commons.', }, 'qqq': { 'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', 'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.', }, }
msg = { 'en': { 'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].', 'commons-file-now-available' : u'File is now available on Wikimedia Commons.', - 'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}', }, 'qqq': { - 'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', + 'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.', ? + - 'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.', + 'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.', ? + - 'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.', }, }
b6978852775bb48e400a31a1e464d7b596db13f2
fsictools.py
fsictools.py
# Version number keeps track with the main `fsic` module from fsic import __version__ import re from typing import List from fsic import BaseModel, Symbol def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame': """Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame return DataFrame([s._asdict() for s in symbols]) def model_to_dataframe(model: BaseModel) -> 'DataFrame': """Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame df = DataFrame(model.values.T, index=model.span, columns=model.names) df['status'] = model.status df['iterations'] = model.iterations return df
# Version number keeps track with the main `fsic` module from fsic import __version__ import re from typing import List from fsic import BaseModel, Symbol def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame': """Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame return DataFrame([s._asdict() for s in symbols]) def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame': """Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame df = DataFrame(model.values.T, index=model.span, columns=model.names) df['status'] = model.status df['iterations'] = model.iterations return df
Update type annotations to reference package names in full
TYP: Update type annotations to reference package names in full
Python
mit
ChrisThoung/fsic
# Version number keeps track with the main `fsic` module from fsic import __version__ import re from typing import List from fsic import BaseModel, Symbol - def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame': + def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame': """Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame return DataFrame([s._asdict() for s in symbols]) - def model_to_dataframe(model: BaseModel) -> 'DataFrame': + def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame': """Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame df = DataFrame(model.values.T, index=model.span, columns=model.names) df['status'] = model.status df['iterations'] = model.iterations return df
Update type annotations to reference package names in full
## Code Before: # Version number keeps track with the main `fsic` module from fsic import __version__ import re from typing import List from fsic import BaseModel, Symbol def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame': """Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame return DataFrame([s._asdict() for s in symbols]) def model_to_dataframe(model: BaseModel) -> 'DataFrame': """Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame df = DataFrame(model.values.T, index=model.span, columns=model.names) df['status'] = model.status df['iterations'] = model.iterations return df ## Instruction: Update type annotations to reference package names in full ## Code After: # Version number keeps track with the main `fsic` module from fsic import __version__ import re from typing import List from fsic import BaseModel, Symbol def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame': """Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame return DataFrame([s._asdict() for s in symbols]) def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame': """Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame df = DataFrame(model.values.T, index=model.span, columns=model.names) df['status'] = model.status df['iterations'] = model.iterations return df
# Version number keeps track with the main `fsic` module from fsic import __version__ import re from typing import List from fsic import BaseModel, Symbol - def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame': + def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame': ? +++++++ """Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame return DataFrame([s._asdict() for s in symbols]) - def model_to_dataframe(model: BaseModel) -> 'DataFrame': + def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame': ? +++++++ """Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**.""" from pandas import DataFrame df = DataFrame(model.values.T, index=model.span, columns=model.names) df['status'] = model.status df['iterations'] = model.iterations return df
4965511fdb9843233e84a8aa9aa0414bf1c02133
mail/views.py
mail/views.py
from django.shortcuts import redirect from django.http import JsonResponse from django.core.mail import EmailMessage from django.middleware import csrf from rest_framework.decorators import api_view @api_view(['POST', 'GET']) def send_contact_message(request): if request.method == 'POST': to_address = request.POST.get("to_address", "").split(',') from_name = request.POST.get("from_name", "") from_address = request.POST.get("from_address", "") from_string = '{} <{}>'.format(from_name, from_address) subject = request.POST.get("subject", "") message_body = request.POST.get("message_body", "") csrf_token = request.POST.get("csrfmiddlewaretoken", "") email = EmailMessage(subject, message_body, 'noreply@openstax.org', to_address, reply_to=[from_string]) email.send(fail_silently=False) #return redirect('/contact-thank-you') data = {'subject': subject, 'message_body': message_body, 'to_address': to_address, 'reply_to': [from_string], 'from_address': 'noreply@openstax.org', 'csrf_token': csrf_token, } return JsonResponse(data) # if this is not posting a message, let's send the csfr token back else: csrf_token = csrf.get_token(request) data = {'csrf_token': csrf_token} return JsonResponse(data)
from django.shortcuts import redirect from django.http import JsonResponse from django.core.mail import EmailMessage from django.middleware import csrf from rest_framework.decorators import api_view @api_view(['POST', 'GET']) def send_contact_message(request): if request.method == 'POST': to_address = request.POST.get("to_address", "").split(',') from_name = request.POST.get("from_name", "") from_address = request.POST.get("from_address", "") from_string = '{} <{}>'.format(from_name, from_address) subject = request.POST.get("subject", "") message_body = request.POST.get("message_body", "") email = EmailMessage(subject, message_body, 'noreply@openstax.org', to_address, reply_to=[from_string]) email.send() return redirect('/contact-thank-you') # if this is not posting a message, let's send the csfr token back else: csrf_token = csrf.get_token(request) data = {'csrf_token': csrf_token} return JsonResponse(data)
Revert "return json of message being sent to debug mail issue"
Revert "return json of message being sent to debug mail issue"
Python
agpl-3.0
openstax/openstax-cms,Connexions/openstax-cms,openstax/openstax-cms,openstax/openstax-cms,openstax/openstax-cms,Connexions/openstax-cms
from django.shortcuts import redirect from django.http import JsonResponse from django.core.mail import EmailMessage from django.middleware import csrf from rest_framework.decorators import api_view @api_view(['POST', 'GET']) def send_contact_message(request): if request.method == 'POST': to_address = request.POST.get("to_address", "").split(',') from_name = request.POST.get("from_name", "") from_address = request.POST.get("from_address", "") from_string = '{} <{}>'.format(from_name, from_address) subject = request.POST.get("subject", "") message_body = request.POST.get("message_body", "") - csrf_token = request.POST.get("csrfmiddlewaretoken", "") email = EmailMessage(subject, message_body, 'noreply@openstax.org', to_address, reply_to=[from_string]) - email.send(fail_silently=False) + email.send() - #return redirect('/contact-thank-you') + return redirect('/contact-thank-you') - data = {'subject': subject, - 'message_body': message_body, - 'to_address': to_address, - 'reply_to': [from_string], - 'from_address': 'noreply@openstax.org', - 'csrf_token': csrf_token, - } - - return JsonResponse(data) # if this is not posting a message, let's send the csfr token back else: csrf_token = csrf.get_token(request) data = {'csrf_token': csrf_token} return JsonResponse(data)
Revert "return json of message being sent to debug mail issue"
## Code Before: from django.shortcuts import redirect from django.http import JsonResponse from django.core.mail import EmailMessage from django.middleware import csrf from rest_framework.decorators import api_view @api_view(['POST', 'GET']) def send_contact_message(request): if request.method == 'POST': to_address = request.POST.get("to_address", "").split(',') from_name = request.POST.get("from_name", "") from_address = request.POST.get("from_address", "") from_string = '{} <{}>'.format(from_name, from_address) subject = request.POST.get("subject", "") message_body = request.POST.get("message_body", "") csrf_token = request.POST.get("csrfmiddlewaretoken", "") email = EmailMessage(subject, message_body, 'noreply@openstax.org', to_address, reply_to=[from_string]) email.send(fail_silently=False) #return redirect('/contact-thank-you') data = {'subject': subject, 'message_body': message_body, 'to_address': to_address, 'reply_to': [from_string], 'from_address': 'noreply@openstax.org', 'csrf_token': csrf_token, } return JsonResponse(data) # if this is not posting a message, let's send the csfr token back else: csrf_token = csrf.get_token(request) data = {'csrf_token': csrf_token} return JsonResponse(data) ## Instruction: Revert "return json of message being sent to debug mail issue" ## Code After: from django.shortcuts import redirect from django.http import JsonResponse from django.core.mail import EmailMessage from django.middleware import csrf from rest_framework.decorators import api_view @api_view(['POST', 'GET']) def send_contact_message(request): if request.method == 'POST': to_address = request.POST.get("to_address", "").split(',') from_name = request.POST.get("from_name", "") from_address = request.POST.get("from_address", "") from_string = '{} <{}>'.format(from_name, from_address) subject = request.POST.get("subject", "") message_body = request.POST.get("message_body", "") email = EmailMessage(subject, message_body, 'noreply@openstax.org', to_address, reply_to=[from_string]) email.send() return redirect('/contact-thank-you') # if this is not posting a message, let's send the csfr token back else: csrf_token = csrf.get_token(request) data = {'csrf_token': csrf_token} return JsonResponse(data)
from django.shortcuts import redirect from django.http import JsonResponse from django.core.mail import EmailMessage from django.middleware import csrf from rest_framework.decorators import api_view @api_view(['POST', 'GET']) def send_contact_message(request): if request.method == 'POST': to_address = request.POST.get("to_address", "").split(',') from_name = request.POST.get("from_name", "") from_address = request.POST.get("from_address", "") from_string = '{} <{}>'.format(from_name, from_address) subject = request.POST.get("subject", "") message_body = request.POST.get("message_body", "") - csrf_token = request.POST.get("csrfmiddlewaretoken", "") email = EmailMessage(subject, message_body, 'noreply@openstax.org', to_address, reply_to=[from_string]) - email.send(fail_silently=False) + email.send() - #return redirect('/contact-thank-you') ? - + return redirect('/contact-thank-you') - data = {'subject': subject, - 'message_body': message_body, - 'to_address': to_address, - 'reply_to': [from_string], - 'from_address': 'noreply@openstax.org', - 'csrf_token': csrf_token, - } - - return JsonResponse(data) # if this is not posting a message, let's send the csfr token back else: csrf_token = csrf.get_token(request) data = {'csrf_token': csrf_token} return JsonResponse(data)
94e070ec33dbc86e38de4839be9461db3a301685
inonemonth/challenges/serializers.py
inonemonth/challenges/serializers.py
from rest_framework import serializers from .models import Challenge, Role from core.serializers import UserSerializer class RoleSerializer(serializers.ModelSerializer): #user = serializers.RelatedField(many=True) #user = serializers.PrimaryKeyRelatedField() #user = serializers.HyperlinkedRelatedField() user = UserSerializer() #challenge = ChallengeSerializer() challenge = serializers.RelatedField() class Meta: model = Role fields = ("id", "user", "type", "challenge") class ChallengeSerializer(serializers.ModelSerializer): #role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True) #role_set = serializers.RelatedField(many=True) #role_set = serializers.SlugRelatedField(many=True, slug_field="type") role_set = RoleSerializer(many=True) class Meta: model = Challenge fields = ("id", "title", "body", "repo_name", "creation_datetime", "role_set")
from rest_framework import serializers from .models import Challenge, Role from core.serializers import UserSerializer from comments.serializers import CommentSerializer class RoleSerializer(serializers.ModelSerializer): #user = UserSerializer() #challenge = serializers.RelatedField() comment_set = CommentSerializer() class Meta: model = Role fields = ("id", "user", "type", "challenge", "comment_set") class ChallengeSerializer(serializers.ModelSerializer): #role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True) #role_set = serializers.RelatedField(many=True) #role_set = serializers.SlugRelatedField(many=True, slug_field="type") role_set = RoleSerializer(many=True) class Meta: model = Challenge fields = ("id", "title", "body", "repo_name", "creation_datetime", "role_set")
Include comments in Role serializer
Include comments in Role serializer
Python
mit
robrechtdr/inonemonth,robrechtdr/inonemonth,robrechtdr/inonemonth,robrechtdr/inonemonth
from rest_framework import serializers from .models import Challenge, Role from core.serializers import UserSerializer + from comments.serializers import CommentSerializer class RoleSerializer(serializers.ModelSerializer): - #user = serializers.RelatedField(many=True) - #user = serializers.PrimaryKeyRelatedField() - #user = serializers.HyperlinkedRelatedField() - user = UserSerializer() + #user = UserSerializer() - #challenge = ChallengeSerializer() - challenge = serializers.RelatedField() + #challenge = serializers.RelatedField() + comment_set = CommentSerializer() class Meta: model = Role - fields = ("id", "user", "type", "challenge") + fields = ("id", "user", "type", "challenge", "comment_set") class ChallengeSerializer(serializers.ModelSerializer): #role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True) #role_set = serializers.RelatedField(many=True) #role_set = serializers.SlugRelatedField(many=True, slug_field="type") role_set = RoleSerializer(many=True) class Meta: model = Challenge fields = ("id", "title", "body", "repo_name", "creation_datetime", "role_set")
Include comments in Role serializer
## Code Before: from rest_framework import serializers from .models import Challenge, Role from core.serializers import UserSerializer class RoleSerializer(serializers.ModelSerializer): #user = serializers.RelatedField(many=True) #user = serializers.PrimaryKeyRelatedField() #user = serializers.HyperlinkedRelatedField() user = UserSerializer() #challenge = ChallengeSerializer() challenge = serializers.RelatedField() class Meta: model = Role fields = ("id", "user", "type", "challenge") class ChallengeSerializer(serializers.ModelSerializer): #role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True) #role_set = serializers.RelatedField(many=True) #role_set = serializers.SlugRelatedField(many=True, slug_field="type") role_set = RoleSerializer(many=True) class Meta: model = Challenge fields = ("id", "title", "body", "repo_name", "creation_datetime", "role_set") ## Instruction: Include comments in Role serializer ## Code After: from rest_framework import serializers from .models import Challenge, Role from core.serializers import UserSerializer from comments.serializers import CommentSerializer class RoleSerializer(serializers.ModelSerializer): #user = UserSerializer() #challenge = serializers.RelatedField() comment_set = CommentSerializer() class Meta: model = Role fields = ("id", "user", "type", "challenge", "comment_set") class ChallengeSerializer(serializers.ModelSerializer): #role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True) #role_set = serializers.RelatedField(many=True) #role_set = serializers.SlugRelatedField(many=True, slug_field="type") role_set = RoleSerializer(many=True) class Meta: model = Challenge fields = ("id", "title", "body", "repo_name", "creation_datetime", "role_set")
from rest_framework import serializers from .models import Challenge, Role from core.serializers import UserSerializer + from comments.serializers import CommentSerializer class RoleSerializer(serializers.ModelSerializer): - #user = serializers.RelatedField(many=True) - #user = serializers.PrimaryKeyRelatedField() - #user = serializers.HyperlinkedRelatedField() - user = UserSerializer() + #user = UserSerializer() ? + - #challenge = ChallengeSerializer() - challenge = serializers.RelatedField() + #challenge = serializers.RelatedField() ? + + comment_set = CommentSerializer() class Meta: model = Role - fields = ("id", "user", "type", "challenge") + fields = ("id", "user", "type", "challenge", "comment_set") ? +++++++++++++++ class ChallengeSerializer(serializers.ModelSerializer): #role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True) #role_set = serializers.RelatedField(many=True) #role_set = serializers.SlugRelatedField(many=True, slug_field="type") role_set = RoleSerializer(many=True) class Meta: model = Challenge fields = ("id", "title", "body", "repo_name", "creation_datetime", "role_set")
1813413b33170f87cc9fb721c7b5a8cdecfab722
ckanext/googleanalytics/tests/conftest.py
ckanext/googleanalytics/tests/conftest.py
import pytest import factory from factory.alchemy import SQLAlchemyModelFactory from pytest_factoryboy import register import ckan.model as model from ckanext.googleanalytics.model import PackageStats, ResourceStats @pytest.fixture() def clean_db(reset_db, migrate_db_for): reset_db() migrate_db_for("googleanalytics") @register class PackageStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = PackageStats package_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint") @register class ResourceStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = ResourceStats resource_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint")
import pytest import factory from factory.alchemy import SQLAlchemyModelFactory from pytest_factoryboy import register from ckan.plugins import toolkit import ckan.model as model from ckanext.googleanalytics.model import PackageStats, ResourceStats if toolkit.requires_ckan_version("2.9"): @pytest.fixture() def clean_db(reset_db, migrate_db_for): reset_db() migrate_db_for("googleanalytics") else: from dbutil import init_tables @pytest.fixture() def clean_db(reset_db): reset_db() init_tables() @register class PackageStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = PackageStats package_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint") @register class ResourceStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = ResourceStats resource_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint")
Fix fixture for older versions
Fix fixture for older versions
Python
agpl-3.0
ckan/ckanext-googleanalytics,ckan/ckanext-googleanalytics,ckan/ckanext-googleanalytics
import pytest import factory from factory.alchemy import SQLAlchemyModelFactory from pytest_factoryboy import register + from ckan.plugins import toolkit import ckan.model as model from ckanext.googleanalytics.model import PackageStats, ResourceStats + if toolkit.requires_ckan_version("2.9"): - @pytest.fixture() + @pytest.fixture() - def clean_db(reset_db, migrate_db_for): + def clean_db(reset_db, migrate_db_for): - reset_db() + reset_db() - migrate_db_for("googleanalytics") + migrate_db_for("googleanalytics") + else: + from dbutil import init_tables + @pytest.fixture() + def clean_db(reset_db): + reset_db() + init_tables() + @register class PackageStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = PackageStats package_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint") @register class ResourceStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = ResourceStats resource_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint")
Fix fixture for older versions
## Code Before: import pytest import factory from factory.alchemy import SQLAlchemyModelFactory from pytest_factoryboy import register import ckan.model as model from ckanext.googleanalytics.model import PackageStats, ResourceStats @pytest.fixture() def clean_db(reset_db, migrate_db_for): reset_db() migrate_db_for("googleanalytics") @register class PackageStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = PackageStats package_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint") @register class ResourceStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = ResourceStats resource_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint") ## Instruction: Fix fixture for older versions ## Code After: import pytest import factory from factory.alchemy import SQLAlchemyModelFactory from pytest_factoryboy import register from ckan.plugins import toolkit import ckan.model as model from ckanext.googleanalytics.model import PackageStats, ResourceStats if toolkit.requires_ckan_version("2.9"): @pytest.fixture() def clean_db(reset_db, migrate_db_for): reset_db() migrate_db_for("googleanalytics") else: from dbutil import init_tables @pytest.fixture() def clean_db(reset_db): reset_db() init_tables() @register class PackageStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = PackageStats package_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint") @register class ResourceStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = ResourceStats resource_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint")
import pytest import factory from factory.alchemy import SQLAlchemyModelFactory from pytest_factoryboy import register + from ckan.plugins import toolkit import ckan.model as model from ckanext.googleanalytics.model import PackageStats, ResourceStats + if toolkit.requires_ckan_version("2.9"): - @pytest.fixture() + @pytest.fixture() ? ++++ - def clean_db(reset_db, migrate_db_for): + def clean_db(reset_db, migrate_db_for): ? ++++ - reset_db() + reset_db() ? ++++ - migrate_db_for("googleanalytics") + migrate_db_for("googleanalytics") ? ++++ + else: + from dbutil import init_tables + @pytest.fixture() + def clean_db(reset_db): + reset_db() + init_tables() + @register class PackageStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = PackageStats package_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint") @register class ResourceStatsFactory(SQLAlchemyModelFactory): class Meta: sqlalchemy_session = model.Session model = ResourceStats resource_id = factory.Faker("uuid4") visits_recently = factory.Faker("pyint") visits_ever = factory.Faker("pyint")
552283714c329e3a304cd8a8bc14e5370fa6a879
cosmo_tester/framework/constants.py
cosmo_tester/framework/constants.py
CLOUDIFY_TENANT_HEADER = 'Tenant' SUPPORTED_RELEASES = [ '5.0.5', '5.1.0', '5.1.1', '5.1.2', '5.1.3', '5.1.4', '5.2.0', '5.2.1', '6.0.0', 'master', ] SUPPORTED_FOR_RPM_UPGRADE = [ version + '-ga' for version in SUPPORTED_RELEASES if version not in ('master', '5.0.5', '5.1.0') ]
CLOUDIFY_TENANT_HEADER = 'Tenant' SUPPORTED_RELEASES = [ '5.0.5', '5.1.0', '5.1.1', '5.1.2', '5.1.3', '5.1.4', '5.2.0', '5.2.1', '5.2.2', '6.0.0', 'master', ] SUPPORTED_FOR_RPM_UPGRADE = [ version + '-ga' for version in SUPPORTED_RELEASES if version not in ('master', '5.0.5', '5.1.0') ]
Add 5.2.2 to supported versions
Add 5.2.2 to supported versions
Python
apache-2.0
cloudify-cosmo/cloudify-system-tests,cloudify-cosmo/cloudify-system-tests
CLOUDIFY_TENANT_HEADER = 'Tenant' SUPPORTED_RELEASES = [ '5.0.5', '5.1.0', '5.1.1', '5.1.2', '5.1.3', '5.1.4', '5.2.0', '5.2.1', + '5.2.2', '6.0.0', 'master', ] SUPPORTED_FOR_RPM_UPGRADE = [ version + '-ga' for version in SUPPORTED_RELEASES if version not in ('master', '5.0.5', '5.1.0') ]
Add 5.2.2 to supported versions
## Code Before: CLOUDIFY_TENANT_HEADER = 'Tenant' SUPPORTED_RELEASES = [ '5.0.5', '5.1.0', '5.1.1', '5.1.2', '5.1.3', '5.1.4', '5.2.0', '5.2.1', '6.0.0', 'master', ] SUPPORTED_FOR_RPM_UPGRADE = [ version + '-ga' for version in SUPPORTED_RELEASES if version not in ('master', '5.0.5', '5.1.0') ] ## Instruction: Add 5.2.2 to supported versions ## Code After: CLOUDIFY_TENANT_HEADER = 'Tenant' SUPPORTED_RELEASES = [ '5.0.5', '5.1.0', '5.1.1', '5.1.2', '5.1.3', '5.1.4', '5.2.0', '5.2.1', '5.2.2', '6.0.0', 'master', ] SUPPORTED_FOR_RPM_UPGRADE = [ version + '-ga' for version in SUPPORTED_RELEASES if version not in ('master', '5.0.5', '5.1.0') ]
CLOUDIFY_TENANT_HEADER = 'Tenant' SUPPORTED_RELEASES = [ '5.0.5', '5.1.0', '5.1.1', '5.1.2', '5.1.3', '5.1.4', '5.2.0', '5.2.1', + '5.2.2', '6.0.0', 'master', ] SUPPORTED_FOR_RPM_UPGRADE = [ version + '-ga' for version in SUPPORTED_RELEASES if version not in ('master', '5.0.5', '5.1.0') ]
3bddeade05ca5ddc799733baa1545aa2b8b68060
hoomd/tune/custom_tuner.py
hoomd/tune/custom_tuner.py
from hoomd import _hoomd from hoomd.custom import ( _CustomOperation, _InternalCustomOperation, Action) from hoomd.operation import _Tuner class _TunerProperty: @property def updater(self): return self._action @updater.setter def updater(self, updater): if isinstance(updater, Action): self._action = updater else: raise ValueError( "updater must be an instance of hoomd.custom.Action") class CustomTuner(_CustomOperation, _TunerProperty, _Tuner): """Tuner wrapper for `hoomd.custom.Action` objects. For usage see `hoomd.custom._CustomOperation`. """ _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) super().attach(simulation) self._action.attach(simulation) class _InternalCustomTuner( _InternalCustomOperation, _TunerProperty, _Tuner): _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner'
from hoomd import _hoomd from hoomd.operation import _Operation from hoomd.custom import ( _CustomOperation, _InternalCustomOperation, Action) from hoomd.operation import _Tuner class _TunerProperty: @property def tuner(self): return self._action @tuner.setter def tuner(self, tuner): if isinstance(tuner, Action): self._action = tuner else: raise ValueError( "updater must be an instance of hoomd.custom.Action") class CustomTuner(_CustomOperation, _TunerProperty, _Tuner): """Tuner wrapper for `hoomd.custom.Action` objects. For usage see `hoomd.custom._CustomOperation`. """ _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) self._action.attach(simulation) _Operation.attach(self, simulation) class _InternalCustomTuner( _InternalCustomOperation, _TunerProperty, _Tuner): _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) self._action.attach(simulation) _Operation.attach(self, simulation)
Fix attaching on custom tuners
Fix attaching on custom tuners
Python
bsd-3-clause
joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue
from hoomd import _hoomd + from hoomd.operation import _Operation from hoomd.custom import ( _CustomOperation, _InternalCustomOperation, Action) from hoomd.operation import _Tuner class _TunerProperty: @property - def updater(self): + def tuner(self): return self._action - @updater.setter + @tuner.setter - def updater(self, updater): + def tuner(self, tuner): - if isinstance(updater, Action): + if isinstance(tuner, Action): - self._action = updater + self._action = tuner else: raise ValueError( "updater must be an instance of hoomd.custom.Action") class CustomTuner(_CustomOperation, _TunerProperty, _Tuner): """Tuner wrapper for `hoomd.custom.Action` objects. For usage see `hoomd.custom._CustomOperation`. """ _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) - super().attach(simulation) self._action.attach(simulation) + _Operation.attach(self, simulation) class _InternalCustomTuner( _InternalCustomOperation, _TunerProperty, _Tuner): _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' + def attach(self, simulation): + self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( + simulation.state._cpp_sys_def, self.trigger, self._action) + self._action.attach(simulation) + _Operation.attach(self, simulation) +
Fix attaching on custom tuners
## Code Before: from hoomd import _hoomd from hoomd.custom import ( _CustomOperation, _InternalCustomOperation, Action) from hoomd.operation import _Tuner class _TunerProperty: @property def updater(self): return self._action @updater.setter def updater(self, updater): if isinstance(updater, Action): self._action = updater else: raise ValueError( "updater must be an instance of hoomd.custom.Action") class CustomTuner(_CustomOperation, _TunerProperty, _Tuner): """Tuner wrapper for `hoomd.custom.Action` objects. For usage see `hoomd.custom._CustomOperation`. """ _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) super().attach(simulation) self._action.attach(simulation) class _InternalCustomTuner( _InternalCustomOperation, _TunerProperty, _Tuner): _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' ## Instruction: Fix attaching on custom tuners ## Code After: from hoomd import _hoomd from hoomd.operation import _Operation from hoomd.custom import ( _CustomOperation, _InternalCustomOperation, Action) from hoomd.operation import _Tuner class _TunerProperty: @property def tuner(self): return self._action @tuner.setter def tuner(self, tuner): if isinstance(tuner, Action): self._action = tuner else: raise ValueError( "updater must be an instance of hoomd.custom.Action") class CustomTuner(_CustomOperation, _TunerProperty, _Tuner): """Tuner wrapper for `hoomd.custom.Action` objects. For usage see `hoomd.custom._CustomOperation`. """ _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) self._action.attach(simulation) _Operation.attach(self, simulation) class _InternalCustomTuner( _InternalCustomOperation, _TunerProperty, _Tuner): _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) self._action.attach(simulation) _Operation.attach(self, simulation)
from hoomd import _hoomd + from hoomd.operation import _Operation from hoomd.custom import ( _CustomOperation, _InternalCustomOperation, Action) from hoomd.operation import _Tuner class _TunerProperty: @property - def updater(self): ? ^^^^ + def tuner(self): ? + ^ return self._action - @updater.setter ? ^^^^ + @tuner.setter ? + ^ - def updater(self, updater): ? ^^^^ ^^^^ + def tuner(self, tuner): ? + ^ + ^ - if isinstance(updater, Action): ? ^^^^ + if isinstance(tuner, Action): ? + ^ - self._action = updater ? ^^^^ + self._action = tuner ? + ^ else: raise ValueError( "updater must be an instance of hoomd.custom.Action") class CustomTuner(_CustomOperation, _TunerProperty, _Tuner): """Tuner wrapper for `hoomd.custom.Action` objects. For usage see `hoomd.custom._CustomOperation`. """ _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' def attach(self, simulation): self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( simulation.state._cpp_sys_def, self.trigger, self._action) - super().attach(simulation) self._action.attach(simulation) + _Operation.attach(self, simulation) class _InternalCustomTuner( _InternalCustomOperation, _TunerProperty, _Tuner): _cpp_list_name = 'tuners' _cpp_class_name = 'PythonTuner' + + def attach(self, simulation): + self._cpp_obj = getattr(_hoomd, self._cpp_class_name)( + simulation.state._cpp_sys_def, self.trigger, self._action) + self._action.attach(simulation) + _Operation.attach(self, simulation)
77af150756021ac4027e290b5d538e0525d812b9
mopidy/settings.py
mopidy/settings.py
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 SPOTIFY_USERNAME = u'' SPOTIFY_PASSWORD = u'' try: from mopidy.local_settings import * except ImportError: pass
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 SPOTIFY_USERNAME = u'' SPOTIFY_PASSWORD = u'' try: from mopidy.local_settings import * except ImportError: pass
Add threadName to log format
Add threadName to log format
Python
apache-2.0
bencevans/mopidy,pacificIT/mopidy,quartz55/mopidy,SuperStarPL/mopidy,priestd09/mopidy,mokieyue/mopidy,abarisain/mopidy,hkariti/mopidy,swak/mopidy,adamcik/mopidy,quartz55/mopidy,priestd09/mopidy,pacificIT/mopidy,dbrgn/mopidy,jmarsik/mopidy,bencevans/mopidy,tkem/mopidy,abarisain/mopidy,liamw9534/mopidy,hkariti/mopidy,pacificIT/mopidy,jcass77/mopidy,diandiankan/mopidy,pacificIT/mopidy,tkem/mopidy,dbrgn/mopidy,mopidy/mopidy,jodal/mopidy,rawdlite/mopidy,vrs01/mopidy,tkem/mopidy,woutervanwijk/mopidy,ali/mopidy,hkariti/mopidy,vrs01/mopidy,woutervanwijk/mopidy,ali/mopidy,hkariti/mopidy,glogiotatidis/mopidy,mokieyue/mopidy,SuperStarPL/mopidy,mokieyue/mopidy,ZenithDK/mopidy,jmarsik/mopidy,dbrgn/mopidy,dbrgn/mopidy,kingosticks/mopidy,diandiankan/mopidy,kingosticks/mopidy,ZenithDK/mopidy,mopidy/mopidy,SuperStarPL/mopidy,glogiotatidis/mopidy,jodal/mopidy,kingosticks/mopidy,mokieyue/mopidy,glogiotatidis/mopidy,bencevans/mopidy,jmarsik/mopidy,swak/mopidy,swak/mopidy,rawdlite/mopidy,vrs01/mopidy,glogiotatidis/mopidy,SuperStarPL/mopidy,bencevans/mopidy,mopidy/mopidy,tkem/mopidy,ali/mopidy,ZenithDK/mopidy,adamcik/mopidy,liamw9534/mopidy,jcass77/mopidy,jcass77/mopidy,jmarsik/mopidy,adamcik/mopidy,rawdlite/mopidy,ali/mopidy,ZenithDK/mopidy,quartz55/mopidy,jodal/mopidy,swak/mopidy,vrs01/mopidy,rawdlite/mopidy,bacontext/mopidy,bacontext/mopidy,diandiankan/mopidy,bacontext/mopidy,quartz55/mopidy,bacontext/mopidy,priestd09/mopidy,diandiankan/mopidy
- CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' + CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 SPOTIFY_USERNAME = u'' SPOTIFY_PASSWORD = u'' try: from mopidy.local_settings import * except ImportError: pass
Add threadName to log format
## Code Before: CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 SPOTIFY_USERNAME = u'' SPOTIFY_PASSWORD = u'' try: from mopidy.local_settings import * except ImportError: pass ## Instruction: Add threadName to log format ## Code After: CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 SPOTIFY_USERNAME = u'' SPOTIFY_PASSWORD = u'' try: from mopidy.local_settings import * except ImportError: pass
- CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' + CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s' ? +++++++++++++++++ MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 SPOTIFY_USERNAME = u'' SPOTIFY_PASSWORD = u'' try: from mopidy.local_settings import * except ImportError: pass
f5bbdea74c0f8a0cc8ac4331ea8adc45c3f266c8
converter.py
converter.py
def main(): command = "" while command != "exit": command = input("> ").strip() if __name__ == "__main__": main()
import sys def main(): if len(sys.argv) > 1: for i in range(1, len(sys.argv)): with open(sys.argv[i]) as f: lines = f.read().splitlines() for line in lines: evaluate(line); command = "" while command != "exit": command = input("> ").strip() def evaluate(string): print(string) if __name__ == "__main__": main()
Read files defined as argument
Read files defined as argument
Python
mit
hampustagerud/colorconverter
+ + import sys def main(): + if len(sys.argv) > 1: + for i in range(1, len(sys.argv)): + with open(sys.argv[i]) as f: + lines = f.read().splitlines() + for line in lines: + evaluate(line); + command = "" while command != "exit": command = input("> ").strip() + + def evaluate(string): + print(string) if __name__ == "__main__": main()
Read files defined as argument
## Code Before: def main(): command = "" while command != "exit": command = input("> ").strip() if __name__ == "__main__": main() ## Instruction: Read files defined as argument ## Code After: import sys def main(): if len(sys.argv) > 1: for i in range(1, len(sys.argv)): with open(sys.argv[i]) as f: lines = f.read().splitlines() for line in lines: evaluate(line); command = "" while command != "exit": command = input("> ").strip() def evaluate(string): print(string) if __name__ == "__main__": main()
+ + import sys def main(): + if len(sys.argv) > 1: + for i in range(1, len(sys.argv)): + with open(sys.argv[i]) as f: + lines = f.read().splitlines() + for line in lines: + evaluate(line); + command = "" while command != "exit": command = input("> ").strip() + def evaluate(string): + print(string) + if __name__ == "__main__": main()
d7e9eba6fb3628f0736bd468ae76e05099b9d651
space/decorators.py
space/decorators.py
from django.http import HttpResponseBadRequest, HttpResponseForbidden from django.views.decorators.csrf import csrf_exempt from incubator.settings import STATUS_SECRETS def one_or_zero(arg): """Typecast to 1 or 0""" if arg == '1': return 1 elif arg == '0': return 0 raise ValueError("not one or zero") def private_api(**required_params): """ Filter incoming private API requests, and perform parameter validation and extraction """ def outer(some_view): @csrf_exempt def inner(request, *args, **kwargs): if request.method != 'POST': return HttpResponseBadRequest("Only POST is allowed") if 'secret' not in request.POST.keys(): return HttpResponseBadRequest( "You must query this endpoint with a secret.") if request.POST['secret'] not in STATUS_SECRETS: message = 'Bad secret {} is not in the allowed list'.format( request.POST['secret']) return HttpResponseForbidden(message) params = {} for name, typecast in required_params.items(): if name not in request.POST.keys(): return HttpResponseBadRequest( "Parameter %s is required" % name) try: params[name] = typecast(request.POST[name]) except ValueError: return HttpResponseBadRequest( "Did not understood %s=%s" % (name, request.POST[name])) return some_view(request, **params) return inner return outer
from django.http import HttpResponseBadRequest, HttpResponseForbidden from django.views.decorators.csrf import csrf_exempt from django.conf import settings def one_or_zero(arg): """Typecast to 1 or 0""" if arg == '1': return 1 elif arg == '0': return 0 raise ValueError("not one or zero") def private_api(**required_params): """ Filter incoming private API requests, and perform parameter validation and extraction """ def outer(some_view): @csrf_exempt def inner(request, *args, **kwargs): if request.method != 'POST': return HttpResponseBadRequest("Only POST is allowed") if 'secret' not in request.POST.keys(): return HttpResponseBadRequest( "You must query this endpoint with a secret.") if request.POST['secret'] not in settings.STATUS_SECRETS: message = 'Bad secret {} is not in the allowed list'.format( request.POST['secret']) return HttpResponseForbidden(message) params = {} for name, typecast in required_params.items(): if name not in request.POST.keys(): return HttpResponseBadRequest( "Parameter %s is required" % name) try: params[name] = typecast(request.POST[name]) except ValueError: return HttpResponseBadRequest( "Did not understood %s=%s" % (name, request.POST[name])) return some_view(request, **params) return inner return outer
Use from django.conf import settings
Use from django.conf import settings
Python
agpl-3.0
UrLab/incubator,UrLab/incubator,UrLab/incubator,UrLab/incubator
from django.http import HttpResponseBadRequest, HttpResponseForbidden from django.views.decorators.csrf import csrf_exempt - from incubator.settings import STATUS_SECRETS + from django.conf import settings def one_or_zero(arg): """Typecast to 1 or 0""" if arg == '1': return 1 elif arg == '0': return 0 raise ValueError("not one or zero") def private_api(**required_params): """ Filter incoming private API requests, and perform parameter validation and extraction """ def outer(some_view): @csrf_exempt def inner(request, *args, **kwargs): if request.method != 'POST': return HttpResponseBadRequest("Only POST is allowed") if 'secret' not in request.POST.keys(): return HttpResponseBadRequest( "You must query this endpoint with a secret.") - if request.POST['secret'] not in STATUS_SECRETS: + if request.POST['secret'] not in settings.STATUS_SECRETS: message = 'Bad secret {} is not in the allowed list'.format( request.POST['secret']) return HttpResponseForbidden(message) params = {} for name, typecast in required_params.items(): if name not in request.POST.keys(): return HttpResponseBadRequest( "Parameter %s is required" % name) try: params[name] = typecast(request.POST[name]) except ValueError: return HttpResponseBadRequest( "Did not understood %s=%s" % (name, request.POST[name])) return some_view(request, **params) return inner return outer
Use from django.conf import settings
## Code Before: from django.http import HttpResponseBadRequest, HttpResponseForbidden from django.views.decorators.csrf import csrf_exempt from incubator.settings import STATUS_SECRETS def one_or_zero(arg): """Typecast to 1 or 0""" if arg == '1': return 1 elif arg == '0': return 0 raise ValueError("not one or zero") def private_api(**required_params): """ Filter incoming private API requests, and perform parameter validation and extraction """ def outer(some_view): @csrf_exempt def inner(request, *args, **kwargs): if request.method != 'POST': return HttpResponseBadRequest("Only POST is allowed") if 'secret' not in request.POST.keys(): return HttpResponseBadRequest( "You must query this endpoint with a secret.") if request.POST['secret'] not in STATUS_SECRETS: message = 'Bad secret {} is not in the allowed list'.format( request.POST['secret']) return HttpResponseForbidden(message) params = {} for name, typecast in required_params.items(): if name not in request.POST.keys(): return HttpResponseBadRequest( "Parameter %s is required" % name) try: params[name] = typecast(request.POST[name]) except ValueError: return HttpResponseBadRequest( "Did not understood %s=%s" % (name, request.POST[name])) return some_view(request, **params) return inner return outer ## Instruction: Use from django.conf import settings ## Code After: from django.http import HttpResponseBadRequest, HttpResponseForbidden from django.views.decorators.csrf import csrf_exempt from django.conf import settings def one_or_zero(arg): """Typecast to 1 or 0""" if arg == '1': return 1 elif arg == '0': return 0 raise ValueError("not one or zero") def private_api(**required_params): """ Filter incoming private API requests, and perform parameter validation and extraction """ def outer(some_view): @csrf_exempt def inner(request, *args, **kwargs): if request.method != 'POST': return HttpResponseBadRequest("Only POST is allowed") if 'secret' not in request.POST.keys(): return HttpResponseBadRequest( "You must query this endpoint with a secret.") if request.POST['secret'] not in settings.STATUS_SECRETS: message = 'Bad secret {} is not in the allowed list'.format( request.POST['secret']) return HttpResponseForbidden(message) params = {} for name, typecast in required_params.items(): if name not in request.POST.keys(): return HttpResponseBadRequest( "Parameter %s is required" % name) try: params[name] = typecast(request.POST[name]) except ValueError: return HttpResponseBadRequest( "Did not understood %s=%s" % (name, request.POST[name])) return some_view(request, **params) return inner return outer
from django.http import HttpResponseBadRequest, HttpResponseForbidden from django.views.decorators.csrf import csrf_exempt - from incubator.settings import STATUS_SECRETS + from django.conf import settings def one_or_zero(arg): """Typecast to 1 or 0""" if arg == '1': return 1 elif arg == '0': return 0 raise ValueError("not one or zero") def private_api(**required_params): """ Filter incoming private API requests, and perform parameter validation and extraction """ def outer(some_view): @csrf_exempt def inner(request, *args, **kwargs): if request.method != 'POST': return HttpResponseBadRequest("Only POST is allowed") if 'secret' not in request.POST.keys(): return HttpResponseBadRequest( "You must query this endpoint with a secret.") - if request.POST['secret'] not in STATUS_SECRETS: + if request.POST['secret'] not in settings.STATUS_SECRETS: ? +++++++++ message = 'Bad secret {} is not in the allowed list'.format( request.POST['secret']) return HttpResponseForbidden(message) params = {} for name, typecast in required_params.items(): if name not in request.POST.keys(): return HttpResponseBadRequest( "Parameter %s is required" % name) try: params[name] = typecast(request.POST[name]) except ValueError: return HttpResponseBadRequest( "Did not understood %s=%s" % (name, request.POST[name])) return some_view(request, **params) return inner return outer
07bc7efb756e2bc99f59c59476379bc186f36143
sktracker/io/__init__.py
sktracker/io/__init__.py
from .tifffile import imsave from .tifffile import imread from .tifffile import imshow from .tifffile import TiffFile from .tifffile import TiffSequence from .ome import OMEModel from .tiff_metadata import get_metadata_from_tiff __all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow', 'TiffFile', 'TiffSequence']
# Remove warnings for tifffile.py import warnings warnings.filterwarnings("ignore") from .tifffile import imsave from .tifffile import imread from .tifffile import imshow from .tifffile import TiffFile from .tifffile import TiffSequence from .ome import OMEModel from .tiff_metadata import get_metadata_from_tiff __all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow', 'TiffFile', 'TiffSequence']
Remove warning messages for tifffile.py
Remove warning messages for tifffile.py
Python
bsd-3-clause
bnoi/scikit-tracker,bnoi/scikit-tracker,bnoi/scikit-tracker
+ + # Remove warnings for tifffile.py + import warnings + warnings.filterwarnings("ignore") from .tifffile import imsave from .tifffile import imread from .tifffile import imshow from .tifffile import TiffFile from .tifffile import TiffSequence from .ome import OMEModel from .tiff_metadata import get_metadata_from_tiff __all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow', 'TiffFile', 'TiffSequence']
Remove warning messages for tifffile.py
## Code Before: from .tifffile import imsave from .tifffile import imread from .tifffile import imshow from .tifffile import TiffFile from .tifffile import TiffSequence from .ome import OMEModel from .tiff_metadata import get_metadata_from_tiff __all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow', 'TiffFile', 'TiffSequence'] ## Instruction: Remove warning messages for tifffile.py ## Code After: # Remove warnings for tifffile.py import warnings warnings.filterwarnings("ignore") from .tifffile import imsave from .tifffile import imread from .tifffile import imshow from .tifffile import TiffFile from .tifffile import TiffSequence from .ome import OMEModel from .tiff_metadata import get_metadata_from_tiff __all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow', 'TiffFile', 'TiffSequence']
+ + # Remove warnings for tifffile.py + import warnings + warnings.filterwarnings("ignore") from .tifffile import imsave from .tifffile import imread from .tifffile import imshow from .tifffile import TiffFile from .tifffile import TiffSequence from .ome import OMEModel from .tiff_metadata import get_metadata_from_tiff __all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow', 'TiffFile', 'TiffSequence']
df01d83a35c59ee2295b358d8eceeb842adb568d
listings/admin.py
listings/admin.py
from django.contrib import admin from .models import Region, City, GatheringCenter, Resource admin.site.register(Region) admin.site.register(City) class PublishMixin(object): actions = ('publish', 'unpublish') def publish(self, request, queryset): queryset.update(published=True) def unpublish(self, request, queryset): queryset.update(published=False) class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin): list_filter = ('published', 'city', 'created') list_editable = ('published', ) list_display = ('location_name', 'created', 'published', 'author', 'city') raw_id_fields = ('author', ) admin.site.register(GatheringCenter, GatheringCenterAdmin) class ResourceAdmin(PublishMixin, admin.ModelAdmin): list_filter = ('published', 'created') list_editable = ('published', ) list_display = ('name', 'created', 'published', 'author', 'url', 'country') raw_id_fields = ('author', ) admin.site.register(Resource, ResourceAdmin)
from django.contrib import admin from .models import Region, City, GatheringCenter, Resource admin.site.register(Region) admin.site.register(City) class PublishMixin(object): actions = ('publish', 'unpublish') def publish(self, request, queryset): queryset.update(published=True) def unpublish(self, request, queryset): queryset.update(published=False) class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin): date_hierarchy = 'created' list_filter = ('published', 'city', ) list_editable = ('published', ) list_display = ('location_name', 'created', 'published', 'author', 'city') raw_id_fields = ('author', ) admin.site.register(GatheringCenter, GatheringCenterAdmin) class ResourceAdmin(PublishMixin, admin.ModelAdmin): date_hierarchy = 'created' list_filter = ('published', ) list_editable = ('published', ) list_display = ('name', 'created', 'published', 'author', 'url', 'country') raw_id_fields = ('author', ) admin.site.register(Resource, ResourceAdmin)
Order by create, don't filter
Order by create, don't filter
Python
apache-2.0
pony-revolution/helpothers,pony-revolution/helpothers,pony-revolution/helpothers
from django.contrib import admin from .models import Region, City, GatheringCenter, Resource admin.site.register(Region) admin.site.register(City) class PublishMixin(object): actions = ('publish', 'unpublish') def publish(self, request, queryset): queryset.update(published=True) def unpublish(self, request, queryset): queryset.update(published=False) class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin): + date_hierarchy = 'created' - list_filter = ('published', 'city', 'created') + list_filter = ('published', 'city', ) list_editable = ('published', ) list_display = ('location_name', 'created', 'published', 'author', 'city') raw_id_fields = ('author', ) admin.site.register(GatheringCenter, GatheringCenterAdmin) class ResourceAdmin(PublishMixin, admin.ModelAdmin): + date_hierarchy = 'created' - list_filter = ('published', 'created') + list_filter = ('published', ) list_editable = ('published', ) list_display = ('name', 'created', 'published', 'author', 'url', 'country') raw_id_fields = ('author', ) admin.site.register(Resource, ResourceAdmin)
Order by create, don't filter
## Code Before: from django.contrib import admin from .models import Region, City, GatheringCenter, Resource admin.site.register(Region) admin.site.register(City) class PublishMixin(object): actions = ('publish', 'unpublish') def publish(self, request, queryset): queryset.update(published=True) def unpublish(self, request, queryset): queryset.update(published=False) class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin): list_filter = ('published', 'city', 'created') list_editable = ('published', ) list_display = ('location_name', 'created', 'published', 'author', 'city') raw_id_fields = ('author', ) admin.site.register(GatheringCenter, GatheringCenterAdmin) class ResourceAdmin(PublishMixin, admin.ModelAdmin): list_filter = ('published', 'created') list_editable = ('published', ) list_display = ('name', 'created', 'published', 'author', 'url', 'country') raw_id_fields = ('author', ) admin.site.register(Resource, ResourceAdmin) ## Instruction: Order by create, don't filter ## Code After: from django.contrib import admin from .models import Region, City, GatheringCenter, Resource admin.site.register(Region) admin.site.register(City) class PublishMixin(object): actions = ('publish', 'unpublish') def publish(self, request, queryset): queryset.update(published=True) def unpublish(self, request, queryset): queryset.update(published=False) class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin): date_hierarchy = 'created' list_filter = ('published', 'city', ) list_editable = ('published', ) list_display = ('location_name', 'created', 'published', 'author', 'city') raw_id_fields = ('author', ) admin.site.register(GatheringCenter, GatheringCenterAdmin) class ResourceAdmin(PublishMixin, admin.ModelAdmin): date_hierarchy = 'created' list_filter = ('published', ) list_editable = ('published', ) list_display = ('name', 'created', 'published', 'author', 'url', 'country') raw_id_fields = ('author', ) admin.site.register(Resource, ResourceAdmin)
from django.contrib import admin from .models import Region, City, GatheringCenter, Resource admin.site.register(Region) admin.site.register(City) class PublishMixin(object): actions = ('publish', 'unpublish') def publish(self, request, queryset): queryset.update(published=True) def unpublish(self, request, queryset): queryset.update(published=False) class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin): + date_hierarchy = 'created' - list_filter = ('published', 'city', 'created') ? --------- + list_filter = ('published', 'city', ) list_editable = ('published', ) list_display = ('location_name', 'created', 'published', 'author', 'city') raw_id_fields = ('author', ) admin.site.register(GatheringCenter, GatheringCenterAdmin) class ResourceAdmin(PublishMixin, admin.ModelAdmin): + date_hierarchy = 'created' - list_filter = ('published', 'created') ? --------- + list_filter = ('published', ) list_editable = ('published', ) list_display = ('name', 'created', 'published', 'author', 'url', 'country') raw_id_fields = ('author', ) admin.site.register(Resource, ResourceAdmin)
0024b8b921d788a0539bc242bd1600c0da666bd6
panoptes/state_machine/states/core.py
panoptes/state_machine/states/core.py
import time import transitions from panoptes.utils.logger import has_logger @has_logger class PanState(transitions.State): """ Base class for PANOPTES transitions """ def __init__(self, *args, **kwargs): name = kwargs.get('name', self.__class__) self.panoptes = kwargs.get('panoptes', None) super().__init__(name=name, on_enter=['execute']) self._sleep_delay = 3 # seconds def main(self): assert self.panoptes is not None msg = "Must implement `main` method inside class {}. Exiting".format(self.name) self.panoptes.logger.warning(msg) return 'exit' def sleep(self, seconds=None): """ sleep for `seconds` or `_sleep_delay` seconds This puts the state into a loop that is responsive to outside messages. Args: seconds(float): Seconds to sleep for, defaults to `_sleep_delay`. """ assert self.panoptes is not None if seconds is None: seconds = self._sleep_delay if seconds > 10: step_time = seconds / 4 while seconds: seconds = seconds - step_time # NOTE: DO SOMETHING RESPONSIVE HERE time.sleep(step_time) else: time.sleep(seconds)
import time import transitions from panoptes.utils.logger import has_logger @has_logger class PanState(transitions.State): """ Base class for PANOPTES transitions """ def __init__(self, *args, **kwargs): name = kwargs.get('name', self.__class__) self.panoptes = kwargs.get('panoptes', None) super().__init__(name=name, on_enter=['execute']) self._sleep_delay = 3 # seconds def main(self): assert self.panoptes is not None msg = "Must implement `main` method inside class {}. Exiting".format(self.name) self.panoptes.logger.warning(msg) def sleep(self, seconds=None): """ sleep for `seconds` or `_sleep_delay` seconds This puts the state into a loop that is responsive to outside messages. Args: seconds(float): Seconds to sleep for, defaults to `_sleep_delay`. """ assert self.panoptes is not None if seconds is None: seconds = self._sleep_delay if seconds > 10: step_time = seconds / 4 while seconds: seconds = seconds - step_time # NOTE: DO SOMETHING RESPONSIVE HERE time.sleep(step_time) else: time.sleep(seconds)
Remove return state from main `main`
Remove return state from main `main`
Python
mit
panoptes/POCS,panoptes/POCS,joshwalawender/POCS,AstroHuntsman/POCS,joshwalawender/POCS,joshwalawender/POCS,panoptes/POCS,panoptes/POCS,AstroHuntsman/POCS,AstroHuntsman/POCS,AstroHuntsman/POCS
import time import transitions from panoptes.utils.logger import has_logger @has_logger class PanState(transitions.State): """ Base class for PANOPTES transitions """ def __init__(self, *args, **kwargs): name = kwargs.get('name', self.__class__) self.panoptes = kwargs.get('panoptes', None) super().__init__(name=name, on_enter=['execute']) self._sleep_delay = 3 # seconds def main(self): assert self.panoptes is not None msg = "Must implement `main` method inside class {}. Exiting".format(self.name) self.panoptes.logger.warning(msg) - return 'exit' def sleep(self, seconds=None): """ sleep for `seconds` or `_sleep_delay` seconds This puts the state into a loop that is responsive to outside messages. Args: seconds(float): Seconds to sleep for, defaults to `_sleep_delay`. """ assert self.panoptes is not None if seconds is None: seconds = self._sleep_delay if seconds > 10: step_time = seconds / 4 while seconds: seconds = seconds - step_time # NOTE: DO SOMETHING RESPONSIVE HERE time.sleep(step_time) else: time.sleep(seconds)
Remove return state from main `main`
## Code Before: import time import transitions from panoptes.utils.logger import has_logger @has_logger class PanState(transitions.State): """ Base class for PANOPTES transitions """ def __init__(self, *args, **kwargs): name = kwargs.get('name', self.__class__) self.panoptes = kwargs.get('panoptes', None) super().__init__(name=name, on_enter=['execute']) self._sleep_delay = 3 # seconds def main(self): assert self.panoptes is not None msg = "Must implement `main` method inside class {}. Exiting".format(self.name) self.panoptes.logger.warning(msg) return 'exit' def sleep(self, seconds=None): """ sleep for `seconds` or `_sleep_delay` seconds This puts the state into a loop that is responsive to outside messages. Args: seconds(float): Seconds to sleep for, defaults to `_sleep_delay`. """ assert self.panoptes is not None if seconds is None: seconds = self._sleep_delay if seconds > 10: step_time = seconds / 4 while seconds: seconds = seconds - step_time # NOTE: DO SOMETHING RESPONSIVE HERE time.sleep(step_time) else: time.sleep(seconds) ## Instruction: Remove return state from main `main` ## Code After: import time import transitions from panoptes.utils.logger import has_logger @has_logger class PanState(transitions.State): """ Base class for PANOPTES transitions """ def __init__(self, *args, **kwargs): name = kwargs.get('name', self.__class__) self.panoptes = kwargs.get('panoptes', None) super().__init__(name=name, on_enter=['execute']) self._sleep_delay = 3 # seconds def main(self): assert self.panoptes is not None msg = "Must implement `main` method inside class {}. Exiting".format(self.name) self.panoptes.logger.warning(msg) def sleep(self, seconds=None): """ sleep for `seconds` or `_sleep_delay` seconds This puts the state into a loop that is responsive to outside messages. Args: seconds(float): Seconds to sleep for, defaults to `_sleep_delay`. """ assert self.panoptes is not None if seconds is None: seconds = self._sleep_delay if seconds > 10: step_time = seconds / 4 while seconds: seconds = seconds - step_time # NOTE: DO SOMETHING RESPONSIVE HERE time.sleep(step_time) else: time.sleep(seconds)
import time import transitions from panoptes.utils.logger import has_logger @has_logger class PanState(transitions.State): """ Base class for PANOPTES transitions """ def __init__(self, *args, **kwargs): name = kwargs.get('name', self.__class__) self.panoptes = kwargs.get('panoptes', None) super().__init__(name=name, on_enter=['execute']) self._sleep_delay = 3 # seconds def main(self): assert self.panoptes is not None msg = "Must implement `main` method inside class {}. Exiting".format(self.name) self.panoptes.logger.warning(msg) - return 'exit' def sleep(self, seconds=None): """ sleep for `seconds` or `_sleep_delay` seconds This puts the state into a loop that is responsive to outside messages. Args: seconds(float): Seconds to sleep for, defaults to `_sleep_delay`. """ assert self.panoptes is not None if seconds is None: seconds = self._sleep_delay if seconds > 10: step_time = seconds / 4 while seconds: seconds = seconds - step_time # NOTE: DO SOMETHING RESPONSIVE HERE time.sleep(step_time) else: time.sleep(seconds)
ce25cea7e8d10f9c318e2e7ef1dc1013921ed062
clint/textui/prompt.py
clint/textui/prompt.py
from __future__ import absolute_import from re import match, I def yn(prompt, default='y', batch=False): # A sanity check against default value # If not y/n then y is assumed if default not in ['y', 'n']: default = 'y' # Let's build the prompt choicebox = '[Y/n]' if default == 'y' else '[y/N]' prompt = prompt + ' ' + choicebox + ' ' # If input is not a yes/no variant or empty # keep asking while True: # If batch option is True then auto reply # with default input if not batch: input = raw_input(prompt).strip() else: print prompt input = '' # If input is empty default choice is assumed # so we return True if input == '': return True # Given 'yes' as input if default choice is y # then return True, False otherwise if match('y(?:es)?', input, I): return True if default == 'y' else False # Given 'no' as input if default choice is n # then return True, False otherwise elif match('n(?:o)?', input, I): return True if default == 'n' else False
from __future__ import absolute_import, print_function from re import match, I def yn(prompt, default='y', batch=False): # A sanity check against default value # If not y/n then y is assumed if default not in ['y', 'n']: default = 'y' # Let's build the prompt choicebox = '[Y/n]' if default == 'y' else '[y/N]' prompt = prompt + ' ' + choicebox + ' ' # If input is not a yes/no variant or empty # keep asking while True: # If batch option is True then auto reply # with default input if not batch: input = raw_input(prompt).strip() else: print(prompt) input = '' # If input is empty default choice is assumed # so we return True if input == '': return True # Given 'yes' as input if default choice is y # then return True, False otherwise if match('y(?:es)?', input, I): return True if default == 'y' else False # Given 'no' as input if default choice is n # then return True, False otherwise elif match('n(?:o)?', input, I): return True if default == 'n' else False
Use print() function to fix install on python 3
Use print() function to fix install on python 3 clint 0.3.2 can't be installed on python 3.3 because of a print statement.
Python
isc
1gitGrey/clint,thusoy/clint,wkentaro/clint,1gitGrey/clint,glorizen/clint,wkentaro/clint,tz70s/clint,Lh4cKg/clint,nathancahill/clint,kennethreitz/clint,nathancahill/clint
- from __future__ import absolute_import + from __future__ import absolute_import, print_function from re import match, I def yn(prompt, default='y', batch=False): # A sanity check against default value # If not y/n then y is assumed if default not in ['y', 'n']: default = 'y' # Let's build the prompt choicebox = '[Y/n]' if default == 'y' else '[y/N]' prompt = prompt + ' ' + choicebox + ' ' # If input is not a yes/no variant or empty # keep asking while True: # If batch option is True then auto reply # with default input if not batch: input = raw_input(prompt).strip() else: - print prompt + print(prompt) input = '' # If input is empty default choice is assumed # so we return True if input == '': return True # Given 'yes' as input if default choice is y # then return True, False otherwise if match('y(?:es)?', input, I): return True if default == 'y' else False # Given 'no' as input if default choice is n # then return True, False otherwise elif match('n(?:o)?', input, I): return True if default == 'n' else False
Use print() function to fix install on python 3
## Code Before: from __future__ import absolute_import from re import match, I def yn(prompt, default='y', batch=False): # A sanity check against default value # If not y/n then y is assumed if default not in ['y', 'n']: default = 'y' # Let's build the prompt choicebox = '[Y/n]' if default == 'y' else '[y/N]' prompt = prompt + ' ' + choicebox + ' ' # If input is not a yes/no variant or empty # keep asking while True: # If batch option is True then auto reply # with default input if not batch: input = raw_input(prompt).strip() else: print prompt input = '' # If input is empty default choice is assumed # so we return True if input == '': return True # Given 'yes' as input if default choice is y # then return True, False otherwise if match('y(?:es)?', input, I): return True if default == 'y' else False # Given 'no' as input if default choice is n # then return True, False otherwise elif match('n(?:o)?', input, I): return True if default == 'n' else False ## Instruction: Use print() function to fix install on python 3 ## Code After: from __future__ import absolute_import, print_function from re import match, I def yn(prompt, default='y', batch=False): # A sanity check against default value # If not y/n then y is assumed if default not in ['y', 'n']: default = 'y' # Let's build the prompt choicebox = '[Y/n]' if default == 'y' else '[y/N]' prompt = prompt + ' ' + choicebox + ' ' # If input is not a yes/no variant or empty # keep asking while True: # If batch option is True then auto reply # with default input if not batch: input = raw_input(prompt).strip() else: print(prompt) input = '' # If input is empty default choice is assumed # so we return True if input == '': return True # Given 'yes' as input if default choice is y # then return True, False otherwise if match('y(?:es)?', input, I): return True if default == 'y' else False # Given 'no' as input if default choice is n # then return True, False otherwise elif match('n(?:o)?', input, I): return True if default == 'n' else False
- from __future__ import absolute_import + from __future__ import absolute_import, print_function ? ++++++++++++++++ from re import match, I def yn(prompt, default='y', batch=False): # A sanity check against default value # If not y/n then y is assumed if default not in ['y', 'n']: default = 'y' # Let's build the prompt choicebox = '[Y/n]' if default == 'y' else '[y/N]' prompt = prompt + ' ' + choicebox + ' ' # If input is not a yes/no variant or empty # keep asking while True: # If batch option is True then auto reply # with default input if not batch: input = raw_input(prompt).strip() else: - print prompt ? ^ + print(prompt) ? ^ + input = '' # If input is empty default choice is assumed # so we return True if input == '': return True # Given 'yes' as input if default choice is y # then return True, False otherwise if match('y(?:es)?', input, I): return True if default == 'y' else False # Given 'no' as input if default choice is n # then return True, False otherwise elif match('n(?:o)?', input, I): return True if default == 'n' else False
692234e72862839d8c14fb0f1a6ebe7259b15413
core/report.py
core/report.py
import config def sendMail(): print config.config print "Sent e-mail" def sendToGrapite(): pass
from config import config as conf from email.mime.multipart import MIMEMultipart from email.mime.text import MIMEText import smtplib import time import pushnotify def sendMail(recipients, subject, body): if not isinstance( recipients, list ): recipients = [ recipients ] session = smtplib.SMTP( conf.get( 'gmail', 'server' ), conf.getint( 'gmail', 'port' ) ) session.ehlo() session.starttls() session.login( conf.get( 'gmail', 'username' ), conf.get( 'gmail', 'password' ) ) for recipient in recipients: headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ), "subject: " + subject, "to: " + recipient, "mime-version: 1.0", "content-type: text/html" ] ) content = headers + "\r\n\r\n" + body session.sendmail( conf.get( 'gmail', 'from' ), recipient, content ) def sendNotification(application, desc, event): client = pushnotify.get_client('nma', application=application ) client.add_key( conf.get( 'notifymyandroid', 'api_key' ) ) try: client.notify( desc, event, split=True ) except: pass def sendToGrapite(path, value): message = '%s %s %d\n' % ( path, value, int( time.time() ) ) sock = socket.socket() graphite_address = ( conf.get( 'graphite', 'server' ), conf.get( 'graphite', 'port' ) ) sock.connect( graphite_address ) sock.sendall( message ) sock.close()
Complete e-mail, Graphite and push notification support
Complete e-mail, Graphite and push notification support
Python
mit
nlindblad/ocarina,nlindblad/ocarina
- import config + from config import config as conf + from email.mime.multipart import MIMEMultipart + from email.mime.text import MIMEText + import smtplib + import time + import pushnotify - def sendMail(): - print config.config - print "Sent e-mail" - def sendToGrapite(): + def sendMail(recipients, subject, body): + if not isinstance( recipients, list ): + recipients = [ recipients ] + session = smtplib.SMTP( conf.get( 'gmail', 'server' ), + conf.getint( 'gmail', 'port' ) ) + session.ehlo() + session.starttls() + session.login( conf.get( 'gmail', 'username' ), + conf.get( 'gmail', 'password' ) ) + for recipient in recipients: + headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ), + "subject: " + subject, + "to: " + recipient, + "mime-version: 1.0", + "content-type: text/html" ] ) + content = headers + "\r\n\r\n" + body + session.sendmail( conf.get( 'gmail', 'from' ), recipient, content ) + + def sendNotification(application, desc, event): + client = pushnotify.get_client('nma', application=application ) + client.add_key( conf.get( 'notifymyandroid', 'api_key' ) ) + try: + client.notify( desc, event, split=True ) + except: - pass + pass + def sendToGrapite(path, value): + message = '%s %s %d\n' % ( path, value, int( time.time() ) ) + sock = socket.socket() + graphite_address = ( conf.get( 'graphite', 'server' ), + conf.get( 'graphite', 'port' ) ) + sock.connect( graphite_address ) + sock.sendall( message ) + sock.close() + +
Complete e-mail, Graphite and push notification support
## Code Before: import config def sendMail(): print config.config print "Sent e-mail" def sendToGrapite(): pass ## Instruction: Complete e-mail, Graphite and push notification support ## Code After: from config import config as conf from email.mime.multipart import MIMEMultipart from email.mime.text import MIMEText import smtplib import time import pushnotify def sendMail(recipients, subject, body): if not isinstance( recipients, list ): recipients = [ recipients ] session = smtplib.SMTP( conf.get( 'gmail', 'server' ), conf.getint( 'gmail', 'port' ) ) session.ehlo() session.starttls() session.login( conf.get( 'gmail', 'username' ), conf.get( 'gmail', 'password' ) ) for recipient in recipients: headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ), "subject: " + subject, "to: " + recipient, "mime-version: 1.0", "content-type: text/html" ] ) content = headers + "\r\n\r\n" + body session.sendmail( conf.get( 'gmail', 'from' ), recipient, content ) def sendNotification(application, desc, event): client = pushnotify.get_client('nma', application=application ) client.add_key( conf.get( 'notifymyandroid', 'api_key' ) ) try: client.notify( desc, event, split=True ) except: pass def sendToGrapite(path, value): message = '%s %s %d\n' % ( path, value, int( time.time() ) ) sock = socket.socket() graphite_address = ( conf.get( 'graphite', 'server' ), conf.get( 'graphite', 'port' ) ) sock.connect( graphite_address ) sock.sendall( message ) sock.close()
- import config + from config import config as conf + from email.mime.multipart import MIMEMultipart + from email.mime.text import MIMEText + import smtplib + import time + import pushnotify - def sendMail(): - print config.config - print "Sent e-mail" - def sendToGrapite(): + def sendMail(recipients, subject, body): + if not isinstance( recipients, list ): + recipients = [ recipients ] + session = smtplib.SMTP( conf.get( 'gmail', 'server' ), + conf.getint( 'gmail', 'port' ) ) + session.ehlo() + session.starttls() + session.login( conf.get( 'gmail', 'username' ), + conf.get( 'gmail', 'password' ) ) + for recipient in recipients: + headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ), + "subject: " + subject, + "to: " + recipient, + "mime-version: 1.0", + "content-type: text/html" ] ) + content = headers + "\r\n\r\n" + body + session.sendmail( conf.get( 'gmail', 'from' ), recipient, content ) + + def sendNotification(application, desc, event): + client = pushnotify.get_client('nma', application=application ) + client.add_key( conf.get( 'notifymyandroid', 'api_key' ) ) + try: + client.notify( desc, event, split=True ) + except: - pass + pass ? ++++ + + def sendToGrapite(path, value): + message = '%s %s %d\n' % ( path, value, int( time.time() ) ) + sock = socket.socket() + graphite_address = ( conf.get( 'graphite', 'server' ), + conf.get( 'graphite', 'port' ) ) + sock.connect( graphite_address ) + sock.sendall( message ) + sock.close() +
5cd3b53f677fd6ab6e77bee5b7d42cf2ac85e47f
feincms/apps.py
feincms/apps.py
from feincms.content.application.models import *
def __getattr__(key): # Work around Django 3.2's autoloading of *.apps modules (AppConfig # autodiscovery) if key in { "ApplicationContent", "app_reverse", "app_reverse_lazy", "permalink", "UnpackTemplateResponse", "standalone", "unpack", }: from feincms.content.application import models return getattr(models, key) raise AttributeError("Unknown attribute '%s'" % key)
Add a workaround for the AppConfig autodiscovery crashes with Django 3.2
Add a workaround for the AppConfig autodiscovery crashes with Django 3.2
Python
bsd-3-clause
mjl/feincms,feincms/feincms,mjl/feincms,feincms/feincms,feincms/feincms,mjl/feincms
+ def __getattr__(key): + # Work around Django 3.2's autoloading of *.apps modules (AppConfig + # autodiscovery) + if key in { + "ApplicationContent", + "app_reverse", + "app_reverse_lazy", + "permalink", + "UnpackTemplateResponse", + "standalone", + "unpack", + }: + from feincms.content.application import models - from feincms.content.application.models import * + return getattr(models, key) + raise AttributeError("Unknown attribute '%s'" % key) +
Add a workaround for the AppConfig autodiscovery crashes with Django 3.2
## Code Before: from feincms.content.application.models import * ## Instruction: Add a workaround for the AppConfig autodiscovery crashes with Django 3.2 ## Code After: def __getattr__(key): # Work around Django 3.2's autoloading of *.apps modules (AppConfig # autodiscovery) if key in { "ApplicationContent", "app_reverse", "app_reverse_lazy", "permalink", "UnpackTemplateResponse", "standalone", "unpack", }: from feincms.content.application import models return getattr(models, key) raise AttributeError("Unknown attribute '%s'" % key)
+ def __getattr__(key): + # Work around Django 3.2's autoloading of *.apps modules (AppConfig + # autodiscovery) + if key in { + "ApplicationContent", + "app_reverse", + "app_reverse_lazy", + "permalink", + "UnpackTemplateResponse", + "standalone", + "unpack", + }: + from feincms.content.application import models - from feincms.content.application.models import * + return getattr(models, key) + + raise AttributeError("Unknown attribute '%s'" % key)