commit stringlengths 40 40 | old_file stringlengths 4 106 | new_file stringlengths 4 106 | old_contents stringlengths 10 2.94k | new_contents stringlengths 21 2.95k | subject stringlengths 16 444 | message stringlengths 17 2.63k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 7 43k | ndiff stringlengths 52 3.31k | instruction stringlengths 16 444 | content stringlengths 133 4.32k | diff stringlengths 49 3.61k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6ccc85832aeff2ca9800cd9e2af8461515ff680d | cartography/midi_utils.py | cartography/midi_utils.py | import mido
def open_output():
return open_steinberg_output()
def get_steinberg_device_name():
output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()]
if len(output_names) != 1:
raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one")
return output_names[0]
def open_steinberg_output():
return mido.open_output(get_steinberg_device_name())
def open_steinberg_input():
return mido.open_input(get_steinberg_device_name())
| import mido
def open_output():
return open_steinberg_output()
def get_steinberg_device_name():
output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()]
if len(output_names) != 1:
raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one")
return output_names[0]
def open_steinberg_output():
return mido.open_output(get_steinberg_device_name(), autoreset=True)
def open_steinberg_input():
return mido.open_input(get_steinberg_device_name())
| Add dump presets and utils | Add dump presets and utils
| Python | mit | tingled/synthetic-cartography,tingled/synthetic-cartography | import mido
def open_output():
return open_steinberg_output()
def get_steinberg_device_name():
output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()]
if len(output_names) != 1:
raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one")
return output_names[0]
def open_steinberg_output():
- return mido.open_output(get_steinberg_device_name())
+ return mido.open_output(get_steinberg_device_name(), autoreset=True)
def open_steinberg_input():
return mido.open_input(get_steinberg_device_name())
| Add dump presets and utils | ## Code Before:
import mido
def open_output():
return open_steinberg_output()
def get_steinberg_device_name():
output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()]
if len(output_names) != 1:
raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one")
return output_names[0]
def open_steinberg_output():
return mido.open_output(get_steinberg_device_name())
def open_steinberg_input():
return mido.open_input(get_steinberg_device_name())
## Instruction:
Add dump presets and utils
## Code After:
import mido
def open_output():
return open_steinberg_output()
def get_steinberg_device_name():
output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()]
if len(output_names) != 1:
raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one")
return output_names[0]
def open_steinberg_output():
return mido.open_output(get_steinberg_device_name(), autoreset=True)
def open_steinberg_input():
return mido.open_input(get_steinberg_device_name())
| import mido
def open_output():
return open_steinberg_output()
def get_steinberg_device_name():
output_names = [n for n in mido.get_output_names() if 'steinberg' in n.lower()]
if len(output_names) != 1:
raise Exception(f"Found the following steinberg MIDI devices: {output_names}. Expected only one")
return output_names[0]
def open_steinberg_output():
- return mido.open_output(get_steinberg_device_name())
+ return mido.open_output(get_steinberg_device_name(), autoreset=True)
? ++++++++++++++++
def open_steinberg_input():
return mido.open_input(get_steinberg_device_name()) |
04ff248c628e5523bc22c532cf4518f210376307 | setup.py | setup.py |
import os
import sys
sys.path.insert(0, os.path.abspath('lib'))
from ansible import __version__, __author__
from distutils.core import setup
setup(name='ansible',
version=__version__,
description='Minimal SSH command and control',
author=__author__,
author_email='michael.dehaan@gmail.com',
url='http://ansible.github.com/',
license='GPLv3',
install_requires=['paramiko', 'jinja2', "PyYAML"],
package_dir={ 'ansible': 'lib/ansible' },
packages=[
'ansible',
'ansible.inventory',
'ansible.playbook',
'ansible.runner',
'ansible.runner.connection',
],
scripts=[
'bin/ansible',
'bin/ansible-playbook'
]
)
|
import os
import sys
sys.path.insert(0, os.path.abspath('lib'))
from ansible import __version__, __author__
from distutils.core import setup
setup(name='ansible',
version=__version__,
description='Minimal SSH command and control',
author=__author__,
author_email='michael.dehaan@gmail.com',
url='http://ansible.github.com/',
license='GPLv3',
install_requires=['paramiko', 'jinja2', "PyYAML"],
package_dir={ 'ansible': 'lib/ansible' },
packages=[
'ansible',
'ansible.inventory',
'ansible.playbook',
'ansible.runner',
'ansible.runner.connection',
],
scripts=[
'bin/ansible',
'bin/ansible-playbook',
'bin/ansible-pull'
]
)
| Include bin/ansible-pull as part of the sdist in distutils. | Include bin/ansible-pull as part of the sdist in distutils.
| Python | mit | thaim/ansible,thaim/ansible |
import os
import sys
sys.path.insert(0, os.path.abspath('lib'))
from ansible import __version__, __author__
from distutils.core import setup
setup(name='ansible',
version=__version__,
description='Minimal SSH command and control',
author=__author__,
author_email='michael.dehaan@gmail.com',
url='http://ansible.github.com/',
license='GPLv3',
install_requires=['paramiko', 'jinja2', "PyYAML"],
package_dir={ 'ansible': 'lib/ansible' },
packages=[
'ansible',
'ansible.inventory',
'ansible.playbook',
'ansible.runner',
'ansible.runner.connection',
],
scripts=[
'bin/ansible',
- 'bin/ansible-playbook'
+ 'bin/ansible-playbook',
+ 'bin/ansible-pull'
]
)
| Include bin/ansible-pull as part of the sdist in distutils. | ## Code Before:
import os
import sys
sys.path.insert(0, os.path.abspath('lib'))
from ansible import __version__, __author__
from distutils.core import setup
setup(name='ansible',
version=__version__,
description='Minimal SSH command and control',
author=__author__,
author_email='michael.dehaan@gmail.com',
url='http://ansible.github.com/',
license='GPLv3',
install_requires=['paramiko', 'jinja2', "PyYAML"],
package_dir={ 'ansible': 'lib/ansible' },
packages=[
'ansible',
'ansible.inventory',
'ansible.playbook',
'ansible.runner',
'ansible.runner.connection',
],
scripts=[
'bin/ansible',
'bin/ansible-playbook'
]
)
## Instruction:
Include bin/ansible-pull as part of the sdist in distutils.
## Code After:
import os
import sys
sys.path.insert(0, os.path.abspath('lib'))
from ansible import __version__, __author__
from distutils.core import setup
setup(name='ansible',
version=__version__,
description='Minimal SSH command and control',
author=__author__,
author_email='michael.dehaan@gmail.com',
url='http://ansible.github.com/',
license='GPLv3',
install_requires=['paramiko', 'jinja2', "PyYAML"],
package_dir={ 'ansible': 'lib/ansible' },
packages=[
'ansible',
'ansible.inventory',
'ansible.playbook',
'ansible.runner',
'ansible.runner.connection',
],
scripts=[
'bin/ansible',
'bin/ansible-playbook',
'bin/ansible-pull'
]
)
|
import os
import sys
sys.path.insert(0, os.path.abspath('lib'))
from ansible import __version__, __author__
from distutils.core import setup
setup(name='ansible',
version=__version__,
description='Minimal SSH command and control',
author=__author__,
author_email='michael.dehaan@gmail.com',
url='http://ansible.github.com/',
license='GPLv3',
install_requires=['paramiko', 'jinja2', "PyYAML"],
package_dir={ 'ansible': 'lib/ansible' },
packages=[
'ansible',
'ansible.inventory',
'ansible.playbook',
'ansible.runner',
'ansible.runner.connection',
],
scripts=[
'bin/ansible',
- 'bin/ansible-playbook'
+ 'bin/ansible-playbook',
? +
+ 'bin/ansible-pull'
]
) |
a4d1659197c0c3da706065d5362fd3b060223c87 | newaccount/views.py | newaccount/views.py | from django.shortcuts import render
from django.http import JsonResponse
import common.render
from common.settings import get_page_config
def form(request):
''' The signup form webpage '''
context = get_page_config(title='New User Sign Up')
context['form'] = [
{'label': 'User Name', 'name': 'username'},
{'label': 'Email Address', 'name': 'email'},
{'label': 'Password', 'type': 'password', 'name':'password'},
{'label': 'Re-enter Password', 'type': 'password', 'id':'repass'}
]
context['validators'] = [ 'signup/newaccount_validator.js' ]
return common.render.singleform(request, context)
def validate(request):
''' Signup form validation handler '''
return JsonResponse({"status": "error", "message": "Test Case"})
def submit(request):
''' Signup form submission handler '''
return HttpResponse('')
| from django.http import JsonResponse
from django.contrib.auth.models import User
from django.core.validators import validate_email
from django.core.exceptions import ValidationError
from django.shortcuts import render
import urllib
import common.render
from common.settings import get_page_config
def form(request):
''' The signup form webpage '''
context = get_page_config(title='New User Sign Up')
context['form'] = [
{'label': 'User Name', 'name': 'username'},
{'label': 'Email Address', 'name': 'email'},
{'label': 'Password', 'type': 'password', 'name':'password'},
{'label': 'Re-enter Password', 'type': 'password', 'id':'repass'}
]
context['validators'] = [ 'signup/newaccount_validator.js' ]
return common.render.singleform(request, context)
def validate(request):
''' Signup form validation handler '''
msg = ''
if not('username' in request.GET):
msg = 'Name not given!'
elif not('email' in request.GET):
msg = 'Email not given!'
name = urllib.unquote(request.GET['username'])
email = urllib.unquote(request.GET['email'])
if not msg: msg = __validate_name__(name)
if not msg: msg = __validate_email__(email)
status = 'error' if msg else 'success'
return JsonResponse({"status": status, "message": msg})
def submit(request):
''' Signup form submission handler '''
return HttpResponse('')
def __validate_name__(name):
''' Internal validation function for username '''
lname = len(name)
if lname < 5:
return 'User name must be at least 5 characters long'
if lname > 64:
return 'User name must not be longer than 64 characters'
if len(User.objects.filter(username=name)):
return 'User name already in use'
return ''
def __validate_email__(email):
''' Internal validation function for email '''
try:
validate_email(email)
except ValidationError:
return 'Invalid email address: '+email
return ''
| Implement backend newaccount form verification | Implement backend newaccount form verification
| Python | mit | NicolasKiely/Robit-Tracker,NicolasKiely/Robit-Tracker,NicolasKiely/Robit-Tracker | + from django.http import JsonResponse
+ from django.contrib.auth.models import User
+ from django.core.validators import validate_email
+ from django.core.exceptions import ValidationError
from django.shortcuts import render
- from django.http import JsonResponse
+ import urllib
import common.render
from common.settings import get_page_config
def form(request):
''' The signup form webpage '''
context = get_page_config(title='New User Sign Up')
context['form'] = [
{'label': 'User Name', 'name': 'username'},
{'label': 'Email Address', 'name': 'email'},
{'label': 'Password', 'type': 'password', 'name':'password'},
{'label': 'Re-enter Password', 'type': 'password', 'id':'repass'}
]
context['validators'] = [ 'signup/newaccount_validator.js' ]
return common.render.singleform(request, context)
def validate(request):
''' Signup form validation handler '''
+ msg = ''
+ if not('username' in request.GET):
+ msg = 'Name not given!'
+ elif not('email' in request.GET):
+ msg = 'Email not given!'
+ name = urllib.unquote(request.GET['username'])
+ email = urllib.unquote(request.GET['email'])
+ if not msg: msg = __validate_name__(name)
+ if not msg: msg = __validate_email__(email)
+
+ status = 'error' if msg else 'success'
- return JsonResponse({"status": "error", "message": "Test Case"})
+ return JsonResponse({"status": status, "message": msg})
def submit(request):
''' Signup form submission handler '''
return HttpResponse('')
+
+ def __validate_name__(name):
+ ''' Internal validation function for username '''
+ lname = len(name)
+ if lname < 5:
+ return 'User name must be at least 5 characters long'
+ if lname > 64:
+ return 'User name must not be longer than 64 characters'
+ if len(User.objects.filter(username=name)):
+ return 'User name already in use'
+ return ''
+
+ def __validate_email__(email):
+ ''' Internal validation function for email '''
+ try:
+ validate_email(email)
+ except ValidationError:
+ return 'Invalid email address: '+email
+ return ''
+ | Implement backend newaccount form verification | ## Code Before:
from django.shortcuts import render
from django.http import JsonResponse
import common.render
from common.settings import get_page_config
def form(request):
''' The signup form webpage '''
context = get_page_config(title='New User Sign Up')
context['form'] = [
{'label': 'User Name', 'name': 'username'},
{'label': 'Email Address', 'name': 'email'},
{'label': 'Password', 'type': 'password', 'name':'password'},
{'label': 'Re-enter Password', 'type': 'password', 'id':'repass'}
]
context['validators'] = [ 'signup/newaccount_validator.js' ]
return common.render.singleform(request, context)
def validate(request):
''' Signup form validation handler '''
return JsonResponse({"status": "error", "message": "Test Case"})
def submit(request):
''' Signup form submission handler '''
return HttpResponse('')
## Instruction:
Implement backend newaccount form verification
## Code After:
from django.http import JsonResponse
from django.contrib.auth.models import User
from django.core.validators import validate_email
from django.core.exceptions import ValidationError
from django.shortcuts import render
import urllib
import common.render
from common.settings import get_page_config
def form(request):
''' The signup form webpage '''
context = get_page_config(title='New User Sign Up')
context['form'] = [
{'label': 'User Name', 'name': 'username'},
{'label': 'Email Address', 'name': 'email'},
{'label': 'Password', 'type': 'password', 'name':'password'},
{'label': 'Re-enter Password', 'type': 'password', 'id':'repass'}
]
context['validators'] = [ 'signup/newaccount_validator.js' ]
return common.render.singleform(request, context)
def validate(request):
''' Signup form validation handler '''
msg = ''
if not('username' in request.GET):
msg = 'Name not given!'
elif not('email' in request.GET):
msg = 'Email not given!'
name = urllib.unquote(request.GET['username'])
email = urllib.unquote(request.GET['email'])
if not msg: msg = __validate_name__(name)
if not msg: msg = __validate_email__(email)
status = 'error' if msg else 'success'
return JsonResponse({"status": status, "message": msg})
def submit(request):
''' Signup form submission handler '''
return HttpResponse('')
def __validate_name__(name):
''' Internal validation function for username '''
lname = len(name)
if lname < 5:
return 'User name must be at least 5 characters long'
if lname > 64:
return 'User name must not be longer than 64 characters'
if len(User.objects.filter(username=name)):
return 'User name already in use'
return ''
def __validate_email__(email):
''' Internal validation function for email '''
try:
validate_email(email)
except ValidationError:
return 'Invalid email address: '+email
return ''
| + from django.http import JsonResponse
+ from django.contrib.auth.models import User
+ from django.core.validators import validate_email
+ from django.core.exceptions import ValidationError
from django.shortcuts import render
- from django.http import JsonResponse
+ import urllib
import common.render
from common.settings import get_page_config
def form(request):
''' The signup form webpage '''
context = get_page_config(title='New User Sign Up')
context['form'] = [
{'label': 'User Name', 'name': 'username'},
{'label': 'Email Address', 'name': 'email'},
{'label': 'Password', 'type': 'password', 'name':'password'},
{'label': 'Re-enter Password', 'type': 'password', 'id':'repass'}
]
context['validators'] = [ 'signup/newaccount_validator.js' ]
return common.render.singleform(request, context)
def validate(request):
''' Signup form validation handler '''
+ msg = ''
+ if not('username' in request.GET):
+ msg = 'Name not given!'
+ elif not('email' in request.GET):
+ msg = 'Email not given!'
+ name = urllib.unquote(request.GET['username'])
+ email = urllib.unquote(request.GET['email'])
+ if not msg: msg = __validate_name__(name)
+ if not msg: msg = __validate_email__(email)
+
+ status = 'error' if msg else 'success'
- return JsonResponse({"status": "error", "message": "Test Case"})
? ^^^^^^^ ^^^ ^^^^^^^
+ return JsonResponse({"status": status, "message": msg})
? ^^^^^^ ^ ^
def submit(request):
''' Signup form submission handler '''
return HttpResponse('')
+
+
+ def __validate_name__(name):
+ ''' Internal validation function for username '''
+ lname = len(name)
+ if lname < 5:
+ return 'User name must be at least 5 characters long'
+ if lname > 64:
+ return 'User name must not be longer than 64 characters'
+ if len(User.objects.filter(username=name)):
+ return 'User name already in use'
+ return ''
+
+ def __validate_email__(email):
+ ''' Internal validation function for email '''
+ try:
+ validate_email(email)
+ except ValidationError:
+ return 'Invalid email address: '+email
+ return '' |
0d2079b1dcb97708dc55c32d9e2c1a0f12595875 | salt/runners/launchd.py | salt/runners/launchd.py | '''
Manage launchd plist files
'''
# Import python libs
import os
import sys
def write_launchd_plist(program):
'''
Write a launchd plist for managing salt-master or salt-minion
CLI Example:
.. code-block:: bash
salt-run launchd.write_launchd_plist salt-master
'''
plist_sample_text = """
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key>
<string>org.saltstack.{program}</string>
<key>ProgramArguments</key>
<array>
<string>{python}</string>
<string>{script}</string>
</array>
<key>RunAtLoad</key>
<true/>
</dict>
</plist>
""".strip()
supported_programs = ['salt-master', 'salt-minion']
if program not in supported_programs:
sys.stderr.write("Supported programs: %r\n" % supported_programs)
sys.exit(-1)
sys.stdout.write(
plist_sample_text.format(
program=program,
python=sys.executable,
script=os.path.join(os.path.dirname(sys.executable), program)
)
)
| '''
Manage launchd plist files
'''
# Import python libs
import os
import sys
def write_launchd_plist(program):
'''
Write a launchd plist for managing salt-master or salt-minion
CLI Example:
.. code-block:: bash
salt-run launchd.write_launchd_plist salt-master
'''
plist_sample_text = '''
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key>
<string>org.saltstack.{program}</string>
<key>ProgramArguments</key>
<array>
<string>{python}</string>
<string>{script}</string>
</array>
<key>RunAtLoad</key>
<true/>
</dict>
</plist>
'''.strip()
supported_programs = ['salt-master', 'salt-minion']
if program not in supported_programs:
sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs))
sys.exit(-1)
sys.stdout.write(
plist_sample_text.format(
program=program,
python=sys.executable,
script=os.path.join(os.path.dirname(sys.executable), program)
)
)
| Replace string substitution with string formatting | Replace string substitution with string formatting
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | '''
Manage launchd plist files
'''
# Import python libs
import os
import sys
def write_launchd_plist(program):
'''
Write a launchd plist for managing salt-master or salt-minion
CLI Example:
.. code-block:: bash
salt-run launchd.write_launchd_plist salt-master
'''
- plist_sample_text = """
+ plist_sample_text = '''
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key>
<string>org.saltstack.{program}</string>
<key>ProgramArguments</key>
<array>
<string>{python}</string>
<string>{script}</string>
</array>
<key>RunAtLoad</key>
<true/>
</dict>
</plist>
- """.strip()
+ '''.strip()
supported_programs = ['salt-master', 'salt-minion']
if program not in supported_programs:
- sys.stderr.write("Supported programs: %r\n" % supported_programs)
+ sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs))
sys.exit(-1)
sys.stdout.write(
plist_sample_text.format(
program=program,
python=sys.executable,
script=os.path.join(os.path.dirname(sys.executable), program)
)
)
| Replace string substitution with string formatting | ## Code Before:
'''
Manage launchd plist files
'''
# Import python libs
import os
import sys
def write_launchd_plist(program):
'''
Write a launchd plist for managing salt-master or salt-minion
CLI Example:
.. code-block:: bash
salt-run launchd.write_launchd_plist salt-master
'''
plist_sample_text = """
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key>
<string>org.saltstack.{program}</string>
<key>ProgramArguments</key>
<array>
<string>{python}</string>
<string>{script}</string>
</array>
<key>RunAtLoad</key>
<true/>
</dict>
</plist>
""".strip()
supported_programs = ['salt-master', 'salt-minion']
if program not in supported_programs:
sys.stderr.write("Supported programs: %r\n" % supported_programs)
sys.exit(-1)
sys.stdout.write(
plist_sample_text.format(
program=program,
python=sys.executable,
script=os.path.join(os.path.dirname(sys.executable), program)
)
)
## Instruction:
Replace string substitution with string formatting
## Code After:
'''
Manage launchd plist files
'''
# Import python libs
import os
import sys
def write_launchd_plist(program):
'''
Write a launchd plist for managing salt-master or salt-minion
CLI Example:
.. code-block:: bash
salt-run launchd.write_launchd_plist salt-master
'''
plist_sample_text = '''
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key>
<string>org.saltstack.{program}</string>
<key>ProgramArguments</key>
<array>
<string>{python}</string>
<string>{script}</string>
</array>
<key>RunAtLoad</key>
<true/>
</dict>
</plist>
'''.strip()
supported_programs = ['salt-master', 'salt-minion']
if program not in supported_programs:
sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs))
sys.exit(-1)
sys.stdout.write(
plist_sample_text.format(
program=program,
python=sys.executable,
script=os.path.join(os.path.dirname(sys.executable), program)
)
)
| '''
Manage launchd plist files
'''
# Import python libs
import os
import sys
def write_launchd_plist(program):
'''
Write a launchd plist for managing salt-master or salt-minion
CLI Example:
.. code-block:: bash
salt-run launchd.write_launchd_plist salt-master
'''
- plist_sample_text = """
? ^^^
+ plist_sample_text = '''
? ^^^
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key>
<string>org.saltstack.{program}</string>
<key>ProgramArguments</key>
<array>
<string>{python}</string>
<string>{script}</string>
</array>
<key>RunAtLoad</key>
<true/>
</dict>
</plist>
- """.strip()
? ^^^
+ '''.strip()
? ^^^
supported_programs = ['salt-master', 'salt-minion']
if program not in supported_programs:
- sys.stderr.write("Supported programs: %r\n" % supported_programs)
? ^ ^ ^^^^
+ sys.stderr.write('Supported programs: {0!r}\n'.format(supported_programs))
? ^ ^^^ + ^^^^^^^^^ +
sys.exit(-1)
sys.stdout.write(
plist_sample_text.format(
program=program,
python=sys.executable,
script=os.path.join(os.path.dirname(sys.executable), program)
)
) |
ad415f26eec5c6a20c26123ccb6ce3e320ea9a69 | zou/app/blueprints/crud/asset_instance.py | zou/app/blueprints/crud/asset_instance.py | from zou.app.models.asset_instance import AssetInstance
from zou.app.services import assets_service, user_service
from zou.app.utils import permissions
from .base import BaseModelResource, BaseModelsResource
class AssetInstancesResource(BaseModelsResource):
def __init__(self):
BaseModelsResource.__init__(self, AssetInstance)
class AssetInstanceResource(BaseModelResource):
def __init__(self):
BaseModelResource.__init__(self, AssetInstance)
def check_read_permissions(self, instance):
if permissions.has_manager_permissions():
return True
else:
asset_instance = self.get_model_or_404(instance["id"])
asset = assets_service.get_asset(asset_instance.asset_id)
return user_service.check_has_task_related(asset["project_id"])
def check_update_permissions(self, asset_instance, data):
if permissions.has_manager_permissions():
return True
else:
return user_service.check_working_on_entity(
asset_instance["entity_id"]
)
| from zou.app.models.asset_instance import AssetInstance
from zou.app.services import assets_service, user_service
from zou.app.utils import permissions
from .base import BaseModelResource, BaseModelsResource
class AssetInstancesResource(BaseModelsResource):
def __init__(self):
BaseModelsResource.__init__(self, AssetInstance)
class AssetInstanceResource(BaseModelResource):
def __init__(self):
BaseModelResource.__init__(self, AssetInstance)
self.protected_fields.append(["number"])
def check_read_permissions(self, instance):
if permissions.has_manager_permissions():
return True
else:
asset_instance = self.get_model_or_404(instance["id"])
asset = assets_service.get_asset(asset_instance.asset_id)
return user_service.check_has_task_related(asset["project_id"])
def check_update_permissions(self, asset_instance, data):
if permissions.has_manager_permissions():
return True
else:
asset = assets_service.get_asset(asset_instance["asset_id"])
return user_service.check_has_task_related(asset["project_id"])
| Change asset instance update permissions | Change asset instance update permissions
* Do not allow to change instance number
* Allow to change instance name by a CG artist
| Python | agpl-3.0 | cgwire/zou | from zou.app.models.asset_instance import AssetInstance
from zou.app.services import assets_service, user_service
from zou.app.utils import permissions
from .base import BaseModelResource, BaseModelsResource
class AssetInstancesResource(BaseModelsResource):
def __init__(self):
BaseModelsResource.__init__(self, AssetInstance)
class AssetInstanceResource(BaseModelResource):
def __init__(self):
BaseModelResource.__init__(self, AssetInstance)
+ self.protected_fields.append(["number"])
def check_read_permissions(self, instance):
if permissions.has_manager_permissions():
return True
else:
asset_instance = self.get_model_or_404(instance["id"])
asset = assets_service.get_asset(asset_instance.asset_id)
return user_service.check_has_task_related(asset["project_id"])
def check_update_permissions(self, asset_instance, data):
if permissions.has_manager_permissions():
return True
else:
+ asset = assets_service.get_asset(asset_instance["asset_id"])
+ return user_service.check_has_task_related(asset["project_id"])
- return user_service.check_working_on_entity(
- asset_instance["entity_id"]
- )
| Change asset instance update permissions | ## Code Before:
from zou.app.models.asset_instance import AssetInstance
from zou.app.services import assets_service, user_service
from zou.app.utils import permissions
from .base import BaseModelResource, BaseModelsResource
class AssetInstancesResource(BaseModelsResource):
def __init__(self):
BaseModelsResource.__init__(self, AssetInstance)
class AssetInstanceResource(BaseModelResource):
def __init__(self):
BaseModelResource.__init__(self, AssetInstance)
def check_read_permissions(self, instance):
if permissions.has_manager_permissions():
return True
else:
asset_instance = self.get_model_or_404(instance["id"])
asset = assets_service.get_asset(asset_instance.asset_id)
return user_service.check_has_task_related(asset["project_id"])
def check_update_permissions(self, asset_instance, data):
if permissions.has_manager_permissions():
return True
else:
return user_service.check_working_on_entity(
asset_instance["entity_id"]
)
## Instruction:
Change asset instance update permissions
## Code After:
from zou.app.models.asset_instance import AssetInstance
from zou.app.services import assets_service, user_service
from zou.app.utils import permissions
from .base import BaseModelResource, BaseModelsResource
class AssetInstancesResource(BaseModelsResource):
def __init__(self):
BaseModelsResource.__init__(self, AssetInstance)
class AssetInstanceResource(BaseModelResource):
def __init__(self):
BaseModelResource.__init__(self, AssetInstance)
self.protected_fields.append(["number"])
def check_read_permissions(self, instance):
if permissions.has_manager_permissions():
return True
else:
asset_instance = self.get_model_or_404(instance["id"])
asset = assets_service.get_asset(asset_instance.asset_id)
return user_service.check_has_task_related(asset["project_id"])
def check_update_permissions(self, asset_instance, data):
if permissions.has_manager_permissions():
return True
else:
asset = assets_service.get_asset(asset_instance["asset_id"])
return user_service.check_has_task_related(asset["project_id"])
| from zou.app.models.asset_instance import AssetInstance
from zou.app.services import assets_service, user_service
from zou.app.utils import permissions
from .base import BaseModelResource, BaseModelsResource
class AssetInstancesResource(BaseModelsResource):
def __init__(self):
BaseModelsResource.__init__(self, AssetInstance)
class AssetInstanceResource(BaseModelResource):
def __init__(self):
BaseModelResource.__init__(self, AssetInstance)
+ self.protected_fields.append(["number"])
def check_read_permissions(self, instance):
if permissions.has_manager_permissions():
return True
else:
asset_instance = self.get_model_or_404(instance["id"])
asset = assets_service.get_asset(asset_instance.asset_id)
return user_service.check_has_task_related(asset["project_id"])
def check_update_permissions(self, asset_instance, data):
if permissions.has_manager_permissions():
return True
else:
+ asset = assets_service.get_asset(asset_instance["asset_id"])
+ return user_service.check_has_task_related(asset["project_id"])
- return user_service.check_working_on_entity(
- asset_instance["entity_id"]
- ) |
8ef4ca2166167f6370dd8c2f724e752210adf067 | sirius/SI_V07/__init__.py | sirius/SI_V07/__init__.py | from . import lattice as _lattice
from . import accelerator as _accelerator
from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = _accelerator.create_accelerator
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| Fix bug when family_data.py was deleted | Fix bug when family_data.py was deleted
| Python | mit | lnls-fac/sirius | from . import lattice as _lattice
from . import accelerator as _accelerator
- from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
- get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| Fix bug when family_data.py was deleted | ## Code Before:
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
## Instruction:
Fix bug when family_data.py was deleted
## Code After:
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = _accelerator.create_accelerator
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| from . import lattice as _lattice
from . import accelerator as _accelerator
- from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
- get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar] |
baaeb4fe0998bac8e0cb853d8124aa6134f55996 | poradnia/letters/admin.py | poradnia/letters/admin.py | from django.contrib import admin
from .models import Attachment, Letter
class AttachmentInline(admin.StackedInline):
'''
Stacked Inline View for Attachment
'''
model = Attachment
class LetterAdmin(admin.ModelAdmin):
'''
Admin View for Letter
'''
inlines = [
AttachmentInline,
]
admin.site.register(Letter, LetterAdmin)
| from django.contrib import admin
from .models import Attachment, Letter
class AttachmentInline(admin.StackedInline):
'''
Stacked Inline View for Attachment
'''
model = Attachment
@admin.register(Letter)
class LetterAdmin(admin.ModelAdmin):
'''
Admin View for Letter
'''
inlines = [
AttachmentInline,
]
| Rewrite DjangoAdmin in letters for decorators | Rewrite DjangoAdmin in letters for decorators
| Python | mit | rwakulszowa/poradnia,rwakulszowa/poradnia,watchdogpolska/poradnia.siecobywatelska.pl,watchdogpolska/poradnia.siecobywatelska.pl,watchdogpolska/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia.siecobywatelska.pl,rwakulszowa/poradnia,rwakulszowa/poradnia | from django.contrib import admin
from .models import Attachment, Letter
class AttachmentInline(admin.StackedInline):
'''
Stacked Inline View for Attachment
'''
model = Attachment
+ @admin.register(Letter)
class LetterAdmin(admin.ModelAdmin):
'''
Admin View for Letter
'''
inlines = [
AttachmentInline,
]
- admin.site.register(Letter, LetterAdmin)
- | Rewrite DjangoAdmin in letters for decorators | ## Code Before:
from django.contrib import admin
from .models import Attachment, Letter
class AttachmentInline(admin.StackedInline):
'''
Stacked Inline View for Attachment
'''
model = Attachment
class LetterAdmin(admin.ModelAdmin):
'''
Admin View for Letter
'''
inlines = [
AttachmentInline,
]
admin.site.register(Letter, LetterAdmin)
## Instruction:
Rewrite DjangoAdmin in letters for decorators
## Code After:
from django.contrib import admin
from .models import Attachment, Letter
class AttachmentInline(admin.StackedInline):
'''
Stacked Inline View for Attachment
'''
model = Attachment
@admin.register(Letter)
class LetterAdmin(admin.ModelAdmin):
'''
Admin View for Letter
'''
inlines = [
AttachmentInline,
]
| from django.contrib import admin
from .models import Attachment, Letter
class AttachmentInline(admin.StackedInline):
'''
Stacked Inline View for Attachment
'''
model = Attachment
+ @admin.register(Letter)
class LetterAdmin(admin.ModelAdmin):
'''
Admin View for Letter
'''
inlines = [
AttachmentInline,
]
-
- admin.site.register(Letter, LetterAdmin) |
aacb30d2d87cf83c3f09225532c06d81e399cae2 | spanky/commands/cmd_roster.py | spanky/commands/cmd_roster.py | import json
import click
from spanky.cli import pass_context
from spanky.lib.enroll import roster
@click.command('roster', short_help='Enroll / leave')
@click.argument('name')
@pass_context
def cli(ctx, name):
config = ctx.config.load('enroll.yml')
click.echo(json.dumps(list(roster(config, name))))
| import json
import click
from spanky.cli import pass_context
from spanky.lib.enroll import roster
@click.command('roster', short_help='Enroll / leave')
@click.argument('name')
@pass_context
def cli(ctx, name):
config = ctx.config.load('enroll.yml')
members = list(roster(config, name))
click.echo(json.dumps({'memebers': members}))
| Return members in roster cmd out. | Return members in roster cmd out.
| Python | bsd-3-clause | pglbutt/spanky,pglbutt/spanky,pglbutt/spanky | import json
import click
from spanky.cli import pass_context
from spanky.lib.enroll import roster
@click.command('roster', short_help='Enroll / leave')
@click.argument('name')
@pass_context
def cli(ctx, name):
config = ctx.config.load('enroll.yml')
- click.echo(json.dumps(list(roster(config, name))))
+ members = list(roster(config, name))
+ click.echo(json.dumps({'memebers': members}))
| Return members in roster cmd out. | ## Code Before:
import json
import click
from spanky.cli import pass_context
from spanky.lib.enroll import roster
@click.command('roster', short_help='Enroll / leave')
@click.argument('name')
@pass_context
def cli(ctx, name):
config = ctx.config.load('enroll.yml')
click.echo(json.dumps(list(roster(config, name))))
## Instruction:
Return members in roster cmd out.
## Code After:
import json
import click
from spanky.cli import pass_context
from spanky.lib.enroll import roster
@click.command('roster', short_help='Enroll / leave')
@click.argument('name')
@pass_context
def cli(ctx, name):
config = ctx.config.load('enroll.yml')
members = list(roster(config, name))
click.echo(json.dumps({'memebers': members}))
| import json
import click
from spanky.cli import pass_context
from spanky.lib.enroll import roster
@click.command('roster', short_help='Enroll / leave')
@click.argument('name')
@pass_context
def cli(ctx, name):
config = ctx.config.load('enroll.yml')
- click.echo(json.dumps(list(roster(config, name))))
+ members = list(roster(config, name))
+ click.echo(json.dumps({'memebers': members})) |
0177066012b3373753cba8baf86f00a365d7147b | findaconf/tests/config.py | findaconf/tests/config.py |
from decouple import config
from findaconf.tests.fake_data import fake_conference, seed
def set_app(app, db=False):
unset_app(db)
app.config['TESTING'] = True
app.config['WTF_CSRF_ENABLED'] = False
if db:
app.config['SQLALCHEMY_DATABASE_URI'] = config(
'DATABASE_URL_TEST',
default='sqlite:///' + app.config['BASEDIR'].child('findaconf',
'tests',
'tests.db')
)
test_app = app.test_client()
if db:
db.create_all()
seed(app, db)
[db.session.add(fake_conference(db)) for i in range(1, 43)]
db.session.commit()
return test_app
def unset_app(db=False):
if db:
db.session.remove()
db.drop_all()
|
from decouple import config
from findaconf.tests.fake_data import fake_conference, seed
def set_app(app, db=False):
# set test vars
app.config['TESTING'] = True
app.config['WTF_CSRF_ENABLED'] = False
# set test db
if db:
app.config['SQLALCHEMY_DATABASE_URI'] = config(
'DATABASE_URL_TEST',
default='sqlite:///' + app.config['BASEDIR'].child('findaconf',
'tests',
'tests.db')
)
# create test app
test_app = app.test_client()
# create and feed db tables
if db:
# start from a clean db
db.session.remove()
db.drop_all()
# create tables and feed them
db.create_all()
seed(app, db)
[db.session.add(fake_conference(db)) for i in range(1, 43)]
db.session.commit()
# return test app
return test_app
def unset_app(db=False):
if db:
db.session.remove()
db.drop_all()
| Fix bug that used dev db instead of test db | Fix bug that used dev db instead of test db
| Python | mit | cuducos/findaconf,koorukuroo/findaconf,cuducos/findaconf,koorukuroo/findaconf,koorukuroo/findaconf,cuducos/findaconf |
from decouple import config
from findaconf.tests.fake_data import fake_conference, seed
def set_app(app, db=False):
- unset_app(db)
+
+ # set test vars
app.config['TESTING'] = True
app.config['WTF_CSRF_ENABLED'] = False
+
+ # set test db
if db:
app.config['SQLALCHEMY_DATABASE_URI'] = config(
'DATABASE_URL_TEST',
default='sqlite:///' + app.config['BASEDIR'].child('findaconf',
'tests',
'tests.db')
)
+
+ # create test app
test_app = app.test_client()
+
+ # create and feed db tables
if db:
+
+ # start from a clean db
+ db.session.remove()
+ db.drop_all()
+
+ # create tables and feed them
db.create_all()
seed(app, db)
[db.session.add(fake_conference(db)) for i in range(1, 43)]
db.session.commit()
+
+ # return test app
return test_app
def unset_app(db=False):
if db:
db.session.remove()
db.drop_all()
| Fix bug that used dev db instead of test db | ## Code Before:
from decouple import config
from findaconf.tests.fake_data import fake_conference, seed
def set_app(app, db=False):
unset_app(db)
app.config['TESTING'] = True
app.config['WTF_CSRF_ENABLED'] = False
if db:
app.config['SQLALCHEMY_DATABASE_URI'] = config(
'DATABASE_URL_TEST',
default='sqlite:///' + app.config['BASEDIR'].child('findaconf',
'tests',
'tests.db')
)
test_app = app.test_client()
if db:
db.create_all()
seed(app, db)
[db.session.add(fake_conference(db)) for i in range(1, 43)]
db.session.commit()
return test_app
def unset_app(db=False):
if db:
db.session.remove()
db.drop_all()
## Instruction:
Fix bug that used dev db instead of test db
## Code After:
from decouple import config
from findaconf.tests.fake_data import fake_conference, seed
def set_app(app, db=False):
# set test vars
app.config['TESTING'] = True
app.config['WTF_CSRF_ENABLED'] = False
# set test db
if db:
app.config['SQLALCHEMY_DATABASE_URI'] = config(
'DATABASE_URL_TEST',
default='sqlite:///' + app.config['BASEDIR'].child('findaconf',
'tests',
'tests.db')
)
# create test app
test_app = app.test_client()
# create and feed db tables
if db:
# start from a clean db
db.session.remove()
db.drop_all()
# create tables and feed them
db.create_all()
seed(app, db)
[db.session.add(fake_conference(db)) for i in range(1, 43)]
db.session.commit()
# return test app
return test_app
def unset_app(db=False):
if db:
db.session.remove()
db.drop_all()
|
from decouple import config
from findaconf.tests.fake_data import fake_conference, seed
def set_app(app, db=False):
- unset_app(db)
+
+ # set test vars
app.config['TESTING'] = True
app.config['WTF_CSRF_ENABLED'] = False
+
+ # set test db
if db:
app.config['SQLALCHEMY_DATABASE_URI'] = config(
'DATABASE_URL_TEST',
default='sqlite:///' + app.config['BASEDIR'].child('findaconf',
'tests',
'tests.db')
)
+
+ # create test app
test_app = app.test_client()
+
+ # create and feed db tables
if db:
+
+ # start from a clean db
+ db.session.remove()
+ db.drop_all()
+
+ # create tables and feed them
db.create_all()
seed(app, db)
[db.session.add(fake_conference(db)) for i in range(1, 43)]
db.session.commit()
+
+ # return test app
return test_app
def unset_app(db=False):
if db:
db.session.remove()
db.drop_all() |
1726a73b81c8a7dfc3610690fe9272776e930f0f | aero/adapters/bower.py | aero/adapters/bower.py | __author__ = 'oliveiraev'
__all__ = ['Bower']
from re import sub
from re import split
from aero.__version__ import enc
from .base import BaseAdapter
class Bower(BaseAdapter):
"""
Twitter Bower - Browser package manager - Adapter
"""
def search(self, query):
return {}
response = self.command('search', query, ['--no-color'])[0].decode(*enc)
lst = dict([(self.package_name(k), v) for k, v in [
line.lstrip(' -').split(' ') for line in response.splitlines()
if line.startswith(' - ')]
])
if lst:
return lst
def install(self, query):
return self.shell('install', query)
def info(self, query):
response = self.command('info', query, ['--no-color'])[0].decode(*enc)
return response or ['Aborted: No info available']
| __author__ = 'oliveiraev'
__all__ = ['Bower']
from re import sub
from re import split
from aero.__version__ import enc
from .base import BaseAdapter
class Bower(BaseAdapter):
"""
Twitter Bower - Browser package manager - Adapter
"""
def search(self, query):
response = self.command('search', query, ['--no-color'])[0].decode(*enc)
lst = dict([(self.package_name(k), v) for k, v in [
line.lstrip(' -').split(' ') for line in response.splitlines()
if line.startswith(' - ')]
])
return lst or {}
def install(self, query):
return self.shell('install', query)
def info(self, query):
response = self.command('info', query, ['--no-color'])[0].decode(*enc)
return response or ['Aborted: No info available']
| Simplify return while we're at it | Simplify return while we're at it
| Python | bsd-3-clause | Aeronautics/aero | __author__ = 'oliveiraev'
__all__ = ['Bower']
from re import sub
from re import split
from aero.__version__ import enc
from .base import BaseAdapter
class Bower(BaseAdapter):
"""
Twitter Bower - Browser package manager - Adapter
"""
def search(self, query):
- return {}
-
response = self.command('search', query, ['--no-color'])[0].decode(*enc)
lst = dict([(self.package_name(k), v) for k, v in [
line.lstrip(' -').split(' ') for line in response.splitlines()
if line.startswith(' - ')]
])
- if lst:
- return lst
+ return lst or {}
def install(self, query):
return self.shell('install', query)
def info(self, query):
response = self.command('info', query, ['--no-color'])[0].decode(*enc)
return response or ['Aborted: No info available']
| Simplify return while we're at it | ## Code Before:
__author__ = 'oliveiraev'
__all__ = ['Bower']
from re import sub
from re import split
from aero.__version__ import enc
from .base import BaseAdapter
class Bower(BaseAdapter):
"""
Twitter Bower - Browser package manager - Adapter
"""
def search(self, query):
return {}
response = self.command('search', query, ['--no-color'])[0].decode(*enc)
lst = dict([(self.package_name(k), v) for k, v in [
line.lstrip(' -').split(' ') for line in response.splitlines()
if line.startswith(' - ')]
])
if lst:
return lst
def install(self, query):
return self.shell('install', query)
def info(self, query):
response = self.command('info', query, ['--no-color'])[0].decode(*enc)
return response or ['Aborted: No info available']
## Instruction:
Simplify return while we're at it
## Code After:
__author__ = 'oliveiraev'
__all__ = ['Bower']
from re import sub
from re import split
from aero.__version__ import enc
from .base import BaseAdapter
class Bower(BaseAdapter):
"""
Twitter Bower - Browser package manager - Adapter
"""
def search(self, query):
response = self.command('search', query, ['--no-color'])[0].decode(*enc)
lst = dict([(self.package_name(k), v) for k, v in [
line.lstrip(' -').split(' ') for line in response.splitlines()
if line.startswith(' - ')]
])
return lst or {}
def install(self, query):
return self.shell('install', query)
def info(self, query):
response = self.command('info', query, ['--no-color'])[0].decode(*enc)
return response or ['Aborted: No info available']
| __author__ = 'oliveiraev'
__all__ = ['Bower']
from re import sub
from re import split
from aero.__version__ import enc
from .base import BaseAdapter
class Bower(BaseAdapter):
"""
Twitter Bower - Browser package manager - Adapter
"""
def search(self, query):
- return {}
-
response = self.command('search', query, ['--no-color'])[0].decode(*enc)
lst = dict([(self.package_name(k), v) for k, v in [
line.lstrip(' -').split(' ') for line in response.splitlines()
if line.startswith(' - ')]
])
- if lst:
- return lst
? ----
+ return lst or {}
? ++++++
def install(self, query):
return self.shell('install', query)
def info(self, query):
response = self.command('info', query, ['--no-color'])[0].decode(*enc)
return response or ['Aborted: No info available'] |
2158edb92cba6c19fa258f19445191d0308c4153 | utils/async_tasks.py | utils/async_tasks.py | from utils.redis_store import store
def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60):
# Get task results previously stored in store
output, elapsed_time = store.get(store_key, include_elapsed_time=True)
# If there are no previously stored results (elapsed_time will be a magically big number) or
# if the previously stored results are older than refresh_time, then we trigger recompute of the
# task so that results are ready for next load.
if elapsed_time > refresh_time:
task_func.delay(store_key, *task_args, **task_kwargs)
return output
| from utils.redis_store import store
from celery.signals import task_postrun, task_prerun
def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True):
# Get task results previously stored in store
output, elapsed_time = store.get(store_key, include_elapsed_time=True)
# If there are no previously stored results (elapsed_time will be a magically big number) or
# if the previously stored results are older than refresh_time, then we trigger recompute of the
# task so that results are ready for next load.
# If run_once=True, we only trigger the recompute if the task is not already running
if elapsed_time > refresh_time:
if run_once:
# Check that it is not already running
computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__)
if store.get(computing_store_key):
# Task is already running, don't trigger running again
print('Skip computing data for {0}, already running'.format(store_key))
return output
task_func.delay(store_key, *task_args, **task_kwargs)
return output
@task_prerun.connect()
def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None):
# Set computing key
computing_store_key = 'computing-{0}'.format(task.name)
store.set(computing_store_key, {'running': True})
@task_postrun.connect()
def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None):
# Delete computing key (if present)
computing_store_key = 'computing-{0}'.format(task.name)
store.delete(computing_store_key)
| Add option to run async tasks only on at a time | Add option to run async tasks only on at a time
This is implemented with a simple lock like mechanism using redis.
| Python | agpl-3.0 | MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets | from utils.redis_store import store
+ from celery.signals import task_postrun, task_prerun
- def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60):
+ def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True):
+
# Get task results previously stored in store
output, elapsed_time = store.get(store_key, include_elapsed_time=True)
# If there are no previously stored results (elapsed_time will be a magically big number) or
# if the previously stored results are older than refresh_time, then we trigger recompute of the
# task so that results are ready for next load.
+ # If run_once=True, we only trigger the recompute if the task is not already running
if elapsed_time > refresh_time:
+ if run_once:
+ # Check that it is not already running
+ computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__)
+ if store.get(computing_store_key):
+ # Task is already running, don't trigger running again
+ print('Skip computing data for {0}, already running'.format(store_key))
+ return output
- task_func.delay(store_key, *task_args, **task_kwargs)
+ task_func.delay(store_key, *task_args, **task_kwargs)
return output
+
+ @task_prerun.connect()
+ def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None):
+ # Set computing key
+ computing_store_key = 'computing-{0}'.format(task.name)
+ store.set(computing_store_key, {'running': True})
+
+
+ @task_postrun.connect()
+ def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None):
+ # Delete computing key (if present)
+ computing_store_key = 'computing-{0}'.format(task.name)
+ store.delete(computing_store_key)
+ | Add option to run async tasks only on at a time | ## Code Before:
from utils.redis_store import store
def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60):
# Get task results previously stored in store
output, elapsed_time = store.get(store_key, include_elapsed_time=True)
# If there are no previously stored results (elapsed_time will be a magically big number) or
# if the previously stored results are older than refresh_time, then we trigger recompute of the
# task so that results are ready for next load.
if elapsed_time > refresh_time:
task_func.delay(store_key, *task_args, **task_kwargs)
return output
## Instruction:
Add option to run async tasks only on at a time
## Code After:
from utils.redis_store import store
from celery.signals import task_postrun, task_prerun
def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True):
# Get task results previously stored in store
output, elapsed_time = store.get(store_key, include_elapsed_time=True)
# If there are no previously stored results (elapsed_time will be a magically big number) or
# if the previously stored results are older than refresh_time, then we trigger recompute of the
# task so that results are ready for next load.
# If run_once=True, we only trigger the recompute if the task is not already running
if elapsed_time > refresh_time:
if run_once:
# Check that it is not already running
computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__)
if store.get(computing_store_key):
# Task is already running, don't trigger running again
print('Skip computing data for {0}, already running'.format(store_key))
return output
task_func.delay(store_key, *task_args, **task_kwargs)
return output
@task_prerun.connect()
def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None):
# Set computing key
computing_store_key = 'computing-{0}'.format(task.name)
store.set(computing_store_key, {'running': True})
@task_postrun.connect()
def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None):
# Delete computing key (if present)
computing_store_key = 'computing-{0}'.format(task.name)
store.delete(computing_store_key)
| from utils.redis_store import store
+ from celery.signals import task_postrun, task_prerun
- def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60):
+ def data_from_async_task(task_func, task_args, task_kwargs, store_key, refresh_time=60, run_once=True):
? +++++++++++++++
+
# Get task results previously stored in store
output, elapsed_time = store.get(store_key, include_elapsed_time=True)
# If there are no previously stored results (elapsed_time will be a magically big number) or
# if the previously stored results are older than refresh_time, then we trigger recompute of the
# task so that results are ready for next load.
+ # If run_once=True, we only trigger the recompute if the task is not already running
if elapsed_time > refresh_time:
+ if run_once:
+ # Check that it is not already running
+ computing_store_key = 'computing-{0}.{1}'.format(task_func.__module__, task_func.__name__)
+ if store.get(computing_store_key):
+ # Task is already running, don't trigger running again
+ print('Skip computing data for {0}, already running'.format(store_key))
+ return output
- task_func.delay(store_key, *task_args, **task_kwargs)
? -
+ task_func.delay(store_key, *task_args, **task_kwargs)
return output
+
+
+ @task_prerun.connect()
+ def task_prerun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None):
+ # Set computing key
+ computing_store_key = 'computing-{0}'.format(task.name)
+ store.set(computing_store_key, {'running': True})
+
+
+ @task_postrun.connect()
+ def task_postrun(signal=None, sender=None, task_id=None, task=None, args=None, kwargs=None, retval=None, state=None):
+ # Delete computing key (if present)
+ computing_store_key = 'computing-{0}'.format(task.name)
+ store.delete(computing_store_key) |
b5fa8ff1d86485c7f00ddecaef040ca66a817dfc | setup.py | setup.py | from distutils.core import setup
setup(
name='freki',
version='0.3.0-develop',
description='PDF-Extraction helper for RiPLEs pipeline.',
author='Michael Goodman, Ryan Georgi',
author_email='goodmami@uw.edu, rgeorgi@uw.edu',
url='https://github.com/xigt/freki',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Text Processing :: Linguistic',
'Topic :: Utilities'
],
keywords='nlp pdf ie text',
packages=['freki', 'freki.readers', 'freki.analyzers'],
install_requires=[
'numpy',
'matplotlib'
],
entry_points={
'console_scripts': [
'freki=freki.main:main'
]
},
)
| from distutils.core import setup
setup(
name='freki',
version='0.3.0-develop',
description='PDF-Extraction helper for RiPLEs pipeline.',
author='Michael Goodman, Ryan Georgi',
author_email='goodmami@uw.edu, rgeorgi@uw.edu',
url='https://github.com/xigt/freki',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Text Processing :: Linguistic',
'Topic :: Utilities'
],
keywords='nlp pdf ie text',
packages=['freki', 'freki.readers', 'freki.analyzers'],
install_requires=[
'numpy',
'matplotlib',
'chardet'
],
entry_points={
'console_scripts': [
'freki=freki.main:main'
]
},
)
| Add Chardet as installation dependency | Add Chardet as installation dependency
| Python | mit | xigt/freki,xigt/freki | from distutils.core import setup
setup(
name='freki',
version='0.3.0-develop',
description='PDF-Extraction helper for RiPLEs pipeline.',
author='Michael Goodman, Ryan Georgi',
author_email='goodmami@uw.edu, rgeorgi@uw.edu',
url='https://github.com/xigt/freki',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Text Processing :: Linguistic',
'Topic :: Utilities'
],
keywords='nlp pdf ie text',
packages=['freki', 'freki.readers', 'freki.analyzers'],
install_requires=[
'numpy',
- 'matplotlib'
+ 'matplotlib',
+ 'chardet'
],
entry_points={
'console_scripts': [
'freki=freki.main:main'
]
},
)
| Add Chardet as installation dependency | ## Code Before:
from distutils.core import setup
setup(
name='freki',
version='0.3.0-develop',
description='PDF-Extraction helper for RiPLEs pipeline.',
author='Michael Goodman, Ryan Georgi',
author_email='goodmami@uw.edu, rgeorgi@uw.edu',
url='https://github.com/xigt/freki',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Text Processing :: Linguistic',
'Topic :: Utilities'
],
keywords='nlp pdf ie text',
packages=['freki', 'freki.readers', 'freki.analyzers'],
install_requires=[
'numpy',
'matplotlib'
],
entry_points={
'console_scripts': [
'freki=freki.main:main'
]
},
)
## Instruction:
Add Chardet as installation dependency
## Code After:
from distutils.core import setup
setup(
name='freki',
version='0.3.0-develop',
description='PDF-Extraction helper for RiPLEs pipeline.',
author='Michael Goodman, Ryan Georgi',
author_email='goodmami@uw.edu, rgeorgi@uw.edu',
url='https://github.com/xigt/freki',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Text Processing :: Linguistic',
'Topic :: Utilities'
],
keywords='nlp pdf ie text',
packages=['freki', 'freki.readers', 'freki.analyzers'],
install_requires=[
'numpy',
'matplotlib',
'chardet'
],
entry_points={
'console_scripts': [
'freki=freki.main:main'
]
},
)
| from distutils.core import setup
setup(
name='freki',
version='0.3.0-develop',
description='PDF-Extraction helper for RiPLEs pipeline.',
author='Michael Goodman, Ryan Georgi',
author_email='goodmami@uw.edu, rgeorgi@uw.edu',
url='https://github.com/xigt/freki',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Text Processing :: Linguistic',
'Topic :: Utilities'
],
keywords='nlp pdf ie text',
packages=['freki', 'freki.readers', 'freki.analyzers'],
install_requires=[
'numpy',
- 'matplotlib'
+ 'matplotlib',
? +
+ 'chardet'
],
entry_points={
'console_scripts': [
'freki=freki.main:main'
]
},
) |
d6ff777c7fb3f645c021da1319bb5d78d13aa9db | meshnet/interface.py | meshnet/interface.py | import serial
import struct
from siphashc import siphash
def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes):
packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data)
return struct.pack("Q", siphash(key, packed_data))
class SerialMessage(object):
def __init__(self):
pass
def serialize(self):
pass
class Connection(object):
def __init__(self, device):
self._device = device
self._conn = None
def connect(self):
self._conn = serial.Serial(self._device, 115200)
| import serial
import struct
from siphashc import siphash
def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes):
packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data
return struct.pack(">Q", siphash(key, packed_data))
class SerialMessage(object):
def __init__(self):
pass
def serialize(self):
pass
class Connection(object):
def __init__(self, device):
self._device = device
self._conn = None
def connect(self):
self._conn = serial.Serial(self._device, 115200)
| Fix python siphashing to match c implementation | Fix python siphashing to match c implementation
Signed-off-by: Jan Losinski <577c4104c61edf9f052c616c0c23e67bef4a9955@wh2.tu-dresden.de>
| Python | bsd-3-clause | janLo/automation_mesh,janLo/automation_mesh,janLo/automation_mesh | import serial
import struct
from siphashc import siphash
+
- def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes):
+ def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes):
- packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data)
+ packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data
- return struct.pack("Q", siphash(key, packed_data))
+ return struct.pack(">Q", siphash(key, packed_data))
class SerialMessage(object):
def __init__(self):
pass
def serialize(self):
pass
-
-
class Connection(object):
-
def __init__(self, device):
self._device = device
self._conn = None
def connect(self):
self._conn = serial.Serial(self._device, 115200)
| Fix python siphashing to match c implementation | ## Code Before:
import serial
import struct
from siphashc import siphash
def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes):
packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data)
return struct.pack("Q", siphash(key, packed_data))
class SerialMessage(object):
def __init__(self):
pass
def serialize(self):
pass
class Connection(object):
def __init__(self, device):
self._device = device
self._conn = None
def connect(self):
self._conn = serial.Serial(self._device, 115200)
## Instruction:
Fix python siphashing to match c implementation
## Code After:
import serial
import struct
from siphashc import siphash
def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes):
packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data
return struct.pack(">Q", siphash(key, packed_data))
class SerialMessage(object):
def __init__(self):
pass
def serialize(self):
pass
class Connection(object):
def __init__(self, device):
self._device = device
self._conn = None
def connect(self):
self._conn = serial.Serial(self._device, 115200)
| import serial
import struct
from siphashc import siphash
+
- def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes):
? --
+ def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes):
? ++++
- packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data)
? - - ^ -
+ packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data
? ^^^
- return struct.pack("Q", siphash(key, packed_data))
+ return struct.pack(">Q", siphash(key, packed_data))
? +
class SerialMessage(object):
def __init__(self):
pass
def serialize(self):
pass
-
-
class Connection(object):
-
def __init__(self, device):
self._device = device
self._conn = None
def connect(self):
self._conn = serial.Serial(self._device, 115200) |
1a761c9360f185d6bd07be9f16ea2cfa239f4bdd | groupy/api/base.py | groupy/api/base.py | from groupy import utils
class Manager:
"""Class for interacting with the endpoint for a resource.
:param session: the requests session
:type session: :class:`~groupy.session.Session`
:param str path: path relative to the base URL
"""
#: the base URL
base_url = 'https://api.groupme.com/v3/'
def __init__(self, session, path=None):
self.session = session
self.url = utils.urljoin(self.base_url, path)
class Resource:
def __init__(self, **data):
self.data = data
def __getattr__(self, attr):
if attr not in self.data:
error_message = 'this {!s} resource does not have a {!r} field'
raise AttributeError(error_message.format(self.__class__.__name__,
attr))
return self.data[attr]
class ManagedResource(Resource):
"""Class to represent an API object."""
def __init__(self, manager, **data):
"""Create an instance of the resource.
:param manager: the resource's manager
:type manager: :class:`~groupy.api.base.Manager`
:param kwargs data: the resource data
"""
super().__init__(**data)
self.manager = manager
| from groupy import utils
class Manager:
"""Class for interacting with the endpoint for a resource.
:param session: the requests session
:type session: :class:`~groupy.session.Session`
:param str path: path relative to the base URL
"""
#: the base URL
base_url = 'https://api.groupme.com/v3/'
def __init__(self, session, path=None):
self.session = session
self.url = utils.urljoin(self.base_url, path)
class Resource:
def __init__(self, **data):
self.data = data
def __getattr__(self, attr):
if attr not in self.data:
error_message = 'this {!s} resource does not have a {!r} field'
raise AttributeError(error_message.format(self.__class__.__name__,
attr))
return self.data[attr]
def __getstate__(self):
return self.__dict__
def __setstate__(self, d):
self.__dict__.update(d)
class ManagedResource(Resource):
"""Class to represent an API object."""
def __init__(self, manager, **data):
"""Create an instance of the resource.
:param manager: the resource's manager
:type manager: :class:`~groupy.api.base.Manager`
:param kwargs data: the resource data
"""
super().__init__(**data)
self.manager = manager
| Fix pickling/unpickling of Resource objects | Fix pickling/unpickling of Resource objects
Add __getstate__ and __setstate__ methods to the Resource class to avoid hitting the recursion limit when trying to pickle/unpickle Resource objects.
A similar issue/solution can be found here: https://stackoverflow.com/a/12102691
| Python | apache-2.0 | rhgrant10/Groupy | from groupy import utils
class Manager:
"""Class for interacting with the endpoint for a resource.
:param session: the requests session
:type session: :class:`~groupy.session.Session`
:param str path: path relative to the base URL
"""
#: the base URL
base_url = 'https://api.groupme.com/v3/'
def __init__(self, session, path=None):
self.session = session
self.url = utils.urljoin(self.base_url, path)
class Resource:
def __init__(self, **data):
self.data = data
def __getattr__(self, attr):
if attr not in self.data:
error_message = 'this {!s} resource does not have a {!r} field'
raise AttributeError(error_message.format(self.__class__.__name__,
attr))
return self.data[attr]
+ def __getstate__(self):
+ return self.__dict__
+
+ def __setstate__(self, d):
+ self.__dict__.update(d)
+
class ManagedResource(Resource):
"""Class to represent an API object."""
def __init__(self, manager, **data):
"""Create an instance of the resource.
:param manager: the resource's manager
:type manager: :class:`~groupy.api.base.Manager`
:param kwargs data: the resource data
"""
super().__init__(**data)
self.manager = manager
| Fix pickling/unpickling of Resource objects | ## Code Before:
from groupy import utils
class Manager:
"""Class for interacting with the endpoint for a resource.
:param session: the requests session
:type session: :class:`~groupy.session.Session`
:param str path: path relative to the base URL
"""
#: the base URL
base_url = 'https://api.groupme.com/v3/'
def __init__(self, session, path=None):
self.session = session
self.url = utils.urljoin(self.base_url, path)
class Resource:
def __init__(self, **data):
self.data = data
def __getattr__(self, attr):
if attr not in self.data:
error_message = 'this {!s} resource does not have a {!r} field'
raise AttributeError(error_message.format(self.__class__.__name__,
attr))
return self.data[attr]
class ManagedResource(Resource):
"""Class to represent an API object."""
def __init__(self, manager, **data):
"""Create an instance of the resource.
:param manager: the resource's manager
:type manager: :class:`~groupy.api.base.Manager`
:param kwargs data: the resource data
"""
super().__init__(**data)
self.manager = manager
## Instruction:
Fix pickling/unpickling of Resource objects
## Code After:
from groupy import utils
class Manager:
"""Class for interacting with the endpoint for a resource.
:param session: the requests session
:type session: :class:`~groupy.session.Session`
:param str path: path relative to the base URL
"""
#: the base URL
base_url = 'https://api.groupme.com/v3/'
def __init__(self, session, path=None):
self.session = session
self.url = utils.urljoin(self.base_url, path)
class Resource:
def __init__(self, **data):
self.data = data
def __getattr__(self, attr):
if attr not in self.data:
error_message = 'this {!s} resource does not have a {!r} field'
raise AttributeError(error_message.format(self.__class__.__name__,
attr))
return self.data[attr]
def __getstate__(self):
return self.__dict__
def __setstate__(self, d):
self.__dict__.update(d)
class ManagedResource(Resource):
"""Class to represent an API object."""
def __init__(self, manager, **data):
"""Create an instance of the resource.
:param manager: the resource's manager
:type manager: :class:`~groupy.api.base.Manager`
:param kwargs data: the resource data
"""
super().__init__(**data)
self.manager = manager
| from groupy import utils
class Manager:
"""Class for interacting with the endpoint for a resource.
:param session: the requests session
:type session: :class:`~groupy.session.Session`
:param str path: path relative to the base URL
"""
#: the base URL
base_url = 'https://api.groupme.com/v3/'
def __init__(self, session, path=None):
self.session = session
self.url = utils.urljoin(self.base_url, path)
class Resource:
def __init__(self, **data):
self.data = data
def __getattr__(self, attr):
if attr not in self.data:
error_message = 'this {!s} resource does not have a {!r} field'
raise AttributeError(error_message.format(self.__class__.__name__,
attr))
return self.data[attr]
+ def __getstate__(self):
+ return self.__dict__
+
+ def __setstate__(self, d):
+ self.__dict__.update(d)
+
class ManagedResource(Resource):
"""Class to represent an API object."""
def __init__(self, manager, **data):
"""Create an instance of the resource.
:param manager: the resource's manager
:type manager: :class:`~groupy.api.base.Manager`
:param kwargs data: the resource data
"""
super().__init__(**data)
self.manager = manager |
93870152b4afb04f1547378184e2cee0bd0dd45f | kobo/apps/languages/serializers/base.py | kobo/apps/languages/serializers/base.py | from collections import defaultdict, OrderedDict
from django.db import models
from rest_framework import serializers
class BaseServiceSerializer(serializers.ModelSerializer):
class Meta:
fields = [
'name',
'code',
]
class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer):
region = serializers.SerializerMethodField()
service = serializers.SerializerMethodField()
language = serializers.SerializerMethodField()
class Meta:
fields = '__all__'
def get_list_serializer_class(self):
pass
def get_region(self, through_instance):
if through_instance.region:
return through_instance.region.code
return None
def get_service(self, through_instance):
return through_instance.service.code
def get_language(self, through_instance):
return through_instance.language.code
class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer):
def to_representation(self, data):
"""
Override `ListSerializer` behaviour to display services as a dictionary
instead of a list
"""
iterable = data.all() if isinstance(data, models.Manager) else data
formatted_data = self._get_formatted_data(iterable)
representation = []
for service_code, service in formatted_data.items():
representation.append({service_code: service})
return representation
| from collections import defaultdict, OrderedDict
from django.db import models
from rest_framework import serializers
class BaseServiceSerializer(serializers.ModelSerializer):
class Meta:
fields = [
'name',
'code',
]
class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer):
region = serializers.SerializerMethodField()
service = serializers.SerializerMethodField()
language = serializers.SerializerMethodField()
class Meta:
fields = '__all__'
def get_list_serializer_class(self):
pass
def get_region(self, through_instance):
if through_instance.region:
return through_instance.region.code
return None
def get_service(self, through_instance):
return through_instance.service.code
def get_language(self, through_instance):
return through_instance.language.code
class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer):
# Force `ListSerializer` to return a dict, not a list
@property
def data(self):
ret = serializers.BaseSerializer.data.fget(self)
return serializers.ReturnDict(ret, serializer=self)
def to_representation(self, data):
"""
Override `ListSerializer` behaviour to display services as a dictionary
instead of a list
"""
iterable = data.all() if isinstance(data, models.Manager) else data
formatted_data = self._get_formatted_data(iterable)
return formatted_data
| Return a dictionary for transcription/translation services (instead of list) | Return a dictionary for transcription/translation services (instead of list)
| Python | agpl-3.0 | kobotoolbox/kpi,kobotoolbox/kpi,kobotoolbox/kpi,kobotoolbox/kpi,kobotoolbox/kpi | from collections import defaultdict, OrderedDict
from django.db import models
from rest_framework import serializers
class BaseServiceSerializer(serializers.ModelSerializer):
class Meta:
fields = [
'name',
'code',
]
class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer):
region = serializers.SerializerMethodField()
service = serializers.SerializerMethodField()
language = serializers.SerializerMethodField()
class Meta:
fields = '__all__'
def get_list_serializer_class(self):
pass
def get_region(self, through_instance):
if through_instance.region:
return through_instance.region.code
return None
def get_service(self, through_instance):
return through_instance.service.code
def get_language(self, through_instance):
return through_instance.language.code
class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer):
+ # Force `ListSerializer` to return a dict, not a list
+ @property
+ def data(self):
+ ret = serializers.BaseSerializer.data.fget(self)
+ return serializers.ReturnDict(ret, serializer=self)
+
def to_representation(self, data):
"""
Override `ListSerializer` behaviour to display services as a dictionary
instead of a list
"""
iterable = data.all() if isinstance(data, models.Manager) else data
formatted_data = self._get_formatted_data(iterable)
+ return formatted_data
- representation = []
- for service_code, service in formatted_data.items():
- representation.append({service_code: service})
- return representation
| Return a dictionary for transcription/translation services (instead of list) | ## Code Before:
from collections import defaultdict, OrderedDict
from django.db import models
from rest_framework import serializers
class BaseServiceSerializer(serializers.ModelSerializer):
class Meta:
fields = [
'name',
'code',
]
class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer):
region = serializers.SerializerMethodField()
service = serializers.SerializerMethodField()
language = serializers.SerializerMethodField()
class Meta:
fields = '__all__'
def get_list_serializer_class(self):
pass
def get_region(self, through_instance):
if through_instance.region:
return through_instance.region.code
return None
def get_service(self, through_instance):
return through_instance.service.code
def get_language(self, through_instance):
return through_instance.language.code
class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer):
def to_representation(self, data):
"""
Override `ListSerializer` behaviour to display services as a dictionary
instead of a list
"""
iterable = data.all() if isinstance(data, models.Manager) else data
formatted_data = self._get_formatted_data(iterable)
representation = []
for service_code, service in formatted_data.items():
representation.append({service_code: service})
return representation
## Instruction:
Return a dictionary for transcription/translation services (instead of list)
## Code After:
from collections import defaultdict, OrderedDict
from django.db import models
from rest_framework import serializers
class BaseServiceSerializer(serializers.ModelSerializer):
class Meta:
fields = [
'name',
'code',
]
class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer):
region = serializers.SerializerMethodField()
service = serializers.SerializerMethodField()
language = serializers.SerializerMethodField()
class Meta:
fields = '__all__'
def get_list_serializer_class(self):
pass
def get_region(self, through_instance):
if through_instance.region:
return through_instance.region.code
return None
def get_service(self, through_instance):
return through_instance.service.code
def get_language(self, through_instance):
return through_instance.language.code
class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer):
# Force `ListSerializer` to return a dict, not a list
@property
def data(self):
ret = serializers.BaseSerializer.data.fget(self)
return serializers.ReturnDict(ret, serializer=self)
def to_representation(self, data):
"""
Override `ListSerializer` behaviour to display services as a dictionary
instead of a list
"""
iterable = data.all() if isinstance(data, models.Manager) else data
formatted_data = self._get_formatted_data(iterable)
return formatted_data
| from collections import defaultdict, OrderedDict
from django.db import models
from rest_framework import serializers
class BaseServiceSerializer(serializers.ModelSerializer):
class Meta:
fields = [
'name',
'code',
]
class BaseServiceLanguageM2MSerializer(serializers.ModelSerializer):
region = serializers.SerializerMethodField()
service = serializers.SerializerMethodField()
language = serializers.SerializerMethodField()
class Meta:
fields = '__all__'
def get_list_serializer_class(self):
pass
def get_region(self, through_instance):
if through_instance.region:
return through_instance.region.code
return None
def get_service(self, through_instance):
return through_instance.service.code
def get_language(self, through_instance):
return through_instance.language.code
class BaseServiceLanguageM2MListSerializer(serializers.ListSerializer):
+ # Force `ListSerializer` to return a dict, not a list
+ @property
+ def data(self):
+ ret = serializers.BaseSerializer.data.fget(self)
+ return serializers.ReturnDict(ret, serializer=self)
+
def to_representation(self, data):
"""
Override `ListSerializer` behaviour to display services as a dictionary
instead of a list
"""
iterable = data.all() if isinstance(data, models.Manager) else data
formatted_data = self._get_formatted_data(iterable)
+ return formatted_data
- representation = []
- for service_code, service in formatted_data.items():
- representation.append({service_code: service})
- return representation |
78f049ce9713dabd3eec544494dadcab7ff93d4c | sui_hei/templatetags/markdown.py | sui_hei/templatetags/markdown.py | import re
from bs4 import BeautifulSoup
from django import template
from django.template.defaultfilters import stringfilter
from markdown import markdown as md
register = template.Library()
@stringfilter
@register.filter(is_safe=True)
def text2md(value):
'''
convert markdown-like text to html.
strip header <p> and footer </p> if p is True.
'''
returns = md(value, ['markdown.extensions.extra'])
returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8")
return returns.decode('utf8')
@register.filter(is_safe=True)
@stringfilter
def line2md(value, p=True):
returns = BeautifulSoup(value, 'html5lib').get_text()
returns = re.sub("^([*+-]) ", r"\\\1 ", returns)
returns = md(returns, ['markdown.extensions.extra'])
if p: returns = returns[3:-4]
return returns
| import re
from bs4 import BeautifulSoup
from django import template
from django.template.defaultfilters import stringfilter
from markdown import markdown as md
from markdown.extensions.headerid import HeaderIdExtension
register = template.Library()
@stringfilter
@register.filter(is_safe=True)
def text2md(value):
'''
convert markdown-like text to html.
strip header <p> and footer </p> if p is True.
'''
returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)])
returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8")
return returns.decode('utf8')
@register.filter(is_safe=True)
@stringfilter
def line2md(value, p=True):
returns = BeautifulSoup(value, 'html5lib').get_text()
returns = re.sub("^([*+-]) ", r"\\\1 ", returns)
returns = md(returns, ['markdown.extensions.extra'])
if p: returns = returns[3:-4]
return returns
| Add header id extension for github preferences | Add header id extension for github preferences
| Python | mit | heyrict/cindy,heyrict/cindy,heyrict/cindy | import re
from bs4 import BeautifulSoup
from django import template
from django.template.defaultfilters import stringfilter
from markdown import markdown as md
+ from markdown.extensions.headerid import HeaderIdExtension
register = template.Library()
@stringfilter
@register.filter(is_safe=True)
def text2md(value):
'''
convert markdown-like text to html.
strip header <p> and footer </p> if p is True.
'''
- returns = md(value, ['markdown.extensions.extra'])
+ returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)])
returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8")
return returns.decode('utf8')
@register.filter(is_safe=True)
@stringfilter
def line2md(value, p=True):
returns = BeautifulSoup(value, 'html5lib').get_text()
returns = re.sub("^([*+-]) ", r"\\\1 ", returns)
returns = md(returns, ['markdown.extensions.extra'])
if p: returns = returns[3:-4]
return returns
| Add header id extension for github preferences | ## Code Before:
import re
from bs4 import BeautifulSoup
from django import template
from django.template.defaultfilters import stringfilter
from markdown import markdown as md
register = template.Library()
@stringfilter
@register.filter(is_safe=True)
def text2md(value):
'''
convert markdown-like text to html.
strip header <p> and footer </p> if p is True.
'''
returns = md(value, ['markdown.extensions.extra'])
returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8")
return returns.decode('utf8')
@register.filter(is_safe=True)
@stringfilter
def line2md(value, p=True):
returns = BeautifulSoup(value, 'html5lib').get_text()
returns = re.sub("^([*+-]) ", r"\\\1 ", returns)
returns = md(returns, ['markdown.extensions.extra'])
if p: returns = returns[3:-4]
return returns
## Instruction:
Add header id extension for github preferences
## Code After:
import re
from bs4 import BeautifulSoup
from django import template
from django.template.defaultfilters import stringfilter
from markdown import markdown as md
from markdown.extensions.headerid import HeaderIdExtension
register = template.Library()
@stringfilter
@register.filter(is_safe=True)
def text2md(value):
'''
convert markdown-like text to html.
strip header <p> and footer </p> if p is True.
'''
returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)])
returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8")
return returns.decode('utf8')
@register.filter(is_safe=True)
@stringfilter
def line2md(value, p=True):
returns = BeautifulSoup(value, 'html5lib').get_text()
returns = re.sub("^([*+-]) ", r"\\\1 ", returns)
returns = md(returns, ['markdown.extensions.extra'])
if p: returns = returns[3:-4]
return returns
| import re
from bs4 import BeautifulSoup
from django import template
from django.template.defaultfilters import stringfilter
from markdown import markdown as md
+ from markdown.extensions.headerid import HeaderIdExtension
register = template.Library()
@stringfilter
@register.filter(is_safe=True)
def text2md(value):
'''
convert markdown-like text to html.
strip header <p> and footer </p> if p is True.
'''
- returns = md(value, ['markdown.extensions.extra'])
+ returns = md(value, ['markdown.extensions.extra', HeaderIdExtension(level=4)])
? ++++++++++++++++++++++++++++
returns = BeautifulSoup(returns, 'html5lib').prettify(encoding="utf8")
return returns.decode('utf8')
@register.filter(is_safe=True)
@stringfilter
def line2md(value, p=True):
returns = BeautifulSoup(value, 'html5lib').get_text()
returns = re.sub("^([*+-]) ", r"\\\1 ", returns)
returns = md(returns, ['markdown.extensions.extra'])
if p: returns = returns[3:-4]
return returns |
7aaa385da78bef57c8b6339f6db04044ace08807 | api/taxonomies/serializers.py | api/taxonomies/serializers.py | from rest_framework import serializers as ser
from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField
class TaxonomyField(ser.Field):
def to_representation(self, obj):
if obj is not None:
return {'id': obj._id,
'text': obj.text}
return None
def to_internal_value(self, data):
return data
class TaxonomySerializer(JSONAPISerializer):
filterable_fields = frozenset([
'text',
'parents',
'id'
])
id = ser.CharField(source='_id', required=True)
text = ser.CharField(max_length=200)
parents = JSONAPIListField(child=TaxonomyField())
links = LinksField({
'parents': 'get_parent_urls',
'self': 'get_absolute_url',
})
def get_parent_urls(self, obj):
return [p.get_absolute_url() for p in obj.parents]
def get_absolute_url(self, obj):
return obj.get_absolute_url()
class Meta:
type_ = 'taxonomies'
| from rest_framework import serializers as ser
from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField
class TaxonomyField(ser.Field):
def to_representation(self, obj):
if obj is not None:
return {'id': obj._id,
'text': obj.text}
return None
def to_internal_value(self, data):
return data
class TaxonomySerializer(JSONAPISerializer):
filterable_fields = frozenset([
'text',
'parents',
'id'
])
id = ser.CharField(source='_id', required=True)
text = ser.CharField(max_length=200)
parents = JSONAPIListField(child=TaxonomyField())
child_count = ser.IntegerField()
links = LinksField({
'parents': 'get_parent_urls',
'self': 'get_absolute_url',
})
def get_parent_urls(self, obj):
return [p.get_absolute_url() for p in obj.parents]
def get_absolute_url(self, obj):
return obj.get_absolute_url()
class Meta:
type_ = 'taxonomies'
| Add child_count taken from new Subject property | Add child_count taken from new Subject property
| Python | apache-2.0 | adlius/osf.io,brianjgeiger/osf.io,chrisseto/osf.io,rdhyee/osf.io,sloria/osf.io,brianjgeiger/osf.io,sloria/osf.io,binoculars/osf.io,mattclark/osf.io,saradbowman/osf.io,aaxelb/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,laurenrevere/osf.io,samchrisinger/osf.io,binoculars/osf.io,aaxelb/osf.io,rdhyee/osf.io,alexschiller/osf.io,HalcyonChimera/osf.io,mfraezz/osf.io,cwisecarver/osf.io,CenterForOpenScience/osf.io,acshi/osf.io,cslzchen/osf.io,mattclark/osf.io,crcresearch/osf.io,TomBaxter/osf.io,acshi/osf.io,samchrisinger/osf.io,monikagrabowska/osf.io,laurenrevere/osf.io,adlius/osf.io,acshi/osf.io,mluo613/osf.io,HalcyonChimera/osf.io,felliott/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,felliott/osf.io,mattclark/osf.io,saradbowman/osf.io,erinspace/osf.io,hmoco/osf.io,emetsger/osf.io,alexschiller/osf.io,icereval/osf.io,mfraezz/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,chrisseto/osf.io,erinspace/osf.io,monikagrabowska/osf.io,leb2dg/osf.io,pattisdr/osf.io,CenterForOpenScience/osf.io,monikagrabowska/osf.io,laurenrevere/osf.io,Nesiehr/osf.io,acshi/osf.io,cslzchen/osf.io,caseyrollins/osf.io,monikagrabowska/osf.io,emetsger/osf.io,adlius/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,crcresearch/osf.io,cwisecarver/osf.io,brianjgeiger/osf.io,mluo613/osf.io,Nesiehr/osf.io,sloria/osf.io,cslzchen/osf.io,icereval/osf.io,caseyrollins/osf.io,binoculars/osf.io,aaxelb/osf.io,caseyrollins/osf.io,cslzchen/osf.io,icereval/osf.io,CenterForOpenScience/osf.io,erinspace/osf.io,acshi/osf.io,brianjgeiger/osf.io,adlius/osf.io,caneruguz/osf.io,pattisdr/osf.io,cwisecarver/osf.io,mluo613/osf.io,chrisseto/osf.io,leb2dg/osf.io,mluo613/osf.io,alexschiller/osf.io,samchrisinger/osf.io,chennan47/osf.io,caneruguz/osf.io,Nesiehr/osf.io,CenterForOpenScience/osf.io,rdhyee/osf.io,chrisseto/osf.io,mluo613/osf.io,Johnetordoff/osf.io,rdhyee/osf.io,felliott/osf.io,mfraezz/osf.io,aaxelb/osf.io,chennan47/osf.io,crcresearch/osf.io,alexschiller/osf.io,pattisdr/osf.io,HalcyonChimera/osf.io,emetsger/osf.io,leb2dg/osf.io,emetsger/osf.io,cwisecarver/osf.io,alexschiller/osf.io,Nesiehr/osf.io,hmoco/osf.io,hmoco/osf.io,felliott/osf.io,chennan47/osf.io,samchrisinger/osf.io,mfraezz/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,TomBaxter/osf.io,baylee-d/osf.io | from rest_framework import serializers as ser
from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField
class TaxonomyField(ser.Field):
def to_representation(self, obj):
if obj is not None:
return {'id': obj._id,
'text': obj.text}
return None
def to_internal_value(self, data):
return data
class TaxonomySerializer(JSONAPISerializer):
filterable_fields = frozenset([
'text',
'parents',
'id'
])
id = ser.CharField(source='_id', required=True)
text = ser.CharField(max_length=200)
parents = JSONAPIListField(child=TaxonomyField())
+ child_count = ser.IntegerField()
links = LinksField({
'parents': 'get_parent_urls',
'self': 'get_absolute_url',
})
def get_parent_urls(self, obj):
return [p.get_absolute_url() for p in obj.parents]
def get_absolute_url(self, obj):
return obj.get_absolute_url()
class Meta:
type_ = 'taxonomies'
| Add child_count taken from new Subject property | ## Code Before:
from rest_framework import serializers as ser
from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField
class TaxonomyField(ser.Field):
def to_representation(self, obj):
if obj is not None:
return {'id': obj._id,
'text': obj.text}
return None
def to_internal_value(self, data):
return data
class TaxonomySerializer(JSONAPISerializer):
filterable_fields = frozenset([
'text',
'parents',
'id'
])
id = ser.CharField(source='_id', required=True)
text = ser.CharField(max_length=200)
parents = JSONAPIListField(child=TaxonomyField())
links = LinksField({
'parents': 'get_parent_urls',
'self': 'get_absolute_url',
})
def get_parent_urls(self, obj):
return [p.get_absolute_url() for p in obj.parents]
def get_absolute_url(self, obj):
return obj.get_absolute_url()
class Meta:
type_ = 'taxonomies'
## Instruction:
Add child_count taken from new Subject property
## Code After:
from rest_framework import serializers as ser
from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField
class TaxonomyField(ser.Field):
def to_representation(self, obj):
if obj is not None:
return {'id': obj._id,
'text': obj.text}
return None
def to_internal_value(self, data):
return data
class TaxonomySerializer(JSONAPISerializer):
filterable_fields = frozenset([
'text',
'parents',
'id'
])
id = ser.CharField(source='_id', required=True)
text = ser.CharField(max_length=200)
parents = JSONAPIListField(child=TaxonomyField())
child_count = ser.IntegerField()
links = LinksField({
'parents': 'get_parent_urls',
'self': 'get_absolute_url',
})
def get_parent_urls(self, obj):
return [p.get_absolute_url() for p in obj.parents]
def get_absolute_url(self, obj):
return obj.get_absolute_url()
class Meta:
type_ = 'taxonomies'
| from rest_framework import serializers as ser
from api.base.serializers import JSONAPISerializer, LinksField, JSONAPIListField
class TaxonomyField(ser.Field):
def to_representation(self, obj):
if obj is not None:
return {'id': obj._id,
'text': obj.text}
return None
def to_internal_value(self, data):
return data
class TaxonomySerializer(JSONAPISerializer):
filterable_fields = frozenset([
'text',
'parents',
'id'
])
id = ser.CharField(source='_id', required=True)
text = ser.CharField(max_length=200)
parents = JSONAPIListField(child=TaxonomyField())
+ child_count = ser.IntegerField()
links = LinksField({
'parents': 'get_parent_urls',
'self': 'get_absolute_url',
})
def get_parent_urls(self, obj):
return [p.get_absolute_url() for p in obj.parents]
def get_absolute_url(self, obj):
return obj.get_absolute_url()
class Meta:
type_ = 'taxonomies' |
a222d268ec1c12466db48bbfcd58d8ecf2907805 | echo_server.py | echo_server.py | import socket
class EchoServer(object):
"""a simple EchoServer"""
def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5):
self.ip = ip
self.port = port
self.backlog = backlog
self.socket = socket.socket(
socket.AF_INET,
socket.SOCK_STREAM,
socket.IPPROTO_IP)
self.socket.bind((self.ip, self.port))
self.socket.listen(self.backlog)
def start_listening(self):
while True:
self.connection, self.addr = self.socket.accept()
words = self.connection.recv(32)
if words:
self.connection.sendall(unicode(words))
self.connection.close()
self.socket.close()
break
if __name__ == "__main__":
server = EchoServer()
server.start_listening()
| import socket
class EchoServer(object):
"""a simple EchoServer"""
def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5):
self.ip = ip
self.port = port
self.backlog = backlog
self.socket = socket.socket(
socket.AF_INET,
socket.SOCK_STREAM,
socket.IPPROTO_IP)
self.socket.bind((self.ip, self.port))
self.socket.listen(self.backlog)
def start_listening(self):
while True:
request = []
self.connection, self.addr = self.socket.accept()
while True:
buffer_ = self.connection.recv(32)
if buffer_:
request.append(buffer_)
else:
break
self.connection.sendall(" ".join(request))
self.connection.close()
if __name__ == "__main__":
server = EchoServer()
server.start_listening()
| Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests | Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests
| Python | mit | jefrailey/network_tools | import socket
class EchoServer(object):
"""a simple EchoServer"""
def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5):
self.ip = ip
self.port = port
self.backlog = backlog
self.socket = socket.socket(
socket.AF_INET,
socket.SOCK_STREAM,
socket.IPPROTO_IP)
self.socket.bind((self.ip, self.port))
self.socket.listen(self.backlog)
def start_listening(self):
while True:
+ request = []
self.connection, self.addr = self.socket.accept()
+
+ while True:
- words = self.connection.recv(32)
+ buffer_ = self.connection.recv(32)
- if words:
+ if buffer_:
+ request.append(buffer_)
+ else:
+ break
- self.connection.sendall(unicode(words))
+ self.connection.sendall(" ".join(request))
self.connection.close()
+
- self.socket.close()
- break
if __name__ == "__main__":
server = EchoServer()
server.start_listening()
| Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests | ## Code Before:
import socket
class EchoServer(object):
"""a simple EchoServer"""
def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5):
self.ip = ip
self.port = port
self.backlog = backlog
self.socket = socket.socket(
socket.AF_INET,
socket.SOCK_STREAM,
socket.IPPROTO_IP)
self.socket.bind((self.ip, self.port))
self.socket.listen(self.backlog)
def start_listening(self):
while True:
self.connection, self.addr = self.socket.accept()
words = self.connection.recv(32)
if words:
self.connection.sendall(unicode(words))
self.connection.close()
self.socket.close()
break
if __name__ == "__main__":
server = EchoServer()
server.start_listening()
## Instruction:
Update EchoServer to keep connection open until client shutsdown connection in order to collect all requests
## Code After:
import socket
class EchoServer(object):
"""a simple EchoServer"""
def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5):
self.ip = ip
self.port = port
self.backlog = backlog
self.socket = socket.socket(
socket.AF_INET,
socket.SOCK_STREAM,
socket.IPPROTO_IP)
self.socket.bind((self.ip, self.port))
self.socket.listen(self.backlog)
def start_listening(self):
while True:
request = []
self.connection, self.addr = self.socket.accept()
while True:
buffer_ = self.connection.recv(32)
if buffer_:
request.append(buffer_)
else:
break
self.connection.sendall(" ".join(request))
self.connection.close()
if __name__ == "__main__":
server = EchoServer()
server.start_listening()
| import socket
class EchoServer(object):
"""a simple EchoServer"""
def __init__(self, ip=u'127.0.0.1', port=50000, backlog=5):
self.ip = ip
self.port = port
self.backlog = backlog
self.socket = socket.socket(
socket.AF_INET,
socket.SOCK_STREAM,
socket.IPPROTO_IP)
self.socket.bind((self.ip, self.port))
self.socket.listen(self.backlog)
def start_listening(self):
while True:
+ request = []
self.connection, self.addr = self.socket.accept()
+
+ while True:
- words = self.connection.recv(32)
? ^^ ^^
+ buffer_ = self.connection.recv(32)
? ^^^^^^^^^ ^
- if words:
+ if buffer_:
+ request.append(buffer_)
+ else:
+ break
- self.connection.sendall(unicode(words))
? ---- ----- -----
+ self.connection.sendall(" ".join(request))
? ++++++++++++ +
self.connection.close()
+
- self.socket.close()
- break
if __name__ == "__main__":
server = EchoServer()
server.start_listening() |
c7a424cb3fb0a037cda04c30d44606515aed829d | chrome/test/functional/test_pyauto.py | chrome/test/functional/test_pyauto.py |
import unittest
import pyauto_functional # Must be imported before pyauto
import pyauto
class PyAutoTest(pyauto.PyUITest):
"""Test functionality of the PyAuto framework."""
_EXTRA_CHROME_FLAGS = [
'--scooby-doo=123',
'--donald-duck=cool',
'--super-mario',
'--marvin-the-martian',
]
def ExtraChromeFlags(self):
"""Ensures Chrome is launched with some custom flags.
Overrides the default list of extra flags passed to Chrome. See
ExtraChromeFlags() in pyauto.py.
"""
return self._EXTRA_CHROME_FLAGS
def testSetCustomChromeFlags(self):
"""Ensures that Chrome can be launched with custom flags."""
self.NavigateToURL('about://version')
for flag in self._EXTRA_CHROME_FLAGS:
self.assertEqual(self.FindInPage(flag)['match_count'], 1,
msg='Missing expected Chrome flag "%s"' % flag)
if __name__ == '__main__':
pyauto_functional.Main()
|
import unittest
import pyauto_functional # Must be imported before pyauto
import pyauto
class PyAutoTest(pyauto.PyUITest):
"""Test functionality of the PyAuto framework."""
_EXTRA_CHROME_FLAGS = [
'--scooby-doo=123',
'--donald-duck=cool',
'--super-mario',
'--marvin-the-martian',
]
def ExtraChromeFlags(self):
"""Ensures Chrome is launched with some custom flags.
Overrides the default list of extra flags passed to Chrome. See
ExtraChromeFlags() in pyauto.py.
"""
return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS
def testSetCustomChromeFlags(self):
"""Ensures that Chrome can be launched with custom flags."""
self.NavigateToURL('about://version')
for flag in self._EXTRA_CHROME_FLAGS:
self.assertEqual(self.FindInPage(flag)['match_count'], 1,
msg='Missing expected Chrome flag "%s"' % flag)
if __name__ == '__main__':
pyauto_functional.Main()
| Update testSetCustomChromeFlags to only append to chrome flags, not override | Update testSetCustomChromeFlags to only append to chrome flags, not override
This is in keeping with the spirit of the ExtraChromeFlags() method. Whenever
it's overridden, it should only ever append to the list of chrome flags,
never completely override it.
BUG=None
TEST=None
R=dennisjeffrey@chromium.org
Review URL: https://chromiumcodereview.appspot.com/10041001
git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@131597 0039d316-1c4b-4281-b951-d872f2087c98
| Python | bsd-3-clause | Jonekee/chromium.src,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,patrickm/chromium.src,littlstar/chromium.src,hujiajie/pa-chromium,jaruba/chromium.src,hgl888/chromium-crosswalk,hujiajie/pa-chromium,junmin-zhu/chromium-rivertrail,PeterWangIntel/chromium-crosswalk,keishi/chromium,jaruba/chromium.src,junmin-zhu/chromium-rivertrail,bright-sparks/chromium-spacewalk,jaruba/chromium.src,crosswalk-project/chromium-crosswalk-efl,littlstar/chromium.src,jaruba/chromium.src,anirudhSK/chromium,crosswalk-project/chromium-crosswalk-efl,jaruba/chromium.src,robclark/chromium,ChromiumWebApps/chromium,TheTypoMaster/chromium-crosswalk,timopulkkinen/BubbleFish,markYoungH/chromium.src,hgl888/chromium-crosswalk,ondra-novak/chromium.src,chuan9/chromium-crosswalk,robclark/chromium,hujiajie/pa-chromium,hgl888/chromium-crosswalk,markYoungH/chromium.src,chuan9/chromium-crosswalk,Jonekee/chromium.src,ltilve/chromium,hgl888/chromium-crosswalk-efl,ChromiumWebApps/chromium,PeterWangIntel/chromium-crosswalk,hgl888/chromium-crosswalk,keishi/chromium,Chilledheart/chromium,jaruba/chromium.src,zcbenz/cefode-chromium,axinging/chromium-crosswalk,fujunwei/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,pozdnyakov/chromium-crosswalk,hujiajie/pa-chromium,nacl-webkit/chrome_deps,hgl888/chromium-crosswalk-efl,PeterWangIntel/chromium-crosswalk,dushu1203/chromium.src,M4sse/chromium.src,PeterWangIntel/chromium-crosswalk,patrickm/chromium.src,mohamed--abdel-maksoud/chromium.src,krieger-od/nwjs_chromium.src,mohamed--abdel-maksoud/chromium.src,chuan9/chromium-crosswalk,junmin-zhu/chromium-rivertrail,PeterWangIntel/chromium-crosswalk,krieger-od/nwjs_chromium.src,anirudhSK/chromium,mohamed--abdel-maksoud/chromium.src,chuan9/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,axinging/chromium-crosswalk,axinging/chromium-crosswalk,timopulkkinen/BubbleFish,mohamed--abdel-maksoud/chromium.src,Fireblend/chromium-crosswalk,hujiajie/pa-chromium,anirudhSK/chromium,markYoungH/chromium.src,mogoweb/chromium-crosswalk,Chilledheart/chromium,nacl-webkit/chrome_deps,hgl888/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,Just-D/chromium-1,ChromiumWebApps/chromium,hujiajie/pa-chromium,axinging/chromium-crosswalk,ChromiumWebApps/chromium,markYoungH/chromium.src,ondra-novak/chromium.src,keishi/chromium,ondra-novak/chromium.src,pozdnyakov/chromium-crosswalk,bright-sparks/chromium-spacewalk,Pluto-tv/chromium-crosswalk,M4sse/chromium.src,timopulkkinen/BubbleFish,timopulkkinen/BubbleFish,Jonekee/chromium.src,Just-D/chromium-1,dushu1203/chromium.src,krieger-od/nwjs_chromium.src,keishi/chromium,junmin-zhu/chromium-rivertrail,hgl888/chromium-crosswalk-efl,Chilledheart/chromium,M4sse/chromium.src,chuan9/chromium-crosswalk,dednal/chromium.src,ChromiumWebApps/chromium,mohamed--abdel-maksoud/chromium.src,nacl-webkit/chrome_deps,ondra-novak/chromium.src,Pluto-tv/chromium-crosswalk,Just-D/chromium-1,PeterWangIntel/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,ondra-novak/chromium.src,markYoungH/chromium.src,littlstar/chromium.src,robclark/chromium,ltilve/chromium,ondra-novak/chromium.src,markYoungH/chromium.src,Chilledheart/chromium,robclark/chromium,hgl888/chromium-crosswalk-efl,Fireblend/chromium-crosswalk,axinging/chromium-crosswalk,mogoweb/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Jonekee/chromium.src,M4sse/chromium.src,pozdnyakov/chromium-crosswalk,hgl888/chromium-crosswalk,bright-sparks/chromium-spacewalk,dednal/chromium.src,zcbenz/cefode-chromium,Chilledheart/chromium,crosswalk-project/chromium-crosswalk-efl,mogoweb/chromium-crosswalk,Chilledheart/chromium,nacl-webkit/chrome_deps,dushu1203/chromium.src,zcbenz/cefode-chromium,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,junmin-zhu/chromium-rivertrail,Fireblend/chromium-crosswalk,patrickm/chromium.src,fujunwei/chromium-crosswalk,timopulkkinen/BubbleFish,PeterWangIntel/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,nacl-webkit/chrome_deps,M4sse/chromium.src,zcbenz/cefode-chromium,ondra-novak/chromium.src,dednal/chromium.src,bright-sparks/chromium-spacewalk,axinging/chromium-crosswalk,keishi/chromium,Fireblend/chromium-crosswalk,junmin-zhu/chromium-rivertrail,junmin-zhu/chromium-rivertrail,dushu1203/chromium.src,keishi/chromium,robclark/chromium,fujunwei/chromium-crosswalk,markYoungH/chromium.src,hgl888/chromium-crosswalk,ltilve/chromium,patrickm/chromium.src,Just-D/chromium-1,mogoweb/chromium-crosswalk,hujiajie/pa-chromium,PeterWangIntel/chromium-crosswalk,hujiajie/pa-chromium,chuan9/chromium-crosswalk,dednal/chromium.src,Chilledheart/chromium,crosswalk-project/chromium-crosswalk-efl,anirudhSK/chromium,mohamed--abdel-maksoud/chromium.src,junmin-zhu/chromium-rivertrail,littlstar/chromium.src,anirudhSK/chromium,zcbenz/cefode-chromium,krieger-od/nwjs_chromium.src,keishi/chromium,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,M4sse/chromium.src,dushu1203/chromium.src,ChromiumWebApps/chromium,jaruba/chromium.src,Pluto-tv/chromium-crosswalk,nacl-webkit/chrome_deps,fujunwei/chromium-crosswalk,zcbenz/cefode-chromium,M4sse/chromium.src,krieger-od/nwjs_chromium.src,chuan9/chromium-crosswalk,hgl888/chromium-crosswalk-efl,nacl-webkit/chrome_deps,patrickm/chromium.src,axinging/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,patrickm/chromium.src,Jonekee/chromium.src,anirudhSK/chromium,hgl888/chromium-crosswalk-efl,pozdnyakov/chromium-crosswalk,Fireblend/chromium-crosswalk,axinging/chromium-crosswalk,dushu1203/chromium.src,patrickm/chromium.src,Fireblend/chromium-crosswalk,bright-sparks/chromium-spacewalk,ChromiumWebApps/chromium,keishi/chromium,TheTypoMaster/chromium-crosswalk,jaruba/chromium.src,anirudhSK/chromium,hgl888/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,anirudhSK/chromium,Chilledheart/chromium,hujiajie/pa-chromium,keishi/chromium,robclark/chromium,robclark/chromium,Fireblend/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,keishi/chromium,robclark/chromium,hgl888/chromium-crosswalk-efl,Pluto-tv/chromium-crosswalk,dednal/chromium.src,mogoweb/chromium-crosswalk,hgl888/chromium-crosswalk-efl,fujunwei/chromium-crosswalk,Fireblend/chromium-crosswalk,ltilve/chromium,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,littlstar/chromium.src,Jonekee/chromium.src,Just-D/chromium-1,M4sse/chromium.src,Pluto-tv/chromium-crosswalk,mogoweb/chromium-crosswalk,nacl-webkit/chrome_deps,ChromiumWebApps/chromium,keishi/chromium,mogoweb/chromium-crosswalk,Jonekee/chromium.src,fujunwei/chromium-crosswalk,dushu1203/chromium.src,anirudhSK/chromium,zcbenz/cefode-chromium,mohamed--abdel-maksoud/chromium.src,littlstar/chromium.src,ltilve/chromium,dednal/chromium.src,axinging/chromium-crosswalk,jaruba/chromium.src,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,mogoweb/chromium-crosswalk,chuan9/chromium-crosswalk,pozdnyakov/chromium-crosswalk,Pluto-tv/chromium-crosswalk,ChromiumWebApps/chromium,bright-sparks/chromium-spacewalk,Jonekee/chromium.src,jaruba/chromium.src,Just-D/chromium-1,littlstar/chromium.src,mogoweb/chromium-crosswalk,markYoungH/chromium.src,nacl-webkit/chrome_deps,pozdnyakov/chromium-crosswalk,ChromiumWebApps/chromium,Just-D/chromium-1,mogoweb/chromium-crosswalk,dushu1203/chromium.src,Jonekee/chromium.src,bright-sparks/chromium-spacewalk,junmin-zhu/chromium-rivertrail,TheTypoMaster/chromium-crosswalk,dushu1203/chromium.src,ChromiumWebApps/chromium,junmin-zhu/chromium-rivertrail,dushu1203/chromium.src,zcbenz/cefode-chromium,hgl888/chromium-crosswalk-efl,littlstar/chromium.src,crosswalk-project/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,dednal/chromium.src,nacl-webkit/chrome_deps,anirudhSK/chromium,markYoungH/chromium.src,M4sse/chromium.src,TheTypoMaster/chromium-crosswalk,krieger-od/nwjs_chromium.src,M4sse/chromium.src,chuan9/chromium-crosswalk,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,zcbenz/cefode-chromium,Just-D/chromium-1,ltilve/chromium,ondra-novak/chromium.src,dednal/chromium.src,ltilve/chromium,M4sse/chromium.src,ondra-novak/chromium.src,jaruba/chromium.src,fujunwei/chromium-crosswalk,Jonekee/chromium.src,TheTypoMaster/chromium-crosswalk,markYoungH/chromium.src,Fireblend/chromium-crosswalk,ltilve/chromium,robclark/chromium,patrickm/chromium.src,dednal/chromium.src,anirudhSK/chromium,Jonekee/chromium.src,pozdnyakov/chromium-crosswalk,Chilledheart/chromium,ChromiumWebApps/chromium,zcbenz/cefode-chromium,timopulkkinen/BubbleFish,axinging/chromium-crosswalk,krieger-od/nwjs_chromium.src,anirudhSK/chromium,Pluto-tv/chromium-crosswalk,nacl-webkit/chrome_deps,ltilve/chromium,crosswalk-project/chromium-crosswalk-efl,fujunwei/chromium-crosswalk,timopulkkinen/BubbleFish,axinging/chromium-crosswalk,dednal/chromium.src,hujiajie/pa-chromium,timopulkkinen/BubbleFish,pozdnyakov/chromium-crosswalk,Pluto-tv/chromium-crosswalk,junmin-zhu/chromium-rivertrail,krieger-od/nwjs_chromium.src,zcbenz/cefode-chromium,robclark/chromium,markYoungH/chromium.src,hujiajie/pa-chromium,dushu1203/chromium.src,fujunwei/chromium-crosswalk,Pluto-tv/chromium-crosswalk,bright-sparks/chromium-spacewalk,patrickm/chromium.src,dednal/chromium.src,Just-D/chromium-1,bright-sparks/chromium-spacewalk |
import unittest
import pyauto_functional # Must be imported before pyauto
import pyauto
class PyAutoTest(pyauto.PyUITest):
"""Test functionality of the PyAuto framework."""
_EXTRA_CHROME_FLAGS = [
'--scooby-doo=123',
'--donald-duck=cool',
'--super-mario',
'--marvin-the-martian',
]
def ExtraChromeFlags(self):
"""Ensures Chrome is launched with some custom flags.
Overrides the default list of extra flags passed to Chrome. See
ExtraChromeFlags() in pyauto.py.
"""
- return self._EXTRA_CHROME_FLAGS
+ return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS
def testSetCustomChromeFlags(self):
"""Ensures that Chrome can be launched with custom flags."""
self.NavigateToURL('about://version')
for flag in self._EXTRA_CHROME_FLAGS:
self.assertEqual(self.FindInPage(flag)['match_count'], 1,
msg='Missing expected Chrome flag "%s"' % flag)
if __name__ == '__main__':
pyauto_functional.Main()
| Update testSetCustomChromeFlags to only append to chrome flags, not override | ## Code Before:
import unittest
import pyauto_functional # Must be imported before pyauto
import pyauto
class PyAutoTest(pyauto.PyUITest):
"""Test functionality of the PyAuto framework."""
_EXTRA_CHROME_FLAGS = [
'--scooby-doo=123',
'--donald-duck=cool',
'--super-mario',
'--marvin-the-martian',
]
def ExtraChromeFlags(self):
"""Ensures Chrome is launched with some custom flags.
Overrides the default list of extra flags passed to Chrome. See
ExtraChromeFlags() in pyauto.py.
"""
return self._EXTRA_CHROME_FLAGS
def testSetCustomChromeFlags(self):
"""Ensures that Chrome can be launched with custom flags."""
self.NavigateToURL('about://version')
for flag in self._EXTRA_CHROME_FLAGS:
self.assertEqual(self.FindInPage(flag)['match_count'], 1,
msg='Missing expected Chrome flag "%s"' % flag)
if __name__ == '__main__':
pyauto_functional.Main()
## Instruction:
Update testSetCustomChromeFlags to only append to chrome flags, not override
## Code After:
import unittest
import pyauto_functional # Must be imported before pyauto
import pyauto
class PyAutoTest(pyauto.PyUITest):
"""Test functionality of the PyAuto framework."""
_EXTRA_CHROME_FLAGS = [
'--scooby-doo=123',
'--donald-duck=cool',
'--super-mario',
'--marvin-the-martian',
]
def ExtraChromeFlags(self):
"""Ensures Chrome is launched with some custom flags.
Overrides the default list of extra flags passed to Chrome. See
ExtraChromeFlags() in pyauto.py.
"""
return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS
def testSetCustomChromeFlags(self):
"""Ensures that Chrome can be launched with custom flags."""
self.NavigateToURL('about://version')
for flag in self._EXTRA_CHROME_FLAGS:
self.assertEqual(self.FindInPage(flag)['match_count'], 1,
msg='Missing expected Chrome flag "%s"' % flag)
if __name__ == '__main__':
pyauto_functional.Main()
|
import unittest
import pyauto_functional # Must be imported before pyauto
import pyauto
class PyAutoTest(pyauto.PyUITest):
"""Test functionality of the PyAuto framework."""
_EXTRA_CHROME_FLAGS = [
'--scooby-doo=123',
'--donald-duck=cool',
'--super-mario',
'--marvin-the-martian',
]
def ExtraChromeFlags(self):
"""Ensures Chrome is launched with some custom flags.
Overrides the default list of extra flags passed to Chrome. See
ExtraChromeFlags() in pyauto.py.
"""
- return self._EXTRA_CHROME_FLAGS
+ return pyauto.PyUITest.ExtraChromeFlags(self) + self._EXTRA_CHROME_FLAGS
def testSetCustomChromeFlags(self):
"""Ensures that Chrome can be launched with custom flags."""
self.NavigateToURL('about://version')
for flag in self._EXTRA_CHROME_FLAGS:
self.assertEqual(self.FindInPage(flag)['match_count'], 1,
msg='Missing expected Chrome flag "%s"' % flag)
if __name__ == '__main__':
pyauto_functional.Main() |
f2139cad673ee50f027164bda80d86979d5ce7a0 | passenger_wsgi.py | passenger_wsgi.py | import os
import sys
try:
from flask import Flask, render_template, send_file, Response
import requests
except ImportError:
INTERP = "venv/bin/python"
if os.path.relpath(sys.executable, os.getcwd()) != INTERP:
try:
os.execl(INTERP, INTERP, *sys.argv)
except OSError:
sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`")
else:
sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`")
application = Flask(__name__)
@application.route("/")
def index():
return "Hello, world!"
| import os
import sys
try:
from flask import Flask
import flask_login
from flask_restless import APIManager
from flask_sqlalchemy import SQLAlchemy
import requests
except ImportError:
INTERP = "venv/bin/python"
if os.path.relpath(sys.executable, os.getcwd()) != INTERP:
try:
os.execl(INTERP, INTERP, *sys.argv)
except OSError:
sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`")
else:
sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`")
application = Flask(__name__)
@application.route("/")
def index():
return "Hello, world!"
| Add more imports for further functionality | Add more imports for further functionality
`flask_login`, `flask_restless`, `flask_sqlalchemy` | Python | mit | GregBrimble/boilerplate-web-service,GregBrimble/boilerplate-web-service | import os
import sys
try:
- from flask import Flask, render_template, send_file, Response
+ from flask import Flask
+ import flask_login
+ from flask_restless import APIManager
+ from flask_sqlalchemy import SQLAlchemy
import requests
except ImportError:
INTERP = "venv/bin/python"
if os.path.relpath(sys.executable, os.getcwd()) != INTERP:
try:
os.execl(INTERP, INTERP, *sys.argv)
except OSError:
sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`")
else:
sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`")
application = Flask(__name__)
@application.route("/")
def index():
return "Hello, world!"
| Add more imports for further functionality | ## Code Before:
import os
import sys
try:
from flask import Flask, render_template, send_file, Response
import requests
except ImportError:
INTERP = "venv/bin/python"
if os.path.relpath(sys.executable, os.getcwd()) != INTERP:
try:
os.execl(INTERP, INTERP, *sys.argv)
except OSError:
sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`")
else:
sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`")
application = Flask(__name__)
@application.route("/")
def index():
return "Hello, world!"
## Instruction:
Add more imports for further functionality
## Code After:
import os
import sys
try:
from flask import Flask
import flask_login
from flask_restless import APIManager
from flask_sqlalchemy import SQLAlchemy
import requests
except ImportError:
INTERP = "venv/bin/python"
if os.path.relpath(sys.executable, os.getcwd()) != INTERP:
try:
os.execl(INTERP, INTERP, *sys.argv)
except OSError:
sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`")
else:
sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`")
application = Flask(__name__)
@application.route("/")
def index():
return "Hello, world!"
| import os
import sys
try:
- from flask import Flask, render_template, send_file, Response
+ from flask import Flask
+ import flask_login
+ from flask_restless import APIManager
+ from flask_sqlalchemy import SQLAlchemy
import requests
except ImportError:
INTERP = "venv/bin/python"
if os.path.relpath(sys.executable, os.getcwd()) != INTERP:
try:
os.execl(INTERP, INTERP, *sys.argv)
except OSError:
sys.exit("Could not find virtual environment. Run `:~$ ./setup.sh`")
else:
sys.exit("Could not find requirements. Are they all included in requirements.txt? Run `:~$ ./setup.sh`")
application = Flask(__name__)
@application.route("/")
def index():
return "Hello, world!" |
b16c49cfd6a0ee659e4493ef959e0483e93d350a | os_client_config/defaults.py | os_client_config/defaults.py |
_defaults = dict(
auth_type='password',
compute_api_version='2',
floating_ip_source='neutron',
identity_api_version='2',
image_api_use_tasks=False,
image_api_version='1',
network_api_version='2',
object_api_version='1',
volume_api_version='1',
)
|
_defaults = dict(
auth_type='password',
baremetal_api_version='1',
compute_api_version='2',
database_api_version='1.0',
floating_ip_source='neutron',
identity_api_version='2',
image_api_use_tasks=False,
image_api_version='1',
network_api_version='2',
object_api_version='1',
volume_api_version='1',
)
| Add default versions for trove and ironic | Add default versions for trove and ironic
Change-Id: Ib7af38664cfbe75c78c70693117f1193c4beb7e6
| Python | apache-2.0 | openstack/python-openstacksdk,stackforge/python-openstacksdk,redhat-openstack/os-client-config,dtroyer/python-openstacksdk,openstack/os-client-config,dtroyer/python-openstacksdk,switch-ch/os-client-config,stackforge/python-openstacksdk,dtroyer/os-client-config,openstack/python-openstacksdk |
_defaults = dict(
auth_type='password',
+ baremetal_api_version='1',
compute_api_version='2',
+ database_api_version='1.0',
floating_ip_source='neutron',
identity_api_version='2',
image_api_use_tasks=False,
image_api_version='1',
network_api_version='2',
object_api_version='1',
volume_api_version='1',
)
| Add default versions for trove and ironic | ## Code Before:
_defaults = dict(
auth_type='password',
compute_api_version='2',
floating_ip_source='neutron',
identity_api_version='2',
image_api_use_tasks=False,
image_api_version='1',
network_api_version='2',
object_api_version='1',
volume_api_version='1',
)
## Instruction:
Add default versions for trove and ironic
## Code After:
_defaults = dict(
auth_type='password',
baremetal_api_version='1',
compute_api_version='2',
database_api_version='1.0',
floating_ip_source='neutron',
identity_api_version='2',
image_api_use_tasks=False,
image_api_version='1',
network_api_version='2',
object_api_version='1',
volume_api_version='1',
)
|
_defaults = dict(
auth_type='password',
+ baremetal_api_version='1',
compute_api_version='2',
+ database_api_version='1.0',
floating_ip_source='neutron',
identity_api_version='2',
image_api_use_tasks=False,
image_api_version='1',
network_api_version='2',
object_api_version='1',
volume_api_version='1',
) |
216f0bb3680b86ac2dfc8c506b791db4e34eeee6 | nextactions/board.py | nextactions/board.py | from nextactions.list import List
class Board:
def __init__(self, trello, json):
self._trello = trello
self.id = json['id']
self.name = json['name']
self.nextActionList = []
def getLists(self):
json = self._trello.get(
'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)'
)
return [List(j) for j in json]
def getListByName(self, name):
for l in self.getLists():
if l.name == name:
return l
return None
| from nextactions.list import List
class Board:
def __init__(self, trello, json):
self._trello = trello
self.id = json['id']
self.name = json['name']
self.nextActionList = []
def getLists(self):
json = self._trello.get(
'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)'
)
return [List(j) for j in json]
def getListByName(self, name):
matches = [l for l in self.getLists() if l.name == name]
return matches[0] if len(matches) else None
| Tidy matching lists by name | Tidy matching lists by name
| Python | mit | stevecshanks/trello-next-actions | from nextactions.list import List
class Board:
def __init__(self, trello, json):
self._trello = trello
self.id = json['id']
self.name = json['name']
self.nextActionList = []
def getLists(self):
json = self._trello.get(
'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)'
)
return [List(j) for j in json]
def getListByName(self, name):
+ matches = [l for l in self.getLists() if l.name == name]
+ return matches[0] if len(matches) else None
- for l in self.getLists():
- if l.name == name:
- return l
- return None
| Tidy matching lists by name | ## Code Before:
from nextactions.list import List
class Board:
def __init__(self, trello, json):
self._trello = trello
self.id = json['id']
self.name = json['name']
self.nextActionList = []
def getLists(self):
json = self._trello.get(
'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)'
)
return [List(j) for j in json]
def getListByName(self, name):
for l in self.getLists():
if l.name == name:
return l
return None
## Instruction:
Tidy matching lists by name
## Code After:
from nextactions.list import List
class Board:
def __init__(self, trello, json):
self._trello = trello
self.id = json['id']
self.name = json['name']
self.nextActionList = []
def getLists(self):
json = self._trello.get(
'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)'
)
return [List(j) for j in json]
def getListByName(self, name):
matches = [l for l in self.getLists() if l.name == name]
return matches[0] if len(matches) else None
| from nextactions.list import List
class Board:
def __init__(self, trello, json):
self._trello = trello
self.id = json['id']
self.name = json['name']
self.nextActionList = []
def getLists(self):
json = self._trello.get(
'https://api.trello.com/1/boards/' + self.id + '/lists?cards=none)'
)
return [List(j) for j in json]
def getListByName(self, name):
+ matches = [l for l in self.getLists() if l.name == name]
+ return matches[0] if len(matches) else None
- for l in self.getLists():
- if l.name == name:
- return l
- return None |
2facb0c8794c9529ccb17631a90b0ee181c4eb5b | xml_json_import/__init__.py | xml_json_import/__init__.py | from django.conf import settings
from os import path
class XmlJsonImportModuleException(Exception):
pass
if not hasattr(settings, 'XSLT_FILES_DIR'):
raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter')
if not path.exists(settings.XSLT_FILES_DIR):
raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist')
| from django.conf import settings
from os import path, listdir
from lxml import etree
class XmlJsonImportModuleException(Exception):
pass
if not hasattr(settings, 'XSLT_FILES_DIR'):
raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter')
if not path.exists(settings.XSLT_FILES_DIR):
raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist')
for filename in listdir(settings.XSLT_FILES_DIR):
filepath = path.join(settings.XSLT_FILES_DIR, filename)
if path.isfile(filepath):
try:
xslt_etree = etree.parse(filepath)
except etree.XMLSyntaxError as er:
raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er))
try:
transform = etree.XSLT(xslt_etree)
except etree.XSLTParseError as er:
raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er))
| Add exception handling for invalid XSLT files | Add exception handling for invalid XSLT files
| Python | mit | lev-veshnyakov/django-import-data,lev-veshnyakov/django-import-data | from django.conf import settings
- from os import path
+ from os import path, listdir
+ from lxml import etree
class XmlJsonImportModuleException(Exception):
pass
if not hasattr(settings, 'XSLT_FILES_DIR'):
raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter')
if not path.exists(settings.XSLT_FILES_DIR):
raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist')
+
+ for filename in listdir(settings.XSLT_FILES_DIR):
+ filepath = path.join(settings.XSLT_FILES_DIR, filename)
+ if path.isfile(filepath):
+ try:
+ xslt_etree = etree.parse(filepath)
+ except etree.XMLSyntaxError as er:
+ raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er))
+ try:
+ transform = etree.XSLT(xslt_etree)
+ except etree.XSLTParseError as er:
+ raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er))
| Add exception handling for invalid XSLT files | ## Code Before:
from django.conf import settings
from os import path
class XmlJsonImportModuleException(Exception):
pass
if not hasattr(settings, 'XSLT_FILES_DIR'):
raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter')
if not path.exists(settings.XSLT_FILES_DIR):
raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist')
## Instruction:
Add exception handling for invalid XSLT files
## Code After:
from django.conf import settings
from os import path, listdir
from lxml import etree
class XmlJsonImportModuleException(Exception):
pass
if not hasattr(settings, 'XSLT_FILES_DIR'):
raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter')
if not path.exists(settings.XSLT_FILES_DIR):
raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist')
for filename in listdir(settings.XSLT_FILES_DIR):
filepath = path.join(settings.XSLT_FILES_DIR, filename)
if path.isfile(filepath):
try:
xslt_etree = etree.parse(filepath)
except etree.XMLSyntaxError as er:
raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er))
try:
transform = etree.XSLT(xslt_etree)
except etree.XSLTParseError as er:
raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er))
| from django.conf import settings
- from os import path
+ from os import path, listdir
? +++++++++
+ from lxml import etree
class XmlJsonImportModuleException(Exception):
pass
if not hasattr(settings, 'XSLT_FILES_DIR'):
raise XmlJsonImportModuleException('Settings must contain XSLT_FILES_DIR parameter')
if not path.exists(settings.XSLT_FILES_DIR):
raise XmlJsonImportModuleException('Directory specified by XSLT_FILES_DIR does not exist')
+
+ for filename in listdir(settings.XSLT_FILES_DIR):
+ filepath = path.join(settings.XSLT_FILES_DIR, filename)
+ if path.isfile(filepath):
+ try:
+ xslt_etree = etree.parse(filepath)
+ except etree.XMLSyntaxError as er:
+ raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XML file: ' + str(er))
+ try:
+ transform = etree.XSLT(xslt_etree)
+ except etree.XSLTParseError as er:
+ raise XmlJsonImportModuleException('File ' + filepath + ' is not a valid XSLT file: ' + str(er)) |
849fbdf724528df99f2ac53d389274f7c2631f11 | invitation/admin.py | invitation/admin.py | from django.contrib import admin
from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode
class InvitationKeyAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired')
class InvitationUserAdmin(admin.ModelAdmin):
list_display = ('inviter', 'invitations_remaining')
def invite_user(modeladmin, request, queryset):
for invitation_request in queryset.all():
invitation = InvitationKey.objects.create_invitation(request.user)
invitation.send_to(invitation_request.email)
invitation_request.invited = True
invitation_request.save()
invite_user.short_description = "Invite this user"
class InvitationRequestAdmin(admin.ModelAdmin):
list_display = ('email', 'invited')
actions = [invite_user]
admin.site.register(InvitationKey, InvitationKeyAdmin)
admin.site.register(InvitationUser, InvitationUserAdmin)
admin.site.register(InvitationRequest, InvitationRequestAdmin)
admin.site.register(InvitationCode)
| from django.contrib import admin
from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode
class InvitationKeyAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired')
class InvitationUserAdmin(admin.ModelAdmin):
list_display = ('inviter', 'invitations_remaining')
def invite_user(modeladmin, request, queryset):
for invitation_request in queryset.all():
invitation = InvitationKey.objects.create_invitation(request.user)
invitation.send_to(invitation_request.email)
invitation_request.invited = True
invitation_request.save()
invite_user.short_description = "Invite selected invitation requests"
class InvitationRequestAdmin(admin.ModelAdmin):
list_display = ('email', 'invited')
actions = [invite_user]
admin.site.register(InvitationKey, InvitationKeyAdmin)
admin.site.register(InvitationUser, InvitationUserAdmin)
admin.site.register(InvitationRequest, InvitationRequestAdmin)
admin.site.register(InvitationCode)
| Improve the invite_user action name. | Improve the invite_user action name. | Python | bsd-3-clause | adieu/django-invitation | from django.contrib import admin
from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode
class InvitationKeyAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired')
class InvitationUserAdmin(admin.ModelAdmin):
list_display = ('inviter', 'invitations_remaining')
def invite_user(modeladmin, request, queryset):
for invitation_request in queryset.all():
invitation = InvitationKey.objects.create_invitation(request.user)
invitation.send_to(invitation_request.email)
invitation_request.invited = True
invitation_request.save()
- invite_user.short_description = "Invite this user"
+ invite_user.short_description = "Invite selected invitation requests"
class InvitationRequestAdmin(admin.ModelAdmin):
list_display = ('email', 'invited')
actions = [invite_user]
admin.site.register(InvitationKey, InvitationKeyAdmin)
admin.site.register(InvitationUser, InvitationUserAdmin)
admin.site.register(InvitationRequest, InvitationRequestAdmin)
admin.site.register(InvitationCode)
| Improve the invite_user action name. | ## Code Before:
from django.contrib import admin
from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode
class InvitationKeyAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired')
class InvitationUserAdmin(admin.ModelAdmin):
list_display = ('inviter', 'invitations_remaining')
def invite_user(modeladmin, request, queryset):
for invitation_request in queryset.all():
invitation = InvitationKey.objects.create_invitation(request.user)
invitation.send_to(invitation_request.email)
invitation_request.invited = True
invitation_request.save()
invite_user.short_description = "Invite this user"
class InvitationRequestAdmin(admin.ModelAdmin):
list_display = ('email', 'invited')
actions = [invite_user]
admin.site.register(InvitationKey, InvitationKeyAdmin)
admin.site.register(InvitationUser, InvitationUserAdmin)
admin.site.register(InvitationRequest, InvitationRequestAdmin)
admin.site.register(InvitationCode)
## Instruction:
Improve the invite_user action name.
## Code After:
from django.contrib import admin
from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode
class InvitationKeyAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired')
class InvitationUserAdmin(admin.ModelAdmin):
list_display = ('inviter', 'invitations_remaining')
def invite_user(modeladmin, request, queryset):
for invitation_request in queryset.all():
invitation = InvitationKey.objects.create_invitation(request.user)
invitation.send_to(invitation_request.email)
invitation_request.invited = True
invitation_request.save()
invite_user.short_description = "Invite selected invitation requests"
class InvitationRequestAdmin(admin.ModelAdmin):
list_display = ('email', 'invited')
actions = [invite_user]
admin.site.register(InvitationKey, InvitationKeyAdmin)
admin.site.register(InvitationUser, InvitationUserAdmin)
admin.site.register(InvitationRequest, InvitationRequestAdmin)
admin.site.register(InvitationCode)
| from django.contrib import admin
from invitation.models import InvitationKey, InvitationUser, InvitationRequest, InvitationCode
class InvitationKeyAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'from_user', 'date_invited', 'key_expired')
class InvitationUserAdmin(admin.ModelAdmin):
list_display = ('inviter', 'invitations_remaining')
def invite_user(modeladmin, request, queryset):
for invitation_request in queryset.all():
invitation = InvitationKey.objects.create_invitation(request.user)
invitation.send_to(invitation_request.email)
invitation_request.invited = True
invitation_request.save()
- invite_user.short_description = "Invite this user"
+ invite_user.short_description = "Invite selected invitation requests"
class InvitationRequestAdmin(admin.ModelAdmin):
list_display = ('email', 'invited')
actions = [invite_user]
admin.site.register(InvitationKey, InvitationKeyAdmin)
admin.site.register(InvitationUser, InvitationUserAdmin)
admin.site.register(InvitationRequest, InvitationRequestAdmin)
admin.site.register(InvitationCode) |
2261b3c6cb579ae65c1119db45f291e246f536c2 | examples/main.py | examples/main.py | import asyncio
import sys
from contextlib import suppress
sys.path.append("..")
from asynccmd import Cmd
class Commander(Cmd):
def __init__(self, intro, prompt):
if sys.platform == 'win32':
super().__init__(mode="Run", run_loop=False)
else:
super().__init__(mode="Reader", run_loop=False)
self.intro = intro
self.prompt = prompt
self.loop = None
def do_tasks(self, arg):
"""
Fake command. Type "prodigy {arg}"
:param arg: args occurred from cmd after command
:return:
"""
print(print(asyncio.Task.all_tasks(loop=self.loop)))
def start(self, loop=None):
self.loop = loop
super().cmdloop(loop)
loop = asyncio.ProactorEventLoop()
#loop = asyncio.get_event_loop()
cmd = Commander(intro="This is example", prompt="example> ")
cmd.start(loop)
try:
loop.run_forever()
except KeyboardInterrupt:
loop.stop()
pending = asyncio.Task.all_tasks(loop=loop)
for task in pending:
task.cancel()
with suppress(asyncio.CancelledError):
loop.run_until_complete(task)
| import asyncio
import sys
from contextlib import suppress
sys.path.append("..")
from asynccmd import Cmd
class Commander(Cmd):
def __init__(self, intro, prompt):
if sys.platform == 'win32':
super().__init__(mode="Run", run_loop=False)
else:
super().__init__(mode="Reader", run_loop=False)
self.intro = intro
self.prompt = prompt
self.loop = None
def do_tasks(self, arg):
"""
Fake command. Type "prodigy {arg}"
:param arg: args occurred from cmd after command
:return:
"""
print(print(asyncio.Task.all_tasks(loop=self.loop)))
def start(self, loop=None):
self.loop = loop
super().cmdloop(loop)
if sys.platform == 'win32':
loop = asyncio.ProactorEventLoop()
else:
loop = asyncio.get_event_loop()
cmd = Commander(intro="This is example", prompt="example> ")
cmd.start(loop)
try:
loop.run_forever()
except KeyboardInterrupt:
loop.stop()
pending = asyncio.Task.all_tasks(loop=loop)
for task in pending:
task.cancel()
with suppress(asyncio.CancelledError):
loop.run_until_complete(task)
| FIX example for both Win and NIX | FIX example for both Win and NIX
TODO: tasks wont work
| Python | apache-2.0 | valentinmk/asynccmd | import asyncio
import sys
from contextlib import suppress
sys.path.append("..")
from asynccmd import Cmd
class Commander(Cmd):
def __init__(self, intro, prompt):
if sys.platform == 'win32':
super().__init__(mode="Run", run_loop=False)
else:
super().__init__(mode="Reader", run_loop=False)
self.intro = intro
self.prompt = prompt
self.loop = None
def do_tasks(self, arg):
"""
Fake command. Type "prodigy {arg}"
:param arg: args occurred from cmd after command
:return:
"""
print(print(asyncio.Task.all_tasks(loop=self.loop)))
def start(self, loop=None):
self.loop = loop
super().cmdloop(loop)
+ if sys.platform == 'win32':
- loop = asyncio.ProactorEventLoop()
+ loop = asyncio.ProactorEventLoop()
+ else:
- #loop = asyncio.get_event_loop()
+ loop = asyncio.get_event_loop()
cmd = Commander(intro="This is example", prompt="example> ")
cmd.start(loop)
try:
loop.run_forever()
except KeyboardInterrupt:
loop.stop()
pending = asyncio.Task.all_tasks(loop=loop)
for task in pending:
task.cancel()
with suppress(asyncio.CancelledError):
loop.run_until_complete(task)
| FIX example for both Win and NIX | ## Code Before:
import asyncio
import sys
from contextlib import suppress
sys.path.append("..")
from asynccmd import Cmd
class Commander(Cmd):
def __init__(self, intro, prompt):
if sys.platform == 'win32':
super().__init__(mode="Run", run_loop=False)
else:
super().__init__(mode="Reader", run_loop=False)
self.intro = intro
self.prompt = prompt
self.loop = None
def do_tasks(self, arg):
"""
Fake command. Type "prodigy {arg}"
:param arg: args occurred from cmd after command
:return:
"""
print(print(asyncio.Task.all_tasks(loop=self.loop)))
def start(self, loop=None):
self.loop = loop
super().cmdloop(loop)
loop = asyncio.ProactorEventLoop()
#loop = asyncio.get_event_loop()
cmd = Commander(intro="This is example", prompt="example> ")
cmd.start(loop)
try:
loop.run_forever()
except KeyboardInterrupt:
loop.stop()
pending = asyncio.Task.all_tasks(loop=loop)
for task in pending:
task.cancel()
with suppress(asyncio.CancelledError):
loop.run_until_complete(task)
## Instruction:
FIX example for both Win and NIX
## Code After:
import asyncio
import sys
from contextlib import suppress
sys.path.append("..")
from asynccmd import Cmd
class Commander(Cmd):
def __init__(self, intro, prompt):
if sys.platform == 'win32':
super().__init__(mode="Run", run_loop=False)
else:
super().__init__(mode="Reader", run_loop=False)
self.intro = intro
self.prompt = prompt
self.loop = None
def do_tasks(self, arg):
"""
Fake command. Type "prodigy {arg}"
:param arg: args occurred from cmd after command
:return:
"""
print(print(asyncio.Task.all_tasks(loop=self.loop)))
def start(self, loop=None):
self.loop = loop
super().cmdloop(loop)
if sys.platform == 'win32':
loop = asyncio.ProactorEventLoop()
else:
loop = asyncio.get_event_loop()
cmd = Commander(intro="This is example", prompt="example> ")
cmd.start(loop)
try:
loop.run_forever()
except KeyboardInterrupt:
loop.stop()
pending = asyncio.Task.all_tasks(loop=loop)
for task in pending:
task.cancel()
with suppress(asyncio.CancelledError):
loop.run_until_complete(task)
| import asyncio
import sys
from contextlib import suppress
sys.path.append("..")
from asynccmd import Cmd
class Commander(Cmd):
def __init__(self, intro, prompt):
if sys.platform == 'win32':
super().__init__(mode="Run", run_loop=False)
else:
super().__init__(mode="Reader", run_loop=False)
self.intro = intro
self.prompt = prompt
self.loop = None
def do_tasks(self, arg):
"""
Fake command. Type "prodigy {arg}"
:param arg: args occurred from cmd after command
:return:
"""
print(print(asyncio.Task.all_tasks(loop=self.loop)))
def start(self, loop=None):
self.loop = loop
super().cmdloop(loop)
+ if sys.platform == 'win32':
- loop = asyncio.ProactorEventLoop()
+ loop = asyncio.ProactorEventLoop()
? +++
+ else:
- #loop = asyncio.get_event_loop()
? ^
+ loop = asyncio.get_event_loop()
? ^^^
cmd = Commander(intro="This is example", prompt="example> ")
cmd.start(loop)
try:
loop.run_forever()
except KeyboardInterrupt:
loop.stop()
pending = asyncio.Task.all_tasks(loop=loop)
for task in pending:
task.cancel()
with suppress(asyncio.CancelledError):
loop.run_until_complete(task) |
85f8d0662901047115f2d852489a3a5be1a01226 | datafilters/views.py | datafilters/views.py | try:
from django.views.generic.base import ContextMixin as mixin_base
except ImportError:
mixin_base = object
__all__ = ('FilterFormMixin',)
class FilterFormMixin(mixin_base):
"""
Mixin that adds filtering behaviour for Class Based Views.
Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and
get_context_data(self, **kwargs) method.
"""
filter_form_cls = None
use_filter_chaining = False
def get_filter(self):
return self.filter_form_cls(self.request.GET,
runtime_context=self.get_runtime_context(),
use_filter_chaining=self.use_filter_chaining)
def get_queryset(self):
qs = super(FilterFormMixin, self).get_queryset()
qs = self.get_filter().filter(qs).distinct()
return qs
def get_context_data(self, **kwargs):
context = super(FilterFormMixin, self).get_context_data(**kwargs)
context['filterform'] = self.get_filter()
return context
def get_runtime_context(self):
return {'user': self.request.user}
| from django.views.generic.list import MultipleObjectMixin
__all__ = ('FilterFormMixin',)
class FilterFormMixin(MultipleObjectMixin):
"""
Mixin that adds filtering behaviour for Class Based Views.
Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and
get_context_data(self, **kwargs) method.
"""
filter_form_cls = None
use_filter_chaining = False
def get_filter(self):
return self.filter_form_cls(self.request.GET,
runtime_context=self.get_runtime_context(),
use_filter_chaining=self.use_filter_chaining)
def get_queryset(self):
qs = super(FilterFormMixin, self).get_queryset()
qs = self.get_filter().filter(qs).distinct()
return qs
def get_context_data(self, **kwargs):
context = super(FilterFormMixin, self).get_context_data(**kwargs)
context['filterform'] = self.get_filter()
return context
def get_runtime_context(self):
return {'user': self.request.user}
| Set base class for view mixin to MultipleObjectMixin | Set base class for view mixin to MultipleObjectMixin
| Python | mit | freevoid/django-datafilters,zorainc/django-datafilters,zorainc/django-datafilters | + from django.views.generic.list import MultipleObjectMixin
- try:
- from django.views.generic.base import ContextMixin as mixin_base
- except ImportError:
- mixin_base = object
__all__ = ('FilterFormMixin',)
- class FilterFormMixin(mixin_base):
+ class FilterFormMixin(MultipleObjectMixin):
"""
Mixin that adds filtering behaviour for Class Based Views.
Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and
get_context_data(self, **kwargs) method.
"""
filter_form_cls = None
use_filter_chaining = False
def get_filter(self):
return self.filter_form_cls(self.request.GET,
runtime_context=self.get_runtime_context(),
use_filter_chaining=self.use_filter_chaining)
def get_queryset(self):
qs = super(FilterFormMixin, self).get_queryset()
qs = self.get_filter().filter(qs).distinct()
return qs
def get_context_data(self, **kwargs):
context = super(FilterFormMixin, self).get_context_data(**kwargs)
context['filterform'] = self.get_filter()
return context
def get_runtime_context(self):
return {'user': self.request.user}
| Set base class for view mixin to MultipleObjectMixin | ## Code Before:
try:
from django.views.generic.base import ContextMixin as mixin_base
except ImportError:
mixin_base = object
__all__ = ('FilterFormMixin',)
class FilterFormMixin(mixin_base):
"""
Mixin that adds filtering behaviour for Class Based Views.
Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and
get_context_data(self, **kwargs) method.
"""
filter_form_cls = None
use_filter_chaining = False
def get_filter(self):
return self.filter_form_cls(self.request.GET,
runtime_context=self.get_runtime_context(),
use_filter_chaining=self.use_filter_chaining)
def get_queryset(self):
qs = super(FilterFormMixin, self).get_queryset()
qs = self.get_filter().filter(qs).distinct()
return qs
def get_context_data(self, **kwargs):
context = super(FilterFormMixin, self).get_context_data(**kwargs)
context['filterform'] = self.get_filter()
return context
def get_runtime_context(self):
return {'user': self.request.user}
## Instruction:
Set base class for view mixin to MultipleObjectMixin
## Code After:
from django.views.generic.list import MultipleObjectMixin
__all__ = ('FilterFormMixin',)
class FilterFormMixin(MultipleObjectMixin):
"""
Mixin that adds filtering behaviour for Class Based Views.
Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and
get_context_data(self, **kwargs) method.
"""
filter_form_cls = None
use_filter_chaining = False
def get_filter(self):
return self.filter_form_cls(self.request.GET,
runtime_context=self.get_runtime_context(),
use_filter_chaining=self.use_filter_chaining)
def get_queryset(self):
qs = super(FilterFormMixin, self).get_queryset()
qs = self.get_filter().filter(qs).distinct()
return qs
def get_context_data(self, **kwargs):
context = super(FilterFormMixin, self).get_context_data(**kwargs)
context['filterform'] = self.get_filter()
return context
def get_runtime_context(self):
return {'user': self.request.user}
| + from django.views.generic.list import MultipleObjectMixin
- try:
- from django.views.generic.base import ContextMixin as mixin_base
- except ImportError:
- mixin_base = object
__all__ = ('FilterFormMixin',)
- class FilterFormMixin(mixin_base):
+ class FilterFormMixin(MultipleObjectMixin):
"""
Mixin that adds filtering behaviour for Class Based Views.
Changed in a way that can play nicely with other CBV simply by overriding the get_queryset(self) and
get_context_data(self, **kwargs) method.
"""
filter_form_cls = None
use_filter_chaining = False
def get_filter(self):
return self.filter_form_cls(self.request.GET,
runtime_context=self.get_runtime_context(),
use_filter_chaining=self.use_filter_chaining)
def get_queryset(self):
qs = super(FilterFormMixin, self).get_queryset()
qs = self.get_filter().filter(qs).distinct()
return qs
def get_context_data(self, **kwargs):
context = super(FilterFormMixin, self).get_context_data(**kwargs)
context['filterform'] = self.get_filter()
return context
def get_runtime_context(self):
return {'user': self.request.user} |
937a5e32c77ca57917d60a891616fbcf19ab19f9 | respite/utils.py | respite/utils.py | from django import forms
def generate_form(model):
"""
Generate a form from a model.
Arguments:
model -- A Django model.
"""
_model = model
class Form(forms.ModelForm):
class Meta:
model = _model
return Form
def parse_http_accept_header(header):
"""
Return a list of content types listed in the HTTP Accept header
ordered by quality.
Arguments:
header -- A string describing the contents of the HTTP Accept header.
"""
components = header.split(',')
l = []
for component in components:
if ';' in component:
subcomponents = component.split(';')
l.append(
(
subcomponents[0], # eg. 'text/html'
subcomponents[1][2:] # eg. 'q=0.9'
)
)
else:
l.append((component, '1'))
l.sort(
key = lambda i: i[1],
reverse = True
)
content_types = []
for i in l:
content_types.append(i[0])
return content_types
| from django import forms
def generate_form(model):
"""
Generate a form from a model.
Arguments:
model -- A Django model.
"""
_model = model
class Form(forms.ModelForm):
class Meta:
model = _model
return Form
def parse_http_accept_header(header):
"""
Return a list of content types listed in the HTTP Accept header
ordered by quality.
Arguments:
header -- A string describing the contents of the HTTP Accept header.
"""
components = [item.strip() for item in header.split(',')]
l = []
for component in components:
if ';' in component:
subcomponents = [item.strip() for item in component.split(';')]
l.append(
(
subcomponents[0], # eg. 'text/html'
subcomponents[1][2:] # eg. 'q=0.9'
)
)
else:
l.append((component, '1'))
l.sort(
key = lambda i: i[1],
reverse = True
)
content_types = []
for i in l:
content_types.append(i[0])
return content_types
| Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly | Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly
| Python | mit | jgorset/django-respite,jgorset/django-respite,jgorset/django-respite | from django import forms
def generate_form(model):
"""
Generate a form from a model.
Arguments:
model -- A Django model.
"""
_model = model
class Form(forms.ModelForm):
class Meta:
model = _model
return Form
def parse_http_accept_header(header):
"""
Return a list of content types listed in the HTTP Accept header
ordered by quality.
Arguments:
header -- A string describing the contents of the HTTP Accept header.
"""
- components = header.split(',')
+ components = [item.strip() for item in header.split(',')]
l = []
for component in components:
if ';' in component:
- subcomponents = component.split(';')
+ subcomponents = [item.strip() for item in component.split(';')]
l.append(
(
subcomponents[0], # eg. 'text/html'
subcomponents[1][2:] # eg. 'q=0.9'
)
)
else:
l.append((component, '1'))
l.sort(
key = lambda i: i[1],
reverse = True
)
content_types = []
for i in l:
content_types.append(i[0])
return content_types
| Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly | ## Code Before:
from django import forms
def generate_form(model):
"""
Generate a form from a model.
Arguments:
model -- A Django model.
"""
_model = model
class Form(forms.ModelForm):
class Meta:
model = _model
return Form
def parse_http_accept_header(header):
"""
Return a list of content types listed in the HTTP Accept header
ordered by quality.
Arguments:
header -- A string describing the contents of the HTTP Accept header.
"""
components = header.split(',')
l = []
for component in components:
if ';' in component:
subcomponents = component.split(';')
l.append(
(
subcomponents[0], # eg. 'text/html'
subcomponents[1][2:] # eg. 'q=0.9'
)
)
else:
l.append((component, '1'))
l.sort(
key = lambda i: i[1],
reverse = True
)
content_types = []
for i in l:
content_types.append(i[0])
return content_types
## Instruction:
Fix a bug that caused HTTP Accept headers with whitespace to be parsed incorrectly
## Code After:
from django import forms
def generate_form(model):
"""
Generate a form from a model.
Arguments:
model -- A Django model.
"""
_model = model
class Form(forms.ModelForm):
class Meta:
model = _model
return Form
def parse_http_accept_header(header):
"""
Return a list of content types listed in the HTTP Accept header
ordered by quality.
Arguments:
header -- A string describing the contents of the HTTP Accept header.
"""
components = [item.strip() for item in header.split(',')]
l = []
for component in components:
if ';' in component:
subcomponents = [item.strip() for item in component.split(';')]
l.append(
(
subcomponents[0], # eg. 'text/html'
subcomponents[1][2:] # eg. 'q=0.9'
)
)
else:
l.append((component, '1'))
l.sort(
key = lambda i: i[1],
reverse = True
)
content_types = []
for i in l:
content_types.append(i[0])
return content_types
| from django import forms
def generate_form(model):
"""
Generate a form from a model.
Arguments:
model -- A Django model.
"""
_model = model
class Form(forms.ModelForm):
class Meta:
model = _model
return Form
def parse_http_accept_header(header):
"""
Return a list of content types listed in the HTTP Accept header
ordered by quality.
Arguments:
header -- A string describing the contents of the HTTP Accept header.
"""
- components = header.split(',')
+ components = [item.strip() for item in header.split(',')]
l = []
for component in components:
if ';' in component:
- subcomponents = component.split(';')
+ subcomponents = [item.strip() for item in component.split(';')]
? ++++++++++++++++++++++++++ +
l.append(
(
subcomponents[0], # eg. 'text/html'
subcomponents[1][2:] # eg. 'q=0.9'
)
)
else:
l.append((component, '1'))
l.sort(
key = lambda i: i[1],
reverse = True
)
content_types = []
for i in l:
content_types.append(i[0])
return content_types |
edc5564d4c3677dc8b545e9c9a6a51b481247eab | contentcuration/contentcuration/tests/test_makemessages.py | contentcuration/contentcuration/tests/test_makemessages.py | import os
import subprocess
import pathlib
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
# this test can make changes to committed files, so only run it
# on the CI server
if 'CI' not in os.environ or not os.environ['CI']:
return
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
)
| import os
import subprocess
import pathlib
import pytest
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
# this test can make changes to committed files, so only run it
# on the CI server
@pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server")
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
)
| Use pytest.skip so we can check the test wasn't skipped on the CI. | Use pytest.skip so we can check the test wasn't skipped on the CI.
| Python | mit | DXCanas/content-curation,DXCanas/content-curation,DXCanas/content-curation,DXCanas/content-curation | import os
import subprocess
import pathlib
+ import pytest
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
+ # this test can make changes to committed files, so only run it
+ # on the CI server
+ @pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server")
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
-
- # this test can make changes to committed files, so only run it
- # on the CI server
- if 'CI' not in os.environ or not os.environ['CI']:
- return
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
)
| Use pytest.skip so we can check the test wasn't skipped on the CI. | ## Code Before:
import os
import subprocess
import pathlib
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
# this test can make changes to committed files, so only run it
# on the CI server
if 'CI' not in os.environ or not os.environ['CI']:
return
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
)
## Instruction:
Use pytest.skip so we can check the test wasn't skipped on the CI.
## Code After:
import os
import subprocess
import pathlib
import pytest
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
# this test can make changes to committed files, so only run it
# on the CI server
@pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server")
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
)
| import os
import subprocess
import pathlib
+ import pytest
from django.conf import settings
from django.test import TestCase
class MakeMessagesCommandRunTestCase(TestCase):
"""
Sanity check to make sure makemessages runs to completion.
"""
+ # this test can make changes to committed files, so only run it
+ # on the CI server
+ @pytest.mark.skipif('CI' not in os.environ or not os.environ['CI'], reason="runs only on CI server")
def test_command_succeeds_without_postgres(self):
"""
Test that we can run makemessages when postgres is not activated.
"""
-
- # this test can make changes to committed files, so only run it
- # on the CI server
- if 'CI' not in os.environ or not os.environ['CI']:
- return
repo_root = pathlib.Path(settings.BASE_DIR).parent
cmd = ["make", "makemessages"]
env = os.environ.copy()
# We fake postgres not being available, by setting the wrong IP address.
# hopefully postgres isn't running at 127.0.0.2!
env.update({"DATA_DB_HOST": "127.0.0.2"})
subprocess.check_output(
cmd,
env=env,
cwd=str(repo_root)
) |
dafde564f3ea18655b1e15f410df70d05b3eb8f5 | beets/util/collections.py | beets/util/collections.py |
class IdentityFallbackDict(dict):
"""A dictionary which is "transparent" (maps keys to themselves) for all
keys not in it.
"""
def __getitem__(self, key):
try:
return dict.__getitem__(self, key)
except KeyError:
return key
|
from __future__ import division, absolute_import, print_function
class IdentityFallbackDict(dict):
"""A dictionary which is "transparent" (maps keys to themselves) for all
keys not in it.
"""
def __getitem__(self, key):
try:
return dict.__getitem__(self, key)
except KeyError:
return key
| Add __future__ imports to a new module | Add __future__ imports to a new module
| Python | mit | mosesfistos1/beetbox,ibmibmibm/beets,mosesfistos1/beetbox,MyTunesFreeMusic/privacy-policy,artemutin/beets,jackwilsdon/beets,sampsyo/beets,pkess/beets,xsteadfastx/beets,shamangeorge/beets,diego-plan9/beets,MyTunesFreeMusic/privacy-policy,jackwilsdon/beets,beetbox/beets,sampsyo/beets,beetbox/beets,madmouser1/beets,beetbox/beets,madmouser1/beets,pkess/beets,shamangeorge/beets,diego-plan9/beets,MyTunesFreeMusic/privacy-policy,xsteadfastx/beets,Kraymer/beets,MyTunesFreeMusic/privacy-policy,SusannaMaria/beets,jackwilsdon/beets,lengtche/beets,Kraymer/beets,lengtche/beets,jackwilsdon/beets,artemutin/beets,madmouser1/beets,sampsyo/beets,ibmibmibm/beets,pkess/beets,Kraymer/beets,mosesfistos1/beetbox,shamangeorge/beets,Kraymer/beets,lengtche/beets,diego-plan9/beets,ibmibmibm/beets,SusannaMaria/beets,beetbox/beets,xsteadfastx/beets,madmouser1/beets,ibmibmibm/beets,artemutin/beets,pkess/beets,xsteadfastx/beets,artemutin/beets,SusannaMaria/beets,SusannaMaria/beets,lengtche/beets,shamangeorge/beets,mosesfistos1/beetbox,sampsyo/beets,diego-plan9/beets | +
+ from __future__ import division, absolute_import, print_function
class IdentityFallbackDict(dict):
"""A dictionary which is "transparent" (maps keys to themselves) for all
keys not in it.
"""
def __getitem__(self, key):
try:
return dict.__getitem__(self, key)
except KeyError:
return key
| Add __future__ imports to a new module | ## Code Before:
class IdentityFallbackDict(dict):
"""A dictionary which is "transparent" (maps keys to themselves) for all
keys not in it.
"""
def __getitem__(self, key):
try:
return dict.__getitem__(self, key)
except KeyError:
return key
## Instruction:
Add __future__ imports to a new module
## Code After:
from __future__ import division, absolute_import, print_function
class IdentityFallbackDict(dict):
"""A dictionary which is "transparent" (maps keys to themselves) for all
keys not in it.
"""
def __getitem__(self, key):
try:
return dict.__getitem__(self, key)
except KeyError:
return key
| +
+ from __future__ import division, absolute_import, print_function
class IdentityFallbackDict(dict):
"""A dictionary which is "transparent" (maps keys to themselves) for all
keys not in it.
"""
def __getitem__(self, key):
try:
return dict.__getitem__(self, key)
except KeyError:
return key |
8a821cb62a35547417fcd56d02486e5cc2d8494f | xzarr.py | xzarr.py | from .base import DataSourceMixin
class ZarrSource(DataSourceMixin):
"""Open a xarray dataset.
Parameters
----------
urlpath: str
Path to source. This can be a local directory or a remote data
service (i.e., with a protocol specifier like ``'s3://``).
storage_options: dict
Parameters passed to the backend file-system
kwargs:
Further parameters are passed to xr.open_zarr
"""
name = 'zarr'
def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs):
super(ZarrSource, self).__init__(metadata=metadata)
self.urlpath = urlpath
self.storage_options = storage_options
self.kwargs = kwargs
self._ds = None
def _open_dataset(self):
import xarray as xr
from dask.bytes.core import get_fs, infer_options, \
update_storage_options
urlpath, protocol, options = infer_options(self.urlpath)
update_storage_options(options, self.storage_options)
self._fs, _ = get_fs(protocol, options)
if protocol != 'file':
self._mapper = get_mapper(protocol, self._fs, urlpath)
self._ds = xr.open_zarr(self._mapper, **self.kwargs)
else:
self._ds = xr.open_zarr(self.urlpath, **self.kwargs)
def close(self):
super(ZarrSource, self).close()
self._fs = None
self._mapper = None
def get_mapper(protocol, fs, path):
if protocol == 's3':
from s3fs.mapping import S3Map
return S3Map(path, fs)
elif protocol == 'gcs':
from gcsfs.mapping import GCSMap
return GCSMap(path, fs)
else:
raise NotImplementedError
| from .base import DataSourceMixin
class ZarrSource(DataSourceMixin):
"""Open a xarray dataset.
Parameters
----------
urlpath: str
Path to source. This can be a local directory or a remote data
service (i.e., with a protocol specifier like ``'s3://``).
storage_options: dict
Parameters passed to the backend file-system
kwargs:
Further parameters are passed to xr.open_zarr
"""
name = 'zarr'
def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs):
super(ZarrSource, self).__init__(metadata=metadata)
self.urlpath = urlpath
self.storage_options = storage_options or {}
self.kwargs = kwargs
self._ds = None
def _open_dataset(self):
import xarray as xr
from fsspec import get_mapper
self._mapper = get_mapper(self.urlpath, **self.storage_options)
self._ds = xr.open_zarr(self._mapper, **self.kwargs)
def close(self):
super(ZarrSource, self).close()
self._fs = None
self._mapper = None
| Make work with any filesystem | Make work with any filesystem
| Python | bsd-3-clause | ericdill/databroker,ericdill/databroker | from .base import DataSourceMixin
class ZarrSource(DataSourceMixin):
"""Open a xarray dataset.
Parameters
----------
urlpath: str
Path to source. This can be a local directory or a remote data
service (i.e., with a protocol specifier like ``'s3://``).
storage_options: dict
Parameters passed to the backend file-system
kwargs:
Further parameters are passed to xr.open_zarr
"""
name = 'zarr'
def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs):
super(ZarrSource, self).__init__(metadata=metadata)
self.urlpath = urlpath
- self.storage_options = storage_options
+ self.storage_options = storage_options or {}
self.kwargs = kwargs
self._ds = None
def _open_dataset(self):
import xarray as xr
+ from fsspec import get_mapper
- from dask.bytes.core import get_fs, infer_options, \
- update_storage_options
- urlpath, protocol, options = infer_options(self.urlpath)
- update_storage_options(options, self.storage_options)
+ self._mapper = get_mapper(self.urlpath, **self.storage_options)
- self._fs, _ = get_fs(protocol, options)
- if protocol != 'file':
- self._mapper = get_mapper(protocol, self._fs, urlpath)
- self._ds = xr.open_zarr(self._mapper, **self.kwargs)
+ self._ds = xr.open_zarr(self._mapper, **self.kwargs)
- else:
- self._ds = xr.open_zarr(self.urlpath, **self.kwargs)
def close(self):
super(ZarrSource, self).close()
self._fs = None
self._mapper = None
-
- def get_mapper(protocol, fs, path):
- if protocol == 's3':
- from s3fs.mapping import S3Map
- return S3Map(path, fs)
- elif protocol == 'gcs':
- from gcsfs.mapping import GCSMap
- return GCSMap(path, fs)
- else:
- raise NotImplementedError
- | Make work with any filesystem | ## Code Before:
from .base import DataSourceMixin
class ZarrSource(DataSourceMixin):
"""Open a xarray dataset.
Parameters
----------
urlpath: str
Path to source. This can be a local directory or a remote data
service (i.e., with a protocol specifier like ``'s3://``).
storage_options: dict
Parameters passed to the backend file-system
kwargs:
Further parameters are passed to xr.open_zarr
"""
name = 'zarr'
def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs):
super(ZarrSource, self).__init__(metadata=metadata)
self.urlpath = urlpath
self.storage_options = storage_options
self.kwargs = kwargs
self._ds = None
def _open_dataset(self):
import xarray as xr
from dask.bytes.core import get_fs, infer_options, \
update_storage_options
urlpath, protocol, options = infer_options(self.urlpath)
update_storage_options(options, self.storage_options)
self._fs, _ = get_fs(protocol, options)
if protocol != 'file':
self._mapper = get_mapper(protocol, self._fs, urlpath)
self._ds = xr.open_zarr(self._mapper, **self.kwargs)
else:
self._ds = xr.open_zarr(self.urlpath, **self.kwargs)
def close(self):
super(ZarrSource, self).close()
self._fs = None
self._mapper = None
def get_mapper(protocol, fs, path):
if protocol == 's3':
from s3fs.mapping import S3Map
return S3Map(path, fs)
elif protocol == 'gcs':
from gcsfs.mapping import GCSMap
return GCSMap(path, fs)
else:
raise NotImplementedError
## Instruction:
Make work with any filesystem
## Code After:
from .base import DataSourceMixin
class ZarrSource(DataSourceMixin):
"""Open a xarray dataset.
Parameters
----------
urlpath: str
Path to source. This can be a local directory or a remote data
service (i.e., with a protocol specifier like ``'s3://``).
storage_options: dict
Parameters passed to the backend file-system
kwargs:
Further parameters are passed to xr.open_zarr
"""
name = 'zarr'
def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs):
super(ZarrSource, self).__init__(metadata=metadata)
self.urlpath = urlpath
self.storage_options = storage_options or {}
self.kwargs = kwargs
self._ds = None
def _open_dataset(self):
import xarray as xr
from fsspec import get_mapper
self._mapper = get_mapper(self.urlpath, **self.storage_options)
self._ds = xr.open_zarr(self._mapper, **self.kwargs)
def close(self):
super(ZarrSource, self).close()
self._fs = None
self._mapper = None
| from .base import DataSourceMixin
class ZarrSource(DataSourceMixin):
"""Open a xarray dataset.
Parameters
----------
urlpath: str
Path to source. This can be a local directory or a remote data
service (i.e., with a protocol specifier like ``'s3://``).
storage_options: dict
Parameters passed to the backend file-system
kwargs:
Further parameters are passed to xr.open_zarr
"""
name = 'zarr'
def __init__(self, urlpath, storage_options=None, metadata=None, **kwargs):
super(ZarrSource, self).__init__(metadata=metadata)
self.urlpath = urlpath
- self.storage_options = storage_options
+ self.storage_options = storage_options or {}
? ++++++
self.kwargs = kwargs
self._ds = None
def _open_dataset(self):
import xarray as xr
+ from fsspec import get_mapper
- from dask.bytes.core import get_fs, infer_options, \
- update_storage_options
- urlpath, protocol, options = infer_options(self.urlpath)
- update_storage_options(options, self.storage_options)
+ self._mapper = get_mapper(self.urlpath, **self.storage_options)
- self._fs, _ = get_fs(protocol, options)
- if protocol != 'file':
- self._mapper = get_mapper(protocol, self._fs, urlpath)
- self._ds = xr.open_zarr(self._mapper, **self.kwargs)
? ----
+ self._ds = xr.open_zarr(self._mapper, **self.kwargs)
- else:
- self._ds = xr.open_zarr(self.urlpath, **self.kwargs)
def close(self):
super(ZarrSource, self).close()
self._fs = None
self._mapper = None
-
-
- def get_mapper(protocol, fs, path):
- if protocol == 's3':
- from s3fs.mapping import S3Map
- return S3Map(path, fs)
- elif protocol == 'gcs':
- from gcsfs.mapping import GCSMap
- return GCSMap(path, fs)
- else:
- raise NotImplementedError |
07fd8bf23917e18ba419859d788d9f51735f3b39 | conda_gitenv/__init__.py | conda_gitenv/__init__.py | from __future__ import absolute_import, division, print_function, unicode_literals
from distutils.version import StrictVersion
from conda import __version__ as CONDA_VERSION
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
_conda_base = StrictVersion('4.3.0')
_conda_support = StrictVersion(CONDA_VERSION) >= _conda_base
assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base)
manifest_branch_prefix = 'manifest/'
| from __future__ import absolute_import, division, print_function, unicode_literals
from distutils.version import StrictVersion
from conda import __version__ as CONDA_VERSION
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
_conda_base = StrictVersion('4.3.0')
_conda_version = StrictVersion(CONDA_VERSION)
_conda_supported = _conda_version >= _conda_base
assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version)
manifest_branch_prefix = 'manifest/'
| Update minimum conda version diagnostic | Update minimum conda version diagnostic
| Python | bsd-3-clause | SciTools/conda-gitenv | from __future__ import absolute_import, division, print_function, unicode_literals
from distutils.version import StrictVersion
from conda import __version__ as CONDA_VERSION
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
_conda_base = StrictVersion('4.3.0')
- _conda_support = StrictVersion(CONDA_VERSION) >= _conda_base
+ _conda_version = StrictVersion(CONDA_VERSION)
+ _conda_supported = _conda_version >= _conda_base
- assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base)
+ assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version)
manifest_branch_prefix = 'manifest/'
| Update minimum conda version diagnostic | ## Code Before:
from __future__ import absolute_import, division, print_function, unicode_literals
from distutils.version import StrictVersion
from conda import __version__ as CONDA_VERSION
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
_conda_base = StrictVersion('4.3.0')
_conda_support = StrictVersion(CONDA_VERSION) >= _conda_base
assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base)
manifest_branch_prefix = 'manifest/'
## Instruction:
Update minimum conda version diagnostic
## Code After:
from __future__ import absolute_import, division, print_function, unicode_literals
from distutils.version import StrictVersion
from conda import __version__ as CONDA_VERSION
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
_conda_base = StrictVersion('4.3.0')
_conda_version = StrictVersion(CONDA_VERSION)
_conda_supported = _conda_version >= _conda_base
assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version)
manifest_branch_prefix = 'manifest/'
| from __future__ import absolute_import, division, print_function, unicode_literals
from distutils.version import StrictVersion
from conda import __version__ as CONDA_VERSION
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
_conda_base = StrictVersion('4.3.0')
- _conda_support = StrictVersion(CONDA_VERSION) >= _conda_base
? ^^^ ^^ ---------------
+ _conda_version = StrictVersion(CONDA_VERSION)
? +++ ^ ^
+ _conda_supported = _conda_version >= _conda_base
- assert _conda_support, 'Minimum supported conda version is {}.'.format(_conda_base)
+ assert _conda_support, 'Minimum supported conda version is {}, got {}.'.format(_conda_base, _conda_version)
? ++++++++ ++++++++++++++++
manifest_branch_prefix = 'manifest/' |
b5d812504924af2e2781f4be63a6191e5c47879d | test_project/urls.py | test_project/urls.py | from django.conf.urls import url, include
from django.conf import settings
from django.contrib import admin
from django.views.generic import TemplateView
TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html')
urlpatterns = [
url('^accounts/', include('django.contrib.auth.urls')),
url(r'^admin/', admin.site.urls),
url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)),
]
| from django.conf.urls import url, include
from django.conf import settings
from django.contrib import admin
from django.views.generic import TemplateView
urlpatterns = [
url('^accounts/', include('django.contrib.auth.urls')),
url(r'^admin/', admin.site.urls),
]
TEST_TEMPLATES = getattr(
settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')])
for path, template in TEST_TEMPLATES:
urlpatterns.append(url(path, TemplateView.as_view(template_name=template)))
| Support multiple templates in TEST_TEMPLATES setting. | Support multiple templates in TEST_TEMPLATES setting.
Unit tests need to be able to test redirects and other features
involving multiple web pages. This commit changes the singleton
TEST_TEMPLATE setting to TEST_TEMPLATES, which is a list of
path, template tuples.
| Python | bsd-3-clause | nimbis/django-selenium-testcase,nimbis/django-selenium-testcase | from django.conf.urls import url, include
from django.conf import settings
from django.contrib import admin
from django.views.generic import TemplateView
- TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html')
-
urlpatterns = [
url('^accounts/', include('django.contrib.auth.urls')),
url(r'^admin/', admin.site.urls),
- url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)),
]
+ TEST_TEMPLATES = getattr(
+ settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')])
+
+ for path, template in TEST_TEMPLATES:
+ urlpatterns.append(url(path, TemplateView.as_view(template_name=template)))
+ | Support multiple templates in TEST_TEMPLATES setting. | ## Code Before:
from django.conf.urls import url, include
from django.conf import settings
from django.contrib import admin
from django.views.generic import TemplateView
TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html')
urlpatterns = [
url('^accounts/', include('django.contrib.auth.urls')),
url(r'^admin/', admin.site.urls),
url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)),
]
## Instruction:
Support multiple templates in TEST_TEMPLATES setting.
## Code After:
from django.conf.urls import url, include
from django.conf import settings
from django.contrib import admin
from django.views.generic import TemplateView
urlpatterns = [
url('^accounts/', include('django.contrib.auth.urls')),
url(r'^admin/', admin.site.urls),
]
TEST_TEMPLATES = getattr(
settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')])
for path, template in TEST_TEMPLATES:
urlpatterns.append(url(path, TemplateView.as_view(template_name=template)))
| from django.conf.urls import url, include
from django.conf import settings
from django.contrib import admin
from django.views.generic import TemplateView
- TEST_TEMPLATE = getattr(settings, 'TEST_TEMPLATE', 'test.html')
-
urlpatterns = [
url('^accounts/', include('django.contrib.auth.urls')),
url(r'^admin/', admin.site.urls),
- url(r'^$', TemplateView.as_view(template_name=TEST_TEMPLATE)),
]
+
+ TEST_TEMPLATES = getattr(
+ settings, 'TEST_TEMPLATES', [(r'^$', 'test.html')])
+
+ for path, template in TEST_TEMPLATES:
+ urlpatterns.append(url(path, TemplateView.as_view(template_name=template))) |
836845abde53ee55bca93f098ece78880ab6b5c6 | examples/events/create_massive_dummy_events.py | examples/events/create_massive_dummy_events.py |
from pymisp import PyMISP
from keys import misp_url, misp_key, misp_verifycert
import argparse
import tools
def init(url, key):
return PyMISP(url, key, misp_verifycert, 'json')
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.')
parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)")
parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)")
args = parser.parse_args()
misp = init(misp_url, misp_key)
if args.limit is None:
args.limit = 1
if args.attribute is None:
args.attribute = 3000
for i in range(args.limit):
tools.create_massive_dummy_events(misp, args.attribute)
|
from pymisp import PyMISP
from keys import url, key
import argparse
import tools
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.')
parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)")
parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)")
args = parser.parse_args()
misp = PyMISP(url, key, True, 'json')
if args.limit is None:
args.limit = 1
if args.attribute is None:
args.attribute = 3000
for i in range(args.limit):
tools.create_massive_dummy_events(misp, args.attribute)
| Use same variable names as testing environment | Use same variable names as testing environment
| Python | bsd-2-clause | pombredanne/PyMISP,iglocska/PyMISP |
from pymisp import PyMISP
- from keys import misp_url, misp_key, misp_verifycert
+ from keys import url, key
import argparse
import tools
- def init(url, key):
- return PyMISP(url, key, misp_verifycert, 'json')
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.')
parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)")
parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)")
args = parser.parse_args()
- misp = init(misp_url, misp_key)
+ misp = PyMISP(url, key, True, 'json')
if args.limit is None:
args.limit = 1
if args.attribute is None:
args.attribute = 3000
for i in range(args.limit):
tools.create_massive_dummy_events(misp, args.attribute)
| Use same variable names as testing environment | ## Code Before:
from pymisp import PyMISP
from keys import misp_url, misp_key, misp_verifycert
import argparse
import tools
def init(url, key):
return PyMISP(url, key, misp_verifycert, 'json')
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.')
parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)")
parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)")
args = parser.parse_args()
misp = init(misp_url, misp_key)
if args.limit is None:
args.limit = 1
if args.attribute is None:
args.attribute = 3000
for i in range(args.limit):
tools.create_massive_dummy_events(misp, args.attribute)
## Instruction:
Use same variable names as testing environment
## Code After:
from pymisp import PyMISP
from keys import url, key
import argparse
import tools
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.')
parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)")
parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)")
args = parser.parse_args()
misp = PyMISP(url, key, True, 'json')
if args.limit is None:
args.limit = 1
if args.attribute is None:
args.attribute = 3000
for i in range(args.limit):
tools.create_massive_dummy_events(misp, args.attribute)
|
from pymisp import PyMISP
- from keys import misp_url, misp_key, misp_verifycert
+ from keys import url, key
import argparse
import tools
- def init(url, key):
- return PyMISP(url, key, misp_verifycert, 'json')
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Create a given number of event containing a given number of attributes eachh.')
parser.add_argument("-l", "--limit", type=int, help="Number of events to create (default 1)")
parser.add_argument("-a", "--attribute", type=int, help="Number of attributes per event (default 3000)")
args = parser.parse_args()
- misp = init(misp_url, misp_key)
+ misp = PyMISP(url, key, True, 'json')
if args.limit is None:
args.limit = 1
if args.attribute is None:
args.attribute = 3000
for i in range(args.limit):
tools.create_massive_dummy_events(misp, args.attribute) |
1cbab715a647689aeda4679d7dcf4e60ff9ab5b1 | api/webview/models.py | api/webview/models.py | from django.db import models
from django_pgjson.fields import JsonField
class Document(models.Model):
source = models.CharField(max_length=100)
docID = models.CharField(max_length=100)
providerUpdatedDateTime = models.DateTimeField(null=True)
raw = JsonField()
normalized = JsonField()
| import json
import six
from requests.structures import CaseInsensitiveDict
from django.db import models
from django_pgjson.fields import JsonField
class Document(models.Model):
source = models.CharField(max_length=100)
docID = models.CharField(max_length=100)
providerUpdatedDateTime = models.DateTimeField(null=True)
raw = JsonField()
normalized = JsonField()
class HarvesterResponse(models.Model):
method = models.TextField(primary_key=True)
url = models.TextField(primary_key=True, required=True)
# Raw request data
ok = models.BooleanField()
content = models.BinaryField()
encoding = models.TextField()
headers_str = models.TextField()
status_code = models.IntegerField()
time_made = models.DateTimeField(auto_now=True)
def json(self):
return json.loads(self.content)
@property
def headers(self):
return CaseInsensitiveDict(json.loads(self.headers_str))
@property
def text(self):
return six.u(self.content)
| Add harvester response model in django ORM | Add harvester response model in django ORM
| Python | apache-2.0 | felliott/scrapi,fabianvf/scrapi,erinspace/scrapi,mehanig/scrapi,felliott/scrapi,erinspace/scrapi,CenterForOpenScience/scrapi,CenterForOpenScience/scrapi,mehanig/scrapi,fabianvf/scrapi | + import json
+
+ import six
+ from requests.structures import CaseInsensitiveDict
+
from django.db import models
-
from django_pgjson.fields import JsonField
class Document(models.Model):
source = models.CharField(max_length=100)
docID = models.CharField(max_length=100)
providerUpdatedDateTime = models.DateTimeField(null=True)
raw = JsonField()
normalized = JsonField()
+
+ class HarvesterResponse(models.Model):
+
+ method = models.TextField(primary_key=True)
+ url = models.TextField(primary_key=True, required=True)
+
+ # Raw request data
+ ok = models.BooleanField()
+ content = models.BinaryField()
+ encoding = models.TextField()
+ headers_str = models.TextField()
+ status_code = models.IntegerField()
+ time_made = models.DateTimeField(auto_now=True)
+
+ def json(self):
+ return json.loads(self.content)
+
+ @property
+ def headers(self):
+ return CaseInsensitiveDict(json.loads(self.headers_str))
+
+ @property
+ def text(self):
+ return six.u(self.content)
+ | Add harvester response model in django ORM | ## Code Before:
from django.db import models
from django_pgjson.fields import JsonField
class Document(models.Model):
source = models.CharField(max_length=100)
docID = models.CharField(max_length=100)
providerUpdatedDateTime = models.DateTimeField(null=True)
raw = JsonField()
normalized = JsonField()
## Instruction:
Add harvester response model in django ORM
## Code After:
import json
import six
from requests.structures import CaseInsensitiveDict
from django.db import models
from django_pgjson.fields import JsonField
class Document(models.Model):
source = models.CharField(max_length=100)
docID = models.CharField(max_length=100)
providerUpdatedDateTime = models.DateTimeField(null=True)
raw = JsonField()
normalized = JsonField()
class HarvesterResponse(models.Model):
method = models.TextField(primary_key=True)
url = models.TextField(primary_key=True, required=True)
# Raw request data
ok = models.BooleanField()
content = models.BinaryField()
encoding = models.TextField()
headers_str = models.TextField()
status_code = models.IntegerField()
time_made = models.DateTimeField(auto_now=True)
def json(self):
return json.loads(self.content)
@property
def headers(self):
return CaseInsensitiveDict(json.loads(self.headers_str))
@property
def text(self):
return six.u(self.content)
| + import json
+
+ import six
+ from requests.structures import CaseInsensitiveDict
+
from django.db import models
-
from django_pgjson.fields import JsonField
class Document(models.Model):
source = models.CharField(max_length=100)
docID = models.CharField(max_length=100)
providerUpdatedDateTime = models.DateTimeField(null=True)
raw = JsonField()
normalized = JsonField()
+
+
+ class HarvesterResponse(models.Model):
+
+ method = models.TextField(primary_key=True)
+ url = models.TextField(primary_key=True, required=True)
+
+ # Raw request data
+ ok = models.BooleanField()
+ content = models.BinaryField()
+ encoding = models.TextField()
+ headers_str = models.TextField()
+ status_code = models.IntegerField()
+ time_made = models.DateTimeField(auto_now=True)
+
+ def json(self):
+ return json.loads(self.content)
+
+ @property
+ def headers(self):
+ return CaseInsensitiveDict(json.loads(self.headers_str))
+
+ @property
+ def text(self):
+ return six.u(self.content) |
11c30f5dd765475a9f5f0f847f31c47af8c40a39 | user_agent/device.py | user_agent/device.py | import os.path
import json
PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__))
with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f:
SMARTPHONE_DEV_IDS = json.load(open(f))
with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f:
TABLET_DEV_IDS = json.load(open())
| import os.path
import json
PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__))
with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f:
SMARTPHONE_DEV_IDS = json.load(f)
with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f:
TABLET_DEV_IDS = json.load(f)
| Fix uses of file objects | Fix uses of file objects | Python | mit | lorien/user_agent | import os.path
import json
PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__))
with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f:
- SMARTPHONE_DEV_IDS = json.load(open(f))
+ SMARTPHONE_DEV_IDS = json.load(f)
with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f:
- TABLET_DEV_IDS = json.load(open())
+ TABLET_DEV_IDS = json.load(f)
| Fix uses of file objects | ## Code Before:
import os.path
import json
PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__))
with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f:
SMARTPHONE_DEV_IDS = json.load(open(f))
with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f:
TABLET_DEV_IDS = json.load(open())
## Instruction:
Fix uses of file objects
## Code After:
import os.path
import json
PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__))
with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f:
SMARTPHONE_DEV_IDS = json.load(f)
with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f:
TABLET_DEV_IDS = json.load(f)
| import os.path
import json
PACKAGE_DIR = os.path.dirname(os.path.realpath(__file__))
with open(os.path.join(PACKAGE_DIR, 'data/smartphone_dev_id.json')) as f:
- SMARTPHONE_DEV_IDS = json.load(open(f))
? ----- -
+ SMARTPHONE_DEV_IDS = json.load(f)
with open(os.path.join(PACKAGE_DIR, 'data/tablet_dev_id.json')) as f:
- TABLET_DEV_IDS = json.load(open())
? ^^^^^ -
+ TABLET_DEV_IDS = json.load(f)
? ^
|
ea315b018fb3fab6925f1194fcd3e341166ab6fb | opt/resource/common.py | opt/resource/common.py | import json
from urllib.parse import urljoin
import sys
def get_payload():
return json.load(sys.stdin)
def get_index_url(payload):
source = payload['source']
uri = source['uri']
index = source['index']
if not uri.endswith('/'):
uri += '/'
if not index.endswith('/'):
index += '/'
return urljoin(uri, index)
def get_package_url(payload):
package = payload['source']['package']
return get_index_url(payload) + package
def get_auth(payload):
source = payload['source']
return source['username'], source['password']
def get_version(payload):
if 'version' in payload:
version = payload['version']['version']
else:
version = None
return version
| import json
from urllib.parse import urljoin
import sys
def get_payload():
return json.load(sys.stdin)
def get_index_url(payload):
source = payload['source']
uri = source['uri']
index = source['index']
if not uri.endswith('/'):
uri += '/'
if not index.endswith('/'):
index += '/'
return urljoin(uri, index)
def get_package_url(payload):
package = payload['source']['package']
return get_index_url(payload) + package
def get_auth(payload):
source = payload['source']
return source['username'], source['password']
def get_version(payload):
try:
version = payload['version']['version']
except TypeError:
version = None
return version
| Handle missing version in payload. | Handle missing version in payload.
| Python | mit | mdomke/concourse-devpi-resource | import json
from urllib.parse import urljoin
import sys
def get_payload():
return json.load(sys.stdin)
def get_index_url(payload):
source = payload['source']
uri = source['uri']
index = source['index']
if not uri.endswith('/'):
uri += '/'
if not index.endswith('/'):
index += '/'
return urljoin(uri, index)
def get_package_url(payload):
package = payload['source']['package']
return get_index_url(payload) + package
def get_auth(payload):
source = payload['source']
return source['username'], source['password']
def get_version(payload):
- if 'version' in payload:
+ try:
version = payload['version']['version']
- else:
+ except TypeError:
version = None
return version
| Handle missing version in payload. | ## Code Before:
import json
from urllib.parse import urljoin
import sys
def get_payload():
return json.load(sys.stdin)
def get_index_url(payload):
source = payload['source']
uri = source['uri']
index = source['index']
if not uri.endswith('/'):
uri += '/'
if not index.endswith('/'):
index += '/'
return urljoin(uri, index)
def get_package_url(payload):
package = payload['source']['package']
return get_index_url(payload) + package
def get_auth(payload):
source = payload['source']
return source['username'], source['password']
def get_version(payload):
if 'version' in payload:
version = payload['version']['version']
else:
version = None
return version
## Instruction:
Handle missing version in payload.
## Code After:
import json
from urllib.parse import urljoin
import sys
def get_payload():
return json.load(sys.stdin)
def get_index_url(payload):
source = payload['source']
uri = source['uri']
index = source['index']
if not uri.endswith('/'):
uri += '/'
if not index.endswith('/'):
index += '/'
return urljoin(uri, index)
def get_package_url(payload):
package = payload['source']['package']
return get_index_url(payload) + package
def get_auth(payload):
source = payload['source']
return source['username'], source['password']
def get_version(payload):
try:
version = payload['version']['version']
except TypeError:
version = None
return version
| import json
from urllib.parse import urljoin
import sys
def get_payload():
return json.load(sys.stdin)
def get_index_url(payload):
source = payload['source']
uri = source['uri']
index = source['index']
if not uri.endswith('/'):
uri += '/'
if not index.endswith('/'):
index += '/'
return urljoin(uri, index)
def get_package_url(payload):
package = payload['source']['package']
return get_index_url(payload) + package
def get_auth(payload):
source = payload['source']
return source['username'], source['password']
def get_version(payload):
- if 'version' in payload:
+ try:
version = payload['version']['version']
- else:
+ except TypeError:
version = None
return version |
0138eacf0d518b86e819a70000b7b527434a6b35 | libretto/signals.py | libretto/signals.py |
from __future__ import unicode_literals
from celery_haystack.signals import CelerySignalProcessor
from django.contrib.admin.models import LogEntry
from reversion.models import Version, Revision
from .tasks import auto_invalidate
class CeleryAutoInvalidator(CelerySignalProcessor):
def enqueue(self, action, instance, sender, **kwargs):
if sender in (LogEntry, Revision, Version):
return
auto_invalidate.delay(action, instance)
|
from __future__ import unicode_literals
from celery_haystack.signals import CelerySignalProcessor
from django.contrib.admin.models import LogEntry
from django.contrib.sessions.models import Session
from reversion.models import Version, Revision
from .tasks import auto_invalidate
class CeleryAutoInvalidator(CelerySignalProcessor):
def enqueue(self, action, instance, sender, **kwargs):
if sender in (LogEntry, Session, Revision, Version):
return
auto_invalidate.delay(action, instance.__class__, instance.pk)
| Change les arguments passés à celery pour gérer la sérialisation JSON. | Change les arguments passés à celery pour gérer la sérialisation JSON.
| Python | bsd-3-clause | dezede/dezede,dezede/dezede,dezede/dezede,dezede/dezede |
from __future__ import unicode_literals
from celery_haystack.signals import CelerySignalProcessor
from django.contrib.admin.models import LogEntry
+ from django.contrib.sessions.models import Session
from reversion.models import Version, Revision
from .tasks import auto_invalidate
class CeleryAutoInvalidator(CelerySignalProcessor):
def enqueue(self, action, instance, sender, **kwargs):
- if sender in (LogEntry, Revision, Version):
+ if sender in (LogEntry, Session, Revision, Version):
return
- auto_invalidate.delay(action, instance)
+ auto_invalidate.delay(action, instance.__class__, instance.pk)
| Change les arguments passés à celery pour gérer la sérialisation JSON. | ## Code Before:
from __future__ import unicode_literals
from celery_haystack.signals import CelerySignalProcessor
from django.contrib.admin.models import LogEntry
from reversion.models import Version, Revision
from .tasks import auto_invalidate
class CeleryAutoInvalidator(CelerySignalProcessor):
def enqueue(self, action, instance, sender, **kwargs):
if sender in (LogEntry, Revision, Version):
return
auto_invalidate.delay(action, instance)
## Instruction:
Change les arguments passés à celery pour gérer la sérialisation JSON.
## Code After:
from __future__ import unicode_literals
from celery_haystack.signals import CelerySignalProcessor
from django.contrib.admin.models import LogEntry
from django.contrib.sessions.models import Session
from reversion.models import Version, Revision
from .tasks import auto_invalidate
class CeleryAutoInvalidator(CelerySignalProcessor):
def enqueue(self, action, instance, sender, **kwargs):
if sender in (LogEntry, Session, Revision, Version):
return
auto_invalidate.delay(action, instance.__class__, instance.pk)
|
from __future__ import unicode_literals
from celery_haystack.signals import CelerySignalProcessor
from django.contrib.admin.models import LogEntry
+ from django.contrib.sessions.models import Session
from reversion.models import Version, Revision
from .tasks import auto_invalidate
class CeleryAutoInvalidator(CelerySignalProcessor):
def enqueue(self, action, instance, sender, **kwargs):
- if sender in (LogEntry, Revision, Version):
+ if sender in (LogEntry, Session, Revision, Version):
? +++++++++
return
- auto_invalidate.delay(action, instance)
+ auto_invalidate.delay(action, instance.__class__, instance.pk)
? +++++++++++++++++++++++
|
396df5eac473fccc16e103d3d3316aefd653789a | changeling/models.py | changeling/models.py | import uuid
import jsonschema
import changeling.exception
class Change(object):
schema = {
'name': 'change',
'properties': {
'id': {'type': 'string'},
'name': {'type': 'string'},
'description': {'type': 'string'},
},
'additionalProperties': False,
}
def __init__(self, id=None, name=None, description=None):
self.id = id or str(uuid.uuid4())
self.name = name
self.description = description
@classmethod
def from_dict(self, data):
self.validate(data)
return Change(**data)
def to_dict(self):
def _generate_set_attributes():
for k in Change.schema['properties'].keys():
val = getattr(self, k)
if val is not None:
yield (k, val)
return dict(_generate_set_attributes())
def __str__(self):
return "<Change id=%s name=%s>" % (self.id, self.name)
@classmethod
def validate(cls, data):
try:
jsonschema.validate(data, cls.schema)
except jsonschema.ValidationError as exc:
raise changeling.exception.ValidationError(exc)
def is_valid(self):
try:
self.validate(self.to_dict())
except changeling.exception.ValidationError:
return False
else:
return True
| import uuid
import jsonschema
import changeling.exception
class Change(object):
schema = {
'name': 'change',
'properties': {
'id': {'type': 'string'},
'name': {'type': 'string'},
'description': {'type': 'string'},
'tags': {'type': 'array'},
},
'additionalProperties': False,
}
def __init__(self, id=None, name=None, description=None, tags=None):
self.id = id or str(uuid.uuid4())
self.name = name
self.description = description
self.tags = tags or []
@classmethod
def from_dict(self, data):
self.validate(data)
return Change(**data)
def to_dict(self):
def _generate_set_attributes():
for k in Change.schema['properties'].keys():
val = getattr(self, k)
if val is not None:
yield (k, val)
return dict(_generate_set_attributes())
def __str__(self):
return "<Change id=%s name=%s>" % (self.id, self.name)
@classmethod
def validate(cls, data):
try:
jsonschema.validate(data, cls.schema)
except jsonschema.ValidationError as exc:
raise changeling.exception.ValidationError(exc)
def is_valid(self):
try:
self.validate(self.to_dict())
except changeling.exception.ValidationError:
return False
else:
return True
| Add tags - that was too easy | Add tags - that was too easy
| Python | apache-2.0 | bcwaldon/changeling,bcwaldon/changeling | import uuid
import jsonschema
import changeling.exception
class Change(object):
schema = {
'name': 'change',
'properties': {
'id': {'type': 'string'},
'name': {'type': 'string'},
'description': {'type': 'string'},
+ 'tags': {'type': 'array'},
},
'additionalProperties': False,
}
- def __init__(self, id=None, name=None, description=None):
+ def __init__(self, id=None, name=None, description=None, tags=None):
self.id = id or str(uuid.uuid4())
self.name = name
self.description = description
+ self.tags = tags or []
@classmethod
def from_dict(self, data):
self.validate(data)
return Change(**data)
def to_dict(self):
def _generate_set_attributes():
for k in Change.schema['properties'].keys():
val = getattr(self, k)
if val is not None:
yield (k, val)
return dict(_generate_set_attributes())
def __str__(self):
return "<Change id=%s name=%s>" % (self.id, self.name)
@classmethod
def validate(cls, data):
try:
jsonschema.validate(data, cls.schema)
except jsonschema.ValidationError as exc:
raise changeling.exception.ValidationError(exc)
def is_valid(self):
try:
self.validate(self.to_dict())
except changeling.exception.ValidationError:
return False
else:
return True
| Add tags - that was too easy | ## Code Before:
import uuid
import jsonschema
import changeling.exception
class Change(object):
schema = {
'name': 'change',
'properties': {
'id': {'type': 'string'},
'name': {'type': 'string'},
'description': {'type': 'string'},
},
'additionalProperties': False,
}
def __init__(self, id=None, name=None, description=None):
self.id = id or str(uuid.uuid4())
self.name = name
self.description = description
@classmethod
def from_dict(self, data):
self.validate(data)
return Change(**data)
def to_dict(self):
def _generate_set_attributes():
for k in Change.schema['properties'].keys():
val = getattr(self, k)
if val is not None:
yield (k, val)
return dict(_generate_set_attributes())
def __str__(self):
return "<Change id=%s name=%s>" % (self.id, self.name)
@classmethod
def validate(cls, data):
try:
jsonschema.validate(data, cls.schema)
except jsonschema.ValidationError as exc:
raise changeling.exception.ValidationError(exc)
def is_valid(self):
try:
self.validate(self.to_dict())
except changeling.exception.ValidationError:
return False
else:
return True
## Instruction:
Add tags - that was too easy
## Code After:
import uuid
import jsonschema
import changeling.exception
class Change(object):
schema = {
'name': 'change',
'properties': {
'id': {'type': 'string'},
'name': {'type': 'string'},
'description': {'type': 'string'},
'tags': {'type': 'array'},
},
'additionalProperties': False,
}
def __init__(self, id=None, name=None, description=None, tags=None):
self.id = id or str(uuid.uuid4())
self.name = name
self.description = description
self.tags = tags or []
@classmethod
def from_dict(self, data):
self.validate(data)
return Change(**data)
def to_dict(self):
def _generate_set_attributes():
for k in Change.schema['properties'].keys():
val = getattr(self, k)
if val is not None:
yield (k, val)
return dict(_generate_set_attributes())
def __str__(self):
return "<Change id=%s name=%s>" % (self.id, self.name)
@classmethod
def validate(cls, data):
try:
jsonschema.validate(data, cls.schema)
except jsonschema.ValidationError as exc:
raise changeling.exception.ValidationError(exc)
def is_valid(self):
try:
self.validate(self.to_dict())
except changeling.exception.ValidationError:
return False
else:
return True
| import uuid
import jsonschema
import changeling.exception
class Change(object):
schema = {
'name': 'change',
'properties': {
'id': {'type': 'string'},
'name': {'type': 'string'},
'description': {'type': 'string'},
+ 'tags': {'type': 'array'},
},
'additionalProperties': False,
}
- def __init__(self, id=None, name=None, description=None):
+ def __init__(self, id=None, name=None, description=None, tags=None):
? +++++++++++
self.id = id or str(uuid.uuid4())
self.name = name
self.description = description
+ self.tags = tags or []
@classmethod
def from_dict(self, data):
self.validate(data)
return Change(**data)
def to_dict(self):
def _generate_set_attributes():
for k in Change.schema['properties'].keys():
val = getattr(self, k)
if val is not None:
yield (k, val)
return dict(_generate_set_attributes())
def __str__(self):
return "<Change id=%s name=%s>" % (self.id, self.name)
@classmethod
def validate(cls, data):
try:
jsonschema.validate(data, cls.schema)
except jsonschema.ValidationError as exc:
raise changeling.exception.ValidationError(exc)
def is_valid(self):
try:
self.validate(self.to_dict())
except changeling.exception.ValidationError:
return False
else:
return True |
1cafb39b6204010d3e17b059254af6042f4a9efc | apts/__init__.py | apts/__init__.py | import os
import shutil
import configparser
from .equipment import Equipment
from .observations import Observation
from .place import Place
from .weather import Weather
from .notify import Notify
from .catalogs import Catalogs
from .utils import Utils
user_config = os.path.expanduser("~") + "/.config/apts/apts.ini"
config = configparser.ConfigParser()
config.read(user_config)
# Load static fields from config
Weather.API_KEY = config['weather']['api_key']
Weather.API_URL = config['weather']['api_url']
Notify.EMAIL_ADDRESS = config['notification']['email_address']
Notify.EMAIL_PASSWORD = config['notification']['email_password']
__version__ = '0.2.9'
| import os
import shutil
import configparser
from .equipment import Equipment
from .observations import Observation
from .place import Place
from .weather import Weather
from .notify import Notify
from .catalogs import Catalogs
from .utils import Utils
# Default values for configuration values
DEFAULTS = {
'weather': {
'api_key': 'unknown',
'api_url': 'unknown'
},
'notification': {
'email_address': 'unknown',
'email_password': 'unknown'
}
}
# Init config with default values
config = configparser.ConfigParser(DEFAULTS)
# Read users configuration
user_config = os.path.expanduser("~") + "/.config/apts/apts.ini"
config.read(user_config)
# Load static fields from config
Weather.API_KEY = config['weather']['api_key']
Weather.API_URL = config['weather']['api_url']
Notify.EMAIL_ADDRESS = config['notification']['email_address']
Notify.EMAIL_PASSWORD = config['notification']['email_password']
__version__ = '0.2.9'
| Add default values for config | Add default values for config
| Python | apache-2.0 | pozar87/apts | import os
import shutil
import configparser
from .equipment import Equipment
from .observations import Observation
from .place import Place
from .weather import Weather
from .notify import Notify
from .catalogs import Catalogs
from .utils import Utils
+ # Default values for configuration values
+ DEFAULTS = {
+ 'weather': {
+ 'api_key': 'unknown',
+ 'api_url': 'unknown'
+ },
+ 'notification': {
+ 'email_address': 'unknown',
+ 'email_password': 'unknown'
+ }
+ }
+
+ # Init config with default values
+ config = configparser.ConfigParser(DEFAULTS)
+
+ # Read users configuration
user_config = os.path.expanduser("~") + "/.config/apts/apts.ini"
-
- config = configparser.ConfigParser()
config.read(user_config)
# Load static fields from config
Weather.API_KEY = config['weather']['api_key']
Weather.API_URL = config['weather']['api_url']
Notify.EMAIL_ADDRESS = config['notification']['email_address']
Notify.EMAIL_PASSWORD = config['notification']['email_password']
__version__ = '0.2.9'
| Add default values for config | ## Code Before:
import os
import shutil
import configparser
from .equipment import Equipment
from .observations import Observation
from .place import Place
from .weather import Weather
from .notify import Notify
from .catalogs import Catalogs
from .utils import Utils
user_config = os.path.expanduser("~") + "/.config/apts/apts.ini"
config = configparser.ConfigParser()
config.read(user_config)
# Load static fields from config
Weather.API_KEY = config['weather']['api_key']
Weather.API_URL = config['weather']['api_url']
Notify.EMAIL_ADDRESS = config['notification']['email_address']
Notify.EMAIL_PASSWORD = config['notification']['email_password']
__version__ = '0.2.9'
## Instruction:
Add default values for config
## Code After:
import os
import shutil
import configparser
from .equipment import Equipment
from .observations import Observation
from .place import Place
from .weather import Weather
from .notify import Notify
from .catalogs import Catalogs
from .utils import Utils
# Default values for configuration values
DEFAULTS = {
'weather': {
'api_key': 'unknown',
'api_url': 'unknown'
},
'notification': {
'email_address': 'unknown',
'email_password': 'unknown'
}
}
# Init config with default values
config = configparser.ConfigParser(DEFAULTS)
# Read users configuration
user_config = os.path.expanduser("~") + "/.config/apts/apts.ini"
config.read(user_config)
# Load static fields from config
Weather.API_KEY = config['weather']['api_key']
Weather.API_URL = config['weather']['api_url']
Notify.EMAIL_ADDRESS = config['notification']['email_address']
Notify.EMAIL_PASSWORD = config['notification']['email_password']
__version__ = '0.2.9'
| import os
import shutil
import configparser
from .equipment import Equipment
from .observations import Observation
from .place import Place
from .weather import Weather
from .notify import Notify
from .catalogs import Catalogs
from .utils import Utils
+ # Default values for configuration values
+ DEFAULTS = {
+ 'weather': {
+ 'api_key': 'unknown',
+ 'api_url': 'unknown'
+ },
+ 'notification': {
+ 'email_address': 'unknown',
+ 'email_password': 'unknown'
+ }
+ }
+
+ # Init config with default values
+ config = configparser.ConfigParser(DEFAULTS)
+
+ # Read users configuration
user_config = os.path.expanduser("~") + "/.config/apts/apts.ini"
-
- config = configparser.ConfigParser()
config.read(user_config)
# Load static fields from config
Weather.API_KEY = config['weather']['api_key']
Weather.API_URL = config['weather']['api_url']
Notify.EMAIL_ADDRESS = config['notification']['email_address']
Notify.EMAIL_PASSWORD = config['notification']['email_password']
__version__ = '0.2.9' |
8c3e3ec6076d8b9ee858fca00d92717d77c67ade | time_lapse.py | time_lapse.py |
import sys
import time
import picamera
import settings
from settings import IMAGE, SNAP
import uploader
def main():
with picamera.PiCamera() as camera:
camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y)
time.sleep(2)
output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
capture = camera.capture_continuous(output_file, quality=IMAGE.quality)
for i, _ in enumerate(capture):
if i == SNAP.total - 1:
break
time.sleep(SNAP.interval)
if __name__ == '__main__':
while True:
main()
|
import time
import picamera
from settings import Job, IMAGES_DIRECTORY
def main():
job = Job()
if job.exists():
resolution_x = job.image_settings.resolution_x
resolution_y = job.image_settings.resolution_y
image_quality = job.image_settings.quality
snap_interval = job.snap_settings.interval
snap_total = job.snap_settings.total
with picamera.PiCamera() as camera:
camera.resolution = (resolution_x, resolution_y)
time.sleep(2)
output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
capture = camera.capture_continuous(output_file, quality=image_quality)
for i, _ in enumerate(capture):
if i == snap_total - 1:
job.archive()
break
time.sleep(snap_interval)
if __name__ == '__main__':
while True:
main()
| Check for job in main loop | Check for job in main loop
| Python | mit | projectweekend/Pi-Camera-Time-Lapse,projectweekend/Pi-Camera-Time-Lapse |
- import sys
import time
import picamera
+ from settings import Job, IMAGES_DIRECTORY
- import settings
- from settings import IMAGE, SNAP
- import uploader
def main():
+ job = Job()
+ if job.exists():
+ resolution_x = job.image_settings.resolution_x
+ resolution_y = job.image_settings.resolution_y
+ image_quality = job.image_settings.quality
+ snap_interval = job.snap_settings.interval
+ snap_total = job.snap_settings.total
- with picamera.PiCamera() as camera:
+ with picamera.PiCamera() as camera:
- camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y)
+ camera.resolution = (resolution_x, resolution_y)
- time.sleep(2)
+ time.sleep(2)
- output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
+ output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
- capture = camera.capture_continuous(output_file, quality=IMAGE.quality)
+ capture = camera.capture_continuous(output_file, quality=image_quality)
- for i, _ in enumerate(capture):
+ for i, _ in enumerate(capture):
- if i == SNAP.total - 1:
+ if i == snap_total - 1:
+ job.archive()
- break
+ break
- time.sleep(SNAP.interval)
+ time.sleep(snap_interval)
if __name__ == '__main__':
while True:
main()
| Check for job in main loop | ## Code Before:
import sys
import time
import picamera
import settings
from settings import IMAGE, SNAP
import uploader
def main():
with picamera.PiCamera() as camera:
camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y)
time.sleep(2)
output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
capture = camera.capture_continuous(output_file, quality=IMAGE.quality)
for i, _ in enumerate(capture):
if i == SNAP.total - 1:
break
time.sleep(SNAP.interval)
if __name__ == '__main__':
while True:
main()
## Instruction:
Check for job in main loop
## Code After:
import time
import picamera
from settings import Job, IMAGES_DIRECTORY
def main():
job = Job()
if job.exists():
resolution_x = job.image_settings.resolution_x
resolution_y = job.image_settings.resolution_y
image_quality = job.image_settings.quality
snap_interval = job.snap_settings.interval
snap_total = job.snap_settings.total
with picamera.PiCamera() as camera:
camera.resolution = (resolution_x, resolution_y)
time.sleep(2)
output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
capture = camera.capture_continuous(output_file, quality=image_quality)
for i, _ in enumerate(capture):
if i == snap_total - 1:
job.archive()
break
time.sleep(snap_interval)
if __name__ == '__main__':
while True:
main()
|
- import sys
import time
import picamera
+ from settings import Job, IMAGES_DIRECTORY
- import settings
- from settings import IMAGE, SNAP
- import uploader
def main():
+ job = Job()
+ if job.exists():
+ resolution_x = job.image_settings.resolution_x
+ resolution_y = job.image_settings.resolution_y
+ image_quality = job.image_settings.quality
+ snap_interval = job.snap_settings.interval
+ snap_total = job.snap_settings.total
- with picamera.PiCamera() as camera:
+ with picamera.PiCamera() as camera:
? ++++
- camera.resolution = (IMAGE.resolution_x, IMAGE.resolution_y)
? ------ ------
+ camera.resolution = (resolution_x, resolution_y)
? ++++
- time.sleep(2)
+ time.sleep(2)
? ++++
- output_file = settings.IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
? ---------
+ output_file = IMAGES_DIRECTORY + '/img{counter:03d}.jpg'
? ++++
- capture = camera.capture_continuous(output_file, quality=IMAGE.quality)
? ^^^^^^
+ capture = camera.capture_continuous(output_file, quality=image_quality)
? ++++ ^^^^^^
- for i, _ in enumerate(capture):
+ for i, _ in enumerate(capture):
? ++++
- if i == SNAP.total - 1:
? ^^^^^
+ if i == snap_total - 1:
? ++++ ^^^^^
+ job.archive()
- break
+ break
? ++++
- time.sleep(SNAP.interval)
? ^^^^^
+ time.sleep(snap_interval)
? ++++ ^^^^^
if __name__ == '__main__':
while True:
main() |
ebac72a3753205d3e45041c6db636a378187e3cf | pylua/tests/test_compiled.py | pylua/tests/test_compiled.py | import os
import subprocess
from pylua.tests.helpers import test_file
class TestCompiled(object):
"""
Tests compiled binary
"""
def test_addition(self, capsys):
f = test_file(src="""
-- short add
x = 10
y = 5
z = y + y + x
print(z)
print(z+y)
--a = 100+y
lx = 1234567890
ly = 99999999
print(lx+ly)
--print(lx+1234567890)
""", suffix=".l"
)
out = subprocess.check_output(['bin/pylua', f.name])
assert out == "20.000000\n25.000000\n1334567889.000000\n"
| import os
import subprocess
from pylua.tests.helpers import test_file
class TestCompiled(object):
"""
Tests compiled binary
"""
PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua'))
def test_addition(self, capsys):
f = test_file(src="""
-- short add
x = 10
y = 5
z = y + y + x
print(z)
print(z+y)
--a = 100+y
lx = 1234567890
ly = 99999999
print(lx+ly)
--print(lx+1234567890)
""", suffix=".l"
)
out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name])
assert out == "20.000000\n25.000000\n1334567889.000000\n"
| Use absolute path for lua binary in tests | Use absolute path for lua binary in tests
| Python | bsd-3-clause | fhahn/luna,fhahn/luna | import os
import subprocess
from pylua.tests.helpers import test_file
class TestCompiled(object):
"""
Tests compiled binary
"""
+
+ PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua'))
def test_addition(self, capsys):
f = test_file(src="""
-- short add
x = 10
y = 5
z = y + y + x
print(z)
print(z+y)
--a = 100+y
lx = 1234567890
ly = 99999999
print(lx+ly)
--print(lx+1234567890)
""", suffix=".l"
)
- out = subprocess.check_output(['bin/pylua', f.name])
+ out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name])
assert out == "20.000000\n25.000000\n1334567889.000000\n"
| Use absolute path for lua binary in tests | ## Code Before:
import os
import subprocess
from pylua.tests.helpers import test_file
class TestCompiled(object):
"""
Tests compiled binary
"""
def test_addition(self, capsys):
f = test_file(src="""
-- short add
x = 10
y = 5
z = y + y + x
print(z)
print(z+y)
--a = 100+y
lx = 1234567890
ly = 99999999
print(lx+ly)
--print(lx+1234567890)
""", suffix=".l"
)
out = subprocess.check_output(['bin/pylua', f.name])
assert out == "20.000000\n25.000000\n1334567889.000000\n"
## Instruction:
Use absolute path for lua binary in tests
## Code After:
import os
import subprocess
from pylua.tests.helpers import test_file
class TestCompiled(object):
"""
Tests compiled binary
"""
PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua'))
def test_addition(self, capsys):
f = test_file(src="""
-- short add
x = 10
y = 5
z = y + y + x
print(z)
print(z+y)
--a = 100+y
lx = 1234567890
ly = 99999999
print(lx+ly)
--print(lx+1234567890)
""", suffix=".l"
)
out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name])
assert out == "20.000000\n25.000000\n1334567889.000000\n"
| import os
import subprocess
from pylua.tests.helpers import test_file
class TestCompiled(object):
"""
Tests compiled binary
"""
+
+ PYLUA_BIN = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('../../bin/pylua'))
def test_addition(self, capsys):
f = test_file(src="""
-- short add
x = 10
y = 5
z = y + y + x
print(z)
print(z+y)
--a = 100+y
lx = 1234567890
ly = 99999999
print(lx+ly)
--print(lx+1234567890)
""", suffix=".l"
)
- out = subprocess.check_output(['bin/pylua', f.name])
? ^^ ---- ^^^
+ out = subprocess.check_output([TestCompiled.PYLUA_BIN, f.name])
? ^^^^^^^^ ^^^^^^^^^^^^
assert out == "20.000000\n25.000000\n1334567889.000000\n" |
46816c4d8470192e76e730969ddcedeb8391fdcf | setup.py | setup.py |
from distutils.core import setup
setup(name="Neighborhoodize",
version='0.9',
description='Utility for translating lat, long coordinates into '
'neighborhoods in various cities',
author='Brian Lange',
author_email='brian.lange@datascopeanalytics.com',
url='https://github.com/bjlange/neighborhoodize',
packages=['neighborhoodize', ],
package_data={'neighborhoodize': ['data/*']},
install_requires=[
"Shapely >= 1.5.7",
"beautifulsoup4 >= 4.3.2",
"fastkml >= 0.9"],
)
|
from distutils.core import setup
setup(name="Neighborhoodize",
version='0.9',
description='Utility for translating lat, long coordinates into '
'neighborhoods in various cities',
author='Brian Lange',
author_email='brian.lange@datascopeanalytics.com',
url='https://github.com/bjlange/neighborhoodize',
packages=['neighborhoodize', ],
package_data={'neighborhoodize': ['data/*']},
download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9',
install_requires=[
"Shapely >= 1.5.7",
"beautifulsoup4 >= 4.3.2",
"fastkml >= 0.9"],
)
| Add download url for pypi | Add download url for pypi
| Python | mit | bjlange/neighborhoodize |
from distutils.core import setup
setup(name="Neighborhoodize",
version='0.9',
description='Utility for translating lat, long coordinates into '
'neighborhoods in various cities',
author='Brian Lange',
author_email='brian.lange@datascopeanalytics.com',
url='https://github.com/bjlange/neighborhoodize',
packages=['neighborhoodize', ],
package_data={'neighborhoodize': ['data/*']},
+ download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9',
install_requires=[
"Shapely >= 1.5.7",
"beautifulsoup4 >= 4.3.2",
"fastkml >= 0.9"],
)
| Add download url for pypi | ## Code Before:
from distutils.core import setup
setup(name="Neighborhoodize",
version='0.9',
description='Utility for translating lat, long coordinates into '
'neighborhoods in various cities',
author='Brian Lange',
author_email='brian.lange@datascopeanalytics.com',
url='https://github.com/bjlange/neighborhoodize',
packages=['neighborhoodize', ],
package_data={'neighborhoodize': ['data/*']},
install_requires=[
"Shapely >= 1.5.7",
"beautifulsoup4 >= 4.3.2",
"fastkml >= 0.9"],
)
## Instruction:
Add download url for pypi
## Code After:
from distutils.core import setup
setup(name="Neighborhoodize",
version='0.9',
description='Utility for translating lat, long coordinates into '
'neighborhoods in various cities',
author='Brian Lange',
author_email='brian.lange@datascopeanalytics.com',
url='https://github.com/bjlange/neighborhoodize',
packages=['neighborhoodize', ],
package_data={'neighborhoodize': ['data/*']},
download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9',
install_requires=[
"Shapely >= 1.5.7",
"beautifulsoup4 >= 4.3.2",
"fastkml >= 0.9"],
)
|
from distutils.core import setup
setup(name="Neighborhoodize",
version='0.9',
description='Utility for translating lat, long coordinates into '
'neighborhoods in various cities',
author='Brian Lange',
author_email='brian.lange@datascopeanalytics.com',
url='https://github.com/bjlange/neighborhoodize',
packages=['neighborhoodize', ],
package_data={'neighborhoodize': ['data/*']},
+ download_url = 'https://github.com/bjlange/neighborhoodize/tarball/0.9',
install_requires=[
"Shapely >= 1.5.7",
"beautifulsoup4 >= 4.3.2",
"fastkml >= 0.9"],
) |
011949b266ab33df8c0f9bec29ba693824e7d8ef | setup.py | setup.py | import sys
from distutils.core import setup
setup(
name='pathlib',
version=open('VERSION.txt').read().strip(),
py_modules=['pathlib'],
license='MIT License',
description='Object-oriented filesystem paths',
long_description=open('README.txt').read(),
author='Antoine Pitrou',
author_email='solipsis@pitrou.net',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.2',
'Topic :: Software Development :: Libraries',
'Topic :: System :: Filesystems',
],
download_url='https://pypi.python.org/pypi/pathlib/',
url='http://readthedocs.org/docs/pathlib/',
)
| import sys
from distutils.core import setup
setup(
name='pathlib',
version=open('VERSION.txt').read().strip(),
py_modules=['pathlib'],
license='MIT License',
description='Object-oriented filesystem paths',
long_description=open('README.txt').read(),
author='Antoine Pitrou',
author_email='solipsis@pitrou.net',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Topic :: Software Development :: Libraries',
'Topic :: System :: Filesystems',
],
download_url='https://pypi.python.org/pypi/pathlib/',
url='http://readthedocs.org/docs/pathlib/',
)
| Add classifier for Python 3.3 | Add classifier for Python 3.3
| Python | mit | pombreda/pathlib | import sys
from distutils.core import setup
setup(
name='pathlib',
version=open('VERSION.txt').read().strip(),
py_modules=['pathlib'],
license='MIT License',
description='Object-oriented filesystem paths',
long_description=open('README.txt').read(),
author='Antoine Pitrou',
author_email='solipsis@pitrou.net',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.2',
+ 'Programming Language :: Python :: 3.3',
'Topic :: Software Development :: Libraries',
'Topic :: System :: Filesystems',
],
download_url='https://pypi.python.org/pypi/pathlib/',
url='http://readthedocs.org/docs/pathlib/',
)
| Add classifier for Python 3.3 | ## Code Before:
import sys
from distutils.core import setup
setup(
name='pathlib',
version=open('VERSION.txt').read().strip(),
py_modules=['pathlib'],
license='MIT License',
description='Object-oriented filesystem paths',
long_description=open('README.txt').read(),
author='Antoine Pitrou',
author_email='solipsis@pitrou.net',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.2',
'Topic :: Software Development :: Libraries',
'Topic :: System :: Filesystems',
],
download_url='https://pypi.python.org/pypi/pathlib/',
url='http://readthedocs.org/docs/pathlib/',
)
## Instruction:
Add classifier for Python 3.3
## Code After:
import sys
from distutils.core import setup
setup(
name='pathlib',
version=open('VERSION.txt').read().strip(),
py_modules=['pathlib'],
license='MIT License',
description='Object-oriented filesystem paths',
long_description=open('README.txt').read(),
author='Antoine Pitrou',
author_email='solipsis@pitrou.net',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Topic :: Software Development :: Libraries',
'Topic :: System :: Filesystems',
],
download_url='https://pypi.python.org/pypi/pathlib/',
url='http://readthedocs.org/docs/pathlib/',
)
| import sys
from distutils.core import setup
setup(
name='pathlib',
version=open('VERSION.txt').read().strip(),
py_modules=['pathlib'],
license='MIT License',
description='Object-oriented filesystem paths',
long_description=open('README.txt').read(),
author='Antoine Pitrou',
author_email='solipsis@pitrou.net',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.2',
+ 'Programming Language :: Python :: 3.3',
'Topic :: Software Development :: Libraries',
'Topic :: System :: Filesystems',
],
download_url='https://pypi.python.org/pypi/pathlib/',
url='http://readthedocs.org/docs/pathlib/',
) |
d4412f8573dbfc1b06f2a298cc5c3042c6c468e6 | tests/test_api.py | tests/test_api.py | from django.test import TestCase
from django_snooze import apis
class APITestCase(TestCase):
def setUp(self):
"""Sets up an API object to play with.
:returns: None
"""
self.api = apis.api
self.api.discover_models()
def test_apps(self):
"""Test if the right apps are present.
:returns: None
"""
self.assertIn('tests', self.api._resources.keys())
self.assertIn('auth', self.api._resources.keys())
| from django.test import TestCase
from django_snooze import apis
class APITestCase(TestCase):
def setUp(self):
"""Sets up an API object to play with.
:returns: None
"""
self.api = apis.api
self.api.discover_models()
def test_apps(self):
"""Test if the right apps are present.
:returns: None
"""
self.assertIn('tests', self.api._resources.keys())
self.assertIn('auth', self.api._resources.keys())
tests_models = [x.model_name for x in self.api._resources['tests']]
self.assertNotIn('abstract', tests_models)
| Test to see if abstract classes sneak in. | Test to see if abstract classes sneak in.
Now that get_models has been found to skip abstract classes, we want to test
for this in case this behaviour ever changes.
| Python | bsd-3-clause | ainmosni/django-snooze,ainmosni/django-snooze | from django.test import TestCase
from django_snooze import apis
class APITestCase(TestCase):
def setUp(self):
"""Sets up an API object to play with.
:returns: None
"""
self.api = apis.api
self.api.discover_models()
def test_apps(self):
"""Test if the right apps are present.
:returns: None
"""
self.assertIn('tests', self.api._resources.keys())
self.assertIn('auth', self.api._resources.keys())
+ tests_models = [x.model_name for x in self.api._resources['tests']]
+ self.assertNotIn('abstract', tests_models)
| Test to see if abstract classes sneak in. | ## Code Before:
from django.test import TestCase
from django_snooze import apis
class APITestCase(TestCase):
def setUp(self):
"""Sets up an API object to play with.
:returns: None
"""
self.api = apis.api
self.api.discover_models()
def test_apps(self):
"""Test if the right apps are present.
:returns: None
"""
self.assertIn('tests', self.api._resources.keys())
self.assertIn('auth', self.api._resources.keys())
## Instruction:
Test to see if abstract classes sneak in.
## Code After:
from django.test import TestCase
from django_snooze import apis
class APITestCase(TestCase):
def setUp(self):
"""Sets up an API object to play with.
:returns: None
"""
self.api = apis.api
self.api.discover_models()
def test_apps(self):
"""Test if the right apps are present.
:returns: None
"""
self.assertIn('tests', self.api._resources.keys())
self.assertIn('auth', self.api._resources.keys())
tests_models = [x.model_name for x in self.api._resources['tests']]
self.assertNotIn('abstract', tests_models)
| from django.test import TestCase
from django_snooze import apis
class APITestCase(TestCase):
def setUp(self):
"""Sets up an API object to play with.
:returns: None
"""
self.api = apis.api
self.api.discover_models()
def test_apps(self):
"""Test if the right apps are present.
:returns: None
"""
self.assertIn('tests', self.api._resources.keys())
self.assertIn('auth', self.api._resources.keys())
+ tests_models = [x.model_name for x in self.api._resources['tests']]
+ self.assertNotIn('abstract', tests_models) |
63a7b11d3ae51a944bf2e70637dea503e455c2f5 | fontdump/cli.py | fontdump/cli.py | from collections import OrderedDict
import requests
import cssutils
USER_AGENTS = OrderedDict()
USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
def main():
font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400'
sheets={}
for (format, ua) in USER_AGENTS.items():
headers = {
'User-Agent': ua,
}
r =requests.get(font_url, headers=headers)
sheets[format] = cssutils.parseString(r.content)
if __name__ == '__main__':
main() | import requests
import cssutils
USER_AGENTS = {
'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
}
def main():
font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400'
sheets={}
for (format, ua) in USER_AGENTS.items():
headers = {
'User-Agent': ua,
}
r =requests.get(font_url, headers=headers)
sheets[format] = cssutils.parseString(r.content)
if __name__ == '__main__':
main() | Revert "The order of the formats matters. Use OrderedDict instead of dict" | Revert "The order of the formats matters. Use OrderedDict instead of dict"
I can't rely on the order of dict. The control flow is more complex.
This reverts commit 3389ed71971ddacd185bbbf8fe667a8651108c70.
| Python | mit | glasslion/fontdump | - from collections import OrderedDict
-
import requests
import cssutils
- USER_AGENTS = OrderedDict()
+ USER_AGENTS = {
- USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
+ 'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
+ 'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
- USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
+ 'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
- USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
- USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
+ 'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
-
+ }
def main():
font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400'
sheets={}
for (format, ua) in USER_AGENTS.items():
headers = {
'User-Agent': ua,
}
r =requests.get(font_url, headers=headers)
sheets[format] = cssutils.parseString(r.content)
if __name__ == '__main__':
main() | Revert "The order of the formats matters. Use OrderedDict instead of dict" | ## Code Before:
from collections import OrderedDict
import requests
import cssutils
USER_AGENTS = OrderedDict()
USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
def main():
font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400'
sheets={}
for (format, ua) in USER_AGENTS.items():
headers = {
'User-Agent': ua,
}
r =requests.get(font_url, headers=headers)
sheets[format] = cssutils.parseString(r.content)
if __name__ == '__main__':
main()
## Instruction:
Revert "The order of the formats matters. Use OrderedDict instead of dict"
## Code After:
import requests
import cssutils
USER_AGENTS = {
'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
}
def main():
font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400'
sheets={}
for (format, ua) in USER_AGENTS.items():
headers = {
'User-Agent': ua,
}
r =requests.get(font_url, headers=headers)
sheets[format] = cssutils.parseString(r.content)
if __name__ == '__main__':
main() | - from collections import OrderedDict
-
import requests
import cssutils
- USER_AGENTS = OrderedDict()
+ USER_AGENTS = {
- USER_AGENTS['woff'] = 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
? ^^^^^^^^^^^^ ^^^
+ 'woff': 'Mozilla/5.0 AppleWebKit/537.36 Chrome/30', # Chrome
? ^^^^ ^
+ 'eot': 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
- USER_AGENTS['ttf'] = 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
? ^^^^^^^^^^^^ ^^^
+ 'ttf': 'Mozilla/5.0 (Linux; U; Android 2.1-update1;)', #Andord 2
? ^^^^ ^
- USER_AGENTS['eot'] = 'Mozilla/4.0 (compatible; MSIE 6.0;)', # IE6
- USER_AGENTS['woff'] = 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
? ^^^^^^^^^^^^ ^^^^ ^^^
+ 'svg': 'Mozilla/4.0 (iPad; CPU OS 4_0_1 ) AppleWebKit', #iOS<4.2
? ^^^^ ^^^ ^
-
+ }
def main():
font_url = 'http://fonts.googleapis.com/css?family=Open+Sans:300,400,700,800|Dosis:300,400'
sheets={}
for (format, ua) in USER_AGENTS.items():
headers = {
'User-Agent': ua,
}
r =requests.get(font_url, headers=headers)
sheets[format] = cssutils.parseString(r.content)
if __name__ == '__main__':
main() |
eb5294f0df32442dbd7431fd9200388ca4c63d62 | tests/builtins/test_reversed.py | tests/builtins/test_reversed.py | from .. utils import TranspileTestCase, BuiltinFunctionTestCase
class ReversedTests(TranspileTestCase):
pass
class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase):
functions = ["reversed"]
not_implemented = [
'test_range',
]
| from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase
def _iterate_test(datatype):
def test_func(self):
code = '\n'.join([
'\nfor x in {value}:\n print(x)\n'.format(value=value)
for value in SAMPLE_DATA[datatype]
])
self.assertCodeExecution(code)
return test_func
class ReversedTests(TranspileTestCase):
# test_iterate_bytearray = _iterate_test('bytearray')
test_iterate_bytes = _iterate_test('bytes')
test_iterate_list = _iterate_test('list')
test_iterate_range = _iterate_test('range')
test_iterate_str = _iterate_test('str')
test_iterate_tuple = _iterate_test('tuple')
class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase):
functions = ["reversed"]
not_implemented = [
'test_range',
]
| Add iteration tests for reversed type | Add iteration tests for reversed type
| Python | bsd-3-clause | cflee/voc,cflee/voc,freakboy3742/voc,freakboy3742/voc | - from .. utils import TranspileTestCase, BuiltinFunctionTestCase
+ from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase
+
+
+ def _iterate_test(datatype):
+
+ def test_func(self):
+ code = '\n'.join([
+ '\nfor x in {value}:\n print(x)\n'.format(value=value)
+ for value in SAMPLE_DATA[datatype]
+ ])
+ self.assertCodeExecution(code)
+
+ return test_func
class ReversedTests(TranspileTestCase):
- pass
+ # test_iterate_bytearray = _iterate_test('bytearray')
+ test_iterate_bytes = _iterate_test('bytes')
+ test_iterate_list = _iterate_test('list')
+ test_iterate_range = _iterate_test('range')
+ test_iterate_str = _iterate_test('str')
+ test_iterate_tuple = _iterate_test('tuple')
class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase):
functions = ["reversed"]
not_implemented = [
'test_range',
]
| Add iteration tests for reversed type | ## Code Before:
from .. utils import TranspileTestCase, BuiltinFunctionTestCase
class ReversedTests(TranspileTestCase):
pass
class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase):
functions = ["reversed"]
not_implemented = [
'test_range',
]
## Instruction:
Add iteration tests for reversed type
## Code After:
from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase
def _iterate_test(datatype):
def test_func(self):
code = '\n'.join([
'\nfor x in {value}:\n print(x)\n'.format(value=value)
for value in SAMPLE_DATA[datatype]
])
self.assertCodeExecution(code)
return test_func
class ReversedTests(TranspileTestCase):
# test_iterate_bytearray = _iterate_test('bytearray')
test_iterate_bytes = _iterate_test('bytes')
test_iterate_list = _iterate_test('list')
test_iterate_range = _iterate_test('range')
test_iterate_str = _iterate_test('str')
test_iterate_tuple = _iterate_test('tuple')
class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase):
functions = ["reversed"]
not_implemented = [
'test_range',
]
| - from .. utils import TranspileTestCase, BuiltinFunctionTestCase
+ from .. utils import SAMPLE_DATA, TranspileTestCase, BuiltinFunctionTestCase
? +++++++++++++
+
+
+ def _iterate_test(datatype):
+
+ def test_func(self):
+ code = '\n'.join([
+ '\nfor x in {value}:\n print(x)\n'.format(value=value)
+ for value in SAMPLE_DATA[datatype]
+ ])
+ self.assertCodeExecution(code)
+
+ return test_func
class ReversedTests(TranspileTestCase):
- pass
+ # test_iterate_bytearray = _iterate_test('bytearray')
+ test_iterate_bytes = _iterate_test('bytes')
+ test_iterate_list = _iterate_test('list')
+ test_iterate_range = _iterate_test('range')
+ test_iterate_str = _iterate_test('str')
+ test_iterate_tuple = _iterate_test('tuple')
class BuiltinReversedFunctionTests(BuiltinFunctionTestCase, TranspileTestCase):
functions = ["reversed"]
not_implemented = [
'test_range',
] |
4c1237d2969d735cfcf9f3c10cf27cb801996e32 | tests/test_integration.py | tests/test_integration.py | """Unit test module for Selenium testing"""
from selenium import webdriver
from flask.ext.testing import LiveServerTestCase
from tests import TestCase
from pages import LoginPage
class TestUI(TestCase, LiveServerTestCase):
"""Test class for UI integration/workflow testing"""
def setUp(self):
"""Reset all tables before testing."""
super(TestUI, self).setUp()
self.driver = webdriver.Firefox()
self.driver.implicitly_wait(60)
self.driver.root_uri = self.get_server_url()
def tearDown(self):
"""Clean db session, drop all tables."""
self.driver.quit()
super(TestUI, self).tearDown()
def test_login_page(self):
"""Ensure login page loads successfully"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text)
def test_login_form_facebook_exists(self):
"""Ensure Facebook button present on login form"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertIsNotNone(page.facebook_button)
| """Unit test module for Selenium testing"""
import os
from selenium import webdriver
from flask.ext.testing import LiveServerTestCase
from tests import TestCase
from pages import LoginPage
class TestUI(TestCase, LiveServerTestCase):
"""Test class for UI integration/workflow testing"""
def setUp(self):
"""Reset all tables before testing."""
super(TestUI, self).setUp()
if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ:
capabilities = {
"tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"],
"build": os.environ["TRAVIS_BUILD_NUMBER"],
"tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"],
}
url = "http://{username}:{access_key}@localhost:4445/wd/hub".format(
username=os.environ["SAUCE_USERNAME"],
access_key=os.environ["SAUCE_ACCESS_KEY"],
)
self.driver = webdriver.Remote(
desired_capabilities=capabilities,
command_executor=url
)
else:
self.driver = webdriver.Firefox()
self.driver.implicitly_wait(60)
self.driver.root_uri = self.get_server_url()
def tearDown(self):
"""Clean db session, drop all tables."""
self.driver.quit()
super(TestUI, self).tearDown()
def test_login_page(self):
"""Ensure login page loads successfully"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text)
def test_login_form_facebook_exists(self):
"""Ensure Facebook button present on login form"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertIsNotNone(page.facebook_button)
| Use Sauce Labs for selenium testing when available | Use Sauce Labs for selenium testing when available
| Python | bsd-3-clause | uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal | """Unit test module for Selenium testing"""
+ import os
from selenium import webdriver
from flask.ext.testing import LiveServerTestCase
from tests import TestCase
from pages import LoginPage
class TestUI(TestCase, LiveServerTestCase):
"""Test class for UI integration/workflow testing"""
def setUp(self):
"""Reset all tables before testing."""
super(TestUI, self).setUp()
+ if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ:
+
+ capabilities = {
+ "tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"],
+ "build": os.environ["TRAVIS_BUILD_NUMBER"],
+ "tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"],
+ }
+ url = "http://{username}:{access_key}@localhost:4445/wd/hub".format(
+ username=os.environ["SAUCE_USERNAME"],
+ access_key=os.environ["SAUCE_ACCESS_KEY"],
+ )
+
+ self.driver = webdriver.Remote(
+ desired_capabilities=capabilities,
+ command_executor=url
+ )
+
+ else:
- self.driver = webdriver.Firefox()
+ self.driver = webdriver.Firefox()
+
self.driver.implicitly_wait(60)
self.driver.root_uri = self.get_server_url()
def tearDown(self):
"""Clean db session, drop all tables."""
self.driver.quit()
super(TestUI, self).tearDown()
def test_login_page(self):
"""Ensure login page loads successfully"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text)
def test_login_form_facebook_exists(self):
"""Ensure Facebook button present on login form"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertIsNotNone(page.facebook_button)
| Use Sauce Labs for selenium testing when available | ## Code Before:
"""Unit test module for Selenium testing"""
from selenium import webdriver
from flask.ext.testing import LiveServerTestCase
from tests import TestCase
from pages import LoginPage
class TestUI(TestCase, LiveServerTestCase):
"""Test class for UI integration/workflow testing"""
def setUp(self):
"""Reset all tables before testing."""
super(TestUI, self).setUp()
self.driver = webdriver.Firefox()
self.driver.implicitly_wait(60)
self.driver.root_uri = self.get_server_url()
def tearDown(self):
"""Clean db session, drop all tables."""
self.driver.quit()
super(TestUI, self).tearDown()
def test_login_page(self):
"""Ensure login page loads successfully"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text)
def test_login_form_facebook_exists(self):
"""Ensure Facebook button present on login form"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertIsNotNone(page.facebook_button)
## Instruction:
Use Sauce Labs for selenium testing when available
## Code After:
"""Unit test module for Selenium testing"""
import os
from selenium import webdriver
from flask.ext.testing import LiveServerTestCase
from tests import TestCase
from pages import LoginPage
class TestUI(TestCase, LiveServerTestCase):
"""Test class for UI integration/workflow testing"""
def setUp(self):
"""Reset all tables before testing."""
super(TestUI, self).setUp()
if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ:
capabilities = {
"tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"],
"build": os.environ["TRAVIS_BUILD_NUMBER"],
"tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"],
}
url = "http://{username}:{access_key}@localhost:4445/wd/hub".format(
username=os.environ["SAUCE_USERNAME"],
access_key=os.environ["SAUCE_ACCESS_KEY"],
)
self.driver = webdriver.Remote(
desired_capabilities=capabilities,
command_executor=url
)
else:
self.driver = webdriver.Firefox()
self.driver.implicitly_wait(60)
self.driver.root_uri = self.get_server_url()
def tearDown(self):
"""Clean db session, drop all tables."""
self.driver.quit()
super(TestUI, self).tearDown()
def test_login_page(self):
"""Ensure login page loads successfully"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text)
def test_login_form_facebook_exists(self):
"""Ensure Facebook button present on login form"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertIsNotNone(page.facebook_button)
| """Unit test module for Selenium testing"""
+ import os
from selenium import webdriver
from flask.ext.testing import LiveServerTestCase
from tests import TestCase
from pages import LoginPage
class TestUI(TestCase, LiveServerTestCase):
"""Test class for UI integration/workflow testing"""
def setUp(self):
"""Reset all tables before testing."""
super(TestUI, self).setUp()
+ if "SAUCE_USERNAME" in os.environ and "SAUCE_ACCESS_KEY" in os.environ:
+
+ capabilities = {
+ "tunnel-identifier": os.environ["TRAVIS_JOB_NUMBER"],
+ "build": os.environ["TRAVIS_BUILD_NUMBER"],
+ "tags": [os.environ["TRAVIS_PYTHON_VERSION"], "CI"],
+ }
+ url = "http://{username}:{access_key}@localhost:4445/wd/hub".format(
+ username=os.environ["SAUCE_USERNAME"],
+ access_key=os.environ["SAUCE_ACCESS_KEY"],
+ )
+
+ self.driver = webdriver.Remote(
+ desired_capabilities=capabilities,
+ command_executor=url
+ )
+
+ else:
- self.driver = webdriver.Firefox()
+ self.driver = webdriver.Firefox()
? ++++
+
self.driver.implicitly_wait(60)
self.driver.root_uri = self.get_server_url()
def tearDown(self):
"""Clean db session, drop all tables."""
self.driver.quit()
super(TestUI, self).tearDown()
def test_login_page(self):
"""Ensure login page loads successfully"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertNotIn("Uh-oh", page.w.find_element_by_tag_name("body").text)
def test_login_form_facebook_exists(self):
"""Ensure Facebook button present on login form"""
page = LoginPage(self.driver)
page.get("/user/sign-in")
self.assertIsNotNone(page.facebook_button) |
9f484e6eb4fcf37f5515d62eb80be6a0f8d5a097 | swteams/views.py | swteams/views.py | import iris.views
from teams.models import Team
def activity(request, slug, *args, **kw):
template_name = 'teams/activity.html'
team = get_object_or_404(Team, slug=slug)
template_context = {
'group': team,
}
return render_to_response(template_name, template_context, RequestContext(request))
def topics_with_team_slug(*args, **kwargs):
# Insert the team into the template context as 'group' so that
# breadcrumbs, etc. render properly.
slug = kwargs['slug']
extra_context = kwargs.setdefault('extra_context', {})
extra_context['group'] = Team.objects.get(slug=slug)
return iris.views.topics(*args, **kwargs)
| from django.shortcuts import get_object_or_404, render_to_response
from django.template.context import RequestContext
import iris.views
from teams.models import Team
def activity(request, slug, *args, **kw):
template_name = 'teams/activity.html'
team = get_object_or_404(Team, slug=slug)
template_context = {
'group': team,
}
return render_to_response(template_name, template_context, RequestContext(request))
def topics_with_team_slug(*args, **kwargs):
# Insert the team into the template context as 'group' so that
# breadcrumbs, etc. render properly.
slug = kwargs['slug']
extra_context = kwargs.setdefault('extra_context', {})
extra_context['group'] = Team.objects.get(slug=slug)
return iris.views.topics(*args, **kwargs)
| Fix team activity history view (was missing imports) | Fix team activity history view (was missing imports)
| Python | apache-2.0 | snswa/swsites,snswa/swsites,snswa/swsites | + from django.shortcuts import get_object_or_404, render_to_response
+ from django.template.context import RequestContext
+
import iris.views
from teams.models import Team
def activity(request, slug, *args, **kw):
template_name = 'teams/activity.html'
team = get_object_or_404(Team, slug=slug)
template_context = {
'group': team,
}
return render_to_response(template_name, template_context, RequestContext(request))
def topics_with_team_slug(*args, **kwargs):
# Insert the team into the template context as 'group' so that
# breadcrumbs, etc. render properly.
slug = kwargs['slug']
extra_context = kwargs.setdefault('extra_context', {})
extra_context['group'] = Team.objects.get(slug=slug)
return iris.views.topics(*args, **kwargs)
| Fix team activity history view (was missing imports) | ## Code Before:
import iris.views
from teams.models import Team
def activity(request, slug, *args, **kw):
template_name = 'teams/activity.html'
team = get_object_or_404(Team, slug=slug)
template_context = {
'group': team,
}
return render_to_response(template_name, template_context, RequestContext(request))
def topics_with_team_slug(*args, **kwargs):
# Insert the team into the template context as 'group' so that
# breadcrumbs, etc. render properly.
slug = kwargs['slug']
extra_context = kwargs.setdefault('extra_context', {})
extra_context['group'] = Team.objects.get(slug=slug)
return iris.views.topics(*args, **kwargs)
## Instruction:
Fix team activity history view (was missing imports)
## Code After:
from django.shortcuts import get_object_or_404, render_to_response
from django.template.context import RequestContext
import iris.views
from teams.models import Team
def activity(request, slug, *args, **kw):
template_name = 'teams/activity.html'
team = get_object_or_404(Team, slug=slug)
template_context = {
'group': team,
}
return render_to_response(template_name, template_context, RequestContext(request))
def topics_with_team_slug(*args, **kwargs):
# Insert the team into the template context as 'group' so that
# breadcrumbs, etc. render properly.
slug = kwargs['slug']
extra_context = kwargs.setdefault('extra_context', {})
extra_context['group'] = Team.objects.get(slug=slug)
return iris.views.topics(*args, **kwargs)
| + from django.shortcuts import get_object_or_404, render_to_response
+ from django.template.context import RequestContext
+
import iris.views
from teams.models import Team
def activity(request, slug, *args, **kw):
template_name = 'teams/activity.html'
team = get_object_or_404(Team, slug=slug)
template_context = {
'group': team,
}
return render_to_response(template_name, template_context, RequestContext(request))
def topics_with_team_slug(*args, **kwargs):
# Insert the team into the template context as 'group' so that
# breadcrumbs, etc. render properly.
slug = kwargs['slug']
extra_context = kwargs.setdefault('extra_context', {})
extra_context['group'] = Team.objects.get(slug=slug)
return iris.views.topics(*args, **kwargs) |
f0871086222be67ba4add4574cf03d3cb9d39d63 | tests/circular/template/test_interpolatedstr.py | tests/circular/template/test_interpolatedstr.py | from src.circular.template.interpolatedstr import InterpolatedStr
from src.circular.template.context import Context
from tests.utils import TObserver
def test_string_interp():
ctx = Context()
ctx.name = "James"
s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.")
s.bind_ctx(ctx)
t = TObserver(s)
assert s.value == "My name is , James ."
ctx.surname = "Bond"
data = t.events.pop().data
assert s.value == "My name is Bond, James Bond."
| from src.circular.template.interpolatedstr import InterpolatedStr
from src.circular.template.context import Context
from tests.utils import TObserver
def test_string_interp():
ctx = Context()
ctx.name = "James"
s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.")
s.bind_ctx(ctx)
t = TObserver(s)
assert s.value == "My name is , James ."
ctx.surname = "Bond"
data = t.events.pop().data
assert s.value == "My name is Bond, James Bond."
# Should correctly interpolate two immediately succeeding expressions
ctx.sur="B"
s = InterpolatedStr('{{name}}{{sur}}')
s.bind_ctx(ctx)
assert s.value == "JamesB"
| Add test for run-together interpolated expressions. | Add test for run-together interpolated expressions.
| Python | mit | jonathanverner/circular,jonathanverner/circular,jonathanverner/circular | from src.circular.template.interpolatedstr import InterpolatedStr
from src.circular.template.context import Context
from tests.utils import TObserver
def test_string_interp():
ctx = Context()
ctx.name = "James"
s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.")
s.bind_ctx(ctx)
t = TObserver(s)
assert s.value == "My name is , James ."
ctx.surname = "Bond"
data = t.events.pop().data
assert s.value == "My name is Bond, James Bond."
+ # Should correctly interpolate two immediately succeeding expressions
+ ctx.sur="B"
+ s = InterpolatedStr('{{name}}{{sur}}')
+ s.bind_ctx(ctx)
+ assert s.value == "JamesB"
+ | Add test for run-together interpolated expressions. | ## Code Before:
from src.circular.template.interpolatedstr import InterpolatedStr
from src.circular.template.context import Context
from tests.utils import TObserver
def test_string_interp():
ctx = Context()
ctx.name = "James"
s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.")
s.bind_ctx(ctx)
t = TObserver(s)
assert s.value == "My name is , James ."
ctx.surname = "Bond"
data = t.events.pop().data
assert s.value == "My name is Bond, James Bond."
## Instruction:
Add test for run-together interpolated expressions.
## Code After:
from src.circular.template.interpolatedstr import InterpolatedStr
from src.circular.template.context import Context
from tests.utils import TObserver
def test_string_interp():
ctx = Context()
ctx.name = "James"
s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.")
s.bind_ctx(ctx)
t = TObserver(s)
assert s.value == "My name is , James ."
ctx.surname = "Bond"
data = t.events.pop().data
assert s.value == "My name is Bond, James Bond."
# Should correctly interpolate two immediately succeeding expressions
ctx.sur="B"
s = InterpolatedStr('{{name}}{{sur}}')
s.bind_ctx(ctx)
assert s.value == "JamesB"
| from src.circular.template.interpolatedstr import InterpolatedStr
from src.circular.template.context import Context
from tests.utils import TObserver
def test_string_interp():
ctx = Context()
ctx.name = "James"
s = InterpolatedStr("My name is {{ surname }}, {{name}} {{ surname}}.")
s.bind_ctx(ctx)
t = TObserver(s)
assert s.value == "My name is , James ."
ctx.surname = "Bond"
data = t.events.pop().data
assert s.value == "My name is Bond, James Bond."
+ # Should correctly interpolate two immediately succeeding expressions
+ ctx.sur="B"
+ s = InterpolatedStr('{{name}}{{sur}}')
+ s.bind_ctx(ctx)
+ assert s.value == "JamesB"
+ |
a287c1e7a6e96a2a2143e9270a5f9b2ec295022e | fireplace/cards/removed/all.py | fireplace/cards/removed/all.py |
from ..utils import *
# Adrenaline Rush
class NEW1_006:
action = drawCard
combo = drawCards(2)
# Bolstered (Bloodsail Corsair)
class NEW1_025e:
Health = 1
|
from ..utils import *
# Dagger Mastery
class CS2_083:
def action(self):
if self.hero.weapon:
self.hero.weapon.buff("CS2_083e")
else:
self.hero.summon("CS2_082")
class CS2_083e:
Atk = 1
# Adrenaline Rush
class NEW1_006:
action = drawCard
combo = drawCards(2)
# Bolstered (Bloodsail Corsair)
class NEW1_025e:
Health = 1
| Implement the old Dagger Mastery | Implement the old Dagger Mastery
| Python | agpl-3.0 | Ragowit/fireplace,beheh/fireplace,NightKev/fireplace,liujimj/fireplace,amw2104/fireplace,butozerca/fireplace,oftc-ftw/fireplace,jleclanche/fireplace,liujimj/fireplace,butozerca/fireplace,oftc-ftw/fireplace,smallnamespace/fireplace,Meerkov/fireplace,smallnamespace/fireplace,amw2104/fireplace,Ragowit/fireplace,Meerkov/fireplace |
from ..utils import *
+
+
+ # Dagger Mastery
+ class CS2_083:
+ def action(self):
+ if self.hero.weapon:
+ self.hero.weapon.buff("CS2_083e")
+ else:
+ self.hero.summon("CS2_082")
+
+ class CS2_083e:
+ Atk = 1
# Adrenaline Rush
class NEW1_006:
action = drawCard
combo = drawCards(2)
# Bolstered (Bloodsail Corsair)
class NEW1_025e:
Health = 1
| Implement the old Dagger Mastery | ## Code Before:
from ..utils import *
# Adrenaline Rush
class NEW1_006:
action = drawCard
combo = drawCards(2)
# Bolstered (Bloodsail Corsair)
class NEW1_025e:
Health = 1
## Instruction:
Implement the old Dagger Mastery
## Code After:
from ..utils import *
# Dagger Mastery
class CS2_083:
def action(self):
if self.hero.weapon:
self.hero.weapon.buff("CS2_083e")
else:
self.hero.summon("CS2_082")
class CS2_083e:
Atk = 1
# Adrenaline Rush
class NEW1_006:
action = drawCard
combo = drawCards(2)
# Bolstered (Bloodsail Corsair)
class NEW1_025e:
Health = 1
|
from ..utils import *
+
+
+ # Dagger Mastery
+ class CS2_083:
+ def action(self):
+ if self.hero.weapon:
+ self.hero.weapon.buff("CS2_083e")
+ else:
+ self.hero.summon("CS2_082")
+
+ class CS2_083e:
+ Atk = 1
# Adrenaline Rush
class NEW1_006:
action = drawCard
combo = drawCards(2)
# Bolstered (Bloodsail Corsair)
class NEW1_025e:
Health = 1 |
07ccbc36fd5148db2efc5f676fd13d4b24aa004f | hackasmlexer/hacklexer.py | hackasmlexer/hacklexer.py | import re
from pygments.lexer import RegexLexer, include
from pygments.token import *
class HackAsmLexer(RegexLexer):
name = 'Hack Assembler'
aliases = ['hack_asm']
filenames = ['*.asm']
identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*'
flags = re.IGNORECASE | re.MULTILINE
tokens = {
'root': [
include('whitespace'),
(r'\(' + identifier + '\)', Name.Label),
(r'[+-=;&|!]+', Operator),
(r'\/\/.+$', Comment),
(r'[\r\n]+', Text),
(r'@[A-Za-z][A-Za-z0-9]+', Name.Variable),
(r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword),
(r'null', Keyword.Pseudo),
(r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin),
(r'@[0-9]+', Name.Constant)
],
'whitespace': [
(r'\n', Text),
(r'\s+', Text),
(r'#.*?\n', Comment)
]
}
| import re
from pygments.lexer import RegexLexer, include
from pygments.token import *
class HackAsmLexer(RegexLexer):
name = 'Hack Assembler'
aliases = ['hack_asm']
filenames = ['*.asm']
identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*'
flags = re.IGNORECASE | re.MULTILINE
tokens = {
'root': [
include('whitespace'),
(r'\(' + identifier + '\)', Name.Label),
(r'[+-=;&|!]+', Operator),
(r'\/\/.+$', Comment),
(r'[\r\n]+', Text),
(r'@[A-Za-z][A-Za-z0-9]+', Name.Variable),
(r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword),
(r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses
(r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses
(r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses
(r'null', Keyword.Pseudo),
(r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin),
(r'@[0-9]+', Name.Constant)
],
'whitespace': [
(r'\n', Text),
(r'\s+', Text),
(r'#.*?\n', Comment)
]
}
| Add register and IO addresses | Add register and IO addresses
| Python | mit | cprieto/pygments_hack_asm | import re
from pygments.lexer import RegexLexer, include
from pygments.token import *
class HackAsmLexer(RegexLexer):
name = 'Hack Assembler'
aliases = ['hack_asm']
filenames = ['*.asm']
identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*'
flags = re.IGNORECASE | re.MULTILINE
tokens = {
'root': [
include('whitespace'),
(r'\(' + identifier + '\)', Name.Label),
(r'[+-=;&|!]+', Operator),
(r'\/\/.+$', Comment),
(r'[\r\n]+', Text),
(r'@[A-Za-z][A-Za-z0-9]+', Name.Variable),
(r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword),
+ (r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses
+ (r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses
+ (r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses
(r'null', Keyword.Pseudo),
(r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin),
(r'@[0-9]+', Name.Constant)
],
'whitespace': [
(r'\n', Text),
(r'\s+', Text),
(r'#.*?\n', Comment)
]
}
| Add register and IO addresses | ## Code Before:
import re
from pygments.lexer import RegexLexer, include
from pygments.token import *
class HackAsmLexer(RegexLexer):
name = 'Hack Assembler'
aliases = ['hack_asm']
filenames = ['*.asm']
identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*'
flags = re.IGNORECASE | re.MULTILINE
tokens = {
'root': [
include('whitespace'),
(r'\(' + identifier + '\)', Name.Label),
(r'[+-=;&|!]+', Operator),
(r'\/\/.+$', Comment),
(r'[\r\n]+', Text),
(r'@[A-Za-z][A-Za-z0-9]+', Name.Variable),
(r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword),
(r'null', Keyword.Pseudo),
(r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin),
(r'@[0-9]+', Name.Constant)
],
'whitespace': [
(r'\n', Text),
(r'\s+', Text),
(r'#.*?\n', Comment)
]
}
## Instruction:
Add register and IO addresses
## Code After:
import re
from pygments.lexer import RegexLexer, include
from pygments.token import *
class HackAsmLexer(RegexLexer):
name = 'Hack Assembler'
aliases = ['hack_asm']
filenames = ['*.asm']
identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*'
flags = re.IGNORECASE | re.MULTILINE
tokens = {
'root': [
include('whitespace'),
(r'\(' + identifier + '\)', Name.Label),
(r'[+-=;&|!]+', Operator),
(r'\/\/.+$', Comment),
(r'[\r\n]+', Text),
(r'@[A-Za-z][A-Za-z0-9]+', Name.Variable),
(r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword),
(r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses
(r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses
(r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses
(r'null', Keyword.Pseudo),
(r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin),
(r'@[0-9]+', Name.Constant)
],
'whitespace': [
(r'\n', Text),
(r'\s+', Text),
(r'#.*?\n', Comment)
]
}
| import re
from pygments.lexer import RegexLexer, include
from pygments.token import *
class HackAsmLexer(RegexLexer):
name = 'Hack Assembler'
aliases = ['hack_asm']
filenames = ['*.asm']
identifier = r'[a-zA-Z$._?][a-zA-Z0-9$._?]*'
flags = re.IGNORECASE | re.MULTILINE
tokens = {
'root': [
include('whitespace'),
(r'\(' + identifier + '\)', Name.Label),
(r'[+-=;&|!]+', Operator),
(r'\/\/.+$', Comment),
(r'[\r\n]+', Text),
(r'@[A-Za-z][A-Za-z0-9]+', Name.Variable),
(r'\b(JGT|JEQ|JGE|JLT|JNE|JLE|JMP)\b', Keyword),
+ (r'\b@(SCREEN|KBD)\b', Name.Builtin.Pseudo), # I/O addresses
+ (r'\b@(R0|R1|R2|R3|R4|R5|R6|R7|R8|R9|R10|R11|R12|R13|R14|R15)\b', Name.Builtin.Pseudo), # RAM Addresses
+ (r'\b@(SP|LCL|ARG|THIS|THAT)\b', Name.Builtin.Pseudo), # Parameter addresses
(r'null', Keyword.Pseudo),
(r'\b(D|M|MD|A|AM|AD|AMD)\b', Name.Builtin),
(r'@[0-9]+', Name.Constant)
],
'whitespace': [
(r'\n', Text),
(r'\s+', Text),
(r'#.*?\n', Comment)
]
} |
8570efd42f35b89d9a97d9aa5a5aa47765cd21f6 | diary/logthread.py | diary/logthread.py | from threading import Thread
try:
from queue import Queue
except ImportError: # python 2
from Queue import Queue
class ElemThread(Thread):
"""A thread for logging as to not disrupt the logged application"""
def __init__(self, elem, name="Elementary Logger"):
"""Construct a thread for logging
:param elem: An Elementary instance to handle logging
:param name: A string to represent this thread
"""
Thread.__init__(self, name=name)
self.daemon = True # py2 constructor requires explicit
self.elem = elem
self.queue = Queue()
self.start()
def add(self, event):
"""Add a logged event to queue for logging"""
self.queue.put(event)
def run(self):
"""Main for thread to run"""
while True:
self.elem.write(self.queue.get())
| from threading import Thread
try:
from queue import Queue
except ImportError: # python 2
from Queue import Queue
class DiaryThread(Thread):
"""A thread for logging as to not disrupt the logged application"""
def __init__(self, diary, name="Diary Logger"):
"""Construct a thread for logging
:param diary: An Diary instance to handle logging
:param name: A string to represent this thread
"""
Thread.__init__(self, name=name)
self.daemon = True # py2 constructor requires explicit
self.diary = diary
self.queue = Queue()
self.start()
def add(self, event):
"""Add a logged event to queue for logging"""
self.queue.put(event)
def run(self):
"""Main for thread to run"""
while True:
self.diary.write(self.queue.get())
| Make last changes over to diary name | Make last changes over to diary name
| Python | mit | GreenVars/diary | from threading import Thread
try:
from queue import Queue
except ImportError: # python 2
from Queue import Queue
- class ElemThread(Thread):
+ class DiaryThread(Thread):
"""A thread for logging as to not disrupt the logged application"""
- def __init__(self, elem, name="Elementary Logger"):
+ def __init__(self, diary, name="Diary Logger"):
"""Construct a thread for logging
- :param elem: An Elementary instance to handle logging
+ :param diary: An Diary instance to handle logging
:param name: A string to represent this thread
"""
Thread.__init__(self, name=name)
self.daemon = True # py2 constructor requires explicit
- self.elem = elem
+ self.diary = diary
self.queue = Queue()
self.start()
def add(self, event):
"""Add a logged event to queue for logging"""
self.queue.put(event)
def run(self):
"""Main for thread to run"""
while True:
- self.elem.write(self.queue.get())
+ self.diary.write(self.queue.get())
| Make last changes over to diary name | ## Code Before:
from threading import Thread
try:
from queue import Queue
except ImportError: # python 2
from Queue import Queue
class ElemThread(Thread):
"""A thread for logging as to not disrupt the logged application"""
def __init__(self, elem, name="Elementary Logger"):
"""Construct a thread for logging
:param elem: An Elementary instance to handle logging
:param name: A string to represent this thread
"""
Thread.__init__(self, name=name)
self.daemon = True # py2 constructor requires explicit
self.elem = elem
self.queue = Queue()
self.start()
def add(self, event):
"""Add a logged event to queue for logging"""
self.queue.put(event)
def run(self):
"""Main for thread to run"""
while True:
self.elem.write(self.queue.get())
## Instruction:
Make last changes over to diary name
## Code After:
from threading import Thread
try:
from queue import Queue
except ImportError: # python 2
from Queue import Queue
class DiaryThread(Thread):
"""A thread for logging as to not disrupt the logged application"""
def __init__(self, diary, name="Diary Logger"):
"""Construct a thread for logging
:param diary: An Diary instance to handle logging
:param name: A string to represent this thread
"""
Thread.__init__(self, name=name)
self.daemon = True # py2 constructor requires explicit
self.diary = diary
self.queue = Queue()
self.start()
def add(self, event):
"""Add a logged event to queue for logging"""
self.queue.put(event)
def run(self):
"""Main for thread to run"""
while True:
self.diary.write(self.queue.get())
| from threading import Thread
try:
from queue import Queue
except ImportError: # python 2
from Queue import Queue
- class ElemThread(Thread):
? ^^^^
+ class DiaryThread(Thread):
? ^^^^^
"""A thread for logging as to not disrupt the logged application"""
- def __init__(self, elem, name="Elementary Logger"):
? ^^^^ ^^^^^^^
+ def __init__(self, diary, name="Diary Logger"):
? ^^^^^ ^^
"""Construct a thread for logging
- :param elem: An Elementary instance to handle logging
? ^^^^ ^^^^^^^
+ :param diary: An Diary instance to handle logging
? ^^^^^ ^^
:param name: A string to represent this thread
"""
Thread.__init__(self, name=name)
self.daemon = True # py2 constructor requires explicit
- self.elem = elem
+ self.diary = diary
self.queue = Queue()
self.start()
def add(self, event):
"""Add a logged event to queue for logging"""
self.queue.put(event)
def run(self):
"""Main for thread to run"""
while True:
- self.elem.write(self.queue.get())
? ^^^^
+ self.diary.write(self.queue.get())
? ^^^^^
|
59daf205869c42b3797aa9dbaaa97930cbca2417 | nanshe_workflow/ipy.py | nanshe_workflow/ipy.py | __author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>"
__date__ = "$Nov 10, 2015 17:09$"
try:
from IPython.utils.shimmodule import ShimWarning
except ImportError:
class ShimWarning(Warning):
"""Warning issued by IPython 4.x regarding deprecated API."""
pass
import warnings
with warnings.catch_warnings():
warnings.filterwarnings('error', '', ShimWarning)
try:
# IPython 3
from IPython.html.widgets import FloatProgress
from IPython.parallel import Client
except ShimWarning:
# IPython 4
from ipywidgets import FloatProgress
from ipyparallel import Client
from IPython.display import display
| __author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>"
__date__ = "$Nov 10, 2015 17:09$"
import json
import re
try:
from IPython.utils.shimmodule import ShimWarning
except ImportError:
class ShimWarning(Warning):
"""Warning issued by IPython 4.x regarding deprecated API."""
pass
import warnings
with warnings.catch_warnings():
warnings.filterwarnings('error', '', ShimWarning)
try:
# IPython 3
from IPython.html.widgets import FloatProgress
from IPython.parallel import Client
except ShimWarning:
# IPython 4
from ipywidgets import FloatProgress
from ipyparallel import Client
from IPython.display import display
import ipykernel
import notebook.notebookapp
import requests
def check_nbserverproxy():
"""
Return the url of the current jupyter notebook server.
"""
kernel_id = re.search(
"kernel-(.*).json",
ipykernel.connect.get_connection_file()
).group(1)
servers = notebook.notebookapp.list_running_servers()
for s in servers:
response = requests.get(
requests.compat.urljoin(s["url"], "api/sessions"),
params={"token": s.get("token", "")}
)
for n in json.loads(response.text):
if n["kernel"]["id"] == kernel_id:
# Found server that is running this Jupyter Notebook.
# Try to requests this servers port through nbserverproxy.
url = requests.compat.urljoin(
s["url"], "proxy/%i" % s["port"]
)
# If the proxy is running, it will redirect.
# If not, it will error out.
try:
requests.get(url).raise_for_status()
except requests.HTTPError:
return False
else:
return True
| Add function to check if nbserverproxy is running | Add function to check if nbserverproxy is running
Provides a simple check to see if the `nbserverproxy` is installed and
running. As this is a Jupyter server extension and this code is run from
the notebook, we can't simply import `nbserverproxy`. In fact that
wouldn't even work when using the Python 2 kernel even though the proxy
server could be running.
Instead to solve this problem try to identify the Jupyter Notebook
server we are running under. Once identified, attempt to query the proxy
server with the port of the Jupyter Notebook server. If the proxy server
is running, this will merely redirect to the Jupyter Notebook server and
return an HTTP 200 status. However if the proxy server is not running,
this will return a HTTP 404 error. There may be other errors that it
could raise. In any event, if the proxy redirects us, we know it is
working and if not we know it doesn't work.
| Python | apache-2.0 | nanshe-org/nanshe_workflow,DudLab/nanshe_workflow | __author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>"
__date__ = "$Nov 10, 2015 17:09$"
+
+ import json
+ import re
try:
from IPython.utils.shimmodule import ShimWarning
except ImportError:
class ShimWarning(Warning):
"""Warning issued by IPython 4.x regarding deprecated API."""
pass
import warnings
with warnings.catch_warnings():
warnings.filterwarnings('error', '', ShimWarning)
try:
# IPython 3
from IPython.html.widgets import FloatProgress
from IPython.parallel import Client
except ShimWarning:
# IPython 4
from ipywidgets import FloatProgress
from ipyparallel import Client
from IPython.display import display
+ import ipykernel
+ import notebook.notebookapp
+
+ import requests
+
+
+ def check_nbserverproxy():
+ """
+ Return the url of the current jupyter notebook server.
+ """
+ kernel_id = re.search(
+ "kernel-(.*).json",
+ ipykernel.connect.get_connection_file()
+ ).group(1)
+ servers = notebook.notebookapp.list_running_servers()
+ for s in servers:
+ response = requests.get(
+ requests.compat.urljoin(s["url"], "api/sessions"),
+ params={"token": s.get("token", "")}
+ )
+ for n in json.loads(response.text):
+ if n["kernel"]["id"] == kernel_id:
+ # Found server that is running this Jupyter Notebook.
+ # Try to requests this servers port through nbserverproxy.
+ url = requests.compat.urljoin(
+ s["url"], "proxy/%i" % s["port"]
+ )
+ # If the proxy is running, it will redirect.
+ # If not, it will error out.
+ try:
+ requests.get(url).raise_for_status()
+ except requests.HTTPError:
+ return False
+ else:
+ return True
+ | Add function to check if nbserverproxy is running | ## Code Before:
__author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>"
__date__ = "$Nov 10, 2015 17:09$"
try:
from IPython.utils.shimmodule import ShimWarning
except ImportError:
class ShimWarning(Warning):
"""Warning issued by IPython 4.x regarding deprecated API."""
pass
import warnings
with warnings.catch_warnings():
warnings.filterwarnings('error', '', ShimWarning)
try:
# IPython 3
from IPython.html.widgets import FloatProgress
from IPython.parallel import Client
except ShimWarning:
# IPython 4
from ipywidgets import FloatProgress
from ipyparallel import Client
from IPython.display import display
## Instruction:
Add function to check if nbserverproxy is running
## Code After:
__author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>"
__date__ = "$Nov 10, 2015 17:09$"
import json
import re
try:
from IPython.utils.shimmodule import ShimWarning
except ImportError:
class ShimWarning(Warning):
"""Warning issued by IPython 4.x regarding deprecated API."""
pass
import warnings
with warnings.catch_warnings():
warnings.filterwarnings('error', '', ShimWarning)
try:
# IPython 3
from IPython.html.widgets import FloatProgress
from IPython.parallel import Client
except ShimWarning:
# IPython 4
from ipywidgets import FloatProgress
from ipyparallel import Client
from IPython.display import display
import ipykernel
import notebook.notebookapp
import requests
def check_nbserverproxy():
"""
Return the url of the current jupyter notebook server.
"""
kernel_id = re.search(
"kernel-(.*).json",
ipykernel.connect.get_connection_file()
).group(1)
servers = notebook.notebookapp.list_running_servers()
for s in servers:
response = requests.get(
requests.compat.urljoin(s["url"], "api/sessions"),
params={"token": s.get("token", "")}
)
for n in json.loads(response.text):
if n["kernel"]["id"] == kernel_id:
# Found server that is running this Jupyter Notebook.
# Try to requests this servers port through nbserverproxy.
url = requests.compat.urljoin(
s["url"], "proxy/%i" % s["port"]
)
# If the proxy is running, it will redirect.
# If not, it will error out.
try:
requests.get(url).raise_for_status()
except requests.HTTPError:
return False
else:
return True
| __author__ = "John Kirkham <kirkhamj@janelia.hhmi.org>"
__date__ = "$Nov 10, 2015 17:09$"
+
+ import json
+ import re
try:
from IPython.utils.shimmodule import ShimWarning
except ImportError:
class ShimWarning(Warning):
"""Warning issued by IPython 4.x regarding deprecated API."""
pass
import warnings
with warnings.catch_warnings():
warnings.filterwarnings('error', '', ShimWarning)
try:
# IPython 3
from IPython.html.widgets import FloatProgress
from IPython.parallel import Client
except ShimWarning:
# IPython 4
from ipywidgets import FloatProgress
from ipyparallel import Client
from IPython.display import display
+
+ import ipykernel
+ import notebook.notebookapp
+
+ import requests
+
+
+ def check_nbserverproxy():
+ """
+ Return the url of the current jupyter notebook server.
+ """
+ kernel_id = re.search(
+ "kernel-(.*).json",
+ ipykernel.connect.get_connection_file()
+ ).group(1)
+ servers = notebook.notebookapp.list_running_servers()
+ for s in servers:
+ response = requests.get(
+ requests.compat.urljoin(s["url"], "api/sessions"),
+ params={"token": s.get("token", "")}
+ )
+ for n in json.loads(response.text):
+ if n["kernel"]["id"] == kernel_id:
+ # Found server that is running this Jupyter Notebook.
+ # Try to requests this servers port through nbserverproxy.
+ url = requests.compat.urljoin(
+ s["url"], "proxy/%i" % s["port"]
+ )
+ # If the proxy is running, it will redirect.
+ # If not, it will error out.
+ try:
+ requests.get(url).raise_for_status()
+ except requests.HTTPError:
+ return False
+ else:
+ return True |
eeac557b77a3a63a3497791a2716706801b20e37 | kodos/main.py | kodos/main.py |
def run(args=None):
"""Main entry point of the application."""
pass
| import sys
from PyQt4.QtGui import QApplication, QMainWindow
from kodos.ui.ui_main import Ui_MainWindow
class KodosMainWindow(QMainWindow, Ui_MainWindow):
def __init__(self, parent=None):
super(KodosMainWindow, self).__init__(parent)
self.setupUi(self)
self.connectActions()
# Trigger the textChanged signal
for widget in [self.regexText, self.searchText, self.replaceText]:
widget.setPlainText('')
def connectActions(self):
# Connect input widgets to update the GUI when their text change
for widget in [self.regexText, self.searchText, self.replaceText]:
widget.textChanged.connect(self.on_compute_regex)
def on_compute_regex(self):
regex = self.regexText.toPlainText()
search = self.searchText.toPlainText()
replace = self.replaceText.toPlainText()
if regex == "" or search == "":
self.statusbar.showMessage(
"Please enter a regex and a search to work on")
else:
self.statusbar.clearMessage()
def run(args=None):
"""Main entry point of the application."""
app = QApplication(sys.argv)
kodos = KodosMainWindow()
kodos.show()
app.exec_()
| Connect the UI to the code and start to connect slots to actions. | Connect the UI to the code and start to connect slots to actions.
| Python | bsd-2-clause | multani/kodos-qt4 | + import sys
+ from PyQt4.QtGui import QApplication, QMainWindow
+
+ from kodos.ui.ui_main import Ui_MainWindow
+
+
+ class KodosMainWindow(QMainWindow, Ui_MainWindow):
+ def __init__(self, parent=None):
+ super(KodosMainWindow, self).__init__(parent)
+ self.setupUi(self)
+ self.connectActions()
+
+ # Trigger the textChanged signal
+ for widget in [self.regexText, self.searchText, self.replaceText]:
+ widget.setPlainText('')
+
+ def connectActions(self):
+ # Connect input widgets to update the GUI when their text change
+ for widget in [self.regexText, self.searchText, self.replaceText]:
+ widget.textChanged.connect(self.on_compute_regex)
+
+ def on_compute_regex(self):
+ regex = self.regexText.toPlainText()
+ search = self.searchText.toPlainText()
+ replace = self.replaceText.toPlainText()
+
+ if regex == "" or search == "":
+ self.statusbar.showMessage(
+ "Please enter a regex and a search to work on")
+ else:
+ self.statusbar.clearMessage()
def run(args=None):
"""Main entry point of the application."""
- pass
+ app = QApplication(sys.argv)
+ kodos = KodosMainWindow()
+ kodos.show()
+ app.exec_()
+ | Connect the UI to the code and start to connect slots to actions. | ## Code Before:
def run(args=None):
"""Main entry point of the application."""
pass
## Instruction:
Connect the UI to the code and start to connect slots to actions.
## Code After:
import sys
from PyQt4.QtGui import QApplication, QMainWindow
from kodos.ui.ui_main import Ui_MainWindow
class KodosMainWindow(QMainWindow, Ui_MainWindow):
def __init__(self, parent=None):
super(KodosMainWindow, self).__init__(parent)
self.setupUi(self)
self.connectActions()
# Trigger the textChanged signal
for widget in [self.regexText, self.searchText, self.replaceText]:
widget.setPlainText('')
def connectActions(self):
# Connect input widgets to update the GUI when their text change
for widget in [self.regexText, self.searchText, self.replaceText]:
widget.textChanged.connect(self.on_compute_regex)
def on_compute_regex(self):
regex = self.regexText.toPlainText()
search = self.searchText.toPlainText()
replace = self.replaceText.toPlainText()
if regex == "" or search == "":
self.statusbar.showMessage(
"Please enter a regex and a search to work on")
else:
self.statusbar.clearMessage()
def run(args=None):
"""Main entry point of the application."""
app = QApplication(sys.argv)
kodos = KodosMainWindow()
kodos.show()
app.exec_()
| + import sys
+ from PyQt4.QtGui import QApplication, QMainWindow
+
+ from kodos.ui.ui_main import Ui_MainWindow
+
+
+ class KodosMainWindow(QMainWindow, Ui_MainWindow):
+ def __init__(self, parent=None):
+ super(KodosMainWindow, self).__init__(parent)
+ self.setupUi(self)
+ self.connectActions()
+
+ # Trigger the textChanged signal
+ for widget in [self.regexText, self.searchText, self.replaceText]:
+ widget.setPlainText('')
+
+ def connectActions(self):
+ # Connect input widgets to update the GUI when their text change
+ for widget in [self.regexText, self.searchText, self.replaceText]:
+ widget.textChanged.connect(self.on_compute_regex)
+
+ def on_compute_regex(self):
+ regex = self.regexText.toPlainText()
+ search = self.searchText.toPlainText()
+ replace = self.replaceText.toPlainText()
+
+ if regex == "" or search == "":
+ self.statusbar.showMessage(
+ "Please enter a regex and a search to work on")
+ else:
+ self.statusbar.clearMessage()
def run(args=None):
"""Main entry point of the application."""
- pass
+
+ app = QApplication(sys.argv)
+ kodos = KodosMainWindow()
+ kodos.show()
+ app.exec_() |
7aa89902f8af2ca1f4b3c9e356a62062cc74696b | bot/anime_searcher.py | bot/anime_searcher.py | from itertools import chain
from typing import Iterable
from minoshiro import Medium, Minoshiro, Site
from minoshiro.helpers import get_synonyms
class AnimeSearcher(Minoshiro):
async def get(self, query: str, medium: Medium,
sites: Iterable[Site] = None, *, timeout=3):
sites = sites if sites else list(Site)
cached_data, cached_id = await self.__get_cached(query, medium)
to_be_cached = {}
names = []
return_val = {}
for site in sites:
res, id_ = await self.__get_result(
cached_data, cached_id, query, names, site, medium, timeout
)
if res:
return_val[site] = res
for title in get_synonyms(res, site):
names.append(title)
if id_:
to_be_cached[site] = id_
return return_val, to_be_cached, names, medium
async def cache(self, to_be_cached, names, medium):
"""
Cache search results into the db.
:param to_be_cached: items to be cached.
:param names: all names for the item.
:param medium: the medium type.
"""
itere = set(chain(*names))
for site, id_ in to_be_cached.items():
await self.cache_one(site, id_, medium, itere)
async def cache_one(self, site, id_, medium, iterator):
"""
Cache one id.
:param site: the site.
:param id_: the id.
:param medium: the medium type.
:param iterator: an iterator for all names.
"""
for name in iterator:
if name:
await self.db_controller.set_identifier(
name, medium, site, id_
)
| from typing import Iterable
from minoshiro import Medium, Minoshiro, Site
from minoshiro.helpers import get_synonyms
class AnimeSearcher(Minoshiro):
async def get(self, query: str, medium: Medium,
sites: Iterable[Site] = None, *, timeout=3):
sites = sites if sites else list(Site)
cached_data, cached_id = await self._get_cached(query, medium)
to_be_cached = {}
names = []
return_val = {}
for site in sites:
res, id_ = await self._get_result(
cached_data, cached_id, query, names, site, medium, timeout
)
if res:
return_val[site] = res
for title in get_synonyms(res, site):
names.append(title)
if id_:
to_be_cached[site] = id_
return return_val, to_be_cached, names, medium
async def cache(self, to_be_cached, names, medium):
await super()._cache(to_be_cached, names, medium)
| Update anime searcher implementation to use super class methods | Update anime searcher implementation to use super class methods
| Python | apache-2.0 | MaT1g3R/YasenBaka | - from itertools import chain
from typing import Iterable
from minoshiro import Medium, Minoshiro, Site
from minoshiro.helpers import get_synonyms
class AnimeSearcher(Minoshiro):
async def get(self, query: str, medium: Medium,
sites: Iterable[Site] = None, *, timeout=3):
sites = sites if sites else list(Site)
- cached_data, cached_id = await self.__get_cached(query, medium)
+ cached_data, cached_id = await self._get_cached(query, medium)
to_be_cached = {}
names = []
return_val = {}
for site in sites:
- res, id_ = await self.__get_result(
+ res, id_ = await self._get_result(
cached_data, cached_id, query, names, site, medium, timeout
)
if res:
return_val[site] = res
for title in get_synonyms(res, site):
names.append(title)
if id_:
to_be_cached[site] = id_
return return_val, to_be_cached, names, medium
async def cache(self, to_be_cached, names, medium):
+ await super()._cache(to_be_cached, names, medium)
- """
- Cache search results into the db.
- :param to_be_cached: items to be cached.
- :param names: all names for the item.
- :param medium: the medium type.
- """
- itere = set(chain(*names))
- for site, id_ in to_be_cached.items():
- await self.cache_one(site, id_, medium, itere)
- async def cache_one(self, site, id_, medium, iterator):
- """
- Cache one id.
- :param site: the site.
- :param id_: the id.
- :param medium: the medium type.
- :param iterator: an iterator for all names.
- """
- for name in iterator:
- if name:
- await self.db_controller.set_identifier(
- name, medium, site, id_
- )
- | Update anime searcher implementation to use super class methods | ## Code Before:
from itertools import chain
from typing import Iterable
from minoshiro import Medium, Minoshiro, Site
from minoshiro.helpers import get_synonyms
class AnimeSearcher(Minoshiro):
async def get(self, query: str, medium: Medium,
sites: Iterable[Site] = None, *, timeout=3):
sites = sites if sites else list(Site)
cached_data, cached_id = await self.__get_cached(query, medium)
to_be_cached = {}
names = []
return_val = {}
for site in sites:
res, id_ = await self.__get_result(
cached_data, cached_id, query, names, site, medium, timeout
)
if res:
return_val[site] = res
for title in get_synonyms(res, site):
names.append(title)
if id_:
to_be_cached[site] = id_
return return_val, to_be_cached, names, medium
async def cache(self, to_be_cached, names, medium):
"""
Cache search results into the db.
:param to_be_cached: items to be cached.
:param names: all names for the item.
:param medium: the medium type.
"""
itere = set(chain(*names))
for site, id_ in to_be_cached.items():
await self.cache_one(site, id_, medium, itere)
async def cache_one(self, site, id_, medium, iterator):
"""
Cache one id.
:param site: the site.
:param id_: the id.
:param medium: the medium type.
:param iterator: an iterator for all names.
"""
for name in iterator:
if name:
await self.db_controller.set_identifier(
name, medium, site, id_
)
## Instruction:
Update anime searcher implementation to use super class methods
## Code After:
from typing import Iterable
from minoshiro import Medium, Minoshiro, Site
from minoshiro.helpers import get_synonyms
class AnimeSearcher(Minoshiro):
async def get(self, query: str, medium: Medium,
sites: Iterable[Site] = None, *, timeout=3):
sites = sites if sites else list(Site)
cached_data, cached_id = await self._get_cached(query, medium)
to_be_cached = {}
names = []
return_val = {}
for site in sites:
res, id_ = await self._get_result(
cached_data, cached_id, query, names, site, medium, timeout
)
if res:
return_val[site] = res
for title in get_synonyms(res, site):
names.append(title)
if id_:
to_be_cached[site] = id_
return return_val, to_be_cached, names, medium
async def cache(self, to_be_cached, names, medium):
await super()._cache(to_be_cached, names, medium)
| - from itertools import chain
from typing import Iterable
from minoshiro import Medium, Minoshiro, Site
from minoshiro.helpers import get_synonyms
class AnimeSearcher(Minoshiro):
async def get(self, query: str, medium: Medium,
sites: Iterable[Site] = None, *, timeout=3):
sites = sites if sites else list(Site)
- cached_data, cached_id = await self.__get_cached(query, medium)
? -
+ cached_data, cached_id = await self._get_cached(query, medium)
to_be_cached = {}
names = []
return_val = {}
for site in sites:
- res, id_ = await self.__get_result(
? -
+ res, id_ = await self._get_result(
cached_data, cached_id, query, names, site, medium, timeout
)
if res:
return_val[site] = res
for title in get_synonyms(res, site):
names.append(title)
if id_:
to_be_cached[site] = id_
return return_val, to_be_cached, names, medium
async def cache(self, to_be_cached, names, medium):
+ await super()._cache(to_be_cached, names, medium)
- """
- Cache search results into the db.
- :param to_be_cached: items to be cached.
- :param names: all names for the item.
- :param medium: the medium type.
- """
- itere = set(chain(*names))
- for site, id_ in to_be_cached.items():
- await self.cache_one(site, id_, medium, itere)
-
- async def cache_one(self, site, id_, medium, iterator):
- """
- Cache one id.
- :param site: the site.
- :param id_: the id.
- :param medium: the medium type.
- :param iterator: an iterator for all names.
- """
- for name in iterator:
- if name:
- await self.db_controller.set_identifier(
- name, medium, site, id_
- ) |
930ad35cee818e2d0b97f840ff0b3b772bd51af3 | post_office/management/commands/send_queued_mail.py | post_office/management/commands/send_queued_mail.py | import tempfile
import sys
from optparse import make_option
from django.core.management.base import BaseCommand
from ...lockfile import FileLock, FileLocked
from ...mail import send_queued
from ...logutils import setup_loghandlers
logger = setup_loghandlers()
default_lockfile = tempfile.gettempdir() + "/post_office"
class Command(BaseCommand):
option_list = BaseCommand.option_list + (
make_option('-p', '--processes', type='int',
help='Number of processes used to send emails', default=1),
make_option('-L', '--lockfile', type='string', default=default_lockfile,
help='Absolute path of lockfile to acquire'),
make_option('-l', '--log-level', type='int',
help='"0" to log nothing, "1" to only log errors'),
)
def handle(self, *args, **options):
logger.info('Acquiring lock for sending queued emails at %s.lock' %
options['lockfile'])
try:
with FileLock(options['lockfile']):
try:
send_queued(options['processes'], options.get('log_level'))
except Exception as e:
logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500})
raise
except FileLocked:
logger.info('Failed to acquire lock, terminating now.')
| import tempfile
import sys
from django.core.management.base import BaseCommand
from ...lockfile import FileLock, FileLocked
from ...mail import send_queued
from ...logutils import setup_loghandlers
logger = setup_loghandlers()
default_lockfile = tempfile.gettempdir() + "/post_office"
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument('-p', '--processes',
type='int',
help='Number of processes used to send emails',
default=1
)
parser.add_argument('-L', '--lockfile',
type='string',
default=default_lockfile,
help='Absolute path of lockfile to acquire'
)
parser.add_argument('-l', '--log-level',
type='int',
help='"0" to log nothing, "1" to only log errors'
)
def handle(self, *args, **options):
logger.info('Acquiring lock for sending queued emails at %s.lock' %
options['lockfile'])
try:
with FileLock(options['lockfile']):
try:
send_queued(options['processes'], options.get('log_level'))
except Exception as e:
logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500})
raise
except FileLocked:
logger.info('Failed to acquire lock, terminating now.')
| Switch to using the `add_arguments` method. | Switch to using the `add_arguments` method.
This is an alternative to using the `option_list` and
`optparse.make_option`. Django deprecated the use of `optparse` in
management commands in Django 1.8 and removed it in Django 1.10.
| Python | mit | jrief/django-post_office,ui/django-post_office,ui/django-post_office | import tempfile
import sys
- from optparse import make_option
from django.core.management.base import BaseCommand
from ...lockfile import FileLock, FileLocked
from ...mail import send_queued
from ...logutils import setup_loghandlers
logger = setup_loghandlers()
default_lockfile = tempfile.gettempdir() + "/post_office"
class Command(BaseCommand):
-
- option_list = BaseCommand.option_list + (
- make_option('-p', '--processes', type='int',
+ def add_arguments(self, parser):
+ parser.add_argument('-p', '--processes',
+ type='int',
- help='Number of processes used to send emails', default=1),
+ help='Number of processes used to send emails',
- make_option('-L', '--lockfile', type='string', default=default_lockfile,
+ default=1
+ )
+ parser.add_argument('-L', '--lockfile',
+ type='string',
+ default=default_lockfile,
- help='Absolute path of lockfile to acquire'),
+ help='Absolute path of lockfile to acquire'
- make_option('-l', '--log-level', type='int',
+ )
+ parser.add_argument('-l', '--log-level',
+ type='int',
- help='"0" to log nothing, "1" to only log errors'),
+ help='"0" to log nothing, "1" to only log errors'
- )
+ )
def handle(self, *args, **options):
logger.info('Acquiring lock for sending queued emails at %s.lock' %
options['lockfile'])
try:
with FileLock(options['lockfile']):
try:
send_queued(options['processes'], options.get('log_level'))
except Exception as e:
logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500})
raise
except FileLocked:
logger.info('Failed to acquire lock, terminating now.')
| Switch to using the `add_arguments` method. | ## Code Before:
import tempfile
import sys
from optparse import make_option
from django.core.management.base import BaseCommand
from ...lockfile import FileLock, FileLocked
from ...mail import send_queued
from ...logutils import setup_loghandlers
logger = setup_loghandlers()
default_lockfile = tempfile.gettempdir() + "/post_office"
class Command(BaseCommand):
option_list = BaseCommand.option_list + (
make_option('-p', '--processes', type='int',
help='Number of processes used to send emails', default=1),
make_option('-L', '--lockfile', type='string', default=default_lockfile,
help='Absolute path of lockfile to acquire'),
make_option('-l', '--log-level', type='int',
help='"0" to log nothing, "1" to only log errors'),
)
def handle(self, *args, **options):
logger.info('Acquiring lock for sending queued emails at %s.lock' %
options['lockfile'])
try:
with FileLock(options['lockfile']):
try:
send_queued(options['processes'], options.get('log_level'))
except Exception as e:
logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500})
raise
except FileLocked:
logger.info('Failed to acquire lock, terminating now.')
## Instruction:
Switch to using the `add_arguments` method.
## Code After:
import tempfile
import sys
from django.core.management.base import BaseCommand
from ...lockfile import FileLock, FileLocked
from ...mail import send_queued
from ...logutils import setup_loghandlers
logger = setup_loghandlers()
default_lockfile = tempfile.gettempdir() + "/post_office"
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument('-p', '--processes',
type='int',
help='Number of processes used to send emails',
default=1
)
parser.add_argument('-L', '--lockfile',
type='string',
default=default_lockfile,
help='Absolute path of lockfile to acquire'
)
parser.add_argument('-l', '--log-level',
type='int',
help='"0" to log nothing, "1" to only log errors'
)
def handle(self, *args, **options):
logger.info('Acquiring lock for sending queued emails at %s.lock' %
options['lockfile'])
try:
with FileLock(options['lockfile']):
try:
send_queued(options['processes'], options.get('log_level'))
except Exception as e:
logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500})
raise
except FileLocked:
logger.info('Failed to acquire lock, terminating now.')
| import tempfile
import sys
- from optparse import make_option
from django.core.management.base import BaseCommand
from ...lockfile import FileLock, FileLocked
from ...mail import send_queued
from ...logutils import setup_loghandlers
logger = setup_loghandlers()
default_lockfile = tempfile.gettempdir() + "/post_office"
class Command(BaseCommand):
-
- option_list = BaseCommand.option_list + (
- make_option('-p', '--processes', type='int',
+ def add_arguments(self, parser):
+ parser.add_argument('-p', '--processes',
+ type='int',
- help='Number of processes used to send emails', default=1),
? -------- ------------
+ help='Number of processes used to send emails',
- make_option('-L', '--lockfile', type='string', default=default_lockfile,
+ default=1
+ )
+ parser.add_argument('-L', '--lockfile',
+ type='string',
+ default=default_lockfile,
- help='Absolute path of lockfile to acquire'),
? -------- --
+ help='Absolute path of lockfile to acquire'
- make_option('-l', '--log-level', type='int',
+ )
+ parser.add_argument('-l', '--log-level',
+ type='int',
- help='"0" to log nothing, "1" to only log errors'),
? -------- --
+ help='"0" to log nothing, "1" to only log errors'
- )
+ )
def handle(self, *args, **options):
logger.info('Acquiring lock for sending queued emails at %s.lock' %
options['lockfile'])
try:
with FileLock(options['lockfile']):
try:
send_queued(options['processes'], options.get('log_level'))
except Exception as e:
logger.error(e, exc_info=sys.exc_info(), extra={'status_code': 500})
raise
except FileLocked:
logger.info('Failed to acquire lock, terminating now.') |
e2be9eb27d6fc7cfa424cbf908347796ab595526 | groundstation/broadcast_announcer.py | groundstation/broadcast_announcer.py | import socket
import logger
from groundstation.broadcast_socket import BroadcastSocket
import logger
log = logger.getLogger(__name__)
class BroadcastAnnouncer(BroadcastSocket):
def __init__(self, port):
super(BroadcastAnnouncer, self).__init__()
self._addr = '255.255.255.255', port
self._name = None
self.broadcast_payload = "PING None"
@property
def name(self):
return self._name
@name.setter
def name(self, value):
self._name = value
self.broadcast_payload = "PING %s" % (self._name)
def ping(self):
log.info("ping payload: %s" % (self.broadcast_payload))
transmitted = self.socket.sendto(self.broadcast_payload, self._addr)
if transmitted != len(self.broadcast_payload):
log.warning("ping wasn't successfully broadcast")
| import socket
import logger
from sockets.broadcast_socket import BroadcastSocket
import logger
log = logger.getLogger(__name__)
class BroadcastAnnouncer(BroadcastSocket):
def __init__(self, port):
super(BroadcastAnnouncer, self).__init__()
self._addr = '255.255.255.255', port
self._name = None
self.broadcast_payload = "PING None"
@property
def name(self):
return self._name
@name.setter
def name(self, value):
self._name = value
self.broadcast_payload = "PING %s" % (self._name)
def ping(self):
log.info("ping payload: %s" % (self.broadcast_payload))
transmitted = self.socket.sendto(self.broadcast_payload, self._addr)
if transmitted != len(self.broadcast_payload):
log.warning("ping wasn't successfully broadcast")
| Fix an import path bug masked by remaining .pyc files | Fix an import path bug masked by remaining .pyc files
| Python | mit | richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation | import socket
import logger
- from groundstation.broadcast_socket import BroadcastSocket
+ from sockets.broadcast_socket import BroadcastSocket
import logger
log = logger.getLogger(__name__)
class BroadcastAnnouncer(BroadcastSocket):
def __init__(self, port):
super(BroadcastAnnouncer, self).__init__()
self._addr = '255.255.255.255', port
self._name = None
self.broadcast_payload = "PING None"
@property
def name(self):
return self._name
@name.setter
def name(self, value):
self._name = value
self.broadcast_payload = "PING %s" % (self._name)
def ping(self):
log.info("ping payload: %s" % (self.broadcast_payload))
transmitted = self.socket.sendto(self.broadcast_payload, self._addr)
if transmitted != len(self.broadcast_payload):
log.warning("ping wasn't successfully broadcast")
| Fix an import path bug masked by remaining .pyc files | ## Code Before:
import socket
import logger
from groundstation.broadcast_socket import BroadcastSocket
import logger
log = logger.getLogger(__name__)
class BroadcastAnnouncer(BroadcastSocket):
def __init__(self, port):
super(BroadcastAnnouncer, self).__init__()
self._addr = '255.255.255.255', port
self._name = None
self.broadcast_payload = "PING None"
@property
def name(self):
return self._name
@name.setter
def name(self, value):
self._name = value
self.broadcast_payload = "PING %s" % (self._name)
def ping(self):
log.info("ping payload: %s" % (self.broadcast_payload))
transmitted = self.socket.sendto(self.broadcast_payload, self._addr)
if transmitted != len(self.broadcast_payload):
log.warning("ping wasn't successfully broadcast")
## Instruction:
Fix an import path bug masked by remaining .pyc files
## Code After:
import socket
import logger
from sockets.broadcast_socket import BroadcastSocket
import logger
log = logger.getLogger(__name__)
class BroadcastAnnouncer(BroadcastSocket):
def __init__(self, port):
super(BroadcastAnnouncer, self).__init__()
self._addr = '255.255.255.255', port
self._name = None
self.broadcast_payload = "PING None"
@property
def name(self):
return self._name
@name.setter
def name(self, value):
self._name = value
self.broadcast_payload = "PING %s" % (self._name)
def ping(self):
log.info("ping payload: %s" % (self.broadcast_payload))
transmitted = self.socket.sendto(self.broadcast_payload, self._addr)
if transmitted != len(self.broadcast_payload):
log.warning("ping wasn't successfully broadcast")
| import socket
import logger
- from groundstation.broadcast_socket import BroadcastSocket
? ^^ ^^^ ------
+ from sockets.broadcast_socket import BroadcastSocket
? ^ ^^^^
import logger
log = logger.getLogger(__name__)
class BroadcastAnnouncer(BroadcastSocket):
def __init__(self, port):
super(BroadcastAnnouncer, self).__init__()
self._addr = '255.255.255.255', port
self._name = None
self.broadcast_payload = "PING None"
@property
def name(self):
return self._name
@name.setter
def name(self, value):
self._name = value
self.broadcast_payload = "PING %s" % (self._name)
def ping(self):
log.info("ping payload: %s" % (self.broadcast_payload))
transmitted = self.socket.sendto(self.broadcast_payload, self._addr)
if transmitted != len(self.broadcast_payload):
log.warning("ping wasn't successfully broadcast") |
608325c33cb2d446b89c263ba0bb02ced5c4ffe8 | portal/views.py | portal/views.py | import csv
from django.shortcuts import render
from django.http import HttpResponse
from . import services
def index(request):
data = services.overview()
return render(request, 'index.html', data)
def meter_runs(request):
"""Render the table of exported MeterRun results in html"""
data = services.meterruns_export()
return render(request, 'project_run_table.html', data)
def csv_export(request):
"""Return a dump of all the MeterRuns in CSV form"""
response = HttpResponse(content_type="text/csv")
response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"'
writer = csv.writer(response)
writer.writerow(['First row', 'Foo', 'Bar'])
return response | import csv
from django.shortcuts import render
from django.http import HttpResponse
from . import services
def index(request):
data = services.overview()
return render(request, 'index.html', data)
def meter_runs(request):
"""Render the table of exported MeterRun results in html"""
data = services.meterruns_export()
return render(request, 'project_run_table.html', data)
def csv_export(request):
"""Return a dump of all the MeterRuns in CSV form"""
response = HttpResponse(content_type="text/csv")
response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"'
data = services.meterruns_export()
writer = csv.DictWriter(response, fieldnames=data['headers'])
writer.writeheader()
for meter_run in data['meter_runs']:
writer.writerow(meter_run)
return response | Use the meterrun_export service to power csv export | Use the meterrun_export service to power csv export
| Python | mit | impactlab/oeem-energy-datastore,impactlab/oeem-energy-datastore,impactlab/oeem-energy-datastore | import csv
from django.shortcuts import render
from django.http import HttpResponse
from . import services
def index(request):
data = services.overview()
return render(request, 'index.html', data)
def meter_runs(request):
"""Render the table of exported MeterRun results in html"""
data = services.meterruns_export()
return render(request, 'project_run_table.html', data)
def csv_export(request):
"""Return a dump of all the MeterRuns in CSV form"""
response = HttpResponse(content_type="text/csv")
response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"'
- writer = csv.writer(response)
- writer.writerow(['First row', 'Foo', 'Bar'])
+ data = services.meterruns_export()
+
+ writer = csv.DictWriter(response, fieldnames=data['headers'])
+ writer.writeheader()
+ for meter_run in data['meter_runs']:
+ writer.writerow(meter_run)
return response | Use the meterrun_export service to power csv export | ## Code Before:
import csv
from django.shortcuts import render
from django.http import HttpResponse
from . import services
def index(request):
data = services.overview()
return render(request, 'index.html', data)
def meter_runs(request):
"""Render the table of exported MeterRun results in html"""
data = services.meterruns_export()
return render(request, 'project_run_table.html', data)
def csv_export(request):
"""Return a dump of all the MeterRuns in CSV form"""
response = HttpResponse(content_type="text/csv")
response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"'
writer = csv.writer(response)
writer.writerow(['First row', 'Foo', 'Bar'])
return response
## Instruction:
Use the meterrun_export service to power csv export
## Code After:
import csv
from django.shortcuts import render
from django.http import HttpResponse
from . import services
def index(request):
data = services.overview()
return render(request, 'index.html', data)
def meter_runs(request):
"""Render the table of exported MeterRun results in html"""
data = services.meterruns_export()
return render(request, 'project_run_table.html', data)
def csv_export(request):
"""Return a dump of all the MeterRuns in CSV form"""
response = HttpResponse(content_type="text/csv")
response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"'
data = services.meterruns_export()
writer = csv.DictWriter(response, fieldnames=data['headers'])
writer.writeheader()
for meter_run in data['meter_runs']:
writer.writerow(meter_run)
return response | import csv
from django.shortcuts import render
from django.http import HttpResponse
from . import services
def index(request):
data = services.overview()
return render(request, 'index.html', data)
def meter_runs(request):
"""Render the table of exported MeterRun results in html"""
data = services.meterruns_export()
return render(request, 'project_run_table.html', data)
def csv_export(request):
"""Return a dump of all the MeterRuns in CSV form"""
response = HttpResponse(content_type="text/csv")
response["Content-Disposition"] = 'attachment; filename="meter_runs.csv"'
- writer = csv.writer(response)
- writer.writerow(['First row', 'Foo', 'Bar'])
+ data = services.meterruns_export()
+
+ writer = csv.DictWriter(response, fieldnames=data['headers'])
+ writer.writeheader()
+ for meter_run in data['meter_runs']:
+ writer.writerow(meter_run)
return response |
29110323469d20ff1e481ab2267812afd8e0a3a4 | more/chameleon/main.py | more/chameleon/main.py | import morepath
import chameleon
class ChameleonApp(morepath.App):
pass
@ChameleonApp.setting_section(section='chameleon')
def get_setting_section():
return {'auto_reload': False}
@ChameleonApp.template_engine(extension='.pt')
def get_chameleon_render(path, original_render, settings):
config = settings.chameleon.__dict__
template = chameleon.PageTemplateFile(path, **config)
def render(content, request):
variables = {'request': request}
variables.update(content)
return original_render(template.render(**variables), request)
return render
| import os
import morepath
import chameleon
class ChameleonApp(morepath.App):
pass
@ChameleonApp.setting_section(section='chameleon')
def get_setting_section():
return {'auto_reload': False}
@ChameleonApp.template_engine(extension='.pt')
def get_chameleon_render(name, original_render, registry, search_path):
config = registry.settings.chameleon.__dict__
template = chameleon.PageTemplateFile(os.path.join(search_path, name),
**config)
def render(content, request):
variables = {'request': request}
variables.update(content)
return original_render(template.render(**variables), request)
return render
| Adjust to modifications in Morepath. But now to enable real explicit file support. | Adjust to modifications in Morepath. But now to enable real
explicit file support.
| Python | bsd-3-clause | morepath/more.chameleon | + import os
import morepath
import chameleon
class ChameleonApp(morepath.App):
pass
@ChameleonApp.setting_section(section='chameleon')
def get_setting_section():
return {'auto_reload': False}
@ChameleonApp.template_engine(extension='.pt')
- def get_chameleon_render(path, original_render, settings):
+ def get_chameleon_render(name, original_render, registry, search_path):
- config = settings.chameleon.__dict__
+ config = registry.settings.chameleon.__dict__
- template = chameleon.PageTemplateFile(path, **config)
+ template = chameleon.PageTemplateFile(os.path.join(search_path, name),
+ **config)
def render(content, request):
variables = {'request': request}
variables.update(content)
return original_render(template.render(**variables), request)
return render
| Adjust to modifications in Morepath. But now to enable real explicit file support. | ## Code Before:
import morepath
import chameleon
class ChameleonApp(morepath.App):
pass
@ChameleonApp.setting_section(section='chameleon')
def get_setting_section():
return {'auto_reload': False}
@ChameleonApp.template_engine(extension='.pt')
def get_chameleon_render(path, original_render, settings):
config = settings.chameleon.__dict__
template = chameleon.PageTemplateFile(path, **config)
def render(content, request):
variables = {'request': request}
variables.update(content)
return original_render(template.render(**variables), request)
return render
## Instruction:
Adjust to modifications in Morepath. But now to enable real explicit file support.
## Code After:
import os
import morepath
import chameleon
class ChameleonApp(morepath.App):
pass
@ChameleonApp.setting_section(section='chameleon')
def get_setting_section():
return {'auto_reload': False}
@ChameleonApp.template_engine(extension='.pt')
def get_chameleon_render(name, original_render, registry, search_path):
config = registry.settings.chameleon.__dict__
template = chameleon.PageTemplateFile(os.path.join(search_path, name),
**config)
def render(content, request):
variables = {'request': request}
variables.update(content)
return original_render(template.render(**variables), request)
return render
| + import os
import morepath
import chameleon
class ChameleonApp(morepath.App):
pass
@ChameleonApp.setting_section(section='chameleon')
def get_setting_section():
return {'auto_reload': False}
@ChameleonApp.template_engine(extension='.pt')
- def get_chameleon_render(path, original_render, settings):
? ^ ^^ ^^^^^
+ def get_chameleon_render(name, original_render, registry, search_path):
? ^ ^^ ++++++++++ +++++++ ^
- config = settings.chameleon.__dict__
+ config = registry.settings.chameleon.__dict__
? +++++++++
- template = chameleon.PageTemplateFile(path, **config)
? ---- ^^^
+ template = chameleon.PageTemplateFile(os.path.join(search_path, name),
? ++++++++++++++++++++ ^^^ +
+ **config)
def render(content, request):
variables = {'request': request}
variables.update(content)
return original_render(template.render(**variables), request)
return render |
73c842af63a09add43c0e33336dd4eb21153fda1 | bin/database.py | bin/database.py |
import json
from api import config
CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new
def confirm_schema_match():
"""
Checks version of database schema
Returns (0) if DB schema version matches requirements.
Returns (42) if DB schema version does not match
requirements and can be upgraded.
Returns (43) if DB schema version does not match
requirements and cannot be upgraded,
perhaps because code is at lower version
than the DB schema version.
"""
version = config.db.version.find_one({"_id": "version"})
if version is None or version.get('database', None) is None:
return 42 # At version 0
db_version = version.get('database', 0)
if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
return 43
elif db_version < CURRENT_DATABASE_VERSION:
return 42
else:
return 0
def upgrade_schema():
"""
Upgrades db to the current schema version
Returns (0) if upgrade is successful
"""
# In progress
# db_version = version.get('database',0)
# if db_version < 1:
# # rename the metadata fields
# config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
# config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
return 0
|
import json
from api import config
CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made
def confirm_schema_match():
"""
Checks version of database schema
Returns (0) if DB schema version matches requirements.
Returns (42) if DB schema version does not match
requirements and can be upgraded.
Returns (43) if DB schema version does not match
requirements and cannot be upgraded,
perhaps because code is at lower version
than the DB schema version.
"""
version = config.db.version.find_one({"_id": "version"})
if version is None or version.get('database', None) is None:
return 42 # At version 0
db_version = version.get('database', 0)
if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
return 43
elif db_version < CURRENT_DATABASE_VERSION:
return 42
else:
return 0
def upgrade_schema():
"""
Upgrades db to the current schema version
Returns (0) if upgrade is successful
"""
# In progress
# db_version = version.get('database',0)
# if db_version < 1:
# # rename the metadata fields
# config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
# config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
return 0
| Fix tab vs spaces issue | Fix tab vs spaces issue
| Python | mit | scitran/api,scitran/api,scitran/core,scitran/core,scitran/core,scitran/core |
import json
from api import config
- CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new
+ CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made
def confirm_schema_match():
- """
+ """
- Checks version of database schema
+ Checks version of database schema
- Returns (0) if DB schema version matches requirements.
+ Returns (0) if DB schema version matches requirements.
- Returns (42) if DB schema version does not match
+ Returns (42) if DB schema version does not match
- requirements and can be upgraded.
+ requirements and can be upgraded.
- Returns (43) if DB schema version does not match
+ Returns (43) if DB schema version does not match
- requirements and cannot be upgraded,
+ requirements and cannot be upgraded,
- perhaps because code is at lower version
+ perhaps because code is at lower version
- than the DB schema version.
+ than the DB schema version.
- """
+ """
- version = config.db.version.find_one({"_id": "version"})
+ version = config.db.version.find_one({"_id": "version"})
- if version is None or version.get('database', None) is None:
+ if version is None or version.get('database', None) is None:
- return 42 # At version 0
+ return 42 # At version 0
- db_version = version.get('database', 0)
+ db_version = version.get('database', 0)
- if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
+ if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
- return 43
+ return 43
- elif db_version < CURRENT_DATABASE_VERSION:
+ elif db_version < CURRENT_DATABASE_VERSION:
- return 42
- else:
- return 0
+ return 42
+ else:
+ return 0
def upgrade_schema():
- """
+ """
- Upgrades db to the current schema version
+ Upgrades db to the current schema version
Returns (0) if upgrade is successful
"""
- # In progress
+ # In progress
- # db_version = version.get('database',0)
+ # db_version = version.get('database',0)
-
+
- # if db_version < 1:
+ # if db_version < 1:
- # # rename the metadata fields
+ # # rename the metadata fields
- # config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
+ # config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
- # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
+ # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
- return 0
+ return 0
| Fix tab vs spaces issue | ## Code Before:
import json
from api import config
CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new
def confirm_schema_match():
"""
Checks version of database schema
Returns (0) if DB schema version matches requirements.
Returns (42) if DB schema version does not match
requirements and can be upgraded.
Returns (43) if DB schema version does not match
requirements and cannot be upgraded,
perhaps because code is at lower version
than the DB schema version.
"""
version = config.db.version.find_one({"_id": "version"})
if version is None or version.get('database', None) is None:
return 42 # At version 0
db_version = version.get('database', 0)
if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
return 43
elif db_version < CURRENT_DATABASE_VERSION:
return 42
else:
return 0
def upgrade_schema():
"""
Upgrades db to the current schema version
Returns (0) if upgrade is successful
"""
# In progress
# db_version = version.get('database',0)
# if db_version < 1:
# # rename the metadata fields
# config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
# config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
return 0
## Instruction:
Fix tab vs spaces issue
## Code After:
import json
from api import config
CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made
def confirm_schema_match():
"""
Checks version of database schema
Returns (0) if DB schema version matches requirements.
Returns (42) if DB schema version does not match
requirements and can be upgraded.
Returns (43) if DB schema version does not match
requirements and cannot be upgraded,
perhaps because code is at lower version
than the DB schema version.
"""
version = config.db.version.find_one({"_id": "version"})
if version is None or version.get('database', None) is None:
return 42 # At version 0
db_version = version.get('database', 0)
if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
return 43
elif db_version < CURRENT_DATABASE_VERSION:
return 42
else:
return 0
def upgrade_schema():
"""
Upgrades db to the current schema version
Returns (0) if upgrade is successful
"""
# In progress
# db_version = version.get('database',0)
# if db_version < 1:
# # rename the metadata fields
# config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
# config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
return 0
|
import json
from api import config
- CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new
+ CURRENT_DATABASE_VERSION = 1 # An int that is bumped when a new schema change is made
? +++++++++++++++++++++
def confirm_schema_match():
- """
+ """
- Checks version of database schema
? ^
+ Checks version of database schema
? ^^^^
- Returns (0) if DB schema version matches requirements.
? ^
+ Returns (0) if DB schema version matches requirements.
? ^^^^
- Returns (42) if DB schema version does not match
? ^
+ Returns (42) if DB schema version does not match
? ^^^^
- requirements and can be upgraded.
? ^
+ requirements and can be upgraded.
? ^^^^
- Returns (43) if DB schema version does not match
? ^
+ Returns (43) if DB schema version does not match
? ^^^^
- requirements and cannot be upgraded,
? ^
+ requirements and cannot be upgraded,
? ^^^^
- perhaps because code is at lower version
? ^
+ perhaps because code is at lower version
? ^^^^
- than the DB schema version.
? ^
+ than the DB schema version.
? ^^^^
- """
+ """
- version = config.db.version.find_one({"_id": "version"})
? ^
+ version = config.db.version.find_one({"_id": "version"})
? ^^^^
- if version is None or version.get('database', None) is None:
? ^
+ if version is None or version.get('database', None) is None:
? ^^^^
- return 42 # At version 0
? ^^
+ return 42 # At version 0
? ^^^^^^^^
- db_version = version.get('database', 0)
? ^
+ db_version = version.get('database', 0)
? ^^^^
- if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
? ^
+ if not isinstance(db_version, int) or db_version > CURRENT_DATABASE_VERSION:
? ^^^^
- return 43
+ return 43
- elif db_version < CURRENT_DATABASE_VERSION:
? ^
+ elif db_version < CURRENT_DATABASE_VERSION:
? ^^^^
- return 42
- else:
- return 0
+ return 42
+ else:
+ return 0
def upgrade_schema():
- """
+ """
- Upgrades db to the current schema version
? ^
+ Upgrades db to the current schema version
? ^^^^
Returns (0) if upgrade is successful
"""
- # In progress
? ^
+ # In progress
? ^^^^
- # db_version = version.get('database',0)
? ^
+ # db_version = version.get('database',0)
? ^^^^
-
+
- # if db_version < 1:
? ^
+ # if db_version < 1:
? ^^^^
- # # rename the metadata fields
? ^ ^
+ # # rename the metadata fields
? ^^^^ ^^
- # config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
? ^ ^
+ # config.db.container.update_many({}, {"$rename": {"metadata": "info"}})
? ^^^^ ^^
- # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
? ^
+ # config.db.version.update_one({"_id": "version"}, {"$set": {"database": CURRENT_DATABASE_VERSION}})
? ^^^^
- return 0
? ^
+ return 0
? ^^^^
|
592c6550255793772add694cb941a0db0883713b | kamboo/core.py | kamboo/core.py |
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
session = botocore.session.get_session()
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
|
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.session = botocore.session.get_session()
self.service = service_name
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
def __repr__(self):
return "KambooConnection: [%s, %s, %s]" % (self.account_id,
self.region,
self.service)
| Fix the issue: "session" shared in different connections | Fix the issue: "session" shared in different connections
| Python | apache-2.0 | henrysher/kamboo,henrysher/kamboo |
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
- session = botocore.session.get_session()
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
+ self.session = botocore.session.get_session()
+ self.service = service_name
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
+ def __repr__(self):
+ return "KambooConnection: [%s, %s, %s]" % (self.account_id,
+ self.region,
+ self.service)
+ | Fix the issue: "session" shared in different connections | ## Code Before:
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
session = botocore.session.get_session()
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
## Instruction:
Fix the issue: "session" shared in different connections
## Code After:
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.session = botocore.session.get_session()
self.service = service_name
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
def __repr__(self):
return "KambooConnection: [%s, %s, %s]" % (self.account_id,
self.region,
self.service)
|
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
- session = botocore.session.get_session()
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
+ self.session = botocore.session.get_session()
+ self.service = service_name
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
+
+ def __repr__(self):
+ return "KambooConnection: [%s, %s, %s]" % (self.account_id,
+ self.region,
+ self.service) |
017ba0d18acb83a5135dd7a23c085b3c93d539b3 | linkatos/message.py | linkatos/message.py | import re
link_re = re.compile("https?://\S+(\s|$)")
def extract_url(message):
"""
Returns the first url in a message. If there aren't any returns None
"""
answer = link_re.search(message)
if answer is not None:
answer = answer.group(0).strip()
return answer
| import re
link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)")
def extract_url(message):
"""
Returns the first url in a message. If there aren't any returns None
"""
answer = link_re.search(message)
if answer is not None:
answer = answer.group(2).strip()
return answer
| Change regex to adapt to the <url> format | fix: Change regex to adapt to the <url> format
| Python | mit | iwi/linkatos,iwi/linkatos | import re
- link_re = re.compile("https?://\S+(\s|$)")
+ link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)")
def extract_url(message):
"""
Returns the first url in a message. If there aren't any returns None
"""
answer = link_re.search(message)
if answer is not None:
- answer = answer.group(0).strip()
+ answer = answer.group(2).strip()
return answer
| Change regex to adapt to the <url> format | ## Code Before:
import re
link_re = re.compile("https?://\S+(\s|$)")
def extract_url(message):
"""
Returns the first url in a message. If there aren't any returns None
"""
answer = link_re.search(message)
if answer is not None:
answer = answer.group(0).strip()
return answer
## Instruction:
Change regex to adapt to the <url> format
## Code After:
import re
link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)")
def extract_url(message):
"""
Returns the first url in a message. If there aren't any returns None
"""
answer = link_re.search(message)
if answer is not None:
answer = answer.group(2).strip()
return answer
| import re
- link_re = re.compile("https?://\S+(\s|$)")
? ^
+ link_re = re.compile("(\s|^)<(https?://[\w./?+]+)>(\s|$)")
? ++++++++ + ^^^^ ++++
def extract_url(message):
"""
Returns the first url in a message. If there aren't any returns None
"""
answer = link_re.search(message)
if answer is not None:
- answer = answer.group(0).strip()
? ^
+ answer = answer.group(2).strip()
? ^
return answer |
84af44868ea742bb5f6d08991526a98c8c78a931 | tellurium/teconverters/__init__.py | tellurium/teconverters/__init__.py |
from __future__ import absolute_import
# converts Antimony to/from SBML
from .convert_antimony import antimonyConverter
from .convert_omex import inlineOmexImporter, OmexFormatDetector
try:
from .convert_phrasedml import phrasedmlImporter
except:
pass
from .antimony_sbo import SBOError
from .inline_omex import inlineOmex, saveInlineOMEX
|
from __future__ import absolute_import
# converts Antimony to/from SBML
from .convert_antimony import antimonyConverter
from .convert_omex import inlineOmexImporter, OmexFormatDetector
try:
from .convert_phrasedml import phrasedmlImporter
from .inline_omex import inlineOmex, saveInlineOMEX
except:
pass
from .antimony_sbo import SBOError
| Drop inline omex if it fails. | Drop inline omex if it fails.
| Python | apache-2.0 | sys-bio/tellurium,sys-bio/tellurium |
from __future__ import absolute_import
# converts Antimony to/from SBML
from .convert_antimony import antimonyConverter
from .convert_omex import inlineOmexImporter, OmexFormatDetector
try:
from .convert_phrasedml import phrasedmlImporter
+ from .inline_omex import inlineOmex, saveInlineOMEX
except:
pass
from .antimony_sbo import SBOError
- from .inline_omex import inlineOmex, saveInlineOMEX
| Drop inline omex if it fails. | ## Code Before:
from __future__ import absolute_import
# converts Antimony to/from SBML
from .convert_antimony import antimonyConverter
from .convert_omex import inlineOmexImporter, OmexFormatDetector
try:
from .convert_phrasedml import phrasedmlImporter
except:
pass
from .antimony_sbo import SBOError
from .inline_omex import inlineOmex, saveInlineOMEX
## Instruction:
Drop inline omex if it fails.
## Code After:
from __future__ import absolute_import
# converts Antimony to/from SBML
from .convert_antimony import antimonyConverter
from .convert_omex import inlineOmexImporter, OmexFormatDetector
try:
from .convert_phrasedml import phrasedmlImporter
from .inline_omex import inlineOmex, saveInlineOMEX
except:
pass
from .antimony_sbo import SBOError
|
from __future__ import absolute_import
# converts Antimony to/from SBML
from .convert_antimony import antimonyConverter
from .convert_omex import inlineOmexImporter, OmexFormatDetector
try:
from .convert_phrasedml import phrasedmlImporter
+ from .inline_omex import inlineOmex, saveInlineOMEX
except:
pass
from .antimony_sbo import SBOError
- from .inline_omex import inlineOmex, saveInlineOMEX |
ae5626eaf36c6be94860d2a9570a777ff7f4e148 | apps/client_config.py | apps/client_config.py |
import superdesk
from flask import current_app as app
from superdesk.utils import ListCursor
class ClientConfigResource(superdesk.Resource):
item_methods = []
public_methods = ['GET']
resource_methods = ['GET']
class ClientConfigService(superdesk.Service):
def get(self, req, lookup):
return ListCursor()
def on_fetched(self, docs):
docs['config'] = getattr(app, 'client_config', {})
def init_app(app):
superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app)
app.client_config.update({
'schema': app.config.get('SCHEMA'),
'editor': app.config.get('EDITOR'),
})
|
import superdesk
from flask import current_app as app
from superdesk.utils import ListCursor
class ClientConfigResource(superdesk.Resource):
item_methods = []
public_methods = ['GET']
resource_methods = ['GET']
class ClientConfigService(superdesk.Service):
def get(self, req, lookup):
return ListCursor()
def on_fetched(self, docs):
docs['config'] = getattr(app, 'client_config', {})
def init_app(app):
superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app)
app.client_config.update({
'schema': app.config.get('SCHEMA'),
'editor': app.config.get('EDITOR'),
'feedback_url': app.config.get('FEEDBACK_URL'),
})
| Add feedback url to served client configuration | [SDESK-2128] Add feedback url to served client configuration
| Python | agpl-3.0 | superdesk/superdesk-core,petrjasek/superdesk-core,mugurrus/superdesk-core,mdhaman/superdesk-core,ioanpocol/superdesk-core,ioanpocol/superdesk-core,mugurrus/superdesk-core,ioanpocol/superdesk-core,superdesk/superdesk-core,mdhaman/superdesk-core,petrjasek/superdesk-core,petrjasek/superdesk-core,hlmnrmr/superdesk-core,mugurrus/superdesk-core,superdesk/superdesk-core,superdesk/superdesk-core,mdhaman/superdesk-core,hlmnrmr/superdesk-core,petrjasek/superdesk-core |
import superdesk
from flask import current_app as app
from superdesk.utils import ListCursor
class ClientConfigResource(superdesk.Resource):
item_methods = []
public_methods = ['GET']
resource_methods = ['GET']
class ClientConfigService(superdesk.Service):
def get(self, req, lookup):
return ListCursor()
def on_fetched(self, docs):
docs['config'] = getattr(app, 'client_config', {})
def init_app(app):
superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app)
app.client_config.update({
'schema': app.config.get('SCHEMA'),
'editor': app.config.get('EDITOR'),
+ 'feedback_url': app.config.get('FEEDBACK_URL'),
})
| Add feedback url to served client configuration | ## Code Before:
import superdesk
from flask import current_app as app
from superdesk.utils import ListCursor
class ClientConfigResource(superdesk.Resource):
item_methods = []
public_methods = ['GET']
resource_methods = ['GET']
class ClientConfigService(superdesk.Service):
def get(self, req, lookup):
return ListCursor()
def on_fetched(self, docs):
docs['config'] = getattr(app, 'client_config', {})
def init_app(app):
superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app)
app.client_config.update({
'schema': app.config.get('SCHEMA'),
'editor': app.config.get('EDITOR'),
})
## Instruction:
Add feedback url to served client configuration
## Code After:
import superdesk
from flask import current_app as app
from superdesk.utils import ListCursor
class ClientConfigResource(superdesk.Resource):
item_methods = []
public_methods = ['GET']
resource_methods = ['GET']
class ClientConfigService(superdesk.Service):
def get(self, req, lookup):
return ListCursor()
def on_fetched(self, docs):
docs['config'] = getattr(app, 'client_config', {})
def init_app(app):
superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app)
app.client_config.update({
'schema': app.config.get('SCHEMA'),
'editor': app.config.get('EDITOR'),
'feedback_url': app.config.get('FEEDBACK_URL'),
})
|
import superdesk
from flask import current_app as app
from superdesk.utils import ListCursor
class ClientConfigResource(superdesk.Resource):
item_methods = []
public_methods = ['GET']
resource_methods = ['GET']
class ClientConfigService(superdesk.Service):
def get(self, req, lookup):
return ListCursor()
def on_fetched(self, docs):
docs['config'] = getattr(app, 'client_config', {})
def init_app(app):
superdesk.register_resource('client_config', ClientConfigResource, ClientConfigService, _app=app)
app.client_config.update({
'schema': app.config.get('SCHEMA'),
'editor': app.config.get('EDITOR'),
+ 'feedback_url': app.config.get('FEEDBACK_URL'),
}) |
67d4f376586c912f852b98c75f7de04aeb05979a | pag/words.py | pag/words.py | """Get words from files in "src/dictionary/"."""
import os
def get_word_list(filepath):
"""
Get a list of words from a file.
Input: file name
Output: dict with formula {word: [synonym, synonym]}"""
filepath = os.path.abspath(filepath)
assert os.path.isfile(filepath), 'Must be a file'
f = open(filepath, 'r')
contents = f.read()
txt = contents.strip().split('\n')
if ':' in contents:
ntxt = txt[:]
for line in txt:
if line[0] == '#':
ntxt.remove(ntxt[ntxt.index(line)])
elif ':' not in line:
ntxt[ntxt.index(line)] = line + ':'
txt = ntxt
words = {}
for line in txt:
index = line.split(':')[0]
words[index] = line.split(':')[1].split(',')
for syn in words[index]:
if syn == '':
words[index].remove(syn)
else:
words = [word.strip() for word in txt]
f.close()
return words
verbs = get_word_list('dictionary/verbs.txt')
nouns = get_word_list('dictionary/nouns.txt')
extras = get_word_list('dictionary/extras.txt')
directions = get_word_list('dictionary/directions.txt')
| """Get words from files in "src/dictionary/"."""
import os
def get_word_list(filepath):
"""
Get a list of words from a file.
Input: file name
Output: dict with formula {word: [synonym, synonym]}"""
filepath = os.path.abspath(filepath)
assert os.path.isfile(filepath), 'Must be a file'
f = open(filepath, 'r')
contents = f.read()
txt = contents.strip().split('\n')
ntxt = txt[:]
for line in txt:
if line[0] == '#':
ntxt.remove(ntxt[ntxt.index(line)])
elif ':' not in line:
ntxt[ntxt.index(line)] = line + ':'
txt = ntxt
words = {}
for line in txt:
index = line.split(':')[0]
words[index] = line.split(':')[1].split(',')
for syn in words[index]:
if syn == '':
words[index].remove(syn)
f.close()
return words
verbs = get_word_list('dictionary/verbs.txt')
nouns = get_word_list('dictionary/nouns.txt')
extras = get_word_list('dictionary/extras.txt')
directions = get_word_list('dictionary/directions.txt')
| Remove useless and confusing code | Remove useless and confusing code
| Python | mit | allanburleson/python-adventure-game,disorientedperson/python-adventure-game | """Get words from files in "src/dictionary/"."""
import os
def get_word_list(filepath):
"""
Get a list of words from a file.
Input: file name
Output: dict with formula {word: [synonym, synonym]}"""
filepath = os.path.abspath(filepath)
assert os.path.isfile(filepath), 'Must be a file'
f = open(filepath, 'r')
contents = f.read()
txt = contents.strip().split('\n')
- if ':' in contents:
- ntxt = txt[:]
+ ntxt = txt[:]
- for line in txt:
+ for line in txt:
- if line[0] == '#':
+ if line[0] == '#':
- ntxt.remove(ntxt[ntxt.index(line)])
+ ntxt.remove(ntxt[ntxt.index(line)])
- elif ':' not in line:
+ elif ':' not in line:
- ntxt[ntxt.index(line)] = line + ':'
+ ntxt[ntxt.index(line)] = line + ':'
- txt = ntxt
+ txt = ntxt
- words = {}
+ words = {}
- for line in txt:
+ for line in txt:
- index = line.split(':')[0]
+ index = line.split(':')[0]
- words[index] = line.split(':')[1].split(',')
+ words[index] = line.split(':')[1].split(',')
- for syn in words[index]:
+ for syn in words[index]:
- if syn == '':
+ if syn == '':
- words[index].remove(syn)
+ words[index].remove(syn)
- else:
- words = [word.strip() for word in txt]
f.close()
return words
verbs = get_word_list('dictionary/verbs.txt')
nouns = get_word_list('dictionary/nouns.txt')
extras = get_word_list('dictionary/extras.txt')
directions = get_word_list('dictionary/directions.txt')
| Remove useless and confusing code | ## Code Before:
"""Get words from files in "src/dictionary/"."""
import os
def get_word_list(filepath):
"""
Get a list of words from a file.
Input: file name
Output: dict with formula {word: [synonym, synonym]}"""
filepath = os.path.abspath(filepath)
assert os.path.isfile(filepath), 'Must be a file'
f = open(filepath, 'r')
contents = f.read()
txt = contents.strip().split('\n')
if ':' in contents:
ntxt = txt[:]
for line in txt:
if line[0] == '#':
ntxt.remove(ntxt[ntxt.index(line)])
elif ':' not in line:
ntxt[ntxt.index(line)] = line + ':'
txt = ntxt
words = {}
for line in txt:
index = line.split(':')[0]
words[index] = line.split(':')[1].split(',')
for syn in words[index]:
if syn == '':
words[index].remove(syn)
else:
words = [word.strip() for word in txt]
f.close()
return words
verbs = get_word_list('dictionary/verbs.txt')
nouns = get_word_list('dictionary/nouns.txt')
extras = get_word_list('dictionary/extras.txt')
directions = get_word_list('dictionary/directions.txt')
## Instruction:
Remove useless and confusing code
## Code After:
"""Get words from files in "src/dictionary/"."""
import os
def get_word_list(filepath):
"""
Get a list of words from a file.
Input: file name
Output: dict with formula {word: [synonym, synonym]}"""
filepath = os.path.abspath(filepath)
assert os.path.isfile(filepath), 'Must be a file'
f = open(filepath, 'r')
contents = f.read()
txt = contents.strip().split('\n')
ntxt = txt[:]
for line in txt:
if line[0] == '#':
ntxt.remove(ntxt[ntxt.index(line)])
elif ':' not in line:
ntxt[ntxt.index(line)] = line + ':'
txt = ntxt
words = {}
for line in txt:
index = line.split(':')[0]
words[index] = line.split(':')[1].split(',')
for syn in words[index]:
if syn == '':
words[index].remove(syn)
f.close()
return words
verbs = get_word_list('dictionary/verbs.txt')
nouns = get_word_list('dictionary/nouns.txt')
extras = get_word_list('dictionary/extras.txt')
directions = get_word_list('dictionary/directions.txt')
| """Get words from files in "src/dictionary/"."""
import os
def get_word_list(filepath):
"""
Get a list of words from a file.
Input: file name
Output: dict with formula {word: [synonym, synonym]}"""
filepath = os.path.abspath(filepath)
assert os.path.isfile(filepath), 'Must be a file'
f = open(filepath, 'r')
contents = f.read()
txt = contents.strip().split('\n')
- if ':' in contents:
- ntxt = txt[:]
? ----
+ ntxt = txt[:]
- for line in txt:
? ----
+ for line in txt:
- if line[0] == '#':
? ----
+ if line[0] == '#':
- ntxt.remove(ntxt[ntxt.index(line)])
? ----
+ ntxt.remove(ntxt[ntxt.index(line)])
- elif ':' not in line:
? ----
+ elif ':' not in line:
- ntxt[ntxt.index(line)] = line + ':'
? ----
+ ntxt[ntxt.index(line)] = line + ':'
- txt = ntxt
? ----
+ txt = ntxt
- words = {}
? ----
+ words = {}
- for line in txt:
? ----
+ for line in txt:
- index = line.split(':')[0]
? ----
+ index = line.split(':')[0]
- words[index] = line.split(':')[1].split(',')
? ----
+ words[index] = line.split(':')[1].split(',')
- for syn in words[index]:
? ----
+ for syn in words[index]:
- if syn == '':
? ----
+ if syn == '':
- words[index].remove(syn)
? ----
+ words[index].remove(syn)
- else:
- words = [word.strip() for word in txt]
f.close()
return words
verbs = get_word_list('dictionary/verbs.txt')
nouns = get_word_list('dictionary/nouns.txt')
extras = get_word_list('dictionary/extras.txt')
directions = get_word_list('dictionary/directions.txt') |
4848baf76e4972401530b624816ba48cb08d9398 | appconf/utils.py | appconf/utils.py | import sys
def import_attribute(import_path, exception_handler=None):
from django.utils.importlib import import_module
module_name, object_name = import_path.rsplit('.', 1)
try:
module = import_module(module_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
try:
return getattr(module, object_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
| import sys
def import_attribute(import_path, exception_handler=None):
try:
from importlib import import_module
except ImportError:
from django.utils.importlib import import_module
module_name, object_name = import_path.rsplit('.', 1)
try:
module = import_module(module_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
try:
return getattr(module, object_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
| Use import_module from standard library if exists | Use import_module from standard library if exists
Django 1.8+ drops `django.utils.importlib`. I imagine because that is because an older version of Python (either 2.5 and/or 2.6) is being dropped. I haven't checked older versions but `importlib` exists in Python 2.7. | Python | bsd-3-clause | diox/django-appconf,carltongibson/django-appconf,django-compressor/django-appconf,jezdez/django-appconf,jessehon/django-appconf,treyhunner/django-appconf,jezdez-archive/django-appconf | import sys
def import_attribute(import_path, exception_handler=None):
+ try:
+ from importlib import import_module
+ except ImportError:
- from django.utils.importlib import import_module
+ from django.utils.importlib import import_module
module_name, object_name = import_path.rsplit('.', 1)
try:
module = import_module(module_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
try:
return getattr(module, object_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
| Use import_module from standard library if exists | ## Code Before:
import sys
def import_attribute(import_path, exception_handler=None):
from django.utils.importlib import import_module
module_name, object_name = import_path.rsplit('.', 1)
try:
module = import_module(module_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
try:
return getattr(module, object_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
## Instruction:
Use import_module from standard library if exists
## Code After:
import sys
def import_attribute(import_path, exception_handler=None):
try:
from importlib import import_module
except ImportError:
from django.utils.importlib import import_module
module_name, object_name = import_path.rsplit('.', 1)
try:
module = import_module(module_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
try:
return getattr(module, object_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
| import sys
def import_attribute(import_path, exception_handler=None):
+ try:
+ from importlib import import_module
+ except ImportError:
- from django.utils.importlib import import_module
+ from django.utils.importlib import import_module
? ++++
module_name, object_name = import_path.rsplit('.', 1)
try:
module = import_module(module_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise
try:
return getattr(module, object_name)
except: # pragma: no cover
if callable(exception_handler):
exctype, excvalue, tb = sys.exc_info()
return exception_handler(import_path, exctype, excvalue, tb)
else:
raise |
9cd3bb79126fa2431ba4ae03811ac30fb77b9b46 | netcat.py | netcat.py | import argparse
import socket
import sys
parser = argparse.ArgumentParser(description='Simple netcat in pure python.')
parser.add_argument('-z', '--scan', action='store_true')
parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int)
parser.add_argument('-v', '--verbose', action='store_true')
parser.add_argument('host')
parser.add_argument('port', type=int)
args = parser.parse_args()
if args.scan:
try:
connection = socket.create_connection((args.host, args.port), args.timeout)
if args.verbose:
print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)
sys.exit(0)
except socket.error as msg:
if args.verbose:
print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)
sys.exit(1)
else:
print 'Not implemented'
| import argparse
import socket
import sys
parser = argparse.ArgumentParser(description='Simple netcat in pure python.')
parser.add_argument('-s', '--source', metavar='ADDRESS')
parser.add_argument('-v', '--verbose', action='store_true')
parser.add_argument('-w', '--wait', metavar='SECONDS', type=int)
parser.add_argument('-z', '--zero', action='store_true')
parser.add_argument('host')
parser.add_argument('port')
args = parser.parse_args()
# Set a souce address for socket connection
source = ('', 0)
if args.source:
source = (args.source, 0) # port 0 specifies that the OS will choose a port
# exit successfully if the connection succeeds
if args.zero:
try:
connection = socket.create_connection((args.host, args.port), args.wait, source)
if args.verbose:
print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port))
sys.exit(0)
except socket.error as msg:
if args.verbose:
print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg))
sys.exit(1)
else:
print('Not implemented')
| Support python 2 and 3 | Support python 2 and 3
Add source argument.
Update arguments to use long names from GNU netcat.
| Python | unlicense | benformosa/Toolbox,benformosa/Toolbox | import argparse
import socket
import sys
parser = argparse.ArgumentParser(description='Simple netcat in pure python.')
+ parser.add_argument('-s', '--source', metavar='ADDRESS')
- parser.add_argument('-z', '--scan', action='store_true')
- parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int)
parser.add_argument('-v', '--verbose', action='store_true')
+ parser.add_argument('-w', '--wait', metavar='SECONDS', type=int)
+ parser.add_argument('-z', '--zero', action='store_true')
parser.add_argument('host')
- parser.add_argument('port', type=int)
+ parser.add_argument('port')
args = parser.parse_args()
+
+ # Set a souce address for socket connection
+ source = ('', 0)
- if args.scan:
+ if args.source:
+ source = (args.source, 0) # port 0 specifies that the OS will choose a port
+
+ # exit successfully if the connection succeeds
+ if args.zero:
try:
- connection = socket.create_connection((args.host, args.port), args.timeout)
+ connection = socket.create_connection((args.host, args.port), args.wait, source)
if args.verbose:
- print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)
+ print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port))
sys.exit(0)
except socket.error as msg:
if args.verbose:
- print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)
+ print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg))
sys.exit(1)
else:
- print 'Not implemented'
+ print('Not implemented')
| Support python 2 and 3 | ## Code Before:
import argparse
import socket
import sys
parser = argparse.ArgumentParser(description='Simple netcat in pure python.')
parser.add_argument('-z', '--scan', action='store_true')
parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int)
parser.add_argument('-v', '--verbose', action='store_true')
parser.add_argument('host')
parser.add_argument('port', type=int)
args = parser.parse_args()
if args.scan:
try:
connection = socket.create_connection((args.host, args.port), args.timeout)
if args.verbose:
print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)
sys.exit(0)
except socket.error as msg:
if args.verbose:
print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)
sys.exit(1)
else:
print 'Not implemented'
## Instruction:
Support python 2 and 3
## Code After:
import argparse
import socket
import sys
parser = argparse.ArgumentParser(description='Simple netcat in pure python.')
parser.add_argument('-s', '--source', metavar='ADDRESS')
parser.add_argument('-v', '--verbose', action='store_true')
parser.add_argument('-w', '--wait', metavar='SECONDS', type=int)
parser.add_argument('-z', '--zero', action='store_true')
parser.add_argument('host')
parser.add_argument('port')
args = parser.parse_args()
# Set a souce address for socket connection
source = ('', 0)
if args.source:
source = (args.source, 0) # port 0 specifies that the OS will choose a port
# exit successfully if the connection succeeds
if args.zero:
try:
connection = socket.create_connection((args.host, args.port), args.wait, source)
if args.verbose:
print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port))
sys.exit(0)
except socket.error as msg:
if args.verbose:
print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg))
sys.exit(1)
else:
print('Not implemented')
| import argparse
import socket
import sys
parser = argparse.ArgumentParser(description='Simple netcat in pure python.')
+ parser.add_argument('-s', '--source', metavar='ADDRESS')
- parser.add_argument('-z', '--scan', action='store_true')
- parser.add_argument('-w', '--timeout', metavar='SECONDS', type=int)
parser.add_argument('-v', '--verbose', action='store_true')
+ parser.add_argument('-w', '--wait', metavar='SECONDS', type=int)
+ parser.add_argument('-z', '--zero', action='store_true')
parser.add_argument('host')
- parser.add_argument('port', type=int)
? ----------
+ parser.add_argument('port')
args = parser.parse_args()
+
+ # Set a souce address for socket connection
+ source = ('', 0)
- if args.scan:
? ^^
+ if args.source:
? +++ ^
+ source = (args.source, 0) # port 0 specifies that the OS will choose a port
+
+ # exit successfully if the connection succeeds
+ if args.zero:
try:
- connection = socket.create_connection((args.host, args.port), args.timeout)
? ^^^ ^
+ connection = socket.create_connection((args.host, args.port), args.wait, source)
? +++ ^^^ ^^^
if args.verbose:
- print "Connection to {} {} port (tcp) succeeded!".format(args.host, args.port)
? ^
+ print("Connection to {} {} port (tcp) succeeded!".format(args.host, args.port))
? ^ +
sys.exit(0)
except socket.error as msg:
if args.verbose:
- print "Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg)
? ^
+ print("Connection to {} {} port (tcp) failed. {}".format(args.host, args.port, msg))
? ^ +
sys.exit(1)
else:
- print 'Not implemented'
? ^
+ print('Not implemented')
? ^ +
|
7ace27a6a114e381a30ac9760880b68277a868fc | python_scripts/mc_config.py | python_scripts/mc_config.py |
import yaml
def read_config():
yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb')
config_file = yaml.load( yml_file )
return config_file
|
import yaml
import os.path
_config_file_base_name = 'mediawords.yml'
_config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml'))
def read_config():
yml_file = open(_config_file_name, 'rb')
config_file = yaml.load( yml_file )
return config_file
| Use relative path location for mediawords.yml. | Use relative path location for mediawords.yml.
| Python | agpl-3.0 | berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud |
import yaml
+ import os.path
+
+ _config_file_base_name = 'mediawords.yml'
+ _config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml'))
def read_config():
- yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb')
+ yml_file = open(_config_file_name, 'rb')
config_file = yaml.load( yml_file )
return config_file
| Use relative path location for mediawords.yml. | ## Code Before:
import yaml
def read_config():
yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb')
config_file = yaml.load( yml_file )
return config_file
## Instruction:
Use relative path location for mediawords.yml.
## Code After:
import yaml
import os.path
_config_file_base_name = 'mediawords.yml'
_config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml'))
def read_config():
yml_file = open(_config_file_name, 'rb')
config_file = yaml.load( yml_file )
return config_file
|
import yaml
+ import os.path
+
+ _config_file_base_name = 'mediawords.yml'
+ _config_file_name = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'mediawords.yml'))
def read_config():
- yml_file = open('/home/dlarochelle/git_dev/mediacloud/mediawords.yml', 'rb')
+ yml_file = open(_config_file_name, 'rb')
config_file = yaml.load( yml_file )
return config_file
|
c1ed5befe3081f6812fc77fc694ea3e82d90f39c | telemetry/telemetry/core/backends/facebook_credentials_backend.py | telemetry/telemetry/core/backends/facebook_credentials_backend.py |
from telemetry.core.backends import form_based_credentials_backend
class FacebookCredentialsBackend(
form_based_credentials_backend.FormBasedCredentialsBackend):
@property
def logged_in_javascript(self):
"""Evaluates to true iff already logged in."""
return ('document.getElementById("fbNotificationsList")!== null || '
'document.getElementById("m_home_notice")!== null')
@property
def credentials_type(self):
return 'facebook'
@property
def url(self):
return 'http://www.facebook.com/'
@property
def login_form_id(self):
return 'login_form'
@property
def login_input_id(self):
return 'email'
@property
def password_input_id(self):
return 'pass'
class FacebookCredentialsBackend2(FacebookCredentialsBackend):
@property
def credentials_type(self):
return 'facebook2'
|
from telemetry.core.backends import form_based_credentials_backend
class FacebookCredentialsBackend(
form_based_credentials_backend.FormBasedCredentialsBackend):
@property
def logged_in_javascript(self):
"""Evaluates to true iff already logged in."""
return ('document.getElementById("fbNotificationsList")!== null || '
'document.getElementById("m_home_notice")!== null')
@property
def credentials_type(self):
return 'facebook'
@property
def url(self):
return 'http://www.facebook.com/'
@property
def login_form_id(self):
return 'login_form'
@property
def login_input_id(self):
return 'email'
@property
def password_input_id(self):
return 'pass'
class FacebookCredentialsBackend2(FacebookCredentialsBackend):
""" Facebook credential backend for https client. """
@property
def credentials_type(self):
return 'facebook2'
@property
def url(self):
return 'https://www.facebook.com/'
| Set facebook_crendentials_backend_2's url to https | [Telemetry] Set facebook_crendentials_backend_2's url to https
TBR=tonyg@chromium.org
BUG=428098
Review URL: https://codereview.chromium.org/688113003
Cr-Commit-Position: 972c6d2dc6dd5efdad1377c0d224e03eb8f276f7@{#301945}
| Python | bsd-3-clause | benschmaus/catapult,SummerLW/Perf-Insight-Report,sahiljain/catapult,SummerLW/Perf-Insight-Report,sahiljain/catapult,catapult-project/catapult,benschmaus/catapult,SummerLW/Perf-Insight-Report,catapult-project/catapult,sahiljain/catapult,catapult-project/catapult-csm,benschmaus/catapult,catapult-project/catapult-csm,sahiljain/catapult,catapult-project/catapult-csm,catapult-project/catapult-csm,SummerLW/Perf-Insight-Report,catapult-project/catapult,benschmaus/catapult,SummerLW/Perf-Insight-Report,catapult-project/catapult,SummerLW/Perf-Insight-Report,benschmaus/catapult,benschmaus/catapult,sahiljain/catapult,catapult-project/catapult-csm,catapult-project/catapult,catapult-project/catapult-csm,sahiljain/catapult,catapult-project/catapult-csm,catapult-project/catapult,benschmaus/catapult,catapult-project/catapult |
from telemetry.core.backends import form_based_credentials_backend
class FacebookCredentialsBackend(
form_based_credentials_backend.FormBasedCredentialsBackend):
@property
def logged_in_javascript(self):
"""Evaluates to true iff already logged in."""
return ('document.getElementById("fbNotificationsList")!== null || '
'document.getElementById("m_home_notice")!== null')
@property
def credentials_type(self):
return 'facebook'
@property
def url(self):
return 'http://www.facebook.com/'
@property
def login_form_id(self):
return 'login_form'
@property
def login_input_id(self):
return 'email'
@property
def password_input_id(self):
return 'pass'
class FacebookCredentialsBackend2(FacebookCredentialsBackend):
+ """ Facebook credential backend for https client. """
@property
def credentials_type(self):
return 'facebook2'
+ @property
+ def url(self):
+ return 'https://www.facebook.com/'
+ | Set facebook_crendentials_backend_2's url to https | ## Code Before:
from telemetry.core.backends import form_based_credentials_backend
class FacebookCredentialsBackend(
form_based_credentials_backend.FormBasedCredentialsBackend):
@property
def logged_in_javascript(self):
"""Evaluates to true iff already logged in."""
return ('document.getElementById("fbNotificationsList")!== null || '
'document.getElementById("m_home_notice")!== null')
@property
def credentials_type(self):
return 'facebook'
@property
def url(self):
return 'http://www.facebook.com/'
@property
def login_form_id(self):
return 'login_form'
@property
def login_input_id(self):
return 'email'
@property
def password_input_id(self):
return 'pass'
class FacebookCredentialsBackend2(FacebookCredentialsBackend):
@property
def credentials_type(self):
return 'facebook2'
## Instruction:
Set facebook_crendentials_backend_2's url to https
## Code After:
from telemetry.core.backends import form_based_credentials_backend
class FacebookCredentialsBackend(
form_based_credentials_backend.FormBasedCredentialsBackend):
@property
def logged_in_javascript(self):
"""Evaluates to true iff already logged in."""
return ('document.getElementById("fbNotificationsList")!== null || '
'document.getElementById("m_home_notice")!== null')
@property
def credentials_type(self):
return 'facebook'
@property
def url(self):
return 'http://www.facebook.com/'
@property
def login_form_id(self):
return 'login_form'
@property
def login_input_id(self):
return 'email'
@property
def password_input_id(self):
return 'pass'
class FacebookCredentialsBackend2(FacebookCredentialsBackend):
""" Facebook credential backend for https client. """
@property
def credentials_type(self):
return 'facebook2'
@property
def url(self):
return 'https://www.facebook.com/'
|
from telemetry.core.backends import form_based_credentials_backend
class FacebookCredentialsBackend(
form_based_credentials_backend.FormBasedCredentialsBackend):
@property
def logged_in_javascript(self):
"""Evaluates to true iff already logged in."""
return ('document.getElementById("fbNotificationsList")!== null || '
'document.getElementById("m_home_notice")!== null')
@property
def credentials_type(self):
return 'facebook'
@property
def url(self):
return 'http://www.facebook.com/'
@property
def login_form_id(self):
return 'login_form'
@property
def login_input_id(self):
return 'email'
@property
def password_input_id(self):
return 'pass'
class FacebookCredentialsBackend2(FacebookCredentialsBackend):
+ """ Facebook credential backend for https client. """
@property
def credentials_type(self):
return 'facebook2'
+
+ @property
+ def url(self):
+ return 'https://www.facebook.com/' |
18fec1124bb86f90183350e7b9c86eb946a01884 | whatchanged/main.py | whatchanged/main.py |
from __future__ import absolute_import, print_function
# Standard library
from os import walk
from os.path import exists, isdir, join
# Local library
from .util import is_py_file
from .diff import diff_files
def main():
import sys
if sys.argv < 3:
print('Usage: %s <module1> <module2>' % sys.argv[0])
sys.exit(1)
old, new = sys.argv[1:3]
diff = set([])
if isdir(old):
assert isdir(new)
for dirpath, dirnames, filenames in walk(new):
for file_ in filenames:
if is_py_file(file_):
new_file = join(dirpath, file_)
old_file = new_file.replace(new, old)
if exists(old_file):
mdiff = diff_files(old_file, new_file)
if mdiff is not None:
diff.add(mdiff)
else:
diff.add(diff_files(old, new))
for module in diff:
print(module)
if __name__ == '__main__':
main()
|
from __future__ import absolute_import, print_function
# Standard library
from os import walk
from os.path import exists, isdir, join
# Local library
from .util import is_py_file
from .diff import diff_files
def main():
import sys
if len(sys.argv) < 3:
print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0])
sys.exit(1)
old, new = sys.argv[1:3]
diff = set([])
if isdir(old):
assert isdir(new)
for dirpath, dirnames, filenames in walk(new):
for file_ in filenames:
if is_py_file(file_):
new_file = join(dirpath, file_)
old_file = new_file.replace(new, old)
if exists(old_file):
mdiff = diff_files(old_file, new_file)
if mdiff is not None:
diff.add(mdiff)
else:
diff.add(diff_files(old, new))
for module in diff:
print(module)
if __name__ == '__main__':
main()
| Fix minor bug in length comparison. | Fix minor bug in length comparison.
| Python | bsd-2-clause | punchagan/what-changed |
from __future__ import absolute_import, print_function
# Standard library
from os import walk
from os.path import exists, isdir, join
# Local library
from .util import is_py_file
from .diff import diff_files
def main():
import sys
- if sys.argv < 3:
+ if len(sys.argv) < 3:
- print('Usage: %s <module1> <module2>' % sys.argv[0])
+ print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0])
sys.exit(1)
old, new = sys.argv[1:3]
diff = set([])
if isdir(old):
assert isdir(new)
for dirpath, dirnames, filenames in walk(new):
for file_ in filenames:
if is_py_file(file_):
new_file = join(dirpath, file_)
old_file = new_file.replace(new, old)
if exists(old_file):
mdiff = diff_files(old_file, new_file)
if mdiff is not None:
diff.add(mdiff)
else:
diff.add(diff_files(old, new))
for module in diff:
print(module)
if __name__ == '__main__':
main()
| Fix minor bug in length comparison. | ## Code Before:
from __future__ import absolute_import, print_function
# Standard library
from os import walk
from os.path import exists, isdir, join
# Local library
from .util import is_py_file
from .diff import diff_files
def main():
import sys
if sys.argv < 3:
print('Usage: %s <module1> <module2>' % sys.argv[0])
sys.exit(1)
old, new = sys.argv[1:3]
diff = set([])
if isdir(old):
assert isdir(new)
for dirpath, dirnames, filenames in walk(new):
for file_ in filenames:
if is_py_file(file_):
new_file = join(dirpath, file_)
old_file = new_file.replace(new, old)
if exists(old_file):
mdiff = diff_files(old_file, new_file)
if mdiff is not None:
diff.add(mdiff)
else:
diff.add(diff_files(old, new))
for module in diff:
print(module)
if __name__ == '__main__':
main()
## Instruction:
Fix minor bug in length comparison.
## Code After:
from __future__ import absolute_import, print_function
# Standard library
from os import walk
from os.path import exists, isdir, join
# Local library
from .util import is_py_file
from .diff import diff_files
def main():
import sys
if len(sys.argv) < 3:
print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0])
sys.exit(1)
old, new = sys.argv[1:3]
diff = set([])
if isdir(old):
assert isdir(new)
for dirpath, dirnames, filenames in walk(new):
for file_ in filenames:
if is_py_file(file_):
new_file = join(dirpath, file_)
old_file = new_file.replace(new, old)
if exists(old_file):
mdiff = diff_files(old_file, new_file)
if mdiff is not None:
diff.add(mdiff)
else:
diff.add(diff_files(old, new))
for module in diff:
print(module)
if __name__ == '__main__':
main()
|
from __future__ import absolute_import, print_function
# Standard library
from os import walk
from os.path import exists, isdir, join
# Local library
from .util import is_py_file
from .diff import diff_files
def main():
import sys
- if sys.argv < 3:
+ if len(sys.argv) < 3:
? ++++ +
- print('Usage: %s <module1> <module2>' % sys.argv[0])
+ print('Usage: %s <package1/module1> <package2/module2>' % sys.argv[0])
? +++++++++ +++++++++
sys.exit(1)
old, new = sys.argv[1:3]
diff = set([])
if isdir(old):
assert isdir(new)
for dirpath, dirnames, filenames in walk(new):
for file_ in filenames:
if is_py_file(file_):
new_file = join(dirpath, file_)
old_file = new_file.replace(new, old)
if exists(old_file):
mdiff = diff_files(old_file, new_file)
if mdiff is not None:
diff.add(mdiff)
else:
diff.add(diff_files(old, new))
for module in diff:
print(module)
if __name__ == '__main__':
main() |
d7fdebdc4ce52e59c126a27ea06171994a6c846b | src/config/common/ssl_adapter.py | src/config/common/ssl_adapter.py | # -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
# @author: Sanju Abraham, Juniper Networks, OpenContrail
from requests.adapters import HTTPAdapter
try:
# This is required for RDO, which installs both python-requests
# and python-urllib3, but symlinks python-request's internally packaged
# urllib3 to the site installed one.
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
# Fallback to standard installation methods
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that can be configured with SSL/TLS
version.'''
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
self.poolmanager = None
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
| # -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
# @author: Sanju Abraham, Juniper Networks, OpenContrail
from requests.adapters import HTTPAdapter
try:
# This is required for RDO, which installs both python-requests
# and python-urllib3, but symlinks python-request's internally packaged
# urllib3 to the site installed one.
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
# Fallback to standard installation methods
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that can be configured with SSL/TLS
version.'''
HTTPAdapter.__attrs__.extend(['ssl_version'])
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
self.poolmanager = None
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
| Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module. This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__. | Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module.
This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__.
Courtesy: https://github.com/sigmavirus24/requests-toolbelt/commit/decadbd3512444889feb30cf1ff2f1448a3ecfca
Closes-Bug:#1604247
Change-Id: Iee9e0348c005e88c535f4da33cf98149a8c1b19d
| Python | apache-2.0 | codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller,codilime/contrail-controller | # -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
# @author: Sanju Abraham, Juniper Networks, OpenContrail
from requests.adapters import HTTPAdapter
try:
# This is required for RDO, which installs both python-requests
# and python-urllib3, but symlinks python-request's internally packaged
# urllib3 to the site installed one.
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
# Fallback to standard installation methods
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that can be configured with SSL/TLS
version.'''
+ HTTPAdapter.__attrs__.extend(['ssl_version'])
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
self.poolmanager = None
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
| Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module. This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__. | ## Code Before:
# -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
# @author: Sanju Abraham, Juniper Networks, OpenContrail
from requests.adapters import HTTPAdapter
try:
# This is required for RDO, which installs both python-requests
# and python-urllib3, but symlinks python-request's internally packaged
# urllib3 to the site installed one.
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
# Fallback to standard installation methods
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that can be configured with SSL/TLS
version.'''
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
self.poolmanager = None
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
## Instruction:
Add ssl_version to the list of attributes, required when vnc_api gets called via multiprocessing module. This will ensure ssl_version gets included when pickle calls __getstate__ and __setstate__.
## Code After:
# -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
# @author: Sanju Abraham, Juniper Networks, OpenContrail
from requests.adapters import HTTPAdapter
try:
# This is required for RDO, which installs both python-requests
# and python-urllib3, but symlinks python-request's internally packaged
# urllib3 to the site installed one.
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
# Fallback to standard installation methods
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that can be configured with SSL/TLS
version.'''
HTTPAdapter.__attrs__.extend(['ssl_version'])
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
self.poolmanager = None
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
| # -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
# @author: Sanju Abraham, Juniper Networks, OpenContrail
from requests.adapters import HTTPAdapter
try:
# This is required for RDO, which installs both python-requests
# and python-urllib3, but symlinks python-request's internally packaged
# urllib3 to the site installed one.
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
# Fallback to standard installation methods
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that can be configured with SSL/TLS
version.'''
+ HTTPAdapter.__attrs__.extend(['ssl_version'])
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
self.poolmanager = None
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version) |
a0e8c92a9d12846c8cfe6819ea26d1e08dd4098a | example/models.py | example/models.py | import i18n
from i18n.models import TranslatableModel
class Document(TranslatableModel):
charfield = i18n.LocalizedCharField(max_length=50)
textfield = i18n.LocalizedTextField(max_length=512)
filefield = i18n.LocalizedFileField(null=True, upload_to='files')
imagefield = i18n.LocalizedImageField(null=True, upload_to='images')
booleanfield = i18n.LocalizedBooleanField()
datefield = i18n.LocalizedDateField()
fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True,
related_name='+')
urlfied = i18n.LocalizedURLField()
decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2)
integerfield = i18n.LocalizedIntegerField()
def __str__(self):
return '%d, %s' % (self.pk, self.charfield)
class Meta:
app_label = 'example'
| from django.db import models
import i18n
from i18n.models import TranslatableModel
class Document(TranslatableModel):
untranslated_charfield = models.CharField(max_length=50, blank=True)
charfield = i18n.LocalizedCharField(max_length=50)
textfield = i18n.LocalizedTextField(max_length=500, blank=True)
filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True)
imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True)
booleanfield = i18n.LocalizedBooleanField()
datefield = i18n.LocalizedDateField(blank=True, null=True)
fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True,
related_name='+')
urlfied = i18n.LocalizedURLField(null=True, blank=True)
decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True,
blank=True)
integerfield = i18n.LocalizedIntegerField(null=True, blank=True)
def __str__(self):
return '%d, %s' % (self.pk, self.charfield)
class Meta:
app_label = 'example'
| Make fields in example app non required | Make fields in example app non required
| Python | bsd-3-clause | jonasundderwolf/django-localizedfields,jonasundderwolf/django-localizedfields | + from django.db import models
import i18n
from i18n.models import TranslatableModel
class Document(TranslatableModel):
-
+ untranslated_charfield = models.CharField(max_length=50, blank=True)
charfield = i18n.LocalizedCharField(max_length=50)
- textfield = i18n.LocalizedTextField(max_length=512)
+ textfield = i18n.LocalizedTextField(max_length=500, blank=True)
- filefield = i18n.LocalizedFileField(null=True, upload_to='files')
+ filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True)
- imagefield = i18n.LocalizedImageField(null=True, upload_to='images')
+ imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True)
booleanfield = i18n.LocalizedBooleanField()
- datefield = i18n.LocalizedDateField()
+ datefield = i18n.LocalizedDateField(blank=True, null=True)
fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True,
related_name='+')
- urlfied = i18n.LocalizedURLField()
+ urlfied = i18n.LocalizedURLField(null=True, blank=True)
- decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2)
+ decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True,
+ blank=True)
- integerfield = i18n.LocalizedIntegerField()
+ integerfield = i18n.LocalizedIntegerField(null=True, blank=True)
def __str__(self):
return '%d, %s' % (self.pk, self.charfield)
class Meta:
app_label = 'example'
| Make fields in example app non required | ## Code Before:
import i18n
from i18n.models import TranslatableModel
class Document(TranslatableModel):
charfield = i18n.LocalizedCharField(max_length=50)
textfield = i18n.LocalizedTextField(max_length=512)
filefield = i18n.LocalizedFileField(null=True, upload_to='files')
imagefield = i18n.LocalizedImageField(null=True, upload_to='images')
booleanfield = i18n.LocalizedBooleanField()
datefield = i18n.LocalizedDateField()
fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True,
related_name='+')
urlfied = i18n.LocalizedURLField()
decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2)
integerfield = i18n.LocalizedIntegerField()
def __str__(self):
return '%d, %s' % (self.pk, self.charfield)
class Meta:
app_label = 'example'
## Instruction:
Make fields in example app non required
## Code After:
from django.db import models
import i18n
from i18n.models import TranslatableModel
class Document(TranslatableModel):
untranslated_charfield = models.CharField(max_length=50, blank=True)
charfield = i18n.LocalizedCharField(max_length=50)
textfield = i18n.LocalizedTextField(max_length=500, blank=True)
filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True)
imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True)
booleanfield = i18n.LocalizedBooleanField()
datefield = i18n.LocalizedDateField(blank=True, null=True)
fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True,
related_name='+')
urlfied = i18n.LocalizedURLField(null=True, blank=True)
decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True,
blank=True)
integerfield = i18n.LocalizedIntegerField(null=True, blank=True)
def __str__(self):
return '%d, %s' % (self.pk, self.charfield)
class Meta:
app_label = 'example'
| + from django.db import models
import i18n
from i18n.models import TranslatableModel
class Document(TranslatableModel):
-
+ untranslated_charfield = models.CharField(max_length=50, blank=True)
charfield = i18n.LocalizedCharField(max_length=50)
- textfield = i18n.LocalizedTextField(max_length=512)
? ^^
+ textfield = i18n.LocalizedTextField(max_length=500, blank=True)
? ^^^^^^^^^^^^^^
- filefield = i18n.LocalizedFileField(null=True, upload_to='files')
+ filefield = i18n.LocalizedFileField(null=True, upload_to='files', blank=True)
? ++++++++++++
- imagefield = i18n.LocalizedImageField(null=True, upload_to='images')
+ imagefield = i18n.LocalizedImageField(null=True, upload_to='images', blank=True)
? ++++++++++++
booleanfield = i18n.LocalizedBooleanField()
- datefield = i18n.LocalizedDateField()
+ datefield = i18n.LocalizedDateField(blank=True, null=True)
? +++++++++++++++++++++
fkfield = i18n.LocalizedForeignKey('self', null=True, blank=True,
related_name='+')
- urlfied = i18n.LocalizedURLField()
+ urlfied = i18n.LocalizedURLField(null=True, blank=True)
? +++++++++++++++++++++
- decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2)
? ^
+ decimalfield = i18n.LocalizedDecimalField(max_digits=4, decimal_places=2, null=True,
? ^^^^^^^^^^^^
+ blank=True)
- integerfield = i18n.LocalizedIntegerField()
+ integerfield = i18n.LocalizedIntegerField(null=True, blank=True)
? +++++++++++++++++++++
def __str__(self):
return '%d, %s' % (self.pk, self.charfield)
class Meta:
app_label = 'example' |
a31e62f2a981f7662aee8a35ad195252a542d08d | plugins/say.py | plugins/say.py | from motobot import command, action
@command('say')
def say_command(bot, message, database):
masters = [
"Moto-chan",
"Motoko11",
"Akahige",
"betholas",
"Baradium",
"Cold_slither",
"Drahken"
]
if message.nick.lower() not in [x.lower() for x in masters]:
return "Check your privilege!"
else:
args = message.message.split(' ')[1:]
if len(args) < 2:
return "You must specify both a channel and a message"
else:
channel = args[0]
message = ' '.join(args[1:])
if message.startswith('/me '):
message = action(message[4:])
bot.send('PRIVMSG {} :{}'.format(channel, message))
| from motobot import command, action
@command('say')
def say_command(bot, message, database):
masters = [
"Moto-chan",
"Motoko11",
"MotoNyan",
"Akahige",
"betholas",
"Baradium",
"Cold_slither",
"Drahken"
]
if message.nick.lower() not in [x.lower() for x in masters]:
return "Check your privilege!"
else:
args = message.message.split(' ')[1:]
if len(args) < 2:
return "You must specify both a channel and a message"
else:
channel = args[0]
message = ' '.join(args[1:])
if message.startswith('/me '):
message = action(message[4:])
bot.send('PRIVMSG {} :{}'.format(channel, message))
| Add MotoNyan to mad hax | Add MotoNyan to mad hax
| Python | mit | Motoko11/MotoBot | from motobot import command, action
@command('say')
def say_command(bot, message, database):
masters = [
"Moto-chan",
"Motoko11",
+ "MotoNyan",
"Akahige",
"betholas",
"Baradium",
"Cold_slither",
"Drahken"
]
if message.nick.lower() not in [x.lower() for x in masters]:
return "Check your privilege!"
else:
args = message.message.split(' ')[1:]
if len(args) < 2:
return "You must specify both a channel and a message"
else:
channel = args[0]
message = ' '.join(args[1:])
if message.startswith('/me '):
message = action(message[4:])
bot.send('PRIVMSG {} :{}'.format(channel, message))
| Add MotoNyan to mad hax | ## Code Before:
from motobot import command, action
@command('say')
def say_command(bot, message, database):
masters = [
"Moto-chan",
"Motoko11",
"Akahige",
"betholas",
"Baradium",
"Cold_slither",
"Drahken"
]
if message.nick.lower() not in [x.lower() for x in masters]:
return "Check your privilege!"
else:
args = message.message.split(' ')[1:]
if len(args) < 2:
return "You must specify both a channel and a message"
else:
channel = args[0]
message = ' '.join(args[1:])
if message.startswith('/me '):
message = action(message[4:])
bot.send('PRIVMSG {} :{}'.format(channel, message))
## Instruction:
Add MotoNyan to mad hax
## Code After:
from motobot import command, action
@command('say')
def say_command(bot, message, database):
masters = [
"Moto-chan",
"Motoko11",
"MotoNyan",
"Akahige",
"betholas",
"Baradium",
"Cold_slither",
"Drahken"
]
if message.nick.lower() not in [x.lower() for x in masters]:
return "Check your privilege!"
else:
args = message.message.split(' ')[1:]
if len(args) < 2:
return "You must specify both a channel and a message"
else:
channel = args[0]
message = ' '.join(args[1:])
if message.startswith('/me '):
message = action(message[4:])
bot.send('PRIVMSG {} :{}'.format(channel, message))
| from motobot import command, action
@command('say')
def say_command(bot, message, database):
masters = [
"Moto-chan",
"Motoko11",
+ "MotoNyan",
"Akahige",
"betholas",
"Baradium",
"Cold_slither",
"Drahken"
]
if message.nick.lower() not in [x.lower() for x in masters]:
return "Check your privilege!"
else:
args = message.message.split(' ')[1:]
if len(args) < 2:
return "You must specify both a channel and a message"
else:
channel = args[0]
message = ' '.join(args[1:])
if message.startswith('/me '):
message = action(message[4:])
bot.send('PRIVMSG {} :{}'.format(channel, message)) |
ef0a6968dedad74ddd40bd4ae81595be6092f24f | wrapper/__init__.py | wrapper/__init__.py | __version__ = '2.2.0'
from libsbol import *
import unit_tests | from __future__ import absolute_import
__version__ = '2.2.0'
from sbol.libsbol import *
import sbol.unit_tests | Fix import issue with Python 3.6/Support future Python by forcing absolute import | Fix import issue with Python 3.6/Support future Python by forcing absolute import
| Python | apache-2.0 | SynBioDex/libSBOL,SynBioDex/libSBOL,SynBioDex/libSBOL,SynBioDex/libSBOL | + from __future__ import absolute_import
+
__version__ = '2.2.0'
- from libsbol import *
+ from sbol.libsbol import *
- import unit_tests
+ import sbol.unit_tests | Fix import issue with Python 3.6/Support future Python by forcing absolute import | ## Code Before:
__version__ = '2.2.0'
from libsbol import *
import unit_tests
## Instruction:
Fix import issue with Python 3.6/Support future Python by forcing absolute import
## Code After:
from __future__ import absolute_import
__version__ = '2.2.0'
from sbol.libsbol import *
import sbol.unit_tests | + from __future__ import absolute_import
+
__version__ = '2.2.0'
- from libsbol import *
+ from sbol.libsbol import *
? +++++
- import unit_tests
+ import sbol.unit_tests
? +++++
|
312c0d463940257cb1f777d3720778550b5bdb2d | bluebottle/organizations/serializers.py | bluebottle/organizations/serializers.py | from rest_framework import serializers
from bluebottle.organizations.models import Organization
from bluebottle.utils.serializers import URLField
class OrganizationSerializer(serializers.ModelSerializer):
class Meta:
model = Organization
fields = ('id', 'name', 'slug', 'address_line1', 'address_line2',
'city', 'state', 'country', 'postal_code', 'phone_number',
'website', 'email')
class ManageOrganizationSerializer(serializers.ModelSerializer):
slug = serializers.SlugField(required=False, allow_null=True)
name = serializers.CharField(required=True)
website = URLField(required=False, allow_blank=True)
email = serializers.EmailField(required=False, allow_blank=True)
class Meta:
model = Organization
fields = OrganizationSerializer.Meta.fields + ('partner_organizations',
'created', 'updated')
| from rest_framework import serializers
from bluebottle.organizations.models import Organization
from bluebottle.utils.serializers import URLField
class OrganizationSerializer(serializers.ModelSerializer):
class Meta:
model = Organization
fields = ('id', 'name', 'slug', 'address_line1', 'address_line2',
'city', 'state', 'country', 'postal_code', 'phone_number',
'website', 'email')
class ManageOrganizationSerializer(serializers.ModelSerializer):
slug = serializers.SlugField(required=False, allow_null=True)
name = serializers.CharField(required=True, allow_blank=True)
website = URLField(required=False, allow_blank=True)
email = serializers.EmailField(required=False, allow_blank=True)
class Meta:
model = Organization
fields = OrganizationSerializer.Meta.fields + ('partner_organizations',
'created', 'updated')
| Revert "Make the name of an organization required" | Revert "Make the name of an organization required"
This reverts commit 02140561a29a2b7fe50f7bf2402da566e60be641.
| Python | bsd-3-clause | jfterpstra/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle | from rest_framework import serializers
from bluebottle.organizations.models import Organization
from bluebottle.utils.serializers import URLField
class OrganizationSerializer(serializers.ModelSerializer):
class Meta:
model = Organization
fields = ('id', 'name', 'slug', 'address_line1', 'address_line2',
'city', 'state', 'country', 'postal_code', 'phone_number',
'website', 'email')
class ManageOrganizationSerializer(serializers.ModelSerializer):
slug = serializers.SlugField(required=False, allow_null=True)
- name = serializers.CharField(required=True)
+ name = serializers.CharField(required=True, allow_blank=True)
website = URLField(required=False, allow_blank=True)
email = serializers.EmailField(required=False, allow_blank=True)
class Meta:
model = Organization
fields = OrganizationSerializer.Meta.fields + ('partner_organizations',
'created', 'updated')
| Revert "Make the name of an organization required" | ## Code Before:
from rest_framework import serializers
from bluebottle.organizations.models import Organization
from bluebottle.utils.serializers import URLField
class OrganizationSerializer(serializers.ModelSerializer):
class Meta:
model = Organization
fields = ('id', 'name', 'slug', 'address_line1', 'address_line2',
'city', 'state', 'country', 'postal_code', 'phone_number',
'website', 'email')
class ManageOrganizationSerializer(serializers.ModelSerializer):
slug = serializers.SlugField(required=False, allow_null=True)
name = serializers.CharField(required=True)
website = URLField(required=False, allow_blank=True)
email = serializers.EmailField(required=False, allow_blank=True)
class Meta:
model = Organization
fields = OrganizationSerializer.Meta.fields + ('partner_organizations',
'created', 'updated')
## Instruction:
Revert "Make the name of an organization required"
## Code After:
from rest_framework import serializers
from bluebottle.organizations.models import Organization
from bluebottle.utils.serializers import URLField
class OrganizationSerializer(serializers.ModelSerializer):
class Meta:
model = Organization
fields = ('id', 'name', 'slug', 'address_line1', 'address_line2',
'city', 'state', 'country', 'postal_code', 'phone_number',
'website', 'email')
class ManageOrganizationSerializer(serializers.ModelSerializer):
slug = serializers.SlugField(required=False, allow_null=True)
name = serializers.CharField(required=True, allow_blank=True)
website = URLField(required=False, allow_blank=True)
email = serializers.EmailField(required=False, allow_blank=True)
class Meta:
model = Organization
fields = OrganizationSerializer.Meta.fields + ('partner_organizations',
'created', 'updated')
| from rest_framework import serializers
from bluebottle.organizations.models import Organization
from bluebottle.utils.serializers import URLField
class OrganizationSerializer(serializers.ModelSerializer):
class Meta:
model = Organization
fields = ('id', 'name', 'slug', 'address_line1', 'address_line2',
'city', 'state', 'country', 'postal_code', 'phone_number',
'website', 'email')
class ManageOrganizationSerializer(serializers.ModelSerializer):
slug = serializers.SlugField(required=False, allow_null=True)
- name = serializers.CharField(required=True)
+ name = serializers.CharField(required=True, allow_blank=True)
? ++++++++++++++++++
website = URLField(required=False, allow_blank=True)
email = serializers.EmailField(required=False, allow_blank=True)
class Meta:
model = Organization
fields = OrganizationSerializer.Meta.fields + ('partner_organizations',
'created', 'updated') |
5a785f725d68733561a7e5e82c57655e25439ec8 | indra/tests/test_grounding_resources.py | indra/tests/test_grounding_resources.py | import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
# Namespaces that are not currently handled but still appear in statements
exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_exceptional_unhandled():
"""Test that exceptional namespaces actually aren't handled.
This will catch if we make an update that makes an exceptional namespace
become a handled namespace. That way we can update the tests.
"""
actually_handled = []
for ns in exceptions:
if validate_ns(ns):
actually_handled.append(ns)
assert not actually_handled, actually_handled
| import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
| Remove exceptional namespaces from test | Remove exceptional namespaces from test
| Python | bsd-2-clause | johnbachman/indra,sorgerlab/indra,sorgerlab/belpy,sorgerlab/indra,bgyori/indra,sorgerlab/indra,bgyori/indra,bgyori/indra,johnbachman/indra,johnbachman/indra,sorgerlab/belpy,sorgerlab/belpy | import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
-
- # Namespaces that are not currently handled but still appear in statements
- exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
-
- def test_exceptional_unhandled():
- """Test that exceptional namespaces actually aren't handled.
-
- This will catch if we make an update that makes an exceptional namespace
- become a handled namespace. That way we can update the tests.
- """
- actually_handled = []
- for ns in exceptions:
- if validate_ns(ns):
- actually_handled.append(ns)
- assert not actually_handled, actually_handled
-
-
-
-
-
-
- | Remove exceptional namespaces from test | ## Code Before:
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
# Namespaces that are not currently handled but still appear in statements
exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_exceptional_unhandled():
"""Test that exceptional namespaces actually aren't handled.
This will catch if we make an update that makes an exceptional namespace
become a handled namespace. That way we can update the tests.
"""
actually_handled = []
for ns in exceptions:
if validate_ns(ns):
actually_handled.append(ns)
assert not actually_handled, actually_handled
## Instruction:
Remove exceptional namespaces from test
## Code After:
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
| import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
-
- # Namespaces that are not currently handled but still appear in statements
- exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
-
-
- def test_exceptional_unhandled():
- """Test that exceptional namespaces actually aren't handled.
-
- This will catch if we make an update that makes an exceptional namespace
- become a handled namespace. That way we can update the tests.
- """
- actually_handled = []
- for ns in exceptions:
- if validate_ns(ns):
- actually_handled.append(ns)
- assert not actually_handled, actually_handled
-
-
-
-
-
- |
0c1caf49a18bcd862247cdca7a4efe2f6fc02d93 | wafer/management/commands/wafer_talk_video_reviewers.py | wafer/management/commands/wafer_talk_video_reviewers.py | import sys
import csv
from django.core.management.base import BaseCommand
from django.contrib.auth import get_user_model
from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL
class Command(BaseCommand):
help = ("List talks and the associated video_reviewer emails."
" Only reviewers for accepted talks are listed")
def _video_reviewers(self, options):
talks = Talk.objects.filter(status=ACCEPTED)
csv_file = csv.writer(sys.stdout)
for talk in talks:
reviewer = talk.video_reviewer
if not reviewer:
reviewer = 'NO REVIEWER'
row = [x.encode("utf-8") for x in (
talk.title,
talk.get_authors_display_name(),
reviewer,
)]
csv_file.writerow(row)
def handle(self, *args, **options):
self._video_reviewers(options)
| import sys
import csv
from django.core.management.base import BaseCommand
from django.contrib.auth import get_user_model
from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL
class Command(BaseCommand):
help = ("List talks and the associated video_reviewer emails."
" Only reviewers for accepted talks are listed")
def _video_reviewers(self, options):
talks = Talk.objects.filter(status=ACCEPTED)
csv_file = csv.writer(sys.stdout)
for talk in talks:
reviewer = talk.video_reviewer
if not reviewer:
reviewer = 'NO REVIEWER'
row = [talk.title,
talk.get_authors_display_name(),
reviewer,
]
csv_file.writerow(row)
def handle(self, *args, **options):
self._video_reviewers(options)
| Drop python2-era manual encode dance | Drop python2-era manual encode dance
| Python | isc | CTPUG/wafer,CTPUG/wafer,CTPUG/wafer,CTPUG/wafer | import sys
import csv
from django.core.management.base import BaseCommand
from django.contrib.auth import get_user_model
from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL
class Command(BaseCommand):
help = ("List talks and the associated video_reviewer emails."
" Only reviewers for accepted talks are listed")
def _video_reviewers(self, options):
talks = Talk.objects.filter(status=ACCEPTED)
csv_file = csv.writer(sys.stdout)
for talk in talks:
reviewer = talk.video_reviewer
if not reviewer:
reviewer = 'NO REVIEWER'
- row = [x.encode("utf-8") for x in (
- talk.title,
+ row = [talk.title,
- talk.get_authors_display_name(),
+ talk.get_authors_display_name(),
- reviewer,
+ reviewer,
- )]
+ ]
csv_file.writerow(row)
def handle(self, *args, **options):
self._video_reviewers(options)
| Drop python2-era manual encode dance | ## Code Before:
import sys
import csv
from django.core.management.base import BaseCommand
from django.contrib.auth import get_user_model
from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL
class Command(BaseCommand):
help = ("List talks and the associated video_reviewer emails."
" Only reviewers for accepted talks are listed")
def _video_reviewers(self, options):
talks = Talk.objects.filter(status=ACCEPTED)
csv_file = csv.writer(sys.stdout)
for talk in talks:
reviewer = talk.video_reviewer
if not reviewer:
reviewer = 'NO REVIEWER'
row = [x.encode("utf-8") for x in (
talk.title,
talk.get_authors_display_name(),
reviewer,
)]
csv_file.writerow(row)
def handle(self, *args, **options):
self._video_reviewers(options)
## Instruction:
Drop python2-era manual encode dance
## Code After:
import sys
import csv
from django.core.management.base import BaseCommand
from django.contrib.auth import get_user_model
from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL
class Command(BaseCommand):
help = ("List talks and the associated video_reviewer emails."
" Only reviewers for accepted talks are listed")
def _video_reviewers(self, options):
talks = Talk.objects.filter(status=ACCEPTED)
csv_file = csv.writer(sys.stdout)
for talk in talks:
reviewer = talk.video_reviewer
if not reviewer:
reviewer = 'NO REVIEWER'
row = [talk.title,
talk.get_authors_display_name(),
reviewer,
]
csv_file.writerow(row)
def handle(self, *args, **options):
self._video_reviewers(options)
| import sys
import csv
from django.core.management.base import BaseCommand
from django.contrib.auth import get_user_model
from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL
class Command(BaseCommand):
help = ("List talks and the associated video_reviewer emails."
" Only reviewers for accepted talks are listed")
def _video_reviewers(self, options):
talks = Talk.objects.filter(status=ACCEPTED)
csv_file = csv.writer(sys.stdout)
for talk in talks:
reviewer = talk.video_reviewer
if not reviewer:
reviewer = 'NO REVIEWER'
- row = [x.encode("utf-8") for x in (
- talk.title,
? ^^^^^^
+ row = [talk.title,
? +++ + ^
- talk.get_authors_display_name(),
? -
+ talk.get_authors_display_name(),
- reviewer,
? -
+ reviewer,
- )]
? -
+ ]
csv_file.writerow(row)
def handle(self, *args, **options):
self._video_reviewers(options) |
fd054790ce32c3918f6edbe824540c09d7efce59 | stagehand/providers/__init__.py | stagehand/providers/__init__.py | import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| Remove tvrage from active providers as site is shut down | Remove tvrage from active providers as site is shut down
| Python | mit | jtackaberry/stagehand,jtackaberry/stagehand | import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
- plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
+ plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| Remove tvrage from active providers as site is shut down | ## Code Before:
import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
## Instruction:
Remove tvrage from active providers as site is shut down
## Code After:
import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
- plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
? ----------
+ plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error) |
64bc8ff452d03c7bb026be0b2edd9a047a88b386 | foyer/forcefields/forcefields.py | foyer/forcefields/forcefields.py | import os
import glob
from pkg_resources import resource_filename
from foyer import Forcefield
def get_ff_path():
return [resource_filename('foyer', 'forcefields')]
def get_forcefield_paths(forcefield_name=None):
for dir_path in get_ff_path():
file_pattern = os.path.join(dir_path, 'xml/*.xml')
file_paths = [file_path for file_path in glob.glob(file_pattern)]
return file_paths
def get_forcefield(name=None):
if name is None:
raise ValueError('Need a force field name')
file_paths = get_forcefield_paths()
try:
ff_path = next(val for val in file_paths if name in val)
except StopIteration:
raise ValueError('Could not find force field with name {}'
' in path {}'.format(name, get_ff_path()))
return Forcefield(ff_path)
load_OPLSAA = get_forcefield(name='oplsaa')
load_TRAPPE_UA = get_forcefield(name='trappe-ua')
| import os
import glob
from pkg_resources import resource_filename
from foyer import Forcefield
def get_ff_path():
return [resource_filename('foyer', 'forcefields')]
def get_forcefield_paths(forcefield_name=None):
for dir_path in get_ff_path():
file_pattern = os.path.join(dir_path, 'xml/*.xml')
file_paths = [file_path for file_path in glob.glob(file_pattern)]
return file_paths
def get_forcefield(name=None):
if name is None:
raise ValueError('Need a force field name')
file_paths = get_forcefield_paths()
try:
ff_path = next(val for val in file_paths if name in val)
except StopIteration:
raise ValueError('Could not find force field with name {}'
' in path {}'.format(name, get_ff_path()))
return Forcefield(ff_path)
def load_OPLSAA():
return get_forcefield(name='oplsaa')
def load_TRAPPE_UA():
return get_forcefield(name='trappe-ua')
load_OPLSAA = load_OPLSAA
load_TRAPPE_UA = load_TRAPPE_UA
| Make discrete functions for each force field | Make discrete functions for each force field
| Python | mit | mosdef-hub/foyer,mosdef-hub/foyer,iModels/foyer,iModels/foyer | import os
import glob
from pkg_resources import resource_filename
from foyer import Forcefield
def get_ff_path():
return [resource_filename('foyer', 'forcefields')]
def get_forcefield_paths(forcefield_name=None):
for dir_path in get_ff_path():
file_pattern = os.path.join(dir_path, 'xml/*.xml')
file_paths = [file_path for file_path in glob.glob(file_pattern)]
return file_paths
def get_forcefield(name=None):
if name is None:
raise ValueError('Need a force field name')
file_paths = get_forcefield_paths()
try:
ff_path = next(val for val in file_paths if name in val)
except StopIteration:
raise ValueError('Could not find force field with name {}'
' in path {}'.format(name, get_ff_path()))
return Forcefield(ff_path)
- load_OPLSAA = get_forcefield(name='oplsaa')
- load_TRAPPE_UA = get_forcefield(name='trappe-ua')
+ def load_OPLSAA():
+ return get_forcefield(name='oplsaa')
+
+
+ def load_TRAPPE_UA():
+ return get_forcefield(name='trappe-ua')
+
+
+ load_OPLSAA = load_OPLSAA
+ load_TRAPPE_UA = load_TRAPPE_UA
+ | Make discrete functions for each force field | ## Code Before:
import os
import glob
from pkg_resources import resource_filename
from foyer import Forcefield
def get_ff_path():
return [resource_filename('foyer', 'forcefields')]
def get_forcefield_paths(forcefield_name=None):
for dir_path in get_ff_path():
file_pattern = os.path.join(dir_path, 'xml/*.xml')
file_paths = [file_path for file_path in glob.glob(file_pattern)]
return file_paths
def get_forcefield(name=None):
if name is None:
raise ValueError('Need a force field name')
file_paths = get_forcefield_paths()
try:
ff_path = next(val for val in file_paths if name in val)
except StopIteration:
raise ValueError('Could not find force field with name {}'
' in path {}'.format(name, get_ff_path()))
return Forcefield(ff_path)
load_OPLSAA = get_forcefield(name='oplsaa')
load_TRAPPE_UA = get_forcefield(name='trappe-ua')
## Instruction:
Make discrete functions for each force field
## Code After:
import os
import glob
from pkg_resources import resource_filename
from foyer import Forcefield
def get_ff_path():
return [resource_filename('foyer', 'forcefields')]
def get_forcefield_paths(forcefield_name=None):
for dir_path in get_ff_path():
file_pattern = os.path.join(dir_path, 'xml/*.xml')
file_paths = [file_path for file_path in glob.glob(file_pattern)]
return file_paths
def get_forcefield(name=None):
if name is None:
raise ValueError('Need a force field name')
file_paths = get_forcefield_paths()
try:
ff_path = next(val for val in file_paths if name in val)
except StopIteration:
raise ValueError('Could not find force field with name {}'
' in path {}'.format(name, get_ff_path()))
return Forcefield(ff_path)
def load_OPLSAA():
return get_forcefield(name='oplsaa')
def load_TRAPPE_UA():
return get_forcefield(name='trappe-ua')
load_OPLSAA = load_OPLSAA
load_TRAPPE_UA = load_TRAPPE_UA
| import os
import glob
from pkg_resources import resource_filename
from foyer import Forcefield
def get_ff_path():
return [resource_filename('foyer', 'forcefields')]
def get_forcefield_paths(forcefield_name=None):
for dir_path in get_ff_path():
file_pattern = os.path.join(dir_path, 'xml/*.xml')
file_paths = [file_path for file_path in glob.glob(file_pattern)]
return file_paths
def get_forcefield(name=None):
if name is None:
raise ValueError('Need a force field name')
file_paths = get_forcefield_paths()
try:
ff_path = next(val for val in file_paths if name in val)
except StopIteration:
raise ValueError('Could not find force field with name {}'
' in path {}'.format(name, get_ff_path()))
return Forcefield(ff_path)
- load_OPLSAA = get_forcefield(name='oplsaa')
- load_TRAPPE_UA = get_forcefield(name='trappe-ua')
+
+ def load_OPLSAA():
+ return get_forcefield(name='oplsaa')
+
+
+ def load_TRAPPE_UA():
+ return get_forcefield(name='trappe-ua')
+
+
+ load_OPLSAA = load_OPLSAA
+ load_TRAPPE_UA = load_TRAPPE_UA |
5dcec96b7af384f7f753cb2d67d7cbd0c361c504 | tests/helpers.py | tests/helpers.py | from __future__ import unicode_literals
import json
from elasticsearch import (
Elasticsearch,
TransportError
)
ELASTICSEARCH_URL = "localhost"
conn = Elasticsearch(ELASTICSEARCH_URL)
def homogeneous(a, b):
json.dumps(a).should.equal(json.dumps(b))
def heterogeneous(a, b):
json.dumps(a).shouldnt.equal(json.dumps(b))
def add_document(index, document, **kwargs):
document_type = "my_doc_type"
conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs)
def clean_elasticsearch(context):
_delete_es_index("foo")
def prepare_elasticsearch(context):
clean_elasticsearch(context)
_create_foo_index()
conn.cluster.health(wait_for_status='yellow')
def _create_foo_index():
conn.indices.create(index="foo", ignore=400)
def _delete_es_index(index):
conn.indices.delete(index=index, ignore=[400, 404])
prepare_data = [
prepare_elasticsearch
]
cleanup_data = [
clean_elasticsearch
]
| from __future__ import unicode_literals
import json
from elasticsearch import (
Elasticsearch,
TransportError
)
ELASTICSEARCH_URL = "localhost"
conn = Elasticsearch(ELASTICSEARCH_URL)
def homogeneous(a, b):
json.dumps(a).should.equal(json.dumps(b))
def heterogeneous(a, b):
json.dumps(a).shouldnt.equal(json.dumps(b))
def add_document(index, document, **kwargs):
if "doc_type" not in kwargs:
# Allow overriding doc type defaults
doc_type = "my_doc_type"
kwargs["doc_type"] = doc_type
conn.create(index=index, body=document, refresh=True, **kwargs)
def clean_elasticsearch(context):
_delete_es_index("foo")
def prepare_elasticsearch(context):
clean_elasticsearch(context)
_create_foo_index()
conn.cluster.health(wait_for_status='yellow')
def _create_foo_index():
conn.indices.create(index="foo", ignore=400)
def _delete_es_index(index):
conn.indices.delete(index=index, ignore=[400, 404])
prepare_data = [
prepare_elasticsearch
]
cleanup_data = [
clean_elasticsearch
]
| Allow overriding doc type defaults | Allow overriding doc type defaults
| Python | mit | Yipit/pyeqs | from __future__ import unicode_literals
import json
from elasticsearch import (
Elasticsearch,
TransportError
)
ELASTICSEARCH_URL = "localhost"
conn = Elasticsearch(ELASTICSEARCH_URL)
def homogeneous(a, b):
json.dumps(a).should.equal(json.dumps(b))
def heterogeneous(a, b):
json.dumps(a).shouldnt.equal(json.dumps(b))
def add_document(index, document, **kwargs):
+ if "doc_type" not in kwargs:
+ # Allow overriding doc type defaults
- document_type = "my_doc_type"
+ doc_type = "my_doc_type"
+ kwargs["doc_type"] = doc_type
- conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs)
+ conn.create(index=index, body=document, refresh=True, **kwargs)
def clean_elasticsearch(context):
_delete_es_index("foo")
def prepare_elasticsearch(context):
clean_elasticsearch(context)
_create_foo_index()
conn.cluster.health(wait_for_status='yellow')
def _create_foo_index():
conn.indices.create(index="foo", ignore=400)
def _delete_es_index(index):
conn.indices.delete(index=index, ignore=[400, 404])
prepare_data = [
prepare_elasticsearch
]
cleanup_data = [
clean_elasticsearch
]
| Allow overriding doc type defaults | ## Code Before:
from __future__ import unicode_literals
import json
from elasticsearch import (
Elasticsearch,
TransportError
)
ELASTICSEARCH_URL = "localhost"
conn = Elasticsearch(ELASTICSEARCH_URL)
def homogeneous(a, b):
json.dumps(a).should.equal(json.dumps(b))
def heterogeneous(a, b):
json.dumps(a).shouldnt.equal(json.dumps(b))
def add_document(index, document, **kwargs):
document_type = "my_doc_type"
conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs)
def clean_elasticsearch(context):
_delete_es_index("foo")
def prepare_elasticsearch(context):
clean_elasticsearch(context)
_create_foo_index()
conn.cluster.health(wait_for_status='yellow')
def _create_foo_index():
conn.indices.create(index="foo", ignore=400)
def _delete_es_index(index):
conn.indices.delete(index=index, ignore=[400, 404])
prepare_data = [
prepare_elasticsearch
]
cleanup_data = [
clean_elasticsearch
]
## Instruction:
Allow overriding doc type defaults
## Code After:
from __future__ import unicode_literals
import json
from elasticsearch import (
Elasticsearch,
TransportError
)
ELASTICSEARCH_URL = "localhost"
conn = Elasticsearch(ELASTICSEARCH_URL)
def homogeneous(a, b):
json.dumps(a).should.equal(json.dumps(b))
def heterogeneous(a, b):
json.dumps(a).shouldnt.equal(json.dumps(b))
def add_document(index, document, **kwargs):
if "doc_type" not in kwargs:
# Allow overriding doc type defaults
doc_type = "my_doc_type"
kwargs["doc_type"] = doc_type
conn.create(index=index, body=document, refresh=True, **kwargs)
def clean_elasticsearch(context):
_delete_es_index("foo")
def prepare_elasticsearch(context):
clean_elasticsearch(context)
_create_foo_index()
conn.cluster.health(wait_for_status='yellow')
def _create_foo_index():
conn.indices.create(index="foo", ignore=400)
def _delete_es_index(index):
conn.indices.delete(index=index, ignore=[400, 404])
prepare_data = [
prepare_elasticsearch
]
cleanup_data = [
clean_elasticsearch
]
| from __future__ import unicode_literals
import json
from elasticsearch import (
Elasticsearch,
TransportError
)
ELASTICSEARCH_URL = "localhost"
conn = Elasticsearch(ELASTICSEARCH_URL)
def homogeneous(a, b):
json.dumps(a).should.equal(json.dumps(b))
def heterogeneous(a, b):
json.dumps(a).shouldnt.equal(json.dumps(b))
def add_document(index, document, **kwargs):
+ if "doc_type" not in kwargs:
+ # Allow overriding doc type defaults
- document_type = "my_doc_type"
? -----
+ doc_type = "my_doc_type"
? ++++
+ kwargs["doc_type"] = doc_type
- conn.create(index=index, doc_type=document_type, body=document, refresh=True, **kwargs)
? ------------------------
+ conn.create(index=index, body=document, refresh=True, **kwargs)
def clean_elasticsearch(context):
_delete_es_index("foo")
def prepare_elasticsearch(context):
clean_elasticsearch(context)
_create_foo_index()
conn.cluster.health(wait_for_status='yellow')
def _create_foo_index():
conn.indices.create(index="foo", ignore=400)
def _delete_es_index(index):
conn.indices.delete(index=index, ignore=[400, 404])
prepare_data = [
prepare_elasticsearch
]
cleanup_data = [
clean_elasticsearch
] |
146e35f48774173c2000b8a9790cdbe6925ba94a | opps/contrib/multisite/admin.py | opps/contrib/multisite/admin.py | from django.contrib import admin
from .models import SitePermission
admin.site.register(SitePermission)
| from django.contrib import admin
from django.utils import timezone
from .models import SitePermission
class AdminViewPermission(admin.ModelAdmin):
def queryset(self, request):
queryset = super(AdminViewPermission, self).queryset(request)
try:
sitepermission = SitePermission.objects.get(
user=request.user,
date_available__lte=timezone.now(),
published=True)
return queryset.filter(site_iid=sitepermission.site_iid)
except SitePermission.DoesNotExist:
pass
return queryset
admin.site.register(SitePermission)
| Create AdminViewPermission on contrib multisite | Create AdminViewPermission on contrib multisite
| Python | mit | opps/opps,YACOWS/opps,williamroot/opps,jeanmask/opps,opps/opps,williamroot/opps,jeanmask/opps,YACOWS/opps,YACOWS/opps,YACOWS/opps,opps/opps,jeanmask/opps,williamroot/opps,opps/opps,jeanmask/opps,williamroot/opps | from django.contrib import admin
+ from django.utils import timezone
+
from .models import SitePermission
+
+
+ class AdminViewPermission(admin.ModelAdmin):
+
+ def queryset(self, request):
+ queryset = super(AdminViewPermission, self).queryset(request)
+ try:
+ sitepermission = SitePermission.objects.get(
+ user=request.user,
+ date_available__lte=timezone.now(),
+ published=True)
+ return queryset.filter(site_iid=sitepermission.site_iid)
+ except SitePermission.DoesNotExist:
+ pass
+ return queryset
admin.site.register(SitePermission)
| Create AdminViewPermission on contrib multisite | ## Code Before:
from django.contrib import admin
from .models import SitePermission
admin.site.register(SitePermission)
## Instruction:
Create AdminViewPermission on contrib multisite
## Code After:
from django.contrib import admin
from django.utils import timezone
from .models import SitePermission
class AdminViewPermission(admin.ModelAdmin):
def queryset(self, request):
queryset = super(AdminViewPermission, self).queryset(request)
try:
sitepermission = SitePermission.objects.get(
user=request.user,
date_available__lte=timezone.now(),
published=True)
return queryset.filter(site_iid=sitepermission.site_iid)
except SitePermission.DoesNotExist:
pass
return queryset
admin.site.register(SitePermission)
| from django.contrib import admin
+ from django.utils import timezone
+
from .models import SitePermission
+ class AdminViewPermission(admin.ModelAdmin):
+
+ def queryset(self, request):
+ queryset = super(AdminViewPermission, self).queryset(request)
+ try:
+ sitepermission = SitePermission.objects.get(
+ user=request.user,
+ date_available__lte=timezone.now(),
+ published=True)
+ return queryset.filter(site_iid=sitepermission.site_iid)
+ except SitePermission.DoesNotExist:
+ pass
+ return queryset
+
+
admin.site.register(SitePermission) |
38ce0d6b0433a68787c18691407c815d4eb1fdb2 | txscrypt/__init__.py | txscrypt/__init__.py | from txscrypt.wrapper import computeKey, verifyPassword
from txscrypt._version import __version__
__all__ = ["computeKey", "verifyPassword"]
| from txscrypt.wrapper import checkPassword, computeKey
from txscrypt._version import __version__
__all__ = ["verifyPassword", "computeKey"]
| Make checkPassword the only public API, remove verifyPassword | Make checkPassword the only public API, remove verifyPassword
| Python | isc | lvh/txscrypt | - from txscrypt.wrapper import computeKey, verifyPassword
+ from txscrypt.wrapper import checkPassword, computeKey
from txscrypt._version import __version__
- __all__ = ["computeKey", "verifyPassword"]
+ __all__ = ["verifyPassword", "computeKey"]
| Make checkPassword the only public API, remove verifyPassword | ## Code Before:
from txscrypt.wrapper import computeKey, verifyPassword
from txscrypt._version import __version__
__all__ = ["computeKey", "verifyPassword"]
## Instruction:
Make checkPassword the only public API, remove verifyPassword
## Code After:
from txscrypt.wrapper import checkPassword, computeKey
from txscrypt._version import __version__
__all__ = ["verifyPassword", "computeKey"]
| - from txscrypt.wrapper import computeKey, verifyPassword
+ from txscrypt.wrapper import checkPassword, computeKey
from txscrypt._version import __version__
- __all__ = ["computeKey", "verifyPassword"]
+ __all__ = ["verifyPassword", "computeKey"] |
c39260e64c8820bad9243c35f10b352419425810 | marble/tests/test_exposure.py | marble/tests/test_exposure.py | """ Tests for the exposure computation """
from nose.tools import *
import marble as mb
# Test maximum value of exposure
# Test maximum value of isolation
# Test minimum of exposure
# Test minimum of isolation
| """ Tests for the exposure computation """
from __future__ import division
from nose.tools import *
import itertools
import marble as mb
#
# Synthetic data for tests
#
def segregated_city():
""" perfect segregation """
city = {"A":{1:7, 2:0, 3:0},
"B":{1:0, 2:0, 3:14},
"C":{1:0, 2:42, 3:0}}
return city
def two_way_city():
""" perfect two-way exposure for 1 and 2 """
city = {"A":{1:7, 2:13, 3:0},
"B":{1:7, 2:13, 3:0},
"C":{1:0, 2:0, 3:37}}
return city
def uniform_city():
""" Uniform representation """
city = {"A":{1:1, 2:10, 3:7},
"B":{1:2, 2:20, 3:14},
"C":{1:4, 2:40, 3:28}}
return city
#
# Test
#
class TestExposure(object):
def test_maximum_isolation(city):
city = segregated_city()
exp = mb.exposure(city)
N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
N_tot = sum(N_cl.values())
for c in exp:
assert_almost_equal(exp[c][c][0],
N_tot/N_cl[c],
places=3)
def test_minimum_exposure(city):
city = segregated_city()
exp = mb.exposure(city)
for c0,c1 in itertools.permutations([1,2,3], 2):
assert_almost_equal(exp[c0][c1][0],
0.0)
def test_maximum_exposure(city):
city = two_way_city()
exp = mb.exposure(city)
N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
N_tot = sum(N_cl.values())
assert_almost_equal(exp[2][1][0],
N_tot/(N_cl[1]+N_cl[2]),
places=3)
def test_minimum_isolation(city):
city = uniform_city()
exp = mb.exposure(city)
for c in [1,2,3]:
assert_almost_equal(exp[c][c][0],
1.0,
places=3)
| Write tests for the exposure | Write tests for the exposure
| Python | bsd-3-clause | walkerke/marble,scities/marble | """ Tests for the exposure computation """
+ from __future__ import division
from nose.tools import *
+ import itertools
import marble as mb
- # Test maximum value of exposure
- # Test maximum value of isolation
- # Test minimum of exposure
- # Test minimum of isolation
+ #
+ # Synthetic data for tests
+ #
+ def segregated_city():
+ """ perfect segregation """
+ city = {"A":{1:7, 2:0, 3:0},
+ "B":{1:0, 2:0, 3:14},
+ "C":{1:0, 2:42, 3:0}}
+ return city
+
+ def two_way_city():
+ """ perfect two-way exposure for 1 and 2 """
+ city = {"A":{1:7, 2:13, 3:0},
+ "B":{1:7, 2:13, 3:0},
+ "C":{1:0, 2:0, 3:37}}
+ return city
+
+ def uniform_city():
+ """ Uniform representation """
+ city = {"A":{1:1, 2:10, 3:7},
+ "B":{1:2, 2:20, 3:14},
+ "C":{1:4, 2:40, 3:28}}
+ return city
+
+
+
+ #
+ # Test
+ #
+ class TestExposure(object):
+
+ def test_maximum_isolation(city):
+ city = segregated_city()
+ exp = mb.exposure(city)
+ N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
+ N_tot = sum(N_cl.values())
+ for c in exp:
+ assert_almost_equal(exp[c][c][0],
+ N_tot/N_cl[c],
+ places=3)
+
+ def test_minimum_exposure(city):
+ city = segregated_city()
+ exp = mb.exposure(city)
+ for c0,c1 in itertools.permutations([1,2,3], 2):
+ assert_almost_equal(exp[c0][c1][0],
+ 0.0)
+
+ def test_maximum_exposure(city):
+ city = two_way_city()
+ exp = mb.exposure(city)
+ N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
+ N_tot = sum(N_cl.values())
+ assert_almost_equal(exp[2][1][0],
+ N_tot/(N_cl[1]+N_cl[2]),
+ places=3)
+
+ def test_minimum_isolation(city):
+ city = uniform_city()
+ exp = mb.exposure(city)
+ for c in [1,2,3]:
+ assert_almost_equal(exp[c][c][0],
+ 1.0,
+ places=3)
+
+ | Write tests for the exposure | ## Code Before:
""" Tests for the exposure computation """
from nose.tools import *
import marble as mb
# Test maximum value of exposure
# Test maximum value of isolation
# Test minimum of exposure
# Test minimum of isolation
## Instruction:
Write tests for the exposure
## Code After:
""" Tests for the exposure computation """
from __future__ import division
from nose.tools import *
import itertools
import marble as mb
#
# Synthetic data for tests
#
def segregated_city():
""" perfect segregation """
city = {"A":{1:7, 2:0, 3:0},
"B":{1:0, 2:0, 3:14},
"C":{1:0, 2:42, 3:0}}
return city
def two_way_city():
""" perfect two-way exposure for 1 and 2 """
city = {"A":{1:7, 2:13, 3:0},
"B":{1:7, 2:13, 3:0},
"C":{1:0, 2:0, 3:37}}
return city
def uniform_city():
""" Uniform representation """
city = {"A":{1:1, 2:10, 3:7},
"B":{1:2, 2:20, 3:14},
"C":{1:4, 2:40, 3:28}}
return city
#
# Test
#
class TestExposure(object):
def test_maximum_isolation(city):
city = segregated_city()
exp = mb.exposure(city)
N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
N_tot = sum(N_cl.values())
for c in exp:
assert_almost_equal(exp[c][c][0],
N_tot/N_cl[c],
places=3)
def test_minimum_exposure(city):
city = segregated_city()
exp = mb.exposure(city)
for c0,c1 in itertools.permutations([1,2,3], 2):
assert_almost_equal(exp[c0][c1][0],
0.0)
def test_maximum_exposure(city):
city = two_way_city()
exp = mb.exposure(city)
N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
N_tot = sum(N_cl.values())
assert_almost_equal(exp[2][1][0],
N_tot/(N_cl[1]+N_cl[2]),
places=3)
def test_minimum_isolation(city):
city = uniform_city()
exp = mb.exposure(city)
for c in [1,2,3]:
assert_almost_equal(exp[c][c][0],
1.0,
places=3)
| """ Tests for the exposure computation """
+ from __future__ import division
from nose.tools import *
+ import itertools
import marble as mb
- # Test maximum value of exposure
- # Test maximum value of isolation
- # Test minimum of exposure
- # Test minimum of isolation
+
+ #
+ # Synthetic data for tests
+ #
+ def segregated_city():
+ """ perfect segregation """
+ city = {"A":{1:7, 2:0, 3:0},
+ "B":{1:0, 2:0, 3:14},
+ "C":{1:0, 2:42, 3:0}}
+ return city
+
+ def two_way_city():
+ """ perfect two-way exposure for 1 and 2 """
+ city = {"A":{1:7, 2:13, 3:0},
+ "B":{1:7, 2:13, 3:0},
+ "C":{1:0, 2:0, 3:37}}
+ return city
+
+ def uniform_city():
+ """ Uniform representation """
+ city = {"A":{1:1, 2:10, 3:7},
+ "B":{1:2, 2:20, 3:14},
+ "C":{1:4, 2:40, 3:28}}
+ return city
+
+
+
+ #
+ # Test
+ #
+ class TestExposure(object):
+
+ def test_maximum_isolation(city):
+ city = segregated_city()
+ exp = mb.exposure(city)
+ N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
+ N_tot = sum(N_cl.values())
+ for c in exp:
+ assert_almost_equal(exp[c][c][0],
+ N_tot/N_cl[c],
+ places=3)
+
+ def test_minimum_exposure(city):
+ city = segregated_city()
+ exp = mb.exposure(city)
+ for c0,c1 in itertools.permutations([1,2,3], 2):
+ assert_almost_equal(exp[c0][c1][0],
+ 0.0)
+
+ def test_maximum_exposure(city):
+ city = two_way_city()
+ exp = mb.exposure(city)
+ N_cl = {i: sum([city[au][i] for au in city]) for i in [1,2,3]}
+ N_tot = sum(N_cl.values())
+ assert_almost_equal(exp[2][1][0],
+ N_tot/(N_cl[1]+N_cl[2]),
+ places=3)
+
+ def test_minimum_isolation(city):
+ city = uniform_city()
+ exp = mb.exposure(city)
+ for c in [1,2,3]:
+ assert_almost_equal(exp[c][c][0],
+ 1.0,
+ places=3)
+ |
a9d4fab047249fbf5db26385779902d0f7483057 | qsimcirq/__init__.py | qsimcirq/__init__.py | from .qsim_circuit import *
from .qsim_simulator import *
from .qsimh_simulator import *
| from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit
from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator
from .qsimh_simulator import QSimhSimulator
| Replace star imports to fix mypy issue. | Replace star imports to fix mypy issue.
| Python | apache-2.0 | quantumlib/qsim,quantumlib/qsim,quantumlib/qsim,quantumlib/qsim | - from .qsim_circuit import *
- from .qsim_simulator import *
+ from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit
+ from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator
- from .qsimh_simulator import *
+ from .qsimh_simulator import QSimhSimulator
-
- | Replace star imports to fix mypy issue. | ## Code Before:
from .qsim_circuit import *
from .qsim_simulator import *
from .qsimh_simulator import *
## Instruction:
Replace star imports to fix mypy issue.
## Code After:
from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit
from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator
from .qsimh_simulator import QSimhSimulator
| - from .qsim_circuit import *
- from .qsim_simulator import *
+ from .qsim_circuit import add_op_to_opstring, add_op_to_circuit, QSimCircuit
+ from .qsim_simulator import QSimSimulatorState, QSimSimulatorTrialResult, QSimSimulator
- from .qsimh_simulator import *
? ^
+ from .qsimh_simulator import QSimhSimulator
? ^^^^^^^^^^^^^^
-
- |
5b50b96b35c678ca17b069630875a9d86e2cbca3 | scripts/i18n/commons.py | scripts/i18n/commons.py | msg = {
'en': {
'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].',
'commons-file-now-available' : u'File is now available on Wikimedia Commons.',
'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}',
},
'qqq': {
'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.',
'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.',
},
}
| msg = {
'en': {
'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].',
'commons-file-now-available' : u'File is now available on Wikimedia Commons.',
},
'qqq': {
'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.',
},
}
| Remove the template for now. | Remove the template for now.
git-svn-id: 9a050473c2aca1e14f53d73349e19b938c2cf203@9344 6a7f98fc-eeb0-4dc1-a6e2-c2c589a08aa6
| Python | mit | legoktm/pywikipedia-rewrite | msg = {
'en': {
'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].',
'commons-file-now-available' : u'File is now available on Wikimedia Commons.',
- 'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}',
},
'qqq': {
- 'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
+ 'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
- 'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.',
+ 'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.',
- 'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.',
},
}
| Remove the template for now. | ## Code Before:
msg = {
'en': {
'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].',
'commons-file-now-available' : u'File is now available on Wikimedia Commons.',
'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}',
},
'qqq': {
'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.',
'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.',
},
}
## Instruction:
Remove the template for now.
## Code After:
msg = {
'en': {
'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].',
'commons-file-now-available' : u'File is now available on Wikimedia Commons.',
},
'qqq': {
'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.',
},
}
| msg = {
'en': {
'commons-file-moved' : u'[[:File:%s|File]] moved to [[:commons:File:%s|commons]].',
'commons-file-now-available' : u'File is now available on Wikimedia Commons.',
- 'commons-nowcommons-template' : 'en': u'{{subst:ncd|%s}}',
},
'qqq': {
- 'commons-file-now-available': u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
+ 'commons-file-now-available' : u'Edit summary when the bot has moved a file to Commons and adds the NowCommons template.',
? +
- 'commons-file-moved': u'Edit summary when the bot replaces usage of an image available under a different name.',
+ 'commons-file-moved' : u'Edit summary when the bot replaces usage of an image available under a different name.',
? +
- 'commons-nowcommons-template' : 'en': u'The template to be added by the bot when a file is available at Commons. %s is the filename at Commons.',
},
} |
b6978852775bb48e400a31a1e464d7b596db13f2 | fsictools.py | fsictools.py |
# Version number keeps track with the main `fsic` module
from fsic import __version__
import re
from typing import List
from fsic import BaseModel, Symbol
def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame':
"""Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
return DataFrame([s._asdict() for s in symbols])
def model_to_dataframe(model: BaseModel) -> 'DataFrame':
"""Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
df = DataFrame(model.values.T, index=model.span, columns=model.names)
df['status'] = model.status
df['iterations'] = model.iterations
return df
|
# Version number keeps track with the main `fsic` module
from fsic import __version__
import re
from typing import List
from fsic import BaseModel, Symbol
def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame':
"""Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
return DataFrame([s._asdict() for s in symbols])
def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame':
"""Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
df = DataFrame(model.values.T, index=model.span, columns=model.names)
df['status'] = model.status
df['iterations'] = model.iterations
return df
| Update type annotations to reference package names in full | TYP: Update type annotations to reference package names in full
| Python | mit | ChrisThoung/fsic |
# Version number keeps track with the main `fsic` module
from fsic import __version__
import re
from typing import List
from fsic import BaseModel, Symbol
- def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame':
+ def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame':
"""Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
return DataFrame([s._asdict() for s in symbols])
- def model_to_dataframe(model: BaseModel) -> 'DataFrame':
+ def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame':
"""Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
df = DataFrame(model.values.T, index=model.span, columns=model.names)
df['status'] = model.status
df['iterations'] = model.iterations
return df
| Update type annotations to reference package names in full | ## Code Before:
# Version number keeps track with the main `fsic` module
from fsic import __version__
import re
from typing import List
from fsic import BaseModel, Symbol
def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame':
"""Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
return DataFrame([s._asdict() for s in symbols])
def model_to_dataframe(model: BaseModel) -> 'DataFrame':
"""Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
df = DataFrame(model.values.T, index=model.span, columns=model.names)
df['status'] = model.status
df['iterations'] = model.iterations
return df
## Instruction:
Update type annotations to reference package names in full
## Code After:
# Version number keeps track with the main `fsic` module
from fsic import __version__
import re
from typing import List
from fsic import BaseModel, Symbol
def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame':
"""Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
return DataFrame([s._asdict() for s in symbols])
def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame':
"""Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
df = DataFrame(model.values.T, index=model.span, columns=model.names)
df['status'] = model.status
df['iterations'] = model.iterations
return df
|
# Version number keeps track with the main `fsic` module
from fsic import __version__
import re
from typing import List
from fsic import BaseModel, Symbol
- def symbols_to_dataframe(symbols: List[Symbol]) -> 'DataFrame':
+ def symbols_to_dataframe(symbols: List[Symbol]) -> 'pandas.DataFrame':
? +++++++
"""Convert the list of symbols to a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
return DataFrame([s._asdict() for s in symbols])
- def model_to_dataframe(model: BaseModel) -> 'DataFrame':
+ def model_to_dataframe(model: BaseModel) -> 'pandas.DataFrame':
? +++++++
"""Return the values and solution information from the model as a `pandas` DataFrame. **Requires `pandas`**."""
from pandas import DataFrame
df = DataFrame(model.values.T, index=model.span, columns=model.names)
df['status'] = model.status
df['iterations'] = model.iterations
return df |
4965511fdb9843233e84a8aa9aa0414bf1c02133 | mail/views.py | mail/views.py | from django.shortcuts import redirect
from django.http import JsonResponse
from django.core.mail import EmailMessage
from django.middleware import csrf
from rest_framework.decorators import api_view
@api_view(['POST', 'GET'])
def send_contact_message(request):
if request.method == 'POST':
to_address = request.POST.get("to_address", "").split(',')
from_name = request.POST.get("from_name", "")
from_address = request.POST.get("from_address", "")
from_string = '{} <{}>'.format(from_name, from_address)
subject = request.POST.get("subject", "")
message_body = request.POST.get("message_body", "")
csrf_token = request.POST.get("csrfmiddlewaretoken", "")
email = EmailMessage(subject,
message_body,
'noreply@openstax.org',
to_address,
reply_to=[from_string])
email.send(fail_silently=False)
#return redirect('/contact-thank-you')
data = {'subject': subject,
'message_body': message_body,
'to_address': to_address,
'reply_to': [from_string],
'from_address': 'noreply@openstax.org',
'csrf_token': csrf_token,
}
return JsonResponse(data)
# if this is not posting a message, let's send the csfr token back
else:
csrf_token = csrf.get_token(request)
data = {'csrf_token': csrf_token}
return JsonResponse(data)
| from django.shortcuts import redirect
from django.http import JsonResponse
from django.core.mail import EmailMessage
from django.middleware import csrf
from rest_framework.decorators import api_view
@api_view(['POST', 'GET'])
def send_contact_message(request):
if request.method == 'POST':
to_address = request.POST.get("to_address", "").split(',')
from_name = request.POST.get("from_name", "")
from_address = request.POST.get("from_address", "")
from_string = '{} <{}>'.format(from_name, from_address)
subject = request.POST.get("subject", "")
message_body = request.POST.get("message_body", "")
email = EmailMessage(subject,
message_body,
'noreply@openstax.org',
to_address,
reply_to=[from_string])
email.send()
return redirect('/contact-thank-you')
# if this is not posting a message, let's send the csfr token back
else:
csrf_token = csrf.get_token(request)
data = {'csrf_token': csrf_token}
return JsonResponse(data)
| Revert "return json of message being sent to debug mail issue" | Revert "return json of message being sent to debug mail issue"
| Python | agpl-3.0 | openstax/openstax-cms,Connexions/openstax-cms,openstax/openstax-cms,openstax/openstax-cms,openstax/openstax-cms,Connexions/openstax-cms | from django.shortcuts import redirect
from django.http import JsonResponse
from django.core.mail import EmailMessage
from django.middleware import csrf
from rest_framework.decorators import api_view
@api_view(['POST', 'GET'])
def send_contact_message(request):
if request.method == 'POST':
to_address = request.POST.get("to_address", "").split(',')
from_name = request.POST.get("from_name", "")
from_address = request.POST.get("from_address", "")
from_string = '{} <{}>'.format(from_name, from_address)
subject = request.POST.get("subject", "")
message_body = request.POST.get("message_body", "")
- csrf_token = request.POST.get("csrfmiddlewaretoken", "")
email = EmailMessage(subject,
message_body,
'noreply@openstax.org',
to_address,
reply_to=[from_string])
- email.send(fail_silently=False)
+ email.send()
- #return redirect('/contact-thank-you')
+ return redirect('/contact-thank-you')
- data = {'subject': subject,
- 'message_body': message_body,
- 'to_address': to_address,
- 'reply_to': [from_string],
- 'from_address': 'noreply@openstax.org',
- 'csrf_token': csrf_token,
- }
-
- return JsonResponse(data)
# if this is not posting a message, let's send the csfr token back
else:
csrf_token = csrf.get_token(request)
data = {'csrf_token': csrf_token}
return JsonResponse(data)
| Revert "return json of message being sent to debug mail issue" | ## Code Before:
from django.shortcuts import redirect
from django.http import JsonResponse
from django.core.mail import EmailMessage
from django.middleware import csrf
from rest_framework.decorators import api_view
@api_view(['POST', 'GET'])
def send_contact_message(request):
if request.method == 'POST':
to_address = request.POST.get("to_address", "").split(',')
from_name = request.POST.get("from_name", "")
from_address = request.POST.get("from_address", "")
from_string = '{} <{}>'.format(from_name, from_address)
subject = request.POST.get("subject", "")
message_body = request.POST.get("message_body", "")
csrf_token = request.POST.get("csrfmiddlewaretoken", "")
email = EmailMessage(subject,
message_body,
'noreply@openstax.org',
to_address,
reply_to=[from_string])
email.send(fail_silently=False)
#return redirect('/contact-thank-you')
data = {'subject': subject,
'message_body': message_body,
'to_address': to_address,
'reply_to': [from_string],
'from_address': 'noreply@openstax.org',
'csrf_token': csrf_token,
}
return JsonResponse(data)
# if this is not posting a message, let's send the csfr token back
else:
csrf_token = csrf.get_token(request)
data = {'csrf_token': csrf_token}
return JsonResponse(data)
## Instruction:
Revert "return json of message being sent to debug mail issue"
## Code After:
from django.shortcuts import redirect
from django.http import JsonResponse
from django.core.mail import EmailMessage
from django.middleware import csrf
from rest_framework.decorators import api_view
@api_view(['POST', 'GET'])
def send_contact_message(request):
if request.method == 'POST':
to_address = request.POST.get("to_address", "").split(',')
from_name = request.POST.get("from_name", "")
from_address = request.POST.get("from_address", "")
from_string = '{} <{}>'.format(from_name, from_address)
subject = request.POST.get("subject", "")
message_body = request.POST.get("message_body", "")
email = EmailMessage(subject,
message_body,
'noreply@openstax.org',
to_address,
reply_to=[from_string])
email.send()
return redirect('/contact-thank-you')
# if this is not posting a message, let's send the csfr token back
else:
csrf_token = csrf.get_token(request)
data = {'csrf_token': csrf_token}
return JsonResponse(data)
| from django.shortcuts import redirect
from django.http import JsonResponse
from django.core.mail import EmailMessage
from django.middleware import csrf
from rest_framework.decorators import api_view
@api_view(['POST', 'GET'])
def send_contact_message(request):
if request.method == 'POST':
to_address = request.POST.get("to_address", "").split(',')
from_name = request.POST.get("from_name", "")
from_address = request.POST.get("from_address", "")
from_string = '{} <{}>'.format(from_name, from_address)
subject = request.POST.get("subject", "")
message_body = request.POST.get("message_body", "")
- csrf_token = request.POST.get("csrfmiddlewaretoken", "")
email = EmailMessage(subject,
message_body,
'noreply@openstax.org',
to_address,
reply_to=[from_string])
- email.send(fail_silently=False)
+ email.send()
- #return redirect('/contact-thank-you')
? -
+ return redirect('/contact-thank-you')
- data = {'subject': subject,
- 'message_body': message_body,
- 'to_address': to_address,
- 'reply_to': [from_string],
- 'from_address': 'noreply@openstax.org',
- 'csrf_token': csrf_token,
- }
-
- return JsonResponse(data)
# if this is not posting a message, let's send the csfr token back
else:
csrf_token = csrf.get_token(request)
data = {'csrf_token': csrf_token}
return JsonResponse(data) |
94e070ec33dbc86e38de4839be9461db3a301685 | inonemonth/challenges/serializers.py | inonemonth/challenges/serializers.py | from rest_framework import serializers
from .models import Challenge, Role
from core.serializers import UserSerializer
class RoleSerializer(serializers.ModelSerializer):
#user = serializers.RelatedField(many=True)
#user = serializers.PrimaryKeyRelatedField()
#user = serializers.HyperlinkedRelatedField()
user = UserSerializer()
#challenge = ChallengeSerializer()
challenge = serializers.RelatedField()
class Meta:
model = Role
fields = ("id", "user", "type", "challenge")
class ChallengeSerializer(serializers.ModelSerializer):
#role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True)
#role_set = serializers.RelatedField(many=True)
#role_set = serializers.SlugRelatedField(many=True, slug_field="type")
role_set = RoleSerializer(many=True)
class Meta:
model = Challenge
fields = ("id", "title", "body", "repo_name", "creation_datetime",
"role_set")
| from rest_framework import serializers
from .models import Challenge, Role
from core.serializers import UserSerializer
from comments.serializers import CommentSerializer
class RoleSerializer(serializers.ModelSerializer):
#user = UserSerializer()
#challenge = serializers.RelatedField()
comment_set = CommentSerializer()
class Meta:
model = Role
fields = ("id", "user", "type", "challenge", "comment_set")
class ChallengeSerializer(serializers.ModelSerializer):
#role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True)
#role_set = serializers.RelatedField(many=True)
#role_set = serializers.SlugRelatedField(many=True, slug_field="type")
role_set = RoleSerializer(many=True)
class Meta:
model = Challenge
fields = ("id", "title", "body", "repo_name", "creation_datetime",
"role_set")
| Include comments in Role serializer | Include comments in Role serializer
| Python | mit | robrechtdr/inonemonth,robrechtdr/inonemonth,robrechtdr/inonemonth,robrechtdr/inonemonth | from rest_framework import serializers
from .models import Challenge, Role
from core.serializers import UserSerializer
+ from comments.serializers import CommentSerializer
class RoleSerializer(serializers.ModelSerializer):
- #user = serializers.RelatedField(many=True)
- #user = serializers.PrimaryKeyRelatedField()
- #user = serializers.HyperlinkedRelatedField()
- user = UserSerializer()
+ #user = UserSerializer()
- #challenge = ChallengeSerializer()
- challenge = serializers.RelatedField()
+ #challenge = serializers.RelatedField()
+ comment_set = CommentSerializer()
class Meta:
model = Role
- fields = ("id", "user", "type", "challenge")
+ fields = ("id", "user", "type", "challenge", "comment_set")
class ChallengeSerializer(serializers.ModelSerializer):
#role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True)
#role_set = serializers.RelatedField(many=True)
#role_set = serializers.SlugRelatedField(many=True, slug_field="type")
role_set = RoleSerializer(many=True)
class Meta:
model = Challenge
fields = ("id", "title", "body", "repo_name", "creation_datetime",
"role_set")
| Include comments in Role serializer | ## Code Before:
from rest_framework import serializers
from .models import Challenge, Role
from core.serializers import UserSerializer
class RoleSerializer(serializers.ModelSerializer):
#user = serializers.RelatedField(many=True)
#user = serializers.PrimaryKeyRelatedField()
#user = serializers.HyperlinkedRelatedField()
user = UserSerializer()
#challenge = ChallengeSerializer()
challenge = serializers.RelatedField()
class Meta:
model = Role
fields = ("id", "user", "type", "challenge")
class ChallengeSerializer(serializers.ModelSerializer):
#role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True)
#role_set = serializers.RelatedField(many=True)
#role_set = serializers.SlugRelatedField(many=True, slug_field="type")
role_set = RoleSerializer(many=True)
class Meta:
model = Challenge
fields = ("id", "title", "body", "repo_name", "creation_datetime",
"role_set")
## Instruction:
Include comments in Role serializer
## Code After:
from rest_framework import serializers
from .models import Challenge, Role
from core.serializers import UserSerializer
from comments.serializers import CommentSerializer
class RoleSerializer(serializers.ModelSerializer):
#user = UserSerializer()
#challenge = serializers.RelatedField()
comment_set = CommentSerializer()
class Meta:
model = Role
fields = ("id", "user", "type", "challenge", "comment_set")
class ChallengeSerializer(serializers.ModelSerializer):
#role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True)
#role_set = serializers.RelatedField(many=True)
#role_set = serializers.SlugRelatedField(many=True, slug_field="type")
role_set = RoleSerializer(many=True)
class Meta:
model = Challenge
fields = ("id", "title", "body", "repo_name", "creation_datetime",
"role_set")
| from rest_framework import serializers
from .models import Challenge, Role
from core.serializers import UserSerializer
+ from comments.serializers import CommentSerializer
class RoleSerializer(serializers.ModelSerializer):
- #user = serializers.RelatedField(many=True)
- #user = serializers.PrimaryKeyRelatedField()
- #user = serializers.HyperlinkedRelatedField()
- user = UserSerializer()
+ #user = UserSerializer()
? +
- #challenge = ChallengeSerializer()
- challenge = serializers.RelatedField()
+ #challenge = serializers.RelatedField()
? +
+ comment_set = CommentSerializer()
class Meta:
model = Role
- fields = ("id", "user", "type", "challenge")
+ fields = ("id", "user", "type", "challenge", "comment_set")
? +++++++++++++++
class ChallengeSerializer(serializers.ModelSerializer):
#role_set = serializers.HyperlinkedRelatedField(view_name="role_api_retrieve", many=True)
#role_set = serializers.RelatedField(many=True)
#role_set = serializers.SlugRelatedField(many=True, slug_field="type")
role_set = RoleSerializer(many=True)
class Meta:
model = Challenge
fields = ("id", "title", "body", "repo_name", "creation_datetime",
"role_set") |
1813413b33170f87cc9fb721c7b5a8cdecfab722 | ckanext/googleanalytics/tests/conftest.py | ckanext/googleanalytics/tests/conftest.py | import pytest
import factory
from factory.alchemy import SQLAlchemyModelFactory
from pytest_factoryboy import register
import ckan.model as model
from ckanext.googleanalytics.model import PackageStats, ResourceStats
@pytest.fixture()
def clean_db(reset_db, migrate_db_for):
reset_db()
migrate_db_for("googleanalytics")
@register
class PackageStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = PackageStats
package_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
@register
class ResourceStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = ResourceStats
resource_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
| import pytest
import factory
from factory.alchemy import SQLAlchemyModelFactory
from pytest_factoryboy import register
from ckan.plugins import toolkit
import ckan.model as model
from ckanext.googleanalytics.model import PackageStats, ResourceStats
if toolkit.requires_ckan_version("2.9"):
@pytest.fixture()
def clean_db(reset_db, migrate_db_for):
reset_db()
migrate_db_for("googleanalytics")
else:
from dbutil import init_tables
@pytest.fixture()
def clean_db(reset_db):
reset_db()
init_tables()
@register
class PackageStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = PackageStats
package_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
@register
class ResourceStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = ResourceStats
resource_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
| Fix fixture for older versions | Fix fixture for older versions
| Python | agpl-3.0 | ckan/ckanext-googleanalytics,ckan/ckanext-googleanalytics,ckan/ckanext-googleanalytics | import pytest
import factory
from factory.alchemy import SQLAlchemyModelFactory
from pytest_factoryboy import register
+ from ckan.plugins import toolkit
import ckan.model as model
from ckanext.googleanalytics.model import PackageStats, ResourceStats
+ if toolkit.requires_ckan_version("2.9"):
- @pytest.fixture()
+ @pytest.fixture()
- def clean_db(reset_db, migrate_db_for):
+ def clean_db(reset_db, migrate_db_for):
- reset_db()
+ reset_db()
- migrate_db_for("googleanalytics")
+ migrate_db_for("googleanalytics")
+ else:
+ from dbutil import init_tables
+ @pytest.fixture()
+ def clean_db(reset_db):
+ reset_db()
+ init_tables()
+
@register
class PackageStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = PackageStats
package_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
@register
class ResourceStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = ResourceStats
resource_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
| Fix fixture for older versions | ## Code Before:
import pytest
import factory
from factory.alchemy import SQLAlchemyModelFactory
from pytest_factoryboy import register
import ckan.model as model
from ckanext.googleanalytics.model import PackageStats, ResourceStats
@pytest.fixture()
def clean_db(reset_db, migrate_db_for):
reset_db()
migrate_db_for("googleanalytics")
@register
class PackageStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = PackageStats
package_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
@register
class ResourceStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = ResourceStats
resource_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
## Instruction:
Fix fixture for older versions
## Code After:
import pytest
import factory
from factory.alchemy import SQLAlchemyModelFactory
from pytest_factoryboy import register
from ckan.plugins import toolkit
import ckan.model as model
from ckanext.googleanalytics.model import PackageStats, ResourceStats
if toolkit.requires_ckan_version("2.9"):
@pytest.fixture()
def clean_db(reset_db, migrate_db_for):
reset_db()
migrate_db_for("googleanalytics")
else:
from dbutil import init_tables
@pytest.fixture()
def clean_db(reset_db):
reset_db()
init_tables()
@register
class PackageStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = PackageStats
package_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
@register
class ResourceStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = ResourceStats
resource_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
| import pytest
import factory
from factory.alchemy import SQLAlchemyModelFactory
from pytest_factoryboy import register
+ from ckan.plugins import toolkit
import ckan.model as model
from ckanext.googleanalytics.model import PackageStats, ResourceStats
+ if toolkit.requires_ckan_version("2.9"):
- @pytest.fixture()
+ @pytest.fixture()
? ++++
- def clean_db(reset_db, migrate_db_for):
+ def clean_db(reset_db, migrate_db_for):
? ++++
- reset_db()
+ reset_db()
? ++++
- migrate_db_for("googleanalytics")
+ migrate_db_for("googleanalytics")
? ++++
+ else:
+ from dbutil import init_tables
+ @pytest.fixture()
+ def clean_db(reset_db):
+ reset_db()
+ init_tables()
+
@register
class PackageStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = PackageStats
package_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint")
@register
class ResourceStatsFactory(SQLAlchemyModelFactory):
class Meta:
sqlalchemy_session = model.Session
model = ResourceStats
resource_id = factory.Faker("uuid4")
visits_recently = factory.Faker("pyint")
visits_ever = factory.Faker("pyint") |
552283714c329e3a304cd8a8bc14e5370fa6a879 | cosmo_tester/framework/constants.py | cosmo_tester/framework/constants.py | CLOUDIFY_TENANT_HEADER = 'Tenant'
SUPPORTED_RELEASES = [
'5.0.5',
'5.1.0',
'5.1.1',
'5.1.2',
'5.1.3',
'5.1.4',
'5.2.0',
'5.2.1',
'6.0.0',
'master',
]
SUPPORTED_FOR_RPM_UPGRADE = [
version + '-ga'
for version in SUPPORTED_RELEASES
if version not in ('master', '5.0.5', '5.1.0')
]
| CLOUDIFY_TENANT_HEADER = 'Tenant'
SUPPORTED_RELEASES = [
'5.0.5',
'5.1.0',
'5.1.1',
'5.1.2',
'5.1.3',
'5.1.4',
'5.2.0',
'5.2.1',
'5.2.2',
'6.0.0',
'master',
]
SUPPORTED_FOR_RPM_UPGRADE = [
version + '-ga'
for version in SUPPORTED_RELEASES
if version not in ('master', '5.0.5', '5.1.0')
]
| Add 5.2.2 to supported versions | Add 5.2.2 to supported versions
| Python | apache-2.0 | cloudify-cosmo/cloudify-system-tests,cloudify-cosmo/cloudify-system-tests | CLOUDIFY_TENANT_HEADER = 'Tenant'
SUPPORTED_RELEASES = [
'5.0.5',
'5.1.0',
'5.1.1',
'5.1.2',
'5.1.3',
'5.1.4',
'5.2.0',
'5.2.1',
+ '5.2.2',
'6.0.0',
'master',
]
SUPPORTED_FOR_RPM_UPGRADE = [
version + '-ga'
for version in SUPPORTED_RELEASES
if version not in ('master', '5.0.5', '5.1.0')
]
| Add 5.2.2 to supported versions | ## Code Before:
CLOUDIFY_TENANT_HEADER = 'Tenant'
SUPPORTED_RELEASES = [
'5.0.5',
'5.1.0',
'5.1.1',
'5.1.2',
'5.1.3',
'5.1.4',
'5.2.0',
'5.2.1',
'6.0.0',
'master',
]
SUPPORTED_FOR_RPM_UPGRADE = [
version + '-ga'
for version in SUPPORTED_RELEASES
if version not in ('master', '5.0.5', '5.1.0')
]
## Instruction:
Add 5.2.2 to supported versions
## Code After:
CLOUDIFY_TENANT_HEADER = 'Tenant'
SUPPORTED_RELEASES = [
'5.0.5',
'5.1.0',
'5.1.1',
'5.1.2',
'5.1.3',
'5.1.4',
'5.2.0',
'5.2.1',
'5.2.2',
'6.0.0',
'master',
]
SUPPORTED_FOR_RPM_UPGRADE = [
version + '-ga'
for version in SUPPORTED_RELEASES
if version not in ('master', '5.0.5', '5.1.0')
]
| CLOUDIFY_TENANT_HEADER = 'Tenant'
SUPPORTED_RELEASES = [
'5.0.5',
'5.1.0',
'5.1.1',
'5.1.2',
'5.1.3',
'5.1.4',
'5.2.0',
'5.2.1',
+ '5.2.2',
'6.0.0',
'master',
]
SUPPORTED_FOR_RPM_UPGRADE = [
version + '-ga'
for version in SUPPORTED_RELEASES
if version not in ('master', '5.0.5', '5.1.0')
] |
3bddeade05ca5ddc799733baa1545aa2b8b68060 | hoomd/tune/custom_tuner.py | hoomd/tune/custom_tuner.py | from hoomd import _hoomd
from hoomd.custom import (
_CustomOperation, _InternalCustomOperation, Action)
from hoomd.operation import _Tuner
class _TunerProperty:
@property
def updater(self):
return self._action
@updater.setter
def updater(self, updater):
if isinstance(updater, Action):
self._action = updater
else:
raise ValueError(
"updater must be an instance of hoomd.custom.Action")
class CustomTuner(_CustomOperation, _TunerProperty, _Tuner):
"""Tuner wrapper for `hoomd.custom.Action` objects.
For usage see `hoomd.custom._CustomOperation`.
"""
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
super().attach(simulation)
self._action.attach(simulation)
class _InternalCustomTuner(
_InternalCustomOperation, _TunerProperty, _Tuner):
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
| from hoomd import _hoomd
from hoomd.operation import _Operation
from hoomd.custom import (
_CustomOperation, _InternalCustomOperation, Action)
from hoomd.operation import _Tuner
class _TunerProperty:
@property
def tuner(self):
return self._action
@tuner.setter
def tuner(self, tuner):
if isinstance(tuner, Action):
self._action = tuner
else:
raise ValueError(
"updater must be an instance of hoomd.custom.Action")
class CustomTuner(_CustomOperation, _TunerProperty, _Tuner):
"""Tuner wrapper for `hoomd.custom.Action` objects.
For usage see `hoomd.custom._CustomOperation`.
"""
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
self._action.attach(simulation)
_Operation.attach(self, simulation)
class _InternalCustomTuner(
_InternalCustomOperation, _TunerProperty, _Tuner):
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
self._action.attach(simulation)
_Operation.attach(self, simulation)
| Fix attaching on custom tuners | Fix attaching on custom tuners
| Python | bsd-3-clause | joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue,joaander/hoomd-blue | from hoomd import _hoomd
+ from hoomd.operation import _Operation
from hoomd.custom import (
_CustomOperation, _InternalCustomOperation, Action)
from hoomd.operation import _Tuner
class _TunerProperty:
@property
- def updater(self):
+ def tuner(self):
return self._action
- @updater.setter
+ @tuner.setter
- def updater(self, updater):
+ def tuner(self, tuner):
- if isinstance(updater, Action):
+ if isinstance(tuner, Action):
- self._action = updater
+ self._action = tuner
else:
raise ValueError(
"updater must be an instance of hoomd.custom.Action")
class CustomTuner(_CustomOperation, _TunerProperty, _Tuner):
"""Tuner wrapper for `hoomd.custom.Action` objects.
For usage see `hoomd.custom._CustomOperation`.
"""
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
- super().attach(simulation)
self._action.attach(simulation)
+ _Operation.attach(self, simulation)
class _InternalCustomTuner(
_InternalCustomOperation, _TunerProperty, _Tuner):
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
+ def attach(self, simulation):
+ self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
+ simulation.state._cpp_sys_def, self.trigger, self._action)
+ self._action.attach(simulation)
+ _Operation.attach(self, simulation)
+ | Fix attaching on custom tuners | ## Code Before:
from hoomd import _hoomd
from hoomd.custom import (
_CustomOperation, _InternalCustomOperation, Action)
from hoomd.operation import _Tuner
class _TunerProperty:
@property
def updater(self):
return self._action
@updater.setter
def updater(self, updater):
if isinstance(updater, Action):
self._action = updater
else:
raise ValueError(
"updater must be an instance of hoomd.custom.Action")
class CustomTuner(_CustomOperation, _TunerProperty, _Tuner):
"""Tuner wrapper for `hoomd.custom.Action` objects.
For usage see `hoomd.custom._CustomOperation`.
"""
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
super().attach(simulation)
self._action.attach(simulation)
class _InternalCustomTuner(
_InternalCustomOperation, _TunerProperty, _Tuner):
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
## Instruction:
Fix attaching on custom tuners
## Code After:
from hoomd import _hoomd
from hoomd.operation import _Operation
from hoomd.custom import (
_CustomOperation, _InternalCustomOperation, Action)
from hoomd.operation import _Tuner
class _TunerProperty:
@property
def tuner(self):
return self._action
@tuner.setter
def tuner(self, tuner):
if isinstance(tuner, Action):
self._action = tuner
else:
raise ValueError(
"updater must be an instance of hoomd.custom.Action")
class CustomTuner(_CustomOperation, _TunerProperty, _Tuner):
"""Tuner wrapper for `hoomd.custom.Action` objects.
For usage see `hoomd.custom._CustomOperation`.
"""
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
self._action.attach(simulation)
_Operation.attach(self, simulation)
class _InternalCustomTuner(
_InternalCustomOperation, _TunerProperty, _Tuner):
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
self._action.attach(simulation)
_Operation.attach(self, simulation)
| from hoomd import _hoomd
+ from hoomd.operation import _Operation
from hoomd.custom import (
_CustomOperation, _InternalCustomOperation, Action)
from hoomd.operation import _Tuner
class _TunerProperty:
@property
- def updater(self):
? ^^^^
+ def tuner(self):
? + ^
return self._action
- @updater.setter
? ^^^^
+ @tuner.setter
? + ^
- def updater(self, updater):
? ^^^^ ^^^^
+ def tuner(self, tuner):
? + ^ + ^
- if isinstance(updater, Action):
? ^^^^
+ if isinstance(tuner, Action):
? + ^
- self._action = updater
? ^^^^
+ self._action = tuner
? + ^
else:
raise ValueError(
"updater must be an instance of hoomd.custom.Action")
class CustomTuner(_CustomOperation, _TunerProperty, _Tuner):
"""Tuner wrapper for `hoomd.custom.Action` objects.
For usage see `hoomd.custom._CustomOperation`.
"""
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
def attach(self, simulation):
self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
simulation.state._cpp_sys_def, self.trigger, self._action)
- super().attach(simulation)
self._action.attach(simulation)
+ _Operation.attach(self, simulation)
class _InternalCustomTuner(
_InternalCustomOperation, _TunerProperty, _Tuner):
_cpp_list_name = 'tuners'
_cpp_class_name = 'PythonTuner'
+
+ def attach(self, simulation):
+ self._cpp_obj = getattr(_hoomd, self._cpp_class_name)(
+ simulation.state._cpp_sys_def, self.trigger, self._action)
+ self._action.attach(simulation)
+ _Operation.attach(self, simulation) |
77af150756021ac4027e290b5d538e0525d812b9 | mopidy/settings.py | mopidy/settings.py | CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
SPOTIFY_USERNAME = u''
SPOTIFY_PASSWORD = u''
try:
from mopidy.local_settings import *
except ImportError:
pass
| CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
SPOTIFY_USERNAME = u''
SPOTIFY_PASSWORD = u''
try:
from mopidy.local_settings import *
except ImportError:
pass
| Add threadName to log format | Add threadName to log format
| Python | apache-2.0 | bencevans/mopidy,pacificIT/mopidy,quartz55/mopidy,SuperStarPL/mopidy,priestd09/mopidy,mokieyue/mopidy,abarisain/mopidy,hkariti/mopidy,swak/mopidy,adamcik/mopidy,quartz55/mopidy,priestd09/mopidy,pacificIT/mopidy,dbrgn/mopidy,jmarsik/mopidy,bencevans/mopidy,tkem/mopidy,abarisain/mopidy,liamw9534/mopidy,hkariti/mopidy,pacificIT/mopidy,jcass77/mopidy,diandiankan/mopidy,pacificIT/mopidy,tkem/mopidy,dbrgn/mopidy,mopidy/mopidy,jodal/mopidy,rawdlite/mopidy,vrs01/mopidy,tkem/mopidy,woutervanwijk/mopidy,ali/mopidy,hkariti/mopidy,vrs01/mopidy,woutervanwijk/mopidy,ali/mopidy,hkariti/mopidy,glogiotatidis/mopidy,mokieyue/mopidy,SuperStarPL/mopidy,mokieyue/mopidy,ZenithDK/mopidy,jmarsik/mopidy,dbrgn/mopidy,dbrgn/mopidy,kingosticks/mopidy,diandiankan/mopidy,kingosticks/mopidy,ZenithDK/mopidy,mopidy/mopidy,SuperStarPL/mopidy,glogiotatidis/mopidy,jodal/mopidy,kingosticks/mopidy,mokieyue/mopidy,glogiotatidis/mopidy,bencevans/mopidy,jmarsik/mopidy,swak/mopidy,swak/mopidy,rawdlite/mopidy,vrs01/mopidy,glogiotatidis/mopidy,SuperStarPL/mopidy,bencevans/mopidy,mopidy/mopidy,tkem/mopidy,ali/mopidy,ZenithDK/mopidy,adamcik/mopidy,liamw9534/mopidy,jcass77/mopidy,jcass77/mopidy,jmarsik/mopidy,adamcik/mopidy,rawdlite/mopidy,ali/mopidy,ZenithDK/mopidy,quartz55/mopidy,jodal/mopidy,swak/mopidy,vrs01/mopidy,rawdlite/mopidy,bacontext/mopidy,bacontext/mopidy,diandiankan/mopidy,bacontext/mopidy,quartz55/mopidy,bacontext/mopidy,priestd09/mopidy,diandiankan/mopidy | - CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
+ CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
SPOTIFY_USERNAME = u''
SPOTIFY_PASSWORD = u''
try:
from mopidy.local_settings import *
except ImportError:
pass
| Add threadName to log format | ## Code Before:
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
SPOTIFY_USERNAME = u''
SPOTIFY_PASSWORD = u''
try:
from mopidy.local_settings import *
except ImportError:
pass
## Instruction:
Add threadName to log format
## Code After:
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
SPOTIFY_USERNAME = u''
SPOTIFY_PASSWORD = u''
try:
from mopidy.local_settings import *
except ImportError:
pass
| - CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
+ CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s [%(threadName)s] %(name)s\n %(message)s'
? +++++++++++++++++
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
SPOTIFY_USERNAME = u''
SPOTIFY_PASSWORD = u''
try:
from mopidy.local_settings import *
except ImportError:
pass
|
f5bbdea74c0f8a0cc8ac4331ea8adc45c3f266c8 | converter.py | converter.py |
def main():
command = ""
while command != "exit":
command = input("> ").strip()
if __name__ == "__main__":
main()
|
import sys
def main():
if len(sys.argv) > 1:
for i in range(1, len(sys.argv)):
with open(sys.argv[i]) as f:
lines = f.read().splitlines()
for line in lines:
evaluate(line);
command = ""
while command != "exit":
command = input("> ").strip()
def evaluate(string):
print(string)
if __name__ == "__main__":
main()
| Read files defined as argument | Read files defined as argument
| Python | mit | hampustagerud/colorconverter | +
+ import sys
def main():
+ if len(sys.argv) > 1:
+ for i in range(1, len(sys.argv)):
+ with open(sys.argv[i]) as f:
+ lines = f.read().splitlines()
+ for line in lines:
+ evaluate(line);
+
command = ""
while command != "exit":
command = input("> ").strip()
+
+ def evaluate(string):
+ print(string)
if __name__ == "__main__":
main()
| Read files defined as argument | ## Code Before:
def main():
command = ""
while command != "exit":
command = input("> ").strip()
if __name__ == "__main__":
main()
## Instruction:
Read files defined as argument
## Code After:
import sys
def main():
if len(sys.argv) > 1:
for i in range(1, len(sys.argv)):
with open(sys.argv[i]) as f:
lines = f.read().splitlines()
for line in lines:
evaluate(line);
command = ""
while command != "exit":
command = input("> ").strip()
def evaluate(string):
print(string)
if __name__ == "__main__":
main()
| +
+ import sys
def main():
+ if len(sys.argv) > 1:
+ for i in range(1, len(sys.argv)):
+ with open(sys.argv[i]) as f:
+ lines = f.read().splitlines()
+ for line in lines:
+ evaluate(line);
+
command = ""
while command != "exit":
command = input("> ").strip()
+ def evaluate(string):
+ print(string)
+
if __name__ == "__main__":
main()
|
d7e9eba6fb3628f0736bd468ae76e05099b9d651 | space/decorators.py | space/decorators.py | from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from incubator.settings import STATUS_SECRETS
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| Use from django.conf import settings | Use from django.conf import settings
| Python | agpl-3.0 | UrLab/incubator,UrLab/incubator,UrLab/incubator,UrLab/incubator | from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
- from incubator.settings import STATUS_SECRETS
+ from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
- if request.POST['secret'] not in STATUS_SECRETS:
+ if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| Use from django.conf import settings | ## Code Before:
from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from incubator.settings import STATUS_SECRETS
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
## Instruction:
Use from django.conf import settings
## Code After:
from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
- from incubator.settings import STATUS_SECRETS
+ from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
- if request.POST['secret'] not in STATUS_SECRETS:
+ if request.POST['secret'] not in settings.STATUS_SECRETS:
? +++++++++
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer |
07bc7efb756e2bc99f59c59476379bc186f36143 | sktracker/io/__init__.py | sktracker/io/__init__.py |
from .tifffile import imsave
from .tifffile import imread
from .tifffile import imshow
from .tifffile import TiffFile
from .tifffile import TiffSequence
from .ome import OMEModel
from .tiff_metadata import get_metadata_from_tiff
__all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow',
'TiffFile', 'TiffSequence']
|
# Remove warnings for tifffile.py
import warnings
warnings.filterwarnings("ignore")
from .tifffile import imsave
from .tifffile import imread
from .tifffile import imshow
from .tifffile import TiffFile
from .tifffile import TiffSequence
from .ome import OMEModel
from .tiff_metadata import get_metadata_from_tiff
__all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow',
'TiffFile', 'TiffSequence']
| Remove warning messages for tifffile.py | Remove warning messages for tifffile.py
| Python | bsd-3-clause | bnoi/scikit-tracker,bnoi/scikit-tracker,bnoi/scikit-tracker | +
+ # Remove warnings for tifffile.py
+ import warnings
+ warnings.filterwarnings("ignore")
from .tifffile import imsave
from .tifffile import imread
from .tifffile import imshow
from .tifffile import TiffFile
from .tifffile import TiffSequence
from .ome import OMEModel
from .tiff_metadata import get_metadata_from_tiff
__all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow',
'TiffFile', 'TiffSequence']
| Remove warning messages for tifffile.py | ## Code Before:
from .tifffile import imsave
from .tifffile import imread
from .tifffile import imshow
from .tifffile import TiffFile
from .tifffile import TiffSequence
from .ome import OMEModel
from .tiff_metadata import get_metadata_from_tiff
__all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow',
'TiffFile', 'TiffSequence']
## Instruction:
Remove warning messages for tifffile.py
## Code After:
# Remove warnings for tifffile.py
import warnings
warnings.filterwarnings("ignore")
from .tifffile import imsave
from .tifffile import imread
from .tifffile import imshow
from .tifffile import TiffFile
from .tifffile import TiffSequence
from .ome import OMEModel
from .tiff_metadata import get_metadata_from_tiff
__all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow',
'TiffFile', 'TiffSequence']
| +
+ # Remove warnings for tifffile.py
+ import warnings
+ warnings.filterwarnings("ignore")
from .tifffile import imsave
from .tifffile import imread
from .tifffile import imshow
from .tifffile import TiffFile
from .tifffile import TiffSequence
from .ome import OMEModel
from .tiff_metadata import get_metadata_from_tiff
__all__ = ['get_metadata_from_tiff', 'OMEModel', 'imsave', 'imread', 'imshow',
'TiffFile', 'TiffSequence'] |
df01d83a35c59ee2295b358d8eceeb842adb568d | listings/admin.py | listings/admin.py | from django.contrib import admin
from .models import Region, City, GatheringCenter, Resource
admin.site.register(Region)
admin.site.register(City)
class PublishMixin(object):
actions = ('publish', 'unpublish')
def publish(self, request, queryset):
queryset.update(published=True)
def unpublish(self, request, queryset):
queryset.update(published=False)
class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin):
list_filter = ('published', 'city', 'created')
list_editable = ('published', )
list_display = ('location_name', 'created', 'published', 'author', 'city')
raw_id_fields = ('author', )
admin.site.register(GatheringCenter, GatheringCenterAdmin)
class ResourceAdmin(PublishMixin, admin.ModelAdmin):
list_filter = ('published', 'created')
list_editable = ('published', )
list_display = ('name', 'created', 'published', 'author', 'url', 'country')
raw_id_fields = ('author', )
admin.site.register(Resource, ResourceAdmin)
| from django.contrib import admin
from .models import Region, City, GatheringCenter, Resource
admin.site.register(Region)
admin.site.register(City)
class PublishMixin(object):
actions = ('publish', 'unpublish')
def publish(self, request, queryset):
queryset.update(published=True)
def unpublish(self, request, queryset):
queryset.update(published=False)
class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin):
date_hierarchy = 'created'
list_filter = ('published', 'city', )
list_editable = ('published', )
list_display = ('location_name', 'created', 'published', 'author', 'city')
raw_id_fields = ('author', )
admin.site.register(GatheringCenter, GatheringCenterAdmin)
class ResourceAdmin(PublishMixin, admin.ModelAdmin):
date_hierarchy = 'created'
list_filter = ('published', )
list_editable = ('published', )
list_display = ('name', 'created', 'published', 'author', 'url', 'country')
raw_id_fields = ('author', )
admin.site.register(Resource, ResourceAdmin)
| Order by create, don't filter | Order by create, don't filter
| Python | apache-2.0 | pony-revolution/helpothers,pony-revolution/helpothers,pony-revolution/helpothers | from django.contrib import admin
from .models import Region, City, GatheringCenter, Resource
admin.site.register(Region)
admin.site.register(City)
class PublishMixin(object):
actions = ('publish', 'unpublish')
def publish(self, request, queryset):
queryset.update(published=True)
def unpublish(self, request, queryset):
queryset.update(published=False)
class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin):
+ date_hierarchy = 'created'
- list_filter = ('published', 'city', 'created')
+ list_filter = ('published', 'city', )
list_editable = ('published', )
list_display = ('location_name', 'created', 'published', 'author', 'city')
raw_id_fields = ('author', )
admin.site.register(GatheringCenter, GatheringCenterAdmin)
class ResourceAdmin(PublishMixin, admin.ModelAdmin):
+ date_hierarchy = 'created'
- list_filter = ('published', 'created')
+ list_filter = ('published', )
list_editable = ('published', )
list_display = ('name', 'created', 'published', 'author', 'url', 'country')
raw_id_fields = ('author', )
admin.site.register(Resource, ResourceAdmin)
| Order by create, don't filter | ## Code Before:
from django.contrib import admin
from .models import Region, City, GatheringCenter, Resource
admin.site.register(Region)
admin.site.register(City)
class PublishMixin(object):
actions = ('publish', 'unpublish')
def publish(self, request, queryset):
queryset.update(published=True)
def unpublish(self, request, queryset):
queryset.update(published=False)
class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin):
list_filter = ('published', 'city', 'created')
list_editable = ('published', )
list_display = ('location_name', 'created', 'published', 'author', 'city')
raw_id_fields = ('author', )
admin.site.register(GatheringCenter, GatheringCenterAdmin)
class ResourceAdmin(PublishMixin, admin.ModelAdmin):
list_filter = ('published', 'created')
list_editable = ('published', )
list_display = ('name', 'created', 'published', 'author', 'url', 'country')
raw_id_fields = ('author', )
admin.site.register(Resource, ResourceAdmin)
## Instruction:
Order by create, don't filter
## Code After:
from django.contrib import admin
from .models import Region, City, GatheringCenter, Resource
admin.site.register(Region)
admin.site.register(City)
class PublishMixin(object):
actions = ('publish', 'unpublish')
def publish(self, request, queryset):
queryset.update(published=True)
def unpublish(self, request, queryset):
queryset.update(published=False)
class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin):
date_hierarchy = 'created'
list_filter = ('published', 'city', )
list_editable = ('published', )
list_display = ('location_name', 'created', 'published', 'author', 'city')
raw_id_fields = ('author', )
admin.site.register(GatheringCenter, GatheringCenterAdmin)
class ResourceAdmin(PublishMixin, admin.ModelAdmin):
date_hierarchy = 'created'
list_filter = ('published', )
list_editable = ('published', )
list_display = ('name', 'created', 'published', 'author', 'url', 'country')
raw_id_fields = ('author', )
admin.site.register(Resource, ResourceAdmin)
| from django.contrib import admin
from .models import Region, City, GatheringCenter, Resource
admin.site.register(Region)
admin.site.register(City)
class PublishMixin(object):
actions = ('publish', 'unpublish')
def publish(self, request, queryset):
queryset.update(published=True)
def unpublish(self, request, queryset):
queryset.update(published=False)
class GatheringCenterAdmin(PublishMixin, admin.ModelAdmin):
+ date_hierarchy = 'created'
- list_filter = ('published', 'city', 'created')
? ---------
+ list_filter = ('published', 'city', )
list_editable = ('published', )
list_display = ('location_name', 'created', 'published', 'author', 'city')
raw_id_fields = ('author', )
admin.site.register(GatheringCenter, GatheringCenterAdmin)
class ResourceAdmin(PublishMixin, admin.ModelAdmin):
+ date_hierarchy = 'created'
- list_filter = ('published', 'created')
? ---------
+ list_filter = ('published', )
list_editable = ('published', )
list_display = ('name', 'created', 'published', 'author', 'url', 'country')
raw_id_fields = ('author', )
admin.site.register(Resource, ResourceAdmin)
|
0024b8b921d788a0539bc242bd1600c0da666bd6 | panoptes/state_machine/states/core.py | panoptes/state_machine/states/core.py | import time
import transitions
from panoptes.utils.logger import has_logger
@has_logger
class PanState(transitions.State):
""" Base class for PANOPTES transitions """
def __init__(self, *args, **kwargs):
name = kwargs.get('name', self.__class__)
self.panoptes = kwargs.get('panoptes', None)
super().__init__(name=name, on_enter=['execute'])
self._sleep_delay = 3 # seconds
def main(self):
assert self.panoptes is not None
msg = "Must implement `main` method inside class {}. Exiting".format(self.name)
self.panoptes.logger.warning(msg)
return 'exit'
def sleep(self, seconds=None):
""" sleep for `seconds` or `_sleep_delay` seconds
This puts the state into a loop that is responsive to outside messages.
Args:
seconds(float): Seconds to sleep for, defaults to `_sleep_delay`.
"""
assert self.panoptes is not None
if seconds is None:
seconds = self._sleep_delay
if seconds > 10:
step_time = seconds / 4
while seconds:
seconds = seconds - step_time
# NOTE: DO SOMETHING RESPONSIVE HERE
time.sleep(step_time)
else:
time.sleep(seconds)
| import time
import transitions
from panoptes.utils.logger import has_logger
@has_logger
class PanState(transitions.State):
""" Base class for PANOPTES transitions """
def __init__(self, *args, **kwargs):
name = kwargs.get('name', self.__class__)
self.panoptes = kwargs.get('panoptes', None)
super().__init__(name=name, on_enter=['execute'])
self._sleep_delay = 3 # seconds
def main(self):
assert self.panoptes is not None
msg = "Must implement `main` method inside class {}. Exiting".format(self.name)
self.panoptes.logger.warning(msg)
def sleep(self, seconds=None):
""" sleep for `seconds` or `_sleep_delay` seconds
This puts the state into a loop that is responsive to outside messages.
Args:
seconds(float): Seconds to sleep for, defaults to `_sleep_delay`.
"""
assert self.panoptes is not None
if seconds is None:
seconds = self._sleep_delay
if seconds > 10:
step_time = seconds / 4
while seconds:
seconds = seconds - step_time
# NOTE: DO SOMETHING RESPONSIVE HERE
time.sleep(step_time)
else:
time.sleep(seconds)
| Remove return state from main `main` | Remove return state from main `main`
| Python | mit | panoptes/POCS,panoptes/POCS,joshwalawender/POCS,AstroHuntsman/POCS,joshwalawender/POCS,joshwalawender/POCS,panoptes/POCS,panoptes/POCS,AstroHuntsman/POCS,AstroHuntsman/POCS,AstroHuntsman/POCS | import time
import transitions
from panoptes.utils.logger import has_logger
@has_logger
class PanState(transitions.State):
""" Base class for PANOPTES transitions """
def __init__(self, *args, **kwargs):
name = kwargs.get('name', self.__class__)
self.panoptes = kwargs.get('panoptes', None)
super().__init__(name=name, on_enter=['execute'])
self._sleep_delay = 3 # seconds
def main(self):
assert self.panoptes is not None
msg = "Must implement `main` method inside class {}. Exiting".format(self.name)
self.panoptes.logger.warning(msg)
- return 'exit'
def sleep(self, seconds=None):
""" sleep for `seconds` or `_sleep_delay` seconds
This puts the state into a loop that is responsive to outside messages.
Args:
seconds(float): Seconds to sleep for, defaults to `_sleep_delay`.
"""
assert self.panoptes is not None
if seconds is None:
seconds = self._sleep_delay
if seconds > 10:
step_time = seconds / 4
while seconds:
seconds = seconds - step_time
# NOTE: DO SOMETHING RESPONSIVE HERE
time.sleep(step_time)
else:
time.sleep(seconds)
| Remove return state from main `main` | ## Code Before:
import time
import transitions
from panoptes.utils.logger import has_logger
@has_logger
class PanState(transitions.State):
""" Base class for PANOPTES transitions """
def __init__(self, *args, **kwargs):
name = kwargs.get('name', self.__class__)
self.panoptes = kwargs.get('panoptes', None)
super().__init__(name=name, on_enter=['execute'])
self._sleep_delay = 3 # seconds
def main(self):
assert self.panoptes is not None
msg = "Must implement `main` method inside class {}. Exiting".format(self.name)
self.panoptes.logger.warning(msg)
return 'exit'
def sleep(self, seconds=None):
""" sleep for `seconds` or `_sleep_delay` seconds
This puts the state into a loop that is responsive to outside messages.
Args:
seconds(float): Seconds to sleep for, defaults to `_sleep_delay`.
"""
assert self.panoptes is not None
if seconds is None:
seconds = self._sleep_delay
if seconds > 10:
step_time = seconds / 4
while seconds:
seconds = seconds - step_time
# NOTE: DO SOMETHING RESPONSIVE HERE
time.sleep(step_time)
else:
time.sleep(seconds)
## Instruction:
Remove return state from main `main`
## Code After:
import time
import transitions
from panoptes.utils.logger import has_logger
@has_logger
class PanState(transitions.State):
""" Base class for PANOPTES transitions """
def __init__(self, *args, **kwargs):
name = kwargs.get('name', self.__class__)
self.panoptes = kwargs.get('panoptes', None)
super().__init__(name=name, on_enter=['execute'])
self._sleep_delay = 3 # seconds
def main(self):
assert self.panoptes is not None
msg = "Must implement `main` method inside class {}. Exiting".format(self.name)
self.panoptes.logger.warning(msg)
def sleep(self, seconds=None):
""" sleep for `seconds` or `_sleep_delay` seconds
This puts the state into a loop that is responsive to outside messages.
Args:
seconds(float): Seconds to sleep for, defaults to `_sleep_delay`.
"""
assert self.panoptes is not None
if seconds is None:
seconds = self._sleep_delay
if seconds > 10:
step_time = seconds / 4
while seconds:
seconds = seconds - step_time
# NOTE: DO SOMETHING RESPONSIVE HERE
time.sleep(step_time)
else:
time.sleep(seconds)
| import time
import transitions
from panoptes.utils.logger import has_logger
@has_logger
class PanState(transitions.State):
""" Base class for PANOPTES transitions """
def __init__(self, *args, **kwargs):
name = kwargs.get('name', self.__class__)
self.panoptes = kwargs.get('panoptes', None)
super().__init__(name=name, on_enter=['execute'])
self._sleep_delay = 3 # seconds
def main(self):
assert self.panoptes is not None
msg = "Must implement `main` method inside class {}. Exiting".format(self.name)
self.panoptes.logger.warning(msg)
- return 'exit'
def sleep(self, seconds=None):
""" sleep for `seconds` or `_sleep_delay` seconds
This puts the state into a loop that is responsive to outside messages.
Args:
seconds(float): Seconds to sleep for, defaults to `_sleep_delay`.
"""
assert self.panoptes is not None
if seconds is None:
seconds = self._sleep_delay
if seconds > 10:
step_time = seconds / 4
while seconds:
seconds = seconds - step_time
# NOTE: DO SOMETHING RESPONSIVE HERE
time.sleep(step_time)
else:
time.sleep(seconds) |
ce25cea7e8d10f9c318e2e7ef1dc1013921ed062 | clint/textui/prompt.py | clint/textui/prompt.py |
from __future__ import absolute_import
from re import match, I
def yn(prompt, default='y', batch=False):
# A sanity check against default value
# If not y/n then y is assumed
if default not in ['y', 'n']:
default = 'y'
# Let's build the prompt
choicebox = '[Y/n]' if default == 'y' else '[y/N]'
prompt = prompt + ' ' + choicebox + ' '
# If input is not a yes/no variant or empty
# keep asking
while True:
# If batch option is True then auto reply
# with default input
if not batch:
input = raw_input(prompt).strip()
else:
print prompt
input = ''
# If input is empty default choice is assumed
# so we return True
if input == '':
return True
# Given 'yes' as input if default choice is y
# then return True, False otherwise
if match('y(?:es)?', input, I):
return True if default == 'y' else False
# Given 'no' as input if default choice is n
# then return True, False otherwise
elif match('n(?:o)?', input, I):
return True if default == 'n' else False
|
from __future__ import absolute_import, print_function
from re import match, I
def yn(prompt, default='y', batch=False):
# A sanity check against default value
# If not y/n then y is assumed
if default not in ['y', 'n']:
default = 'y'
# Let's build the prompt
choicebox = '[Y/n]' if default == 'y' else '[y/N]'
prompt = prompt + ' ' + choicebox + ' '
# If input is not a yes/no variant or empty
# keep asking
while True:
# If batch option is True then auto reply
# with default input
if not batch:
input = raw_input(prompt).strip()
else:
print(prompt)
input = ''
# If input is empty default choice is assumed
# so we return True
if input == '':
return True
# Given 'yes' as input if default choice is y
# then return True, False otherwise
if match('y(?:es)?', input, I):
return True if default == 'y' else False
# Given 'no' as input if default choice is n
# then return True, False otherwise
elif match('n(?:o)?', input, I):
return True if default == 'n' else False
| Use print() function to fix install on python 3 | Use print() function to fix install on python 3
clint 0.3.2 can't be installed on python 3.3 because of a print statement. | Python | isc | 1gitGrey/clint,thusoy/clint,wkentaro/clint,1gitGrey/clint,glorizen/clint,wkentaro/clint,tz70s/clint,Lh4cKg/clint,nathancahill/clint,kennethreitz/clint,nathancahill/clint |
- from __future__ import absolute_import
+ from __future__ import absolute_import, print_function
from re import match, I
def yn(prompt, default='y', batch=False):
# A sanity check against default value
# If not y/n then y is assumed
if default not in ['y', 'n']:
default = 'y'
# Let's build the prompt
choicebox = '[Y/n]' if default == 'y' else '[y/N]'
prompt = prompt + ' ' + choicebox + ' '
# If input is not a yes/no variant or empty
# keep asking
while True:
# If batch option is True then auto reply
# with default input
if not batch:
input = raw_input(prompt).strip()
else:
- print prompt
+ print(prompt)
input = ''
# If input is empty default choice is assumed
# so we return True
if input == '':
return True
# Given 'yes' as input if default choice is y
# then return True, False otherwise
if match('y(?:es)?', input, I):
return True if default == 'y' else False
# Given 'no' as input if default choice is n
# then return True, False otherwise
elif match('n(?:o)?', input, I):
return True if default == 'n' else False
| Use print() function to fix install on python 3 | ## Code Before:
from __future__ import absolute_import
from re import match, I
def yn(prompt, default='y', batch=False):
# A sanity check against default value
# If not y/n then y is assumed
if default not in ['y', 'n']:
default = 'y'
# Let's build the prompt
choicebox = '[Y/n]' if default == 'y' else '[y/N]'
prompt = prompt + ' ' + choicebox + ' '
# If input is not a yes/no variant or empty
# keep asking
while True:
# If batch option is True then auto reply
# with default input
if not batch:
input = raw_input(prompt).strip()
else:
print prompt
input = ''
# If input is empty default choice is assumed
# so we return True
if input == '':
return True
# Given 'yes' as input if default choice is y
# then return True, False otherwise
if match('y(?:es)?', input, I):
return True if default == 'y' else False
# Given 'no' as input if default choice is n
# then return True, False otherwise
elif match('n(?:o)?', input, I):
return True if default == 'n' else False
## Instruction:
Use print() function to fix install on python 3
## Code After:
from __future__ import absolute_import, print_function
from re import match, I
def yn(prompt, default='y', batch=False):
# A sanity check against default value
# If not y/n then y is assumed
if default not in ['y', 'n']:
default = 'y'
# Let's build the prompt
choicebox = '[Y/n]' if default == 'y' else '[y/N]'
prompt = prompt + ' ' + choicebox + ' '
# If input is not a yes/no variant or empty
# keep asking
while True:
# If batch option is True then auto reply
# with default input
if not batch:
input = raw_input(prompt).strip()
else:
print(prompt)
input = ''
# If input is empty default choice is assumed
# so we return True
if input == '':
return True
# Given 'yes' as input if default choice is y
# then return True, False otherwise
if match('y(?:es)?', input, I):
return True if default == 'y' else False
# Given 'no' as input if default choice is n
# then return True, False otherwise
elif match('n(?:o)?', input, I):
return True if default == 'n' else False
|
- from __future__ import absolute_import
+ from __future__ import absolute_import, print_function
? ++++++++++++++++
from re import match, I
def yn(prompt, default='y', batch=False):
# A sanity check against default value
# If not y/n then y is assumed
if default not in ['y', 'n']:
default = 'y'
# Let's build the prompt
choicebox = '[Y/n]' if default == 'y' else '[y/N]'
prompt = prompt + ' ' + choicebox + ' '
# If input is not a yes/no variant or empty
# keep asking
while True:
# If batch option is True then auto reply
# with default input
if not batch:
input = raw_input(prompt).strip()
else:
- print prompt
? ^
+ print(prompt)
? ^ +
input = ''
# If input is empty default choice is assumed
# so we return True
if input == '':
return True
# Given 'yes' as input if default choice is y
# then return True, False otherwise
if match('y(?:es)?', input, I):
return True if default == 'y' else False
# Given 'no' as input if default choice is n
# then return True, False otherwise
elif match('n(?:o)?', input, I):
return True if default == 'n' else False |
692234e72862839d8c14fb0f1a6ebe7259b15413 | core/report.py | core/report.py | import config
def sendMail():
print config.config
print "Sent e-mail"
def sendToGrapite():
pass
| from config import config as conf
from email.mime.multipart import MIMEMultipart
from email.mime.text import MIMEText
import smtplib
import time
import pushnotify
def sendMail(recipients, subject, body):
if not isinstance( recipients, list ):
recipients = [ recipients ]
session = smtplib.SMTP( conf.get( 'gmail', 'server' ),
conf.getint( 'gmail', 'port' ) )
session.ehlo()
session.starttls()
session.login( conf.get( 'gmail', 'username' ),
conf.get( 'gmail', 'password' ) )
for recipient in recipients:
headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ),
"subject: " + subject,
"to: " + recipient,
"mime-version: 1.0",
"content-type: text/html" ] )
content = headers + "\r\n\r\n" + body
session.sendmail( conf.get( 'gmail', 'from' ), recipient, content )
def sendNotification(application, desc, event):
client = pushnotify.get_client('nma', application=application )
client.add_key( conf.get( 'notifymyandroid', 'api_key' ) )
try:
client.notify( desc, event, split=True )
except:
pass
def sendToGrapite(path, value):
message = '%s %s %d\n' % ( path, value, int( time.time() ) )
sock = socket.socket()
graphite_address = ( conf.get( 'graphite', 'server' ),
conf.get( 'graphite', 'port' ) )
sock.connect( graphite_address )
sock.sendall( message )
sock.close()
| Complete e-mail, Graphite and push notification support | Complete e-mail, Graphite and push notification support
| Python | mit | nlindblad/ocarina,nlindblad/ocarina | - import config
+ from config import config as conf
+ from email.mime.multipart import MIMEMultipart
+ from email.mime.text import MIMEText
+ import smtplib
+ import time
+ import pushnotify
- def sendMail():
- print config.config
- print "Sent e-mail"
- def sendToGrapite():
+ def sendMail(recipients, subject, body):
+ if not isinstance( recipients, list ):
+ recipients = [ recipients ]
+ session = smtplib.SMTP( conf.get( 'gmail', 'server' ),
+ conf.getint( 'gmail', 'port' ) )
+ session.ehlo()
+ session.starttls()
+ session.login( conf.get( 'gmail', 'username' ),
+ conf.get( 'gmail', 'password' ) )
+ for recipient in recipients:
+ headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ),
+ "subject: " + subject,
+ "to: " + recipient,
+ "mime-version: 1.0",
+ "content-type: text/html" ] )
+ content = headers + "\r\n\r\n" + body
+ session.sendmail( conf.get( 'gmail', 'from' ), recipient, content )
+
+ def sendNotification(application, desc, event):
+ client = pushnotify.get_client('nma', application=application )
+ client.add_key( conf.get( 'notifymyandroid', 'api_key' ) )
+ try:
+ client.notify( desc, event, split=True )
+ except:
- pass
+ pass
+ def sendToGrapite(path, value):
+ message = '%s %s %d\n' % ( path, value, int( time.time() ) )
+ sock = socket.socket()
+ graphite_address = ( conf.get( 'graphite', 'server' ),
+ conf.get( 'graphite', 'port' ) )
+ sock.connect( graphite_address )
+ sock.sendall( message )
+ sock.close()
+
+ | Complete e-mail, Graphite and push notification support | ## Code Before:
import config
def sendMail():
print config.config
print "Sent e-mail"
def sendToGrapite():
pass
## Instruction:
Complete e-mail, Graphite and push notification support
## Code After:
from config import config as conf
from email.mime.multipart import MIMEMultipart
from email.mime.text import MIMEText
import smtplib
import time
import pushnotify
def sendMail(recipients, subject, body):
if not isinstance( recipients, list ):
recipients = [ recipients ]
session = smtplib.SMTP( conf.get( 'gmail', 'server' ),
conf.getint( 'gmail', 'port' ) )
session.ehlo()
session.starttls()
session.login( conf.get( 'gmail', 'username' ),
conf.get( 'gmail', 'password' ) )
for recipient in recipients:
headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ),
"subject: " + subject,
"to: " + recipient,
"mime-version: 1.0",
"content-type: text/html" ] )
content = headers + "\r\n\r\n" + body
session.sendmail( conf.get( 'gmail', 'from' ), recipient, content )
def sendNotification(application, desc, event):
client = pushnotify.get_client('nma', application=application )
client.add_key( conf.get( 'notifymyandroid', 'api_key' ) )
try:
client.notify( desc, event, split=True )
except:
pass
def sendToGrapite(path, value):
message = '%s %s %d\n' % ( path, value, int( time.time() ) )
sock = socket.socket()
graphite_address = ( conf.get( 'graphite', 'server' ),
conf.get( 'graphite', 'port' ) )
sock.connect( graphite_address )
sock.sendall( message )
sock.close()
| - import config
+ from config import config as conf
+ from email.mime.multipart import MIMEMultipart
+ from email.mime.text import MIMEText
+ import smtplib
+ import time
+ import pushnotify
- def sendMail():
- print config.config
- print "Sent e-mail"
- def sendToGrapite():
+ def sendMail(recipients, subject, body):
+ if not isinstance( recipients, list ):
+ recipients = [ recipients ]
+ session = smtplib.SMTP( conf.get( 'gmail', 'server' ),
+ conf.getint( 'gmail', 'port' ) )
+ session.ehlo()
+ session.starttls()
+ session.login( conf.get( 'gmail', 'username' ),
+ conf.get( 'gmail', 'password' ) )
+ for recipient in recipients:
+ headers = "\r\n".join( [ "from: " + conf.get( 'gmail', 'from' ),
+ "subject: " + subject,
+ "to: " + recipient,
+ "mime-version: 1.0",
+ "content-type: text/html" ] )
+ content = headers + "\r\n\r\n" + body
+ session.sendmail( conf.get( 'gmail', 'from' ), recipient, content )
+
+ def sendNotification(application, desc, event):
+ client = pushnotify.get_client('nma', application=application )
+ client.add_key( conf.get( 'notifymyandroid', 'api_key' ) )
+ try:
+ client.notify( desc, event, split=True )
+ except:
- pass
+ pass
? ++++
+
+ def sendToGrapite(path, value):
+ message = '%s %s %d\n' % ( path, value, int( time.time() ) )
+ sock = socket.socket()
+ graphite_address = ( conf.get( 'graphite', 'server' ),
+ conf.get( 'graphite', 'port' ) )
+ sock.connect( graphite_address )
+ sock.sendall( message )
+ sock.close()
+ |
5cd3b53f677fd6ab6e77bee5b7d42cf2ac85e47f | feincms/apps.py | feincms/apps.py |
from feincms.content.application.models import *
| def __getattr__(key):
# Work around Django 3.2's autoloading of *.apps modules (AppConfig
# autodiscovery)
if key in {
"ApplicationContent",
"app_reverse",
"app_reverse_lazy",
"permalink",
"UnpackTemplateResponse",
"standalone",
"unpack",
}:
from feincms.content.application import models
return getattr(models, key)
raise AttributeError("Unknown attribute '%s'" % key)
| Add a workaround for the AppConfig autodiscovery crashes with Django 3.2 | Add a workaround for the AppConfig autodiscovery crashes with Django 3.2
| Python | bsd-3-clause | mjl/feincms,feincms/feincms,mjl/feincms,feincms/feincms,feincms/feincms,mjl/feincms | + def __getattr__(key):
+ # Work around Django 3.2's autoloading of *.apps modules (AppConfig
+ # autodiscovery)
+ if key in {
+ "ApplicationContent",
+ "app_reverse",
+ "app_reverse_lazy",
+ "permalink",
+ "UnpackTemplateResponse",
+ "standalone",
+ "unpack",
+ }:
+ from feincms.content.application import models
- from feincms.content.application.models import *
+ return getattr(models, key)
+ raise AttributeError("Unknown attribute '%s'" % key)
+ | Add a workaround for the AppConfig autodiscovery crashes with Django 3.2 | ## Code Before:
from feincms.content.application.models import *
## Instruction:
Add a workaround for the AppConfig autodiscovery crashes with Django 3.2
## Code After:
def __getattr__(key):
# Work around Django 3.2's autoloading of *.apps modules (AppConfig
# autodiscovery)
if key in {
"ApplicationContent",
"app_reverse",
"app_reverse_lazy",
"permalink",
"UnpackTemplateResponse",
"standalone",
"unpack",
}:
from feincms.content.application import models
return getattr(models, key)
raise AttributeError("Unknown attribute '%s'" % key)
| + def __getattr__(key):
+ # Work around Django 3.2's autoloading of *.apps modules (AppConfig
+ # autodiscovery)
+ if key in {
+ "ApplicationContent",
+ "app_reverse",
+ "app_reverse_lazy",
+ "permalink",
+ "UnpackTemplateResponse",
+ "standalone",
+ "unpack",
+ }:
+ from feincms.content.application import models
- from feincms.content.application.models import *
+ return getattr(models, key)
+
+ raise AttributeError("Unknown attribute '%s'" % key) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.