commit stringlengths 40 40 | old_file stringlengths 4 106 | new_file stringlengths 4 106 | old_contents stringlengths 10 2.94k | new_contents stringlengths 21 2.95k | subject stringlengths 16 444 | message stringlengths 17 2.63k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 7 43k | ndiff stringlengths 52 3.31k | instruction stringlengths 16 444 | content stringlengths 133 4.32k | diff stringlengths 49 3.61k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4522de348aab4cc99904b0bc210c223b2477b4b7 | tests/config.py | tests/config.py |
import os
local_path = os.path.dirname(__file__)
xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml'))
csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv'))
bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif'))
tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx'))
|
import os
local_path = os.path.dirname(__file__)
xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml'))
csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv'))
bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif'))
pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
| Use pyt file instead of stand-alone tbx for testing. | Use pyt file instead of stand-alone tbx for testing.
| Python | mpl-2.0 | EsriOceans/btm |
import os
local_path = os.path.dirname(__file__)
xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml'))
csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv'))
bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif'))
- tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx'))
+ pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
| Use pyt file instead of stand-alone tbx for testing. | ## Code Before:
import os
local_path = os.path.dirname(__file__)
xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml'))
csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv'))
bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif'))
tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx'))
## Instruction:
Use pyt file instead of stand-alone tbx for testing.
## Code After:
import os
local_path = os.path.dirname(__file__)
xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml'))
csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv'))
bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif'))
pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
|
import os
local_path = os.path.dirname(__file__)
xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml'))
csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv'))
bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif'))
- tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx'))
? -- ------ --
+ pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
? ++ ++
|
07d2ffe3c14a6c908a7bf138f40ba8d49bf7b2c3 | examples/plot_grow.py | examples/plot_grow.py |
# Code source: Óscar Nájera
# License: BSD 3 clause
import numpy as np
import matplotlib.pyplot as plt
x = np.linspace(-1, 2, 100)
y = np.exp(x)
plt.plot(x, y)
plt.xlabel('$x$')
plt.ylabel('$exp(x)$')
plt.show()
|
# Code source: Óscar Nájera
# License: BSD 3 clause
import numpy as np
import matplotlib.pyplot as plt
x = np.linspace(-1, 2, 100)
y = np.exp(x)
plt.figure()
plt.plot(x, y)
plt.xlabel('$x$')
plt.ylabel('$\exp(x)$')
plt.figure()
plt.plot(x, -np.exp(-x))
plt.xlabel('$x$')
plt.ylabel('$-\exp(-x)$')
plt.show()
| Update example for image stacking CSS instuction | Update example for image stacking CSS instuction
| Python | bsd-3-clause | lesteve/sphinx-gallery,Eric89GXL/sphinx-gallery,sphinx-gallery/sphinx-gallery,Titan-C/sphinx-gallery,lesteve/sphinx-gallery,Titan-C/sphinx-gallery,Eric89GXL/sphinx-gallery,sphinx-gallery/sphinx-gallery |
# Code source: Óscar Nájera
# License: BSD 3 clause
import numpy as np
import matplotlib.pyplot as plt
x = np.linspace(-1, 2, 100)
y = np.exp(x)
+ plt.figure()
plt.plot(x, y)
plt.xlabel('$x$')
- plt.ylabel('$exp(x)$')
+ plt.ylabel('$\exp(x)$')
+
+ plt.figure()
+ plt.plot(x, -np.exp(-x))
+ plt.xlabel('$x$')
+ plt.ylabel('$-\exp(-x)$')
+
+
plt.show()
| Update example for image stacking CSS instuction | ## Code Before:
# Code source: Óscar Nájera
# License: BSD 3 clause
import numpy as np
import matplotlib.pyplot as plt
x = np.linspace(-1, 2, 100)
y = np.exp(x)
plt.plot(x, y)
plt.xlabel('$x$')
plt.ylabel('$exp(x)$')
plt.show()
## Instruction:
Update example for image stacking CSS instuction
## Code After:
# Code source: Óscar Nájera
# License: BSD 3 clause
import numpy as np
import matplotlib.pyplot as plt
x = np.linspace(-1, 2, 100)
y = np.exp(x)
plt.figure()
plt.plot(x, y)
plt.xlabel('$x$')
plt.ylabel('$\exp(x)$')
plt.figure()
plt.plot(x, -np.exp(-x))
plt.xlabel('$x$')
plt.ylabel('$-\exp(-x)$')
plt.show()
|
# Code source: Óscar Nájera
# License: BSD 3 clause
import numpy as np
import matplotlib.pyplot as plt
x = np.linspace(-1, 2, 100)
y = np.exp(x)
+ plt.figure()
plt.plot(x, y)
plt.xlabel('$x$')
- plt.ylabel('$exp(x)$')
+ plt.ylabel('$\exp(x)$')
? +
+
+ plt.figure()
+ plt.plot(x, -np.exp(-x))
+ plt.xlabel('$x$')
+ plt.ylabel('$-\exp(-x)$')
+
+
plt.show() |
6b49f7b1948ab94631c79304c91f8d5590d03e40 | addons/project/models/project_config_settings.py | addons/project/models/project_config_settings.py |
from odoo import api, fields, models
class ProjectConfiguration(models.TransientModel):
_name = 'project.config.settings'
_inherit = 'res.config.settings'
company_id = fields.Many2one('res.company', string='Company', required=True,
default=lambda self: self.env.user.company_id)
module_pad = fields.Boolean("Collaborative Pads")
module_hr_timesheet = fields.Boolean("Timesheets")
module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet")
module_rating_project = fields.Boolean(string="Rating on Tasks")
module_project_forecast = fields.Boolean(string="Forecasts")
module_hr_holidays = fields.Boolean("Leave Management")
module_hr_timesheet_attendance = fields.Boolean("Attendances")
module_sale_timesheet = fields.Boolean("Time Billing")
module_hr_expense = fields.Boolean("Expenses")
group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project")
|
from odoo import api, fields, models
class ProjectConfiguration(models.TransientModel):
_name = 'project.config.settings'
_inherit = 'res.config.settings'
company_id = fields.Many2one('res.company', string='Company', required=True,
default=lambda self: self.env.user.company_id)
module_pad = fields.Boolean("Collaborative Pads")
module_hr_timesheet = fields.Boolean("Timesheets")
module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet")
module_rating_project = fields.Boolean(string="Rating on Tasks")
module_project_forecast = fields.Boolean(string="Forecasts")
module_hr_holidays = fields.Boolean("Leave Management")
module_hr_timesheet_attendance = fields.Boolean("Attendances")
module_sale_timesheet = fields.Boolean("Time Billing")
module_hr_expense = fields.Boolean("Expenses")
group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project")
@api.onchange('module_sale_timesheet')
def _onchange_module_sale_timesheet(self):
if self.module_sale_timesheet:
self.module_hr_timesheet = True
| Enable `Timesheets` option if `Time Billing` is enabled | [IMP] project: Enable `Timesheets` option if `Time Billing` is enabled
| Python | agpl-3.0 | ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo |
from odoo import api, fields, models
class ProjectConfiguration(models.TransientModel):
_name = 'project.config.settings'
_inherit = 'res.config.settings'
company_id = fields.Many2one('res.company', string='Company', required=True,
default=lambda self: self.env.user.company_id)
module_pad = fields.Boolean("Collaborative Pads")
module_hr_timesheet = fields.Boolean("Timesheets")
module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet")
module_rating_project = fields.Boolean(string="Rating on Tasks")
module_project_forecast = fields.Boolean(string="Forecasts")
module_hr_holidays = fields.Boolean("Leave Management")
module_hr_timesheet_attendance = fields.Boolean("Attendances")
module_sale_timesheet = fields.Boolean("Time Billing")
module_hr_expense = fields.Boolean("Expenses")
group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project")
+ @api.onchange('module_sale_timesheet')
+ def _onchange_module_sale_timesheet(self):
+ if self.module_sale_timesheet:
+ self.module_hr_timesheet = True
+ | Enable `Timesheets` option if `Time Billing` is enabled | ## Code Before:
from odoo import api, fields, models
class ProjectConfiguration(models.TransientModel):
_name = 'project.config.settings'
_inherit = 'res.config.settings'
company_id = fields.Many2one('res.company', string='Company', required=True,
default=lambda self: self.env.user.company_id)
module_pad = fields.Boolean("Collaborative Pads")
module_hr_timesheet = fields.Boolean("Timesheets")
module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet")
module_rating_project = fields.Boolean(string="Rating on Tasks")
module_project_forecast = fields.Boolean(string="Forecasts")
module_hr_holidays = fields.Boolean("Leave Management")
module_hr_timesheet_attendance = fields.Boolean("Attendances")
module_sale_timesheet = fields.Boolean("Time Billing")
module_hr_expense = fields.Boolean("Expenses")
group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project")
## Instruction:
Enable `Timesheets` option if `Time Billing` is enabled
## Code After:
from odoo import api, fields, models
class ProjectConfiguration(models.TransientModel):
_name = 'project.config.settings'
_inherit = 'res.config.settings'
company_id = fields.Many2one('res.company', string='Company', required=True,
default=lambda self: self.env.user.company_id)
module_pad = fields.Boolean("Collaborative Pads")
module_hr_timesheet = fields.Boolean("Timesheets")
module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet")
module_rating_project = fields.Boolean(string="Rating on Tasks")
module_project_forecast = fields.Boolean(string="Forecasts")
module_hr_holidays = fields.Boolean("Leave Management")
module_hr_timesheet_attendance = fields.Boolean("Attendances")
module_sale_timesheet = fields.Boolean("Time Billing")
module_hr_expense = fields.Boolean("Expenses")
group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project")
@api.onchange('module_sale_timesheet')
def _onchange_module_sale_timesheet(self):
if self.module_sale_timesheet:
self.module_hr_timesheet = True
|
from odoo import api, fields, models
class ProjectConfiguration(models.TransientModel):
_name = 'project.config.settings'
_inherit = 'res.config.settings'
company_id = fields.Many2one('res.company', string='Company', required=True,
default=lambda self: self.env.user.company_id)
module_pad = fields.Boolean("Collaborative Pads")
module_hr_timesheet = fields.Boolean("Timesheets")
module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet")
module_rating_project = fields.Boolean(string="Rating on Tasks")
module_project_forecast = fields.Boolean(string="Forecasts")
module_hr_holidays = fields.Boolean("Leave Management")
module_hr_timesheet_attendance = fields.Boolean("Attendances")
module_sale_timesheet = fields.Boolean("Time Billing")
module_hr_expense = fields.Boolean("Expenses")
group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project")
+
+ @api.onchange('module_sale_timesheet')
+ def _onchange_module_sale_timesheet(self):
+ if self.module_sale_timesheet:
+ self.module_hr_timesheet = True |
9e22082a280babb1e0880fe24fa17c45aac09515 | docker-nodev.py | docker-nodev.py |
from __future__ import print_function
import subprocess
import sys
DOCKER_CREATE_IN = 'docker create -it nodev {}'
DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}'
def nodev(argv=()):
container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip()
print('creating container: {container_id}'.format(**locals()))
try:
subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True)
subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True)
finally:
print('removing container: {container_id}'.format(**locals()))
subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True)
if __name__ == '__main__':
nodev(sys.argv)
|
from __future__ import print_function
import subprocess
import sys
DOCKER_CREATE_IN = 'docker create -it nodev {}'
DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}'
def nodev(argv=()):
container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip()
print('creating container: {container_id}'.format(**locals()))
try:
subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True)
subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True)
finally:
print('removing container: {container_id}'.format(**locals()))
subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True)
if __name__ == '__main__':
try:
nodev(sys.argv)
except subprocess.CalledProcessError as ex:
print(ex.args)
sys.exit(1)
| Fix python3 crash and cleaner error reporting. | Fix python3 crash and cleaner error reporting.
| Python | mit | nodev-io/nodev-starter-kit,nodev-io/nodev-tutorial,nodev-io/nodev-starter-kit |
from __future__ import print_function
import subprocess
import sys
DOCKER_CREATE_IN = 'docker create -it nodev {}'
DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}'
def nodev(argv=()):
- container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip()
+ container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip()
print('creating container: {container_id}'.format(**locals()))
try:
subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True)
subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True)
finally:
print('removing container: {container_id}'.format(**locals()))
subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True)
if __name__ == '__main__':
+ try:
- nodev(sys.argv)
+ nodev(sys.argv)
+ except subprocess.CalledProcessError as ex:
+ print(ex.args)
+ sys.exit(1)
| Fix python3 crash and cleaner error reporting. | ## Code Before:
from __future__ import print_function
import subprocess
import sys
DOCKER_CREATE_IN = 'docker create -it nodev {}'
DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}'
def nodev(argv=()):
container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip()
print('creating container: {container_id}'.format(**locals()))
try:
subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True)
subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True)
finally:
print('removing container: {container_id}'.format(**locals()))
subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True)
if __name__ == '__main__':
nodev(sys.argv)
## Instruction:
Fix python3 crash and cleaner error reporting.
## Code After:
from __future__ import print_function
import subprocess
import sys
DOCKER_CREATE_IN = 'docker create -it nodev {}'
DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}'
def nodev(argv=()):
container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip()
print('creating container: {container_id}'.format(**locals()))
try:
subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True)
subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True)
finally:
print('removing container: {container_id}'.format(**locals()))
subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True)
if __name__ == '__main__':
try:
nodev(sys.argv)
except subprocess.CalledProcessError as ex:
print(ex.args)
sys.exit(1)
|
from __future__ import print_function
import subprocess
import sys
DOCKER_CREATE_IN = 'docker create -it nodev {}'
DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}'
def nodev(argv=()):
- container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip()
+ container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip()
? ++++++++++++++++
print('creating container: {container_id}'.format(**locals()))
try:
subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True)
subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True)
finally:
print('removing container: {container_id}'.format(**locals()))
subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True)
if __name__ == '__main__':
+ try:
- nodev(sys.argv)
+ nodev(sys.argv)
? ++++
+ except subprocess.CalledProcessError as ex:
+ print(ex.args)
+ sys.exit(1) |
db4355ce0345df9dd23b937370f5f0d4cb2164e9 | zc_common/remote_resource/filters.py | zc_common/remote_resource/filters.py | import re
from django.db.models.fields.related import ManyToManyField
from rest_framework import filters
class JSONAPIFilterBackend(filters.DjangoFilterBackend):
def filter_queryset(self, request, queryset, view):
filter_class = self.get_filter_class(view, queryset)
primary_key = queryset.model._meta.pk.name
query_params = {}
for param, value in request.query_params.iteritems():
match = re.search(r'^filter\[(\w+)\]$', param)
if match:
field_name = match.group(1)
try:
name, extra = field_name.split('__')
except ValueError:
name = field_name
extra = None
if name not in view.filter_fields.keys():
return queryset.none()
if len(field_name) > 1 and field_name[:2] == 'id':
query_params['{0}__{1}'.format(primary_key, extra)] = value
if hasattr(queryset.model, field_name)\
and isinstance(getattr(queryset.model, field_name).field, ManyToManyField):
value = value.split(',')
query_params[field_name] = value
if filter_class:
return filter_class(query_params, queryset=queryset).qs
return queryset
| import re
from distutils.util import strtobool
from django.db.models import BooleanField, FieldDoesNotExist
from django.db.models.fields.related import ManyToManyField
from rest_framework import filters
class JSONAPIFilterBackend(filters.DjangoFilterBackend):
def filter_queryset(self, request, queryset, view):
filter_class = self.get_filter_class(view, queryset)
primary_key = queryset.model._meta.pk.name
query_params = {}
for param, value in request.query_params.iteritems():
match = re.search(r'^filter\[(\w+)\]$', param)
if match:
field_name = match.group(1)
try:
name, extra = field_name.split('__')
except ValueError:
name = field_name
extra = None
if name not in view.filter_fields.keys():
return queryset.none()
if len(field_name) > 1 and field_name[:2] == 'id':
query_params['{0}__{1}'.format(primary_key, extra)] = value
if hasattr(queryset.model, field_name)\
and isinstance(getattr(queryset.model, field_name).field, ManyToManyField):
value = value.split(',')
# Allow 'true' or 'false' as values for boolean fields
try:
if isinstance(queryset.model._meta.get_field(field_name), BooleanField):
value = bool(strtobool(value))
except FieldDoesNotExist:
pass
query_params[field_name] = value
if filter_class:
return filter_class(query_params, queryset=queryset).qs
return queryset
| Use 'true' while filtering a boolean as opposed to 'True' | Use 'true' while filtering a boolean as opposed to 'True'
| Python | mit | ZeroCater/zc_common,ZeroCater/zc_common | import re
+ from distutils.util import strtobool
+ from django.db.models import BooleanField, FieldDoesNotExist
from django.db.models.fields.related import ManyToManyField
from rest_framework import filters
class JSONAPIFilterBackend(filters.DjangoFilterBackend):
def filter_queryset(self, request, queryset, view):
filter_class = self.get_filter_class(view, queryset)
primary_key = queryset.model._meta.pk.name
query_params = {}
for param, value in request.query_params.iteritems():
match = re.search(r'^filter\[(\w+)\]$', param)
if match:
field_name = match.group(1)
try:
name, extra = field_name.split('__')
except ValueError:
name = field_name
extra = None
if name not in view.filter_fields.keys():
return queryset.none()
if len(field_name) > 1 and field_name[:2] == 'id':
query_params['{0}__{1}'.format(primary_key, extra)] = value
if hasattr(queryset.model, field_name)\
and isinstance(getattr(queryset.model, field_name).field, ManyToManyField):
value = value.split(',')
+
+ # Allow 'true' or 'false' as values for boolean fields
+ try:
+ if isinstance(queryset.model._meta.get_field(field_name), BooleanField):
+ value = bool(strtobool(value))
+ except FieldDoesNotExist:
+ pass
+
query_params[field_name] = value
if filter_class:
return filter_class(query_params, queryset=queryset).qs
return queryset
| Use 'true' while filtering a boolean as opposed to 'True' | ## Code Before:
import re
from django.db.models.fields.related import ManyToManyField
from rest_framework import filters
class JSONAPIFilterBackend(filters.DjangoFilterBackend):
def filter_queryset(self, request, queryset, view):
filter_class = self.get_filter_class(view, queryset)
primary_key = queryset.model._meta.pk.name
query_params = {}
for param, value in request.query_params.iteritems():
match = re.search(r'^filter\[(\w+)\]$', param)
if match:
field_name = match.group(1)
try:
name, extra = field_name.split('__')
except ValueError:
name = field_name
extra = None
if name not in view.filter_fields.keys():
return queryset.none()
if len(field_name) > 1 and field_name[:2] == 'id':
query_params['{0}__{1}'.format(primary_key, extra)] = value
if hasattr(queryset.model, field_name)\
and isinstance(getattr(queryset.model, field_name).field, ManyToManyField):
value = value.split(',')
query_params[field_name] = value
if filter_class:
return filter_class(query_params, queryset=queryset).qs
return queryset
## Instruction:
Use 'true' while filtering a boolean as opposed to 'True'
## Code After:
import re
from distutils.util import strtobool
from django.db.models import BooleanField, FieldDoesNotExist
from django.db.models.fields.related import ManyToManyField
from rest_framework import filters
class JSONAPIFilterBackend(filters.DjangoFilterBackend):
def filter_queryset(self, request, queryset, view):
filter_class = self.get_filter_class(view, queryset)
primary_key = queryset.model._meta.pk.name
query_params = {}
for param, value in request.query_params.iteritems():
match = re.search(r'^filter\[(\w+)\]$', param)
if match:
field_name = match.group(1)
try:
name, extra = field_name.split('__')
except ValueError:
name = field_name
extra = None
if name not in view.filter_fields.keys():
return queryset.none()
if len(field_name) > 1 and field_name[:2] == 'id':
query_params['{0}__{1}'.format(primary_key, extra)] = value
if hasattr(queryset.model, field_name)\
and isinstance(getattr(queryset.model, field_name).field, ManyToManyField):
value = value.split(',')
# Allow 'true' or 'false' as values for boolean fields
try:
if isinstance(queryset.model._meta.get_field(field_name), BooleanField):
value = bool(strtobool(value))
except FieldDoesNotExist:
pass
query_params[field_name] = value
if filter_class:
return filter_class(query_params, queryset=queryset).qs
return queryset
| import re
+ from distutils.util import strtobool
+ from django.db.models import BooleanField, FieldDoesNotExist
from django.db.models.fields.related import ManyToManyField
from rest_framework import filters
class JSONAPIFilterBackend(filters.DjangoFilterBackend):
def filter_queryset(self, request, queryset, view):
filter_class = self.get_filter_class(view, queryset)
primary_key = queryset.model._meta.pk.name
query_params = {}
for param, value in request.query_params.iteritems():
match = re.search(r'^filter\[(\w+)\]$', param)
if match:
field_name = match.group(1)
try:
name, extra = field_name.split('__')
except ValueError:
name = field_name
extra = None
if name not in view.filter_fields.keys():
return queryset.none()
if len(field_name) > 1 and field_name[:2] == 'id':
query_params['{0}__{1}'.format(primary_key, extra)] = value
if hasattr(queryset.model, field_name)\
and isinstance(getattr(queryset.model, field_name).field, ManyToManyField):
value = value.split(',')
+
+ # Allow 'true' or 'false' as values for boolean fields
+ try:
+ if isinstance(queryset.model._meta.get_field(field_name), BooleanField):
+ value = bool(strtobool(value))
+ except FieldDoesNotExist:
+ pass
+
query_params[field_name] = value
if filter_class:
return filter_class(query_params, queryset=queryset).qs
return queryset |
0e99654d606038098d45fb83cc40405742e43ae8 | readthedocs/builds/filters.py | readthedocs/builds/filters.py | from django.utils.translation import ugettext_lazy as _
import django_filters
from builds import constants
from builds.models import Build, Version
ANY_REPO = (
('', _('Any')),
)
BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES
class VersionFilter(django_filters.FilterSet):
project = django_filters.CharFilter(name='project__name', lookup_type="icontains")
slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains')
class Meta:
model = Version
fields = ['project', 'slug']
class BuildFilter(django_filters.FilterSet):
date = django_filters.DateRangeFilter(label=_("Build Date"), name="date")
type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES)
class Meta:
model = Build
fields = ['type', 'date', 'version', 'success']
| from django.utils.translation import ugettext_lazy as _
import django_filters
from builds import constants
from builds.models import Build, Version
ANY_REPO = (
('', _('Any')),
)
BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES
class VersionFilter(django_filters.FilterSet):
project = django_filters.CharFilter(name='project__name', lookup_type="icontains")
slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains')
class Meta:
model = Version
fields = ['project', 'slug']
class BuildFilter(django_filters.FilterSet):
date = django_filters.DateRangeFilter(label=_("Build Date"), name="date")
type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES)
class Meta:
model = Build
fields = ['type', 'date', 'success']
| Remove version from Build filter. | Remove version from Build filter.
| Python | mit | agjohnson/readthedocs.org,fujita-shintaro/readthedocs.org,GovReady/readthedocs.org,nyergler/pythonslides,Tazer/readthedocs.org,techtonik/readthedocs.org,takluyver/readthedocs.org,nyergler/pythonslides,GovReady/readthedocs.org,nikolas/readthedocs.org,gjtorikian/readthedocs.org,cgourlay/readthedocs.org,d0ugal/readthedocs.org,sid-kap/readthedocs.org,CedarLogic/readthedocs.org,gjtorikian/readthedocs.org,sils1297/readthedocs.org,singingwolfboy/readthedocs.org,kdkeyser/readthedocs.org,kenshinthebattosai/readthedocs.org,VishvajitP/readthedocs.org,hach-que/readthedocs.org,laplaceliu/readthedocs.org,KamranMackey/readthedocs.org,techtonik/readthedocs.org,kdkeyser/readthedocs.org,ojii/readthedocs.org,asampat3090/readthedocs.org,michaelmcandrew/readthedocs.org,dirn/readthedocs.org,attakei/readthedocs-oauth,LukasBoersma/readthedocs.org,rtfd/readthedocs.org,techtonik/readthedocs.org,tddv/readthedocs.org,sunnyzwh/readthedocs.org,emawind84/readthedocs.org,davidfischer/readthedocs.org,jerel/readthedocs.org,clarkperkins/readthedocs.org,mhils/readthedocs.org,takluyver/readthedocs.org,sils1297/readthedocs.org,Carreau/readthedocs.org,wijerasa/readthedocs.org,SteveViss/readthedocs.org,singingwolfboy/readthedocs.org,Tazer/readthedocs.org,pombredanne/readthedocs.org,rtfd/readthedocs.org,nyergler/pythonslides,nikolas/readthedocs.org,michaelmcandrew/readthedocs.org,CedarLogic/readthedocs.org,safwanrahman/readthedocs.org,takluyver/readthedocs.org,royalwang/readthedocs.org,espdev/readthedocs.org,hach-que/readthedocs.org,Carreau/readthedocs.org,takluyver/readthedocs.org,dirn/readthedocs.org,d0ugal/readthedocs.org,sid-kap/readthedocs.org,fujita-shintaro/readthedocs.org,kdkeyser/readthedocs.org,tddv/readthedocs.org,espdev/readthedocs.org,asampat3090/readthedocs.org,LukasBoersma/readthedocs.org,kenshinthebattosai/readthedocs.org,nyergler/pythonslides,titiushko/readthedocs.org,techtonik/readthedocs.org,wanghaven/readthedocs.org,stevepiercy/readthedocs.org,rtfd/readthedocs.org,sunnyzwh/readthedocs.org,mhils/readthedocs.org,d0ugal/readthedocs.org,titiushko/readthedocs.org,istresearch/readthedocs.org,SteveViss/readthedocs.org,mrshoki/readthedocs.org,raven47git/readthedocs.org,atsuyim/readthedocs.org,wanghaven/readthedocs.org,soulshake/readthedocs.org,agjohnson/readthedocs.org,Carreau/readthedocs.org,tddv/readthedocs.org,KamranMackey/readthedocs.org,laplaceliu/readthedocs.org,nikolas/readthedocs.org,kenwang76/readthedocs.org,clarkperkins/readthedocs.org,VishvajitP/readthedocs.org,jerel/readthedocs.org,kenshinthebattosai/readthedocs.org,hach-que/readthedocs.org,asampat3090/readthedocs.org,stevepiercy/readthedocs.org,VishvajitP/readthedocs.org,ojii/readthedocs.org,agjohnson/readthedocs.org,agjohnson/readthedocs.org,jerel/readthedocs.org,attakei/readthedocs-oauth,michaelmcandrew/readthedocs.org,istresearch/readthedocs.org,safwanrahman/readthedocs.org,asampat3090/readthedocs.org,wanghaven/readthedocs.org,nikolas/readthedocs.org,emawind84/readthedocs.org,rtfd/readthedocs.org,wijerasa/readthedocs.org,jerel/readthedocs.org,cgourlay/readthedocs.org,sils1297/readthedocs.org,ojii/readthedocs.org,stevepiercy/readthedocs.org,LukasBoersma/readthedocs.org,raven47git/readthedocs.org,clarkperkins/readthedocs.org,wijerasa/readthedocs.org,CedarLogic/readthedocs.org,GovReady/readthedocs.org,singingwolfboy/readthedocs.org,kenwang76/readthedocs.org,mrshoki/readthedocs.org,istresearch/readthedocs.org,dirn/readthedocs.org,mrshoki/readthedocs.org,kenwang76/readthedocs.org,mhils/readthedocs.org,Tazer/readthedocs.org,sils1297/readthedocs.org,cgourlay/readthedocs.org,royalwang/readthedocs.org,sid-kap/readthedocs.org,emawind84/readthedocs.org,mrshoki/readthedocs.org,pombredanne/readthedocs.org,SteveViss/readthedocs.org,attakei/readthedocs-oauth,ojii/readthedocs.org,davidfischer/readthedocs.org,istresearch/readthedocs.org,hach-que/readthedocs.org,raven47git/readthedocs.org,fujita-shintaro/readthedocs.org,titiushko/readthedocs.org,LukasBoersma/readthedocs.org,singingwolfboy/readthedocs.org,royalwang/readthedocs.org,atsuyim/readthedocs.org,safwanrahman/readthedocs.org,VishvajitP/readthedocs.org,wanghaven/readthedocs.org,davidfischer/readthedocs.org,GovReady/readthedocs.org,clarkperkins/readthedocs.org,royalwang/readthedocs.org,atsuyim/readthedocs.org,dirn/readthedocs.org,sunnyzwh/readthedocs.org,stevepiercy/readthedocs.org,kdkeyser/readthedocs.org,gjtorikian/readthedocs.org,fujita-shintaro/readthedocs.org,sunnyzwh/readthedocs.org,wijerasa/readthedocs.org,attakei/readthedocs-oauth,michaelmcandrew/readthedocs.org,davidfischer/readthedocs.org,kenshinthebattosai/readthedocs.org,Tazer/readthedocs.org,atsuyim/readthedocs.org,KamranMackey/readthedocs.org,laplaceliu/readthedocs.org,pombredanne/readthedocs.org,CedarLogic/readthedocs.org,soulshake/readthedocs.org,cgourlay/readthedocs.org,d0ugal/readthedocs.org,emawind84/readthedocs.org,soulshake/readthedocs.org,espdev/readthedocs.org,kenwang76/readthedocs.org,espdev/readthedocs.org,raven47git/readthedocs.org,SteveViss/readthedocs.org,mhils/readthedocs.org,Carreau/readthedocs.org,laplaceliu/readthedocs.org,KamranMackey/readthedocs.org,sid-kap/readthedocs.org,soulshake/readthedocs.org,safwanrahman/readthedocs.org,espdev/readthedocs.org,gjtorikian/readthedocs.org,titiushko/readthedocs.org | from django.utils.translation import ugettext_lazy as _
import django_filters
from builds import constants
from builds.models import Build, Version
ANY_REPO = (
('', _('Any')),
)
BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES
class VersionFilter(django_filters.FilterSet):
project = django_filters.CharFilter(name='project__name', lookup_type="icontains")
slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains')
class Meta:
model = Version
fields = ['project', 'slug']
class BuildFilter(django_filters.FilterSet):
date = django_filters.DateRangeFilter(label=_("Build Date"), name="date")
type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES)
class Meta:
model = Build
- fields = ['type', 'date', 'version', 'success']
+ fields = ['type', 'date', 'success']
- | Remove version from Build filter. | ## Code Before:
from django.utils.translation import ugettext_lazy as _
import django_filters
from builds import constants
from builds.models import Build, Version
ANY_REPO = (
('', _('Any')),
)
BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES
class VersionFilter(django_filters.FilterSet):
project = django_filters.CharFilter(name='project__name', lookup_type="icontains")
slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains')
class Meta:
model = Version
fields = ['project', 'slug']
class BuildFilter(django_filters.FilterSet):
date = django_filters.DateRangeFilter(label=_("Build Date"), name="date")
type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES)
class Meta:
model = Build
fields = ['type', 'date', 'version', 'success']
## Instruction:
Remove version from Build filter.
## Code After:
from django.utils.translation import ugettext_lazy as _
import django_filters
from builds import constants
from builds.models import Build, Version
ANY_REPO = (
('', _('Any')),
)
BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES
class VersionFilter(django_filters.FilterSet):
project = django_filters.CharFilter(name='project__name', lookup_type="icontains")
slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains')
class Meta:
model = Version
fields = ['project', 'slug']
class BuildFilter(django_filters.FilterSet):
date = django_filters.DateRangeFilter(label=_("Build Date"), name="date")
type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES)
class Meta:
model = Build
fields = ['type', 'date', 'success']
| from django.utils.translation import ugettext_lazy as _
import django_filters
from builds import constants
from builds.models import Build, Version
ANY_REPO = (
('', _('Any')),
)
BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES
class VersionFilter(django_filters.FilterSet):
project = django_filters.CharFilter(name='project__name', lookup_type="icontains")
slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains')
class Meta:
model = Version
fields = ['project', 'slug']
class BuildFilter(django_filters.FilterSet):
date = django_filters.DateRangeFilter(label=_("Build Date"), name="date")
type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES)
class Meta:
model = Build
- fields = ['type', 'date', 'version', 'success']
? -----------
+ fields = ['type', 'date', 'success']
- |
7c894c716cb712bbcb137df3a5df5548bdca9d93 | wafer/sponsors/migrations/0005_sponsorshippackage_symbol.py | wafer/sponsors/migrations/0005_sponsorshippackage_symbol.py | from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('sponsors', '0004_auto_20160813_1328'),
]
operations = [
migrations.AddField(
model_name='sponsorshippackage',
name='symbol',
field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True),
),
]
| from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('sponsors', '0004_auto_20160813_1328'),
]
operations = [
migrations.AddField(
model_name='sponsorshippackage',
name='symbol',
field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1),
),
]
| Update the migration to changed text | Update the migration to changed text
| Python | isc | CTPUG/wafer,CTPUG/wafer,CTPUG/wafer,CTPUG/wafer | from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('sponsors', '0004_auto_20160813_1328'),
]
operations = [
migrations.AddField(
model_name='sponsorshippackage',
name='symbol',
- field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True),
+ field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1),
),
]
| Update the migration to changed text | ## Code Before:
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('sponsors', '0004_auto_20160813_1328'),
]
operations = [
migrations.AddField(
model_name='sponsorshippackage',
name='symbol',
field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True),
),
]
## Instruction:
Update the migration to changed text
## Code After:
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('sponsors', '0004_auto_20160813_1328'),
]
operations = [
migrations.AddField(
model_name='sponsorshippackage',
name='symbol',
field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1),
),
]
| from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('sponsors', '0004_auto_20160813_1328'),
]
operations = [
migrations.AddField(
model_name='sponsorshippackage',
name='symbol',
- field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True),
+ field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1),
),
] |
54b2a6953a4da2b217052d166ad1f069f683b9ee | scripts/nomenclature/nomenclature_map.py | scripts/nomenclature/nomenclature_map.py | import pandas as pd
itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1")
| import pandas as pd
from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder
# Globals
INPUT_FILE_NAME = "search_results.csv"
# Input
nomenclature_folder = locate_nomenclature_folder()
itis_results = pd.read_csv(
str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1")
| Implement locator in nomenclature map | Implement locator in nomenclature map
| Python | mit | jnfrye/local_plants_book | import pandas as pd
- itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1")
+ from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder
+
+ # Globals
+ INPUT_FILE_NAME = "search_results.csv"
+
+ # Input
+ nomenclature_folder = locate_nomenclature_folder()
+ itis_results = pd.read_csv(
+ str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1")
+ | Implement locator in nomenclature map | ## Code Before:
import pandas as pd
itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1")
## Instruction:
Implement locator in nomenclature map
## Code After:
import pandas as pd
from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder
# Globals
INPUT_FILE_NAME = "search_results.csv"
# Input
nomenclature_folder = locate_nomenclature_folder()
itis_results = pd.read_csv(
str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1")
| import pandas as pd
- itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1")
+ from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder
+
+
+ # Globals
+ INPUT_FILE_NAME = "search_results.csv"
+
+ # Input
+ nomenclature_folder = locate_nomenclature_folder()
+ itis_results = pd.read_csv(
+ str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1") |
6353a3d1443c717b2d2e804190153f8be605c2f1 | setup.py | setup.py | from distutils.core import setup
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='udiskie',
version='0.4.2',
description='Removable disk automounter for udisks',
long_description=long_description,
author='Byron Clark',
author_email='byron@theclarkfamily.name',
maintainer='Thomas Gläßle',
maintainer_email='t_glaessle@gmx.de',
url='https://github.com/coldfix/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| from distutils.core import setup
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='udiskie',
version='0.4.2',
description='Removable disk automounter for udisks',
long_description=long_description,
author='Byron Clark',
author_email='byron@theclarkfamily.name',
maintainer='Thomas Gläßle',
maintainer_email='t_glaessle@gmx.de',
url='https://github.com/coldfix/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
'bin/udiskie-mount'
],
)
| Include udiskie-mount in binary distribution | Include udiskie-mount in binary distribution
| Python | mit | khardix/udiskie,pstray/udiskie,coldfix/udiskie,coldfix/udiskie,mathstuf/udiskie,pstray/udiskie | from distutils.core import setup
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='udiskie',
version='0.4.2',
description='Removable disk automounter for udisks',
long_description=long_description,
author='Byron Clark',
author_email='byron@theclarkfamily.name',
maintainer='Thomas Gläßle',
maintainer_email='t_glaessle@gmx.de',
url='https://github.com/coldfix/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
+ 'bin/udiskie-mount'
],
)
| Include udiskie-mount in binary distribution | ## Code Before:
from distutils.core import setup
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='udiskie',
version='0.4.2',
description='Removable disk automounter for udisks',
long_description=long_description,
author='Byron Clark',
author_email='byron@theclarkfamily.name',
maintainer='Thomas Gläßle',
maintainer_email='t_glaessle@gmx.de',
url='https://github.com/coldfix/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
## Instruction:
Include udiskie-mount in binary distribution
## Code After:
from distutils.core import setup
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='udiskie',
version='0.4.2',
description='Removable disk automounter for udisks',
long_description=long_description,
author='Byron Clark',
author_email='byron@theclarkfamily.name',
maintainer='Thomas Gläßle',
maintainer_email='t_glaessle@gmx.de',
url='https://github.com/coldfix/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
'bin/udiskie-mount'
],
)
| from distutils.core import setup
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='udiskie',
version='0.4.2',
description='Removable disk automounter for udisks',
long_description=long_description,
author='Byron Clark',
author_email='byron@theclarkfamily.name',
maintainer='Thomas Gläßle',
maintainer_email='t_glaessle@gmx.de',
url='https://github.com/coldfix/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
+ 'bin/udiskie-mount'
],
) |
1cf1da043ceab767d9d0dbdbed62c2f1c5ff36e9 | test_http.py | test_http.py | from http_server import HttpServer
import socket
def test_200_ok():
s = HttpServer()
assert s.ok() == "HTTP/1.1 200 OK"
def test_200_ok_byte():
s = HttpServer()
assert isinstance(s.ok(), bytes)
def test_socket_is_socket():
s = HttpServer()
s.open_socket()
assert isinstance(s._socket, socket.socket)
def test_open_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
assert s._socket.getsockname() == ('127.0.0.1', 50000) | from http_server import HttpServer
import socket
def test_200_ok():
s = HttpServer()
assert s.ok() == "HTTP/1.1 200 OK"
def test_200_ok_byte():
s = HttpServer()
assert isinstance(s.ok(), bytes)
def test_socket_is_socket():
s = HttpServer()
s.open_socket()
assert isinstance(s._socket, socket.socket)
def test_open_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
assert s._socket.getsockname() == ('127.0.0.1', 50000)
def test_close_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
s.close_socket()
assert s._socket is None | Add tests for closing a socket | Add tests for closing a socket
| Python | mit | jefrailey/network_tools | from http_server import HttpServer
import socket
def test_200_ok():
s = HttpServer()
assert s.ok() == "HTTP/1.1 200 OK"
def test_200_ok_byte():
s = HttpServer()
assert isinstance(s.ok(), bytes)
def test_socket_is_socket():
s = HttpServer()
s.open_socket()
assert isinstance(s._socket, socket.socket)
def test_open_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
assert s._socket.getsockname() == ('127.0.0.1', 50000)
+
+
+ def test_close_socket():
+ s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
+ s.open_socket()
+ s.close_socket()
+ assert s._socket is None | Add tests for closing a socket | ## Code Before:
from http_server import HttpServer
import socket
def test_200_ok():
s = HttpServer()
assert s.ok() == "HTTP/1.1 200 OK"
def test_200_ok_byte():
s = HttpServer()
assert isinstance(s.ok(), bytes)
def test_socket_is_socket():
s = HttpServer()
s.open_socket()
assert isinstance(s._socket, socket.socket)
def test_open_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
assert s._socket.getsockname() == ('127.0.0.1', 50000)
## Instruction:
Add tests for closing a socket
## Code After:
from http_server import HttpServer
import socket
def test_200_ok():
s = HttpServer()
assert s.ok() == "HTTP/1.1 200 OK"
def test_200_ok_byte():
s = HttpServer()
assert isinstance(s.ok(), bytes)
def test_socket_is_socket():
s = HttpServer()
s.open_socket()
assert isinstance(s._socket, socket.socket)
def test_open_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
assert s._socket.getsockname() == ('127.0.0.1', 50000)
def test_close_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
s.close_socket()
assert s._socket is None | from http_server import HttpServer
import socket
def test_200_ok():
s = HttpServer()
assert s.ok() == "HTTP/1.1 200 OK"
def test_200_ok_byte():
s = HttpServer()
assert isinstance(s.ok(), bytes)
def test_socket_is_socket():
s = HttpServer()
s.open_socket()
assert isinstance(s._socket, socket.socket)
def test_open_socket():
s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
s.open_socket()
assert s._socket.getsockname() == ('127.0.0.1', 50000)
+
+
+ def test_close_socket():
+ s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5)
+ s.open_socket()
+ s.close_socket()
+ assert s._socket is None |
3f166b110d4e8623966ca29c71445973da4876f9 | armstrong/hatband/forms.py | armstrong/hatband/forms.py | from django import forms
from django.db import models
from . import widgets
RICH_TEXT_DBFIELD_OVERRIDES = {
models.TextField: {'widget': widgets.RichTextWidget},
}
class BackboneFormMixin(object):
class Media:
js = (
'hatband/js/jquery-1.6.2.min.js',
'hatband/js/underscore.js',
'hatband/js/backbone.js',
'hatband/js/backbone-inline-base.js')
class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm):
class Meta:
widgets = {
"content_type": forms.HiddenInput(),
"object_id": widgets.GenericKeyWidget(),
"order": forms.HiddenInput(),
}
| from django import forms
from django.conf import settings
from django.db import models
from . import widgets
RICH_TEXT_DBFIELD_OVERRIDES = {
models.TextField: {'widget': widgets.RichTextWidget},
}
class BackboneFormMixin(object):
if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True):
class Media:
js = (
'hatband/js/jquery-1.6.2.min.js',
'hatband/js/underscore.js',
'hatband/js/backbone.js',
'hatband/js/backbone-inline-base.js')
class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm):
class Meta:
widgets = {
"content_type": forms.HiddenInput(),
"object_id": widgets.GenericKeyWidget(),
"order": forms.HiddenInput(),
}
| Make it possible to turn off admin JS | Make it possible to turn off admin JS
| Python | apache-2.0 | armstrong/armstrong.hatband,armstrong/armstrong.hatband,texastribune/armstrong.hatband,armstrong/armstrong.hatband,texastribune/armstrong.hatband,texastribune/armstrong.hatband | from django import forms
+ from django.conf import settings
from django.db import models
from . import widgets
RICH_TEXT_DBFIELD_OVERRIDES = {
models.TextField: {'widget': widgets.RichTextWidget},
}
class BackboneFormMixin(object):
+ if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True):
- class Media:
+ class Media:
- js = (
+ js = (
- 'hatband/js/jquery-1.6.2.min.js',
+ 'hatband/js/jquery-1.6.2.min.js',
- 'hatband/js/underscore.js',
+ 'hatband/js/underscore.js',
- 'hatband/js/backbone.js',
+ 'hatband/js/backbone.js',
- 'hatband/js/backbone-inline-base.js')
+ 'hatband/js/backbone-inline-base.js')
class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm):
class Meta:
widgets = {
"content_type": forms.HiddenInput(),
"object_id": widgets.GenericKeyWidget(),
"order": forms.HiddenInput(),
}
| Make it possible to turn off admin JS | ## Code Before:
from django import forms
from django.db import models
from . import widgets
RICH_TEXT_DBFIELD_OVERRIDES = {
models.TextField: {'widget': widgets.RichTextWidget},
}
class BackboneFormMixin(object):
class Media:
js = (
'hatband/js/jquery-1.6.2.min.js',
'hatband/js/underscore.js',
'hatband/js/backbone.js',
'hatband/js/backbone-inline-base.js')
class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm):
class Meta:
widgets = {
"content_type": forms.HiddenInput(),
"object_id": widgets.GenericKeyWidget(),
"order": forms.HiddenInput(),
}
## Instruction:
Make it possible to turn off admin JS
## Code After:
from django import forms
from django.conf import settings
from django.db import models
from . import widgets
RICH_TEXT_DBFIELD_OVERRIDES = {
models.TextField: {'widget': widgets.RichTextWidget},
}
class BackboneFormMixin(object):
if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True):
class Media:
js = (
'hatband/js/jquery-1.6.2.min.js',
'hatband/js/underscore.js',
'hatband/js/backbone.js',
'hatband/js/backbone-inline-base.js')
class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm):
class Meta:
widgets = {
"content_type": forms.HiddenInput(),
"object_id": widgets.GenericKeyWidget(),
"order": forms.HiddenInput(),
}
| from django import forms
+ from django.conf import settings
from django.db import models
from . import widgets
RICH_TEXT_DBFIELD_OVERRIDES = {
models.TextField: {'widget': widgets.RichTextWidget},
}
class BackboneFormMixin(object):
+ if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True):
- class Media:
+ class Media:
? ++++
- js = (
+ js = (
? ++++
- 'hatband/js/jquery-1.6.2.min.js',
+ 'hatband/js/jquery-1.6.2.min.js',
? ++++
- 'hatband/js/underscore.js',
+ 'hatband/js/underscore.js',
? ++++
- 'hatband/js/backbone.js',
+ 'hatband/js/backbone.js',
? ++++
- 'hatband/js/backbone-inline-base.js')
+ 'hatband/js/backbone-inline-base.js')
? ++++
class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm):
class Meta:
widgets = {
"content_type": forms.HiddenInput(),
"object_id": widgets.GenericKeyWidget(),
"order": forms.HiddenInput(),
} |
6e6bffc19873260696822bb3f4a821ce4ea6f4a3 | consulrest/keyvalue.py | consulrest/keyvalue.py | import json
import re
import requests
class KeyValue(object):
def __init__(self, url):
self._url = "%s/kv" % url
def _get(self, key, recurse=None, keys=None):
url = self._url + '/' + key
if recurse is not None:
url += '?recurse'
if keys is not None:
url += '?keys'
r = requests.get(url)
if r.status_code == 200:
return json.loads(r.text)
else:
return None
def get(self, key, recurse=None):
return self._get(key, recurse=recurse)
def list(self, key=''):
return self._get(key, keys=True)
def set(self, key, value):
r = requests.put(self._url + '/' + key, data=value)
if r.status_code == 200 and re.match(r"true", r.text) is not None:
return True
else:
return False
def delete(self, key, recurse=None):
url = self._url + '/' + key
if recurse is not None:
url += '?recurse'
requests.delete(url)
| import json
import re
import requests
class KeyValue(object):
def __init__(self, url):
self._url = "%s/kv" % url
def _get(self, key, recurse=None, keys=None):
url = self._url + '/' + key
params = dict()
if recurse is not None:
params['recurse'] = True
if keys is not None:
params['keys'] = True
r = requests.get(url, params=params)
if r.status_code == 200:
return json.loads(r.text)
else:
return None
def get(self, key, recurse=None):
return self._get(key, recurse=recurse)
def list(self, key=''):
return self._get(key, keys=True)
def set(self, key, value):
r = requests.put(self._url + '/' + key, data=value)
if r.status_code == 200 and re.match(r"true", r.text) is not None:
return True
else:
return False
def delete(self, key, recurse=None):
url = self._url + '/' + key
params = dict()
if recurse is not None:
params['recurse'] = True
requests.delete(url, params=params)
| Use params dictionary instead of appending to the end of URL string | Use params dictionary instead of appending to the end of URL string
| Python | mit | vcoque/consul-ri | import json
import re
import requests
class KeyValue(object):
def __init__(self, url):
self._url = "%s/kv" % url
def _get(self, key, recurse=None, keys=None):
url = self._url + '/' + key
+ params = dict()
if recurse is not None:
- url += '?recurse'
+ params['recurse'] = True
+ if keys is not None:
+ params['keys'] = True
- if keys is not None:
- url += '?keys'
-
- r = requests.get(url)
+ r = requests.get(url, params=params)
if r.status_code == 200:
return json.loads(r.text)
else:
return None
def get(self, key, recurse=None):
return self._get(key, recurse=recurse)
def list(self, key=''):
return self._get(key, keys=True)
def set(self, key, value):
r = requests.put(self._url + '/' + key, data=value)
if r.status_code == 200 and re.match(r"true", r.text) is not None:
return True
else:
return False
def delete(self, key, recurse=None):
url = self._url + '/' + key
+
+ params = dict()
if recurse is not None:
- url += '?recurse'
+ params['recurse'] = True
- requests.delete(url)
+ requests.delete(url, params=params)
| Use params dictionary instead of appending to the end of URL string | ## Code Before:
import json
import re
import requests
class KeyValue(object):
def __init__(self, url):
self._url = "%s/kv" % url
def _get(self, key, recurse=None, keys=None):
url = self._url + '/' + key
if recurse is not None:
url += '?recurse'
if keys is not None:
url += '?keys'
r = requests.get(url)
if r.status_code == 200:
return json.loads(r.text)
else:
return None
def get(self, key, recurse=None):
return self._get(key, recurse=recurse)
def list(self, key=''):
return self._get(key, keys=True)
def set(self, key, value):
r = requests.put(self._url + '/' + key, data=value)
if r.status_code == 200 and re.match(r"true", r.text) is not None:
return True
else:
return False
def delete(self, key, recurse=None):
url = self._url + '/' + key
if recurse is not None:
url += '?recurse'
requests.delete(url)
## Instruction:
Use params dictionary instead of appending to the end of URL string
## Code After:
import json
import re
import requests
class KeyValue(object):
def __init__(self, url):
self._url = "%s/kv" % url
def _get(self, key, recurse=None, keys=None):
url = self._url + '/' + key
params = dict()
if recurse is not None:
params['recurse'] = True
if keys is not None:
params['keys'] = True
r = requests.get(url, params=params)
if r.status_code == 200:
return json.loads(r.text)
else:
return None
def get(self, key, recurse=None):
return self._get(key, recurse=recurse)
def list(self, key=''):
return self._get(key, keys=True)
def set(self, key, value):
r = requests.put(self._url + '/' + key, data=value)
if r.status_code == 200 and re.match(r"true", r.text) is not None:
return True
else:
return False
def delete(self, key, recurse=None):
url = self._url + '/' + key
params = dict()
if recurse is not None:
params['recurse'] = True
requests.delete(url, params=params)
| import json
import re
import requests
class KeyValue(object):
def __init__(self, url):
self._url = "%s/kv" % url
def _get(self, key, recurse=None, keys=None):
url = self._url + '/' + key
+ params = dict()
if recurse is not None:
- url += '?recurse'
+ params['recurse'] = True
+ if keys is not None:
+ params['keys'] = True
- if keys is not None:
- url += '?keys'
-
- r = requests.get(url)
+ r = requests.get(url, params=params)
? +++++++++++++++
if r.status_code == 200:
return json.loads(r.text)
else:
return None
def get(self, key, recurse=None):
return self._get(key, recurse=recurse)
def list(self, key=''):
return self._get(key, keys=True)
def set(self, key, value):
r = requests.put(self._url + '/' + key, data=value)
if r.status_code == 200 and re.match(r"true", r.text) is not None:
return True
else:
return False
def delete(self, key, recurse=None):
url = self._url + '/' + key
+
+ params = dict()
if recurse is not None:
- url += '?recurse'
+ params['recurse'] = True
- requests.delete(url)
+ requests.delete(url, params=params)
? +++++++++++++++
|
f3b9cc6392e4c271ae11417357ecdc196f1c3ae7 | python_scripts/extractor_python_readability_server.py | python_scripts/extractor_python_readability_server.py |
import sys
import os
import glob
#sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py"))
sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/"))
sys.path.append(os.path.dirname(__file__) )
from thrift.transport import TSocket
from thrift.server import TServer
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
#print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
#print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
server = TServer.TThreadPoolServer(processor, listening_socket)
print ("[Server] Started")
server.serve()
|
import sys
import os
import glob
#sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py"))
sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/"))
sys.path.append(os.path.dirname(__file__) )
from thrift.transport import TSocket
from thrift.transport import TTransport
from thrift.protocol import TBinaryProtocol
from thrift.server import TServer
from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
#print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
#print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
tfactory = TTransport.TBufferedTransportFactory()
#pfactory = TBinaryProtocol.TBinaryProtocolFactory()
pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory()
server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory)
print ("[Server] Started")
server.serve()
| Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance. | Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance.
| Python | agpl-3.0 | AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud |
import sys
import os
import glob
#sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py"))
sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/"))
sys.path.append(os.path.dirname(__file__) )
from thrift.transport import TSocket
+ from thrift.transport import TTransport
+ from thrift.protocol import TBinaryProtocol
from thrift.server import TServer
+ from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated
+
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
#print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
#print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
+ tfactory = TTransport.TBufferedTransportFactory()
+ #pfactory = TBinaryProtocol.TBinaryProtocolFactory()
+ pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory()
+
- server = TServer.TThreadPoolServer(processor, listening_socket)
+ server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory)
print ("[Server] Started")
server.serve()
| Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance. | ## Code Before:
import sys
import os
import glob
#sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py"))
sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/"))
sys.path.append(os.path.dirname(__file__) )
from thrift.transport import TSocket
from thrift.server import TServer
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
#print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
#print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
server = TServer.TThreadPoolServer(processor, listening_socket)
print ("[Server] Started")
server.serve()
## Instruction:
Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance.
## Code After:
import sys
import os
import glob
#sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py"))
sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/"))
sys.path.append(os.path.dirname(__file__) )
from thrift.transport import TSocket
from thrift.transport import TTransport
from thrift.protocol import TBinaryProtocol
from thrift.server import TServer
from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
#print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
#print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
tfactory = TTransport.TBufferedTransportFactory()
#pfactory = TBinaryProtocol.TBinaryProtocolFactory()
pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory()
server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory)
print ("[Server] Started")
server.serve()
|
import sys
import os
import glob
#sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py"))
sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/"))
sys.path.append(os.path.dirname(__file__) )
from thrift.transport import TSocket
+ from thrift.transport import TTransport
+ from thrift.protocol import TBinaryProtocol
from thrift.server import TServer
+ from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated
+
#import thrift_solr
import ExtractorService
import sys
import readability
import readability
def extract_with_python_readability( raw_content ):
doc = readability.Document( raw_content )
return [ u'' + doc.short_title(),
u'' + doc.summary() ]
class ExtractorHandler:
def extract_html( self, raw_html ):
#print raw_html
#raw_html = raw_html.encode( 'utf-8' )
ret = extract_with_python_readability( raw_html )
#print ret[1]
return ret
handler = ExtractorHandler()
processor = ExtractorService.Processor(handler)
listening_socket = TSocket.TServerSocket(port=9090)
+ tfactory = TTransport.TBufferedTransportFactory()
+ #pfactory = TBinaryProtocol.TBinaryProtocolFactory()
+ pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory()
+
- server = TServer.TThreadPoolServer(processor, listening_socket)
+ server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory)
? ++++++++++++++++++++
print ("[Server] Started")
server.serve() |
106fc0f8bae7c776a8f6c7dcec2947420492d118 | homographynet/callbacks.py | homographynet/callbacks.py |
from keras.callbacks import Callback
import keras.backend as K
class LearningRateScheduler(Callback):
"""Learning rate scheduler.
See Caffe SGD docs
"""
def __init__(self, base_lr, gamma, step_size):
super().__init__()
self._lr = base_lr
self._gamma = gamma
self._step_size = step_size
self._iteration = 1
def on_batch_begin(self, batch, logs=None):
if self._iteration % self._step_size == 0:
self._lr *= self._gamma
K.set_value(self.model.optimizer.lr, self._lr)
print('New learning rate:', self._lr)
self._iteration += 1
|
from keras.callbacks import Callback
import keras.backend as K
class LearningRateScheduler(Callback):
"""Learning rate scheduler.
See Caffe SGD docs
"""
def __init__(self, base_lr, gamma, step_size):
super().__init__()
self._base_lr = base_lr
self._gamma = gamma
self._step_size = step_size
self._steps = 0
def on_epoch_begin(self, epoch, logs=None):
self._steps = epoch * self.params['steps']
def on_batch_begin(self, batch, logs=None):
self._steps += 1
if self._steps % self._step_size == 0:
exp = int(self._steps / self._step_size)
lr = self._base_lr * (self._gamma ** exp)
K.set_value(self.model.optimizer.lr, lr)
print('New learning rate:', lr)
| Fix calculation of current steps when starting with epoch != 0 | Fix calculation of current steps when starting with epoch != 0
| Python | apache-2.0 | baudm/HomographyNet |
from keras.callbacks import Callback
import keras.backend as K
class LearningRateScheduler(Callback):
"""Learning rate scheduler.
See Caffe SGD docs
"""
def __init__(self, base_lr, gamma, step_size):
super().__init__()
- self._lr = base_lr
+ self._base_lr = base_lr
self._gamma = gamma
self._step_size = step_size
- self._iteration = 1
+ self._steps = 0
+
+ def on_epoch_begin(self, epoch, logs=None):
+ self._steps = epoch * self.params['steps']
def on_batch_begin(self, batch, logs=None):
+ self._steps += 1
- if self._iteration % self._step_size == 0:
+ if self._steps % self._step_size == 0:
+ exp = int(self._steps / self._step_size)
- self._lr *= self._gamma
+ lr = self._base_lr * (self._gamma ** exp)
- K.set_value(self.model.optimizer.lr, self._lr)
+ K.set_value(self.model.optimizer.lr, lr)
- print('New learning rate:', self._lr)
+ print('New learning rate:', lr)
- self._iteration += 1
| Fix calculation of current steps when starting with epoch != 0 | ## Code Before:
from keras.callbacks import Callback
import keras.backend as K
class LearningRateScheduler(Callback):
"""Learning rate scheduler.
See Caffe SGD docs
"""
def __init__(self, base_lr, gamma, step_size):
super().__init__()
self._lr = base_lr
self._gamma = gamma
self._step_size = step_size
self._iteration = 1
def on_batch_begin(self, batch, logs=None):
if self._iteration % self._step_size == 0:
self._lr *= self._gamma
K.set_value(self.model.optimizer.lr, self._lr)
print('New learning rate:', self._lr)
self._iteration += 1
## Instruction:
Fix calculation of current steps when starting with epoch != 0
## Code After:
from keras.callbacks import Callback
import keras.backend as K
class LearningRateScheduler(Callback):
"""Learning rate scheduler.
See Caffe SGD docs
"""
def __init__(self, base_lr, gamma, step_size):
super().__init__()
self._base_lr = base_lr
self._gamma = gamma
self._step_size = step_size
self._steps = 0
def on_epoch_begin(self, epoch, logs=None):
self._steps = epoch * self.params['steps']
def on_batch_begin(self, batch, logs=None):
self._steps += 1
if self._steps % self._step_size == 0:
exp = int(self._steps / self._step_size)
lr = self._base_lr * (self._gamma ** exp)
K.set_value(self.model.optimizer.lr, lr)
print('New learning rate:', lr)
|
from keras.callbacks import Callback
import keras.backend as K
class LearningRateScheduler(Callback):
"""Learning rate scheduler.
See Caffe SGD docs
"""
def __init__(self, base_lr, gamma, step_size):
super().__init__()
- self._lr = base_lr
+ self._base_lr = base_lr
? +++++
self._gamma = gamma
self._step_size = step_size
- self._iteration = 1
? ^ ^^^^^^ ^
+ self._steps = 0
? ^ ^^ ^
+
+ def on_epoch_begin(self, epoch, logs=None):
+ self._steps = epoch * self.params['steps']
def on_batch_begin(self, batch, logs=None):
+ self._steps += 1
- if self._iteration % self._step_size == 0:
? ^ ^^^^^^
+ if self._steps % self._step_size == 0:
? ^ ^^
+ exp = int(self._steps / self._step_size)
- self._lr *= self._gamma
? -
+ lr = self._base_lr * (self._gamma ** exp)
? +++++ +++++ + ++++++++
- K.set_value(self.model.optimizer.lr, self._lr)
? ------
+ K.set_value(self.model.optimizer.lr, lr)
- print('New learning rate:', self._lr)
? ------
+ print('New learning rate:', lr)
- self._iteration += 1 |
e70537eb2c1a8a68a6a66550e6714816e048bb5e | tests/integration/modules/git.py | tests/integration/modules/git.py |
import shutil
import subprocess
import tempfile
# Import Salt Testing libs
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import salt libs
import integration
class GitModuleTest(integration.ModuleCase):
@classmethod
def setUpClass(cls):
from salt.utils import which
git = which('git')
if not git:
self.skipTest('The git binary is not available')
def setUp(self):
self.repos = tempfile.mkdtemp(dir=integration.TMP)
self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True)
subprocess.check_call(['git', 'init', '--quiet', self.repos])
def test_config_set_value_has_space_characters(self):
'''
git.config_set
'''
config_key = "user.name"
config_value = "foo bar"
ret = self.run_function(
'git.config_set',
cwd=self.repos,
setting_name=config_key,
setting_value=config_value,
)
self.assertEqual("", ret)
output = subprocess.check_output(
['git', 'config', '--local', config_key],
cwd=self.repos)
self.assertEqual(config_value + "\n", output)
| import shutil
import subprocess
import tempfile
# Import Salt Testing libs
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import salt libs
import integration
class GitModuleTest(integration.ModuleCase):
'''
Integration tests for the git module
'''
@classmethod
def setUpClass(cls):
'''
Check if git is installed. If it isn't, skip everything in this class.
'''
from salt.utils import which
git = which('git')
if not git:
cls.skipTest('The git binary is not available')
def setUp(self):
self.repos = tempfile.mkdtemp(dir=integration.TMP)
self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True)
subprocess.check_call(['git', 'init', '--quiet', self.repos])
def test_config_set_value_has_space_characters(self):
'''
Tests the git.config_set function
'''
config_key = "user.name"
config_value = "foo bar"
ret = self.run_function(
'git.config_set',
cwd=self.repos,
setting_name=config_key,
setting_value=config_value,
)
self.assertEqual("", ret)
output = subprocess.check_output(
['git', 'config', '--local', config_key],
cwd=self.repos)
self.assertEqual(config_value + "\n", output)
| Fix missing cls variable and add some docstring info | Fix missing cls variable and add some docstring info
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | -
import shutil
import subprocess
import tempfile
# Import Salt Testing libs
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import salt libs
import integration
class GitModuleTest(integration.ModuleCase):
+ '''
+ Integration tests for the git module
+ '''
+
@classmethod
def setUpClass(cls):
+ '''
+ Check if git is installed. If it isn't, skip everything in this class.
+ '''
from salt.utils import which
git = which('git')
if not git:
- self.skipTest('The git binary is not available')
+ cls.skipTest('The git binary is not available')
def setUp(self):
self.repos = tempfile.mkdtemp(dir=integration.TMP)
self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True)
subprocess.check_call(['git', 'init', '--quiet', self.repos])
def test_config_set_value_has_space_characters(self):
'''
- git.config_set
+ Tests the git.config_set function
'''
config_key = "user.name"
config_value = "foo bar"
ret = self.run_function(
'git.config_set',
cwd=self.repos,
setting_name=config_key,
setting_value=config_value,
)
self.assertEqual("", ret)
output = subprocess.check_output(
['git', 'config', '--local', config_key],
cwd=self.repos)
self.assertEqual(config_value + "\n", output)
| Fix missing cls variable and add some docstring info | ## Code Before:
import shutil
import subprocess
import tempfile
# Import Salt Testing libs
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import salt libs
import integration
class GitModuleTest(integration.ModuleCase):
@classmethod
def setUpClass(cls):
from salt.utils import which
git = which('git')
if not git:
self.skipTest('The git binary is not available')
def setUp(self):
self.repos = tempfile.mkdtemp(dir=integration.TMP)
self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True)
subprocess.check_call(['git', 'init', '--quiet', self.repos])
def test_config_set_value_has_space_characters(self):
'''
git.config_set
'''
config_key = "user.name"
config_value = "foo bar"
ret = self.run_function(
'git.config_set',
cwd=self.repos,
setting_name=config_key,
setting_value=config_value,
)
self.assertEqual("", ret)
output = subprocess.check_output(
['git', 'config', '--local', config_key],
cwd=self.repos)
self.assertEqual(config_value + "\n", output)
## Instruction:
Fix missing cls variable and add some docstring info
## Code After:
import shutil
import subprocess
import tempfile
# Import Salt Testing libs
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import salt libs
import integration
class GitModuleTest(integration.ModuleCase):
'''
Integration tests for the git module
'''
@classmethod
def setUpClass(cls):
'''
Check if git is installed. If it isn't, skip everything in this class.
'''
from salt.utils import which
git = which('git')
if not git:
cls.skipTest('The git binary is not available')
def setUp(self):
self.repos = tempfile.mkdtemp(dir=integration.TMP)
self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True)
subprocess.check_call(['git', 'init', '--quiet', self.repos])
def test_config_set_value_has_space_characters(self):
'''
Tests the git.config_set function
'''
config_key = "user.name"
config_value = "foo bar"
ret = self.run_function(
'git.config_set',
cwd=self.repos,
setting_name=config_key,
setting_value=config_value,
)
self.assertEqual("", ret)
output = subprocess.check_output(
['git', 'config', '--local', config_key],
cwd=self.repos)
self.assertEqual(config_value + "\n", output)
| -
import shutil
import subprocess
import tempfile
# Import Salt Testing libs
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import salt libs
import integration
class GitModuleTest(integration.ModuleCase):
+ '''
+ Integration tests for the git module
+ '''
+
@classmethod
def setUpClass(cls):
+ '''
+ Check if git is installed. If it isn't, skip everything in this class.
+ '''
from salt.utils import which
git = which('git')
if not git:
- self.skipTest('The git binary is not available')
? ---
+ cls.skipTest('The git binary is not available')
? ++
def setUp(self):
self.repos = tempfile.mkdtemp(dir=integration.TMP)
self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True)
subprocess.check_call(['git', 'init', '--quiet', self.repos])
def test_config_set_value_has_space_characters(self):
'''
- git.config_set
+ Tests the git.config_set function
'''
config_key = "user.name"
config_value = "foo bar"
ret = self.run_function(
'git.config_set',
cwd=self.repos,
setting_name=config_key,
setting_value=config_value,
)
self.assertEqual("", ret)
output = subprocess.check_output(
['git', 'config', '--local', config_key],
cwd=self.repos)
self.assertEqual(config_value + "\n", output) |
9771381323e4eb44a13ffc8742615fba61ad2b85 | lino/modlib/notify/consumers.py | lino/modlib/notify/consumers.py | from channels import Group
def ws_echo(message):
Group(str(message.content['text'])).add(message.reply_channel)
message.reply_channel.send({
"text": message.content['text'],
})
| import json
from channels import Channel
from channels import Group
from channels.auth import channel_session_user, channel_session_user_from_http
from django.utils import timezone
from lino.modlib.notify.models import Notification
# This decorator copies the user from the HTTP session (only available in
# websocket.connect or http.request messages) to the channel session (available
# in all consumers with the same reply_channel, so all three here)
@channel_session_user_from_http
def ws_connect(message):
pass
def ws_receive(message):
# All WebSocket frames have either a text or binary payload; we decode the
# text part here assuming it's JSON.
# You could easily build up a basic framework that did this encoding/decoding
# for you as well as handling common errors.
payload = json.loads(message['text'])
payload['reply_channel'] = message.content['reply_channel']
Channel("notify.receive").send(payload)
@channel_session_user
def set_notification_as_seen(message):
notification_id = message['notification_id']
notif = Notification.objects.get(pk=notification_id)
notif.seen = timezone.now()
notif.save()
@channel_session_user
def user_connected(message):
username = message['username']
Group(username).add(message.reply_channel)
message.reply_channel.send({
"text": username,
})
| Update receive and send functions according to the new requirements | Update receive and send functions according to the new requirements
| Python | unknown | lsaffre/lino,lsaffre/lino,khchine5/lino,khchine5/lino,khchine5/lino,lino-framework/lino,lino-framework/lino,lsaffre/lino,lsaffre/lino,lino-framework/lino,lino-framework/lino,lsaffre/lino,khchine5/lino,khchine5/lino,lino-framework/lino | + import json
+
+ from channels import Channel
from channels import Group
+ from channels.auth import channel_session_user, channel_session_user_from_http
+ from django.utils import timezone
+ from lino.modlib.notify.models import Notification
+ # This decorator copies the user from the HTTP session (only available in
+ # websocket.connect or http.request messages) to the channel session (available
+ # in all consumers with the same reply_channel, so all three here)
+ @channel_session_user_from_http
- def ws_echo(message):
+ def ws_connect(message):
+ pass
+
+
+ def ws_receive(message):
+ # All WebSocket frames have either a text or binary payload; we decode the
+ # text part here assuming it's JSON.
+ # You could easily build up a basic framework that did this encoding/decoding
+ # for you as well as handling common errors.
+ payload = json.loads(message['text'])
+ payload['reply_channel'] = message.content['reply_channel']
+ Channel("notify.receive").send(payload)
+
+
+ @channel_session_user
+ def set_notification_as_seen(message):
+ notification_id = message['notification_id']
+ notif = Notification.objects.get(pk=notification_id)
+ notif.seen = timezone.now()
+ notif.save()
+
+
+ @channel_session_user
+ def user_connected(message):
+ username = message['username']
- Group(str(message.content['text'])).add(message.reply_channel)
+ Group(username).add(message.reply_channel)
message.reply_channel.send({
- "text": message.content['text'],
+ "text": username,
})
| Update receive and send functions according to the new requirements | ## Code Before:
from channels import Group
def ws_echo(message):
Group(str(message.content['text'])).add(message.reply_channel)
message.reply_channel.send({
"text": message.content['text'],
})
## Instruction:
Update receive and send functions according to the new requirements
## Code After:
import json
from channels import Channel
from channels import Group
from channels.auth import channel_session_user, channel_session_user_from_http
from django.utils import timezone
from lino.modlib.notify.models import Notification
# This decorator copies the user from the HTTP session (only available in
# websocket.connect or http.request messages) to the channel session (available
# in all consumers with the same reply_channel, so all three here)
@channel_session_user_from_http
def ws_connect(message):
pass
def ws_receive(message):
# All WebSocket frames have either a text or binary payload; we decode the
# text part here assuming it's JSON.
# You could easily build up a basic framework that did this encoding/decoding
# for you as well as handling common errors.
payload = json.loads(message['text'])
payload['reply_channel'] = message.content['reply_channel']
Channel("notify.receive").send(payload)
@channel_session_user
def set_notification_as_seen(message):
notification_id = message['notification_id']
notif = Notification.objects.get(pk=notification_id)
notif.seen = timezone.now()
notif.save()
@channel_session_user
def user_connected(message):
username = message['username']
Group(username).add(message.reply_channel)
message.reply_channel.send({
"text": username,
})
| + import json
+
+ from channels import Channel
from channels import Group
+ from channels.auth import channel_session_user, channel_session_user_from_http
+ from django.utils import timezone
+ from lino.modlib.notify.models import Notification
+ # This decorator copies the user from the HTTP session (only available in
+ # websocket.connect or http.request messages) to the channel session (available
+ # in all consumers with the same reply_channel, so all three here)
+ @channel_session_user_from_http
- def ws_echo(message):
? ^^
+ def ws_connect(message):
? ++++ ^
+ pass
+
+
+ def ws_receive(message):
+ # All WebSocket frames have either a text or binary payload; we decode the
+ # text part here assuming it's JSON.
+ # You could easily build up a basic framework that did this encoding/decoding
+ # for you as well as handling common errors.
+ payload = json.loads(message['text'])
+ payload['reply_channel'] = message.content['reply_channel']
+ Channel("notify.receive").send(payload)
+
+
+ @channel_session_user
+ def set_notification_as_seen(message):
+ notification_id = message['notification_id']
+ notif = Notification.objects.get(pk=notification_id)
+ notif.seen = timezone.now()
+ notif.save()
+
+
+ @channel_session_user
+ def user_connected(message):
+ username = message['username']
- Group(str(message.content['text'])).add(message.reply_channel)
? ^ ^ ----------------------
+ Group(username).add(message.reply_channel)
? + ^ ^^
message.reply_channel.send({
- "text": message.content['text'],
+ "text": username,
}) |
50ab2ed3d8e50e5106dc486e4d20c889d6b18e82 | spkg/base/package_database.py | spkg/base/package_database.py |
from os.path import split, splitext
from json import load
f = open("packages.json")
data = load(f)
g = []
for p in data:
pkg = {
"name": p["name"],
"dependencies": p["dependencies"],
"version": p["version"],
"download": p["download"],
}
g.append(pkg)
from json import dump
from StringIO import StringIO
s = StringIO()
dump(g, s, sort_keys=True, indent=4)
s.seek(0)
s = s.read()
# Remove the trailing space
s = s.replace(" \n", "\n")
f = open("packages.json", "w")
f.write(s)
|
from os.path import split, splitext
from json import load
f = open("packages.json")
data = load(f)
g = []
for p in data:
pkg = {
"name": p["name"],
"dependencies": p["dependencies"],
"version": p["version"],
"download": p["download"],
}
g.append(pkg)
from json import dump
from StringIO import StringIO
s = StringIO()
dump(g, s, sort_keys=True, indent=4)
s.seek(0)
s = s.read()
# Remove the trailing space
s = s.replace(" \n", "\n")
f = open("packages.json", "w")
f.write(s)
f.write("\n")
| Add a new line at the end of the file | Add a new line at the end of the file
| Python | bsd-3-clause | qsnake/qsnake,qsnake/qsnake |
from os.path import split, splitext
from json import load
f = open("packages.json")
data = load(f)
g = []
for p in data:
pkg = {
"name": p["name"],
"dependencies": p["dependencies"],
"version": p["version"],
"download": p["download"],
}
g.append(pkg)
from json import dump
from StringIO import StringIO
s = StringIO()
dump(g, s, sort_keys=True, indent=4)
s.seek(0)
s = s.read()
# Remove the trailing space
s = s.replace(" \n", "\n")
f = open("packages.json", "w")
f.write(s)
+ f.write("\n")
| Add a new line at the end of the file | ## Code Before:
from os.path import split, splitext
from json import load
f = open("packages.json")
data = load(f)
g = []
for p in data:
pkg = {
"name": p["name"],
"dependencies": p["dependencies"],
"version": p["version"],
"download": p["download"],
}
g.append(pkg)
from json import dump
from StringIO import StringIO
s = StringIO()
dump(g, s, sort_keys=True, indent=4)
s.seek(0)
s = s.read()
# Remove the trailing space
s = s.replace(" \n", "\n")
f = open("packages.json", "w")
f.write(s)
## Instruction:
Add a new line at the end of the file
## Code After:
from os.path import split, splitext
from json import load
f = open("packages.json")
data = load(f)
g = []
for p in data:
pkg = {
"name": p["name"],
"dependencies": p["dependencies"],
"version": p["version"],
"download": p["download"],
}
g.append(pkg)
from json import dump
from StringIO import StringIO
s = StringIO()
dump(g, s, sort_keys=True, indent=4)
s.seek(0)
s = s.read()
# Remove the trailing space
s = s.replace(" \n", "\n")
f = open("packages.json", "w")
f.write(s)
f.write("\n")
|
from os.path import split, splitext
from json import load
f = open("packages.json")
data = load(f)
g = []
for p in data:
pkg = {
"name": p["name"],
"dependencies": p["dependencies"],
"version": p["version"],
"download": p["download"],
}
g.append(pkg)
from json import dump
from StringIO import StringIO
s = StringIO()
dump(g, s, sort_keys=True, indent=4)
s.seek(0)
s = s.read()
# Remove the trailing space
s = s.replace(" \n", "\n")
f = open("packages.json", "w")
f.write(s)
+ f.write("\n") |
def66bc381f03970640a61d64b49ad5de9ef3879 | ocaml/build-in.py | ocaml/build-in.py | import sys
import os
from os.path import relpath
ocaml_build_dir = relpath(sys.argv[1], '.')
os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
| import sys
import os
from os.path import relpath
ocaml_build_dir = relpath(sys.argv[1], '.')
# Hack: when we can depend on a full OCaml feed with the build tools, we can remove this.
# Until then, we need to avoid trying to compile against the limited runtime environment.
if 'OCAMLLIB' in os.environ:
del os.environ['OCAMLLIB']
os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
| Remove OCAMLLIB from build environment | Remove OCAMLLIB from build environment
This is a temporary hack: when we can depend on a full OCaml feed with
the build tools, we can remove this. Until then, we need to avoid trying
to compile against the limited runtime environment.
| Python | lgpl-2.1 | 0install/0install,afb/0install,afb/0install,afb/0install,gasche/0install,bastianeicher/0install,bhilton/0install,fdopen/0install,gasche/0install,0install/0install,jaychoo/0install,dbenamy/0install,gfxmonk/0install,jaychoo/0install,dbenamy/0install,DarkGreising/0install,bastianeicher/0install,fdopen/0install,bhilton/0install,bhilton/0install,bartbes/0install,gasche/0install,bastianeicher/0install,dbenamy/0install,DarkGreising/0install,bartbes/0install,fdopen/0install,gasche/0install,HoMeCracKeR/0install,jaychoo/0install,pombreda/0install,gfxmonk/0install,bartbes/0install,afb/0install,HoMeCracKeR/0install,HoMeCracKeR/0install,gfxmonk/0install,DarkGreising/0install,pombreda/0install,pombreda/0install,0install/0install | import sys
import os
from os.path import relpath
ocaml_build_dir = relpath(sys.argv[1], '.')
+
+ # Hack: when we can depend on a full OCaml feed with the build tools, we can remove this.
+ # Until then, we need to avoid trying to compile against the limited runtime environment.
+ if 'OCAMLLIB' in os.environ:
+ del os.environ['OCAMLLIB']
+
os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
| Remove OCAMLLIB from build environment | ## Code Before:
import sys
import os
from os.path import relpath
ocaml_build_dir = relpath(sys.argv[1], '.')
os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
## Instruction:
Remove OCAMLLIB from build environment
## Code After:
import sys
import os
from os.path import relpath
ocaml_build_dir = relpath(sys.argv[1], '.')
# Hack: when we can depend on a full OCaml feed with the build tools, we can remove this.
# Until then, we need to avoid trying to compile against the limited runtime environment.
if 'OCAMLLIB' in os.environ:
del os.environ['OCAMLLIB']
os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
| import sys
import os
from os.path import relpath
ocaml_build_dir = relpath(sys.argv[1], '.')
+
+ # Hack: when we can depend on a full OCaml feed with the build tools, we can remove this.
+ # Until then, we need to avoid trying to compile against the limited runtime environment.
+ if 'OCAMLLIB' in os.environ:
+ del os.environ['OCAMLLIB']
+
os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"]) |
0aa5741ce05dcd4926be9c74af18f6fe46f4aded | etl_framework/utilities/DatetimeConverter.py | etl_framework/utilities/DatetimeConverter.py | """class to convert datetime values"""
import datetime
class DatetimeConverter(object):
"""stuff"""
_EPOCH_0 = datetime.datetime(1970, 1, 1)
def __init__(self):
"""stuff"""
pass
@staticmethod
def get_tomorrow():
"""stuff"""
return datetime.datetime.today() + datetime.timedelta(days=1)
@classmethod
def get_timestamp(cls, datetime_obj):
"""helper method to return timestamp fo datetime object"""
return (datetime_obj - cls._EPOCH_0).total_seconds()
@classmethod
def get_tomorrow_timestamp(cls):
"""stuff"""
return cls.get_timestamp(cls.get_tomorrow())
| """class to convert datetime values"""
import datetime
class DatetimeConverter(object):
"""stuff"""
_EPOCH_0 = datetime.datetime(1970, 1, 1)
def __init__(self):
"""stuff"""
pass
@staticmethod
def get_tomorrow():
"""stuff"""
return datetime.datetime.today() + datetime.timedelta(days=1)
@staticmethod
def get_yesterday():
return datetime.datetime.today() - datetime.timedelta(days=1)
@classmethod
def get_timestamp(cls, datetime_obj):
"""helper method to return timestamp fo datetime object"""
return (datetime_obj - cls._EPOCH_0).total_seconds()
@classmethod
def get_tomorrow_timestamp(cls):
"""stuff"""
return cls.get_timestamp(cls.get_tomorrow())
@classmethod
def get_yesterday_timestamp(cls):
return cls.get_timestamp(cls.get_yesterday())
| Add utility methods for yesterday's date | Add utility methods for yesterday's date
| Python | mit | pantheon-systems/etl-framework | """class to convert datetime values"""
import datetime
class DatetimeConverter(object):
"""stuff"""
_EPOCH_0 = datetime.datetime(1970, 1, 1)
def __init__(self):
"""stuff"""
pass
@staticmethod
def get_tomorrow():
"""stuff"""
return datetime.datetime.today() + datetime.timedelta(days=1)
+ @staticmethod
+ def get_yesterday():
+
+ return datetime.datetime.today() - datetime.timedelta(days=1)
+
@classmethod
def get_timestamp(cls, datetime_obj):
"""helper method to return timestamp fo datetime object"""
return (datetime_obj - cls._EPOCH_0).total_seconds()
@classmethod
def get_tomorrow_timestamp(cls):
"""stuff"""
return cls.get_timestamp(cls.get_tomorrow())
+ @classmethod
+ def get_yesterday_timestamp(cls):
+
+ return cls.get_timestamp(cls.get_yesterday())
+ | Add utility methods for yesterday's date | ## Code Before:
"""class to convert datetime values"""
import datetime
class DatetimeConverter(object):
"""stuff"""
_EPOCH_0 = datetime.datetime(1970, 1, 1)
def __init__(self):
"""stuff"""
pass
@staticmethod
def get_tomorrow():
"""stuff"""
return datetime.datetime.today() + datetime.timedelta(days=1)
@classmethod
def get_timestamp(cls, datetime_obj):
"""helper method to return timestamp fo datetime object"""
return (datetime_obj - cls._EPOCH_0).total_seconds()
@classmethod
def get_tomorrow_timestamp(cls):
"""stuff"""
return cls.get_timestamp(cls.get_tomorrow())
## Instruction:
Add utility methods for yesterday's date
## Code After:
"""class to convert datetime values"""
import datetime
class DatetimeConverter(object):
"""stuff"""
_EPOCH_0 = datetime.datetime(1970, 1, 1)
def __init__(self):
"""stuff"""
pass
@staticmethod
def get_tomorrow():
"""stuff"""
return datetime.datetime.today() + datetime.timedelta(days=1)
@staticmethod
def get_yesterday():
return datetime.datetime.today() - datetime.timedelta(days=1)
@classmethod
def get_timestamp(cls, datetime_obj):
"""helper method to return timestamp fo datetime object"""
return (datetime_obj - cls._EPOCH_0).total_seconds()
@classmethod
def get_tomorrow_timestamp(cls):
"""stuff"""
return cls.get_timestamp(cls.get_tomorrow())
@classmethod
def get_yesterday_timestamp(cls):
return cls.get_timestamp(cls.get_yesterday())
| """class to convert datetime values"""
import datetime
class DatetimeConverter(object):
"""stuff"""
_EPOCH_0 = datetime.datetime(1970, 1, 1)
def __init__(self):
"""stuff"""
pass
@staticmethod
def get_tomorrow():
"""stuff"""
return datetime.datetime.today() + datetime.timedelta(days=1)
+ @staticmethod
+ def get_yesterday():
+
+ return datetime.datetime.today() - datetime.timedelta(days=1)
+
@classmethod
def get_timestamp(cls, datetime_obj):
"""helper method to return timestamp fo datetime object"""
return (datetime_obj - cls._EPOCH_0).total_seconds()
@classmethod
def get_tomorrow_timestamp(cls):
"""stuff"""
return cls.get_timestamp(cls.get_tomorrow())
+
+ @classmethod
+ def get_yesterday_timestamp(cls):
+
+ return cls.get_timestamp(cls.get_yesterday()) |
7079614f35de60def5f4e1cc1cb17cf3e5b4d9c6 | setup.py | setup.py | from distutils.core import setup
import os
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name='facebook-sdk',
version='0.3.2',
description='This client library is designed to support the Facebook '
'Graph API and the official Facebook JavaScript SDK, which '
'is the canonical way to implement Facebook authentication.',
author='Facebook',
maintainer='Martey Dodoo',
maintainer_email='facebook-sdk@marteydodoo.com',
url='https://github.com/pythonforfacebook/facebook-sdk',
license='Apache',
py_modules=[
'facebook',
],
long_description=read("README.rst"),
classifiers=[
'License :: OSI Approved :: Apache Software License',
],
)
| from distutils.core import setup
setup(
name='facebook-sdk',
version='0.3.2',
description='This client library is designed to support the Facebook '
'Graph API and the official Facebook JavaScript SDK, which '
'is the canonical way to implement Facebook authentication.',
author='Facebook',
maintainer='Martey Dodoo',
maintainer_email='facebook-sdk@marteydodoo.com',
url='https://github.com/pythonforfacebook/facebook-sdk',
license='Apache',
py_modules=[
'facebook',
],
long_description=open("README.rst").read(),
classifiers=[
'License :: OSI Approved :: Apache Software License',
],
)
| Change way README is imported. | Change way README is imported.
The custom read function is unnecessary since only one file is being
accessed. Removing it reduces the amount of code.
| Python | apache-2.0 | Aloomaio/facebook-sdk,mobolic/facebook-sdk | from distutils.core import setup
- import os
-
- def read(fname):
- return open(os.path.join(os.path.dirname(__file__), fname)).read()
-
setup(
name='facebook-sdk',
version='0.3.2',
description='This client library is designed to support the Facebook '
'Graph API and the official Facebook JavaScript SDK, which '
'is the canonical way to implement Facebook authentication.',
author='Facebook',
maintainer='Martey Dodoo',
maintainer_email='facebook-sdk@marteydodoo.com',
url='https://github.com/pythonforfacebook/facebook-sdk',
license='Apache',
py_modules=[
'facebook',
],
- long_description=read("README.rst"),
+ long_description=open("README.rst").read(),
classifiers=[
'License :: OSI Approved :: Apache Software License',
],
)
| Change way README is imported. | ## Code Before:
from distutils.core import setup
import os
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name='facebook-sdk',
version='0.3.2',
description='This client library is designed to support the Facebook '
'Graph API and the official Facebook JavaScript SDK, which '
'is the canonical way to implement Facebook authentication.',
author='Facebook',
maintainer='Martey Dodoo',
maintainer_email='facebook-sdk@marteydodoo.com',
url='https://github.com/pythonforfacebook/facebook-sdk',
license='Apache',
py_modules=[
'facebook',
],
long_description=read("README.rst"),
classifiers=[
'License :: OSI Approved :: Apache Software License',
],
)
## Instruction:
Change way README is imported.
## Code After:
from distutils.core import setup
setup(
name='facebook-sdk',
version='0.3.2',
description='This client library is designed to support the Facebook '
'Graph API and the official Facebook JavaScript SDK, which '
'is the canonical way to implement Facebook authentication.',
author='Facebook',
maintainer='Martey Dodoo',
maintainer_email='facebook-sdk@marteydodoo.com',
url='https://github.com/pythonforfacebook/facebook-sdk',
license='Apache',
py_modules=[
'facebook',
],
long_description=open("README.rst").read(),
classifiers=[
'License :: OSI Approved :: Apache Software License',
],
)
| from distutils.core import setup
- import os
-
- def read(fname):
- return open(os.path.join(os.path.dirname(__file__), fname)).read()
-
setup(
name='facebook-sdk',
version='0.3.2',
description='This client library is designed to support the Facebook '
'Graph API and the official Facebook JavaScript SDK, which '
'is the canonical way to implement Facebook authentication.',
author='Facebook',
maintainer='Martey Dodoo',
maintainer_email='facebook-sdk@marteydodoo.com',
url='https://github.com/pythonforfacebook/facebook-sdk',
license='Apache',
py_modules=[
'facebook',
],
- long_description=read("README.rst"),
? ^ ^^
+ long_description=open("README.rst").read(),
? ^^ ^ +++++++
classifiers=[
'License :: OSI Approved :: Apache Software License',
],
) |
d8cc0fdaea848ed5d626ba6ba4292fd3cb906da3 | project7/TrackParser.py | project7/TrackParser.py | """ Created by Max 12/2/2017 """
import pprint
import numpy as np
class TrackParser:
@staticmethod
def parse_track(path_to_track_file: str) -> np.ndarray:
track = None
with open(path_to_track_file, 'r') as track_file:
lines = track_file.readlines()
dimensions_str = lines[0]
dims = dimensions_str.split(',')
track = np.zeros((int(dims[0]), int(dims[1])))
for line_index in range(1, len(lines)):
line = lines[line_index]
for char_index in range(len(line) - 1):
track_value = TrackParser.get_char_value(line[char_index])
track[line_index-1][char_index] = track_value
return track
@staticmethod
def get_char_value(char: str):
if char == '#':
return -1
elif char == '.':
return 0
elif char == 'S':
return 1
elif char == 'F':
return 2
else:
return -1
np.set_printoptions(linewidth=500)
pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
| """ Created by Max 12/2/2017 """
import pprint
import numpy as np
class TrackParser:
@staticmethod
def parse_track(path_to_track_file: str) -> np.ndarray:
track = None
with open(path_to_track_file, 'r') as track_file:
lines = track_file.readlines()
dimensions_str = lines[0]
dims = dimensions_str.split(',')
# (Y,X) coords.
track = np.zeros((int(dims[0]), int(dims[1])))
for line_index in range(1, len(lines)):
line = lines[line_index]
for char_index in range(len(line) - 1):
track_value = TrackParser.get_char_value(line[char_index])
track[line_index-1][char_index] = track_value
return track
@staticmethod
def get_char_value(char: str):
if char == '#':
return -1
elif char == '.':
return 0
elif char == 'S':
return 1
elif char == 'F':
return 2
else:
return -1
# np.set_printoptions(linewidth=500)
# pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
| Add comment to clarify coordinate order | Add comment to clarify coordinate order
coordinates are (y,x) in the track.
| Python | apache-2.0 | MaxRobinson/CS449,MaxRobinson/CS449,MaxRobinson/CS449 | """ Created by Max 12/2/2017 """
import pprint
import numpy as np
class TrackParser:
@staticmethod
def parse_track(path_to_track_file: str) -> np.ndarray:
track = None
with open(path_to_track_file, 'r') as track_file:
lines = track_file.readlines()
dimensions_str = lines[0]
dims = dimensions_str.split(',')
+ # (Y,X) coords.
track = np.zeros((int(dims[0]), int(dims[1])))
for line_index in range(1, len(lines)):
line = lines[line_index]
for char_index in range(len(line) - 1):
track_value = TrackParser.get_char_value(line[char_index])
track[line_index-1][char_index] = track_value
return track
@staticmethod
def get_char_value(char: str):
if char == '#':
return -1
elif char == '.':
return 0
elif char == 'S':
return 1
elif char == 'F':
return 2
else:
return -1
- np.set_printoptions(linewidth=500)
+ # np.set_printoptions(linewidth=500)
- pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
+ # pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
| Add comment to clarify coordinate order | ## Code Before:
""" Created by Max 12/2/2017 """
import pprint
import numpy as np
class TrackParser:
@staticmethod
def parse_track(path_to_track_file: str) -> np.ndarray:
track = None
with open(path_to_track_file, 'r') as track_file:
lines = track_file.readlines()
dimensions_str = lines[0]
dims = dimensions_str.split(',')
track = np.zeros((int(dims[0]), int(dims[1])))
for line_index in range(1, len(lines)):
line = lines[line_index]
for char_index in range(len(line) - 1):
track_value = TrackParser.get_char_value(line[char_index])
track[line_index-1][char_index] = track_value
return track
@staticmethod
def get_char_value(char: str):
if char == '#':
return -1
elif char == '.':
return 0
elif char == 'S':
return 1
elif char == 'F':
return 2
else:
return -1
np.set_printoptions(linewidth=500)
pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
## Instruction:
Add comment to clarify coordinate order
## Code After:
""" Created by Max 12/2/2017 """
import pprint
import numpy as np
class TrackParser:
@staticmethod
def parse_track(path_to_track_file: str) -> np.ndarray:
track = None
with open(path_to_track_file, 'r') as track_file:
lines = track_file.readlines()
dimensions_str = lines[0]
dims = dimensions_str.split(',')
# (Y,X) coords.
track = np.zeros((int(dims[0]), int(dims[1])))
for line_index in range(1, len(lines)):
line = lines[line_index]
for char_index in range(len(line) - 1):
track_value = TrackParser.get_char_value(line[char_index])
track[line_index-1][char_index] = track_value
return track
@staticmethod
def get_char_value(char: str):
if char == '#':
return -1
elif char == '.':
return 0
elif char == 'S':
return 1
elif char == 'F':
return 2
else:
return -1
# np.set_printoptions(linewidth=500)
# pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
| """ Created by Max 12/2/2017 """
import pprint
import numpy as np
class TrackParser:
@staticmethod
def parse_track(path_to_track_file: str) -> np.ndarray:
track = None
with open(path_to_track_file, 'r') as track_file:
lines = track_file.readlines()
dimensions_str = lines[0]
dims = dimensions_str.split(',')
+ # (Y,X) coords.
track = np.zeros((int(dims[0]), int(dims[1])))
for line_index in range(1, len(lines)):
line = lines[line_index]
for char_index in range(len(line) - 1):
track_value = TrackParser.get_char_value(line[char_index])
track[line_index-1][char_index] = track_value
return track
@staticmethod
def get_char_value(char: str):
if char == '#':
return -1
elif char == '.':
return 0
elif char == 'S':
return 1
elif char == 'F':
return 2
else:
return -1
- np.set_printoptions(linewidth=500)
+ # np.set_printoptions(linewidth=500)
? ++
- pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
+ # pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
? ++
|
e1ffdcc5f12be623633e2abab2041fcb574173ea | homeassistant/components/zeroconf.py | homeassistant/components/zeroconf.py | import logging
import socket
from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__)
REQUIREMENTS = ["zeroconf==0.17.5"]
_LOGGER = logging.getLogger(__name__)
DOMAIN = "zeroconf"
ZEROCONF_TYPE = "_home-assistant._tcp.local."
DEPENDENCIES = ["http"]
def setup(hass, config):
"""Set up Zeroconf and make Home Assistant discoverable."""
from zeroconf import Zeroconf, ServiceInfo
zeroconf = Zeroconf()
zeroconf_name = "{}.{}".format(hass.config.location_name,
ZEROCONF_TYPE)
params = {"version": __version__, "base_url": hass.http.base_url,
"needs_auth": (hass.http.api_password != "")}
info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name,
socket.inet_aton(hass.http.routable_address),
hass.http.server_address[1], 0, 0, params)
zeroconf.register_service(info)
def stop_zeroconf(event):
"""Stop Zeroconf."""
zeroconf.unregister_service(info)
hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf)
return True
| import logging
import socket
from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__)
REQUIREMENTS = ["zeroconf==0.17.5"]
DEPENDENCIES = ["api"]
_LOGGER = logging.getLogger(__name__)
DOMAIN = "zeroconf"
ZEROCONF_TYPE = "_home-assistant._tcp.local."
def setup(hass, config):
"""Set up Zeroconf and make Home Assistant discoverable."""
from zeroconf import Zeroconf, ServiceInfo
zeroconf = Zeroconf()
zeroconf_name = "{}.{}".format(hass.config.location_name,
ZEROCONF_TYPE)
params = {"version": __version__, "base_url": hass.config.api.base_url,
"needs_auth": (hass.config.api.api_password != "")}
info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name,
socket.inet_aton(hass.config.api.host),
hass.config.api.port, 0, 0, params)
zeroconf.register_service(info)
def stop_zeroconf(event):
"""Stop Zeroconf."""
zeroconf.unregister_service(info)
hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf)
return True
| Use hass.config.api instead of hass.http | Use hass.config.api instead of hass.http
| Python | mit | miniconfig/home-assistant,Julian/home-assistant,toddeye/home-assistant,ct-23/home-assistant,deisi/home-assistant,tchellomello/home-assistant,rohitranjan1991/home-assistant,Julian/home-assistant,Duoxilian/home-assistant,betrisey/home-assistant,keerts/home-assistant,ct-23/home-assistant,tboyce021/home-assistant,kyvinh/home-assistant,leoc/home-assistant,jawilson/home-assistant,JshWright/home-assistant,turbokongen/home-assistant,varunr047/homefile,open-homeautomation/home-assistant,betrisey/home-assistant,GenericStudent/home-assistant,alexmogavero/home-assistant,fbradyirl/home-assistant,eagleamon/home-assistant,kyvinh/home-assistant,toddeye/home-assistant,auduny/home-assistant,aronsky/home-assistant,hmronline/home-assistant,bdfoster/blumate,jaharkes/home-assistant,fbradyirl/home-assistant,srcLurker/home-assistant,morphis/home-assistant,deisi/home-assistant,LinuxChristian/home-assistant,hexxter/home-assistant,sdague/home-assistant,LinuxChristian/home-assistant,shaftoe/home-assistant,stefan-jonasson/home-assistant,adrienbrault/home-assistant,w1ll1am23/home-assistant,tinloaf/home-assistant,soldag/home-assistant,molobrakos/home-assistant,Zyell/home-assistant,HydrelioxGitHub/home-assistant,keerts/home-assistant,molobrakos/home-assistant,varunr047/homefile,jaharkes/home-assistant,xifle/home-assistant,Smart-Torvy/torvy-home-assistant,stefan-jonasson/home-assistant,tboyce1/home-assistant,kennedyshead/home-assistant,sffjunkie/home-assistant,kennedyshead/home-assistant,jnewland/home-assistant,tinloaf/home-assistant,dmeulen/home-assistant,xifle/home-assistant,mikaelboman/home-assistant,Zac-HD/home-assistant,hmronline/home-assistant,partofthething/home-assistant,robjohnson189/home-assistant,lukas-hetzenecker/home-assistant,nugget/home-assistant,aequitas/home-assistant,oandrew/home-assistant,morphis/home-assistant,eagleamon/home-assistant,HydrelioxGitHub/home-assistant,Teagan42/home-assistant,ct-23/home-assistant,Zyell/home-assistant,sffjunkie/home-assistant,tinloaf/home-assistant,philipbl/home-assistant,kyvinh/home-assistant,Julian/home-assistant,PetePriority/home-assistant,sffjunkie/home-assistant,bdfoster/blumate,Teagan42/home-assistant,rohitranjan1991/home-assistant,deisi/home-assistant,miniconfig/home-assistant,robjohnson189/home-assistant,happyleavesaoc/home-assistant,morphis/home-assistant,GenericStudent/home-assistant,soldag/home-assistant,deisi/home-assistant,Duoxilian/home-assistant,persandstrom/home-assistant,w1ll1am23/home-assistant,oandrew/home-assistant,mezz64/home-assistant,DavidLP/home-assistant,morphis/home-assistant,tboyce1/home-assistant,florianholzapfel/home-assistant,Danielhiversen/home-assistant,home-assistant/home-assistant,florianholzapfel/home-assistant,srcLurker/home-assistant,stefan-jonasson/home-assistant,Smart-Torvy/torvy-home-assistant,sdague/home-assistant,kyvinh/home-assistant,balloob/home-assistant,happyleavesaoc/home-assistant,MartinHjelmare/home-assistant,happyleavesaoc/home-assistant,adrienbrault/home-assistant,ewandor/home-assistant,nugget/home-assistant,xifle/home-assistant,MungoRae/home-assistant,robbiet480/home-assistant,MartinHjelmare/home-assistant,alexmogavero/home-assistant,leppa/home-assistant,open-homeautomation/home-assistant,leoc/home-assistant,LinuxChristian/home-assistant,leoc/home-assistant,MartinHjelmare/home-assistant,Julian/home-assistant,miniconfig/home-assistant,lukas-hetzenecker/home-assistant,mikaelboman/home-assistant,joopert/home-assistant,Zac-HD/home-assistant,auduny/home-assistant,Zac-HD/home-assistant,joopert/home-assistant,ma314smith/home-assistant,Zyell/home-assistant,devdelay/home-assistant,srcLurker/home-assistant,hexxter/home-assistant,HydrelioxGitHub/home-assistant,mikaelboman/home-assistant,aequitas/home-assistant,qedi-r/home-assistant,nkgilley/home-assistant,jnewland/home-assistant,deisi/home-assistant,leoc/home-assistant,jaharkes/home-assistant,varunr047/homefile,nugget/home-assistant,hexxter/home-assistant,sander76/home-assistant,fbradyirl/home-assistant,partofthething/home-assistant,jamespcole/home-assistant,persandstrom/home-assistant,jnewland/home-assistant,balloob/home-assistant,leppa/home-assistant,bdfoster/blumate,keerts/home-assistant,shaftoe/home-assistant,pschmitt/home-assistant,philipbl/home-assistant,tboyce1/home-assistant,betrisey/home-assistant,hmronline/home-assistant,sander76/home-assistant,xifle/home-assistant,JshWright/home-assistant,oandrew/home-assistant,PetePriority/home-assistant,varunr047/homefile,dmeulen/home-assistant,tboyce021/home-assistant,alexmogavero/home-assistant,DavidLP/home-assistant,persandstrom/home-assistant,LinuxChristian/home-assistant,philipbl/home-assistant,JshWright/home-assistant,FreekingDean/home-assistant,robbiet480/home-assistant,jabesq/home-assistant,ct-23/home-assistant,hmronline/home-assistant,ma314smith/home-assistant,Smart-Torvy/torvy-home-assistant,JshWright/home-assistant,home-assistant/home-assistant,shaftoe/home-assistant,devdelay/home-assistant,emilhetty/home-assistant,emilhetty/home-assistant,Smart-Torvy/torvy-home-assistant,mKeRix/home-assistant,mKeRix/home-assistant,Danielhiversen/home-assistant,stefan-jonasson/home-assistant,bdfoster/blumate,sffjunkie/home-assistant,bdfoster/blumate,jawilson/home-assistant,florianholzapfel/home-assistant,MungoRae/home-assistant,jaharkes/home-assistant,Cinntax/home-assistant,aronsky/home-assistant,auduny/home-assistant,robjohnson189/home-assistant,florianholzapfel/home-assistant,emilhetty/home-assistant,devdelay/home-assistant,sffjunkie/home-assistant,MungoRae/home-assistant,mikaelboman/home-assistant,Duoxilian/home-assistant,Duoxilian/home-assistant,dmeulen/home-assistant,betrisey/home-assistant,ma314smith/home-assistant,open-homeautomation/home-assistant,DavidLP/home-assistant,robjohnson189/home-assistant,dmeulen/home-assistant,mikaelboman/home-assistant,jamespcole/home-assistant,philipbl/home-assistant,jabesq/home-assistant,emilhetty/home-assistant,oandrew/home-assistant,rohitranjan1991/home-assistant,MungoRae/home-assistant,mKeRix/home-assistant,shaftoe/home-assistant,titilambert/home-assistant,MungoRae/home-assistant,varunr047/homefile,balloob/home-assistant,molobrakos/home-assistant,postlund/home-assistant,happyleavesaoc/home-assistant,mezz64/home-assistant,Zac-HD/home-assistant,ct-23/home-assistant,open-homeautomation/home-assistant,ma314smith/home-assistant,FreekingDean/home-assistant,mKeRix/home-assistant,qedi-r/home-assistant,miniconfig/home-assistant,ewandor/home-assistant,srcLurker/home-assistant,ewandor/home-assistant,PetePriority/home-assistant,postlund/home-assistant,jabesq/home-assistant,alexmogavero/home-assistant,aequitas/home-assistant,devdelay/home-assistant,hexxter/home-assistant,Cinntax/home-assistant,titilambert/home-assistant,emilhetty/home-assistant,tchellomello/home-assistant,keerts/home-assistant,LinuxChristian/home-assistant,eagleamon/home-assistant,nkgilley/home-assistant,hmronline/home-assistant,jamespcole/home-assistant,pschmitt/home-assistant,tboyce1/home-assistant,eagleamon/home-assistant,turbokongen/home-assistant | import logging
import socket
from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__)
REQUIREMENTS = ["zeroconf==0.17.5"]
+ DEPENDENCIES = ["api"]
+
_LOGGER = logging.getLogger(__name__)
DOMAIN = "zeroconf"
ZEROCONF_TYPE = "_home-assistant._tcp.local."
-
- DEPENDENCIES = ["http"]
def setup(hass, config):
"""Set up Zeroconf and make Home Assistant discoverable."""
from zeroconf import Zeroconf, ServiceInfo
zeroconf = Zeroconf()
zeroconf_name = "{}.{}".format(hass.config.location_name,
ZEROCONF_TYPE)
- params = {"version": __version__, "base_url": hass.http.base_url,
+ params = {"version": __version__, "base_url": hass.config.api.base_url,
- "needs_auth": (hass.http.api_password != "")}
+ "needs_auth": (hass.config.api.api_password != "")}
info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name,
- socket.inet_aton(hass.http.routable_address),
+ socket.inet_aton(hass.config.api.host),
- hass.http.server_address[1], 0, 0, params)
+ hass.config.api.port, 0, 0, params)
zeroconf.register_service(info)
def stop_zeroconf(event):
"""Stop Zeroconf."""
zeroconf.unregister_service(info)
hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf)
return True
| Use hass.config.api instead of hass.http | ## Code Before:
import logging
import socket
from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__)
REQUIREMENTS = ["zeroconf==0.17.5"]
_LOGGER = logging.getLogger(__name__)
DOMAIN = "zeroconf"
ZEROCONF_TYPE = "_home-assistant._tcp.local."
DEPENDENCIES = ["http"]
def setup(hass, config):
"""Set up Zeroconf and make Home Assistant discoverable."""
from zeroconf import Zeroconf, ServiceInfo
zeroconf = Zeroconf()
zeroconf_name = "{}.{}".format(hass.config.location_name,
ZEROCONF_TYPE)
params = {"version": __version__, "base_url": hass.http.base_url,
"needs_auth": (hass.http.api_password != "")}
info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name,
socket.inet_aton(hass.http.routable_address),
hass.http.server_address[1], 0, 0, params)
zeroconf.register_service(info)
def stop_zeroconf(event):
"""Stop Zeroconf."""
zeroconf.unregister_service(info)
hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf)
return True
## Instruction:
Use hass.config.api instead of hass.http
## Code After:
import logging
import socket
from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__)
REQUIREMENTS = ["zeroconf==0.17.5"]
DEPENDENCIES = ["api"]
_LOGGER = logging.getLogger(__name__)
DOMAIN = "zeroconf"
ZEROCONF_TYPE = "_home-assistant._tcp.local."
def setup(hass, config):
"""Set up Zeroconf and make Home Assistant discoverable."""
from zeroconf import Zeroconf, ServiceInfo
zeroconf = Zeroconf()
zeroconf_name = "{}.{}".format(hass.config.location_name,
ZEROCONF_TYPE)
params = {"version": __version__, "base_url": hass.config.api.base_url,
"needs_auth": (hass.config.api.api_password != "")}
info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name,
socket.inet_aton(hass.config.api.host),
hass.config.api.port, 0, 0, params)
zeroconf.register_service(info)
def stop_zeroconf(event):
"""Stop Zeroconf."""
zeroconf.unregister_service(info)
hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf)
return True
| import logging
import socket
from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__)
REQUIREMENTS = ["zeroconf==0.17.5"]
+ DEPENDENCIES = ["api"]
+
_LOGGER = logging.getLogger(__name__)
DOMAIN = "zeroconf"
ZEROCONF_TYPE = "_home-assistant._tcp.local."
-
- DEPENDENCIES = ["http"]
def setup(hass, config):
"""Set up Zeroconf and make Home Assistant discoverable."""
from zeroconf import Zeroconf, ServiceInfo
zeroconf = Zeroconf()
zeroconf_name = "{}.{}".format(hass.config.location_name,
ZEROCONF_TYPE)
- params = {"version": __version__, "base_url": hass.http.base_url,
? ^^^
+ params = {"version": __version__, "base_url": hass.config.api.base_url,
? ^^^^^^^^ +
- "needs_auth": (hass.http.api_password != "")}
? ^^^
+ "needs_auth": (hass.config.api.api_password != "")}
? ^^^^^^^^ +
info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name,
- socket.inet_aton(hass.http.routable_address),
? -------------------
+ socket.inet_aton(hass.config.api.host),
? +++++++++++ ++
- hass.http.server_address[1], 0, 0, params)
+ hass.config.api.port, 0, 0, params)
zeroconf.register_service(info)
def stop_zeroconf(event):
"""Stop Zeroconf."""
zeroconf.unregister_service(info)
hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf)
return True |
69b0e1c60eafff596ebb494a7e79a22c6bea374b | polling_stations/apps/data_collection/management/commands/import_hart.py | polling_stations/apps/data_collection/management/commands/import_hart.py | from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = 'E07000089'
addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
elections = ['parl.2017-06-08']
csv_delimiter = '\t'
| from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = 'E07000089'
addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
elections = ['parl.2017-06-08']
csv_delimiter = '\t'
def station_record_to_dict(self, record):
if record.polling_place_id == '1914':
record = record._replace(polling_place_easting = '479224')
record = record._replace(polling_place_northing = '154016')
return super().station_record_to_dict(record)
| Fix dodgy point in Hart | Fix dodgy point in Hart
| Python | bsd-3-clause | DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations,chris48s/UK-Polling-Stations,chris48s/UK-Polling-Stations,chris48s/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations | from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = 'E07000089'
addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
elections = ['parl.2017-06-08']
csv_delimiter = '\t'
+ def station_record_to_dict(self, record):
+
+ if record.polling_place_id == '1914':
+ record = record._replace(polling_place_easting = '479224')
+ record = record._replace(polling_place_northing = '154016')
+
+ return super().station_record_to_dict(record)
+ | Fix dodgy point in Hart | ## Code Before:
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = 'E07000089'
addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
elections = ['parl.2017-06-08']
csv_delimiter = '\t'
## Instruction:
Fix dodgy point in Hart
## Code After:
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = 'E07000089'
addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
elections = ['parl.2017-06-08']
csv_delimiter = '\t'
def station_record_to_dict(self, record):
if record.polling_place_id == '1914':
record = record._replace(polling_place_easting = '479224')
record = record._replace(polling_place_northing = '154016')
return super().station_record_to_dict(record)
| from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = 'E07000089'
addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV'
elections = ['parl.2017-06-08']
csv_delimiter = '\t'
+
+ def station_record_to_dict(self, record):
+
+ if record.polling_place_id == '1914':
+ record = record._replace(polling_place_easting = '479224')
+ record = record._replace(polling_place_northing = '154016')
+
+ return super().station_record_to_dict(record) |
55e316a45256d054d19425015ef13868a84c5ff1 | src/pip/_internal/resolution/resolvelib/reporter.py | src/pip/_internal/resolution/resolvelib/reporter.py | from collections import defaultdict
from logging import getLogger
from pip._vendor.resolvelib.reporters import BaseReporter
from pip._internal.utils.typing import MYPY_CHECK_RUNNING
if MYPY_CHECK_RUNNING:
from typing import DefaultDict
from .base import Candidate
logger = getLogger(__name__)
class PipReporter(BaseReporter):
def __init__(self):
# type: () -> None
self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int]
self._messages_at_backtrack = {
1: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
8: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
13: (
"This is taking longer than usual. You might need to provide the "
"dependency resolver with stricter constraints to reduce runtime."
"If you want to abort this run, you can press Ctrl + C to do so."
)
}
def backtracking(self, candidate):
# type: (Candidate) -> None
self.backtracks_by_package[candidate.name] += 1
count = self.backtracks_by_package[candidate.name]
if count not in self._messages_at_backtrack:
return
message = self._messages_at_backtrack[count]
logger.info("INFO: %s", message)
| from collections import defaultdict
from logging import getLogger
from pip._vendor.resolvelib.reporters import BaseReporter
from pip._internal.utils.typing import MYPY_CHECK_RUNNING
if MYPY_CHECK_RUNNING:
from typing import DefaultDict
from .base import Candidate
logger = getLogger(__name__)
class PipReporter(BaseReporter):
def __init__(self):
# type: () -> None
self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int]
self._messages_at_backtrack = {
1: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
8: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
13: (
"This is taking longer than usual. You might need to provide the "
"dependency resolver with stricter constraints to reduce runtime."
"If you want to abort this run, you can press Ctrl + C to do so."
"To improve how pip performs, tell us that this happened here: "
"https://pip.pypa.io/surveys/backtracking"
)
}
def backtracking(self, candidate):
# type: (Candidate) -> None
self.backtracks_by_package[candidate.name] += 1
count = self.backtracks_by_package[candidate.name]
if count not in self._messages_at_backtrack:
return
message = self._messages_at_backtrack[count]
logger.info("INFO: %s", message)
| Add the last line to the info message | Add the last line to the info message
| Python | mit | sbidoul/pip,pradyunsg/pip,pypa/pip,pypa/pip,sbidoul/pip,pfmoore/pip,pfmoore/pip,pradyunsg/pip | from collections import defaultdict
from logging import getLogger
from pip._vendor.resolvelib.reporters import BaseReporter
from pip._internal.utils.typing import MYPY_CHECK_RUNNING
if MYPY_CHECK_RUNNING:
from typing import DefaultDict
from .base import Candidate
logger = getLogger(__name__)
class PipReporter(BaseReporter):
def __init__(self):
# type: () -> None
self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int]
self._messages_at_backtrack = {
1: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
8: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
13: (
"This is taking longer than usual. You might need to provide the "
"dependency resolver with stricter constraints to reduce runtime."
"If you want to abort this run, you can press Ctrl + C to do so."
+ "To improve how pip performs, tell us that this happened here: "
+ "https://pip.pypa.io/surveys/backtracking"
)
}
def backtracking(self, candidate):
# type: (Candidate) -> None
self.backtracks_by_package[candidate.name] += 1
count = self.backtracks_by_package[candidate.name]
if count not in self._messages_at_backtrack:
return
message = self._messages_at_backtrack[count]
logger.info("INFO: %s", message)
| Add the last line to the info message | ## Code Before:
from collections import defaultdict
from logging import getLogger
from pip._vendor.resolvelib.reporters import BaseReporter
from pip._internal.utils.typing import MYPY_CHECK_RUNNING
if MYPY_CHECK_RUNNING:
from typing import DefaultDict
from .base import Candidate
logger = getLogger(__name__)
class PipReporter(BaseReporter):
def __init__(self):
# type: () -> None
self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int]
self._messages_at_backtrack = {
1: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
8: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
13: (
"This is taking longer than usual. You might need to provide the "
"dependency resolver with stricter constraints to reduce runtime."
"If you want to abort this run, you can press Ctrl + C to do so."
)
}
def backtracking(self, candidate):
# type: (Candidate) -> None
self.backtracks_by_package[candidate.name] += 1
count = self.backtracks_by_package[candidate.name]
if count not in self._messages_at_backtrack:
return
message = self._messages_at_backtrack[count]
logger.info("INFO: %s", message)
## Instruction:
Add the last line to the info message
## Code After:
from collections import defaultdict
from logging import getLogger
from pip._vendor.resolvelib.reporters import BaseReporter
from pip._internal.utils.typing import MYPY_CHECK_RUNNING
if MYPY_CHECK_RUNNING:
from typing import DefaultDict
from .base import Candidate
logger = getLogger(__name__)
class PipReporter(BaseReporter):
def __init__(self):
# type: () -> None
self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int]
self._messages_at_backtrack = {
1: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
8: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
13: (
"This is taking longer than usual. You might need to provide the "
"dependency resolver with stricter constraints to reduce runtime."
"If you want to abort this run, you can press Ctrl + C to do so."
"To improve how pip performs, tell us that this happened here: "
"https://pip.pypa.io/surveys/backtracking"
)
}
def backtracking(self, candidate):
# type: (Candidate) -> None
self.backtracks_by_package[candidate.name] += 1
count = self.backtracks_by_package[candidate.name]
if count not in self._messages_at_backtrack:
return
message = self._messages_at_backtrack[count]
logger.info("INFO: %s", message)
| from collections import defaultdict
from logging import getLogger
from pip._vendor.resolvelib.reporters import BaseReporter
from pip._internal.utils.typing import MYPY_CHECK_RUNNING
if MYPY_CHECK_RUNNING:
from typing import DefaultDict
from .base import Candidate
logger = getLogger(__name__)
class PipReporter(BaseReporter):
def __init__(self):
# type: () -> None
self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int]
self._messages_at_backtrack = {
1: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
8: (
"pip is looking at multiple versions of this package to determine "
"which version is compatible with other requirements. "
"This could take a while."
),
13: (
"This is taking longer than usual. You might need to provide the "
"dependency resolver with stricter constraints to reduce runtime."
"If you want to abort this run, you can press Ctrl + C to do so."
+ "To improve how pip performs, tell us that this happened here: "
+ "https://pip.pypa.io/surveys/backtracking"
)
}
def backtracking(self, candidate):
# type: (Candidate) -> None
self.backtracks_by_package[candidate.name] += 1
count = self.backtracks_by_package[candidate.name]
if count not in self._messages_at_backtrack:
return
message = self._messages_at_backtrack[count]
logger.info("INFO: %s", message) |
d81a6930d21262464ee06ae8afb51b65920f378c | tap/tests/test_pytest_plugin.py | tap/tests/test_pytest_plugin.py |
try:
from unittest import mock
except ImportError:
import mock
from tap.plugins import pytest
from tap.tests import TestCase
from tap.tracker import Tracker
class TestPytestPlugin(TestCase):
def setUp(self):
"""The pytest plugin uses module scope so a fresh tracker
must be installed each time."""
pytest.tracker = Tracker()
def test_includes_options(self):
group = mock.Mock()
parser = mock.Mock()
parser.getgroup.return_value = group
pytest.pytest_addoption(parser)
self.assertEqual(group.addoption.call_count, 1)
def test_tracker_outdir_set(self):
config = mock.Mock()
config.option.tap_outdir = 'fakeout'
pytest.pytest_configure(config)
self.assertEqual(pytest.tracker.outdir, 'fakeout')
|
try:
from unittest import mock
except ImportError:
import mock
import tempfile
from tap.plugins import pytest
from tap.tests import TestCase
from tap.tracker import Tracker
class TestPytestPlugin(TestCase):
def setUp(self):
"""The pytest plugin uses module scope so a fresh tracker
must be installed each time."""
pytest.tracker = Tracker()
def test_includes_options(self):
group = mock.Mock()
parser = mock.Mock()
parser.getgroup.return_value = group
pytest.pytest_addoption(parser)
self.assertEqual(group.addoption.call_count, 1)
def test_tracker_outdir_set(self):
outdir = tempfile.mkdtemp()
config = mock.Mock()
config.option.tap_outdir = outdir
pytest.pytest_configure(config)
self.assertEqual(pytest.tracker.outdir, outdir)
| Fix test to not create a new directory in the project. | Fix test to not create a new directory in the project.
| Python | bsd-2-clause | mblayman/tappy,python-tap/tappy,Mark-E-Hamilton/tappy |
try:
from unittest import mock
except ImportError:
import mock
+ import tempfile
from tap.plugins import pytest
from tap.tests import TestCase
from tap.tracker import Tracker
class TestPytestPlugin(TestCase):
def setUp(self):
"""The pytest plugin uses module scope so a fresh tracker
must be installed each time."""
pytest.tracker = Tracker()
def test_includes_options(self):
group = mock.Mock()
parser = mock.Mock()
parser.getgroup.return_value = group
pytest.pytest_addoption(parser)
self.assertEqual(group.addoption.call_count, 1)
def test_tracker_outdir_set(self):
+ outdir = tempfile.mkdtemp()
config = mock.Mock()
- config.option.tap_outdir = 'fakeout'
+ config.option.tap_outdir = outdir
pytest.pytest_configure(config)
- self.assertEqual(pytest.tracker.outdir, 'fakeout')
+ self.assertEqual(pytest.tracker.outdir, outdir)
| Fix test to not create a new directory in the project. | ## Code Before:
try:
from unittest import mock
except ImportError:
import mock
from tap.plugins import pytest
from tap.tests import TestCase
from tap.tracker import Tracker
class TestPytestPlugin(TestCase):
def setUp(self):
"""The pytest plugin uses module scope so a fresh tracker
must be installed each time."""
pytest.tracker = Tracker()
def test_includes_options(self):
group = mock.Mock()
parser = mock.Mock()
parser.getgroup.return_value = group
pytest.pytest_addoption(parser)
self.assertEqual(group.addoption.call_count, 1)
def test_tracker_outdir_set(self):
config = mock.Mock()
config.option.tap_outdir = 'fakeout'
pytest.pytest_configure(config)
self.assertEqual(pytest.tracker.outdir, 'fakeout')
## Instruction:
Fix test to not create a new directory in the project.
## Code After:
try:
from unittest import mock
except ImportError:
import mock
import tempfile
from tap.plugins import pytest
from tap.tests import TestCase
from tap.tracker import Tracker
class TestPytestPlugin(TestCase):
def setUp(self):
"""The pytest plugin uses module scope so a fresh tracker
must be installed each time."""
pytest.tracker = Tracker()
def test_includes_options(self):
group = mock.Mock()
parser = mock.Mock()
parser.getgroup.return_value = group
pytest.pytest_addoption(parser)
self.assertEqual(group.addoption.call_count, 1)
def test_tracker_outdir_set(self):
outdir = tempfile.mkdtemp()
config = mock.Mock()
config.option.tap_outdir = outdir
pytest.pytest_configure(config)
self.assertEqual(pytest.tracker.outdir, outdir)
|
try:
from unittest import mock
except ImportError:
import mock
+ import tempfile
from tap.plugins import pytest
from tap.tests import TestCase
from tap.tracker import Tracker
class TestPytestPlugin(TestCase):
def setUp(self):
"""The pytest plugin uses module scope so a fresh tracker
must be installed each time."""
pytest.tracker = Tracker()
def test_includes_options(self):
group = mock.Mock()
parser = mock.Mock()
parser.getgroup.return_value = group
pytest.pytest_addoption(parser)
self.assertEqual(group.addoption.call_count, 1)
def test_tracker_outdir_set(self):
+ outdir = tempfile.mkdtemp()
config = mock.Mock()
- config.option.tap_outdir = 'fakeout'
? ----- ^
+ config.option.tap_outdir = outdir
? ^^^
pytest.pytest_configure(config)
- self.assertEqual(pytest.tracker.outdir, 'fakeout')
? ----- ^
+ self.assertEqual(pytest.tracker.outdir, outdir)
? ^^^
|
b56c5ca12f9806ecedc531e1f00ec1d7f2162b46 | src-django/authentication/urls.py | src-django/authentication/urls.py | from django.conf.urls import url
from views import login, logout, signup
urlpatterns = [
url(r'^login', login),
url(r'^logout', logout),
url(r'^signup', signup),
]
| from django.conf.urls import url
from views import login, logout, signup, confirm_email
urlpatterns = [
url(r'^login', login),
url(r'^logout', logout),
url(r'^signup', signup),
url(r'^confirm_email/(?P<key>\w+)', confirm_email),
]
| Add an endpoint for email confirmation | Add an endpoint for email confirmation
| Python | bsd-3-clause | SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder | from django.conf.urls import url
- from views import login, logout, signup
+ from views import login, logout, signup, confirm_email
urlpatterns = [
url(r'^login', login),
url(r'^logout', logout),
url(r'^signup', signup),
+ url(r'^confirm_email/(?P<key>\w+)', confirm_email),
]
| Add an endpoint for email confirmation | ## Code Before:
from django.conf.urls import url
from views import login, logout, signup
urlpatterns = [
url(r'^login', login),
url(r'^logout', logout),
url(r'^signup', signup),
]
## Instruction:
Add an endpoint for email confirmation
## Code After:
from django.conf.urls import url
from views import login, logout, signup, confirm_email
urlpatterns = [
url(r'^login', login),
url(r'^logout', logout),
url(r'^signup', signup),
url(r'^confirm_email/(?P<key>\w+)', confirm_email),
]
| from django.conf.urls import url
- from views import login, logout, signup
+ from views import login, logout, signup, confirm_email
? +++++++++++++++
urlpatterns = [
url(r'^login', login),
url(r'^logout', logout),
url(r'^signup', signup),
+ url(r'^confirm_email/(?P<key>\w+)', confirm_email),
] |
943699de02c3d8f4f8e26370fbbff2dec8a2d5ea | api/identifiers/urls.py | api/identifiers/urls.py | from django.conf.urls import url
from api.identifiers import views
urlpatterns = [
url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name),
]
| from django.conf.urls import url
from api.identifiers import views
urlpatterns = [
url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name),
url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name),
]
| Add identifier list to identifier views for use with embeds in registrations | Add identifier list to identifier views for use with embeds in registrations
[#OSF-6628]
| Python | apache-2.0 | saradbowman/osf.io,alexschiller/osf.io,wearpants/osf.io,erinspace/osf.io,alexschiller/osf.io,mluo613/osf.io,rdhyee/osf.io,icereval/osf.io,chrisseto/osf.io,mluo613/osf.io,chennan47/osf.io,emetsger/osf.io,hmoco/osf.io,hmoco/osf.io,hmoco/osf.io,baylee-d/osf.io,baylee-d/osf.io,sloria/osf.io,SSJohns/osf.io,erinspace/osf.io,leb2dg/osf.io,felliott/osf.io,brianjgeiger/osf.io,SSJohns/osf.io,alexschiller/osf.io,HalcyonChimera/osf.io,felliott/osf.io,Nesiehr/osf.io,baylee-d/osf.io,TomBaxter/osf.io,wearpants/osf.io,acshi/osf.io,cslzchen/osf.io,SSJohns/osf.io,leb2dg/osf.io,CenterForOpenScience/osf.io,laurenrevere/osf.io,cslzchen/osf.io,acshi/osf.io,wearpants/osf.io,monikagrabowska/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,Nesiehr/osf.io,binoculars/osf.io,adlius/osf.io,chrisseto/osf.io,crcresearch/osf.io,icereval/osf.io,emetsger/osf.io,chennan47/osf.io,SSJohns/osf.io,pattisdr/osf.io,samchrisinger/osf.io,rdhyee/osf.io,mattclark/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,caneruguz/osf.io,chrisseto/osf.io,caseyrollins/osf.io,brianjgeiger/osf.io,pattisdr/osf.io,acshi/osf.io,saradbowman/osf.io,felliott/osf.io,icereval/osf.io,cwisecarver/osf.io,laurenrevere/osf.io,monikagrabowska/osf.io,mfraezz/osf.io,mattclark/osf.io,rdhyee/osf.io,caneruguz/osf.io,acshi/osf.io,amyshi188/osf.io,DanielSBrown/osf.io,amyshi188/osf.io,adlius/osf.io,adlius/osf.io,amyshi188/osf.io,DanielSBrown/osf.io,cwisecarver/osf.io,HalcyonChimera/osf.io,erinspace/osf.io,sloria/osf.io,rdhyee/osf.io,amyshi188/osf.io,leb2dg/osf.io,cslzchen/osf.io,mfraezz/osf.io,binoculars/osf.io,mfraezz/osf.io,cwisecarver/osf.io,leb2dg/osf.io,pattisdr/osf.io,cslzchen/osf.io,adlius/osf.io,emetsger/osf.io,aaxelb/osf.io,DanielSBrown/osf.io,cwisecarver/osf.io,emetsger/osf.io,mattclark/osf.io,Johnetordoff/osf.io,aaxelb/osf.io,alexschiller/osf.io,crcresearch/osf.io,brianjgeiger/osf.io,mluo613/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,chennan47/osf.io,CenterForOpenScience/osf.io,Nesiehr/osf.io,CenterForOpenScience/osf.io,binoculars/osf.io,monikagrabowska/osf.io,caseyrollins/osf.io,samchrisinger/osf.io,Nesiehr/osf.io,caseyrollins/osf.io,alexschiller/osf.io,TomBaxter/osf.io,mluo613/osf.io,samchrisinger/osf.io,crcresearch/osf.io,TomBaxter/osf.io,Johnetordoff/osf.io,DanielSBrown/osf.io,mluo613/osf.io,aaxelb/osf.io,felliott/osf.io,laurenrevere/osf.io,CenterForOpenScience/osf.io,HalcyonChimera/osf.io,mfraezz/osf.io,chrisseto/osf.io,acshi/osf.io,caneruguz/osf.io,aaxelb/osf.io,samchrisinger/osf.io,sloria/osf.io,wearpants/osf.io | from django.conf.urls import url
from api.identifiers import views
urlpatterns = [
url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name),
+ url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name),
]
| Add identifier list to identifier views for use with embeds in registrations | ## Code Before:
from django.conf.urls import url
from api.identifiers import views
urlpatterns = [
url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name),
]
## Instruction:
Add identifier list to identifier views for use with embeds in registrations
## Code After:
from django.conf.urls import url
from api.identifiers import views
urlpatterns = [
url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name),
url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name),
]
| from django.conf.urls import url
from api.identifiers import views
urlpatterns = [
url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name),
+ url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name),
] |
a6ae4171de33dd77e9109523380c1330d4037f9f | gengine/app/tests/runner.py | gengine/app/tests/runner.py | from gengine.app.tests import db as db
from gengine.metadata import init_declarative_base, init_session
import unittest
import os
import pkgutil
import testing.redis
import logging
log = logging.getLogger(__name__)
init_session()
init_declarative_base()
__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]
def create_test_suite():
suite = unittest.TestSuite()
for imp, modname, _ in pkgutil.walk_packages(__path__):
mod = imp.find_module(modname).load_module(modname)
for test in unittest.defaultTestLoader.loadTestsFromModule(mod):
suite.addTests(test)
return suite
if __name__=="__main__":
exit = 1
try:
redis = testing.redis.RedisServer()
from gengine.base.cache import setup_redis_cache
dsn = redis.dsn()
setup_redis_cache(dsn["host"], dsn["port"], dsn["db"])
from gengine.app.cache import init_caches
init_caches()
db.setupDB()
testSuite = create_test_suite()
text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)
if text_runner.wasSuccessful():
exit = 0
finally:
try:
db.unsetupDB()
except:
log.exception()
try:
redis.stop()
except:
log.exception()
sys.exit(exit)
| from gengine.app.tests import db as db
from gengine.metadata import init_declarative_base, init_session
import unittest
import os
import pkgutil
import testing.redis
import logging
import sys
log = logging.getLogger(__name__)
init_session()
init_declarative_base()
__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]
def create_test_suite():
suite = unittest.TestSuite()
for imp, modname, _ in pkgutil.walk_packages(__path__):
mod = imp.find_module(modname).load_module(modname)
for test in unittest.defaultTestLoader.loadTestsFromModule(mod):
suite.addTests(test)
return suite
if __name__=="__main__":
exit = 1
try:
redis = testing.redis.RedisServer()
from gengine.base.cache import setup_redis_cache
dsn = redis.dsn()
setup_redis_cache(dsn["host"], dsn["port"], dsn["db"])
from gengine.app.cache import init_caches
init_caches()
db.setupDB()
testSuite = create_test_suite()
text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)
if text_runner.wasSuccessful():
exit = 0
finally:
try:
db.unsetupDB()
except:
log.exception()
try:
redis.stop()
except:
log.exception()
sys.exit(exit)
| Add missing import for sys | Add missing import for sys | Python | mit | ActiDoo/gamification-engine,ActiDoo/gamification-engine,ActiDoo/gamification-engine,ActiDoo/gamification-engine | from gengine.app.tests import db as db
from gengine.metadata import init_declarative_base, init_session
import unittest
import os
import pkgutil
import testing.redis
import logging
+ import sys
log = logging.getLogger(__name__)
init_session()
init_declarative_base()
__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]
def create_test_suite():
suite = unittest.TestSuite()
for imp, modname, _ in pkgutil.walk_packages(__path__):
mod = imp.find_module(modname).load_module(modname)
for test in unittest.defaultTestLoader.loadTestsFromModule(mod):
suite.addTests(test)
return suite
if __name__=="__main__":
exit = 1
try:
redis = testing.redis.RedisServer()
from gengine.base.cache import setup_redis_cache
dsn = redis.dsn()
setup_redis_cache(dsn["host"], dsn["port"], dsn["db"])
from gengine.app.cache import init_caches
init_caches()
db.setupDB()
testSuite = create_test_suite()
text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)
if text_runner.wasSuccessful():
exit = 0
finally:
try:
db.unsetupDB()
except:
log.exception()
try:
redis.stop()
except:
log.exception()
sys.exit(exit)
| Add missing import for sys | ## Code Before:
from gengine.app.tests import db as db
from gengine.metadata import init_declarative_base, init_session
import unittest
import os
import pkgutil
import testing.redis
import logging
log = logging.getLogger(__name__)
init_session()
init_declarative_base()
__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]
def create_test_suite():
suite = unittest.TestSuite()
for imp, modname, _ in pkgutil.walk_packages(__path__):
mod = imp.find_module(modname).load_module(modname)
for test in unittest.defaultTestLoader.loadTestsFromModule(mod):
suite.addTests(test)
return suite
if __name__=="__main__":
exit = 1
try:
redis = testing.redis.RedisServer()
from gengine.base.cache import setup_redis_cache
dsn = redis.dsn()
setup_redis_cache(dsn["host"], dsn["port"], dsn["db"])
from gengine.app.cache import init_caches
init_caches()
db.setupDB()
testSuite = create_test_suite()
text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)
if text_runner.wasSuccessful():
exit = 0
finally:
try:
db.unsetupDB()
except:
log.exception()
try:
redis.stop()
except:
log.exception()
sys.exit(exit)
## Instruction:
Add missing import for sys
## Code After:
from gengine.app.tests import db as db
from gengine.metadata import init_declarative_base, init_session
import unittest
import os
import pkgutil
import testing.redis
import logging
import sys
log = logging.getLogger(__name__)
init_session()
init_declarative_base()
__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]
def create_test_suite():
suite = unittest.TestSuite()
for imp, modname, _ in pkgutil.walk_packages(__path__):
mod = imp.find_module(modname).load_module(modname)
for test in unittest.defaultTestLoader.loadTestsFromModule(mod):
suite.addTests(test)
return suite
if __name__=="__main__":
exit = 1
try:
redis = testing.redis.RedisServer()
from gengine.base.cache import setup_redis_cache
dsn = redis.dsn()
setup_redis_cache(dsn["host"], dsn["port"], dsn["db"])
from gengine.app.cache import init_caches
init_caches()
db.setupDB()
testSuite = create_test_suite()
text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)
if text_runner.wasSuccessful():
exit = 0
finally:
try:
db.unsetupDB()
except:
log.exception()
try:
redis.stop()
except:
log.exception()
sys.exit(exit)
| from gengine.app.tests import db as db
from gengine.metadata import init_declarative_base, init_session
import unittest
import os
import pkgutil
import testing.redis
import logging
+ import sys
log = logging.getLogger(__name__)
init_session()
init_declarative_base()
__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]
def create_test_suite():
suite = unittest.TestSuite()
for imp, modname, _ in pkgutil.walk_packages(__path__):
mod = imp.find_module(modname).load_module(modname)
for test in unittest.defaultTestLoader.loadTestsFromModule(mod):
suite.addTests(test)
return suite
if __name__=="__main__":
exit = 1
try:
redis = testing.redis.RedisServer()
from gengine.base.cache import setup_redis_cache
dsn = redis.dsn()
setup_redis_cache(dsn["host"], dsn["port"], dsn["db"])
from gengine.app.cache import init_caches
init_caches()
db.setupDB()
testSuite = create_test_suite()
text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)
if text_runner.wasSuccessful():
exit = 0
finally:
try:
db.unsetupDB()
except:
log.exception()
try:
redis.stop()
except:
log.exception()
sys.exit(exit) |
32410e639f3202c10d9c75083319a9ab81932b82 | client/api.py | client/api.py | import os
import sys
import json
import urllib
import urllib2
import cookielib
sys.path.append((os.path.dirname(__file__) or ".") + "/../")
import config
cj = cookielib.CookieJar()
def callapi(action, postdata={}):
postdata.update({"action": action})
opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
opener.addheaders = [("User-Agent", "pyWebCash Scraper")]
f = opener.open(config.apiurl,urllib.urlencode(postdata))
data = f.read()
return json.loads(data)
| import os
import sys
import json
import urllib
import httplib
import urllib2
import cookielib
sys.path.append((os.path.dirname(__file__) or ".") + "/../")
import config
cj = cookielib.CookieJar()
class HTTPSClientAuthHandler(urllib2.HTTPSHandler):
def __init__(self, key):
urllib2.HTTPSHandler.__init__(self)
self.key = key
def https_open(self, req):
# Rather than pass in a reference to a connection class, we pass in
# a reference to a function which, for all intents and purposes,
# will behave as a constructor
return self.do_open(self.getConnection, req)
def getConnection(self, host, timeout=300):
return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key)
def callapi(action, postdata={}):
postdata.update({"action": action})
if config.certfile:
opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj))
else:
opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
opener.addheaders = [("User-Agent", "pyWebCash Scraper")]
f = opener.open(config.apiurl,urllib.urlencode(postdata))
data = f.read()
return json.loads(data)
| Use ssl client cert if given. | Use ssl client cert if given. | Python | agpl-3.0 | vincebusam/pyWebCash,vincebusam/pyWebCash,vincebusam/pyWebCash | import os
import sys
import json
import urllib
+ import httplib
import urllib2
import cookielib
sys.path.append((os.path.dirname(__file__) or ".") + "/../")
import config
cj = cookielib.CookieJar()
+ class HTTPSClientAuthHandler(urllib2.HTTPSHandler):
+ def __init__(self, key):
+ urllib2.HTTPSHandler.__init__(self)
+ self.key = key
+
+ def https_open(self, req):
+ # Rather than pass in a reference to a connection class, we pass in
+ # a reference to a function which, for all intents and purposes,
+ # will behave as a constructor
+ return self.do_open(self.getConnection, req)
+
+ def getConnection(self, host, timeout=300):
+ return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key)
+
def callapi(action, postdata={}):
postdata.update({"action": action})
+ if config.certfile:
+ opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj))
+ else:
- opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
+ opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
opener.addheaders = [("User-Agent", "pyWebCash Scraper")]
f = opener.open(config.apiurl,urllib.urlencode(postdata))
data = f.read()
return json.loads(data)
| Use ssl client cert if given. | ## Code Before:
import os
import sys
import json
import urllib
import urllib2
import cookielib
sys.path.append((os.path.dirname(__file__) or ".") + "/../")
import config
cj = cookielib.CookieJar()
def callapi(action, postdata={}):
postdata.update({"action": action})
opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
opener.addheaders = [("User-Agent", "pyWebCash Scraper")]
f = opener.open(config.apiurl,urllib.urlencode(postdata))
data = f.read()
return json.loads(data)
## Instruction:
Use ssl client cert if given.
## Code After:
import os
import sys
import json
import urllib
import httplib
import urllib2
import cookielib
sys.path.append((os.path.dirname(__file__) or ".") + "/../")
import config
cj = cookielib.CookieJar()
class HTTPSClientAuthHandler(urllib2.HTTPSHandler):
def __init__(self, key):
urllib2.HTTPSHandler.__init__(self)
self.key = key
def https_open(self, req):
# Rather than pass in a reference to a connection class, we pass in
# a reference to a function which, for all intents and purposes,
# will behave as a constructor
return self.do_open(self.getConnection, req)
def getConnection(self, host, timeout=300):
return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key)
def callapi(action, postdata={}):
postdata.update({"action": action})
if config.certfile:
opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj))
else:
opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
opener.addheaders = [("User-Agent", "pyWebCash Scraper")]
f = opener.open(config.apiurl,urllib.urlencode(postdata))
data = f.read()
return json.loads(data)
| import os
import sys
import json
import urllib
+ import httplib
import urllib2
import cookielib
sys.path.append((os.path.dirname(__file__) or ".") + "/../")
import config
cj = cookielib.CookieJar()
+ class HTTPSClientAuthHandler(urllib2.HTTPSHandler):
+ def __init__(self, key):
+ urllib2.HTTPSHandler.__init__(self)
+ self.key = key
+
+ def https_open(self, req):
+ # Rather than pass in a reference to a connection class, we pass in
+ # a reference to a function which, for all intents and purposes,
+ # will behave as a constructor
+ return self.do_open(self.getConnection, req)
+
+ def getConnection(self, host, timeout=300):
+ return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key)
+
def callapi(action, postdata={}):
postdata.update({"action": action})
+ if config.certfile:
+ opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj))
+ else:
- opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
+ opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj))
? ++++
opener.addheaders = [("User-Agent", "pyWebCash Scraper")]
f = opener.open(config.apiurl,urllib.urlencode(postdata))
data = f.read()
return json.loads(data) |
06645a637c0d34270f88f9a6b96133da5c415dd7 | froide/publicbody/admin.py | froide/publicbody/admin.py | from django.contrib import admin
from froide.publicbody.models import PublicBody, FoiLaw
class PublicBodyAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("geography", "name",)}
list_display = ('name', 'classification', 'geography')
list_filter = ('classification',)
search_fields = ['name', "description"]
exclude = ('confirmed',)
class FoiLawAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("jurisdiction", "name",)}
admin.site.register(PublicBody, PublicBodyAdmin)
admin.site.register(FoiLaw, FoiLawAdmin)
| from django.contrib import admin
from froide.publicbody.models import PublicBody, FoiLaw
class PublicBodyAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("geography", "name",)}
list_display = ('name', 'classification', 'topic', 'geography')
list_filter = ('classification', 'topic',)
search_fields = ['name', "description"]
exclude = ('confirmed',)
class FoiLawAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("jurisdiction", "name",)}
admin.site.register(PublicBody, PublicBodyAdmin)
admin.site.register(FoiLaw, FoiLawAdmin)
| Add topic to PublicBodyAdmin list_filter and list_display | Add topic to PublicBodyAdmin list_filter and list_display | Python | mit | catcosmo/froide,ryankanno/froide,stefanw/froide,catcosmo/froide,fin/froide,CodeforHawaii/froide,fin/froide,okfse/froide,LilithWittmann/froide,okfse/froide,stefanw/froide,ryankanno/froide,LilithWittmann/froide,LilithWittmann/froide,CodeforHawaii/froide,fin/froide,catcosmo/froide,catcosmo/froide,okfse/froide,ryankanno/froide,okfse/froide,catcosmo/froide,CodeforHawaii/froide,ryankanno/froide,LilithWittmann/froide,fin/froide,LilithWittmann/froide,stefanw/froide,stefanw/froide,stefanw/froide,ryankanno/froide,CodeforHawaii/froide,CodeforHawaii/froide,okfse/froide | from django.contrib import admin
from froide.publicbody.models import PublicBody, FoiLaw
class PublicBodyAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("geography", "name",)}
- list_display = ('name', 'classification', 'geography')
+ list_display = ('name', 'classification', 'topic', 'geography')
- list_filter = ('classification',)
+ list_filter = ('classification', 'topic',)
search_fields = ['name', "description"]
exclude = ('confirmed',)
class FoiLawAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("jurisdiction", "name",)}
admin.site.register(PublicBody, PublicBodyAdmin)
admin.site.register(FoiLaw, FoiLawAdmin)
| Add topic to PublicBodyAdmin list_filter and list_display | ## Code Before:
from django.contrib import admin
from froide.publicbody.models import PublicBody, FoiLaw
class PublicBodyAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("geography", "name",)}
list_display = ('name', 'classification', 'geography')
list_filter = ('classification',)
search_fields = ['name', "description"]
exclude = ('confirmed',)
class FoiLawAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("jurisdiction", "name",)}
admin.site.register(PublicBody, PublicBodyAdmin)
admin.site.register(FoiLaw, FoiLawAdmin)
## Instruction:
Add topic to PublicBodyAdmin list_filter and list_display
## Code After:
from django.contrib import admin
from froide.publicbody.models import PublicBody, FoiLaw
class PublicBodyAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("geography", "name",)}
list_display = ('name', 'classification', 'topic', 'geography')
list_filter = ('classification', 'topic',)
search_fields = ['name', "description"]
exclude = ('confirmed',)
class FoiLawAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("jurisdiction", "name",)}
admin.site.register(PublicBody, PublicBodyAdmin)
admin.site.register(FoiLaw, FoiLawAdmin)
| from django.contrib import admin
from froide.publicbody.models import PublicBody, FoiLaw
class PublicBodyAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("geography", "name",)}
- list_display = ('name', 'classification', 'geography')
+ list_display = ('name', 'classification', 'topic', 'geography')
? +++++++++
- list_filter = ('classification',)
+ list_filter = ('classification', 'topic',)
? +++++++++
search_fields = ['name', "description"]
exclude = ('confirmed',)
class FoiLawAdmin(admin.ModelAdmin):
prepopulated_fields = {"slug": ("jurisdiction", "name",)}
admin.site.register(PublicBody, PublicBodyAdmin)
admin.site.register(FoiLaw, FoiLawAdmin) |
c27a1fc4c0251b896667e21a0a88fb44a403242f | cistern/migrations.py | cistern/migrations.py | import os
from playhouse.migrate import *
cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
migrator = SqliteMigrator(db)
date_added = DateTimeField(default=None)
migrate(
migrator.add_column('torrent', 'date_added', date_added)
)
| import datetime
import os
from playhouse.migrate import *
def update():
cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
migrator = SqliteMigrator(db)
date_added = DateTimeField(default=datetime.datetime.now)
migrate(
migrator.add_column('torrent', 'date_added', date_added)
)
| Move migration to a function | Move migration to a function
| Python | mit | archangelic/cistern | + import datetime
import os
from playhouse.migrate import *
+ def update():
- cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
+ cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
- db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
+ db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
- migrator = SqliteMigrator(db)
+ migrator = SqliteMigrator(db)
- date_added = DateTimeField(default=None)
+ date_added = DateTimeField(default=datetime.datetime.now)
- migrate(
+ migrate(
- migrator.add_column('torrent', 'date_added', date_added)
+ migrator.add_column('torrent', 'date_added', date_added)
- )
+ )
| Move migration to a function | ## Code Before:
import os
from playhouse.migrate import *
cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
migrator = SqliteMigrator(db)
date_added = DateTimeField(default=None)
migrate(
migrator.add_column('torrent', 'date_added', date_added)
)
## Instruction:
Move migration to a function
## Code After:
import datetime
import os
from playhouse.migrate import *
def update():
cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
migrator = SqliteMigrator(db)
date_added = DateTimeField(default=datetime.datetime.now)
migrate(
migrator.add_column('torrent', 'date_added', date_added)
)
| + import datetime
import os
from playhouse.migrate import *
+ def update():
- cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
+ cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern'))
? ++++
- db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
+ db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db'))
? ++++
- migrator = SqliteMigrator(db)
+ migrator = SqliteMigrator(db)
? ++++
- date_added = DateTimeField(default=None)
+ date_added = DateTimeField(default=datetime.datetime.now)
- migrate(
+ migrate(
? ++++
- migrator.add_column('torrent', 'date_added', date_added)
+ migrator.add_column('torrent', 'date_added', date_added)
? ++++
- )
+ ) |
6fe5a416ed229e7ec8efab9d6b3dac43f16515b6 | corehq/apps/domain/__init__.py | corehq/apps/domain/__init__.py | from corehq.preindex import ExtraPreindexPlugin
from django.conf import settings
ExtraPreindexPlugin.register('domain', __file__, (
settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta'))
| from corehq.preindex import ExtraPreindexPlugin
from django.conf import settings
ExtraPreindexPlugin.register('domain', __file__, (
settings.NEW_DOMAINS_DB,
settings.NEW_USERS_GROUPS_DB,
settings.NEW_FIXTURES_DB,
'meta',
))
| Add the new domains db | Add the new domains db
| Python | bsd-3-clause | dimagi/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq | from corehq.preindex import ExtraPreindexPlugin
from django.conf import settings
ExtraPreindexPlugin.register('domain', __file__, (
- settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta'))
+ settings.NEW_DOMAINS_DB,
+ settings.NEW_USERS_GROUPS_DB,
+ settings.NEW_FIXTURES_DB,
+ 'meta',
+ ))
| Add the new domains db | ## Code Before:
from corehq.preindex import ExtraPreindexPlugin
from django.conf import settings
ExtraPreindexPlugin.register('domain', __file__, (
settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta'))
## Instruction:
Add the new domains db
## Code After:
from corehq.preindex import ExtraPreindexPlugin
from django.conf import settings
ExtraPreindexPlugin.register('domain', __file__, (
settings.NEW_DOMAINS_DB,
settings.NEW_USERS_GROUPS_DB,
settings.NEW_FIXTURES_DB,
'meta',
))
| from corehq.preindex import ExtraPreindexPlugin
from django.conf import settings
ExtraPreindexPlugin.register('domain', __file__, (
- settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta'))
+ settings.NEW_DOMAINS_DB,
+ settings.NEW_USERS_GROUPS_DB,
+ settings.NEW_FIXTURES_DB,
+ 'meta',
+ )) |
a52c84f092d89f89130c2696c98779e955f083dc | tests/test_version_parser.py | tests/test_version_parser.py | import pytest
from leak.version_parser import versions_split
def test_versions_split():
pass
def test_wrong_versions_split():
# too many dots
assert versions_split('1.2.3.4') == [0, 0, 0]
# test missing numeric version
with pytest.raises(ValueError):
versions_split('not.numeric')
# test not string provided
with pytest.raises(AttributeError):
versions_split(12345)
| import pytest
from leak.version_parser import versions_split
def test_versions_split():
assert versions_split('1.8.1') == [1, 8, 1]
assert versions_split('1.4') == [1, 4, 0]
assert versions_split('2') == [2, 0, 0]
def test_versions_split_str_mapping():
assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0']
assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0']
assert versions_split('text', type_applyer=str) == ['text', '0', '0']
def test_wrong_versions_split():
# too many dots
assert versions_split('1.2.3.4') == [0, 0, 0]
# test missing numeric version
with pytest.raises(ValueError):
versions_split('not.numeric')
# test not string provided
with pytest.raises(AttributeError):
versions_split(12345)
| Add tests for version splitting | Add tests for version splitting
| Python | mit | bmwant21/leak | import pytest
from leak.version_parser import versions_split
def test_versions_split():
- pass
+ assert versions_split('1.8.1') == [1, 8, 1]
+ assert versions_split('1.4') == [1, 4, 0]
+ assert versions_split('2') == [2, 0, 0]
+
+
+ def test_versions_split_str_mapping():
+ assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0']
+ assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0']
+ assert versions_split('text', type_applyer=str) == ['text', '0', '0']
def test_wrong_versions_split():
# too many dots
assert versions_split('1.2.3.4') == [0, 0, 0]
# test missing numeric version
with pytest.raises(ValueError):
versions_split('not.numeric')
# test not string provided
with pytest.raises(AttributeError):
versions_split(12345)
| Add tests for version splitting | ## Code Before:
import pytest
from leak.version_parser import versions_split
def test_versions_split():
pass
def test_wrong_versions_split():
# too many dots
assert versions_split('1.2.3.4') == [0, 0, 0]
# test missing numeric version
with pytest.raises(ValueError):
versions_split('not.numeric')
# test not string provided
with pytest.raises(AttributeError):
versions_split(12345)
## Instruction:
Add tests for version splitting
## Code After:
import pytest
from leak.version_parser import versions_split
def test_versions_split():
assert versions_split('1.8.1') == [1, 8, 1]
assert versions_split('1.4') == [1, 4, 0]
assert versions_split('2') == [2, 0, 0]
def test_versions_split_str_mapping():
assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0']
assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0']
assert versions_split('text', type_applyer=str) == ['text', '0', '0']
def test_wrong_versions_split():
# too many dots
assert versions_split('1.2.3.4') == [0, 0, 0]
# test missing numeric version
with pytest.raises(ValueError):
versions_split('not.numeric')
# test not string provided
with pytest.raises(AttributeError):
versions_split(12345)
| import pytest
from leak.version_parser import versions_split
def test_versions_split():
- pass
+ assert versions_split('1.8.1') == [1, 8, 1]
+ assert versions_split('1.4') == [1, 4, 0]
+ assert versions_split('2') == [2, 0, 0]
+
+
+ def test_versions_split_str_mapping():
+ assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0']
+ assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0']
+ assert versions_split('text', type_applyer=str) == ['text', '0', '0']
def test_wrong_versions_split():
# too many dots
assert versions_split('1.2.3.4') == [0, 0, 0]
# test missing numeric version
with pytest.raises(ValueError):
versions_split('not.numeric')
# test not string provided
with pytest.raises(AttributeError):
versions_split(12345) |
4bc31e675659af54ee26fe5df16a0ee3ebeb5947 | firefed/__main__.py | firefed/__main__.py | import argparse
import os
import re
from firefed import Firefed
from feature import feature_map, Summary
def feature_type(val):
try:
return feature_map()[val]
except KeyError as key:
raise argparse.ArgumentTypeError(
'Feature %s not found. Choose from: {%s}' %
(key, ', '.join(feature_map())))
def profile_dir(dirname):
if dirname is None:
dirname = 'default'
if os.path.isdir(dirname):
return dirname
if re.match('^[\\w-]+$', dirname):
home = os.path.expanduser('~/.mozilla/firefox')
profile_names = os.listdir(home)
for name in profile_names:
if name.endswith('.%s' % dirname):
return os.path.join(home, name)
raise argparse.ArgumentTypeError('Profile %s not found.' % dirname)
def main():
parser = argparse.ArgumentParser(
'firefed',
description=
'Firefed is a Firefox profile analyzer focusing on privacy and security.',
)
parser.add_argument(
'-p',
'--profile',
help='profile name or directory',
type=profile_dir,
required=True)
parser.add_argument(
'-f',
'--feature',
type=feature_type,
default=Summary,
help='{%s}' % ', '.join(feature_map()))
parser.add_argument(
'-s', '--summarize', action='store_true', help='summarize results')
args = parser.parse_args()
Firefed(args)
if __name__ == '__main__':
main()
| import argparse
import os
import re
from firefed import Firefed
from feature import feature_map, Summary
def feature_type(val):
try:
return feature_map()[val]
except KeyError as key:
raise argparse.ArgumentTypeError(
'Feature %s not found. Choose from: {%s}' %
(key, ', '.join(feature_map())))
def profile_dir(dirname):
if dirname is None:
dirname = 'default'
if os.path.isdir(dirname):
return dirname
if re.match('^[\\w-]+$', dirname):
home = os.path.expanduser('~/.mozilla/firefox')
profile_names = os.listdir(home)
for name in profile_names:
if name.endswith('.%s' % dirname):
return os.path.join(home, name)
raise argparse.ArgumentTypeError('Profile %s not found.' % dirname)
def main():
parser = argparse.ArgumentParser(
'firefed',
description=
'Firefed is a Firefox profile analyzer focusing on privacy and security.',
)
parser.add_argument(
'-p',
'--profile',
help='profile name or directory',
type=profile_dir,
default='default')
parser.add_argument(
'-f',
'--feature',
type=feature_type,
default=Summary,
help='{%s}' % ', '.join(feature_map()))
parser.add_argument(
'-s', '--summarize', action='store_true', help='summarize results')
args = parser.parse_args()
Firefed(args)
if __name__ == '__main__':
main()
| Add default argument for profile | Add default argument for profile
| Python | mit | numirias/firefed | import argparse
import os
import re
from firefed import Firefed
from feature import feature_map, Summary
def feature_type(val):
try:
return feature_map()[val]
except KeyError as key:
raise argparse.ArgumentTypeError(
'Feature %s not found. Choose from: {%s}' %
(key, ', '.join(feature_map())))
def profile_dir(dirname):
if dirname is None:
dirname = 'default'
if os.path.isdir(dirname):
return dirname
if re.match('^[\\w-]+$', dirname):
home = os.path.expanduser('~/.mozilla/firefox')
profile_names = os.listdir(home)
for name in profile_names:
if name.endswith('.%s' % dirname):
return os.path.join(home, name)
raise argparse.ArgumentTypeError('Profile %s not found.' % dirname)
def main():
parser = argparse.ArgumentParser(
'firefed',
description=
'Firefed is a Firefox profile analyzer focusing on privacy and security.',
)
parser.add_argument(
'-p',
'--profile',
help='profile name or directory',
type=profile_dir,
- required=True)
+ default='default')
parser.add_argument(
'-f',
'--feature',
type=feature_type,
default=Summary,
help='{%s}' % ', '.join(feature_map()))
parser.add_argument(
'-s', '--summarize', action='store_true', help='summarize results')
args = parser.parse_args()
Firefed(args)
if __name__ == '__main__':
main()
| Add default argument for profile | ## Code Before:
import argparse
import os
import re
from firefed import Firefed
from feature import feature_map, Summary
def feature_type(val):
try:
return feature_map()[val]
except KeyError as key:
raise argparse.ArgumentTypeError(
'Feature %s not found. Choose from: {%s}' %
(key, ', '.join(feature_map())))
def profile_dir(dirname):
if dirname is None:
dirname = 'default'
if os.path.isdir(dirname):
return dirname
if re.match('^[\\w-]+$', dirname):
home = os.path.expanduser('~/.mozilla/firefox')
profile_names = os.listdir(home)
for name in profile_names:
if name.endswith('.%s' % dirname):
return os.path.join(home, name)
raise argparse.ArgumentTypeError('Profile %s not found.' % dirname)
def main():
parser = argparse.ArgumentParser(
'firefed',
description=
'Firefed is a Firefox profile analyzer focusing on privacy and security.',
)
parser.add_argument(
'-p',
'--profile',
help='profile name or directory',
type=profile_dir,
required=True)
parser.add_argument(
'-f',
'--feature',
type=feature_type,
default=Summary,
help='{%s}' % ', '.join(feature_map()))
parser.add_argument(
'-s', '--summarize', action='store_true', help='summarize results')
args = parser.parse_args()
Firefed(args)
if __name__ == '__main__':
main()
## Instruction:
Add default argument for profile
## Code After:
import argparse
import os
import re
from firefed import Firefed
from feature import feature_map, Summary
def feature_type(val):
try:
return feature_map()[val]
except KeyError as key:
raise argparse.ArgumentTypeError(
'Feature %s not found. Choose from: {%s}' %
(key, ', '.join(feature_map())))
def profile_dir(dirname):
if dirname is None:
dirname = 'default'
if os.path.isdir(dirname):
return dirname
if re.match('^[\\w-]+$', dirname):
home = os.path.expanduser('~/.mozilla/firefox')
profile_names = os.listdir(home)
for name in profile_names:
if name.endswith('.%s' % dirname):
return os.path.join(home, name)
raise argparse.ArgumentTypeError('Profile %s not found.' % dirname)
def main():
parser = argparse.ArgumentParser(
'firefed',
description=
'Firefed is a Firefox profile analyzer focusing on privacy and security.',
)
parser.add_argument(
'-p',
'--profile',
help='profile name or directory',
type=profile_dir,
default='default')
parser.add_argument(
'-f',
'--feature',
type=feature_type,
default=Summary,
help='{%s}' % ', '.join(feature_map()))
parser.add_argument(
'-s', '--summarize', action='store_true', help='summarize results')
args = parser.parse_args()
Firefed(args)
if __name__ == '__main__':
main()
| import argparse
import os
import re
from firefed import Firefed
from feature import feature_map, Summary
def feature_type(val):
try:
return feature_map()[val]
except KeyError as key:
raise argparse.ArgumentTypeError(
'Feature %s not found. Choose from: {%s}' %
(key, ', '.join(feature_map())))
def profile_dir(dirname):
if dirname is None:
dirname = 'default'
if os.path.isdir(dirname):
return dirname
if re.match('^[\\w-]+$', dirname):
home = os.path.expanduser('~/.mozilla/firefox')
profile_names = os.listdir(home)
for name in profile_names:
if name.endswith('.%s' % dirname):
return os.path.join(home, name)
raise argparse.ArgumentTypeError('Profile %s not found.' % dirname)
def main():
parser = argparse.ArgumentParser(
'firefed',
description=
'Firefed is a Firefox profile analyzer focusing on privacy and security.',
)
parser.add_argument(
'-p',
'--profile',
help='profile name or directory',
type=profile_dir,
- required=True)
+ default='default')
parser.add_argument(
'-f',
'--feature',
type=feature_type,
default=Summary,
help='{%s}' % ', '.join(feature_map()))
parser.add_argument(
'-s', '--summarize', action='store_true', help='summarize results')
args = parser.parse_args()
Firefed(args)
if __name__ == '__main__':
main() |
dd19012ed8bb6ec702d84abe400bc3dec47044f3 | sortedm2m_tests/__init__.py | sortedm2m_tests/__init__.py | import os
# django-setuptest
import setuptest
TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support']
class TestSuite(setuptest.SetupTestSuite):
def resolve_packages(self):
packages = super(TestSuite, self).resolve_packages()
for test_app in TEST_APPS:
if test_app not in packages:
packages.append(test_app)
return packages
| import os
# django-setuptest
import setuptest
TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support']
class TestSuite(setuptest.SetupTestSuite):
def __init__(self, *args, **kwargs):
os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings'
from south.management.commands import patch_for_test_db_setup
patch_for_test_db_setup()
super(TestSuite, self).__init__(*args, **kwargs)
def resolve_packages(self):
packages = super(TestSuite, self).resolve_packages()
for test_app in TEST_APPS:
if test_app not in packages:
packages.append(test_app)
return packages
| Fix to allow tests with South migrations to run. | Fix to allow tests with South migrations to run.
| Python | bsd-3-clause | gregmuellegger/django-sortedm2m,fabrique/django-sortedm2m,gradel/django-sortedm2m,MathieuDuponchelle/django-sortedm2m,fabrique/django-sortedm2m,gradel/django-sortedm2m,fabrique/django-sortedm2m,gradel/django-sortedm2m,gregmuellegger/django-sortedm2m,MathieuDuponchelle/django-sortedm2m,gregmuellegger/django-sortedm2m | import os
# django-setuptest
import setuptest
TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support']
class TestSuite(setuptest.SetupTestSuite):
-
+
+ def __init__(self, *args, **kwargs):
+ os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings'
+ from south.management.commands import patch_for_test_db_setup
+ patch_for_test_db_setup()
+ super(TestSuite, self).__init__(*args, **kwargs)
+
def resolve_packages(self):
packages = super(TestSuite, self).resolve_packages()
for test_app in TEST_APPS:
if test_app not in packages:
packages.append(test_app)
return packages
+ | Fix to allow tests with South migrations to run. | ## Code Before:
import os
# django-setuptest
import setuptest
TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support']
class TestSuite(setuptest.SetupTestSuite):
def resolve_packages(self):
packages = super(TestSuite, self).resolve_packages()
for test_app in TEST_APPS:
if test_app not in packages:
packages.append(test_app)
return packages
## Instruction:
Fix to allow tests with South migrations to run.
## Code After:
import os
# django-setuptest
import setuptest
TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support']
class TestSuite(setuptest.SetupTestSuite):
def __init__(self, *args, **kwargs):
os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings'
from south.management.commands import patch_for_test_db_setup
patch_for_test_db_setup()
super(TestSuite, self).__init__(*args, **kwargs)
def resolve_packages(self):
packages = super(TestSuite, self).resolve_packages()
for test_app in TEST_APPS:
if test_app not in packages:
packages.append(test_app)
return packages
| import os
# django-setuptest
import setuptest
TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support']
class TestSuite(setuptest.SetupTestSuite):
-
+
+ def __init__(self, *args, **kwargs):
+ os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings'
+ from south.management.commands import patch_for_test_db_setup
+ patch_for_test_db_setup()
+ super(TestSuite, self).__init__(*args, **kwargs)
+
def resolve_packages(self):
packages = super(TestSuite, self).resolve_packages()
for test_app in TEST_APPS:
if test_app not in packages:
packages.append(test_app)
return packages
+
+ |
84b48b9be466ac72bddf5ee6288ff48be26eed62 | tests/classifier/RandomForestClassifier/RandomForestClassifierPHPTest.py | tests/classifier/RandomForestClassifier/RandomForestClassifierPHPTest.py |
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=100, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
@unittest.skip('The generated code would be too large.')
def test_existing_features_w_digits_data(self):
pass
@unittest.skip('The generated code would be too large.')
def test_random_features_w_digits_data(self):
pass
|
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=20, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
| Reduce the number of trees | Reduce the number of trees
| Python | bsd-3-clause | nok/sklearn-porter |
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
- self.mdl = RandomForestClassifier(n_estimators=100, random_state=0)
+ self.mdl = RandomForestClassifier(n_estimators=20, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
- @unittest.skip('The generated code would be too large.')
- def test_existing_features_w_digits_data(self):
- pass
-
- @unittest.skip('The generated code would be too large.')
- def test_random_features_w_digits_data(self):
- pass
- | Reduce the number of trees | ## Code Before:
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=100, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
@unittest.skip('The generated code would be too large.')
def test_existing_features_w_digits_data(self):
pass
@unittest.skip('The generated code would be too large.')
def test_random_features_w_digits_data(self):
pass
## Instruction:
Reduce the number of trees
## Code After:
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=20, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
|
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
- self.mdl = RandomForestClassifier(n_estimators=100, random_state=0)
? ^^
+ self.mdl = RandomForestClassifier(n_estimators=20, random_state=0)
? ^
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
-
- @unittest.skip('The generated code would be too large.')
- def test_existing_features_w_digits_data(self):
- pass
-
- @unittest.skip('The generated code would be too large.')
- def test_random_features_w_digits_data(self):
- pass |
9dad4f997371011ee7fe9f6ecd0c1a58cbba6d27 | html_parse.py | html_parse.py | from bs4 import BeautifulSoup
def parse(html):
soup = BeautifulSoup(html, features="html.parser")
return soup.get_text() | import imp
import logging
log = logging.getLogger(__name__)
def module_exists(module_name):
try:
imp.find_module(module_name)
return True
except ImportError:
return False
if module_exists("bs4"):
log.info("Parsing HTML using beautifulsoup4")
from bs4 import BeautifulSoup
def parse(html):
soup = BeautifulSoup(html, features="html.parser")
return soup.get_text()
elif module_exists("html2text"):
log.info("Parsing HTML using html2text")
import html2text
def parse(html):
h = html2text.HTML2Text()
h.single_line_break = True
return h.handle(html)
else:
warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text"
log.warning(warning_msg)
def parse(html):
raise ImportWarning(warning_msg) | Add support for html2text or no parser Will still prefer beautifulsoup4 if installed | Add support for html2text or no parser
Will still prefer beautifulsoup4 if installed
| Python | mit | idiotandrobot/heathergraph | - from bs4 import BeautifulSoup
+ import imp
+ import logging
+ log = logging.getLogger(__name__)
+ def module_exists(module_name):
+ try:
+ imp.find_module(module_name)
+ return True
+ except ImportError:
+ return False
+
+ if module_exists("bs4"):
+ log.info("Parsing HTML using beautifulsoup4")
+ from bs4 import BeautifulSoup
+
- def parse(html):
+ def parse(html):
- soup = BeautifulSoup(html, features="html.parser")
+ soup = BeautifulSoup(html, features="html.parser")
- return soup.get_text()
+ return soup.get_text()
+ elif module_exists("html2text"):
+ log.info("Parsing HTML using html2text")
+ import html2text
+
+ def parse(html):
+ h = html2text.HTML2Text()
+ h.single_line_break = True
+ return h.handle(html)
+ else:
+ warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text"
+ log.warning(warning_msg)
+ def parse(html):
+ raise ImportWarning(warning_msg) | Add support for html2text or no parser Will still prefer beautifulsoup4 if installed | ## Code Before:
from bs4 import BeautifulSoup
def parse(html):
soup = BeautifulSoup(html, features="html.parser")
return soup.get_text()
## Instruction:
Add support for html2text or no parser Will still prefer beautifulsoup4 if installed
## Code After:
import imp
import logging
log = logging.getLogger(__name__)
def module_exists(module_name):
try:
imp.find_module(module_name)
return True
except ImportError:
return False
if module_exists("bs4"):
log.info("Parsing HTML using beautifulsoup4")
from bs4 import BeautifulSoup
def parse(html):
soup = BeautifulSoup(html, features="html.parser")
return soup.get_text()
elif module_exists("html2text"):
log.info("Parsing HTML using html2text")
import html2text
def parse(html):
h = html2text.HTML2Text()
h.single_line_break = True
return h.handle(html)
else:
warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text"
log.warning(warning_msg)
def parse(html):
raise ImportWarning(warning_msg) | - from bs4 import BeautifulSoup
+ import imp
+ import logging
+ log = logging.getLogger(__name__)
+ def module_exists(module_name):
+ try:
+ imp.find_module(module_name)
+ return True
+ except ImportError:
+ return False
+
+ if module_exists("bs4"):
+ log.info("Parsing HTML using beautifulsoup4")
+ from bs4 import BeautifulSoup
+
- def parse(html):
+ def parse(html):
? ++++
- soup = BeautifulSoup(html, features="html.parser")
+ soup = BeautifulSoup(html, features="html.parser")
? ++++
- return soup.get_text()
+ return soup.get_text()
? ++++
+ elif module_exists("html2text"):
+ log.info("Parsing HTML using html2text")
+ import html2text
+
+ def parse(html):
+ h = html2text.HTML2Text()
+ h.single_line_break = True
+ return h.handle(html)
+ else:
+ warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text"
+ log.warning(warning_msg)
+ def parse(html):
+ raise ImportWarning(warning_msg) |
2dc56ab04ea17bea05654eaec12bb27b48b0b225 | robotd/cvcapture.py | robotd/cvcapture.py | import threading
from robotd.native import _cvcapture
class CaptureDevice(object):
def __init__(self, path=None):
if path is not None:
argument_c = _cvcapture.ffi.new(
'char[]',
path.encode('utf-8'),
)
else:
argument_c = _cvcapture.ffi.NULL
self.instance = _cvcapture.lib.cvopen(argument_c)
self.lock = threading.Lock()
def capture(self, width, height):
if self.instance is None:
raise RuntimeError("capture device is closed")
capture_buffer = _cvcapture.ffi.new(
'uint8_t[{}]'.format(width * height),
)
with self.lock:
status = _cvcapture.lib.cvcapture(
self.instance,
capture_buffer,
width,
height,
)
if status == 0:
raise RuntimeError("cvcapture() failed")
return bytes(_cvcapture.ffi.buffer(capture_buffer))
def __enter__(self):
return self
def __exit__(self, exc_value, exc_type, exc_traceback):
self.close()
def close(self):
if self.instance is not None:
with self.lock:
_cvcapture.lib.cvclose(self.instance)
self.instance = None
__del__ = close
| import threading
from robotd.native import _cvcapture
class CaptureDevice(object):
def __init__(self, path=None):
if path is not None:
argument_c = _cvcapture.ffi.new(
'char[]',
path.encode('utf-8'),
)
else:
argument_c = _cvcapture.ffi.NULL
self.instance = _cvcapture.lib.cvopen(argument_c)
if self.instance == _cvcapture.ffi.NULL:
raise RuntimeError("Unable to open capture device")
self.lock = threading.Lock()
def capture(self, width, height):
if self.instance is None:
raise RuntimeError("capture device is closed")
capture_buffer = _cvcapture.ffi.new(
'uint8_t[{}]'.format(width * height),
)
with self.lock:
status = _cvcapture.lib.cvcapture(
self.instance,
capture_buffer,
width,
height,
)
if status == 0:
raise RuntimeError("cvcapture() failed")
return bytes(_cvcapture.ffi.buffer(capture_buffer))
def __enter__(self):
return self
def __exit__(self, exc_value, exc_type, exc_traceback):
self.close()
def close(self):
if self.instance is not None:
with self.lock:
_cvcapture.lib.cvclose(self.instance)
self.instance = None
__del__ = close
| Raise a `RuntimeError` if the device cannot be opened | Raise a `RuntimeError` if the device cannot be opened
| Python | mit | sourcebots/robotd,sourcebots/robotd | import threading
from robotd.native import _cvcapture
class CaptureDevice(object):
def __init__(self, path=None):
if path is not None:
argument_c = _cvcapture.ffi.new(
'char[]',
path.encode('utf-8'),
)
else:
argument_c = _cvcapture.ffi.NULL
self.instance = _cvcapture.lib.cvopen(argument_c)
+ if self.instance == _cvcapture.ffi.NULL:
+ raise RuntimeError("Unable to open capture device")
self.lock = threading.Lock()
def capture(self, width, height):
if self.instance is None:
raise RuntimeError("capture device is closed")
capture_buffer = _cvcapture.ffi.new(
'uint8_t[{}]'.format(width * height),
)
with self.lock:
status = _cvcapture.lib.cvcapture(
self.instance,
capture_buffer,
width,
height,
)
if status == 0:
raise RuntimeError("cvcapture() failed")
return bytes(_cvcapture.ffi.buffer(capture_buffer))
def __enter__(self):
return self
def __exit__(self, exc_value, exc_type, exc_traceback):
self.close()
def close(self):
if self.instance is not None:
with self.lock:
_cvcapture.lib.cvclose(self.instance)
self.instance = None
__del__ = close
| Raise a `RuntimeError` if the device cannot be opened | ## Code Before:
import threading
from robotd.native import _cvcapture
class CaptureDevice(object):
def __init__(self, path=None):
if path is not None:
argument_c = _cvcapture.ffi.new(
'char[]',
path.encode('utf-8'),
)
else:
argument_c = _cvcapture.ffi.NULL
self.instance = _cvcapture.lib.cvopen(argument_c)
self.lock = threading.Lock()
def capture(self, width, height):
if self.instance is None:
raise RuntimeError("capture device is closed")
capture_buffer = _cvcapture.ffi.new(
'uint8_t[{}]'.format(width * height),
)
with self.lock:
status = _cvcapture.lib.cvcapture(
self.instance,
capture_buffer,
width,
height,
)
if status == 0:
raise RuntimeError("cvcapture() failed")
return bytes(_cvcapture.ffi.buffer(capture_buffer))
def __enter__(self):
return self
def __exit__(self, exc_value, exc_type, exc_traceback):
self.close()
def close(self):
if self.instance is not None:
with self.lock:
_cvcapture.lib.cvclose(self.instance)
self.instance = None
__del__ = close
## Instruction:
Raise a `RuntimeError` if the device cannot be opened
## Code After:
import threading
from robotd.native import _cvcapture
class CaptureDevice(object):
def __init__(self, path=None):
if path is not None:
argument_c = _cvcapture.ffi.new(
'char[]',
path.encode('utf-8'),
)
else:
argument_c = _cvcapture.ffi.NULL
self.instance = _cvcapture.lib.cvopen(argument_c)
if self.instance == _cvcapture.ffi.NULL:
raise RuntimeError("Unable to open capture device")
self.lock = threading.Lock()
def capture(self, width, height):
if self.instance is None:
raise RuntimeError("capture device is closed")
capture_buffer = _cvcapture.ffi.new(
'uint8_t[{}]'.format(width * height),
)
with self.lock:
status = _cvcapture.lib.cvcapture(
self.instance,
capture_buffer,
width,
height,
)
if status == 0:
raise RuntimeError("cvcapture() failed")
return bytes(_cvcapture.ffi.buffer(capture_buffer))
def __enter__(self):
return self
def __exit__(self, exc_value, exc_type, exc_traceback):
self.close()
def close(self):
if self.instance is not None:
with self.lock:
_cvcapture.lib.cvclose(self.instance)
self.instance = None
__del__ = close
| import threading
from robotd.native import _cvcapture
class CaptureDevice(object):
def __init__(self, path=None):
if path is not None:
argument_c = _cvcapture.ffi.new(
'char[]',
path.encode('utf-8'),
)
else:
argument_c = _cvcapture.ffi.NULL
self.instance = _cvcapture.lib.cvopen(argument_c)
+ if self.instance == _cvcapture.ffi.NULL:
+ raise RuntimeError("Unable to open capture device")
self.lock = threading.Lock()
def capture(self, width, height):
if self.instance is None:
raise RuntimeError("capture device is closed")
capture_buffer = _cvcapture.ffi.new(
'uint8_t[{}]'.format(width * height),
)
with self.lock:
status = _cvcapture.lib.cvcapture(
self.instance,
capture_buffer,
width,
height,
)
if status == 0:
raise RuntimeError("cvcapture() failed")
return bytes(_cvcapture.ffi.buffer(capture_buffer))
def __enter__(self):
return self
def __exit__(self, exc_value, exc_type, exc_traceback):
self.close()
def close(self):
if self.instance is not None:
with self.lock:
_cvcapture.lib.cvclose(self.instance)
self.instance = None
__del__ = close |
71df45002746b162e04a125403cad390accb949e | backend/main.py | backend/main.py | import logging
from firebase import firebase
from flask import Flask, jsonify, request
import flask_cors
from google.appengine.ext import ndb
import google.auth.transport.requests
import google.oauth2.id_token
import requests_toolbelt.adapters.appengine
requests_toolbelt.adapters.appengine.monkeypatch()
HTTP_REQUEST = google.auth.transport.requests.Request()
app = Flask(__name__)
firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None)
flask_cors.CORS(app)
@app.route('/')
def index():
return "<h1>Welcome To Google HVZ (backend)!</h1>"
@app.route('/test', methods=['GET'])
def get_testdata():
testdata = firebase.get('testdata', None)
return jsonify(testdata)
| import logging
from firebase import firebase
from flask import Flask, jsonify, request
import flask_cors
from google.appengine.ext import ndb
import google.auth.transport.requests
import google.oauth2.id_token
import requests_toolbelt.adapters.appengine
requests_toolbelt.adapters.appengine.monkeypatch()
HTTP_REQUEST = google.auth.transport.requests.Request()
# Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database
FIREBASE_SECRET = ""
FIREBASE_EMAIL = ""
app = Flask(__name__)
auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True)
firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth)
flask_cors.CORS(app)
@app.route('/')
def index():
return "<h1>Welcome To Google HVZ (backend)!</h1>"
@app.route('/test', methods=['GET'])
def get_testdata():
testdata = firebase.get('testdata', None)
return jsonify(testdata)
| Add proper authentication for db (without actual key). | Add proper authentication for db (without actual key).
| Python | apache-2.0 | google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz | import logging
from firebase import firebase
from flask import Flask, jsonify, request
import flask_cors
from google.appengine.ext import ndb
import google.auth.transport.requests
import google.oauth2.id_token
import requests_toolbelt.adapters.appengine
requests_toolbelt.adapters.appengine.monkeypatch()
HTTP_REQUEST = google.auth.transport.requests.Request()
+ # Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database
+ FIREBASE_SECRET = ""
+ FIREBASE_EMAIL = ""
+
app = Flask(__name__)
+ auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True)
- firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None)
+ firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth)
flask_cors.CORS(app)
@app.route('/')
def index():
return "<h1>Welcome To Google HVZ (backend)!</h1>"
@app.route('/test', methods=['GET'])
def get_testdata():
testdata = firebase.get('testdata', None)
return jsonify(testdata)
| Add proper authentication for db (without actual key). | ## Code Before:
import logging
from firebase import firebase
from flask import Flask, jsonify, request
import flask_cors
from google.appengine.ext import ndb
import google.auth.transport.requests
import google.oauth2.id_token
import requests_toolbelt.adapters.appengine
requests_toolbelt.adapters.appengine.monkeypatch()
HTTP_REQUEST = google.auth.transport.requests.Request()
app = Flask(__name__)
firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None)
flask_cors.CORS(app)
@app.route('/')
def index():
return "<h1>Welcome To Google HVZ (backend)!</h1>"
@app.route('/test', methods=['GET'])
def get_testdata():
testdata = firebase.get('testdata', None)
return jsonify(testdata)
## Instruction:
Add proper authentication for db (without actual key).
## Code After:
import logging
from firebase import firebase
from flask import Flask, jsonify, request
import flask_cors
from google.appengine.ext import ndb
import google.auth.transport.requests
import google.oauth2.id_token
import requests_toolbelt.adapters.appengine
requests_toolbelt.adapters.appengine.monkeypatch()
HTTP_REQUEST = google.auth.transport.requests.Request()
# Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database
FIREBASE_SECRET = ""
FIREBASE_EMAIL = ""
app = Flask(__name__)
auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True)
firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth)
flask_cors.CORS(app)
@app.route('/')
def index():
return "<h1>Welcome To Google HVZ (backend)!</h1>"
@app.route('/test', methods=['GET'])
def get_testdata():
testdata = firebase.get('testdata', None)
return jsonify(testdata)
| import logging
from firebase import firebase
from flask import Flask, jsonify, request
import flask_cors
from google.appengine.ext import ndb
import google.auth.transport.requests
import google.oauth2.id_token
import requests_toolbelt.adapters.appengine
requests_toolbelt.adapters.appengine.monkeypatch()
HTTP_REQUEST = google.auth.transport.requests.Request()
+ # Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database
+ FIREBASE_SECRET = ""
+ FIREBASE_EMAIL = ""
+
app = Flask(__name__)
+ auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True)
- firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None)
? ^ ^
+ firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth)
? ^^^^^^^^^^^^ ^^^^^
flask_cors.CORS(app)
@app.route('/')
def index():
return "<h1>Welcome To Google HVZ (backend)!</h1>"
@app.route('/test', methods=['GET'])
def get_testdata():
testdata = firebase.get('testdata', None)
return jsonify(testdata) |
ca953b2ef7662e4a70eba386e66ed6d66fad4eec | setup.py | setup.py |
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.1",
description = "Store and access your passwords safely.",
url = "http://keyring-python.org/",
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README.txt').read(),
platforms = ["Many"],
packages = ['keyring'],
ext_modules = get_extensions()
)
|
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.1",
description = "Store and access your passwords safely.",
url = "http://home.python-keyring.org/",
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README.txt').read(),
platforms = ["Many"],
packages = ['keyring'],
ext_modules = get_extensions()
)
| Fix the error in the home page URL. | Fix the error in the home page URL.
| Python | mit | jaraco/keyring |
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.1",
description = "Store and access your passwords safely.",
- url = "http://keyring-python.org/",
+ url = "http://home.python-keyring.org/",
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README.txt').read(),
platforms = ["Many"],
packages = ['keyring'],
ext_modules = get_extensions()
)
| Fix the error in the home page URL. | ## Code Before:
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.1",
description = "Store and access your passwords safely.",
url = "http://keyring-python.org/",
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README.txt').read(),
platforms = ["Many"],
packages = ['keyring'],
ext_modules = get_extensions()
)
## Instruction:
Fix the error in the home page URL.
## Code After:
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.1",
description = "Store and access your passwords safely.",
url = "http://home.python-keyring.org/",
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README.txt').read(),
platforms = ["Many"],
packages = ['keyring'],
ext_modules = get_extensions()
)
|
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.1",
description = "Store and access your passwords safely.",
- url = "http://keyring-python.org/",
? -------
+ url = "http://home.python-keyring.org/",
? ++++++++++++
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README.txt').read(),
platforms = ["Many"],
packages = ['keyring'],
ext_modules = get_extensions()
)
|
0ee2b337b61155044a66ae1f6f173492a51c1150 | dipsim/fluorophore.py | dipsim/fluorophore.py | import numpy as np
class Fluorophore:
"""A single fluorophore is specified by its 3D position, (unit) absorption
dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi).
"""
def __init__(self, position=np.array([0, 0, 0]),
mu_abs=np.array([0, 0]),
mu_em=np.array([0, 0])):
self.position = position
self.mu_abs = mu_abs
self.mu_em = mu_em
self.mu_ind = 0
| import numpy as np
class Fluorophore:
"""A fluorophore is specified by its orientation (in theta and phi spherical
coordinates), it distribution (using a kappa watson distribution), and a
constant (c) proportional to the fluorohphore's brightness.
"""
def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0):
self.theta = theta
self.phi = phi
self.kappa = kappa
self.c = c
| Modify Fluorophore for more convenient coordinates. | Modify Fluorophore for more convenient coordinates.
| Python | mit | talonchandler/dipsim,talonchandler/dipsim | import numpy as np
class Fluorophore:
- """A single fluorophore is specified by its 3D position, (unit) absorption
- dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi).
+ """A fluorophore is specified by its orientation (in theta and phi spherical
+ coordinates), it distribution (using a kappa watson distribution), and a
+ constant (c) proportional to the fluorohphore's brightness.
"""
+ def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0):
+ self.theta = theta
- def __init__(self, position=np.array([0, 0, 0]),
- mu_abs=np.array([0, 0]),
- mu_em=np.array([0, 0])):
- self.position = position
- self.mu_abs = mu_abs
- self.mu_em = mu_em
- self.mu_ind = 0
+ self.phi = phi
+ self.kappa = kappa
+ self.c = c
| Modify Fluorophore for more convenient coordinates. | ## Code Before:
import numpy as np
class Fluorophore:
"""A single fluorophore is specified by its 3D position, (unit) absorption
dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi).
"""
def __init__(self, position=np.array([0, 0, 0]),
mu_abs=np.array([0, 0]),
mu_em=np.array([0, 0])):
self.position = position
self.mu_abs = mu_abs
self.mu_em = mu_em
self.mu_ind = 0
## Instruction:
Modify Fluorophore for more convenient coordinates.
## Code After:
import numpy as np
class Fluorophore:
"""A fluorophore is specified by its orientation (in theta and phi spherical
coordinates), it distribution (using a kappa watson distribution), and a
constant (c) proportional to the fluorohphore's brightness.
"""
def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0):
self.theta = theta
self.phi = phi
self.kappa = kappa
self.c = c
| import numpy as np
class Fluorophore:
- """A single fluorophore is specified by its 3D position, (unit) absorption
- dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi).
+ """A fluorophore is specified by its orientation (in theta and phi spherical
+ coordinates), it distribution (using a kappa watson distribution), and a
+ constant (c) proportional to the fluorohphore's brightness.
"""
- def __init__(self, position=np.array([0, 0, 0]),
- mu_abs=np.array([0, 0]),
- mu_em=np.array([0, 0])):
- self.position = position
+ def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0):
+ self.theta = theta
+ self.phi = phi
+ self.kappa = kappa
- self.mu_abs = mu_abs
? ^^^^^^ ^^^^^^
+ self.c = c
? ^ ^
- self.mu_em = mu_em
- self.mu_ind = 0 |
075d6f1b8f232c1ae7cb7d288da8f8d1040f49c9 | hooks/pre_gen_project.py | hooks/pre_gen_project.py | repo_name = '{{ cookiecutter.repo_name }}'
assert_msg = 'Repo name should be valid Python identifier!'
if hasattr(repo_name, 'isidentifier'):
assert repo_name.isidentifier(), assert_msg
else:
import re
identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$")
assert bool(identifier_re.match(repo_name)), assert_msg
| import sys
import cookiecutter
# Ensure cookiecutter is recent enough
cookiecutter_min_version = '1.3.0'
if cookiecutter.__version__ < cookiecutter_min_version:
print("--------------------------------------------------------------")
print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version)
print("--------------------------------------------------------------")
sys.exit(1)
# Ensure the selected repo name is usable
repo_name = '{{ cookiecutter.repo_name }}'
assert_msg = 'Repo name should be valid Python identifier!'
if hasattr(repo_name, 'isidentifier'):
assert repo_name.isidentifier(), assert_msg
else:
import re
identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$")
assert bool(identifier_re.match(repo_name)), assert_msg
| Add check for cookiecutter version - at least 1.3.0 is required now | Add check for cookiecutter version - at least 1.3.0 is required now
| Python | isc | thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template | + import sys
+
+ import cookiecutter
+
+
+ # Ensure cookiecutter is recent enough
+ cookiecutter_min_version = '1.3.0'
+ if cookiecutter.__version__ < cookiecutter_min_version:
+ print("--------------------------------------------------------------")
+ print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version)
+ print("--------------------------------------------------------------")
+ sys.exit(1)
+
+
+ # Ensure the selected repo name is usable
repo_name = '{{ cookiecutter.repo_name }}'
assert_msg = 'Repo name should be valid Python identifier!'
if hasattr(repo_name, 'isidentifier'):
assert repo_name.isidentifier(), assert_msg
else:
import re
identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$")
assert bool(identifier_re.match(repo_name)), assert_msg
| Add check for cookiecutter version - at least 1.3.0 is required now | ## Code Before:
repo_name = '{{ cookiecutter.repo_name }}'
assert_msg = 'Repo name should be valid Python identifier!'
if hasattr(repo_name, 'isidentifier'):
assert repo_name.isidentifier(), assert_msg
else:
import re
identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$")
assert bool(identifier_re.match(repo_name)), assert_msg
## Instruction:
Add check for cookiecutter version - at least 1.3.0 is required now
## Code After:
import sys
import cookiecutter
# Ensure cookiecutter is recent enough
cookiecutter_min_version = '1.3.0'
if cookiecutter.__version__ < cookiecutter_min_version:
print("--------------------------------------------------------------")
print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version)
print("--------------------------------------------------------------")
sys.exit(1)
# Ensure the selected repo name is usable
repo_name = '{{ cookiecutter.repo_name }}'
assert_msg = 'Repo name should be valid Python identifier!'
if hasattr(repo_name, 'isidentifier'):
assert repo_name.isidentifier(), assert_msg
else:
import re
identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$")
assert bool(identifier_re.match(repo_name)), assert_msg
| + import sys
+
+ import cookiecutter
+
+
+ # Ensure cookiecutter is recent enough
+ cookiecutter_min_version = '1.3.0'
+ if cookiecutter.__version__ < cookiecutter_min_version:
+ print("--------------------------------------------------------------")
+ print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version)
+ print("--------------------------------------------------------------")
+ sys.exit(1)
+
+
+ # Ensure the selected repo name is usable
repo_name = '{{ cookiecutter.repo_name }}'
assert_msg = 'Repo name should be valid Python identifier!'
if hasattr(repo_name, 'isidentifier'):
assert repo_name.isidentifier(), assert_msg
else:
import re
identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$")
assert bool(identifier_re.match(repo_name)), assert_msg |
f1d48525f1e8cde2af9a49636f38360b87b0ecb6 | function/univariate_function.py | function/univariate_function.py | from abc import ABC, abstractmethod
class UnivariateFunction(ABC):
"""
Class that defines a generic (abstract) univariate function.
"""
def __init(self):
super().__init__()
@abstractmethod
def eval(self, v):
pass
@property
@abstractmethod
def domain_start(self):
pass
@property
@abstractmethod
def domain_end(self):
pass
| from abc import ABC, abstractmethod
class UnivariateFunction(ABC):
"""
Class that defines a generic (abstract) univariate function.
"""
def __init(self):
super().__init__()
@abstractmethod
def eval(self, v):
"""
Evaluate the univariate function with input v, and return that value
:param v: Typically some kind of numeric.
:return:
"""
pass
@property
@abstractmethod
def domain_start(self):
"""
Return the start value of the domain.
:return:
"""
pass
@property
@abstractmethod
def domain_end(self):
"""
Return the end value of the domain.
:return:
"""
pass
| Add comments to abstract methods. | Add comments to abstract methods.
| Python | mit | dpazel/music_rep | from abc import ABC, abstractmethod
class UnivariateFunction(ABC):
"""
Class that defines a generic (abstract) univariate function.
"""
def __init(self):
super().__init__()
@abstractmethod
def eval(self, v):
+ """
+ Evaluate the univariate function with input v, and return that value
+ :param v: Typically some kind of numeric.
+ :return:
+ """
pass
@property
@abstractmethod
def domain_start(self):
+ """
+ Return the start value of the domain.
+ :return:
+ """
pass
@property
@abstractmethod
def domain_end(self):
+ """
+ Return the end value of the domain.
+ :return:
+ """
pass
| Add comments to abstract methods. | ## Code Before:
from abc import ABC, abstractmethod
class UnivariateFunction(ABC):
"""
Class that defines a generic (abstract) univariate function.
"""
def __init(self):
super().__init__()
@abstractmethod
def eval(self, v):
pass
@property
@abstractmethod
def domain_start(self):
pass
@property
@abstractmethod
def domain_end(self):
pass
## Instruction:
Add comments to abstract methods.
## Code After:
from abc import ABC, abstractmethod
class UnivariateFunction(ABC):
"""
Class that defines a generic (abstract) univariate function.
"""
def __init(self):
super().__init__()
@abstractmethod
def eval(self, v):
"""
Evaluate the univariate function with input v, and return that value
:param v: Typically some kind of numeric.
:return:
"""
pass
@property
@abstractmethod
def domain_start(self):
"""
Return the start value of the domain.
:return:
"""
pass
@property
@abstractmethod
def domain_end(self):
"""
Return the end value of the domain.
:return:
"""
pass
| from abc import ABC, abstractmethod
class UnivariateFunction(ABC):
"""
Class that defines a generic (abstract) univariate function.
"""
def __init(self):
super().__init__()
@abstractmethod
def eval(self, v):
+ """
+ Evaluate the univariate function with input v, and return that value
+ :param v: Typically some kind of numeric.
+ :return:
+ """
pass
@property
@abstractmethod
def domain_start(self):
+ """
+ Return the start value of the domain.
+ :return:
+ """
pass
@property
@abstractmethod
def domain_end(self):
+ """
+ Return the end value of the domain.
+ :return:
+ """
pass |
dbfc1a11c0ced8ab7a997944e943a17ff0069199 | lexer.py | lexer.py | import ply.lex as lex
import re
tokens = ( 'SECTION',
'IDENTIFIER',
'LBRACE',
'RBRACE',
'SEMI',
'TRUE',
'FALSE' )
def t_SECTION(t):
r'section'
return t
def t_TRUE(t):
r'(true|1)'
return t
def t_FALSE(t):
r'(false|0)'
return t
def t_IDENTIFIER(t):
r'[a-zA-Z\-0-9]+'
return t
def t_LBRACE(t):
r'{'
return t
def t_RBRACE(t):
r'}'
return t
def t_SEMI(t):
r';'
return t
def t_NEWLINE(t):
r'\n+'
t.lexer.lineno += len(t.value)
return t
t_ignore = ' \t\n'
# Error handling rule
def t_error(t):
print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno))
t.lexer.skip(1)
lexer = lex.lex()
| import ply.lex as lex
import re
tokens = ( 'SECTION',
'IDENTIFIER',
'STRING',
'LBRACE',
'RBRACE',
'SEMI',
'EQU',
'TRUE',
'FALSE' )
def t_SECTION(t):
r'section'
return t
def t_TRUE(t):
r'(true|1)'
return t
def t_FALSE(t):
r'(false|0)'
return t
def t_IDENTIFIER(t):
r'[a-zA-Z\-0-9]+'
return t
def t_STRING(t):
r'(\".*\"|\'.*\')'
t.value = t.value[1:-1]
return t
def t_LBRACE(t):
r'{'
return t
def t_EQU(t):
r'='
return t
def t_RBRACE(t):
r'}'
return t
def t_SEMI(t):
r';'
return t
def t_NEWLINE(t):
r'\n+'
t.lexer.lineno += len(t.value)
return t
t_ignore = ' \t\n'
# Error handling rule
def t_error(t):
print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno))
t.lexer.skip(1)
lexer = lex.lex()
| Add rudimentary string support and a token for '=' sign. | Add rudimentary string support and a token for '=' sign.
| Python | cc0-1.0 | dmbaturin/ply-example | import ply.lex as lex
import re
tokens = ( 'SECTION',
'IDENTIFIER',
+ 'STRING',
'LBRACE',
'RBRACE',
'SEMI',
+ 'EQU',
'TRUE',
'FALSE' )
def t_SECTION(t):
r'section'
return t
def t_TRUE(t):
r'(true|1)'
return t
def t_FALSE(t):
r'(false|0)'
return t
def t_IDENTIFIER(t):
r'[a-zA-Z\-0-9]+'
return t
+ def t_STRING(t):
+ r'(\".*\"|\'.*\')'
+ t.value = t.value[1:-1]
+ return t
+
def t_LBRACE(t):
r'{'
+ return t
+
+ def t_EQU(t):
+ r'='
return t
def t_RBRACE(t):
r'}'
return t
def t_SEMI(t):
r';'
return t
def t_NEWLINE(t):
r'\n+'
t.lexer.lineno += len(t.value)
return t
t_ignore = ' \t\n'
# Error handling rule
def t_error(t):
print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno))
t.lexer.skip(1)
lexer = lex.lex()
| Add rudimentary string support and a token for '=' sign. | ## Code Before:
import ply.lex as lex
import re
tokens = ( 'SECTION',
'IDENTIFIER',
'LBRACE',
'RBRACE',
'SEMI',
'TRUE',
'FALSE' )
def t_SECTION(t):
r'section'
return t
def t_TRUE(t):
r'(true|1)'
return t
def t_FALSE(t):
r'(false|0)'
return t
def t_IDENTIFIER(t):
r'[a-zA-Z\-0-9]+'
return t
def t_LBRACE(t):
r'{'
return t
def t_RBRACE(t):
r'}'
return t
def t_SEMI(t):
r';'
return t
def t_NEWLINE(t):
r'\n+'
t.lexer.lineno += len(t.value)
return t
t_ignore = ' \t\n'
# Error handling rule
def t_error(t):
print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno))
t.lexer.skip(1)
lexer = lex.lex()
## Instruction:
Add rudimentary string support and a token for '=' sign.
## Code After:
import ply.lex as lex
import re
tokens = ( 'SECTION',
'IDENTIFIER',
'STRING',
'LBRACE',
'RBRACE',
'SEMI',
'EQU',
'TRUE',
'FALSE' )
def t_SECTION(t):
r'section'
return t
def t_TRUE(t):
r'(true|1)'
return t
def t_FALSE(t):
r'(false|0)'
return t
def t_IDENTIFIER(t):
r'[a-zA-Z\-0-9]+'
return t
def t_STRING(t):
r'(\".*\"|\'.*\')'
t.value = t.value[1:-1]
return t
def t_LBRACE(t):
r'{'
return t
def t_EQU(t):
r'='
return t
def t_RBRACE(t):
r'}'
return t
def t_SEMI(t):
r';'
return t
def t_NEWLINE(t):
r'\n+'
t.lexer.lineno += len(t.value)
return t
t_ignore = ' \t\n'
# Error handling rule
def t_error(t):
print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno))
t.lexer.skip(1)
lexer = lex.lex()
| import ply.lex as lex
import re
tokens = ( 'SECTION',
'IDENTIFIER',
+ 'STRING',
'LBRACE',
'RBRACE',
'SEMI',
+ 'EQU',
'TRUE',
'FALSE' )
def t_SECTION(t):
r'section'
return t
def t_TRUE(t):
r'(true|1)'
return t
def t_FALSE(t):
r'(false|0)'
return t
def t_IDENTIFIER(t):
r'[a-zA-Z\-0-9]+'
return t
+ def t_STRING(t):
+ r'(\".*\"|\'.*\')'
+ t.value = t.value[1:-1]
+ return t
+
def t_LBRACE(t):
r'{'
+ return t
+
+ def t_EQU(t):
+ r'='
return t
def t_RBRACE(t):
r'}'
return t
def t_SEMI(t):
r';'
return t
def t_NEWLINE(t):
r'\n+'
t.lexer.lineno += len(t.value)
return t
t_ignore = ' \t\n'
# Error handling rule
def t_error(t):
print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno))
t.lexer.skip(1)
lexer = lex.lex() |
67ca9f09cd2cfb5e646b9a09b540c5ff88276201 | pydirections/models/models.py | pydirections/models/models.py | from schematics.models import Model
from schematics.types import StringType
class Step(Model):
"""
Represents an individual step
"""
html_instructions = StringType()
class Leg(Model):
"""
Represents an individual leg
"""
start_address = StringType()
end_address = StringType()
class Route(Model):
"""
Represents an individual route whose attributes include
"""
summary = StringType(required=True)
copyrights = StringType()
@property
def summary():
return summary | from schematics.models import Model
from schematics.types import StringType, DecimalType
from schematics.types.compound import ListType
class Distance(Model):
"""
Represents the duration of a leg/step
"""
value = DecimalType()
text = StringType()
class Duration(Model):
"""
Represents the duration of a leg/step
"""
value = DecimalType()
text = StringType()
class Step(Model):
"""
Represents an individual step
"""
html_instructions = StringType()
class Leg(Model):
"""
Represents an individual leg
"""
start_address = StringType()
end_address = StringType()
steps = ListType(ModelType(Step))
class Route(Model):
"""
Represents an individual route whose attributes include
"""
summary = StringType(required=True)
legs = ListType(ModelType(Leg))
copyrights = StringType()
@property
def summary():
return summary
| Add more details to routes | Add more details to routes
| Python | apache-2.0 | apranav19/pydirections | from schematics.models import Model
- from schematics.types import StringType
+ from schematics.types import StringType, DecimalType
+ from schematics.types.compound import ListType
+
+ class Distance(Model):
+ """
+ Represents the duration of a leg/step
+ """
+ value = DecimalType()
+ text = StringType()
+
+ class Duration(Model):
+ """
+ Represents the duration of a leg/step
+ """
+ value = DecimalType()
+ text = StringType()
+
class Step(Model):
"""
Represents an individual step
"""
html_instructions = StringType()
class Leg(Model):
"""
Represents an individual leg
"""
start_address = StringType()
end_address = StringType()
-
+ steps = ListType(ModelType(Step))
+
class Route(Model):
"""
Represents an individual route whose attributes include
"""
summary = StringType(required=True)
+ legs = ListType(ModelType(Leg))
copyrights = StringType()
@property
def summary():
return summary
+ | Add more details to routes | ## Code Before:
from schematics.models import Model
from schematics.types import StringType
class Step(Model):
"""
Represents an individual step
"""
html_instructions = StringType()
class Leg(Model):
"""
Represents an individual leg
"""
start_address = StringType()
end_address = StringType()
class Route(Model):
"""
Represents an individual route whose attributes include
"""
summary = StringType(required=True)
copyrights = StringType()
@property
def summary():
return summary
## Instruction:
Add more details to routes
## Code After:
from schematics.models import Model
from schematics.types import StringType, DecimalType
from schematics.types.compound import ListType
class Distance(Model):
"""
Represents the duration of a leg/step
"""
value = DecimalType()
text = StringType()
class Duration(Model):
"""
Represents the duration of a leg/step
"""
value = DecimalType()
text = StringType()
class Step(Model):
"""
Represents an individual step
"""
html_instructions = StringType()
class Leg(Model):
"""
Represents an individual leg
"""
start_address = StringType()
end_address = StringType()
steps = ListType(ModelType(Step))
class Route(Model):
"""
Represents an individual route whose attributes include
"""
summary = StringType(required=True)
legs = ListType(ModelType(Leg))
copyrights = StringType()
@property
def summary():
return summary
| from schematics.models import Model
- from schematics.types import StringType
+ from schematics.types import StringType, DecimalType
? +++++++++++++
+ from schematics.types.compound import ListType
+
+ class Distance(Model):
+ """
+ Represents the duration of a leg/step
+ """
+ value = DecimalType()
+ text = StringType()
+
+ class Duration(Model):
+ """
+ Represents the duration of a leg/step
+ """
+ value = DecimalType()
+ text = StringType()
+
class Step(Model):
"""
Represents an individual step
"""
html_instructions = StringType()
class Leg(Model):
"""
Represents an individual leg
"""
start_address = StringType()
end_address = StringType()
-
+ steps = ListType(ModelType(Step))
+
class Route(Model):
"""
Represents an individual route whose attributes include
"""
summary = StringType(required=True)
+ legs = ListType(ModelType(Leg))
copyrights = StringType()
@property
def summary():
return summary |
c87c4a972f0f2d4966142fa666a900112762ed76 | scipy/constants/tests/test_codata.py | scipy/constants/tests/test_codata.py |
import warnings
from scipy.constants import find
from numpy.testing import assert_equal
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
|
import warnings
from scipy.constants import find
from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
if __name__ == "__main__":
run_module_suite()
| Allow codata tests to be run as script. | ENH: Allow codata tests to be run as script.
| Python | bsd-3-clause | zerothi/scipy,zxsted/scipy,josephcslater/scipy,rgommers/scipy,grlee77/scipy,sargas/scipy,dch312/scipy,ilayn/scipy,apbard/scipy,jakevdp/scipy,niknow/scipy,vanpact/scipy,jakevdp/scipy,rmcgibbo/scipy,zxsted/scipy,pnedunuri/scipy,raoulbq/scipy,lhilt/scipy,mgaitan/scipy,mingwpy/scipy,maciejkula/scipy,njwilson23/scipy,Dapid/scipy,woodscn/scipy,perimosocordiae/scipy,aarchiba/scipy,pbrod/scipy,aarchiba/scipy,trankmichael/scipy,Srisai85/scipy,aarchiba/scipy,pschella/scipy,tylerjereddy/scipy,minhlongdo/scipy,mortonjt/scipy,dominicelse/scipy,jsilter/scipy,sargas/scipy,vanpact/scipy,anielsen001/scipy,mingwpy/scipy,matthew-brett/scipy,aeklant/scipy,mikebenfield/scipy,Srisai85/scipy,sauliusl/scipy,FRidh/scipy,andim/scipy,ortylp/scipy,josephcslater/scipy,mhogg/scipy,WarrenWeckesser/scipy,ndchorley/scipy,josephcslater/scipy,mortonjt/scipy,felipebetancur/scipy,jonycgn/scipy,mtrbean/scipy,juliantaylor/scipy,mdhaber/scipy,mdhaber/scipy,kleskjr/scipy,ogrisel/scipy,teoliphant/scipy,petebachant/scipy,lukauskas/scipy,Gillu13/scipy,sonnyhu/scipy,WillieMaddox/scipy,arokem/scipy,jonycgn/scipy,gdooper/scipy,ChanderG/scipy,pbrod/scipy,surhudm/scipy,rgommers/scipy,anntzer/scipy,gfyoung/scipy,sriki18/scipy,nmayorov/scipy,ogrisel/scipy,nvoron23/scipy,giorgiop/scipy,trankmichael/scipy,scipy/scipy,dominicelse/scipy,petebachant/scipy,ChanderG/scipy,ndchorley/scipy,scipy/scipy,Stefan-Endres/scipy,futurulus/scipy,giorgiop/scipy,behzadnouri/scipy,sargas/scipy,haudren/scipy,Gillu13/scipy,teoliphant/scipy,WillieMaddox/scipy,cpaulik/scipy,juliantaylor/scipy,sriki18/scipy,ortylp/scipy,zerothi/scipy,gdooper/scipy,befelix/scipy,raoulbq/scipy,sargas/scipy,newemailjdm/scipy,piyush0609/scipy,vhaasteren/scipy,sriki18/scipy,sauliusl/scipy,rgommers/scipy,zaxliu/scipy,Stefan-Endres/scipy,dch312/scipy,scipy/scipy,aeklant/scipy,cpaulik/scipy,maniteja123/scipy,zerothi/scipy,cpaulik/scipy,Eric89GXL/scipy,mgaitan/scipy,bkendzior/scipy,jonycgn/scipy,juliantaylor/scipy,felipebetancur/scipy,giorgiop/scipy,gertingold/scipy,befelix/scipy,maciejkula/scipy,vberaudi/scipy,lhilt/scipy,ChanderG/scipy,mgaitan/scipy,hainm/scipy,endolith/scipy,lukauskas/scipy,gdooper/scipy,kalvdans/scipy,richardotis/scipy,jor-/scipy,mgaitan/scipy,mtrbean/scipy,Gillu13/scipy,raoulbq/scipy,lukauskas/scipy,tylerjereddy/scipy,person142/scipy,mortada/scipy,vhaasteren/scipy,Shaswat27/scipy,vigna/scipy,pschella/scipy,surhudm/scipy,kalvdans/scipy,witcxc/scipy,piyush0609/scipy,woodscn/scipy,matthew-brett/scipy,nvoron23/scipy,matthewalbani/scipy,mdhaber/scipy,gfyoung/scipy,pnedunuri/scipy,vanpact/scipy,ilayn/scipy,FRidh/scipy,mingwpy/scipy,efiring/scipy,jakevdp/scipy,Shaswat27/scipy,matthewalbani/scipy,pizzathief/scipy,mdhaber/scipy,jonycgn/scipy,pyramania/scipy,apbard/scipy,Dapid/scipy,rmcgibbo/scipy,vhaasteren/scipy,niknow/scipy,apbard/scipy,sonnyhu/scipy,larsmans/scipy,bkendzior/scipy,perimosocordiae/scipy,maniteja123/scipy,mortada/scipy,felipebetancur/scipy,andyfaff/scipy,mikebenfield/scipy,jseabold/scipy,mtrbean/scipy,jamestwebber/scipy,anielsen001/scipy,ortylp/scipy,Gillu13/scipy,woodscn/scipy,witcxc/scipy,haudren/scipy,witcxc/scipy,andyfaff/scipy,newemailjdm/scipy,vhaasteren/scipy,njwilson23/scipy,grlee77/scipy,surhudm/scipy,zaxliu/scipy,ortylp/scipy,njwilson23/scipy,dch312/scipy,fredrikw/scipy,pnedunuri/scipy,mortada/scipy,mortada/scipy,fredrikw/scipy,gef756/scipy,zaxliu/scipy,chatcannon/scipy,matthew-brett/scipy,jsilter/scipy,ilayn/scipy,mortada/scipy,Stefan-Endres/scipy,jseabold/scipy,minhlongdo/scipy,anntzer/scipy,andyfaff/scipy,futurulus/scipy,sauliusl/scipy,e-q/scipy,mingwpy/scipy,rmcgibbo/scipy,Dapid/scipy,vigna/scipy,minhlongdo/scipy,mdhaber/scipy,raoulbq/scipy,pyramania/scipy,richardotis/scipy,jor-/scipy,befelix/scipy,sriki18/scipy,mortonjt/scipy,endolith/scipy,haudren/scipy,juliantaylor/scipy,hainm/scipy,e-q/scipy,pizzathief/scipy,endolith/scipy,FRidh/scipy,argriffing/scipy,perimosocordiae/scipy,tylerjereddy/scipy,fernand/scipy,tylerjereddy/scipy,argriffing/scipy,vigna/scipy,mhogg/scipy,vanpact/scipy,woodscn/scipy,Eric89GXL/scipy,vberaudi/scipy,ChanderG/scipy,raoulbq/scipy,gef756/scipy,sauliusl/scipy,anntzer/scipy,jseabold/scipy,pbrod/scipy,argriffing/scipy,nvoron23/scipy,ales-erjavec/scipy,ortylp/scipy,Newman101/scipy,jjhelmus/scipy,petebachant/scipy,mingwpy/scipy,chatcannon/scipy,Eric89GXL/scipy,Shaswat27/scipy,ogrisel/scipy,anntzer/scipy,andyfaff/scipy,aman-iitj/scipy,cpaulik/scipy,sauliusl/scipy,zerothi/scipy,hainm/scipy,Srisai85/scipy,scipy/scipy,mtrbean/scipy,fredrikw/scipy,endolith/scipy,Dapid/scipy,ndchorley/scipy,giorgiop/scipy,felipebetancur/scipy,Gillu13/scipy,maciejkula/scipy,jonycgn/scipy,vanpact/scipy,fernand/scipy,argriffing/scipy,aarchiba/scipy,jjhelmus/scipy,newemailjdm/scipy,matthewalbani/scipy,behzadnouri/scipy,jamestwebber/scipy,behzadnouri/scipy,andim/scipy,pschella/scipy,nvoron23/scipy,ndchorley/scipy,Shaswat27/scipy,argriffing/scipy,andyfaff/scipy,juliantaylor/scipy,felipebetancur/scipy,mtrbean/scipy,fredrikw/scipy,sonnyhu/scipy,efiring/scipy,e-q/scipy,jsilter/scipy,mtrbean/scipy,behzadnouri/scipy,grlee77/scipy,ChanderG/scipy,maciejkula/scipy,ogrisel/scipy,maciejkula/scipy,Srisai85/scipy,jamestwebber/scipy,woodscn/scipy,pschella/scipy,vigna/scipy,dch312/scipy,lhilt/scipy,haudren/scipy,nmayorov/scipy,futurulus/scipy,Newman101/scipy,larsmans/scipy,Eric89GXL/scipy,larsmans/scipy,maniteja123/scipy,maniteja123/scipy,gfyoung/scipy,fernand/scipy,sargas/scipy,nmayorov/scipy,lhilt/scipy,richardotis/scipy,vanpact/scipy,teoliphant/scipy,arokem/scipy,aeklant/scipy,pyramania/scipy,jamestwebber/scipy,vberaudi/scipy,dominicelse/scipy,efiring/scipy,rmcgibbo/scipy,ales-erjavec/scipy,Eric89GXL/scipy,niknow/scipy,zaxliu/scipy,surhudm/scipy,maniteja123/scipy,mikebenfield/scipy,gef756/scipy,vberaudi/scipy,efiring/scipy,sriki18/scipy,FRidh/scipy,gdooper/scipy,futurulus/scipy,cpaulik/scipy,WarrenWeckesser/scipy,larsmans/scipy,person142/scipy,teoliphant/scipy,kleskjr/scipy,befelix/scipy,arokem/scipy,Kamp9/scipy,kleskjr/scipy,teoliphant/scipy,kleskjr/scipy,dominicelse/scipy,ndchorley/scipy,anntzer/scipy,jjhelmus/scipy,pbrod/scipy,ilayn/scipy,Shaswat27/scipy,efiring/scipy,njwilson23/scipy,niknow/scipy,dominicelse/scipy,behzadnouri/scipy,mingwpy/scipy,rgommers/scipy,FRidh/scipy,person142/scipy,mortonjt/scipy,vhaasteren/scipy,pbrod/scipy,nvoron23/scipy,jor-/scipy,niknow/scipy,trankmichael/scipy,fernand/scipy,haudren/scipy,njwilson23/scipy,matthew-brett/scipy,nvoron23/scipy,endolith/scipy,e-q/scipy,jonycgn/scipy,dch312/scipy,WillieMaddox/scipy,rmcgibbo/scipy,lukauskas/scipy,andim/scipy,jakevdp/scipy,zerothi/scipy,witcxc/scipy,ales-erjavec/scipy,ales-erjavec/scipy,jor-/scipy,grlee77/scipy,nonhermitian/scipy,aman-iitj/scipy,vhaasteren/scipy,chatcannon/scipy,arokem/scipy,giorgiop/scipy,futurulus/scipy,e-q/scipy,ChanderG/scipy,ilayn/scipy,WarrenWeckesser/scipy,mikebenfield/scipy,Newman101/scipy,richardotis/scipy,jsilter/scipy,perimosocordiae/scipy,gfyoung/scipy,scipy/scipy,kalvdans/scipy,sonnyhu/scipy,lukauskas/scipy,WillieMaddox/scipy,aman-iitj/scipy,Newman101/scipy,zerothi/scipy,apbard/scipy,jsilter/scipy,hainm/scipy,Stefan-Endres/scipy,maniteja123/scipy,petebachant/scipy,efiring/scipy,mhogg/scipy,Dapid/scipy,nmayorov/scipy,anielsen001/scipy,Dapid/scipy,aeklant/scipy,nonhermitian/scipy,perimosocordiae/scipy,gertingold/scipy,chatcannon/scipy,zxsted/scipy,trankmichael/scipy,behzadnouri/scipy,WillieMaddox/scipy,newemailjdm/scipy,bkendzior/scipy,vberaudi/scipy,anielsen001/scipy,argriffing/scipy,minhlongdo/scipy,aman-iitj/scipy,futurulus/scipy,pizzathief/scipy,aeklant/scipy,anielsen001/scipy,cpaulik/scipy,njwilson23/scipy,zaxliu/scipy,surhudm/scipy,Kamp9/scipy,Eric89GXL/scipy,mhogg/scipy,scipy/scipy,gfyoung/scipy,befelix/scipy,Newman101/scipy,ales-erjavec/scipy,raoulbq/scipy,zaxliu/scipy,pyramania/scipy,larsmans/scipy,mortada/scipy,chatcannon/scipy,grlee77/scipy,person142/scipy,Kamp9/scipy,giorgiop/scipy,kalvdans/scipy,Stefan-Endres/scipy,mhogg/scipy,mortonjt/scipy,Kamp9/scipy,zxsted/scipy,hainm/scipy,lhilt/scipy,haudren/scipy,sauliusl/scipy,aman-iitj/scipy,aarchiba/scipy,kleskjr/scipy,andyfaff/scipy,gef756/scipy,WarrenWeckesser/scipy,fredrikw/scipy,gertingold/scipy,petebachant/scipy,petebachant/scipy,endolith/scipy,Newman101/scipy,sonnyhu/scipy,trankmichael/scipy,josephcslater/scipy,vigna/scipy,kleskjr/scipy,matthewalbani/scipy,ortylp/scipy,Gillu13/scipy,gdooper/scipy,Kamp9/scipy,larsmans/scipy,mikebenfield/scipy,matthewalbani/scipy,Shaswat27/scipy,lukauskas/scipy,pizzathief/scipy,josephcslater/scipy,pnedunuri/scipy,matthew-brett/scipy,chatcannon/scipy,bkendzior/scipy,ilayn/scipy,jjhelmus/scipy,perimosocordiae/scipy,apbard/scipy,ogrisel/scipy,anntzer/scipy,rmcgibbo/scipy,surhudm/scipy,mdhaber/scipy,bkendzior/scipy,jseabold/scipy,sriki18/scipy,arokem/scipy,kalvdans/scipy,Kamp9/scipy,Srisai85/scipy,pnedunuri/scipy,pyramania/scipy,andim/scipy,fernand/scipy,gef756/scipy,trankmichael/scipy,nonhermitian/scipy,pnedunuri/scipy,niknow/scipy,piyush0609/scipy,richardotis/scipy,jamestwebber/scipy,Stefan-Endres/scipy,woodscn/scipy,minhlongdo/scipy,fredrikw/scipy,newemailjdm/scipy,WarrenWeckesser/scipy,ales-erjavec/scipy,andim/scipy,richardotis/scipy,piyush0609/scipy,andim/scipy,jor-/scipy,FRidh/scipy,nonhermitian/scipy,tylerjereddy/scipy,jseabold/scipy,gertingold/scipy,pizzathief/scipy,jseabold/scipy,zxsted/scipy,rgommers/scipy,gertingold/scipy,WarrenWeckesser/scipy,jakevdp/scipy,mortonjt/scipy,hainm/scipy,jjhelmus/scipy,piyush0609/scipy,WillieMaddox/scipy,ndchorley/scipy,zxsted/scipy,mgaitan/scipy,person142/scipy,minhlongdo/scipy,fernand/scipy,gef756/scipy,pbrod/scipy,pschella/scipy,anielsen001/scipy,vberaudi/scipy,aman-iitj/scipy,mhogg/scipy,witcxc/scipy,sonnyhu/scipy,piyush0609/scipy,mgaitan/scipy,nmayorov/scipy,newemailjdm/scipy,felipebetancur/scipy,Srisai85/scipy,nonhermitian/scipy |
import warnings
from scipy.constants import find
- from numpy.testing import assert_equal
+ from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
+ if __name__ == "__main__":
+ run_module_suite()
+ | Allow codata tests to be run as script. | ## Code Before:
import warnings
from scipy.constants import find
from numpy.testing import assert_equal
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
## Instruction:
Allow codata tests to be run as script.
## Code After:
import warnings
from scipy.constants import find
from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
if __name__ == "__main__":
run_module_suite()
|
import warnings
from scipy.constants import find
- from numpy.testing import assert_equal
+ from numpy.testing import assert_equal, run_module_suite
? ++++++++++++++++++
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
+
+ if __name__ == "__main__":
+ run_module_suite() |
380baa34af7e8a704780f0ec535b626f4a286e23 | deflect/admin.py | deflect/admin.py | from django.contrib import admin
from .models import RedirectURL
class RedirectURLAdmin(admin.ModelAdmin):
list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',)
list_filter = ('creator__username', 'campaign', 'medium',)
ordering = ('-last_used',)
readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',)
search_fields = ['url', 'campaign']
fieldsets = ((None, {'fields': ('url', 'short_url',)}),
('Google', {'fields': ('campaign', 'medium', 'content',)}),
('Additional info', {'fields': ('description', 'qr_code',)}),
('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),)
def save_model(self, request, obj, form, change):
obj.user = request.user
obj.save()
admin.site.register(RedirectURL, RedirectURLAdmin)
| from django.contrib import admin
from .models import RedirectURL
class RedirectURLAdmin(admin.ModelAdmin):
list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',)
list_filter = ('creator__username', 'campaign', 'medium',)
ordering = ('-last_used',)
readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',)
search_fields = ['url', 'campaign']
fieldsets = ((None, {'fields': ('url', 'short_url',)}),
('Google', {'fields': ('campaign', 'medium', 'content',)}),
('Additional info', {'fields': ('description', 'qr_code',)}),
('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),)
def save_model(self, request, obj, form, change):
if not change:
obj.creator = request.user
obj.save()
admin.site.register(RedirectURL, RedirectURLAdmin)
| Fix model creator updating on change event | Fix model creator updating on change event
| Python | bsd-3-clause | jbittel/django-deflect | from django.contrib import admin
from .models import RedirectURL
class RedirectURLAdmin(admin.ModelAdmin):
list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',)
list_filter = ('creator__username', 'campaign', 'medium',)
ordering = ('-last_used',)
readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',)
search_fields = ['url', 'campaign']
fieldsets = ((None, {'fields': ('url', 'short_url',)}),
('Google', {'fields': ('campaign', 'medium', 'content',)}),
('Additional info', {'fields': ('description', 'qr_code',)}),
('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),)
def save_model(self, request, obj, form, change):
+ if not change:
- obj.user = request.user
+ obj.creator = request.user
obj.save()
admin.site.register(RedirectURL, RedirectURLAdmin)
| Fix model creator updating on change event | ## Code Before:
from django.contrib import admin
from .models import RedirectURL
class RedirectURLAdmin(admin.ModelAdmin):
list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',)
list_filter = ('creator__username', 'campaign', 'medium',)
ordering = ('-last_used',)
readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',)
search_fields = ['url', 'campaign']
fieldsets = ((None, {'fields': ('url', 'short_url',)}),
('Google', {'fields': ('campaign', 'medium', 'content',)}),
('Additional info', {'fields': ('description', 'qr_code',)}),
('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),)
def save_model(self, request, obj, form, change):
obj.user = request.user
obj.save()
admin.site.register(RedirectURL, RedirectURLAdmin)
## Instruction:
Fix model creator updating on change event
## Code After:
from django.contrib import admin
from .models import RedirectURL
class RedirectURLAdmin(admin.ModelAdmin):
list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',)
list_filter = ('creator__username', 'campaign', 'medium',)
ordering = ('-last_used',)
readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',)
search_fields = ['url', 'campaign']
fieldsets = ((None, {'fields': ('url', 'short_url',)}),
('Google', {'fields': ('campaign', 'medium', 'content',)}),
('Additional info', {'fields': ('description', 'qr_code',)}),
('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),)
def save_model(self, request, obj, form, change):
if not change:
obj.creator = request.user
obj.save()
admin.site.register(RedirectURL, RedirectURLAdmin)
| from django.contrib import admin
from .models import RedirectURL
class RedirectURLAdmin(admin.ModelAdmin):
list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',)
list_filter = ('creator__username', 'campaign', 'medium',)
ordering = ('-last_used',)
readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',)
search_fields = ['url', 'campaign']
fieldsets = ((None, {'fields': ('url', 'short_url',)}),
('Google', {'fields': ('campaign', 'medium', 'content',)}),
('Additional info', {'fields': ('description', 'qr_code',)}),
('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),)
def save_model(self, request, obj, form, change):
+ if not change:
- obj.user = request.user
? ^^
+ obj.creator = request.user
? ++++ ^^ +++
obj.save()
admin.site.register(RedirectURL, RedirectURLAdmin) |
d46d0b5a5392b6ca047b519a9d6280b5b0581e81 | system_maintenance/tests/functional/tests.py | system_maintenance/tests/functional/tests.py | from django.test import LiveServerTestCase
from selenium import webdriver
class FunctionalTest(LiveServerTestCase):
def setUp(self):
self.browser = webdriver.Firefox()
self.browser.implicitly_wait(3)
def tearDown(self):
self.browser.quit()
def test_app_home_title(self):
self.browser.get('http://localhost:8000/system_maintenance')
self.assertIn('System Maintenance', self.browser.title)
| from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from selenium import webdriver
class FunctionalTest(StaticLiveServerTestCase):
def setUp(self):
self.browser = webdriver.Firefox()
self.browser.implicitly_wait(3)
def tearDown(self):
self.browser.quit()
def test_app_home_title(self):
self.browser.get('http://localhost:8000/system_maintenance')
self.assertIn('System Maintenance', self.browser.title)
| Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT' | Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT'
| Python | bsd-3-clause | mfcovington/django-system-maintenance,mfcovington/django-system-maintenance,mfcovington/django-system-maintenance | - from django.test import LiveServerTestCase
+ from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from selenium import webdriver
- class FunctionalTest(LiveServerTestCase):
+ class FunctionalTest(StaticLiveServerTestCase):
def setUp(self):
self.browser = webdriver.Firefox()
self.browser.implicitly_wait(3)
def tearDown(self):
self.browser.quit()
def test_app_home_title(self):
self.browser.get('http://localhost:8000/system_maintenance')
self.assertIn('System Maintenance', self.browser.title)
| Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT' | ## Code Before:
from django.test import LiveServerTestCase
from selenium import webdriver
class FunctionalTest(LiveServerTestCase):
def setUp(self):
self.browser = webdriver.Firefox()
self.browser.implicitly_wait(3)
def tearDown(self):
self.browser.quit()
def test_app_home_title(self):
self.browser.get('http://localhost:8000/system_maintenance')
self.assertIn('System Maintenance', self.browser.title)
## Instruction:
Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT'
## Code After:
from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from selenium import webdriver
class FunctionalTest(StaticLiveServerTestCase):
def setUp(self):
self.browser = webdriver.Firefox()
self.browser.implicitly_wait(3)
def tearDown(self):
self.browser.quit()
def test_app_home_title(self):
self.browser.get('http://localhost:8000/system_maintenance')
self.assertIn('System Maintenance', self.browser.title)
| - from django.test import LiveServerTestCase
+ from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from selenium import webdriver
- class FunctionalTest(LiveServerTestCase):
+ class FunctionalTest(StaticLiveServerTestCase):
? ++++++
def setUp(self):
self.browser = webdriver.Firefox()
self.browser.implicitly_wait(3)
def tearDown(self):
self.browser.quit()
def test_app_home_title(self):
self.browser.get('http://localhost:8000/system_maintenance')
self.assertIn('System Maintenance', self.browser.title) |
09099ab106ae4c0695502e3632e4ac1c2f459566 | apps/teams/bulk_actions.py | apps/teams/bulk_actions.py | from django.contrib.contenttypes.models import ContentType
from subtitles.models import SubtitleLanguage
from teams.signals import api_subtitles_approved
from utils.csv_parser import UnicodeReader
from videos.tasks import video_changed_tasks
def complete_approve_tasks(tasks):
lang_ct = ContentType.objects.get_for_model(SubtitleLanguage)
video_ids = set()
for task in tasks:
task.do_complete_approve(lang_ct=lang_ct)
api_subtitles_approved.send(task.get_subtitle_version())
video_ids.add(task.team_video.video_id)
for video_id in video_ids:
video_changed_tasks.delay(video_id)
def add_videos_from_csv(team, user, csv_file):
from .tasks import add_team_videos
videos = []
fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript']
num_fields = len(fields)
try:
reader = UnicodeReader(csv_file)
header = reader.next()
if len(header) != num_fields:
raise Exception()
except:
raise ValueError(u'CSV format is not valid')
for row in reader:
videos.append(dict(zip(fields, row)))
add_team_videos.delay(team.pk, user.pk, videos)
| from django.contrib.contenttypes.models import ContentType
from subtitles.models import SubtitleLanguage
from subtitles.signals import subtitles_published
from teams.signals import api_subtitles_approved
from utils.csv_parser import UnicodeReader
from videos.tasks import video_changed_tasks
def complete_approve_tasks(tasks):
lang_ct = ContentType.objects.get_for_model(SubtitleLanguage)
video_ids = set()
for task in tasks:
task.do_complete_approve(lang_ct=lang_ct)
version = task.get_subtitle_version()
api_subtitles_approved.send(version)
if version.is_public():
subtitles_published.send(version.subtitle_language, version=version)
video_ids.add(task.team_video.video_id)
for video_id in video_ids:
video_changed_tasks.delay(video_id)
def add_videos_from_csv(team, user, csv_file):
from .tasks import add_team_videos
videos = []
fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript']
num_fields = len(fields)
try:
reader = UnicodeReader(csv_file)
header = reader.next()
if len(header) != num_fields:
raise Exception()
except:
raise ValueError(u'CSV format is not valid')
for row in reader:
videos.append(dict(zip(fields, row)))
add_team_videos.delay(team.pk, user.pk, videos)
| Send subtitles_published signal for bulk approvals | Send subtitles_published signal for bulk approvals
This fixes pculture/amara-enterprise#608
| Python | agpl-3.0 | pculture/unisubs,pculture/unisubs,wevoice/wesub,pculture/unisubs,wevoice/wesub,wevoice/wesub,pculture/unisubs,wevoice/wesub | from django.contrib.contenttypes.models import ContentType
from subtitles.models import SubtitleLanguage
+ from subtitles.signals import subtitles_published
from teams.signals import api_subtitles_approved
from utils.csv_parser import UnicodeReader
from videos.tasks import video_changed_tasks
def complete_approve_tasks(tasks):
lang_ct = ContentType.objects.get_for_model(SubtitleLanguage)
video_ids = set()
for task in tasks:
task.do_complete_approve(lang_ct=lang_ct)
+ version = task.get_subtitle_version()
- api_subtitles_approved.send(task.get_subtitle_version())
+ api_subtitles_approved.send(version)
+ if version.is_public():
+ subtitles_published.send(version.subtitle_language, version=version)
video_ids.add(task.team_video.video_id)
for video_id in video_ids:
video_changed_tasks.delay(video_id)
def add_videos_from_csv(team, user, csv_file):
from .tasks import add_team_videos
videos = []
fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript']
num_fields = len(fields)
try:
reader = UnicodeReader(csv_file)
header = reader.next()
if len(header) != num_fields:
raise Exception()
except:
raise ValueError(u'CSV format is not valid')
for row in reader:
videos.append(dict(zip(fields, row)))
add_team_videos.delay(team.pk, user.pk, videos)
| Send subtitles_published signal for bulk approvals | ## Code Before:
from django.contrib.contenttypes.models import ContentType
from subtitles.models import SubtitleLanguage
from teams.signals import api_subtitles_approved
from utils.csv_parser import UnicodeReader
from videos.tasks import video_changed_tasks
def complete_approve_tasks(tasks):
lang_ct = ContentType.objects.get_for_model(SubtitleLanguage)
video_ids = set()
for task in tasks:
task.do_complete_approve(lang_ct=lang_ct)
api_subtitles_approved.send(task.get_subtitle_version())
video_ids.add(task.team_video.video_id)
for video_id in video_ids:
video_changed_tasks.delay(video_id)
def add_videos_from_csv(team, user, csv_file):
from .tasks import add_team_videos
videos = []
fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript']
num_fields = len(fields)
try:
reader = UnicodeReader(csv_file)
header = reader.next()
if len(header) != num_fields:
raise Exception()
except:
raise ValueError(u'CSV format is not valid')
for row in reader:
videos.append(dict(zip(fields, row)))
add_team_videos.delay(team.pk, user.pk, videos)
## Instruction:
Send subtitles_published signal for bulk approvals
## Code After:
from django.contrib.contenttypes.models import ContentType
from subtitles.models import SubtitleLanguage
from subtitles.signals import subtitles_published
from teams.signals import api_subtitles_approved
from utils.csv_parser import UnicodeReader
from videos.tasks import video_changed_tasks
def complete_approve_tasks(tasks):
lang_ct = ContentType.objects.get_for_model(SubtitleLanguage)
video_ids = set()
for task in tasks:
task.do_complete_approve(lang_ct=lang_ct)
version = task.get_subtitle_version()
api_subtitles_approved.send(version)
if version.is_public():
subtitles_published.send(version.subtitle_language, version=version)
video_ids.add(task.team_video.video_id)
for video_id in video_ids:
video_changed_tasks.delay(video_id)
def add_videos_from_csv(team, user, csv_file):
from .tasks import add_team_videos
videos = []
fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript']
num_fields = len(fields)
try:
reader = UnicodeReader(csv_file)
header = reader.next()
if len(header) != num_fields:
raise Exception()
except:
raise ValueError(u'CSV format is not valid')
for row in reader:
videos.append(dict(zip(fields, row)))
add_team_videos.delay(team.pk, user.pk, videos)
| from django.contrib.contenttypes.models import ContentType
from subtitles.models import SubtitleLanguage
+ from subtitles.signals import subtitles_published
from teams.signals import api_subtitles_approved
from utils.csv_parser import UnicodeReader
from videos.tasks import video_changed_tasks
def complete_approve_tasks(tasks):
lang_ct = ContentType.objects.get_for_model(SubtitleLanguage)
video_ids = set()
for task in tasks:
task.do_complete_approve(lang_ct=lang_ct)
+ version = task.get_subtitle_version()
- api_subtitles_approved.send(task.get_subtitle_version())
? ------------------ - -
+ api_subtitles_approved.send(version)
+ if version.is_public():
+ subtitles_published.send(version.subtitle_language, version=version)
video_ids.add(task.team_video.video_id)
for video_id in video_ids:
video_changed_tasks.delay(video_id)
def add_videos_from_csv(team, user, csv_file):
from .tasks import add_team_videos
videos = []
fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript']
num_fields = len(fields)
try:
reader = UnicodeReader(csv_file)
header = reader.next()
if len(header) != num_fields:
raise Exception()
except:
raise ValueError(u'CSV format is not valid')
for row in reader:
videos.append(dict(zip(fields, row)))
add_team_videos.delay(team.pk, user.pk, videos) |
d3f0d83b0c783d2f15a6f5eaf6fd4ace426307a6 | tests/__init__.py | tests/__init__.py | import os
import sys
import unittest
def suite():
MODULE_DIR = os.path.join(os.path.dirname(__file__), '..')
MODULE_DIR = os.path.abspath(MODULE_DIR)
sys.path.insert(0, MODULE_DIR)
sys.path.insert(0, os.path.dirname(__file__))
SUB_UNITS = os.path.dirname(__file__)
SUB_UNITS = os.listdir(SUB_UNITS)
SUB_UNITS = [
filename[:-3]
for filename in SUB_UNITS
if filename.startswith('test_')
]
os.chdir(os.path.dirname(__file__))
loader = unittest.TestLoader()
return loader.loadTestsFromNames(SUB_UNITS)
| from os import walk, chdir
from os.path import join, dirname, splitext, abspath, relpath
import sys
import unittest
MODULE_DIR = join(dirname(__file__), '..')
MODULE_DIR = abspath(MODULE_DIR)
def walker(opath='.'):
for path, folders, files in walk(opath):
for filename in files:
if filename.startswith('test_') and filename.endswith('.py'):
rpath = relpath(path, opath)
yield (rpath + '.' + splitext(filename)[0]).strip('.')
def suite():
sys.path.insert(0, MODULE_DIR)
sys.path.insert(0, dirname(__file__))
SUB_UNITS = dirname(__file__)
SUB_UNITS = walker(SUB_UNITS)
chdir(dirname(__file__))
return unittest.TestLoader().loadTestsFromNames(SUB_UNITS)
| Rework for tests in subdirectories | Rework for tests in subdirectories
| Python | mit | Mause/pytransperth,Mause/pytransperth | - import os
+ from os import walk, chdir
+ from os.path import join, dirname, splitext, abspath, relpath
import sys
import unittest
+ MODULE_DIR = join(dirname(__file__), '..')
+ MODULE_DIR = abspath(MODULE_DIR)
+
+
+ def walker(opath='.'):
+ for path, folders, files in walk(opath):
+ for filename in files:
+ if filename.startswith('test_') and filename.endswith('.py'):
+ rpath = relpath(path, opath)
+
+ yield (rpath + '.' + splitext(filename)[0]).strip('.')
+
def suite():
- MODULE_DIR = os.path.join(os.path.dirname(__file__), '..')
- MODULE_DIR = os.path.abspath(MODULE_DIR)
sys.path.insert(0, MODULE_DIR)
- sys.path.insert(0, os.path.dirname(__file__))
+ sys.path.insert(0, dirname(__file__))
- SUB_UNITS = os.path.dirname(__file__)
+ SUB_UNITS = dirname(__file__)
- SUB_UNITS = os.listdir(SUB_UNITS)
+ SUB_UNITS = walker(SUB_UNITS)
- SUB_UNITS = [
- filename[:-3]
- for filename in SUB_UNITS
- if filename.startswith('test_')
- ]
- os.chdir(os.path.dirname(__file__))
+ chdir(dirname(__file__))
- loader = unittest.TestLoader()
- return loader.loadTestsFromNames(SUB_UNITS)
+ return unittest.TestLoader().loadTestsFromNames(SUB_UNITS)
+ | Rework for tests in subdirectories | ## Code Before:
import os
import sys
import unittest
def suite():
MODULE_DIR = os.path.join(os.path.dirname(__file__), '..')
MODULE_DIR = os.path.abspath(MODULE_DIR)
sys.path.insert(0, MODULE_DIR)
sys.path.insert(0, os.path.dirname(__file__))
SUB_UNITS = os.path.dirname(__file__)
SUB_UNITS = os.listdir(SUB_UNITS)
SUB_UNITS = [
filename[:-3]
for filename in SUB_UNITS
if filename.startswith('test_')
]
os.chdir(os.path.dirname(__file__))
loader = unittest.TestLoader()
return loader.loadTestsFromNames(SUB_UNITS)
## Instruction:
Rework for tests in subdirectories
## Code After:
from os import walk, chdir
from os.path import join, dirname, splitext, abspath, relpath
import sys
import unittest
MODULE_DIR = join(dirname(__file__), '..')
MODULE_DIR = abspath(MODULE_DIR)
def walker(opath='.'):
for path, folders, files in walk(opath):
for filename in files:
if filename.startswith('test_') and filename.endswith('.py'):
rpath = relpath(path, opath)
yield (rpath + '.' + splitext(filename)[0]).strip('.')
def suite():
sys.path.insert(0, MODULE_DIR)
sys.path.insert(0, dirname(__file__))
SUB_UNITS = dirname(__file__)
SUB_UNITS = walker(SUB_UNITS)
chdir(dirname(__file__))
return unittest.TestLoader().loadTestsFromNames(SUB_UNITS)
| - import os
+ from os import walk, chdir
+ from os.path import join, dirname, splitext, abspath, relpath
import sys
import unittest
+ MODULE_DIR = join(dirname(__file__), '..')
+ MODULE_DIR = abspath(MODULE_DIR)
+
+
+ def walker(opath='.'):
+ for path, folders, files in walk(opath):
+ for filename in files:
+ if filename.startswith('test_') and filename.endswith('.py'):
+ rpath = relpath(path, opath)
+
+ yield (rpath + '.' + splitext(filename)[0]).strip('.')
+
def suite():
- MODULE_DIR = os.path.join(os.path.dirname(__file__), '..')
- MODULE_DIR = os.path.abspath(MODULE_DIR)
sys.path.insert(0, MODULE_DIR)
- sys.path.insert(0, os.path.dirname(__file__))
? --------
+ sys.path.insert(0, dirname(__file__))
- SUB_UNITS = os.path.dirname(__file__)
? --------
+ SUB_UNITS = dirname(__file__)
- SUB_UNITS = os.listdir(SUB_UNITS)
? ^^^ ^^^^^
+ SUB_UNITS = walker(SUB_UNITS)
? ^^ ^^
- SUB_UNITS = [
- filename[:-3]
- for filename in SUB_UNITS
- if filename.startswith('test_')
- ]
- os.chdir(os.path.dirname(__file__))
? --- --------
+ chdir(dirname(__file__))
- loader = unittest.TestLoader()
+
- return loader.loadTestsFromNames(SUB_UNITS)
? ^
+ return unittest.TestLoader().loadTestsFromNames(SUB_UNITS)
? ^^^^^^^^^^^^^^ ++
|
725b3a9db33c90187b913123deefeb180c7fee4c | client/app.py | client/app.py |
import argparse
from server import *
from commandRunner import *
class App:
def __init__(self, baseurl, clientid):
self.server = Server(baseurl, clientid)
def run(self):
runner = CommandRunner()
command = self.server.get()
while command is not None:
response = runner.run(command)
self.server.send(response)
command = self.server.get()
def parseCommandLine():
parser = argparse.ArgumentParser()
parser.add_argument("--baseurl", required=True)
parser.add_argument("--clientid", required=True)
return parser.parse_args()
if __name__ == '__main__':
args = parseCommandLine()
app = App(args.baseurl, args.clientid);
app.run() |
import argparse
from server import *
from commandRunner import *
class App:
server = None
runner = None
def __init__(self, baseurl, clientid):
self.server = Server(baseurl, clientid)
self.runner = CommandRunner()
def run(self):
command = self.server.get()
while command is not None:
response = self.runner.run(command)
self.server.send(response)
command = self.server.get()
def parseCommandLine():
parser = argparse.ArgumentParser()
parser.add_argument("--baseurl", required=True)
parser.add_argument("--clientid", required=True)
return parser.parse_args()
if __name__ == '__main__':
args = parseCommandLine()
app = App(args.baseurl, args.clientid);
app.run() | Add DI to App object | Add DI to App object
| Python | mit | CaminsTECH/owncloud-test |
import argparse
from server import *
from commandRunner import *
class App:
+ server = None
+ runner = None
+
def __init__(self, baseurl, clientid):
self.server = Server(baseurl, clientid)
-
+ self.runner = CommandRunner()
+
- def run(self):
+ def run(self):
- runner = CommandRunner()
command = self.server.get()
while command is not None:
- response = runner.run(command)
+ response = self.runner.run(command)
self.server.send(response)
command = self.server.get()
def parseCommandLine():
parser = argparse.ArgumentParser()
parser.add_argument("--baseurl", required=True)
parser.add_argument("--clientid", required=True)
return parser.parse_args()
if __name__ == '__main__':
args = parseCommandLine()
app = App(args.baseurl, args.clientid);
app.run() | Add DI to App object | ## Code Before:
import argparse
from server import *
from commandRunner import *
class App:
def __init__(self, baseurl, clientid):
self.server = Server(baseurl, clientid)
def run(self):
runner = CommandRunner()
command = self.server.get()
while command is not None:
response = runner.run(command)
self.server.send(response)
command = self.server.get()
def parseCommandLine():
parser = argparse.ArgumentParser()
parser.add_argument("--baseurl", required=True)
parser.add_argument("--clientid", required=True)
return parser.parse_args()
if __name__ == '__main__':
args = parseCommandLine()
app = App(args.baseurl, args.clientid);
app.run()
## Instruction:
Add DI to App object
## Code After:
import argparse
from server import *
from commandRunner import *
class App:
server = None
runner = None
def __init__(self, baseurl, clientid):
self.server = Server(baseurl, clientid)
self.runner = CommandRunner()
def run(self):
command = self.server.get()
while command is not None:
response = self.runner.run(command)
self.server.send(response)
command = self.server.get()
def parseCommandLine():
parser = argparse.ArgumentParser()
parser.add_argument("--baseurl", required=True)
parser.add_argument("--clientid", required=True)
return parser.parse_args()
if __name__ == '__main__':
args = parseCommandLine()
app = App(args.baseurl, args.clientid);
app.run() |
import argparse
from server import *
from commandRunner import *
class App:
+ server = None
+ runner = None
+
def __init__(self, baseurl, clientid):
self.server = Server(baseurl, clientid)
-
+ self.runner = CommandRunner()
+
- def run(self):
+ def run(self):
? ++
- runner = CommandRunner()
command = self.server.get()
while command is not None:
- response = runner.run(command)
+ response = self.runner.run(command)
? +++++
self.server.send(response)
command = self.server.get()
def parseCommandLine():
parser = argparse.ArgumentParser()
parser.add_argument("--baseurl", required=True)
parser.add_argument("--clientid", required=True)
return parser.parse_args()
if __name__ == '__main__':
args = parseCommandLine()
app = App(args.baseurl, args.clientid);
app.run() |
926d5333c1556850a3eda6025ac8cf471b67c0a3 | condor/probes/setup.py | condor/probes/setup.py |
from distutils.core import setup
setup(name='htcondor-es-probes',
version='0.6.3',
description='HTCondor probes for Elasticsearch analytics',
author='Suchandra Thapa',
author_email='sthapa@ci.uchicago.edu',
url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor',
packages=['probe_libs'],
scripts=['collect_history_info.py', 'get_job_status.py'],
data_files=[('/etc/init.d/', ['scripts/collect_history']),
('/etc/cron.d/', ['config/schedd_probe']),
('/etc/sysconfig', ['config/collect_history'])],
license = 'Apache 2.0'
)
|
from distutils.core import setup
setup(name='htcondor-es-probes',
version='0.6.3',
description='HTCondor probes for Elasticsearch analytics',
author='Suchandra Thapa',
author_email='sthapa@ci.uchicago.edu',
url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor',
packages=['probe_libs'],
scripts=['collect_history_info.py', 'get_job_status.py'],
data_files=[('/etc/init.d/', ['scripts/collect_history']),
('/etc/cron.d/', ['config/schedd_probe']),
('/var/lib/collect_history', []),
('/etc/sysconfig', ['config/collect_history'])],
license='Apache 2.0'
)
| Add directory for state files | Add directory for state files
| Python | apache-2.0 | DHTC-Tools/logstash-confs,DHTC-Tools/logstash-confs,DHTC-Tools/logstash-confs |
from distutils.core import setup
setup(name='htcondor-es-probes',
version='0.6.3',
description='HTCondor probes for Elasticsearch analytics',
author='Suchandra Thapa',
author_email='sthapa@ci.uchicago.edu',
url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor',
packages=['probe_libs'],
scripts=['collect_history_info.py', 'get_job_status.py'],
data_files=[('/etc/init.d/', ['scripts/collect_history']),
('/etc/cron.d/', ['config/schedd_probe']),
+ ('/var/lib/collect_history', []),
('/etc/sysconfig', ['config/collect_history'])],
- license = 'Apache 2.0'
+ license='Apache 2.0'
)
| Add directory for state files | ## Code Before:
from distutils.core import setup
setup(name='htcondor-es-probes',
version='0.6.3',
description='HTCondor probes for Elasticsearch analytics',
author='Suchandra Thapa',
author_email='sthapa@ci.uchicago.edu',
url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor',
packages=['probe_libs'],
scripts=['collect_history_info.py', 'get_job_status.py'],
data_files=[('/etc/init.d/', ['scripts/collect_history']),
('/etc/cron.d/', ['config/schedd_probe']),
('/etc/sysconfig', ['config/collect_history'])],
license = 'Apache 2.0'
)
## Instruction:
Add directory for state files
## Code After:
from distutils.core import setup
setup(name='htcondor-es-probes',
version='0.6.3',
description='HTCondor probes for Elasticsearch analytics',
author='Suchandra Thapa',
author_email='sthapa@ci.uchicago.edu',
url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor',
packages=['probe_libs'],
scripts=['collect_history_info.py', 'get_job_status.py'],
data_files=[('/etc/init.d/', ['scripts/collect_history']),
('/etc/cron.d/', ['config/schedd_probe']),
('/var/lib/collect_history', []),
('/etc/sysconfig', ['config/collect_history'])],
license='Apache 2.0'
)
|
from distutils.core import setup
setup(name='htcondor-es-probes',
version='0.6.3',
description='HTCondor probes for Elasticsearch analytics',
author='Suchandra Thapa',
author_email='sthapa@ci.uchicago.edu',
url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor',
packages=['probe_libs'],
scripts=['collect_history_info.py', 'get_job_status.py'],
data_files=[('/etc/init.d/', ['scripts/collect_history']),
('/etc/cron.d/', ['config/schedd_probe']),
+ ('/var/lib/collect_history', []),
('/etc/sysconfig', ['config/collect_history'])],
- license = 'Apache 2.0'
? - -
+ license='Apache 2.0'
) |
9c4ecf0b72d86ae113fa13f210c543120635b73e | board.py | board.py | import numpy
"""
Board represents a four in a row game board.
Author: Isaac Arvestad
"""
class Board:
"""
Initializes the game with a certain number of rows
and columns.
"""
def __init(self, rows, columns):
self.rows = rows
self.columns = columns
self.boardMatrix = numpy.zeros((rows, columns))
"""
Attempts to add a piece to a certain column. If the column is
full the move is illegal and false is returned, otherwise true
is returned.
"""
def addPiece(self, column, value):
"Check if column is full."
if self.boardMatrix.item(0,column) != 0:
return false
"Place piece."
for y in range(self.rows):
currentValue = self.boardMatrix.item(y, column)
if currentValue == 0:
if y == rows - 1:
self.boardMatrix.itemset((y, column), value)
else:
continue
return true
| import numpy
"""
Board represents a four in a row game board.
Author: Isaac Arvestad
"""
class Board:
"""
Initializes the game with a certain number of rows
and columns.
"""
def __init__(self, rows, columns):
self.rows = rows
self.columns = columns
self.boardMatrix = numpy.zeros((rows, columns))
"""
Attempts to add a piece to a certain column. If the column is
full the move is illegal and false is returned, otherwise true
is returned.
"""
def addPiece(self, column, value):
"Check if column is full."
if self.boardMatrix.item(0,column) != 0:
return False
"Place piece."
for y in range(self.rows):
currentValue = self.boardMatrix.item(y, column)
if currentValue == 0:
if y == self.rows - 1:
self.boardMatrix.itemset((y, column), value)
else:
continue
return True
| Fix init method, self.rows and capital booleans. | Fix init method, self.rows and capital booleans.
| Python | mit | isaacarvestad/four-in-a-row | import numpy
"""
Board represents a four in a row game board.
Author: Isaac Arvestad
"""
class Board:
"""
Initializes the game with a certain number of rows
and columns.
"""
- def __init(self, rows, columns):
+ def __init__(self, rows, columns):
self.rows = rows
self.columns = columns
self.boardMatrix = numpy.zeros((rows, columns))
"""
Attempts to add a piece to a certain column. If the column is
full the move is illegal and false is returned, otherwise true
is returned.
"""
def addPiece(self, column, value):
"Check if column is full."
if self.boardMatrix.item(0,column) != 0:
- return false
+ return False
"Place piece."
for y in range(self.rows):
currentValue = self.boardMatrix.item(y, column)
if currentValue == 0:
- if y == rows - 1:
+ if y == self.rows - 1:
self.boardMatrix.itemset((y, column), value)
else:
continue
- return true
+ return True
-
| Fix init method, self.rows and capital booleans. | ## Code Before:
import numpy
"""
Board represents a four in a row game board.
Author: Isaac Arvestad
"""
class Board:
"""
Initializes the game with a certain number of rows
and columns.
"""
def __init(self, rows, columns):
self.rows = rows
self.columns = columns
self.boardMatrix = numpy.zeros((rows, columns))
"""
Attempts to add a piece to a certain column. If the column is
full the move is illegal and false is returned, otherwise true
is returned.
"""
def addPiece(self, column, value):
"Check if column is full."
if self.boardMatrix.item(0,column) != 0:
return false
"Place piece."
for y in range(self.rows):
currentValue = self.boardMatrix.item(y, column)
if currentValue == 0:
if y == rows - 1:
self.boardMatrix.itemset((y, column), value)
else:
continue
return true
## Instruction:
Fix init method, self.rows and capital booleans.
## Code After:
import numpy
"""
Board represents a four in a row game board.
Author: Isaac Arvestad
"""
class Board:
"""
Initializes the game with a certain number of rows
and columns.
"""
def __init__(self, rows, columns):
self.rows = rows
self.columns = columns
self.boardMatrix = numpy.zeros((rows, columns))
"""
Attempts to add a piece to a certain column. If the column is
full the move is illegal and false is returned, otherwise true
is returned.
"""
def addPiece(self, column, value):
"Check if column is full."
if self.boardMatrix.item(0,column) != 0:
return False
"Place piece."
for y in range(self.rows):
currentValue = self.boardMatrix.item(y, column)
if currentValue == 0:
if y == self.rows - 1:
self.boardMatrix.itemset((y, column), value)
else:
continue
return True
| import numpy
"""
Board represents a four in a row game board.
Author: Isaac Arvestad
"""
class Board:
"""
Initializes the game with a certain number of rows
and columns.
"""
- def __init(self, rows, columns):
+ def __init__(self, rows, columns):
? ++
self.rows = rows
self.columns = columns
self.boardMatrix = numpy.zeros((rows, columns))
"""
Attempts to add a piece to a certain column. If the column is
full the move is illegal and false is returned, otherwise true
is returned.
"""
def addPiece(self, column, value):
"Check if column is full."
if self.boardMatrix.item(0,column) != 0:
- return false
? ^
+ return False
? ^
"Place piece."
for y in range(self.rows):
currentValue = self.boardMatrix.item(y, column)
if currentValue == 0:
- if y == rows - 1:
+ if y == self.rows - 1:
? +++++
self.boardMatrix.itemset((y, column), value)
else:
continue
- return true
? ^
+ return True
? ^
- |
29562b08e436abc8465404e49d9193537721b717 | src/odin/contrib/money/fields.py | src/odin/contrib/money/fields.py | from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ('AmountField', )
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
'invalid': "'%s' value must be a (amount, currency).",
'invalid_currency': "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages['invalid'] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if self.allowed_currencies and value not in EMPTY_VALUES:
if value.currency not in self.allowed_currencies:
msg = self.error_messages['invalid_currency'] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
"invalid": "'%s' value must be a (amount, currency).",
"invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if (
self.allowed_currencies
and (value not in EMPTY_VALUES)
and (value.currency not in self.allowed_currencies)
):
msg = self.error_messages["invalid_currency"] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| Correct issue from Sonar (and black file) | Correct issue from Sonar (and black file)
| Python | bsd-3-clause | python-odin/odin | from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
- __all__ = ('AmountField', )
+ __all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
+
default_error_messages = {
- 'invalid': "'%s' value must be a (amount, currency).",
+ "invalid": "'%s' value must be a (amount, currency).",
- 'invalid_currency': "'%s' currency is not supported.",
+ "invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
- msg = self.error_messages['invalid'] % value
+ msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
- if self.allowed_currencies and value not in EMPTY_VALUES:
+ if (
+ self.allowed_currencies
+ and (value not in EMPTY_VALUES)
- if value.currency not in self.allowed_currencies:
+ and (value.currency not in self.allowed_currencies)
+ ):
- msg = self.error_messages['invalid_currency'] % str(value.currency)
+ msg = self.error_messages["invalid_currency"] % str(value.currency)
- raise exceptions.ValidationError(msg)
+ raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| Correct issue from Sonar (and black file) | ## Code Before:
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ('AmountField', )
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
'invalid': "'%s' value must be a (amount, currency).",
'invalid_currency': "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages['invalid'] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if self.allowed_currencies and value not in EMPTY_VALUES:
if value.currency not in self.allowed_currencies:
msg = self.error_messages['invalid_currency'] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
## Instruction:
Correct issue from Sonar (and black file)
## Code After:
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
"invalid": "'%s' value must be a (amount, currency).",
"invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if (
self.allowed_currencies
and (value not in EMPTY_VALUES)
and (value.currency not in self.allowed_currencies)
):
msg = self.error_messages["invalid_currency"] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
- __all__ = ('AmountField', )
? ^ ^ -
+ __all__ = ("AmountField",)
? ^ ^
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
+
default_error_messages = {
- 'invalid': "'%s' value must be a (amount, currency).",
? ^ ^
+ "invalid": "'%s' value must be a (amount, currency).",
? ^ ^
- 'invalid_currency': "'%s' currency is not supported.",
? ^ ^
+ "invalid_currency": "'%s' currency is not supported.",
? ^ ^
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
- msg = self.error_messages['invalid'] % value
? ^ ^
+ msg = self.error_messages["invalid"] % value
? ^ ^
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
- if self.allowed_currencies and value not in EMPTY_VALUES:
+ if (
+ self.allowed_currencies
+ and (value not in EMPTY_VALUES)
- if value.currency not in self.allowed_currencies:
? ^^ ^
+ and (value.currency not in self.allowed_currencies)
? ^^^ + ^
+ ):
- msg = self.error_messages['invalid_currency'] % str(value.currency)
? ---- ^ ^
+ msg = self.error_messages["invalid_currency"] % str(value.currency)
? ^ ^
- raise exceptions.ValidationError(msg)
? ----
+ raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code |
3043a2400e46648f01921aad265816d2bcf18211 | test/test_conjunctive_graph.py | test/test_conjunctive_graph.py | from rdflib.graph import ConjunctiveGraph
from rdflib.term import Identifier, URIRef
from rdflib.parser import StringInputSource
from os import path
DATA = u"""
<http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> .
"""
PUBLIC_ID = u"http://example.org/record/1"
def test_graph_ids():
def check(kws):
cg = ConjunctiveGraph()
cg.parse(**kws)
for g in cg.contexts():
gid = g.identifier
assert isinstance(gid, Identifier)
yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle")
source = StringInputSource(DATA)
source.setPublicId(PUBLIC_ID)
yield check, dict(source=source, format='turtle')
if __name__ == '__main__':
import nose
nose.main(defaultTest=__name__)
| from rdflib.graph import ConjunctiveGraph
from rdflib.term import Identifier, URIRef
from rdflib.parser import StringInputSource
from os import path
DATA = u"""
<http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> .
"""
PUBLIC_ID = u"http://example.org/record/1"
def test_graph_ids():
def check(kws):
cg = ConjunctiveGraph()
cg.parse(**kws)
for g in cg.contexts():
gid = g.identifier
assert isinstance(gid, Identifier)
yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle")
source = StringInputSource(DATA.encode('utf8'))
source.setPublicId(PUBLIC_ID)
yield check, dict(source=source, format='turtle')
if __name__ == '__main__':
import nose
nose.main(defaultTest=__name__)
| Fix py3-incompatible test code that causes PY3 test failure. | Fix py3-incompatible test code that causes PY3 test failure.
| Python | bsd-3-clause | avorio/rdflib,yingerj/rdflib,avorio/rdflib,RDFLib/rdflib,ssssam/rdflib,marma/rdflib,marma/rdflib,avorio/rdflib,armandobs14/rdflib,RDFLib/rdflib,dbs/rdflib,armandobs14/rdflib,yingerj/rdflib,yingerj/rdflib,ssssam/rdflib,RDFLib/rdflib,dbs/rdflib,yingerj/rdflib,marma/rdflib,dbs/rdflib,armandobs14/rdflib,RDFLib/rdflib,avorio/rdflib,marma/rdflib,ssssam/rdflib,armandobs14/rdflib,ssssam/rdflib,dbs/rdflib | from rdflib.graph import ConjunctiveGraph
from rdflib.term import Identifier, URIRef
from rdflib.parser import StringInputSource
from os import path
DATA = u"""
<http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> .
"""
PUBLIC_ID = u"http://example.org/record/1"
def test_graph_ids():
def check(kws):
cg = ConjunctiveGraph()
cg.parse(**kws)
for g in cg.contexts():
gid = g.identifier
assert isinstance(gid, Identifier)
yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle")
- source = StringInputSource(DATA)
+ source = StringInputSource(DATA.encode('utf8'))
source.setPublicId(PUBLIC_ID)
yield check, dict(source=source, format='turtle')
if __name__ == '__main__':
import nose
nose.main(defaultTest=__name__)
| Fix py3-incompatible test code that causes PY3 test failure. | ## Code Before:
from rdflib.graph import ConjunctiveGraph
from rdflib.term import Identifier, URIRef
from rdflib.parser import StringInputSource
from os import path
DATA = u"""
<http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> .
"""
PUBLIC_ID = u"http://example.org/record/1"
def test_graph_ids():
def check(kws):
cg = ConjunctiveGraph()
cg.parse(**kws)
for g in cg.contexts():
gid = g.identifier
assert isinstance(gid, Identifier)
yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle")
source = StringInputSource(DATA)
source.setPublicId(PUBLIC_ID)
yield check, dict(source=source, format='turtle')
if __name__ == '__main__':
import nose
nose.main(defaultTest=__name__)
## Instruction:
Fix py3-incompatible test code that causes PY3 test failure.
## Code After:
from rdflib.graph import ConjunctiveGraph
from rdflib.term import Identifier, URIRef
from rdflib.parser import StringInputSource
from os import path
DATA = u"""
<http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> .
"""
PUBLIC_ID = u"http://example.org/record/1"
def test_graph_ids():
def check(kws):
cg = ConjunctiveGraph()
cg.parse(**kws)
for g in cg.contexts():
gid = g.identifier
assert isinstance(gid, Identifier)
yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle")
source = StringInputSource(DATA.encode('utf8'))
source.setPublicId(PUBLIC_ID)
yield check, dict(source=source, format='turtle')
if __name__ == '__main__':
import nose
nose.main(defaultTest=__name__)
| from rdflib.graph import ConjunctiveGraph
from rdflib.term import Identifier, URIRef
from rdflib.parser import StringInputSource
from os import path
DATA = u"""
<http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> .
"""
PUBLIC_ID = u"http://example.org/record/1"
def test_graph_ids():
def check(kws):
cg = ConjunctiveGraph()
cg.parse(**kws)
for g in cg.contexts():
gid = g.identifier
assert isinstance(gid, Identifier)
yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle")
- source = StringInputSource(DATA)
+ source = StringInputSource(DATA.encode('utf8'))
? ++++++++++++++ +
source.setPublicId(PUBLIC_ID)
yield check, dict(source=source, format='turtle')
if __name__ == '__main__':
import nose
nose.main(defaultTest=__name__) |
c1756ab481f3bf72ab33465c8eb1d5a3e729ce4e | model_logging/migrations/0003_data_migration.py | model_logging/migrations/0003_data_migration.py | from django.db import migrations
app = 'model_logging'
model = 'LogEntry'
def move_data(apps, schema_editor):
LogEntry = apps.get_model(app, model)
for entry in LogEntry.objects.all():
entry.data_temp = entry.data
entry.save()
class Migration(migrations.Migration):
dependencies = [
('model_logging', '0002_add_new_data_field'),
]
operations = [
migrations.RunPython(move_data),
]
| from django.db import migrations
app = 'model_logging'
model = 'LogEntry'
def move_data(apps, schema_editor):
try:
from pgcrypto.fields import TextPGPPublicKeyField
except ImportError:
raise ImportError('Please install django-pgcrypto-fields to perform migration')
LogEntry = apps.get_model(app, model)
for entry in LogEntry.objects.all():
entry.data_temp = entry.data
entry.save()
class Migration(migrations.Migration):
dependencies = [
('model_logging', '0002_add_new_data_field'),
]
operations = [
migrations.RunPython(move_data),
]
| Add try, catch statement to ensure data migration can be performed. | Add try, catch statement to ensure data migration can be performed.
| Python | bsd-2-clause | incuna/django-model-logging | from django.db import migrations
app = 'model_logging'
model = 'LogEntry'
def move_data(apps, schema_editor):
+ try:
+ from pgcrypto.fields import TextPGPPublicKeyField
+ except ImportError:
+ raise ImportError('Please install django-pgcrypto-fields to perform migration')
+
LogEntry = apps.get_model(app, model)
for entry in LogEntry.objects.all():
entry.data_temp = entry.data
entry.save()
class Migration(migrations.Migration):
dependencies = [
('model_logging', '0002_add_new_data_field'),
]
operations = [
migrations.RunPython(move_data),
]
| Add try, catch statement to ensure data migration can be performed. | ## Code Before:
from django.db import migrations
app = 'model_logging'
model = 'LogEntry'
def move_data(apps, schema_editor):
LogEntry = apps.get_model(app, model)
for entry in LogEntry.objects.all():
entry.data_temp = entry.data
entry.save()
class Migration(migrations.Migration):
dependencies = [
('model_logging', '0002_add_new_data_field'),
]
operations = [
migrations.RunPython(move_data),
]
## Instruction:
Add try, catch statement to ensure data migration can be performed.
## Code After:
from django.db import migrations
app = 'model_logging'
model = 'LogEntry'
def move_data(apps, schema_editor):
try:
from pgcrypto.fields import TextPGPPublicKeyField
except ImportError:
raise ImportError('Please install django-pgcrypto-fields to perform migration')
LogEntry = apps.get_model(app, model)
for entry in LogEntry.objects.all():
entry.data_temp = entry.data
entry.save()
class Migration(migrations.Migration):
dependencies = [
('model_logging', '0002_add_new_data_field'),
]
operations = [
migrations.RunPython(move_data),
]
| from django.db import migrations
app = 'model_logging'
model = 'LogEntry'
def move_data(apps, schema_editor):
+ try:
+ from pgcrypto.fields import TextPGPPublicKeyField
+ except ImportError:
+ raise ImportError('Please install django-pgcrypto-fields to perform migration')
+
LogEntry = apps.get_model(app, model)
for entry in LogEntry.objects.all():
entry.data_temp = entry.data
entry.save()
class Migration(migrations.Migration):
dependencies = [
('model_logging', '0002_add_new_data_field'),
]
operations = [
migrations.RunPython(move_data),
] |
1090acb35ea4ce5c8d17db716539d3354feabc12 | nodeconductor/iaas/migrations/0038_securitygroup_state.py | nodeconductor/iaas/migrations/0038_securitygroup_state.py | from __future__ import unicode_literals
from django.db import models, migrations
import django_fsm
class Migration(migrations.Migration):
dependencies = [
('iaas', '0037_init_security_groups_quotas'),
]
operations = [
migrations.AddField(
model_name='securitygroup',
name='state',
field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]),
preserve_default=True,
),
]
| from __future__ import unicode_literals
from django.db import models, migrations
import django_fsm
def mark_security_groups_as_synced(apps, schema_editor):
SecurityGroup = apps.get_model('iaas', 'SecurityGroup')
SecurityGroup.objects.all().update(state=3)
class Migration(migrations.Migration):
dependencies = [
('iaas', '0037_init_security_groups_quotas'),
]
operations = [
migrations.AddField(
model_name='securitygroup',
name='state',
field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]),
preserve_default=True,
),
migrations.RunPython(mark_security_groups_as_synced),
]
| Mark all exist security groups as synced | Mark all exist security groups as synced
- itacloud-4843
| Python | mit | opennode/nodeconductor,opennode/nodeconductor,opennode/nodeconductor | from __future__ import unicode_literals
from django.db import models, migrations
import django_fsm
+
+
+ def mark_security_groups_as_synced(apps, schema_editor):
+ SecurityGroup = apps.get_model('iaas', 'SecurityGroup')
+ SecurityGroup.objects.all().update(state=3)
class Migration(migrations.Migration):
dependencies = [
('iaas', '0037_init_security_groups_quotas'),
]
operations = [
migrations.AddField(
model_name='securitygroup',
name='state',
field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]),
preserve_default=True,
),
+ migrations.RunPython(mark_security_groups_as_synced),
]
| Mark all exist security groups as synced | ## Code Before:
from __future__ import unicode_literals
from django.db import models, migrations
import django_fsm
class Migration(migrations.Migration):
dependencies = [
('iaas', '0037_init_security_groups_quotas'),
]
operations = [
migrations.AddField(
model_name='securitygroup',
name='state',
field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]),
preserve_default=True,
),
]
## Instruction:
Mark all exist security groups as synced
## Code After:
from __future__ import unicode_literals
from django.db import models, migrations
import django_fsm
def mark_security_groups_as_synced(apps, schema_editor):
SecurityGroup = apps.get_model('iaas', 'SecurityGroup')
SecurityGroup.objects.all().update(state=3)
class Migration(migrations.Migration):
dependencies = [
('iaas', '0037_init_security_groups_quotas'),
]
operations = [
migrations.AddField(
model_name='securitygroup',
name='state',
field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]),
preserve_default=True,
),
migrations.RunPython(mark_security_groups_as_synced),
]
| from __future__ import unicode_literals
from django.db import models, migrations
import django_fsm
+
+
+ def mark_security_groups_as_synced(apps, schema_editor):
+ SecurityGroup = apps.get_model('iaas', 'SecurityGroup')
+ SecurityGroup.objects.all().update(state=3)
class Migration(migrations.Migration):
dependencies = [
('iaas', '0037_init_security_groups_quotas'),
]
operations = [
migrations.AddField(
model_name='securitygroup',
name='state',
field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]),
preserve_default=True,
),
+ migrations.RunPython(mark_security_groups_as_synced),
] |
1cda977eff5a2edaa0de82882ef2e7d1611329b7 | tests/test_protocol.py | tests/test_protocol.py | import pytest
class TestProtocol:
@pytest.mark.asyncio
def test_server_hello(self, ws_client_factory, get_unencrypted_packet):
"""
The server must send a valid `server-hello` on connection.
"""
client = yield from ws_client_factory()
receiver, message = yield from get_unencrypted_packet(client)
assert receiver == 0x00
assert message['type'] == 'server-hello'
assert len(message['key']) == 32
assert len(message['my-cookie']) == 16
yield from client.close()
| import asyncio
import pytest
import saltyrtc
class TestProtocol:
@pytest.mark.asyncio
def test_no_subprotocols(self, ws_client_factory):
"""
The server must drop the client after the connection has been
established with a close code of *1002*.
"""
client = yield from ws_client_factory(subprotocols=None)
yield from asyncio.sleep(0.05)
assert not client.open
assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
@pytest.mark.asyncio
def test_invalid_subprotocols(self, ws_client_factory):
"""
The server must drop the client after the connection has been
established with a close code of *1002*.
"""
client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000'])
yield from asyncio.sleep(0.05)
assert not client.open
assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
@pytest.mark.asyncio
def test_server_hello(self, ws_client_factory, get_unencrypted_packet):
"""
The server must send a valid `server-hello` on connection.
"""
client = yield from ws_client_factory()
receiver, message = yield from get_unencrypted_packet(client)
assert receiver == 0x00
assert message['type'] == 'server-hello'
assert len(message['key']) == 32
assert len(message['my-cookie']) == 16
yield from client.close()
| Add tests for invalid and no provided sub-protocols | Add tests for invalid and no provided sub-protocols
| Python | mit | saltyrtc/saltyrtc-server-python,saltyrtc/saltyrtc-server-python | + import asyncio
+
import pytest
+
+ import saltyrtc
class TestProtocol:
+ @pytest.mark.asyncio
+ def test_no_subprotocols(self, ws_client_factory):
+ """
+ The server must drop the client after the connection has been
+ established with a close code of *1002*.
+ """
+ client = yield from ws_client_factory(subprotocols=None)
+ yield from asyncio.sleep(0.05)
+ assert not client.open
+ assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
+
+ @pytest.mark.asyncio
+ def test_invalid_subprotocols(self, ws_client_factory):
+ """
+ The server must drop the client after the connection has been
+ established with a close code of *1002*.
+ """
+ client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000'])
+ yield from asyncio.sleep(0.05)
+ assert not client.open
+ assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
+
@pytest.mark.asyncio
def test_server_hello(self, ws_client_factory, get_unencrypted_packet):
"""
The server must send a valid `server-hello` on connection.
"""
client = yield from ws_client_factory()
receiver, message = yield from get_unencrypted_packet(client)
assert receiver == 0x00
assert message['type'] == 'server-hello'
assert len(message['key']) == 32
assert len(message['my-cookie']) == 16
yield from client.close()
| Add tests for invalid and no provided sub-protocols | ## Code Before:
import pytest
class TestProtocol:
@pytest.mark.asyncio
def test_server_hello(self, ws_client_factory, get_unencrypted_packet):
"""
The server must send a valid `server-hello` on connection.
"""
client = yield from ws_client_factory()
receiver, message = yield from get_unencrypted_packet(client)
assert receiver == 0x00
assert message['type'] == 'server-hello'
assert len(message['key']) == 32
assert len(message['my-cookie']) == 16
yield from client.close()
## Instruction:
Add tests for invalid and no provided sub-protocols
## Code After:
import asyncio
import pytest
import saltyrtc
class TestProtocol:
@pytest.mark.asyncio
def test_no_subprotocols(self, ws_client_factory):
"""
The server must drop the client after the connection has been
established with a close code of *1002*.
"""
client = yield from ws_client_factory(subprotocols=None)
yield from asyncio.sleep(0.05)
assert not client.open
assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
@pytest.mark.asyncio
def test_invalid_subprotocols(self, ws_client_factory):
"""
The server must drop the client after the connection has been
established with a close code of *1002*.
"""
client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000'])
yield from asyncio.sleep(0.05)
assert not client.open
assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
@pytest.mark.asyncio
def test_server_hello(self, ws_client_factory, get_unencrypted_packet):
"""
The server must send a valid `server-hello` on connection.
"""
client = yield from ws_client_factory()
receiver, message = yield from get_unencrypted_packet(client)
assert receiver == 0x00
assert message['type'] == 'server-hello'
assert len(message['key']) == 32
assert len(message['my-cookie']) == 16
yield from client.close()
| + import asyncio
+
import pytest
+
+ import saltyrtc
class TestProtocol:
+ @pytest.mark.asyncio
+ def test_no_subprotocols(self, ws_client_factory):
+ """
+ The server must drop the client after the connection has been
+ established with a close code of *1002*.
+ """
+ client = yield from ws_client_factory(subprotocols=None)
+ yield from asyncio.sleep(0.05)
+ assert not client.open
+ assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
+
+ @pytest.mark.asyncio
+ def test_invalid_subprotocols(self, ws_client_factory):
+ """
+ The server must drop the client after the connection has been
+ established with a close code of *1002*.
+ """
+ client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000'])
+ yield from asyncio.sleep(0.05)
+ assert not client.open
+ assert client.close_code == saltyrtc.CloseCode.sub_protocol_error
+
@pytest.mark.asyncio
def test_server_hello(self, ws_client_factory, get_unencrypted_packet):
"""
The server must send a valid `server-hello` on connection.
"""
client = yield from ws_client_factory()
receiver, message = yield from get_unencrypted_packet(client)
assert receiver == 0x00
assert message['type'] == 'server-hello'
assert len(message['key']) == 32
assert len(message['my-cookie']) == 16
yield from client.close() |
e7cb5b0be49bc5e811809c56eb4ad3c0dc861cdf | examples/child_watcher.py | examples/child_watcher.py | import logging
import random
from tornado import gen
from zoonado import exc
log = logging.getLogger()
def arguments(parser):
parser.add_argument(
"--path", "-p", type=str, default="/examplewatcher",
help="ZNode path to use for the example."
)
def watcher_callback(children):
children.sort()
log.info("There are %d items now: %s", len(children), children)
@gen.coroutine
def run(client, args):
yield client.start()
try:
yield client.create(args.path)
except exc.NodeExists:
pass
watcher = client.recipes.ChildrenWatcher()
watcher.add_callback(args.path, watcher_callback)
to_make = ["cat", "dog", "mouse", "human"]
random.shuffle(to_make)
for item in to_make:
yield client.create(args.path + "/" + item, ephemeral=True)
yield gen.sleep(1)
for item in to_make:
yield client.delete(args.path + "/" + item)
| import logging
import random
from tornado import gen
from zoonado import exc
log = logging.getLogger()
def arguments(parser):
parser.add_argument(
"--path", "-p", type=str, default="/examplewatcher",
help="ZNode path to use for the example."
)
def watcher_callback(children):
children.sort()
log.info("There are %d items now: %s", len(children), ", ".join(children))
@gen.coroutine
def run(client, args):
yield client.start()
try:
yield client.create(args.path)
except exc.NodeExists:
pass
watcher = client.recipes.ChildrenWatcher()
watcher.add_callback(args.path, watcher_callback)
to_make = ["cat", "dog", "mouse", "human"]
random.shuffle(to_make)
for item in to_make:
yield client.create(args.path + "/" + item, ephemeral=True)
yield gen.sleep(1)
for item in to_make:
yield client.delete(args.path + "/" + item)
yield gen.sleep(1)
| Fix up to the child watcher example. | Fix up to the child watcher example.
Without yielding to the ioloop after each call to client.delete() the child
znodes would be deleted but that would never be reported.
| Python | apache-2.0 | wglass/zoonado | import logging
import random
from tornado import gen
from zoonado import exc
log = logging.getLogger()
def arguments(parser):
parser.add_argument(
"--path", "-p", type=str, default="/examplewatcher",
help="ZNode path to use for the example."
)
def watcher_callback(children):
children.sort()
- log.info("There are %d items now: %s", len(children), children)
+ log.info("There are %d items now: %s", len(children), ", ".join(children))
@gen.coroutine
def run(client, args):
yield client.start()
try:
yield client.create(args.path)
except exc.NodeExists:
pass
watcher = client.recipes.ChildrenWatcher()
watcher.add_callback(args.path, watcher_callback)
to_make = ["cat", "dog", "mouse", "human"]
random.shuffle(to_make)
for item in to_make:
yield client.create(args.path + "/" + item, ephemeral=True)
yield gen.sleep(1)
for item in to_make:
yield client.delete(args.path + "/" + item)
+ yield gen.sleep(1)
| Fix up to the child watcher example. | ## Code Before:
import logging
import random
from tornado import gen
from zoonado import exc
log = logging.getLogger()
def arguments(parser):
parser.add_argument(
"--path", "-p", type=str, default="/examplewatcher",
help="ZNode path to use for the example."
)
def watcher_callback(children):
children.sort()
log.info("There are %d items now: %s", len(children), children)
@gen.coroutine
def run(client, args):
yield client.start()
try:
yield client.create(args.path)
except exc.NodeExists:
pass
watcher = client.recipes.ChildrenWatcher()
watcher.add_callback(args.path, watcher_callback)
to_make = ["cat", "dog", "mouse", "human"]
random.shuffle(to_make)
for item in to_make:
yield client.create(args.path + "/" + item, ephemeral=True)
yield gen.sleep(1)
for item in to_make:
yield client.delete(args.path + "/" + item)
## Instruction:
Fix up to the child watcher example.
## Code After:
import logging
import random
from tornado import gen
from zoonado import exc
log = logging.getLogger()
def arguments(parser):
parser.add_argument(
"--path", "-p", type=str, default="/examplewatcher",
help="ZNode path to use for the example."
)
def watcher_callback(children):
children.sort()
log.info("There are %d items now: %s", len(children), ", ".join(children))
@gen.coroutine
def run(client, args):
yield client.start()
try:
yield client.create(args.path)
except exc.NodeExists:
pass
watcher = client.recipes.ChildrenWatcher()
watcher.add_callback(args.path, watcher_callback)
to_make = ["cat", "dog", "mouse", "human"]
random.shuffle(to_make)
for item in to_make:
yield client.create(args.path + "/" + item, ephemeral=True)
yield gen.sleep(1)
for item in to_make:
yield client.delete(args.path + "/" + item)
yield gen.sleep(1)
| import logging
import random
from tornado import gen
from zoonado import exc
log = logging.getLogger()
def arguments(parser):
parser.add_argument(
"--path", "-p", type=str, default="/examplewatcher",
help="ZNode path to use for the example."
)
def watcher_callback(children):
children.sort()
- log.info("There are %d items now: %s", len(children), children)
+ log.info("There are %d items now: %s", len(children), ", ".join(children))
? ++++++++++ +
@gen.coroutine
def run(client, args):
yield client.start()
try:
yield client.create(args.path)
except exc.NodeExists:
pass
watcher = client.recipes.ChildrenWatcher()
watcher.add_callback(args.path, watcher_callback)
to_make = ["cat", "dog", "mouse", "human"]
random.shuffle(to_make)
for item in to_make:
yield client.create(args.path + "/" + item, ephemeral=True)
yield gen.sleep(1)
for item in to_make:
yield client.delete(args.path + "/" + item)
+ yield gen.sleep(1) |
347545cc7ece8c0763ef194654fbaa34d16efe54 | styleguide/views.py | styleguide/views.py | from django.shortcuts import get_object_or_404, render
from django import forms
class ExampleForm(forms.Form):
text = forms.CharField()
disabled_text = forms.CharField(disabled=True)
readonly_text = forms.CharField(
widget=forms.TextInput(attrs={'readonly':'readonly'})
)
checkbox1 = forms.BooleanField()
checkbox2 = forms.BooleanField()
select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')])
radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect())
form_initial = {
"text": "",
"disabled_text": "This field can't be changed",
}
def styleguide(request):
return render(request, "styleguide/styleguide.html", {
})
def styleguide_page(request, name):
return render(request, "styleguide/styleguide-%s.html" % name, {
"example_form": ExampleForm(initial=form_initial),
})
def styleguide_sub_page(request, name, sub_page):
return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), {
"example_form": ExampleForm(initial=form_initial),
})
| from django.shortcuts import get_object_or_404, render
from django import forms
class ExampleForm(forms.Form):
text = forms.CharField()
disabled_text = forms.CharField(disabled=True)
readonly_text = forms.CharField(
widget=forms.TextInput(attrs={'readonly':'readonly'})
)
checkbox1 = forms.BooleanField()
checkbox2 = forms.BooleanField()
select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')])
radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect())
form_initial = {
"text": "",
"disabled_text": "This field can't be changed",
"readonly_text": "This field is read only",
}
def styleguide(request):
return render(request, "styleguide/styleguide.html", {
})
def styleguide_page(request, name):
return render(request, "styleguide/styleguide-%s.html" % name, {
"example_form": ExampleForm(initial=form_initial),
})
def styleguide_sub_page(request, name, sub_page):
return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), {
"example_form": ExampleForm(initial=form_initial),
})
| Add readonly text to form | Add readonly text to form
| Python | bsd-3-clause | caktus/django-styleguide,caktus/django-styleguide,caktus/django-styleguide | from django.shortcuts import get_object_or_404, render
from django import forms
class ExampleForm(forms.Form):
text = forms.CharField()
disabled_text = forms.CharField(disabled=True)
readonly_text = forms.CharField(
widget=forms.TextInput(attrs={'readonly':'readonly'})
)
checkbox1 = forms.BooleanField()
checkbox2 = forms.BooleanField()
select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')])
radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect())
form_initial = {
"text": "",
"disabled_text": "This field can't be changed",
+ "readonly_text": "This field is read only",
}
def styleguide(request):
return render(request, "styleguide/styleguide.html", {
})
def styleguide_page(request, name):
return render(request, "styleguide/styleguide-%s.html" % name, {
"example_form": ExampleForm(initial=form_initial),
})
def styleguide_sub_page(request, name, sub_page):
return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), {
"example_form": ExampleForm(initial=form_initial),
})
| Add readonly text to form | ## Code Before:
from django.shortcuts import get_object_or_404, render
from django import forms
class ExampleForm(forms.Form):
text = forms.CharField()
disabled_text = forms.CharField(disabled=True)
readonly_text = forms.CharField(
widget=forms.TextInput(attrs={'readonly':'readonly'})
)
checkbox1 = forms.BooleanField()
checkbox2 = forms.BooleanField()
select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')])
radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect())
form_initial = {
"text": "",
"disabled_text": "This field can't be changed",
}
def styleguide(request):
return render(request, "styleguide/styleguide.html", {
})
def styleguide_page(request, name):
return render(request, "styleguide/styleguide-%s.html" % name, {
"example_form": ExampleForm(initial=form_initial),
})
def styleguide_sub_page(request, name, sub_page):
return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), {
"example_form": ExampleForm(initial=form_initial),
})
## Instruction:
Add readonly text to form
## Code After:
from django.shortcuts import get_object_or_404, render
from django import forms
class ExampleForm(forms.Form):
text = forms.CharField()
disabled_text = forms.CharField(disabled=True)
readonly_text = forms.CharField(
widget=forms.TextInput(attrs={'readonly':'readonly'})
)
checkbox1 = forms.BooleanField()
checkbox2 = forms.BooleanField()
select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')])
radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect())
form_initial = {
"text": "",
"disabled_text": "This field can't be changed",
"readonly_text": "This field is read only",
}
def styleguide(request):
return render(request, "styleguide/styleguide.html", {
})
def styleguide_page(request, name):
return render(request, "styleguide/styleguide-%s.html" % name, {
"example_form": ExampleForm(initial=form_initial),
})
def styleguide_sub_page(request, name, sub_page):
return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), {
"example_form": ExampleForm(initial=form_initial),
})
| from django.shortcuts import get_object_or_404, render
from django import forms
class ExampleForm(forms.Form):
text = forms.CharField()
disabled_text = forms.CharField(disabled=True)
readonly_text = forms.CharField(
widget=forms.TextInput(attrs={'readonly':'readonly'})
)
checkbox1 = forms.BooleanField()
checkbox2 = forms.BooleanField()
select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')])
radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect())
form_initial = {
"text": "",
"disabled_text": "This field can't be changed",
+ "readonly_text": "This field is read only",
}
def styleguide(request):
return render(request, "styleguide/styleguide.html", {
})
def styleguide_page(request, name):
return render(request, "styleguide/styleguide-%s.html" % name, {
"example_form": ExampleForm(initial=form_initial),
})
def styleguide_sub_page(request, name, sub_page):
return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), {
"example_form": ExampleForm(initial=form_initial),
}) |
7a448c4df3feb717d0b1d8abbf9d32237751aab5 | nbgrader/tests/apps/test_nbgrader_extension.py | nbgrader/tests/apps/test_nbgrader_extension.py | import os
import nbgrader
def test_nbextension():
from nbgrader import _jupyter_nbextension_paths
nbexts = _jupyter_nbextension_paths()
assert len(nbexts) == 3
assert nbexts[0]['section'] == 'tree'
assert nbexts[1]['section'] == 'notebook'
assert nbexts[2]['section'] == 'tree'
paths = [ext['src'] for ext in nbexts]
for path in paths:
assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path))
def test_serverextension():
from nbgrader import _jupyter_server_extension_paths
serverexts = _jupyter_server_extension_paths()
assert len(serverexts) == 2
assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list'
assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader'
| import os
import nbgrader
def test_nbextension():
from nbgrader import _jupyter_nbextension_paths
nbexts = _jupyter_nbextension_paths()
assert len(nbexts) == 4
assert nbexts[0]['section'] == 'tree'
assert nbexts[1]['section'] == 'notebook'
assert nbexts[2]['section'] == 'tree'
assert nbexts[3]['section'] == 'notebook'
paths = [ext['src'] for ext in nbexts]
for path in paths:
assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path))
def test_serverextension():
from nbgrader import _jupyter_server_extension_paths
serverexts = _jupyter_server_extension_paths()
assert len(serverexts) == 3
assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list'
assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader'
assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment'
| Fix tests for nbgrader extensions | Fix tests for nbgrader extensions
| Python | bsd-3-clause | jhamrick/nbgrader,jhamrick/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jhamrick/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jhamrick/nbgrader | import os
import nbgrader
def test_nbextension():
from nbgrader import _jupyter_nbextension_paths
nbexts = _jupyter_nbextension_paths()
- assert len(nbexts) == 3
+ assert len(nbexts) == 4
assert nbexts[0]['section'] == 'tree'
assert nbexts[1]['section'] == 'notebook'
assert nbexts[2]['section'] == 'tree'
+ assert nbexts[3]['section'] == 'notebook'
paths = [ext['src'] for ext in nbexts]
for path in paths:
assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path))
def test_serverextension():
from nbgrader import _jupyter_server_extension_paths
serverexts = _jupyter_server_extension_paths()
- assert len(serverexts) == 2
+ assert len(serverexts) == 3
assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list'
assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader'
+ assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment'
| Fix tests for nbgrader extensions | ## Code Before:
import os
import nbgrader
def test_nbextension():
from nbgrader import _jupyter_nbextension_paths
nbexts = _jupyter_nbextension_paths()
assert len(nbexts) == 3
assert nbexts[0]['section'] == 'tree'
assert nbexts[1]['section'] == 'notebook'
assert nbexts[2]['section'] == 'tree'
paths = [ext['src'] for ext in nbexts]
for path in paths:
assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path))
def test_serverextension():
from nbgrader import _jupyter_server_extension_paths
serverexts = _jupyter_server_extension_paths()
assert len(serverexts) == 2
assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list'
assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader'
## Instruction:
Fix tests for nbgrader extensions
## Code After:
import os
import nbgrader
def test_nbextension():
from nbgrader import _jupyter_nbextension_paths
nbexts = _jupyter_nbextension_paths()
assert len(nbexts) == 4
assert nbexts[0]['section'] == 'tree'
assert nbexts[1]['section'] == 'notebook'
assert nbexts[2]['section'] == 'tree'
assert nbexts[3]['section'] == 'notebook'
paths = [ext['src'] for ext in nbexts]
for path in paths:
assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path))
def test_serverextension():
from nbgrader import _jupyter_server_extension_paths
serverexts = _jupyter_server_extension_paths()
assert len(serverexts) == 3
assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list'
assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader'
assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment'
| import os
import nbgrader
def test_nbextension():
from nbgrader import _jupyter_nbextension_paths
nbexts = _jupyter_nbextension_paths()
- assert len(nbexts) == 3
? ^
+ assert len(nbexts) == 4
? ^
assert nbexts[0]['section'] == 'tree'
assert nbexts[1]['section'] == 'notebook'
assert nbexts[2]['section'] == 'tree'
+ assert nbexts[3]['section'] == 'notebook'
paths = [ext['src'] for ext in nbexts]
for path in paths:
assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path))
def test_serverextension():
from nbgrader import _jupyter_server_extension_paths
serverexts = _jupyter_server_extension_paths()
- assert len(serverexts) == 2
? ^
+ assert len(serverexts) == 3
? ^
assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list'
assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader'
+ assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment' |
a1f5a392d5270dd6f80a40e45c5e25b6ae04b7c3 | embed_video/fields.py | embed_video/fields.py | from django.db import models
from django import forms
from django.utils.translation import ugettext_lazy as _
from .backends import detect_backend, UnknownIdException, \
UnknownBackendException
__all__ = ('EmbedVideoField', 'EmbedVideoFormField')
class EmbedVideoField(models.URLField):
"""
Model field for embeded video. Descendant of
:py:class:`django.db.models.URLField`.
"""
def formfield(self, **kwargs):
defaults = {'form_class': EmbedVideoFormField}
defaults.update(kwargs)
return super(EmbedVideoField, self).formfield(**defaults)
def south_field_triple(self):
from south.modelsinspector import introspector
cls_name = '%s.%s' % (
self.__class__.__module__,
self.__class__.__name__
)
args, kwargs = introspector(self)
return (cls_name, args, kwargs)
class EmbedVideoFormField(forms.URLField):
"""
Form field for embeded video. Descendant of
:py:class:`django.forms.URLField`
"""
def validate(self, url):
super(EmbedVideoFormField, self).validate(url)
if url:
try:
detect_backend(url)
except UnknownBackendException:
raise forms.ValidationError(_(u'URL could not be recognized.'))
except UnknownIdException:
raise forms.ValidationError(_(u'ID of this video could not be \
recognized.'))
return url
| from django.db import models
from django import forms
from django.utils.translation import ugettext_lazy as _
from .backends import detect_backend, UnknownIdException, \
UnknownBackendException
__all__ = ('EmbedVideoField', 'EmbedVideoFormField')
class EmbedVideoField(models.URLField):
"""
Model field for embeded video. Descendant of
:py:class:`django.db.models.URLField`.
"""
def formfield(self, **kwargs):
defaults = {'form_class': EmbedVideoFormField}
defaults.update(kwargs)
return super(EmbedVideoField, self).formfield(**defaults)
def south_field_triple(self):
from south.modelsinspector import introspector
cls_name = '%s.%s' % (
self.__class__.__module__,
self.__class__.__name__
)
args, kwargs = introspector(self)
return (cls_name, args, kwargs)
class EmbedVideoFormField(forms.URLField):
"""
Form field for embeded video. Descendant of
:py:class:`django.forms.URLField`
"""
def validate(self, url):
# if empty url is not allowed throws an exception
super(EmbedVideoFormField, self).validate(url)
if not url:
return
try:
detect_backend(url)
except UnknownBackendException:
raise forms.ValidationError(_(u'URL could not be recognized.'))
except UnknownIdException:
raise forms.ValidationError(_(u'ID of this video could not be \
recognized.'))
return url
| Simplify validate method in FormField. | Simplify validate method in FormField. | Python | mit | yetty/django-embed-video,jazzband/django-embed-video,jazzband/django-embed-video,mpachas/django-embed-video,yetty/django-embed-video,mpachas/django-embed-video | from django.db import models
from django import forms
from django.utils.translation import ugettext_lazy as _
from .backends import detect_backend, UnknownIdException, \
UnknownBackendException
__all__ = ('EmbedVideoField', 'EmbedVideoFormField')
class EmbedVideoField(models.URLField):
"""
Model field for embeded video. Descendant of
:py:class:`django.db.models.URLField`.
"""
def formfield(self, **kwargs):
defaults = {'form_class': EmbedVideoFormField}
defaults.update(kwargs)
return super(EmbedVideoField, self).formfield(**defaults)
def south_field_triple(self):
from south.modelsinspector import introspector
cls_name = '%s.%s' % (
self.__class__.__module__,
self.__class__.__name__
)
args, kwargs = introspector(self)
return (cls_name, args, kwargs)
class EmbedVideoFormField(forms.URLField):
"""
Form field for embeded video. Descendant of
:py:class:`django.forms.URLField`
"""
def validate(self, url):
+ # if empty url is not allowed throws an exception
super(EmbedVideoFormField, self).validate(url)
+
+ if not url:
+ return
- if url:
- try:
+ try:
- detect_backend(url)
+ detect_backend(url)
- except UnknownBackendException:
+ except UnknownBackendException:
- raise forms.ValidationError(_(u'URL could not be recognized.'))
+ raise forms.ValidationError(_(u'URL could not be recognized.'))
- except UnknownIdException:
+ except UnknownIdException:
- raise forms.ValidationError(_(u'ID of this video could not be \
+ raise forms.ValidationError(_(u'ID of this video could not be \
- recognized.'))
+ recognized.'))
-
return url
| Simplify validate method in FormField. | ## Code Before:
from django.db import models
from django import forms
from django.utils.translation import ugettext_lazy as _
from .backends import detect_backend, UnknownIdException, \
UnknownBackendException
__all__ = ('EmbedVideoField', 'EmbedVideoFormField')
class EmbedVideoField(models.URLField):
"""
Model field for embeded video. Descendant of
:py:class:`django.db.models.URLField`.
"""
def formfield(self, **kwargs):
defaults = {'form_class': EmbedVideoFormField}
defaults.update(kwargs)
return super(EmbedVideoField, self).formfield(**defaults)
def south_field_triple(self):
from south.modelsinspector import introspector
cls_name = '%s.%s' % (
self.__class__.__module__,
self.__class__.__name__
)
args, kwargs = introspector(self)
return (cls_name, args, kwargs)
class EmbedVideoFormField(forms.URLField):
"""
Form field for embeded video. Descendant of
:py:class:`django.forms.URLField`
"""
def validate(self, url):
super(EmbedVideoFormField, self).validate(url)
if url:
try:
detect_backend(url)
except UnknownBackendException:
raise forms.ValidationError(_(u'URL could not be recognized.'))
except UnknownIdException:
raise forms.ValidationError(_(u'ID of this video could not be \
recognized.'))
return url
## Instruction:
Simplify validate method in FormField.
## Code After:
from django.db import models
from django import forms
from django.utils.translation import ugettext_lazy as _
from .backends import detect_backend, UnknownIdException, \
UnknownBackendException
__all__ = ('EmbedVideoField', 'EmbedVideoFormField')
class EmbedVideoField(models.URLField):
"""
Model field for embeded video. Descendant of
:py:class:`django.db.models.URLField`.
"""
def formfield(self, **kwargs):
defaults = {'form_class': EmbedVideoFormField}
defaults.update(kwargs)
return super(EmbedVideoField, self).formfield(**defaults)
def south_field_triple(self):
from south.modelsinspector import introspector
cls_name = '%s.%s' % (
self.__class__.__module__,
self.__class__.__name__
)
args, kwargs = introspector(self)
return (cls_name, args, kwargs)
class EmbedVideoFormField(forms.URLField):
"""
Form field for embeded video. Descendant of
:py:class:`django.forms.URLField`
"""
def validate(self, url):
# if empty url is not allowed throws an exception
super(EmbedVideoFormField, self).validate(url)
if not url:
return
try:
detect_backend(url)
except UnknownBackendException:
raise forms.ValidationError(_(u'URL could not be recognized.'))
except UnknownIdException:
raise forms.ValidationError(_(u'ID of this video could not be \
recognized.'))
return url
| from django.db import models
from django import forms
from django.utils.translation import ugettext_lazy as _
from .backends import detect_backend, UnknownIdException, \
UnknownBackendException
__all__ = ('EmbedVideoField', 'EmbedVideoFormField')
class EmbedVideoField(models.URLField):
"""
Model field for embeded video. Descendant of
:py:class:`django.db.models.URLField`.
"""
def formfield(self, **kwargs):
defaults = {'form_class': EmbedVideoFormField}
defaults.update(kwargs)
return super(EmbedVideoField, self).formfield(**defaults)
def south_field_triple(self):
from south.modelsinspector import introspector
cls_name = '%s.%s' % (
self.__class__.__module__,
self.__class__.__name__
)
args, kwargs = introspector(self)
return (cls_name, args, kwargs)
class EmbedVideoFormField(forms.URLField):
"""
Form field for embeded video. Descendant of
:py:class:`django.forms.URLField`
"""
def validate(self, url):
+ # if empty url is not allowed throws an exception
super(EmbedVideoFormField, self).validate(url)
+
+ if not url:
+ return
- if url:
- try:
? ----
+ try:
- detect_backend(url)
? ----
+ detect_backend(url)
- except UnknownBackendException:
? ----
+ except UnknownBackendException:
- raise forms.ValidationError(_(u'URL could not be recognized.'))
? ----
+ raise forms.ValidationError(_(u'URL could not be recognized.'))
- except UnknownIdException:
? ----
+ except UnknownIdException:
- raise forms.ValidationError(_(u'ID of this video could not be \
? ----
+ raise forms.ValidationError(_(u'ID of this video could not be \
- recognized.'))
? ----
+ recognized.'))
-
return url |
1f697a2c7bcf0f7769a9fc4f81be676ed5ee97c6 | examples/flask/flask_seguro/cart.py | examples/flask/flask_seguro/cart.py | from flask_seguro.products import Products
from flask import current_app as app
class Cart:
def __init__(self, cart_dict={}):
if cart_dict == {}:
self.total = 0
self.subtotal = 0
self.items = []
else:
self.total = cart_dict["total"]
self.subtotal = cart_dict["subtotal"]
self.items = cart_dict["items"]
self.extra_amount = float(app.config['EXTRA_AMOUNT'])
def to_dict(self):
return {"total": self.total,
"subtotal": self.subtotal,
"items": self.items,
"extra_amount": self.extra_amount}
def change_item(self, item_id, operation):
product = Products().get_one(item_id)
if product:
if operation == 'add':
self.items.append(product)
elif operation == 'remove':
cart_product = filter(
lambda x: x['id'] == product['id'], self.items)
self.items.remove(cart_product[0])
self.update()
return True
else:
return False
def update(self):
subtotal = float(0)
total = float(0)
for product in self.items:
subtotal += float(product["price"])
if subtotal > 0:
total = subtotal + self.extra_amount
self.subtotal = subtotal
self.total = total
| from flask_seguro.products import Products
from flask import current_app as app
class Cart:
def __init__(self, cart_dict=None):
cart_dict = cart_dict or {}
if cart_dict == {}:
self.total = 0
self.subtotal = 0
self.items = []
else:
self.total = cart_dict["total"]
self.subtotal = cart_dict["subtotal"]
self.items = cart_dict["items"]
self.extra_amount = float(app.config['EXTRA_AMOUNT'])
def to_dict(self):
return {"total": self.total,
"subtotal": self.subtotal,
"items": self.items,
"extra_amount": self.extra_amount}
def change_item(self, item_id, operation):
product = Products().get_one(item_id)
if product:
if operation == 'add':
self.items.append(product)
elif operation == 'remove':
cart_product = filter(
lambda x: x['id'] == product['id'], self.items)
self.items.remove(cart_product[0])
self.update()
return True
else:
return False
def update(self):
subtotal = float(0)
total = float(0)
for product in self.items:
subtotal += float(product["price"])
if subtotal > 0:
total = subtotal + self.extra_amount
self.subtotal = subtotal
self.total = total
| Fix dangerous default mutable value | Fix dangerous default mutable value | Python | mit | rgcarrasqueira/python-pagseguro,vintasoftware/python-pagseguro,rochacbruno/python-pagseguro | from flask_seguro.products import Products
from flask import current_app as app
class Cart:
- def __init__(self, cart_dict={}):
+ def __init__(self, cart_dict=None):
+ cart_dict = cart_dict or {}
if cart_dict == {}:
self.total = 0
self.subtotal = 0
self.items = []
else:
self.total = cart_dict["total"]
self.subtotal = cart_dict["subtotal"]
self.items = cart_dict["items"]
self.extra_amount = float(app.config['EXTRA_AMOUNT'])
def to_dict(self):
return {"total": self.total,
"subtotal": self.subtotal,
"items": self.items,
"extra_amount": self.extra_amount}
def change_item(self, item_id, operation):
product = Products().get_one(item_id)
if product:
if operation == 'add':
self.items.append(product)
elif operation == 'remove':
cart_product = filter(
lambda x: x['id'] == product['id'], self.items)
self.items.remove(cart_product[0])
self.update()
return True
else:
return False
def update(self):
subtotal = float(0)
total = float(0)
for product in self.items:
subtotal += float(product["price"])
if subtotal > 0:
total = subtotal + self.extra_amount
self.subtotal = subtotal
self.total = total
| Fix dangerous default mutable value | ## Code Before:
from flask_seguro.products import Products
from flask import current_app as app
class Cart:
def __init__(self, cart_dict={}):
if cart_dict == {}:
self.total = 0
self.subtotal = 0
self.items = []
else:
self.total = cart_dict["total"]
self.subtotal = cart_dict["subtotal"]
self.items = cart_dict["items"]
self.extra_amount = float(app.config['EXTRA_AMOUNT'])
def to_dict(self):
return {"total": self.total,
"subtotal": self.subtotal,
"items": self.items,
"extra_amount": self.extra_amount}
def change_item(self, item_id, operation):
product = Products().get_one(item_id)
if product:
if operation == 'add':
self.items.append(product)
elif operation == 'remove':
cart_product = filter(
lambda x: x['id'] == product['id'], self.items)
self.items.remove(cart_product[0])
self.update()
return True
else:
return False
def update(self):
subtotal = float(0)
total = float(0)
for product in self.items:
subtotal += float(product["price"])
if subtotal > 0:
total = subtotal + self.extra_amount
self.subtotal = subtotal
self.total = total
## Instruction:
Fix dangerous default mutable value
## Code After:
from flask_seguro.products import Products
from flask import current_app as app
class Cart:
def __init__(self, cart_dict=None):
cart_dict = cart_dict or {}
if cart_dict == {}:
self.total = 0
self.subtotal = 0
self.items = []
else:
self.total = cart_dict["total"]
self.subtotal = cart_dict["subtotal"]
self.items = cart_dict["items"]
self.extra_amount = float(app.config['EXTRA_AMOUNT'])
def to_dict(self):
return {"total": self.total,
"subtotal": self.subtotal,
"items": self.items,
"extra_amount": self.extra_amount}
def change_item(self, item_id, operation):
product = Products().get_one(item_id)
if product:
if operation == 'add':
self.items.append(product)
elif operation == 'remove':
cart_product = filter(
lambda x: x['id'] == product['id'], self.items)
self.items.remove(cart_product[0])
self.update()
return True
else:
return False
def update(self):
subtotal = float(0)
total = float(0)
for product in self.items:
subtotal += float(product["price"])
if subtotal > 0:
total = subtotal + self.extra_amount
self.subtotal = subtotal
self.total = total
| from flask_seguro.products import Products
from flask import current_app as app
class Cart:
- def __init__(self, cart_dict={}):
? ^^
+ def __init__(self, cart_dict=None):
? ^^^^
+ cart_dict = cart_dict or {}
if cart_dict == {}:
self.total = 0
self.subtotal = 0
self.items = []
else:
self.total = cart_dict["total"]
self.subtotal = cart_dict["subtotal"]
self.items = cart_dict["items"]
self.extra_amount = float(app.config['EXTRA_AMOUNT'])
def to_dict(self):
return {"total": self.total,
"subtotal": self.subtotal,
"items": self.items,
"extra_amount": self.extra_amount}
def change_item(self, item_id, operation):
product = Products().get_one(item_id)
if product:
if operation == 'add':
self.items.append(product)
elif operation == 'remove':
cart_product = filter(
lambda x: x['id'] == product['id'], self.items)
self.items.remove(cart_product[0])
self.update()
return True
else:
return False
def update(self):
subtotal = float(0)
total = float(0)
for product in self.items:
subtotal += float(product["price"])
if subtotal > 0:
total = subtotal + self.extra_amount
self.subtotal = subtotal
self.total = total |
fff0b4af89e02ff834221ef056b7dcb979dc6cd7 | webpay/webpay.py | webpay/webpay.py | from .api import Account, Charges, Customers
import requests
class WebPay:
def __init__(self, key, api_base = 'https://api.webpay.jp/v1'):
self.key = key
self.api_base = api_base
self.account = Account(self)
self.charges = Charges(self)
self.customers = Customers(self)
def post(self, path, params):
r = requests.post(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def get(self, path, params = {}):
r = requests.get(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def delete(self, path, params = {}):
r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
| from .api import Account, Charges, Customers
import requests
import json
class WebPay:
def __init__(self, key, api_base = 'https://api.webpay.jp/v1'):
self.key = key
self.api_base = api_base
self.account = Account(self)
self.charges = Charges(self)
self.customers = Customers(self)
def post(self, path, params):
r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
return r.json()
def get(self, path, params = {}):
r = requests.get(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def delete(self, path, params = {}):
r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
return r.json()
| Use JSON for other than GET request | Use JSON for other than GET request
Because internal dict parameters is not handled as expected.
>>> payload = {'key1': 'value1', 'key2': 'value2', 'set': {'a': 'x', 'b': 'y'}}
>>> r = requests.post("http://httpbin.org/post", data=payload)
>>> r.json()
{...
'form': {'key2': 'value2', 'key1': 'value1', 'set': ['a', 'b']}
...}
| Python | mit | yamaneko1212/webpay-python | from .api import Account, Charges, Customers
import requests
+ import json
class WebPay:
def __init__(self, key, api_base = 'https://api.webpay.jp/v1'):
self.key = key
self.api_base = api_base
self.account = Account(self)
self.charges = Charges(self)
self.customers = Customers(self)
def post(self, path, params):
- r = requests.post(self.api_base + path, auth = (self.key, ''), params = params)
+ r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
return r.json()
def get(self, path, params = {}):
r = requests.get(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def delete(self, path, params = {}):
- r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params)
+ r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
return r.json()
| Use JSON for other than GET request | ## Code Before:
from .api import Account, Charges, Customers
import requests
class WebPay:
def __init__(self, key, api_base = 'https://api.webpay.jp/v1'):
self.key = key
self.api_base = api_base
self.account = Account(self)
self.charges = Charges(self)
self.customers = Customers(self)
def post(self, path, params):
r = requests.post(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def get(self, path, params = {}):
r = requests.get(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def delete(self, path, params = {}):
r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
## Instruction:
Use JSON for other than GET request
## Code After:
from .api import Account, Charges, Customers
import requests
import json
class WebPay:
def __init__(self, key, api_base = 'https://api.webpay.jp/v1'):
self.key = key
self.api_base = api_base
self.account = Account(self)
self.charges = Charges(self)
self.customers = Customers(self)
def post(self, path, params):
r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
return r.json()
def get(self, path, params = {}):
r = requests.get(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def delete(self, path, params = {}):
r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
return r.json()
| from .api import Account, Charges, Customers
import requests
+ import json
class WebPay:
def __init__(self, key, api_base = 'https://api.webpay.jp/v1'):
self.key = key
self.api_base = api_base
self.account = Account(self)
self.charges = Charges(self)
self.customers = Customers(self)
def post(self, path, params):
- r = requests.post(self.api_base + path, auth = (self.key, ''), params = params)
? ^ ^ --
+ r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
? ^ ^ +++++++++++ +
return r.json()
def get(self, path, params = {}):
r = requests.get(self.api_base + path, auth = (self.key, ''), params = params)
return r.json()
def delete(self, path, params = {}):
- r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params)
? ^ ^ --
+ r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params))
? ^ ^ +++++++++++ +
return r.json() |
d953055801c8d618c70cea81e3e35684122c66a7 | setuptools/config/__init__.py | setuptools/config/__init__.py | import warnings
from functools import wraps
from textwrap import dedent
from typing import Callable, TypeVar, cast
from .._deprecation_warning import SetuptoolsDeprecationWarning
from . import setupcfg
Fn = TypeVar("Fn", bound=Callable)
__all__ = ('parse_configuration', 'read_configuration')
def _deprecation_notice(fn: Fn) -> Fn:
@wraps(fn)
def _wrapper(*args, **kwargs):
msg = f"""\
As setuptools moves its configuration towards `pyproject.toml`,
`{__name__}.{fn.__name__}` became deprecated.
For the time being, you can use the `{setupcfg.__name__}` module
to access a backward compatible API, but this module is provisional
and might be removed in the future.
"""
warnings.warn(dedent(msg), SetuptoolsDeprecationWarning)
return fn(*args, **kwargs)
return cast(Fn, _wrapper)
read_configuration = _deprecation_notice(setupcfg.read_configuration)
parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
| import warnings
from functools import wraps
from textwrap import dedent
from typing import Callable, TypeVar, cast
from .._deprecation_warning import SetuptoolsDeprecationWarning
from . import setupcfg
Fn = TypeVar("Fn", bound=Callable)
__all__ = ('parse_configuration', 'read_configuration')
def _deprecation_notice(fn: Fn) -> Fn:
@wraps(fn)
def _wrapper(*args, **kwargs):
msg = f"""\
As setuptools moves its configuration towards `pyproject.toml`,
`{__name__}.{fn.__name__}` became deprecated.
For the time being, you can use the `{setupcfg.__name__}` module
to access a backward compatible API, but this module is provisional
and might be removed in the future.
"""
warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2)
return fn(*args, **kwargs)
return cast(Fn, _wrapper)
read_configuration = _deprecation_notice(setupcfg.read_configuration)
parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
| Add stacklevel=2 to make calling code clear | Add stacklevel=2 to make calling code clear | Python | mit | pypa/setuptools,pypa/setuptools,pypa/setuptools | import warnings
from functools import wraps
from textwrap import dedent
from typing import Callable, TypeVar, cast
from .._deprecation_warning import SetuptoolsDeprecationWarning
from . import setupcfg
Fn = TypeVar("Fn", bound=Callable)
__all__ = ('parse_configuration', 'read_configuration')
def _deprecation_notice(fn: Fn) -> Fn:
@wraps(fn)
def _wrapper(*args, **kwargs):
msg = f"""\
As setuptools moves its configuration towards `pyproject.toml`,
`{__name__}.{fn.__name__}` became deprecated.
For the time being, you can use the `{setupcfg.__name__}` module
to access a backward compatible API, but this module is provisional
and might be removed in the future.
"""
- warnings.warn(dedent(msg), SetuptoolsDeprecationWarning)
+ warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2)
return fn(*args, **kwargs)
return cast(Fn, _wrapper)
read_configuration = _deprecation_notice(setupcfg.read_configuration)
parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
| Add stacklevel=2 to make calling code clear | ## Code Before:
import warnings
from functools import wraps
from textwrap import dedent
from typing import Callable, TypeVar, cast
from .._deprecation_warning import SetuptoolsDeprecationWarning
from . import setupcfg
Fn = TypeVar("Fn", bound=Callable)
__all__ = ('parse_configuration', 'read_configuration')
def _deprecation_notice(fn: Fn) -> Fn:
@wraps(fn)
def _wrapper(*args, **kwargs):
msg = f"""\
As setuptools moves its configuration towards `pyproject.toml`,
`{__name__}.{fn.__name__}` became deprecated.
For the time being, you can use the `{setupcfg.__name__}` module
to access a backward compatible API, but this module is provisional
and might be removed in the future.
"""
warnings.warn(dedent(msg), SetuptoolsDeprecationWarning)
return fn(*args, **kwargs)
return cast(Fn, _wrapper)
read_configuration = _deprecation_notice(setupcfg.read_configuration)
parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
## Instruction:
Add stacklevel=2 to make calling code clear
## Code After:
import warnings
from functools import wraps
from textwrap import dedent
from typing import Callable, TypeVar, cast
from .._deprecation_warning import SetuptoolsDeprecationWarning
from . import setupcfg
Fn = TypeVar("Fn", bound=Callable)
__all__ = ('parse_configuration', 'read_configuration')
def _deprecation_notice(fn: Fn) -> Fn:
@wraps(fn)
def _wrapper(*args, **kwargs):
msg = f"""\
As setuptools moves its configuration towards `pyproject.toml`,
`{__name__}.{fn.__name__}` became deprecated.
For the time being, you can use the `{setupcfg.__name__}` module
to access a backward compatible API, but this module is provisional
and might be removed in the future.
"""
warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2)
return fn(*args, **kwargs)
return cast(Fn, _wrapper)
read_configuration = _deprecation_notice(setupcfg.read_configuration)
parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
| import warnings
from functools import wraps
from textwrap import dedent
from typing import Callable, TypeVar, cast
from .._deprecation_warning import SetuptoolsDeprecationWarning
from . import setupcfg
Fn = TypeVar("Fn", bound=Callable)
__all__ = ('parse_configuration', 'read_configuration')
def _deprecation_notice(fn: Fn) -> Fn:
@wraps(fn)
def _wrapper(*args, **kwargs):
msg = f"""\
As setuptools moves its configuration towards `pyproject.toml`,
`{__name__}.{fn.__name__}` became deprecated.
For the time being, you can use the `{setupcfg.__name__}` module
to access a backward compatible API, but this module is provisional
and might be removed in the future.
"""
- warnings.warn(dedent(msg), SetuptoolsDeprecationWarning)
+ warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2)
? ++++++++++++++
return fn(*args, **kwargs)
return cast(Fn, _wrapper)
read_configuration = _deprecation_notice(setupcfg.read_configuration)
parse_configuration = _deprecation_notice(setupcfg.parse_configuration) |
261393eb46cdc082b60d9ea11ec862f508632ad2 | audit_log/models/__init__.py | audit_log/models/__init__.py | from django.db.models import Model
from django.utils.translation import ugettext_lazy as _
from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField
class AuthStampedModel(Model):
"""
An abstract base class model that provides auth and session information
fields.
"""
created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
created_with_session_key = CreatingSessionKeyField(_("created with session key"))
modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
modified_with_session_key = LastSessionKeyField(_("modified with session key"))
class Meta:
abstract = True
| from django.db.models import Model, SET_NULL
from django.utils.translation import ugettext_lazy as _
from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField
class AuthStampedModel(Model):
"""
An abstract base class model that provides auth and session information
fields.
"""
created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
created_with_session_key = CreatingSessionKeyField(_("created with session key"))
modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
modified_with_session_key = LastSessionKeyField(_("modified with session key"))
class Meta:
abstract = True
| Add mandatory `on_delete` and allow nulls | Add mandatory `on_delete` and allow nulls
I assume it's better to allow nulls than to have auth stamped models
disappear with deleted users.
| Python | bsd-3-clause | Atomidata/django-audit-log,Atomidata/django-audit-log | - from django.db.models import Model
+ from django.db.models import Model, SET_NULL
from django.utils.translation import ugettext_lazy as _
from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField
class AuthStampedModel(Model):
"""
An abstract base class model that provides auth and session information
fields.
"""
- created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
+ created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
created_with_session_key = CreatingSessionKeyField(_("created with session key"))
- modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
+ modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
modified_with_session_key = LastSessionKeyField(_("modified with session key"))
class Meta:
abstract = True
| Add mandatory `on_delete` and allow nulls | ## Code Before:
from django.db.models import Model
from django.utils.translation import ugettext_lazy as _
from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField
class AuthStampedModel(Model):
"""
An abstract base class model that provides auth and session information
fields.
"""
created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
created_with_session_key = CreatingSessionKeyField(_("created with session key"))
modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
modified_with_session_key = LastSessionKeyField(_("modified with session key"))
class Meta:
abstract = True
## Instruction:
Add mandatory `on_delete` and allow nulls
## Code After:
from django.db.models import Model, SET_NULL
from django.utils.translation import ugettext_lazy as _
from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField
class AuthStampedModel(Model):
"""
An abstract base class model that provides auth and session information
fields.
"""
created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
created_with_session_key = CreatingSessionKeyField(_("created with session key"))
modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
modified_with_session_key = LastSessionKeyField(_("modified with session key"))
class Meta:
abstract = True
| - from django.db.models import Model
+ from django.db.models import Model, SET_NULL
? ++++++++++
from django.utils.translation import ugettext_lazy as _
from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField
class AuthStampedModel(Model):
"""
An abstract base class model that provides auth and session information
fields.
"""
- created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
+ created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set")
? +++++++++++++++++++++++++++++++
created_with_session_key = CreatingSessionKeyField(_("created with session key"))
- modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
+ modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set")
? +++++++++++++++++++++++++++++++
modified_with_session_key = LastSessionKeyField(_("modified with session key"))
class Meta:
abstract = True |
c21fe453911af190f3cbd93356396d4f5e65195e | mopidy/backends/gstreamer.py | mopidy/backends/gstreamer.py | import logging
import gst
from mopidy import config
from mopidy.backends import BaseBackend
from mopidy.models import Artist, Album, Track, Playlist
logger = logging.getLogger(u'backends.gstreamer')
class GStreamerBackend(BaseBackend):
def __init__(self, *args, **kwargs):
super(GStreamerBackend, self).__init__(*args, **kwargs)
playlist = []
player = gst.element_factory_make("playbin2", "player")
fakesink = gst.element_factory_make("fakesink", "fakesink")
player.set_property("video-sink", fakesink)
self.player = player
def _play_current_track(self):
self.player.set_property("uri", self._current_track.uri)
self.player.set_state(gst.STATE_PLAYING)
def _play(self):
if self._current_track is not None:
self._play_current_track()
return True
else:
return False
def _stop(self):
self.player.set_state(gst.STATE_NULL)
return True
| import logging
import gst
from mopidy import config
from mopidy.backends import BaseBackend
from mopidy.models import Artist, Album, Track, Playlist
logger = logging.getLogger(u'backends.gstreamer')
class GStreamerBackend(BaseBackend):
def __init__(self, *args, **kwargs):
super(GStreamerBackend, self).__init__(*args, **kwargs)
playlist = []
player = gst.element_factory_make("playbin2", "player")
fakesink = gst.element_factory_make("fakesink", "fakesink")
player.set_property("video-sink", fakesink)
self.player = player
def _play(self):
if self._current_track is None:
return False
self.player.set_property("uri", self._current_track.uri)
self.player.set_state(gst.STATE_PLAYING)
return True
def _stop(self):
self.player.set_state(gst.STATE_NULL)
return True
| Clean play code for GStreamer | Clean play code for GStreamer
| Python | apache-2.0 | vrs01/mopidy,bacontext/mopidy,vrs01/mopidy,quartz55/mopidy,priestd09/mopidy,swak/mopidy,liamw9534/mopidy,hkariti/mopidy,pacificIT/mopidy,ZenithDK/mopidy,hkariti/mopidy,dbrgn/mopidy,kingosticks/mopidy,ZenithDK/mopidy,diandiankan/mopidy,vrs01/mopidy,hkariti/mopidy,pacificIT/mopidy,ali/mopidy,SuperStarPL/mopidy,adamcik/mopidy,swak/mopidy,glogiotatidis/mopidy,adamcik/mopidy,SuperStarPL/mopidy,mokieyue/mopidy,rawdlite/mopidy,priestd09/mopidy,diandiankan/mopidy,swak/mopidy,tkem/mopidy,diandiankan/mopidy,dbrgn/mopidy,mopidy/mopidy,woutervanwijk/mopidy,jodal/mopidy,bacontext/mopidy,jcass77/mopidy,ali/mopidy,jcass77/mopidy,bacontext/mopidy,ali/mopidy,SuperStarPL/mopidy,abarisain/mopidy,mopidy/mopidy,mopidy/mopidy,bacontext/mopidy,quartz55/mopidy,quartz55/mopidy,quartz55/mopidy,ZenithDK/mopidy,rawdlite/mopidy,ali/mopidy,abarisain/mopidy,hkariti/mopidy,tkem/mopidy,ZenithDK/mopidy,tkem/mopidy,rawdlite/mopidy,jmarsik/mopidy,jodal/mopidy,diandiankan/mopidy,priestd09/mopidy,tkem/mopidy,liamw9534/mopidy,bencevans/mopidy,dbrgn/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,swak/mopidy,glogiotatidis/mopidy,glogiotatidis/mopidy,mokieyue/mopidy,dbrgn/mopidy,bencevans/mopidy,SuperStarPL/mopidy,adamcik/mopidy,jmarsik/mopidy,jcass77/mopidy,kingosticks/mopidy,rawdlite/mopidy,vrs01/mopidy,mokieyue/mopidy,bencevans/mopidy,bencevans/mopidy,jmarsik/mopidy,jmarsik/mopidy,kingosticks/mopidy,woutervanwijk/mopidy,mokieyue/mopidy,jodal/mopidy,pacificIT/mopidy | import logging
import gst
from mopidy import config
from mopidy.backends import BaseBackend
from mopidy.models import Artist, Album, Track, Playlist
logger = logging.getLogger(u'backends.gstreamer')
class GStreamerBackend(BaseBackend):
def __init__(self, *args, **kwargs):
super(GStreamerBackend, self).__init__(*args, **kwargs)
playlist = []
player = gst.element_factory_make("playbin2", "player")
fakesink = gst.element_factory_make("fakesink", "fakesink")
player.set_property("video-sink", fakesink)
self.player = player
+ def _play(self):
+ if self._current_track is None:
+ return False
- def _play_current_track(self):
self.player.set_property("uri", self._current_track.uri)
self.player.set_state(gst.STATE_PLAYING)
- def _play(self):
- if self._current_track is not None:
- self._play_current_track()
- return True
+ return True
- else:
- return False
def _stop(self):
self.player.set_state(gst.STATE_NULL)
return True
| Clean play code for GStreamer | ## Code Before:
import logging
import gst
from mopidy import config
from mopidy.backends import BaseBackend
from mopidy.models import Artist, Album, Track, Playlist
logger = logging.getLogger(u'backends.gstreamer')
class GStreamerBackend(BaseBackend):
def __init__(self, *args, **kwargs):
super(GStreamerBackend, self).__init__(*args, **kwargs)
playlist = []
player = gst.element_factory_make("playbin2", "player")
fakesink = gst.element_factory_make("fakesink", "fakesink")
player.set_property("video-sink", fakesink)
self.player = player
def _play_current_track(self):
self.player.set_property("uri", self._current_track.uri)
self.player.set_state(gst.STATE_PLAYING)
def _play(self):
if self._current_track is not None:
self._play_current_track()
return True
else:
return False
def _stop(self):
self.player.set_state(gst.STATE_NULL)
return True
## Instruction:
Clean play code for GStreamer
## Code After:
import logging
import gst
from mopidy import config
from mopidy.backends import BaseBackend
from mopidy.models import Artist, Album, Track, Playlist
logger = logging.getLogger(u'backends.gstreamer')
class GStreamerBackend(BaseBackend):
def __init__(self, *args, **kwargs):
super(GStreamerBackend, self).__init__(*args, **kwargs)
playlist = []
player = gst.element_factory_make("playbin2", "player")
fakesink = gst.element_factory_make("fakesink", "fakesink")
player.set_property("video-sink", fakesink)
self.player = player
def _play(self):
if self._current_track is None:
return False
self.player.set_property("uri", self._current_track.uri)
self.player.set_state(gst.STATE_PLAYING)
return True
def _stop(self):
self.player.set_state(gst.STATE_NULL)
return True
| import logging
import gst
from mopidy import config
from mopidy.backends import BaseBackend
from mopidy.models import Artist, Album, Track, Playlist
logger = logging.getLogger(u'backends.gstreamer')
class GStreamerBackend(BaseBackend):
def __init__(self, *args, **kwargs):
super(GStreamerBackend, self).__init__(*args, **kwargs)
playlist = []
player = gst.element_factory_make("playbin2", "player")
fakesink = gst.element_factory_make("fakesink", "fakesink")
player.set_property("video-sink", fakesink)
self.player = player
+ def _play(self):
+ if self._current_track is None:
+ return False
- def _play_current_track(self):
self.player.set_property("uri", self._current_track.uri)
self.player.set_state(gst.STATE_PLAYING)
- def _play(self):
- if self._current_track is not None:
- self._play_current_track()
- return True
? ----
+ return True
- else:
- return False
def _stop(self):
self.player.set_state(gst.STATE_NULL)
return True |
b7c967ad0f45cc1144a8713c6513bae5bca89242 | LiSE/LiSE/test_proxy.py | LiSE/LiSE/test_proxy.py | from LiSE.proxy import EngineProcessManager
import allegedb.test
class ProxyTest(allegedb.test.AllegedTest):
def setUp(self):
self.manager = EngineProcessManager()
self.engine = self.manager.start('sqlite:///:memory:')
self.graphmakers = (self.engine.new_character,)
def tearDown(self):
self.manager.shutdown()
class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest):
pass
class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest):
pass
class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest):
pass
class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest):
pass
class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest):
pass
| from LiSE.proxy import EngineProcessManager
import allegedb.test
class ProxyTest(allegedb.test.AllegedTest):
def setUp(self):
self.manager = EngineProcessManager()
self.engine = self.manager.start('sqlite:///:memory:')
self.graphmakers = (self.engine.new_character,)
def tearDown(self):
self.manager.shutdown()
class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest):
pass
class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest):
pass
class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest):
pass
class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest):
pass
| Delete BranchLineageTest, which assumes bidirectional graphs exist | Delete BranchLineageTest, which assumes bidirectional graphs exist
| Python | agpl-3.0 | LogicalDash/LiSE,LogicalDash/LiSE | from LiSE.proxy import EngineProcessManager
import allegedb.test
class ProxyTest(allegedb.test.AllegedTest):
def setUp(self):
self.manager = EngineProcessManager()
self.engine = self.manager.start('sqlite:///:memory:')
self.graphmakers = (self.engine.new_character,)
def tearDown(self):
self.manager.shutdown()
class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest):
pass
- class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest):
- pass
-
-
class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest):
pass
class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest):
pass
class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest):
pass
| Delete BranchLineageTest, which assumes bidirectional graphs exist | ## Code Before:
from LiSE.proxy import EngineProcessManager
import allegedb.test
class ProxyTest(allegedb.test.AllegedTest):
def setUp(self):
self.manager = EngineProcessManager()
self.engine = self.manager.start('sqlite:///:memory:')
self.graphmakers = (self.engine.new_character,)
def tearDown(self):
self.manager.shutdown()
class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest):
pass
class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest):
pass
class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest):
pass
class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest):
pass
class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest):
pass
## Instruction:
Delete BranchLineageTest, which assumes bidirectional graphs exist
## Code After:
from LiSE.proxy import EngineProcessManager
import allegedb.test
class ProxyTest(allegedb.test.AllegedTest):
def setUp(self):
self.manager = EngineProcessManager()
self.engine = self.manager.start('sqlite:///:memory:')
self.graphmakers = (self.engine.new_character,)
def tearDown(self):
self.manager.shutdown()
class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest):
pass
class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest):
pass
class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest):
pass
class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest):
pass
| from LiSE.proxy import EngineProcessManager
import allegedb.test
class ProxyTest(allegedb.test.AllegedTest):
def setUp(self):
self.manager = EngineProcessManager()
self.engine = self.manager.start('sqlite:///:memory:')
self.graphmakers = (self.engine.new_character,)
def tearDown(self):
self.manager.shutdown()
class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest):
pass
- class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest):
- pass
-
-
class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest):
pass
class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest):
pass
class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest):
pass |
f0f3c50a65aae1393928579ca0e48891d1ac8f18 | app/access_control.py | app/access_control.py | from functools import wraps
from flask import flash, redirect, url_for, session
from app import views
def login_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
if 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Please login to continue.", "danger")
return redirect(url_for("login"))
return decorated_function | from functools import wraps
from flask import flash, redirect, url_for, session
from app import views
def login_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
if 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Please login to continue.", "danger")
return redirect(url_for("login"))
return decorated_function
def for_guests(f):
@wraps(f)
def decorated_function(*args, **kwrags):
if not 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Invalid Action.", "danger")
return redirect(url_for("dashboard"))
return decorated_function | Create a decorator `for_guest` for access control on pages for guests. | Create a decorator `for_guest` for access control on pages for guests.
| Python | mit | alchermd/flask-todo-app,alchermd/flask-todo-app | from functools import wraps
from flask import flash, redirect, url_for, session
from app import views
def login_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
if 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Please login to continue.", "danger")
return redirect(url_for("login"))
return decorated_function
+
+ def for_guests(f):
+ @wraps(f)
+ def decorated_function(*args, **kwrags):
+ if not 'logged_in' in session:
+ return f(*args, **kwargs)
+ else:
+ flash("Invalid Action.", "danger")
+ return redirect(url_for("dashboard"))
+ return decorated_function | Create a decorator `for_guest` for access control on pages for guests. | ## Code Before:
from functools import wraps
from flask import flash, redirect, url_for, session
from app import views
def login_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
if 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Please login to continue.", "danger")
return redirect(url_for("login"))
return decorated_function
## Instruction:
Create a decorator `for_guest` for access control on pages for guests.
## Code After:
from functools import wraps
from flask import flash, redirect, url_for, session
from app import views
def login_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
if 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Please login to continue.", "danger")
return redirect(url_for("login"))
return decorated_function
def for_guests(f):
@wraps(f)
def decorated_function(*args, **kwrags):
if not 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Invalid Action.", "danger")
return redirect(url_for("dashboard"))
return decorated_function | from functools import wraps
from flask import flash, redirect, url_for, session
from app import views
def login_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
if 'logged_in' in session:
return f(*args, **kwargs)
else:
flash("Please login to continue.", "danger")
return redirect(url_for("login"))
return decorated_function
+
+ def for_guests(f):
+ @wraps(f)
+ def decorated_function(*args, **kwrags):
+ if not 'logged_in' in session:
+ return f(*args, **kwargs)
+ else:
+ flash("Invalid Action.", "danger")
+ return redirect(url_for("dashboard"))
+ return decorated_function |
37333506e6866e7d0859c5068f115a3e1b9dec3a | test/test_coordinate.py | test/test_coordinate.py | import unittest
from src import coordinate
class TestRules(unittest.TestCase):
""" Tests for the coordinate module """
def test_get_x_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 4
actual_result = board_location.get_x_board()
self.assertEqual(actual_result, expected_result)
def test_get_y_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 6
actual_result = board_location.get_y_board()
self.assertEqual(actual_result, expected_result)
def test_get_x_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 3
actual_result = board_location.get_x_array()
self.assertEqual(actual_result, expected_result)
def test_get_y_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 5
actual_result = board_location.get_y_array()
self.assertEqual(actual_result, expected_result) | import unittest
from src import coordinate
class TestRules(unittest.TestCase):
""" Tests for the coordinate module """
def test_get_x_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 4
actual_result = board_location.get_x_board()
self.assertEqual(actual_result, expected_result)
def test_get_y_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 6
actual_result = board_location.get_y_board()
self.assertEqual(actual_result, expected_result)
def test_get_x_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 3
actual_result = board_location.get_x_array()
self.assertEqual(actual_result, expected_result)
def test_get_y_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 5
actual_result = board_location.get_y_array()
self.assertEqual(actual_result, expected_result)
def test_coordinate_bad_x(self):
self.assertRaises(TypeError, coordinate.Coordinate, "4", 6)
def test_coordinate_bad_y(self):
self.assertRaises(TypeError, coordinate.Coordinate, 4, "6")
def test_coordinate_bad_location(self):
self.assertRaises(ValueError, coordinate.Coordinate, 50, 100)
| Add unit tests for fail fast logic in convertCharToInt() | Add unit tests for fail fast logic in convertCharToInt()
| Python | mit | blairck/jaeger | import unittest
from src import coordinate
class TestRules(unittest.TestCase):
""" Tests for the coordinate module """
def test_get_x_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 4
actual_result = board_location.get_x_board()
self.assertEqual(actual_result, expected_result)
def test_get_y_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 6
actual_result = board_location.get_y_board()
self.assertEqual(actual_result, expected_result)
def test_get_x_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 3
actual_result = board_location.get_x_array()
self.assertEqual(actual_result, expected_result)
def test_get_y_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 5
actual_result = board_location.get_y_array()
self.assertEqual(actual_result, expected_result)
+
+ def test_coordinate_bad_x(self):
+ self.assertRaises(TypeError, coordinate.Coordinate, "4", 6)
+
+ def test_coordinate_bad_y(self):
+ self.assertRaises(TypeError, coordinate.Coordinate, 4, "6")
+
+ def test_coordinate_bad_location(self):
+ self.assertRaises(ValueError, coordinate.Coordinate, 50, 100)
+ | Add unit tests for fail fast logic in convertCharToInt() | ## Code Before:
import unittest
from src import coordinate
class TestRules(unittest.TestCase):
""" Tests for the coordinate module """
def test_get_x_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 4
actual_result = board_location.get_x_board()
self.assertEqual(actual_result, expected_result)
def test_get_y_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 6
actual_result = board_location.get_y_board()
self.assertEqual(actual_result, expected_result)
def test_get_x_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 3
actual_result = board_location.get_x_array()
self.assertEqual(actual_result, expected_result)
def test_get_y_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 5
actual_result = board_location.get_y_array()
self.assertEqual(actual_result, expected_result)
## Instruction:
Add unit tests for fail fast logic in convertCharToInt()
## Code After:
import unittest
from src import coordinate
class TestRules(unittest.TestCase):
""" Tests for the coordinate module """
def test_get_x_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 4
actual_result = board_location.get_x_board()
self.assertEqual(actual_result, expected_result)
def test_get_y_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 6
actual_result = board_location.get_y_board()
self.assertEqual(actual_result, expected_result)
def test_get_x_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 3
actual_result = board_location.get_x_array()
self.assertEqual(actual_result, expected_result)
def test_get_y_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 5
actual_result = board_location.get_y_array()
self.assertEqual(actual_result, expected_result)
def test_coordinate_bad_x(self):
self.assertRaises(TypeError, coordinate.Coordinate, "4", 6)
def test_coordinate_bad_y(self):
self.assertRaises(TypeError, coordinate.Coordinate, 4, "6")
def test_coordinate_bad_location(self):
self.assertRaises(ValueError, coordinate.Coordinate, 50, 100)
| import unittest
from src import coordinate
class TestRules(unittest.TestCase):
""" Tests for the coordinate module """
def test_get_x_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 4
actual_result = board_location.get_x_board()
self.assertEqual(actual_result, expected_result)
def test_get_y_board(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 6
actual_result = board_location.get_y_board()
self.assertEqual(actual_result, expected_result)
def test_get_x_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 3
actual_result = board_location.get_x_array()
self.assertEqual(actual_result, expected_result)
def test_get_y_array(self):
board_location = coordinate.Coordinate(4, 6)
expected_result = 5
actual_result = board_location.get_y_array()
self.assertEqual(actual_result, expected_result)
+
+ def test_coordinate_bad_x(self):
+ self.assertRaises(TypeError, coordinate.Coordinate, "4", 6)
+
+ def test_coordinate_bad_y(self):
+ self.assertRaises(TypeError, coordinate.Coordinate, 4, "6")
+
+ def test_coordinate_bad_location(self):
+ self.assertRaises(ValueError, coordinate.Coordinate, 50, 100) |
0afdab2f6feced873c88ba1e73fdde0dad5f041e | skytap/Quotas.py | skytap/Quotas.py | import json
import sys
from skytap.models.Quota import Quota
from skytap.models.SkytapGroup import SkytapGroup
class Quotas(SkytapGroup):
"""Company/account quotas object."""
def __init__(self):
"""Load the quotas from Skytap."""
super(Quotas, self).__init__()
quota_rest = self.rest('/v2/company/quotas')
quota_json = json.loads(quota_rest)
for qu in quota_json:
self.data[qu] = Quota(quota_json[qu][0])
if __name__ == '__main__':
print(Quotas().main(sys.argv[1:]))
| import json
import sys
from skytap.models.Quota import Quota
from skytap.models.SkytapGroup import SkytapGroup
class Quotas(SkytapGroup):
"""Company/account quotas object.
Note: This code assumes that you have regional limits on your account.
The return is different if you don't (see the /v2 API doc). We should get
each piece of the return and sort it into type-and-region (whether you
have regional limits or not) and can then access things uniformly. Doing
so will also require smartly accessing the API on demand more, since
accounts with regional limits may require multiple calls to get the info
desired.
"""
def __init__(self):
"""Load the quotas from Skytap."""
super(Quotas, self).__init__()
quota_rest = self.rest('/v2/company/quotas')
quota_json = json.loads(quota_rest)
for qu in quota_json:
self.data[qu] = Quota(quota_json[qu][0])
if __name__ == '__main__':
print(Quotas().main(sys.argv[1:]))
| Comment re: API usage to clarify quotas. | Comment re: API usage to clarify quotas.
| Python | mit | mapledyne/skytap,FulcrumIT/skytap | import json
import sys
from skytap.models.Quota import Quota
from skytap.models.SkytapGroup import SkytapGroup
class Quotas(SkytapGroup):
- """Company/account quotas object."""
+ """Company/account quotas object.
+
+ Note: This code assumes that you have regional limits on your account.
+ The return is different if you don't (see the /v2 API doc). We should get
+ each piece of the return and sort it into type-and-region (whether you
+ have regional limits or not) and can then access things uniformly. Doing
+ so will also require smartly accessing the API on demand more, since
+ accounts with regional limits may require multiple calls to get the info
+ desired.
+ """
def __init__(self):
"""Load the quotas from Skytap."""
super(Quotas, self).__init__()
quota_rest = self.rest('/v2/company/quotas')
quota_json = json.loads(quota_rest)
for qu in quota_json:
self.data[qu] = Quota(quota_json[qu][0])
if __name__ == '__main__':
print(Quotas().main(sys.argv[1:]))
| Comment re: API usage to clarify quotas. | ## Code Before:
import json
import sys
from skytap.models.Quota import Quota
from skytap.models.SkytapGroup import SkytapGroup
class Quotas(SkytapGroup):
"""Company/account quotas object."""
def __init__(self):
"""Load the quotas from Skytap."""
super(Quotas, self).__init__()
quota_rest = self.rest('/v2/company/quotas')
quota_json = json.loads(quota_rest)
for qu in quota_json:
self.data[qu] = Quota(quota_json[qu][0])
if __name__ == '__main__':
print(Quotas().main(sys.argv[1:]))
## Instruction:
Comment re: API usage to clarify quotas.
## Code After:
import json
import sys
from skytap.models.Quota import Quota
from skytap.models.SkytapGroup import SkytapGroup
class Quotas(SkytapGroup):
"""Company/account quotas object.
Note: This code assumes that you have regional limits on your account.
The return is different if you don't (see the /v2 API doc). We should get
each piece of the return and sort it into type-and-region (whether you
have regional limits or not) and can then access things uniformly. Doing
so will also require smartly accessing the API on demand more, since
accounts with regional limits may require multiple calls to get the info
desired.
"""
def __init__(self):
"""Load the quotas from Skytap."""
super(Quotas, self).__init__()
quota_rest = self.rest('/v2/company/quotas')
quota_json = json.loads(quota_rest)
for qu in quota_json:
self.data[qu] = Quota(quota_json[qu][0])
if __name__ == '__main__':
print(Quotas().main(sys.argv[1:]))
| import json
import sys
from skytap.models.Quota import Quota
from skytap.models.SkytapGroup import SkytapGroup
class Quotas(SkytapGroup):
- """Company/account quotas object."""
? ---
+ """Company/account quotas object.
+
+ Note: This code assumes that you have regional limits on your account.
+ The return is different if you don't (see the /v2 API doc). We should get
+ each piece of the return and sort it into type-and-region (whether you
+ have regional limits or not) and can then access things uniformly. Doing
+ so will also require smartly accessing the API on demand more, since
+ accounts with regional limits may require multiple calls to get the info
+ desired.
+ """
def __init__(self):
"""Load the quotas from Skytap."""
super(Quotas, self).__init__()
quota_rest = self.rest('/v2/company/quotas')
quota_json = json.loads(quota_rest)
for qu in quota_json:
self.data[qu] = Quota(quota_json[qu][0])
if __name__ == '__main__':
print(Quotas().main(sys.argv[1:])) |
038978f87883247a14e9bec08708452c98c91285 | test/test_chimera.py | test/test_chimera.py | import unittest
import utils
import os
import sys
import re
import shutil
import subprocess
TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
utils.set_search_paths(TOPDIR)
import cryptosite.chimera
class Tests(unittest.TestCase):
def test_bad(self):
"""Test wrong arguments to chimera"""
for args in ([], ['x'] * 4):
out = utils.check_output(['cryptosite', 'chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
out = utils.check_output(['python', '-m',
'cryptosite.chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
def test_make_chimera_file(self):
"""Test make_chimera_file() function"""
cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax')
os.unlink('out.chimerax')
if __name__ == '__main__':
unittest.main()
| import unittest
import utils
import os
import sys
import re
import shutil
import subprocess
TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
utils.set_search_paths(TOPDIR)
import cryptosite.chimera
class Tests(unittest.TestCase):
def test_bad(self):
"""Test wrong arguments to chimera"""
for args in ([], ['x'] * 4):
out = utils.check_output(['cryptosite', 'chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
out = utils.check_output(['python', '-m',
'cryptosite.chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
def test_make_chimera_file(self):
"""Test make_chimera_file() function"""
cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax')
with open('out.chimerax') as fh:
lines = fh.readlines()
self.assertEqual(lines[-4], 'open_files("url1", "url2")\n')
os.unlink('out.chimerax')
if __name__ == '__main__':
unittest.main()
| Check generated file for sanity. | Check generated file for sanity.
| Python | lgpl-2.1 | salilab/cryptosite,salilab/cryptosite,salilab/cryptosite | import unittest
import utils
import os
import sys
import re
import shutil
import subprocess
TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
utils.set_search_paths(TOPDIR)
import cryptosite.chimera
class Tests(unittest.TestCase):
def test_bad(self):
"""Test wrong arguments to chimera"""
for args in ([], ['x'] * 4):
out = utils.check_output(['cryptosite', 'chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
out = utils.check_output(['python', '-m',
'cryptosite.chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
def test_make_chimera_file(self):
"""Test make_chimera_file() function"""
cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax')
+ with open('out.chimerax') as fh:
+ lines = fh.readlines()
+ self.assertEqual(lines[-4], 'open_files("url1", "url2")\n')
os.unlink('out.chimerax')
if __name__ == '__main__':
unittest.main()
| Check generated file for sanity. | ## Code Before:
import unittest
import utils
import os
import sys
import re
import shutil
import subprocess
TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
utils.set_search_paths(TOPDIR)
import cryptosite.chimera
class Tests(unittest.TestCase):
def test_bad(self):
"""Test wrong arguments to chimera"""
for args in ([], ['x'] * 4):
out = utils.check_output(['cryptosite', 'chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
out = utils.check_output(['python', '-m',
'cryptosite.chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
def test_make_chimera_file(self):
"""Test make_chimera_file() function"""
cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax')
os.unlink('out.chimerax')
if __name__ == '__main__':
unittest.main()
## Instruction:
Check generated file for sanity.
## Code After:
import unittest
import utils
import os
import sys
import re
import shutil
import subprocess
TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
utils.set_search_paths(TOPDIR)
import cryptosite.chimera
class Tests(unittest.TestCase):
def test_bad(self):
"""Test wrong arguments to chimera"""
for args in ([], ['x'] * 4):
out = utils.check_output(['cryptosite', 'chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
out = utils.check_output(['python', '-m',
'cryptosite.chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
def test_make_chimera_file(self):
"""Test make_chimera_file() function"""
cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax')
with open('out.chimerax') as fh:
lines = fh.readlines()
self.assertEqual(lines[-4], 'open_files("url1", "url2")\n')
os.unlink('out.chimerax')
if __name__ == '__main__':
unittest.main()
| import unittest
import utils
import os
import sys
import re
import shutil
import subprocess
TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
utils.set_search_paths(TOPDIR)
import cryptosite.chimera
class Tests(unittest.TestCase):
def test_bad(self):
"""Test wrong arguments to chimera"""
for args in ([], ['x'] * 4):
out = utils.check_output(['cryptosite', 'chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
out = utils.check_output(['python', '-m',
'cryptosite.chimera'] + args,
stderr=subprocess.STDOUT, retcode=2)
def test_make_chimera_file(self):
"""Test make_chimera_file() function"""
cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax')
+ with open('out.chimerax') as fh:
+ lines = fh.readlines()
+ self.assertEqual(lines[-4], 'open_files("url1", "url2")\n')
os.unlink('out.chimerax')
if __name__ == '__main__':
unittest.main() |
42f0c76664337af80d692fe7649f3643c237cc47 | Tests/MathFunctionsTest.py | Tests/MathFunctionsTest.py | from Math.MathFunctions import *
def pointTest():
point1 = (0, 0)
point2 = (2, 4)
print("Point 1: {}".format(point1))
print("Point 2: {}".format(point2))
print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1])))
angle = pointAngle(point1[0],point1[1],point2[0],point2[1]);
print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG))
pointTest()
| from Math.MathFunctions import *
import unittest
class TestPointMethods(unittest.TestCase):
def test_point(self):
point1 = (0, 0)
point2 = (2, 4)
angle = pointAngle(point1[0], point1[1], point2[0], point2[1])
dist = pointDistance(point1[0], point1[1], point2[0], point2[1])
self.assertAlmostEqual(angle, 1.1071487177940904)
self.assertAlmostEqual(dist, 4.47213595499958)
class TestHelperMethods(unittest.TestCase):
def test_clamp(self):
self.assertEqual(clamp(10, 1, 5), 5)
self.assertEqual(clamp(0, 1, 5), 1)
self.assertEqual(clamp(3, 1, 5), 3)
self.assertEqual(clamp(5, 1, 5), 5)
if __name__ == '__main__':
unittest.main()
| Use python's unit testing framework | Use python's unit testing framework
| Python | mit | turtles/PythonScripts | from Math.MathFunctions import *
+ import unittest
- def pointTest():
+ class TestPointMethods(unittest.TestCase):
+ def test_point(self):
- point1 = (0, 0)
+ point1 = (0, 0)
- point2 = (2, 4)
+ point2 = (2, 4)
- print("Point 1: {}".format(point1))
- print("Point 2: {}".format(point2))
- print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1])))
- angle = pointAngle(point1[0],point1[1],point2[0],point2[1]);
+ angle = pointAngle(point1[0], point1[1], point2[0], point2[1])
- print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG))
+ dist = pointDistance(point1[0], point1[1], point2[0], point2[1])
- pointTest()
+ self.assertAlmostEqual(angle, 1.1071487177940904)
+ self.assertAlmostEqual(dist, 4.47213595499958)
+ class TestHelperMethods(unittest.TestCase):
+ def test_clamp(self):
+ self.assertEqual(clamp(10, 1, 5), 5)
+ self.assertEqual(clamp(0, 1, 5), 1)
+ self.assertEqual(clamp(3, 1, 5), 3)
+ self.assertEqual(clamp(5, 1, 5), 5)
+
+ if __name__ == '__main__':
+ unittest.main()
+ | Use python's unit testing framework | ## Code Before:
from Math.MathFunctions import *
def pointTest():
point1 = (0, 0)
point2 = (2, 4)
print("Point 1: {}".format(point1))
print("Point 2: {}".format(point2))
print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1])))
angle = pointAngle(point1[0],point1[1],point2[0],point2[1]);
print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG))
pointTest()
## Instruction:
Use python's unit testing framework
## Code After:
from Math.MathFunctions import *
import unittest
class TestPointMethods(unittest.TestCase):
def test_point(self):
point1 = (0, 0)
point2 = (2, 4)
angle = pointAngle(point1[0], point1[1], point2[0], point2[1])
dist = pointDistance(point1[0], point1[1], point2[0], point2[1])
self.assertAlmostEqual(angle, 1.1071487177940904)
self.assertAlmostEqual(dist, 4.47213595499958)
class TestHelperMethods(unittest.TestCase):
def test_clamp(self):
self.assertEqual(clamp(10, 1, 5), 5)
self.assertEqual(clamp(0, 1, 5), 1)
self.assertEqual(clamp(3, 1, 5), 3)
self.assertEqual(clamp(5, 1, 5), 5)
if __name__ == '__main__':
unittest.main()
| from Math.MathFunctions import *
+ import unittest
- def pointTest():
+ class TestPointMethods(unittest.TestCase):
+ def test_point(self):
- point1 = (0, 0)
+ point1 = (0, 0)
? ++++
- point2 = (2, 4)
+ point2 = (2, 4)
? ++++
- print("Point 1: {}".format(point1))
- print("Point 2: {}".format(point2))
- print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1])))
- angle = pointAngle(point1[0],point1[1],point2[0],point2[1]);
? -
+ angle = pointAngle(point1[0], point1[1], point2[0], point2[1])
? ++++ + + +
- print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG))
+ dist = pointDistance(point1[0], point1[1], point2[0], point2[1])
- pointTest()
+ self.assertAlmostEqual(angle, 1.1071487177940904)
+ self.assertAlmostEqual(dist, 4.47213595499958)
+
+ class TestHelperMethods(unittest.TestCase):
+ def test_clamp(self):
+ self.assertEqual(clamp(10, 1, 5), 5)
+ self.assertEqual(clamp(0, 1, 5), 1)
+ self.assertEqual(clamp(3, 1, 5), 3)
+ self.assertEqual(clamp(5, 1, 5), 5)
+
+ if __name__ == '__main__':
+ unittest.main() |
45381a1ce6e271cc06ce130cb35a93f14eceba90 | troposphere/utils.py | troposphere/utils.py | import time
def _tail_print(e):
print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id))
def get_events(conn, stackname):
"""Get the events in batches and return in chronological order"""
next = None
event_list = []
while 1:
events = conn.describe_stack_events(stackname, next)
event_list.append(events)
if events.next_token is None:
break
next = events.next_token
time.sleep(1)
return reversed(sum(event_list, []))
def tail(conn, stack_name, log_func=_tail_print, sleep_time=5):
"""Show and then tail the event log"""
# First dump the full list of events in chronological order and keep
# track of the events we've seen already
seen = set()
initial_events = get_events(conn, stack_name)
for e in initial_events:
log_func(e)
seen.add(e.event_id)
# Now keep looping through and dump the new events
while 1:
events = get_events(conn, stack_name)
for e in events:
if e.event_id not in seen:
log_func(e)
seen.add(e.event_id)
time.sleep(sleep_time)
| import time
def _tail_print(e):
print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id))
def get_events(conn, stackname):
"""Get the events in batches and return in chronological order"""
next = None
event_list = []
while 1:
events = conn.describe_stack_events(stackname, next)
event_list.append(events)
if events.next_token is None:
break
next = events.next_token
time.sleep(1)
return reversed(sum(event_list, []))
def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True):
"""Show and then tail the event log"""
# First dump the full list of events in chronological order and keep
# track of the events we've seen already
seen = set()
initial_events = get_events(conn, stack_name)
for e in initial_events:
if include_initial:
log_func(e)
seen.add(e.event_id)
# Now keep looping through and dump the new events
while 1:
events = get_events(conn, stack_name)
for e in events:
if e.event_id not in seen:
log_func(e)
seen.add(e.event_id)
time.sleep(sleep_time)
| Add "include_initial" kwarg to support tailing stack updates | Add "include_initial" kwarg to support tailing stack updates
`get_events` will return all events that have occurred for a stack. This
is useless if we're tailing an update to a stack.
| Python | bsd-2-clause | mhahn/troposphere | import time
def _tail_print(e):
print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id))
def get_events(conn, stackname):
"""Get the events in batches and return in chronological order"""
next = None
event_list = []
while 1:
events = conn.describe_stack_events(stackname, next)
event_list.append(events)
if events.next_token is None:
break
next = events.next_token
time.sleep(1)
return reversed(sum(event_list, []))
- def tail(conn, stack_name, log_func=_tail_print, sleep_time=5):
+ def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True):
"""Show and then tail the event log"""
# First dump the full list of events in chronological order and keep
# track of the events we've seen already
seen = set()
initial_events = get_events(conn, stack_name)
for e in initial_events:
+ if include_initial:
- log_func(e)
+ log_func(e)
seen.add(e.event_id)
# Now keep looping through and dump the new events
while 1:
events = get_events(conn, stack_name)
for e in events:
if e.event_id not in seen:
log_func(e)
seen.add(e.event_id)
time.sleep(sleep_time)
| Add "include_initial" kwarg to support tailing stack updates | ## Code Before:
import time
def _tail_print(e):
print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id))
def get_events(conn, stackname):
"""Get the events in batches and return in chronological order"""
next = None
event_list = []
while 1:
events = conn.describe_stack_events(stackname, next)
event_list.append(events)
if events.next_token is None:
break
next = events.next_token
time.sleep(1)
return reversed(sum(event_list, []))
def tail(conn, stack_name, log_func=_tail_print, sleep_time=5):
"""Show and then tail the event log"""
# First dump the full list of events in chronological order and keep
# track of the events we've seen already
seen = set()
initial_events = get_events(conn, stack_name)
for e in initial_events:
log_func(e)
seen.add(e.event_id)
# Now keep looping through and dump the new events
while 1:
events = get_events(conn, stack_name)
for e in events:
if e.event_id not in seen:
log_func(e)
seen.add(e.event_id)
time.sleep(sleep_time)
## Instruction:
Add "include_initial" kwarg to support tailing stack updates
## Code After:
import time
def _tail_print(e):
print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id))
def get_events(conn, stackname):
"""Get the events in batches and return in chronological order"""
next = None
event_list = []
while 1:
events = conn.describe_stack_events(stackname, next)
event_list.append(events)
if events.next_token is None:
break
next = events.next_token
time.sleep(1)
return reversed(sum(event_list, []))
def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True):
"""Show and then tail the event log"""
# First dump the full list of events in chronological order and keep
# track of the events we've seen already
seen = set()
initial_events = get_events(conn, stack_name)
for e in initial_events:
if include_initial:
log_func(e)
seen.add(e.event_id)
# Now keep looping through and dump the new events
while 1:
events = get_events(conn, stack_name)
for e in events:
if e.event_id not in seen:
log_func(e)
seen.add(e.event_id)
time.sleep(sleep_time)
| import time
def _tail_print(e):
print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id))
def get_events(conn, stackname):
"""Get the events in batches and return in chronological order"""
next = None
event_list = []
while 1:
events = conn.describe_stack_events(stackname, next)
event_list.append(events)
if events.next_token is None:
break
next = events.next_token
time.sleep(1)
return reversed(sum(event_list, []))
- def tail(conn, stack_name, log_func=_tail_print, sleep_time=5):
+ def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True):
? ++++++++++++++++++++++
"""Show and then tail the event log"""
# First dump the full list of events in chronological order and keep
# track of the events we've seen already
seen = set()
initial_events = get_events(conn, stack_name)
for e in initial_events:
+ if include_initial:
- log_func(e)
+ log_func(e)
? ++++
seen.add(e.event_id)
# Now keep looping through and dump the new events
while 1:
events = get_events(conn, stack_name)
for e in events:
if e.event_id not in seen:
log_func(e)
seen.add(e.event_id)
time.sleep(sleep_time) |
5a76457e2b9596ad3497b0145410a2f4090a5c54 | tests/mixins.py | tests/mixins.py | class RedisCleanupMixin(object):
client = None
prefix = None
def setUp(self):
super(RedisCleanupMixin, self).setUp()
self.assertIsNotNone(self.client, "Need a redis client to be provided")
def tearDown(self):
root = '*'
if self.prefix is not None:
root = '{}*'.format(self.prefix)
keys = self.client.keys(root)
for key in keys:
self.client.delete(key)
super(RedisCleanupMixin, self).tearDown()
| class RedisCleanupMixin(object):
client = None
prefix = NotImplemented # type: str
def setUp(self):
super(RedisCleanupMixin, self).setUp()
self.assertIsNotNone(self.client, "Need a redis client to be provided")
def tearDown(self):
root = '*'
if self.prefix is not None:
root = '{}*'.format(self.prefix)
keys = self.client.keys(root)
for key in keys:
self.client.delete(key)
super(RedisCleanupMixin, self).tearDown()
| Add annotation required for mypy | Add annotation required for mypy
| Python | bsd-3-clause | thread/django-lightweight-queue,thread/django-lightweight-queue | class RedisCleanupMixin(object):
client = None
- prefix = None
+ prefix = NotImplemented # type: str
def setUp(self):
super(RedisCleanupMixin, self).setUp()
self.assertIsNotNone(self.client, "Need a redis client to be provided")
def tearDown(self):
root = '*'
if self.prefix is not None:
root = '{}*'.format(self.prefix)
keys = self.client.keys(root)
for key in keys:
self.client.delete(key)
super(RedisCleanupMixin, self).tearDown()
| Add annotation required for mypy | ## Code Before:
class RedisCleanupMixin(object):
client = None
prefix = None
def setUp(self):
super(RedisCleanupMixin, self).setUp()
self.assertIsNotNone(self.client, "Need a redis client to be provided")
def tearDown(self):
root = '*'
if self.prefix is not None:
root = '{}*'.format(self.prefix)
keys = self.client.keys(root)
for key in keys:
self.client.delete(key)
super(RedisCleanupMixin, self).tearDown()
## Instruction:
Add annotation required for mypy
## Code After:
class RedisCleanupMixin(object):
client = None
prefix = NotImplemented # type: str
def setUp(self):
super(RedisCleanupMixin, self).setUp()
self.assertIsNotNone(self.client, "Need a redis client to be provided")
def tearDown(self):
root = '*'
if self.prefix is not None:
root = '{}*'.format(self.prefix)
keys = self.client.keys(root)
for key in keys:
self.client.delete(key)
super(RedisCleanupMixin, self).tearDown()
| class RedisCleanupMixin(object):
client = None
- prefix = None
+ prefix = NotImplemented # type: str
def setUp(self):
super(RedisCleanupMixin, self).setUp()
self.assertIsNotNone(self.client, "Need a redis client to be provided")
def tearDown(self):
root = '*'
if self.prefix is not None:
root = '{}*'.format(self.prefix)
keys = self.client.keys(root)
for key in keys:
self.client.delete(key)
super(RedisCleanupMixin, self).tearDown() |
28c88cbc34dcf2af5c98ce3f3eed3774dd5be15e | lcapy/discretetime.py | lcapy/discretetime.py |
import sympy as sym
from .sym import sympify
from .nexpr import nexpr, n
from .kexpr import kexpr, k
from .zexpr import zexpr, z
from .dsym import nsym, ksym, zsym, dt, df
from .expr import expr as expr1
from .transform import transform as transform1
from .transform import call as call1
from .functions import Function
from .ztransform import *
from .seq import seq
def expr(arg, **assumptions):
expr = sympify(arg, **assumptions)
symbols = expr.free_symbols
if nsym in symbols:
return nexpr(expr, **assumptions)
elif ksym in symbols:
return kexpr(expr, **assumptions)
elif zsym in symbols:
return zexpr(expr, **assumptions)
return expr1(arg, **assumptions)
|
import sympy as sym
from .sym import sympify
from .nexpr import nexpr, n
from .kexpr import kexpr, k
from .zexpr import zexpr, z
from .dsym import nsym, ksym, zsym, dt, df
from .expr import expr as expr1
from .transform import transform as transform1
from .transform import call as call1
from .functions import Function
from .ztransform import *
from .seq import seq
def expr(arg, **assumptions):
# Handle container args.
if not isinstance(arg, str) and hasattr(arg, '__iter__'):
return expr1(arg, **assumptions)
expr = sympify(arg, **assumptions)
symbols = expr.free_symbols
if nsym in symbols:
return nexpr(expr, **assumptions)
elif ksym in symbols:
return kexpr(expr, **assumptions)
elif zsym in symbols:
return zexpr(expr, **assumptions)
return expr1(arg, **assumptions)
| Handle container types for discrete-time expr | Handle container types for discrete-time expr
| Python | lgpl-2.1 | mph-/lcapy |
import sympy as sym
from .sym import sympify
from .nexpr import nexpr, n
from .kexpr import kexpr, k
from .zexpr import zexpr, z
from .dsym import nsym, ksym, zsym, dt, df
from .expr import expr as expr1
from .transform import transform as transform1
from .transform import call as call1
from .functions import Function
from .ztransform import *
from .seq import seq
def expr(arg, **assumptions):
+ # Handle container args.
+ if not isinstance(arg, str) and hasattr(arg, '__iter__'):
+ return expr1(arg, **assumptions)
+
expr = sympify(arg, **assumptions)
symbols = expr.free_symbols
if nsym in symbols:
return nexpr(expr, **assumptions)
elif ksym in symbols:
return kexpr(expr, **assumptions)
elif zsym in symbols:
return zexpr(expr, **assumptions)
return expr1(arg, **assumptions)
| Handle container types for discrete-time expr | ## Code Before:
import sympy as sym
from .sym import sympify
from .nexpr import nexpr, n
from .kexpr import kexpr, k
from .zexpr import zexpr, z
from .dsym import nsym, ksym, zsym, dt, df
from .expr import expr as expr1
from .transform import transform as transform1
from .transform import call as call1
from .functions import Function
from .ztransform import *
from .seq import seq
def expr(arg, **assumptions):
expr = sympify(arg, **assumptions)
symbols = expr.free_symbols
if nsym in symbols:
return nexpr(expr, **assumptions)
elif ksym in symbols:
return kexpr(expr, **assumptions)
elif zsym in symbols:
return zexpr(expr, **assumptions)
return expr1(arg, **assumptions)
## Instruction:
Handle container types for discrete-time expr
## Code After:
import sympy as sym
from .sym import sympify
from .nexpr import nexpr, n
from .kexpr import kexpr, k
from .zexpr import zexpr, z
from .dsym import nsym, ksym, zsym, dt, df
from .expr import expr as expr1
from .transform import transform as transform1
from .transform import call as call1
from .functions import Function
from .ztransform import *
from .seq import seq
def expr(arg, **assumptions):
# Handle container args.
if not isinstance(arg, str) and hasattr(arg, '__iter__'):
return expr1(arg, **assumptions)
expr = sympify(arg, **assumptions)
symbols = expr.free_symbols
if nsym in symbols:
return nexpr(expr, **assumptions)
elif ksym in symbols:
return kexpr(expr, **assumptions)
elif zsym in symbols:
return zexpr(expr, **assumptions)
return expr1(arg, **assumptions)
|
import sympy as sym
from .sym import sympify
from .nexpr import nexpr, n
from .kexpr import kexpr, k
from .zexpr import zexpr, z
from .dsym import nsym, ksym, zsym, dt, df
from .expr import expr as expr1
from .transform import transform as transform1
from .transform import call as call1
from .functions import Function
from .ztransform import *
from .seq import seq
def expr(arg, **assumptions):
+ # Handle container args.
+ if not isinstance(arg, str) and hasattr(arg, '__iter__'):
+ return expr1(arg, **assumptions)
+
expr = sympify(arg, **assumptions)
symbols = expr.free_symbols
if nsym in symbols:
return nexpr(expr, **assumptions)
elif ksym in symbols:
return kexpr(expr, **assumptions)
elif zsym in symbols:
return zexpr(expr, **assumptions)
return expr1(arg, **assumptions)
|
521ebf29990de4d997c90f4168ea300d75776cfc | components/utilities.py | components/utilities.py | """Utilities for general operations."""
def IsNumeric(num_str):
try:
val = int(num_str)
except ValueError:
return False
else:
return True
| """Utilities for general operations."""
def IsNumeric(num_str):
try:
val = int(num_str)
except ValueError:
return False
else:
return True
def GuaranteeUnicode(obj):
if type(obj) == unicode:
return obj
elif type(obj) == str:
return unicode(obj, "utf-8")
else:
return unicode(str(obj), "utf-8")
| Add GuranteeUnicode function which always returns a unicode object | Add GuranteeUnicode function which always returns a unicode object
| Python | mit | lnishan/SQLGitHub | """Utilities for general operations."""
def IsNumeric(num_str):
try:
val = int(num_str)
except ValueError:
return False
else:
return True
+ def GuaranteeUnicode(obj):
+ if type(obj) == unicode:
+ return obj
+ elif type(obj) == str:
+ return unicode(obj, "utf-8")
+ else:
+ return unicode(str(obj), "utf-8")
+ | Add GuranteeUnicode function which always returns a unicode object | ## Code Before:
"""Utilities for general operations."""
def IsNumeric(num_str):
try:
val = int(num_str)
except ValueError:
return False
else:
return True
## Instruction:
Add GuranteeUnicode function which always returns a unicode object
## Code After:
"""Utilities for general operations."""
def IsNumeric(num_str):
try:
val = int(num_str)
except ValueError:
return False
else:
return True
def GuaranteeUnicode(obj):
if type(obj) == unicode:
return obj
elif type(obj) == str:
return unicode(obj, "utf-8")
else:
return unicode(str(obj), "utf-8")
| """Utilities for general operations."""
def IsNumeric(num_str):
try:
val = int(num_str)
except ValueError:
return False
else:
return True
+
+ def GuaranteeUnicode(obj):
+ if type(obj) == unicode:
+ return obj
+ elif type(obj) == str:
+ return unicode(obj, "utf-8")
+ else:
+ return unicode(str(obj), "utf-8") |
ff02f0e8a7b62d5afdc88730129a5e0811fb5a82 | monitor/monitor_test.py | monitor/monitor_test.py |
import unittest
import monitor
class TestMonitor(unittest.TestCase):
def test_attr(self):
mon = monitor.Monitor()
assert mon.pvprefix == "MON-CONTROL:"
assert mon.monitorname == "PI1"
def test_testimage(self):
mon = monitor.Monitor()
image = mon.testimage()
assert len(image) != 0
# print image
def test_readPV(self):
mon = monitor.Monitor()
pv_result = mon.readPV()
assert pv_result == "CAM1"
if __name__ == '__main__':
unittest.main()
|
import unittest
from mock import patch, MagicMock, PropertyMock
from monitor import Monitor
class MonitorUpdateTest(unittest.TestCase):
def setUp(self):
with patch('monitor.pv.PV'):
mock_plotter = MagicMock()
self.monitor = Monitor("MYNAME", mock_plotter)
def test_does_nothing_if_camera_is_None(self):
self.monitor.camera = None
try:
self.monitor._update_image()
self.assertTrue(True)
except Exception as ex:
self.fail("Unexpected expection thrown" + str(ex))
def test_gets_image_data_from_camera_when_not_None(self):
mock_camera = MagicMock()
self.monitor.camera = mock_camera
self.monitor._update_image()
mock_camera.get_image_data.assert_called_once_with()
def test_get_size_data_from_camera_when_not_None(self):
mock_camera = MagicMock()
mock_xsize = PropertyMock(return_value=100)
mock_ysize = PropertyMock(return_value=200)
type(mock_camera).xsize = mock_xsize
type(mock_camera).ysize = mock_ysize
self.monitor.camera = mock_camera
self.monitor._update_image()
mock_xsize.assert_called_once_with()
mock_ysize.assert_called_once_with()
def test_calls_plotter_with_image_and_size_data(self):
data = 111
xsize = 100
ysize = 200
mock_camera = MagicMock(xsize=xsize, ysize=ysize)
mock_camera.get_image_data = MagicMock(return_value=data)
self.monitor.camera = mock_camera
self.monitor._update_image()
self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize)
if __name__ == '__main__':
unittest.main()
| Rewrite monitortests. Add tests for update_image | Rewrite monitortests. Add tests for update_image
| Python | apache-2.0 | nickbattam/picamon,nickbattam/picamon,nickbattam/picamon,nickbattam/picamon |
import unittest
- import monitor
+ from mock import patch, MagicMock, PropertyMock
- class TestMonitor(unittest.TestCase):
+ from monitor import Monitor
- def test_attr(self):
- mon = monitor.Monitor()
- assert mon.pvprefix == "MON-CONTROL:"
- assert mon.monitorname == "PI1"
+ class MonitorUpdateTest(unittest.TestCase):
- def test_testimage(self):
- mon = monitor.Monitor()
- image = mon.testimage()
- assert len(image) != 0
- # print image
- def test_readPV(self):
- mon = monitor.Monitor()
- pv_result = mon.readPV()
- assert pv_result == "CAM1"
+ def setUp(self):
+ with patch('monitor.pv.PV'):
+ mock_plotter = MagicMock()
+ self.monitor = Monitor("MYNAME", mock_plotter)
+
+ def test_does_nothing_if_camera_is_None(self):
+ self.monitor.camera = None
+
+ try:
+ self.monitor._update_image()
+ self.assertTrue(True)
+ except Exception as ex:
+ self.fail("Unexpected expection thrown" + str(ex))
+
+
+ def test_gets_image_data_from_camera_when_not_None(self):
+ mock_camera = MagicMock()
+ self.monitor.camera = mock_camera
+
+ self.monitor._update_image()
+
+ mock_camera.get_image_data.assert_called_once_with()
+
+ def test_get_size_data_from_camera_when_not_None(self):
+ mock_camera = MagicMock()
+
+ mock_xsize = PropertyMock(return_value=100)
+ mock_ysize = PropertyMock(return_value=200)
+ type(mock_camera).xsize = mock_xsize
+ type(mock_camera).ysize = mock_ysize
+
+ self.monitor.camera = mock_camera
+
+ self.monitor._update_image()
+
+ mock_xsize.assert_called_once_with()
+ mock_ysize.assert_called_once_with()
+
+ def test_calls_plotter_with_image_and_size_data(self):
+ data = 111
+ xsize = 100
+ ysize = 200
+ mock_camera = MagicMock(xsize=xsize, ysize=ysize)
+ mock_camera.get_image_data = MagicMock(return_value=data)
+ self.monitor.camera = mock_camera
+
+ self.monitor._update_image()
+
+ self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize)
if __name__ == '__main__':
- unittest.main()
+ unittest.main()
| Rewrite monitortests. Add tests for update_image | ## Code Before:
import unittest
import monitor
class TestMonitor(unittest.TestCase):
def test_attr(self):
mon = monitor.Monitor()
assert mon.pvprefix == "MON-CONTROL:"
assert mon.monitorname == "PI1"
def test_testimage(self):
mon = monitor.Monitor()
image = mon.testimage()
assert len(image) != 0
# print image
def test_readPV(self):
mon = monitor.Monitor()
pv_result = mon.readPV()
assert pv_result == "CAM1"
if __name__ == '__main__':
unittest.main()
## Instruction:
Rewrite monitortests. Add tests for update_image
## Code After:
import unittest
from mock import patch, MagicMock, PropertyMock
from monitor import Monitor
class MonitorUpdateTest(unittest.TestCase):
def setUp(self):
with patch('monitor.pv.PV'):
mock_plotter = MagicMock()
self.monitor = Monitor("MYNAME", mock_plotter)
def test_does_nothing_if_camera_is_None(self):
self.monitor.camera = None
try:
self.monitor._update_image()
self.assertTrue(True)
except Exception as ex:
self.fail("Unexpected expection thrown" + str(ex))
def test_gets_image_data_from_camera_when_not_None(self):
mock_camera = MagicMock()
self.monitor.camera = mock_camera
self.monitor._update_image()
mock_camera.get_image_data.assert_called_once_with()
def test_get_size_data_from_camera_when_not_None(self):
mock_camera = MagicMock()
mock_xsize = PropertyMock(return_value=100)
mock_ysize = PropertyMock(return_value=200)
type(mock_camera).xsize = mock_xsize
type(mock_camera).ysize = mock_ysize
self.monitor.camera = mock_camera
self.monitor._update_image()
mock_xsize.assert_called_once_with()
mock_ysize.assert_called_once_with()
def test_calls_plotter_with_image_and_size_data(self):
data = 111
xsize = 100
ysize = 200
mock_camera = MagicMock(xsize=xsize, ysize=ysize)
mock_camera.get_image_data = MagicMock(return_value=data)
self.monitor.camera = mock_camera
self.monitor._update_image()
self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize)
if __name__ == '__main__':
unittest.main()
|
import unittest
- import monitor
+ from mock import patch, MagicMock, PropertyMock
- class TestMonitor(unittest.TestCase):
+ from monitor import Monitor
- def test_attr(self):
- mon = monitor.Monitor()
- assert mon.pvprefix == "MON-CONTROL:"
- assert mon.monitorname == "PI1"
+ class MonitorUpdateTest(unittest.TestCase):
- def test_testimage(self):
- mon = monitor.Monitor()
- image = mon.testimage()
- assert len(image) != 0
- # print image
- def test_readPV(self):
- mon = monitor.Monitor()
- pv_result = mon.readPV()
- assert pv_result == "CAM1"
+ def setUp(self):
+ with patch('monitor.pv.PV'):
+ mock_plotter = MagicMock()
+ self.monitor = Monitor("MYNAME", mock_plotter)
+
+ def test_does_nothing_if_camera_is_None(self):
+ self.monitor.camera = None
+
+ try:
+ self.monitor._update_image()
+ self.assertTrue(True)
+ except Exception as ex:
+ self.fail("Unexpected expection thrown" + str(ex))
+
+
+ def test_gets_image_data_from_camera_when_not_None(self):
+ mock_camera = MagicMock()
+ self.monitor.camera = mock_camera
+
+ self.monitor._update_image()
+
+ mock_camera.get_image_data.assert_called_once_with()
+
+ def test_get_size_data_from_camera_when_not_None(self):
+ mock_camera = MagicMock()
+
+ mock_xsize = PropertyMock(return_value=100)
+ mock_ysize = PropertyMock(return_value=200)
+ type(mock_camera).xsize = mock_xsize
+ type(mock_camera).ysize = mock_ysize
+
+ self.monitor.camera = mock_camera
+
+ self.monitor._update_image()
+
+ mock_xsize.assert_called_once_with()
+ mock_ysize.assert_called_once_with()
+
+ def test_calls_plotter_with_image_and_size_data(self):
+ data = 111
+ xsize = 100
+ ysize = 200
+ mock_camera = MagicMock(xsize=xsize, ysize=ysize)
+ mock_camera.get_image_data = MagicMock(return_value=data)
+ self.monitor.camera = mock_camera
+
+ self.monitor._update_image()
+
+ self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize)
if __name__ == '__main__':
- unittest.main()
? ^
+ unittest.main()
? ^^^^
|
e9c4881ee29ba104caf9fc8330583c254fe52c06 | scripts/examples/Arduino/Portenta-H7/19-Low-Power/deep_sleep.py | scripts/examples/Arduino/Portenta-H7/19-Low-Power/deep_sleep.py | import pyb, machine, sensor
# Create and init RTC object.
rtc = pyb.RTC()
# (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]])
rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0))
# Print RTC info.
print(rtc.datetime())
sensor.reset()
# Enable sensor softsleep
sensor.sleep(True)
# Optionally bypass the regulator on OV7725
# for the lowest possible power consumption.
if (sensor.get_id() == sensor.OV7725):
# Bypass internal regulator
sensor.__write_reg(0x4F, 0x18)
# Shutdown the sensor (pulls PWDN high).
sensor.shutdown(True)
# Enable RTC interrupts every 30 seconds.
# Note the camera will RESET after wakeup from Deepsleep Mode.
rtc.wakeup(30000)
# Enter Deepsleep Mode.
machine.deepsleep()
| import pyb, machine, sensor
# Create and init RTC object.
rtc = pyb.RTC()
# (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]])
rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0))
# Print RTC info.
print(rtc.datetime())
sensor.reset()
# Shutdown the sensor (pulls PWDN high).
sensor.shutdown(True)
# Enable RTC interrupts every 30 seconds.
# Note the camera will RESET after wakeup from Deepsleep Mode.
rtc.wakeup(30000)
# Enter Deepsleep Mode.
machine.deepsleep()
| Remove sensor setting from deep sleep example | Remove sensor setting from deep sleep example
| Python | mit | iabdalkader/openmv,openmv/openmv,kwagyeman/openmv,iabdalkader/openmv,openmv/openmv,kwagyeman/openmv,openmv/openmv,iabdalkader/openmv,kwagyeman/openmv,iabdalkader/openmv,kwagyeman/openmv,openmv/openmv | import pyb, machine, sensor
# Create and init RTC object.
rtc = pyb.RTC()
# (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]])
rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0))
# Print RTC info.
print(rtc.datetime())
sensor.reset()
- # Enable sensor softsleep
- sensor.sleep(True)
-
- # Optionally bypass the regulator on OV7725
- # for the lowest possible power consumption.
- if (sensor.get_id() == sensor.OV7725):
- # Bypass internal regulator
- sensor.__write_reg(0x4F, 0x18)
-
# Shutdown the sensor (pulls PWDN high).
sensor.shutdown(True)
# Enable RTC interrupts every 30 seconds.
# Note the camera will RESET after wakeup from Deepsleep Mode.
rtc.wakeup(30000)
# Enter Deepsleep Mode.
machine.deepsleep()
| Remove sensor setting from deep sleep example | ## Code Before:
import pyb, machine, sensor
# Create and init RTC object.
rtc = pyb.RTC()
# (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]])
rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0))
# Print RTC info.
print(rtc.datetime())
sensor.reset()
# Enable sensor softsleep
sensor.sleep(True)
# Optionally bypass the regulator on OV7725
# for the lowest possible power consumption.
if (sensor.get_id() == sensor.OV7725):
# Bypass internal regulator
sensor.__write_reg(0x4F, 0x18)
# Shutdown the sensor (pulls PWDN high).
sensor.shutdown(True)
# Enable RTC interrupts every 30 seconds.
# Note the camera will RESET after wakeup from Deepsleep Mode.
rtc.wakeup(30000)
# Enter Deepsleep Mode.
machine.deepsleep()
## Instruction:
Remove sensor setting from deep sleep example
## Code After:
import pyb, machine, sensor
# Create and init RTC object.
rtc = pyb.RTC()
# (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]])
rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0))
# Print RTC info.
print(rtc.datetime())
sensor.reset()
# Shutdown the sensor (pulls PWDN high).
sensor.shutdown(True)
# Enable RTC interrupts every 30 seconds.
# Note the camera will RESET after wakeup from Deepsleep Mode.
rtc.wakeup(30000)
# Enter Deepsleep Mode.
machine.deepsleep()
| import pyb, machine, sensor
# Create and init RTC object.
rtc = pyb.RTC()
# (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]])
rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0))
# Print RTC info.
print(rtc.datetime())
sensor.reset()
- # Enable sensor softsleep
- sensor.sleep(True)
-
- # Optionally bypass the regulator on OV7725
- # for the lowest possible power consumption.
- if (sensor.get_id() == sensor.OV7725):
- # Bypass internal regulator
- sensor.__write_reg(0x4F, 0x18)
-
# Shutdown the sensor (pulls PWDN high).
sensor.shutdown(True)
# Enable RTC interrupts every 30 seconds.
# Note the camera will RESET after wakeup from Deepsleep Mode.
rtc.wakeup(30000)
# Enter Deepsleep Mode.
machine.deepsleep() |
d491aea2da5d52245001f4da24331f33e4a3a299 | importlib_metadata/_meta.py | importlib_metadata/_meta.py | from ._compat import Protocol
from typing import Any, Dict, Iterator, List, TypeVar, Union
_T = TypeVar("_T")
class PackageMetadata(Protocol):
def __len__(self) -> int:
... # pragma: no cover
def __contains__(self, item: str) -> bool:
... # pragma: no cover
def __getitem__(self, key: str) -> str:
... # pragma: no cover
def __iter__(self) -> Iterator[str]:
... # pragma: no cover
def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]:
"""
Return all values associated with a possibly multi-valued key.
"""
@property
def json(self) -> Dict[str, Union[str, List[str]]]:
"""
A JSON-compatible form of the metadata.
"""
class SimplePath(Protocol):
"""
A minimal subset of pathlib.Path required by PathDistribution.
"""
def joinpath(self) -> 'SimplePath':
... # pragma: no cover
def __div__(self) -> 'SimplePath':
... # pragma: no cover
def parent(self) -> 'SimplePath':
... # pragma: no cover
def read_text(self) -> str:
... # pragma: no cover
| from ._compat import Protocol
from typing import Any, Dict, Iterator, List, TypeVar, Union
_T = TypeVar("_T")
class PackageMetadata(Protocol):
def __len__(self) -> int:
... # pragma: no cover
def __contains__(self, item: str) -> bool:
... # pragma: no cover
def __getitem__(self, key: str) -> str:
... # pragma: no cover
def __iter__(self) -> Iterator[str]:
... # pragma: no cover
def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]:
"""
Return all values associated with a possibly multi-valued key.
"""
@property
def json(self) -> Dict[str, Union[str, List[str]]]:
"""
A JSON-compatible form of the metadata.
"""
class SimplePath(Protocol):
"""
A minimal subset of pathlib.Path required by PathDistribution.
>>> import pathlib
>>> import typing
>>> _: SimplePath = typing.cast(pathlib.Path, None)
"""
def joinpath(self) -> 'SimplePath':
... # pragma: no cover
def __div__(self) -> 'SimplePath':
... # pragma: no cover
def parent(self) -> 'SimplePath':
... # pragma: no cover
def read_text(self) -> str:
... # pragma: no cover
| Add test purported to capture the failure, but it still passes. | Add test purported to capture the failure, but it still passes.
| Python | apache-2.0 | python/importlib_metadata | from ._compat import Protocol
from typing import Any, Dict, Iterator, List, TypeVar, Union
_T = TypeVar("_T")
class PackageMetadata(Protocol):
def __len__(self) -> int:
... # pragma: no cover
def __contains__(self, item: str) -> bool:
... # pragma: no cover
def __getitem__(self, key: str) -> str:
... # pragma: no cover
def __iter__(self) -> Iterator[str]:
... # pragma: no cover
def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]:
"""
Return all values associated with a possibly multi-valued key.
"""
@property
def json(self) -> Dict[str, Union[str, List[str]]]:
"""
A JSON-compatible form of the metadata.
"""
class SimplePath(Protocol):
"""
A minimal subset of pathlib.Path required by PathDistribution.
+
+ >>> import pathlib
+ >>> import typing
+ >>> _: SimplePath = typing.cast(pathlib.Path, None)
"""
def joinpath(self) -> 'SimplePath':
... # pragma: no cover
def __div__(self) -> 'SimplePath':
... # pragma: no cover
def parent(self) -> 'SimplePath':
... # pragma: no cover
def read_text(self) -> str:
... # pragma: no cover
| Add test purported to capture the failure, but it still passes. | ## Code Before:
from ._compat import Protocol
from typing import Any, Dict, Iterator, List, TypeVar, Union
_T = TypeVar("_T")
class PackageMetadata(Protocol):
def __len__(self) -> int:
... # pragma: no cover
def __contains__(self, item: str) -> bool:
... # pragma: no cover
def __getitem__(self, key: str) -> str:
... # pragma: no cover
def __iter__(self) -> Iterator[str]:
... # pragma: no cover
def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]:
"""
Return all values associated with a possibly multi-valued key.
"""
@property
def json(self) -> Dict[str, Union[str, List[str]]]:
"""
A JSON-compatible form of the metadata.
"""
class SimplePath(Protocol):
"""
A minimal subset of pathlib.Path required by PathDistribution.
"""
def joinpath(self) -> 'SimplePath':
... # pragma: no cover
def __div__(self) -> 'SimplePath':
... # pragma: no cover
def parent(self) -> 'SimplePath':
... # pragma: no cover
def read_text(self) -> str:
... # pragma: no cover
## Instruction:
Add test purported to capture the failure, but it still passes.
## Code After:
from ._compat import Protocol
from typing import Any, Dict, Iterator, List, TypeVar, Union
_T = TypeVar("_T")
class PackageMetadata(Protocol):
def __len__(self) -> int:
... # pragma: no cover
def __contains__(self, item: str) -> bool:
... # pragma: no cover
def __getitem__(self, key: str) -> str:
... # pragma: no cover
def __iter__(self) -> Iterator[str]:
... # pragma: no cover
def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]:
"""
Return all values associated with a possibly multi-valued key.
"""
@property
def json(self) -> Dict[str, Union[str, List[str]]]:
"""
A JSON-compatible form of the metadata.
"""
class SimplePath(Protocol):
"""
A minimal subset of pathlib.Path required by PathDistribution.
>>> import pathlib
>>> import typing
>>> _: SimplePath = typing.cast(pathlib.Path, None)
"""
def joinpath(self) -> 'SimplePath':
... # pragma: no cover
def __div__(self) -> 'SimplePath':
... # pragma: no cover
def parent(self) -> 'SimplePath':
... # pragma: no cover
def read_text(self) -> str:
... # pragma: no cover
| from ._compat import Protocol
from typing import Any, Dict, Iterator, List, TypeVar, Union
_T = TypeVar("_T")
class PackageMetadata(Protocol):
def __len__(self) -> int:
... # pragma: no cover
def __contains__(self, item: str) -> bool:
... # pragma: no cover
def __getitem__(self, key: str) -> str:
... # pragma: no cover
def __iter__(self) -> Iterator[str]:
... # pragma: no cover
def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]:
"""
Return all values associated with a possibly multi-valued key.
"""
@property
def json(self) -> Dict[str, Union[str, List[str]]]:
"""
A JSON-compatible form of the metadata.
"""
class SimplePath(Protocol):
"""
A minimal subset of pathlib.Path required by PathDistribution.
+
+ >>> import pathlib
+ >>> import typing
+ >>> _: SimplePath = typing.cast(pathlib.Path, None)
"""
def joinpath(self) -> 'SimplePath':
... # pragma: no cover
def __div__(self) -> 'SimplePath':
... # pragma: no cover
def parent(self) -> 'SimplePath':
... # pragma: no cover
def read_text(self) -> str:
... # pragma: no cover |
ed350a7387c376538f51a8a7a8cfde5469baba8a | tests/testutils.py | tests/testutils.py | import psycopg2
import os
import getpass
def get_pg_connection():
return psycopg2.connect(
"dbname=bedquilt_test user={}".format(getpass.getuser())
)
| import psycopg2
import os
import getpass
# CREATE DATABASE bedquilt_test
# WITH OWNER = {{owner}}
# ENCODING = 'UTF8'
# TABLESPACE = pg_default
# LC_COLLATE = 'en_GB.UTF-8'
# LC_CTYPE = 'en_GB.UTF-8'
# CONNECTION LIMIT = -1;
def get_pg_connection():
return psycopg2.connect(
"dbname=bedquilt_test user={}".format(getpass.getuser())
)
| Add the sql to create the test database | Add the sql to create the test database
| Python | mit | BedquiltDB/bedquilt-core | import psycopg2
import os
import getpass
+
+
+ # CREATE DATABASE bedquilt_test
+ # WITH OWNER = {{owner}}
+ # ENCODING = 'UTF8'
+ # TABLESPACE = pg_default
+ # LC_COLLATE = 'en_GB.UTF-8'
+ # LC_CTYPE = 'en_GB.UTF-8'
+ # CONNECTION LIMIT = -1;
def get_pg_connection():
return psycopg2.connect(
"dbname=bedquilt_test user={}".format(getpass.getuser())
)
| Add the sql to create the test database | ## Code Before:
import psycopg2
import os
import getpass
def get_pg_connection():
return psycopg2.connect(
"dbname=bedquilt_test user={}".format(getpass.getuser())
)
## Instruction:
Add the sql to create the test database
## Code After:
import psycopg2
import os
import getpass
# CREATE DATABASE bedquilt_test
# WITH OWNER = {{owner}}
# ENCODING = 'UTF8'
# TABLESPACE = pg_default
# LC_COLLATE = 'en_GB.UTF-8'
# LC_CTYPE = 'en_GB.UTF-8'
# CONNECTION LIMIT = -1;
def get_pg_connection():
return psycopg2.connect(
"dbname=bedquilt_test user={}".format(getpass.getuser())
)
| import psycopg2
import os
import getpass
+
+
+ # CREATE DATABASE bedquilt_test
+ # WITH OWNER = {{owner}}
+ # ENCODING = 'UTF8'
+ # TABLESPACE = pg_default
+ # LC_COLLATE = 'en_GB.UTF-8'
+ # LC_CTYPE = 'en_GB.UTF-8'
+ # CONNECTION LIMIT = -1;
def get_pg_connection():
return psycopg2.connect(
"dbname=bedquilt_test user={}".format(getpass.getuser())
) |
8194f327032c064fe71ba3dc918e28ee2a586b12 | sqlalchemy_mixins/serialize.py | sqlalchemy_mixins/serialize.py | from collections.abc import Iterable
from .inspection import InspectionMixin
class SerializeMixin(InspectionMixin):
"""Mixin to make model serializable."""
__abstract__ = True
def to_dict(self,nested = False, hybrid_attributes = False, exclude = None):
"""Return dict object with model's data.
:param nested: flag to return nested relationships' data if true
:type: bool
:param include_hybrid: flag to include hybrid attributes if true
:return: dict
"""
result = dict()
if exclude is None:
view_cols = self.columns
else :
view_cols = filter(lambda e: e not in exclude, self.columns)
for key in view_cols :
result[key] = getattr(self, key)
if hybrid_attributes:
for key in self.hybrid_properties:
result[key] = getattr(self, key)
if nested:
for key in self.relations:
obj = getattr(self, key)
if isinstance(obj, SerializeMixin):
result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes)
elif isinstance(obj, Iterable):
result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj]
return result
| from collections.abc import Iterable
from .inspection import InspectionMixin
class SerializeMixin(InspectionMixin):
"""Mixin to make model serializable."""
__abstract__ = True
def to_dict(self,nested = False, hybrid_attributes = False, exclude = None):
"""Return dict object with model's data.
:param nested: flag to return nested relationships' data if true
:type: bool
:param include_hybrid: flag to include hybrid attributes if true
:return: dict
"""
result = dict()
if exclude is None:
view_cols = self.columns
else :
view_cols = filter(lambda e: e not in exclude, self.columns)
for key in view_cols :
result[key] = getattr(self, key)
if hybrid_attributes:
for key in self.hybrid_properties:
result[key] = getattr(self, key)
if nested:
for key in self.relations:
obj = getattr(self, key)
if isinstance(obj, SerializeMixin):
result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes)
elif isinstance(obj, Iterable):
result[key] = [
o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj
if isinstance(o, SerializeMixin)
]
return result
| Check if relation objects are class of SerializeMixin | Check if relation objects are class of SerializeMixin
| Python | mit | absent1706/sqlalchemy-mixins | from collections.abc import Iterable
from .inspection import InspectionMixin
class SerializeMixin(InspectionMixin):
"""Mixin to make model serializable."""
__abstract__ = True
def to_dict(self,nested = False, hybrid_attributes = False, exclude = None):
"""Return dict object with model's data.
:param nested: flag to return nested relationships' data if true
:type: bool
:param include_hybrid: flag to include hybrid attributes if true
:return: dict
"""
result = dict()
if exclude is None:
view_cols = self.columns
else :
view_cols = filter(lambda e: e not in exclude, self.columns)
for key in view_cols :
result[key] = getattr(self, key)
if hybrid_attributes:
for key in self.hybrid_properties:
result[key] = getattr(self, key)
if nested:
for key in self.relations:
obj = getattr(self, key)
if isinstance(obj, SerializeMixin):
result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes)
elif isinstance(obj, Iterable):
+ result[key] = [
- result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj]
+ o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj
+ if isinstance(o, SerializeMixin)
+ ]
return result
| Check if relation objects are class of SerializeMixin | ## Code Before:
from collections.abc import Iterable
from .inspection import InspectionMixin
class SerializeMixin(InspectionMixin):
"""Mixin to make model serializable."""
__abstract__ = True
def to_dict(self,nested = False, hybrid_attributes = False, exclude = None):
"""Return dict object with model's data.
:param nested: flag to return nested relationships' data if true
:type: bool
:param include_hybrid: flag to include hybrid attributes if true
:return: dict
"""
result = dict()
if exclude is None:
view_cols = self.columns
else :
view_cols = filter(lambda e: e not in exclude, self.columns)
for key in view_cols :
result[key] = getattr(self, key)
if hybrid_attributes:
for key in self.hybrid_properties:
result[key] = getattr(self, key)
if nested:
for key in self.relations:
obj = getattr(self, key)
if isinstance(obj, SerializeMixin):
result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes)
elif isinstance(obj, Iterable):
result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj]
return result
## Instruction:
Check if relation objects are class of SerializeMixin
## Code After:
from collections.abc import Iterable
from .inspection import InspectionMixin
class SerializeMixin(InspectionMixin):
"""Mixin to make model serializable."""
__abstract__ = True
def to_dict(self,nested = False, hybrid_attributes = False, exclude = None):
"""Return dict object with model's data.
:param nested: flag to return nested relationships' data if true
:type: bool
:param include_hybrid: flag to include hybrid attributes if true
:return: dict
"""
result = dict()
if exclude is None:
view_cols = self.columns
else :
view_cols = filter(lambda e: e not in exclude, self.columns)
for key in view_cols :
result[key] = getattr(self, key)
if hybrid_attributes:
for key in self.hybrid_properties:
result[key] = getattr(self, key)
if nested:
for key in self.relations:
obj = getattr(self, key)
if isinstance(obj, SerializeMixin):
result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes)
elif isinstance(obj, Iterable):
result[key] = [
o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj
if isinstance(o, SerializeMixin)
]
return result
| from collections.abc import Iterable
from .inspection import InspectionMixin
class SerializeMixin(InspectionMixin):
"""Mixin to make model serializable."""
__abstract__ = True
def to_dict(self,nested = False, hybrid_attributes = False, exclude = None):
"""Return dict object with model's data.
:param nested: flag to return nested relationships' data if true
:type: bool
:param include_hybrid: flag to include hybrid attributes if true
:return: dict
"""
result = dict()
if exclude is None:
view_cols = self.columns
else :
view_cols = filter(lambda e: e not in exclude, self.columns)
for key in view_cols :
result[key] = getattr(self, key)
if hybrid_attributes:
for key in self.hybrid_properties:
result[key] = getattr(self, key)
if nested:
for key in self.relations:
obj = getattr(self, key)
if isinstance(obj, SerializeMixin):
result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes)
elif isinstance(obj, Iterable):
+ result[key] = [
- result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj]
? ----------- - ^ -
+ o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj
? ^^
+ if isinstance(o, SerializeMixin)
+ ]
return result |
63a539ff4a3a832286136c40a74b1a8b3db1a5c0 | falcom/api/uri/api_querier.py | falcom/api/uri/api_querier.py | from time import sleep
class APIQuerier:
def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0):
self.uri = uri
self.url_opener = url_opener
self.sleep_time = sleep_time
self.max_tries = max_tries
def get (self, **kwargs):
try:
return self.__open_uri(kwargs)
except ConnectionError:
sleep(self.sleep_time)
i = 1
while i != self.max_tries:
i += 1
try:
return self.__open_uri(kwargs)
except ConnectionError:
sleep(self.sleep_time)
return b""
@staticmethod
def utf8 (str_or_bytes):
if isinstance(str_or_bytes, bytes):
return str_or_bytes.decode("utf_8")
else:
return str_or_bytes
def __open_uri (self, kwargs):
with self.url_opener(self.uri(**kwargs)) as response:
result = self.utf8(response.read())
return result
| from time import sleep
class APIQuerier:
def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0):
self.uri = uri
self.url_opener = url_opener
self.sleep_time = sleep_time
self.max_tries = max_tries
def get (self, **kwargs):
class SpecialNull: pass
result = SpecialNull
i = 1
while result is SpecialNull:
try:
result = self.__open_uri(kwargs)
except ConnectionError:
sleep(self.sleep_time)
if i == self.max_tries:
result = b""
else:
i += 1
return result
@staticmethod
def utf8 (str_or_bytes):
if isinstance(str_or_bytes, bytes):
return str_or_bytes.decode("utf_8")
else:
return str_or_bytes
def __open_uri (self, kwargs):
with self.url_opener(self.uri(**kwargs)) as response:
result = self.utf8(response.read())
return result
| Rewrite get() to be less repetitive but still stupid | Rewrite get() to be less repetitive but still stupid
| Python | bsd-3-clause | mlibrary/image-conversion-and-validation,mlibrary/image-conversion-and-validation | from time import sleep
class APIQuerier:
def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0):
self.uri = uri
self.url_opener = url_opener
self.sleep_time = sleep_time
self.max_tries = max_tries
def get (self, **kwargs):
- try:
- return self.__open_uri(kwargs)
+ class SpecialNull: pass
+ result = SpecialNull
+ i = 1
+ while result is SpecialNull:
- except ConnectionError:
- sleep(self.sleep_time)
-
- i = 1
- while i != self.max_tries:
- i += 1
-
try:
- return self.__open_uri(kwargs)
+ result = self.__open_uri(kwargs)
except ConnectionError:
sleep(self.sleep_time)
+ if i == self.max_tries:
+ result = b""
+
+ else:
+ i += 1
+
- return b""
+ return result
@staticmethod
def utf8 (str_or_bytes):
if isinstance(str_or_bytes, bytes):
return str_or_bytes.decode("utf_8")
else:
return str_or_bytes
def __open_uri (self, kwargs):
with self.url_opener(self.uri(**kwargs)) as response:
result = self.utf8(response.read())
return result
| Rewrite get() to be less repetitive but still stupid | ## Code Before:
from time import sleep
class APIQuerier:
def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0):
self.uri = uri
self.url_opener = url_opener
self.sleep_time = sleep_time
self.max_tries = max_tries
def get (self, **kwargs):
try:
return self.__open_uri(kwargs)
except ConnectionError:
sleep(self.sleep_time)
i = 1
while i != self.max_tries:
i += 1
try:
return self.__open_uri(kwargs)
except ConnectionError:
sleep(self.sleep_time)
return b""
@staticmethod
def utf8 (str_or_bytes):
if isinstance(str_or_bytes, bytes):
return str_or_bytes.decode("utf_8")
else:
return str_or_bytes
def __open_uri (self, kwargs):
with self.url_opener(self.uri(**kwargs)) as response:
result = self.utf8(response.read())
return result
## Instruction:
Rewrite get() to be less repetitive but still stupid
## Code After:
from time import sleep
class APIQuerier:
def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0):
self.uri = uri
self.url_opener = url_opener
self.sleep_time = sleep_time
self.max_tries = max_tries
def get (self, **kwargs):
class SpecialNull: pass
result = SpecialNull
i = 1
while result is SpecialNull:
try:
result = self.__open_uri(kwargs)
except ConnectionError:
sleep(self.sleep_time)
if i == self.max_tries:
result = b""
else:
i += 1
return result
@staticmethod
def utf8 (str_or_bytes):
if isinstance(str_or_bytes, bytes):
return str_or_bytes.decode("utf_8")
else:
return str_or_bytes
def __open_uri (self, kwargs):
with self.url_opener(self.uri(**kwargs)) as response:
result = self.utf8(response.read())
return result
| from time import sleep
class APIQuerier:
def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0):
self.uri = uri
self.url_opener = url_opener
self.sleep_time = sleep_time
self.max_tries = max_tries
def get (self, **kwargs):
- try:
- return self.__open_uri(kwargs)
+ class SpecialNull: pass
+ result = SpecialNull
+ i = 1
+ while result is SpecialNull:
- except ConnectionError:
- sleep(self.sleep_time)
-
- i = 1
- while i != self.max_tries:
- i += 1
-
try:
- return self.__open_uri(kwargs)
? ^^^
+ result = self.__open_uri(kwargs)
? +++ ^^
except ConnectionError:
sleep(self.sleep_time)
+ if i == self.max_tries:
+ result = b""
+
+ else:
+ i += 1
+
- return b""
? ^^^
+ return result
? ^^^^^^
@staticmethod
def utf8 (str_or_bytes):
if isinstance(str_or_bytes, bytes):
return str_or_bytes.decode("utf_8")
else:
return str_or_bytes
def __open_uri (self, kwargs):
with self.url_opener(self.uri(**kwargs)) as response:
result = self.utf8(response.read())
return result |
02ac5dcfa6bdaf9b8152ef2f49fd61afe9faf8ab | client/python/plot_request_times.py | client/python/plot_request_times.py | import requests
from plotly.offline import plot
import plotly.graph_objs as go
r = requests.get('http://localhost:8081/monitor_results/1')
print(r.json())
# build traces for plotting from monitoring data
request_times = list()
timestamps = list()
timestamp = 0
url = r.json()[0]['urlToMonitor']['url']
for monitoring_data in r.json():
request_time = monitoring_data['timeNeededForRequest']
request_times.append(request_time)
timestamps.append(timestamp)
timestamp = timestamp + 1
plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name =
'THE OTHER NAME')], filename='request_times.html')
| import requests
from plotly.offline import plot
import plotly.graph_objs as go
def build_data_for_monitored_url(id):
'''Fetches and prepares data for plotting for the given URL id'''
r = requests.get('http://localhost:8081/monitor_results/' + str(id))
# build traces for plotting from monitoring data
request_times = list()
timestamps = list()
timestamp = 0
url = r.json()[0]['urlToMonitor']['url']
for monitoring_data in r.json():
request_time = monitoring_data['timeNeededForRequest']
request_times.append(request_time)
timestamps.append(timestamp)
timestamp = timestamp + 1
return go.Scatter(x = timestamps, y = request_times, name = url)
# get all monitored sites and fetch data for it
r = requests.get('http://localhost:8081/monitored-sites')
plotting_data = list()
for monitored_site in r.json():
print('Fetching data for ' + monitored_site['url'])
data_for_site = build_data_for_monitored_url(monitored_site['id'])
plotting_data.append(data_for_site)
plot(plotting_data, filename='request_times.html')
| Implement fetching all monitored data | Implement fetching all monitored data
| Python | mit | gernd/simple-site-mon | import requests
from plotly.offline import plot
import plotly.graph_objs as go
+ def build_data_for_monitored_url(id):
+ '''Fetches and prepares data for plotting for the given URL id'''
- r = requests.get('http://localhost:8081/monitor_results/1')
+ r = requests.get('http://localhost:8081/monitor_results/' + str(id))
- print(r.json())
- # build traces for plotting from monitoring data
+ # build traces for plotting from monitoring data
- request_times = list()
+ request_times = list()
- timestamps = list()
+ timestamps = list()
- timestamp = 0
+ timestamp = 0
- url = r.json()[0]['urlToMonitor']['url']
+ url = r.json()[0]['urlToMonitor']['url']
- for monitoring_data in r.json():
+ for monitoring_data in r.json():
- request_time = monitoring_data['timeNeededForRequest']
+ request_time = monitoring_data['timeNeededForRequest']
- request_times.append(request_time)
+ request_times.append(request_time)
- timestamps.append(timestamp)
+ timestamps.append(timestamp)
- timestamp = timestamp + 1
+ timestamp = timestamp + 1
+ return go.Scatter(x = timestamps, y = request_times, name = url)
- plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name =
- 'THE OTHER NAME')], filename='request_times.html')
+ # get all monitored sites and fetch data for it
+ r = requests.get('http://localhost:8081/monitored-sites')
+ plotting_data = list()
+ for monitored_site in r.json():
+ print('Fetching data for ' + monitored_site['url'])
+ data_for_site = build_data_for_monitored_url(monitored_site['id'])
+ plotting_data.append(data_for_site)
+ plot(plotting_data, filename='request_times.html')
+ | Implement fetching all monitored data | ## Code Before:
import requests
from plotly.offline import plot
import plotly.graph_objs as go
r = requests.get('http://localhost:8081/monitor_results/1')
print(r.json())
# build traces for plotting from monitoring data
request_times = list()
timestamps = list()
timestamp = 0
url = r.json()[0]['urlToMonitor']['url']
for monitoring_data in r.json():
request_time = monitoring_data['timeNeededForRequest']
request_times.append(request_time)
timestamps.append(timestamp)
timestamp = timestamp + 1
plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name =
'THE OTHER NAME')], filename='request_times.html')
## Instruction:
Implement fetching all monitored data
## Code After:
import requests
from plotly.offline import plot
import plotly.graph_objs as go
def build_data_for_monitored_url(id):
'''Fetches and prepares data for plotting for the given URL id'''
r = requests.get('http://localhost:8081/monitor_results/' + str(id))
# build traces for plotting from monitoring data
request_times = list()
timestamps = list()
timestamp = 0
url = r.json()[0]['urlToMonitor']['url']
for monitoring_data in r.json():
request_time = monitoring_data['timeNeededForRequest']
request_times.append(request_time)
timestamps.append(timestamp)
timestamp = timestamp + 1
return go.Scatter(x = timestamps, y = request_times, name = url)
# get all monitored sites and fetch data for it
r = requests.get('http://localhost:8081/monitored-sites')
plotting_data = list()
for monitored_site in r.json():
print('Fetching data for ' + monitored_site['url'])
data_for_site = build_data_for_monitored_url(monitored_site['id'])
plotting_data.append(data_for_site)
plot(plotting_data, filename='request_times.html')
| import requests
from plotly.offline import plot
import plotly.graph_objs as go
+ def build_data_for_monitored_url(id):
+ '''Fetches and prepares data for plotting for the given URL id'''
- r = requests.get('http://localhost:8081/monitor_results/1')
? -
+ r = requests.get('http://localhost:8081/monitor_results/' + str(id))
? ++++ +++++++++ +
- print(r.json())
- # build traces for plotting from monitoring data
+ # build traces for plotting from monitoring data
? ++++
- request_times = list()
+ request_times = list()
? ++++
- timestamps = list()
+ timestamps = list()
? ++++
- timestamp = 0
+ timestamp = 0
? ++++
- url = r.json()[0]['urlToMonitor']['url']
+ url = r.json()[0]['urlToMonitor']['url']
? ++++
- for monitoring_data in r.json():
+ for monitoring_data in r.json():
? ++++
- request_time = monitoring_data['timeNeededForRequest']
+ request_time = monitoring_data['timeNeededForRequest']
? ++++
- request_times.append(request_time)
+ request_times.append(request_time)
? ++++
- timestamps.append(timestamp)
+ timestamps.append(timestamp)
? ++++
- timestamp = timestamp + 1
+ timestamp = timestamp + 1
? ++++
+ return go.Scatter(x = timestamps, y = request_times, name = url)
- plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name =
- 'THE OTHER NAME')], filename='request_times.html')
+ # get all monitored sites and fetch data for it
+ r = requests.get('http://localhost:8081/monitored-sites')
+ plotting_data = list()
+ for monitored_site in r.json():
+ print('Fetching data for ' + monitored_site['url'])
+ data_for_site = build_data_for_monitored_url(monitored_site['id'])
+ plotting_data.append(data_for_site)
+
+ plot(plotting_data, filename='request_times.html') |
976ca1d7f02a0aab7397a6eb1784436593e6c644 | watchman/management/commands/watchman.py | watchman/management/commands/watchman.py | from __future__ import absolute_import
import json
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from watchman.utils import get_checks
class Command(BaseCommand):
help = 'Runs the default django-watchman checks'
option_list = BaseCommand.option_list + (
make_option(
'-c',
'--checks',
dest='checks',
help='A comma-separated list of watchman checks to run (full python dotted paths)'
),
make_option(
'-s',
'--skips',
dest='skips',
help='A comma-separated list of watchman checks to skip (full python dotted paths)'
),
)
def handle(self, *args, **options):
check_list = None
skip_list = None
verbosity = options['verbosity']
print_all_checks = verbosity == '2' or verbosity == '3'
checks = options['checks']
skips = options['skips']
if checks is not None:
check_list = checks.split(',')
if skips is not None:
skip_list = skips.split(',')
for check in get_checks(check_list=check_list, skip_list=skip_list):
if callable(check):
resp = json.dumps(check())
if '"ok": false' in resp:
raise CommandError(resp)
elif print_all_checks:
self.stdout.write(resp)
| from __future__ import absolute_import
import json
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from watchman.utils import get_checks
class Command(BaseCommand):
help = 'Runs the default django-watchman checks'
option_list = BaseCommand.option_list + (
make_option(
'-c',
'--checks',
dest='checks',
help='A comma-separated list of watchman checks to run (full python dotted paths)'
),
make_option(
'-s',
'--skips',
dest='skips',
help='A comma-separated list of watchman checks to skip (full python dotted paths)'
),
)
def handle(self, *args, **options):
check_list = None
skip_list = None
verbosity = options['verbosity']
print_all_checks = verbosity in ['2', '3', ]
checks = options['checks']
skips = options['skips']
if checks is not None:
check_list = checks.split(',')
if skips is not None:
skip_list = skips.split(',')
for check in get_checks(check_list=check_list, skip_list=skip_list):
if callable(check):
resp = json.dumps(check())
if '"ok": false' in resp:
raise CommandError(resp)
elif print_all_checks:
self.stdout.write(resp)
| Swap equality checks for `in` | Swap equality checks for `in`
| Python | bsd-3-clause | mwarkentin/django-watchman,mwarkentin/django-watchman,JBKahn/django-watchman,JBKahn/django-watchman | from __future__ import absolute_import
import json
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from watchman.utils import get_checks
class Command(BaseCommand):
help = 'Runs the default django-watchman checks'
option_list = BaseCommand.option_list + (
make_option(
'-c',
'--checks',
dest='checks',
help='A comma-separated list of watchman checks to run (full python dotted paths)'
),
make_option(
'-s',
'--skips',
dest='skips',
help='A comma-separated list of watchman checks to skip (full python dotted paths)'
),
)
def handle(self, *args, **options):
check_list = None
skip_list = None
verbosity = options['verbosity']
- print_all_checks = verbosity == '2' or verbosity == '3'
+ print_all_checks = verbosity in ['2', '3', ]
checks = options['checks']
skips = options['skips']
if checks is not None:
check_list = checks.split(',')
if skips is not None:
skip_list = skips.split(',')
for check in get_checks(check_list=check_list, skip_list=skip_list):
if callable(check):
resp = json.dumps(check())
if '"ok": false' in resp:
raise CommandError(resp)
elif print_all_checks:
self.stdout.write(resp)
| Swap equality checks for `in` | ## Code Before:
from __future__ import absolute_import
import json
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from watchman.utils import get_checks
class Command(BaseCommand):
help = 'Runs the default django-watchman checks'
option_list = BaseCommand.option_list + (
make_option(
'-c',
'--checks',
dest='checks',
help='A comma-separated list of watchman checks to run (full python dotted paths)'
),
make_option(
'-s',
'--skips',
dest='skips',
help='A comma-separated list of watchman checks to skip (full python dotted paths)'
),
)
def handle(self, *args, **options):
check_list = None
skip_list = None
verbosity = options['verbosity']
print_all_checks = verbosity == '2' or verbosity == '3'
checks = options['checks']
skips = options['skips']
if checks is not None:
check_list = checks.split(',')
if skips is not None:
skip_list = skips.split(',')
for check in get_checks(check_list=check_list, skip_list=skip_list):
if callable(check):
resp = json.dumps(check())
if '"ok": false' in resp:
raise CommandError(resp)
elif print_all_checks:
self.stdout.write(resp)
## Instruction:
Swap equality checks for `in`
## Code After:
from __future__ import absolute_import
import json
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from watchman.utils import get_checks
class Command(BaseCommand):
help = 'Runs the default django-watchman checks'
option_list = BaseCommand.option_list + (
make_option(
'-c',
'--checks',
dest='checks',
help='A comma-separated list of watchman checks to run (full python dotted paths)'
),
make_option(
'-s',
'--skips',
dest='skips',
help='A comma-separated list of watchman checks to skip (full python dotted paths)'
),
)
def handle(self, *args, **options):
check_list = None
skip_list = None
verbosity = options['verbosity']
print_all_checks = verbosity in ['2', '3', ]
checks = options['checks']
skips = options['skips']
if checks is not None:
check_list = checks.split(',')
if skips is not None:
skip_list = skips.split(',')
for check in get_checks(check_list=check_list, skip_list=skip_list):
if callable(check):
resp = json.dumps(check())
if '"ok": false' in resp:
raise CommandError(resp)
elif print_all_checks:
self.stdout.write(resp)
| from __future__ import absolute_import
import json
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from watchman.utils import get_checks
class Command(BaseCommand):
help = 'Runs the default django-watchman checks'
option_list = BaseCommand.option_list + (
make_option(
'-c',
'--checks',
dest='checks',
help='A comma-separated list of watchman checks to run (full python dotted paths)'
),
make_option(
'-s',
'--skips',
dest='skips',
help='A comma-separated list of watchman checks to skip (full python dotted paths)'
),
)
def handle(self, *args, **options):
check_list = None
skip_list = None
verbosity = options['verbosity']
- print_all_checks = verbosity == '2' or verbosity == '3'
? ^^ ^^^^^^^^^^^^^^^^
+ print_all_checks = verbosity in ['2', '3', ]
? ^^ + ^ +++
checks = options['checks']
skips = options['skips']
if checks is not None:
check_list = checks.split(',')
if skips is not None:
skip_list = skips.split(',')
for check in get_checks(check_list=check_list, skip_list=skip_list):
if callable(check):
resp = json.dumps(check())
if '"ok": false' in resp:
raise CommandError(resp)
elif print_all_checks:
self.stdout.write(resp) |
9c176de1fd280e72dd06c9eaa64060e52abca746 | python/prebuild.py | python/prebuild.py |
def extract_function_names(module):
'''
extract function names from attributes of 'module'.
'''
from importlib import import_module
mod = import_module(module.__name__)
attr_list = dir(mod)
scope = locals()
def iscallable(name):
return eval('callable(mod.{})'.format(name), scope)
return filter(iscallable, attr_list)
def gen_pyplot_functions(dub_root):
'''
generate 'pyplot_functions.txt' for matplotlibd.pyplot.
'''
import matplotlib.pyplot
from string import lowercase
functions = filter(lambda i: i[0] != '_' or i[0] in lowercase,
extract_function_names(matplotlib.pyplot))
with open(dub_root + "/views/pyplot_functions.txt", "w") as f:
f.write("\n".join(functions))
if __name__ == '__main__':
from sys import argv
gen_pyplot_functions(argv[1])
|
def extract_function_names(module):
'''
extract function names from attributes of 'module'.
'''
from importlib import import_module
mod = import_module(module.__name__)
attr_list = dir(mod)
scope = locals()
def iscallable(name):
ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated']
return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators
return filter(iscallable, attr_list)
def gen_pyplot_functions(dub_root):
'''
generate 'pyplot_functions.txt' for matplotlibd.pyplot.
'''
import matplotlib.pyplot
from string import lowercase
functions = filter(lambda i: i[0] != '_' or i[0] in lowercase,
extract_function_names(matplotlib.pyplot))
with open(dub_root + "/views/pyplot_functions.txt", "w") as f:
f.write("\n".join(functions))
if __name__ == '__main__':
from sys import argv
gen_pyplot_functions(argv[1])
| Remove python decorators from list | Remove python decorators from list
| Python | mit | koji-kojiro/matplotlib-d |
def extract_function_names(module):
'''
extract function names from attributes of 'module'.
'''
from importlib import import_module
mod = import_module(module.__name__)
attr_list = dir(mod)
scope = locals()
def iscallable(name):
+ ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated']
- return eval('callable(mod.{})'.format(name), scope)
+ return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators
return filter(iscallable, attr_list)
def gen_pyplot_functions(dub_root):
'''
generate 'pyplot_functions.txt' for matplotlibd.pyplot.
'''
import matplotlib.pyplot
from string import lowercase
functions = filter(lambda i: i[0] != '_' or i[0] in lowercase,
extract_function_names(matplotlib.pyplot))
with open(dub_root + "/views/pyplot_functions.txt", "w") as f:
f.write("\n".join(functions))
if __name__ == '__main__':
from sys import argv
gen_pyplot_functions(argv[1])
| Remove python decorators from list | ## Code Before:
def extract_function_names(module):
'''
extract function names from attributes of 'module'.
'''
from importlib import import_module
mod = import_module(module.__name__)
attr_list = dir(mod)
scope = locals()
def iscallable(name):
return eval('callable(mod.{})'.format(name), scope)
return filter(iscallable, attr_list)
def gen_pyplot_functions(dub_root):
'''
generate 'pyplot_functions.txt' for matplotlibd.pyplot.
'''
import matplotlib.pyplot
from string import lowercase
functions = filter(lambda i: i[0] != '_' or i[0] in lowercase,
extract_function_names(matplotlib.pyplot))
with open(dub_root + "/views/pyplot_functions.txt", "w") as f:
f.write("\n".join(functions))
if __name__ == '__main__':
from sys import argv
gen_pyplot_functions(argv[1])
## Instruction:
Remove python decorators from list
## Code After:
def extract_function_names(module):
'''
extract function names from attributes of 'module'.
'''
from importlib import import_module
mod = import_module(module.__name__)
attr_list = dir(mod)
scope = locals()
def iscallable(name):
ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated']
return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators
return filter(iscallable, attr_list)
def gen_pyplot_functions(dub_root):
'''
generate 'pyplot_functions.txt' for matplotlibd.pyplot.
'''
import matplotlib.pyplot
from string import lowercase
functions = filter(lambda i: i[0] != '_' or i[0] in lowercase,
extract_function_names(matplotlib.pyplot))
with open(dub_root + "/views/pyplot_functions.txt", "w") as f:
f.write("\n".join(functions))
if __name__ == '__main__':
from sys import argv
gen_pyplot_functions(argv[1])
|
def extract_function_names(module):
'''
extract function names from attributes of 'module'.
'''
from importlib import import_module
mod = import_module(module.__name__)
attr_list = dir(mod)
scope = locals()
def iscallable(name):
+ ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated']
- return eval('callable(mod.{})'.format(name), scope)
+ return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators
? ++++++++++++++++++++++++++++++++++
return filter(iscallable, attr_list)
def gen_pyplot_functions(dub_root):
'''
generate 'pyplot_functions.txt' for matplotlibd.pyplot.
'''
import matplotlib.pyplot
from string import lowercase
functions = filter(lambda i: i[0] != '_' or i[0] in lowercase,
extract_function_names(matplotlib.pyplot))
with open(dub_root + "/views/pyplot_functions.txt", "w") as f:
f.write("\n".join(functions))
if __name__ == '__main__':
from sys import argv
gen_pyplot_functions(argv[1]) |
7c0d68b1bce27d026b69e3a069c549ab560b0f3d | spillway/mixins.py | spillway/mixins.py | class FormMixin(object):
"""Mixin to provide form validation and data cleaning of GET or POST
requests.
"""
form_class = None
@property
def form(self):
"""Returns a validated form dict or an empty dict."""
_form = getattr(self, '_form', False)
if not _form:
self._form = self.form_class(self.request.GET or self.request.POST,
self.request.FILES or None)
valid = self._form.is_valid()
return self._form
| class FormMixin(object):
"""Mixin to provide form validation and data cleaning of GET or POST
requests.
"""
form_class = None
@property
def form(self):
"""Returns a validated form dict or an empty dict."""
_form = getattr(self, '_form', False)
if not _form:
self._form = self.form_class(
self.request.QUERY_PARAMS or self.request.DATA,
self.request.FILES or None)
valid = self._form.is_valid()
return self._form
| Use DRF query params and data request attrs | Use DRF query params and data request attrs
| Python | bsd-3-clause | barseghyanartur/django-spillway,kuzmich/django-spillway,bkg/django-spillway | class FormMixin(object):
"""Mixin to provide form validation and data cleaning of GET or POST
requests.
"""
form_class = None
@property
def form(self):
"""Returns a validated form dict or an empty dict."""
_form = getattr(self, '_form', False)
if not _form:
- self._form = self.form_class(self.request.GET or self.request.POST,
+ self._form = self.form_class(
+ self.request.QUERY_PARAMS or self.request.DATA,
- self.request.FILES or None)
+ self.request.FILES or None)
valid = self._form.is_valid()
return self._form
| Use DRF query params and data request attrs | ## Code Before:
class FormMixin(object):
"""Mixin to provide form validation and data cleaning of GET or POST
requests.
"""
form_class = None
@property
def form(self):
"""Returns a validated form dict or an empty dict."""
_form = getattr(self, '_form', False)
if not _form:
self._form = self.form_class(self.request.GET or self.request.POST,
self.request.FILES or None)
valid = self._form.is_valid()
return self._form
## Instruction:
Use DRF query params and data request attrs
## Code After:
class FormMixin(object):
"""Mixin to provide form validation and data cleaning of GET or POST
requests.
"""
form_class = None
@property
def form(self):
"""Returns a validated form dict or an empty dict."""
_form = getattr(self, '_form', False)
if not _form:
self._form = self.form_class(
self.request.QUERY_PARAMS or self.request.DATA,
self.request.FILES or None)
valid = self._form.is_valid()
return self._form
| class FormMixin(object):
"""Mixin to provide form validation and data cleaning of GET or POST
requests.
"""
form_class = None
@property
def form(self):
"""Returns a validated form dict or an empty dict."""
_form = getattr(self, '_form', False)
if not _form:
- self._form = self.form_class(self.request.GET or self.request.POST,
+ self._form = self.form_class(
+ self.request.QUERY_PARAMS or self.request.DATA,
- self.request.FILES or None)
? -------------------------
+ self.request.FILES or None)
valid = self._form.is_valid()
return self._form |
d5c59c018ba7558a9d21370d7eb58ab590779cf1 | plugins/autojoin/plugin_tests/autojoin_test.py | plugins/autojoin/plugin_tests/autojoin_test.py | from tests import base
def setUpModule():
base.enabledPlugins.append('autojoin')
base.startServer()
def tearDownModule():
base.stopServer()
class AutoJoinTest(base.TestCase):
def setUp(self):
base.TestCase.setUp(self)
| from girder.constants import AccessType
from tests import base
import json
def setUpModule():
base.enabledPlugins.append('autojoin')
base.startServer()
def tearDownModule():
base.stopServer()
class AutoJoinTest(base.TestCase):
def setUp(self):
base.TestCase.setUp(self)
self.users = [self.model('user').createUser(
'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num)
for num in [0, 1]]
def testCuration(self):
admin, user = self.users
# create some groups
g1 = self.model('group').createGroup('g1', admin)
g2 = self.model('group').createGroup('g2', admin)
g3 = self.model('group').createGroup('g3', admin)
# set auto join rules
rules = [
{
'pattern': '@kitware.com',
'groupId': str(g1['_id']),
'level': AccessType.ADMIN
},
{
'pattern': '@example.com',
'groupId': str(g2['_id']),
'level': AccessType.READ
},
{
'pattern': '@example.com',
'groupId': str(g3['_id']),
'level': AccessType.WRITE
},
]
params = {
'list': json.dumps([{'key': 'autojoin', 'value': rules}])
}
resp = self.request(
'/system/setting', user=admin, method='PUT', params=params)
self.assertStatusOk(resp)
# create users
user1 = self.model('user').createUser(
'user1', 'password', 'John', 'Doe', 'user1@example.com')
user2 = self.model('user').createUser(
'user2', 'password', 'John', 'Doe', 'user2@kitware.com')
user3 = self.model('user').createUser(
'user3', 'password', 'John', 'Doe', 'user3@kitware.co')
# check correct groups were joined
self.assertEqual(user1['groups'], [g2['_id'], g3['_id']])
self.assertEqual(user2['groups'], [g1['_id']])
self.assertEqual(user3['groups'], [])
# check correct access levels
g1 = self.model('group').load(g1['_id'], force=True)
g2 = self.model('group').load(g2['_id'], force=True)
g3 = self.model('group').load(g3['_id'], force=True)
self.assertTrue(
{u'id': user2['_id'], u'level': AccessType.ADMIN} in
g1['access']['users'])
self.assertTrue(
{u'id': user1['_id'], u'level': AccessType.WRITE} in
g3['access']['users'])
| Add server tests for auto join plugin | Add server tests for auto join plugin
| Python | apache-2.0 | kotfic/girder,kotfic/girder,adsorensen/girder,jbeezley/girder,data-exp-lab/girder,girder/girder,sutartmelson/girder,Kitware/girder,girder/girder,sutartmelson/girder,jbeezley/girder,RafaelPalomar/girder,girder/girder,adsorensen/girder,RafaelPalomar/girder,manthey/girder,manthey/girder,data-exp-lab/girder,RafaelPalomar/girder,Kitware/girder,RafaelPalomar/girder,manthey/girder,Kitware/girder,data-exp-lab/girder,sutartmelson/girder,data-exp-lab/girder,kotfic/girder,girder/girder,jbeezley/girder,manthey/girder,adsorensen/girder,Xarthisius/girder,kotfic/girder,Kitware/girder,adsorensen/girder,data-exp-lab/girder,RafaelPalomar/girder,jbeezley/girder,Xarthisius/girder,sutartmelson/girder,Xarthisius/girder,adsorensen/girder,kotfic/girder,Xarthisius/girder,sutartmelson/girder,Xarthisius/girder | + from girder.constants import AccessType
from tests import base
+ import json
def setUpModule():
base.enabledPlugins.append('autojoin')
base.startServer()
def tearDownModule():
base.stopServer()
class AutoJoinTest(base.TestCase):
def setUp(self):
base.TestCase.setUp(self)
+ self.users = [self.model('user').createUser(
+ 'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num)
+ for num in [0, 1]]
+
+ def testCuration(self):
+ admin, user = self.users
+
+ # create some groups
+ g1 = self.model('group').createGroup('g1', admin)
+ g2 = self.model('group').createGroup('g2', admin)
+ g3 = self.model('group').createGroup('g3', admin)
+
+ # set auto join rules
+ rules = [
+ {
+ 'pattern': '@kitware.com',
+ 'groupId': str(g1['_id']),
+ 'level': AccessType.ADMIN
+ },
+ {
+ 'pattern': '@example.com',
+ 'groupId': str(g2['_id']),
+ 'level': AccessType.READ
+ },
+ {
+ 'pattern': '@example.com',
+ 'groupId': str(g3['_id']),
+ 'level': AccessType.WRITE
+ },
+ ]
+ params = {
+ 'list': json.dumps([{'key': 'autojoin', 'value': rules}])
+ }
+ resp = self.request(
+ '/system/setting', user=admin, method='PUT', params=params)
+ self.assertStatusOk(resp)
+
+ # create users
+ user1 = self.model('user').createUser(
+ 'user1', 'password', 'John', 'Doe', 'user1@example.com')
+ user2 = self.model('user').createUser(
+ 'user2', 'password', 'John', 'Doe', 'user2@kitware.com')
+ user3 = self.model('user').createUser(
+ 'user3', 'password', 'John', 'Doe', 'user3@kitware.co')
+
+ # check correct groups were joined
+ self.assertEqual(user1['groups'], [g2['_id'], g3['_id']])
+ self.assertEqual(user2['groups'], [g1['_id']])
+ self.assertEqual(user3['groups'], [])
+
+ # check correct access levels
+ g1 = self.model('group').load(g1['_id'], force=True)
+ g2 = self.model('group').load(g2['_id'], force=True)
+ g3 = self.model('group').load(g3['_id'], force=True)
+ self.assertTrue(
+ {u'id': user2['_id'], u'level': AccessType.ADMIN} in
+ g1['access']['users'])
+ self.assertTrue(
+ {u'id': user1['_id'], u'level': AccessType.WRITE} in
+ g3['access']['users'])
+ | Add server tests for auto join plugin | ## Code Before:
from tests import base
def setUpModule():
base.enabledPlugins.append('autojoin')
base.startServer()
def tearDownModule():
base.stopServer()
class AutoJoinTest(base.TestCase):
def setUp(self):
base.TestCase.setUp(self)
## Instruction:
Add server tests for auto join plugin
## Code After:
from girder.constants import AccessType
from tests import base
import json
def setUpModule():
base.enabledPlugins.append('autojoin')
base.startServer()
def tearDownModule():
base.stopServer()
class AutoJoinTest(base.TestCase):
def setUp(self):
base.TestCase.setUp(self)
self.users = [self.model('user').createUser(
'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num)
for num in [0, 1]]
def testCuration(self):
admin, user = self.users
# create some groups
g1 = self.model('group').createGroup('g1', admin)
g2 = self.model('group').createGroup('g2', admin)
g3 = self.model('group').createGroup('g3', admin)
# set auto join rules
rules = [
{
'pattern': '@kitware.com',
'groupId': str(g1['_id']),
'level': AccessType.ADMIN
},
{
'pattern': '@example.com',
'groupId': str(g2['_id']),
'level': AccessType.READ
},
{
'pattern': '@example.com',
'groupId': str(g3['_id']),
'level': AccessType.WRITE
},
]
params = {
'list': json.dumps([{'key': 'autojoin', 'value': rules}])
}
resp = self.request(
'/system/setting', user=admin, method='PUT', params=params)
self.assertStatusOk(resp)
# create users
user1 = self.model('user').createUser(
'user1', 'password', 'John', 'Doe', 'user1@example.com')
user2 = self.model('user').createUser(
'user2', 'password', 'John', 'Doe', 'user2@kitware.com')
user3 = self.model('user').createUser(
'user3', 'password', 'John', 'Doe', 'user3@kitware.co')
# check correct groups were joined
self.assertEqual(user1['groups'], [g2['_id'], g3['_id']])
self.assertEqual(user2['groups'], [g1['_id']])
self.assertEqual(user3['groups'], [])
# check correct access levels
g1 = self.model('group').load(g1['_id'], force=True)
g2 = self.model('group').load(g2['_id'], force=True)
g3 = self.model('group').load(g3['_id'], force=True)
self.assertTrue(
{u'id': user2['_id'], u'level': AccessType.ADMIN} in
g1['access']['users'])
self.assertTrue(
{u'id': user1['_id'], u'level': AccessType.WRITE} in
g3['access']['users'])
| + from girder.constants import AccessType
from tests import base
+ import json
def setUpModule():
base.enabledPlugins.append('autojoin')
base.startServer()
def tearDownModule():
base.stopServer()
class AutoJoinTest(base.TestCase):
def setUp(self):
base.TestCase.setUp(self)
+
+ self.users = [self.model('user').createUser(
+ 'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num)
+ for num in [0, 1]]
+
+ def testCuration(self):
+ admin, user = self.users
+
+ # create some groups
+ g1 = self.model('group').createGroup('g1', admin)
+ g2 = self.model('group').createGroup('g2', admin)
+ g3 = self.model('group').createGroup('g3', admin)
+
+ # set auto join rules
+ rules = [
+ {
+ 'pattern': '@kitware.com',
+ 'groupId': str(g1['_id']),
+ 'level': AccessType.ADMIN
+ },
+ {
+ 'pattern': '@example.com',
+ 'groupId': str(g2['_id']),
+ 'level': AccessType.READ
+ },
+ {
+ 'pattern': '@example.com',
+ 'groupId': str(g3['_id']),
+ 'level': AccessType.WRITE
+ },
+ ]
+ params = {
+ 'list': json.dumps([{'key': 'autojoin', 'value': rules}])
+ }
+ resp = self.request(
+ '/system/setting', user=admin, method='PUT', params=params)
+ self.assertStatusOk(resp)
+
+ # create users
+ user1 = self.model('user').createUser(
+ 'user1', 'password', 'John', 'Doe', 'user1@example.com')
+ user2 = self.model('user').createUser(
+ 'user2', 'password', 'John', 'Doe', 'user2@kitware.com')
+ user3 = self.model('user').createUser(
+ 'user3', 'password', 'John', 'Doe', 'user3@kitware.co')
+
+ # check correct groups were joined
+ self.assertEqual(user1['groups'], [g2['_id'], g3['_id']])
+ self.assertEqual(user2['groups'], [g1['_id']])
+ self.assertEqual(user3['groups'], [])
+
+ # check correct access levels
+ g1 = self.model('group').load(g1['_id'], force=True)
+ g2 = self.model('group').load(g2['_id'], force=True)
+ g3 = self.model('group').load(g3['_id'], force=True)
+ self.assertTrue(
+ {u'id': user2['_id'], u'level': AccessType.ADMIN} in
+ g1['access']['users'])
+ self.assertTrue(
+ {u'id': user1['_id'], u'level': AccessType.WRITE} in
+ g3['access']['users']) |
21c7232081483c05752e6db3d60692a04d482d24 | dakota/tests/test_dakota_base.py | dakota/tests/test_dakota_base.py |
import os
import filecmp
from nose.tools import *
from dakota.dakota_base import DakotaBase
# Fixtures -------------------------------------------------------------
def setup_module():
"""Called before any tests are performed."""
print('\n*** DakotaBase tests')
def teardown_module():
"""Called after all tests have completed."""
pass
# Tests ----------------------------------------------------------------
@raises(TypeError)
def test_instantiate():
"""Test whether DakotaBase fails to instantiate."""
d = DakotaBase()
|
from nose.tools import *
from dakota.dakota_base import DakotaBase
# Helpers --------------------------------------------------------------
class Concrete(DakotaBase):
"""A subclass of DakotaBase used for testing."""
def __init__(self):
DakotaBase.__init__(self)
# Fixtures -------------------------------------------------------------
def setup_module():
"""Called before any tests are performed."""
print('\n*** DakotaBase tests')
global c
c = Concrete()
def teardown_module():
"""Called after all tests have completed."""
pass
# Tests ----------------------------------------------------------------
@raises(TypeError)
def test_instantiate():
"""Test whether DakotaBase fails to instantiate."""
d = DakotaBase()
def test_environment_block():
"""Test type of environment_block method results."""
s = c.environment_block()
assert_true(type(s) is str)
def test_method_block():
"""Test type of method_block method results."""
s = c.method_block()
assert_true(type(s) is str)
def test_variables_block():
"""Test type of variables_block method results."""
s = c.variables_block()
assert_true(type(s) is str)
def test_interface_block():
"""Test type of interface_block method results."""
s = c.interface_block()
assert_true(type(s) is str)
def test_responses_block():
"""Test type of responses_block method results."""
s = c.responses_block()
assert_true(type(s) is str)
def test_autogenerate_descriptors():
"""Test autogenerate_descriptors method."""
c.n_variables, c.n_responses = 1, 1
c.autogenerate_descriptors()
assert_true(len(c.variable_descriptors) == 1)
assert_true(len(c.response_descriptors) == 1)
| Add tests for dakota.dakota_base module | Add tests for dakota.dakota_base module
Make a subclass of DakotaBase to use for testing. Add tests for the
"block" sections used to define an input file.
| Python | mit | csdms/dakota,csdms/dakota |
- import os
- import filecmp
from nose.tools import *
from dakota.dakota_base import DakotaBase
+ # Helpers --------------------------------------------------------------
+
+ class Concrete(DakotaBase):
+ """A subclass of DakotaBase used for testing."""
+ def __init__(self):
+ DakotaBase.__init__(self)
# Fixtures -------------------------------------------------------------
def setup_module():
"""Called before any tests are performed."""
print('\n*** DakotaBase tests')
+ global c
+ c = Concrete()
def teardown_module():
"""Called after all tests have completed."""
pass
# Tests ----------------------------------------------------------------
@raises(TypeError)
def test_instantiate():
"""Test whether DakotaBase fails to instantiate."""
d = DakotaBase()
+ def test_environment_block():
+ """Test type of environment_block method results."""
+ s = c.environment_block()
+ assert_true(type(s) is str)
+
+ def test_method_block():
+ """Test type of method_block method results."""
+ s = c.method_block()
+ assert_true(type(s) is str)
+
+ def test_variables_block():
+ """Test type of variables_block method results."""
+ s = c.variables_block()
+ assert_true(type(s) is str)
+
+ def test_interface_block():
+ """Test type of interface_block method results."""
+ s = c.interface_block()
+ assert_true(type(s) is str)
+
+ def test_responses_block():
+ """Test type of responses_block method results."""
+ s = c.responses_block()
+ assert_true(type(s) is str)
+
+ def test_autogenerate_descriptors():
+ """Test autogenerate_descriptors method."""
+ c.n_variables, c.n_responses = 1, 1
+ c.autogenerate_descriptors()
+ assert_true(len(c.variable_descriptors) == 1)
+ assert_true(len(c.response_descriptors) == 1)
+ | Add tests for dakota.dakota_base module | ## Code Before:
import os
import filecmp
from nose.tools import *
from dakota.dakota_base import DakotaBase
# Fixtures -------------------------------------------------------------
def setup_module():
"""Called before any tests are performed."""
print('\n*** DakotaBase tests')
def teardown_module():
"""Called after all tests have completed."""
pass
# Tests ----------------------------------------------------------------
@raises(TypeError)
def test_instantiate():
"""Test whether DakotaBase fails to instantiate."""
d = DakotaBase()
## Instruction:
Add tests for dakota.dakota_base module
## Code After:
from nose.tools import *
from dakota.dakota_base import DakotaBase
# Helpers --------------------------------------------------------------
class Concrete(DakotaBase):
"""A subclass of DakotaBase used for testing."""
def __init__(self):
DakotaBase.__init__(self)
# Fixtures -------------------------------------------------------------
def setup_module():
"""Called before any tests are performed."""
print('\n*** DakotaBase tests')
global c
c = Concrete()
def teardown_module():
"""Called after all tests have completed."""
pass
# Tests ----------------------------------------------------------------
@raises(TypeError)
def test_instantiate():
"""Test whether DakotaBase fails to instantiate."""
d = DakotaBase()
def test_environment_block():
"""Test type of environment_block method results."""
s = c.environment_block()
assert_true(type(s) is str)
def test_method_block():
"""Test type of method_block method results."""
s = c.method_block()
assert_true(type(s) is str)
def test_variables_block():
"""Test type of variables_block method results."""
s = c.variables_block()
assert_true(type(s) is str)
def test_interface_block():
"""Test type of interface_block method results."""
s = c.interface_block()
assert_true(type(s) is str)
def test_responses_block():
"""Test type of responses_block method results."""
s = c.responses_block()
assert_true(type(s) is str)
def test_autogenerate_descriptors():
"""Test autogenerate_descriptors method."""
c.n_variables, c.n_responses = 1, 1
c.autogenerate_descriptors()
assert_true(len(c.variable_descriptors) == 1)
assert_true(len(c.response_descriptors) == 1)
|
- import os
- import filecmp
from nose.tools import *
from dakota.dakota_base import DakotaBase
+ # Helpers --------------------------------------------------------------
+
+ class Concrete(DakotaBase):
+ """A subclass of DakotaBase used for testing."""
+ def __init__(self):
+ DakotaBase.__init__(self)
# Fixtures -------------------------------------------------------------
def setup_module():
"""Called before any tests are performed."""
print('\n*** DakotaBase tests')
+ global c
+ c = Concrete()
def teardown_module():
"""Called after all tests have completed."""
pass
# Tests ----------------------------------------------------------------
@raises(TypeError)
def test_instantiate():
"""Test whether DakotaBase fails to instantiate."""
d = DakotaBase()
+
+ def test_environment_block():
+ """Test type of environment_block method results."""
+ s = c.environment_block()
+ assert_true(type(s) is str)
+
+ def test_method_block():
+ """Test type of method_block method results."""
+ s = c.method_block()
+ assert_true(type(s) is str)
+
+ def test_variables_block():
+ """Test type of variables_block method results."""
+ s = c.variables_block()
+ assert_true(type(s) is str)
+
+ def test_interface_block():
+ """Test type of interface_block method results."""
+ s = c.interface_block()
+ assert_true(type(s) is str)
+
+ def test_responses_block():
+ """Test type of responses_block method results."""
+ s = c.responses_block()
+ assert_true(type(s) is str)
+
+ def test_autogenerate_descriptors():
+ """Test autogenerate_descriptors method."""
+ c.n_variables, c.n_responses = 1, 1
+ c.autogenerate_descriptors()
+ assert_true(len(c.variable_descriptors) == 1)
+ assert_true(len(c.response_descriptors) == 1) |
53646da453a4aa6d0e559ee3069626458f2fef78 | common/urls.py | common/urls.py | import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
return re.sub("\\(.+\\)", "{id}", pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True) | import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
_pattern = pattern.replace('^', '').replace('$', '')
return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
| Fix backend home page url generator | Fix backend home page url generator
| Python | mit | DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange | import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
+ _pattern = pattern.replace('^', '').replace('$', '')
- return re.sub("\\(.+\\)", "{id}", pattern)
+ return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
+ | Fix backend home page url generator | ## Code Before:
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
return re.sub("\\(.+\\)", "{id}", pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
## Instruction:
Fix backend home page url generator
## Code After:
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
_pattern = pattern.replace('^', '').replace('$', '')
return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
| import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
+ _pattern = pattern.replace('^', '').replace('$', '')
- return re.sub("\\(.+\\)", "{id}", pattern)
+ return re.sub("\\(.+\\)", "{id}", _pattern)
? +
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True) |
26b1845419cd7ea55cf1f40f26812eb9c124299f | kolibri/core/content/signals.py | kolibri/core/content/signals.py | from django.db.models import F
from django.db.models.signals import pre_delete
from django.dispatch import receiver
from .models import ChannelMetadata
from .models import ContentNode
from kolibri.core.notifications.models import LearnerProgressNotification
@receiver(pre_delete, sender=ContentNode)
def cascade_delete_node(sender, instance=None, *args, **kwargs):
"""
For a given node, we delete all notifications
objects whose contentnode is the instance's node..
"""
LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete()
@receiver(pre_delete, sender=ChannelMetadata)
def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs):
"""
For a given channel, decrement the order of all channels that come after this channel.
"""
ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1)
| from django.db.models import F
from django.db.models.signals import pre_delete
from django.dispatch import receiver
from .models import ChannelMetadata
from .models import ContentNode
from kolibri.core.notifications.models import LearnerProgressNotification
from kolibri.core.lessons.models import Lesson
@receiver(pre_delete, sender=ContentNode)
def cascade_delete_node(sender, instance=None, *args, **kwargs):
"""
For a given node, we delete all notifications
objects whose contentnode is the instance's node..
"""
LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete()
@receiver(pre_delete, sender=ChannelMetadata)
def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs):
"""
For a given channel, decrement the order of all channels that come after this channel.
"""
ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1)
@receiver(pre_delete, sender=ChannelMetadata)
def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs):
# Update the resources array of all lessons to ensure they don't have
# any deleted content
lessons = Lesson.objects.all()
for lesson in lessons:
updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id]
if len(updated_resources) < len(lesson.resources):
lesson.resources = updated_resources
lesson.save()
| Add Channel deletion side-effect that updates affected lessons | Add Channel deletion side-effect that updates affected lessons
| Python | mit | learningequality/kolibri,lyw07/kolibri,mrpau/kolibri,learningequality/kolibri,indirectlylit/kolibri,lyw07/kolibri,mrpau/kolibri,mrpau/kolibri,indirectlylit/kolibri,indirectlylit/kolibri,learningequality/kolibri,indirectlylit/kolibri,lyw07/kolibri,learningequality/kolibri,lyw07/kolibri,mrpau/kolibri | from django.db.models import F
from django.db.models.signals import pre_delete
from django.dispatch import receiver
from .models import ChannelMetadata
from .models import ContentNode
from kolibri.core.notifications.models import LearnerProgressNotification
+ from kolibri.core.lessons.models import Lesson
@receiver(pre_delete, sender=ContentNode)
def cascade_delete_node(sender, instance=None, *args, **kwargs):
"""
For a given node, we delete all notifications
objects whose contentnode is the instance's node..
"""
LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete()
@receiver(pre_delete, sender=ChannelMetadata)
def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs):
"""
For a given channel, decrement the order of all channels that come after this channel.
"""
ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1)
+
+ @receiver(pre_delete, sender=ChannelMetadata)
+ def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs):
+ # Update the resources array of all lessons to ensure they don't have
+ # any deleted content
+ lessons = Lesson.objects.all()
+ for lesson in lessons:
+ updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id]
+ if len(updated_resources) < len(lesson.resources):
+ lesson.resources = updated_resources
+ lesson.save()
+ | Add Channel deletion side-effect that updates affected lessons | ## Code Before:
from django.db.models import F
from django.db.models.signals import pre_delete
from django.dispatch import receiver
from .models import ChannelMetadata
from .models import ContentNode
from kolibri.core.notifications.models import LearnerProgressNotification
@receiver(pre_delete, sender=ContentNode)
def cascade_delete_node(sender, instance=None, *args, **kwargs):
"""
For a given node, we delete all notifications
objects whose contentnode is the instance's node..
"""
LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete()
@receiver(pre_delete, sender=ChannelMetadata)
def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs):
"""
For a given channel, decrement the order of all channels that come after this channel.
"""
ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1)
## Instruction:
Add Channel deletion side-effect that updates affected lessons
## Code After:
from django.db.models import F
from django.db.models.signals import pre_delete
from django.dispatch import receiver
from .models import ChannelMetadata
from .models import ContentNode
from kolibri.core.notifications.models import LearnerProgressNotification
from kolibri.core.lessons.models import Lesson
@receiver(pre_delete, sender=ContentNode)
def cascade_delete_node(sender, instance=None, *args, **kwargs):
"""
For a given node, we delete all notifications
objects whose contentnode is the instance's node..
"""
LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete()
@receiver(pre_delete, sender=ChannelMetadata)
def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs):
"""
For a given channel, decrement the order of all channels that come after this channel.
"""
ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1)
@receiver(pre_delete, sender=ChannelMetadata)
def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs):
# Update the resources array of all lessons to ensure they don't have
# any deleted content
lessons = Lesson.objects.all()
for lesson in lessons:
updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id]
if len(updated_resources) < len(lesson.resources):
lesson.resources = updated_resources
lesson.save()
| from django.db.models import F
from django.db.models.signals import pre_delete
from django.dispatch import receiver
from .models import ChannelMetadata
from .models import ContentNode
from kolibri.core.notifications.models import LearnerProgressNotification
+ from kolibri.core.lessons.models import Lesson
@receiver(pre_delete, sender=ContentNode)
def cascade_delete_node(sender, instance=None, *args, **kwargs):
"""
For a given node, we delete all notifications
objects whose contentnode is the instance's node..
"""
LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete()
@receiver(pre_delete, sender=ChannelMetadata)
def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs):
"""
For a given channel, decrement the order of all channels that come after this channel.
"""
ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1)
+
+
+ @receiver(pre_delete, sender=ChannelMetadata)
+ def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs):
+ # Update the resources array of all lessons to ensure they don't have
+ # any deleted content
+ lessons = Lesson.objects.all()
+ for lesson in lessons:
+ updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id]
+ if len(updated_resources) < len(lesson.resources):
+ lesson.resources = updated_resources
+ lesson.save() |
1a871cf3bf1fd40342e490599361d57017cdcc65 | backend/breach/tests/test_strategy.py | backend/breach/tests/test_strategy.py | from mock import patch
from breach.tests.base import RuptureTestCase
from breach.strategy import Strategy
class StrategyTestCase(RuptureTestCase):
@patch('breach.strategy.Sniffer')
def test_first_round(self, Sniffer):
strategy0 = Strategy(self.victim)
work0 = strategy0.get_work()
self.assertEqual(
work0['url'],
'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^'
)
self.assertTrue('amount' in work0)
self.assertTrue('timeout' in work0)
strategy1 = Strategy(self.victim)
work1 = strategy1.get_work()
self.assertEqual(
work1['url'],
'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^'
)
def test_same_round_same_batch(self):
pass
def test_same_round_different_batch(self):
pass
def test_advance_round(self):
pass
| from mock import patch
from breach.tests.base import RuptureTestCase
from breach.strategy import Strategy
class StrategyTestCase(RuptureTestCase):
@patch('breach.strategy.Sniffer')
def test_first_round(self, Sniffer):
strategy0 = Strategy(self.victim)
work0 = strategy0.get_work()
self.assertEqual(
work0['url'],
'https://di.uoa.gr/?breach=^testsecret0^1^'
)
self.assertTrue('amount' in work0)
self.assertTrue('timeout' in work0)
strategy1 = Strategy(self.victim)
work1 = strategy1.get_work()
self.assertEqual(
work1['url'],
'https://di.uoa.gr/?breach=^testsecret1^0^'
)
def test_same_round_same_batch(self):
pass
def test_same_round_different_batch(self):
pass
def test_advance_round(self):
pass
| Update first round test, create huffman based on knownalphabet | Update first round test, create huffman based on knownalphabet
| Python | mit | dionyziz/rupture,dimriou/rupture,esarafianou/rupture,esarafianou/rupture,dimkarakostas/rupture,dimkarakostas/rupture,esarafianou/rupture,dionyziz/rupture,dimkarakostas/rupture,dionyziz/rupture,esarafianou/rupture,dimkarakostas/rupture,dimkarakostas/rupture,dimriou/rupture,dimriou/rupture,dimriou/rupture,dimriou/rupture,dionyziz/rupture,dionyziz/rupture | from mock import patch
from breach.tests.base import RuptureTestCase
from breach.strategy import Strategy
class StrategyTestCase(RuptureTestCase):
@patch('breach.strategy.Sniffer')
def test_first_round(self, Sniffer):
strategy0 = Strategy(self.victim)
work0 = strategy0.get_work()
self.assertEqual(
work0['url'],
- 'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^'
+ 'https://di.uoa.gr/?breach=^testsecret0^1^'
)
self.assertTrue('amount' in work0)
self.assertTrue('timeout' in work0)
strategy1 = Strategy(self.victim)
work1 = strategy1.get_work()
self.assertEqual(
work1['url'],
- 'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^'
+ 'https://di.uoa.gr/?breach=^testsecret1^0^'
)
def test_same_round_same_batch(self):
pass
def test_same_round_different_batch(self):
pass
def test_advance_round(self):
pass
| Update first round test, create huffman based on knownalphabet | ## Code Before:
from mock import patch
from breach.tests.base import RuptureTestCase
from breach.strategy import Strategy
class StrategyTestCase(RuptureTestCase):
@patch('breach.strategy.Sniffer')
def test_first_round(self, Sniffer):
strategy0 = Strategy(self.victim)
work0 = strategy0.get_work()
self.assertEqual(
work0['url'],
'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^'
)
self.assertTrue('amount' in work0)
self.assertTrue('timeout' in work0)
strategy1 = Strategy(self.victim)
work1 = strategy1.get_work()
self.assertEqual(
work1['url'],
'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^'
)
def test_same_round_same_batch(self):
pass
def test_same_round_different_batch(self):
pass
def test_advance_round(self):
pass
## Instruction:
Update first round test, create huffman based on knownalphabet
## Code After:
from mock import patch
from breach.tests.base import RuptureTestCase
from breach.strategy import Strategy
class StrategyTestCase(RuptureTestCase):
@patch('breach.strategy.Sniffer')
def test_first_round(self, Sniffer):
strategy0 = Strategy(self.victim)
work0 = strategy0.get_work()
self.assertEqual(
work0['url'],
'https://di.uoa.gr/?breach=^testsecret0^1^'
)
self.assertTrue('amount' in work0)
self.assertTrue('timeout' in work0)
strategy1 = Strategy(self.victim)
work1 = strategy1.get_work()
self.assertEqual(
work1['url'],
'https://di.uoa.gr/?breach=^testsecret1^0^'
)
def test_same_round_same_batch(self):
pass
def test_same_round_different_batch(self):
pass
def test_advance_round(self):
pass
| from mock import patch
from breach.tests.base import RuptureTestCase
from breach.strategy import Strategy
class StrategyTestCase(RuptureTestCase):
@patch('breach.strategy.Sniffer')
def test_first_round(self, Sniffer):
strategy0 = Strategy(self.victim)
work0 = strategy0.get_work()
self.assertEqual(
work0['url'],
- 'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^'
? ----------------
+ 'https://di.uoa.gr/?breach=^testsecret0^1^'
)
self.assertTrue('amount' in work0)
self.assertTrue('timeout' in work0)
strategy1 = Strategy(self.victim)
work1 = strategy1.get_work()
self.assertEqual(
work1['url'],
- 'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^'
? ----------------
+ 'https://di.uoa.gr/?breach=^testsecret1^0^'
)
def test_same_round_same_batch(self):
pass
def test_same_round_different_batch(self):
pass
def test_advance_round(self):
pass |
bea258e2affc165f610de83248d9f958eec1ef4e | cmsplugin_markdown/models.py | cmsplugin_markdown/models.py | from django.db import models
from cms.models import CMSPlugin
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
| from django.db import models
from cms.models import CMSPlugin
from cms.utils.compat.dj import python_2_unicode_compatible
@python_2_unicode_compatible
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
def __str__(self):
text = self.markdown_text
return (text[:50] + '...') if len(text) > 53 else text
| Add __str__ method for better representation in frontend | Add __str__ method for better representation in frontend
| Python | mit | bitmazk/cmsplugin-markdown,bitmazk/cmsplugin-markdown,bitmazk/cmsplugin-markdown | from django.db import models
from cms.models import CMSPlugin
+ from cms.utils.compat.dj import python_2_unicode_compatible
+ @python_2_unicode_compatible
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
+ def __str__(self):
+ text = self.markdown_text
+ return (text[:50] + '...') if len(text) > 53 else text
+ | Add __str__ method for better representation in frontend | ## Code Before:
from django.db import models
from cms.models import CMSPlugin
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
## Instruction:
Add __str__ method for better representation in frontend
## Code After:
from django.db import models
from cms.models import CMSPlugin
from cms.utils.compat.dj import python_2_unicode_compatible
@python_2_unicode_compatible
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
def __str__(self):
text = self.markdown_text
return (text[:50] + '...') if len(text) > 53 else text
| from django.db import models
from cms.models import CMSPlugin
+ from cms.utils.compat.dj import python_2_unicode_compatible
+ @python_2_unicode_compatible
class MarkdownPlugin(CMSPlugin):
markdown_text = models.TextField(max_length=8000)
+
+ def __str__(self):
+ text = self.markdown_text
+ return (text[:50] + '...') if len(text) > 53 else text |
208760340d3314f666d7e6437817cc96e0e16194 | organizer/urls/tag.py | organizer/urls/tag.py | from django.conf.urls import url
from ..views import (
TagCreate, TagDelete, TagDetail, TagList,
TagUpdate)
urlpatterns = [
url(r'^$',
TagList.as_view(),
name='organizer_tag_list'),
url(r'^create/$',
TagCreate.as_view(),
name='organizer_tag_create'),
url(r'^(?P<slug>[\w\-]+)/$',
TagDetail.as_view(),
name='organizer_tag_detail'),
url(r'^(?P<slug>[\w-]+)/delete/$',
TagDelete.as_view(),
name='organizer_tag_delete'),
url(r'^(?P<slug>[\w\-]+)/update/$',
TagUpdate.as_view(),
name='organizer_tag_update'),
]
| from django.conf.urls import url
from django.contrib.auth.decorators import \
login_required
from ..views import (
TagCreate, TagDelete, TagDetail, TagList,
TagUpdate)
urlpatterns = [
url(r'^$',
TagList.as_view(),
name='organizer_tag_list'),
url(r'^create/$',
login_required(
TagCreate.as_view()),
name='organizer_tag_create'),
url(r'^(?P<slug>[\w\-]+)/$',
TagDetail.as_view(),
name='organizer_tag_detail'),
url(r'^(?P<slug>[\w-]+)/delete/$',
TagDelete.as_view(),
name='organizer_tag_delete'),
url(r'^(?P<slug>[\w\-]+)/update/$',
TagUpdate.as_view(),
name='organizer_tag_update'),
]
| Use login_required decorator in URL pattern. | Ch20: Use login_required decorator in URL pattern.
| Python | bsd-2-clause | jambonrose/DjangoUnleashed-1.8,jambonrose/DjangoUnleashed-1.8 | from django.conf.urls import url
+ from django.contrib.auth.decorators import \
+ login_required
from ..views import (
TagCreate, TagDelete, TagDetail, TagList,
TagUpdate)
urlpatterns = [
url(r'^$',
TagList.as_view(),
name='organizer_tag_list'),
url(r'^create/$',
+ login_required(
- TagCreate.as_view(),
+ TagCreate.as_view()),
name='organizer_tag_create'),
url(r'^(?P<slug>[\w\-]+)/$',
TagDetail.as_view(),
name='organizer_tag_detail'),
url(r'^(?P<slug>[\w-]+)/delete/$',
TagDelete.as_view(),
name='organizer_tag_delete'),
url(r'^(?P<slug>[\w\-]+)/update/$',
TagUpdate.as_view(),
name='organizer_tag_update'),
]
| Use login_required decorator in URL pattern. | ## Code Before:
from django.conf.urls import url
from ..views import (
TagCreate, TagDelete, TagDetail, TagList,
TagUpdate)
urlpatterns = [
url(r'^$',
TagList.as_view(),
name='organizer_tag_list'),
url(r'^create/$',
TagCreate.as_view(),
name='organizer_tag_create'),
url(r'^(?P<slug>[\w\-]+)/$',
TagDetail.as_view(),
name='organizer_tag_detail'),
url(r'^(?P<slug>[\w-]+)/delete/$',
TagDelete.as_view(),
name='organizer_tag_delete'),
url(r'^(?P<slug>[\w\-]+)/update/$',
TagUpdate.as_view(),
name='organizer_tag_update'),
]
## Instruction:
Use login_required decorator in URL pattern.
## Code After:
from django.conf.urls import url
from django.contrib.auth.decorators import \
login_required
from ..views import (
TagCreate, TagDelete, TagDetail, TagList,
TagUpdate)
urlpatterns = [
url(r'^$',
TagList.as_view(),
name='organizer_tag_list'),
url(r'^create/$',
login_required(
TagCreate.as_view()),
name='organizer_tag_create'),
url(r'^(?P<slug>[\w\-]+)/$',
TagDetail.as_view(),
name='organizer_tag_detail'),
url(r'^(?P<slug>[\w-]+)/delete/$',
TagDelete.as_view(),
name='organizer_tag_delete'),
url(r'^(?P<slug>[\w\-]+)/update/$',
TagUpdate.as_view(),
name='organizer_tag_update'),
]
| from django.conf.urls import url
+ from django.contrib.auth.decorators import \
+ login_required
from ..views import (
TagCreate, TagDelete, TagDetail, TagList,
TagUpdate)
urlpatterns = [
url(r'^$',
TagList.as_view(),
name='organizer_tag_list'),
url(r'^create/$',
+ login_required(
- TagCreate.as_view(),
+ TagCreate.as_view()),
? ++++ +
name='organizer_tag_create'),
url(r'^(?P<slug>[\w\-]+)/$',
TagDetail.as_view(),
name='organizer_tag_detail'),
url(r'^(?P<slug>[\w-]+)/delete/$',
TagDelete.as_view(),
name='organizer_tag_delete'),
url(r'^(?P<slug>[\w\-]+)/update/$',
TagUpdate.as_view(),
name='organizer_tag_update'),
] |
7e0024878352ba544a8b40d2c8b5741aedf05a70 | Code/login_proxy.py | Code/login_proxy.py |
import sys
import datetime
from colors import farben
def request(flow):
now = datetime.datetime.now()
content = flow.request.get_text()
host = flow.request.pretty_host
method = flow.request.method
if method == "POST" and ("pass" in content) or ("password" in content) :
with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
myfile.write(farben.AUF + str(now) +" // " + farben.END)
myfile.write(farben.LD + host + farben.END)
myfile.write("\n")
myfile.write(farben.IN + content + farben.END)
myfile.write("\n")
myfile.write("\n")
|
import re
import sys
import datetime
from colors import farben
def request(flow):
now = datetime.datetime.now()
content = flow.request.get_text()
host = flow.request.pretty_host
method = flow.request.method
if method == "POST" and ("pass" in content) or ("password" in content):
with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
myfile.write(farben.AUF + str(now) + " // " + farben.END)
myfile.write(farben.LD + host + farben.END)
myfile.write("\n")
passwords = re.findall(r"(?:pass|password)=([^&]*)", content)
myfile.write(farben.IN + passwords[0] + farben.END)
myfile.write("\n")
myfile.write("\n")
| Add regex for parsing password url parameter | Add regex for parsing password url parameter
| Python | apache-2.0 | sarah314/SpyPi |
+ import re
import sys
import datetime
from colors import farben
+
def request(flow):
now = datetime.datetime.now()
content = flow.request.get_text()
host = flow.request.pretty_host
method = flow.request.method
- if method == "POST" and ("pass" in content) or ("password" in content) :
+ if method == "POST" and ("pass" in content) or ("password" in content):
- with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
+ with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
- myfile.write(farben.AUF + str(now) +" // " + farben.END)
+ myfile.write(farben.AUF + str(now) + " // " + farben.END)
myfile.write(farben.LD + host + farben.END)
myfile.write("\n")
+ passwords = re.findall(r"(?:pass|password)=([^&]*)", content)
- myfile.write(farben.IN + content + farben.END)
+ myfile.write(farben.IN + passwords[0] + farben.END)
myfile.write("\n")
myfile.write("\n")
| Add regex for parsing password url parameter | ## Code Before:
import sys
import datetime
from colors import farben
def request(flow):
now = datetime.datetime.now()
content = flow.request.get_text()
host = flow.request.pretty_host
method = flow.request.method
if method == "POST" and ("pass" in content) or ("password" in content) :
with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
myfile.write(farben.AUF + str(now) +" // " + farben.END)
myfile.write(farben.LD + host + farben.END)
myfile.write("\n")
myfile.write(farben.IN + content + farben.END)
myfile.write("\n")
myfile.write("\n")
## Instruction:
Add regex for parsing password url parameter
## Code After:
import re
import sys
import datetime
from colors import farben
def request(flow):
now = datetime.datetime.now()
content = flow.request.get_text()
host = flow.request.pretty_host
method = flow.request.method
if method == "POST" and ("pass" in content) or ("password" in content):
with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
myfile.write(farben.AUF + str(now) + " // " + farben.END)
myfile.write(farben.LD + host + farben.END)
myfile.write("\n")
passwords = re.findall(r"(?:pass|password)=([^&]*)", content)
myfile.write(farben.IN + passwords[0] + farben.END)
myfile.write("\n")
myfile.write("\n")
|
+ import re
import sys
import datetime
from colors import farben
+
def request(flow):
now = datetime.datetime.now()
content = flow.request.get_text()
host = flow.request.pretty_host
method = flow.request.method
- if method == "POST" and ("pass" in content) or ("password" in content) :
? -
+ if method == "POST" and ("pass" in content) or ("password" in content):
- with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
? -
+ with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile:
- myfile.write(farben.AUF + str(now) +" // " + farben.END)
+ myfile.write(farben.AUF + str(now) + " // " + farben.END)
? +
myfile.write(farben.LD + host + farben.END)
myfile.write("\n")
+ passwords = re.findall(r"(?:pass|password)=([^&]*)", content)
- myfile.write(farben.IN + content + farben.END)
? ^ ^^^^^
+ myfile.write(farben.IN + passwords[0] + farben.END)
? ^^^^^ ^^^^^^
myfile.write("\n")
myfile.write("\n") |
1a0339b85d852526c184eeace73021fc7d68b2c6 | python_dispatcher.py | python_dispatcher.py | import traceback
from routes import Mapper
import ppp_core
import example_ppp_module as flower
import ppp_questionparsing_grammatical as qp_grammatical
import ppp_cas
#import ppp_nlp_ml_standalone
class Application:
def __init__(self):
self.mapper = Mapper()
self.mapper.connect('core', '/core/', app=ppp_core.app)
self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app)
self.mapper.connect('flower', '/flower/', app=flower.app)
self.mapper.connect('cas', '/cas/', app=ppp_cas.app)
self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app)
#self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app)
def __call__(self, environ, start_response):
match = self.mapper.routematch(environ=environ)
app = match[0]['app'] if match else self.not_found
try:
return app(environ, start_response)
except KeyboardInterrupt:
raise
except Exception as e:
traceback.print_exc(e)
def not_found(self, environ, start_response):
headers = [('Content-Type', 'text/plain')]
start_response('404 Not Found', headers)
return [b'Not found.']
app = Application()
| import traceback
from routes import Mapper
import ppp_core
import example_ppp_module as flower
import ppp_questionparsing_grammatical as qp_grammatical
import ppp_cas
import ppp_spell_checker
#import ppp_nlp_ml_standalone
class Application:
def __init__(self):
self.mapper = Mapper()
self.mapper.connect('core', '/core/', app=ppp_core.app)
self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app)
self.mapper.connect('flower', '/flower/', app=flower.app)
self.mapper.connect('cas', '/cas/', app=ppp_cas.app)
self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app)
#self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app)
def __call__(self, environ, start_response):
match = self.mapper.routematch(environ=environ)
app = match[0]['app'] if match else self.not_found
try:
return app(environ, start_response)
except KeyboardInterrupt:
raise
except Exception as e:
traceback.print_exc(e)
def not_found(self, environ, start_response):
headers = [('Content-Type', 'text/plain')]
start_response('404 Not Found', headers)
return [b'Not found.']
app = Application()
| Fix name of spell checker. | Fix name of spell checker.
| Python | cc0-1.0 | ProjetPP/Deployment,ProjetPP/Deployment,ProjetPP/Deployment | import traceback
from routes import Mapper
import ppp_core
import example_ppp_module as flower
import ppp_questionparsing_grammatical as qp_grammatical
import ppp_cas
+ import ppp_spell_checker
#import ppp_nlp_ml_standalone
class Application:
def __init__(self):
self.mapper = Mapper()
self.mapper.connect('core', '/core/', app=ppp_core.app)
self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app)
self.mapper.connect('flower', '/flower/', app=flower.app)
self.mapper.connect('cas', '/cas/', app=ppp_cas.app)
- self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app)
+ self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app)
#self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app)
def __call__(self, environ, start_response):
match = self.mapper.routematch(environ=environ)
app = match[0]['app'] if match else self.not_found
try:
return app(environ, start_response)
except KeyboardInterrupt:
raise
except Exception as e:
traceback.print_exc(e)
def not_found(self, environ, start_response):
headers = [('Content-Type', 'text/plain')]
start_response('404 Not Found', headers)
return [b'Not found.']
app = Application()
| Fix name of spell checker. | ## Code Before:
import traceback
from routes import Mapper
import ppp_core
import example_ppp_module as flower
import ppp_questionparsing_grammatical as qp_grammatical
import ppp_cas
#import ppp_nlp_ml_standalone
class Application:
def __init__(self):
self.mapper = Mapper()
self.mapper.connect('core', '/core/', app=ppp_core.app)
self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app)
self.mapper.connect('flower', '/flower/', app=flower.app)
self.mapper.connect('cas', '/cas/', app=ppp_cas.app)
self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app)
#self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app)
def __call__(self, environ, start_response):
match = self.mapper.routematch(environ=environ)
app = match[0]['app'] if match else self.not_found
try:
return app(environ, start_response)
except KeyboardInterrupt:
raise
except Exception as e:
traceback.print_exc(e)
def not_found(self, environ, start_response):
headers = [('Content-Type', 'text/plain')]
start_response('404 Not Found', headers)
return [b'Not found.']
app = Application()
## Instruction:
Fix name of spell checker.
## Code After:
import traceback
from routes import Mapper
import ppp_core
import example_ppp_module as flower
import ppp_questionparsing_grammatical as qp_grammatical
import ppp_cas
import ppp_spell_checker
#import ppp_nlp_ml_standalone
class Application:
def __init__(self):
self.mapper = Mapper()
self.mapper.connect('core', '/core/', app=ppp_core.app)
self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app)
self.mapper.connect('flower', '/flower/', app=flower.app)
self.mapper.connect('cas', '/cas/', app=ppp_cas.app)
self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app)
#self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app)
def __call__(self, environ, start_response):
match = self.mapper.routematch(environ=environ)
app = match[0]['app'] if match else self.not_found
try:
return app(environ, start_response)
except KeyboardInterrupt:
raise
except Exception as e:
traceback.print_exc(e)
def not_found(self, environ, start_response):
headers = [('Content-Type', 'text/plain')]
start_response('404 Not Found', headers)
return [b'Not found.']
app = Application()
| import traceback
from routes import Mapper
import ppp_core
import example_ppp_module as flower
import ppp_questionparsing_grammatical as qp_grammatical
import ppp_cas
+ import ppp_spell_checker
#import ppp_nlp_ml_standalone
class Application:
def __init__(self):
self.mapper = Mapper()
self.mapper.connect('core', '/core/', app=ppp_core.app)
self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app)
self.mapper.connect('flower', '/flower/', app=flower.app)
self.mapper.connect('cas', '/cas/', app=ppp_cas.app)
- self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app)
? ^^
+ self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app)
? ++++++ ^^^^^^
#self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app)
def __call__(self, environ, start_response):
match = self.mapper.routematch(environ=environ)
app = match[0]['app'] if match else self.not_found
try:
return app(environ, start_response)
except KeyboardInterrupt:
raise
except Exception as e:
traceback.print_exc(e)
def not_found(self, environ, start_response):
headers = [('Content-Type', 'text/plain')]
start_response('404 Not Found', headers)
return [b'Not found.']
app = Application() |
1bf6211f2fd5aef99e529fdc0e714b1a36ace346 | gallery/util.py | gallery/util.py | import os
from addict import Dict
from gallery.models import File
def get_dir_file_contents(dir_id):
print(File.query.filter(File.parent == dir_id).all())
return File.query.filter(File.parent == dir_id).all()
def get_dir_tree_dict():
path = os.path.normpath("/gallery-data/root")
file_tree = Dict()
for root, _, files in os.walk(path, topdown=True):
path = root.split('/')
path.pop(0)
file_tree_fd = file_tree
for part in path:
file_tree_fd = file_tree_fd[part]
file_tree_fd['.'] = files
return file_tree
def convert_bytes_to_utf8(dic):
for key in dic:
if isinstance(key, bytes):
k = key.decode('utf-8')
v = dic[key]
del dic[key]
dic[k] = v
if isinstance(dic[key], bytes):
v = dic[key].decode('utf-8')
dic[key] = v
return dic
def allowed_file(filename):
return '.' in filename and filename.lower().rsplit('.', 1)[1] in \
[
'txt',
'png',
'jpg',
'jpeg',
'mpg',
'mp4',
'avi'
]
| import os
from addict import Dict
from gallery.models import File
def get_dir_file_contents(dir_id):
print(File.query.filter(File.parent == dir_id).all())
return File.query.filter(File.parent == dir_id).all()
def get_dir_tree_dict():
path = os.path.normpath("/gallery-data/root")
file_tree = Dict()
for root, _, files in os.walk(path, topdown=True):
path = root.split('/')
path.pop(0)
file_tree_fd = file_tree
for part in path:
file_tree_fd = file_tree_fd[part]
file_tree_fd['.'] = files
return file_tree
def convert_bytes_to_utf8(dic):
for key in dic:
if isinstance(key, bytes):
k = key.decode('utf-8')
v = dic[key]
del dic[key]
dic[k] = v
if isinstance(dic[key], bytes):
v = dic[key].decode('utf-8')
dic[key] = v
return dic
def allowed_file(filename):
return '.' in filename and filename.lower().rsplit('.', 1)[1] in \
[
'txt',
'png',
'jpg',
'jpeg',
'mpg',
'mp4',
'avi',
'cr2'
]
| Add CR2 to allowed files | Add CR2 to allowed files
| Python | mit | liam-middlebrook/gallery,liam-middlebrook/gallery,liam-middlebrook/gallery,liam-middlebrook/gallery | import os
from addict import Dict
from gallery.models import File
def get_dir_file_contents(dir_id):
print(File.query.filter(File.parent == dir_id).all())
return File.query.filter(File.parent == dir_id).all()
def get_dir_tree_dict():
path = os.path.normpath("/gallery-data/root")
file_tree = Dict()
for root, _, files in os.walk(path, topdown=True):
path = root.split('/')
path.pop(0)
file_tree_fd = file_tree
for part in path:
file_tree_fd = file_tree_fd[part]
file_tree_fd['.'] = files
return file_tree
def convert_bytes_to_utf8(dic):
for key in dic:
if isinstance(key, bytes):
k = key.decode('utf-8')
v = dic[key]
del dic[key]
dic[k] = v
if isinstance(dic[key], bytes):
v = dic[key].decode('utf-8')
dic[key] = v
return dic
def allowed_file(filename):
return '.' in filename and filename.lower().rsplit('.', 1)[1] in \
[
'txt',
'png',
'jpg',
'jpeg',
'mpg',
'mp4',
- 'avi'
+ 'avi',
+ 'cr2'
]
| Add CR2 to allowed files | ## Code Before:
import os
from addict import Dict
from gallery.models import File
def get_dir_file_contents(dir_id):
print(File.query.filter(File.parent == dir_id).all())
return File.query.filter(File.parent == dir_id).all()
def get_dir_tree_dict():
path = os.path.normpath("/gallery-data/root")
file_tree = Dict()
for root, _, files in os.walk(path, topdown=True):
path = root.split('/')
path.pop(0)
file_tree_fd = file_tree
for part in path:
file_tree_fd = file_tree_fd[part]
file_tree_fd['.'] = files
return file_tree
def convert_bytes_to_utf8(dic):
for key in dic:
if isinstance(key, bytes):
k = key.decode('utf-8')
v = dic[key]
del dic[key]
dic[k] = v
if isinstance(dic[key], bytes):
v = dic[key].decode('utf-8')
dic[key] = v
return dic
def allowed_file(filename):
return '.' in filename and filename.lower().rsplit('.', 1)[1] in \
[
'txt',
'png',
'jpg',
'jpeg',
'mpg',
'mp4',
'avi'
]
## Instruction:
Add CR2 to allowed files
## Code After:
import os
from addict import Dict
from gallery.models import File
def get_dir_file_contents(dir_id):
print(File.query.filter(File.parent == dir_id).all())
return File.query.filter(File.parent == dir_id).all()
def get_dir_tree_dict():
path = os.path.normpath("/gallery-data/root")
file_tree = Dict()
for root, _, files in os.walk(path, topdown=True):
path = root.split('/')
path.pop(0)
file_tree_fd = file_tree
for part in path:
file_tree_fd = file_tree_fd[part]
file_tree_fd['.'] = files
return file_tree
def convert_bytes_to_utf8(dic):
for key in dic:
if isinstance(key, bytes):
k = key.decode('utf-8')
v = dic[key]
del dic[key]
dic[k] = v
if isinstance(dic[key], bytes):
v = dic[key].decode('utf-8')
dic[key] = v
return dic
def allowed_file(filename):
return '.' in filename and filename.lower().rsplit('.', 1)[1] in \
[
'txt',
'png',
'jpg',
'jpeg',
'mpg',
'mp4',
'avi',
'cr2'
]
| import os
from addict import Dict
from gallery.models import File
def get_dir_file_contents(dir_id):
print(File.query.filter(File.parent == dir_id).all())
return File.query.filter(File.parent == dir_id).all()
def get_dir_tree_dict():
path = os.path.normpath("/gallery-data/root")
file_tree = Dict()
for root, _, files in os.walk(path, topdown=True):
path = root.split('/')
path.pop(0)
file_tree_fd = file_tree
for part in path:
file_tree_fd = file_tree_fd[part]
file_tree_fd['.'] = files
return file_tree
def convert_bytes_to_utf8(dic):
for key in dic:
if isinstance(key, bytes):
k = key.decode('utf-8')
v = dic[key]
del dic[key]
dic[k] = v
if isinstance(dic[key], bytes):
v = dic[key].decode('utf-8')
dic[key] = v
return dic
def allowed_file(filename):
return '.' in filename and filename.lower().rsplit('.', 1)[1] in \
[
'txt',
'png',
'jpg',
'jpeg',
'mpg',
'mp4',
- 'avi'
+ 'avi',
? +
+ 'cr2'
] |
1c60cf7082672335279d5b96e83f3cb2eb57424f | purchase_supplier_minimum_order/models/__init__.py | purchase_supplier_minimum_order/models/__init__.py |
from . import (
res_partner,
)
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
|
from . import (
res_partner,
purchase,
)
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
| Enforce minimum PO value for supplier. | Enforce minimum PO value for supplier.
| Python | agpl-3.0 | OpusVL/odoo-purchase-min-order |
from . import (
res_partner,
+ purchase,
)
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
| Enforce minimum PO value for supplier. | ## Code Before:
from . import (
res_partner,
)
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
## Instruction:
Enforce minimum PO value for supplier.
## Code After:
from . import (
res_partner,
purchase,
)
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
|
from . import (
res_partner,
+ purchase,
)
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: |
a04a8c7d8e1087df39025d6e798d83438ac35f77 | setup.py | setup.py |
from distutils.core import setup
setup(name='hpswitch',
version='0.1',
description="A library for interacting with HP Networking switches",
packages=['hpswitch', ],
url='https://github.com/leonhandreke/hpswitch',
license="MIT License",
)
|
from distutils.core import setup
setup(name='hpswitch',
version='0.1',
description="A library for interacting with HP Networking switches",
packages=['hpswitch', ],
url='https://github.com/leonhandreke/hpswitch',
license="MIT License",
requires=['pysnmp']
)
| Add pysnmp as a dependency | Add pysnmp as a dependency
| Python | mit | leonhandreke/hpswitch,thechristschn/hpswitch |
from distutils.core import setup
setup(name='hpswitch',
version='0.1',
description="A library for interacting with HP Networking switches",
packages=['hpswitch', ],
url='https://github.com/leonhandreke/hpswitch',
license="MIT License",
+ requires=['pysnmp']
)
| Add pysnmp as a dependency | ## Code Before:
from distutils.core import setup
setup(name='hpswitch',
version='0.1',
description="A library for interacting with HP Networking switches",
packages=['hpswitch', ],
url='https://github.com/leonhandreke/hpswitch',
license="MIT License",
)
## Instruction:
Add pysnmp as a dependency
## Code After:
from distutils.core import setup
setup(name='hpswitch',
version='0.1',
description="A library for interacting with HP Networking switches",
packages=['hpswitch', ],
url='https://github.com/leonhandreke/hpswitch',
license="MIT License",
requires=['pysnmp']
)
|
from distutils.core import setup
setup(name='hpswitch',
version='0.1',
description="A library for interacting with HP Networking switches",
packages=['hpswitch', ],
url='https://github.com/leonhandreke/hpswitch',
license="MIT License",
+ requires=['pysnmp']
) |
e183578b6211d7311d62100ad643cbaf8408de99 | tests/__init__.py | tests/__init__.py | import unittest.mock
def _test_module_init(module, main_name="main"):
with unittest.mock.patch.object(module, main_name, return_value=0):
with unittest.mock.patch.object(module, "__name__", "__main__"):
with unittest.mock.patch.object(module.sys, "exit") as exit:
module.module_init()
return exit.call_args[0][0] == 0
| import unittest.mock
def _test_module_init(module, main_name="main"):
with unittest.mock.patch.object(
module, main_name, return_value=0
), unittest.mock.patch.object(
module, "__name__", "__main__"
), unittest.mock.patch.object(
module.sys, "exit"
) as exit:
module.module_init()
return exit.call_args[0][0] == 0
| Use multiple context managers on one with statement (thanks Anna) | Use multiple context managers on one with statement (thanks Anna)
| Python | mpl-2.0 | rfinnie/2ping,rfinnie/2ping | import unittest.mock
def _test_module_init(module, main_name="main"):
- with unittest.mock.patch.object(module, main_name, return_value=0):
- with unittest.mock.patch.object(module, "__name__", "__main__"):
- with unittest.mock.patch.object(module.sys, "exit") as exit:
+ with unittest.mock.patch.object(
+ module, main_name, return_value=0
+ ), unittest.mock.patch.object(
+ module, "__name__", "__main__"
+ ), unittest.mock.patch.object(
+ module.sys, "exit"
+ ) as exit:
- module.module_init()
+ module.module_init()
- return exit.call_args[0][0] == 0
+ return exit.call_args[0][0] == 0
| Use multiple context managers on one with statement (thanks Anna) | ## Code Before:
import unittest.mock
def _test_module_init(module, main_name="main"):
with unittest.mock.patch.object(module, main_name, return_value=0):
with unittest.mock.patch.object(module, "__name__", "__main__"):
with unittest.mock.patch.object(module.sys, "exit") as exit:
module.module_init()
return exit.call_args[0][0] == 0
## Instruction:
Use multiple context managers on one with statement (thanks Anna)
## Code After:
import unittest.mock
def _test_module_init(module, main_name="main"):
with unittest.mock.patch.object(
module, main_name, return_value=0
), unittest.mock.patch.object(
module, "__name__", "__main__"
), unittest.mock.patch.object(
module.sys, "exit"
) as exit:
module.module_init()
return exit.call_args[0][0] == 0
| import unittest.mock
def _test_module_init(module, main_name="main"):
- with unittest.mock.patch.object(module, main_name, return_value=0):
- with unittest.mock.patch.object(module, "__name__", "__main__"):
- with unittest.mock.patch.object(module.sys, "exit") as exit:
+ with unittest.mock.patch.object(
+ module, main_name, return_value=0
+ ), unittest.mock.patch.object(
+ module, "__name__", "__main__"
+ ), unittest.mock.patch.object(
+ module.sys, "exit"
+ ) as exit:
- module.module_init()
? --------
+ module.module_init()
- return exit.call_args[0][0] == 0
? --------
+ return exit.call_args[0][0] == 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.