commit
stringlengths
40
40
old_file
stringlengths
4
106
new_file
stringlengths
4
106
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
2.95k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.31k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
diff
stringlengths
49
3.61k
4522de348aab4cc99904b0bc210c223b2477b4b7
tests/config.py
tests/config.py
import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx'))
import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
Use pyt file instead of stand-alone tbx for testing.
Use pyt file instead of stand-alone tbx for testing.
Python
mpl-2.0
EsriOceans/btm
import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) - tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx')) + pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
Use pyt file instead of stand-alone tbx for testing.
## Code Before: import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx')) ## Instruction: Use pyt file instead of stand-alone tbx for testing. ## Code After: import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) - tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx')) ? -- ------ -- + pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt')) ? ++ ++
07d2ffe3c14a6c908a7bf138f40ba8d49bf7b2c3
examples/plot_grow.py
examples/plot_grow.py
# Code source: Óscar Nájera # License: BSD 3 clause import numpy as np import matplotlib.pyplot as plt x = np.linspace(-1, 2, 100) y = np.exp(x) plt.plot(x, y) plt.xlabel('$x$') plt.ylabel('$exp(x)$') plt.show()
# Code source: Óscar Nájera # License: BSD 3 clause import numpy as np import matplotlib.pyplot as plt x = np.linspace(-1, 2, 100) y = np.exp(x) plt.figure() plt.plot(x, y) plt.xlabel('$x$') plt.ylabel('$\exp(x)$') plt.figure() plt.plot(x, -np.exp(-x)) plt.xlabel('$x$') plt.ylabel('$-\exp(-x)$') plt.show()
Update example for image stacking CSS instuction
Update example for image stacking CSS instuction
Python
bsd-3-clause
lesteve/sphinx-gallery,Eric89GXL/sphinx-gallery,sphinx-gallery/sphinx-gallery,Titan-C/sphinx-gallery,lesteve/sphinx-gallery,Titan-C/sphinx-gallery,Eric89GXL/sphinx-gallery,sphinx-gallery/sphinx-gallery
# Code source: Óscar Nájera # License: BSD 3 clause import numpy as np import matplotlib.pyplot as plt x = np.linspace(-1, 2, 100) y = np.exp(x) + plt.figure() plt.plot(x, y) plt.xlabel('$x$') - plt.ylabel('$exp(x)$') + plt.ylabel('$\exp(x)$') + + plt.figure() + plt.plot(x, -np.exp(-x)) + plt.xlabel('$x$') + plt.ylabel('$-\exp(-x)$') + + plt.show()
Update example for image stacking CSS instuction
## Code Before: # Code source: Óscar Nájera # License: BSD 3 clause import numpy as np import matplotlib.pyplot as plt x = np.linspace(-1, 2, 100) y = np.exp(x) plt.plot(x, y) plt.xlabel('$x$') plt.ylabel('$exp(x)$') plt.show() ## Instruction: Update example for image stacking CSS instuction ## Code After: # Code source: Óscar Nájera # License: BSD 3 clause import numpy as np import matplotlib.pyplot as plt x = np.linspace(-1, 2, 100) y = np.exp(x) plt.figure() plt.plot(x, y) plt.xlabel('$x$') plt.ylabel('$\exp(x)$') plt.figure() plt.plot(x, -np.exp(-x)) plt.xlabel('$x$') plt.ylabel('$-\exp(-x)$') plt.show()
# Code source: Óscar Nájera # License: BSD 3 clause import numpy as np import matplotlib.pyplot as plt x = np.linspace(-1, 2, 100) y = np.exp(x) + plt.figure() plt.plot(x, y) plt.xlabel('$x$') - plt.ylabel('$exp(x)$') + plt.ylabel('$\exp(x)$') ? + + + plt.figure() + plt.plot(x, -np.exp(-x)) + plt.xlabel('$x$') + plt.ylabel('$-\exp(-x)$') + + plt.show()
6b49f7b1948ab94631c79304c91f8d5590d03e40
addons/project/models/project_config_settings.py
addons/project/models/project_config_settings.py
from odoo import api, fields, models class ProjectConfiguration(models.TransientModel): _name = 'project.config.settings' _inherit = 'res.config.settings' company_id = fields.Many2one('res.company', string='Company', required=True, default=lambda self: self.env.user.company_id) module_pad = fields.Boolean("Collaborative Pads") module_hr_timesheet = fields.Boolean("Timesheets") module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet") module_rating_project = fields.Boolean(string="Rating on Tasks") module_project_forecast = fields.Boolean(string="Forecasts") module_hr_holidays = fields.Boolean("Leave Management") module_hr_timesheet_attendance = fields.Boolean("Attendances") module_sale_timesheet = fields.Boolean("Time Billing") module_hr_expense = fields.Boolean("Expenses") group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project")
from odoo import api, fields, models class ProjectConfiguration(models.TransientModel): _name = 'project.config.settings' _inherit = 'res.config.settings' company_id = fields.Many2one('res.company', string='Company', required=True, default=lambda self: self.env.user.company_id) module_pad = fields.Boolean("Collaborative Pads") module_hr_timesheet = fields.Boolean("Timesheets") module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet") module_rating_project = fields.Boolean(string="Rating on Tasks") module_project_forecast = fields.Boolean(string="Forecasts") module_hr_holidays = fields.Boolean("Leave Management") module_hr_timesheet_attendance = fields.Boolean("Attendances") module_sale_timesheet = fields.Boolean("Time Billing") module_hr_expense = fields.Boolean("Expenses") group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project") @api.onchange('module_sale_timesheet') def _onchange_module_sale_timesheet(self): if self.module_sale_timesheet: self.module_hr_timesheet = True
Enable `Timesheets` option if `Time Billing` is enabled
[IMP] project: Enable `Timesheets` option if `Time Billing` is enabled
Python
agpl-3.0
ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo
from odoo import api, fields, models class ProjectConfiguration(models.TransientModel): _name = 'project.config.settings' _inherit = 'res.config.settings' company_id = fields.Many2one('res.company', string='Company', required=True, default=lambda self: self.env.user.company_id) module_pad = fields.Boolean("Collaborative Pads") module_hr_timesheet = fields.Boolean("Timesheets") module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet") module_rating_project = fields.Boolean(string="Rating on Tasks") module_project_forecast = fields.Boolean(string="Forecasts") module_hr_holidays = fields.Boolean("Leave Management") module_hr_timesheet_attendance = fields.Boolean("Attendances") module_sale_timesheet = fields.Boolean("Time Billing") module_hr_expense = fields.Boolean("Expenses") group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project") + @api.onchange('module_sale_timesheet') + def _onchange_module_sale_timesheet(self): + if self.module_sale_timesheet: + self.module_hr_timesheet = True +
Enable `Timesheets` option if `Time Billing` is enabled
## Code Before: from odoo import api, fields, models class ProjectConfiguration(models.TransientModel): _name = 'project.config.settings' _inherit = 'res.config.settings' company_id = fields.Many2one('res.company', string='Company', required=True, default=lambda self: self.env.user.company_id) module_pad = fields.Boolean("Collaborative Pads") module_hr_timesheet = fields.Boolean("Timesheets") module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet") module_rating_project = fields.Boolean(string="Rating on Tasks") module_project_forecast = fields.Boolean(string="Forecasts") module_hr_holidays = fields.Boolean("Leave Management") module_hr_timesheet_attendance = fields.Boolean("Attendances") module_sale_timesheet = fields.Boolean("Time Billing") module_hr_expense = fields.Boolean("Expenses") group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project") ## Instruction: Enable `Timesheets` option if `Time Billing` is enabled ## Code After: from odoo import api, fields, models class ProjectConfiguration(models.TransientModel): _name = 'project.config.settings' _inherit = 'res.config.settings' company_id = fields.Many2one('res.company', string='Company', required=True, default=lambda self: self.env.user.company_id) module_pad = fields.Boolean("Collaborative Pads") module_hr_timesheet = fields.Boolean("Timesheets") module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet") module_rating_project = fields.Boolean(string="Rating on Tasks") module_project_forecast = fields.Boolean(string="Forecasts") module_hr_holidays = fields.Boolean("Leave Management") module_hr_timesheet_attendance = fields.Boolean("Attendances") module_sale_timesheet = fields.Boolean("Time Billing") module_hr_expense = fields.Boolean("Expenses") group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project") @api.onchange('module_sale_timesheet') def _onchange_module_sale_timesheet(self): if self.module_sale_timesheet: self.module_hr_timesheet = True
from odoo import api, fields, models class ProjectConfiguration(models.TransientModel): _name = 'project.config.settings' _inherit = 'res.config.settings' company_id = fields.Many2one('res.company', string='Company', required=True, default=lambda self: self.env.user.company_id) module_pad = fields.Boolean("Collaborative Pads") module_hr_timesheet = fields.Boolean("Timesheets") module_project_timesheet_synchro = fields.Boolean("Awesome Timesheet") module_rating_project = fields.Boolean(string="Rating on Tasks") module_project_forecast = fields.Boolean(string="Forecasts") module_hr_holidays = fields.Boolean("Leave Management") module_hr_timesheet_attendance = fields.Boolean("Attendances") module_sale_timesheet = fields.Boolean("Time Billing") module_hr_expense = fields.Boolean("Expenses") group_subtask_project = fields.Boolean("Sub-tasks", implied_group="project.group_subtask_project") + + @api.onchange('module_sale_timesheet') + def _onchange_module_sale_timesheet(self): + if self.module_sale_timesheet: + self.module_hr_timesheet = True
9e22082a280babb1e0880fe24fa17c45aac09515
docker-nodev.py
docker-nodev.py
from __future__ import print_function import subprocess import sys DOCKER_CREATE_IN = 'docker create -it nodev {}' DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}' def nodev(argv=()): container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip() print('creating container: {container_id}'.format(**locals())) try: subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True) subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True) finally: print('removing container: {container_id}'.format(**locals())) subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True) if __name__ == '__main__': nodev(sys.argv)
from __future__ import print_function import subprocess import sys DOCKER_CREATE_IN = 'docker create -it nodev {}' DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}' def nodev(argv=()): container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip() print('creating container: {container_id}'.format(**locals())) try: subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True) subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True) finally: print('removing container: {container_id}'.format(**locals())) subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True) if __name__ == '__main__': try: nodev(sys.argv) except subprocess.CalledProcessError as ex: print(ex.args) sys.exit(1)
Fix python3 crash and cleaner error reporting.
Fix python3 crash and cleaner error reporting.
Python
mit
nodev-io/nodev-starter-kit,nodev-io/nodev-tutorial,nodev-io/nodev-starter-kit
from __future__ import print_function import subprocess import sys DOCKER_CREATE_IN = 'docker create -it nodev {}' DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}' def nodev(argv=()): - container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip() + container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip() print('creating container: {container_id}'.format(**locals())) try: subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True) subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True) finally: print('removing container: {container_id}'.format(**locals())) subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True) if __name__ == '__main__': + try: - nodev(sys.argv) + nodev(sys.argv) + except subprocess.CalledProcessError as ex: + print(ex.args) + sys.exit(1)
Fix python3 crash and cleaner error reporting.
## Code Before: from __future__ import print_function import subprocess import sys DOCKER_CREATE_IN = 'docker create -it nodev {}' DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}' def nodev(argv=()): container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip() print('creating container: {container_id}'.format(**locals())) try: subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True) subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True) finally: print('removing container: {container_id}'.format(**locals())) subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True) if __name__ == '__main__': nodev(sys.argv) ## Instruction: Fix python3 crash and cleaner error reporting. ## Code After: from __future__ import print_function import subprocess import sys DOCKER_CREATE_IN = 'docker create -it nodev {}' DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}' def nodev(argv=()): container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip() print('creating container: {container_id}'.format(**locals())) try: subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True) subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True) finally: print('removing container: {container_id}'.format(**locals())) subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True) if __name__ == '__main__': try: nodev(sys.argv) except subprocess.CalledProcessError as ex: print(ex.args) sys.exit(1)
from __future__ import print_function import subprocess import sys DOCKER_CREATE_IN = 'docker create -it nodev {}' DOCKER_SIMPLE_CMD_IN = 'docker {} {container_id}' def nodev(argv=()): - container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).strip() + container_id = subprocess.check_output(DOCKER_CREATE_IN.format(' '.join(argv)), shell=True).decode('utf-8').strip() ? ++++++++++++++++ print('creating container: {container_id}'.format(**locals())) try: subprocess.check_call('docker cp . {container_id}:/src '.format(**locals()), shell=True) subprocess.check_call('docker start -ai {container_id}'.format(**locals()), shell=True) finally: print('removing container: {container_id}'.format(**locals())) subprocess.check_output(DOCKER_SIMPLE_CMD_IN.format('rm -f', **locals()), shell=True) if __name__ == '__main__': + try: - nodev(sys.argv) + nodev(sys.argv) ? ++++ + except subprocess.CalledProcessError as ex: + print(ex.args) + sys.exit(1)
db4355ce0345df9dd23b937370f5f0d4cb2164e9
zc_common/remote_resource/filters.py
zc_common/remote_resource/filters.py
import re from django.db.models.fields.related import ManyToManyField from rest_framework import filters class JSONAPIFilterBackend(filters.DjangoFilterBackend): def filter_queryset(self, request, queryset, view): filter_class = self.get_filter_class(view, queryset) primary_key = queryset.model._meta.pk.name query_params = {} for param, value in request.query_params.iteritems(): match = re.search(r'^filter\[(\w+)\]$', param) if match: field_name = match.group(1) try: name, extra = field_name.split('__') except ValueError: name = field_name extra = None if name not in view.filter_fields.keys(): return queryset.none() if len(field_name) > 1 and field_name[:2] == 'id': query_params['{0}__{1}'.format(primary_key, extra)] = value if hasattr(queryset.model, field_name)\ and isinstance(getattr(queryset.model, field_name).field, ManyToManyField): value = value.split(',') query_params[field_name] = value if filter_class: return filter_class(query_params, queryset=queryset).qs return queryset
import re from distutils.util import strtobool from django.db.models import BooleanField, FieldDoesNotExist from django.db.models.fields.related import ManyToManyField from rest_framework import filters class JSONAPIFilterBackend(filters.DjangoFilterBackend): def filter_queryset(self, request, queryset, view): filter_class = self.get_filter_class(view, queryset) primary_key = queryset.model._meta.pk.name query_params = {} for param, value in request.query_params.iteritems(): match = re.search(r'^filter\[(\w+)\]$', param) if match: field_name = match.group(1) try: name, extra = field_name.split('__') except ValueError: name = field_name extra = None if name not in view.filter_fields.keys(): return queryset.none() if len(field_name) > 1 and field_name[:2] == 'id': query_params['{0}__{1}'.format(primary_key, extra)] = value if hasattr(queryset.model, field_name)\ and isinstance(getattr(queryset.model, field_name).field, ManyToManyField): value = value.split(',') # Allow 'true' or 'false' as values for boolean fields try: if isinstance(queryset.model._meta.get_field(field_name), BooleanField): value = bool(strtobool(value)) except FieldDoesNotExist: pass query_params[field_name] = value if filter_class: return filter_class(query_params, queryset=queryset).qs return queryset
Use 'true' while filtering a boolean as opposed to 'True'
Use 'true' while filtering a boolean as opposed to 'True'
Python
mit
ZeroCater/zc_common,ZeroCater/zc_common
import re + from distutils.util import strtobool + from django.db.models import BooleanField, FieldDoesNotExist from django.db.models.fields.related import ManyToManyField from rest_framework import filters class JSONAPIFilterBackend(filters.DjangoFilterBackend): def filter_queryset(self, request, queryset, view): filter_class = self.get_filter_class(view, queryset) primary_key = queryset.model._meta.pk.name query_params = {} for param, value in request.query_params.iteritems(): match = re.search(r'^filter\[(\w+)\]$', param) if match: field_name = match.group(1) try: name, extra = field_name.split('__') except ValueError: name = field_name extra = None if name not in view.filter_fields.keys(): return queryset.none() if len(field_name) > 1 and field_name[:2] == 'id': query_params['{0}__{1}'.format(primary_key, extra)] = value if hasattr(queryset.model, field_name)\ and isinstance(getattr(queryset.model, field_name).field, ManyToManyField): value = value.split(',') + + # Allow 'true' or 'false' as values for boolean fields + try: + if isinstance(queryset.model._meta.get_field(field_name), BooleanField): + value = bool(strtobool(value)) + except FieldDoesNotExist: + pass + query_params[field_name] = value if filter_class: return filter_class(query_params, queryset=queryset).qs return queryset
Use 'true' while filtering a boolean as opposed to 'True'
## Code Before: import re from django.db.models.fields.related import ManyToManyField from rest_framework import filters class JSONAPIFilterBackend(filters.DjangoFilterBackend): def filter_queryset(self, request, queryset, view): filter_class = self.get_filter_class(view, queryset) primary_key = queryset.model._meta.pk.name query_params = {} for param, value in request.query_params.iteritems(): match = re.search(r'^filter\[(\w+)\]$', param) if match: field_name = match.group(1) try: name, extra = field_name.split('__') except ValueError: name = field_name extra = None if name not in view.filter_fields.keys(): return queryset.none() if len(field_name) > 1 and field_name[:2] == 'id': query_params['{0}__{1}'.format(primary_key, extra)] = value if hasattr(queryset.model, field_name)\ and isinstance(getattr(queryset.model, field_name).field, ManyToManyField): value = value.split(',') query_params[field_name] = value if filter_class: return filter_class(query_params, queryset=queryset).qs return queryset ## Instruction: Use 'true' while filtering a boolean as opposed to 'True' ## Code After: import re from distutils.util import strtobool from django.db.models import BooleanField, FieldDoesNotExist from django.db.models.fields.related import ManyToManyField from rest_framework import filters class JSONAPIFilterBackend(filters.DjangoFilterBackend): def filter_queryset(self, request, queryset, view): filter_class = self.get_filter_class(view, queryset) primary_key = queryset.model._meta.pk.name query_params = {} for param, value in request.query_params.iteritems(): match = re.search(r'^filter\[(\w+)\]$', param) if match: field_name = match.group(1) try: name, extra = field_name.split('__') except ValueError: name = field_name extra = None if name not in view.filter_fields.keys(): return queryset.none() if len(field_name) > 1 and field_name[:2] == 'id': query_params['{0}__{1}'.format(primary_key, extra)] = value if hasattr(queryset.model, field_name)\ and isinstance(getattr(queryset.model, field_name).field, ManyToManyField): value = value.split(',') # Allow 'true' or 'false' as values for boolean fields try: if isinstance(queryset.model._meta.get_field(field_name), BooleanField): value = bool(strtobool(value)) except FieldDoesNotExist: pass query_params[field_name] = value if filter_class: return filter_class(query_params, queryset=queryset).qs return queryset
import re + from distutils.util import strtobool + from django.db.models import BooleanField, FieldDoesNotExist from django.db.models.fields.related import ManyToManyField from rest_framework import filters class JSONAPIFilterBackend(filters.DjangoFilterBackend): def filter_queryset(self, request, queryset, view): filter_class = self.get_filter_class(view, queryset) primary_key = queryset.model._meta.pk.name query_params = {} for param, value in request.query_params.iteritems(): match = re.search(r'^filter\[(\w+)\]$', param) if match: field_name = match.group(1) try: name, extra = field_name.split('__') except ValueError: name = field_name extra = None if name not in view.filter_fields.keys(): return queryset.none() if len(field_name) > 1 and field_name[:2] == 'id': query_params['{0}__{1}'.format(primary_key, extra)] = value if hasattr(queryset.model, field_name)\ and isinstance(getattr(queryset.model, field_name).field, ManyToManyField): value = value.split(',') + + # Allow 'true' or 'false' as values for boolean fields + try: + if isinstance(queryset.model._meta.get_field(field_name), BooleanField): + value = bool(strtobool(value)) + except FieldDoesNotExist: + pass + query_params[field_name] = value if filter_class: return filter_class(query_params, queryset=queryset).qs return queryset
0e99654d606038098d45fb83cc40405742e43ae8
readthedocs/builds/filters.py
readthedocs/builds/filters.py
from django.utils.translation import ugettext_lazy as _ import django_filters from builds import constants from builds.models import Build, Version ANY_REPO = ( ('', _('Any')), ) BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES class VersionFilter(django_filters.FilterSet): project = django_filters.CharFilter(name='project__name', lookup_type="icontains") slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains') class Meta: model = Version fields = ['project', 'slug'] class BuildFilter(django_filters.FilterSet): date = django_filters.DateRangeFilter(label=_("Build Date"), name="date") type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES) class Meta: model = Build fields = ['type', 'date', 'version', 'success']
from django.utils.translation import ugettext_lazy as _ import django_filters from builds import constants from builds.models import Build, Version ANY_REPO = ( ('', _('Any')), ) BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES class VersionFilter(django_filters.FilterSet): project = django_filters.CharFilter(name='project__name', lookup_type="icontains") slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains') class Meta: model = Version fields = ['project', 'slug'] class BuildFilter(django_filters.FilterSet): date = django_filters.DateRangeFilter(label=_("Build Date"), name="date") type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES) class Meta: model = Build fields = ['type', 'date', 'success']
Remove version from Build filter.
Remove version from Build filter.
Python
mit
agjohnson/readthedocs.org,fujita-shintaro/readthedocs.org,GovReady/readthedocs.org,nyergler/pythonslides,Tazer/readthedocs.org,techtonik/readthedocs.org,takluyver/readthedocs.org,nyergler/pythonslides,GovReady/readthedocs.org,nikolas/readthedocs.org,gjtorikian/readthedocs.org,cgourlay/readthedocs.org,d0ugal/readthedocs.org,sid-kap/readthedocs.org,CedarLogic/readthedocs.org,gjtorikian/readthedocs.org,sils1297/readthedocs.org,singingwolfboy/readthedocs.org,kdkeyser/readthedocs.org,kenshinthebattosai/readthedocs.org,VishvajitP/readthedocs.org,hach-que/readthedocs.org,laplaceliu/readthedocs.org,KamranMackey/readthedocs.org,techtonik/readthedocs.org,kdkeyser/readthedocs.org,ojii/readthedocs.org,asampat3090/readthedocs.org,michaelmcandrew/readthedocs.org,dirn/readthedocs.org,attakei/readthedocs-oauth,LukasBoersma/readthedocs.org,rtfd/readthedocs.org,techtonik/readthedocs.org,tddv/readthedocs.org,sunnyzwh/readthedocs.org,emawind84/readthedocs.org,davidfischer/readthedocs.org,jerel/readthedocs.org,clarkperkins/readthedocs.org,mhils/readthedocs.org,takluyver/readthedocs.org,sils1297/readthedocs.org,Carreau/readthedocs.org,wijerasa/readthedocs.org,SteveViss/readthedocs.org,singingwolfboy/readthedocs.org,Tazer/readthedocs.org,pombredanne/readthedocs.org,rtfd/readthedocs.org,nyergler/pythonslides,nikolas/readthedocs.org,michaelmcandrew/readthedocs.org,CedarLogic/readthedocs.org,safwanrahman/readthedocs.org,takluyver/readthedocs.org,royalwang/readthedocs.org,espdev/readthedocs.org,hach-que/readthedocs.org,Carreau/readthedocs.org,takluyver/readthedocs.org,dirn/readthedocs.org,d0ugal/readthedocs.org,sid-kap/readthedocs.org,fujita-shintaro/readthedocs.org,kdkeyser/readthedocs.org,tddv/readthedocs.org,espdev/readthedocs.org,asampat3090/readthedocs.org,LukasBoersma/readthedocs.org,kenshinthebattosai/readthedocs.org,nyergler/pythonslides,titiushko/readthedocs.org,techtonik/readthedocs.org,wanghaven/readthedocs.org,stevepiercy/readthedocs.org,rtfd/readthedocs.org,sunnyzwh/readthedocs.org,mhils/readthedocs.org,d0ugal/readthedocs.org,titiushko/readthedocs.org,istresearch/readthedocs.org,SteveViss/readthedocs.org,mrshoki/readthedocs.org,raven47git/readthedocs.org,atsuyim/readthedocs.org,wanghaven/readthedocs.org,soulshake/readthedocs.org,agjohnson/readthedocs.org,Carreau/readthedocs.org,tddv/readthedocs.org,KamranMackey/readthedocs.org,laplaceliu/readthedocs.org,nikolas/readthedocs.org,kenwang76/readthedocs.org,clarkperkins/readthedocs.org,VishvajitP/readthedocs.org,jerel/readthedocs.org,kenshinthebattosai/readthedocs.org,hach-que/readthedocs.org,asampat3090/readthedocs.org,stevepiercy/readthedocs.org,VishvajitP/readthedocs.org,ojii/readthedocs.org,agjohnson/readthedocs.org,agjohnson/readthedocs.org,jerel/readthedocs.org,attakei/readthedocs-oauth,michaelmcandrew/readthedocs.org,istresearch/readthedocs.org,safwanrahman/readthedocs.org,asampat3090/readthedocs.org,wanghaven/readthedocs.org,nikolas/readthedocs.org,emawind84/readthedocs.org,rtfd/readthedocs.org,wijerasa/readthedocs.org,jerel/readthedocs.org,cgourlay/readthedocs.org,sils1297/readthedocs.org,ojii/readthedocs.org,stevepiercy/readthedocs.org,LukasBoersma/readthedocs.org,raven47git/readthedocs.org,clarkperkins/readthedocs.org,wijerasa/readthedocs.org,CedarLogic/readthedocs.org,GovReady/readthedocs.org,singingwolfboy/readthedocs.org,kenwang76/readthedocs.org,mrshoki/readthedocs.org,istresearch/readthedocs.org,dirn/readthedocs.org,mrshoki/readthedocs.org,kenwang76/readthedocs.org,mhils/readthedocs.org,Tazer/readthedocs.org,sils1297/readthedocs.org,cgourlay/readthedocs.org,royalwang/readthedocs.org,sid-kap/readthedocs.org,emawind84/readthedocs.org,mrshoki/readthedocs.org,pombredanne/readthedocs.org,SteveViss/readthedocs.org,attakei/readthedocs-oauth,ojii/readthedocs.org,davidfischer/readthedocs.org,istresearch/readthedocs.org,hach-que/readthedocs.org,raven47git/readthedocs.org,fujita-shintaro/readthedocs.org,titiushko/readthedocs.org,LukasBoersma/readthedocs.org,singingwolfboy/readthedocs.org,royalwang/readthedocs.org,atsuyim/readthedocs.org,safwanrahman/readthedocs.org,VishvajitP/readthedocs.org,wanghaven/readthedocs.org,davidfischer/readthedocs.org,GovReady/readthedocs.org,clarkperkins/readthedocs.org,royalwang/readthedocs.org,atsuyim/readthedocs.org,dirn/readthedocs.org,sunnyzwh/readthedocs.org,stevepiercy/readthedocs.org,kdkeyser/readthedocs.org,gjtorikian/readthedocs.org,fujita-shintaro/readthedocs.org,sunnyzwh/readthedocs.org,wijerasa/readthedocs.org,attakei/readthedocs-oauth,michaelmcandrew/readthedocs.org,davidfischer/readthedocs.org,kenshinthebattosai/readthedocs.org,Tazer/readthedocs.org,atsuyim/readthedocs.org,KamranMackey/readthedocs.org,laplaceliu/readthedocs.org,pombredanne/readthedocs.org,CedarLogic/readthedocs.org,soulshake/readthedocs.org,cgourlay/readthedocs.org,d0ugal/readthedocs.org,emawind84/readthedocs.org,soulshake/readthedocs.org,espdev/readthedocs.org,kenwang76/readthedocs.org,espdev/readthedocs.org,raven47git/readthedocs.org,SteveViss/readthedocs.org,mhils/readthedocs.org,Carreau/readthedocs.org,laplaceliu/readthedocs.org,KamranMackey/readthedocs.org,sid-kap/readthedocs.org,soulshake/readthedocs.org,safwanrahman/readthedocs.org,espdev/readthedocs.org,gjtorikian/readthedocs.org,titiushko/readthedocs.org
from django.utils.translation import ugettext_lazy as _ import django_filters from builds import constants from builds.models import Build, Version ANY_REPO = ( ('', _('Any')), ) BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES class VersionFilter(django_filters.FilterSet): project = django_filters.CharFilter(name='project__name', lookup_type="icontains") slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains') class Meta: model = Version fields = ['project', 'slug'] class BuildFilter(django_filters.FilterSet): date = django_filters.DateRangeFilter(label=_("Build Date"), name="date") type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES) class Meta: model = Build - fields = ['type', 'date', 'version', 'success'] + fields = ['type', 'date', 'success'] -
Remove version from Build filter.
## Code Before: from django.utils.translation import ugettext_lazy as _ import django_filters from builds import constants from builds.models import Build, Version ANY_REPO = ( ('', _('Any')), ) BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES class VersionFilter(django_filters.FilterSet): project = django_filters.CharFilter(name='project__name', lookup_type="icontains") slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains') class Meta: model = Version fields = ['project', 'slug'] class BuildFilter(django_filters.FilterSet): date = django_filters.DateRangeFilter(label=_("Build Date"), name="date") type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES) class Meta: model = Build fields = ['type', 'date', 'version', 'success'] ## Instruction: Remove version from Build filter. ## Code After: from django.utils.translation import ugettext_lazy as _ import django_filters from builds import constants from builds.models import Build, Version ANY_REPO = ( ('', _('Any')), ) BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES class VersionFilter(django_filters.FilterSet): project = django_filters.CharFilter(name='project__name', lookup_type="icontains") slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains') class Meta: model = Version fields = ['project', 'slug'] class BuildFilter(django_filters.FilterSet): date = django_filters.DateRangeFilter(label=_("Build Date"), name="date") type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES) class Meta: model = Build fields = ['type', 'date', 'success']
from django.utils.translation import ugettext_lazy as _ import django_filters from builds import constants from builds.models import Build, Version ANY_REPO = ( ('', _('Any')), ) BUILD_TYPES = ANY_REPO + constants.BUILD_TYPES class VersionFilter(django_filters.FilterSet): project = django_filters.CharFilter(name='project__name', lookup_type="icontains") slug= django_filters.CharFilter(label=_("Slug"), name='slug', lookup_type='icontains') class Meta: model = Version fields = ['project', 'slug'] class BuildFilter(django_filters.FilterSet): date = django_filters.DateRangeFilter(label=_("Build Date"), name="date") type = django_filters.ChoiceFilter(label=_("Build Type"), choices=BUILD_TYPES) class Meta: model = Build - fields = ['type', 'date', 'version', 'success'] ? ----------- + fields = ['type', 'date', 'success'] -
7c894c716cb712bbcb137df3a5df5548bdca9d93
wafer/sponsors/migrations/0005_sponsorshippackage_symbol.py
wafer/sponsors/migrations/0005_sponsorshippackage_symbol.py
from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('sponsors', '0004_auto_20160813_1328'), ] operations = [ migrations.AddField( model_name='sponsorshippackage', name='symbol', field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True), ), ]
from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('sponsors', '0004_auto_20160813_1328'), ] operations = [ migrations.AddField( model_name='sponsorshippackage', name='symbol', field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1), ), ]
Update the migration to changed text
Update the migration to changed text
Python
isc
CTPUG/wafer,CTPUG/wafer,CTPUG/wafer,CTPUG/wafer
from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('sponsors', '0004_auto_20160813_1328'), ] operations = [ migrations.AddField( model_name='sponsorshippackage', name='symbol', - field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True), + field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1), ), ]
Update the migration to changed text
## Code Before: from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('sponsors', '0004_auto_20160813_1328'), ] operations = [ migrations.AddField( model_name='sponsorshippackage', name='symbol', field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True), ), ] ## Instruction: Update the migration to changed text ## Code After: from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('sponsors', '0004_auto_20160813_1328'), ] operations = [ migrations.AddField( model_name='sponsorshippackage', name='symbol', field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1), ), ]
from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('sponsors', '0004_auto_20160813_1328'), ] operations = [ migrations.AddField( model_name='sponsorshippackage', name='symbol', - field=models.CharField(help_text='Optional symbol to display next to sponsors backing at this level sponsors list', max_length=1, blank=True), + field=models.CharField(blank=True, help_text='Optional symbol to display in the sponsors list next to sponsors who have sponsored at this list, (for example *).', max_length=1), ), ]
54b2a6953a4da2b217052d166ad1f069f683b9ee
scripts/nomenclature/nomenclature_map.py
scripts/nomenclature/nomenclature_map.py
import pandas as pd itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1")
import pandas as pd from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder # Globals INPUT_FILE_NAME = "search_results.csv" # Input nomenclature_folder = locate_nomenclature_folder() itis_results = pd.read_csv( str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1")
Implement locator in nomenclature map
Implement locator in nomenclature map
Python
mit
jnfrye/local_plants_book
import pandas as pd - itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1") + from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder + + # Globals + INPUT_FILE_NAME = "search_results.csv" + + # Input + nomenclature_folder = locate_nomenclature_folder() + itis_results = pd.read_csv( + str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1") +
Implement locator in nomenclature map
## Code Before: import pandas as pd itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1") ## Instruction: Implement locator in nomenclature map ## Code After: import pandas as pd from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder # Globals INPUT_FILE_NAME = "search_results.csv" # Input nomenclature_folder = locate_nomenclature_folder() itis_results = pd.read_csv( str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1")
import pandas as pd - itis_results = pd.read_csv("search_result.csv", encoding = "ISO-8859-1") + from PyFloraBook.in_out.data_coordinator import locate_nomenclature_folder + + + # Globals + INPUT_FILE_NAME = "search_results.csv" + + # Input + nomenclature_folder = locate_nomenclature_folder() + itis_results = pd.read_csv( + str(nomenclature_folder / INPUT_FILE_NAME), encoding="ISO-8859-1")
6353a3d1443c717b2d2e804190153f8be605c2f1
setup.py
setup.py
from distutils.core import setup with open('README.rst') as readme: long_description = readme.read() setup( name='udiskie', version='0.4.2', description='Removable disk automounter for udisks', long_description=long_description, author='Byron Clark', author_email='byron@theclarkfamily.name', maintainer='Thomas Gläßle', maintainer_email='t_glaessle@gmx.de', url='https://github.com/coldfix/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', ], )
from distutils.core import setup with open('README.rst') as readme: long_description = readme.read() setup( name='udiskie', version='0.4.2', description='Removable disk automounter for udisks', long_description=long_description, author='Byron Clark', author_email='byron@theclarkfamily.name', maintainer='Thomas Gläßle', maintainer_email='t_glaessle@gmx.de', url='https://github.com/coldfix/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', 'bin/udiskie-mount' ], )
Include udiskie-mount in binary distribution
Include udiskie-mount in binary distribution
Python
mit
khardix/udiskie,pstray/udiskie,coldfix/udiskie,coldfix/udiskie,mathstuf/udiskie,pstray/udiskie
from distutils.core import setup with open('README.rst') as readme: long_description = readme.read() setup( name='udiskie', version='0.4.2', description='Removable disk automounter for udisks', long_description=long_description, author='Byron Clark', author_email='byron@theclarkfamily.name', maintainer='Thomas Gläßle', maintainer_email='t_glaessle@gmx.de', url='https://github.com/coldfix/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', + 'bin/udiskie-mount' ], )
Include udiskie-mount in binary distribution
## Code Before: from distutils.core import setup with open('README.rst') as readme: long_description = readme.read() setup( name='udiskie', version='0.4.2', description='Removable disk automounter for udisks', long_description=long_description, author='Byron Clark', author_email='byron@theclarkfamily.name', maintainer='Thomas Gläßle', maintainer_email='t_glaessle@gmx.de', url='https://github.com/coldfix/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', ], ) ## Instruction: Include udiskie-mount in binary distribution ## Code After: from distutils.core import setup with open('README.rst') as readme: long_description = readme.read() setup( name='udiskie', version='0.4.2', description='Removable disk automounter for udisks', long_description=long_description, author='Byron Clark', author_email='byron@theclarkfamily.name', maintainer='Thomas Gläßle', maintainer_email='t_glaessle@gmx.de', url='https://github.com/coldfix/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', 'bin/udiskie-mount' ], )
from distutils.core import setup with open('README.rst') as readme: long_description = readme.read() setup( name='udiskie', version='0.4.2', description='Removable disk automounter for udisks', long_description=long_description, author='Byron Clark', author_email='byron@theclarkfamily.name', maintainer='Thomas Gläßle', maintainer_email='t_glaessle@gmx.de', url='https://github.com/coldfix/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', + 'bin/udiskie-mount' ], )
1cf1da043ceab767d9d0dbdbed62c2f1c5ff36e9
test_http.py
test_http.py
from http_server import HttpServer import socket def test_200_ok(): s = HttpServer() assert s.ok() == "HTTP/1.1 200 OK" def test_200_ok_byte(): s = HttpServer() assert isinstance(s.ok(), bytes) def test_socket_is_socket(): s = HttpServer() s.open_socket() assert isinstance(s._socket, socket.socket) def test_open_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() assert s._socket.getsockname() == ('127.0.0.1', 50000)
from http_server import HttpServer import socket def test_200_ok(): s = HttpServer() assert s.ok() == "HTTP/1.1 200 OK" def test_200_ok_byte(): s = HttpServer() assert isinstance(s.ok(), bytes) def test_socket_is_socket(): s = HttpServer() s.open_socket() assert isinstance(s._socket, socket.socket) def test_open_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() assert s._socket.getsockname() == ('127.0.0.1', 50000) def test_close_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() s.close_socket() assert s._socket is None
Add tests for closing a socket
Add tests for closing a socket
Python
mit
jefrailey/network_tools
from http_server import HttpServer import socket def test_200_ok(): s = HttpServer() assert s.ok() == "HTTP/1.1 200 OK" def test_200_ok_byte(): s = HttpServer() assert isinstance(s.ok(), bytes) def test_socket_is_socket(): s = HttpServer() s.open_socket() assert isinstance(s._socket, socket.socket) def test_open_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() assert s._socket.getsockname() == ('127.0.0.1', 50000) + + + def test_close_socket(): + s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) + s.open_socket() + s.close_socket() + assert s._socket is None
Add tests for closing a socket
## Code Before: from http_server import HttpServer import socket def test_200_ok(): s = HttpServer() assert s.ok() == "HTTP/1.1 200 OK" def test_200_ok_byte(): s = HttpServer() assert isinstance(s.ok(), bytes) def test_socket_is_socket(): s = HttpServer() s.open_socket() assert isinstance(s._socket, socket.socket) def test_open_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() assert s._socket.getsockname() == ('127.0.0.1', 50000) ## Instruction: Add tests for closing a socket ## Code After: from http_server import HttpServer import socket def test_200_ok(): s = HttpServer() assert s.ok() == "HTTP/1.1 200 OK" def test_200_ok_byte(): s = HttpServer() assert isinstance(s.ok(), bytes) def test_socket_is_socket(): s = HttpServer() s.open_socket() assert isinstance(s._socket, socket.socket) def test_open_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() assert s._socket.getsockname() == ('127.0.0.1', 50000) def test_close_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() s.close_socket() assert s._socket is None
from http_server import HttpServer import socket def test_200_ok(): s = HttpServer() assert s.ok() == "HTTP/1.1 200 OK" def test_200_ok_byte(): s = HttpServer() assert isinstance(s.ok(), bytes) def test_socket_is_socket(): s = HttpServer() s.open_socket() assert isinstance(s._socket, socket.socket) def test_open_socket(): s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) s.open_socket() assert s._socket.getsockname() == ('127.0.0.1', 50000) + + + def test_close_socket(): + s = HttpServer(ip=u'127.0.0.1', port=50000, backlog=5) + s.open_socket() + s.close_socket() + assert s._socket is None
3f166b110d4e8623966ca29c71445973da4876f9
armstrong/hatband/forms.py
armstrong/hatband/forms.py
from django import forms from django.db import models from . import widgets RICH_TEXT_DBFIELD_OVERRIDES = { models.TextField: {'widget': widgets.RichTextWidget}, } class BackboneFormMixin(object): class Media: js = ( 'hatband/js/jquery-1.6.2.min.js', 'hatband/js/underscore.js', 'hatband/js/backbone.js', 'hatband/js/backbone-inline-base.js') class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm): class Meta: widgets = { "content_type": forms.HiddenInput(), "object_id": widgets.GenericKeyWidget(), "order": forms.HiddenInput(), }
from django import forms from django.conf import settings from django.db import models from . import widgets RICH_TEXT_DBFIELD_OVERRIDES = { models.TextField: {'widget': widgets.RichTextWidget}, } class BackboneFormMixin(object): if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True): class Media: js = ( 'hatband/js/jquery-1.6.2.min.js', 'hatband/js/underscore.js', 'hatband/js/backbone.js', 'hatband/js/backbone-inline-base.js') class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm): class Meta: widgets = { "content_type": forms.HiddenInput(), "object_id": widgets.GenericKeyWidget(), "order": forms.HiddenInput(), }
Make it possible to turn off admin JS
Make it possible to turn off admin JS
Python
apache-2.0
armstrong/armstrong.hatband,armstrong/armstrong.hatband,texastribune/armstrong.hatband,armstrong/armstrong.hatband,texastribune/armstrong.hatband,texastribune/armstrong.hatband
from django import forms + from django.conf import settings from django.db import models from . import widgets RICH_TEXT_DBFIELD_OVERRIDES = { models.TextField: {'widget': widgets.RichTextWidget}, } class BackboneFormMixin(object): + if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True): - class Media: + class Media: - js = ( + js = ( - 'hatband/js/jquery-1.6.2.min.js', + 'hatband/js/jquery-1.6.2.min.js', - 'hatband/js/underscore.js', + 'hatband/js/underscore.js', - 'hatband/js/backbone.js', + 'hatband/js/backbone.js', - 'hatband/js/backbone-inline-base.js') + 'hatband/js/backbone-inline-base.js') class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm): class Meta: widgets = { "content_type": forms.HiddenInput(), "object_id": widgets.GenericKeyWidget(), "order": forms.HiddenInput(), }
Make it possible to turn off admin JS
## Code Before: from django import forms from django.db import models from . import widgets RICH_TEXT_DBFIELD_OVERRIDES = { models.TextField: {'widget': widgets.RichTextWidget}, } class BackboneFormMixin(object): class Media: js = ( 'hatband/js/jquery-1.6.2.min.js', 'hatband/js/underscore.js', 'hatband/js/backbone.js', 'hatband/js/backbone-inline-base.js') class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm): class Meta: widgets = { "content_type": forms.HiddenInput(), "object_id": widgets.GenericKeyWidget(), "order": forms.HiddenInput(), } ## Instruction: Make it possible to turn off admin JS ## Code After: from django import forms from django.conf import settings from django.db import models from . import widgets RICH_TEXT_DBFIELD_OVERRIDES = { models.TextField: {'widget': widgets.RichTextWidget}, } class BackboneFormMixin(object): if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True): class Media: js = ( 'hatband/js/jquery-1.6.2.min.js', 'hatband/js/underscore.js', 'hatband/js/backbone.js', 'hatband/js/backbone-inline-base.js') class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm): class Meta: widgets = { "content_type": forms.HiddenInput(), "object_id": widgets.GenericKeyWidget(), "order": forms.HiddenInput(), }
from django import forms + from django.conf import settings from django.db import models from . import widgets RICH_TEXT_DBFIELD_OVERRIDES = { models.TextField: {'widget': widgets.RichTextWidget}, } class BackboneFormMixin(object): + if getattr(settings, "ARMSTRONG_ADMIN_PROVIDE_STATIC", True): - class Media: + class Media: ? ++++ - js = ( + js = ( ? ++++ - 'hatband/js/jquery-1.6.2.min.js', + 'hatband/js/jquery-1.6.2.min.js', ? ++++ - 'hatband/js/underscore.js', + 'hatband/js/underscore.js', ? ++++ - 'hatband/js/backbone.js', + 'hatband/js/backbone.js', ? ++++ - 'hatband/js/backbone-inline-base.js') + 'hatband/js/backbone-inline-base.js') ? ++++ class OrderableGenericKeyLookupForm(BackboneFormMixin, forms.ModelForm): class Meta: widgets = { "content_type": forms.HiddenInput(), "object_id": widgets.GenericKeyWidget(), "order": forms.HiddenInput(), }
6e6bffc19873260696822bb3f4a821ce4ea6f4a3
consulrest/keyvalue.py
consulrest/keyvalue.py
import json import re import requests class KeyValue(object): def __init__(self, url): self._url = "%s/kv" % url def _get(self, key, recurse=None, keys=None): url = self._url + '/' + key if recurse is not None: url += '?recurse' if keys is not None: url += '?keys' r = requests.get(url) if r.status_code == 200: return json.loads(r.text) else: return None def get(self, key, recurse=None): return self._get(key, recurse=recurse) def list(self, key=''): return self._get(key, keys=True) def set(self, key, value): r = requests.put(self._url + '/' + key, data=value) if r.status_code == 200 and re.match(r"true", r.text) is not None: return True else: return False def delete(self, key, recurse=None): url = self._url + '/' + key if recurse is not None: url += '?recurse' requests.delete(url)
import json import re import requests class KeyValue(object): def __init__(self, url): self._url = "%s/kv" % url def _get(self, key, recurse=None, keys=None): url = self._url + '/' + key params = dict() if recurse is not None: params['recurse'] = True if keys is not None: params['keys'] = True r = requests.get(url, params=params) if r.status_code == 200: return json.loads(r.text) else: return None def get(self, key, recurse=None): return self._get(key, recurse=recurse) def list(self, key=''): return self._get(key, keys=True) def set(self, key, value): r = requests.put(self._url + '/' + key, data=value) if r.status_code == 200 and re.match(r"true", r.text) is not None: return True else: return False def delete(self, key, recurse=None): url = self._url + '/' + key params = dict() if recurse is not None: params['recurse'] = True requests.delete(url, params=params)
Use params dictionary instead of appending to the end of URL string
Use params dictionary instead of appending to the end of URL string
Python
mit
vcoque/consul-ri
import json import re import requests class KeyValue(object): def __init__(self, url): self._url = "%s/kv" % url def _get(self, key, recurse=None, keys=None): url = self._url + '/' + key + params = dict() if recurse is not None: - url += '?recurse' + params['recurse'] = True + if keys is not None: + params['keys'] = True - if keys is not None: - url += '?keys' - - r = requests.get(url) + r = requests.get(url, params=params) if r.status_code == 200: return json.loads(r.text) else: return None def get(self, key, recurse=None): return self._get(key, recurse=recurse) def list(self, key=''): return self._get(key, keys=True) def set(self, key, value): r = requests.put(self._url + '/' + key, data=value) if r.status_code == 200 and re.match(r"true", r.text) is not None: return True else: return False def delete(self, key, recurse=None): url = self._url + '/' + key + + params = dict() if recurse is not None: - url += '?recurse' + params['recurse'] = True - requests.delete(url) + requests.delete(url, params=params)
Use params dictionary instead of appending to the end of URL string
## Code Before: import json import re import requests class KeyValue(object): def __init__(self, url): self._url = "%s/kv" % url def _get(self, key, recurse=None, keys=None): url = self._url + '/' + key if recurse is not None: url += '?recurse' if keys is not None: url += '?keys' r = requests.get(url) if r.status_code == 200: return json.loads(r.text) else: return None def get(self, key, recurse=None): return self._get(key, recurse=recurse) def list(self, key=''): return self._get(key, keys=True) def set(self, key, value): r = requests.put(self._url + '/' + key, data=value) if r.status_code == 200 and re.match(r"true", r.text) is not None: return True else: return False def delete(self, key, recurse=None): url = self._url + '/' + key if recurse is not None: url += '?recurse' requests.delete(url) ## Instruction: Use params dictionary instead of appending to the end of URL string ## Code After: import json import re import requests class KeyValue(object): def __init__(self, url): self._url = "%s/kv" % url def _get(self, key, recurse=None, keys=None): url = self._url + '/' + key params = dict() if recurse is not None: params['recurse'] = True if keys is not None: params['keys'] = True r = requests.get(url, params=params) if r.status_code == 200: return json.loads(r.text) else: return None def get(self, key, recurse=None): return self._get(key, recurse=recurse) def list(self, key=''): return self._get(key, keys=True) def set(self, key, value): r = requests.put(self._url + '/' + key, data=value) if r.status_code == 200 and re.match(r"true", r.text) is not None: return True else: return False def delete(self, key, recurse=None): url = self._url + '/' + key params = dict() if recurse is not None: params['recurse'] = True requests.delete(url, params=params)
import json import re import requests class KeyValue(object): def __init__(self, url): self._url = "%s/kv" % url def _get(self, key, recurse=None, keys=None): url = self._url + '/' + key + params = dict() if recurse is not None: - url += '?recurse' + params['recurse'] = True + if keys is not None: + params['keys'] = True - if keys is not None: - url += '?keys' - - r = requests.get(url) + r = requests.get(url, params=params) ? +++++++++++++++ if r.status_code == 200: return json.loads(r.text) else: return None def get(self, key, recurse=None): return self._get(key, recurse=recurse) def list(self, key=''): return self._get(key, keys=True) def set(self, key, value): r = requests.put(self._url + '/' + key, data=value) if r.status_code == 200 and re.match(r"true", r.text) is not None: return True else: return False def delete(self, key, recurse=None): url = self._url + '/' + key + + params = dict() if recurse is not None: - url += '?recurse' + params['recurse'] = True - requests.delete(url) + requests.delete(url, params=params) ? +++++++++++++++
f3b9cc6392e4c271ae11417357ecdc196f1c3ae7
python_scripts/extractor_python_readability_server.py
python_scripts/extractor_python_readability_server.py
import sys import os import glob #sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py")) sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/")) sys.path.append(os.path.dirname(__file__) ) from thrift.transport import TSocket from thrift.server import TServer #import thrift_solr import ExtractorService import sys import readability import readability def extract_with_python_readability( raw_content ): doc = readability.Document( raw_content ) return [ u'' + doc.short_title(), u'' + doc.summary() ] class ExtractorHandler: def extract_html( self, raw_html ): #print raw_html #raw_html = raw_html.encode( 'utf-8' ) ret = extract_with_python_readability( raw_html ) #print ret[1] return ret handler = ExtractorHandler() processor = ExtractorService.Processor(handler) listening_socket = TSocket.TServerSocket(port=9090) server = TServer.TThreadPoolServer(processor, listening_socket) print ("[Server] Started") server.serve()
import sys import os import glob #sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py")) sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/")) sys.path.append(os.path.dirname(__file__) ) from thrift.transport import TSocket from thrift.transport import TTransport from thrift.protocol import TBinaryProtocol from thrift.server import TServer from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated #import thrift_solr import ExtractorService import sys import readability import readability def extract_with_python_readability( raw_content ): doc = readability.Document( raw_content ) return [ u'' + doc.short_title(), u'' + doc.summary() ] class ExtractorHandler: def extract_html( self, raw_html ): #print raw_html #raw_html = raw_html.encode( 'utf-8' ) ret = extract_with_python_readability( raw_html ) #print ret[1] return ret handler = ExtractorHandler() processor = ExtractorService.Processor(handler) listening_socket = TSocket.TServerSocket(port=9090) tfactory = TTransport.TBufferedTransportFactory() #pfactory = TBinaryProtocol.TBinaryProtocolFactory() pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory() server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory) print ("[Server] Started") server.serve()
Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance.
Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance.
Python
agpl-3.0
AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,berkmancenter/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud,AchyuthIIIT/mediacloud
import sys import os import glob #sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py")) sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/")) sys.path.append(os.path.dirname(__file__) ) from thrift.transport import TSocket + from thrift.transport import TTransport + from thrift.protocol import TBinaryProtocol from thrift.server import TServer + from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated + #import thrift_solr import ExtractorService import sys import readability import readability def extract_with_python_readability( raw_content ): doc = readability.Document( raw_content ) return [ u'' + doc.short_title(), u'' + doc.summary() ] class ExtractorHandler: def extract_html( self, raw_html ): #print raw_html #raw_html = raw_html.encode( 'utf-8' ) ret = extract_with_python_readability( raw_html ) #print ret[1] return ret handler = ExtractorHandler() processor = ExtractorService.Processor(handler) listening_socket = TSocket.TServerSocket(port=9090) + tfactory = TTransport.TBufferedTransportFactory() + #pfactory = TBinaryProtocol.TBinaryProtocolFactory() + pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory() + - server = TServer.TThreadPoolServer(processor, listening_socket) + server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory) print ("[Server] Started") server.serve()
Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance.
## Code Before: import sys import os import glob #sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py")) sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/")) sys.path.append(os.path.dirname(__file__) ) from thrift.transport import TSocket from thrift.server import TServer #import thrift_solr import ExtractorService import sys import readability import readability def extract_with_python_readability( raw_content ): doc = readability.Document( raw_content ) return [ u'' + doc.short_title(), u'' + doc.summary() ] class ExtractorHandler: def extract_html( self, raw_html ): #print raw_html #raw_html = raw_html.encode( 'utf-8' ) ret = extract_with_python_readability( raw_html ) #print ret[1] return ret handler = ExtractorHandler() processor = ExtractorService.Processor(handler) listening_socket = TSocket.TServerSocket(port=9090) server = TServer.TThreadPoolServer(processor, listening_socket) print ("[Server] Started") server.serve() ## Instruction: Use the TBinaryProtocolAccelerated protocol instead of TBinaryProtocol to improve performance. ## Code After: import sys import os import glob #sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py")) sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/")) sys.path.append(os.path.dirname(__file__) ) from thrift.transport import TSocket from thrift.transport import TTransport from thrift.protocol import TBinaryProtocol from thrift.server import TServer from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated #import thrift_solr import ExtractorService import sys import readability import readability def extract_with_python_readability( raw_content ): doc = readability.Document( raw_content ) return [ u'' + doc.short_title(), u'' + doc.summary() ] class ExtractorHandler: def extract_html( self, raw_html ): #print raw_html #raw_html = raw_html.encode( 'utf-8' ) ret = extract_with_python_readability( raw_html ) #print ret[1] return ret handler = ExtractorHandler() processor = ExtractorService.Processor(handler) listening_socket = TSocket.TServerSocket(port=9090) tfactory = TTransport.TBufferedTransportFactory() #pfactory = TBinaryProtocol.TBinaryProtocolFactory() pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory() server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory) print ("[Server] Started") server.serve()
import sys import os import glob #sys.path.append(os.path.join(os.path.dirname(__file__), "gen-py")) sys.path.append(os.path.join(os.path.dirname(__file__),"gen-py/thrift_solr/")) sys.path.append(os.path.dirname(__file__) ) from thrift.transport import TSocket + from thrift.transport import TTransport + from thrift.protocol import TBinaryProtocol from thrift.server import TServer + from thrift.protocol.TBinaryProtocol import TBinaryProtocolAccelerated + #import thrift_solr import ExtractorService import sys import readability import readability def extract_with_python_readability( raw_content ): doc = readability.Document( raw_content ) return [ u'' + doc.short_title(), u'' + doc.summary() ] class ExtractorHandler: def extract_html( self, raw_html ): #print raw_html #raw_html = raw_html.encode( 'utf-8' ) ret = extract_with_python_readability( raw_html ) #print ret[1] return ret handler = ExtractorHandler() processor = ExtractorService.Processor(handler) listening_socket = TSocket.TServerSocket(port=9090) + tfactory = TTransport.TBufferedTransportFactory() + #pfactory = TBinaryProtocol.TBinaryProtocolFactory() + pfactory = TBinaryProtocol.TBinaryProtocolAcceleratedFactory() + - server = TServer.TThreadPoolServer(processor, listening_socket) + server = TServer.TThreadPoolServer(processor, listening_socket, tfactory, pfactory) ? ++++++++++++++++++++ print ("[Server] Started") server.serve()
106fc0f8bae7c776a8f6c7dcec2947420492d118
homographynet/callbacks.py
homographynet/callbacks.py
from keras.callbacks import Callback import keras.backend as K class LearningRateScheduler(Callback): """Learning rate scheduler. See Caffe SGD docs """ def __init__(self, base_lr, gamma, step_size): super().__init__() self._lr = base_lr self._gamma = gamma self._step_size = step_size self._iteration = 1 def on_batch_begin(self, batch, logs=None): if self._iteration % self._step_size == 0: self._lr *= self._gamma K.set_value(self.model.optimizer.lr, self._lr) print('New learning rate:', self._lr) self._iteration += 1
from keras.callbacks import Callback import keras.backend as K class LearningRateScheduler(Callback): """Learning rate scheduler. See Caffe SGD docs """ def __init__(self, base_lr, gamma, step_size): super().__init__() self._base_lr = base_lr self._gamma = gamma self._step_size = step_size self._steps = 0 def on_epoch_begin(self, epoch, logs=None): self._steps = epoch * self.params['steps'] def on_batch_begin(self, batch, logs=None): self._steps += 1 if self._steps % self._step_size == 0: exp = int(self._steps / self._step_size) lr = self._base_lr * (self._gamma ** exp) K.set_value(self.model.optimizer.lr, lr) print('New learning rate:', lr)
Fix calculation of current steps when starting with epoch != 0
Fix calculation of current steps when starting with epoch != 0
Python
apache-2.0
baudm/HomographyNet
from keras.callbacks import Callback import keras.backend as K class LearningRateScheduler(Callback): """Learning rate scheduler. See Caffe SGD docs """ def __init__(self, base_lr, gamma, step_size): super().__init__() - self._lr = base_lr + self._base_lr = base_lr self._gamma = gamma self._step_size = step_size - self._iteration = 1 + self._steps = 0 + + def on_epoch_begin(self, epoch, logs=None): + self._steps = epoch * self.params['steps'] def on_batch_begin(self, batch, logs=None): + self._steps += 1 - if self._iteration % self._step_size == 0: + if self._steps % self._step_size == 0: + exp = int(self._steps / self._step_size) - self._lr *= self._gamma + lr = self._base_lr * (self._gamma ** exp) - K.set_value(self.model.optimizer.lr, self._lr) + K.set_value(self.model.optimizer.lr, lr) - print('New learning rate:', self._lr) + print('New learning rate:', lr) - self._iteration += 1
Fix calculation of current steps when starting with epoch != 0
## Code Before: from keras.callbacks import Callback import keras.backend as K class LearningRateScheduler(Callback): """Learning rate scheduler. See Caffe SGD docs """ def __init__(self, base_lr, gamma, step_size): super().__init__() self._lr = base_lr self._gamma = gamma self._step_size = step_size self._iteration = 1 def on_batch_begin(self, batch, logs=None): if self._iteration % self._step_size == 0: self._lr *= self._gamma K.set_value(self.model.optimizer.lr, self._lr) print('New learning rate:', self._lr) self._iteration += 1 ## Instruction: Fix calculation of current steps when starting with epoch != 0 ## Code After: from keras.callbacks import Callback import keras.backend as K class LearningRateScheduler(Callback): """Learning rate scheduler. See Caffe SGD docs """ def __init__(self, base_lr, gamma, step_size): super().__init__() self._base_lr = base_lr self._gamma = gamma self._step_size = step_size self._steps = 0 def on_epoch_begin(self, epoch, logs=None): self._steps = epoch * self.params['steps'] def on_batch_begin(self, batch, logs=None): self._steps += 1 if self._steps % self._step_size == 0: exp = int(self._steps / self._step_size) lr = self._base_lr * (self._gamma ** exp) K.set_value(self.model.optimizer.lr, lr) print('New learning rate:', lr)
from keras.callbacks import Callback import keras.backend as K class LearningRateScheduler(Callback): """Learning rate scheduler. See Caffe SGD docs """ def __init__(self, base_lr, gamma, step_size): super().__init__() - self._lr = base_lr + self._base_lr = base_lr ? +++++ self._gamma = gamma self._step_size = step_size - self._iteration = 1 ? ^ ^^^^^^ ^ + self._steps = 0 ? ^ ^^ ^ + + def on_epoch_begin(self, epoch, logs=None): + self._steps = epoch * self.params['steps'] def on_batch_begin(self, batch, logs=None): + self._steps += 1 - if self._iteration % self._step_size == 0: ? ^ ^^^^^^ + if self._steps % self._step_size == 0: ? ^ ^^ + exp = int(self._steps / self._step_size) - self._lr *= self._gamma ? - + lr = self._base_lr * (self._gamma ** exp) ? +++++ +++++ + ++++++++ - K.set_value(self.model.optimizer.lr, self._lr) ? ------ + K.set_value(self.model.optimizer.lr, lr) - print('New learning rate:', self._lr) ? ------ + print('New learning rate:', lr) - self._iteration += 1
e70537eb2c1a8a68a6a66550e6714816e048bb5e
tests/integration/modules/git.py
tests/integration/modules/git.py
import shutil import subprocess import tempfile # Import Salt Testing libs from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import salt libs import integration class GitModuleTest(integration.ModuleCase): @classmethod def setUpClass(cls): from salt.utils import which git = which('git') if not git: self.skipTest('The git binary is not available') def setUp(self): self.repos = tempfile.mkdtemp(dir=integration.TMP) self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True) subprocess.check_call(['git', 'init', '--quiet', self.repos]) def test_config_set_value_has_space_characters(self): ''' git.config_set ''' config_key = "user.name" config_value = "foo bar" ret = self.run_function( 'git.config_set', cwd=self.repos, setting_name=config_key, setting_value=config_value, ) self.assertEqual("", ret) output = subprocess.check_output( ['git', 'config', '--local', config_key], cwd=self.repos) self.assertEqual(config_value + "\n", output)
import shutil import subprocess import tempfile # Import Salt Testing libs from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import salt libs import integration class GitModuleTest(integration.ModuleCase): ''' Integration tests for the git module ''' @classmethod def setUpClass(cls): ''' Check if git is installed. If it isn't, skip everything in this class. ''' from salt.utils import which git = which('git') if not git: cls.skipTest('The git binary is not available') def setUp(self): self.repos = tempfile.mkdtemp(dir=integration.TMP) self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True) subprocess.check_call(['git', 'init', '--quiet', self.repos]) def test_config_set_value_has_space_characters(self): ''' Tests the git.config_set function ''' config_key = "user.name" config_value = "foo bar" ret = self.run_function( 'git.config_set', cwd=self.repos, setting_name=config_key, setting_value=config_value, ) self.assertEqual("", ret) output = subprocess.check_output( ['git', 'config', '--local', config_key], cwd=self.repos) self.assertEqual(config_value + "\n", output)
Fix missing cls variable and add some docstring info
Fix missing cls variable and add some docstring info
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
- import shutil import subprocess import tempfile # Import Salt Testing libs from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import salt libs import integration class GitModuleTest(integration.ModuleCase): + ''' + Integration tests for the git module + ''' + @classmethod def setUpClass(cls): + ''' + Check if git is installed. If it isn't, skip everything in this class. + ''' from salt.utils import which git = which('git') if not git: - self.skipTest('The git binary is not available') + cls.skipTest('The git binary is not available') def setUp(self): self.repos = tempfile.mkdtemp(dir=integration.TMP) self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True) subprocess.check_call(['git', 'init', '--quiet', self.repos]) def test_config_set_value_has_space_characters(self): ''' - git.config_set + Tests the git.config_set function ''' config_key = "user.name" config_value = "foo bar" ret = self.run_function( 'git.config_set', cwd=self.repos, setting_name=config_key, setting_value=config_value, ) self.assertEqual("", ret) output = subprocess.check_output( ['git', 'config', '--local', config_key], cwd=self.repos) self.assertEqual(config_value + "\n", output)
Fix missing cls variable and add some docstring info
## Code Before: import shutil import subprocess import tempfile # Import Salt Testing libs from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import salt libs import integration class GitModuleTest(integration.ModuleCase): @classmethod def setUpClass(cls): from salt.utils import which git = which('git') if not git: self.skipTest('The git binary is not available') def setUp(self): self.repos = tempfile.mkdtemp(dir=integration.TMP) self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True) subprocess.check_call(['git', 'init', '--quiet', self.repos]) def test_config_set_value_has_space_characters(self): ''' git.config_set ''' config_key = "user.name" config_value = "foo bar" ret = self.run_function( 'git.config_set', cwd=self.repos, setting_name=config_key, setting_value=config_value, ) self.assertEqual("", ret) output = subprocess.check_output( ['git', 'config', '--local', config_key], cwd=self.repos) self.assertEqual(config_value + "\n", output) ## Instruction: Fix missing cls variable and add some docstring info ## Code After: import shutil import subprocess import tempfile # Import Salt Testing libs from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import salt libs import integration class GitModuleTest(integration.ModuleCase): ''' Integration tests for the git module ''' @classmethod def setUpClass(cls): ''' Check if git is installed. If it isn't, skip everything in this class. ''' from salt.utils import which git = which('git') if not git: cls.skipTest('The git binary is not available') def setUp(self): self.repos = tempfile.mkdtemp(dir=integration.TMP) self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True) subprocess.check_call(['git', 'init', '--quiet', self.repos]) def test_config_set_value_has_space_characters(self): ''' Tests the git.config_set function ''' config_key = "user.name" config_value = "foo bar" ret = self.run_function( 'git.config_set', cwd=self.repos, setting_name=config_key, setting_value=config_value, ) self.assertEqual("", ret) output = subprocess.check_output( ['git', 'config', '--local', config_key], cwd=self.repos) self.assertEqual(config_value + "\n", output)
- import shutil import subprocess import tempfile # Import Salt Testing libs from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import salt libs import integration class GitModuleTest(integration.ModuleCase): + ''' + Integration tests for the git module + ''' + @classmethod def setUpClass(cls): + ''' + Check if git is installed. If it isn't, skip everything in this class. + ''' from salt.utils import which git = which('git') if not git: - self.skipTest('The git binary is not available') ? --- + cls.skipTest('The git binary is not available') ? ++ def setUp(self): self.repos = tempfile.mkdtemp(dir=integration.TMP) self.addCleanup(shutil.rmtree, self.repos, ignore_errors=True) subprocess.check_call(['git', 'init', '--quiet', self.repos]) def test_config_set_value_has_space_characters(self): ''' - git.config_set + Tests the git.config_set function ''' config_key = "user.name" config_value = "foo bar" ret = self.run_function( 'git.config_set', cwd=self.repos, setting_name=config_key, setting_value=config_value, ) self.assertEqual("", ret) output = subprocess.check_output( ['git', 'config', '--local', config_key], cwd=self.repos) self.assertEqual(config_value + "\n", output)
9771381323e4eb44a13ffc8742615fba61ad2b85
lino/modlib/notify/consumers.py
lino/modlib/notify/consumers.py
from channels import Group def ws_echo(message): Group(str(message.content['text'])).add(message.reply_channel) message.reply_channel.send({ "text": message.content['text'], })
import json from channels import Channel from channels import Group from channels.auth import channel_session_user, channel_session_user_from_http from django.utils import timezone from lino.modlib.notify.models import Notification # This decorator copies the user from the HTTP session (only available in # websocket.connect or http.request messages) to the channel session (available # in all consumers with the same reply_channel, so all three here) @channel_session_user_from_http def ws_connect(message): pass def ws_receive(message): # All WebSocket frames have either a text or binary payload; we decode the # text part here assuming it's JSON. # You could easily build up a basic framework that did this encoding/decoding # for you as well as handling common errors. payload = json.loads(message['text']) payload['reply_channel'] = message.content['reply_channel'] Channel("notify.receive").send(payload) @channel_session_user def set_notification_as_seen(message): notification_id = message['notification_id'] notif = Notification.objects.get(pk=notification_id) notif.seen = timezone.now() notif.save() @channel_session_user def user_connected(message): username = message['username'] Group(username).add(message.reply_channel) message.reply_channel.send({ "text": username, })
Update receive and send functions according to the new requirements
Update receive and send functions according to the new requirements
Python
unknown
lsaffre/lino,lsaffre/lino,khchine5/lino,khchine5/lino,khchine5/lino,lino-framework/lino,lino-framework/lino,lsaffre/lino,lsaffre/lino,lino-framework/lino,lino-framework/lino,lsaffre/lino,khchine5/lino,khchine5/lino,lino-framework/lino
+ import json + + from channels import Channel from channels import Group + from channels.auth import channel_session_user, channel_session_user_from_http + from django.utils import timezone + from lino.modlib.notify.models import Notification + # This decorator copies the user from the HTTP session (only available in + # websocket.connect or http.request messages) to the channel session (available + # in all consumers with the same reply_channel, so all three here) + @channel_session_user_from_http - def ws_echo(message): + def ws_connect(message): + pass + + + def ws_receive(message): + # All WebSocket frames have either a text or binary payload; we decode the + # text part here assuming it's JSON. + # You could easily build up a basic framework that did this encoding/decoding + # for you as well as handling common errors. + payload = json.loads(message['text']) + payload['reply_channel'] = message.content['reply_channel'] + Channel("notify.receive").send(payload) + + + @channel_session_user + def set_notification_as_seen(message): + notification_id = message['notification_id'] + notif = Notification.objects.get(pk=notification_id) + notif.seen = timezone.now() + notif.save() + + + @channel_session_user + def user_connected(message): + username = message['username'] - Group(str(message.content['text'])).add(message.reply_channel) + Group(username).add(message.reply_channel) message.reply_channel.send({ - "text": message.content['text'], + "text": username, })
Update receive and send functions according to the new requirements
## Code Before: from channels import Group def ws_echo(message): Group(str(message.content['text'])).add(message.reply_channel) message.reply_channel.send({ "text": message.content['text'], }) ## Instruction: Update receive and send functions according to the new requirements ## Code After: import json from channels import Channel from channels import Group from channels.auth import channel_session_user, channel_session_user_from_http from django.utils import timezone from lino.modlib.notify.models import Notification # This decorator copies the user from the HTTP session (only available in # websocket.connect or http.request messages) to the channel session (available # in all consumers with the same reply_channel, so all three here) @channel_session_user_from_http def ws_connect(message): pass def ws_receive(message): # All WebSocket frames have either a text or binary payload; we decode the # text part here assuming it's JSON. # You could easily build up a basic framework that did this encoding/decoding # for you as well as handling common errors. payload = json.loads(message['text']) payload['reply_channel'] = message.content['reply_channel'] Channel("notify.receive").send(payload) @channel_session_user def set_notification_as_seen(message): notification_id = message['notification_id'] notif = Notification.objects.get(pk=notification_id) notif.seen = timezone.now() notif.save() @channel_session_user def user_connected(message): username = message['username'] Group(username).add(message.reply_channel) message.reply_channel.send({ "text": username, })
+ import json + + from channels import Channel from channels import Group + from channels.auth import channel_session_user, channel_session_user_from_http + from django.utils import timezone + from lino.modlib.notify.models import Notification + # This decorator copies the user from the HTTP session (only available in + # websocket.connect or http.request messages) to the channel session (available + # in all consumers with the same reply_channel, so all three here) + @channel_session_user_from_http - def ws_echo(message): ? ^^ + def ws_connect(message): ? ++++ ^ + pass + + + def ws_receive(message): + # All WebSocket frames have either a text or binary payload; we decode the + # text part here assuming it's JSON. + # You could easily build up a basic framework that did this encoding/decoding + # for you as well as handling common errors. + payload = json.loads(message['text']) + payload['reply_channel'] = message.content['reply_channel'] + Channel("notify.receive").send(payload) + + + @channel_session_user + def set_notification_as_seen(message): + notification_id = message['notification_id'] + notif = Notification.objects.get(pk=notification_id) + notif.seen = timezone.now() + notif.save() + + + @channel_session_user + def user_connected(message): + username = message['username'] - Group(str(message.content['text'])).add(message.reply_channel) ? ^ ^ ---------------------- + Group(username).add(message.reply_channel) ? + ^ ^^ message.reply_channel.send({ - "text": message.content['text'], + "text": username, })
50ab2ed3d8e50e5106dc486e4d20c889d6b18e82
spkg/base/package_database.py
spkg/base/package_database.py
from os.path import split, splitext from json import load f = open("packages.json") data = load(f) g = [] for p in data: pkg = { "name": p["name"], "dependencies": p["dependencies"], "version": p["version"], "download": p["download"], } g.append(pkg) from json import dump from StringIO import StringIO s = StringIO() dump(g, s, sort_keys=True, indent=4) s.seek(0) s = s.read() # Remove the trailing space s = s.replace(" \n", "\n") f = open("packages.json", "w") f.write(s)
from os.path import split, splitext from json import load f = open("packages.json") data = load(f) g = [] for p in data: pkg = { "name": p["name"], "dependencies": p["dependencies"], "version": p["version"], "download": p["download"], } g.append(pkg) from json import dump from StringIO import StringIO s = StringIO() dump(g, s, sort_keys=True, indent=4) s.seek(0) s = s.read() # Remove the trailing space s = s.replace(" \n", "\n") f = open("packages.json", "w") f.write(s) f.write("\n")
Add a new line at the end of the file
Add a new line at the end of the file
Python
bsd-3-clause
qsnake/qsnake,qsnake/qsnake
from os.path import split, splitext from json import load f = open("packages.json") data = load(f) g = [] for p in data: pkg = { "name": p["name"], "dependencies": p["dependencies"], "version": p["version"], "download": p["download"], } g.append(pkg) from json import dump from StringIO import StringIO s = StringIO() dump(g, s, sort_keys=True, indent=4) s.seek(0) s = s.read() # Remove the trailing space s = s.replace(" \n", "\n") f = open("packages.json", "w") f.write(s) + f.write("\n")
Add a new line at the end of the file
## Code Before: from os.path import split, splitext from json import load f = open("packages.json") data = load(f) g = [] for p in data: pkg = { "name": p["name"], "dependencies": p["dependencies"], "version": p["version"], "download": p["download"], } g.append(pkg) from json import dump from StringIO import StringIO s = StringIO() dump(g, s, sort_keys=True, indent=4) s.seek(0) s = s.read() # Remove the trailing space s = s.replace(" \n", "\n") f = open("packages.json", "w") f.write(s) ## Instruction: Add a new line at the end of the file ## Code After: from os.path import split, splitext from json import load f = open("packages.json") data = load(f) g = [] for p in data: pkg = { "name": p["name"], "dependencies": p["dependencies"], "version": p["version"], "download": p["download"], } g.append(pkg) from json import dump from StringIO import StringIO s = StringIO() dump(g, s, sort_keys=True, indent=4) s.seek(0) s = s.read() # Remove the trailing space s = s.replace(" \n", "\n") f = open("packages.json", "w") f.write(s) f.write("\n")
from os.path import split, splitext from json import load f = open("packages.json") data = load(f) g = [] for p in data: pkg = { "name": p["name"], "dependencies": p["dependencies"], "version": p["version"], "download": p["download"], } g.append(pkg) from json import dump from StringIO import StringIO s = StringIO() dump(g, s, sort_keys=True, indent=4) s.seek(0) s = s.read() # Remove the trailing space s = s.replace(" \n", "\n") f = open("packages.json", "w") f.write(s) + f.write("\n")
def66bc381f03970640a61d64b49ad5de9ef3879
ocaml/build-in.py
ocaml/build-in.py
import sys import os from os.path import relpath ocaml_build_dir = relpath(sys.argv[1], '.') os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
import sys import os from os.path import relpath ocaml_build_dir = relpath(sys.argv[1], '.') # Hack: when we can depend on a full OCaml feed with the build tools, we can remove this. # Until then, we need to avoid trying to compile against the limited runtime environment. if 'OCAMLLIB' in os.environ: del os.environ['OCAMLLIB'] os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
Remove OCAMLLIB from build environment
Remove OCAMLLIB from build environment This is a temporary hack: when we can depend on a full OCaml feed with the build tools, we can remove this. Until then, we need to avoid trying to compile against the limited runtime environment.
Python
lgpl-2.1
0install/0install,afb/0install,afb/0install,afb/0install,gasche/0install,bastianeicher/0install,bhilton/0install,fdopen/0install,gasche/0install,0install/0install,jaychoo/0install,dbenamy/0install,gfxmonk/0install,jaychoo/0install,dbenamy/0install,DarkGreising/0install,bastianeicher/0install,fdopen/0install,bhilton/0install,bhilton/0install,bartbes/0install,gasche/0install,bastianeicher/0install,dbenamy/0install,DarkGreising/0install,bartbes/0install,fdopen/0install,gasche/0install,HoMeCracKeR/0install,jaychoo/0install,pombreda/0install,gfxmonk/0install,bartbes/0install,afb/0install,HoMeCracKeR/0install,HoMeCracKeR/0install,gfxmonk/0install,DarkGreising/0install,pombreda/0install,pombreda/0install,0install/0install
import sys import os from os.path import relpath ocaml_build_dir = relpath(sys.argv[1], '.') + + # Hack: when we can depend on a full OCaml feed with the build tools, we can remove this. + # Until then, we need to avoid trying to compile against the limited runtime environment. + if 'OCAMLLIB' in os.environ: + del os.environ['OCAMLLIB'] + os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
Remove OCAMLLIB from build environment
## Code Before: import sys import os from os.path import relpath ocaml_build_dir = relpath(sys.argv[1], '.') os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"]) ## Instruction: Remove OCAMLLIB from build environment ## Code After: import sys import os from os.path import relpath ocaml_build_dir = relpath(sys.argv[1], '.') # Hack: when we can depend on a full OCaml feed with the build tools, we can remove this. # Until then, we need to avoid trying to compile against the limited runtime environment. if 'OCAMLLIB' in os.environ: del os.environ['OCAMLLIB'] os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
import sys import os from os.path import relpath ocaml_build_dir = relpath(sys.argv[1], '.') + + # Hack: when we can depend on a full OCaml feed with the build tools, we can remove this. + # Until then, we need to avoid trying to compile against the limited runtime environment. + if 'OCAMLLIB' in os.environ: + del os.environ['OCAMLLIB'] + os.execvp("make", ["make", 'OCAML_BUILDDIR=' + ocaml_build_dir, "ocaml"])
0aa5741ce05dcd4926be9c74af18f6fe46f4aded
etl_framework/utilities/DatetimeConverter.py
etl_framework/utilities/DatetimeConverter.py
"""class to convert datetime values""" import datetime class DatetimeConverter(object): """stuff""" _EPOCH_0 = datetime.datetime(1970, 1, 1) def __init__(self): """stuff""" pass @staticmethod def get_tomorrow(): """stuff""" return datetime.datetime.today() + datetime.timedelta(days=1) @classmethod def get_timestamp(cls, datetime_obj): """helper method to return timestamp fo datetime object""" return (datetime_obj - cls._EPOCH_0).total_seconds() @classmethod def get_tomorrow_timestamp(cls): """stuff""" return cls.get_timestamp(cls.get_tomorrow())
"""class to convert datetime values""" import datetime class DatetimeConverter(object): """stuff""" _EPOCH_0 = datetime.datetime(1970, 1, 1) def __init__(self): """stuff""" pass @staticmethod def get_tomorrow(): """stuff""" return datetime.datetime.today() + datetime.timedelta(days=1) @staticmethod def get_yesterday(): return datetime.datetime.today() - datetime.timedelta(days=1) @classmethod def get_timestamp(cls, datetime_obj): """helper method to return timestamp fo datetime object""" return (datetime_obj - cls._EPOCH_0).total_seconds() @classmethod def get_tomorrow_timestamp(cls): """stuff""" return cls.get_timestamp(cls.get_tomorrow()) @classmethod def get_yesterday_timestamp(cls): return cls.get_timestamp(cls.get_yesterday())
Add utility methods for yesterday's date
Add utility methods for yesterday's date
Python
mit
pantheon-systems/etl-framework
"""class to convert datetime values""" import datetime class DatetimeConverter(object): """stuff""" _EPOCH_0 = datetime.datetime(1970, 1, 1) def __init__(self): """stuff""" pass @staticmethod def get_tomorrow(): """stuff""" return datetime.datetime.today() + datetime.timedelta(days=1) + @staticmethod + def get_yesterday(): + + return datetime.datetime.today() - datetime.timedelta(days=1) + @classmethod def get_timestamp(cls, datetime_obj): """helper method to return timestamp fo datetime object""" return (datetime_obj - cls._EPOCH_0).total_seconds() @classmethod def get_tomorrow_timestamp(cls): """stuff""" return cls.get_timestamp(cls.get_tomorrow()) + @classmethod + def get_yesterday_timestamp(cls): + + return cls.get_timestamp(cls.get_yesterday()) +
Add utility methods for yesterday's date
## Code Before: """class to convert datetime values""" import datetime class DatetimeConverter(object): """stuff""" _EPOCH_0 = datetime.datetime(1970, 1, 1) def __init__(self): """stuff""" pass @staticmethod def get_tomorrow(): """stuff""" return datetime.datetime.today() + datetime.timedelta(days=1) @classmethod def get_timestamp(cls, datetime_obj): """helper method to return timestamp fo datetime object""" return (datetime_obj - cls._EPOCH_0).total_seconds() @classmethod def get_tomorrow_timestamp(cls): """stuff""" return cls.get_timestamp(cls.get_tomorrow()) ## Instruction: Add utility methods for yesterday's date ## Code After: """class to convert datetime values""" import datetime class DatetimeConverter(object): """stuff""" _EPOCH_0 = datetime.datetime(1970, 1, 1) def __init__(self): """stuff""" pass @staticmethod def get_tomorrow(): """stuff""" return datetime.datetime.today() + datetime.timedelta(days=1) @staticmethod def get_yesterday(): return datetime.datetime.today() - datetime.timedelta(days=1) @classmethod def get_timestamp(cls, datetime_obj): """helper method to return timestamp fo datetime object""" return (datetime_obj - cls._EPOCH_0).total_seconds() @classmethod def get_tomorrow_timestamp(cls): """stuff""" return cls.get_timestamp(cls.get_tomorrow()) @classmethod def get_yesterday_timestamp(cls): return cls.get_timestamp(cls.get_yesterday())
"""class to convert datetime values""" import datetime class DatetimeConverter(object): """stuff""" _EPOCH_0 = datetime.datetime(1970, 1, 1) def __init__(self): """stuff""" pass @staticmethod def get_tomorrow(): """stuff""" return datetime.datetime.today() + datetime.timedelta(days=1) + @staticmethod + def get_yesterday(): + + return datetime.datetime.today() - datetime.timedelta(days=1) + @classmethod def get_timestamp(cls, datetime_obj): """helper method to return timestamp fo datetime object""" return (datetime_obj - cls._EPOCH_0).total_seconds() @classmethod def get_tomorrow_timestamp(cls): """stuff""" return cls.get_timestamp(cls.get_tomorrow()) + + @classmethod + def get_yesterday_timestamp(cls): + + return cls.get_timestamp(cls.get_yesterday())
7079614f35de60def5f4e1cc1cb17cf3e5b4d9c6
setup.py
setup.py
from distutils.core import setup import os def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup( name='facebook-sdk', version='0.3.2', description='This client library is designed to support the Facebook ' 'Graph API and the official Facebook JavaScript SDK, which ' 'is the canonical way to implement Facebook authentication.', author='Facebook', maintainer='Martey Dodoo', maintainer_email='facebook-sdk@marteydodoo.com', url='https://github.com/pythonforfacebook/facebook-sdk', license='Apache', py_modules=[ 'facebook', ], long_description=read("README.rst"), classifiers=[ 'License :: OSI Approved :: Apache Software License', ], )
from distutils.core import setup setup( name='facebook-sdk', version='0.3.2', description='This client library is designed to support the Facebook ' 'Graph API and the official Facebook JavaScript SDK, which ' 'is the canonical way to implement Facebook authentication.', author='Facebook', maintainer='Martey Dodoo', maintainer_email='facebook-sdk@marteydodoo.com', url='https://github.com/pythonforfacebook/facebook-sdk', license='Apache', py_modules=[ 'facebook', ], long_description=open("README.rst").read(), classifiers=[ 'License :: OSI Approved :: Apache Software License', ], )
Change way README is imported.
Change way README is imported. The custom read function is unnecessary since only one file is being accessed. Removing it reduces the amount of code.
Python
apache-2.0
Aloomaio/facebook-sdk,mobolic/facebook-sdk
from distutils.core import setup - import os - - def read(fname): - return open(os.path.join(os.path.dirname(__file__), fname)).read() - setup( name='facebook-sdk', version='0.3.2', description='This client library is designed to support the Facebook ' 'Graph API and the official Facebook JavaScript SDK, which ' 'is the canonical way to implement Facebook authentication.', author='Facebook', maintainer='Martey Dodoo', maintainer_email='facebook-sdk@marteydodoo.com', url='https://github.com/pythonforfacebook/facebook-sdk', license='Apache', py_modules=[ 'facebook', ], - long_description=read("README.rst"), + long_description=open("README.rst").read(), classifiers=[ 'License :: OSI Approved :: Apache Software License', ], )
Change way README is imported.
## Code Before: from distutils.core import setup import os def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup( name='facebook-sdk', version='0.3.2', description='This client library is designed to support the Facebook ' 'Graph API and the official Facebook JavaScript SDK, which ' 'is the canonical way to implement Facebook authentication.', author='Facebook', maintainer='Martey Dodoo', maintainer_email='facebook-sdk@marteydodoo.com', url='https://github.com/pythonforfacebook/facebook-sdk', license='Apache', py_modules=[ 'facebook', ], long_description=read("README.rst"), classifiers=[ 'License :: OSI Approved :: Apache Software License', ], ) ## Instruction: Change way README is imported. ## Code After: from distutils.core import setup setup( name='facebook-sdk', version='0.3.2', description='This client library is designed to support the Facebook ' 'Graph API and the official Facebook JavaScript SDK, which ' 'is the canonical way to implement Facebook authentication.', author='Facebook', maintainer='Martey Dodoo', maintainer_email='facebook-sdk@marteydodoo.com', url='https://github.com/pythonforfacebook/facebook-sdk', license='Apache', py_modules=[ 'facebook', ], long_description=open("README.rst").read(), classifiers=[ 'License :: OSI Approved :: Apache Software License', ], )
from distutils.core import setup - import os - - def read(fname): - return open(os.path.join(os.path.dirname(__file__), fname)).read() - setup( name='facebook-sdk', version='0.3.2', description='This client library is designed to support the Facebook ' 'Graph API and the official Facebook JavaScript SDK, which ' 'is the canonical way to implement Facebook authentication.', author='Facebook', maintainer='Martey Dodoo', maintainer_email='facebook-sdk@marteydodoo.com', url='https://github.com/pythonforfacebook/facebook-sdk', license='Apache', py_modules=[ 'facebook', ], - long_description=read("README.rst"), ? ^ ^^ + long_description=open("README.rst").read(), ? ^^ ^ +++++++ classifiers=[ 'License :: OSI Approved :: Apache Software License', ], )
d8cc0fdaea848ed5d626ba6ba4292fd3cb906da3
project7/TrackParser.py
project7/TrackParser.py
""" Created by Max 12/2/2017 """ import pprint import numpy as np class TrackParser: @staticmethod def parse_track(path_to_track_file: str) -> np.ndarray: track = None with open(path_to_track_file, 'r') as track_file: lines = track_file.readlines() dimensions_str = lines[0] dims = dimensions_str.split(',') track = np.zeros((int(dims[0]), int(dims[1]))) for line_index in range(1, len(lines)): line = lines[line_index] for char_index in range(len(line) - 1): track_value = TrackParser.get_char_value(line[char_index]) track[line_index-1][char_index] = track_value return track @staticmethod def get_char_value(char: str): if char == '#': return -1 elif char == '.': return 0 elif char == 'S': return 1 elif char == 'F': return 2 else: return -1 np.set_printoptions(linewidth=500) pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
""" Created by Max 12/2/2017 """ import pprint import numpy as np class TrackParser: @staticmethod def parse_track(path_to_track_file: str) -> np.ndarray: track = None with open(path_to_track_file, 'r') as track_file: lines = track_file.readlines() dimensions_str = lines[0] dims = dimensions_str.split(',') # (Y,X) coords. track = np.zeros((int(dims[0]), int(dims[1]))) for line_index in range(1, len(lines)): line = lines[line_index] for char_index in range(len(line) - 1): track_value = TrackParser.get_char_value(line[char_index]) track[line_index-1][char_index] = track_value return track @staticmethod def get_char_value(char: str): if char == '#': return -1 elif char == '.': return 0 elif char == 'S': return 1 elif char == 'F': return 2 else: return -1 # np.set_printoptions(linewidth=500) # pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
Add comment to clarify coordinate order
Add comment to clarify coordinate order coordinates are (y,x) in the track.
Python
apache-2.0
MaxRobinson/CS449,MaxRobinson/CS449,MaxRobinson/CS449
""" Created by Max 12/2/2017 """ import pprint import numpy as np class TrackParser: @staticmethod def parse_track(path_to_track_file: str) -> np.ndarray: track = None with open(path_to_track_file, 'r') as track_file: lines = track_file.readlines() dimensions_str = lines[0] dims = dimensions_str.split(',') + # (Y,X) coords. track = np.zeros((int(dims[0]), int(dims[1]))) for line_index in range(1, len(lines)): line = lines[line_index] for char_index in range(len(line) - 1): track_value = TrackParser.get_char_value(line[char_index]) track[line_index-1][char_index] = track_value return track @staticmethod def get_char_value(char: str): if char == '#': return -1 elif char == '.': return 0 elif char == 'S': return 1 elif char == 'F': return 2 else: return -1 - np.set_printoptions(linewidth=500) + # np.set_printoptions(linewidth=500) - pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500) + # pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
Add comment to clarify coordinate order
## Code Before: """ Created by Max 12/2/2017 """ import pprint import numpy as np class TrackParser: @staticmethod def parse_track(path_to_track_file: str) -> np.ndarray: track = None with open(path_to_track_file, 'r') as track_file: lines = track_file.readlines() dimensions_str = lines[0] dims = dimensions_str.split(',') track = np.zeros((int(dims[0]), int(dims[1]))) for line_index in range(1, len(lines)): line = lines[line_index] for char_index in range(len(line) - 1): track_value = TrackParser.get_char_value(line[char_index]) track[line_index-1][char_index] = track_value return track @staticmethod def get_char_value(char: str): if char == '#': return -1 elif char == '.': return 0 elif char == 'S': return 1 elif char == 'F': return 2 else: return -1 np.set_printoptions(linewidth=500) pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500) ## Instruction: Add comment to clarify coordinate order ## Code After: """ Created by Max 12/2/2017 """ import pprint import numpy as np class TrackParser: @staticmethod def parse_track(path_to_track_file: str) -> np.ndarray: track = None with open(path_to_track_file, 'r') as track_file: lines = track_file.readlines() dimensions_str = lines[0] dims = dimensions_str.split(',') # (Y,X) coords. track = np.zeros((int(dims[0]), int(dims[1]))) for line_index in range(1, len(lines)): line = lines[line_index] for char_index in range(len(line) - 1): track_value = TrackParser.get_char_value(line[char_index]) track[line_index-1][char_index] = track_value return track @staticmethod def get_char_value(char: str): if char == '#': return -1 elif char == '.': return 0 elif char == 'S': return 1 elif char == 'F': return 2 else: return -1 # np.set_printoptions(linewidth=500) # pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500)
""" Created by Max 12/2/2017 """ import pprint import numpy as np class TrackParser: @staticmethod def parse_track(path_to_track_file: str) -> np.ndarray: track = None with open(path_to_track_file, 'r') as track_file: lines = track_file.readlines() dimensions_str = lines[0] dims = dimensions_str.split(',') + # (Y,X) coords. track = np.zeros((int(dims[0]), int(dims[1]))) for line_index in range(1, len(lines)): line = lines[line_index] for char_index in range(len(line) - 1): track_value = TrackParser.get_char_value(line[char_index]) track[line_index-1][char_index] = track_value return track @staticmethod def get_char_value(char: str): if char == '#': return -1 elif char == '.': return 0 elif char == 'S': return 1 elif char == 'F': return 2 else: return -1 - np.set_printoptions(linewidth=500) + # np.set_printoptions(linewidth=500) ? ++ - pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500) + # pprint.pprint(TrackParser.parse_track("tracks/L-track.txt"), width=500) ? ++
e1ffdcc5f12be623633e2abab2041fcb574173ea
homeassistant/components/zeroconf.py
homeassistant/components/zeroconf.py
import logging import socket from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__) REQUIREMENTS = ["zeroconf==0.17.5"] _LOGGER = logging.getLogger(__name__) DOMAIN = "zeroconf" ZEROCONF_TYPE = "_home-assistant._tcp.local." DEPENDENCIES = ["http"] def setup(hass, config): """Set up Zeroconf and make Home Assistant discoverable.""" from zeroconf import Zeroconf, ServiceInfo zeroconf = Zeroconf() zeroconf_name = "{}.{}".format(hass.config.location_name, ZEROCONF_TYPE) params = {"version": __version__, "base_url": hass.http.base_url, "needs_auth": (hass.http.api_password != "")} info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name, socket.inet_aton(hass.http.routable_address), hass.http.server_address[1], 0, 0, params) zeroconf.register_service(info) def stop_zeroconf(event): """Stop Zeroconf.""" zeroconf.unregister_service(info) hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf) return True
import logging import socket from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__) REQUIREMENTS = ["zeroconf==0.17.5"] DEPENDENCIES = ["api"] _LOGGER = logging.getLogger(__name__) DOMAIN = "zeroconf" ZEROCONF_TYPE = "_home-assistant._tcp.local." def setup(hass, config): """Set up Zeroconf and make Home Assistant discoverable.""" from zeroconf import Zeroconf, ServiceInfo zeroconf = Zeroconf() zeroconf_name = "{}.{}".format(hass.config.location_name, ZEROCONF_TYPE) params = {"version": __version__, "base_url": hass.config.api.base_url, "needs_auth": (hass.config.api.api_password != "")} info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name, socket.inet_aton(hass.config.api.host), hass.config.api.port, 0, 0, params) zeroconf.register_service(info) def stop_zeroconf(event): """Stop Zeroconf.""" zeroconf.unregister_service(info) hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf) return True
Use hass.config.api instead of hass.http
Use hass.config.api instead of hass.http
Python
mit
miniconfig/home-assistant,Julian/home-assistant,toddeye/home-assistant,ct-23/home-assistant,deisi/home-assistant,tchellomello/home-assistant,rohitranjan1991/home-assistant,Julian/home-assistant,Duoxilian/home-assistant,betrisey/home-assistant,keerts/home-assistant,ct-23/home-assistant,tboyce021/home-assistant,kyvinh/home-assistant,leoc/home-assistant,jawilson/home-assistant,JshWright/home-assistant,turbokongen/home-assistant,varunr047/homefile,open-homeautomation/home-assistant,betrisey/home-assistant,GenericStudent/home-assistant,alexmogavero/home-assistant,fbradyirl/home-assistant,eagleamon/home-assistant,kyvinh/home-assistant,toddeye/home-assistant,auduny/home-assistant,aronsky/home-assistant,hmronline/home-assistant,bdfoster/blumate,jaharkes/home-assistant,fbradyirl/home-assistant,srcLurker/home-assistant,morphis/home-assistant,deisi/home-assistant,LinuxChristian/home-assistant,hexxter/home-assistant,sdague/home-assistant,LinuxChristian/home-assistant,shaftoe/home-assistant,stefan-jonasson/home-assistant,adrienbrault/home-assistant,w1ll1am23/home-assistant,tinloaf/home-assistant,soldag/home-assistant,molobrakos/home-assistant,Zyell/home-assistant,HydrelioxGitHub/home-assistant,keerts/home-assistant,molobrakos/home-assistant,varunr047/homefile,jaharkes/home-assistant,xifle/home-assistant,Smart-Torvy/torvy-home-assistant,stefan-jonasson/home-assistant,tboyce1/home-assistant,kennedyshead/home-assistant,sffjunkie/home-assistant,kennedyshead/home-assistant,jnewland/home-assistant,tinloaf/home-assistant,dmeulen/home-assistant,xifle/home-assistant,mikaelboman/home-assistant,Zac-HD/home-assistant,hmronline/home-assistant,partofthething/home-assistant,robjohnson189/home-assistant,lukas-hetzenecker/home-assistant,nugget/home-assistant,aequitas/home-assistant,oandrew/home-assistant,morphis/home-assistant,eagleamon/home-assistant,HydrelioxGitHub/home-assistant,Teagan42/home-assistant,ct-23/home-assistant,Zyell/home-assistant,sffjunkie/home-assistant,tinloaf/home-assistant,philipbl/home-assistant,kyvinh/home-assistant,Julian/home-assistant,PetePriority/home-assistant,sffjunkie/home-assistant,bdfoster/blumate,Teagan42/home-assistant,rohitranjan1991/home-assistant,deisi/home-assistant,miniconfig/home-assistant,robjohnson189/home-assistant,happyleavesaoc/home-assistant,morphis/home-assistant,GenericStudent/home-assistant,soldag/home-assistant,deisi/home-assistant,Duoxilian/home-assistant,persandstrom/home-assistant,w1ll1am23/home-assistant,oandrew/home-assistant,mezz64/home-assistant,DavidLP/home-assistant,morphis/home-assistant,tboyce1/home-assistant,florianholzapfel/home-assistant,Danielhiversen/home-assistant,home-assistant/home-assistant,florianholzapfel/home-assistant,srcLurker/home-assistant,stefan-jonasson/home-assistant,Smart-Torvy/torvy-home-assistant,sdague/home-assistant,kyvinh/home-assistant,balloob/home-assistant,happyleavesaoc/home-assistant,MartinHjelmare/home-assistant,happyleavesaoc/home-assistant,adrienbrault/home-assistant,ewandor/home-assistant,nugget/home-assistant,xifle/home-assistant,MungoRae/home-assistant,robbiet480/home-assistant,MartinHjelmare/home-assistant,alexmogavero/home-assistant,leppa/home-assistant,open-homeautomation/home-assistant,leoc/home-assistant,LinuxChristian/home-assistant,leoc/home-assistant,MartinHjelmare/home-assistant,Julian/home-assistant,miniconfig/home-assistant,lukas-hetzenecker/home-assistant,mikaelboman/home-assistant,joopert/home-assistant,Zac-HD/home-assistant,auduny/home-assistant,Zac-HD/home-assistant,joopert/home-assistant,ma314smith/home-assistant,Zyell/home-assistant,devdelay/home-assistant,srcLurker/home-assistant,hexxter/home-assistant,HydrelioxGitHub/home-assistant,mikaelboman/home-assistant,aequitas/home-assistant,qedi-r/home-assistant,nkgilley/home-assistant,jnewland/home-assistant,deisi/home-assistant,leoc/home-assistant,jaharkes/home-assistant,varunr047/homefile,nugget/home-assistant,hexxter/home-assistant,sander76/home-assistant,fbradyirl/home-assistant,partofthething/home-assistant,jamespcole/home-assistant,persandstrom/home-assistant,jnewland/home-assistant,balloob/home-assistant,leppa/home-assistant,bdfoster/blumate,keerts/home-assistant,shaftoe/home-assistant,pschmitt/home-assistant,philipbl/home-assistant,tboyce1/home-assistant,betrisey/home-assistant,hmronline/home-assistant,sander76/home-assistant,xifle/home-assistant,JshWright/home-assistant,oandrew/home-assistant,PetePriority/home-assistant,varunr047/homefile,dmeulen/home-assistant,tboyce021/home-assistant,alexmogavero/home-assistant,DavidLP/home-assistant,persandstrom/home-assistant,LinuxChristian/home-assistant,philipbl/home-assistant,JshWright/home-assistant,FreekingDean/home-assistant,robbiet480/home-assistant,jabesq/home-assistant,ct-23/home-assistant,hmronline/home-assistant,ma314smith/home-assistant,Smart-Torvy/torvy-home-assistant,JshWright/home-assistant,home-assistant/home-assistant,shaftoe/home-assistant,devdelay/home-assistant,emilhetty/home-assistant,emilhetty/home-assistant,Smart-Torvy/torvy-home-assistant,mKeRix/home-assistant,mKeRix/home-assistant,Danielhiversen/home-assistant,stefan-jonasson/home-assistant,bdfoster/blumate,sffjunkie/home-assistant,bdfoster/blumate,jawilson/home-assistant,florianholzapfel/home-assistant,MungoRae/home-assistant,jaharkes/home-assistant,Cinntax/home-assistant,aronsky/home-assistant,auduny/home-assistant,robjohnson189/home-assistant,florianholzapfel/home-assistant,emilhetty/home-assistant,devdelay/home-assistant,sffjunkie/home-assistant,MungoRae/home-assistant,mikaelboman/home-assistant,Duoxilian/home-assistant,Duoxilian/home-assistant,dmeulen/home-assistant,betrisey/home-assistant,ma314smith/home-assistant,open-homeautomation/home-assistant,DavidLP/home-assistant,robjohnson189/home-assistant,dmeulen/home-assistant,mikaelboman/home-assistant,jamespcole/home-assistant,philipbl/home-assistant,jabesq/home-assistant,emilhetty/home-assistant,oandrew/home-assistant,rohitranjan1991/home-assistant,MungoRae/home-assistant,mKeRix/home-assistant,shaftoe/home-assistant,titilambert/home-assistant,MungoRae/home-assistant,varunr047/homefile,balloob/home-assistant,molobrakos/home-assistant,postlund/home-assistant,happyleavesaoc/home-assistant,mezz64/home-assistant,Zac-HD/home-assistant,ct-23/home-assistant,open-homeautomation/home-assistant,ma314smith/home-assistant,FreekingDean/home-assistant,mKeRix/home-assistant,qedi-r/home-assistant,miniconfig/home-assistant,ewandor/home-assistant,srcLurker/home-assistant,ewandor/home-assistant,PetePriority/home-assistant,postlund/home-assistant,jabesq/home-assistant,alexmogavero/home-assistant,aequitas/home-assistant,devdelay/home-assistant,hexxter/home-assistant,Cinntax/home-assistant,titilambert/home-assistant,emilhetty/home-assistant,tchellomello/home-assistant,keerts/home-assistant,LinuxChristian/home-assistant,eagleamon/home-assistant,nkgilley/home-assistant,hmronline/home-assistant,jamespcole/home-assistant,pschmitt/home-assistant,tboyce1/home-assistant,eagleamon/home-assistant,turbokongen/home-assistant
import logging import socket from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__) REQUIREMENTS = ["zeroconf==0.17.5"] + DEPENDENCIES = ["api"] + _LOGGER = logging.getLogger(__name__) DOMAIN = "zeroconf" ZEROCONF_TYPE = "_home-assistant._tcp.local." - - DEPENDENCIES = ["http"] def setup(hass, config): """Set up Zeroconf and make Home Assistant discoverable.""" from zeroconf import Zeroconf, ServiceInfo zeroconf = Zeroconf() zeroconf_name = "{}.{}".format(hass.config.location_name, ZEROCONF_TYPE) - params = {"version": __version__, "base_url": hass.http.base_url, + params = {"version": __version__, "base_url": hass.config.api.base_url, - "needs_auth": (hass.http.api_password != "")} + "needs_auth": (hass.config.api.api_password != "")} info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name, - socket.inet_aton(hass.http.routable_address), + socket.inet_aton(hass.config.api.host), - hass.http.server_address[1], 0, 0, params) + hass.config.api.port, 0, 0, params) zeroconf.register_service(info) def stop_zeroconf(event): """Stop Zeroconf.""" zeroconf.unregister_service(info) hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf) return True
Use hass.config.api instead of hass.http
## Code Before: import logging import socket from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__) REQUIREMENTS = ["zeroconf==0.17.5"] _LOGGER = logging.getLogger(__name__) DOMAIN = "zeroconf" ZEROCONF_TYPE = "_home-assistant._tcp.local." DEPENDENCIES = ["http"] def setup(hass, config): """Set up Zeroconf and make Home Assistant discoverable.""" from zeroconf import Zeroconf, ServiceInfo zeroconf = Zeroconf() zeroconf_name = "{}.{}".format(hass.config.location_name, ZEROCONF_TYPE) params = {"version": __version__, "base_url": hass.http.base_url, "needs_auth": (hass.http.api_password != "")} info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name, socket.inet_aton(hass.http.routable_address), hass.http.server_address[1], 0, 0, params) zeroconf.register_service(info) def stop_zeroconf(event): """Stop Zeroconf.""" zeroconf.unregister_service(info) hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf) return True ## Instruction: Use hass.config.api instead of hass.http ## Code After: import logging import socket from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__) REQUIREMENTS = ["zeroconf==0.17.5"] DEPENDENCIES = ["api"] _LOGGER = logging.getLogger(__name__) DOMAIN = "zeroconf" ZEROCONF_TYPE = "_home-assistant._tcp.local." def setup(hass, config): """Set up Zeroconf and make Home Assistant discoverable.""" from zeroconf import Zeroconf, ServiceInfo zeroconf = Zeroconf() zeroconf_name = "{}.{}".format(hass.config.location_name, ZEROCONF_TYPE) params = {"version": __version__, "base_url": hass.config.api.base_url, "needs_auth": (hass.config.api.api_password != "")} info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name, socket.inet_aton(hass.config.api.host), hass.config.api.port, 0, 0, params) zeroconf.register_service(info) def stop_zeroconf(event): """Stop Zeroconf.""" zeroconf.unregister_service(info) hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf) return True
import logging import socket from homeassistant.const import (EVENT_HOMEASSISTANT_STOP, __version__) REQUIREMENTS = ["zeroconf==0.17.5"] + DEPENDENCIES = ["api"] + _LOGGER = logging.getLogger(__name__) DOMAIN = "zeroconf" ZEROCONF_TYPE = "_home-assistant._tcp.local." - - DEPENDENCIES = ["http"] def setup(hass, config): """Set up Zeroconf and make Home Assistant discoverable.""" from zeroconf import Zeroconf, ServiceInfo zeroconf = Zeroconf() zeroconf_name = "{}.{}".format(hass.config.location_name, ZEROCONF_TYPE) - params = {"version": __version__, "base_url": hass.http.base_url, ? ^^^ + params = {"version": __version__, "base_url": hass.config.api.base_url, ? ^^^^^^^^ + - "needs_auth": (hass.http.api_password != "")} ? ^^^ + "needs_auth": (hass.config.api.api_password != "")} ? ^^^^^^^^ + info = ServiceInfo(ZEROCONF_TYPE, zeroconf_name, - socket.inet_aton(hass.http.routable_address), ? ------------------- + socket.inet_aton(hass.config.api.host), ? +++++++++++ ++ - hass.http.server_address[1], 0, 0, params) + hass.config.api.port, 0, 0, params) zeroconf.register_service(info) def stop_zeroconf(event): """Stop Zeroconf.""" zeroconf.unregister_service(info) hass.bus.listen_once(EVENT_HOMEASSISTANT_STOP, stop_zeroconf) return True
69b0e1c60eafff596ebb494a7e79a22c6bea374b
polling_stations/apps/data_collection/management/commands/import_hart.py
polling_stations/apps/data_collection/management/commands/import_hart.py
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = 'E07000089' addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' elections = ['parl.2017-06-08'] csv_delimiter = '\t'
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = 'E07000089' addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' elections = ['parl.2017-06-08'] csv_delimiter = '\t' def station_record_to_dict(self, record): if record.polling_place_id == '1914': record = record._replace(polling_place_easting = '479224') record = record._replace(polling_place_northing = '154016') return super().station_record_to_dict(record)
Fix dodgy point in Hart
Fix dodgy point in Hart
Python
bsd-3-clause
DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations,chris48s/UK-Polling-Stations,chris48s/UK-Polling-Stations,chris48s/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = 'E07000089' addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' elections = ['parl.2017-06-08'] csv_delimiter = '\t' + def station_record_to_dict(self, record): + + if record.polling_place_id == '1914': + record = record._replace(polling_place_easting = '479224') + record = record._replace(polling_place_northing = '154016') + + return super().station_record_to_dict(record) +
Fix dodgy point in Hart
## Code Before: from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = 'E07000089' addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' elections = ['parl.2017-06-08'] csv_delimiter = '\t' ## Instruction: Fix dodgy point in Hart ## Code After: from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = 'E07000089' addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' elections = ['parl.2017-06-08'] csv_delimiter = '\t' def station_record_to_dict(self, record): if record.polling_place_id == '1914': record = record._replace(polling_place_easting = '479224') record = record._replace(polling_place_northing = '154016') return super().station_record_to_dict(record)
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = 'E07000089' addresses_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' stations_name = 'parl.2017-06-08/Version 1/Hart DC General Election polling place 120517.TSV' elections = ['parl.2017-06-08'] csv_delimiter = '\t' + + def station_record_to_dict(self, record): + + if record.polling_place_id == '1914': + record = record._replace(polling_place_easting = '479224') + record = record._replace(polling_place_northing = '154016') + + return super().station_record_to_dict(record)
55e316a45256d054d19425015ef13868a84c5ff1
src/pip/_internal/resolution/resolvelib/reporter.py
src/pip/_internal/resolution/resolvelib/reporter.py
from collections import defaultdict from logging import getLogger from pip._vendor.resolvelib.reporters import BaseReporter from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import DefaultDict from .base import Candidate logger = getLogger(__name__) class PipReporter(BaseReporter): def __init__(self): # type: () -> None self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int] self._messages_at_backtrack = { 1: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 8: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 13: ( "This is taking longer than usual. You might need to provide the " "dependency resolver with stricter constraints to reduce runtime." "If you want to abort this run, you can press Ctrl + C to do so." ) } def backtracking(self, candidate): # type: (Candidate) -> None self.backtracks_by_package[candidate.name] += 1 count = self.backtracks_by_package[candidate.name] if count not in self._messages_at_backtrack: return message = self._messages_at_backtrack[count] logger.info("INFO: %s", message)
from collections import defaultdict from logging import getLogger from pip._vendor.resolvelib.reporters import BaseReporter from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import DefaultDict from .base import Candidate logger = getLogger(__name__) class PipReporter(BaseReporter): def __init__(self): # type: () -> None self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int] self._messages_at_backtrack = { 1: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 8: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 13: ( "This is taking longer than usual. You might need to provide the " "dependency resolver with stricter constraints to reduce runtime." "If you want to abort this run, you can press Ctrl + C to do so." "To improve how pip performs, tell us that this happened here: " "https://pip.pypa.io/surveys/backtracking" ) } def backtracking(self, candidate): # type: (Candidate) -> None self.backtracks_by_package[candidate.name] += 1 count = self.backtracks_by_package[candidate.name] if count not in self._messages_at_backtrack: return message = self._messages_at_backtrack[count] logger.info("INFO: %s", message)
Add the last line to the info message
Add the last line to the info message
Python
mit
sbidoul/pip,pradyunsg/pip,pypa/pip,pypa/pip,sbidoul/pip,pfmoore/pip,pfmoore/pip,pradyunsg/pip
from collections import defaultdict from logging import getLogger from pip._vendor.resolvelib.reporters import BaseReporter from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import DefaultDict from .base import Candidate logger = getLogger(__name__) class PipReporter(BaseReporter): def __init__(self): # type: () -> None self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int] self._messages_at_backtrack = { 1: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 8: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 13: ( "This is taking longer than usual. You might need to provide the " "dependency resolver with stricter constraints to reduce runtime." "If you want to abort this run, you can press Ctrl + C to do so." + "To improve how pip performs, tell us that this happened here: " + "https://pip.pypa.io/surveys/backtracking" ) } def backtracking(self, candidate): # type: (Candidate) -> None self.backtracks_by_package[candidate.name] += 1 count = self.backtracks_by_package[candidate.name] if count not in self._messages_at_backtrack: return message = self._messages_at_backtrack[count] logger.info("INFO: %s", message)
Add the last line to the info message
## Code Before: from collections import defaultdict from logging import getLogger from pip._vendor.resolvelib.reporters import BaseReporter from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import DefaultDict from .base import Candidate logger = getLogger(__name__) class PipReporter(BaseReporter): def __init__(self): # type: () -> None self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int] self._messages_at_backtrack = { 1: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 8: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 13: ( "This is taking longer than usual. You might need to provide the " "dependency resolver with stricter constraints to reduce runtime." "If you want to abort this run, you can press Ctrl + C to do so." ) } def backtracking(self, candidate): # type: (Candidate) -> None self.backtracks_by_package[candidate.name] += 1 count = self.backtracks_by_package[candidate.name] if count not in self._messages_at_backtrack: return message = self._messages_at_backtrack[count] logger.info("INFO: %s", message) ## Instruction: Add the last line to the info message ## Code After: from collections import defaultdict from logging import getLogger from pip._vendor.resolvelib.reporters import BaseReporter from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import DefaultDict from .base import Candidate logger = getLogger(__name__) class PipReporter(BaseReporter): def __init__(self): # type: () -> None self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int] self._messages_at_backtrack = { 1: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 8: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 13: ( "This is taking longer than usual. You might need to provide the " "dependency resolver with stricter constraints to reduce runtime." "If you want to abort this run, you can press Ctrl + C to do so." "To improve how pip performs, tell us that this happened here: " "https://pip.pypa.io/surveys/backtracking" ) } def backtracking(self, candidate): # type: (Candidate) -> None self.backtracks_by_package[candidate.name] += 1 count = self.backtracks_by_package[candidate.name] if count not in self._messages_at_backtrack: return message = self._messages_at_backtrack[count] logger.info("INFO: %s", message)
from collections import defaultdict from logging import getLogger from pip._vendor.resolvelib.reporters import BaseReporter from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import DefaultDict from .base import Candidate logger = getLogger(__name__) class PipReporter(BaseReporter): def __init__(self): # type: () -> None self.backtracks_by_package = defaultdict(int) # type: DefaultDict[str, int] self._messages_at_backtrack = { 1: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 8: ( "pip is looking at multiple versions of this package to determine " "which version is compatible with other requirements. " "This could take a while." ), 13: ( "This is taking longer than usual. You might need to provide the " "dependency resolver with stricter constraints to reduce runtime." "If you want to abort this run, you can press Ctrl + C to do so." + "To improve how pip performs, tell us that this happened here: " + "https://pip.pypa.io/surveys/backtracking" ) } def backtracking(self, candidate): # type: (Candidate) -> None self.backtracks_by_package[candidate.name] += 1 count = self.backtracks_by_package[candidate.name] if count not in self._messages_at_backtrack: return message = self._messages_at_backtrack[count] logger.info("INFO: %s", message)
d81a6930d21262464ee06ae8afb51b65920f378c
tap/tests/test_pytest_plugin.py
tap/tests/test_pytest_plugin.py
try: from unittest import mock except ImportError: import mock from tap.plugins import pytest from tap.tests import TestCase from tap.tracker import Tracker class TestPytestPlugin(TestCase): def setUp(self): """The pytest plugin uses module scope so a fresh tracker must be installed each time.""" pytest.tracker = Tracker() def test_includes_options(self): group = mock.Mock() parser = mock.Mock() parser.getgroup.return_value = group pytest.pytest_addoption(parser) self.assertEqual(group.addoption.call_count, 1) def test_tracker_outdir_set(self): config = mock.Mock() config.option.tap_outdir = 'fakeout' pytest.pytest_configure(config) self.assertEqual(pytest.tracker.outdir, 'fakeout')
try: from unittest import mock except ImportError: import mock import tempfile from tap.plugins import pytest from tap.tests import TestCase from tap.tracker import Tracker class TestPytestPlugin(TestCase): def setUp(self): """The pytest plugin uses module scope so a fresh tracker must be installed each time.""" pytest.tracker = Tracker() def test_includes_options(self): group = mock.Mock() parser = mock.Mock() parser.getgroup.return_value = group pytest.pytest_addoption(parser) self.assertEqual(group.addoption.call_count, 1) def test_tracker_outdir_set(self): outdir = tempfile.mkdtemp() config = mock.Mock() config.option.tap_outdir = outdir pytest.pytest_configure(config) self.assertEqual(pytest.tracker.outdir, outdir)
Fix test to not create a new directory in the project.
Fix test to not create a new directory in the project.
Python
bsd-2-clause
mblayman/tappy,python-tap/tappy,Mark-E-Hamilton/tappy
try: from unittest import mock except ImportError: import mock + import tempfile from tap.plugins import pytest from tap.tests import TestCase from tap.tracker import Tracker class TestPytestPlugin(TestCase): def setUp(self): """The pytest plugin uses module scope so a fresh tracker must be installed each time.""" pytest.tracker = Tracker() def test_includes_options(self): group = mock.Mock() parser = mock.Mock() parser.getgroup.return_value = group pytest.pytest_addoption(parser) self.assertEqual(group.addoption.call_count, 1) def test_tracker_outdir_set(self): + outdir = tempfile.mkdtemp() config = mock.Mock() - config.option.tap_outdir = 'fakeout' + config.option.tap_outdir = outdir pytest.pytest_configure(config) - self.assertEqual(pytest.tracker.outdir, 'fakeout') + self.assertEqual(pytest.tracker.outdir, outdir)
Fix test to not create a new directory in the project.
## Code Before: try: from unittest import mock except ImportError: import mock from tap.plugins import pytest from tap.tests import TestCase from tap.tracker import Tracker class TestPytestPlugin(TestCase): def setUp(self): """The pytest plugin uses module scope so a fresh tracker must be installed each time.""" pytest.tracker = Tracker() def test_includes_options(self): group = mock.Mock() parser = mock.Mock() parser.getgroup.return_value = group pytest.pytest_addoption(parser) self.assertEqual(group.addoption.call_count, 1) def test_tracker_outdir_set(self): config = mock.Mock() config.option.tap_outdir = 'fakeout' pytest.pytest_configure(config) self.assertEqual(pytest.tracker.outdir, 'fakeout') ## Instruction: Fix test to not create a new directory in the project. ## Code After: try: from unittest import mock except ImportError: import mock import tempfile from tap.plugins import pytest from tap.tests import TestCase from tap.tracker import Tracker class TestPytestPlugin(TestCase): def setUp(self): """The pytest plugin uses module scope so a fresh tracker must be installed each time.""" pytest.tracker = Tracker() def test_includes_options(self): group = mock.Mock() parser = mock.Mock() parser.getgroup.return_value = group pytest.pytest_addoption(parser) self.assertEqual(group.addoption.call_count, 1) def test_tracker_outdir_set(self): outdir = tempfile.mkdtemp() config = mock.Mock() config.option.tap_outdir = outdir pytest.pytest_configure(config) self.assertEqual(pytest.tracker.outdir, outdir)
try: from unittest import mock except ImportError: import mock + import tempfile from tap.plugins import pytest from tap.tests import TestCase from tap.tracker import Tracker class TestPytestPlugin(TestCase): def setUp(self): """The pytest plugin uses module scope so a fresh tracker must be installed each time.""" pytest.tracker = Tracker() def test_includes_options(self): group = mock.Mock() parser = mock.Mock() parser.getgroup.return_value = group pytest.pytest_addoption(parser) self.assertEqual(group.addoption.call_count, 1) def test_tracker_outdir_set(self): + outdir = tempfile.mkdtemp() config = mock.Mock() - config.option.tap_outdir = 'fakeout' ? ----- ^ + config.option.tap_outdir = outdir ? ^^^ pytest.pytest_configure(config) - self.assertEqual(pytest.tracker.outdir, 'fakeout') ? ----- ^ + self.assertEqual(pytest.tracker.outdir, outdir) ? ^^^
b56c5ca12f9806ecedc531e1f00ec1d7f2162b46
src-django/authentication/urls.py
src-django/authentication/urls.py
from django.conf.urls import url from views import login, logout, signup urlpatterns = [ url(r'^login', login), url(r'^logout', logout), url(r'^signup', signup), ]
from django.conf.urls import url from views import login, logout, signup, confirm_email urlpatterns = [ url(r'^login', login), url(r'^logout', logout), url(r'^signup', signup), url(r'^confirm_email/(?P<key>\w+)', confirm_email), ]
Add an endpoint for email confirmation
Add an endpoint for email confirmation
Python
bsd-3-clause
SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder
from django.conf.urls import url - from views import login, logout, signup + from views import login, logout, signup, confirm_email urlpatterns = [ url(r'^login', login), url(r'^logout', logout), url(r'^signup', signup), + url(r'^confirm_email/(?P<key>\w+)', confirm_email), ]
Add an endpoint for email confirmation
## Code Before: from django.conf.urls import url from views import login, logout, signup urlpatterns = [ url(r'^login', login), url(r'^logout', logout), url(r'^signup', signup), ] ## Instruction: Add an endpoint for email confirmation ## Code After: from django.conf.urls import url from views import login, logout, signup, confirm_email urlpatterns = [ url(r'^login', login), url(r'^logout', logout), url(r'^signup', signup), url(r'^confirm_email/(?P<key>\w+)', confirm_email), ]
from django.conf.urls import url - from views import login, logout, signup + from views import login, logout, signup, confirm_email ? +++++++++++++++ urlpatterns = [ url(r'^login', login), url(r'^logout', logout), url(r'^signup', signup), + url(r'^confirm_email/(?P<key>\w+)', confirm_email), ]
943699de02c3d8f4f8e26370fbbff2dec8a2d5ea
api/identifiers/urls.py
api/identifiers/urls.py
from django.conf.urls import url from api.identifiers import views urlpatterns = [ url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name), ]
from django.conf.urls import url from api.identifiers import views urlpatterns = [ url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name), url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name), ]
Add identifier list to identifier views for use with embeds in registrations
Add identifier list to identifier views for use with embeds in registrations [#OSF-6628]
Python
apache-2.0
saradbowman/osf.io,alexschiller/osf.io,wearpants/osf.io,erinspace/osf.io,alexschiller/osf.io,mluo613/osf.io,rdhyee/osf.io,icereval/osf.io,chrisseto/osf.io,mluo613/osf.io,chennan47/osf.io,emetsger/osf.io,hmoco/osf.io,hmoco/osf.io,hmoco/osf.io,baylee-d/osf.io,baylee-d/osf.io,sloria/osf.io,SSJohns/osf.io,erinspace/osf.io,leb2dg/osf.io,felliott/osf.io,brianjgeiger/osf.io,SSJohns/osf.io,alexschiller/osf.io,HalcyonChimera/osf.io,felliott/osf.io,Nesiehr/osf.io,baylee-d/osf.io,TomBaxter/osf.io,wearpants/osf.io,acshi/osf.io,cslzchen/osf.io,SSJohns/osf.io,leb2dg/osf.io,CenterForOpenScience/osf.io,laurenrevere/osf.io,cslzchen/osf.io,acshi/osf.io,wearpants/osf.io,monikagrabowska/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,Nesiehr/osf.io,binoculars/osf.io,adlius/osf.io,chrisseto/osf.io,crcresearch/osf.io,icereval/osf.io,emetsger/osf.io,chennan47/osf.io,SSJohns/osf.io,pattisdr/osf.io,samchrisinger/osf.io,rdhyee/osf.io,mattclark/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,caneruguz/osf.io,chrisseto/osf.io,caseyrollins/osf.io,brianjgeiger/osf.io,pattisdr/osf.io,acshi/osf.io,saradbowman/osf.io,felliott/osf.io,icereval/osf.io,cwisecarver/osf.io,laurenrevere/osf.io,monikagrabowska/osf.io,mfraezz/osf.io,mattclark/osf.io,rdhyee/osf.io,caneruguz/osf.io,acshi/osf.io,amyshi188/osf.io,DanielSBrown/osf.io,amyshi188/osf.io,adlius/osf.io,adlius/osf.io,amyshi188/osf.io,DanielSBrown/osf.io,cwisecarver/osf.io,HalcyonChimera/osf.io,erinspace/osf.io,sloria/osf.io,rdhyee/osf.io,amyshi188/osf.io,leb2dg/osf.io,cslzchen/osf.io,mfraezz/osf.io,binoculars/osf.io,mfraezz/osf.io,cwisecarver/osf.io,leb2dg/osf.io,pattisdr/osf.io,cslzchen/osf.io,adlius/osf.io,emetsger/osf.io,aaxelb/osf.io,DanielSBrown/osf.io,cwisecarver/osf.io,emetsger/osf.io,mattclark/osf.io,Johnetordoff/osf.io,aaxelb/osf.io,alexschiller/osf.io,crcresearch/osf.io,brianjgeiger/osf.io,mluo613/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,chennan47/osf.io,CenterForOpenScience/osf.io,Nesiehr/osf.io,CenterForOpenScience/osf.io,binoculars/osf.io,monikagrabowska/osf.io,caseyrollins/osf.io,samchrisinger/osf.io,Nesiehr/osf.io,caseyrollins/osf.io,alexschiller/osf.io,TomBaxter/osf.io,mluo613/osf.io,samchrisinger/osf.io,crcresearch/osf.io,TomBaxter/osf.io,Johnetordoff/osf.io,DanielSBrown/osf.io,mluo613/osf.io,aaxelb/osf.io,felliott/osf.io,laurenrevere/osf.io,CenterForOpenScience/osf.io,HalcyonChimera/osf.io,mfraezz/osf.io,chrisseto/osf.io,acshi/osf.io,caneruguz/osf.io,aaxelb/osf.io,samchrisinger/osf.io,sloria/osf.io,wearpants/osf.io
from django.conf.urls import url from api.identifiers import views urlpatterns = [ url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name), + url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name), ]
Add identifier list to identifier views for use with embeds in registrations
## Code Before: from django.conf.urls import url from api.identifiers import views urlpatterns = [ url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name), ] ## Instruction: Add identifier list to identifier views for use with embeds in registrations ## Code After: from django.conf.urls import url from api.identifiers import views urlpatterns = [ url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name), url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name), ]
from django.conf.urls import url from api.identifiers import views urlpatterns = [ url(r'^(?P<identifier_id>\w+)/$', views.IdentifierDetail.as_view(), name=views.IdentifierDetail.view_name), + url(r'^(?P<node_id>\w+)/identifiers/$', views.IdentifierList.as_view(), name=views.IdentifierList.view_name), ]
a6ae4171de33dd77e9109523380c1330d4037f9f
gengine/app/tests/runner.py
gengine/app/tests/runner.py
from gengine.app.tests import db as db from gengine.metadata import init_declarative_base, init_session import unittest import os import pkgutil import testing.redis import logging log = logging.getLogger(__name__) init_session() init_declarative_base() __path__ = [x[0] for x in os.walk(os.path.dirname(__file__))] def create_test_suite(): suite = unittest.TestSuite() for imp, modname, _ in pkgutil.walk_packages(__path__): mod = imp.find_module(modname).load_module(modname) for test in unittest.defaultTestLoader.loadTestsFromModule(mod): suite.addTests(test) return suite if __name__=="__main__": exit = 1 try: redis = testing.redis.RedisServer() from gengine.base.cache import setup_redis_cache dsn = redis.dsn() setup_redis_cache(dsn["host"], dsn["port"], dsn["db"]) from gengine.app.cache import init_caches init_caches() db.setupDB() testSuite = create_test_suite() text_runner = unittest.TextTestRunner(failfast=True).run(testSuite) if text_runner.wasSuccessful(): exit = 0 finally: try: db.unsetupDB() except: log.exception() try: redis.stop() except: log.exception() sys.exit(exit)
from gengine.app.tests import db as db from gengine.metadata import init_declarative_base, init_session import unittest import os import pkgutil import testing.redis import logging import sys log = logging.getLogger(__name__) init_session() init_declarative_base() __path__ = [x[0] for x in os.walk(os.path.dirname(__file__))] def create_test_suite(): suite = unittest.TestSuite() for imp, modname, _ in pkgutil.walk_packages(__path__): mod = imp.find_module(modname).load_module(modname) for test in unittest.defaultTestLoader.loadTestsFromModule(mod): suite.addTests(test) return suite if __name__=="__main__": exit = 1 try: redis = testing.redis.RedisServer() from gengine.base.cache import setup_redis_cache dsn = redis.dsn() setup_redis_cache(dsn["host"], dsn["port"], dsn["db"]) from gengine.app.cache import init_caches init_caches() db.setupDB() testSuite = create_test_suite() text_runner = unittest.TextTestRunner(failfast=True).run(testSuite) if text_runner.wasSuccessful(): exit = 0 finally: try: db.unsetupDB() except: log.exception() try: redis.stop() except: log.exception() sys.exit(exit)
Add missing import for sys
Add missing import for sys
Python
mit
ActiDoo/gamification-engine,ActiDoo/gamification-engine,ActiDoo/gamification-engine,ActiDoo/gamification-engine
from gengine.app.tests import db as db from gengine.metadata import init_declarative_base, init_session import unittest import os import pkgutil import testing.redis import logging + import sys log = logging.getLogger(__name__) init_session() init_declarative_base() __path__ = [x[0] for x in os.walk(os.path.dirname(__file__))] def create_test_suite(): suite = unittest.TestSuite() for imp, modname, _ in pkgutil.walk_packages(__path__): mod = imp.find_module(modname).load_module(modname) for test in unittest.defaultTestLoader.loadTestsFromModule(mod): suite.addTests(test) return suite if __name__=="__main__": exit = 1 try: redis = testing.redis.RedisServer() from gengine.base.cache import setup_redis_cache dsn = redis.dsn() setup_redis_cache(dsn["host"], dsn["port"], dsn["db"]) from gengine.app.cache import init_caches init_caches() db.setupDB() testSuite = create_test_suite() text_runner = unittest.TextTestRunner(failfast=True).run(testSuite) if text_runner.wasSuccessful(): exit = 0 finally: try: db.unsetupDB() except: log.exception() try: redis.stop() except: log.exception() sys.exit(exit)
Add missing import for sys
## Code Before: from gengine.app.tests import db as db from gengine.metadata import init_declarative_base, init_session import unittest import os import pkgutil import testing.redis import logging log = logging.getLogger(__name__) init_session() init_declarative_base() __path__ = [x[0] for x in os.walk(os.path.dirname(__file__))] def create_test_suite(): suite = unittest.TestSuite() for imp, modname, _ in pkgutil.walk_packages(__path__): mod = imp.find_module(modname).load_module(modname) for test in unittest.defaultTestLoader.loadTestsFromModule(mod): suite.addTests(test) return suite if __name__=="__main__": exit = 1 try: redis = testing.redis.RedisServer() from gengine.base.cache import setup_redis_cache dsn = redis.dsn() setup_redis_cache(dsn["host"], dsn["port"], dsn["db"]) from gengine.app.cache import init_caches init_caches() db.setupDB() testSuite = create_test_suite() text_runner = unittest.TextTestRunner(failfast=True).run(testSuite) if text_runner.wasSuccessful(): exit = 0 finally: try: db.unsetupDB() except: log.exception() try: redis.stop() except: log.exception() sys.exit(exit) ## Instruction: Add missing import for sys ## Code After: from gengine.app.tests import db as db from gengine.metadata import init_declarative_base, init_session import unittest import os import pkgutil import testing.redis import logging import sys log = logging.getLogger(__name__) init_session() init_declarative_base() __path__ = [x[0] for x in os.walk(os.path.dirname(__file__))] def create_test_suite(): suite = unittest.TestSuite() for imp, modname, _ in pkgutil.walk_packages(__path__): mod = imp.find_module(modname).load_module(modname) for test in unittest.defaultTestLoader.loadTestsFromModule(mod): suite.addTests(test) return suite if __name__=="__main__": exit = 1 try: redis = testing.redis.RedisServer() from gengine.base.cache import setup_redis_cache dsn = redis.dsn() setup_redis_cache(dsn["host"], dsn["port"], dsn["db"]) from gengine.app.cache import init_caches init_caches() db.setupDB() testSuite = create_test_suite() text_runner = unittest.TextTestRunner(failfast=True).run(testSuite) if text_runner.wasSuccessful(): exit = 0 finally: try: db.unsetupDB() except: log.exception() try: redis.stop() except: log.exception() sys.exit(exit)
from gengine.app.tests import db as db from gengine.metadata import init_declarative_base, init_session import unittest import os import pkgutil import testing.redis import logging + import sys log = logging.getLogger(__name__) init_session() init_declarative_base() __path__ = [x[0] for x in os.walk(os.path.dirname(__file__))] def create_test_suite(): suite = unittest.TestSuite() for imp, modname, _ in pkgutil.walk_packages(__path__): mod = imp.find_module(modname).load_module(modname) for test in unittest.defaultTestLoader.loadTestsFromModule(mod): suite.addTests(test) return suite if __name__=="__main__": exit = 1 try: redis = testing.redis.RedisServer() from gengine.base.cache import setup_redis_cache dsn = redis.dsn() setup_redis_cache(dsn["host"], dsn["port"], dsn["db"]) from gengine.app.cache import init_caches init_caches() db.setupDB() testSuite = create_test_suite() text_runner = unittest.TextTestRunner(failfast=True).run(testSuite) if text_runner.wasSuccessful(): exit = 0 finally: try: db.unsetupDB() except: log.exception() try: redis.stop() except: log.exception() sys.exit(exit)
32410e639f3202c10d9c75083319a9ab81932b82
client/api.py
client/api.py
import os import sys import json import urllib import urllib2 import cookielib sys.path.append((os.path.dirname(__file__) or ".") + "/../") import config cj = cookielib.CookieJar() def callapi(action, postdata={}): postdata.update({"action": action}) opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [("User-Agent", "pyWebCash Scraper")] f = opener.open(config.apiurl,urllib.urlencode(postdata)) data = f.read() return json.loads(data)
import os import sys import json import urllib import httplib import urllib2 import cookielib sys.path.append((os.path.dirname(__file__) or ".") + "/../") import config cj = cookielib.CookieJar() class HTTPSClientAuthHandler(urllib2.HTTPSHandler): def __init__(self, key): urllib2.HTTPSHandler.__init__(self) self.key = key def https_open(self, req): # Rather than pass in a reference to a connection class, we pass in # a reference to a function which, for all intents and purposes, # will behave as a constructor return self.do_open(self.getConnection, req) def getConnection(self, host, timeout=300): return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key) def callapi(action, postdata={}): postdata.update({"action": action}) if config.certfile: opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj)) else: opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [("User-Agent", "pyWebCash Scraper")] f = opener.open(config.apiurl,urllib.urlencode(postdata)) data = f.read() return json.loads(data)
Use ssl client cert if given.
Use ssl client cert if given.
Python
agpl-3.0
vincebusam/pyWebCash,vincebusam/pyWebCash,vincebusam/pyWebCash
import os import sys import json import urllib + import httplib import urllib2 import cookielib sys.path.append((os.path.dirname(__file__) or ".") + "/../") import config cj = cookielib.CookieJar() + class HTTPSClientAuthHandler(urllib2.HTTPSHandler): + def __init__(self, key): + urllib2.HTTPSHandler.__init__(self) + self.key = key + + def https_open(self, req): + # Rather than pass in a reference to a connection class, we pass in + # a reference to a function which, for all intents and purposes, + # will behave as a constructor + return self.do_open(self.getConnection, req) + + def getConnection(self, host, timeout=300): + return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key) + def callapi(action, postdata={}): postdata.update({"action": action}) + if config.certfile: + opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj)) + else: - opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) + opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [("User-Agent", "pyWebCash Scraper")] f = opener.open(config.apiurl,urllib.urlencode(postdata)) data = f.read() return json.loads(data)
Use ssl client cert if given.
## Code Before: import os import sys import json import urllib import urllib2 import cookielib sys.path.append((os.path.dirname(__file__) or ".") + "/../") import config cj = cookielib.CookieJar() def callapi(action, postdata={}): postdata.update({"action": action}) opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [("User-Agent", "pyWebCash Scraper")] f = opener.open(config.apiurl,urllib.urlencode(postdata)) data = f.read() return json.loads(data) ## Instruction: Use ssl client cert if given. ## Code After: import os import sys import json import urllib import httplib import urllib2 import cookielib sys.path.append((os.path.dirname(__file__) or ".") + "/../") import config cj = cookielib.CookieJar() class HTTPSClientAuthHandler(urllib2.HTTPSHandler): def __init__(self, key): urllib2.HTTPSHandler.__init__(self) self.key = key def https_open(self, req): # Rather than pass in a reference to a connection class, we pass in # a reference to a function which, for all intents and purposes, # will behave as a constructor return self.do_open(self.getConnection, req) def getConnection(self, host, timeout=300): return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key) def callapi(action, postdata={}): postdata.update({"action": action}) if config.certfile: opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj)) else: opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [("User-Agent", "pyWebCash Scraper")] f = opener.open(config.apiurl,urllib.urlencode(postdata)) data = f.read() return json.loads(data)
import os import sys import json import urllib + import httplib import urllib2 import cookielib sys.path.append((os.path.dirname(__file__) or ".") + "/../") import config cj = cookielib.CookieJar() + class HTTPSClientAuthHandler(urllib2.HTTPSHandler): + def __init__(self, key): + urllib2.HTTPSHandler.__init__(self) + self.key = key + + def https_open(self, req): + # Rather than pass in a reference to a connection class, we pass in + # a reference to a function which, for all intents and purposes, + # will behave as a constructor + return self.do_open(self.getConnection, req) + + def getConnection(self, host, timeout=300): + return httplib.HTTPSConnection(host, key_file=self.key, cert_file=self.key) + def callapi(action, postdata={}): postdata.update({"action": action}) + if config.certfile: + opener = urllib2.build_opener(HTTPSClientAuthHandler(config.certfile), urllib2.HTTPCookieProcessor(cj)) + else: - opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) + opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) ? ++++ opener.addheaders = [("User-Agent", "pyWebCash Scraper")] f = opener.open(config.apiurl,urllib.urlencode(postdata)) data = f.read() return json.loads(data)
06645a637c0d34270f88f9a6b96133da5c415dd7
froide/publicbody/admin.py
froide/publicbody/admin.py
from django.contrib import admin from froide.publicbody.models import PublicBody, FoiLaw class PublicBodyAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("geography", "name",)} list_display = ('name', 'classification', 'geography') list_filter = ('classification',) search_fields = ['name', "description"] exclude = ('confirmed',) class FoiLawAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("jurisdiction", "name",)} admin.site.register(PublicBody, PublicBodyAdmin) admin.site.register(FoiLaw, FoiLawAdmin)
from django.contrib import admin from froide.publicbody.models import PublicBody, FoiLaw class PublicBodyAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("geography", "name",)} list_display = ('name', 'classification', 'topic', 'geography') list_filter = ('classification', 'topic',) search_fields = ['name', "description"] exclude = ('confirmed',) class FoiLawAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("jurisdiction", "name",)} admin.site.register(PublicBody, PublicBodyAdmin) admin.site.register(FoiLaw, FoiLawAdmin)
Add topic to PublicBodyAdmin list_filter and list_display
Add topic to PublicBodyAdmin list_filter and list_display
Python
mit
catcosmo/froide,ryankanno/froide,stefanw/froide,catcosmo/froide,fin/froide,CodeforHawaii/froide,fin/froide,okfse/froide,LilithWittmann/froide,okfse/froide,stefanw/froide,ryankanno/froide,LilithWittmann/froide,LilithWittmann/froide,CodeforHawaii/froide,fin/froide,catcosmo/froide,catcosmo/froide,okfse/froide,ryankanno/froide,okfse/froide,catcosmo/froide,CodeforHawaii/froide,ryankanno/froide,LilithWittmann/froide,fin/froide,LilithWittmann/froide,stefanw/froide,stefanw/froide,stefanw/froide,ryankanno/froide,CodeforHawaii/froide,CodeforHawaii/froide,okfse/froide
from django.contrib import admin from froide.publicbody.models import PublicBody, FoiLaw class PublicBodyAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("geography", "name",)} - list_display = ('name', 'classification', 'geography') + list_display = ('name', 'classification', 'topic', 'geography') - list_filter = ('classification',) + list_filter = ('classification', 'topic',) search_fields = ['name', "description"] exclude = ('confirmed',) class FoiLawAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("jurisdiction", "name",)} admin.site.register(PublicBody, PublicBodyAdmin) admin.site.register(FoiLaw, FoiLawAdmin)
Add topic to PublicBodyAdmin list_filter and list_display
## Code Before: from django.contrib import admin from froide.publicbody.models import PublicBody, FoiLaw class PublicBodyAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("geography", "name",)} list_display = ('name', 'classification', 'geography') list_filter = ('classification',) search_fields = ['name', "description"] exclude = ('confirmed',) class FoiLawAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("jurisdiction", "name",)} admin.site.register(PublicBody, PublicBodyAdmin) admin.site.register(FoiLaw, FoiLawAdmin) ## Instruction: Add topic to PublicBodyAdmin list_filter and list_display ## Code After: from django.contrib import admin from froide.publicbody.models import PublicBody, FoiLaw class PublicBodyAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("geography", "name",)} list_display = ('name', 'classification', 'topic', 'geography') list_filter = ('classification', 'topic',) search_fields = ['name', "description"] exclude = ('confirmed',) class FoiLawAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("jurisdiction", "name",)} admin.site.register(PublicBody, PublicBodyAdmin) admin.site.register(FoiLaw, FoiLawAdmin)
from django.contrib import admin from froide.publicbody.models import PublicBody, FoiLaw class PublicBodyAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("geography", "name",)} - list_display = ('name', 'classification', 'geography') + list_display = ('name', 'classification', 'topic', 'geography') ? +++++++++ - list_filter = ('classification',) + list_filter = ('classification', 'topic',) ? +++++++++ search_fields = ['name', "description"] exclude = ('confirmed',) class FoiLawAdmin(admin.ModelAdmin): prepopulated_fields = {"slug": ("jurisdiction", "name",)} admin.site.register(PublicBody, PublicBodyAdmin) admin.site.register(FoiLaw, FoiLawAdmin)
c27a1fc4c0251b896667e21a0a88fb44a403242f
cistern/migrations.py
cistern/migrations.py
import os from playhouse.migrate import * cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) migrator = SqliteMigrator(db) date_added = DateTimeField(default=None) migrate( migrator.add_column('torrent', 'date_added', date_added) )
import datetime import os from playhouse.migrate import * def update(): cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) migrator = SqliteMigrator(db) date_added = DateTimeField(default=datetime.datetime.now) migrate( migrator.add_column('torrent', 'date_added', date_added) )
Move migration to a function
Move migration to a function
Python
mit
archangelic/cistern
+ import datetime import os from playhouse.migrate import * + def update(): - cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) + cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) - db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) + db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) - migrator = SqliteMigrator(db) + migrator = SqliteMigrator(db) - date_added = DateTimeField(default=None) + date_added = DateTimeField(default=datetime.datetime.now) - migrate( + migrate( - migrator.add_column('torrent', 'date_added', date_added) + migrator.add_column('torrent', 'date_added', date_added) - ) + )
Move migration to a function
## Code Before: import os from playhouse.migrate import * cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) migrator = SqliteMigrator(db) date_added = DateTimeField(default=None) migrate( migrator.add_column('torrent', 'date_added', date_added) ) ## Instruction: Move migration to a function ## Code After: import datetime import os from playhouse.migrate import * def update(): cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) migrator = SqliteMigrator(db) date_added = DateTimeField(default=datetime.datetime.now) migrate( migrator.add_column('torrent', 'date_added', date_added) )
+ import datetime import os from playhouse.migrate import * + def update(): - cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) + cistern_folder = os.getenv('CISTERNHOME', os.path.join(os.environ['HOME'], '.cistern')) ? ++++ - db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) + db = SqliteDatabase(os.path.join(cistern_folder, 'cistern.db')) ? ++++ - migrator = SqliteMigrator(db) + migrator = SqliteMigrator(db) ? ++++ - date_added = DateTimeField(default=None) + date_added = DateTimeField(default=datetime.datetime.now) - migrate( + migrate( ? ++++ - migrator.add_column('torrent', 'date_added', date_added) + migrator.add_column('torrent', 'date_added', date_added) ? ++++ - ) + )
6fe5a416ed229e7ec8efab9d6b3dac43f16515b6
corehq/apps/domain/__init__.py
corehq/apps/domain/__init__.py
from corehq.preindex import ExtraPreindexPlugin from django.conf import settings ExtraPreindexPlugin.register('domain', __file__, ( settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta'))
from corehq.preindex import ExtraPreindexPlugin from django.conf import settings ExtraPreindexPlugin.register('domain', __file__, ( settings.NEW_DOMAINS_DB, settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta', ))
Add the new domains db
Add the new domains db
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq
from corehq.preindex import ExtraPreindexPlugin from django.conf import settings ExtraPreindexPlugin.register('domain', __file__, ( - settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta')) + settings.NEW_DOMAINS_DB, + settings.NEW_USERS_GROUPS_DB, + settings.NEW_FIXTURES_DB, + 'meta', + ))
Add the new domains db
## Code Before: from corehq.preindex import ExtraPreindexPlugin from django.conf import settings ExtraPreindexPlugin.register('domain', __file__, ( settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta')) ## Instruction: Add the new domains db ## Code After: from corehq.preindex import ExtraPreindexPlugin from django.conf import settings ExtraPreindexPlugin.register('domain', __file__, ( settings.NEW_DOMAINS_DB, settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta', ))
from corehq.preindex import ExtraPreindexPlugin from django.conf import settings ExtraPreindexPlugin.register('domain', __file__, ( - settings.NEW_USERS_GROUPS_DB, settings.NEW_FIXTURES_DB, 'meta')) + settings.NEW_DOMAINS_DB, + settings.NEW_USERS_GROUPS_DB, + settings.NEW_FIXTURES_DB, + 'meta', + ))
a52c84f092d89f89130c2696c98779e955f083dc
tests/test_version_parser.py
tests/test_version_parser.py
import pytest from leak.version_parser import versions_split def test_versions_split(): pass def test_wrong_versions_split(): # too many dots assert versions_split('1.2.3.4') == [0, 0, 0] # test missing numeric version with pytest.raises(ValueError): versions_split('not.numeric') # test not string provided with pytest.raises(AttributeError): versions_split(12345)
import pytest from leak.version_parser import versions_split def test_versions_split(): assert versions_split('1.8.1') == [1, 8, 1] assert versions_split('1.4') == [1, 4, 0] assert versions_split('2') == [2, 0, 0] def test_versions_split_str_mapping(): assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0'] assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0'] assert versions_split('text', type_applyer=str) == ['text', '0', '0'] def test_wrong_versions_split(): # too many dots assert versions_split('1.2.3.4') == [0, 0, 0] # test missing numeric version with pytest.raises(ValueError): versions_split('not.numeric') # test not string provided with pytest.raises(AttributeError): versions_split(12345)
Add tests for version splitting
Add tests for version splitting
Python
mit
bmwant21/leak
import pytest from leak.version_parser import versions_split def test_versions_split(): - pass + assert versions_split('1.8.1') == [1, 8, 1] + assert versions_split('1.4') == [1, 4, 0] + assert versions_split('2') == [2, 0, 0] + + + def test_versions_split_str_mapping(): + assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0'] + assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0'] + assert versions_split('text', type_applyer=str) == ['text', '0', '0'] def test_wrong_versions_split(): # too many dots assert versions_split('1.2.3.4') == [0, 0, 0] # test missing numeric version with pytest.raises(ValueError): versions_split('not.numeric') # test not string provided with pytest.raises(AttributeError): versions_split(12345)
Add tests for version splitting
## Code Before: import pytest from leak.version_parser import versions_split def test_versions_split(): pass def test_wrong_versions_split(): # too many dots assert versions_split('1.2.3.4') == [0, 0, 0] # test missing numeric version with pytest.raises(ValueError): versions_split('not.numeric') # test not string provided with pytest.raises(AttributeError): versions_split(12345) ## Instruction: Add tests for version splitting ## Code After: import pytest from leak.version_parser import versions_split def test_versions_split(): assert versions_split('1.8.1') == [1, 8, 1] assert versions_split('1.4') == [1, 4, 0] assert versions_split('2') == [2, 0, 0] def test_versions_split_str_mapping(): assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0'] assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0'] assert versions_split('text', type_applyer=str) == ['text', '0', '0'] def test_wrong_versions_split(): # too many dots assert versions_split('1.2.3.4') == [0, 0, 0] # test missing numeric version with pytest.raises(ValueError): versions_split('not.numeric') # test not string provided with pytest.raises(AttributeError): versions_split(12345)
import pytest from leak.version_parser import versions_split def test_versions_split(): - pass + assert versions_split('1.8.1') == [1, 8, 1] + assert versions_split('1.4') == [1, 4, 0] + assert versions_split('2') == [2, 0, 0] + + + def test_versions_split_str_mapping(): + assert versions_split('1.11rc1', type_applyer=str) == ['1', '11rc1', '0'] + assert versions_split('1.10b1', type_applyer=str) == ['1', '10b1', '0'] + assert versions_split('text', type_applyer=str) == ['text', '0', '0'] def test_wrong_versions_split(): # too many dots assert versions_split('1.2.3.4') == [0, 0, 0] # test missing numeric version with pytest.raises(ValueError): versions_split('not.numeric') # test not string provided with pytest.raises(AttributeError): versions_split(12345)
4bc31e675659af54ee26fe5df16a0ee3ebeb5947
firefed/__main__.py
firefed/__main__.py
import argparse import os import re from firefed import Firefed from feature import feature_map, Summary def feature_type(val): try: return feature_map()[val] except KeyError as key: raise argparse.ArgumentTypeError( 'Feature %s not found. Choose from: {%s}' % (key, ', '.join(feature_map()))) def profile_dir(dirname): if dirname is None: dirname = 'default' if os.path.isdir(dirname): return dirname if re.match('^[\\w-]+$', dirname): home = os.path.expanduser('~/.mozilla/firefox') profile_names = os.listdir(home) for name in profile_names: if name.endswith('.%s' % dirname): return os.path.join(home, name) raise argparse.ArgumentTypeError('Profile %s not found.' % dirname) def main(): parser = argparse.ArgumentParser( 'firefed', description= 'Firefed is a Firefox profile analyzer focusing on privacy and security.', ) parser.add_argument( '-p', '--profile', help='profile name or directory', type=profile_dir, required=True) parser.add_argument( '-f', '--feature', type=feature_type, default=Summary, help='{%s}' % ', '.join(feature_map())) parser.add_argument( '-s', '--summarize', action='store_true', help='summarize results') args = parser.parse_args() Firefed(args) if __name__ == '__main__': main()
import argparse import os import re from firefed import Firefed from feature import feature_map, Summary def feature_type(val): try: return feature_map()[val] except KeyError as key: raise argparse.ArgumentTypeError( 'Feature %s not found. Choose from: {%s}' % (key, ', '.join(feature_map()))) def profile_dir(dirname): if dirname is None: dirname = 'default' if os.path.isdir(dirname): return dirname if re.match('^[\\w-]+$', dirname): home = os.path.expanduser('~/.mozilla/firefox') profile_names = os.listdir(home) for name in profile_names: if name.endswith('.%s' % dirname): return os.path.join(home, name) raise argparse.ArgumentTypeError('Profile %s not found.' % dirname) def main(): parser = argparse.ArgumentParser( 'firefed', description= 'Firefed is a Firefox profile analyzer focusing on privacy and security.', ) parser.add_argument( '-p', '--profile', help='profile name or directory', type=profile_dir, default='default') parser.add_argument( '-f', '--feature', type=feature_type, default=Summary, help='{%s}' % ', '.join(feature_map())) parser.add_argument( '-s', '--summarize', action='store_true', help='summarize results') args = parser.parse_args() Firefed(args) if __name__ == '__main__': main()
Add default argument for profile
Add default argument for profile
Python
mit
numirias/firefed
import argparse import os import re from firefed import Firefed from feature import feature_map, Summary def feature_type(val): try: return feature_map()[val] except KeyError as key: raise argparse.ArgumentTypeError( 'Feature %s not found. Choose from: {%s}' % (key, ', '.join(feature_map()))) def profile_dir(dirname): if dirname is None: dirname = 'default' if os.path.isdir(dirname): return dirname if re.match('^[\\w-]+$', dirname): home = os.path.expanduser('~/.mozilla/firefox') profile_names = os.listdir(home) for name in profile_names: if name.endswith('.%s' % dirname): return os.path.join(home, name) raise argparse.ArgumentTypeError('Profile %s not found.' % dirname) def main(): parser = argparse.ArgumentParser( 'firefed', description= 'Firefed is a Firefox profile analyzer focusing on privacy and security.', ) parser.add_argument( '-p', '--profile', help='profile name or directory', type=profile_dir, - required=True) + default='default') parser.add_argument( '-f', '--feature', type=feature_type, default=Summary, help='{%s}' % ', '.join(feature_map())) parser.add_argument( '-s', '--summarize', action='store_true', help='summarize results') args = parser.parse_args() Firefed(args) if __name__ == '__main__': main()
Add default argument for profile
## Code Before: import argparse import os import re from firefed import Firefed from feature import feature_map, Summary def feature_type(val): try: return feature_map()[val] except KeyError as key: raise argparse.ArgumentTypeError( 'Feature %s not found. Choose from: {%s}' % (key, ', '.join(feature_map()))) def profile_dir(dirname): if dirname is None: dirname = 'default' if os.path.isdir(dirname): return dirname if re.match('^[\\w-]+$', dirname): home = os.path.expanduser('~/.mozilla/firefox') profile_names = os.listdir(home) for name in profile_names: if name.endswith('.%s' % dirname): return os.path.join(home, name) raise argparse.ArgumentTypeError('Profile %s not found.' % dirname) def main(): parser = argparse.ArgumentParser( 'firefed', description= 'Firefed is a Firefox profile analyzer focusing on privacy and security.', ) parser.add_argument( '-p', '--profile', help='profile name or directory', type=profile_dir, required=True) parser.add_argument( '-f', '--feature', type=feature_type, default=Summary, help='{%s}' % ', '.join(feature_map())) parser.add_argument( '-s', '--summarize', action='store_true', help='summarize results') args = parser.parse_args() Firefed(args) if __name__ == '__main__': main() ## Instruction: Add default argument for profile ## Code After: import argparse import os import re from firefed import Firefed from feature import feature_map, Summary def feature_type(val): try: return feature_map()[val] except KeyError as key: raise argparse.ArgumentTypeError( 'Feature %s not found. Choose from: {%s}' % (key, ', '.join(feature_map()))) def profile_dir(dirname): if dirname is None: dirname = 'default' if os.path.isdir(dirname): return dirname if re.match('^[\\w-]+$', dirname): home = os.path.expanduser('~/.mozilla/firefox') profile_names = os.listdir(home) for name in profile_names: if name.endswith('.%s' % dirname): return os.path.join(home, name) raise argparse.ArgumentTypeError('Profile %s not found.' % dirname) def main(): parser = argparse.ArgumentParser( 'firefed', description= 'Firefed is a Firefox profile analyzer focusing on privacy and security.', ) parser.add_argument( '-p', '--profile', help='profile name or directory', type=profile_dir, default='default') parser.add_argument( '-f', '--feature', type=feature_type, default=Summary, help='{%s}' % ', '.join(feature_map())) parser.add_argument( '-s', '--summarize', action='store_true', help='summarize results') args = parser.parse_args() Firefed(args) if __name__ == '__main__': main()
import argparse import os import re from firefed import Firefed from feature import feature_map, Summary def feature_type(val): try: return feature_map()[val] except KeyError as key: raise argparse.ArgumentTypeError( 'Feature %s not found. Choose from: {%s}' % (key, ', '.join(feature_map()))) def profile_dir(dirname): if dirname is None: dirname = 'default' if os.path.isdir(dirname): return dirname if re.match('^[\\w-]+$', dirname): home = os.path.expanduser('~/.mozilla/firefox') profile_names = os.listdir(home) for name in profile_names: if name.endswith('.%s' % dirname): return os.path.join(home, name) raise argparse.ArgumentTypeError('Profile %s not found.' % dirname) def main(): parser = argparse.ArgumentParser( 'firefed', description= 'Firefed is a Firefox profile analyzer focusing on privacy and security.', ) parser.add_argument( '-p', '--profile', help='profile name or directory', type=profile_dir, - required=True) + default='default') parser.add_argument( '-f', '--feature', type=feature_type, default=Summary, help='{%s}' % ', '.join(feature_map())) parser.add_argument( '-s', '--summarize', action='store_true', help='summarize results') args = parser.parse_args() Firefed(args) if __name__ == '__main__': main()
dd19012ed8bb6ec702d84abe400bc3dec47044f3
sortedm2m_tests/__init__.py
sortedm2m_tests/__init__.py
import os # django-setuptest import setuptest TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support'] class TestSuite(setuptest.SetupTestSuite): def resolve_packages(self): packages = super(TestSuite, self).resolve_packages() for test_app in TEST_APPS: if test_app not in packages: packages.append(test_app) return packages
import os # django-setuptest import setuptest TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support'] class TestSuite(setuptest.SetupTestSuite): def __init__(self, *args, **kwargs): os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings' from south.management.commands import patch_for_test_db_setup patch_for_test_db_setup() super(TestSuite, self).__init__(*args, **kwargs) def resolve_packages(self): packages = super(TestSuite, self).resolve_packages() for test_app in TEST_APPS: if test_app not in packages: packages.append(test_app) return packages
Fix to allow tests with South migrations to run.
Fix to allow tests with South migrations to run.
Python
bsd-3-clause
gregmuellegger/django-sortedm2m,fabrique/django-sortedm2m,gradel/django-sortedm2m,MathieuDuponchelle/django-sortedm2m,fabrique/django-sortedm2m,gradel/django-sortedm2m,fabrique/django-sortedm2m,gradel/django-sortedm2m,gregmuellegger/django-sortedm2m,MathieuDuponchelle/django-sortedm2m,gregmuellegger/django-sortedm2m
import os # django-setuptest import setuptest TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support'] class TestSuite(setuptest.SetupTestSuite): - + + def __init__(self, *args, **kwargs): + os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings' + from south.management.commands import patch_for_test_db_setup + patch_for_test_db_setup() + super(TestSuite, self).__init__(*args, **kwargs) + def resolve_packages(self): packages = super(TestSuite, self).resolve_packages() for test_app in TEST_APPS: if test_app not in packages: packages.append(test_app) return packages +
Fix to allow tests with South migrations to run.
## Code Before: import os # django-setuptest import setuptest TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support'] class TestSuite(setuptest.SetupTestSuite): def resolve_packages(self): packages = super(TestSuite, self).resolve_packages() for test_app in TEST_APPS: if test_app not in packages: packages.append(test_app) return packages ## Instruction: Fix to allow tests with South migrations to run. ## Code After: import os # django-setuptest import setuptest TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support'] class TestSuite(setuptest.SetupTestSuite): def __init__(self, *args, **kwargs): os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings' from south.management.commands import patch_for_test_db_setup patch_for_test_db_setup() super(TestSuite, self).__init__(*args, **kwargs) def resolve_packages(self): packages = super(TestSuite, self).resolve_packages() for test_app in TEST_APPS: if test_app not in packages: packages.append(test_app) return packages
import os # django-setuptest import setuptest TEST_APPS = ['sortedm2m_tests', 'sortedm2m_field', 'sortedm2m_form', 'south_support'] class TestSuite(setuptest.SetupTestSuite): - + + def __init__(self, *args, **kwargs): + os.environ['DJANGO_SETTINGS_MODULE'] = 'test_settings' + from south.management.commands import patch_for_test_db_setup + patch_for_test_db_setup() + super(TestSuite, self).__init__(*args, **kwargs) + def resolve_packages(self): packages = super(TestSuite, self).resolve_packages() for test_app in TEST_APPS: if test_app not in packages: packages.append(test_app) return packages + +
84b48b9be466ac72bddf5ee6288ff48be26eed62
tests/classifier/RandomForestClassifier/RandomForestClassifierPHPTest.py
tests/classifier/RandomForestClassifier/RandomForestClassifierPHPTest.py
import unittest from unittest import TestCase from sklearn.ensemble import RandomForestClassifier from ..Classifier import Classifier from ...language.PHP import PHP class RandomForestClassifierPHPTest(PHP, Classifier, TestCase): def setUp(self): super(RandomForestClassifierPHPTest, self).setUp() self.mdl = RandomForestClassifier(n_estimators=100, random_state=0) def tearDown(self): super(RandomForestClassifierPHPTest, self).tearDown() @unittest.skip('The generated code would be too large.') def test_existing_features_w_digits_data(self): pass @unittest.skip('The generated code would be too large.') def test_random_features_w_digits_data(self): pass
import unittest from unittest import TestCase from sklearn.ensemble import RandomForestClassifier from ..Classifier import Classifier from ...language.PHP import PHP class RandomForestClassifierPHPTest(PHP, Classifier, TestCase): def setUp(self): super(RandomForestClassifierPHPTest, self).setUp() self.mdl = RandomForestClassifier(n_estimators=20, random_state=0) def tearDown(self): super(RandomForestClassifierPHPTest, self).tearDown()
Reduce the number of trees
Reduce the number of trees
Python
bsd-3-clause
nok/sklearn-porter
import unittest from unittest import TestCase from sklearn.ensemble import RandomForestClassifier from ..Classifier import Classifier from ...language.PHP import PHP class RandomForestClassifierPHPTest(PHP, Classifier, TestCase): def setUp(self): super(RandomForestClassifierPHPTest, self).setUp() - self.mdl = RandomForestClassifier(n_estimators=100, random_state=0) + self.mdl = RandomForestClassifier(n_estimators=20, random_state=0) def tearDown(self): super(RandomForestClassifierPHPTest, self).tearDown() - @unittest.skip('The generated code would be too large.') - def test_existing_features_w_digits_data(self): - pass - - @unittest.skip('The generated code would be too large.') - def test_random_features_w_digits_data(self): - pass -
Reduce the number of trees
## Code Before: import unittest from unittest import TestCase from sklearn.ensemble import RandomForestClassifier from ..Classifier import Classifier from ...language.PHP import PHP class RandomForestClassifierPHPTest(PHP, Classifier, TestCase): def setUp(self): super(RandomForestClassifierPHPTest, self).setUp() self.mdl = RandomForestClassifier(n_estimators=100, random_state=0) def tearDown(self): super(RandomForestClassifierPHPTest, self).tearDown() @unittest.skip('The generated code would be too large.') def test_existing_features_w_digits_data(self): pass @unittest.skip('The generated code would be too large.') def test_random_features_w_digits_data(self): pass ## Instruction: Reduce the number of trees ## Code After: import unittest from unittest import TestCase from sklearn.ensemble import RandomForestClassifier from ..Classifier import Classifier from ...language.PHP import PHP class RandomForestClassifierPHPTest(PHP, Classifier, TestCase): def setUp(self): super(RandomForestClassifierPHPTest, self).setUp() self.mdl = RandomForestClassifier(n_estimators=20, random_state=0) def tearDown(self): super(RandomForestClassifierPHPTest, self).tearDown()
import unittest from unittest import TestCase from sklearn.ensemble import RandomForestClassifier from ..Classifier import Classifier from ...language.PHP import PHP class RandomForestClassifierPHPTest(PHP, Classifier, TestCase): def setUp(self): super(RandomForestClassifierPHPTest, self).setUp() - self.mdl = RandomForestClassifier(n_estimators=100, random_state=0) ? ^^ + self.mdl = RandomForestClassifier(n_estimators=20, random_state=0) ? ^ def tearDown(self): super(RandomForestClassifierPHPTest, self).tearDown() - - @unittest.skip('The generated code would be too large.') - def test_existing_features_w_digits_data(self): - pass - - @unittest.skip('The generated code would be too large.') - def test_random_features_w_digits_data(self): - pass
9dad4f997371011ee7fe9f6ecd0c1a58cbba6d27
html_parse.py
html_parse.py
from bs4 import BeautifulSoup def parse(html): soup = BeautifulSoup(html, features="html.parser") return soup.get_text()
import imp import logging log = logging.getLogger(__name__) def module_exists(module_name): try: imp.find_module(module_name) return True except ImportError: return False if module_exists("bs4"): log.info("Parsing HTML using beautifulsoup4") from bs4 import BeautifulSoup def parse(html): soup = BeautifulSoup(html, features="html.parser") return soup.get_text() elif module_exists("html2text"): log.info("Parsing HTML using html2text") import html2text def parse(html): h = html2text.HTML2Text() h.single_line_break = True return h.handle(html) else: warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text" log.warning(warning_msg) def parse(html): raise ImportWarning(warning_msg)
Add support for html2text or no parser Will still prefer beautifulsoup4 if installed
Add support for html2text or no parser Will still prefer beautifulsoup4 if installed
Python
mit
idiotandrobot/heathergraph
- from bs4 import BeautifulSoup + import imp + import logging + log = logging.getLogger(__name__) + def module_exists(module_name): + try: + imp.find_module(module_name) + return True + except ImportError: + return False + + if module_exists("bs4"): + log.info("Parsing HTML using beautifulsoup4") + from bs4 import BeautifulSoup + - def parse(html): + def parse(html): - soup = BeautifulSoup(html, features="html.parser") + soup = BeautifulSoup(html, features="html.parser") - return soup.get_text() + return soup.get_text() + elif module_exists("html2text"): + log.info("Parsing HTML using html2text") + import html2text + + def parse(html): + h = html2text.HTML2Text() + h.single_line_break = True + return h.handle(html) + else: + warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text" + log.warning(warning_msg) + def parse(html): + raise ImportWarning(warning_msg)
Add support for html2text or no parser Will still prefer beautifulsoup4 if installed
## Code Before: from bs4 import BeautifulSoup def parse(html): soup = BeautifulSoup(html, features="html.parser") return soup.get_text() ## Instruction: Add support for html2text or no parser Will still prefer beautifulsoup4 if installed ## Code After: import imp import logging log = logging.getLogger(__name__) def module_exists(module_name): try: imp.find_module(module_name) return True except ImportError: return False if module_exists("bs4"): log.info("Parsing HTML using beautifulsoup4") from bs4 import BeautifulSoup def parse(html): soup = BeautifulSoup(html, features="html.parser") return soup.get_text() elif module_exists("html2text"): log.info("Parsing HTML using html2text") import html2text def parse(html): h = html2text.HTML2Text() h.single_line_break = True return h.handle(html) else: warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text" log.warning(warning_msg) def parse(html): raise ImportWarning(warning_msg)
- from bs4 import BeautifulSoup + import imp + import logging + log = logging.getLogger(__name__) + def module_exists(module_name): + try: + imp.find_module(module_name) + return True + except ImportError: + return False + + if module_exists("bs4"): + log.info("Parsing HTML using beautifulsoup4") + from bs4 import BeautifulSoup + - def parse(html): + def parse(html): ? ++++ - soup = BeautifulSoup(html, features="html.parser") + soup = BeautifulSoup(html, features="html.parser") ? ++++ - return soup.get_text() + return soup.get_text() ? ++++ + elif module_exists("html2text"): + log.info("Parsing HTML using html2text") + import html2text + + def parse(html): + h = html2text.HTML2Text() + h.single_line_break = True + return h.handle(html) + else: + warning_msg = "HTML parsing not available. Install beautifulsoup4 or html2text" + log.warning(warning_msg) + def parse(html): + raise ImportWarning(warning_msg)
2dc56ab04ea17bea05654eaec12bb27b48b0b225
robotd/cvcapture.py
robotd/cvcapture.py
import threading from robotd.native import _cvcapture class CaptureDevice(object): def __init__(self, path=None): if path is not None: argument_c = _cvcapture.ffi.new( 'char[]', path.encode('utf-8'), ) else: argument_c = _cvcapture.ffi.NULL self.instance = _cvcapture.lib.cvopen(argument_c) self.lock = threading.Lock() def capture(self, width, height): if self.instance is None: raise RuntimeError("capture device is closed") capture_buffer = _cvcapture.ffi.new( 'uint8_t[{}]'.format(width * height), ) with self.lock: status = _cvcapture.lib.cvcapture( self.instance, capture_buffer, width, height, ) if status == 0: raise RuntimeError("cvcapture() failed") return bytes(_cvcapture.ffi.buffer(capture_buffer)) def __enter__(self): return self def __exit__(self, exc_value, exc_type, exc_traceback): self.close() def close(self): if self.instance is not None: with self.lock: _cvcapture.lib.cvclose(self.instance) self.instance = None __del__ = close
import threading from robotd.native import _cvcapture class CaptureDevice(object): def __init__(self, path=None): if path is not None: argument_c = _cvcapture.ffi.new( 'char[]', path.encode('utf-8'), ) else: argument_c = _cvcapture.ffi.NULL self.instance = _cvcapture.lib.cvopen(argument_c) if self.instance == _cvcapture.ffi.NULL: raise RuntimeError("Unable to open capture device") self.lock = threading.Lock() def capture(self, width, height): if self.instance is None: raise RuntimeError("capture device is closed") capture_buffer = _cvcapture.ffi.new( 'uint8_t[{}]'.format(width * height), ) with self.lock: status = _cvcapture.lib.cvcapture( self.instance, capture_buffer, width, height, ) if status == 0: raise RuntimeError("cvcapture() failed") return bytes(_cvcapture.ffi.buffer(capture_buffer)) def __enter__(self): return self def __exit__(self, exc_value, exc_type, exc_traceback): self.close() def close(self): if self.instance is not None: with self.lock: _cvcapture.lib.cvclose(self.instance) self.instance = None __del__ = close
Raise a `RuntimeError` if the device cannot be opened
Raise a `RuntimeError` if the device cannot be opened
Python
mit
sourcebots/robotd,sourcebots/robotd
import threading from robotd.native import _cvcapture class CaptureDevice(object): def __init__(self, path=None): if path is not None: argument_c = _cvcapture.ffi.new( 'char[]', path.encode('utf-8'), ) else: argument_c = _cvcapture.ffi.NULL self.instance = _cvcapture.lib.cvopen(argument_c) + if self.instance == _cvcapture.ffi.NULL: + raise RuntimeError("Unable to open capture device") self.lock = threading.Lock() def capture(self, width, height): if self.instance is None: raise RuntimeError("capture device is closed") capture_buffer = _cvcapture.ffi.new( 'uint8_t[{}]'.format(width * height), ) with self.lock: status = _cvcapture.lib.cvcapture( self.instance, capture_buffer, width, height, ) if status == 0: raise RuntimeError("cvcapture() failed") return bytes(_cvcapture.ffi.buffer(capture_buffer)) def __enter__(self): return self def __exit__(self, exc_value, exc_type, exc_traceback): self.close() def close(self): if self.instance is not None: with self.lock: _cvcapture.lib.cvclose(self.instance) self.instance = None __del__ = close
Raise a `RuntimeError` if the device cannot be opened
## Code Before: import threading from robotd.native import _cvcapture class CaptureDevice(object): def __init__(self, path=None): if path is not None: argument_c = _cvcapture.ffi.new( 'char[]', path.encode('utf-8'), ) else: argument_c = _cvcapture.ffi.NULL self.instance = _cvcapture.lib.cvopen(argument_c) self.lock = threading.Lock() def capture(self, width, height): if self.instance is None: raise RuntimeError("capture device is closed") capture_buffer = _cvcapture.ffi.new( 'uint8_t[{}]'.format(width * height), ) with self.lock: status = _cvcapture.lib.cvcapture( self.instance, capture_buffer, width, height, ) if status == 0: raise RuntimeError("cvcapture() failed") return bytes(_cvcapture.ffi.buffer(capture_buffer)) def __enter__(self): return self def __exit__(self, exc_value, exc_type, exc_traceback): self.close() def close(self): if self.instance is not None: with self.lock: _cvcapture.lib.cvclose(self.instance) self.instance = None __del__ = close ## Instruction: Raise a `RuntimeError` if the device cannot be opened ## Code After: import threading from robotd.native import _cvcapture class CaptureDevice(object): def __init__(self, path=None): if path is not None: argument_c = _cvcapture.ffi.new( 'char[]', path.encode('utf-8'), ) else: argument_c = _cvcapture.ffi.NULL self.instance = _cvcapture.lib.cvopen(argument_c) if self.instance == _cvcapture.ffi.NULL: raise RuntimeError("Unable to open capture device") self.lock = threading.Lock() def capture(self, width, height): if self.instance is None: raise RuntimeError("capture device is closed") capture_buffer = _cvcapture.ffi.new( 'uint8_t[{}]'.format(width * height), ) with self.lock: status = _cvcapture.lib.cvcapture( self.instance, capture_buffer, width, height, ) if status == 0: raise RuntimeError("cvcapture() failed") return bytes(_cvcapture.ffi.buffer(capture_buffer)) def __enter__(self): return self def __exit__(self, exc_value, exc_type, exc_traceback): self.close() def close(self): if self.instance is not None: with self.lock: _cvcapture.lib.cvclose(self.instance) self.instance = None __del__ = close
import threading from robotd.native import _cvcapture class CaptureDevice(object): def __init__(self, path=None): if path is not None: argument_c = _cvcapture.ffi.new( 'char[]', path.encode('utf-8'), ) else: argument_c = _cvcapture.ffi.NULL self.instance = _cvcapture.lib.cvopen(argument_c) + if self.instance == _cvcapture.ffi.NULL: + raise RuntimeError("Unable to open capture device") self.lock = threading.Lock() def capture(self, width, height): if self.instance is None: raise RuntimeError("capture device is closed") capture_buffer = _cvcapture.ffi.new( 'uint8_t[{}]'.format(width * height), ) with self.lock: status = _cvcapture.lib.cvcapture( self.instance, capture_buffer, width, height, ) if status == 0: raise RuntimeError("cvcapture() failed") return bytes(_cvcapture.ffi.buffer(capture_buffer)) def __enter__(self): return self def __exit__(self, exc_value, exc_type, exc_traceback): self.close() def close(self): if self.instance is not None: with self.lock: _cvcapture.lib.cvclose(self.instance) self.instance = None __del__ = close
71df45002746b162e04a125403cad390accb949e
backend/main.py
backend/main.py
import logging from firebase import firebase from flask import Flask, jsonify, request import flask_cors from google.appengine.ext import ndb import google.auth.transport.requests import google.oauth2.id_token import requests_toolbelt.adapters.appengine requests_toolbelt.adapters.appengine.monkeypatch() HTTP_REQUEST = google.auth.transport.requests.Request() app = Flask(__name__) firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None) flask_cors.CORS(app) @app.route('/') def index(): return "<h1>Welcome To Google HVZ (backend)!</h1>" @app.route('/test', methods=['GET']) def get_testdata(): testdata = firebase.get('testdata', None) return jsonify(testdata)
import logging from firebase import firebase from flask import Flask, jsonify, request import flask_cors from google.appengine.ext import ndb import google.auth.transport.requests import google.oauth2.id_token import requests_toolbelt.adapters.appengine requests_toolbelt.adapters.appengine.monkeypatch() HTTP_REQUEST = google.auth.transport.requests.Request() # Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database FIREBASE_SECRET = "" FIREBASE_EMAIL = "" app = Flask(__name__) auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True) firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth) flask_cors.CORS(app) @app.route('/') def index(): return "<h1>Welcome To Google HVZ (backend)!</h1>" @app.route('/test', methods=['GET']) def get_testdata(): testdata = firebase.get('testdata', None) return jsonify(testdata)
Add proper authentication for db (without actual key).
Add proper authentication for db (without actual key).
Python
apache-2.0
google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz,google/playhvz
import logging from firebase import firebase from flask import Flask, jsonify, request import flask_cors from google.appengine.ext import ndb import google.auth.transport.requests import google.oauth2.id_token import requests_toolbelt.adapters.appengine requests_toolbelt.adapters.appengine.monkeypatch() HTTP_REQUEST = google.auth.transport.requests.Request() + # Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database + FIREBASE_SECRET = "" + FIREBASE_EMAIL = "" + app = Flask(__name__) + auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True) - firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None) + firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth) flask_cors.CORS(app) @app.route('/') def index(): return "<h1>Welcome To Google HVZ (backend)!</h1>" @app.route('/test', methods=['GET']) def get_testdata(): testdata = firebase.get('testdata', None) return jsonify(testdata)
Add proper authentication for db (without actual key).
## Code Before: import logging from firebase import firebase from flask import Flask, jsonify, request import flask_cors from google.appengine.ext import ndb import google.auth.transport.requests import google.oauth2.id_token import requests_toolbelt.adapters.appengine requests_toolbelt.adapters.appengine.monkeypatch() HTTP_REQUEST = google.auth.transport.requests.Request() app = Flask(__name__) firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None) flask_cors.CORS(app) @app.route('/') def index(): return "<h1>Welcome To Google HVZ (backend)!</h1>" @app.route('/test', methods=['GET']) def get_testdata(): testdata = firebase.get('testdata', None) return jsonify(testdata) ## Instruction: Add proper authentication for db (without actual key). ## Code After: import logging from firebase import firebase from flask import Flask, jsonify, request import flask_cors from google.appengine.ext import ndb import google.auth.transport.requests import google.oauth2.id_token import requests_toolbelt.adapters.appengine requests_toolbelt.adapters.appengine.monkeypatch() HTTP_REQUEST = google.auth.transport.requests.Request() # Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database FIREBASE_SECRET = "" FIREBASE_EMAIL = "" app = Flask(__name__) auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True) firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth) flask_cors.CORS(app) @app.route('/') def index(): return "<h1>Welcome To Google HVZ (backend)!</h1>" @app.route('/test', methods=['GET']) def get_testdata(): testdata = firebase.get('testdata', None) return jsonify(testdata)
import logging from firebase import firebase from flask import Flask, jsonify, request import flask_cors from google.appengine.ext import ndb import google.auth.transport.requests import google.oauth2.id_token import requests_toolbelt.adapters.appengine requests_toolbelt.adapters.appengine.monkeypatch() HTTP_REQUEST = google.auth.transport.requests.Request() + # Fill out with value from https://firebase.corp.google.com/project/trogdors-29fa4/settings/database + FIREBASE_SECRET = "" + FIREBASE_EMAIL = "" + app = Flask(__name__) + auth = firebase.FirebaseAuthentication(FIREBASE_SECRET, FIREBASE_EMAIL, admin=True) - firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', None) ? ^ ^ + firebase = firebase.FirebaseApplication('https://trogdors-29fa4.firebaseio.com', authentication=auth) ? ^^^^^^^^^^^^ ^^^^^ flask_cors.CORS(app) @app.route('/') def index(): return "<h1>Welcome To Google HVZ (backend)!</h1>" @app.route('/test', methods=['GET']) def get_testdata(): testdata = firebase.get('testdata', None) return jsonify(testdata)
ca953b2ef7662e4a70eba386e66ed6d66fad4eec
setup.py
setup.py
import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.1", description = "Store and access your passwords safely.", url = "http://keyring-python.org/", keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README.txt').read(), platforms = ["Many"], packages = ['keyring'], ext_modules = get_extensions() )
import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.1", description = "Store and access your passwords safely.", url = "http://home.python-keyring.org/", keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README.txt').read(), platforms = ["Many"], packages = ['keyring'], ext_modules = get_extensions() )
Fix the error in the home page URL.
Fix the error in the home page URL.
Python
mit
jaraco/keyring
import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.1", description = "Store and access your passwords safely.", - url = "http://keyring-python.org/", + url = "http://home.python-keyring.org/", keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README.txt').read(), platforms = ["Many"], packages = ['keyring'], ext_modules = get_extensions() )
Fix the error in the home page URL.
## Code Before: import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.1", description = "Store and access your passwords safely.", url = "http://keyring-python.org/", keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README.txt').read(), platforms = ["Many"], packages = ['keyring'], ext_modules = get_extensions() ) ## Instruction: Fix the error in the home page URL. ## Code After: import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.1", description = "Store and access your passwords safely.", url = "http://home.python-keyring.org/", keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README.txt').read(), platforms = ["Many"], packages = ['keyring'], ext_modules = get_extensions() )
import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.1", description = "Store and access your passwords safely.", - url = "http://keyring-python.org/", ? ------- + url = "http://home.python-keyring.org/", ? ++++++++++++ keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README.txt').read(), platforms = ["Many"], packages = ['keyring'], ext_modules = get_extensions() )
0ee2b337b61155044a66ae1f6f173492a51c1150
dipsim/fluorophore.py
dipsim/fluorophore.py
import numpy as np class Fluorophore: """A single fluorophore is specified by its 3D position, (unit) absorption dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi). """ def __init__(self, position=np.array([0, 0, 0]), mu_abs=np.array([0, 0]), mu_em=np.array([0, 0])): self.position = position self.mu_abs = mu_abs self.mu_em = mu_em self.mu_ind = 0
import numpy as np class Fluorophore: """A fluorophore is specified by its orientation (in theta and phi spherical coordinates), it distribution (using a kappa watson distribution), and a constant (c) proportional to the fluorohphore's brightness. """ def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0): self.theta = theta self.phi = phi self.kappa = kappa self.c = c
Modify Fluorophore for more convenient coordinates.
Modify Fluorophore for more convenient coordinates.
Python
mit
talonchandler/dipsim,talonchandler/dipsim
import numpy as np class Fluorophore: - """A single fluorophore is specified by its 3D position, (unit) absorption - dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi). + """A fluorophore is specified by its orientation (in theta and phi spherical + coordinates), it distribution (using a kappa watson distribution), and a + constant (c) proportional to the fluorohphore's brightness. """ + def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0): + self.theta = theta - def __init__(self, position=np.array([0, 0, 0]), - mu_abs=np.array([0, 0]), - mu_em=np.array([0, 0])): - self.position = position - self.mu_abs = mu_abs - self.mu_em = mu_em - self.mu_ind = 0 + self.phi = phi + self.kappa = kappa + self.c = c
Modify Fluorophore for more convenient coordinates.
## Code Before: import numpy as np class Fluorophore: """A single fluorophore is specified by its 3D position, (unit) absorption dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi). """ def __init__(self, position=np.array([0, 0, 0]), mu_abs=np.array([0, 0]), mu_em=np.array([0, 0])): self.position = position self.mu_abs = mu_abs self.mu_em = mu_em self.mu_ind = 0 ## Instruction: Modify Fluorophore for more convenient coordinates. ## Code After: import numpy as np class Fluorophore: """A fluorophore is specified by its orientation (in theta and phi spherical coordinates), it distribution (using a kappa watson distribution), and a constant (c) proportional to the fluorohphore's brightness. """ def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0): self.theta = theta self.phi = phi self.kappa = kappa self.c = c
import numpy as np class Fluorophore: - """A single fluorophore is specified by its 3D position, (unit) absorption - dipole moment (theta, phi), and (unit) emission dipole moment (theta, phi). + """A fluorophore is specified by its orientation (in theta and phi spherical + coordinates), it distribution (using a kappa watson distribution), and a + constant (c) proportional to the fluorohphore's brightness. """ - def __init__(self, position=np.array([0, 0, 0]), - mu_abs=np.array([0, 0]), - mu_em=np.array([0, 0])): - self.position = position + def __init__(self, theta=np.pi/2, phi=0, kappa=None, c=1.0): + self.theta = theta + self.phi = phi + self.kappa = kappa - self.mu_abs = mu_abs ? ^^^^^^ ^^^^^^ + self.c = c ? ^ ^ - self.mu_em = mu_em - self.mu_ind = 0
075d6f1b8f232c1ae7cb7d288da8f8d1040f49c9
hooks/pre_gen_project.py
hooks/pre_gen_project.py
repo_name = '{{ cookiecutter.repo_name }}' assert_msg = 'Repo name should be valid Python identifier!' if hasattr(repo_name, 'isidentifier'): assert repo_name.isidentifier(), assert_msg else: import re identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$") assert bool(identifier_re.match(repo_name)), assert_msg
import sys import cookiecutter # Ensure cookiecutter is recent enough cookiecutter_min_version = '1.3.0' if cookiecutter.__version__ < cookiecutter_min_version: print("--------------------------------------------------------------") print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version) print("--------------------------------------------------------------") sys.exit(1) # Ensure the selected repo name is usable repo_name = '{{ cookiecutter.repo_name }}' assert_msg = 'Repo name should be valid Python identifier!' if hasattr(repo_name, 'isidentifier'): assert repo_name.isidentifier(), assert_msg else: import re identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$") assert bool(identifier_re.match(repo_name)), assert_msg
Add check for cookiecutter version - at least 1.3.0 is required now
Add check for cookiecutter version - at least 1.3.0 is required now
Python
isc
thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template
+ import sys + + import cookiecutter + + + # Ensure cookiecutter is recent enough + cookiecutter_min_version = '1.3.0' + if cookiecutter.__version__ < cookiecutter_min_version: + print("--------------------------------------------------------------") + print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version) + print("--------------------------------------------------------------") + sys.exit(1) + + + # Ensure the selected repo name is usable repo_name = '{{ cookiecutter.repo_name }}' assert_msg = 'Repo name should be valid Python identifier!' if hasattr(repo_name, 'isidentifier'): assert repo_name.isidentifier(), assert_msg else: import re identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$") assert bool(identifier_re.match(repo_name)), assert_msg
Add check for cookiecutter version - at least 1.3.0 is required now
## Code Before: repo_name = '{{ cookiecutter.repo_name }}' assert_msg = 'Repo name should be valid Python identifier!' if hasattr(repo_name, 'isidentifier'): assert repo_name.isidentifier(), assert_msg else: import re identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$") assert bool(identifier_re.match(repo_name)), assert_msg ## Instruction: Add check for cookiecutter version - at least 1.3.0 is required now ## Code After: import sys import cookiecutter # Ensure cookiecutter is recent enough cookiecutter_min_version = '1.3.0' if cookiecutter.__version__ < cookiecutter_min_version: print("--------------------------------------------------------------") print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version) print("--------------------------------------------------------------") sys.exit(1) # Ensure the selected repo name is usable repo_name = '{{ cookiecutter.repo_name }}' assert_msg = 'Repo name should be valid Python identifier!' if hasattr(repo_name, 'isidentifier'): assert repo_name.isidentifier(), assert_msg else: import re identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$") assert bool(identifier_re.match(repo_name)), assert_msg
+ import sys + + import cookiecutter + + + # Ensure cookiecutter is recent enough + cookiecutter_min_version = '1.3.0' + if cookiecutter.__version__ < cookiecutter_min_version: + print("--------------------------------------------------------------") + print("!! Your cookiecutter is too old, at least %s is required !!" % cookiecutter_min_version) + print("--------------------------------------------------------------") + sys.exit(1) + + + # Ensure the selected repo name is usable repo_name = '{{ cookiecutter.repo_name }}' assert_msg = 'Repo name should be valid Python identifier!' if hasattr(repo_name, 'isidentifier'): assert repo_name.isidentifier(), assert_msg else: import re identifier_re = re.compile(r"[a-zA-Z_][a-zA-Z0-9_]*$") assert bool(identifier_re.match(repo_name)), assert_msg
f1d48525f1e8cde2af9a49636f38360b87b0ecb6
function/univariate_function.py
function/univariate_function.py
from abc import ABC, abstractmethod class UnivariateFunction(ABC): """ Class that defines a generic (abstract) univariate function. """ def __init(self): super().__init__() @abstractmethod def eval(self, v): pass @property @abstractmethod def domain_start(self): pass @property @abstractmethod def domain_end(self): pass
from abc import ABC, abstractmethod class UnivariateFunction(ABC): """ Class that defines a generic (abstract) univariate function. """ def __init(self): super().__init__() @abstractmethod def eval(self, v): """ Evaluate the univariate function with input v, and return that value :param v: Typically some kind of numeric. :return: """ pass @property @abstractmethod def domain_start(self): """ Return the start value of the domain. :return: """ pass @property @abstractmethod def domain_end(self): """ Return the end value of the domain. :return: """ pass
Add comments to abstract methods.
Add comments to abstract methods.
Python
mit
dpazel/music_rep
from abc import ABC, abstractmethod class UnivariateFunction(ABC): """ Class that defines a generic (abstract) univariate function. """ def __init(self): super().__init__() @abstractmethod def eval(self, v): + """ + Evaluate the univariate function with input v, and return that value + :param v: Typically some kind of numeric. + :return: + """ pass @property @abstractmethod def domain_start(self): + """ + Return the start value of the domain. + :return: + """ pass @property @abstractmethod def domain_end(self): + """ + Return the end value of the domain. + :return: + """ pass
Add comments to abstract methods.
## Code Before: from abc import ABC, abstractmethod class UnivariateFunction(ABC): """ Class that defines a generic (abstract) univariate function. """ def __init(self): super().__init__() @abstractmethod def eval(self, v): pass @property @abstractmethod def domain_start(self): pass @property @abstractmethod def domain_end(self): pass ## Instruction: Add comments to abstract methods. ## Code After: from abc import ABC, abstractmethod class UnivariateFunction(ABC): """ Class that defines a generic (abstract) univariate function. """ def __init(self): super().__init__() @abstractmethod def eval(self, v): """ Evaluate the univariate function with input v, and return that value :param v: Typically some kind of numeric. :return: """ pass @property @abstractmethod def domain_start(self): """ Return the start value of the domain. :return: """ pass @property @abstractmethod def domain_end(self): """ Return the end value of the domain. :return: """ pass
from abc import ABC, abstractmethod class UnivariateFunction(ABC): """ Class that defines a generic (abstract) univariate function. """ def __init(self): super().__init__() @abstractmethod def eval(self, v): + """ + Evaluate the univariate function with input v, and return that value + :param v: Typically some kind of numeric. + :return: + """ pass @property @abstractmethod def domain_start(self): + """ + Return the start value of the domain. + :return: + """ pass @property @abstractmethod def domain_end(self): + """ + Return the end value of the domain. + :return: + """ pass
dbfc1a11c0ced8ab7a997944e943a17ff0069199
lexer.py
lexer.py
import ply.lex as lex import re tokens = ( 'SECTION', 'IDENTIFIER', 'LBRACE', 'RBRACE', 'SEMI', 'TRUE', 'FALSE' ) def t_SECTION(t): r'section' return t def t_TRUE(t): r'(true|1)' return t def t_FALSE(t): r'(false|0)' return t def t_IDENTIFIER(t): r'[a-zA-Z\-0-9]+' return t def t_LBRACE(t): r'{' return t def t_RBRACE(t): r'}' return t def t_SEMI(t): r';' return t def t_NEWLINE(t): r'\n+' t.lexer.lineno += len(t.value) return t t_ignore = ' \t\n' # Error handling rule def t_error(t): print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno)) t.lexer.skip(1) lexer = lex.lex()
import ply.lex as lex import re tokens = ( 'SECTION', 'IDENTIFIER', 'STRING', 'LBRACE', 'RBRACE', 'SEMI', 'EQU', 'TRUE', 'FALSE' ) def t_SECTION(t): r'section' return t def t_TRUE(t): r'(true|1)' return t def t_FALSE(t): r'(false|0)' return t def t_IDENTIFIER(t): r'[a-zA-Z\-0-9]+' return t def t_STRING(t): r'(\".*\"|\'.*\')' t.value = t.value[1:-1] return t def t_LBRACE(t): r'{' return t def t_EQU(t): r'=' return t def t_RBRACE(t): r'}' return t def t_SEMI(t): r';' return t def t_NEWLINE(t): r'\n+' t.lexer.lineno += len(t.value) return t t_ignore = ' \t\n' # Error handling rule def t_error(t): print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno)) t.lexer.skip(1) lexer = lex.lex()
Add rudimentary string support and a token for '=' sign.
Add rudimentary string support and a token for '=' sign.
Python
cc0-1.0
dmbaturin/ply-example
import ply.lex as lex import re tokens = ( 'SECTION', 'IDENTIFIER', + 'STRING', 'LBRACE', 'RBRACE', 'SEMI', + 'EQU', 'TRUE', 'FALSE' ) def t_SECTION(t): r'section' return t def t_TRUE(t): r'(true|1)' return t def t_FALSE(t): r'(false|0)' return t def t_IDENTIFIER(t): r'[a-zA-Z\-0-9]+' return t + def t_STRING(t): + r'(\".*\"|\'.*\')' + t.value = t.value[1:-1] + return t + def t_LBRACE(t): r'{' + return t + + def t_EQU(t): + r'=' return t def t_RBRACE(t): r'}' return t def t_SEMI(t): r';' return t def t_NEWLINE(t): r'\n+' t.lexer.lineno += len(t.value) return t t_ignore = ' \t\n' # Error handling rule def t_error(t): print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno)) t.lexer.skip(1) lexer = lex.lex()
Add rudimentary string support and a token for '=' sign.
## Code Before: import ply.lex as lex import re tokens = ( 'SECTION', 'IDENTIFIER', 'LBRACE', 'RBRACE', 'SEMI', 'TRUE', 'FALSE' ) def t_SECTION(t): r'section' return t def t_TRUE(t): r'(true|1)' return t def t_FALSE(t): r'(false|0)' return t def t_IDENTIFIER(t): r'[a-zA-Z\-0-9]+' return t def t_LBRACE(t): r'{' return t def t_RBRACE(t): r'}' return t def t_SEMI(t): r';' return t def t_NEWLINE(t): r'\n+' t.lexer.lineno += len(t.value) return t t_ignore = ' \t\n' # Error handling rule def t_error(t): print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno)) t.lexer.skip(1) lexer = lex.lex() ## Instruction: Add rudimentary string support and a token for '=' sign. ## Code After: import ply.lex as lex import re tokens = ( 'SECTION', 'IDENTIFIER', 'STRING', 'LBRACE', 'RBRACE', 'SEMI', 'EQU', 'TRUE', 'FALSE' ) def t_SECTION(t): r'section' return t def t_TRUE(t): r'(true|1)' return t def t_FALSE(t): r'(false|0)' return t def t_IDENTIFIER(t): r'[a-zA-Z\-0-9]+' return t def t_STRING(t): r'(\".*\"|\'.*\')' t.value = t.value[1:-1] return t def t_LBRACE(t): r'{' return t def t_EQU(t): r'=' return t def t_RBRACE(t): r'}' return t def t_SEMI(t): r';' return t def t_NEWLINE(t): r'\n+' t.lexer.lineno += len(t.value) return t t_ignore = ' \t\n' # Error handling rule def t_error(t): print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno)) t.lexer.skip(1) lexer = lex.lex()
import ply.lex as lex import re tokens = ( 'SECTION', 'IDENTIFIER', + 'STRING', 'LBRACE', 'RBRACE', 'SEMI', + 'EQU', 'TRUE', 'FALSE' ) def t_SECTION(t): r'section' return t def t_TRUE(t): r'(true|1)' return t def t_FALSE(t): r'(false|0)' return t def t_IDENTIFIER(t): r'[a-zA-Z\-0-9]+' return t + def t_STRING(t): + r'(\".*\"|\'.*\')' + t.value = t.value[1:-1] + return t + def t_LBRACE(t): r'{' + return t + + def t_EQU(t): + r'=' return t def t_RBRACE(t): r'}' return t def t_SEMI(t): r';' return t def t_NEWLINE(t): r'\n+' t.lexer.lineno += len(t.value) return t t_ignore = ' \t\n' # Error handling rule def t_error(t): print("Illegal character '{0}' at line {1}".format(t.value[0], t.lineno)) t.lexer.skip(1) lexer = lex.lex()
67ca9f09cd2cfb5e646b9a09b540c5ff88276201
pydirections/models/models.py
pydirections/models/models.py
from schematics.models import Model from schematics.types import StringType class Step(Model): """ Represents an individual step """ html_instructions = StringType() class Leg(Model): """ Represents an individual leg """ start_address = StringType() end_address = StringType() class Route(Model): """ Represents an individual route whose attributes include """ summary = StringType(required=True) copyrights = StringType() @property def summary(): return summary
from schematics.models import Model from schematics.types import StringType, DecimalType from schematics.types.compound import ListType class Distance(Model): """ Represents the duration of a leg/step """ value = DecimalType() text = StringType() class Duration(Model): """ Represents the duration of a leg/step """ value = DecimalType() text = StringType() class Step(Model): """ Represents an individual step """ html_instructions = StringType() class Leg(Model): """ Represents an individual leg """ start_address = StringType() end_address = StringType() steps = ListType(ModelType(Step)) class Route(Model): """ Represents an individual route whose attributes include """ summary = StringType(required=True) legs = ListType(ModelType(Leg)) copyrights = StringType() @property def summary(): return summary
Add more details to routes
Add more details to routes
Python
apache-2.0
apranav19/pydirections
from schematics.models import Model - from schematics.types import StringType + from schematics.types import StringType, DecimalType + from schematics.types.compound import ListType + + class Distance(Model): + """ + Represents the duration of a leg/step + """ + value = DecimalType() + text = StringType() + + class Duration(Model): + """ + Represents the duration of a leg/step + """ + value = DecimalType() + text = StringType() + class Step(Model): """ Represents an individual step """ html_instructions = StringType() class Leg(Model): """ Represents an individual leg """ start_address = StringType() end_address = StringType() - + steps = ListType(ModelType(Step)) + class Route(Model): """ Represents an individual route whose attributes include """ summary = StringType(required=True) + legs = ListType(ModelType(Leg)) copyrights = StringType() @property def summary(): return summary +
Add more details to routes
## Code Before: from schematics.models import Model from schematics.types import StringType class Step(Model): """ Represents an individual step """ html_instructions = StringType() class Leg(Model): """ Represents an individual leg """ start_address = StringType() end_address = StringType() class Route(Model): """ Represents an individual route whose attributes include """ summary = StringType(required=True) copyrights = StringType() @property def summary(): return summary ## Instruction: Add more details to routes ## Code After: from schematics.models import Model from schematics.types import StringType, DecimalType from schematics.types.compound import ListType class Distance(Model): """ Represents the duration of a leg/step """ value = DecimalType() text = StringType() class Duration(Model): """ Represents the duration of a leg/step """ value = DecimalType() text = StringType() class Step(Model): """ Represents an individual step """ html_instructions = StringType() class Leg(Model): """ Represents an individual leg """ start_address = StringType() end_address = StringType() steps = ListType(ModelType(Step)) class Route(Model): """ Represents an individual route whose attributes include """ summary = StringType(required=True) legs = ListType(ModelType(Leg)) copyrights = StringType() @property def summary(): return summary
from schematics.models import Model - from schematics.types import StringType + from schematics.types import StringType, DecimalType ? +++++++++++++ + from schematics.types.compound import ListType + + class Distance(Model): + """ + Represents the duration of a leg/step + """ + value = DecimalType() + text = StringType() + + class Duration(Model): + """ + Represents the duration of a leg/step + """ + value = DecimalType() + text = StringType() + class Step(Model): """ Represents an individual step """ html_instructions = StringType() class Leg(Model): """ Represents an individual leg """ start_address = StringType() end_address = StringType() - + steps = ListType(ModelType(Step)) + class Route(Model): """ Represents an individual route whose attributes include """ summary = StringType(required=True) + legs = ListType(ModelType(Leg)) copyrights = StringType() @property def summary(): return summary
c87c4a972f0f2d4966142fa666a900112762ed76
scipy/constants/tests/test_codata.py
scipy/constants/tests/test_codata.py
import warnings from scipy.constants import find from numpy.testing import assert_equal def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time']))
import warnings from scipy.constants import find from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) if __name__ == "__main__": run_module_suite()
Allow codata tests to be run as script.
ENH: Allow codata tests to be run as script.
Python
bsd-3-clause
zerothi/scipy,zxsted/scipy,josephcslater/scipy,rgommers/scipy,grlee77/scipy,sargas/scipy,dch312/scipy,ilayn/scipy,apbard/scipy,jakevdp/scipy,niknow/scipy,vanpact/scipy,jakevdp/scipy,rmcgibbo/scipy,zxsted/scipy,pnedunuri/scipy,raoulbq/scipy,lhilt/scipy,mgaitan/scipy,mingwpy/scipy,maciejkula/scipy,njwilson23/scipy,Dapid/scipy,woodscn/scipy,perimosocordiae/scipy,aarchiba/scipy,pbrod/scipy,aarchiba/scipy,trankmichael/scipy,Srisai85/scipy,aarchiba/scipy,pschella/scipy,tylerjereddy/scipy,minhlongdo/scipy,mortonjt/scipy,dominicelse/scipy,jsilter/scipy,sargas/scipy,vanpact/scipy,anielsen001/scipy,mingwpy/scipy,matthew-brett/scipy,aeklant/scipy,mikebenfield/scipy,Srisai85/scipy,sauliusl/scipy,FRidh/scipy,andim/scipy,ortylp/scipy,josephcslater/scipy,mhogg/scipy,WarrenWeckesser/scipy,ndchorley/scipy,josephcslater/scipy,mortonjt/scipy,felipebetancur/scipy,jonycgn/scipy,mtrbean/scipy,juliantaylor/scipy,mdhaber/scipy,mdhaber/scipy,kleskjr/scipy,ogrisel/scipy,teoliphant/scipy,petebachant/scipy,lukauskas/scipy,Gillu13/scipy,sonnyhu/scipy,WillieMaddox/scipy,arokem/scipy,jonycgn/scipy,gdooper/scipy,ChanderG/scipy,pbrod/scipy,surhudm/scipy,rgommers/scipy,anntzer/scipy,gfyoung/scipy,sriki18/scipy,nmayorov/scipy,ogrisel/scipy,nvoron23/scipy,giorgiop/scipy,trankmichael/scipy,scipy/scipy,dominicelse/scipy,petebachant/scipy,ChanderG/scipy,ndchorley/scipy,scipy/scipy,Stefan-Endres/scipy,futurulus/scipy,giorgiop/scipy,behzadnouri/scipy,sargas/scipy,haudren/scipy,Gillu13/scipy,teoliphant/scipy,WillieMaddox/scipy,cpaulik/scipy,juliantaylor/scipy,sriki18/scipy,ortylp/scipy,zerothi/scipy,gdooper/scipy,befelix/scipy,raoulbq/scipy,sargas/scipy,newemailjdm/scipy,piyush0609/scipy,vhaasteren/scipy,sriki18/scipy,sauliusl/scipy,rgommers/scipy,zaxliu/scipy,Stefan-Endres/scipy,dch312/scipy,scipy/scipy,aeklant/scipy,cpaulik/scipy,maniteja123/scipy,zerothi/scipy,cpaulik/scipy,Eric89GXL/scipy,mgaitan/scipy,bkendzior/scipy,jonycgn/scipy,juliantaylor/scipy,felipebetancur/scipy,giorgiop/scipy,gertingold/scipy,befelix/scipy,maciejkula/scipy,vberaudi/scipy,lhilt/scipy,ChanderG/scipy,mgaitan/scipy,hainm/scipy,endolith/scipy,lukauskas/scipy,gdooper/scipy,kalvdans/scipy,richardotis/scipy,jor-/scipy,mgaitan/scipy,mtrbean/scipy,Gillu13/scipy,raoulbq/scipy,lukauskas/scipy,tylerjereddy/scipy,person142/scipy,mortada/scipy,vhaasteren/scipy,Shaswat27/scipy,vigna/scipy,pschella/scipy,surhudm/scipy,kalvdans/scipy,witcxc/scipy,piyush0609/scipy,woodscn/scipy,matthew-brett/scipy,nvoron23/scipy,matthewalbani/scipy,mdhaber/scipy,gfyoung/scipy,pnedunuri/scipy,vanpact/scipy,ilayn/scipy,FRidh/scipy,mingwpy/scipy,efiring/scipy,jakevdp/scipy,Shaswat27/scipy,matthewalbani/scipy,pizzathief/scipy,mdhaber/scipy,jonycgn/scipy,pyramania/scipy,apbard/scipy,Dapid/scipy,rmcgibbo/scipy,vhaasteren/scipy,niknow/scipy,apbard/scipy,sonnyhu/scipy,larsmans/scipy,bkendzior/scipy,perimosocordiae/scipy,maniteja123/scipy,mortada/scipy,felipebetancur/scipy,andyfaff/scipy,mikebenfield/scipy,jseabold/scipy,mtrbean/scipy,jamestwebber/scipy,anielsen001/scipy,ortylp/scipy,Gillu13/scipy,woodscn/scipy,witcxc/scipy,haudren/scipy,witcxc/scipy,andyfaff/scipy,newemailjdm/scipy,vhaasteren/scipy,njwilson23/scipy,grlee77/scipy,surhudm/scipy,zaxliu/scipy,ortylp/scipy,njwilson23/scipy,dch312/scipy,fredrikw/scipy,pnedunuri/scipy,mortada/scipy,mortada/scipy,fredrikw/scipy,gef756/scipy,zaxliu/scipy,chatcannon/scipy,matthew-brett/scipy,jsilter/scipy,ilayn/scipy,mortada/scipy,Stefan-Endres/scipy,jseabold/scipy,minhlongdo/scipy,anntzer/scipy,andyfaff/scipy,futurulus/scipy,sauliusl/scipy,e-q/scipy,mingwpy/scipy,rmcgibbo/scipy,Dapid/scipy,vigna/scipy,minhlongdo/scipy,mdhaber/scipy,raoulbq/scipy,pyramania/scipy,richardotis/scipy,jor-/scipy,befelix/scipy,sriki18/scipy,mortonjt/scipy,endolith/scipy,haudren/scipy,juliantaylor/scipy,hainm/scipy,e-q/scipy,pizzathief/scipy,endolith/scipy,FRidh/scipy,argriffing/scipy,perimosocordiae/scipy,tylerjereddy/scipy,fernand/scipy,tylerjereddy/scipy,argriffing/scipy,vigna/scipy,mhogg/scipy,vanpact/scipy,woodscn/scipy,Eric89GXL/scipy,vberaudi/scipy,ChanderG/scipy,raoulbq/scipy,gef756/scipy,sauliusl/scipy,anntzer/scipy,jseabold/scipy,pbrod/scipy,argriffing/scipy,nvoron23/scipy,ales-erjavec/scipy,ortylp/scipy,Newman101/scipy,jjhelmus/scipy,petebachant/scipy,mingwpy/scipy,chatcannon/scipy,Eric89GXL/scipy,Shaswat27/scipy,ogrisel/scipy,anntzer/scipy,andyfaff/scipy,aman-iitj/scipy,cpaulik/scipy,sauliusl/scipy,zerothi/scipy,hainm/scipy,Srisai85/scipy,scipy/scipy,mtrbean/scipy,fredrikw/scipy,endolith/scipy,Dapid/scipy,ndchorley/scipy,giorgiop/scipy,felipebetancur/scipy,Gillu13/scipy,maciejkula/scipy,jonycgn/scipy,vanpact/scipy,fernand/scipy,argriffing/scipy,aarchiba/scipy,jjhelmus/scipy,newemailjdm/scipy,matthewalbani/scipy,behzadnouri/scipy,jamestwebber/scipy,behzadnouri/scipy,andim/scipy,pschella/scipy,nvoron23/scipy,ndchorley/scipy,Shaswat27/scipy,argriffing/scipy,andyfaff/scipy,juliantaylor/scipy,felipebetancur/scipy,mtrbean/scipy,fredrikw/scipy,sonnyhu/scipy,efiring/scipy,e-q/scipy,jsilter/scipy,mtrbean/scipy,behzadnouri/scipy,grlee77/scipy,ChanderG/scipy,maciejkula/scipy,ogrisel/scipy,maciejkula/scipy,Srisai85/scipy,jamestwebber/scipy,woodscn/scipy,pschella/scipy,vigna/scipy,dch312/scipy,lhilt/scipy,haudren/scipy,nmayorov/scipy,futurulus/scipy,Newman101/scipy,larsmans/scipy,Eric89GXL/scipy,larsmans/scipy,maniteja123/scipy,maniteja123/scipy,gfyoung/scipy,fernand/scipy,sargas/scipy,nmayorov/scipy,lhilt/scipy,richardotis/scipy,vanpact/scipy,teoliphant/scipy,arokem/scipy,aeklant/scipy,pyramania/scipy,jamestwebber/scipy,vberaudi/scipy,dominicelse/scipy,efiring/scipy,rmcgibbo/scipy,ales-erjavec/scipy,Eric89GXL/scipy,niknow/scipy,zaxliu/scipy,surhudm/scipy,maniteja123/scipy,mikebenfield/scipy,gef756/scipy,vberaudi/scipy,efiring/scipy,sriki18/scipy,FRidh/scipy,gdooper/scipy,futurulus/scipy,cpaulik/scipy,WarrenWeckesser/scipy,larsmans/scipy,person142/scipy,teoliphant/scipy,kleskjr/scipy,befelix/scipy,arokem/scipy,Kamp9/scipy,kleskjr/scipy,teoliphant/scipy,kleskjr/scipy,dominicelse/scipy,ndchorley/scipy,anntzer/scipy,jjhelmus/scipy,pbrod/scipy,ilayn/scipy,Shaswat27/scipy,efiring/scipy,njwilson23/scipy,niknow/scipy,dominicelse/scipy,behzadnouri/scipy,mingwpy/scipy,rgommers/scipy,FRidh/scipy,person142/scipy,mortonjt/scipy,vhaasteren/scipy,pbrod/scipy,nvoron23/scipy,jor-/scipy,niknow/scipy,trankmichael/scipy,fernand/scipy,haudren/scipy,njwilson23/scipy,matthew-brett/scipy,nvoron23/scipy,endolith/scipy,e-q/scipy,jonycgn/scipy,dch312/scipy,WillieMaddox/scipy,rmcgibbo/scipy,lukauskas/scipy,andim/scipy,jakevdp/scipy,zerothi/scipy,witcxc/scipy,ales-erjavec/scipy,ales-erjavec/scipy,jor-/scipy,grlee77/scipy,nonhermitian/scipy,aman-iitj/scipy,vhaasteren/scipy,chatcannon/scipy,arokem/scipy,giorgiop/scipy,futurulus/scipy,e-q/scipy,ChanderG/scipy,ilayn/scipy,WarrenWeckesser/scipy,mikebenfield/scipy,Newman101/scipy,richardotis/scipy,jsilter/scipy,perimosocordiae/scipy,gfyoung/scipy,scipy/scipy,kalvdans/scipy,sonnyhu/scipy,lukauskas/scipy,WillieMaddox/scipy,aman-iitj/scipy,Newman101/scipy,zerothi/scipy,apbard/scipy,jsilter/scipy,hainm/scipy,Stefan-Endres/scipy,maniteja123/scipy,petebachant/scipy,efiring/scipy,mhogg/scipy,Dapid/scipy,nmayorov/scipy,anielsen001/scipy,Dapid/scipy,aeklant/scipy,nonhermitian/scipy,perimosocordiae/scipy,gertingold/scipy,chatcannon/scipy,zxsted/scipy,trankmichael/scipy,behzadnouri/scipy,WillieMaddox/scipy,newemailjdm/scipy,bkendzior/scipy,vberaudi/scipy,anielsen001/scipy,argriffing/scipy,minhlongdo/scipy,aman-iitj/scipy,futurulus/scipy,pizzathief/scipy,aeklant/scipy,anielsen001/scipy,cpaulik/scipy,njwilson23/scipy,zaxliu/scipy,surhudm/scipy,Kamp9/scipy,Eric89GXL/scipy,mhogg/scipy,scipy/scipy,gfyoung/scipy,befelix/scipy,Newman101/scipy,ales-erjavec/scipy,raoulbq/scipy,zaxliu/scipy,pyramania/scipy,larsmans/scipy,mortada/scipy,chatcannon/scipy,grlee77/scipy,person142/scipy,Kamp9/scipy,giorgiop/scipy,kalvdans/scipy,Stefan-Endres/scipy,mhogg/scipy,mortonjt/scipy,Kamp9/scipy,zxsted/scipy,hainm/scipy,lhilt/scipy,haudren/scipy,sauliusl/scipy,aman-iitj/scipy,aarchiba/scipy,kleskjr/scipy,andyfaff/scipy,gef756/scipy,WarrenWeckesser/scipy,fredrikw/scipy,gertingold/scipy,petebachant/scipy,petebachant/scipy,endolith/scipy,Newman101/scipy,sonnyhu/scipy,trankmichael/scipy,josephcslater/scipy,vigna/scipy,kleskjr/scipy,matthewalbani/scipy,ortylp/scipy,Gillu13/scipy,gdooper/scipy,Kamp9/scipy,larsmans/scipy,mikebenfield/scipy,matthewalbani/scipy,Shaswat27/scipy,lukauskas/scipy,pizzathief/scipy,josephcslater/scipy,pnedunuri/scipy,matthew-brett/scipy,chatcannon/scipy,bkendzior/scipy,ilayn/scipy,jjhelmus/scipy,perimosocordiae/scipy,apbard/scipy,ogrisel/scipy,anntzer/scipy,rmcgibbo/scipy,surhudm/scipy,mdhaber/scipy,bkendzior/scipy,jseabold/scipy,sriki18/scipy,arokem/scipy,kalvdans/scipy,Kamp9/scipy,Srisai85/scipy,pnedunuri/scipy,pyramania/scipy,andim/scipy,fernand/scipy,gef756/scipy,trankmichael/scipy,nonhermitian/scipy,pnedunuri/scipy,niknow/scipy,piyush0609/scipy,richardotis/scipy,jamestwebber/scipy,Stefan-Endres/scipy,woodscn/scipy,minhlongdo/scipy,fredrikw/scipy,newemailjdm/scipy,WarrenWeckesser/scipy,ales-erjavec/scipy,andim/scipy,richardotis/scipy,piyush0609/scipy,andim/scipy,jor-/scipy,FRidh/scipy,nonhermitian/scipy,tylerjereddy/scipy,jseabold/scipy,gertingold/scipy,pizzathief/scipy,jseabold/scipy,zxsted/scipy,rgommers/scipy,gertingold/scipy,WarrenWeckesser/scipy,jakevdp/scipy,mortonjt/scipy,hainm/scipy,jjhelmus/scipy,piyush0609/scipy,WillieMaddox/scipy,ndchorley/scipy,zxsted/scipy,mgaitan/scipy,person142/scipy,minhlongdo/scipy,fernand/scipy,gef756/scipy,pbrod/scipy,pschella/scipy,anielsen001/scipy,vberaudi/scipy,aman-iitj/scipy,mhogg/scipy,witcxc/scipy,sonnyhu/scipy,piyush0609/scipy,mgaitan/scipy,nmayorov/scipy,newemailjdm/scipy,felipebetancur/scipy,Srisai85/scipy,nonhermitian/scipy
import warnings from scipy.constants import find - from numpy.testing import assert_equal + from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) + if __name__ == "__main__": + run_module_suite() +
Allow codata tests to be run as script.
## Code Before: import warnings from scipy.constants import find from numpy.testing import assert_equal def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) ## Instruction: Allow codata tests to be run as script. ## Code After: import warnings from scipy.constants import find from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) if __name__ == "__main__": run_module_suite()
import warnings from scipy.constants import find - from numpy.testing import assert_equal + from numpy.testing import assert_equal, run_module_suite ? ++++++++++++++++++ def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) + + if __name__ == "__main__": + run_module_suite()
380baa34af7e8a704780f0ec535b626f4a286e23
deflect/admin.py
deflect/admin.py
from django.contrib import admin from .models import RedirectURL class RedirectURLAdmin(admin.ModelAdmin): list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',) list_filter = ('creator__username', 'campaign', 'medium',) ordering = ('-last_used',) readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',) search_fields = ['url', 'campaign'] fieldsets = ((None, {'fields': ('url', 'short_url',)}), ('Google', {'fields': ('campaign', 'medium', 'content',)}), ('Additional info', {'fields': ('description', 'qr_code',)}), ('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),) def save_model(self, request, obj, form, change): obj.user = request.user obj.save() admin.site.register(RedirectURL, RedirectURLAdmin)
from django.contrib import admin from .models import RedirectURL class RedirectURLAdmin(admin.ModelAdmin): list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',) list_filter = ('creator__username', 'campaign', 'medium',) ordering = ('-last_used',) readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',) search_fields = ['url', 'campaign'] fieldsets = ((None, {'fields': ('url', 'short_url',)}), ('Google', {'fields': ('campaign', 'medium', 'content',)}), ('Additional info', {'fields': ('description', 'qr_code',)}), ('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),) def save_model(self, request, obj, form, change): if not change: obj.creator = request.user obj.save() admin.site.register(RedirectURL, RedirectURLAdmin)
Fix model creator updating on change event
Fix model creator updating on change event
Python
bsd-3-clause
jbittel/django-deflect
from django.contrib import admin from .models import RedirectURL class RedirectURLAdmin(admin.ModelAdmin): list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',) list_filter = ('creator__username', 'campaign', 'medium',) ordering = ('-last_used',) readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',) search_fields = ['url', 'campaign'] fieldsets = ((None, {'fields': ('url', 'short_url',)}), ('Google', {'fields': ('campaign', 'medium', 'content',)}), ('Additional info', {'fields': ('description', 'qr_code',)}), ('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),) def save_model(self, request, obj, form, change): + if not change: - obj.user = request.user + obj.creator = request.user obj.save() admin.site.register(RedirectURL, RedirectURLAdmin)
Fix model creator updating on change event
## Code Before: from django.contrib import admin from .models import RedirectURL class RedirectURLAdmin(admin.ModelAdmin): list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',) list_filter = ('creator__username', 'campaign', 'medium',) ordering = ('-last_used',) readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',) search_fields = ['url', 'campaign'] fieldsets = ((None, {'fields': ('url', 'short_url',)}), ('Google', {'fields': ('campaign', 'medium', 'content',)}), ('Additional info', {'fields': ('description', 'qr_code',)}), ('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),) def save_model(self, request, obj, form, change): obj.user = request.user obj.save() admin.site.register(RedirectURL, RedirectURLAdmin) ## Instruction: Fix model creator updating on change event ## Code After: from django.contrib import admin from .models import RedirectURL class RedirectURLAdmin(admin.ModelAdmin): list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',) list_filter = ('creator__username', 'campaign', 'medium',) ordering = ('-last_used',) readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',) search_fields = ['url', 'campaign'] fieldsets = ((None, {'fields': ('url', 'short_url',)}), ('Google', {'fields': ('campaign', 'medium', 'content',)}), ('Additional info', {'fields': ('description', 'qr_code',)}), ('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),) def save_model(self, request, obj, form, change): if not change: obj.creator = request.user obj.save() admin.site.register(RedirectURL, RedirectURLAdmin)
from django.contrib import admin from .models import RedirectURL class RedirectURLAdmin(admin.ModelAdmin): list_display = ('url', 'short_url', 'hits', 'last_used', 'creator', 'campaign', 'medium',) list_filter = ('creator__username', 'campaign', 'medium',) ordering = ('-last_used',) readonly_fields = ('created', 'short_url', 'qr_code', 'hits', 'last_used',) search_fields = ['url', 'campaign'] fieldsets = ((None, {'fields': ('url', 'short_url',)}), ('Google', {'fields': ('campaign', 'medium', 'content',)}), ('Additional info', {'fields': ('description', 'qr_code',)}), ('Short URL Usage', {'fields': ('hits', 'created', 'last_used',)}),) def save_model(self, request, obj, form, change): + if not change: - obj.user = request.user ? ^^ + obj.creator = request.user ? ++++ ^^ +++ obj.save() admin.site.register(RedirectURL, RedirectURLAdmin)
d46d0b5a5392b6ca047b519a9d6280b5b0581e81
system_maintenance/tests/functional/tests.py
system_maintenance/tests/functional/tests.py
from django.test import LiveServerTestCase from selenium import webdriver class FunctionalTest(LiveServerTestCase): def setUp(self): self.browser = webdriver.Firefox() self.browser.implicitly_wait(3) def tearDown(self): self.browser.quit() def test_app_home_title(self): self.browser.get('http://localhost:8000/system_maintenance') self.assertIn('System Maintenance', self.browser.title)
from django.contrib.staticfiles.testing import StaticLiveServerTestCase from selenium import webdriver class FunctionalTest(StaticLiveServerTestCase): def setUp(self): self.browser = webdriver.Firefox() self.browser.implicitly_wait(3) def tearDown(self): self.browser.quit() def test_app_home_title(self): self.browser.get('http://localhost:8000/system_maintenance') self.assertIn('System Maintenance', self.browser.title)
Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT'
Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT'
Python
bsd-3-clause
mfcovington/django-system-maintenance,mfcovington/django-system-maintenance,mfcovington/django-system-maintenance
- from django.test import LiveServerTestCase + from django.contrib.staticfiles.testing import StaticLiveServerTestCase from selenium import webdriver - class FunctionalTest(LiveServerTestCase): + class FunctionalTest(StaticLiveServerTestCase): def setUp(self): self.browser = webdriver.Firefox() self.browser.implicitly_wait(3) def tearDown(self): self.browser.quit() def test_app_home_title(self): self.browser.get('http://localhost:8000/system_maintenance') self.assertIn('System Maintenance', self.browser.title)
Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT'
## Code Before: from django.test import LiveServerTestCase from selenium import webdriver class FunctionalTest(LiveServerTestCase): def setUp(self): self.browser = webdriver.Firefox() self.browser.implicitly_wait(3) def tearDown(self): self.browser.quit() def test_app_home_title(self): self.browser.get('http://localhost:8000/system_maintenance') self.assertIn('System Maintenance', self.browser.title) ## Instruction: Switch to 'StaticLiveServerTestCase' to avoid having to set 'settings.STATIC_ROOT' ## Code After: from django.contrib.staticfiles.testing import StaticLiveServerTestCase from selenium import webdriver class FunctionalTest(StaticLiveServerTestCase): def setUp(self): self.browser = webdriver.Firefox() self.browser.implicitly_wait(3) def tearDown(self): self.browser.quit() def test_app_home_title(self): self.browser.get('http://localhost:8000/system_maintenance') self.assertIn('System Maintenance', self.browser.title)
- from django.test import LiveServerTestCase + from django.contrib.staticfiles.testing import StaticLiveServerTestCase from selenium import webdriver - class FunctionalTest(LiveServerTestCase): + class FunctionalTest(StaticLiveServerTestCase): ? ++++++ def setUp(self): self.browser = webdriver.Firefox() self.browser.implicitly_wait(3) def tearDown(self): self.browser.quit() def test_app_home_title(self): self.browser.get('http://localhost:8000/system_maintenance') self.assertIn('System Maintenance', self.browser.title)
09099ab106ae4c0695502e3632e4ac1c2f459566
apps/teams/bulk_actions.py
apps/teams/bulk_actions.py
from django.contrib.contenttypes.models import ContentType from subtitles.models import SubtitleLanguage from teams.signals import api_subtitles_approved from utils.csv_parser import UnicodeReader from videos.tasks import video_changed_tasks def complete_approve_tasks(tasks): lang_ct = ContentType.objects.get_for_model(SubtitleLanguage) video_ids = set() for task in tasks: task.do_complete_approve(lang_ct=lang_ct) api_subtitles_approved.send(task.get_subtitle_version()) video_ids.add(task.team_video.video_id) for video_id in video_ids: video_changed_tasks.delay(video_id) def add_videos_from_csv(team, user, csv_file): from .tasks import add_team_videos videos = [] fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript'] num_fields = len(fields) try: reader = UnicodeReader(csv_file) header = reader.next() if len(header) != num_fields: raise Exception() except: raise ValueError(u'CSV format is not valid') for row in reader: videos.append(dict(zip(fields, row))) add_team_videos.delay(team.pk, user.pk, videos)
from django.contrib.contenttypes.models import ContentType from subtitles.models import SubtitleLanguage from subtitles.signals import subtitles_published from teams.signals import api_subtitles_approved from utils.csv_parser import UnicodeReader from videos.tasks import video_changed_tasks def complete_approve_tasks(tasks): lang_ct = ContentType.objects.get_for_model(SubtitleLanguage) video_ids = set() for task in tasks: task.do_complete_approve(lang_ct=lang_ct) version = task.get_subtitle_version() api_subtitles_approved.send(version) if version.is_public(): subtitles_published.send(version.subtitle_language, version=version) video_ids.add(task.team_video.video_id) for video_id in video_ids: video_changed_tasks.delay(video_id) def add_videos_from_csv(team, user, csv_file): from .tasks import add_team_videos videos = [] fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript'] num_fields = len(fields) try: reader = UnicodeReader(csv_file) header = reader.next() if len(header) != num_fields: raise Exception() except: raise ValueError(u'CSV format is not valid') for row in reader: videos.append(dict(zip(fields, row))) add_team_videos.delay(team.pk, user.pk, videos)
Send subtitles_published signal for bulk approvals
Send subtitles_published signal for bulk approvals This fixes pculture/amara-enterprise#608
Python
agpl-3.0
pculture/unisubs,pculture/unisubs,wevoice/wesub,pculture/unisubs,wevoice/wesub,wevoice/wesub,pculture/unisubs,wevoice/wesub
from django.contrib.contenttypes.models import ContentType from subtitles.models import SubtitleLanguage + from subtitles.signals import subtitles_published from teams.signals import api_subtitles_approved from utils.csv_parser import UnicodeReader from videos.tasks import video_changed_tasks def complete_approve_tasks(tasks): lang_ct = ContentType.objects.get_for_model(SubtitleLanguage) video_ids = set() for task in tasks: task.do_complete_approve(lang_ct=lang_ct) + version = task.get_subtitle_version() - api_subtitles_approved.send(task.get_subtitle_version()) + api_subtitles_approved.send(version) + if version.is_public(): + subtitles_published.send(version.subtitle_language, version=version) video_ids.add(task.team_video.video_id) for video_id in video_ids: video_changed_tasks.delay(video_id) def add_videos_from_csv(team, user, csv_file): from .tasks import add_team_videos videos = [] fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript'] num_fields = len(fields) try: reader = UnicodeReader(csv_file) header = reader.next() if len(header) != num_fields: raise Exception() except: raise ValueError(u'CSV format is not valid') for row in reader: videos.append(dict(zip(fields, row))) add_team_videos.delay(team.pk, user.pk, videos)
Send subtitles_published signal for bulk approvals
## Code Before: from django.contrib.contenttypes.models import ContentType from subtitles.models import SubtitleLanguage from teams.signals import api_subtitles_approved from utils.csv_parser import UnicodeReader from videos.tasks import video_changed_tasks def complete_approve_tasks(tasks): lang_ct = ContentType.objects.get_for_model(SubtitleLanguage) video_ids = set() for task in tasks: task.do_complete_approve(lang_ct=lang_ct) api_subtitles_approved.send(task.get_subtitle_version()) video_ids.add(task.team_video.video_id) for video_id in video_ids: video_changed_tasks.delay(video_id) def add_videos_from_csv(team, user, csv_file): from .tasks import add_team_videos videos = [] fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript'] num_fields = len(fields) try: reader = UnicodeReader(csv_file) header = reader.next() if len(header) != num_fields: raise Exception() except: raise ValueError(u'CSV format is not valid') for row in reader: videos.append(dict(zip(fields, row))) add_team_videos.delay(team.pk, user.pk, videos) ## Instruction: Send subtitles_published signal for bulk approvals ## Code After: from django.contrib.contenttypes.models import ContentType from subtitles.models import SubtitleLanguage from subtitles.signals import subtitles_published from teams.signals import api_subtitles_approved from utils.csv_parser import UnicodeReader from videos.tasks import video_changed_tasks def complete_approve_tasks(tasks): lang_ct = ContentType.objects.get_for_model(SubtitleLanguage) video_ids = set() for task in tasks: task.do_complete_approve(lang_ct=lang_ct) version = task.get_subtitle_version() api_subtitles_approved.send(version) if version.is_public(): subtitles_published.send(version.subtitle_language, version=version) video_ids.add(task.team_video.video_id) for video_id in video_ids: video_changed_tasks.delay(video_id) def add_videos_from_csv(team, user, csv_file): from .tasks import add_team_videos videos = [] fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript'] num_fields = len(fields) try: reader = UnicodeReader(csv_file) header = reader.next() if len(header) != num_fields: raise Exception() except: raise ValueError(u'CSV format is not valid') for row in reader: videos.append(dict(zip(fields, row))) add_team_videos.delay(team.pk, user.pk, videos)
from django.contrib.contenttypes.models import ContentType from subtitles.models import SubtitleLanguage + from subtitles.signals import subtitles_published from teams.signals import api_subtitles_approved from utils.csv_parser import UnicodeReader from videos.tasks import video_changed_tasks def complete_approve_tasks(tasks): lang_ct = ContentType.objects.get_for_model(SubtitleLanguage) video_ids = set() for task in tasks: task.do_complete_approve(lang_ct=lang_ct) + version = task.get_subtitle_version() - api_subtitles_approved.send(task.get_subtitle_version()) ? ------------------ - - + api_subtitles_approved.send(version) + if version.is_public(): + subtitles_published.send(version.subtitle_language, version=version) video_ids.add(task.team_video.video_id) for video_id in video_ids: video_changed_tasks.delay(video_id) def add_videos_from_csv(team, user, csv_file): from .tasks import add_team_videos videos = [] fields = ['project', 'url', 'title', 'description', 'duration', 'language', 'transcript'] num_fields = len(fields) try: reader = UnicodeReader(csv_file) header = reader.next() if len(header) != num_fields: raise Exception() except: raise ValueError(u'CSV format is not valid') for row in reader: videos.append(dict(zip(fields, row))) add_team_videos.delay(team.pk, user.pk, videos)
d3f0d83b0c783d2f15a6f5eaf6fd4ace426307a6
tests/__init__.py
tests/__init__.py
import os import sys import unittest def suite(): MODULE_DIR = os.path.join(os.path.dirname(__file__), '..') MODULE_DIR = os.path.abspath(MODULE_DIR) sys.path.insert(0, MODULE_DIR) sys.path.insert(0, os.path.dirname(__file__)) SUB_UNITS = os.path.dirname(__file__) SUB_UNITS = os.listdir(SUB_UNITS) SUB_UNITS = [ filename[:-3] for filename in SUB_UNITS if filename.startswith('test_') ] os.chdir(os.path.dirname(__file__)) loader = unittest.TestLoader() return loader.loadTestsFromNames(SUB_UNITS)
from os import walk, chdir from os.path import join, dirname, splitext, abspath, relpath import sys import unittest MODULE_DIR = join(dirname(__file__), '..') MODULE_DIR = abspath(MODULE_DIR) def walker(opath='.'): for path, folders, files in walk(opath): for filename in files: if filename.startswith('test_') and filename.endswith('.py'): rpath = relpath(path, opath) yield (rpath + '.' + splitext(filename)[0]).strip('.') def suite(): sys.path.insert(0, MODULE_DIR) sys.path.insert(0, dirname(__file__)) SUB_UNITS = dirname(__file__) SUB_UNITS = walker(SUB_UNITS) chdir(dirname(__file__)) return unittest.TestLoader().loadTestsFromNames(SUB_UNITS)
Rework for tests in subdirectories
Rework for tests in subdirectories
Python
mit
Mause/pytransperth,Mause/pytransperth
- import os + from os import walk, chdir + from os.path import join, dirname, splitext, abspath, relpath import sys import unittest + MODULE_DIR = join(dirname(__file__), '..') + MODULE_DIR = abspath(MODULE_DIR) + + + def walker(opath='.'): + for path, folders, files in walk(opath): + for filename in files: + if filename.startswith('test_') and filename.endswith('.py'): + rpath = relpath(path, opath) + + yield (rpath + '.' + splitext(filename)[0]).strip('.') + def suite(): - MODULE_DIR = os.path.join(os.path.dirname(__file__), '..') - MODULE_DIR = os.path.abspath(MODULE_DIR) sys.path.insert(0, MODULE_DIR) - sys.path.insert(0, os.path.dirname(__file__)) + sys.path.insert(0, dirname(__file__)) - SUB_UNITS = os.path.dirname(__file__) + SUB_UNITS = dirname(__file__) - SUB_UNITS = os.listdir(SUB_UNITS) + SUB_UNITS = walker(SUB_UNITS) - SUB_UNITS = [ - filename[:-3] - for filename in SUB_UNITS - if filename.startswith('test_') - ] - os.chdir(os.path.dirname(__file__)) + chdir(dirname(__file__)) - loader = unittest.TestLoader() - return loader.loadTestsFromNames(SUB_UNITS) + return unittest.TestLoader().loadTestsFromNames(SUB_UNITS) +
Rework for tests in subdirectories
## Code Before: import os import sys import unittest def suite(): MODULE_DIR = os.path.join(os.path.dirname(__file__), '..') MODULE_DIR = os.path.abspath(MODULE_DIR) sys.path.insert(0, MODULE_DIR) sys.path.insert(0, os.path.dirname(__file__)) SUB_UNITS = os.path.dirname(__file__) SUB_UNITS = os.listdir(SUB_UNITS) SUB_UNITS = [ filename[:-3] for filename in SUB_UNITS if filename.startswith('test_') ] os.chdir(os.path.dirname(__file__)) loader = unittest.TestLoader() return loader.loadTestsFromNames(SUB_UNITS) ## Instruction: Rework for tests in subdirectories ## Code After: from os import walk, chdir from os.path import join, dirname, splitext, abspath, relpath import sys import unittest MODULE_DIR = join(dirname(__file__), '..') MODULE_DIR = abspath(MODULE_DIR) def walker(opath='.'): for path, folders, files in walk(opath): for filename in files: if filename.startswith('test_') and filename.endswith('.py'): rpath = relpath(path, opath) yield (rpath + '.' + splitext(filename)[0]).strip('.') def suite(): sys.path.insert(0, MODULE_DIR) sys.path.insert(0, dirname(__file__)) SUB_UNITS = dirname(__file__) SUB_UNITS = walker(SUB_UNITS) chdir(dirname(__file__)) return unittest.TestLoader().loadTestsFromNames(SUB_UNITS)
- import os + from os import walk, chdir + from os.path import join, dirname, splitext, abspath, relpath import sys import unittest + MODULE_DIR = join(dirname(__file__), '..') + MODULE_DIR = abspath(MODULE_DIR) + + + def walker(opath='.'): + for path, folders, files in walk(opath): + for filename in files: + if filename.startswith('test_') and filename.endswith('.py'): + rpath = relpath(path, opath) + + yield (rpath + '.' + splitext(filename)[0]).strip('.') + def suite(): - MODULE_DIR = os.path.join(os.path.dirname(__file__), '..') - MODULE_DIR = os.path.abspath(MODULE_DIR) sys.path.insert(0, MODULE_DIR) - sys.path.insert(0, os.path.dirname(__file__)) ? -------- + sys.path.insert(0, dirname(__file__)) - SUB_UNITS = os.path.dirname(__file__) ? -------- + SUB_UNITS = dirname(__file__) - SUB_UNITS = os.listdir(SUB_UNITS) ? ^^^ ^^^^^ + SUB_UNITS = walker(SUB_UNITS) ? ^^ ^^ - SUB_UNITS = [ - filename[:-3] - for filename in SUB_UNITS - if filename.startswith('test_') - ] - os.chdir(os.path.dirname(__file__)) ? --- -------- + chdir(dirname(__file__)) - loader = unittest.TestLoader() + - return loader.loadTestsFromNames(SUB_UNITS) ? ^ + return unittest.TestLoader().loadTestsFromNames(SUB_UNITS) ? ^^^^^^^^^^^^^^ ++
725b3a9db33c90187b913123deefeb180c7fee4c
client/app.py
client/app.py
import argparse from server import * from commandRunner import * class App: def __init__(self, baseurl, clientid): self.server = Server(baseurl, clientid) def run(self): runner = CommandRunner() command = self.server.get() while command is not None: response = runner.run(command) self.server.send(response) command = self.server.get() def parseCommandLine(): parser = argparse.ArgumentParser() parser.add_argument("--baseurl", required=True) parser.add_argument("--clientid", required=True) return parser.parse_args() if __name__ == '__main__': args = parseCommandLine() app = App(args.baseurl, args.clientid); app.run()
import argparse from server import * from commandRunner import * class App: server = None runner = None def __init__(self, baseurl, clientid): self.server = Server(baseurl, clientid) self.runner = CommandRunner() def run(self): command = self.server.get() while command is not None: response = self.runner.run(command) self.server.send(response) command = self.server.get() def parseCommandLine(): parser = argparse.ArgumentParser() parser.add_argument("--baseurl", required=True) parser.add_argument("--clientid", required=True) return parser.parse_args() if __name__ == '__main__': args = parseCommandLine() app = App(args.baseurl, args.clientid); app.run()
Add DI to App object
Add DI to App object
Python
mit
CaminsTECH/owncloud-test
import argparse from server import * from commandRunner import * class App: + server = None + runner = None + def __init__(self, baseurl, clientid): self.server = Server(baseurl, clientid) - + self.runner = CommandRunner() + - def run(self): + def run(self): - runner = CommandRunner() command = self.server.get() while command is not None: - response = runner.run(command) + response = self.runner.run(command) self.server.send(response) command = self.server.get() def parseCommandLine(): parser = argparse.ArgumentParser() parser.add_argument("--baseurl", required=True) parser.add_argument("--clientid", required=True) return parser.parse_args() if __name__ == '__main__': args = parseCommandLine() app = App(args.baseurl, args.clientid); app.run()
Add DI to App object
## Code Before: import argparse from server import * from commandRunner import * class App: def __init__(self, baseurl, clientid): self.server = Server(baseurl, clientid) def run(self): runner = CommandRunner() command = self.server.get() while command is not None: response = runner.run(command) self.server.send(response) command = self.server.get() def parseCommandLine(): parser = argparse.ArgumentParser() parser.add_argument("--baseurl", required=True) parser.add_argument("--clientid", required=True) return parser.parse_args() if __name__ == '__main__': args = parseCommandLine() app = App(args.baseurl, args.clientid); app.run() ## Instruction: Add DI to App object ## Code After: import argparse from server import * from commandRunner import * class App: server = None runner = None def __init__(self, baseurl, clientid): self.server = Server(baseurl, clientid) self.runner = CommandRunner() def run(self): command = self.server.get() while command is not None: response = self.runner.run(command) self.server.send(response) command = self.server.get() def parseCommandLine(): parser = argparse.ArgumentParser() parser.add_argument("--baseurl", required=True) parser.add_argument("--clientid", required=True) return parser.parse_args() if __name__ == '__main__': args = parseCommandLine() app = App(args.baseurl, args.clientid); app.run()
import argparse from server import * from commandRunner import * class App: + server = None + runner = None + def __init__(self, baseurl, clientid): self.server = Server(baseurl, clientid) - + self.runner = CommandRunner() + - def run(self): + def run(self): ? ++ - runner = CommandRunner() command = self.server.get() while command is not None: - response = runner.run(command) + response = self.runner.run(command) ? +++++ self.server.send(response) command = self.server.get() def parseCommandLine(): parser = argparse.ArgumentParser() parser.add_argument("--baseurl", required=True) parser.add_argument("--clientid", required=True) return parser.parse_args() if __name__ == '__main__': args = parseCommandLine() app = App(args.baseurl, args.clientid); app.run()
926d5333c1556850a3eda6025ac8cf471b67c0a3
condor/probes/setup.py
condor/probes/setup.py
from distutils.core import setup setup(name='htcondor-es-probes', version='0.6.3', description='HTCondor probes for Elasticsearch analytics', author='Suchandra Thapa', author_email='sthapa@ci.uchicago.edu', url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor', packages=['probe_libs'], scripts=['collect_history_info.py', 'get_job_status.py'], data_files=[('/etc/init.d/', ['scripts/collect_history']), ('/etc/cron.d/', ['config/schedd_probe']), ('/etc/sysconfig', ['config/collect_history'])], license = 'Apache 2.0' )
from distutils.core import setup setup(name='htcondor-es-probes', version='0.6.3', description='HTCondor probes for Elasticsearch analytics', author='Suchandra Thapa', author_email='sthapa@ci.uchicago.edu', url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor', packages=['probe_libs'], scripts=['collect_history_info.py', 'get_job_status.py'], data_files=[('/etc/init.d/', ['scripts/collect_history']), ('/etc/cron.d/', ['config/schedd_probe']), ('/var/lib/collect_history', []), ('/etc/sysconfig', ['config/collect_history'])], license='Apache 2.0' )
Add directory for state files
Add directory for state files
Python
apache-2.0
DHTC-Tools/logstash-confs,DHTC-Tools/logstash-confs,DHTC-Tools/logstash-confs
from distutils.core import setup setup(name='htcondor-es-probes', version='0.6.3', description='HTCondor probes for Elasticsearch analytics', author='Suchandra Thapa', author_email='sthapa@ci.uchicago.edu', url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor', packages=['probe_libs'], scripts=['collect_history_info.py', 'get_job_status.py'], data_files=[('/etc/init.d/', ['scripts/collect_history']), ('/etc/cron.d/', ['config/schedd_probe']), + ('/var/lib/collect_history', []), ('/etc/sysconfig', ['config/collect_history'])], - license = 'Apache 2.0' + license='Apache 2.0' )
Add directory for state files
## Code Before: from distutils.core import setup setup(name='htcondor-es-probes', version='0.6.3', description='HTCondor probes for Elasticsearch analytics', author='Suchandra Thapa', author_email='sthapa@ci.uchicago.edu', url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor', packages=['probe_libs'], scripts=['collect_history_info.py', 'get_job_status.py'], data_files=[('/etc/init.d/', ['scripts/collect_history']), ('/etc/cron.d/', ['config/schedd_probe']), ('/etc/sysconfig', ['config/collect_history'])], license = 'Apache 2.0' ) ## Instruction: Add directory for state files ## Code After: from distutils.core import setup setup(name='htcondor-es-probes', version='0.6.3', description='HTCondor probes for Elasticsearch analytics', author='Suchandra Thapa', author_email='sthapa@ci.uchicago.edu', url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor', packages=['probe_libs'], scripts=['collect_history_info.py', 'get_job_status.py'], data_files=[('/etc/init.d/', ['scripts/collect_history']), ('/etc/cron.d/', ['config/schedd_probe']), ('/var/lib/collect_history', []), ('/etc/sysconfig', ['config/collect_history'])], license='Apache 2.0' )
from distutils.core import setup setup(name='htcondor-es-probes', version='0.6.3', description='HTCondor probes for Elasticsearch analytics', author='Suchandra Thapa', author_email='sthapa@ci.uchicago.edu', url='https://github.com/DHTC-Tools/logstash-confs/tree/master/condor', packages=['probe_libs'], scripts=['collect_history_info.py', 'get_job_status.py'], data_files=[('/etc/init.d/', ['scripts/collect_history']), ('/etc/cron.d/', ['config/schedd_probe']), + ('/var/lib/collect_history', []), ('/etc/sysconfig', ['config/collect_history'])], - license = 'Apache 2.0' ? - - + license='Apache 2.0' )
9c4ecf0b72d86ae113fa13f210c543120635b73e
board.py
board.py
import numpy """ Board represents a four in a row game board. Author: Isaac Arvestad """ class Board: """ Initializes the game with a certain number of rows and columns. """ def __init(self, rows, columns): self.rows = rows self.columns = columns self.boardMatrix = numpy.zeros((rows, columns)) """ Attempts to add a piece to a certain column. If the column is full the move is illegal and false is returned, otherwise true is returned. """ def addPiece(self, column, value): "Check if column is full." if self.boardMatrix.item(0,column) != 0: return false "Place piece." for y in range(self.rows): currentValue = self.boardMatrix.item(y, column) if currentValue == 0: if y == rows - 1: self.boardMatrix.itemset((y, column), value) else: continue return true
import numpy """ Board represents a four in a row game board. Author: Isaac Arvestad """ class Board: """ Initializes the game with a certain number of rows and columns. """ def __init__(self, rows, columns): self.rows = rows self.columns = columns self.boardMatrix = numpy.zeros((rows, columns)) """ Attempts to add a piece to a certain column. If the column is full the move is illegal and false is returned, otherwise true is returned. """ def addPiece(self, column, value): "Check if column is full." if self.boardMatrix.item(0,column) != 0: return False "Place piece." for y in range(self.rows): currentValue = self.boardMatrix.item(y, column) if currentValue == 0: if y == self.rows - 1: self.boardMatrix.itemset((y, column), value) else: continue return True
Fix init method, self.rows and capital booleans.
Fix init method, self.rows and capital booleans.
Python
mit
isaacarvestad/four-in-a-row
import numpy """ Board represents a four in a row game board. Author: Isaac Arvestad """ class Board: """ Initializes the game with a certain number of rows and columns. """ - def __init(self, rows, columns): + def __init__(self, rows, columns): self.rows = rows self.columns = columns self.boardMatrix = numpy.zeros((rows, columns)) """ Attempts to add a piece to a certain column. If the column is full the move is illegal and false is returned, otherwise true is returned. """ def addPiece(self, column, value): "Check if column is full." if self.boardMatrix.item(0,column) != 0: - return false + return False "Place piece." for y in range(self.rows): currentValue = self.boardMatrix.item(y, column) if currentValue == 0: - if y == rows - 1: + if y == self.rows - 1: self.boardMatrix.itemset((y, column), value) else: continue - return true + return True -
Fix init method, self.rows and capital booleans.
## Code Before: import numpy """ Board represents a four in a row game board. Author: Isaac Arvestad """ class Board: """ Initializes the game with a certain number of rows and columns. """ def __init(self, rows, columns): self.rows = rows self.columns = columns self.boardMatrix = numpy.zeros((rows, columns)) """ Attempts to add a piece to a certain column. If the column is full the move is illegal and false is returned, otherwise true is returned. """ def addPiece(self, column, value): "Check if column is full." if self.boardMatrix.item(0,column) != 0: return false "Place piece." for y in range(self.rows): currentValue = self.boardMatrix.item(y, column) if currentValue == 0: if y == rows - 1: self.boardMatrix.itemset((y, column), value) else: continue return true ## Instruction: Fix init method, self.rows and capital booleans. ## Code After: import numpy """ Board represents a four in a row game board. Author: Isaac Arvestad """ class Board: """ Initializes the game with a certain number of rows and columns. """ def __init__(self, rows, columns): self.rows = rows self.columns = columns self.boardMatrix = numpy.zeros((rows, columns)) """ Attempts to add a piece to a certain column. If the column is full the move is illegal and false is returned, otherwise true is returned. """ def addPiece(self, column, value): "Check if column is full." if self.boardMatrix.item(0,column) != 0: return False "Place piece." for y in range(self.rows): currentValue = self.boardMatrix.item(y, column) if currentValue == 0: if y == self.rows - 1: self.boardMatrix.itemset((y, column), value) else: continue return True
import numpy """ Board represents a four in a row game board. Author: Isaac Arvestad """ class Board: """ Initializes the game with a certain number of rows and columns. """ - def __init(self, rows, columns): + def __init__(self, rows, columns): ? ++ self.rows = rows self.columns = columns self.boardMatrix = numpy.zeros((rows, columns)) """ Attempts to add a piece to a certain column. If the column is full the move is illegal and false is returned, otherwise true is returned. """ def addPiece(self, column, value): "Check if column is full." if self.boardMatrix.item(0,column) != 0: - return false ? ^ + return False ? ^ "Place piece." for y in range(self.rows): currentValue = self.boardMatrix.item(y, column) if currentValue == 0: - if y == rows - 1: + if y == self.rows - 1: ? +++++ self.boardMatrix.itemset((y, column), value) else: continue - return true ? ^ + return True ? ^ -
29562b08e436abc8465404e49d9193537721b717
src/odin/contrib/money/fields.py
src/odin/contrib/money/fields.py
from __future__ import absolute_import, print_function from odin import exceptions from odin.fields import ScalarField from odin.validators import EMPTY_VALUES from .datatypes import Amount __all__ = ('AmountField', ) class AmountField(ScalarField): """ Field that contains a monetary amount (with an optional currency). """ default_error_messages = { 'invalid': "'%s' value must be a (amount, currency).", 'invalid_currency': "'%s' currency is not supported.", } data_type_name = "Amount" def __init__(self, allowed_currencies=None, **kwargs): super(AmountField, self).__init__(**kwargs) self.allowed_currencies = allowed_currencies def to_python(self, value): if value in EMPTY_VALUES: return if isinstance(value, Amount): return value try: return Amount(value) except (ValueError, TypeError): msg = self.error_messages['invalid'] % value raise exceptions.ValidationError(msg) def validate(self, value): super(AmountField, self).validate(value) if self.allowed_currencies and value not in EMPTY_VALUES: if value.currency not in self.allowed_currencies: msg = self.error_messages['invalid_currency'] % str(value.currency) raise exceptions.ValidationError(msg) def prepare(self, value): if value in EMPTY_VALUES: return return float(value), value.currency.code
from __future__ import absolute_import, print_function from odin import exceptions from odin.fields import ScalarField from odin.validators import EMPTY_VALUES from .datatypes import Amount __all__ = ("AmountField",) class AmountField(ScalarField): """ Field that contains a monetary amount (with an optional currency). """ default_error_messages = { "invalid": "'%s' value must be a (amount, currency).", "invalid_currency": "'%s' currency is not supported.", } data_type_name = "Amount" def __init__(self, allowed_currencies=None, **kwargs): super(AmountField, self).__init__(**kwargs) self.allowed_currencies = allowed_currencies def to_python(self, value): if value in EMPTY_VALUES: return if isinstance(value, Amount): return value try: return Amount(value) except (ValueError, TypeError): msg = self.error_messages["invalid"] % value raise exceptions.ValidationError(msg) def validate(self, value): super(AmountField, self).validate(value) if ( self.allowed_currencies and (value not in EMPTY_VALUES) and (value.currency not in self.allowed_currencies) ): msg = self.error_messages["invalid_currency"] % str(value.currency) raise exceptions.ValidationError(msg) def prepare(self, value): if value in EMPTY_VALUES: return return float(value), value.currency.code
Correct issue from Sonar (and black file)
Correct issue from Sonar (and black file)
Python
bsd-3-clause
python-odin/odin
from __future__ import absolute_import, print_function from odin import exceptions from odin.fields import ScalarField from odin.validators import EMPTY_VALUES from .datatypes import Amount - __all__ = ('AmountField', ) + __all__ = ("AmountField",) class AmountField(ScalarField): """ Field that contains a monetary amount (with an optional currency). """ + default_error_messages = { - 'invalid': "'%s' value must be a (amount, currency).", + "invalid": "'%s' value must be a (amount, currency).", - 'invalid_currency': "'%s' currency is not supported.", + "invalid_currency": "'%s' currency is not supported.", } data_type_name = "Amount" def __init__(self, allowed_currencies=None, **kwargs): super(AmountField, self).__init__(**kwargs) self.allowed_currencies = allowed_currencies def to_python(self, value): if value in EMPTY_VALUES: return if isinstance(value, Amount): return value try: return Amount(value) except (ValueError, TypeError): - msg = self.error_messages['invalid'] % value + msg = self.error_messages["invalid"] % value raise exceptions.ValidationError(msg) def validate(self, value): super(AmountField, self).validate(value) - if self.allowed_currencies and value not in EMPTY_VALUES: + if ( + self.allowed_currencies + and (value not in EMPTY_VALUES) - if value.currency not in self.allowed_currencies: + and (value.currency not in self.allowed_currencies) + ): - msg = self.error_messages['invalid_currency'] % str(value.currency) + msg = self.error_messages["invalid_currency"] % str(value.currency) - raise exceptions.ValidationError(msg) + raise exceptions.ValidationError(msg) def prepare(self, value): if value in EMPTY_VALUES: return return float(value), value.currency.code
Correct issue from Sonar (and black file)
## Code Before: from __future__ import absolute_import, print_function from odin import exceptions from odin.fields import ScalarField from odin.validators import EMPTY_VALUES from .datatypes import Amount __all__ = ('AmountField', ) class AmountField(ScalarField): """ Field that contains a monetary amount (with an optional currency). """ default_error_messages = { 'invalid': "'%s' value must be a (amount, currency).", 'invalid_currency': "'%s' currency is not supported.", } data_type_name = "Amount" def __init__(self, allowed_currencies=None, **kwargs): super(AmountField, self).__init__(**kwargs) self.allowed_currencies = allowed_currencies def to_python(self, value): if value in EMPTY_VALUES: return if isinstance(value, Amount): return value try: return Amount(value) except (ValueError, TypeError): msg = self.error_messages['invalid'] % value raise exceptions.ValidationError(msg) def validate(self, value): super(AmountField, self).validate(value) if self.allowed_currencies and value not in EMPTY_VALUES: if value.currency not in self.allowed_currencies: msg = self.error_messages['invalid_currency'] % str(value.currency) raise exceptions.ValidationError(msg) def prepare(self, value): if value in EMPTY_VALUES: return return float(value), value.currency.code ## Instruction: Correct issue from Sonar (and black file) ## Code After: from __future__ import absolute_import, print_function from odin import exceptions from odin.fields import ScalarField from odin.validators import EMPTY_VALUES from .datatypes import Amount __all__ = ("AmountField",) class AmountField(ScalarField): """ Field that contains a monetary amount (with an optional currency). """ default_error_messages = { "invalid": "'%s' value must be a (amount, currency).", "invalid_currency": "'%s' currency is not supported.", } data_type_name = "Amount" def __init__(self, allowed_currencies=None, **kwargs): super(AmountField, self).__init__(**kwargs) self.allowed_currencies = allowed_currencies def to_python(self, value): if value in EMPTY_VALUES: return if isinstance(value, Amount): return value try: return Amount(value) except (ValueError, TypeError): msg = self.error_messages["invalid"] % value raise exceptions.ValidationError(msg) def validate(self, value): super(AmountField, self).validate(value) if ( self.allowed_currencies and (value not in EMPTY_VALUES) and (value.currency not in self.allowed_currencies) ): msg = self.error_messages["invalid_currency"] % str(value.currency) raise exceptions.ValidationError(msg) def prepare(self, value): if value in EMPTY_VALUES: return return float(value), value.currency.code
from __future__ import absolute_import, print_function from odin import exceptions from odin.fields import ScalarField from odin.validators import EMPTY_VALUES from .datatypes import Amount - __all__ = ('AmountField', ) ? ^ ^ - + __all__ = ("AmountField",) ? ^ ^ class AmountField(ScalarField): """ Field that contains a monetary amount (with an optional currency). """ + default_error_messages = { - 'invalid': "'%s' value must be a (amount, currency).", ? ^ ^ + "invalid": "'%s' value must be a (amount, currency).", ? ^ ^ - 'invalid_currency': "'%s' currency is not supported.", ? ^ ^ + "invalid_currency": "'%s' currency is not supported.", ? ^ ^ } data_type_name = "Amount" def __init__(self, allowed_currencies=None, **kwargs): super(AmountField, self).__init__(**kwargs) self.allowed_currencies = allowed_currencies def to_python(self, value): if value in EMPTY_VALUES: return if isinstance(value, Amount): return value try: return Amount(value) except (ValueError, TypeError): - msg = self.error_messages['invalid'] % value ? ^ ^ + msg = self.error_messages["invalid"] % value ? ^ ^ raise exceptions.ValidationError(msg) def validate(self, value): super(AmountField, self).validate(value) - if self.allowed_currencies and value not in EMPTY_VALUES: + if ( + self.allowed_currencies + and (value not in EMPTY_VALUES) - if value.currency not in self.allowed_currencies: ? ^^ ^ + and (value.currency not in self.allowed_currencies) ? ^^^ + ^ + ): - msg = self.error_messages['invalid_currency'] % str(value.currency) ? ---- ^ ^ + msg = self.error_messages["invalid_currency"] % str(value.currency) ? ^ ^ - raise exceptions.ValidationError(msg) ? ---- + raise exceptions.ValidationError(msg) def prepare(self, value): if value in EMPTY_VALUES: return return float(value), value.currency.code
3043a2400e46648f01921aad265816d2bcf18211
test/test_conjunctive_graph.py
test/test_conjunctive_graph.py
from rdflib.graph import ConjunctiveGraph from rdflib.term import Identifier, URIRef from rdflib.parser import StringInputSource from os import path DATA = u""" <http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> . """ PUBLIC_ID = u"http://example.org/record/1" def test_graph_ids(): def check(kws): cg = ConjunctiveGraph() cg.parse(**kws) for g in cg.contexts(): gid = g.identifier assert isinstance(gid, Identifier) yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle") source = StringInputSource(DATA) source.setPublicId(PUBLIC_ID) yield check, dict(source=source, format='turtle') if __name__ == '__main__': import nose nose.main(defaultTest=__name__)
from rdflib.graph import ConjunctiveGraph from rdflib.term import Identifier, URIRef from rdflib.parser import StringInputSource from os import path DATA = u""" <http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> . """ PUBLIC_ID = u"http://example.org/record/1" def test_graph_ids(): def check(kws): cg = ConjunctiveGraph() cg.parse(**kws) for g in cg.contexts(): gid = g.identifier assert isinstance(gid, Identifier) yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle") source = StringInputSource(DATA.encode('utf8')) source.setPublicId(PUBLIC_ID) yield check, dict(source=source, format='turtle') if __name__ == '__main__': import nose nose.main(defaultTest=__name__)
Fix py3-incompatible test code that causes PY3 test failure.
Fix py3-incompatible test code that causes PY3 test failure.
Python
bsd-3-clause
avorio/rdflib,yingerj/rdflib,avorio/rdflib,RDFLib/rdflib,ssssam/rdflib,marma/rdflib,marma/rdflib,avorio/rdflib,armandobs14/rdflib,RDFLib/rdflib,dbs/rdflib,armandobs14/rdflib,yingerj/rdflib,yingerj/rdflib,ssssam/rdflib,RDFLib/rdflib,dbs/rdflib,yingerj/rdflib,marma/rdflib,dbs/rdflib,armandobs14/rdflib,RDFLib/rdflib,avorio/rdflib,marma/rdflib,ssssam/rdflib,armandobs14/rdflib,ssssam/rdflib,dbs/rdflib
from rdflib.graph import ConjunctiveGraph from rdflib.term import Identifier, URIRef from rdflib.parser import StringInputSource from os import path DATA = u""" <http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> . """ PUBLIC_ID = u"http://example.org/record/1" def test_graph_ids(): def check(kws): cg = ConjunctiveGraph() cg.parse(**kws) for g in cg.contexts(): gid = g.identifier assert isinstance(gid, Identifier) yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle") - source = StringInputSource(DATA) + source = StringInputSource(DATA.encode('utf8')) source.setPublicId(PUBLIC_ID) yield check, dict(source=source, format='turtle') if __name__ == '__main__': import nose nose.main(defaultTest=__name__)
Fix py3-incompatible test code that causes PY3 test failure.
## Code Before: from rdflib.graph import ConjunctiveGraph from rdflib.term import Identifier, URIRef from rdflib.parser import StringInputSource from os import path DATA = u""" <http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> . """ PUBLIC_ID = u"http://example.org/record/1" def test_graph_ids(): def check(kws): cg = ConjunctiveGraph() cg.parse(**kws) for g in cg.contexts(): gid = g.identifier assert isinstance(gid, Identifier) yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle") source = StringInputSource(DATA) source.setPublicId(PUBLIC_ID) yield check, dict(source=source, format='turtle') if __name__ == '__main__': import nose nose.main(defaultTest=__name__) ## Instruction: Fix py3-incompatible test code that causes PY3 test failure. ## Code After: from rdflib.graph import ConjunctiveGraph from rdflib.term import Identifier, URIRef from rdflib.parser import StringInputSource from os import path DATA = u""" <http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> . """ PUBLIC_ID = u"http://example.org/record/1" def test_graph_ids(): def check(kws): cg = ConjunctiveGraph() cg.parse(**kws) for g in cg.contexts(): gid = g.identifier assert isinstance(gid, Identifier) yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle") source = StringInputSource(DATA.encode('utf8')) source.setPublicId(PUBLIC_ID) yield check, dict(source=source, format='turtle') if __name__ == '__main__': import nose nose.main(defaultTest=__name__)
from rdflib.graph import ConjunctiveGraph from rdflib.term import Identifier, URIRef from rdflib.parser import StringInputSource from os import path DATA = u""" <http://example.org/record/1> a <http://xmlns.com/foaf/0.1/Document> . """ PUBLIC_ID = u"http://example.org/record/1" def test_graph_ids(): def check(kws): cg = ConjunctiveGraph() cg.parse(**kws) for g in cg.contexts(): gid = g.identifier assert isinstance(gid, Identifier) yield check, dict(data=DATA, publicID=PUBLIC_ID, format="turtle") - source = StringInputSource(DATA) + source = StringInputSource(DATA.encode('utf8')) ? ++++++++++++++ + source.setPublicId(PUBLIC_ID) yield check, dict(source=source, format='turtle') if __name__ == '__main__': import nose nose.main(defaultTest=__name__)
c1756ab481f3bf72ab33465c8eb1d5a3e729ce4e
model_logging/migrations/0003_data_migration.py
model_logging/migrations/0003_data_migration.py
from django.db import migrations app = 'model_logging' model = 'LogEntry' def move_data(apps, schema_editor): LogEntry = apps.get_model(app, model) for entry in LogEntry.objects.all(): entry.data_temp = entry.data entry.save() class Migration(migrations.Migration): dependencies = [ ('model_logging', '0002_add_new_data_field'), ] operations = [ migrations.RunPython(move_data), ]
from django.db import migrations app = 'model_logging' model = 'LogEntry' def move_data(apps, schema_editor): try: from pgcrypto.fields import TextPGPPublicKeyField except ImportError: raise ImportError('Please install django-pgcrypto-fields to perform migration') LogEntry = apps.get_model(app, model) for entry in LogEntry.objects.all(): entry.data_temp = entry.data entry.save() class Migration(migrations.Migration): dependencies = [ ('model_logging', '0002_add_new_data_field'), ] operations = [ migrations.RunPython(move_data), ]
Add try, catch statement to ensure data migration can be performed.
Add try, catch statement to ensure data migration can be performed.
Python
bsd-2-clause
incuna/django-model-logging
from django.db import migrations app = 'model_logging' model = 'LogEntry' def move_data(apps, schema_editor): + try: + from pgcrypto.fields import TextPGPPublicKeyField + except ImportError: + raise ImportError('Please install django-pgcrypto-fields to perform migration') + LogEntry = apps.get_model(app, model) for entry in LogEntry.objects.all(): entry.data_temp = entry.data entry.save() class Migration(migrations.Migration): dependencies = [ ('model_logging', '0002_add_new_data_field'), ] operations = [ migrations.RunPython(move_data), ]
Add try, catch statement to ensure data migration can be performed.
## Code Before: from django.db import migrations app = 'model_logging' model = 'LogEntry' def move_data(apps, schema_editor): LogEntry = apps.get_model(app, model) for entry in LogEntry.objects.all(): entry.data_temp = entry.data entry.save() class Migration(migrations.Migration): dependencies = [ ('model_logging', '0002_add_new_data_field'), ] operations = [ migrations.RunPython(move_data), ] ## Instruction: Add try, catch statement to ensure data migration can be performed. ## Code After: from django.db import migrations app = 'model_logging' model = 'LogEntry' def move_data(apps, schema_editor): try: from pgcrypto.fields import TextPGPPublicKeyField except ImportError: raise ImportError('Please install django-pgcrypto-fields to perform migration') LogEntry = apps.get_model(app, model) for entry in LogEntry.objects.all(): entry.data_temp = entry.data entry.save() class Migration(migrations.Migration): dependencies = [ ('model_logging', '0002_add_new_data_field'), ] operations = [ migrations.RunPython(move_data), ]
from django.db import migrations app = 'model_logging' model = 'LogEntry' def move_data(apps, schema_editor): + try: + from pgcrypto.fields import TextPGPPublicKeyField + except ImportError: + raise ImportError('Please install django-pgcrypto-fields to perform migration') + LogEntry = apps.get_model(app, model) for entry in LogEntry.objects.all(): entry.data_temp = entry.data entry.save() class Migration(migrations.Migration): dependencies = [ ('model_logging', '0002_add_new_data_field'), ] operations = [ migrations.RunPython(move_data), ]
1090acb35ea4ce5c8d17db716539d3354feabc12
nodeconductor/iaas/migrations/0038_securitygroup_state.py
nodeconductor/iaas/migrations/0038_securitygroup_state.py
from __future__ import unicode_literals from django.db import models, migrations import django_fsm class Migration(migrations.Migration): dependencies = [ ('iaas', '0037_init_security_groups_quotas'), ] operations = [ migrations.AddField( model_name='securitygroup', name='state', field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]), preserve_default=True, ), ]
from __future__ import unicode_literals from django.db import models, migrations import django_fsm def mark_security_groups_as_synced(apps, schema_editor): SecurityGroup = apps.get_model('iaas', 'SecurityGroup') SecurityGroup.objects.all().update(state=3) class Migration(migrations.Migration): dependencies = [ ('iaas', '0037_init_security_groups_quotas'), ] operations = [ migrations.AddField( model_name='securitygroup', name='state', field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]), preserve_default=True, ), migrations.RunPython(mark_security_groups_as_synced), ]
Mark all exist security groups as synced
Mark all exist security groups as synced - itacloud-4843
Python
mit
opennode/nodeconductor,opennode/nodeconductor,opennode/nodeconductor
from __future__ import unicode_literals from django.db import models, migrations import django_fsm + + + def mark_security_groups_as_synced(apps, schema_editor): + SecurityGroup = apps.get_model('iaas', 'SecurityGroup') + SecurityGroup.objects.all().update(state=3) class Migration(migrations.Migration): dependencies = [ ('iaas', '0037_init_security_groups_quotas'), ] operations = [ migrations.AddField( model_name='securitygroup', name='state', field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]), preserve_default=True, ), + migrations.RunPython(mark_security_groups_as_synced), ]
Mark all exist security groups as synced
## Code Before: from __future__ import unicode_literals from django.db import models, migrations import django_fsm class Migration(migrations.Migration): dependencies = [ ('iaas', '0037_init_security_groups_quotas'), ] operations = [ migrations.AddField( model_name='securitygroup', name='state', field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]), preserve_default=True, ), ] ## Instruction: Mark all exist security groups as synced ## Code After: from __future__ import unicode_literals from django.db import models, migrations import django_fsm def mark_security_groups_as_synced(apps, schema_editor): SecurityGroup = apps.get_model('iaas', 'SecurityGroup') SecurityGroup.objects.all().update(state=3) class Migration(migrations.Migration): dependencies = [ ('iaas', '0037_init_security_groups_quotas'), ] operations = [ migrations.AddField( model_name='securitygroup', name='state', field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]), preserve_default=True, ), migrations.RunPython(mark_security_groups_as_synced), ]
from __future__ import unicode_literals from django.db import models, migrations import django_fsm + + + def mark_security_groups_as_synced(apps, schema_editor): + SecurityGroup = apps.get_model('iaas', 'SecurityGroup') + SecurityGroup.objects.all().update(state=3) class Migration(migrations.Migration): dependencies = [ ('iaas', '0037_init_security_groups_quotas'), ] operations = [ migrations.AddField( model_name='securitygroup', name='state', field=django_fsm.FSMIntegerField(default=1, choices=[(1, 'Sync Scheduled'), (2, 'Syncing'), (3, 'In Sync'), (4, 'Erred')]), preserve_default=True, ), + migrations.RunPython(mark_security_groups_as_synced), ]
1cda977eff5a2edaa0de82882ef2e7d1611329b7
tests/test_protocol.py
tests/test_protocol.py
import pytest class TestProtocol: @pytest.mark.asyncio def test_server_hello(self, ws_client_factory, get_unencrypted_packet): """ The server must send a valid `server-hello` on connection. """ client = yield from ws_client_factory() receiver, message = yield from get_unencrypted_packet(client) assert receiver == 0x00 assert message['type'] == 'server-hello' assert len(message['key']) == 32 assert len(message['my-cookie']) == 16 yield from client.close()
import asyncio import pytest import saltyrtc class TestProtocol: @pytest.mark.asyncio def test_no_subprotocols(self, ws_client_factory): """ The server must drop the client after the connection has been established with a close code of *1002*. """ client = yield from ws_client_factory(subprotocols=None) yield from asyncio.sleep(0.05) assert not client.open assert client.close_code == saltyrtc.CloseCode.sub_protocol_error @pytest.mark.asyncio def test_invalid_subprotocols(self, ws_client_factory): """ The server must drop the client after the connection has been established with a close code of *1002*. """ client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000']) yield from asyncio.sleep(0.05) assert not client.open assert client.close_code == saltyrtc.CloseCode.sub_protocol_error @pytest.mark.asyncio def test_server_hello(self, ws_client_factory, get_unencrypted_packet): """ The server must send a valid `server-hello` on connection. """ client = yield from ws_client_factory() receiver, message = yield from get_unencrypted_packet(client) assert receiver == 0x00 assert message['type'] == 'server-hello' assert len(message['key']) == 32 assert len(message['my-cookie']) == 16 yield from client.close()
Add tests for invalid and no provided sub-protocols
Add tests for invalid and no provided sub-protocols
Python
mit
saltyrtc/saltyrtc-server-python,saltyrtc/saltyrtc-server-python
+ import asyncio + import pytest + + import saltyrtc class TestProtocol: + @pytest.mark.asyncio + def test_no_subprotocols(self, ws_client_factory): + """ + The server must drop the client after the connection has been + established with a close code of *1002*. + """ + client = yield from ws_client_factory(subprotocols=None) + yield from asyncio.sleep(0.05) + assert not client.open + assert client.close_code == saltyrtc.CloseCode.sub_protocol_error + + @pytest.mark.asyncio + def test_invalid_subprotocols(self, ws_client_factory): + """ + The server must drop the client after the connection has been + established with a close code of *1002*. + """ + client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000']) + yield from asyncio.sleep(0.05) + assert not client.open + assert client.close_code == saltyrtc.CloseCode.sub_protocol_error + @pytest.mark.asyncio def test_server_hello(self, ws_client_factory, get_unencrypted_packet): """ The server must send a valid `server-hello` on connection. """ client = yield from ws_client_factory() receiver, message = yield from get_unencrypted_packet(client) assert receiver == 0x00 assert message['type'] == 'server-hello' assert len(message['key']) == 32 assert len(message['my-cookie']) == 16 yield from client.close()
Add tests for invalid and no provided sub-protocols
## Code Before: import pytest class TestProtocol: @pytest.mark.asyncio def test_server_hello(self, ws_client_factory, get_unencrypted_packet): """ The server must send a valid `server-hello` on connection. """ client = yield from ws_client_factory() receiver, message = yield from get_unencrypted_packet(client) assert receiver == 0x00 assert message['type'] == 'server-hello' assert len(message['key']) == 32 assert len(message['my-cookie']) == 16 yield from client.close() ## Instruction: Add tests for invalid and no provided sub-protocols ## Code After: import asyncio import pytest import saltyrtc class TestProtocol: @pytest.mark.asyncio def test_no_subprotocols(self, ws_client_factory): """ The server must drop the client after the connection has been established with a close code of *1002*. """ client = yield from ws_client_factory(subprotocols=None) yield from asyncio.sleep(0.05) assert not client.open assert client.close_code == saltyrtc.CloseCode.sub_protocol_error @pytest.mark.asyncio def test_invalid_subprotocols(self, ws_client_factory): """ The server must drop the client after the connection has been established with a close code of *1002*. """ client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000']) yield from asyncio.sleep(0.05) assert not client.open assert client.close_code == saltyrtc.CloseCode.sub_protocol_error @pytest.mark.asyncio def test_server_hello(self, ws_client_factory, get_unencrypted_packet): """ The server must send a valid `server-hello` on connection. """ client = yield from ws_client_factory() receiver, message = yield from get_unencrypted_packet(client) assert receiver == 0x00 assert message['type'] == 'server-hello' assert len(message['key']) == 32 assert len(message['my-cookie']) == 16 yield from client.close()
+ import asyncio + import pytest + + import saltyrtc class TestProtocol: + @pytest.mark.asyncio + def test_no_subprotocols(self, ws_client_factory): + """ + The server must drop the client after the connection has been + established with a close code of *1002*. + """ + client = yield from ws_client_factory(subprotocols=None) + yield from asyncio.sleep(0.05) + assert not client.open + assert client.close_code == saltyrtc.CloseCode.sub_protocol_error + + @pytest.mark.asyncio + def test_invalid_subprotocols(self, ws_client_factory): + """ + The server must drop the client after the connection has been + established with a close code of *1002*. + """ + client = yield from ws_client_factory(subprotocols=['kittie-protocol-3000']) + yield from asyncio.sleep(0.05) + assert not client.open + assert client.close_code == saltyrtc.CloseCode.sub_protocol_error + @pytest.mark.asyncio def test_server_hello(self, ws_client_factory, get_unencrypted_packet): """ The server must send a valid `server-hello` on connection. """ client = yield from ws_client_factory() receiver, message = yield from get_unencrypted_packet(client) assert receiver == 0x00 assert message['type'] == 'server-hello' assert len(message['key']) == 32 assert len(message['my-cookie']) == 16 yield from client.close()
e7cb5b0be49bc5e811809c56eb4ad3c0dc861cdf
examples/child_watcher.py
examples/child_watcher.py
import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() log.info("There are %d items now: %s", len(children), children) @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item)
import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() log.info("There are %d items now: %s", len(children), ", ".join(children)) @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item) yield gen.sleep(1)
Fix up to the child watcher example.
Fix up to the child watcher example. Without yielding to the ioloop after each call to client.delete() the child znodes would be deleted but that would never be reported.
Python
apache-2.0
wglass/zoonado
import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() - log.info("There are %d items now: %s", len(children), children) + log.info("There are %d items now: %s", len(children), ", ".join(children)) @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item) + yield gen.sleep(1)
Fix up to the child watcher example.
## Code Before: import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() log.info("There are %d items now: %s", len(children), children) @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item) ## Instruction: Fix up to the child watcher example. ## Code After: import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() log.info("There are %d items now: %s", len(children), ", ".join(children)) @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item) yield gen.sleep(1)
import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() - log.info("There are %d items now: %s", len(children), children) + log.info("There are %d items now: %s", len(children), ", ".join(children)) ? ++++++++++ + @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item) + yield gen.sleep(1)
347545cc7ece8c0763ef194654fbaa34d16efe54
styleguide/views.py
styleguide/views.py
from django.shortcuts import get_object_or_404, render from django import forms class ExampleForm(forms.Form): text = forms.CharField() disabled_text = forms.CharField(disabled=True) readonly_text = forms.CharField( widget=forms.TextInput(attrs={'readonly':'readonly'}) ) checkbox1 = forms.BooleanField() checkbox2 = forms.BooleanField() select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')]) radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect()) form_initial = { "text": "", "disabled_text": "This field can't be changed", } def styleguide(request): return render(request, "styleguide/styleguide.html", { }) def styleguide_page(request, name): return render(request, "styleguide/styleguide-%s.html" % name, { "example_form": ExampleForm(initial=form_initial), }) def styleguide_sub_page(request, name, sub_page): return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), { "example_form": ExampleForm(initial=form_initial), })
from django.shortcuts import get_object_or_404, render from django import forms class ExampleForm(forms.Form): text = forms.CharField() disabled_text = forms.CharField(disabled=True) readonly_text = forms.CharField( widget=forms.TextInput(attrs={'readonly':'readonly'}) ) checkbox1 = forms.BooleanField() checkbox2 = forms.BooleanField() select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')]) radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect()) form_initial = { "text": "", "disabled_text": "This field can't be changed", "readonly_text": "This field is read only", } def styleguide(request): return render(request, "styleguide/styleguide.html", { }) def styleguide_page(request, name): return render(request, "styleguide/styleguide-%s.html" % name, { "example_form": ExampleForm(initial=form_initial), }) def styleguide_sub_page(request, name, sub_page): return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), { "example_form": ExampleForm(initial=form_initial), })
Add readonly text to form
Add readonly text to form
Python
bsd-3-clause
caktus/django-styleguide,caktus/django-styleguide,caktus/django-styleguide
from django.shortcuts import get_object_or_404, render from django import forms class ExampleForm(forms.Form): text = forms.CharField() disabled_text = forms.CharField(disabled=True) readonly_text = forms.CharField( widget=forms.TextInput(attrs={'readonly':'readonly'}) ) checkbox1 = forms.BooleanField() checkbox2 = forms.BooleanField() select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')]) radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect()) form_initial = { "text": "", "disabled_text": "This field can't be changed", + "readonly_text": "This field is read only", } def styleguide(request): return render(request, "styleguide/styleguide.html", { }) def styleguide_page(request, name): return render(request, "styleguide/styleguide-%s.html" % name, { "example_form": ExampleForm(initial=form_initial), }) def styleguide_sub_page(request, name, sub_page): return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), { "example_form": ExampleForm(initial=form_initial), })
Add readonly text to form
## Code Before: from django.shortcuts import get_object_or_404, render from django import forms class ExampleForm(forms.Form): text = forms.CharField() disabled_text = forms.CharField(disabled=True) readonly_text = forms.CharField( widget=forms.TextInput(attrs={'readonly':'readonly'}) ) checkbox1 = forms.BooleanField() checkbox2 = forms.BooleanField() select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')]) radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect()) form_initial = { "text": "", "disabled_text": "This field can't be changed", } def styleguide(request): return render(request, "styleguide/styleguide.html", { }) def styleguide_page(request, name): return render(request, "styleguide/styleguide-%s.html" % name, { "example_form": ExampleForm(initial=form_initial), }) def styleguide_sub_page(request, name, sub_page): return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), { "example_form": ExampleForm(initial=form_initial), }) ## Instruction: Add readonly text to form ## Code After: from django.shortcuts import get_object_or_404, render from django import forms class ExampleForm(forms.Form): text = forms.CharField() disabled_text = forms.CharField(disabled=True) readonly_text = forms.CharField( widget=forms.TextInput(attrs={'readonly':'readonly'}) ) checkbox1 = forms.BooleanField() checkbox2 = forms.BooleanField() select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')]) radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect()) form_initial = { "text": "", "disabled_text": "This field can't be changed", "readonly_text": "This field is read only", } def styleguide(request): return render(request, "styleguide/styleguide.html", { }) def styleguide_page(request, name): return render(request, "styleguide/styleguide-%s.html" % name, { "example_form": ExampleForm(initial=form_initial), }) def styleguide_sub_page(request, name, sub_page): return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), { "example_form": ExampleForm(initial=form_initial), })
from django.shortcuts import get_object_or_404, render from django import forms class ExampleForm(forms.Form): text = forms.CharField() disabled_text = forms.CharField(disabled=True) readonly_text = forms.CharField( widget=forms.TextInput(attrs={'readonly':'readonly'}) ) checkbox1 = forms.BooleanField() checkbox2 = forms.BooleanField() select = forms.ChoiceField(choices=[('', "Select an Option"), (1, 'one'), (2, 'two'), (3, 'three')]) radio = forms.ChoiceField(choices=[(1, 'one'), (2, 'two'), (3, 'three')], widget=forms.RadioSelect()) form_initial = { "text": "", "disabled_text": "This field can't be changed", + "readonly_text": "This field is read only", } def styleguide(request): return render(request, "styleguide/styleguide.html", { }) def styleguide_page(request, name): return render(request, "styleguide/styleguide-%s.html" % name, { "example_form": ExampleForm(initial=form_initial), }) def styleguide_sub_page(request, name, sub_page): return render(request, "styleguide/styleguide-%s-%s.html" % (name, sub_page), { "example_form": ExampleForm(initial=form_initial), })
7a448c4df3feb717d0b1d8abbf9d32237751aab5
nbgrader/tests/apps/test_nbgrader_extension.py
nbgrader/tests/apps/test_nbgrader_extension.py
import os import nbgrader def test_nbextension(): from nbgrader import _jupyter_nbextension_paths nbexts = _jupyter_nbextension_paths() assert len(nbexts) == 3 assert nbexts[0]['section'] == 'tree' assert nbexts[1]['section'] == 'notebook' assert nbexts[2]['section'] == 'tree' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_serverextension(): from nbgrader import _jupyter_server_extension_paths serverexts = _jupyter_server_extension_paths() assert len(serverexts) == 2 assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list' assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader'
import os import nbgrader def test_nbextension(): from nbgrader import _jupyter_nbextension_paths nbexts = _jupyter_nbextension_paths() assert len(nbexts) == 4 assert nbexts[0]['section'] == 'tree' assert nbexts[1]['section'] == 'notebook' assert nbexts[2]['section'] == 'tree' assert nbexts[3]['section'] == 'notebook' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_serverextension(): from nbgrader import _jupyter_server_extension_paths serverexts = _jupyter_server_extension_paths() assert len(serverexts) == 3 assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list' assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader' assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment'
Fix tests for nbgrader extensions
Fix tests for nbgrader extensions
Python
bsd-3-clause
jhamrick/nbgrader,jhamrick/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jhamrick/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jhamrick/nbgrader
import os import nbgrader def test_nbextension(): from nbgrader import _jupyter_nbextension_paths nbexts = _jupyter_nbextension_paths() - assert len(nbexts) == 3 + assert len(nbexts) == 4 assert nbexts[0]['section'] == 'tree' assert nbexts[1]['section'] == 'notebook' assert nbexts[2]['section'] == 'tree' + assert nbexts[3]['section'] == 'notebook' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_serverextension(): from nbgrader import _jupyter_server_extension_paths serverexts = _jupyter_server_extension_paths() - assert len(serverexts) == 2 + assert len(serverexts) == 3 assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list' assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader' + assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment'
Fix tests for nbgrader extensions
## Code Before: import os import nbgrader def test_nbextension(): from nbgrader import _jupyter_nbextension_paths nbexts = _jupyter_nbextension_paths() assert len(nbexts) == 3 assert nbexts[0]['section'] == 'tree' assert nbexts[1]['section'] == 'notebook' assert nbexts[2]['section'] == 'tree' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_serverextension(): from nbgrader import _jupyter_server_extension_paths serverexts = _jupyter_server_extension_paths() assert len(serverexts) == 2 assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list' assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader' ## Instruction: Fix tests for nbgrader extensions ## Code After: import os import nbgrader def test_nbextension(): from nbgrader import _jupyter_nbextension_paths nbexts = _jupyter_nbextension_paths() assert len(nbexts) == 4 assert nbexts[0]['section'] == 'tree' assert nbexts[1]['section'] == 'notebook' assert nbexts[2]['section'] == 'tree' assert nbexts[3]['section'] == 'notebook' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_serverextension(): from nbgrader import _jupyter_server_extension_paths serverexts = _jupyter_server_extension_paths() assert len(serverexts) == 3 assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list' assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader' assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment'
import os import nbgrader def test_nbextension(): from nbgrader import _jupyter_nbextension_paths nbexts = _jupyter_nbextension_paths() - assert len(nbexts) == 3 ? ^ + assert len(nbexts) == 4 ? ^ assert nbexts[0]['section'] == 'tree' assert nbexts[1]['section'] == 'notebook' assert nbexts[2]['section'] == 'tree' + assert nbexts[3]['section'] == 'notebook' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_serverextension(): from nbgrader import _jupyter_server_extension_paths serverexts = _jupyter_server_extension_paths() - assert len(serverexts) == 2 ? ^ + assert len(serverexts) == 3 ? ^ assert serverexts[0]['module'] == 'nbgrader.server_extensions.assignment_list' assert serverexts[1]['module'] == 'nbgrader.server_extensions.formgrader' + assert serverexts[2]['module'] == 'nbgrader.server_extensions.validate_assignment'
a1f5a392d5270dd6f80a40e45c5e25b6ae04b7c3
embed_video/fields.py
embed_video/fields.py
from django.db import models from django import forms from django.utils.translation import ugettext_lazy as _ from .backends import detect_backend, UnknownIdException, \ UnknownBackendException __all__ = ('EmbedVideoField', 'EmbedVideoFormField') class EmbedVideoField(models.URLField): """ Model field for embeded video. Descendant of :py:class:`django.db.models.URLField`. """ def formfield(self, **kwargs): defaults = {'form_class': EmbedVideoFormField} defaults.update(kwargs) return super(EmbedVideoField, self).formfield(**defaults) def south_field_triple(self): from south.modelsinspector import introspector cls_name = '%s.%s' % ( self.__class__.__module__, self.__class__.__name__ ) args, kwargs = introspector(self) return (cls_name, args, kwargs) class EmbedVideoFormField(forms.URLField): """ Form field for embeded video. Descendant of :py:class:`django.forms.URLField` """ def validate(self, url): super(EmbedVideoFormField, self).validate(url) if url: try: detect_backend(url) except UnknownBackendException: raise forms.ValidationError(_(u'URL could not be recognized.')) except UnknownIdException: raise forms.ValidationError(_(u'ID of this video could not be \ recognized.')) return url
from django.db import models from django import forms from django.utils.translation import ugettext_lazy as _ from .backends import detect_backend, UnknownIdException, \ UnknownBackendException __all__ = ('EmbedVideoField', 'EmbedVideoFormField') class EmbedVideoField(models.URLField): """ Model field for embeded video. Descendant of :py:class:`django.db.models.URLField`. """ def formfield(self, **kwargs): defaults = {'form_class': EmbedVideoFormField} defaults.update(kwargs) return super(EmbedVideoField, self).formfield(**defaults) def south_field_triple(self): from south.modelsinspector import introspector cls_name = '%s.%s' % ( self.__class__.__module__, self.__class__.__name__ ) args, kwargs = introspector(self) return (cls_name, args, kwargs) class EmbedVideoFormField(forms.URLField): """ Form field for embeded video. Descendant of :py:class:`django.forms.URLField` """ def validate(self, url): # if empty url is not allowed throws an exception super(EmbedVideoFormField, self).validate(url) if not url: return try: detect_backend(url) except UnknownBackendException: raise forms.ValidationError(_(u'URL could not be recognized.')) except UnknownIdException: raise forms.ValidationError(_(u'ID of this video could not be \ recognized.')) return url
Simplify validate method in FormField.
Simplify validate method in FormField.
Python
mit
yetty/django-embed-video,jazzband/django-embed-video,jazzband/django-embed-video,mpachas/django-embed-video,yetty/django-embed-video,mpachas/django-embed-video
from django.db import models from django import forms from django.utils.translation import ugettext_lazy as _ from .backends import detect_backend, UnknownIdException, \ UnknownBackendException __all__ = ('EmbedVideoField', 'EmbedVideoFormField') class EmbedVideoField(models.URLField): """ Model field for embeded video. Descendant of :py:class:`django.db.models.URLField`. """ def formfield(self, **kwargs): defaults = {'form_class': EmbedVideoFormField} defaults.update(kwargs) return super(EmbedVideoField, self).formfield(**defaults) def south_field_triple(self): from south.modelsinspector import introspector cls_name = '%s.%s' % ( self.__class__.__module__, self.__class__.__name__ ) args, kwargs = introspector(self) return (cls_name, args, kwargs) class EmbedVideoFormField(forms.URLField): """ Form field for embeded video. Descendant of :py:class:`django.forms.URLField` """ def validate(self, url): + # if empty url is not allowed throws an exception super(EmbedVideoFormField, self).validate(url) + + if not url: + return - if url: - try: + try: - detect_backend(url) + detect_backend(url) - except UnknownBackendException: + except UnknownBackendException: - raise forms.ValidationError(_(u'URL could not be recognized.')) + raise forms.ValidationError(_(u'URL could not be recognized.')) - except UnknownIdException: + except UnknownIdException: - raise forms.ValidationError(_(u'ID of this video could not be \ + raise forms.ValidationError(_(u'ID of this video could not be \ - recognized.')) + recognized.')) - return url
Simplify validate method in FormField.
## Code Before: from django.db import models from django import forms from django.utils.translation import ugettext_lazy as _ from .backends import detect_backend, UnknownIdException, \ UnknownBackendException __all__ = ('EmbedVideoField', 'EmbedVideoFormField') class EmbedVideoField(models.URLField): """ Model field for embeded video. Descendant of :py:class:`django.db.models.URLField`. """ def formfield(self, **kwargs): defaults = {'form_class': EmbedVideoFormField} defaults.update(kwargs) return super(EmbedVideoField, self).formfield(**defaults) def south_field_triple(self): from south.modelsinspector import introspector cls_name = '%s.%s' % ( self.__class__.__module__, self.__class__.__name__ ) args, kwargs = introspector(self) return (cls_name, args, kwargs) class EmbedVideoFormField(forms.URLField): """ Form field for embeded video. Descendant of :py:class:`django.forms.URLField` """ def validate(self, url): super(EmbedVideoFormField, self).validate(url) if url: try: detect_backend(url) except UnknownBackendException: raise forms.ValidationError(_(u'URL could not be recognized.')) except UnknownIdException: raise forms.ValidationError(_(u'ID of this video could not be \ recognized.')) return url ## Instruction: Simplify validate method in FormField. ## Code After: from django.db import models from django import forms from django.utils.translation import ugettext_lazy as _ from .backends import detect_backend, UnknownIdException, \ UnknownBackendException __all__ = ('EmbedVideoField', 'EmbedVideoFormField') class EmbedVideoField(models.URLField): """ Model field for embeded video. Descendant of :py:class:`django.db.models.URLField`. """ def formfield(self, **kwargs): defaults = {'form_class': EmbedVideoFormField} defaults.update(kwargs) return super(EmbedVideoField, self).formfield(**defaults) def south_field_triple(self): from south.modelsinspector import introspector cls_name = '%s.%s' % ( self.__class__.__module__, self.__class__.__name__ ) args, kwargs = introspector(self) return (cls_name, args, kwargs) class EmbedVideoFormField(forms.URLField): """ Form field for embeded video. Descendant of :py:class:`django.forms.URLField` """ def validate(self, url): # if empty url is not allowed throws an exception super(EmbedVideoFormField, self).validate(url) if not url: return try: detect_backend(url) except UnknownBackendException: raise forms.ValidationError(_(u'URL could not be recognized.')) except UnknownIdException: raise forms.ValidationError(_(u'ID of this video could not be \ recognized.')) return url
from django.db import models from django import forms from django.utils.translation import ugettext_lazy as _ from .backends import detect_backend, UnknownIdException, \ UnknownBackendException __all__ = ('EmbedVideoField', 'EmbedVideoFormField') class EmbedVideoField(models.URLField): """ Model field for embeded video. Descendant of :py:class:`django.db.models.URLField`. """ def formfield(self, **kwargs): defaults = {'form_class': EmbedVideoFormField} defaults.update(kwargs) return super(EmbedVideoField, self).formfield(**defaults) def south_field_triple(self): from south.modelsinspector import introspector cls_name = '%s.%s' % ( self.__class__.__module__, self.__class__.__name__ ) args, kwargs = introspector(self) return (cls_name, args, kwargs) class EmbedVideoFormField(forms.URLField): """ Form field for embeded video. Descendant of :py:class:`django.forms.URLField` """ def validate(self, url): + # if empty url is not allowed throws an exception super(EmbedVideoFormField, self).validate(url) + + if not url: + return - if url: - try: ? ---- + try: - detect_backend(url) ? ---- + detect_backend(url) - except UnknownBackendException: ? ---- + except UnknownBackendException: - raise forms.ValidationError(_(u'URL could not be recognized.')) ? ---- + raise forms.ValidationError(_(u'URL could not be recognized.')) - except UnknownIdException: ? ---- + except UnknownIdException: - raise forms.ValidationError(_(u'ID of this video could not be \ ? ---- + raise forms.ValidationError(_(u'ID of this video could not be \ - recognized.')) ? ---- + recognized.')) - return url
1f697a2c7bcf0f7769a9fc4f81be676ed5ee97c6
examples/flask/flask_seguro/cart.py
examples/flask/flask_seguro/cart.py
from flask_seguro.products import Products from flask import current_app as app class Cart: def __init__(self, cart_dict={}): if cart_dict == {}: self.total = 0 self.subtotal = 0 self.items = [] else: self.total = cart_dict["total"] self.subtotal = cart_dict["subtotal"] self.items = cart_dict["items"] self.extra_amount = float(app.config['EXTRA_AMOUNT']) def to_dict(self): return {"total": self.total, "subtotal": self.subtotal, "items": self.items, "extra_amount": self.extra_amount} def change_item(self, item_id, operation): product = Products().get_one(item_id) if product: if operation == 'add': self.items.append(product) elif operation == 'remove': cart_product = filter( lambda x: x['id'] == product['id'], self.items) self.items.remove(cart_product[0]) self.update() return True else: return False def update(self): subtotal = float(0) total = float(0) for product in self.items: subtotal += float(product["price"]) if subtotal > 0: total = subtotal + self.extra_amount self.subtotal = subtotal self.total = total
from flask_seguro.products import Products from flask import current_app as app class Cart: def __init__(self, cart_dict=None): cart_dict = cart_dict or {} if cart_dict == {}: self.total = 0 self.subtotal = 0 self.items = [] else: self.total = cart_dict["total"] self.subtotal = cart_dict["subtotal"] self.items = cart_dict["items"] self.extra_amount = float(app.config['EXTRA_AMOUNT']) def to_dict(self): return {"total": self.total, "subtotal": self.subtotal, "items": self.items, "extra_amount": self.extra_amount} def change_item(self, item_id, operation): product = Products().get_one(item_id) if product: if operation == 'add': self.items.append(product) elif operation == 'remove': cart_product = filter( lambda x: x['id'] == product['id'], self.items) self.items.remove(cart_product[0]) self.update() return True else: return False def update(self): subtotal = float(0) total = float(0) for product in self.items: subtotal += float(product["price"]) if subtotal > 0: total = subtotal + self.extra_amount self.subtotal = subtotal self.total = total
Fix dangerous default mutable value
Fix dangerous default mutable value
Python
mit
rgcarrasqueira/python-pagseguro,vintasoftware/python-pagseguro,rochacbruno/python-pagseguro
from flask_seguro.products import Products from flask import current_app as app class Cart: - def __init__(self, cart_dict={}): + def __init__(self, cart_dict=None): + cart_dict = cart_dict or {} if cart_dict == {}: self.total = 0 self.subtotal = 0 self.items = [] else: self.total = cart_dict["total"] self.subtotal = cart_dict["subtotal"] self.items = cart_dict["items"] self.extra_amount = float(app.config['EXTRA_AMOUNT']) def to_dict(self): return {"total": self.total, "subtotal": self.subtotal, "items": self.items, "extra_amount": self.extra_amount} def change_item(self, item_id, operation): product = Products().get_one(item_id) if product: if operation == 'add': self.items.append(product) elif operation == 'remove': cart_product = filter( lambda x: x['id'] == product['id'], self.items) self.items.remove(cart_product[0]) self.update() return True else: return False def update(self): subtotal = float(0) total = float(0) for product in self.items: subtotal += float(product["price"]) if subtotal > 0: total = subtotal + self.extra_amount self.subtotal = subtotal self.total = total
Fix dangerous default mutable value
## Code Before: from flask_seguro.products import Products from flask import current_app as app class Cart: def __init__(self, cart_dict={}): if cart_dict == {}: self.total = 0 self.subtotal = 0 self.items = [] else: self.total = cart_dict["total"] self.subtotal = cart_dict["subtotal"] self.items = cart_dict["items"] self.extra_amount = float(app.config['EXTRA_AMOUNT']) def to_dict(self): return {"total": self.total, "subtotal": self.subtotal, "items": self.items, "extra_amount": self.extra_amount} def change_item(self, item_id, operation): product = Products().get_one(item_id) if product: if operation == 'add': self.items.append(product) elif operation == 'remove': cart_product = filter( lambda x: x['id'] == product['id'], self.items) self.items.remove(cart_product[0]) self.update() return True else: return False def update(self): subtotal = float(0) total = float(0) for product in self.items: subtotal += float(product["price"]) if subtotal > 0: total = subtotal + self.extra_amount self.subtotal = subtotal self.total = total ## Instruction: Fix dangerous default mutable value ## Code After: from flask_seguro.products import Products from flask import current_app as app class Cart: def __init__(self, cart_dict=None): cart_dict = cart_dict or {} if cart_dict == {}: self.total = 0 self.subtotal = 0 self.items = [] else: self.total = cart_dict["total"] self.subtotal = cart_dict["subtotal"] self.items = cart_dict["items"] self.extra_amount = float(app.config['EXTRA_AMOUNT']) def to_dict(self): return {"total": self.total, "subtotal": self.subtotal, "items": self.items, "extra_amount": self.extra_amount} def change_item(self, item_id, operation): product = Products().get_one(item_id) if product: if operation == 'add': self.items.append(product) elif operation == 'remove': cart_product = filter( lambda x: x['id'] == product['id'], self.items) self.items.remove(cart_product[0]) self.update() return True else: return False def update(self): subtotal = float(0) total = float(0) for product in self.items: subtotal += float(product["price"]) if subtotal > 0: total = subtotal + self.extra_amount self.subtotal = subtotal self.total = total
from flask_seguro.products import Products from flask import current_app as app class Cart: - def __init__(self, cart_dict={}): ? ^^ + def __init__(self, cart_dict=None): ? ^^^^ + cart_dict = cart_dict or {} if cart_dict == {}: self.total = 0 self.subtotal = 0 self.items = [] else: self.total = cart_dict["total"] self.subtotal = cart_dict["subtotal"] self.items = cart_dict["items"] self.extra_amount = float(app.config['EXTRA_AMOUNT']) def to_dict(self): return {"total": self.total, "subtotal": self.subtotal, "items": self.items, "extra_amount": self.extra_amount} def change_item(self, item_id, operation): product = Products().get_one(item_id) if product: if operation == 'add': self.items.append(product) elif operation == 'remove': cart_product = filter( lambda x: x['id'] == product['id'], self.items) self.items.remove(cart_product[0]) self.update() return True else: return False def update(self): subtotal = float(0) total = float(0) for product in self.items: subtotal += float(product["price"]) if subtotal > 0: total = subtotal + self.extra_amount self.subtotal = subtotal self.total = total
fff0b4af89e02ff834221ef056b7dcb979dc6cd7
webpay/webpay.py
webpay/webpay.py
from .api import Account, Charges, Customers import requests class WebPay: def __init__(self, key, api_base = 'https://api.webpay.jp/v1'): self.key = key self.api_base = api_base self.account = Account(self) self.charges = Charges(self) self.customers = Customers(self) def post(self, path, params): r = requests.post(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def get(self, path, params = {}): r = requests.get(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def delete(self, path, params = {}): r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params) return r.json()
from .api import Account, Charges, Customers import requests import json class WebPay: def __init__(self, key, api_base = 'https://api.webpay.jp/v1'): self.key = key self.api_base = api_base self.account = Account(self) self.charges = Charges(self) self.customers = Customers(self) def post(self, path, params): r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) return r.json() def get(self, path, params = {}): r = requests.get(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def delete(self, path, params = {}): r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) return r.json()
Use JSON for other than GET request
Use JSON for other than GET request Because internal dict parameters is not handled as expected. >>> payload = {'key1': 'value1', 'key2': 'value2', 'set': {'a': 'x', 'b': 'y'}} >>> r = requests.post("http://httpbin.org/post", data=payload) >>> r.json() {... 'form': {'key2': 'value2', 'key1': 'value1', 'set': ['a', 'b']} ...}
Python
mit
yamaneko1212/webpay-python
from .api import Account, Charges, Customers import requests + import json class WebPay: def __init__(self, key, api_base = 'https://api.webpay.jp/v1'): self.key = key self.api_base = api_base self.account = Account(self) self.charges = Charges(self) self.customers = Customers(self) def post(self, path, params): - r = requests.post(self.api_base + path, auth = (self.key, ''), params = params) + r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) return r.json() def get(self, path, params = {}): r = requests.get(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def delete(self, path, params = {}): - r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params) + r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) return r.json()
Use JSON for other than GET request
## Code Before: from .api import Account, Charges, Customers import requests class WebPay: def __init__(self, key, api_base = 'https://api.webpay.jp/v1'): self.key = key self.api_base = api_base self.account = Account(self) self.charges = Charges(self) self.customers = Customers(self) def post(self, path, params): r = requests.post(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def get(self, path, params = {}): r = requests.get(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def delete(self, path, params = {}): r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params) return r.json() ## Instruction: Use JSON for other than GET request ## Code After: from .api import Account, Charges, Customers import requests import json class WebPay: def __init__(self, key, api_base = 'https://api.webpay.jp/v1'): self.key = key self.api_base = api_base self.account = Account(self) self.charges = Charges(self) self.customers = Customers(self) def post(self, path, params): r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) return r.json() def get(self, path, params = {}): r = requests.get(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def delete(self, path, params = {}): r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) return r.json()
from .api import Account, Charges, Customers import requests + import json class WebPay: def __init__(self, key, api_base = 'https://api.webpay.jp/v1'): self.key = key self.api_base = api_base self.account = Account(self) self.charges = Charges(self) self.customers = Customers(self) def post(self, path, params): - r = requests.post(self.api_base + path, auth = (self.key, ''), params = params) ? ^ ^ -- + r = requests.post(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) ? ^ ^ +++++++++++ + return r.json() def get(self, path, params = {}): r = requests.get(self.api_base + path, auth = (self.key, ''), params = params) return r.json() def delete(self, path, params = {}): - r = requests.delete(self.api_base + path, auth = (self.key, ''), params = params) ? ^ ^ -- + r = requests.delete(self.api_base + path, auth = (self.key, ''), data = json.dumps(params)) ? ^ ^ +++++++++++ + return r.json()
d953055801c8d618c70cea81e3e35684122c66a7
setuptools/config/__init__.py
setuptools/config/__init__.py
import warnings from functools import wraps from textwrap import dedent from typing import Callable, TypeVar, cast from .._deprecation_warning import SetuptoolsDeprecationWarning from . import setupcfg Fn = TypeVar("Fn", bound=Callable) __all__ = ('parse_configuration', 'read_configuration') def _deprecation_notice(fn: Fn) -> Fn: @wraps(fn) def _wrapper(*args, **kwargs): msg = f"""\ As setuptools moves its configuration towards `pyproject.toml`, `{__name__}.{fn.__name__}` became deprecated. For the time being, you can use the `{setupcfg.__name__}` module to access a backward compatible API, but this module is provisional and might be removed in the future. """ warnings.warn(dedent(msg), SetuptoolsDeprecationWarning) return fn(*args, **kwargs) return cast(Fn, _wrapper) read_configuration = _deprecation_notice(setupcfg.read_configuration) parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
import warnings from functools import wraps from textwrap import dedent from typing import Callable, TypeVar, cast from .._deprecation_warning import SetuptoolsDeprecationWarning from . import setupcfg Fn = TypeVar("Fn", bound=Callable) __all__ = ('parse_configuration', 'read_configuration') def _deprecation_notice(fn: Fn) -> Fn: @wraps(fn) def _wrapper(*args, **kwargs): msg = f"""\ As setuptools moves its configuration towards `pyproject.toml`, `{__name__}.{fn.__name__}` became deprecated. For the time being, you can use the `{setupcfg.__name__}` module to access a backward compatible API, but this module is provisional and might be removed in the future. """ warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2) return fn(*args, **kwargs) return cast(Fn, _wrapper) read_configuration = _deprecation_notice(setupcfg.read_configuration) parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
Add stacklevel=2 to make calling code clear
Add stacklevel=2 to make calling code clear
Python
mit
pypa/setuptools,pypa/setuptools,pypa/setuptools
import warnings from functools import wraps from textwrap import dedent from typing import Callable, TypeVar, cast from .._deprecation_warning import SetuptoolsDeprecationWarning from . import setupcfg Fn = TypeVar("Fn", bound=Callable) __all__ = ('parse_configuration', 'read_configuration') def _deprecation_notice(fn: Fn) -> Fn: @wraps(fn) def _wrapper(*args, **kwargs): msg = f"""\ As setuptools moves its configuration towards `pyproject.toml`, `{__name__}.{fn.__name__}` became deprecated. For the time being, you can use the `{setupcfg.__name__}` module to access a backward compatible API, but this module is provisional and might be removed in the future. """ - warnings.warn(dedent(msg), SetuptoolsDeprecationWarning) + warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2) return fn(*args, **kwargs) return cast(Fn, _wrapper) read_configuration = _deprecation_notice(setupcfg.read_configuration) parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
Add stacklevel=2 to make calling code clear
## Code Before: import warnings from functools import wraps from textwrap import dedent from typing import Callable, TypeVar, cast from .._deprecation_warning import SetuptoolsDeprecationWarning from . import setupcfg Fn = TypeVar("Fn", bound=Callable) __all__ = ('parse_configuration', 'read_configuration') def _deprecation_notice(fn: Fn) -> Fn: @wraps(fn) def _wrapper(*args, **kwargs): msg = f"""\ As setuptools moves its configuration towards `pyproject.toml`, `{__name__}.{fn.__name__}` became deprecated. For the time being, you can use the `{setupcfg.__name__}` module to access a backward compatible API, but this module is provisional and might be removed in the future. """ warnings.warn(dedent(msg), SetuptoolsDeprecationWarning) return fn(*args, **kwargs) return cast(Fn, _wrapper) read_configuration = _deprecation_notice(setupcfg.read_configuration) parse_configuration = _deprecation_notice(setupcfg.parse_configuration) ## Instruction: Add stacklevel=2 to make calling code clear ## Code After: import warnings from functools import wraps from textwrap import dedent from typing import Callable, TypeVar, cast from .._deprecation_warning import SetuptoolsDeprecationWarning from . import setupcfg Fn = TypeVar("Fn", bound=Callable) __all__ = ('parse_configuration', 'read_configuration') def _deprecation_notice(fn: Fn) -> Fn: @wraps(fn) def _wrapper(*args, **kwargs): msg = f"""\ As setuptools moves its configuration towards `pyproject.toml`, `{__name__}.{fn.__name__}` became deprecated. For the time being, you can use the `{setupcfg.__name__}` module to access a backward compatible API, but this module is provisional and might be removed in the future. """ warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2) return fn(*args, **kwargs) return cast(Fn, _wrapper) read_configuration = _deprecation_notice(setupcfg.read_configuration) parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
import warnings from functools import wraps from textwrap import dedent from typing import Callable, TypeVar, cast from .._deprecation_warning import SetuptoolsDeprecationWarning from . import setupcfg Fn = TypeVar("Fn", bound=Callable) __all__ = ('parse_configuration', 'read_configuration') def _deprecation_notice(fn: Fn) -> Fn: @wraps(fn) def _wrapper(*args, **kwargs): msg = f"""\ As setuptools moves its configuration towards `pyproject.toml`, `{__name__}.{fn.__name__}` became deprecated. For the time being, you can use the `{setupcfg.__name__}` module to access a backward compatible API, but this module is provisional and might be removed in the future. """ - warnings.warn(dedent(msg), SetuptoolsDeprecationWarning) + warnings.warn(dedent(msg), SetuptoolsDeprecationWarning, stacklevel=2) ? ++++++++++++++ return fn(*args, **kwargs) return cast(Fn, _wrapper) read_configuration = _deprecation_notice(setupcfg.read_configuration) parse_configuration = _deprecation_notice(setupcfg.parse_configuration)
261393eb46cdc082b60d9ea11ec862f508632ad2
audit_log/models/__init__.py
audit_log/models/__init__.py
from django.db.models import Model from django.utils.translation import ugettext_lazy as _ from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField class AuthStampedModel(Model): """ An abstract base class model that provides auth and session information fields. """ created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") created_with_session_key = CreatingSessionKeyField(_("created with session key")) modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") modified_with_session_key = LastSessionKeyField(_("modified with session key")) class Meta: abstract = True
from django.db.models import Model, SET_NULL from django.utils.translation import ugettext_lazy as _ from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField class AuthStampedModel(Model): """ An abstract base class model that provides auth and session information fields. """ created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") created_with_session_key = CreatingSessionKeyField(_("created with session key")) modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") modified_with_session_key = LastSessionKeyField(_("modified with session key")) class Meta: abstract = True
Add mandatory `on_delete` and allow nulls
Add mandatory `on_delete` and allow nulls I assume it's better to allow nulls than to have auth stamped models disappear with deleted users.
Python
bsd-3-clause
Atomidata/django-audit-log,Atomidata/django-audit-log
- from django.db.models import Model + from django.db.models import Model, SET_NULL from django.utils.translation import ugettext_lazy as _ from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField class AuthStampedModel(Model): """ An abstract base class model that provides auth and session information fields. """ - created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") + created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") created_with_session_key = CreatingSessionKeyField(_("created with session key")) - modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") + modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") modified_with_session_key = LastSessionKeyField(_("modified with session key")) class Meta: abstract = True
Add mandatory `on_delete` and allow nulls
## Code Before: from django.db.models import Model from django.utils.translation import ugettext_lazy as _ from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField class AuthStampedModel(Model): """ An abstract base class model that provides auth and session information fields. """ created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") created_with_session_key = CreatingSessionKeyField(_("created with session key")) modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") modified_with_session_key = LastSessionKeyField(_("modified with session key")) class Meta: abstract = True ## Instruction: Add mandatory `on_delete` and allow nulls ## Code After: from django.db.models import Model, SET_NULL from django.utils.translation import ugettext_lazy as _ from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField class AuthStampedModel(Model): """ An abstract base class model that provides auth and session information fields. """ created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") created_with_session_key = CreatingSessionKeyField(_("created with session key")) modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") modified_with_session_key = LastSessionKeyField(_("modified with session key")) class Meta: abstract = True
- from django.db.models import Model + from django.db.models import Model, SET_NULL ? ++++++++++ from django.utils.translation import ugettext_lazy as _ from audit_log.models.fields import CreatingUserField, CreatingSessionKeyField, LastUserField, LastSessionKeyField class AuthStampedModel(Model): """ An abstract base class model that provides auth and session information fields. """ - created_by = CreatingUserField(verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") + created_by = CreatingUserField(on_delete=SET_NULL, null=True, verbose_name = _("created by"), related_name = "created_%(app_label)s_%(class)s_set") ? +++++++++++++++++++++++++++++++ created_with_session_key = CreatingSessionKeyField(_("created with session key")) - modified_by = LastUserField(verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") + modified_by = LastUserField(on_delete=SET_NULL, null=True, verbose_name = _("modified by"), related_name = "modified_%(app_label)s_%(class)s_set") ? +++++++++++++++++++++++++++++++ modified_with_session_key = LastSessionKeyField(_("modified with session key")) class Meta: abstract = True
c21fe453911af190f3cbd93356396d4f5e65195e
mopidy/backends/gstreamer.py
mopidy/backends/gstreamer.py
import logging import gst from mopidy import config from mopidy.backends import BaseBackend from mopidy.models import Artist, Album, Track, Playlist logger = logging.getLogger(u'backends.gstreamer') class GStreamerBackend(BaseBackend): def __init__(self, *args, **kwargs): super(GStreamerBackend, self).__init__(*args, **kwargs) playlist = [] player = gst.element_factory_make("playbin2", "player") fakesink = gst.element_factory_make("fakesink", "fakesink") player.set_property("video-sink", fakesink) self.player = player def _play_current_track(self): self.player.set_property("uri", self._current_track.uri) self.player.set_state(gst.STATE_PLAYING) def _play(self): if self._current_track is not None: self._play_current_track() return True else: return False def _stop(self): self.player.set_state(gst.STATE_NULL) return True
import logging import gst from mopidy import config from mopidy.backends import BaseBackend from mopidy.models import Artist, Album, Track, Playlist logger = logging.getLogger(u'backends.gstreamer') class GStreamerBackend(BaseBackend): def __init__(self, *args, **kwargs): super(GStreamerBackend, self).__init__(*args, **kwargs) playlist = [] player = gst.element_factory_make("playbin2", "player") fakesink = gst.element_factory_make("fakesink", "fakesink") player.set_property("video-sink", fakesink) self.player = player def _play(self): if self._current_track is None: return False self.player.set_property("uri", self._current_track.uri) self.player.set_state(gst.STATE_PLAYING) return True def _stop(self): self.player.set_state(gst.STATE_NULL) return True
Clean play code for GStreamer
Clean play code for GStreamer
Python
apache-2.0
vrs01/mopidy,bacontext/mopidy,vrs01/mopidy,quartz55/mopidy,priestd09/mopidy,swak/mopidy,liamw9534/mopidy,hkariti/mopidy,pacificIT/mopidy,ZenithDK/mopidy,hkariti/mopidy,dbrgn/mopidy,kingosticks/mopidy,ZenithDK/mopidy,diandiankan/mopidy,vrs01/mopidy,hkariti/mopidy,pacificIT/mopidy,ali/mopidy,SuperStarPL/mopidy,adamcik/mopidy,swak/mopidy,glogiotatidis/mopidy,adamcik/mopidy,SuperStarPL/mopidy,mokieyue/mopidy,rawdlite/mopidy,priestd09/mopidy,diandiankan/mopidy,swak/mopidy,tkem/mopidy,diandiankan/mopidy,dbrgn/mopidy,mopidy/mopidy,woutervanwijk/mopidy,jodal/mopidy,bacontext/mopidy,jcass77/mopidy,ali/mopidy,jcass77/mopidy,bacontext/mopidy,ali/mopidy,SuperStarPL/mopidy,abarisain/mopidy,mopidy/mopidy,mopidy/mopidy,bacontext/mopidy,quartz55/mopidy,quartz55/mopidy,quartz55/mopidy,ZenithDK/mopidy,rawdlite/mopidy,ali/mopidy,abarisain/mopidy,hkariti/mopidy,tkem/mopidy,ZenithDK/mopidy,tkem/mopidy,rawdlite/mopidy,jmarsik/mopidy,jodal/mopidy,diandiankan/mopidy,priestd09/mopidy,tkem/mopidy,liamw9534/mopidy,bencevans/mopidy,dbrgn/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,swak/mopidy,glogiotatidis/mopidy,glogiotatidis/mopidy,mokieyue/mopidy,dbrgn/mopidy,bencevans/mopidy,SuperStarPL/mopidy,adamcik/mopidy,jmarsik/mopidy,jcass77/mopidy,kingosticks/mopidy,rawdlite/mopidy,vrs01/mopidy,mokieyue/mopidy,bencevans/mopidy,bencevans/mopidy,jmarsik/mopidy,jmarsik/mopidy,kingosticks/mopidy,woutervanwijk/mopidy,mokieyue/mopidy,jodal/mopidy,pacificIT/mopidy
import logging import gst from mopidy import config from mopidy.backends import BaseBackend from mopidy.models import Artist, Album, Track, Playlist logger = logging.getLogger(u'backends.gstreamer') class GStreamerBackend(BaseBackend): def __init__(self, *args, **kwargs): super(GStreamerBackend, self).__init__(*args, **kwargs) playlist = [] player = gst.element_factory_make("playbin2", "player") fakesink = gst.element_factory_make("fakesink", "fakesink") player.set_property("video-sink", fakesink) self.player = player + def _play(self): + if self._current_track is None: + return False - def _play_current_track(self): self.player.set_property("uri", self._current_track.uri) self.player.set_state(gst.STATE_PLAYING) - def _play(self): - if self._current_track is not None: - self._play_current_track() - return True + return True - else: - return False def _stop(self): self.player.set_state(gst.STATE_NULL) return True
Clean play code for GStreamer
## Code Before: import logging import gst from mopidy import config from mopidy.backends import BaseBackend from mopidy.models import Artist, Album, Track, Playlist logger = logging.getLogger(u'backends.gstreamer') class GStreamerBackend(BaseBackend): def __init__(self, *args, **kwargs): super(GStreamerBackend, self).__init__(*args, **kwargs) playlist = [] player = gst.element_factory_make("playbin2", "player") fakesink = gst.element_factory_make("fakesink", "fakesink") player.set_property("video-sink", fakesink) self.player = player def _play_current_track(self): self.player.set_property("uri", self._current_track.uri) self.player.set_state(gst.STATE_PLAYING) def _play(self): if self._current_track is not None: self._play_current_track() return True else: return False def _stop(self): self.player.set_state(gst.STATE_NULL) return True ## Instruction: Clean play code for GStreamer ## Code After: import logging import gst from mopidy import config from mopidy.backends import BaseBackend from mopidy.models import Artist, Album, Track, Playlist logger = logging.getLogger(u'backends.gstreamer') class GStreamerBackend(BaseBackend): def __init__(self, *args, **kwargs): super(GStreamerBackend, self).__init__(*args, **kwargs) playlist = [] player = gst.element_factory_make("playbin2", "player") fakesink = gst.element_factory_make("fakesink", "fakesink") player.set_property("video-sink", fakesink) self.player = player def _play(self): if self._current_track is None: return False self.player.set_property("uri", self._current_track.uri) self.player.set_state(gst.STATE_PLAYING) return True def _stop(self): self.player.set_state(gst.STATE_NULL) return True
import logging import gst from mopidy import config from mopidy.backends import BaseBackend from mopidy.models import Artist, Album, Track, Playlist logger = logging.getLogger(u'backends.gstreamer') class GStreamerBackend(BaseBackend): def __init__(self, *args, **kwargs): super(GStreamerBackend, self).__init__(*args, **kwargs) playlist = [] player = gst.element_factory_make("playbin2", "player") fakesink = gst.element_factory_make("fakesink", "fakesink") player.set_property("video-sink", fakesink) self.player = player + def _play(self): + if self._current_track is None: + return False - def _play_current_track(self): self.player.set_property("uri", self._current_track.uri) self.player.set_state(gst.STATE_PLAYING) - def _play(self): - if self._current_track is not None: - self._play_current_track() - return True ? ---- + return True - else: - return False def _stop(self): self.player.set_state(gst.STATE_NULL) return True
b7c967ad0f45cc1144a8713c6513bae5bca89242
LiSE/LiSE/test_proxy.py
LiSE/LiSE/test_proxy.py
from LiSE.proxy import EngineProcessManager import allegedb.test class ProxyTest(allegedb.test.AllegedTest): def setUp(self): self.manager = EngineProcessManager() self.engine = self.manager.start('sqlite:///:memory:') self.graphmakers = (self.engine.new_character,) def tearDown(self): self.manager.shutdown() class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest): pass class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest): pass class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest): pass class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest): pass class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest): pass
from LiSE.proxy import EngineProcessManager import allegedb.test class ProxyTest(allegedb.test.AllegedTest): def setUp(self): self.manager = EngineProcessManager() self.engine = self.manager.start('sqlite:///:memory:') self.graphmakers = (self.engine.new_character,) def tearDown(self): self.manager.shutdown() class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest): pass class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest): pass class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest): pass class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest): pass
Delete BranchLineageTest, which assumes bidirectional graphs exist
Delete BranchLineageTest, which assumes bidirectional graphs exist
Python
agpl-3.0
LogicalDash/LiSE,LogicalDash/LiSE
from LiSE.proxy import EngineProcessManager import allegedb.test class ProxyTest(allegedb.test.AllegedTest): def setUp(self): self.manager = EngineProcessManager() self.engine = self.manager.start('sqlite:///:memory:') self.graphmakers = (self.engine.new_character,) def tearDown(self): self.manager.shutdown() class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest): pass - class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest): - pass - - class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest): pass class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest): pass class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest): pass
Delete BranchLineageTest, which assumes bidirectional graphs exist
## Code Before: from LiSE.proxy import EngineProcessManager import allegedb.test class ProxyTest(allegedb.test.AllegedTest): def setUp(self): self.manager = EngineProcessManager() self.engine = self.manager.start('sqlite:///:memory:') self.graphmakers = (self.engine.new_character,) def tearDown(self): self.manager.shutdown() class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest): pass class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest): pass class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest): pass class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest): pass class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest): pass ## Instruction: Delete BranchLineageTest, which assumes bidirectional graphs exist ## Code After: from LiSE.proxy import EngineProcessManager import allegedb.test class ProxyTest(allegedb.test.AllegedTest): def setUp(self): self.manager = EngineProcessManager() self.engine = self.manager.start('sqlite:///:memory:') self.graphmakers = (self.engine.new_character,) def tearDown(self): self.manager.shutdown() class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest): pass class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest): pass class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest): pass class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest): pass
from LiSE.proxy import EngineProcessManager import allegedb.test class ProxyTest(allegedb.test.AllegedTest): def setUp(self): self.manager = EngineProcessManager() self.engine = self.manager.start('sqlite:///:memory:') self.graphmakers = (self.engine.new_character,) def tearDown(self): self.manager.shutdown() class ProxyGraphTest(allegedb.test.AbstractGraphTest, ProxyTest): pass - class BranchLineageTest(ProxyGraphTest, allegedb.test.AbstractBranchLineageTest): - pass - - class DictStorageTest(ProxyTest, allegedb.test.DictStorageTest): pass class ListStorageTest(ProxyTest, allegedb.test.ListStorageTest): pass class SetStorageTest(ProxyTest, allegedb.test.SetStorageTest): pass
f0f3c50a65aae1393928579ca0e48891d1ac8f18
app/access_control.py
app/access_control.py
from functools import wraps from flask import flash, redirect, url_for, session from app import views def login_required(f): @wraps(f) def decorated_function(*args, **kwargs): if 'logged_in' in session: return f(*args, **kwargs) else: flash("Please login to continue.", "danger") return redirect(url_for("login")) return decorated_function
from functools import wraps from flask import flash, redirect, url_for, session from app import views def login_required(f): @wraps(f) def decorated_function(*args, **kwargs): if 'logged_in' in session: return f(*args, **kwargs) else: flash("Please login to continue.", "danger") return redirect(url_for("login")) return decorated_function def for_guests(f): @wraps(f) def decorated_function(*args, **kwrags): if not 'logged_in' in session: return f(*args, **kwargs) else: flash("Invalid Action.", "danger") return redirect(url_for("dashboard")) return decorated_function
Create a decorator `for_guest` for access control on pages for guests.
Create a decorator `for_guest` for access control on pages for guests.
Python
mit
alchermd/flask-todo-app,alchermd/flask-todo-app
from functools import wraps from flask import flash, redirect, url_for, session from app import views def login_required(f): @wraps(f) def decorated_function(*args, **kwargs): if 'logged_in' in session: return f(*args, **kwargs) else: flash("Please login to continue.", "danger") return redirect(url_for("login")) return decorated_function + + def for_guests(f): + @wraps(f) + def decorated_function(*args, **kwrags): + if not 'logged_in' in session: + return f(*args, **kwargs) + else: + flash("Invalid Action.", "danger") + return redirect(url_for("dashboard")) + return decorated_function
Create a decorator `for_guest` for access control on pages for guests.
## Code Before: from functools import wraps from flask import flash, redirect, url_for, session from app import views def login_required(f): @wraps(f) def decorated_function(*args, **kwargs): if 'logged_in' in session: return f(*args, **kwargs) else: flash("Please login to continue.", "danger") return redirect(url_for("login")) return decorated_function ## Instruction: Create a decorator `for_guest` for access control on pages for guests. ## Code After: from functools import wraps from flask import flash, redirect, url_for, session from app import views def login_required(f): @wraps(f) def decorated_function(*args, **kwargs): if 'logged_in' in session: return f(*args, **kwargs) else: flash("Please login to continue.", "danger") return redirect(url_for("login")) return decorated_function def for_guests(f): @wraps(f) def decorated_function(*args, **kwrags): if not 'logged_in' in session: return f(*args, **kwargs) else: flash("Invalid Action.", "danger") return redirect(url_for("dashboard")) return decorated_function
from functools import wraps from flask import flash, redirect, url_for, session from app import views def login_required(f): @wraps(f) def decorated_function(*args, **kwargs): if 'logged_in' in session: return f(*args, **kwargs) else: flash("Please login to continue.", "danger") return redirect(url_for("login")) return decorated_function + + def for_guests(f): + @wraps(f) + def decorated_function(*args, **kwrags): + if not 'logged_in' in session: + return f(*args, **kwargs) + else: + flash("Invalid Action.", "danger") + return redirect(url_for("dashboard")) + return decorated_function
37333506e6866e7d0859c5068f115a3e1b9dec3a
test/test_coordinate.py
test/test_coordinate.py
import unittest from src import coordinate class TestRules(unittest.TestCase): """ Tests for the coordinate module """ def test_get_x_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 4 actual_result = board_location.get_x_board() self.assertEqual(actual_result, expected_result) def test_get_y_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 6 actual_result = board_location.get_y_board() self.assertEqual(actual_result, expected_result) def test_get_x_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 3 actual_result = board_location.get_x_array() self.assertEqual(actual_result, expected_result) def test_get_y_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 5 actual_result = board_location.get_y_array() self.assertEqual(actual_result, expected_result)
import unittest from src import coordinate class TestRules(unittest.TestCase): """ Tests for the coordinate module """ def test_get_x_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 4 actual_result = board_location.get_x_board() self.assertEqual(actual_result, expected_result) def test_get_y_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 6 actual_result = board_location.get_y_board() self.assertEqual(actual_result, expected_result) def test_get_x_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 3 actual_result = board_location.get_x_array() self.assertEqual(actual_result, expected_result) def test_get_y_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 5 actual_result = board_location.get_y_array() self.assertEqual(actual_result, expected_result) def test_coordinate_bad_x(self): self.assertRaises(TypeError, coordinate.Coordinate, "4", 6) def test_coordinate_bad_y(self): self.assertRaises(TypeError, coordinate.Coordinate, 4, "6") def test_coordinate_bad_location(self): self.assertRaises(ValueError, coordinate.Coordinate, 50, 100)
Add unit tests for fail fast logic in convertCharToInt()
Add unit tests for fail fast logic in convertCharToInt()
Python
mit
blairck/jaeger
import unittest from src import coordinate class TestRules(unittest.TestCase): """ Tests for the coordinate module """ def test_get_x_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 4 actual_result = board_location.get_x_board() self.assertEqual(actual_result, expected_result) def test_get_y_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 6 actual_result = board_location.get_y_board() self.assertEqual(actual_result, expected_result) def test_get_x_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 3 actual_result = board_location.get_x_array() self.assertEqual(actual_result, expected_result) def test_get_y_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 5 actual_result = board_location.get_y_array() self.assertEqual(actual_result, expected_result) + + def test_coordinate_bad_x(self): + self.assertRaises(TypeError, coordinate.Coordinate, "4", 6) + + def test_coordinate_bad_y(self): + self.assertRaises(TypeError, coordinate.Coordinate, 4, "6") + + def test_coordinate_bad_location(self): + self.assertRaises(ValueError, coordinate.Coordinate, 50, 100) +
Add unit tests for fail fast logic in convertCharToInt()
## Code Before: import unittest from src import coordinate class TestRules(unittest.TestCase): """ Tests for the coordinate module """ def test_get_x_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 4 actual_result = board_location.get_x_board() self.assertEqual(actual_result, expected_result) def test_get_y_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 6 actual_result = board_location.get_y_board() self.assertEqual(actual_result, expected_result) def test_get_x_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 3 actual_result = board_location.get_x_array() self.assertEqual(actual_result, expected_result) def test_get_y_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 5 actual_result = board_location.get_y_array() self.assertEqual(actual_result, expected_result) ## Instruction: Add unit tests for fail fast logic in convertCharToInt() ## Code After: import unittest from src import coordinate class TestRules(unittest.TestCase): """ Tests for the coordinate module """ def test_get_x_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 4 actual_result = board_location.get_x_board() self.assertEqual(actual_result, expected_result) def test_get_y_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 6 actual_result = board_location.get_y_board() self.assertEqual(actual_result, expected_result) def test_get_x_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 3 actual_result = board_location.get_x_array() self.assertEqual(actual_result, expected_result) def test_get_y_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 5 actual_result = board_location.get_y_array() self.assertEqual(actual_result, expected_result) def test_coordinate_bad_x(self): self.assertRaises(TypeError, coordinate.Coordinate, "4", 6) def test_coordinate_bad_y(self): self.assertRaises(TypeError, coordinate.Coordinate, 4, "6") def test_coordinate_bad_location(self): self.assertRaises(ValueError, coordinate.Coordinate, 50, 100)
import unittest from src import coordinate class TestRules(unittest.TestCase): """ Tests for the coordinate module """ def test_get_x_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 4 actual_result = board_location.get_x_board() self.assertEqual(actual_result, expected_result) def test_get_y_board(self): board_location = coordinate.Coordinate(4, 6) expected_result = 6 actual_result = board_location.get_y_board() self.assertEqual(actual_result, expected_result) def test_get_x_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 3 actual_result = board_location.get_x_array() self.assertEqual(actual_result, expected_result) def test_get_y_array(self): board_location = coordinate.Coordinate(4, 6) expected_result = 5 actual_result = board_location.get_y_array() self.assertEqual(actual_result, expected_result) + + def test_coordinate_bad_x(self): + self.assertRaises(TypeError, coordinate.Coordinate, "4", 6) + + def test_coordinate_bad_y(self): + self.assertRaises(TypeError, coordinate.Coordinate, 4, "6") + + def test_coordinate_bad_location(self): + self.assertRaises(ValueError, coordinate.Coordinate, 50, 100)
0afdab2f6feced873c88ba1e73fdde0dad5f041e
skytap/Quotas.py
skytap/Quotas.py
import json import sys from skytap.models.Quota import Quota from skytap.models.SkytapGroup import SkytapGroup class Quotas(SkytapGroup): """Company/account quotas object.""" def __init__(self): """Load the quotas from Skytap.""" super(Quotas, self).__init__() quota_rest = self.rest('/v2/company/quotas') quota_json = json.loads(quota_rest) for qu in quota_json: self.data[qu] = Quota(quota_json[qu][0]) if __name__ == '__main__': print(Quotas().main(sys.argv[1:]))
import json import sys from skytap.models.Quota import Quota from skytap.models.SkytapGroup import SkytapGroup class Quotas(SkytapGroup): """Company/account quotas object. Note: This code assumes that you have regional limits on your account. The return is different if you don't (see the /v2 API doc). We should get each piece of the return and sort it into type-and-region (whether you have regional limits or not) and can then access things uniformly. Doing so will also require smartly accessing the API on demand more, since accounts with regional limits may require multiple calls to get the info desired. """ def __init__(self): """Load the quotas from Skytap.""" super(Quotas, self).__init__() quota_rest = self.rest('/v2/company/quotas') quota_json = json.loads(quota_rest) for qu in quota_json: self.data[qu] = Quota(quota_json[qu][0]) if __name__ == '__main__': print(Quotas().main(sys.argv[1:]))
Comment re: API usage to clarify quotas.
Comment re: API usage to clarify quotas.
Python
mit
mapledyne/skytap,FulcrumIT/skytap
import json import sys from skytap.models.Quota import Quota from skytap.models.SkytapGroup import SkytapGroup class Quotas(SkytapGroup): - """Company/account quotas object.""" + """Company/account quotas object. + + Note: This code assumes that you have regional limits on your account. + The return is different if you don't (see the /v2 API doc). We should get + each piece of the return and sort it into type-and-region (whether you + have regional limits or not) and can then access things uniformly. Doing + so will also require smartly accessing the API on demand more, since + accounts with regional limits may require multiple calls to get the info + desired. + """ def __init__(self): """Load the quotas from Skytap.""" super(Quotas, self).__init__() quota_rest = self.rest('/v2/company/quotas') quota_json = json.loads(quota_rest) for qu in quota_json: self.data[qu] = Quota(quota_json[qu][0]) if __name__ == '__main__': print(Quotas().main(sys.argv[1:]))
Comment re: API usage to clarify quotas.
## Code Before: import json import sys from skytap.models.Quota import Quota from skytap.models.SkytapGroup import SkytapGroup class Quotas(SkytapGroup): """Company/account quotas object.""" def __init__(self): """Load the quotas from Skytap.""" super(Quotas, self).__init__() quota_rest = self.rest('/v2/company/quotas') quota_json = json.loads(quota_rest) for qu in quota_json: self.data[qu] = Quota(quota_json[qu][0]) if __name__ == '__main__': print(Quotas().main(sys.argv[1:])) ## Instruction: Comment re: API usage to clarify quotas. ## Code After: import json import sys from skytap.models.Quota import Quota from skytap.models.SkytapGroup import SkytapGroup class Quotas(SkytapGroup): """Company/account quotas object. Note: This code assumes that you have regional limits on your account. The return is different if you don't (see the /v2 API doc). We should get each piece of the return and sort it into type-and-region (whether you have regional limits or not) and can then access things uniformly. Doing so will also require smartly accessing the API on demand more, since accounts with regional limits may require multiple calls to get the info desired. """ def __init__(self): """Load the quotas from Skytap.""" super(Quotas, self).__init__() quota_rest = self.rest('/v2/company/quotas') quota_json = json.loads(quota_rest) for qu in quota_json: self.data[qu] = Quota(quota_json[qu][0]) if __name__ == '__main__': print(Quotas().main(sys.argv[1:]))
import json import sys from skytap.models.Quota import Quota from skytap.models.SkytapGroup import SkytapGroup class Quotas(SkytapGroup): - """Company/account quotas object.""" ? --- + """Company/account quotas object. + + Note: This code assumes that you have regional limits on your account. + The return is different if you don't (see the /v2 API doc). We should get + each piece of the return and sort it into type-and-region (whether you + have regional limits or not) and can then access things uniformly. Doing + so will also require smartly accessing the API on demand more, since + accounts with regional limits may require multiple calls to get the info + desired. + """ def __init__(self): """Load the quotas from Skytap.""" super(Quotas, self).__init__() quota_rest = self.rest('/v2/company/quotas') quota_json = json.loads(quota_rest) for qu in quota_json: self.data[qu] = Quota(quota_json[qu][0]) if __name__ == '__main__': print(Quotas().main(sys.argv[1:]))
038978f87883247a14e9bec08708452c98c91285
test/test_chimera.py
test/test_chimera.py
import unittest import utils import os import sys import re import shutil import subprocess TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) utils.set_search_paths(TOPDIR) import cryptosite.chimera class Tests(unittest.TestCase): def test_bad(self): """Test wrong arguments to chimera""" for args in ([], ['x'] * 4): out = utils.check_output(['cryptosite', 'chimera'] + args, stderr=subprocess.STDOUT, retcode=2) out = utils.check_output(['python', '-m', 'cryptosite.chimera'] + args, stderr=subprocess.STDOUT, retcode=2) def test_make_chimera_file(self): """Test make_chimera_file() function""" cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax') os.unlink('out.chimerax') if __name__ == '__main__': unittest.main()
import unittest import utils import os import sys import re import shutil import subprocess TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) utils.set_search_paths(TOPDIR) import cryptosite.chimera class Tests(unittest.TestCase): def test_bad(self): """Test wrong arguments to chimera""" for args in ([], ['x'] * 4): out = utils.check_output(['cryptosite', 'chimera'] + args, stderr=subprocess.STDOUT, retcode=2) out = utils.check_output(['python', '-m', 'cryptosite.chimera'] + args, stderr=subprocess.STDOUT, retcode=2) def test_make_chimera_file(self): """Test make_chimera_file() function""" cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax') with open('out.chimerax') as fh: lines = fh.readlines() self.assertEqual(lines[-4], 'open_files("url1", "url2")\n') os.unlink('out.chimerax') if __name__ == '__main__': unittest.main()
Check generated file for sanity.
Check generated file for sanity.
Python
lgpl-2.1
salilab/cryptosite,salilab/cryptosite,salilab/cryptosite
import unittest import utils import os import sys import re import shutil import subprocess TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) utils.set_search_paths(TOPDIR) import cryptosite.chimera class Tests(unittest.TestCase): def test_bad(self): """Test wrong arguments to chimera""" for args in ([], ['x'] * 4): out = utils.check_output(['cryptosite', 'chimera'] + args, stderr=subprocess.STDOUT, retcode=2) out = utils.check_output(['python', '-m', 'cryptosite.chimera'] + args, stderr=subprocess.STDOUT, retcode=2) def test_make_chimera_file(self): """Test make_chimera_file() function""" cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax') + with open('out.chimerax') as fh: + lines = fh.readlines() + self.assertEqual(lines[-4], 'open_files("url1", "url2")\n') os.unlink('out.chimerax') if __name__ == '__main__': unittest.main()
Check generated file for sanity.
## Code Before: import unittest import utils import os import sys import re import shutil import subprocess TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) utils.set_search_paths(TOPDIR) import cryptosite.chimera class Tests(unittest.TestCase): def test_bad(self): """Test wrong arguments to chimera""" for args in ([], ['x'] * 4): out = utils.check_output(['cryptosite', 'chimera'] + args, stderr=subprocess.STDOUT, retcode=2) out = utils.check_output(['python', '-m', 'cryptosite.chimera'] + args, stderr=subprocess.STDOUT, retcode=2) def test_make_chimera_file(self): """Test make_chimera_file() function""" cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax') os.unlink('out.chimerax') if __name__ == '__main__': unittest.main() ## Instruction: Check generated file for sanity. ## Code After: import unittest import utils import os import sys import re import shutil import subprocess TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) utils.set_search_paths(TOPDIR) import cryptosite.chimera class Tests(unittest.TestCase): def test_bad(self): """Test wrong arguments to chimera""" for args in ([], ['x'] * 4): out = utils.check_output(['cryptosite', 'chimera'] + args, stderr=subprocess.STDOUT, retcode=2) out = utils.check_output(['python', '-m', 'cryptosite.chimera'] + args, stderr=subprocess.STDOUT, retcode=2) def test_make_chimera_file(self): """Test make_chimera_file() function""" cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax') with open('out.chimerax') as fh: lines = fh.readlines() self.assertEqual(lines[-4], 'open_files("url1", "url2")\n') os.unlink('out.chimerax') if __name__ == '__main__': unittest.main()
import unittest import utils import os import sys import re import shutil import subprocess TOPDIR = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) utils.set_search_paths(TOPDIR) import cryptosite.chimera class Tests(unittest.TestCase): def test_bad(self): """Test wrong arguments to chimera""" for args in ([], ['x'] * 4): out = utils.check_output(['cryptosite', 'chimera'] + args, stderr=subprocess.STDOUT, retcode=2) out = utils.check_output(['python', '-m', 'cryptosite.chimera'] + args, stderr=subprocess.STDOUT, retcode=2) def test_make_chimera_file(self): """Test make_chimera_file() function""" cryptosite.chimera.make_chimera_file('url1', 'url2', 'out.chimerax') + with open('out.chimerax') as fh: + lines = fh.readlines() + self.assertEqual(lines[-4], 'open_files("url1", "url2")\n') os.unlink('out.chimerax') if __name__ == '__main__': unittest.main()
42f0c76664337af80d692fe7649f3643c237cc47
Tests/MathFunctionsTest.py
Tests/MathFunctionsTest.py
from Math.MathFunctions import * def pointTest(): point1 = (0, 0) point2 = (2, 4) print("Point 1: {}".format(point1)) print("Point 2: {}".format(point2)) print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1]))) angle = pointAngle(point1[0],point1[1],point2[0],point2[1]); print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG)) pointTest()
from Math.MathFunctions import * import unittest class TestPointMethods(unittest.TestCase): def test_point(self): point1 = (0, 0) point2 = (2, 4) angle = pointAngle(point1[0], point1[1], point2[0], point2[1]) dist = pointDistance(point1[0], point1[1], point2[0], point2[1]) self.assertAlmostEqual(angle, 1.1071487177940904) self.assertAlmostEqual(dist, 4.47213595499958) class TestHelperMethods(unittest.TestCase): def test_clamp(self): self.assertEqual(clamp(10, 1, 5), 5) self.assertEqual(clamp(0, 1, 5), 1) self.assertEqual(clamp(3, 1, 5), 3) self.assertEqual(clamp(5, 1, 5), 5) if __name__ == '__main__': unittest.main()
Use python's unit testing framework
Use python's unit testing framework
Python
mit
turtles/PythonScripts
from Math.MathFunctions import * + import unittest - def pointTest(): + class TestPointMethods(unittest.TestCase): + def test_point(self): - point1 = (0, 0) + point1 = (0, 0) - point2 = (2, 4) + point2 = (2, 4) - print("Point 1: {}".format(point1)) - print("Point 2: {}".format(point2)) - print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1]))) - angle = pointAngle(point1[0],point1[1],point2[0],point2[1]); + angle = pointAngle(point1[0], point1[1], point2[0], point2[1]) - print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG)) + dist = pointDistance(point1[0], point1[1], point2[0], point2[1]) - pointTest() + self.assertAlmostEqual(angle, 1.1071487177940904) + self.assertAlmostEqual(dist, 4.47213595499958) + class TestHelperMethods(unittest.TestCase): + def test_clamp(self): + self.assertEqual(clamp(10, 1, 5), 5) + self.assertEqual(clamp(0, 1, 5), 1) + self.assertEqual(clamp(3, 1, 5), 3) + self.assertEqual(clamp(5, 1, 5), 5) + + if __name__ == '__main__': + unittest.main() +
Use python's unit testing framework
## Code Before: from Math.MathFunctions import * def pointTest(): point1 = (0, 0) point2 = (2, 4) print("Point 1: {}".format(point1)) print("Point 2: {}".format(point2)) print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1]))) angle = pointAngle(point1[0],point1[1],point2[0],point2[1]); print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG)) pointTest() ## Instruction: Use python's unit testing framework ## Code After: from Math.MathFunctions import * import unittest class TestPointMethods(unittest.TestCase): def test_point(self): point1 = (0, 0) point2 = (2, 4) angle = pointAngle(point1[0], point1[1], point2[0], point2[1]) dist = pointDistance(point1[0], point1[1], point2[0], point2[1]) self.assertAlmostEqual(angle, 1.1071487177940904) self.assertAlmostEqual(dist, 4.47213595499958) class TestHelperMethods(unittest.TestCase): def test_clamp(self): self.assertEqual(clamp(10, 1, 5), 5) self.assertEqual(clamp(0, 1, 5), 1) self.assertEqual(clamp(3, 1, 5), 3) self.assertEqual(clamp(5, 1, 5), 5) if __name__ == '__main__': unittest.main()
from Math.MathFunctions import * + import unittest - def pointTest(): + class TestPointMethods(unittest.TestCase): + def test_point(self): - point1 = (0, 0) + point1 = (0, 0) ? ++++ - point2 = (2, 4) + point2 = (2, 4) ? ++++ - print("Point 1: {}".format(point1)) - print("Point 2: {}".format(point2)) - print("Point distance: {}".format(pointDistance(point1[0],point1[1],point2[0],point2[1]))) - angle = pointAngle(point1[0],point1[1],point2[0],point2[1]); ? - + angle = pointAngle(point1[0], point1[1], point2[0], point2[1]) ? ++++ + + + - print("Point angle: {:.3f}, {:.3f} degrees".format(angle, angle*RAD_TO_DEG)) + dist = pointDistance(point1[0], point1[1], point2[0], point2[1]) - pointTest() + self.assertAlmostEqual(angle, 1.1071487177940904) + self.assertAlmostEqual(dist, 4.47213595499958) + + class TestHelperMethods(unittest.TestCase): + def test_clamp(self): + self.assertEqual(clamp(10, 1, 5), 5) + self.assertEqual(clamp(0, 1, 5), 1) + self.assertEqual(clamp(3, 1, 5), 3) + self.assertEqual(clamp(5, 1, 5), 5) + + if __name__ == '__main__': + unittest.main()
45381a1ce6e271cc06ce130cb35a93f14eceba90
troposphere/utils.py
troposphere/utils.py
import time def _tail_print(e): print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id)) def get_events(conn, stackname): """Get the events in batches and return in chronological order""" next = None event_list = [] while 1: events = conn.describe_stack_events(stackname, next) event_list.append(events) if events.next_token is None: break next = events.next_token time.sleep(1) return reversed(sum(event_list, [])) def tail(conn, stack_name, log_func=_tail_print, sleep_time=5): """Show and then tail the event log""" # First dump the full list of events in chronological order and keep # track of the events we've seen already seen = set() initial_events = get_events(conn, stack_name) for e in initial_events: log_func(e) seen.add(e.event_id) # Now keep looping through and dump the new events while 1: events = get_events(conn, stack_name) for e in events: if e.event_id not in seen: log_func(e) seen.add(e.event_id) time.sleep(sleep_time)
import time def _tail_print(e): print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id)) def get_events(conn, stackname): """Get the events in batches and return in chronological order""" next = None event_list = [] while 1: events = conn.describe_stack_events(stackname, next) event_list.append(events) if events.next_token is None: break next = events.next_token time.sleep(1) return reversed(sum(event_list, [])) def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True): """Show and then tail the event log""" # First dump the full list of events in chronological order and keep # track of the events we've seen already seen = set() initial_events = get_events(conn, stack_name) for e in initial_events: if include_initial: log_func(e) seen.add(e.event_id) # Now keep looping through and dump the new events while 1: events = get_events(conn, stack_name) for e in events: if e.event_id not in seen: log_func(e) seen.add(e.event_id) time.sleep(sleep_time)
Add "include_initial" kwarg to support tailing stack updates
Add "include_initial" kwarg to support tailing stack updates `get_events` will return all events that have occurred for a stack. This is useless if we're tailing an update to a stack.
Python
bsd-2-clause
mhahn/troposphere
import time def _tail_print(e): print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id)) def get_events(conn, stackname): """Get the events in batches and return in chronological order""" next = None event_list = [] while 1: events = conn.describe_stack_events(stackname, next) event_list.append(events) if events.next_token is None: break next = events.next_token time.sleep(1) return reversed(sum(event_list, [])) - def tail(conn, stack_name, log_func=_tail_print, sleep_time=5): + def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True): """Show and then tail the event log""" # First dump the full list of events in chronological order and keep # track of the events we've seen already seen = set() initial_events = get_events(conn, stack_name) for e in initial_events: + if include_initial: - log_func(e) + log_func(e) seen.add(e.event_id) # Now keep looping through and dump the new events while 1: events = get_events(conn, stack_name) for e in events: if e.event_id not in seen: log_func(e) seen.add(e.event_id) time.sleep(sleep_time)
Add "include_initial" kwarg to support tailing stack updates
## Code Before: import time def _tail_print(e): print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id)) def get_events(conn, stackname): """Get the events in batches and return in chronological order""" next = None event_list = [] while 1: events = conn.describe_stack_events(stackname, next) event_list.append(events) if events.next_token is None: break next = events.next_token time.sleep(1) return reversed(sum(event_list, [])) def tail(conn, stack_name, log_func=_tail_print, sleep_time=5): """Show and then tail the event log""" # First dump the full list of events in chronological order and keep # track of the events we've seen already seen = set() initial_events = get_events(conn, stack_name) for e in initial_events: log_func(e) seen.add(e.event_id) # Now keep looping through and dump the new events while 1: events = get_events(conn, stack_name) for e in events: if e.event_id not in seen: log_func(e) seen.add(e.event_id) time.sleep(sleep_time) ## Instruction: Add "include_initial" kwarg to support tailing stack updates ## Code After: import time def _tail_print(e): print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id)) def get_events(conn, stackname): """Get the events in batches and return in chronological order""" next = None event_list = [] while 1: events = conn.describe_stack_events(stackname, next) event_list.append(events) if events.next_token is None: break next = events.next_token time.sleep(1) return reversed(sum(event_list, [])) def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True): """Show and then tail the event log""" # First dump the full list of events in chronological order and keep # track of the events we've seen already seen = set() initial_events = get_events(conn, stack_name) for e in initial_events: if include_initial: log_func(e) seen.add(e.event_id) # Now keep looping through and dump the new events while 1: events = get_events(conn, stack_name) for e in events: if e.event_id not in seen: log_func(e) seen.add(e.event_id) time.sleep(sleep_time)
import time def _tail_print(e): print("%s %s %s" % (e.resource_status, e.resource_type, e.event_id)) def get_events(conn, stackname): """Get the events in batches and return in chronological order""" next = None event_list = [] while 1: events = conn.describe_stack_events(stackname, next) event_list.append(events) if events.next_token is None: break next = events.next_token time.sleep(1) return reversed(sum(event_list, [])) - def tail(conn, stack_name, log_func=_tail_print, sleep_time=5): + def tail(conn, stack_name, log_func=_tail_print, sleep_time=5, include_initial=True): ? ++++++++++++++++++++++ """Show and then tail the event log""" # First dump the full list of events in chronological order and keep # track of the events we've seen already seen = set() initial_events = get_events(conn, stack_name) for e in initial_events: + if include_initial: - log_func(e) + log_func(e) ? ++++ seen.add(e.event_id) # Now keep looping through and dump the new events while 1: events = get_events(conn, stack_name) for e in events: if e.event_id not in seen: log_func(e) seen.add(e.event_id) time.sleep(sleep_time)
5a76457e2b9596ad3497b0145410a2f4090a5c54
tests/mixins.py
tests/mixins.py
class RedisCleanupMixin(object): client = None prefix = None def setUp(self): super(RedisCleanupMixin, self).setUp() self.assertIsNotNone(self.client, "Need a redis client to be provided") def tearDown(self): root = '*' if self.prefix is not None: root = '{}*'.format(self.prefix) keys = self.client.keys(root) for key in keys: self.client.delete(key) super(RedisCleanupMixin, self).tearDown()
class RedisCleanupMixin(object): client = None prefix = NotImplemented # type: str def setUp(self): super(RedisCleanupMixin, self).setUp() self.assertIsNotNone(self.client, "Need a redis client to be provided") def tearDown(self): root = '*' if self.prefix is not None: root = '{}*'.format(self.prefix) keys = self.client.keys(root) for key in keys: self.client.delete(key) super(RedisCleanupMixin, self).tearDown()
Add annotation required for mypy
Add annotation required for mypy
Python
bsd-3-clause
thread/django-lightweight-queue,thread/django-lightweight-queue
class RedisCleanupMixin(object): client = None - prefix = None + prefix = NotImplemented # type: str def setUp(self): super(RedisCleanupMixin, self).setUp() self.assertIsNotNone(self.client, "Need a redis client to be provided") def tearDown(self): root = '*' if self.prefix is not None: root = '{}*'.format(self.prefix) keys = self.client.keys(root) for key in keys: self.client.delete(key) super(RedisCleanupMixin, self).tearDown()
Add annotation required for mypy
## Code Before: class RedisCleanupMixin(object): client = None prefix = None def setUp(self): super(RedisCleanupMixin, self).setUp() self.assertIsNotNone(self.client, "Need a redis client to be provided") def tearDown(self): root = '*' if self.prefix is not None: root = '{}*'.format(self.prefix) keys = self.client.keys(root) for key in keys: self.client.delete(key) super(RedisCleanupMixin, self).tearDown() ## Instruction: Add annotation required for mypy ## Code After: class RedisCleanupMixin(object): client = None prefix = NotImplemented # type: str def setUp(self): super(RedisCleanupMixin, self).setUp() self.assertIsNotNone(self.client, "Need a redis client to be provided") def tearDown(self): root = '*' if self.prefix is not None: root = '{}*'.format(self.prefix) keys = self.client.keys(root) for key in keys: self.client.delete(key) super(RedisCleanupMixin, self).tearDown()
class RedisCleanupMixin(object): client = None - prefix = None + prefix = NotImplemented # type: str def setUp(self): super(RedisCleanupMixin, self).setUp() self.assertIsNotNone(self.client, "Need a redis client to be provided") def tearDown(self): root = '*' if self.prefix is not None: root = '{}*'.format(self.prefix) keys = self.client.keys(root) for key in keys: self.client.delete(key) super(RedisCleanupMixin, self).tearDown()
28c88cbc34dcf2af5c98ce3f3eed3774dd5be15e
lcapy/discretetime.py
lcapy/discretetime.py
import sympy as sym from .sym import sympify from .nexpr import nexpr, n from .kexpr import kexpr, k from .zexpr import zexpr, z from .dsym import nsym, ksym, zsym, dt, df from .expr import expr as expr1 from .transform import transform as transform1 from .transform import call as call1 from .functions import Function from .ztransform import * from .seq import seq def expr(arg, **assumptions): expr = sympify(arg, **assumptions) symbols = expr.free_symbols if nsym in symbols: return nexpr(expr, **assumptions) elif ksym in symbols: return kexpr(expr, **assumptions) elif zsym in symbols: return zexpr(expr, **assumptions) return expr1(arg, **assumptions)
import sympy as sym from .sym import sympify from .nexpr import nexpr, n from .kexpr import kexpr, k from .zexpr import zexpr, z from .dsym import nsym, ksym, zsym, dt, df from .expr import expr as expr1 from .transform import transform as transform1 from .transform import call as call1 from .functions import Function from .ztransform import * from .seq import seq def expr(arg, **assumptions): # Handle container args. if not isinstance(arg, str) and hasattr(arg, '__iter__'): return expr1(arg, **assumptions) expr = sympify(arg, **assumptions) symbols = expr.free_symbols if nsym in symbols: return nexpr(expr, **assumptions) elif ksym in symbols: return kexpr(expr, **assumptions) elif zsym in symbols: return zexpr(expr, **assumptions) return expr1(arg, **assumptions)
Handle container types for discrete-time expr
Handle container types for discrete-time expr
Python
lgpl-2.1
mph-/lcapy
import sympy as sym from .sym import sympify from .nexpr import nexpr, n from .kexpr import kexpr, k from .zexpr import zexpr, z from .dsym import nsym, ksym, zsym, dt, df from .expr import expr as expr1 from .transform import transform as transform1 from .transform import call as call1 from .functions import Function from .ztransform import * from .seq import seq def expr(arg, **assumptions): + # Handle container args. + if not isinstance(arg, str) and hasattr(arg, '__iter__'): + return expr1(arg, **assumptions) + expr = sympify(arg, **assumptions) symbols = expr.free_symbols if nsym in symbols: return nexpr(expr, **assumptions) elif ksym in symbols: return kexpr(expr, **assumptions) elif zsym in symbols: return zexpr(expr, **assumptions) return expr1(arg, **assumptions)
Handle container types for discrete-time expr
## Code Before: import sympy as sym from .sym import sympify from .nexpr import nexpr, n from .kexpr import kexpr, k from .zexpr import zexpr, z from .dsym import nsym, ksym, zsym, dt, df from .expr import expr as expr1 from .transform import transform as transform1 from .transform import call as call1 from .functions import Function from .ztransform import * from .seq import seq def expr(arg, **assumptions): expr = sympify(arg, **assumptions) symbols = expr.free_symbols if nsym in symbols: return nexpr(expr, **assumptions) elif ksym in symbols: return kexpr(expr, **assumptions) elif zsym in symbols: return zexpr(expr, **assumptions) return expr1(arg, **assumptions) ## Instruction: Handle container types for discrete-time expr ## Code After: import sympy as sym from .sym import sympify from .nexpr import nexpr, n from .kexpr import kexpr, k from .zexpr import zexpr, z from .dsym import nsym, ksym, zsym, dt, df from .expr import expr as expr1 from .transform import transform as transform1 from .transform import call as call1 from .functions import Function from .ztransform import * from .seq import seq def expr(arg, **assumptions): # Handle container args. if not isinstance(arg, str) and hasattr(arg, '__iter__'): return expr1(arg, **assumptions) expr = sympify(arg, **assumptions) symbols = expr.free_symbols if nsym in symbols: return nexpr(expr, **assumptions) elif ksym in symbols: return kexpr(expr, **assumptions) elif zsym in symbols: return zexpr(expr, **assumptions) return expr1(arg, **assumptions)
import sympy as sym from .sym import sympify from .nexpr import nexpr, n from .kexpr import kexpr, k from .zexpr import zexpr, z from .dsym import nsym, ksym, zsym, dt, df from .expr import expr as expr1 from .transform import transform as transform1 from .transform import call as call1 from .functions import Function from .ztransform import * from .seq import seq def expr(arg, **assumptions): + # Handle container args. + if not isinstance(arg, str) and hasattr(arg, '__iter__'): + return expr1(arg, **assumptions) + expr = sympify(arg, **assumptions) symbols = expr.free_symbols if nsym in symbols: return nexpr(expr, **assumptions) elif ksym in symbols: return kexpr(expr, **assumptions) elif zsym in symbols: return zexpr(expr, **assumptions) return expr1(arg, **assumptions)
521ebf29990de4d997c90f4168ea300d75776cfc
components/utilities.py
components/utilities.py
"""Utilities for general operations.""" def IsNumeric(num_str): try: val = int(num_str) except ValueError: return False else: return True
"""Utilities for general operations.""" def IsNumeric(num_str): try: val = int(num_str) except ValueError: return False else: return True def GuaranteeUnicode(obj): if type(obj) == unicode: return obj elif type(obj) == str: return unicode(obj, "utf-8") else: return unicode(str(obj), "utf-8")
Add GuranteeUnicode function which always returns a unicode object
Add GuranteeUnicode function which always returns a unicode object
Python
mit
lnishan/SQLGitHub
"""Utilities for general operations.""" def IsNumeric(num_str): try: val = int(num_str) except ValueError: return False else: return True + def GuaranteeUnicode(obj): + if type(obj) == unicode: + return obj + elif type(obj) == str: + return unicode(obj, "utf-8") + else: + return unicode(str(obj), "utf-8") +
Add GuranteeUnicode function which always returns a unicode object
## Code Before: """Utilities for general operations.""" def IsNumeric(num_str): try: val = int(num_str) except ValueError: return False else: return True ## Instruction: Add GuranteeUnicode function which always returns a unicode object ## Code After: """Utilities for general operations.""" def IsNumeric(num_str): try: val = int(num_str) except ValueError: return False else: return True def GuaranteeUnicode(obj): if type(obj) == unicode: return obj elif type(obj) == str: return unicode(obj, "utf-8") else: return unicode(str(obj), "utf-8")
"""Utilities for general operations.""" def IsNumeric(num_str): try: val = int(num_str) except ValueError: return False else: return True + + def GuaranteeUnicode(obj): + if type(obj) == unicode: + return obj + elif type(obj) == str: + return unicode(obj, "utf-8") + else: + return unicode(str(obj), "utf-8")
ff02f0e8a7b62d5afdc88730129a5e0811fb5a82
monitor/monitor_test.py
monitor/monitor_test.py
import unittest import monitor class TestMonitor(unittest.TestCase): def test_attr(self): mon = monitor.Monitor() assert mon.pvprefix == "MON-CONTROL:" assert mon.monitorname == "PI1" def test_testimage(self): mon = monitor.Monitor() image = mon.testimage() assert len(image) != 0 # print image def test_readPV(self): mon = monitor.Monitor() pv_result = mon.readPV() assert pv_result == "CAM1" if __name__ == '__main__': unittest.main()
import unittest from mock import patch, MagicMock, PropertyMock from monitor import Monitor class MonitorUpdateTest(unittest.TestCase): def setUp(self): with patch('monitor.pv.PV'): mock_plotter = MagicMock() self.monitor = Monitor("MYNAME", mock_plotter) def test_does_nothing_if_camera_is_None(self): self.monitor.camera = None try: self.monitor._update_image() self.assertTrue(True) except Exception as ex: self.fail("Unexpected expection thrown" + str(ex)) def test_gets_image_data_from_camera_when_not_None(self): mock_camera = MagicMock() self.monitor.camera = mock_camera self.monitor._update_image() mock_camera.get_image_data.assert_called_once_with() def test_get_size_data_from_camera_when_not_None(self): mock_camera = MagicMock() mock_xsize = PropertyMock(return_value=100) mock_ysize = PropertyMock(return_value=200) type(mock_camera).xsize = mock_xsize type(mock_camera).ysize = mock_ysize self.monitor.camera = mock_camera self.monitor._update_image() mock_xsize.assert_called_once_with() mock_ysize.assert_called_once_with() def test_calls_plotter_with_image_and_size_data(self): data = 111 xsize = 100 ysize = 200 mock_camera = MagicMock(xsize=xsize, ysize=ysize) mock_camera.get_image_data = MagicMock(return_value=data) self.monitor.camera = mock_camera self.monitor._update_image() self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize) if __name__ == '__main__': unittest.main()
Rewrite monitortests. Add tests for update_image
Rewrite monitortests. Add tests for update_image
Python
apache-2.0
nickbattam/picamon,nickbattam/picamon,nickbattam/picamon,nickbattam/picamon
import unittest - import monitor + from mock import patch, MagicMock, PropertyMock - class TestMonitor(unittest.TestCase): + from monitor import Monitor - def test_attr(self): - mon = monitor.Monitor() - assert mon.pvprefix == "MON-CONTROL:" - assert mon.monitorname == "PI1" + class MonitorUpdateTest(unittest.TestCase): - def test_testimage(self): - mon = monitor.Monitor() - image = mon.testimage() - assert len(image) != 0 - # print image - def test_readPV(self): - mon = monitor.Monitor() - pv_result = mon.readPV() - assert pv_result == "CAM1" + def setUp(self): + with patch('monitor.pv.PV'): + mock_plotter = MagicMock() + self.monitor = Monitor("MYNAME", mock_plotter) + + def test_does_nothing_if_camera_is_None(self): + self.monitor.camera = None + + try: + self.monitor._update_image() + self.assertTrue(True) + except Exception as ex: + self.fail("Unexpected expection thrown" + str(ex)) + + + def test_gets_image_data_from_camera_when_not_None(self): + mock_camera = MagicMock() + self.monitor.camera = mock_camera + + self.monitor._update_image() + + mock_camera.get_image_data.assert_called_once_with() + + def test_get_size_data_from_camera_when_not_None(self): + mock_camera = MagicMock() + + mock_xsize = PropertyMock(return_value=100) + mock_ysize = PropertyMock(return_value=200) + type(mock_camera).xsize = mock_xsize + type(mock_camera).ysize = mock_ysize + + self.monitor.camera = mock_camera + + self.monitor._update_image() + + mock_xsize.assert_called_once_with() + mock_ysize.assert_called_once_with() + + def test_calls_plotter_with_image_and_size_data(self): + data = 111 + xsize = 100 + ysize = 200 + mock_camera = MagicMock(xsize=xsize, ysize=ysize) + mock_camera.get_image_data = MagicMock(return_value=data) + self.monitor.camera = mock_camera + + self.monitor._update_image() + + self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize) if __name__ == '__main__': - unittest.main() + unittest.main()
Rewrite monitortests. Add tests for update_image
## Code Before: import unittest import monitor class TestMonitor(unittest.TestCase): def test_attr(self): mon = monitor.Monitor() assert mon.pvprefix == "MON-CONTROL:" assert mon.monitorname == "PI1" def test_testimage(self): mon = monitor.Monitor() image = mon.testimage() assert len(image) != 0 # print image def test_readPV(self): mon = monitor.Monitor() pv_result = mon.readPV() assert pv_result == "CAM1" if __name__ == '__main__': unittest.main() ## Instruction: Rewrite monitortests. Add tests for update_image ## Code After: import unittest from mock import patch, MagicMock, PropertyMock from monitor import Monitor class MonitorUpdateTest(unittest.TestCase): def setUp(self): with patch('monitor.pv.PV'): mock_plotter = MagicMock() self.monitor = Monitor("MYNAME", mock_plotter) def test_does_nothing_if_camera_is_None(self): self.monitor.camera = None try: self.monitor._update_image() self.assertTrue(True) except Exception as ex: self.fail("Unexpected expection thrown" + str(ex)) def test_gets_image_data_from_camera_when_not_None(self): mock_camera = MagicMock() self.monitor.camera = mock_camera self.monitor._update_image() mock_camera.get_image_data.assert_called_once_with() def test_get_size_data_from_camera_when_not_None(self): mock_camera = MagicMock() mock_xsize = PropertyMock(return_value=100) mock_ysize = PropertyMock(return_value=200) type(mock_camera).xsize = mock_xsize type(mock_camera).ysize = mock_ysize self.monitor.camera = mock_camera self.monitor._update_image() mock_xsize.assert_called_once_with() mock_ysize.assert_called_once_with() def test_calls_plotter_with_image_and_size_data(self): data = 111 xsize = 100 ysize = 200 mock_camera = MagicMock(xsize=xsize, ysize=ysize) mock_camera.get_image_data = MagicMock(return_value=data) self.monitor.camera = mock_camera self.monitor._update_image() self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize) if __name__ == '__main__': unittest.main()
import unittest - import monitor + from mock import patch, MagicMock, PropertyMock - class TestMonitor(unittest.TestCase): + from monitor import Monitor - def test_attr(self): - mon = monitor.Monitor() - assert mon.pvprefix == "MON-CONTROL:" - assert mon.monitorname == "PI1" + class MonitorUpdateTest(unittest.TestCase): - def test_testimage(self): - mon = monitor.Monitor() - image = mon.testimage() - assert len(image) != 0 - # print image - def test_readPV(self): - mon = monitor.Monitor() - pv_result = mon.readPV() - assert pv_result == "CAM1" + def setUp(self): + with patch('monitor.pv.PV'): + mock_plotter = MagicMock() + self.monitor = Monitor("MYNAME", mock_plotter) + + def test_does_nothing_if_camera_is_None(self): + self.monitor.camera = None + + try: + self.monitor._update_image() + self.assertTrue(True) + except Exception as ex: + self.fail("Unexpected expection thrown" + str(ex)) + + + def test_gets_image_data_from_camera_when_not_None(self): + mock_camera = MagicMock() + self.monitor.camera = mock_camera + + self.monitor._update_image() + + mock_camera.get_image_data.assert_called_once_with() + + def test_get_size_data_from_camera_when_not_None(self): + mock_camera = MagicMock() + + mock_xsize = PropertyMock(return_value=100) + mock_ysize = PropertyMock(return_value=200) + type(mock_camera).xsize = mock_xsize + type(mock_camera).ysize = mock_ysize + + self.monitor.camera = mock_camera + + self.monitor._update_image() + + mock_xsize.assert_called_once_with() + mock_ysize.assert_called_once_with() + + def test_calls_plotter_with_image_and_size_data(self): + data = 111 + xsize = 100 + ysize = 200 + mock_camera = MagicMock(xsize=xsize, ysize=ysize) + mock_camera.get_image_data = MagicMock(return_value=data) + self.monitor.camera = mock_camera + + self.monitor._update_image() + + self.monitor.plotter.show.assert_called_once_with(data, xsize, ysize) if __name__ == '__main__': - unittest.main() ? ^ + unittest.main() ? ^^^^
e9c4881ee29ba104caf9fc8330583c254fe52c06
scripts/examples/Arduino/Portenta-H7/19-Low-Power/deep_sleep.py
scripts/examples/Arduino/Portenta-H7/19-Low-Power/deep_sleep.py
import pyb, machine, sensor # Create and init RTC object. rtc = pyb.RTC() # (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]]) rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0)) # Print RTC info. print(rtc.datetime()) sensor.reset() # Enable sensor softsleep sensor.sleep(True) # Optionally bypass the regulator on OV7725 # for the lowest possible power consumption. if (sensor.get_id() == sensor.OV7725): # Bypass internal regulator sensor.__write_reg(0x4F, 0x18) # Shutdown the sensor (pulls PWDN high). sensor.shutdown(True) # Enable RTC interrupts every 30 seconds. # Note the camera will RESET after wakeup from Deepsleep Mode. rtc.wakeup(30000) # Enter Deepsleep Mode. machine.deepsleep()
import pyb, machine, sensor # Create and init RTC object. rtc = pyb.RTC() # (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]]) rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0)) # Print RTC info. print(rtc.datetime()) sensor.reset() # Shutdown the sensor (pulls PWDN high). sensor.shutdown(True) # Enable RTC interrupts every 30 seconds. # Note the camera will RESET after wakeup from Deepsleep Mode. rtc.wakeup(30000) # Enter Deepsleep Mode. machine.deepsleep()
Remove sensor setting from deep sleep example
Remove sensor setting from deep sleep example
Python
mit
iabdalkader/openmv,openmv/openmv,kwagyeman/openmv,iabdalkader/openmv,openmv/openmv,kwagyeman/openmv,openmv/openmv,iabdalkader/openmv,kwagyeman/openmv,iabdalkader/openmv,kwagyeman/openmv,openmv/openmv
import pyb, machine, sensor # Create and init RTC object. rtc = pyb.RTC() # (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]]) rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0)) # Print RTC info. print(rtc.datetime()) sensor.reset() - # Enable sensor softsleep - sensor.sleep(True) - - # Optionally bypass the regulator on OV7725 - # for the lowest possible power consumption. - if (sensor.get_id() == sensor.OV7725): - # Bypass internal regulator - sensor.__write_reg(0x4F, 0x18) - # Shutdown the sensor (pulls PWDN high). sensor.shutdown(True) # Enable RTC interrupts every 30 seconds. # Note the camera will RESET after wakeup from Deepsleep Mode. rtc.wakeup(30000) # Enter Deepsleep Mode. machine.deepsleep()
Remove sensor setting from deep sleep example
## Code Before: import pyb, machine, sensor # Create and init RTC object. rtc = pyb.RTC() # (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]]) rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0)) # Print RTC info. print(rtc.datetime()) sensor.reset() # Enable sensor softsleep sensor.sleep(True) # Optionally bypass the regulator on OV7725 # for the lowest possible power consumption. if (sensor.get_id() == sensor.OV7725): # Bypass internal regulator sensor.__write_reg(0x4F, 0x18) # Shutdown the sensor (pulls PWDN high). sensor.shutdown(True) # Enable RTC interrupts every 30 seconds. # Note the camera will RESET after wakeup from Deepsleep Mode. rtc.wakeup(30000) # Enter Deepsleep Mode. machine.deepsleep() ## Instruction: Remove sensor setting from deep sleep example ## Code After: import pyb, machine, sensor # Create and init RTC object. rtc = pyb.RTC() # (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]]) rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0)) # Print RTC info. print(rtc.datetime()) sensor.reset() # Shutdown the sensor (pulls PWDN high). sensor.shutdown(True) # Enable RTC interrupts every 30 seconds. # Note the camera will RESET after wakeup from Deepsleep Mode. rtc.wakeup(30000) # Enter Deepsleep Mode. machine.deepsleep()
import pyb, machine, sensor # Create and init RTC object. rtc = pyb.RTC() # (year, month, day[, hour[, minute[, second[, microsecond[, tzinfo]]]]]) rtc.datetime((2014, 5, 1, 4, 13, 0, 0, 0)) # Print RTC info. print(rtc.datetime()) sensor.reset() - # Enable sensor softsleep - sensor.sleep(True) - - # Optionally bypass the regulator on OV7725 - # for the lowest possible power consumption. - if (sensor.get_id() == sensor.OV7725): - # Bypass internal regulator - sensor.__write_reg(0x4F, 0x18) - # Shutdown the sensor (pulls PWDN high). sensor.shutdown(True) # Enable RTC interrupts every 30 seconds. # Note the camera will RESET after wakeup from Deepsleep Mode. rtc.wakeup(30000) # Enter Deepsleep Mode. machine.deepsleep()
d491aea2da5d52245001f4da24331f33e4a3a299
importlib_metadata/_meta.py
importlib_metadata/_meta.py
from ._compat import Protocol from typing import Any, Dict, Iterator, List, TypeVar, Union _T = TypeVar("_T") class PackageMetadata(Protocol): def __len__(self) -> int: ... # pragma: no cover def __contains__(self, item: str) -> bool: ... # pragma: no cover def __getitem__(self, key: str) -> str: ... # pragma: no cover def __iter__(self) -> Iterator[str]: ... # pragma: no cover def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]: """ Return all values associated with a possibly multi-valued key. """ @property def json(self) -> Dict[str, Union[str, List[str]]]: """ A JSON-compatible form of the metadata. """ class SimplePath(Protocol): """ A minimal subset of pathlib.Path required by PathDistribution. """ def joinpath(self) -> 'SimplePath': ... # pragma: no cover def __div__(self) -> 'SimplePath': ... # pragma: no cover def parent(self) -> 'SimplePath': ... # pragma: no cover def read_text(self) -> str: ... # pragma: no cover
from ._compat import Protocol from typing import Any, Dict, Iterator, List, TypeVar, Union _T = TypeVar("_T") class PackageMetadata(Protocol): def __len__(self) -> int: ... # pragma: no cover def __contains__(self, item: str) -> bool: ... # pragma: no cover def __getitem__(self, key: str) -> str: ... # pragma: no cover def __iter__(self) -> Iterator[str]: ... # pragma: no cover def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]: """ Return all values associated with a possibly multi-valued key. """ @property def json(self) -> Dict[str, Union[str, List[str]]]: """ A JSON-compatible form of the metadata. """ class SimplePath(Protocol): """ A minimal subset of pathlib.Path required by PathDistribution. >>> import pathlib >>> import typing >>> _: SimplePath = typing.cast(pathlib.Path, None) """ def joinpath(self) -> 'SimplePath': ... # pragma: no cover def __div__(self) -> 'SimplePath': ... # pragma: no cover def parent(self) -> 'SimplePath': ... # pragma: no cover def read_text(self) -> str: ... # pragma: no cover
Add test purported to capture the failure, but it still passes.
Add test purported to capture the failure, but it still passes.
Python
apache-2.0
python/importlib_metadata
from ._compat import Protocol from typing import Any, Dict, Iterator, List, TypeVar, Union _T = TypeVar("_T") class PackageMetadata(Protocol): def __len__(self) -> int: ... # pragma: no cover def __contains__(self, item: str) -> bool: ... # pragma: no cover def __getitem__(self, key: str) -> str: ... # pragma: no cover def __iter__(self) -> Iterator[str]: ... # pragma: no cover def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]: """ Return all values associated with a possibly multi-valued key. """ @property def json(self) -> Dict[str, Union[str, List[str]]]: """ A JSON-compatible form of the metadata. """ class SimplePath(Protocol): """ A minimal subset of pathlib.Path required by PathDistribution. + + >>> import pathlib + >>> import typing + >>> _: SimplePath = typing.cast(pathlib.Path, None) """ def joinpath(self) -> 'SimplePath': ... # pragma: no cover def __div__(self) -> 'SimplePath': ... # pragma: no cover def parent(self) -> 'SimplePath': ... # pragma: no cover def read_text(self) -> str: ... # pragma: no cover
Add test purported to capture the failure, but it still passes.
## Code Before: from ._compat import Protocol from typing import Any, Dict, Iterator, List, TypeVar, Union _T = TypeVar("_T") class PackageMetadata(Protocol): def __len__(self) -> int: ... # pragma: no cover def __contains__(self, item: str) -> bool: ... # pragma: no cover def __getitem__(self, key: str) -> str: ... # pragma: no cover def __iter__(self) -> Iterator[str]: ... # pragma: no cover def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]: """ Return all values associated with a possibly multi-valued key. """ @property def json(self) -> Dict[str, Union[str, List[str]]]: """ A JSON-compatible form of the metadata. """ class SimplePath(Protocol): """ A minimal subset of pathlib.Path required by PathDistribution. """ def joinpath(self) -> 'SimplePath': ... # pragma: no cover def __div__(self) -> 'SimplePath': ... # pragma: no cover def parent(self) -> 'SimplePath': ... # pragma: no cover def read_text(self) -> str: ... # pragma: no cover ## Instruction: Add test purported to capture the failure, but it still passes. ## Code After: from ._compat import Protocol from typing import Any, Dict, Iterator, List, TypeVar, Union _T = TypeVar("_T") class PackageMetadata(Protocol): def __len__(self) -> int: ... # pragma: no cover def __contains__(self, item: str) -> bool: ... # pragma: no cover def __getitem__(self, key: str) -> str: ... # pragma: no cover def __iter__(self) -> Iterator[str]: ... # pragma: no cover def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]: """ Return all values associated with a possibly multi-valued key. """ @property def json(self) -> Dict[str, Union[str, List[str]]]: """ A JSON-compatible form of the metadata. """ class SimplePath(Protocol): """ A minimal subset of pathlib.Path required by PathDistribution. >>> import pathlib >>> import typing >>> _: SimplePath = typing.cast(pathlib.Path, None) """ def joinpath(self) -> 'SimplePath': ... # pragma: no cover def __div__(self) -> 'SimplePath': ... # pragma: no cover def parent(self) -> 'SimplePath': ... # pragma: no cover def read_text(self) -> str: ... # pragma: no cover
from ._compat import Protocol from typing import Any, Dict, Iterator, List, TypeVar, Union _T = TypeVar("_T") class PackageMetadata(Protocol): def __len__(self) -> int: ... # pragma: no cover def __contains__(self, item: str) -> bool: ... # pragma: no cover def __getitem__(self, key: str) -> str: ... # pragma: no cover def __iter__(self) -> Iterator[str]: ... # pragma: no cover def get_all(self, name: str, failobj: _T = ...) -> Union[List[Any], _T]: """ Return all values associated with a possibly multi-valued key. """ @property def json(self) -> Dict[str, Union[str, List[str]]]: """ A JSON-compatible form of the metadata. """ class SimplePath(Protocol): """ A minimal subset of pathlib.Path required by PathDistribution. + + >>> import pathlib + >>> import typing + >>> _: SimplePath = typing.cast(pathlib.Path, None) """ def joinpath(self) -> 'SimplePath': ... # pragma: no cover def __div__(self) -> 'SimplePath': ... # pragma: no cover def parent(self) -> 'SimplePath': ... # pragma: no cover def read_text(self) -> str: ... # pragma: no cover
ed350a7387c376538f51a8a7a8cfde5469baba8a
tests/testutils.py
tests/testutils.py
import psycopg2 import os import getpass def get_pg_connection(): return psycopg2.connect( "dbname=bedquilt_test user={}".format(getpass.getuser()) )
import psycopg2 import os import getpass # CREATE DATABASE bedquilt_test # WITH OWNER = {{owner}} # ENCODING = 'UTF8' # TABLESPACE = pg_default # LC_COLLATE = 'en_GB.UTF-8' # LC_CTYPE = 'en_GB.UTF-8' # CONNECTION LIMIT = -1; def get_pg_connection(): return psycopg2.connect( "dbname=bedquilt_test user={}".format(getpass.getuser()) )
Add the sql to create the test database
Add the sql to create the test database
Python
mit
BedquiltDB/bedquilt-core
import psycopg2 import os import getpass + + + # CREATE DATABASE bedquilt_test + # WITH OWNER = {{owner}} + # ENCODING = 'UTF8' + # TABLESPACE = pg_default + # LC_COLLATE = 'en_GB.UTF-8' + # LC_CTYPE = 'en_GB.UTF-8' + # CONNECTION LIMIT = -1; def get_pg_connection(): return psycopg2.connect( "dbname=bedquilt_test user={}".format(getpass.getuser()) )
Add the sql to create the test database
## Code Before: import psycopg2 import os import getpass def get_pg_connection(): return psycopg2.connect( "dbname=bedquilt_test user={}".format(getpass.getuser()) ) ## Instruction: Add the sql to create the test database ## Code After: import psycopg2 import os import getpass # CREATE DATABASE bedquilt_test # WITH OWNER = {{owner}} # ENCODING = 'UTF8' # TABLESPACE = pg_default # LC_COLLATE = 'en_GB.UTF-8' # LC_CTYPE = 'en_GB.UTF-8' # CONNECTION LIMIT = -1; def get_pg_connection(): return psycopg2.connect( "dbname=bedquilt_test user={}".format(getpass.getuser()) )
import psycopg2 import os import getpass + + + # CREATE DATABASE bedquilt_test + # WITH OWNER = {{owner}} + # ENCODING = 'UTF8' + # TABLESPACE = pg_default + # LC_COLLATE = 'en_GB.UTF-8' + # LC_CTYPE = 'en_GB.UTF-8' + # CONNECTION LIMIT = -1; def get_pg_connection(): return psycopg2.connect( "dbname=bedquilt_test user={}".format(getpass.getuser()) )
8194f327032c064fe71ba3dc918e28ee2a586b12
sqlalchemy_mixins/serialize.py
sqlalchemy_mixins/serialize.py
from collections.abc import Iterable from .inspection import InspectionMixin class SerializeMixin(InspectionMixin): """Mixin to make model serializable.""" __abstract__ = True def to_dict(self,nested = False, hybrid_attributes = False, exclude = None): """Return dict object with model's data. :param nested: flag to return nested relationships' data if true :type: bool :param include_hybrid: flag to include hybrid attributes if true :return: dict """ result = dict() if exclude is None: view_cols = self.columns else : view_cols = filter(lambda e: e not in exclude, self.columns) for key in view_cols : result[key] = getattr(self, key) if hybrid_attributes: for key in self.hybrid_properties: result[key] = getattr(self, key) if nested: for key in self.relations: obj = getattr(self, key) if isinstance(obj, SerializeMixin): result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes) elif isinstance(obj, Iterable): result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj] return result
from collections.abc import Iterable from .inspection import InspectionMixin class SerializeMixin(InspectionMixin): """Mixin to make model serializable.""" __abstract__ = True def to_dict(self,nested = False, hybrid_attributes = False, exclude = None): """Return dict object with model's data. :param nested: flag to return nested relationships' data if true :type: bool :param include_hybrid: flag to include hybrid attributes if true :return: dict """ result = dict() if exclude is None: view_cols = self.columns else : view_cols = filter(lambda e: e not in exclude, self.columns) for key in view_cols : result[key] = getattr(self, key) if hybrid_attributes: for key in self.hybrid_properties: result[key] = getattr(self, key) if nested: for key in self.relations: obj = getattr(self, key) if isinstance(obj, SerializeMixin): result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes) elif isinstance(obj, Iterable): result[key] = [ o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj if isinstance(o, SerializeMixin) ] return result
Check if relation objects are class of SerializeMixin
Check if relation objects are class of SerializeMixin
Python
mit
absent1706/sqlalchemy-mixins
from collections.abc import Iterable from .inspection import InspectionMixin class SerializeMixin(InspectionMixin): """Mixin to make model serializable.""" __abstract__ = True def to_dict(self,nested = False, hybrid_attributes = False, exclude = None): """Return dict object with model's data. :param nested: flag to return nested relationships' data if true :type: bool :param include_hybrid: flag to include hybrid attributes if true :return: dict """ result = dict() if exclude is None: view_cols = self.columns else : view_cols = filter(lambda e: e not in exclude, self.columns) for key in view_cols : result[key] = getattr(self, key) if hybrid_attributes: for key in self.hybrid_properties: result[key] = getattr(self, key) if nested: for key in self.relations: obj = getattr(self, key) if isinstance(obj, SerializeMixin): result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes) elif isinstance(obj, Iterable): + result[key] = [ - result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj] + o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj + if isinstance(o, SerializeMixin) + ] return result
Check if relation objects are class of SerializeMixin
## Code Before: from collections.abc import Iterable from .inspection import InspectionMixin class SerializeMixin(InspectionMixin): """Mixin to make model serializable.""" __abstract__ = True def to_dict(self,nested = False, hybrid_attributes = False, exclude = None): """Return dict object with model's data. :param nested: flag to return nested relationships' data if true :type: bool :param include_hybrid: flag to include hybrid attributes if true :return: dict """ result = dict() if exclude is None: view_cols = self.columns else : view_cols = filter(lambda e: e not in exclude, self.columns) for key in view_cols : result[key] = getattr(self, key) if hybrid_attributes: for key in self.hybrid_properties: result[key] = getattr(self, key) if nested: for key in self.relations: obj = getattr(self, key) if isinstance(obj, SerializeMixin): result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes) elif isinstance(obj, Iterable): result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj] return result ## Instruction: Check if relation objects are class of SerializeMixin ## Code After: from collections.abc import Iterable from .inspection import InspectionMixin class SerializeMixin(InspectionMixin): """Mixin to make model serializable.""" __abstract__ = True def to_dict(self,nested = False, hybrid_attributes = False, exclude = None): """Return dict object with model's data. :param nested: flag to return nested relationships' data if true :type: bool :param include_hybrid: flag to include hybrid attributes if true :return: dict """ result = dict() if exclude is None: view_cols = self.columns else : view_cols = filter(lambda e: e not in exclude, self.columns) for key in view_cols : result[key] = getattr(self, key) if hybrid_attributes: for key in self.hybrid_properties: result[key] = getattr(self, key) if nested: for key in self.relations: obj = getattr(self, key) if isinstance(obj, SerializeMixin): result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes) elif isinstance(obj, Iterable): result[key] = [ o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj if isinstance(o, SerializeMixin) ] return result
from collections.abc import Iterable from .inspection import InspectionMixin class SerializeMixin(InspectionMixin): """Mixin to make model serializable.""" __abstract__ = True def to_dict(self,nested = False, hybrid_attributes = False, exclude = None): """Return dict object with model's data. :param nested: flag to return nested relationships' data if true :type: bool :param include_hybrid: flag to include hybrid attributes if true :return: dict """ result = dict() if exclude is None: view_cols = self.columns else : view_cols = filter(lambda e: e not in exclude, self.columns) for key in view_cols : result[key] = getattr(self, key) if hybrid_attributes: for key in self.hybrid_properties: result[key] = getattr(self, key) if nested: for key in self.relations: obj = getattr(self, key) if isinstance(obj, SerializeMixin): result[key] = obj.to_dict(hybrid_attributes=hybrid_attributes) elif isinstance(obj, Iterable): + result[key] = [ - result[key] = [o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj] ? ----------- - ^ - + o.to_dict(hybrid_attributes=hybrid_attributes) for o in obj ? ^^ + if isinstance(o, SerializeMixin) + ] return result
63a539ff4a3a832286136c40a74b1a8b3db1a5c0
falcom/api/uri/api_querier.py
falcom/api/uri/api_querier.py
from time import sleep class APIQuerier: def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0): self.uri = uri self.url_opener = url_opener self.sleep_time = sleep_time self.max_tries = max_tries def get (self, **kwargs): try: return self.__open_uri(kwargs) except ConnectionError: sleep(self.sleep_time) i = 1 while i != self.max_tries: i += 1 try: return self.__open_uri(kwargs) except ConnectionError: sleep(self.sleep_time) return b"" @staticmethod def utf8 (str_or_bytes): if isinstance(str_or_bytes, bytes): return str_or_bytes.decode("utf_8") else: return str_or_bytes def __open_uri (self, kwargs): with self.url_opener(self.uri(**kwargs)) as response: result = self.utf8(response.read()) return result
from time import sleep class APIQuerier: def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0): self.uri = uri self.url_opener = url_opener self.sleep_time = sleep_time self.max_tries = max_tries def get (self, **kwargs): class SpecialNull: pass result = SpecialNull i = 1 while result is SpecialNull: try: result = self.__open_uri(kwargs) except ConnectionError: sleep(self.sleep_time) if i == self.max_tries: result = b"" else: i += 1 return result @staticmethod def utf8 (str_or_bytes): if isinstance(str_or_bytes, bytes): return str_or_bytes.decode("utf_8") else: return str_or_bytes def __open_uri (self, kwargs): with self.url_opener(self.uri(**kwargs)) as response: result = self.utf8(response.read()) return result
Rewrite get() to be less repetitive but still stupid
Rewrite get() to be less repetitive but still stupid
Python
bsd-3-clause
mlibrary/image-conversion-and-validation,mlibrary/image-conversion-and-validation
from time import sleep class APIQuerier: def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0): self.uri = uri self.url_opener = url_opener self.sleep_time = sleep_time self.max_tries = max_tries def get (self, **kwargs): - try: - return self.__open_uri(kwargs) + class SpecialNull: pass + result = SpecialNull + i = 1 + while result is SpecialNull: - except ConnectionError: - sleep(self.sleep_time) - - i = 1 - while i != self.max_tries: - i += 1 - try: - return self.__open_uri(kwargs) + result = self.__open_uri(kwargs) except ConnectionError: sleep(self.sleep_time) + if i == self.max_tries: + result = b"" + + else: + i += 1 + - return b"" + return result @staticmethod def utf8 (str_or_bytes): if isinstance(str_or_bytes, bytes): return str_or_bytes.decode("utf_8") else: return str_or_bytes def __open_uri (self, kwargs): with self.url_opener(self.uri(**kwargs)) as response: result = self.utf8(response.read()) return result
Rewrite get() to be less repetitive but still stupid
## Code Before: from time import sleep class APIQuerier: def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0): self.uri = uri self.url_opener = url_opener self.sleep_time = sleep_time self.max_tries = max_tries def get (self, **kwargs): try: return self.__open_uri(kwargs) except ConnectionError: sleep(self.sleep_time) i = 1 while i != self.max_tries: i += 1 try: return self.__open_uri(kwargs) except ConnectionError: sleep(self.sleep_time) return b"" @staticmethod def utf8 (str_or_bytes): if isinstance(str_or_bytes, bytes): return str_or_bytes.decode("utf_8") else: return str_or_bytes def __open_uri (self, kwargs): with self.url_opener(self.uri(**kwargs)) as response: result = self.utf8(response.read()) return result ## Instruction: Rewrite get() to be less repetitive but still stupid ## Code After: from time import sleep class APIQuerier: def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0): self.uri = uri self.url_opener = url_opener self.sleep_time = sleep_time self.max_tries = max_tries def get (self, **kwargs): class SpecialNull: pass result = SpecialNull i = 1 while result is SpecialNull: try: result = self.__open_uri(kwargs) except ConnectionError: sleep(self.sleep_time) if i == self.max_tries: result = b"" else: i += 1 return result @staticmethod def utf8 (str_or_bytes): if isinstance(str_or_bytes, bytes): return str_or_bytes.decode("utf_8") else: return str_or_bytes def __open_uri (self, kwargs): with self.url_opener(self.uri(**kwargs)) as response: result = self.utf8(response.read()) return result
from time import sleep class APIQuerier: def __init__ (self, uri, url_opener, sleep_time=300, max_tries=0): self.uri = uri self.url_opener = url_opener self.sleep_time = sleep_time self.max_tries = max_tries def get (self, **kwargs): - try: - return self.__open_uri(kwargs) + class SpecialNull: pass + result = SpecialNull + i = 1 + while result is SpecialNull: - except ConnectionError: - sleep(self.sleep_time) - - i = 1 - while i != self.max_tries: - i += 1 - try: - return self.__open_uri(kwargs) ? ^^^ + result = self.__open_uri(kwargs) ? +++ ^^ except ConnectionError: sleep(self.sleep_time) + if i == self.max_tries: + result = b"" + + else: + i += 1 + - return b"" ? ^^^ + return result ? ^^^^^^ @staticmethod def utf8 (str_or_bytes): if isinstance(str_or_bytes, bytes): return str_or_bytes.decode("utf_8") else: return str_or_bytes def __open_uri (self, kwargs): with self.url_opener(self.uri(**kwargs)) as response: result = self.utf8(response.read()) return result
02ac5dcfa6bdaf9b8152ef2f49fd61afe9faf8ab
client/python/plot_request_times.py
client/python/plot_request_times.py
import requests from plotly.offline import plot import plotly.graph_objs as go r = requests.get('http://localhost:8081/monitor_results/1') print(r.json()) # build traces for plotting from monitoring data request_times = list() timestamps = list() timestamp = 0 url = r.json()[0]['urlToMonitor']['url'] for monitoring_data in r.json(): request_time = monitoring_data['timeNeededForRequest'] request_times.append(request_time) timestamps.append(timestamp) timestamp = timestamp + 1 plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name = 'THE OTHER NAME')], filename='request_times.html')
import requests from plotly.offline import plot import plotly.graph_objs as go def build_data_for_monitored_url(id): '''Fetches and prepares data for plotting for the given URL id''' r = requests.get('http://localhost:8081/monitor_results/' + str(id)) # build traces for plotting from monitoring data request_times = list() timestamps = list() timestamp = 0 url = r.json()[0]['urlToMonitor']['url'] for monitoring_data in r.json(): request_time = monitoring_data['timeNeededForRequest'] request_times.append(request_time) timestamps.append(timestamp) timestamp = timestamp + 1 return go.Scatter(x = timestamps, y = request_times, name = url) # get all monitored sites and fetch data for it r = requests.get('http://localhost:8081/monitored-sites') plotting_data = list() for monitored_site in r.json(): print('Fetching data for ' + monitored_site['url']) data_for_site = build_data_for_monitored_url(monitored_site['id']) plotting_data.append(data_for_site) plot(plotting_data, filename='request_times.html')
Implement fetching all monitored data
Implement fetching all monitored data
Python
mit
gernd/simple-site-mon
import requests from plotly.offline import plot import plotly.graph_objs as go + def build_data_for_monitored_url(id): + '''Fetches and prepares data for plotting for the given URL id''' - r = requests.get('http://localhost:8081/monitor_results/1') + r = requests.get('http://localhost:8081/monitor_results/' + str(id)) - print(r.json()) - # build traces for plotting from monitoring data + # build traces for plotting from monitoring data - request_times = list() + request_times = list() - timestamps = list() + timestamps = list() - timestamp = 0 + timestamp = 0 - url = r.json()[0]['urlToMonitor']['url'] + url = r.json()[0]['urlToMonitor']['url'] - for monitoring_data in r.json(): + for monitoring_data in r.json(): - request_time = monitoring_data['timeNeededForRequest'] + request_time = monitoring_data['timeNeededForRequest'] - request_times.append(request_time) + request_times.append(request_time) - timestamps.append(timestamp) + timestamps.append(timestamp) - timestamp = timestamp + 1 + timestamp = timestamp + 1 + return go.Scatter(x = timestamps, y = request_times, name = url) - plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name = - 'THE OTHER NAME')], filename='request_times.html') + # get all monitored sites and fetch data for it + r = requests.get('http://localhost:8081/monitored-sites') + plotting_data = list() + for monitored_site in r.json(): + print('Fetching data for ' + monitored_site['url']) + data_for_site = build_data_for_monitored_url(monitored_site['id']) + plotting_data.append(data_for_site) + plot(plotting_data, filename='request_times.html') +
Implement fetching all monitored data
## Code Before: import requests from plotly.offline import plot import plotly.graph_objs as go r = requests.get('http://localhost:8081/monitor_results/1') print(r.json()) # build traces for plotting from monitoring data request_times = list() timestamps = list() timestamp = 0 url = r.json()[0]['urlToMonitor']['url'] for monitoring_data in r.json(): request_time = monitoring_data['timeNeededForRequest'] request_times.append(request_time) timestamps.append(timestamp) timestamp = timestamp + 1 plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name = 'THE OTHER NAME')], filename='request_times.html') ## Instruction: Implement fetching all monitored data ## Code After: import requests from plotly.offline import plot import plotly.graph_objs as go def build_data_for_monitored_url(id): '''Fetches and prepares data for plotting for the given URL id''' r = requests.get('http://localhost:8081/monitor_results/' + str(id)) # build traces for plotting from monitoring data request_times = list() timestamps = list() timestamp = 0 url = r.json()[0]['urlToMonitor']['url'] for monitoring_data in r.json(): request_time = monitoring_data['timeNeededForRequest'] request_times.append(request_time) timestamps.append(timestamp) timestamp = timestamp + 1 return go.Scatter(x = timestamps, y = request_times, name = url) # get all monitored sites and fetch data for it r = requests.get('http://localhost:8081/monitored-sites') plotting_data = list() for monitored_site in r.json(): print('Fetching data for ' + monitored_site['url']) data_for_site = build_data_for_monitored_url(monitored_site['id']) plotting_data.append(data_for_site) plot(plotting_data, filename='request_times.html')
import requests from plotly.offline import plot import plotly.graph_objs as go + def build_data_for_monitored_url(id): + '''Fetches and prepares data for plotting for the given URL id''' - r = requests.get('http://localhost:8081/monitor_results/1') ? - + r = requests.get('http://localhost:8081/monitor_results/' + str(id)) ? ++++ +++++++++ + - print(r.json()) - # build traces for plotting from monitoring data + # build traces for plotting from monitoring data ? ++++ - request_times = list() + request_times = list() ? ++++ - timestamps = list() + timestamps = list() ? ++++ - timestamp = 0 + timestamp = 0 ? ++++ - url = r.json()[0]['urlToMonitor']['url'] + url = r.json()[0]['urlToMonitor']['url'] ? ++++ - for monitoring_data in r.json(): + for monitoring_data in r.json(): ? ++++ - request_time = monitoring_data['timeNeededForRequest'] + request_time = monitoring_data['timeNeededForRequest'] ? ++++ - request_times.append(request_time) + request_times.append(request_time) ? ++++ - timestamps.append(timestamp) + timestamps.append(timestamp) ? ++++ - timestamp = timestamp + 1 + timestamp = timestamp + 1 ? ++++ + return go.Scatter(x = timestamps, y = request_times, name = url) - plot([go.Scatter(x = timestamps, y = request_times, name = 'THE NAME'), go.Scatter(x = timestamps, y = request_times, name = - 'THE OTHER NAME')], filename='request_times.html') + # get all monitored sites and fetch data for it + r = requests.get('http://localhost:8081/monitored-sites') + plotting_data = list() + for monitored_site in r.json(): + print('Fetching data for ' + monitored_site['url']) + data_for_site = build_data_for_monitored_url(monitored_site['id']) + plotting_data.append(data_for_site) + + plot(plotting_data, filename='request_times.html')
976ca1d7f02a0aab7397a6eb1784436593e6c644
watchman/management/commands/watchman.py
watchman/management/commands/watchman.py
from __future__ import absolute_import import json from optparse import make_option from django.core.management.base import BaseCommand, CommandError from watchman.utils import get_checks class Command(BaseCommand): help = 'Runs the default django-watchman checks' option_list = BaseCommand.option_list + ( make_option( '-c', '--checks', dest='checks', help='A comma-separated list of watchman checks to run (full python dotted paths)' ), make_option( '-s', '--skips', dest='skips', help='A comma-separated list of watchman checks to skip (full python dotted paths)' ), ) def handle(self, *args, **options): check_list = None skip_list = None verbosity = options['verbosity'] print_all_checks = verbosity == '2' or verbosity == '3' checks = options['checks'] skips = options['skips'] if checks is not None: check_list = checks.split(',') if skips is not None: skip_list = skips.split(',') for check in get_checks(check_list=check_list, skip_list=skip_list): if callable(check): resp = json.dumps(check()) if '"ok": false' in resp: raise CommandError(resp) elif print_all_checks: self.stdout.write(resp)
from __future__ import absolute_import import json from optparse import make_option from django.core.management.base import BaseCommand, CommandError from watchman.utils import get_checks class Command(BaseCommand): help = 'Runs the default django-watchman checks' option_list = BaseCommand.option_list + ( make_option( '-c', '--checks', dest='checks', help='A comma-separated list of watchman checks to run (full python dotted paths)' ), make_option( '-s', '--skips', dest='skips', help='A comma-separated list of watchman checks to skip (full python dotted paths)' ), ) def handle(self, *args, **options): check_list = None skip_list = None verbosity = options['verbosity'] print_all_checks = verbosity in ['2', '3', ] checks = options['checks'] skips = options['skips'] if checks is not None: check_list = checks.split(',') if skips is not None: skip_list = skips.split(',') for check in get_checks(check_list=check_list, skip_list=skip_list): if callable(check): resp = json.dumps(check()) if '"ok": false' in resp: raise CommandError(resp) elif print_all_checks: self.stdout.write(resp)
Swap equality checks for `in`
Swap equality checks for `in`
Python
bsd-3-clause
mwarkentin/django-watchman,mwarkentin/django-watchman,JBKahn/django-watchman,JBKahn/django-watchman
from __future__ import absolute_import import json from optparse import make_option from django.core.management.base import BaseCommand, CommandError from watchman.utils import get_checks class Command(BaseCommand): help = 'Runs the default django-watchman checks' option_list = BaseCommand.option_list + ( make_option( '-c', '--checks', dest='checks', help='A comma-separated list of watchman checks to run (full python dotted paths)' ), make_option( '-s', '--skips', dest='skips', help='A comma-separated list of watchman checks to skip (full python dotted paths)' ), ) def handle(self, *args, **options): check_list = None skip_list = None verbosity = options['verbosity'] - print_all_checks = verbosity == '2' or verbosity == '3' + print_all_checks = verbosity in ['2', '3', ] checks = options['checks'] skips = options['skips'] if checks is not None: check_list = checks.split(',') if skips is not None: skip_list = skips.split(',') for check in get_checks(check_list=check_list, skip_list=skip_list): if callable(check): resp = json.dumps(check()) if '"ok": false' in resp: raise CommandError(resp) elif print_all_checks: self.stdout.write(resp)
Swap equality checks for `in`
## Code Before: from __future__ import absolute_import import json from optparse import make_option from django.core.management.base import BaseCommand, CommandError from watchman.utils import get_checks class Command(BaseCommand): help = 'Runs the default django-watchman checks' option_list = BaseCommand.option_list + ( make_option( '-c', '--checks', dest='checks', help='A comma-separated list of watchman checks to run (full python dotted paths)' ), make_option( '-s', '--skips', dest='skips', help='A comma-separated list of watchman checks to skip (full python dotted paths)' ), ) def handle(self, *args, **options): check_list = None skip_list = None verbosity = options['verbosity'] print_all_checks = verbosity == '2' or verbosity == '3' checks = options['checks'] skips = options['skips'] if checks is not None: check_list = checks.split(',') if skips is not None: skip_list = skips.split(',') for check in get_checks(check_list=check_list, skip_list=skip_list): if callable(check): resp = json.dumps(check()) if '"ok": false' in resp: raise CommandError(resp) elif print_all_checks: self.stdout.write(resp) ## Instruction: Swap equality checks for `in` ## Code After: from __future__ import absolute_import import json from optparse import make_option from django.core.management.base import BaseCommand, CommandError from watchman.utils import get_checks class Command(BaseCommand): help = 'Runs the default django-watchman checks' option_list = BaseCommand.option_list + ( make_option( '-c', '--checks', dest='checks', help='A comma-separated list of watchman checks to run (full python dotted paths)' ), make_option( '-s', '--skips', dest='skips', help='A comma-separated list of watchman checks to skip (full python dotted paths)' ), ) def handle(self, *args, **options): check_list = None skip_list = None verbosity = options['verbosity'] print_all_checks = verbosity in ['2', '3', ] checks = options['checks'] skips = options['skips'] if checks is not None: check_list = checks.split(',') if skips is not None: skip_list = skips.split(',') for check in get_checks(check_list=check_list, skip_list=skip_list): if callable(check): resp = json.dumps(check()) if '"ok": false' in resp: raise CommandError(resp) elif print_all_checks: self.stdout.write(resp)
from __future__ import absolute_import import json from optparse import make_option from django.core.management.base import BaseCommand, CommandError from watchman.utils import get_checks class Command(BaseCommand): help = 'Runs the default django-watchman checks' option_list = BaseCommand.option_list + ( make_option( '-c', '--checks', dest='checks', help='A comma-separated list of watchman checks to run (full python dotted paths)' ), make_option( '-s', '--skips', dest='skips', help='A comma-separated list of watchman checks to skip (full python dotted paths)' ), ) def handle(self, *args, **options): check_list = None skip_list = None verbosity = options['verbosity'] - print_all_checks = verbosity == '2' or verbosity == '3' ? ^^ ^^^^^^^^^^^^^^^^ + print_all_checks = verbosity in ['2', '3', ] ? ^^ + ^ +++ checks = options['checks'] skips = options['skips'] if checks is not None: check_list = checks.split(',') if skips is not None: skip_list = skips.split(',') for check in get_checks(check_list=check_list, skip_list=skip_list): if callable(check): resp = json.dumps(check()) if '"ok": false' in resp: raise CommandError(resp) elif print_all_checks: self.stdout.write(resp)
9c176de1fd280e72dd06c9eaa64060e52abca746
python/prebuild.py
python/prebuild.py
def extract_function_names(module): ''' extract function names from attributes of 'module'. ''' from importlib import import_module mod = import_module(module.__name__) attr_list = dir(mod) scope = locals() def iscallable(name): return eval('callable(mod.{})'.format(name), scope) return filter(iscallable, attr_list) def gen_pyplot_functions(dub_root): ''' generate 'pyplot_functions.txt' for matplotlibd.pyplot. ''' import matplotlib.pyplot from string import lowercase functions = filter(lambda i: i[0] != '_' or i[0] in lowercase, extract_function_names(matplotlib.pyplot)) with open(dub_root + "/views/pyplot_functions.txt", "w") as f: f.write("\n".join(functions)) if __name__ == '__main__': from sys import argv gen_pyplot_functions(argv[1])
def extract_function_names(module): ''' extract function names from attributes of 'module'. ''' from importlib import import_module mod = import_module(module.__name__) attr_list = dir(mod) scope = locals() def iscallable(name): ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated'] return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators return filter(iscallable, attr_list) def gen_pyplot_functions(dub_root): ''' generate 'pyplot_functions.txt' for matplotlibd.pyplot. ''' import matplotlib.pyplot from string import lowercase functions = filter(lambda i: i[0] != '_' or i[0] in lowercase, extract_function_names(matplotlib.pyplot)) with open(dub_root + "/views/pyplot_functions.txt", "w") as f: f.write("\n".join(functions)) if __name__ == '__main__': from sys import argv gen_pyplot_functions(argv[1])
Remove python decorators from list
Remove python decorators from list
Python
mit
koji-kojiro/matplotlib-d
def extract_function_names(module): ''' extract function names from attributes of 'module'. ''' from importlib import import_module mod = import_module(module.__name__) attr_list = dir(mod) scope = locals() def iscallable(name): + ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated'] - return eval('callable(mod.{})'.format(name), scope) + return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators return filter(iscallable, attr_list) def gen_pyplot_functions(dub_root): ''' generate 'pyplot_functions.txt' for matplotlibd.pyplot. ''' import matplotlib.pyplot from string import lowercase functions = filter(lambda i: i[0] != '_' or i[0] in lowercase, extract_function_names(matplotlib.pyplot)) with open(dub_root + "/views/pyplot_functions.txt", "w") as f: f.write("\n".join(functions)) if __name__ == '__main__': from sys import argv gen_pyplot_functions(argv[1])
Remove python decorators from list
## Code Before: def extract_function_names(module): ''' extract function names from attributes of 'module'. ''' from importlib import import_module mod = import_module(module.__name__) attr_list = dir(mod) scope = locals() def iscallable(name): return eval('callable(mod.{})'.format(name), scope) return filter(iscallable, attr_list) def gen_pyplot_functions(dub_root): ''' generate 'pyplot_functions.txt' for matplotlibd.pyplot. ''' import matplotlib.pyplot from string import lowercase functions = filter(lambda i: i[0] != '_' or i[0] in lowercase, extract_function_names(matplotlib.pyplot)) with open(dub_root + "/views/pyplot_functions.txt", "w") as f: f.write("\n".join(functions)) if __name__ == '__main__': from sys import argv gen_pyplot_functions(argv[1]) ## Instruction: Remove python decorators from list ## Code After: def extract_function_names(module): ''' extract function names from attributes of 'module'. ''' from importlib import import_module mod = import_module(module.__name__) attr_list = dir(mod) scope = locals() def iscallable(name): ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated'] return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators return filter(iscallable, attr_list) def gen_pyplot_functions(dub_root): ''' generate 'pyplot_functions.txt' for matplotlibd.pyplot. ''' import matplotlib.pyplot from string import lowercase functions = filter(lambda i: i[0] != '_' or i[0] in lowercase, extract_function_names(matplotlib.pyplot)) with open(dub_root + "/views/pyplot_functions.txt", "w") as f: f.write("\n".join(functions)) if __name__ == '__main__': from sys import argv gen_pyplot_functions(argv[1])
def extract_function_names(module): ''' extract function names from attributes of 'module'. ''' from importlib import import_module mod = import_module(module.__name__) attr_list = dir(mod) scope = locals() def iscallable(name): + ignore_decorators = ['dedent','deprecated','silent_list', 'warn_deprecated'] - return eval('callable(mod.{})'.format(name), scope) + return eval('callable(mod.{})'.format(name), scope) and name not in ignore_decorators ? ++++++++++++++++++++++++++++++++++ return filter(iscallable, attr_list) def gen_pyplot_functions(dub_root): ''' generate 'pyplot_functions.txt' for matplotlibd.pyplot. ''' import matplotlib.pyplot from string import lowercase functions = filter(lambda i: i[0] != '_' or i[0] in lowercase, extract_function_names(matplotlib.pyplot)) with open(dub_root + "/views/pyplot_functions.txt", "w") as f: f.write("\n".join(functions)) if __name__ == '__main__': from sys import argv gen_pyplot_functions(argv[1])
7c0d68b1bce27d026b69e3a069c549ab560b0f3d
spillway/mixins.py
spillway/mixins.py
class FormMixin(object): """Mixin to provide form validation and data cleaning of GET or POST requests. """ form_class = None @property def form(self): """Returns a validated form dict or an empty dict.""" _form = getattr(self, '_form', False) if not _form: self._form = self.form_class(self.request.GET or self.request.POST, self.request.FILES or None) valid = self._form.is_valid() return self._form
class FormMixin(object): """Mixin to provide form validation and data cleaning of GET or POST requests. """ form_class = None @property def form(self): """Returns a validated form dict or an empty dict.""" _form = getattr(self, '_form', False) if not _form: self._form = self.form_class( self.request.QUERY_PARAMS or self.request.DATA, self.request.FILES or None) valid = self._form.is_valid() return self._form
Use DRF query params and data request attrs
Use DRF query params and data request attrs
Python
bsd-3-clause
barseghyanartur/django-spillway,kuzmich/django-spillway,bkg/django-spillway
class FormMixin(object): """Mixin to provide form validation and data cleaning of GET or POST requests. """ form_class = None @property def form(self): """Returns a validated form dict or an empty dict.""" _form = getattr(self, '_form', False) if not _form: - self._form = self.form_class(self.request.GET or self.request.POST, + self._form = self.form_class( + self.request.QUERY_PARAMS or self.request.DATA, - self.request.FILES or None) + self.request.FILES or None) valid = self._form.is_valid() return self._form
Use DRF query params and data request attrs
## Code Before: class FormMixin(object): """Mixin to provide form validation and data cleaning of GET or POST requests. """ form_class = None @property def form(self): """Returns a validated form dict or an empty dict.""" _form = getattr(self, '_form', False) if not _form: self._form = self.form_class(self.request.GET or self.request.POST, self.request.FILES or None) valid = self._form.is_valid() return self._form ## Instruction: Use DRF query params and data request attrs ## Code After: class FormMixin(object): """Mixin to provide form validation and data cleaning of GET or POST requests. """ form_class = None @property def form(self): """Returns a validated form dict or an empty dict.""" _form = getattr(self, '_form', False) if not _form: self._form = self.form_class( self.request.QUERY_PARAMS or self.request.DATA, self.request.FILES or None) valid = self._form.is_valid() return self._form
class FormMixin(object): """Mixin to provide form validation and data cleaning of GET or POST requests. """ form_class = None @property def form(self): """Returns a validated form dict or an empty dict.""" _form = getattr(self, '_form', False) if not _form: - self._form = self.form_class(self.request.GET or self.request.POST, + self._form = self.form_class( + self.request.QUERY_PARAMS or self.request.DATA, - self.request.FILES or None) ? ------------------------- + self.request.FILES or None) valid = self._form.is_valid() return self._form
d5c59c018ba7558a9d21370d7eb58ab590779cf1
plugins/autojoin/plugin_tests/autojoin_test.py
plugins/autojoin/plugin_tests/autojoin_test.py
from tests import base def setUpModule(): base.enabledPlugins.append('autojoin') base.startServer() def tearDownModule(): base.stopServer() class AutoJoinTest(base.TestCase): def setUp(self): base.TestCase.setUp(self)
from girder.constants import AccessType from tests import base import json def setUpModule(): base.enabledPlugins.append('autojoin') base.startServer() def tearDownModule(): base.stopServer() class AutoJoinTest(base.TestCase): def setUp(self): base.TestCase.setUp(self) self.users = [self.model('user').createUser( 'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num) for num in [0, 1]] def testCuration(self): admin, user = self.users # create some groups g1 = self.model('group').createGroup('g1', admin) g2 = self.model('group').createGroup('g2', admin) g3 = self.model('group').createGroup('g3', admin) # set auto join rules rules = [ { 'pattern': '@kitware.com', 'groupId': str(g1['_id']), 'level': AccessType.ADMIN }, { 'pattern': '@example.com', 'groupId': str(g2['_id']), 'level': AccessType.READ }, { 'pattern': '@example.com', 'groupId': str(g3['_id']), 'level': AccessType.WRITE }, ] params = { 'list': json.dumps([{'key': 'autojoin', 'value': rules}]) } resp = self.request( '/system/setting', user=admin, method='PUT', params=params) self.assertStatusOk(resp) # create users user1 = self.model('user').createUser( 'user1', 'password', 'John', 'Doe', 'user1@example.com') user2 = self.model('user').createUser( 'user2', 'password', 'John', 'Doe', 'user2@kitware.com') user3 = self.model('user').createUser( 'user3', 'password', 'John', 'Doe', 'user3@kitware.co') # check correct groups were joined self.assertEqual(user1['groups'], [g2['_id'], g3['_id']]) self.assertEqual(user2['groups'], [g1['_id']]) self.assertEqual(user3['groups'], []) # check correct access levels g1 = self.model('group').load(g1['_id'], force=True) g2 = self.model('group').load(g2['_id'], force=True) g3 = self.model('group').load(g3['_id'], force=True) self.assertTrue( {u'id': user2['_id'], u'level': AccessType.ADMIN} in g1['access']['users']) self.assertTrue( {u'id': user1['_id'], u'level': AccessType.WRITE} in g3['access']['users'])
Add server tests for auto join plugin
Add server tests for auto join plugin
Python
apache-2.0
kotfic/girder,kotfic/girder,adsorensen/girder,jbeezley/girder,data-exp-lab/girder,girder/girder,sutartmelson/girder,Kitware/girder,girder/girder,sutartmelson/girder,jbeezley/girder,RafaelPalomar/girder,girder/girder,adsorensen/girder,RafaelPalomar/girder,manthey/girder,manthey/girder,data-exp-lab/girder,RafaelPalomar/girder,Kitware/girder,RafaelPalomar/girder,manthey/girder,Kitware/girder,data-exp-lab/girder,sutartmelson/girder,data-exp-lab/girder,kotfic/girder,girder/girder,jbeezley/girder,manthey/girder,adsorensen/girder,Xarthisius/girder,kotfic/girder,Kitware/girder,adsorensen/girder,data-exp-lab/girder,RafaelPalomar/girder,jbeezley/girder,Xarthisius/girder,sutartmelson/girder,Xarthisius/girder,adsorensen/girder,kotfic/girder,Xarthisius/girder,sutartmelson/girder,Xarthisius/girder
+ from girder.constants import AccessType from tests import base + import json def setUpModule(): base.enabledPlugins.append('autojoin') base.startServer() def tearDownModule(): base.stopServer() class AutoJoinTest(base.TestCase): def setUp(self): base.TestCase.setUp(self) + self.users = [self.model('user').createUser( + 'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num) + for num in [0, 1]] + + def testCuration(self): + admin, user = self.users + + # create some groups + g1 = self.model('group').createGroup('g1', admin) + g2 = self.model('group').createGroup('g2', admin) + g3 = self.model('group').createGroup('g3', admin) + + # set auto join rules + rules = [ + { + 'pattern': '@kitware.com', + 'groupId': str(g1['_id']), + 'level': AccessType.ADMIN + }, + { + 'pattern': '@example.com', + 'groupId': str(g2['_id']), + 'level': AccessType.READ + }, + { + 'pattern': '@example.com', + 'groupId': str(g3['_id']), + 'level': AccessType.WRITE + }, + ] + params = { + 'list': json.dumps([{'key': 'autojoin', 'value': rules}]) + } + resp = self.request( + '/system/setting', user=admin, method='PUT', params=params) + self.assertStatusOk(resp) + + # create users + user1 = self.model('user').createUser( + 'user1', 'password', 'John', 'Doe', 'user1@example.com') + user2 = self.model('user').createUser( + 'user2', 'password', 'John', 'Doe', 'user2@kitware.com') + user3 = self.model('user').createUser( + 'user3', 'password', 'John', 'Doe', 'user3@kitware.co') + + # check correct groups were joined + self.assertEqual(user1['groups'], [g2['_id'], g3['_id']]) + self.assertEqual(user2['groups'], [g1['_id']]) + self.assertEqual(user3['groups'], []) + + # check correct access levels + g1 = self.model('group').load(g1['_id'], force=True) + g2 = self.model('group').load(g2['_id'], force=True) + g3 = self.model('group').load(g3['_id'], force=True) + self.assertTrue( + {u'id': user2['_id'], u'level': AccessType.ADMIN} in + g1['access']['users']) + self.assertTrue( + {u'id': user1['_id'], u'level': AccessType.WRITE} in + g3['access']['users']) +
Add server tests for auto join plugin
## Code Before: from tests import base def setUpModule(): base.enabledPlugins.append('autojoin') base.startServer() def tearDownModule(): base.stopServer() class AutoJoinTest(base.TestCase): def setUp(self): base.TestCase.setUp(self) ## Instruction: Add server tests for auto join plugin ## Code After: from girder.constants import AccessType from tests import base import json def setUpModule(): base.enabledPlugins.append('autojoin') base.startServer() def tearDownModule(): base.stopServer() class AutoJoinTest(base.TestCase): def setUp(self): base.TestCase.setUp(self) self.users = [self.model('user').createUser( 'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num) for num in [0, 1]] def testCuration(self): admin, user = self.users # create some groups g1 = self.model('group').createGroup('g1', admin) g2 = self.model('group').createGroup('g2', admin) g3 = self.model('group').createGroup('g3', admin) # set auto join rules rules = [ { 'pattern': '@kitware.com', 'groupId': str(g1['_id']), 'level': AccessType.ADMIN }, { 'pattern': '@example.com', 'groupId': str(g2['_id']), 'level': AccessType.READ }, { 'pattern': '@example.com', 'groupId': str(g3['_id']), 'level': AccessType.WRITE }, ] params = { 'list': json.dumps([{'key': 'autojoin', 'value': rules}]) } resp = self.request( '/system/setting', user=admin, method='PUT', params=params) self.assertStatusOk(resp) # create users user1 = self.model('user').createUser( 'user1', 'password', 'John', 'Doe', 'user1@example.com') user2 = self.model('user').createUser( 'user2', 'password', 'John', 'Doe', 'user2@kitware.com') user3 = self.model('user').createUser( 'user3', 'password', 'John', 'Doe', 'user3@kitware.co') # check correct groups were joined self.assertEqual(user1['groups'], [g2['_id'], g3['_id']]) self.assertEqual(user2['groups'], [g1['_id']]) self.assertEqual(user3['groups'], []) # check correct access levels g1 = self.model('group').load(g1['_id'], force=True) g2 = self.model('group').load(g2['_id'], force=True) g3 = self.model('group').load(g3['_id'], force=True) self.assertTrue( {u'id': user2['_id'], u'level': AccessType.ADMIN} in g1['access']['users']) self.assertTrue( {u'id': user1['_id'], u'level': AccessType.WRITE} in g3['access']['users'])
+ from girder.constants import AccessType from tests import base + import json def setUpModule(): base.enabledPlugins.append('autojoin') base.startServer() def tearDownModule(): base.stopServer() class AutoJoinTest(base.TestCase): def setUp(self): base.TestCase.setUp(self) + + self.users = [self.model('user').createUser( + 'usr%s' % num, 'passwd', 'tst', 'usr', 'u%s@u.com' % num) + for num in [0, 1]] + + def testCuration(self): + admin, user = self.users + + # create some groups + g1 = self.model('group').createGroup('g1', admin) + g2 = self.model('group').createGroup('g2', admin) + g3 = self.model('group').createGroup('g3', admin) + + # set auto join rules + rules = [ + { + 'pattern': '@kitware.com', + 'groupId': str(g1['_id']), + 'level': AccessType.ADMIN + }, + { + 'pattern': '@example.com', + 'groupId': str(g2['_id']), + 'level': AccessType.READ + }, + { + 'pattern': '@example.com', + 'groupId': str(g3['_id']), + 'level': AccessType.WRITE + }, + ] + params = { + 'list': json.dumps([{'key': 'autojoin', 'value': rules}]) + } + resp = self.request( + '/system/setting', user=admin, method='PUT', params=params) + self.assertStatusOk(resp) + + # create users + user1 = self.model('user').createUser( + 'user1', 'password', 'John', 'Doe', 'user1@example.com') + user2 = self.model('user').createUser( + 'user2', 'password', 'John', 'Doe', 'user2@kitware.com') + user3 = self.model('user').createUser( + 'user3', 'password', 'John', 'Doe', 'user3@kitware.co') + + # check correct groups were joined + self.assertEqual(user1['groups'], [g2['_id'], g3['_id']]) + self.assertEqual(user2['groups'], [g1['_id']]) + self.assertEqual(user3['groups'], []) + + # check correct access levels + g1 = self.model('group').load(g1['_id'], force=True) + g2 = self.model('group').load(g2['_id'], force=True) + g3 = self.model('group').load(g3['_id'], force=True) + self.assertTrue( + {u'id': user2['_id'], u'level': AccessType.ADMIN} in + g1['access']['users']) + self.assertTrue( + {u'id': user1['_id'], u'level': AccessType.WRITE} in + g3['access']['users'])
21c7232081483c05752e6db3d60692a04d482d24
dakota/tests/test_dakota_base.py
dakota/tests/test_dakota_base.py
import os import filecmp from nose.tools import * from dakota.dakota_base import DakotaBase # Fixtures ------------------------------------------------------------- def setup_module(): """Called before any tests are performed.""" print('\n*** DakotaBase tests') def teardown_module(): """Called after all tests have completed.""" pass # Tests ---------------------------------------------------------------- @raises(TypeError) def test_instantiate(): """Test whether DakotaBase fails to instantiate.""" d = DakotaBase()
from nose.tools import * from dakota.dakota_base import DakotaBase # Helpers -------------------------------------------------------------- class Concrete(DakotaBase): """A subclass of DakotaBase used for testing.""" def __init__(self): DakotaBase.__init__(self) # Fixtures ------------------------------------------------------------- def setup_module(): """Called before any tests are performed.""" print('\n*** DakotaBase tests') global c c = Concrete() def teardown_module(): """Called after all tests have completed.""" pass # Tests ---------------------------------------------------------------- @raises(TypeError) def test_instantiate(): """Test whether DakotaBase fails to instantiate.""" d = DakotaBase() def test_environment_block(): """Test type of environment_block method results.""" s = c.environment_block() assert_true(type(s) is str) def test_method_block(): """Test type of method_block method results.""" s = c.method_block() assert_true(type(s) is str) def test_variables_block(): """Test type of variables_block method results.""" s = c.variables_block() assert_true(type(s) is str) def test_interface_block(): """Test type of interface_block method results.""" s = c.interface_block() assert_true(type(s) is str) def test_responses_block(): """Test type of responses_block method results.""" s = c.responses_block() assert_true(type(s) is str) def test_autogenerate_descriptors(): """Test autogenerate_descriptors method.""" c.n_variables, c.n_responses = 1, 1 c.autogenerate_descriptors() assert_true(len(c.variable_descriptors) == 1) assert_true(len(c.response_descriptors) == 1)
Add tests for dakota.dakota_base module
Add tests for dakota.dakota_base module Make a subclass of DakotaBase to use for testing. Add tests for the "block" sections used to define an input file.
Python
mit
csdms/dakota,csdms/dakota
- import os - import filecmp from nose.tools import * from dakota.dakota_base import DakotaBase + # Helpers -------------------------------------------------------------- + + class Concrete(DakotaBase): + """A subclass of DakotaBase used for testing.""" + def __init__(self): + DakotaBase.__init__(self) # Fixtures ------------------------------------------------------------- def setup_module(): """Called before any tests are performed.""" print('\n*** DakotaBase tests') + global c + c = Concrete() def teardown_module(): """Called after all tests have completed.""" pass # Tests ---------------------------------------------------------------- @raises(TypeError) def test_instantiate(): """Test whether DakotaBase fails to instantiate.""" d = DakotaBase() + def test_environment_block(): + """Test type of environment_block method results.""" + s = c.environment_block() + assert_true(type(s) is str) + + def test_method_block(): + """Test type of method_block method results.""" + s = c.method_block() + assert_true(type(s) is str) + + def test_variables_block(): + """Test type of variables_block method results.""" + s = c.variables_block() + assert_true(type(s) is str) + + def test_interface_block(): + """Test type of interface_block method results.""" + s = c.interface_block() + assert_true(type(s) is str) + + def test_responses_block(): + """Test type of responses_block method results.""" + s = c.responses_block() + assert_true(type(s) is str) + + def test_autogenerate_descriptors(): + """Test autogenerate_descriptors method.""" + c.n_variables, c.n_responses = 1, 1 + c.autogenerate_descriptors() + assert_true(len(c.variable_descriptors) == 1) + assert_true(len(c.response_descriptors) == 1) +
Add tests for dakota.dakota_base module
## Code Before: import os import filecmp from nose.tools import * from dakota.dakota_base import DakotaBase # Fixtures ------------------------------------------------------------- def setup_module(): """Called before any tests are performed.""" print('\n*** DakotaBase tests') def teardown_module(): """Called after all tests have completed.""" pass # Tests ---------------------------------------------------------------- @raises(TypeError) def test_instantiate(): """Test whether DakotaBase fails to instantiate.""" d = DakotaBase() ## Instruction: Add tests for dakota.dakota_base module ## Code After: from nose.tools import * from dakota.dakota_base import DakotaBase # Helpers -------------------------------------------------------------- class Concrete(DakotaBase): """A subclass of DakotaBase used for testing.""" def __init__(self): DakotaBase.__init__(self) # Fixtures ------------------------------------------------------------- def setup_module(): """Called before any tests are performed.""" print('\n*** DakotaBase tests') global c c = Concrete() def teardown_module(): """Called after all tests have completed.""" pass # Tests ---------------------------------------------------------------- @raises(TypeError) def test_instantiate(): """Test whether DakotaBase fails to instantiate.""" d = DakotaBase() def test_environment_block(): """Test type of environment_block method results.""" s = c.environment_block() assert_true(type(s) is str) def test_method_block(): """Test type of method_block method results.""" s = c.method_block() assert_true(type(s) is str) def test_variables_block(): """Test type of variables_block method results.""" s = c.variables_block() assert_true(type(s) is str) def test_interface_block(): """Test type of interface_block method results.""" s = c.interface_block() assert_true(type(s) is str) def test_responses_block(): """Test type of responses_block method results.""" s = c.responses_block() assert_true(type(s) is str) def test_autogenerate_descriptors(): """Test autogenerate_descriptors method.""" c.n_variables, c.n_responses = 1, 1 c.autogenerate_descriptors() assert_true(len(c.variable_descriptors) == 1) assert_true(len(c.response_descriptors) == 1)
- import os - import filecmp from nose.tools import * from dakota.dakota_base import DakotaBase + # Helpers -------------------------------------------------------------- + + class Concrete(DakotaBase): + """A subclass of DakotaBase used for testing.""" + def __init__(self): + DakotaBase.__init__(self) # Fixtures ------------------------------------------------------------- def setup_module(): """Called before any tests are performed.""" print('\n*** DakotaBase tests') + global c + c = Concrete() def teardown_module(): """Called after all tests have completed.""" pass # Tests ---------------------------------------------------------------- @raises(TypeError) def test_instantiate(): """Test whether DakotaBase fails to instantiate.""" d = DakotaBase() + + def test_environment_block(): + """Test type of environment_block method results.""" + s = c.environment_block() + assert_true(type(s) is str) + + def test_method_block(): + """Test type of method_block method results.""" + s = c.method_block() + assert_true(type(s) is str) + + def test_variables_block(): + """Test type of variables_block method results.""" + s = c.variables_block() + assert_true(type(s) is str) + + def test_interface_block(): + """Test type of interface_block method results.""" + s = c.interface_block() + assert_true(type(s) is str) + + def test_responses_block(): + """Test type of responses_block method results.""" + s = c.responses_block() + assert_true(type(s) is str) + + def test_autogenerate_descriptors(): + """Test autogenerate_descriptors method.""" + c.n_variables, c.n_responses = 1, 1 + c.autogenerate_descriptors() + assert_true(len(c.variable_descriptors) == 1) + assert_true(len(c.response_descriptors) == 1)
53646da453a4aa6d0e559ee3069626458f2fef78
common/urls.py
common/urls.py
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): return re.sub("\\(.+\\)", "{id}", pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): _pattern = pattern.replace('^', '').replace('$', '') return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
Fix backend home page url generator
Fix backend home page url generator
Python
mit
DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): + _pattern = pattern.replace('^', '').replace('$', '') - return re.sub("\\(.+\\)", "{id}", pattern) + return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True) +
Fix backend home page url generator
## Code Before: import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): return re.sub("\\(.+\\)", "{id}", pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True) ## Instruction: Fix backend home page url generator ## Code After: import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): _pattern = pattern.replace('^', '').replace('$', '') return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): + _pattern = pattern.replace('^', '').replace('$', '') - return re.sub("\\(.+\\)", "{id}", pattern) + return re.sub("\\(.+\\)", "{id}", _pattern) ? + def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
26b1845419cd7ea55cf1f40f26812eb9c124299f
kolibri/core/content/signals.py
kolibri/core/content/signals.py
from django.db.models import F from django.db.models.signals import pre_delete from django.dispatch import receiver from .models import ChannelMetadata from .models import ContentNode from kolibri.core.notifications.models import LearnerProgressNotification @receiver(pre_delete, sender=ContentNode) def cascade_delete_node(sender, instance=None, *args, **kwargs): """ For a given node, we delete all notifications objects whose contentnode is the instance's node.. """ LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete() @receiver(pre_delete, sender=ChannelMetadata) def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs): """ For a given channel, decrement the order of all channels that come after this channel. """ ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1)
from django.db.models import F from django.db.models.signals import pre_delete from django.dispatch import receiver from .models import ChannelMetadata from .models import ContentNode from kolibri.core.notifications.models import LearnerProgressNotification from kolibri.core.lessons.models import Lesson @receiver(pre_delete, sender=ContentNode) def cascade_delete_node(sender, instance=None, *args, **kwargs): """ For a given node, we delete all notifications objects whose contentnode is the instance's node.. """ LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete() @receiver(pre_delete, sender=ChannelMetadata) def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs): """ For a given channel, decrement the order of all channels that come after this channel. """ ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1) @receiver(pre_delete, sender=ChannelMetadata) def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs): # Update the resources array of all lessons to ensure they don't have # any deleted content lessons = Lesson.objects.all() for lesson in lessons: updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id] if len(updated_resources) < len(lesson.resources): lesson.resources = updated_resources lesson.save()
Add Channel deletion side-effect that updates affected lessons
Add Channel deletion side-effect that updates affected lessons
Python
mit
learningequality/kolibri,lyw07/kolibri,mrpau/kolibri,learningequality/kolibri,indirectlylit/kolibri,lyw07/kolibri,mrpau/kolibri,mrpau/kolibri,indirectlylit/kolibri,indirectlylit/kolibri,learningequality/kolibri,indirectlylit/kolibri,lyw07/kolibri,learningequality/kolibri,lyw07/kolibri,mrpau/kolibri
from django.db.models import F from django.db.models.signals import pre_delete from django.dispatch import receiver from .models import ChannelMetadata from .models import ContentNode from kolibri.core.notifications.models import LearnerProgressNotification + from kolibri.core.lessons.models import Lesson @receiver(pre_delete, sender=ContentNode) def cascade_delete_node(sender, instance=None, *args, **kwargs): """ For a given node, we delete all notifications objects whose contentnode is the instance's node.. """ LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete() @receiver(pre_delete, sender=ChannelMetadata) def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs): """ For a given channel, decrement the order of all channels that come after this channel. """ ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1) + + @receiver(pre_delete, sender=ChannelMetadata) + def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs): + # Update the resources array of all lessons to ensure they don't have + # any deleted content + lessons = Lesson.objects.all() + for lesson in lessons: + updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id] + if len(updated_resources) < len(lesson.resources): + lesson.resources = updated_resources + lesson.save() +
Add Channel deletion side-effect that updates affected lessons
## Code Before: from django.db.models import F from django.db.models.signals import pre_delete from django.dispatch import receiver from .models import ChannelMetadata from .models import ContentNode from kolibri.core.notifications.models import LearnerProgressNotification @receiver(pre_delete, sender=ContentNode) def cascade_delete_node(sender, instance=None, *args, **kwargs): """ For a given node, we delete all notifications objects whose contentnode is the instance's node.. """ LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete() @receiver(pre_delete, sender=ChannelMetadata) def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs): """ For a given channel, decrement the order of all channels that come after this channel. """ ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1) ## Instruction: Add Channel deletion side-effect that updates affected lessons ## Code After: from django.db.models import F from django.db.models.signals import pre_delete from django.dispatch import receiver from .models import ChannelMetadata from .models import ContentNode from kolibri.core.notifications.models import LearnerProgressNotification from kolibri.core.lessons.models import Lesson @receiver(pre_delete, sender=ContentNode) def cascade_delete_node(sender, instance=None, *args, **kwargs): """ For a given node, we delete all notifications objects whose contentnode is the instance's node.. """ LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete() @receiver(pre_delete, sender=ChannelMetadata) def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs): """ For a given channel, decrement the order of all channels that come after this channel. """ ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1) @receiver(pre_delete, sender=ChannelMetadata) def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs): # Update the resources array of all lessons to ensure they don't have # any deleted content lessons = Lesson.objects.all() for lesson in lessons: updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id] if len(updated_resources) < len(lesson.resources): lesson.resources = updated_resources lesson.save()
from django.db.models import F from django.db.models.signals import pre_delete from django.dispatch import receiver from .models import ChannelMetadata from .models import ContentNode from kolibri.core.notifications.models import LearnerProgressNotification + from kolibri.core.lessons.models import Lesson @receiver(pre_delete, sender=ContentNode) def cascade_delete_node(sender, instance=None, *args, **kwargs): """ For a given node, we delete all notifications objects whose contentnode is the instance's node.. """ LearnerProgressNotification.objects.filter(contentnode_id=instance.id).delete() @receiver(pre_delete, sender=ChannelMetadata) def reorder_channels_upon_deletion(sender, instance=None, *args, **kwargs): """ For a given channel, decrement the order of all channels that come after this channel. """ ChannelMetadata.objects.filter(order__gt=instance.order).update(order=F('order') - 1) + + + @receiver(pre_delete, sender=ChannelMetadata) + def update_lesson_resources_before_delete(sender, instance=None, *args, **kwargs): + # Update the resources array of all lessons to ensure they don't have + # any deleted content + lessons = Lesson.objects.all() + for lesson in lessons: + updated_resources = [r for r in lesson.resources if r['channel_id'] != instance.id] + if len(updated_resources) < len(lesson.resources): + lesson.resources = updated_resources + lesson.save()
1a871cf3bf1fd40342e490599361d57017cdcc65
backend/breach/tests/test_strategy.py
backend/breach/tests/test_strategy.py
from mock import patch from breach.tests.base import RuptureTestCase from breach.strategy import Strategy class StrategyTestCase(RuptureTestCase): @patch('breach.strategy.Sniffer') def test_first_round(self, Sniffer): strategy0 = Strategy(self.victim) work0 = strategy0.get_work() self.assertEqual( work0['url'], 'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^' ) self.assertTrue('amount' in work0) self.assertTrue('timeout' in work0) strategy1 = Strategy(self.victim) work1 = strategy1.get_work() self.assertEqual( work1['url'], 'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^' ) def test_same_round_same_batch(self): pass def test_same_round_different_batch(self): pass def test_advance_round(self): pass
from mock import patch from breach.tests.base import RuptureTestCase from breach.strategy import Strategy class StrategyTestCase(RuptureTestCase): @patch('breach.strategy.Sniffer') def test_first_round(self, Sniffer): strategy0 = Strategy(self.victim) work0 = strategy0.get_work() self.assertEqual( work0['url'], 'https://di.uoa.gr/?breach=^testsecret0^1^' ) self.assertTrue('amount' in work0) self.assertTrue('timeout' in work0) strategy1 = Strategy(self.victim) work1 = strategy1.get_work() self.assertEqual( work1['url'], 'https://di.uoa.gr/?breach=^testsecret1^0^' ) def test_same_round_same_batch(self): pass def test_same_round_different_batch(self): pass def test_advance_round(self): pass
Update first round test, create huffman based on knownalphabet
Update first round test, create huffman based on knownalphabet
Python
mit
dionyziz/rupture,dimriou/rupture,esarafianou/rupture,esarafianou/rupture,dimkarakostas/rupture,dimkarakostas/rupture,esarafianou/rupture,dionyziz/rupture,dimkarakostas/rupture,dionyziz/rupture,esarafianou/rupture,dimkarakostas/rupture,dimkarakostas/rupture,dimriou/rupture,dimriou/rupture,dimriou/rupture,dimriou/rupture,dionyziz/rupture,dionyziz/rupture
from mock import patch from breach.tests.base import RuptureTestCase from breach.strategy import Strategy class StrategyTestCase(RuptureTestCase): @patch('breach.strategy.Sniffer') def test_first_round(self, Sniffer): strategy0 = Strategy(self.victim) work0 = strategy0.get_work() self.assertEqual( work0['url'], - 'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^' + 'https://di.uoa.gr/?breach=^testsecret0^1^' ) self.assertTrue('amount' in work0) self.assertTrue('timeout' in work0) strategy1 = Strategy(self.victim) work1 = strategy1.get_work() self.assertEqual( work1['url'], - 'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^' + 'https://di.uoa.gr/?breach=^testsecret1^0^' ) def test_same_round_same_batch(self): pass def test_same_round_different_batch(self): pass def test_advance_round(self): pass
Update first round test, create huffman based on knownalphabet
## Code Before: from mock import patch from breach.tests.base import RuptureTestCase from breach.strategy import Strategy class StrategyTestCase(RuptureTestCase): @patch('breach.strategy.Sniffer') def test_first_round(self, Sniffer): strategy0 = Strategy(self.victim) work0 = strategy0.get_work() self.assertEqual( work0['url'], 'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^' ) self.assertTrue('amount' in work0) self.assertTrue('timeout' in work0) strategy1 = Strategy(self.victim) work1 = strategy1.get_work() self.assertEqual( work1['url'], 'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^' ) def test_same_round_same_batch(self): pass def test_same_round_different_batch(self): pass def test_advance_round(self): pass ## Instruction: Update first round test, create huffman based on knownalphabet ## Code After: from mock import patch from breach.tests.base import RuptureTestCase from breach.strategy import Strategy class StrategyTestCase(RuptureTestCase): @patch('breach.strategy.Sniffer') def test_first_round(self, Sniffer): strategy0 = Strategy(self.victim) work0 = strategy0.get_work() self.assertEqual( work0['url'], 'https://di.uoa.gr/?breach=^testsecret0^1^' ) self.assertTrue('amount' in work0) self.assertTrue('timeout' in work0) strategy1 = Strategy(self.victim) work1 = strategy1.get_work() self.assertEqual( work1['url'], 'https://di.uoa.gr/?breach=^testsecret1^0^' ) def test_same_round_same_batch(self): pass def test_same_round_different_batch(self): pass def test_advance_round(self): pass
from mock import patch from breach.tests.base import RuptureTestCase from breach.strategy import Strategy class StrategyTestCase(RuptureTestCase): @patch('breach.strategy.Sniffer') def test_first_round(self, Sniffer): strategy0 = Strategy(self.victim) work0 = strategy0.get_work() self.assertEqual( work0['url'], - 'https://di.uoa.gr/?breach=^testsecret0^1^3^2^5^4^7^6^9^8^' ? ---------------- + 'https://di.uoa.gr/?breach=^testsecret0^1^' ) self.assertTrue('amount' in work0) self.assertTrue('timeout' in work0) strategy1 = Strategy(self.victim) work1 = strategy1.get_work() self.assertEqual( work1['url'], - 'https://di.uoa.gr/?breach=^testsecret1^0^3^2^5^4^7^6^9^8^' ? ---------------- + 'https://di.uoa.gr/?breach=^testsecret1^0^' ) def test_same_round_same_batch(self): pass def test_same_round_different_batch(self): pass def test_advance_round(self): pass
bea258e2affc165f610de83248d9f958eec1ef4e
cmsplugin_markdown/models.py
cmsplugin_markdown/models.py
from django.db import models from cms.models import CMSPlugin class MarkdownPlugin(CMSPlugin): markdown_text = models.TextField(max_length=8000)
from django.db import models from cms.models import CMSPlugin from cms.utils.compat.dj import python_2_unicode_compatible @python_2_unicode_compatible class MarkdownPlugin(CMSPlugin): markdown_text = models.TextField(max_length=8000) def __str__(self): text = self.markdown_text return (text[:50] + '...') if len(text) > 53 else text
Add __str__ method for better representation in frontend
Add __str__ method for better representation in frontend
Python
mit
bitmazk/cmsplugin-markdown,bitmazk/cmsplugin-markdown,bitmazk/cmsplugin-markdown
from django.db import models from cms.models import CMSPlugin + from cms.utils.compat.dj import python_2_unicode_compatible + @python_2_unicode_compatible class MarkdownPlugin(CMSPlugin): markdown_text = models.TextField(max_length=8000) + def __str__(self): + text = self.markdown_text + return (text[:50] + '...') if len(text) > 53 else text +
Add __str__ method for better representation in frontend
## Code Before: from django.db import models from cms.models import CMSPlugin class MarkdownPlugin(CMSPlugin): markdown_text = models.TextField(max_length=8000) ## Instruction: Add __str__ method for better representation in frontend ## Code After: from django.db import models from cms.models import CMSPlugin from cms.utils.compat.dj import python_2_unicode_compatible @python_2_unicode_compatible class MarkdownPlugin(CMSPlugin): markdown_text = models.TextField(max_length=8000) def __str__(self): text = self.markdown_text return (text[:50] + '...') if len(text) > 53 else text
from django.db import models from cms.models import CMSPlugin + from cms.utils.compat.dj import python_2_unicode_compatible + @python_2_unicode_compatible class MarkdownPlugin(CMSPlugin): markdown_text = models.TextField(max_length=8000) + + def __str__(self): + text = self.markdown_text + return (text[:50] + '...') if len(text) > 53 else text
208760340d3314f666d7e6437817cc96e0e16194
organizer/urls/tag.py
organizer/urls/tag.py
from django.conf.urls import url from ..views import ( TagCreate, TagDelete, TagDetail, TagList, TagUpdate) urlpatterns = [ url(r'^$', TagList.as_view(), name='organizer_tag_list'), url(r'^create/$', TagCreate.as_view(), name='organizer_tag_create'), url(r'^(?P<slug>[\w\-]+)/$', TagDetail.as_view(), name='organizer_tag_detail'), url(r'^(?P<slug>[\w-]+)/delete/$', TagDelete.as_view(), name='organizer_tag_delete'), url(r'^(?P<slug>[\w\-]+)/update/$', TagUpdate.as_view(), name='organizer_tag_update'), ]
from django.conf.urls import url from django.contrib.auth.decorators import \ login_required from ..views import ( TagCreate, TagDelete, TagDetail, TagList, TagUpdate) urlpatterns = [ url(r'^$', TagList.as_view(), name='organizer_tag_list'), url(r'^create/$', login_required( TagCreate.as_view()), name='organizer_tag_create'), url(r'^(?P<slug>[\w\-]+)/$', TagDetail.as_view(), name='organizer_tag_detail'), url(r'^(?P<slug>[\w-]+)/delete/$', TagDelete.as_view(), name='organizer_tag_delete'), url(r'^(?P<slug>[\w\-]+)/update/$', TagUpdate.as_view(), name='organizer_tag_update'), ]
Use login_required decorator in URL pattern.
Ch20: Use login_required decorator in URL pattern.
Python
bsd-2-clause
jambonrose/DjangoUnleashed-1.8,jambonrose/DjangoUnleashed-1.8
from django.conf.urls import url + from django.contrib.auth.decorators import \ + login_required from ..views import ( TagCreate, TagDelete, TagDetail, TagList, TagUpdate) urlpatterns = [ url(r'^$', TagList.as_view(), name='organizer_tag_list'), url(r'^create/$', + login_required( - TagCreate.as_view(), + TagCreate.as_view()), name='organizer_tag_create'), url(r'^(?P<slug>[\w\-]+)/$', TagDetail.as_view(), name='organizer_tag_detail'), url(r'^(?P<slug>[\w-]+)/delete/$', TagDelete.as_view(), name='organizer_tag_delete'), url(r'^(?P<slug>[\w\-]+)/update/$', TagUpdate.as_view(), name='organizer_tag_update'), ]
Use login_required decorator in URL pattern.
## Code Before: from django.conf.urls import url from ..views import ( TagCreate, TagDelete, TagDetail, TagList, TagUpdate) urlpatterns = [ url(r'^$', TagList.as_view(), name='organizer_tag_list'), url(r'^create/$', TagCreate.as_view(), name='organizer_tag_create'), url(r'^(?P<slug>[\w\-]+)/$', TagDetail.as_view(), name='organizer_tag_detail'), url(r'^(?P<slug>[\w-]+)/delete/$', TagDelete.as_view(), name='organizer_tag_delete'), url(r'^(?P<slug>[\w\-]+)/update/$', TagUpdate.as_view(), name='organizer_tag_update'), ] ## Instruction: Use login_required decorator in URL pattern. ## Code After: from django.conf.urls import url from django.contrib.auth.decorators import \ login_required from ..views import ( TagCreate, TagDelete, TagDetail, TagList, TagUpdate) urlpatterns = [ url(r'^$', TagList.as_view(), name='organizer_tag_list'), url(r'^create/$', login_required( TagCreate.as_view()), name='organizer_tag_create'), url(r'^(?P<slug>[\w\-]+)/$', TagDetail.as_view(), name='organizer_tag_detail'), url(r'^(?P<slug>[\w-]+)/delete/$', TagDelete.as_view(), name='organizer_tag_delete'), url(r'^(?P<slug>[\w\-]+)/update/$', TagUpdate.as_view(), name='organizer_tag_update'), ]
from django.conf.urls import url + from django.contrib.auth.decorators import \ + login_required from ..views import ( TagCreate, TagDelete, TagDetail, TagList, TagUpdate) urlpatterns = [ url(r'^$', TagList.as_view(), name='organizer_tag_list'), url(r'^create/$', + login_required( - TagCreate.as_view(), + TagCreate.as_view()), ? ++++ + name='organizer_tag_create'), url(r'^(?P<slug>[\w\-]+)/$', TagDetail.as_view(), name='organizer_tag_detail'), url(r'^(?P<slug>[\w-]+)/delete/$', TagDelete.as_view(), name='organizer_tag_delete'), url(r'^(?P<slug>[\w\-]+)/update/$', TagUpdate.as_view(), name='organizer_tag_update'), ]
7e0024878352ba544a8b40d2c8b5741aedf05a70
Code/login_proxy.py
Code/login_proxy.py
import sys import datetime from colors import farben def request(flow): now = datetime.datetime.now() content = flow.request.get_text() host = flow.request.pretty_host method = flow.request.method if method == "POST" and ("pass" in content) or ("password" in content) : with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: myfile.write(farben.AUF + str(now) +" // " + farben.END) myfile.write(farben.LD + host + farben.END) myfile.write("\n") myfile.write(farben.IN + content + farben.END) myfile.write("\n") myfile.write("\n")
import re import sys import datetime from colors import farben def request(flow): now = datetime.datetime.now() content = flow.request.get_text() host = flow.request.pretty_host method = flow.request.method if method == "POST" and ("pass" in content) or ("password" in content): with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: myfile.write(farben.AUF + str(now) + " // " + farben.END) myfile.write(farben.LD + host + farben.END) myfile.write("\n") passwords = re.findall(r"(?:pass|password)=([^&]*)", content) myfile.write(farben.IN + passwords[0] + farben.END) myfile.write("\n") myfile.write("\n")
Add regex for parsing password url parameter
Add regex for parsing password url parameter
Python
apache-2.0
sarah314/SpyPi
+ import re import sys import datetime from colors import farben + def request(flow): now = datetime.datetime.now() content = flow.request.get_text() host = flow.request.pretty_host method = flow.request.method - if method == "POST" and ("pass" in content) or ("password" in content) : + if method == "POST" and ("pass" in content) or ("password" in content): - with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: + with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: - myfile.write(farben.AUF + str(now) +" // " + farben.END) + myfile.write(farben.AUF + str(now) + " // " + farben.END) myfile.write(farben.LD + host + farben.END) myfile.write("\n") + passwords = re.findall(r"(?:pass|password)=([^&]*)", content) - myfile.write(farben.IN + content + farben.END) + myfile.write(farben.IN + passwords[0] + farben.END) myfile.write("\n") myfile.write("\n")
Add regex for parsing password url parameter
## Code Before: import sys import datetime from colors import farben def request(flow): now = datetime.datetime.now() content = flow.request.get_text() host = flow.request.pretty_host method = flow.request.method if method == "POST" and ("pass" in content) or ("password" in content) : with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: myfile.write(farben.AUF + str(now) +" // " + farben.END) myfile.write(farben.LD + host + farben.END) myfile.write("\n") myfile.write(farben.IN + content + farben.END) myfile.write("\n") myfile.write("\n") ## Instruction: Add regex for parsing password url parameter ## Code After: import re import sys import datetime from colors import farben def request(flow): now = datetime.datetime.now() content = flow.request.get_text() host = flow.request.pretty_host method = flow.request.method if method == "POST" and ("pass" in content) or ("password" in content): with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: myfile.write(farben.AUF + str(now) + " // " + farben.END) myfile.write(farben.LD + host + farben.END) myfile.write("\n") passwords = re.findall(r"(?:pass|password)=([^&]*)", content) myfile.write(farben.IN + passwords[0] + farben.END) myfile.write("\n") myfile.write("\n")
+ import re import sys import datetime from colors import farben + def request(flow): now = datetime.datetime.now() content = flow.request.get_text() host = flow.request.pretty_host method = flow.request.method - if method == "POST" and ("pass" in content) or ("password" in content) : ? - + if method == "POST" and ("pass" in content) or ("password" in content): - with open ("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: ? - + with open("/home/pi/SpyPi/Code/proxy.txt", "a") as myfile: - myfile.write(farben.AUF + str(now) +" // " + farben.END) + myfile.write(farben.AUF + str(now) + " // " + farben.END) ? + myfile.write(farben.LD + host + farben.END) myfile.write("\n") + passwords = re.findall(r"(?:pass|password)=([^&]*)", content) - myfile.write(farben.IN + content + farben.END) ? ^ ^^^^^ + myfile.write(farben.IN + passwords[0] + farben.END) ? ^^^^^ ^^^^^^ myfile.write("\n") myfile.write("\n")
1a0339b85d852526c184eeace73021fc7d68b2c6
python_dispatcher.py
python_dispatcher.py
import traceback from routes import Mapper import ppp_core import example_ppp_module as flower import ppp_questionparsing_grammatical as qp_grammatical import ppp_cas #import ppp_nlp_ml_standalone class Application: def __init__(self): self.mapper = Mapper() self.mapper.connect('core', '/core/', app=ppp_core.app) self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app) self.mapper.connect('flower', '/flower/', app=flower.app) self.mapper.connect('cas', '/cas/', app=ppp_cas.app) self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app) #self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app) def __call__(self, environ, start_response): match = self.mapper.routematch(environ=environ) app = match[0]['app'] if match else self.not_found try: return app(environ, start_response) except KeyboardInterrupt: raise except Exception as e: traceback.print_exc(e) def not_found(self, environ, start_response): headers = [('Content-Type', 'text/plain')] start_response('404 Not Found', headers) return [b'Not found.'] app = Application()
import traceback from routes import Mapper import ppp_core import example_ppp_module as flower import ppp_questionparsing_grammatical as qp_grammatical import ppp_cas import ppp_spell_checker #import ppp_nlp_ml_standalone class Application: def __init__(self): self.mapper = Mapper() self.mapper.connect('core', '/core/', app=ppp_core.app) self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app) self.mapper.connect('flower', '/flower/', app=flower.app) self.mapper.connect('cas', '/cas/', app=ppp_cas.app) self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app) #self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app) def __call__(self, environ, start_response): match = self.mapper.routematch(environ=environ) app = match[0]['app'] if match else self.not_found try: return app(environ, start_response) except KeyboardInterrupt: raise except Exception as e: traceback.print_exc(e) def not_found(self, environ, start_response): headers = [('Content-Type', 'text/plain')] start_response('404 Not Found', headers) return [b'Not found.'] app = Application()
Fix name of spell checker.
Fix name of spell checker.
Python
cc0-1.0
ProjetPP/Deployment,ProjetPP/Deployment,ProjetPP/Deployment
import traceback from routes import Mapper import ppp_core import example_ppp_module as flower import ppp_questionparsing_grammatical as qp_grammatical import ppp_cas + import ppp_spell_checker #import ppp_nlp_ml_standalone class Application: def __init__(self): self.mapper = Mapper() self.mapper.connect('core', '/core/', app=ppp_core.app) self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app) self.mapper.connect('flower', '/flower/', app=flower.app) self.mapper.connect('cas', '/cas/', app=ppp_cas.app) - self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app) + self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app) #self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app) def __call__(self, environ, start_response): match = self.mapper.routematch(environ=environ) app = match[0]['app'] if match else self.not_found try: return app(environ, start_response) except KeyboardInterrupt: raise except Exception as e: traceback.print_exc(e) def not_found(self, environ, start_response): headers = [('Content-Type', 'text/plain')] start_response('404 Not Found', headers) return [b'Not found.'] app = Application()
Fix name of spell checker.
## Code Before: import traceback from routes import Mapper import ppp_core import example_ppp_module as flower import ppp_questionparsing_grammatical as qp_grammatical import ppp_cas #import ppp_nlp_ml_standalone class Application: def __init__(self): self.mapper = Mapper() self.mapper.connect('core', '/core/', app=ppp_core.app) self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app) self.mapper.connect('flower', '/flower/', app=flower.app) self.mapper.connect('cas', '/cas/', app=ppp_cas.app) self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app) #self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app) def __call__(self, environ, start_response): match = self.mapper.routematch(environ=environ) app = match[0]['app'] if match else self.not_found try: return app(environ, start_response) except KeyboardInterrupt: raise except Exception as e: traceback.print_exc(e) def not_found(self, environ, start_response): headers = [('Content-Type', 'text/plain')] start_response('404 Not Found', headers) return [b'Not found.'] app = Application() ## Instruction: Fix name of spell checker. ## Code After: import traceback from routes import Mapper import ppp_core import example_ppp_module as flower import ppp_questionparsing_grammatical as qp_grammatical import ppp_cas import ppp_spell_checker #import ppp_nlp_ml_standalone class Application: def __init__(self): self.mapper = Mapper() self.mapper.connect('core', '/core/', app=ppp_core.app) self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app) self.mapper.connect('flower', '/flower/', app=flower.app) self.mapper.connect('cas', '/cas/', app=ppp_cas.app) self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app) #self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app) def __call__(self, environ, start_response): match = self.mapper.routematch(environ=environ) app = match[0]['app'] if match else self.not_found try: return app(environ, start_response) except KeyboardInterrupt: raise except Exception as e: traceback.print_exc(e) def not_found(self, environ, start_response): headers = [('Content-Type', 'text/plain')] start_response('404 Not Found', headers) return [b'Not found.'] app = Application()
import traceback from routes import Mapper import ppp_core import example_ppp_module as flower import ppp_questionparsing_grammatical as qp_grammatical import ppp_cas + import ppp_spell_checker #import ppp_nlp_ml_standalone class Application: def __init__(self): self.mapper = Mapper() self.mapper.connect('core', '/core/', app=ppp_core.app) self.mapper.connect('qp_grammatical', '/qp_grammatical/', app=qp_grammatical.app) self.mapper.connect('flower', '/flower/', app=flower.app) self.mapper.connect('cas', '/cas/', app=ppp_cas.app) - self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_cas.app) ? ^^ + self.mapper.connect('spellcheck', '/spell_checker/', app=ppp_spell_checker.app) ? ++++++ ^^^^^^ #self.mapper.connect('nlp_ml_standalone', '/nlp_ml_standalone/', app=ppp_nlp_ml_standalone.app) def __call__(self, environ, start_response): match = self.mapper.routematch(environ=environ) app = match[0]['app'] if match else self.not_found try: return app(environ, start_response) except KeyboardInterrupt: raise except Exception as e: traceback.print_exc(e) def not_found(self, environ, start_response): headers = [('Content-Type', 'text/plain')] start_response('404 Not Found', headers) return [b'Not found.'] app = Application()
1bf6211f2fd5aef99e529fdc0e714b1a36ace346
gallery/util.py
gallery/util.py
import os from addict import Dict from gallery.models import File def get_dir_file_contents(dir_id): print(File.query.filter(File.parent == dir_id).all()) return File.query.filter(File.parent == dir_id).all() def get_dir_tree_dict(): path = os.path.normpath("/gallery-data/root") file_tree = Dict() for root, _, files in os.walk(path, topdown=True): path = root.split('/') path.pop(0) file_tree_fd = file_tree for part in path: file_tree_fd = file_tree_fd[part] file_tree_fd['.'] = files return file_tree def convert_bytes_to_utf8(dic): for key in dic: if isinstance(key, bytes): k = key.decode('utf-8') v = dic[key] del dic[key] dic[k] = v if isinstance(dic[key], bytes): v = dic[key].decode('utf-8') dic[key] = v return dic def allowed_file(filename): return '.' in filename and filename.lower().rsplit('.', 1)[1] in \ [ 'txt', 'png', 'jpg', 'jpeg', 'mpg', 'mp4', 'avi' ]
import os from addict import Dict from gallery.models import File def get_dir_file_contents(dir_id): print(File.query.filter(File.parent == dir_id).all()) return File.query.filter(File.parent == dir_id).all() def get_dir_tree_dict(): path = os.path.normpath("/gallery-data/root") file_tree = Dict() for root, _, files in os.walk(path, topdown=True): path = root.split('/') path.pop(0) file_tree_fd = file_tree for part in path: file_tree_fd = file_tree_fd[part] file_tree_fd['.'] = files return file_tree def convert_bytes_to_utf8(dic): for key in dic: if isinstance(key, bytes): k = key.decode('utf-8') v = dic[key] del dic[key] dic[k] = v if isinstance(dic[key], bytes): v = dic[key].decode('utf-8') dic[key] = v return dic def allowed_file(filename): return '.' in filename and filename.lower().rsplit('.', 1)[1] in \ [ 'txt', 'png', 'jpg', 'jpeg', 'mpg', 'mp4', 'avi', 'cr2' ]
Add CR2 to allowed files
Add CR2 to allowed files
Python
mit
liam-middlebrook/gallery,liam-middlebrook/gallery,liam-middlebrook/gallery,liam-middlebrook/gallery
import os from addict import Dict from gallery.models import File def get_dir_file_contents(dir_id): print(File.query.filter(File.parent == dir_id).all()) return File.query.filter(File.parent == dir_id).all() def get_dir_tree_dict(): path = os.path.normpath("/gallery-data/root") file_tree = Dict() for root, _, files in os.walk(path, topdown=True): path = root.split('/') path.pop(0) file_tree_fd = file_tree for part in path: file_tree_fd = file_tree_fd[part] file_tree_fd['.'] = files return file_tree def convert_bytes_to_utf8(dic): for key in dic: if isinstance(key, bytes): k = key.decode('utf-8') v = dic[key] del dic[key] dic[k] = v if isinstance(dic[key], bytes): v = dic[key].decode('utf-8') dic[key] = v return dic def allowed_file(filename): return '.' in filename and filename.lower().rsplit('.', 1)[1] in \ [ 'txt', 'png', 'jpg', 'jpeg', 'mpg', 'mp4', - 'avi' + 'avi', + 'cr2' ]
Add CR2 to allowed files
## Code Before: import os from addict import Dict from gallery.models import File def get_dir_file_contents(dir_id): print(File.query.filter(File.parent == dir_id).all()) return File.query.filter(File.parent == dir_id).all() def get_dir_tree_dict(): path = os.path.normpath("/gallery-data/root") file_tree = Dict() for root, _, files in os.walk(path, topdown=True): path = root.split('/') path.pop(0) file_tree_fd = file_tree for part in path: file_tree_fd = file_tree_fd[part] file_tree_fd['.'] = files return file_tree def convert_bytes_to_utf8(dic): for key in dic: if isinstance(key, bytes): k = key.decode('utf-8') v = dic[key] del dic[key] dic[k] = v if isinstance(dic[key], bytes): v = dic[key].decode('utf-8') dic[key] = v return dic def allowed_file(filename): return '.' in filename and filename.lower().rsplit('.', 1)[1] in \ [ 'txt', 'png', 'jpg', 'jpeg', 'mpg', 'mp4', 'avi' ] ## Instruction: Add CR2 to allowed files ## Code After: import os from addict import Dict from gallery.models import File def get_dir_file_contents(dir_id): print(File.query.filter(File.parent == dir_id).all()) return File.query.filter(File.parent == dir_id).all() def get_dir_tree_dict(): path = os.path.normpath("/gallery-data/root") file_tree = Dict() for root, _, files in os.walk(path, topdown=True): path = root.split('/') path.pop(0) file_tree_fd = file_tree for part in path: file_tree_fd = file_tree_fd[part] file_tree_fd['.'] = files return file_tree def convert_bytes_to_utf8(dic): for key in dic: if isinstance(key, bytes): k = key.decode('utf-8') v = dic[key] del dic[key] dic[k] = v if isinstance(dic[key], bytes): v = dic[key].decode('utf-8') dic[key] = v return dic def allowed_file(filename): return '.' in filename and filename.lower().rsplit('.', 1)[1] in \ [ 'txt', 'png', 'jpg', 'jpeg', 'mpg', 'mp4', 'avi', 'cr2' ]
import os from addict import Dict from gallery.models import File def get_dir_file_contents(dir_id): print(File.query.filter(File.parent == dir_id).all()) return File.query.filter(File.parent == dir_id).all() def get_dir_tree_dict(): path = os.path.normpath("/gallery-data/root") file_tree = Dict() for root, _, files in os.walk(path, topdown=True): path = root.split('/') path.pop(0) file_tree_fd = file_tree for part in path: file_tree_fd = file_tree_fd[part] file_tree_fd['.'] = files return file_tree def convert_bytes_to_utf8(dic): for key in dic: if isinstance(key, bytes): k = key.decode('utf-8') v = dic[key] del dic[key] dic[k] = v if isinstance(dic[key], bytes): v = dic[key].decode('utf-8') dic[key] = v return dic def allowed_file(filename): return '.' in filename and filename.lower().rsplit('.', 1)[1] in \ [ 'txt', 'png', 'jpg', 'jpeg', 'mpg', 'mp4', - 'avi' + 'avi', ? + + 'cr2' ]
1c60cf7082672335279d5b96e83f3cb2eb57424f
purchase_supplier_minimum_order/models/__init__.py
purchase_supplier_minimum_order/models/__init__.py
from . import ( res_partner, ) # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
from . import ( res_partner, purchase, ) # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
Enforce minimum PO value for supplier.
Enforce minimum PO value for supplier.
Python
agpl-3.0
OpusVL/odoo-purchase-min-order
from . import ( res_partner, + purchase, ) # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
Enforce minimum PO value for supplier.
## Code Before: from . import ( res_partner, ) # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: ## Instruction: Enforce minimum PO value for supplier. ## Code After: from . import ( res_partner, purchase, ) # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
from . import ( res_partner, + purchase, ) # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
a04a8c7d8e1087df39025d6e798d83438ac35f77
setup.py
setup.py
from distutils.core import setup setup(name='hpswitch', version='0.1', description="A library for interacting with HP Networking switches", packages=['hpswitch', ], url='https://github.com/leonhandreke/hpswitch', license="MIT License", )
from distutils.core import setup setup(name='hpswitch', version='0.1', description="A library for interacting with HP Networking switches", packages=['hpswitch', ], url='https://github.com/leonhandreke/hpswitch', license="MIT License", requires=['pysnmp'] )
Add pysnmp as a dependency
Add pysnmp as a dependency
Python
mit
leonhandreke/hpswitch,thechristschn/hpswitch
from distutils.core import setup setup(name='hpswitch', version='0.1', description="A library for interacting with HP Networking switches", packages=['hpswitch', ], url='https://github.com/leonhandreke/hpswitch', license="MIT License", + requires=['pysnmp'] )
Add pysnmp as a dependency
## Code Before: from distutils.core import setup setup(name='hpswitch', version='0.1', description="A library for interacting with HP Networking switches", packages=['hpswitch', ], url='https://github.com/leonhandreke/hpswitch', license="MIT License", ) ## Instruction: Add pysnmp as a dependency ## Code After: from distutils.core import setup setup(name='hpswitch', version='0.1', description="A library for interacting with HP Networking switches", packages=['hpswitch', ], url='https://github.com/leonhandreke/hpswitch', license="MIT License", requires=['pysnmp'] )
from distutils.core import setup setup(name='hpswitch', version='0.1', description="A library for interacting with HP Networking switches", packages=['hpswitch', ], url='https://github.com/leonhandreke/hpswitch', license="MIT License", + requires=['pysnmp'] )
e183578b6211d7311d62100ad643cbaf8408de99
tests/__init__.py
tests/__init__.py
import unittest.mock def _test_module_init(module, main_name="main"): with unittest.mock.patch.object(module, main_name, return_value=0): with unittest.mock.patch.object(module, "__name__", "__main__"): with unittest.mock.patch.object(module.sys, "exit") as exit: module.module_init() return exit.call_args[0][0] == 0
import unittest.mock def _test_module_init(module, main_name="main"): with unittest.mock.patch.object( module, main_name, return_value=0 ), unittest.mock.patch.object( module, "__name__", "__main__" ), unittest.mock.patch.object( module.sys, "exit" ) as exit: module.module_init() return exit.call_args[0][0] == 0
Use multiple context managers on one with statement (thanks Anna)
Use multiple context managers on one with statement (thanks Anna)
Python
mpl-2.0
rfinnie/2ping,rfinnie/2ping
import unittest.mock def _test_module_init(module, main_name="main"): - with unittest.mock.patch.object(module, main_name, return_value=0): - with unittest.mock.patch.object(module, "__name__", "__main__"): - with unittest.mock.patch.object(module.sys, "exit") as exit: + with unittest.mock.patch.object( + module, main_name, return_value=0 + ), unittest.mock.patch.object( + module, "__name__", "__main__" + ), unittest.mock.patch.object( + module.sys, "exit" + ) as exit: - module.module_init() + module.module_init() - return exit.call_args[0][0] == 0 + return exit.call_args[0][0] == 0
Use multiple context managers on one with statement (thanks Anna)
## Code Before: import unittest.mock def _test_module_init(module, main_name="main"): with unittest.mock.patch.object(module, main_name, return_value=0): with unittest.mock.patch.object(module, "__name__", "__main__"): with unittest.mock.patch.object(module.sys, "exit") as exit: module.module_init() return exit.call_args[0][0] == 0 ## Instruction: Use multiple context managers on one with statement (thanks Anna) ## Code After: import unittest.mock def _test_module_init(module, main_name="main"): with unittest.mock.patch.object( module, main_name, return_value=0 ), unittest.mock.patch.object( module, "__name__", "__main__" ), unittest.mock.patch.object( module.sys, "exit" ) as exit: module.module_init() return exit.call_args[0][0] == 0
import unittest.mock def _test_module_init(module, main_name="main"): - with unittest.mock.patch.object(module, main_name, return_value=0): - with unittest.mock.patch.object(module, "__name__", "__main__"): - with unittest.mock.patch.object(module.sys, "exit") as exit: + with unittest.mock.patch.object( + module, main_name, return_value=0 + ), unittest.mock.patch.object( + module, "__name__", "__main__" + ), unittest.mock.patch.object( + module.sys, "exit" + ) as exit: - module.module_init() ? -------- + module.module_init() - return exit.call_args[0][0] == 0 ? -------- + return exit.call_args[0][0] == 0