commit
stringlengths
40
40
old_file
stringlengths
4
106
new_file
stringlengths
4
106
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
2.95k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.31k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
diff
stringlengths
49
3.61k
7325eacc1066970a98be30b56fdf4cd31ecc2f57
db_file_storage/views.py
db_file_storage/views.py
from wsgiref.util import FileWrapper from django.http import HttpResponse, HttpResponseBadRequest from django.utils.translation import ugettext as _ # project from db_file_storage.storage import DatabaseFileStorage storage = DatabaseFileStorage() def get_file(request, add_attachment_headers): name = request.GET.get('name') try: _file = storage.open(name) except Exception: return HttpResponseBadRequest(_('Invalid request')) response = HttpResponse( FileWrapper(_file), content_type=_file.mimetype ) if add_attachment_headers: response['Content-Disposition'] = \ 'attachment; filename=%(name)s' % {'name': _file.filename} return response
from wsgiref.util import FileWrapper from django.http import HttpResponse, HttpResponseBadRequest from django.utils.translation import ugettext as _ # project from db_file_storage.storage import DatabaseFileStorage storage = DatabaseFileStorage() def get_file(request, add_attachment_headers): name = request.GET.get('name') try: _file = storage.open(name) except Exception: return HttpResponseBadRequest(_('Invalid request')) response = HttpResponse( FileWrapper(_file), content_type=_file.mimetype ) response['Content-Length'] = _file.tell() if add_attachment_headers: response['Content-Disposition'] = \ 'attachment; filename=%(name)s' % {'name': _file.filename} return response
Set Content-Length header in get_file view
Set Content-Length header in get_file view
Python
mit
victor-o-silva/db_file_storage,victor-o-silva/db_file_storage
from wsgiref.util import FileWrapper from django.http import HttpResponse, HttpResponseBadRequest from django.utils.translation import ugettext as _ # project from db_file_storage.storage import DatabaseFileStorage storage = DatabaseFileStorage() def get_file(request, add_attachment_headers): name = request.GET.get('name') try: _file = storage.open(name) except Exception: return HttpResponseBadRequest(_('Invalid request')) response = HttpResponse( FileWrapper(_file), content_type=_file.mimetype ) - + response['Content-Length'] = _file.tell() if add_attachment_headers: response['Content-Disposition'] = \ 'attachment; filename=%(name)s' % {'name': _file.filename} return response
Set Content-Length header in get_file view
## Code Before: from wsgiref.util import FileWrapper from django.http import HttpResponse, HttpResponseBadRequest from django.utils.translation import ugettext as _ # project from db_file_storage.storage import DatabaseFileStorage storage = DatabaseFileStorage() def get_file(request, add_attachment_headers): name = request.GET.get('name') try: _file = storage.open(name) except Exception: return HttpResponseBadRequest(_('Invalid request')) response = HttpResponse( FileWrapper(_file), content_type=_file.mimetype ) if add_attachment_headers: response['Content-Disposition'] = \ 'attachment; filename=%(name)s' % {'name': _file.filename} return response ## Instruction: Set Content-Length header in get_file view ## Code After: from wsgiref.util import FileWrapper from django.http import HttpResponse, HttpResponseBadRequest from django.utils.translation import ugettext as _ # project from db_file_storage.storage import DatabaseFileStorage storage = DatabaseFileStorage() def get_file(request, add_attachment_headers): name = request.GET.get('name') try: _file = storage.open(name) except Exception: return HttpResponseBadRequest(_('Invalid request')) response = HttpResponse( FileWrapper(_file), content_type=_file.mimetype ) response['Content-Length'] = _file.tell() if add_attachment_headers: response['Content-Disposition'] = \ 'attachment; filename=%(name)s' % {'name': _file.filename} return response
from wsgiref.util import FileWrapper from django.http import HttpResponse, HttpResponseBadRequest from django.utils.translation import ugettext as _ # project from db_file_storage.storage import DatabaseFileStorage storage = DatabaseFileStorage() def get_file(request, add_attachment_headers): name = request.GET.get('name') try: _file = storage.open(name) except Exception: return HttpResponseBadRequest(_('Invalid request')) response = HttpResponse( FileWrapper(_file), content_type=_file.mimetype ) - + response['Content-Length'] = _file.tell() if add_attachment_headers: response['Content-Disposition'] = \ 'attachment; filename=%(name)s' % {'name': _file.filename} return response
7a8112249de859a5ef73fe07eb6029aeb1266f35
tob-api/tob_api/urls.py
tob-api/tob_api/urls.py
from django.conf.urls import include, url from django.views.generic import RedirectView from . import views # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = [ url(r"^$", RedirectView.as_view(url="api/v2/")), url( r"^api-auth/", include("rest_framework.urls", namespace="rest_framework"), ), # url(r"^api/v1/", include("api.urls")), url(r"^api/v2/", include("api_v2.urls")), url(r"^health$", views.health), ]
from django.conf.urls import include, url from django.views.generic import RedirectView from . import views # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = [ url(r"^$", RedirectView.as_view(url="api/v2/")), url( r"^api-auth/", include("rest_framework.urls", namespace="rest_framework"), ), url(r"^api/v2/", include("api_v2.urls")), url(r"^health$", views.health), ]
Remove commented-out reference to v1
Remove commented-out reference to v1 Signed-off-by: Nicholas Rempel <b7f0f2181f2dc324d159332b253a82a715a40706@gmail.com>
Python
apache-2.0
swcurran/TheOrgBook,swcurran/TheOrgBook,WadeBarnes/TheOrgBook,swcurran/TheOrgBook,WadeBarnes/TheOrgBook,WadeBarnes/TheOrgBook,WadeBarnes/TheOrgBook,WadeBarnes/TheOrgBook,swcurran/TheOrgBook,swcurran/TheOrgBook
from django.conf.urls import include, url from django.views.generic import RedirectView from . import views # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = [ url(r"^$", RedirectView.as_view(url="api/v2/")), url( r"^api-auth/", include("rest_framework.urls", namespace="rest_framework"), ), - # url(r"^api/v1/", include("api.urls")), url(r"^api/v2/", include("api_v2.urls")), url(r"^health$", views.health), ]
Remove commented-out reference to v1
## Code Before: from django.conf.urls import include, url from django.views.generic import RedirectView from . import views # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = [ url(r"^$", RedirectView.as_view(url="api/v2/")), url( r"^api-auth/", include("rest_framework.urls", namespace="rest_framework"), ), # url(r"^api/v1/", include("api.urls")), url(r"^api/v2/", include("api_v2.urls")), url(r"^health$", views.health), ] ## Instruction: Remove commented-out reference to v1 ## Code After: from django.conf.urls import include, url from django.views.generic import RedirectView from . import views # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = [ url(r"^$", RedirectView.as_view(url="api/v2/")), url( r"^api-auth/", include("rest_framework.urls", namespace="rest_framework"), ), url(r"^api/v2/", include("api_v2.urls")), url(r"^health$", views.health), ]
from django.conf.urls import include, url from django.views.generic import RedirectView from . import views # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = [ url(r"^$", RedirectView.as_view(url="api/v2/")), url( r"^api-auth/", include("rest_framework.urls", namespace="rest_framework"), ), - # url(r"^api/v1/", include("api.urls")), url(r"^api/v2/", include("api_v2.urls")), url(r"^health$", views.health), ]
f68b4b9b133d3c8ecb9826af9736c8c1fca64e49
maxims/credentials.py
maxims/credentials.py
from axiom import attributes, item from twisted.cred import credentials class UsernamePassword(item.Item): """ A stored username and password. """ username = attributes.bytes(allowNone=False) password = attributes.bytes(allowNone=False) def instantiate(self): return credentials.UsernamePassword(self.username, self.password)
from axiom import attributes, item from twisted.cred import credentials class UsernamePassword(item.Item): """ A stored username and password. Note that although this class is an ``IUsernamePassword`` implementation, you should still use the ``instantiate`` method to get independent ``IUsernamePassword`` providers. """ username = attributes.bytes(allowNone=False) password = attributes.bytes(allowNone=False) def instantiate(self): return credentials.UsernamePassword(self.username, self.password)
Add caveat about UsernamePassword already being an IUsernamePassword implementation
Add caveat about UsernamePassword already being an IUsernamePassword implementation
Python
isc
lvh/maxims
from axiom import attributes, item from twisted.cred import credentials class UsernamePassword(item.Item): """ A stored username and password. + + Note that although this class is an ``IUsernamePassword`` implementation, + you should still use the ``instantiate`` method to get independent + ``IUsernamePassword`` providers. """ username = attributes.bytes(allowNone=False) password = attributes.bytes(allowNone=False) def instantiate(self): return credentials.UsernamePassword(self.username, self.password)
Add caveat about UsernamePassword already being an IUsernamePassword implementation
## Code Before: from axiom import attributes, item from twisted.cred import credentials class UsernamePassword(item.Item): """ A stored username and password. """ username = attributes.bytes(allowNone=False) password = attributes.bytes(allowNone=False) def instantiate(self): return credentials.UsernamePassword(self.username, self.password) ## Instruction: Add caveat about UsernamePassword already being an IUsernamePassword implementation ## Code After: from axiom import attributes, item from twisted.cred import credentials class UsernamePassword(item.Item): """ A stored username and password. Note that although this class is an ``IUsernamePassword`` implementation, you should still use the ``instantiate`` method to get independent ``IUsernamePassword`` providers. """ username = attributes.bytes(allowNone=False) password = attributes.bytes(allowNone=False) def instantiate(self): return credentials.UsernamePassword(self.username, self.password)
from axiom import attributes, item from twisted.cred import credentials class UsernamePassword(item.Item): """ A stored username and password. + + Note that although this class is an ``IUsernamePassword`` implementation, + you should still use the ``instantiate`` method to get independent + ``IUsernamePassword`` providers. """ username = attributes.bytes(allowNone=False) password = attributes.bytes(allowNone=False) def instantiate(self): return credentials.UsernamePassword(self.username, self.password)
df18229b38a01d87076f3b13aee5bfd1f0f989c2
tunobase/blog/models.py
tunobase/blog/models.py
''' Blog App This module determines how to display the Blog app in Django's admin and lists other model functions. ''' from django.conf import settings from django.core.urlresolvers import reverse from django.db import models from tunobase.core import models as core_models class Blog(core_models.ContentModel): ''' Blogs the Site has ''' class BlogEntry(core_models.ContentModel): ''' Entries per Blog ''' blog = models.ForeignKey(Blog, related_name='entries') author_users = models.ManyToManyField( settings.AUTH_USER_MODEL, related_name='blog_entries_authored', null=True, blank=True ) authors_alternate = models.CharField( max_length=512, blank=True, null=True ) class Meta: verbose_name_plural = 'Blog entries' def get_absolute_url(self): return reverse('blog_entry_detail', args=(self.slug,)) @property def authors(self): ''' Return a list of authors selected as users on the system and a list of alternate authors as not users on the system if either exist ''' authors_dict = {} auth_users = self.author_users.all() if auth_users: authors_dict.update({ 'users': auth_users }) if self.authors_alternate: authors_dict.update({ 'alternate': self.authors_alternate.split(',') }) return authors_dict
''' Blog App This module determines how to display the Blog app in Django's admin and lists other model functions. ''' from django.conf import settings from django.core.urlresolvers import reverse from django.db import models from tunobase.core import models as core_models class Blog(core_models.ContentModel): ''' Blogs the Site has ''' class Meta: verbose_name = 'Blog Category' verbose_name_plural = 'Blog Categories' class BlogEntry(core_models.ContentModel): ''' Entries per Blog ''' blog = models.ForeignKey(Blog, related_name='entries') author_users = models.ManyToManyField( settings.AUTH_USER_MODEL, related_name='blog_entries_authored', null=True, blank=True ) authors_alternate = models.CharField( max_length=512, blank=True, null=True ) class Meta: verbose_name_plural = 'Blog entries' def get_absolute_url(self): return reverse('blog_entry_detail', args=(self.slug,)) @property def authors(self): ''' Return a list of authors selected as users on the system and a list of alternate authors as not users on the system if either exist ''' authors_dict = {} auth_users = self.author_users.all() if auth_users: authors_dict.update({ 'users': auth_users }) if self.authors_alternate: authors_dict.update({ 'alternate': self.authors_alternate.split(',') }) return authors_dict
Update blog model with a more descriptive name
Update blog model with a more descriptive name
Python
bsd-3-clause
unomena/tunobase,unomena/tunobase
''' Blog App This module determines how to display the Blog app in Django's admin and lists other model functions. ''' from django.conf import settings from django.core.urlresolvers import reverse from django.db import models from tunobase.core import models as core_models class Blog(core_models.ContentModel): ''' Blogs the Site has ''' + + class Meta: + verbose_name = 'Blog Category' + verbose_name_plural = 'Blog Categories' class BlogEntry(core_models.ContentModel): ''' Entries per Blog ''' blog = models.ForeignKey(Blog, related_name='entries') author_users = models.ManyToManyField( settings.AUTH_USER_MODEL, related_name='blog_entries_authored', null=True, blank=True ) authors_alternate = models.CharField( max_length=512, blank=True, null=True ) class Meta: verbose_name_plural = 'Blog entries' def get_absolute_url(self): return reverse('blog_entry_detail', args=(self.slug,)) @property def authors(self): ''' Return a list of authors selected as users on the system and a list of alternate authors as not users on the system if either exist ''' authors_dict = {} auth_users = self.author_users.all() if auth_users: authors_dict.update({ 'users': auth_users }) if self.authors_alternate: authors_dict.update({ 'alternate': self.authors_alternate.split(',') }) return authors_dict
Update blog model with a more descriptive name
## Code Before: ''' Blog App This module determines how to display the Blog app in Django's admin and lists other model functions. ''' from django.conf import settings from django.core.urlresolvers import reverse from django.db import models from tunobase.core import models as core_models class Blog(core_models.ContentModel): ''' Blogs the Site has ''' class BlogEntry(core_models.ContentModel): ''' Entries per Blog ''' blog = models.ForeignKey(Blog, related_name='entries') author_users = models.ManyToManyField( settings.AUTH_USER_MODEL, related_name='blog_entries_authored', null=True, blank=True ) authors_alternate = models.CharField( max_length=512, blank=True, null=True ) class Meta: verbose_name_plural = 'Blog entries' def get_absolute_url(self): return reverse('blog_entry_detail', args=(self.slug,)) @property def authors(self): ''' Return a list of authors selected as users on the system and a list of alternate authors as not users on the system if either exist ''' authors_dict = {} auth_users = self.author_users.all() if auth_users: authors_dict.update({ 'users': auth_users }) if self.authors_alternate: authors_dict.update({ 'alternate': self.authors_alternate.split(',') }) return authors_dict ## Instruction: Update blog model with a more descriptive name ## Code After: ''' Blog App This module determines how to display the Blog app in Django's admin and lists other model functions. ''' from django.conf import settings from django.core.urlresolvers import reverse from django.db import models from tunobase.core import models as core_models class Blog(core_models.ContentModel): ''' Blogs the Site has ''' class Meta: verbose_name = 'Blog Category' verbose_name_plural = 'Blog Categories' class BlogEntry(core_models.ContentModel): ''' Entries per Blog ''' blog = models.ForeignKey(Blog, related_name='entries') author_users = models.ManyToManyField( settings.AUTH_USER_MODEL, related_name='blog_entries_authored', null=True, blank=True ) authors_alternate = models.CharField( max_length=512, blank=True, null=True ) class Meta: verbose_name_plural = 'Blog entries' def get_absolute_url(self): return reverse('blog_entry_detail', args=(self.slug,)) @property def authors(self): ''' Return a list of authors selected as users on the system and a list of alternate authors as not users on the system if either exist ''' authors_dict = {} auth_users = self.author_users.all() if auth_users: authors_dict.update({ 'users': auth_users }) if self.authors_alternate: authors_dict.update({ 'alternate': self.authors_alternate.split(',') }) return authors_dict
''' Blog App This module determines how to display the Blog app in Django's admin and lists other model functions. ''' from django.conf import settings from django.core.urlresolvers import reverse from django.db import models from tunobase.core import models as core_models class Blog(core_models.ContentModel): ''' Blogs the Site has ''' + + class Meta: + verbose_name = 'Blog Category' + verbose_name_plural = 'Blog Categories' class BlogEntry(core_models.ContentModel): ''' Entries per Blog ''' blog = models.ForeignKey(Blog, related_name='entries') author_users = models.ManyToManyField( settings.AUTH_USER_MODEL, related_name='blog_entries_authored', null=True, blank=True ) authors_alternate = models.CharField( max_length=512, blank=True, null=True ) class Meta: verbose_name_plural = 'Blog entries' def get_absolute_url(self): return reverse('blog_entry_detail', args=(self.slug,)) @property def authors(self): ''' Return a list of authors selected as users on the system and a list of alternate authors as not users on the system if either exist ''' authors_dict = {} auth_users = self.author_users.all() if auth_users: authors_dict.update({ 'users': auth_users }) if self.authors_alternate: authors_dict.update({ 'alternate': self.authors_alternate.split(',') }) return authors_dict
7a37e3afa29410636c75408bc649e70c519e07f1
test/user_profile_test.py
test/user_profile_test.py
import json from pymessenger.user_profile import UserProfileApi from test_env import * upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET) def test_fields_blank(): user_profile = upa.get(TEST_USER_ID) assert user_profile is not None def test_fields(): fields = ['first_name', 'last_name'] user_profile = upa.get(TEST_USER_ID, fields=fields) assert user_profile is not None assert len(user_profile.keys()) == len(fields)
import json import sys, os sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/..")) from pymessenger.user_profile import UserProfileApi TOKEN = os.environ.get('TOKEN') APP_SECRET = os.environ.get('APP_SECRET') TEST_USER_ID = os.environ.get('RECIPIENT_ID') upa = UserProfileApi(TOKEN, app_secret=APP_SECRET) def test_fields_blank(): user_profile = upa.get(TEST_USER_ID) assert user_profile is not None def test_fields(): fields = ['first_name', 'last_name'] user_profile = upa.get(TEST_USER_ID, fields=fields) assert user_profile is not None assert len(user_profile.keys()) == len(fields)
Fix user profile test to include same environment variables
Fix user profile test to include same environment variables
Python
mit
karlinnolabs/pymessenger,Cretezy/pymessenger2,davidchua/pymessenger
import json + import sys, os + sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/..")) from pymessenger.user_profile import UserProfileApi - from test_env import * + TOKEN = os.environ.get('TOKEN') + APP_SECRET = os.environ.get('APP_SECRET') + TEST_USER_ID = os.environ.get('RECIPIENT_ID') + - upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET) + upa = UserProfileApi(TOKEN, app_secret=APP_SECRET) def test_fields_blank(): user_profile = upa.get(TEST_USER_ID) assert user_profile is not None def test_fields(): fields = ['first_name', 'last_name'] user_profile = upa.get(TEST_USER_ID, fields=fields) assert user_profile is not None assert len(user_profile.keys()) == len(fields)
Fix user profile test to include same environment variables
## Code Before: import json from pymessenger.user_profile import UserProfileApi from test_env import * upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET) def test_fields_blank(): user_profile = upa.get(TEST_USER_ID) assert user_profile is not None def test_fields(): fields = ['first_name', 'last_name'] user_profile = upa.get(TEST_USER_ID, fields=fields) assert user_profile is not None assert len(user_profile.keys()) == len(fields) ## Instruction: Fix user profile test to include same environment variables ## Code After: import json import sys, os sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/..")) from pymessenger.user_profile import UserProfileApi TOKEN = os.environ.get('TOKEN') APP_SECRET = os.environ.get('APP_SECRET') TEST_USER_ID = os.environ.get('RECIPIENT_ID') upa = UserProfileApi(TOKEN, app_secret=APP_SECRET) def test_fields_blank(): user_profile = upa.get(TEST_USER_ID) assert user_profile is not None def test_fields(): fields = ['first_name', 'last_name'] user_profile = upa.get(TEST_USER_ID, fields=fields) assert user_profile is not None assert len(user_profile.keys()) == len(fields)
import json + import sys, os + sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/..")) from pymessenger.user_profile import UserProfileApi - from test_env import * + TOKEN = os.environ.get('TOKEN') + APP_SECRET = os.environ.get('APP_SECRET') + TEST_USER_ID = os.environ.get('RECIPIENT_ID') + - upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET) ? ------------ + upa = UserProfileApi(TOKEN, app_secret=APP_SECRET) def test_fields_blank(): user_profile = upa.get(TEST_USER_ID) assert user_profile is not None def test_fields(): fields = ['first_name', 'last_name'] user_profile = upa.get(TEST_USER_ID, fields=fields) assert user_profile is not None assert len(user_profile.keys()) == len(fields)
f3763c417d745463361b054fd4ffa0ddf35833eb
src/server/Universe.py
src/server/Universe.py
class Universe: def __init__(self, height=100000000, width=100000000): self.entities = [] self.height = height self.width = width self.teams = [] self.state = [] def add(self, entity): self.entities.append(entity) def remove(self, entity): self.entities.remove(entity) # Time passes and position updates during tick def tick(self, duration): for i in self.entities: i.tick(duration) # Position changed, so check for collisions def collide(self): for i in self.entities: for j in self.entities: if i != j: if i.checkCollide(j): i.collide(j) # Now that damage is dealt in collisions, destroy objects and update logic def tock(self): for i in self.entities: i.tock() def dumpState(self): return self.state
class Universe: def __init__(self, height=100000000, width=100000000): self.entities = [] self.height = height self.width = width self.teams = [] self.state = [] self.maxID = 0 def add(self, entity): maxID += 1 entity.id = maxID self.entities.append(entity) def remove(self, entity): self.entities.remove(entity) # Time passes and position updates during tick def tick(self, duration): for i in self.entities: i.tick(duration) # Position changed, so check for collisions def collide(self): for i in self.entities: for j in self.entities: if i != j: if i.checkCollide(j): i.collide(j) # Now that damage is dealt in collisions, destroy objects and update logic def tock(self): for i in self.entities: i.tock() def dumpState(self): return self.state
Add ID to Entity upon creation
Add ID to Entity upon creation
Python
mit
cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim
class Universe: def __init__(self, height=100000000, width=100000000): self.entities = [] self.height = height self.width = width self.teams = [] self.state = [] + self.maxID = 0 def add(self, entity): + maxID += 1 + entity.id = maxID self.entities.append(entity) def remove(self, entity): self.entities.remove(entity) # Time passes and position updates during tick def tick(self, duration): for i in self.entities: i.tick(duration) # Position changed, so check for collisions def collide(self): for i in self.entities: for j in self.entities: if i != j: if i.checkCollide(j): i.collide(j) # Now that damage is dealt in collisions, destroy objects and update logic def tock(self): for i in self.entities: i.tock() def dumpState(self): return self.state
Add ID to Entity upon creation
## Code Before: class Universe: def __init__(self, height=100000000, width=100000000): self.entities = [] self.height = height self.width = width self.teams = [] self.state = [] def add(self, entity): self.entities.append(entity) def remove(self, entity): self.entities.remove(entity) # Time passes and position updates during tick def tick(self, duration): for i in self.entities: i.tick(duration) # Position changed, so check for collisions def collide(self): for i in self.entities: for j in self.entities: if i != j: if i.checkCollide(j): i.collide(j) # Now that damage is dealt in collisions, destroy objects and update logic def tock(self): for i in self.entities: i.tock() def dumpState(self): return self.state ## Instruction: Add ID to Entity upon creation ## Code After: class Universe: def __init__(self, height=100000000, width=100000000): self.entities = [] self.height = height self.width = width self.teams = [] self.state = [] self.maxID = 0 def add(self, entity): maxID += 1 entity.id = maxID self.entities.append(entity) def remove(self, entity): self.entities.remove(entity) # Time passes and position updates during tick def tick(self, duration): for i in self.entities: i.tick(duration) # Position changed, so check for collisions def collide(self): for i in self.entities: for j in self.entities: if i != j: if i.checkCollide(j): i.collide(j) # Now that damage is dealt in collisions, destroy objects and update logic def tock(self): for i in self.entities: i.tock() def dumpState(self): return self.state
class Universe: def __init__(self, height=100000000, width=100000000): self.entities = [] self.height = height self.width = width self.teams = [] self.state = [] + self.maxID = 0 def add(self, entity): + maxID += 1 + entity.id = maxID self.entities.append(entity) def remove(self, entity): self.entities.remove(entity) # Time passes and position updates during tick def tick(self, duration): for i in self.entities: i.tick(duration) # Position changed, so check for collisions def collide(self): for i in self.entities: for j in self.entities: if i != j: if i.checkCollide(j): i.collide(j) # Now that damage is dealt in collisions, destroy objects and update logic def tock(self): for i in self.entities: i.tock() def dumpState(self): return self.state
88752efa9ac2c0f251733e335763cb880da34741
thinglang/parser/definitions/member_definition.py
thinglang/parser/definitions/member_definition.py
from thinglang.lexer.definitions.tags import LexicalPrivateTag from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember from thinglang.lexer.values.identifier import Identifier from thinglang.parser.nodes import BaseNode from thinglang.parser.rule import ParserRule from thinglang.symbols.symbol import Symbol class MemberDefinition(BaseNode): """ A member definition Must be a direct child of a ThingDefinition """ def __init__(self, name, type_name, visibility=Symbol.PUBLIC): super(MemberDefinition, self).__init__([name, type_name]) self.type, self.name, self.visibility = type_name, name, visibility def __repr__(self): return 'has {} {}'.format(self.type, self.name) def symbol(self): return Symbol.member(self.name, self.type, self.visibility) MEMBER_NAME_TYPES = Identifier @staticmethod @ParserRule.mark def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier): return MemberDefinition(name, type_name) @staticmethod @ParserRule.mark def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'): member.visibility = Symbol.PRIVATE return member
from thinglang.lexer.definitions.tags import LexicalPrivateTag from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember from thinglang.lexer.values.identifier import Identifier from thinglang.parser.nodes import BaseNode from thinglang.parser.rule import ParserRule from thinglang.symbols.symbol import Symbol class MemberDefinition(BaseNode): """ A member definition Must be a direct child of a ThingDefinition """ def __init__(self, name, type_name, visibility=Symbol.PUBLIC): super(MemberDefinition, self).__init__([name, type_name]) self.type, self.name, self.visibility = type_name, name, visibility def __repr__(self): return 'has {} {}'.format(self.type, self.name) def symbol(self): return Symbol.member(self.name, self.type, self.visibility) MEMBER_NAME_TYPES = Identifier @staticmethod @ParserRule.mark def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier): return MemberDefinition(name, type_name)
Add visibility tagging to MethoDefinition
Add visibility tagging to MethoDefinition
Python
mit
ytanay/thinglang,ytanay/thinglang,ytanay/thinglang,ytanay/thinglang
from thinglang.lexer.definitions.tags import LexicalPrivateTag from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember from thinglang.lexer.values.identifier import Identifier from thinglang.parser.nodes import BaseNode from thinglang.parser.rule import ParserRule from thinglang.symbols.symbol import Symbol class MemberDefinition(BaseNode): """ A member definition Must be a direct child of a ThingDefinition """ def __init__(self, name, type_name, visibility=Symbol.PUBLIC): super(MemberDefinition, self).__init__([name, type_name]) self.type, self.name, self.visibility = type_name, name, visibility def __repr__(self): return 'has {} {}'.format(self.type, self.name) def symbol(self): return Symbol.member(self.name, self.type, self.visibility) MEMBER_NAME_TYPES = Identifier @staticmethod @ParserRule.mark - def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier): + def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier): return MemberDefinition(name, type_name) - @staticmethod - @ParserRule.mark - def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'): - member.visibility = Symbol.PRIVATE - return member -
Add visibility tagging to MethoDefinition
## Code Before: from thinglang.lexer.definitions.tags import LexicalPrivateTag from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember from thinglang.lexer.values.identifier import Identifier from thinglang.parser.nodes import BaseNode from thinglang.parser.rule import ParserRule from thinglang.symbols.symbol import Symbol class MemberDefinition(BaseNode): """ A member definition Must be a direct child of a ThingDefinition """ def __init__(self, name, type_name, visibility=Symbol.PUBLIC): super(MemberDefinition, self).__init__([name, type_name]) self.type, self.name, self.visibility = type_name, name, visibility def __repr__(self): return 'has {} {}'.format(self.type, self.name) def symbol(self): return Symbol.member(self.name, self.type, self.visibility) MEMBER_NAME_TYPES = Identifier @staticmethod @ParserRule.mark def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier): return MemberDefinition(name, type_name) @staticmethod @ParserRule.mark def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'): member.visibility = Symbol.PRIVATE return member ## Instruction: Add visibility tagging to MethoDefinition ## Code After: from thinglang.lexer.definitions.tags import LexicalPrivateTag from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember from thinglang.lexer.values.identifier import Identifier from thinglang.parser.nodes import BaseNode from thinglang.parser.rule import ParserRule from thinglang.symbols.symbol import Symbol class MemberDefinition(BaseNode): """ A member definition Must be a direct child of a ThingDefinition """ def __init__(self, name, type_name, visibility=Symbol.PUBLIC): super(MemberDefinition, self).__init__([name, type_name]) self.type, self.name, self.visibility = type_name, name, visibility def __repr__(self): return 'has {} {}'.format(self.type, self.name) def symbol(self): return Symbol.member(self.name, self.type, self.visibility) MEMBER_NAME_TYPES = Identifier @staticmethod @ParserRule.mark def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier): return MemberDefinition(name, type_name)
from thinglang.lexer.definitions.tags import LexicalPrivateTag from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember from thinglang.lexer.values.identifier import Identifier from thinglang.parser.nodes import BaseNode from thinglang.parser.rule import ParserRule from thinglang.symbols.symbol import Symbol class MemberDefinition(BaseNode): """ A member definition Must be a direct child of a ThingDefinition """ def __init__(self, name, type_name, visibility=Symbol.PUBLIC): super(MemberDefinition, self).__init__([name, type_name]) self.type, self.name, self.visibility = type_name, name, visibility def __repr__(self): return 'has {} {}'.format(self.type, self.name) def symbol(self): return Symbol.member(self.name, self.type, self.visibility) MEMBER_NAME_TYPES = Identifier @staticmethod @ParserRule.mark - def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier): + def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier): ? + ++++++++++++++++++++ return MemberDefinition(name, type_name) - - @staticmethod - @ParserRule.mark - def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'): - member.visibility = Symbol.PRIVATE - return member
f300f3b31dcdefa91fa8fe46bdaab2d2490ac06a
snd/image_board/serializers.py
snd/image_board/serializers.py
from django.contrib.auth.models import User from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like from rest_framework import serializers class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name') class ProfileSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Profile fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise', 'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo') class ContentItemSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentItem fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by') class HashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Hashtag fields = ('id', 'hashtag_text') class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentHashTag fields = ('id', 'content_id', 'hashtag_id') class LikeSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Like fields = ('id', 'user_id', 'content_id') class CommentSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Comment fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem')
from django.contrib.auth.models import User from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like from rest_framework import serializers class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name') class ProfileSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Profile fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise', 'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo') class ContentItemSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentItem fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by') class HashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Hashtag fields = ('id', 'url', 'hashtag_text') class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentHashTag fields = ('id', 'url', 'content_id', 'hashtag_id') class LikeSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Like fields = ('id', 'url', 'user_id', 'content_id') class CommentSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Comment fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem')
Add URLs to each searializer
Add URLs to each searializer
Python
mit
SNDjango/server,SNDjango/server,SNDjango/server
from django.contrib.auth.models import User from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like from rest_framework import serializers class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name') class ProfileSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Profile - fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise', + fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise', 'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo') class ContentItemSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentItem - fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by') + fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by') class HashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Hashtag - fields = ('id', 'hashtag_text') + fields = ('id', 'url', 'hashtag_text') class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentHashTag - fields = ('id', 'content_id', 'hashtag_id') + fields = ('id', 'url', 'content_id', 'hashtag_id') class LikeSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Like - fields = ('id', 'user_id', 'content_id') + fields = ('id', 'url', 'user_id', 'content_id') class CommentSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Comment - fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem') + fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem')
Add URLs to each searializer
## Code Before: from django.contrib.auth.models import User from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like from rest_framework import serializers class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name') class ProfileSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Profile fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise', 'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo') class ContentItemSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentItem fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by') class HashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Hashtag fields = ('id', 'hashtag_text') class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentHashTag fields = ('id', 'content_id', 'hashtag_id') class LikeSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Like fields = ('id', 'user_id', 'content_id') class CommentSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Comment fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem') ## Instruction: Add URLs to each searializer ## Code After: from django.contrib.auth.models import User from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like from rest_framework import serializers class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name') class ProfileSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Profile fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise', 'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo') class ContentItemSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentItem fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by') class HashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Hashtag fields = ('id', 'url', 'hashtag_text') class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentHashTag fields = ('id', 'url', 'content_id', 'hashtag_id') class LikeSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Like fields = ('id', 'url', 'user_id', 'content_id') class CommentSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Comment fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem')
from django.contrib.auth.models import User from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like from rest_framework import serializers class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name') class ProfileSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Profile - fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise', + fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise', ? +++++++++++++ 'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo') class ContentItemSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentItem - fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by') + fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by') ? +++++++ class HashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Hashtag - fields = ('id', 'hashtag_text') + fields = ('id', 'url', 'hashtag_text') ? +++++++ class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = ContentHashTag - fields = ('id', 'content_id', 'hashtag_id') + fields = ('id', 'url', 'content_id', 'hashtag_id') ? +++++++ class LikeSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Like - fields = ('id', 'user_id', 'content_id') + fields = ('id', 'url', 'user_id', 'content_id') ? +++++++ class CommentSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Comment - fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem') + fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem') ? +++++++
0cb5447de992389be9587d7706637212bfe3b90b
tests/events/tests.py
tests/events/tests.py
from mock import Mock from unittest2 import TestCase from raven.events import Message class MessageTest(TestCase): def test_to_string(self): unformatted_message = 'My message from %s about %s' client = Mock() message = Message(client) message.logger = Mock() data = { 'sentry.interfaces.Message': { 'message': unformatted_message, } } self.assertEqual(message.to_string(data), unformatted_message) self.assertEqual(message.logger.warn.call_count, 1) args, kwargs = message.logger.warn.call_args self.assertEqual(args, ('Unable to find params for message',)) self.assertEqual(kwargs, {'extra': {'msg': {'message': unformatted_message}}}) data['sentry.interfaces.Message']['params'] = (1, 2) self.assertEqual(message.to_string(data), unformatted_message % (1, 2))
from mock import Mock from unittest2 import TestCase from raven.events import Message class MessageTest(TestCase): def test_to_string(self): unformatted_message = 'My message from %s about %s' client = Mock() message = Message(client) message.logger = Mock() data = { 'sentry.interfaces.Message': { 'message': unformatted_message, } } self.assertEqual(message.to_string(data), unformatted_message) data['sentry.interfaces.Message']['params'] = (1, 2) self.assertEqual(message.to_string(data), unformatted_message % (1, 2))
Update test to match current behavior
Update test to match current behavior
Python
bsd-3-clause
johansteffner/raven-python,Photonomie/raven-python,jbarbuto/raven-python,nikolas/raven-python,lepture/raven-python,smarkets/raven-python,arthurlogilab/raven-python,lepture/raven-python,hzy/raven-python,recht/raven-python,inspirehep/raven-python,nikolas/raven-python,openlabs/raven,patrys/opbeat_python,ewdurbin/raven-python,dbravender/raven-python,akalipetis/raven-python,alex/raven,Photonomie/raven-python,jmp0xf/raven-python,akalipetis/raven-python,jbarbuto/raven-python,smarkets/raven-python,nikolas/raven-python,smarkets/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,lopter/raven-python-old,beniwohli/apm-agent-python,percipient/raven-python,icereval/raven-python,percipient/raven-python,ronaldevers/raven-python,danriti/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akalipetis/raven-python,akheron/raven-python,dirtycoder/opbeat_python,hzy/raven-python,hzy/raven-python,patrys/opbeat_python,ewdurbin/raven-python,jmagnusson/raven-python,ronaldevers/raven-python,icereval/raven-python,collective/mr.poe,jmp0xf/raven-python,ticosax/opbeat_python,arthurlogilab/raven-python,tarkatronic/opbeat_python,icereval/raven-python,jbarbuto/raven-python,inspirehep/raven-python,ronaldevers/raven-python,akheron/raven-python,danriti/raven-python,daikeren/opbeat_python,johansteffner/raven-python,dbravender/raven-python,tarkatronic/opbeat_python,daikeren/opbeat_python,johansteffner/raven-python,ticosax/opbeat_python,someonehan/raven-python,Photonomie/raven-python,someonehan/raven-python,inspirehep/raven-python,dirtycoder/opbeat_python,beniwohli/apm-agent-python,beniwohli/apm-agent-python,jmagnusson/raven-python,percipient/raven-python,recht/raven-python,jbarbuto/raven-python,someonehan/raven-python,dirtycoder/opbeat_python,getsentry/raven-python,getsentry/raven-python,jmp0xf/raven-python,tarkatronic/opbeat_python,dbravender/raven-python,inspirehep/raven-python,akheron/raven-python,smarkets/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,patrys/opbeat_python,jmagnusson/raven-python,danriti/raven-python,nikolas/raven-python,getsentry/raven-python,icereval/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,arthurlogilab/raven-python,patrys/opbeat_python,arthurlogilab/raven-python,recht/raven-python,lepture/raven-python,daikeren/opbeat_python,ticosax/opbeat_python,beniwohli/apm-agent-python,ewdurbin/raven-python
from mock import Mock from unittest2 import TestCase from raven.events import Message class MessageTest(TestCase): def test_to_string(self): unformatted_message = 'My message from %s about %s' client = Mock() message = Message(client) message.logger = Mock() data = { 'sentry.interfaces.Message': { 'message': unformatted_message, } } self.assertEqual(message.to_string(data), unformatted_message) - self.assertEqual(message.logger.warn.call_count, 1) - - args, kwargs = message.logger.warn.call_args - self.assertEqual(args, ('Unable to find params for message',)) - self.assertEqual(kwargs, - {'extra': {'msg': {'message': unformatted_message}}}) data['sentry.interfaces.Message']['params'] = (1, 2) self.assertEqual(message.to_string(data), unformatted_message % (1, 2))
Update test to match current behavior
## Code Before: from mock import Mock from unittest2 import TestCase from raven.events import Message class MessageTest(TestCase): def test_to_string(self): unformatted_message = 'My message from %s about %s' client = Mock() message = Message(client) message.logger = Mock() data = { 'sentry.interfaces.Message': { 'message': unformatted_message, } } self.assertEqual(message.to_string(data), unformatted_message) self.assertEqual(message.logger.warn.call_count, 1) args, kwargs = message.logger.warn.call_args self.assertEqual(args, ('Unable to find params for message',)) self.assertEqual(kwargs, {'extra': {'msg': {'message': unformatted_message}}}) data['sentry.interfaces.Message']['params'] = (1, 2) self.assertEqual(message.to_string(data), unformatted_message % (1, 2)) ## Instruction: Update test to match current behavior ## Code After: from mock import Mock from unittest2 import TestCase from raven.events import Message class MessageTest(TestCase): def test_to_string(self): unformatted_message = 'My message from %s about %s' client = Mock() message = Message(client) message.logger = Mock() data = { 'sentry.interfaces.Message': { 'message': unformatted_message, } } self.assertEqual(message.to_string(data), unformatted_message) data['sentry.interfaces.Message']['params'] = (1, 2) self.assertEqual(message.to_string(data), unformatted_message % (1, 2))
from mock import Mock from unittest2 import TestCase from raven.events import Message class MessageTest(TestCase): def test_to_string(self): unformatted_message = 'My message from %s about %s' client = Mock() message = Message(client) message.logger = Mock() data = { 'sentry.interfaces.Message': { 'message': unformatted_message, } } self.assertEqual(message.to_string(data), unformatted_message) - self.assertEqual(message.logger.warn.call_count, 1) - - args, kwargs = message.logger.warn.call_args - self.assertEqual(args, ('Unable to find params for message',)) - self.assertEqual(kwargs, - {'extra': {'msg': {'message': unformatted_message}}}) data['sentry.interfaces.Message']['params'] = (1, 2) self.assertEqual(message.to_string(data), unformatted_message % (1, 2))
af45e43c46a22f3168ab946bf914a45eae9ade19
avatar/urls.py
avatar/urls.py
try: from django.conf.urls import patterns, url except ImportError: # Django < 1.4 from django.conf.urls.defaults import patterns, url from avatar import views urlpatterns = patterns('', url(r'^add/$', views.add, name='avatar_add'), url(r'^change/$', views.change, name='avatar_change'), url(r'^delete/$', views.delete, name='avatar_delete'), url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$', views.render_primary, name='avatar_render_primary'), url(r'^list/(?P<username>[\+\w\@\.]+)/$', views.avatar_gallery, name='avatar_gallery'), url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$', views.avatar, name='avatar'), )
try: from django.conf.urls import patterns, url except ImportError: # Django < 1.4 from django.conf.urls.defaults import url from avatar import views urlpatterns = [ url(r'^add/$', views.add, name='avatar_add'), url(r'^change/$', views.change, name='avatar_change'), url(r'^delete/$', views.delete, name='avatar_delete'), url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$', views.render_primary, name='avatar_render_primary'), url(r'^list/(?P<username>[\+\w\@\.]+)/$', views.avatar_gallery, name='avatar_gallery'), url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$', views.avatar, name='avatar'), ]
Remove replace urlpatterns with simple array, make compatible with Django 1.9
Remove replace urlpatterns with simple array, make compatible with Django 1.9
Python
bsd-3-clause
ad-m/django-avatar,ad-m/django-avatar,grantmcconnaughey/django-avatar,grantmcconnaughey/django-avatar,jezdez/django-avatar,jezdez/django-avatar,MachineandMagic/django-avatar,MachineandMagic/django-avatar
try: from django.conf.urls import patterns, url except ImportError: # Django < 1.4 - from django.conf.urls.defaults import patterns, url + from django.conf.urls.defaults import url from avatar import views - urlpatterns = patterns('', + urlpatterns = [ url(r'^add/$', views.add, name='avatar_add'), url(r'^change/$', views.change, name='avatar_change'), url(r'^delete/$', views.delete, name='avatar_delete'), url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$', views.render_primary, name='avatar_render_primary'), url(r'^list/(?P<username>[\+\w\@\.]+)/$', views.avatar_gallery, name='avatar_gallery'), url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$', views.avatar, name='avatar'), - ) + ]
Remove replace urlpatterns with simple array, make compatible with Django 1.9
## Code Before: try: from django.conf.urls import patterns, url except ImportError: # Django < 1.4 from django.conf.urls.defaults import patterns, url from avatar import views urlpatterns = patterns('', url(r'^add/$', views.add, name='avatar_add'), url(r'^change/$', views.change, name='avatar_change'), url(r'^delete/$', views.delete, name='avatar_delete'), url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$', views.render_primary, name='avatar_render_primary'), url(r'^list/(?P<username>[\+\w\@\.]+)/$', views.avatar_gallery, name='avatar_gallery'), url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$', views.avatar, name='avatar'), ) ## Instruction: Remove replace urlpatterns with simple array, make compatible with Django 1.9 ## Code After: try: from django.conf.urls import patterns, url except ImportError: # Django < 1.4 from django.conf.urls.defaults import url from avatar import views urlpatterns = [ url(r'^add/$', views.add, name='avatar_add'), url(r'^change/$', views.change, name='avatar_change'), url(r'^delete/$', views.delete, name='avatar_delete'), url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$', views.render_primary, name='avatar_render_primary'), url(r'^list/(?P<username>[\+\w\@\.]+)/$', views.avatar_gallery, name='avatar_gallery'), url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$', views.avatar, name='avatar'), ]
try: from django.conf.urls import patterns, url except ImportError: # Django < 1.4 - from django.conf.urls.defaults import patterns, url ? ---------- + from django.conf.urls.defaults import url from avatar import views - urlpatterns = patterns('', + urlpatterns = [ url(r'^add/$', views.add, name='avatar_add'), url(r'^change/$', views.change, name='avatar_change'), url(r'^delete/$', views.delete, name='avatar_delete'), url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$', views.render_primary, name='avatar_render_primary'), url(r'^list/(?P<username>[\+\w\@\.]+)/$', views.avatar_gallery, name='avatar_gallery'), url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$', views.avatar, name='avatar'), - ) + ]
c7ac2b1805bd82bdeef3227fc34149431c161df7
yesimeanit/showoff/newsletter_subscriptions/admin.py
yesimeanit/showoff/newsletter_subscriptions/admin.py
from django.contrib import admin from .models import NewsletterSubscription admin.site.register(NewsletterSubscription, list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'), list_filter=('is_active',), )
from django.contrib import admin from .models import NewsletterSubscription admin.site.register(NewsletterSubscription, list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'), list_filter=('is_active',), search_fields=('code', 'email', 'first_name', 'last_name'), )
Add search fields for newsletter subscriptions
Add search fields for newsletter subscriptions
Python
bsd-3-clause
guetux/django-yesimeanit
from django.contrib import admin from .models import NewsletterSubscription admin.site.register(NewsletterSubscription, list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'), list_filter=('is_active',), + search_fields=('code', 'email', 'first_name', 'last_name'), )
Add search fields for newsletter subscriptions
## Code Before: from django.contrib import admin from .models import NewsletterSubscription admin.site.register(NewsletterSubscription, list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'), list_filter=('is_active',), ) ## Instruction: Add search fields for newsletter subscriptions ## Code After: from django.contrib import admin from .models import NewsletterSubscription admin.site.register(NewsletterSubscription, list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'), list_filter=('is_active',), search_fields=('code', 'email', 'first_name', 'last_name'), )
from django.contrib import admin from .models import NewsletterSubscription admin.site.register(NewsletterSubscription, list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'), list_filter=('is_active',), + search_fields=('code', 'email', 'first_name', 'last_name'), )
c01a858306d31a5b12e42f30ff01bdbdb2240092
froide/publicbody/tests.py
froide/publicbody/tests.py
from django.test import TestCase class SimpleTest(TestCase): def test_basic_addition(self): """ Tests that 1 + 1 always equals 2. """ self.assertEqual(1 + 1, 2)
from django.test import TestCase from django.core.urlresolvers import reverse from publicbody.models import PublicBody class PublicBodyTest(TestCase): fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json'] def test_web_page(self): response = self.client.get(reverse('publicbody-list')) self.assertEqual(response.status_code, 200) pb = PublicBody.objects.all()[0] response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug})) self.assertEqual(response.status_code, 200) response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"})) self.assertEqual(response.status_code, 200) self.assertEqual(response['Content-Type'], 'application/json') self.assertIn('"name":', response.content) self.assertIn('"laws": [{', response.content) response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"})) self.assertEqual(response.status_code, 200) self.assertEqual(response['Content-Type'], 'application/json') def test_csv(self): csv = PublicBody.export_csv() self.assertTrue(csv) def test_search(self): response = self.client.get(reverse('publicbody-search')+"?q=umwelt") self.assertIn("Umweltbundesamt", response.content) self.assertEqual(response['Content-Type'], 'application/json')
Test public body showing, json view and csv export
Test public body showing, json view and csv export
Python
mit
okfse/froide,ryankanno/froide,catcosmo/froide,ryankanno/froide,okfse/froide,LilithWittmann/froide,okfse/froide,LilithWittmann/froide,ryankanno/froide,CodeforHawaii/froide,stefanw/froide,stefanw/froide,LilithWittmann/froide,CodeforHawaii/froide,catcosmo/froide,catcosmo/froide,stefanw/froide,ryankanno/froide,fin/froide,fin/froide,catcosmo/froide,okfse/froide,fin/froide,LilithWittmann/froide,stefanw/froide,catcosmo/froide,ryankanno/froide,LilithWittmann/froide,fin/froide,stefanw/froide,CodeforHawaii/froide,CodeforHawaii/froide,CodeforHawaii/froide,okfse/froide
+ from django.test import TestCase + from django.core.urlresolvers import reverse - from django.test import TestCase + from publicbody.models import PublicBody + class PublicBodyTest(TestCase): + fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json'] - class SimpleTest(TestCase): - def test_basic_addition(self): - """ - Tests that 1 + 1 always equals 2. - """ - self.assertEqual(1 + 1, 2) + def test_web_page(self): + response = self.client.get(reverse('publicbody-list')) + self.assertEqual(response.status_code, 200) + pb = PublicBody.objects.all()[0] + response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug})) + self.assertEqual(response.status_code, 200) + response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"})) + self.assertEqual(response.status_code, 200) + self.assertEqual(response['Content-Type'], 'application/json') + self.assertIn('"name":', response.content) + self.assertIn('"laws": [{', response.content) + response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"})) + self.assertEqual(response.status_code, 200) + self.assertEqual(response['Content-Type'], 'application/json') + def test_csv(self): + csv = PublicBody.export_csv() + self.assertTrue(csv) + + def test_search(self): + response = self.client.get(reverse('publicbody-search')+"?q=umwelt") + self.assertIn("Umweltbundesamt", response.content) + self.assertEqual(response['Content-Type'], 'application/json') +
Test public body showing, json view and csv export
## Code Before: from django.test import TestCase class SimpleTest(TestCase): def test_basic_addition(self): """ Tests that 1 + 1 always equals 2. """ self.assertEqual(1 + 1, 2) ## Instruction: Test public body showing, json view and csv export ## Code After: from django.test import TestCase from django.core.urlresolvers import reverse from publicbody.models import PublicBody class PublicBodyTest(TestCase): fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json'] def test_web_page(self): response = self.client.get(reverse('publicbody-list')) self.assertEqual(response.status_code, 200) pb = PublicBody.objects.all()[0] response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug})) self.assertEqual(response.status_code, 200) response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"})) self.assertEqual(response.status_code, 200) self.assertEqual(response['Content-Type'], 'application/json') self.assertIn('"name":', response.content) self.assertIn('"laws": [{', response.content) response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"})) self.assertEqual(response.status_code, 200) self.assertEqual(response['Content-Type'], 'application/json') def test_csv(self): csv = PublicBody.export_csv() self.assertTrue(csv) def test_search(self): response = self.client.get(reverse('publicbody-search')+"?q=umwelt") self.assertIn("Umweltbundesamt", response.content) self.assertEqual(response['Content-Type'], 'application/json')
+ from django.test import TestCase + from django.core.urlresolvers import reverse - from django.test import TestCase + from publicbody.models import PublicBody + class PublicBodyTest(TestCase): + fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json'] - class SimpleTest(TestCase): - def test_basic_addition(self): - """ - Tests that 1 + 1 always equals 2. - """ - self.assertEqual(1 + 1, 2) + def test_web_page(self): + response = self.client.get(reverse('publicbody-list')) + self.assertEqual(response.status_code, 200) + pb = PublicBody.objects.all()[0] + response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug})) + self.assertEqual(response.status_code, 200) + response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"})) + self.assertEqual(response.status_code, 200) + self.assertEqual(response['Content-Type'], 'application/json') + self.assertIn('"name":', response.content) + self.assertIn('"laws": [{', response.content) + response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"})) + self.assertEqual(response.status_code, 200) + self.assertEqual(response['Content-Type'], 'application/json') + + def test_csv(self): + csv = PublicBody.export_csv() + self.assertTrue(csv) + + def test_search(self): + response = self.client.get(reverse('publicbody-search')+"?q=umwelt") + self.assertIn("Umweltbundesamt", response.content) + self.assertEqual(response['Content-Type'], 'application/json')
d154cd852bdb02743e9752179559a91b9f1a7f8c
example/tests/unit/test_renderer_class_methods.py
example/tests/unit/test_renderer_class_methods.py
from django.contrib.auth import get_user_model from rest_framework_json_api import serializers from rest_framework_json_api.renderers import JSONRenderer class ResourceSerializer(serializers.ModelSerializer): class Meta: fields = ('username',) model = get_user_model() def test_build_json_resource_obj(): resource = { 'pk': 1, 'username': 'Alice', } serializer = ResourceSerializer(data={'username': 'Alice'}) serializer.is_valid() resource_instance = serializer.save() output = { 'type': 'user', 'id': '1', 'attributes': { 'username': 'Alice' }, } assert JSONRenderer.build_json_resource_obj( serializer.fields, resource, resource_instance, 'user') == output def test_extract_attributes(): fields = { 'id': serializers.Field(), 'username': serializers.Field(), 'deleted': serializers.ReadOnlyField(), } resource = {'id': 1, 'deleted': None, 'username': 'jerel'} expected = { 'username': 'jerel', 'deleted': None } assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted' assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted( {'username': ''}), 'Should not extract read_only fields on empty serializer'
from django.contrib.auth import get_user_model from rest_framework_json_api import serializers from rest_framework_json_api.renderers import JSONRenderer pytestmark = pytest.mark.django_db class ResourceSerializer(serializers.ModelSerializer): class Meta: fields = ('username',) model = get_user_model() def test_build_json_resource_obj(): resource = { 'pk': 1, 'username': 'Alice', } serializer = ResourceSerializer(data={'username': 'Alice'}) serializer.is_valid() resource_instance = serializer.save() output = { 'type': 'user', 'id': '1', 'attributes': { 'username': 'Alice' }, } assert JSONRenderer.build_json_resource_obj( serializer.fields, resource, resource_instance, 'user') == output def test_extract_attributes(): fields = { 'id': serializers.Field(), 'username': serializers.Field(), 'deleted': serializers.ReadOnlyField(), } resource = {'id': 1, 'deleted': None, 'username': 'jerel'} expected = { 'username': 'jerel', 'deleted': None } assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted' assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted( {'username': ''}), 'Should not extract read_only fields on empty serializer'
Fix for Database access not allowed, use the "django_db" mark to enable it.
Fix for Database access not allowed, use the "django_db" mark to enable it.
Python
bsd-2-clause
django-json-api/django-rest-framework-json-api,martinmaillard/django-rest-framework-json-api,schtibe/django-rest-framework-json-api,pombredanne/django-rest-framework-json-api,scottfisk/django-rest-framework-json-api,Instawork/django-rest-framework-json-api,leo-naeka/rest_framework_ember,django-json-api/django-rest-framework-json-api,leo-naeka/django-rest-framework-json-api,abdulhaq-e/django-rest-framework-json-api,lukaslundgren/django-rest-framework-json-api,django-json-api/rest_framework_ember
from django.contrib.auth import get_user_model from rest_framework_json_api import serializers from rest_framework_json_api.renderers import JSONRenderer + pytestmark = pytest.mark.django_db class ResourceSerializer(serializers.ModelSerializer): class Meta: fields = ('username',) model = get_user_model() def test_build_json_resource_obj(): resource = { 'pk': 1, 'username': 'Alice', } serializer = ResourceSerializer(data={'username': 'Alice'}) serializer.is_valid() resource_instance = serializer.save() output = { 'type': 'user', 'id': '1', 'attributes': { 'username': 'Alice' }, } assert JSONRenderer.build_json_resource_obj( serializer.fields, resource, resource_instance, 'user') == output def test_extract_attributes(): fields = { 'id': serializers.Field(), 'username': serializers.Field(), 'deleted': serializers.ReadOnlyField(), } resource = {'id': 1, 'deleted': None, 'username': 'jerel'} expected = { 'username': 'jerel', 'deleted': None } assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted' assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted( {'username': ''}), 'Should not extract read_only fields on empty serializer'
Fix for Database access not allowed, use the "django_db" mark to enable it.
## Code Before: from django.contrib.auth import get_user_model from rest_framework_json_api import serializers from rest_framework_json_api.renderers import JSONRenderer class ResourceSerializer(serializers.ModelSerializer): class Meta: fields = ('username',) model = get_user_model() def test_build_json_resource_obj(): resource = { 'pk': 1, 'username': 'Alice', } serializer = ResourceSerializer(data={'username': 'Alice'}) serializer.is_valid() resource_instance = serializer.save() output = { 'type': 'user', 'id': '1', 'attributes': { 'username': 'Alice' }, } assert JSONRenderer.build_json_resource_obj( serializer.fields, resource, resource_instance, 'user') == output def test_extract_attributes(): fields = { 'id': serializers.Field(), 'username': serializers.Field(), 'deleted': serializers.ReadOnlyField(), } resource = {'id': 1, 'deleted': None, 'username': 'jerel'} expected = { 'username': 'jerel', 'deleted': None } assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted' assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted( {'username': ''}), 'Should not extract read_only fields on empty serializer' ## Instruction: Fix for Database access not allowed, use the "django_db" mark to enable it. ## Code After: from django.contrib.auth import get_user_model from rest_framework_json_api import serializers from rest_framework_json_api.renderers import JSONRenderer pytestmark = pytest.mark.django_db class ResourceSerializer(serializers.ModelSerializer): class Meta: fields = ('username',) model = get_user_model() def test_build_json_resource_obj(): resource = { 'pk': 1, 'username': 'Alice', } serializer = ResourceSerializer(data={'username': 'Alice'}) serializer.is_valid() resource_instance = serializer.save() output = { 'type': 'user', 'id': '1', 'attributes': { 'username': 'Alice' }, } assert JSONRenderer.build_json_resource_obj( serializer.fields, resource, resource_instance, 'user') == output def test_extract_attributes(): fields = { 'id': serializers.Field(), 'username': serializers.Field(), 'deleted': serializers.ReadOnlyField(), } resource = {'id': 1, 'deleted': None, 'username': 'jerel'} expected = { 'username': 'jerel', 'deleted': None } assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted' assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted( {'username': ''}), 'Should not extract read_only fields on empty serializer'
from django.contrib.auth import get_user_model from rest_framework_json_api import serializers from rest_framework_json_api.renderers import JSONRenderer + pytestmark = pytest.mark.django_db class ResourceSerializer(serializers.ModelSerializer): class Meta: fields = ('username',) model = get_user_model() def test_build_json_resource_obj(): resource = { 'pk': 1, 'username': 'Alice', } serializer = ResourceSerializer(data={'username': 'Alice'}) serializer.is_valid() resource_instance = serializer.save() output = { 'type': 'user', 'id': '1', 'attributes': { 'username': 'Alice' }, } assert JSONRenderer.build_json_resource_obj( serializer.fields, resource, resource_instance, 'user') == output def test_extract_attributes(): fields = { 'id': serializers.Field(), 'username': serializers.Field(), 'deleted': serializers.ReadOnlyField(), } resource = {'id': 1, 'deleted': None, 'username': 'jerel'} expected = { 'username': 'jerel', 'deleted': None } assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted' assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted( {'username': ''}), 'Should not extract read_only fields on empty serializer'
64dbe1d931edd38b4d731db18408e337d39e42c3
cab/admin.py
cab/admin.py
from django.contrib import admin from cab.models import Language, Snippet, SnippetFlag class LanguageAdmin(admin.ModelAdmin): prepopulated_fields = {'slug': ['name']} class SnippetAdmin(admin.ModelAdmin): list_display = ('id', 'title', 'author', 'rating_score', 'pub_date') list_filter = ('language',) date_hierarchy = 'pub_date' search_fields = ('author__username', 'title', 'description', 'code',) class SnippetFlagAdmin(admin.ModelAdmin): list_display = ('snippet', 'flag') list_filter = ('flag',) actions = ['remove_and_ban'] def remove_and_ban(self, request, queryset): for obj in queryset: obj.remove_and_ban() self.message_user(request, 'Snippets removed successfully') remove_and_ban.short_description = 'Remove snippet and ban user' admin.site.register(Language, LanguageAdmin) admin.site.register(Snippet, SnippetAdmin) admin.site.register(SnippetFlag, SnippetFlagAdmin)
from django.contrib import admin from cab.models import Language, Snippet, SnippetFlag class LanguageAdmin(admin.ModelAdmin): prepopulated_fields = {'slug': ['name']} class SnippetAdmin(admin.ModelAdmin): list_display = ('id', 'title', 'author', 'rating_score', 'pub_date') list_filter = ('language',) date_hierarchy = 'pub_date' search_fields = ('author__username', 'title', 'description', 'code',) raw_id_fields = ('author',) class SnippetFlagAdmin(admin.ModelAdmin): list_display = ('snippet', 'flag') list_filter = ('flag',) actions = ['remove_and_ban'] raw_id_fields = ('snippet', 'user',) def remove_and_ban(self, request, queryset): for obj in queryset: obj.remove_and_ban() self.message_user(request, 'Snippets removed successfully') remove_and_ban.short_description = 'Remove snippet and ban user' admin.site.register(Language, LanguageAdmin) admin.site.register(Snippet, SnippetAdmin) admin.site.register(SnippetFlag, SnippetFlagAdmin)
Use raw_id_fields for users and snippets.
Use raw_id_fields for users and snippets.
Python
bsd-3-clause
django/djangosnippets.org,django/djangosnippets.org,django/djangosnippets.org,django/djangosnippets.org,django-de/djangosnippets.org,django/djangosnippets.org,django-de/djangosnippets.org,django-de/djangosnippets.org,django-de/djangosnippets.org
from django.contrib import admin from cab.models import Language, Snippet, SnippetFlag class LanguageAdmin(admin.ModelAdmin): prepopulated_fields = {'slug': ['name']} class SnippetAdmin(admin.ModelAdmin): list_display = ('id', 'title', 'author', 'rating_score', 'pub_date') list_filter = ('language',) date_hierarchy = 'pub_date' search_fields = ('author__username', 'title', 'description', 'code',) + raw_id_fields = ('author',) class SnippetFlagAdmin(admin.ModelAdmin): list_display = ('snippet', 'flag') list_filter = ('flag',) actions = ['remove_and_ban'] + raw_id_fields = ('snippet', 'user',) def remove_and_ban(self, request, queryset): for obj in queryset: obj.remove_and_ban() self.message_user(request, 'Snippets removed successfully') remove_and_ban.short_description = 'Remove snippet and ban user' admin.site.register(Language, LanguageAdmin) admin.site.register(Snippet, SnippetAdmin) admin.site.register(SnippetFlag, SnippetFlagAdmin)
Use raw_id_fields for users and snippets.
## Code Before: from django.contrib import admin from cab.models import Language, Snippet, SnippetFlag class LanguageAdmin(admin.ModelAdmin): prepopulated_fields = {'slug': ['name']} class SnippetAdmin(admin.ModelAdmin): list_display = ('id', 'title', 'author', 'rating_score', 'pub_date') list_filter = ('language',) date_hierarchy = 'pub_date' search_fields = ('author__username', 'title', 'description', 'code',) class SnippetFlagAdmin(admin.ModelAdmin): list_display = ('snippet', 'flag') list_filter = ('flag',) actions = ['remove_and_ban'] def remove_and_ban(self, request, queryset): for obj in queryset: obj.remove_and_ban() self.message_user(request, 'Snippets removed successfully') remove_and_ban.short_description = 'Remove snippet and ban user' admin.site.register(Language, LanguageAdmin) admin.site.register(Snippet, SnippetAdmin) admin.site.register(SnippetFlag, SnippetFlagAdmin) ## Instruction: Use raw_id_fields for users and snippets. ## Code After: from django.contrib import admin from cab.models import Language, Snippet, SnippetFlag class LanguageAdmin(admin.ModelAdmin): prepopulated_fields = {'slug': ['name']} class SnippetAdmin(admin.ModelAdmin): list_display = ('id', 'title', 'author', 'rating_score', 'pub_date') list_filter = ('language',) date_hierarchy = 'pub_date' search_fields = ('author__username', 'title', 'description', 'code',) raw_id_fields = ('author',) class SnippetFlagAdmin(admin.ModelAdmin): list_display = ('snippet', 'flag') list_filter = ('flag',) actions = ['remove_and_ban'] raw_id_fields = ('snippet', 'user',) def remove_and_ban(self, request, queryset): for obj in queryset: obj.remove_and_ban() self.message_user(request, 'Snippets removed successfully') remove_and_ban.short_description = 'Remove snippet and ban user' admin.site.register(Language, LanguageAdmin) admin.site.register(Snippet, SnippetAdmin) admin.site.register(SnippetFlag, SnippetFlagAdmin)
from django.contrib import admin from cab.models import Language, Snippet, SnippetFlag class LanguageAdmin(admin.ModelAdmin): prepopulated_fields = {'slug': ['name']} class SnippetAdmin(admin.ModelAdmin): list_display = ('id', 'title', 'author', 'rating_score', 'pub_date') list_filter = ('language',) date_hierarchy = 'pub_date' search_fields = ('author__username', 'title', 'description', 'code',) + raw_id_fields = ('author',) class SnippetFlagAdmin(admin.ModelAdmin): list_display = ('snippet', 'flag') list_filter = ('flag',) actions = ['remove_and_ban'] + raw_id_fields = ('snippet', 'user',) def remove_and_ban(self, request, queryset): for obj in queryset: obj.remove_and_ban() self.message_user(request, 'Snippets removed successfully') remove_and_ban.short_description = 'Remove snippet and ban user' admin.site.register(Language, LanguageAdmin) admin.site.register(Snippet, SnippetAdmin) admin.site.register(SnippetFlag, SnippetFlagAdmin)
fdf0daefac50de71a8c4f80a9ef877669ebea48b
byceps/services/tourney/transfer/models.py
byceps/services/tourney/transfer/models.py
from typing import NewType from uuid import UUID from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @attrs(auto_attribs=True, frozen=True, slots=True) class Match: id: MatchID
from dataclasses import dataclass from typing import NewType from uuid import UUID TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @dataclass(frozen=True) class Match: id: MatchID
Change tourney match transfer model from `attrs` to `dataclass`
Change tourney match transfer model from `attrs` to `dataclass`
Python
bsd-3-clause
homeworkprod/byceps,homeworkprod/byceps,m-ober/byceps,homeworkprod/byceps,m-ober/byceps,m-ober/byceps
+ from dataclasses import dataclass from typing import NewType from uuid import UUID - - from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) - @attrs(auto_attribs=True, frozen=True, slots=True) + @dataclass(frozen=True) class Match: id: MatchID
Change tourney match transfer model from `attrs` to `dataclass`
## Code Before: from typing import NewType from uuid import UUID from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @attrs(auto_attribs=True, frozen=True, slots=True) class Match: id: MatchID ## Instruction: Change tourney match transfer model from `attrs` to `dataclass` ## Code After: from dataclasses import dataclass from typing import NewType from uuid import UUID TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @dataclass(frozen=True) class Match: id: MatchID
+ from dataclasses import dataclass from typing import NewType from uuid import UUID - - from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) - @attrs(auto_attribs=True, frozen=True, slots=True) + @dataclass(frozen=True) class Match: id: MatchID
899f28e2cd7dbeb6227e8c56eef541cce1a424f4
alertaclient/commands/cmd_heartbeat.py
alertaclient/commands/cmd_heartbeat.py
import os import platform import sys import click prog = os.path.basename(sys.argv[0]) @click.command('heartbeat', short_help='Send a heartbeat') @click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1])) @click.option('--tag', '-T', 'tags', multiple=True) @click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale') @click.option('--delete', '-D', metavar='ID', help='Delete hearbeat') @click.pass_obj def cli(obj, origin, tags, timeout, delete): """Send or delete a heartbeat.""" client = obj['client'] if delete: if origin or tags or timeout: raise click.UsageError('Option "--delete" is mutually exclusive.') client.delete_heartbeat(delete) else: try: heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout) except Exception as e: click.echo('ERROR: {}'.format(e)) sys.exit(1) click.echo(heartbeat.id)
import os import platform import sys import click prog = os.path.basename(sys.argv[0]) @click.command('heartbeat', short_help='Send a heartbeat') @click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1])) @click.option('--tag', '-T', 'tags', multiple=True) @click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale') @click.option('--delete', '-D', metavar='ID', help='Delete hearbeat') @click.pass_obj def cli(obj, origin, tags, timeout, delete): """Send or delete a heartbeat.""" client = obj['client'] if delete: if origin or tags or timeout: raise click.UsageError('Option "--delete" is mutually exclusive.') client.delete_heartbeat(delete) else: try: heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout) except Exception as e: click.echo('ERROR: {}'.format(e)) sys.exit(1) click.echo(heartbeat.id)
Add check that heartbeat timeout is integer
Add check that heartbeat timeout is integer
Python
apache-2.0
alerta/python-alerta-client,alerta/python-alerta-client,alerta/python-alerta
import os import platform import sys import click prog = os.path.basename(sys.argv[0]) @click.command('heartbeat', short_help='Send a heartbeat') @click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1])) @click.option('--tag', '-T', 'tags', multiple=True) - @click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale') + @click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale') @click.option('--delete', '-D', metavar='ID', help='Delete hearbeat') @click.pass_obj def cli(obj, origin, tags, timeout, delete): """Send or delete a heartbeat.""" client = obj['client'] if delete: if origin or tags or timeout: raise click.UsageError('Option "--delete" is mutually exclusive.') client.delete_heartbeat(delete) else: try: heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout) except Exception as e: click.echo('ERROR: {}'.format(e)) sys.exit(1) click.echo(heartbeat.id)
Add check that heartbeat timeout is integer
## Code Before: import os import platform import sys import click prog = os.path.basename(sys.argv[0]) @click.command('heartbeat', short_help='Send a heartbeat') @click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1])) @click.option('--tag', '-T', 'tags', multiple=True) @click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale') @click.option('--delete', '-D', metavar='ID', help='Delete hearbeat') @click.pass_obj def cli(obj, origin, tags, timeout, delete): """Send or delete a heartbeat.""" client = obj['client'] if delete: if origin or tags or timeout: raise click.UsageError('Option "--delete" is mutually exclusive.') client.delete_heartbeat(delete) else: try: heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout) except Exception as e: click.echo('ERROR: {}'.format(e)) sys.exit(1) click.echo(heartbeat.id) ## Instruction: Add check that heartbeat timeout is integer ## Code After: import os import platform import sys import click prog = os.path.basename(sys.argv[0]) @click.command('heartbeat', short_help='Send a heartbeat') @click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1])) @click.option('--tag', '-T', 'tags', multiple=True) @click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale') @click.option('--delete', '-D', metavar='ID', help='Delete hearbeat') @click.pass_obj def cli(obj, origin, tags, timeout, delete): """Send or delete a heartbeat.""" client = obj['client'] if delete: if origin or tags or timeout: raise click.UsageError('Option "--delete" is mutually exclusive.') client.delete_heartbeat(delete) else: try: heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout) except Exception as e: click.echo('ERROR: {}'.format(e)) sys.exit(1) click.echo(heartbeat.id)
import os import platform import sys import click prog = os.path.basename(sys.argv[0]) @click.command('heartbeat', short_help='Send a heartbeat') @click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1])) @click.option('--tag', '-T', 'tags', multiple=True) - @click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale') + @click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale') ? ++++++++++ @click.option('--delete', '-D', metavar='ID', help='Delete hearbeat') @click.pass_obj def cli(obj, origin, tags, timeout, delete): """Send or delete a heartbeat.""" client = obj['client'] if delete: if origin or tags or timeout: raise click.UsageError('Option "--delete" is mutually exclusive.') client.delete_heartbeat(delete) else: try: heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout) except Exception as e: click.echo('ERROR: {}'.format(e)) sys.exit(1) click.echo(heartbeat.id)
bd7c0a9ac2d357ab635bf2948824256f1e6ddbec
src/carreralib/serial.py
src/carreralib/serial.py
from serial import serial_for_url from .connection import BufferTooShort, Connection, TimeoutError class SerialConnection(Connection): def __init__(self, url, timeout=None): self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout) def close(self): self.__serial.close() def recv(self, maxlength=None): buf = bytearray() while True: c = self.__serial.read() if not c: raise TimeoutError("Timeout waiting for serial data") elif c == b"$" or c == b"#": break elif maxlength is not None and maxlength <= len(buf): raise BufferTooShort("Buffer too short for data received") else: buf.extend(c) return bytes(buf) def send(self, buf, offset=0, size=None): n = len(buf) if offset < 0: raise ValueError("offset is negative") elif n < offset: raise ValueError("buffer length < offset") elif size is None: size = n - offset elif size < 0: raise ValueError("size is negative") elif offset + size > n: raise ValueError("buffer length < offset + size") self.__serial.write(b'"') self.__serial.write(buf[offset : offset + size]) self.__serial.write(b"$") self.__serial.flush() @classmethod def scan(_): from serial.tools.list_ports import comports return ((info.device, info.description) for info in comports())
from serial import serial_for_url from .connection import BufferTooShort, Connection, TimeoutError class SerialConnection(Connection): __serial = None def __init__(self, url, timeout=None): self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout) def close(self): if self.__serial: self.__serial.close() def recv(self, maxlength=None): buf = bytearray() while True: c = self.__serial.read() if not c: raise TimeoutError("Timeout waiting for serial data") elif c == b"$" or c == b"#": break elif maxlength is not None and maxlength <= len(buf): raise BufferTooShort("Buffer too short for data received") else: buf.extend(c) return bytes(buf) def send(self, buf, offset=0, size=None): n = len(buf) if offset < 0: raise ValueError("offset is negative") elif n < offset: raise ValueError("buffer length < offset") elif size is None: size = n - offset elif size < 0: raise ValueError("size is negative") elif offset + size > n: raise ValueError("buffer length < offset + size") self.__serial.write(b'"') self.__serial.write(buf[offset : offset + size]) self.__serial.write(b"$") self.__serial.flush() @classmethod def scan(_): from serial.tools.list_ports import comports return ((info.device, info.description) for info in comports())
Fix SerialConnection.close() with invalid device.
Fix SerialConnection.close() with invalid device.
Python
mit
tkem/carreralib
from serial import serial_for_url from .connection import BufferTooShort, Connection, TimeoutError class SerialConnection(Connection): + + __serial = None + def __init__(self, url, timeout=None): self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout) def close(self): + if self.__serial: - self.__serial.close() + self.__serial.close() def recv(self, maxlength=None): buf = bytearray() while True: c = self.__serial.read() if not c: raise TimeoutError("Timeout waiting for serial data") elif c == b"$" or c == b"#": break elif maxlength is not None and maxlength <= len(buf): raise BufferTooShort("Buffer too short for data received") else: buf.extend(c) return bytes(buf) def send(self, buf, offset=0, size=None): n = len(buf) if offset < 0: raise ValueError("offset is negative") elif n < offset: raise ValueError("buffer length < offset") elif size is None: size = n - offset elif size < 0: raise ValueError("size is negative") elif offset + size > n: raise ValueError("buffer length < offset + size") self.__serial.write(b'"') self.__serial.write(buf[offset : offset + size]) self.__serial.write(b"$") self.__serial.flush() @classmethod def scan(_): from serial.tools.list_ports import comports return ((info.device, info.description) for info in comports())
Fix SerialConnection.close() with invalid device.
## Code Before: from serial import serial_for_url from .connection import BufferTooShort, Connection, TimeoutError class SerialConnection(Connection): def __init__(self, url, timeout=None): self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout) def close(self): self.__serial.close() def recv(self, maxlength=None): buf = bytearray() while True: c = self.__serial.read() if not c: raise TimeoutError("Timeout waiting for serial data") elif c == b"$" or c == b"#": break elif maxlength is not None and maxlength <= len(buf): raise BufferTooShort("Buffer too short for data received") else: buf.extend(c) return bytes(buf) def send(self, buf, offset=0, size=None): n = len(buf) if offset < 0: raise ValueError("offset is negative") elif n < offset: raise ValueError("buffer length < offset") elif size is None: size = n - offset elif size < 0: raise ValueError("size is negative") elif offset + size > n: raise ValueError("buffer length < offset + size") self.__serial.write(b'"') self.__serial.write(buf[offset : offset + size]) self.__serial.write(b"$") self.__serial.flush() @classmethod def scan(_): from serial.tools.list_ports import comports return ((info.device, info.description) for info in comports()) ## Instruction: Fix SerialConnection.close() with invalid device. ## Code After: from serial import serial_for_url from .connection import BufferTooShort, Connection, TimeoutError class SerialConnection(Connection): __serial = None def __init__(self, url, timeout=None): self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout) def close(self): if self.__serial: self.__serial.close() def recv(self, maxlength=None): buf = bytearray() while True: c = self.__serial.read() if not c: raise TimeoutError("Timeout waiting for serial data") elif c == b"$" or c == b"#": break elif maxlength is not None and maxlength <= len(buf): raise BufferTooShort("Buffer too short for data received") else: buf.extend(c) return bytes(buf) def send(self, buf, offset=0, size=None): n = len(buf) if offset < 0: raise ValueError("offset is negative") elif n < offset: raise ValueError("buffer length < offset") elif size is None: size = n - offset elif size < 0: raise ValueError("size is negative") elif offset + size > n: raise ValueError("buffer length < offset + size") self.__serial.write(b'"') self.__serial.write(buf[offset : offset + size]) self.__serial.write(b"$") self.__serial.flush() @classmethod def scan(_): from serial.tools.list_ports import comports return ((info.device, info.description) for info in comports())
from serial import serial_for_url from .connection import BufferTooShort, Connection, TimeoutError class SerialConnection(Connection): + + __serial = None + def __init__(self, url, timeout=None): self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout) def close(self): + if self.__serial: - self.__serial.close() + self.__serial.close() ? ++++ def recv(self, maxlength=None): buf = bytearray() while True: c = self.__serial.read() if not c: raise TimeoutError("Timeout waiting for serial data") elif c == b"$" or c == b"#": break elif maxlength is not None and maxlength <= len(buf): raise BufferTooShort("Buffer too short for data received") else: buf.extend(c) return bytes(buf) def send(self, buf, offset=0, size=None): n = len(buf) if offset < 0: raise ValueError("offset is negative") elif n < offset: raise ValueError("buffer length < offset") elif size is None: size = n - offset elif size < 0: raise ValueError("size is negative") elif offset + size > n: raise ValueError("buffer length < offset + size") self.__serial.write(b'"') self.__serial.write(buf[offset : offset + size]) self.__serial.write(b"$") self.__serial.flush() @classmethod def scan(_): from serial.tools.list_ports import comports return ((info.device, info.description) for info in comports())
73d22cc63a2a37bd3c99774bf098ca12c81d54ae
funnels.py
funnels.py
import pyglet from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels window = pyglet.window.Window()#fullscreen=True) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run()
import pyglet import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels def main(fullscreen): window = pyglet.window.Window(fullscreen=fullscreen) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() if __name__ == '__main__': parser = argparse.ArgumentParser(description="Arithemetic practice game.") parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') parser.set_defaults(fullscreen=True) results = parser.parse_args() main(results.fullscreen)
Add argparse to turn on/off fullscreen behavior
Add argparse to turn on/off fullscreen behavior
Python
mit
simeonf/claire
import pyglet - + import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels + def main(fullscreen): - window = pyglet.window.Window()#fullscreen=True) + window = pyglet.window.Window(fullscreen=fullscreen) - levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) + levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) - pyglet.clock.schedule(levels.clock) + pyglet.clock.schedule(levels.clock) - @window.event + @window.event - def on_key_press(symbol, modifiers): + def on_key_press(symbol, modifiers): - levels.key(symbol, modifiers) + levels.key(symbol, modifiers) - @window.event + @window.event - def on_draw(): + def on_draw(): - levels.draw() + levels.draw() - pyglet.app.run() + pyglet.app.run() + if __name__ == '__main__': + parser = argparse.ArgumentParser(description="Arithemetic practice game.") + parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') + parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') + parser.set_defaults(fullscreen=True) + results = parser.parse_args() + main(results.fullscreen) +
Add argparse to turn on/off fullscreen behavior
## Code Before: import pyglet from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels window = pyglet.window.Window()#fullscreen=True) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() ## Instruction: Add argparse to turn on/off fullscreen behavior ## Code After: import pyglet import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels def main(fullscreen): window = pyglet.window.Window(fullscreen=fullscreen) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() if __name__ == '__main__': parser = argparse.ArgumentParser(description="Arithemetic practice game.") parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') parser.set_defaults(fullscreen=True) results = parser.parse_args() main(results.fullscreen)
import pyglet - + import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels + def main(fullscreen): - window = pyglet.window.Window()#fullscreen=True) ? -- ^ - + window = pyglet.window.Window(fullscreen=fullscreen) ? ++ ^^^^^^ ++ - levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) + levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) ? ++ - pyglet.clock.schedule(levels.clock) + pyglet.clock.schedule(levels.clock) ? ++ - @window.event + @window.event ? ++ - def on_key_press(symbol, modifiers): + def on_key_press(symbol, modifiers): ? ++ - levels.key(symbol, modifiers) + levels.key(symbol, modifiers) ? ++ - @window.event + @window.event ? ++ - def on_draw(): + def on_draw(): ? ++ - levels.draw() + levels.draw() ? ++ - pyglet.app.run() + pyglet.app.run() ? ++ + + if __name__ == '__main__': + parser = argparse.ArgumentParser(description="Arithemetic practice game.") + parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') + parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') + parser.set_defaults(fullscreen=True) + results = parser.parse_args() + main(results.fullscreen)
6b84688c1b5a7f2e8c9e5007455b88cbaa845e9f
tests/test_track_output/results.py
tests/test_track_output/results.py
import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../track.py', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
Fix path to script in test_track_output
Fix path to script in test_track_output
Python
mit
mjlong/openmc,wbinventor/openmc,bhermanmit/openmc,wbinventor/openmc,johnnyliu27/openmc,paulromano/openmc,mjlong/openmc,smharper/openmc,lilulu/openmc,johnnyliu27/openmc,amandalund/openmc,samuelshaner/openmc,mit-crpg/openmc,shikhar413/openmc,johnnyliu27/openmc,mit-crpg/openmc,liangjg/openmc,kellyrowland/openmc,amandalund/openmc,walshjon/openmc,lilulu/openmc,liangjg/openmc,liangjg/openmc,johnnyliu27/openmc,samuelshaner/openmc,mit-crpg/openmc,amandalund/openmc,smharper/openmc,shikhar413/openmc,samuelshaner/openmc,mit-crpg/openmc,shikhar413/openmc,paulromano/openmc,liangjg/openmc,smharper/openmc,shikhar413/openmc,samuelshaner/openmc,paulromano/openmc,walshjon/openmc,walshjon/openmc,smharper/openmc,bhermanmit/openmc,wbinventor/openmc,wbinventor/openmc,walshjon/openmc,paulromano/openmc,lilulu/openmc,amandalund/openmc,kellyrowland/openmc
import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script - call(['../../track.py', '-o', 'poly'] + + call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
Fix path to script in test_track_output
## Code Before: import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../track.py', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat') ## Instruction: Fix path to script in test_track_output ## Code After: import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script - call(['../../track.py', '-o', 'poly'] + + call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
6131430ff7d1e9e8cd95f8d2793e82cc72679d81
auditlog/admin.py
auditlog/admin.py
from django.contrib import admin from .filters import ResourceTypeFilter from .mixins import LogEntryAdminMixin from .models import LogEntry class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin): list_display = ["created", "resource_url", "action", "msg_short", "user_url"] search_fields = [ "timestamp", "object_repr", "changes", "actor__first_name", "actor__last_name", ] list_filter = ["action", ResourceTypeFilter] readonly_fields = ["created", "resource_url", "action", "user_url", "msg"] fieldsets = [ (None, {"fields": ["created", "user_url", "resource_url"]}), ("Changes", {"fields": ["action", "msg"]}), ] admin.site.register(LogEntry, LogEntryAdmin)
from django.contrib import admin from auditlog.filters import ResourceTypeFilter from auditlog.mixins import LogEntryAdminMixin from auditlog.models import LogEntry class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin): list_display = ["created", "resource_url", "action", "msg_short", "user_url"] search_fields = [ "timestamp", "object_repr", "changes", "actor__first_name", "actor__last_name", ] list_filter = ["action", ResourceTypeFilter] readonly_fields = ["created", "resource_url", "action", "user_url", "msg"] fieldsets = [ (None, {"fields": ["created", "user_url", "resource_url"]}), ("Changes", {"fields": ["action", "msg"]}), ] admin.site.register(LogEntry, LogEntryAdmin)
Change relative imports to absolute.
Change relative imports to absolute.
Python
mit
jjkester/django-auditlog
from django.contrib import admin - from .filters import ResourceTypeFilter + from auditlog.filters import ResourceTypeFilter - from .mixins import LogEntryAdminMixin + from auditlog.mixins import LogEntryAdminMixin - from .models import LogEntry + from auditlog.models import LogEntry class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin): list_display = ["created", "resource_url", "action", "msg_short", "user_url"] search_fields = [ "timestamp", "object_repr", "changes", "actor__first_name", "actor__last_name", ] list_filter = ["action", ResourceTypeFilter] readonly_fields = ["created", "resource_url", "action", "user_url", "msg"] fieldsets = [ (None, {"fields": ["created", "user_url", "resource_url"]}), ("Changes", {"fields": ["action", "msg"]}), ] admin.site.register(LogEntry, LogEntryAdmin)
Change relative imports to absolute.
## Code Before: from django.contrib import admin from .filters import ResourceTypeFilter from .mixins import LogEntryAdminMixin from .models import LogEntry class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin): list_display = ["created", "resource_url", "action", "msg_short", "user_url"] search_fields = [ "timestamp", "object_repr", "changes", "actor__first_name", "actor__last_name", ] list_filter = ["action", ResourceTypeFilter] readonly_fields = ["created", "resource_url", "action", "user_url", "msg"] fieldsets = [ (None, {"fields": ["created", "user_url", "resource_url"]}), ("Changes", {"fields": ["action", "msg"]}), ] admin.site.register(LogEntry, LogEntryAdmin) ## Instruction: Change relative imports to absolute. ## Code After: from django.contrib import admin from auditlog.filters import ResourceTypeFilter from auditlog.mixins import LogEntryAdminMixin from auditlog.models import LogEntry class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin): list_display = ["created", "resource_url", "action", "msg_short", "user_url"] search_fields = [ "timestamp", "object_repr", "changes", "actor__first_name", "actor__last_name", ] list_filter = ["action", ResourceTypeFilter] readonly_fields = ["created", "resource_url", "action", "user_url", "msg"] fieldsets = [ (None, {"fields": ["created", "user_url", "resource_url"]}), ("Changes", {"fields": ["action", "msg"]}), ] admin.site.register(LogEntry, LogEntryAdmin)
from django.contrib import admin - from .filters import ResourceTypeFilter + from auditlog.filters import ResourceTypeFilter ? ++++++++ - from .mixins import LogEntryAdminMixin + from auditlog.mixins import LogEntryAdminMixin ? ++++++++ - from .models import LogEntry + from auditlog.models import LogEntry ? ++++++++ class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin): list_display = ["created", "resource_url", "action", "msg_short", "user_url"] search_fields = [ "timestamp", "object_repr", "changes", "actor__first_name", "actor__last_name", ] list_filter = ["action", ResourceTypeFilter] readonly_fields = ["created", "resource_url", "action", "user_url", "msg"] fieldsets = [ (None, {"fields": ["created", "user_url", "resource_url"]}), ("Changes", {"fields": ["action", "msg"]}), ] admin.site.register(LogEntry, LogEntryAdmin)
5d8a09ebff9cc8a8e8bdf4bff3963cee7a1aae6a
tools/skp/page_sets/skia_ebay_desktop.py
tools/skp/page_sets/skia_ebay_desktop.py
from telemetry import story from telemetry.page import page as page_module from telemetry.page import shared_page_state class SkiaDesktopPage(page_module.Page): def __init__(self, url, page_set): super(SkiaDesktopPage, self).__init__( url=url, name=url, page_set=page_set, shared_page_state_class=shared_page_state.SharedDesktopPageState) self.archive_data_file = 'data/skia_ebay_desktop.json' def RunNavigateSteps(self, action_runner): action_runner.Navigate(self.url) action_runner.Wait(15) class SkiaEbayDesktopPageSet(story.StorySet): """ Pages designed to represent the median, not highly optimized web """ def __init__(self): super(SkiaEbayDesktopPageSet, self).__init__( archive_data_file='data/skia_ebay_desktop.json') urls_list = [ # go/skia-skps-3-2019 'http://www.ebay.com', ] for url in urls_list: self.AddStory(SkiaDesktopPage(url, self))
from telemetry import story from telemetry.page import page as page_module from telemetry.page import shared_page_state class SkiaDesktopPage(page_module.Page): def __init__(self, url, page_set): super(SkiaDesktopPage, self).__init__( url=url, name=url, page_set=page_set, shared_page_state_class=shared_page_state.SharedDesktopPageState) self.archive_data_file = 'data/skia_ebay_desktop.json' def RunNavigateSteps(self, action_runner): action_runner.Navigate(self.url, timeout_in_seconds=120) class SkiaEbayDesktopPageSet(story.StorySet): """ Pages designed to represent the median, not highly optimized web """ def __init__(self): super(SkiaEbayDesktopPageSet, self).__init__( archive_data_file='data/skia_ebay_desktop.json') urls_list = [ # go/skia-skps-3-2019 'http://www.ebay.com', ] for url in urls_list: self.AddStory(SkiaDesktopPage(url, self))
Add wait time to ebay pageset
Add wait time to ebay pageset Bug: skia:11898 Change-Id: I0bb58f1d8e9c6ad48148d50b840f152fc158f071 Reviewed-on: https://skia-review.googlesource.com/c/skia/+/400538 Reviewed-by: Ravi Mistry <9fa2e7438b8cb730f96b74865492597170561628@google.com> Commit-Queue: Ravi Mistry <9fa2e7438b8cb730f96b74865492597170561628@google.com>
Python
bsd-3-clause
aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia,google/skia,aosp-mirror/platform_external_skia,google/skia,google/skia,aosp-mirror/platform_external_skia,google/skia,aosp-mirror/platform_external_skia,google/skia,google/skia,aosp-mirror/platform_external_skia,google/skia,google/skia,google/skia,google/skia,aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia
from telemetry import story from telemetry.page import page as page_module from telemetry.page import shared_page_state class SkiaDesktopPage(page_module.Page): def __init__(self, url, page_set): super(SkiaDesktopPage, self).__init__( url=url, name=url, page_set=page_set, shared_page_state_class=shared_page_state.SharedDesktopPageState) self.archive_data_file = 'data/skia_ebay_desktop.json' def RunNavigateSteps(self, action_runner): - action_runner.Navigate(self.url) + action_runner.Navigate(self.url, timeout_in_seconds=120) - action_runner.Wait(15) class SkiaEbayDesktopPageSet(story.StorySet): """ Pages designed to represent the median, not highly optimized web """ def __init__(self): super(SkiaEbayDesktopPageSet, self).__init__( archive_data_file='data/skia_ebay_desktop.json') urls_list = [ # go/skia-skps-3-2019 'http://www.ebay.com', ] for url in urls_list: self.AddStory(SkiaDesktopPage(url, self)) +
Add wait time to ebay pageset
## Code Before: from telemetry import story from telemetry.page import page as page_module from telemetry.page import shared_page_state class SkiaDesktopPage(page_module.Page): def __init__(self, url, page_set): super(SkiaDesktopPage, self).__init__( url=url, name=url, page_set=page_set, shared_page_state_class=shared_page_state.SharedDesktopPageState) self.archive_data_file = 'data/skia_ebay_desktop.json' def RunNavigateSteps(self, action_runner): action_runner.Navigate(self.url) action_runner.Wait(15) class SkiaEbayDesktopPageSet(story.StorySet): """ Pages designed to represent the median, not highly optimized web """ def __init__(self): super(SkiaEbayDesktopPageSet, self).__init__( archive_data_file='data/skia_ebay_desktop.json') urls_list = [ # go/skia-skps-3-2019 'http://www.ebay.com', ] for url in urls_list: self.AddStory(SkiaDesktopPage(url, self)) ## Instruction: Add wait time to ebay pageset ## Code After: from telemetry import story from telemetry.page import page as page_module from telemetry.page import shared_page_state class SkiaDesktopPage(page_module.Page): def __init__(self, url, page_set): super(SkiaDesktopPage, self).__init__( url=url, name=url, page_set=page_set, shared_page_state_class=shared_page_state.SharedDesktopPageState) self.archive_data_file = 'data/skia_ebay_desktop.json' def RunNavigateSteps(self, action_runner): action_runner.Navigate(self.url, timeout_in_seconds=120) class SkiaEbayDesktopPageSet(story.StorySet): """ Pages designed to represent the median, not highly optimized web """ def __init__(self): super(SkiaEbayDesktopPageSet, self).__init__( archive_data_file='data/skia_ebay_desktop.json') urls_list = [ # go/skia-skps-3-2019 'http://www.ebay.com', ] for url in urls_list: self.AddStory(SkiaDesktopPage(url, self))
from telemetry import story from telemetry.page import page as page_module from telemetry.page import shared_page_state class SkiaDesktopPage(page_module.Page): def __init__(self, url, page_set): super(SkiaDesktopPage, self).__init__( url=url, name=url, page_set=page_set, shared_page_state_class=shared_page_state.SharedDesktopPageState) self.archive_data_file = 'data/skia_ebay_desktop.json' def RunNavigateSteps(self, action_runner): - action_runner.Navigate(self.url) + action_runner.Navigate(self.url, timeout_in_seconds=120) ? ++++++++++++++++++++++++ - action_runner.Wait(15) class SkiaEbayDesktopPageSet(story.StorySet): """ Pages designed to represent the median, not highly optimized web """ def __init__(self): super(SkiaEbayDesktopPageSet, self).__init__( archive_data_file='data/skia_ebay_desktop.json') urls_list = [ # go/skia-skps-3-2019 'http://www.ebay.com', ] for url in urls_list: self.AddStory(SkiaDesktopPage(url, self))
618bf6d4c1fc5e60b7e94d1ad1030bf2cf0de5c2
src/main/python/alppaca/server_mock/__init__.py
src/main/python/alppaca/server_mock/__init__.py
from __future__ import print_function, absolute_import, unicode_literals, division from datetime import datetime, timedelta from textwrap import dedent from bottle import Bottle import pytz """ Super simple IMS mock. Just listens on localhost:8080 for the appropriate url, returns a test role and a dummy JSON response. """ def expiration_10s_from_now(): n = datetime.now(tz=pytz.utc) + timedelta(seconds=10) return n.strftime("%Y-%m-%dT%H:%M:%SZ") class MockIms(Bottle): PATH = '/latest/meta-data/iam/security-credentials/' json_response = dedent(""" {"Code": "Success", "AccessKeyId": "ASIAI", "SecretAccessKey": "XXYYZZ", "Token": "0123456789abcdefghijklmnopqrstuvwxyzAB", "Expiration": "%s", "Type": "AWS-HMAC"} """) def __init__(self): super(MockIms, self).__init__() self.route(self.PATH, callback=self.get_roles) self.route(self.PATH + '<role>', callback=self.get_credentials) def get_roles(self): return 'test_role' def get_credentials(self, role): return self.json_response % expiration_10s_from_now() if role == 'test_role' else '' if __name__ == "__main__": MockIms().run()
from __future__ import print_function, absolute_import, unicode_literals, division from datetime import datetime, timedelta from textwrap import dedent from bottle import Bottle import pytz def expiration_10s_from_now(): n = datetime.now(tz=pytz.utc) + timedelta(seconds=10) return n.strftime("%Y-%m-%dT%H:%M:%SZ") class MockIms(Bottle): PATH = '/latest/meta-data/iam/security-credentials/' json_response = dedent(""" {"Code": "Success", "AccessKeyId": "ASIAI", "SecretAccessKey": "XXYYZZ", "Token": "0123456789abcdefghijklmnopqrstuvwxyzAB", "Expiration": "%s", "Type": "AWS-HMAC"} """) def __init__(self): super(MockIms, self).__init__() self.route(self.PATH, callback=self.get_roles) self.route(self.PATH + '<role>', callback=self.get_credentials) def get_roles(self): return 'test_role' def get_credentials(self, role): return self.json_response % expiration_10s_from_now() if role == 'test_role' else '' if __name__ == "__main__": MockIms().run()
Move string above the imports so it becomes a docstring
Move string above the imports so it becomes a docstring
Python
apache-2.0
ImmobilienScout24/afp-alppaca,ImmobilienScout24/alppaca,ImmobilienScout24/alppaca,ImmobilienScout24/afp-alppaca
from __future__ import print_function, absolute_import, unicode_literals, division from datetime import datetime, timedelta from textwrap import dedent from bottle import Bottle import pytz - """ Super simple IMS mock. - - Just listens on localhost:8080 for the appropriate url, returns a test role and - a dummy JSON response. - - """ def expiration_10s_from_now(): n = datetime.now(tz=pytz.utc) + timedelta(seconds=10) return n.strftime("%Y-%m-%dT%H:%M:%SZ") class MockIms(Bottle): PATH = '/latest/meta-data/iam/security-credentials/' json_response = dedent(""" {"Code": "Success", "AccessKeyId": "ASIAI", "SecretAccessKey": "XXYYZZ", "Token": "0123456789abcdefghijklmnopqrstuvwxyzAB", "Expiration": "%s", "Type": "AWS-HMAC"} """) def __init__(self): super(MockIms, self).__init__() self.route(self.PATH, callback=self.get_roles) self.route(self.PATH + '<role>', callback=self.get_credentials) def get_roles(self): return 'test_role' def get_credentials(self, role): return self.json_response % expiration_10s_from_now() if role == 'test_role' else '' if __name__ == "__main__": MockIms().run()
Move string above the imports so it becomes a docstring
## Code Before: from __future__ import print_function, absolute_import, unicode_literals, division from datetime import datetime, timedelta from textwrap import dedent from bottle import Bottle import pytz """ Super simple IMS mock. Just listens on localhost:8080 for the appropriate url, returns a test role and a dummy JSON response. """ def expiration_10s_from_now(): n = datetime.now(tz=pytz.utc) + timedelta(seconds=10) return n.strftime("%Y-%m-%dT%H:%M:%SZ") class MockIms(Bottle): PATH = '/latest/meta-data/iam/security-credentials/' json_response = dedent(""" {"Code": "Success", "AccessKeyId": "ASIAI", "SecretAccessKey": "XXYYZZ", "Token": "0123456789abcdefghijklmnopqrstuvwxyzAB", "Expiration": "%s", "Type": "AWS-HMAC"} """) def __init__(self): super(MockIms, self).__init__() self.route(self.PATH, callback=self.get_roles) self.route(self.PATH + '<role>', callback=self.get_credentials) def get_roles(self): return 'test_role' def get_credentials(self, role): return self.json_response % expiration_10s_from_now() if role == 'test_role' else '' if __name__ == "__main__": MockIms().run() ## Instruction: Move string above the imports so it becomes a docstring ## Code After: from __future__ import print_function, absolute_import, unicode_literals, division from datetime import datetime, timedelta from textwrap import dedent from bottle import Bottle import pytz def expiration_10s_from_now(): n = datetime.now(tz=pytz.utc) + timedelta(seconds=10) return n.strftime("%Y-%m-%dT%H:%M:%SZ") class MockIms(Bottle): PATH = '/latest/meta-data/iam/security-credentials/' json_response = dedent(""" {"Code": "Success", "AccessKeyId": "ASIAI", "SecretAccessKey": "XXYYZZ", "Token": "0123456789abcdefghijklmnopqrstuvwxyzAB", "Expiration": "%s", "Type": "AWS-HMAC"} """) def __init__(self): super(MockIms, self).__init__() self.route(self.PATH, callback=self.get_roles) self.route(self.PATH + '<role>', callback=self.get_credentials) def get_roles(self): return 'test_role' def get_credentials(self, role): return self.json_response % expiration_10s_from_now() if role == 'test_role' else '' if __name__ == "__main__": MockIms().run()
from __future__ import print_function, absolute_import, unicode_literals, division from datetime import datetime, timedelta from textwrap import dedent from bottle import Bottle import pytz - """ Super simple IMS mock. - - Just listens on localhost:8080 for the appropriate url, returns a test role and - a dummy JSON response. - - """ def expiration_10s_from_now(): n = datetime.now(tz=pytz.utc) + timedelta(seconds=10) return n.strftime("%Y-%m-%dT%H:%M:%SZ") class MockIms(Bottle): PATH = '/latest/meta-data/iam/security-credentials/' json_response = dedent(""" {"Code": "Success", "AccessKeyId": "ASIAI", "SecretAccessKey": "XXYYZZ", "Token": "0123456789abcdefghijklmnopqrstuvwxyzAB", "Expiration": "%s", "Type": "AWS-HMAC"} """) def __init__(self): super(MockIms, self).__init__() self.route(self.PATH, callback=self.get_roles) self.route(self.PATH + '<role>', callback=self.get_credentials) def get_roles(self): return 'test_role' def get_credentials(self, role): return self.json_response % expiration_10s_from_now() if role == 'test_role' else '' if __name__ == "__main__": MockIms().run()
f4837fd60ce09b69d334fcad1403b721723d3504
tests/test_conf.py
tests/test_conf.py
import sys from unittest import mock import pytest from bottery.conf import Settings @pytest.fixture def mocked_settings(): settings = mock.MagicMock() sys.modules['settings'] = settings yield settings del sys.modules['settings'] @pytest.mark.skip def test_global_settings(): settings = Settings() assert settings.PLATFORMS == {} assert settings.TEMPLATES == [] @pytest.mark.skip def test_settings_from_module(mocked_settings): mocked_settings.PLATFORM = 'matrix' settings = Settings.from_object('settings') assert settings.PLATFORM == 'matrix' assert settings.PLATFORM == 'matrix'
from unittest import mock import pytest from bottery.conf import Settings @pytest.fixture def mocked_settings(): settings = mock.MagicMock() sys.modules['settings'] = settings yield settings del sys.modules['settings'] @pytest.mark.skip def test_global_settings(): settings = Settings() assert settings.PLATFORMS == {} assert settings.TEMPLATES == [] @pytest.mark.skip def test_settings_from_module(mocked_settings): mocked_settings.PLATFORM = 'matrix' settings = Settings.from_object('settings') assert settings.PLATFORM == 'matrix' assert settings.PLATFORM == 'matrix'
Remove unused sys import from conf tests
Remove unused sys import from conf tests
Python
mit
rougeth/bottery
- import sys from unittest import mock import pytest from bottery.conf import Settings @pytest.fixture def mocked_settings(): settings = mock.MagicMock() sys.modules['settings'] = settings yield settings del sys.modules['settings'] @pytest.mark.skip def test_global_settings(): settings = Settings() assert settings.PLATFORMS == {} assert settings.TEMPLATES == [] @pytest.mark.skip def test_settings_from_module(mocked_settings): mocked_settings.PLATFORM = 'matrix' settings = Settings.from_object('settings') assert settings.PLATFORM == 'matrix' assert settings.PLATFORM == 'matrix'
Remove unused sys import from conf tests
## Code Before: import sys from unittest import mock import pytest from bottery.conf import Settings @pytest.fixture def mocked_settings(): settings = mock.MagicMock() sys.modules['settings'] = settings yield settings del sys.modules['settings'] @pytest.mark.skip def test_global_settings(): settings = Settings() assert settings.PLATFORMS == {} assert settings.TEMPLATES == [] @pytest.mark.skip def test_settings_from_module(mocked_settings): mocked_settings.PLATFORM = 'matrix' settings = Settings.from_object('settings') assert settings.PLATFORM == 'matrix' assert settings.PLATFORM == 'matrix' ## Instruction: Remove unused sys import from conf tests ## Code After: from unittest import mock import pytest from bottery.conf import Settings @pytest.fixture def mocked_settings(): settings = mock.MagicMock() sys.modules['settings'] = settings yield settings del sys.modules['settings'] @pytest.mark.skip def test_global_settings(): settings = Settings() assert settings.PLATFORMS == {} assert settings.TEMPLATES == [] @pytest.mark.skip def test_settings_from_module(mocked_settings): mocked_settings.PLATFORM = 'matrix' settings = Settings.from_object('settings') assert settings.PLATFORM == 'matrix' assert settings.PLATFORM == 'matrix'
- import sys from unittest import mock import pytest from bottery.conf import Settings @pytest.fixture def mocked_settings(): settings = mock.MagicMock() sys.modules['settings'] = settings yield settings del sys.modules['settings'] @pytest.mark.skip def test_global_settings(): settings = Settings() assert settings.PLATFORMS == {} assert settings.TEMPLATES == [] @pytest.mark.skip def test_settings_from_module(mocked_settings): mocked_settings.PLATFORM = 'matrix' settings = Settings.from_object('settings') assert settings.PLATFORM == 'matrix' assert settings.PLATFORM == 'matrix'
fae33cf7d42559384deb7a9949f47b0881b0a29b
Cython/Tests/TestCythonUtils.py
Cython/Tests/TestCythonUtils.py
import unittest from ..Utils import build_hex_version class TestCythonUtils(unittest.TestCase): def test_build_hex_version(self): self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4')) self.assertEqual('0x001D00F0', build_hex_version('0.29')) self.assertEqual('0x040000F0', build_hex_version('4.0'))
import unittest from ..Utils import build_hex_version class TestCythonUtils(unittest.TestCase): def test_build_hex_version(self): self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4')) self.assertEqual('0x001D00F0', build_hex_version('0.29')) self.assertEqual('0x040000F0', build_hex_version('4.0'))
Remove accidentally duplicated test code.
Remove accidentally duplicated test code.
Python
apache-2.0
da-woods/cython,scoder/cython,da-woods/cython,cython/cython,scoder/cython,scoder/cython,da-woods/cython,cython/cython,scoder/cython,cython/cython,da-woods/cython,cython/cython
import unittest from ..Utils import build_hex_version class TestCythonUtils(unittest.TestCase): def test_build_hex_version(self): self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) - self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4')) self.assertEqual('0x001D00F0', build_hex_version('0.29')) self.assertEqual('0x040000F0', build_hex_version('4.0'))
Remove accidentally duplicated test code.
## Code Before: import unittest from ..Utils import build_hex_version class TestCythonUtils(unittest.TestCase): def test_build_hex_version(self): self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4')) self.assertEqual('0x001D00F0', build_hex_version('0.29')) self.assertEqual('0x040000F0', build_hex_version('4.0')) ## Instruction: Remove accidentally duplicated test code. ## Code After: import unittest from ..Utils import build_hex_version class TestCythonUtils(unittest.TestCase): def test_build_hex_version(self): self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4')) self.assertEqual('0x001D00F0', build_hex_version('0.29')) self.assertEqual('0x040000F0', build_hex_version('4.0'))
import unittest from ..Utils import build_hex_version class TestCythonUtils(unittest.TestCase): def test_build_hex_version(self): self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) - self.assertEqual('0x001D00A1', build_hex_version('0.29a1')) self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4')) self.assertEqual('0x001D00F0', build_hex_version('0.29')) self.assertEqual('0x040000F0', build_hex_version('4.0'))
89a8d6021d8ca8a714af018f3168298109013c6f
radio/__init__.py
radio/__init__.py
from django.utils.version import get_version from subprocess import check_output, CalledProcessError VERSION = (0, 0, 3, 'beta', 1) __version__ = get_version(VERSION) try: __git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode() except (FileNotFoundError, CalledProcessError): __git_hash__ = '0' __fullversion__ = '{} #{}'.format(__version__,__git_hash__) print('Trunk-Player Version ' + __fullversion__)
import logging from django.utils.version import get_version from subprocess import check_output, CalledProcessError logger = logging.getLogger(__name__) VERSION = (0, 0, 3, 'beta', 1) __version__ = get_version(VERSION) try: __git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode() except (FileNotFoundError, CalledProcessError): __git_hash__ = '0' __fullversion__ = '{} #{}'.format(__version__,__git_hash__) logger.error('Trunk-Player Version ' + __fullversion__)
Move version print to logger
Move version print to logger
Python
mit
ScanOC/trunk-player,ScanOC/trunk-player,ScanOC/trunk-player,ScanOC/trunk-player
+ import logging + from django.utils.version import get_version from subprocess import check_output, CalledProcessError + + logger = logging.getLogger(__name__) + VERSION = (0, 0, 3, 'beta', 1) __version__ = get_version(VERSION) try: __git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode() except (FileNotFoundError, CalledProcessError): __git_hash__ = '0' __fullversion__ = '{} #{}'.format(__version__,__git_hash__) - print('Trunk-Player Version ' + __fullversion__) + logger.error('Trunk-Player Version ' + __fullversion__)
Move version print to logger
## Code Before: from django.utils.version import get_version from subprocess import check_output, CalledProcessError VERSION = (0, 0, 3, 'beta', 1) __version__ = get_version(VERSION) try: __git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode() except (FileNotFoundError, CalledProcessError): __git_hash__ = '0' __fullversion__ = '{} #{}'.format(__version__,__git_hash__) print('Trunk-Player Version ' + __fullversion__) ## Instruction: Move version print to logger ## Code After: import logging from django.utils.version import get_version from subprocess import check_output, CalledProcessError logger = logging.getLogger(__name__) VERSION = (0, 0, 3, 'beta', 1) __version__ = get_version(VERSION) try: __git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode() except (FileNotFoundError, CalledProcessError): __git_hash__ = '0' __fullversion__ = '{} #{}'.format(__version__,__git_hash__) logger.error('Trunk-Player Version ' + __fullversion__)
+ import logging + from django.utils.version import get_version from subprocess import check_output, CalledProcessError + + logger = logging.getLogger(__name__) + VERSION = (0, 0, 3, 'beta', 1) __version__ = get_version(VERSION) try: __git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode() except (FileNotFoundError, CalledProcessError): __git_hash__ = '0' __fullversion__ = '{} #{}'.format(__version__,__git_hash__) - print('Trunk-Player Version ' + __fullversion__) ? ^ ^^^ + logger.error('Trunk-Player Version ' + __fullversion__) ? ^^^^^ ^^^^^^
41c49a44c5f1bc9747b22b6d1f1088f1354a2cd5
nes/cpu/decoder.py
nes/cpu/decoder.py
from sqlite3 import Connection, Row class Decoder: def __init__(self): self.conn = Connection('nes.sqlite') self.conn.row_factory = Row def __del__(self): self.conn.close() def decode(self, opcode): c = self.conn.cursor() c.execute('select * from instruction where opcode=?', [opcode]) row = c.fetchone() return dict(zip(row.keys(), row))
from sqlite3 import Connection, Row class Decoder: def __init__(self): self.conn = Connection('nes.sqlite') self.conn.row_factory = Row def __del__(self): self.conn.close() def decode(self, opcode): c = self.conn.cursor() c.execute('select * from instruction where opcode=?', [opcode]) row = c.fetchone() if row: return dict(zip(row.keys(), row)) else: raise NotImplementedError('Undocumented Opcode: ' + str(opcode))
Raise an exception when it's an undocumented opcode.
Raise an exception when it's an undocumented opcode.
Python
mit
Hexadorsimal/pynes
from sqlite3 import Connection, Row class Decoder: def __init__(self): self.conn = Connection('nes.sqlite') self.conn.row_factory = Row def __del__(self): self.conn.close() def decode(self, opcode): c = self.conn.cursor() c.execute('select * from instruction where opcode=?', [opcode]) row = c.fetchone() + if row: - return dict(zip(row.keys(), row)) + return dict(zip(row.keys(), row)) + else: + raise NotImplementedError('Undocumented Opcode: ' + str(opcode))
Raise an exception when it's an undocumented opcode.
## Code Before: from sqlite3 import Connection, Row class Decoder: def __init__(self): self.conn = Connection('nes.sqlite') self.conn.row_factory = Row def __del__(self): self.conn.close() def decode(self, opcode): c = self.conn.cursor() c.execute('select * from instruction where opcode=?', [opcode]) row = c.fetchone() return dict(zip(row.keys(), row)) ## Instruction: Raise an exception when it's an undocumented opcode. ## Code After: from sqlite3 import Connection, Row class Decoder: def __init__(self): self.conn = Connection('nes.sqlite') self.conn.row_factory = Row def __del__(self): self.conn.close() def decode(self, opcode): c = self.conn.cursor() c.execute('select * from instruction where opcode=?', [opcode]) row = c.fetchone() if row: return dict(zip(row.keys(), row)) else: raise NotImplementedError('Undocumented Opcode: ' + str(opcode))
from sqlite3 import Connection, Row class Decoder: def __init__(self): self.conn = Connection('nes.sqlite') self.conn.row_factory = Row def __del__(self): self.conn.close() def decode(self, opcode): c = self.conn.cursor() c.execute('select * from instruction where opcode=?', [opcode]) row = c.fetchone() + if row: - return dict(zip(row.keys(), row)) + return dict(zip(row.keys(), row)) ? ++++ + else: + raise NotImplementedError('Undocumented Opcode: ' + str(opcode))
c43820a2e26dd4f87c36b986a9a0af80b409f659
sentence_extractor.py
sentence_extractor.py
import textract import sys import os import re import random ################################### # Extracts text from a pdf file and # selects one sentence, which it # then prints. # # Created by Fredrik Omstedt. ################################### # Extracts texts from pdf files. If given a directory, the # program will return texts from all pdf files in that directory. def extractTexts(): file = sys.argv[1] texts = [] if os.path.isdir(file): for f in os.listdir(file): if re.match(r'^.*\.pdf$', f): texts.append(textract.process(file + "/" + f)) else: texts.append(textract.process(file)) return texts # Chooses one sentence randomly from each of the given texts. def selectSentences(texts): chosen_sentences = [] for text in texts: sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M) sentences = sentence_structure.findall(text) chosen_sentences.append( sentences[random.randint(0, len(sentences)-1)].replace("\n", " ") ) return chosen_sentences def main(): texts = extractTexts() sentences = selectSentences(texts) for sentence in sentences: print(sentence) print("\n") if __name__ == '__main__': main()
import textract import sys import os import re import random ################################### # Extracts text from a pdf file and # selects one sentence, which it # then prints. # # Created by Fredrik Omstedt. ################################### # Extracts texts from pdf files. If given a directory, the # program will return texts from all pdf files in that directory. def extractTexts(): file = sys.argv[1] texts = [] if os.path.isdir(file): for f in os.listdir(file): if re.match(r'^.*\.pdf$', f): texts.append(textract.process(file + "/" + f)) else: texts.append(textract.process(file)) return texts # Chooses one sentence randomly from each of the given texts. def selectSentences(texts): chosen_sentences = [] for text in texts: sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M) sentences = sentence_structure.findall(text) chosen_sentences.append( sentences[random.randint(0, len(sentences)-1)].replace("\n", " ") ) return chosen_sentences def main(): texts = extractTexts() sentences = selectSentences(texts) for sentence in sentences: print(sentence) print("\n") if __name__ == '__main__': main()
Update regex to match sentences starting with ÅÄÖ
Update regex to match sentences starting with ÅÄÖ
Python
mit
Xaril/sentence-extractor,Xaril/sentence-extractor
import textract import sys import os import re import random ################################### # Extracts text from a pdf file and # selects one sentence, which it # then prints. # # Created by Fredrik Omstedt. ################################### # Extracts texts from pdf files. If given a directory, the # program will return texts from all pdf files in that directory. def extractTexts(): file = sys.argv[1] texts = [] if os.path.isdir(file): for f in os.listdir(file): if re.match(r'^.*\.pdf$', f): texts.append(textract.process(file + "/" + f)) else: texts.append(textract.process(file)) return texts # Chooses one sentence randomly from each of the given texts. def selectSentences(texts): chosen_sentences = [] for text in texts: - sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M) + sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M) sentences = sentence_structure.findall(text) chosen_sentences.append( sentences[random.randint(0, len(sentences)-1)].replace("\n", " ") ) return chosen_sentences def main(): texts = extractTexts() sentences = selectSentences(texts) for sentence in sentences: print(sentence) print("\n") if __name__ == '__main__': main()
Update regex to match sentences starting with ÅÄÖ
## Code Before: import textract import sys import os import re import random ################################### # Extracts text from a pdf file and # selects one sentence, which it # then prints. # # Created by Fredrik Omstedt. ################################### # Extracts texts from pdf files. If given a directory, the # program will return texts from all pdf files in that directory. def extractTexts(): file = sys.argv[1] texts = [] if os.path.isdir(file): for f in os.listdir(file): if re.match(r'^.*\.pdf$', f): texts.append(textract.process(file + "/" + f)) else: texts.append(textract.process(file)) return texts # Chooses one sentence randomly from each of the given texts. def selectSentences(texts): chosen_sentences = [] for text in texts: sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M) sentences = sentence_structure.findall(text) chosen_sentences.append( sentences[random.randint(0, len(sentences)-1)].replace("\n", " ") ) return chosen_sentences def main(): texts = extractTexts() sentences = selectSentences(texts) for sentence in sentences: print(sentence) print("\n") if __name__ == '__main__': main() ## Instruction: Update regex to match sentences starting with ÅÄÖ ## Code After: import textract import sys import os import re import random ################################### # Extracts text from a pdf file and # selects one sentence, which it # then prints. # # Created by Fredrik Omstedt. ################################### # Extracts texts from pdf files. If given a directory, the # program will return texts from all pdf files in that directory. def extractTexts(): file = sys.argv[1] texts = [] if os.path.isdir(file): for f in os.listdir(file): if re.match(r'^.*\.pdf$', f): texts.append(textract.process(file + "/" + f)) else: texts.append(textract.process(file)) return texts # Chooses one sentence randomly from each of the given texts. def selectSentences(texts): chosen_sentences = [] for text in texts: sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M) sentences = sentence_structure.findall(text) chosen_sentences.append( sentences[random.randint(0, len(sentences)-1)].replace("\n", " ") ) return chosen_sentences def main(): texts = extractTexts() sentences = selectSentences(texts) for sentence in sentences: print(sentence) print("\n") if __name__ == '__main__': main()
import textract import sys import os import re import random ################################### # Extracts text from a pdf file and # selects one sentence, which it # then prints. # # Created by Fredrik Omstedt. ################################### # Extracts texts from pdf files. If given a directory, the # program will return texts from all pdf files in that directory. def extractTexts(): file = sys.argv[1] texts = [] if os.path.isdir(file): for f in os.listdir(file): if re.match(r'^.*\.pdf$', f): texts.append(textract.process(file + "/" + f)) else: texts.append(textract.process(file)) return texts # Chooses one sentence randomly from each of the given texts. def selectSentences(texts): chosen_sentences = [] for text in texts: - sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M) + sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M) ? ++++++++++++ sentences = sentence_structure.findall(text) chosen_sentences.append( sentences[random.randint(0, len(sentences)-1)].replace("\n", " ") ) return chosen_sentences def main(): texts = extractTexts() sentences = selectSentences(texts) for sentence in sentences: print(sentence) print("\n") if __name__ == '__main__': main()
78c5580d349d6bec0715a36c13437177a726f7ad
tests/test_isim.py
tests/test_isim.py
import pytest def test_isim(): import os import shutil import tempfile import yaml from fusesoc.edatools import get_edatool from edalize_common import compare_files, files, param_gen, tests_dir, vpi (parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam']) work_root = tempfile.mkdtemp(prefix='isim_') eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml') with open(eda_api_file,'w') as f: f.write(yaml.dump({'name' : 'test_isim_0', 'files' : files, 'parameters' : parameters, 'tool_options' : {'isim' : { 'fuse_options' : ['some', 'fuse_options'], 'isim_options' : ['a', 'few', 'isim_options']}}, 'toplevel' : 'top_module', 'vpi' : vpi})) backend = get_edatool('isim')(eda_api_file=eda_api_file) backend.configure(args) ref_dir = os.path.join(tests_dir, __name__) compare_files(ref_dir, work_root, ['config.mk', 'Makefile', 'run_test_isim_0.tcl', 'test_isim_0.prj']) dummy_exe = 'test_isim_0' shutil.copy(os.path.join(ref_dir, dummy_exe), os.path.join(work_root, dummy_exe)) backend.run([]) compare_files(ref_dir, work_root, ['run.cmd'])
import pytest def test_isim(): import os import shutil from edalize_common import compare_files, setup_backend, tests_dir ref_dir = os.path.join(tests_dir, __name__) paramtypes = ['plusarg', 'vlogdefine', 'vlogparam'] name = 'test_isim_0' tool = 'isim' tool_options = { 'fuse_options' : ['some', 'fuse_options'], 'isim_options' : ['a', 'few', 'isim_options'], } (backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options) backend.configure(args) compare_files(ref_dir, work_root, ['config.mk', 'Makefile', 'run_test_isim_0.tcl', 'test_isim_0.prj']) dummy_exe = 'test_isim_0' shutil.copy(os.path.join(ref_dir, dummy_exe), os.path.join(work_root, dummy_exe)) backend.run([]) compare_files(ref_dir, work_root, ['run.cmd'])
Reduce code duplication in isim test
Reduce code duplication in isim test
Python
bsd-2-clause
olofk/fusesoc,olofk/fusesoc,lowRISC/fusesoc,lowRISC/fusesoc
import pytest def test_isim(): import os import shutil - import tempfile - import yaml - from fusesoc.edatools import get_edatool - from edalize_common import compare_files, files, param_gen, tests_dir, vpi + from edalize_common import compare_files, setup_backend, tests_dir + ref_dir = os.path.join(tests_dir, __name__) - (parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam']) + paramtypes = ['plusarg', 'vlogdefine', 'vlogparam'] + name = 'test_isim_0' + tool = 'isim' + tool_options = { + 'fuse_options' : ['some', 'fuse_options'], + 'isim_options' : ['a', 'few', 'isim_options'], + } + (backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options) - work_root = tempfile.mkdtemp(prefix='isim_') - eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml') - with open(eda_api_file,'w') as f: - f.write(yaml.dump({'name' : 'test_isim_0', - 'files' : files, - 'parameters' : parameters, - 'tool_options' : {'isim' : { - 'fuse_options' : ['some', 'fuse_options'], - 'isim_options' : ['a', 'few', 'isim_options']}}, - 'toplevel' : 'top_module', - 'vpi' : vpi})) - - backend = get_edatool('isim')(eda_api_file=eda_api_file) backend.configure(args) - ref_dir = os.path.join(tests_dir, __name__) compare_files(ref_dir, work_root, ['config.mk', 'Makefile', 'run_test_isim_0.tcl', 'test_isim_0.prj']) dummy_exe = 'test_isim_0' shutil.copy(os.path.join(ref_dir, dummy_exe), os.path.join(work_root, dummy_exe)) backend.run([]) compare_files(ref_dir, work_root, ['run.cmd'])
Reduce code duplication in isim test
## Code Before: import pytest def test_isim(): import os import shutil import tempfile import yaml from fusesoc.edatools import get_edatool from edalize_common import compare_files, files, param_gen, tests_dir, vpi (parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam']) work_root = tempfile.mkdtemp(prefix='isim_') eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml') with open(eda_api_file,'w') as f: f.write(yaml.dump({'name' : 'test_isim_0', 'files' : files, 'parameters' : parameters, 'tool_options' : {'isim' : { 'fuse_options' : ['some', 'fuse_options'], 'isim_options' : ['a', 'few', 'isim_options']}}, 'toplevel' : 'top_module', 'vpi' : vpi})) backend = get_edatool('isim')(eda_api_file=eda_api_file) backend.configure(args) ref_dir = os.path.join(tests_dir, __name__) compare_files(ref_dir, work_root, ['config.mk', 'Makefile', 'run_test_isim_0.tcl', 'test_isim_0.prj']) dummy_exe = 'test_isim_0' shutil.copy(os.path.join(ref_dir, dummy_exe), os.path.join(work_root, dummy_exe)) backend.run([]) compare_files(ref_dir, work_root, ['run.cmd']) ## Instruction: Reduce code duplication in isim test ## Code After: import pytest def test_isim(): import os import shutil from edalize_common import compare_files, setup_backend, tests_dir ref_dir = os.path.join(tests_dir, __name__) paramtypes = ['plusarg', 'vlogdefine', 'vlogparam'] name = 'test_isim_0' tool = 'isim' tool_options = { 'fuse_options' : ['some', 'fuse_options'], 'isim_options' : ['a', 'few', 'isim_options'], } (backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options) backend.configure(args) compare_files(ref_dir, work_root, ['config.mk', 'Makefile', 'run_test_isim_0.tcl', 'test_isim_0.prj']) dummy_exe = 'test_isim_0' shutil.copy(os.path.join(ref_dir, dummy_exe), os.path.join(work_root, dummy_exe)) backend.run([]) compare_files(ref_dir, work_root, ['run.cmd'])
import pytest def test_isim(): import os import shutil - import tempfile - import yaml - from fusesoc.edatools import get_edatool - from edalize_common import compare_files, files, param_gen, tests_dir, vpi ? ^^^ ^^^ ^^^^^ ----- + from edalize_common import compare_files, setup_backend, tests_dir ? ^ ^^ ++ ^^ + - (parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam']) + ref_dir = os.path.join(tests_dir, __name__) + paramtypes = ['plusarg', 'vlogdefine', 'vlogparam'] + name = 'test_isim_0' + tool = 'isim' + tool_options = { + 'fuse_options' : ['some', 'fuse_options'], + 'isim_options' : ['a', 'few', 'isim_options'], + } + (backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options) - work_root = tempfile.mkdtemp(prefix='isim_') - eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml') - with open(eda_api_file,'w') as f: - f.write(yaml.dump({'name' : 'test_isim_0', - 'files' : files, - 'parameters' : parameters, - 'tool_options' : {'isim' : { - 'fuse_options' : ['some', 'fuse_options'], - 'isim_options' : ['a', 'few', 'isim_options']}}, - 'toplevel' : 'top_module', - 'vpi' : vpi})) - - backend = get_edatool('isim')(eda_api_file=eda_api_file) backend.configure(args) - ref_dir = os.path.join(tests_dir, __name__) compare_files(ref_dir, work_root, ['config.mk', 'Makefile', 'run_test_isim_0.tcl', 'test_isim_0.prj']) dummy_exe = 'test_isim_0' shutil.copy(os.path.join(ref_dir, dummy_exe), os.path.join(work_root, dummy_exe)) backend.run([]) compare_files(ref_dir, work_root, ['run.cmd'])
8621d6c0826beb4a4b4e920ce27786b01546ed28
impactstoryanalytics/highcharts.py
impactstoryanalytics/highcharts.py
boilerplate = { 'chart': { 'renderTo': 'container', 'plotBackgroundColor': 'none', 'backgroundColor': 'none', 'spacingTop': 5 }, 'title': {'text': None}, 'subtitle': {'text': None}, 'yAxis': { 'title':{ 'text': None }, 'gridLineColor': 'rgba(255, 255, 255, .1)' }, "xAxis": { "lineColor": "rgba(0,0,0,0)" }, 'credits': { 'enabled': False }, 'plotOptions': { 'series': { 'marker': { 'enabled': False } } }, }
boilerplate = { 'chart': { 'renderTo': 'container', 'plotBackgroundColor': 'none', 'backgroundColor': 'none', 'spacingTop': 5 }, 'title': {'text': None}, 'subtitle': {'text': None}, 'yAxis': { 'min': 0, 'title':{ 'text': None }, 'gridLineColor': 'rgba(255, 255, 255, .1)' }, "xAxis": { "lineColor": "rgba(0,0,0,0)" }, 'credits': { 'enabled': False }, 'plotOptions': { 'series': { 'marker': { 'enabled': False } } }, }
Set y-axis min to 0
Set y-axis min to 0
Python
mit
Impactstory/impactstory-analytics,Impactstory/impactstory-analytics,total-impact/impactstory-analytics,total-impact/impactstory-analytics,Impactstory/impactstory-analytics,total-impact/impactstory-analytics,Impactstory/impactstory-analytics,total-impact/impactstory-analytics
boilerplate = { 'chart': { 'renderTo': 'container', 'plotBackgroundColor': 'none', 'backgroundColor': 'none', 'spacingTop': 5 }, 'title': {'text': None}, 'subtitle': {'text': None}, 'yAxis': { + 'min': 0, 'title':{ 'text': None }, 'gridLineColor': 'rgba(255, 255, 255, .1)' }, "xAxis": { "lineColor": "rgba(0,0,0,0)" }, 'credits': { 'enabled': False }, 'plotOptions': { 'series': { 'marker': { 'enabled': False } } }, }
Set y-axis min to 0
## Code Before: boilerplate = { 'chart': { 'renderTo': 'container', 'plotBackgroundColor': 'none', 'backgroundColor': 'none', 'spacingTop': 5 }, 'title': {'text': None}, 'subtitle': {'text': None}, 'yAxis': { 'title':{ 'text': None }, 'gridLineColor': 'rgba(255, 255, 255, .1)' }, "xAxis": { "lineColor": "rgba(0,0,0,0)" }, 'credits': { 'enabled': False }, 'plotOptions': { 'series': { 'marker': { 'enabled': False } } }, } ## Instruction: Set y-axis min to 0 ## Code After: boilerplate = { 'chart': { 'renderTo': 'container', 'plotBackgroundColor': 'none', 'backgroundColor': 'none', 'spacingTop': 5 }, 'title': {'text': None}, 'subtitle': {'text': None}, 'yAxis': { 'min': 0, 'title':{ 'text': None }, 'gridLineColor': 'rgba(255, 255, 255, .1)' }, "xAxis": { "lineColor": "rgba(0,0,0,0)" }, 'credits': { 'enabled': False }, 'plotOptions': { 'series': { 'marker': { 'enabled': False } } }, }
boilerplate = { 'chart': { 'renderTo': 'container', 'plotBackgroundColor': 'none', 'backgroundColor': 'none', 'spacingTop': 5 }, 'title': {'text': None}, 'subtitle': {'text': None}, 'yAxis': { + 'min': 0, 'title':{ 'text': None }, 'gridLineColor': 'rgba(255, 255, 255, .1)' }, "xAxis": { "lineColor": "rgba(0,0,0,0)" }, 'credits': { 'enabled': False }, 'plotOptions': { 'series': { 'marker': { 'enabled': False } } }, }
8befea283830f76dfa41cfd10d7eb916c68f7ef9
intern/views.py
intern/views.py
from django.contrib.auth.decorators import login_required from django.shortcuts import render from filer.models import File from filer.models import Folder @login_required def documents(request): files = File.objects.all() folders = Folder.objects.all() #print(files[0]) return render(request, 'intern/documents.html', {'files': files, 'folders': folders})
from django.contrib.auth.decorators import login_required from django.shortcuts import render from filer.models import File from filer.models import Folder @login_required def documents(request): files = File.objects.all().order_by("-modified_at") folders = Folder.objects.all() #print(files[0]) return render(request, 'intern/documents.html', {'files': files, 'folders': folders})
Sort files by last modification
Sort files by last modification
Python
mit
n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb
from django.contrib.auth.decorators import login_required from django.shortcuts import render from filer.models import File from filer.models import Folder @login_required def documents(request): - files = File.objects.all() + files = File.objects.all().order_by("-modified_at") folders = Folder.objects.all() #print(files[0]) return render(request, 'intern/documents.html', {'files': files, 'folders': folders})
Sort files by last modification
## Code Before: from django.contrib.auth.decorators import login_required from django.shortcuts import render from filer.models import File from filer.models import Folder @login_required def documents(request): files = File.objects.all() folders = Folder.objects.all() #print(files[0]) return render(request, 'intern/documents.html', {'files': files, 'folders': folders}) ## Instruction: Sort files by last modification ## Code After: from django.contrib.auth.decorators import login_required from django.shortcuts import render from filer.models import File from filer.models import Folder @login_required def documents(request): files = File.objects.all().order_by("-modified_at") folders = Folder.objects.all() #print(files[0]) return render(request, 'intern/documents.html', {'files': files, 'folders': folders})
from django.contrib.auth.decorators import login_required from django.shortcuts import render from filer.models import File from filer.models import Folder @login_required def documents(request): - files = File.objects.all() + files = File.objects.all().order_by("-modified_at") folders = Folder.objects.all() #print(files[0]) return render(request, 'intern/documents.html', {'files': files, 'folders': folders})
ee9df63aeaabb4111cea3698a4f0e30b4502e519
test/disable_captcha.py
test/disable_captcha.py
import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text('Spambot countermeasures') # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert 'Enable spambot countermeasures' in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { 'enable_confirm': '0', } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') self.change_acp_knob( link_text='Spambot countermeasures', check_page_text='Enable spambot countermeasures', name='enable_confirm', value='0', ) def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text(link_text) # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert check_page_text in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { name: value, } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
Generalize to allow editing other configuration
Generalize to allow editing other configuration
Python
bsd-2-clause
p/wolis-phpbb,p/wolis-phpbb
import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') + self.change_acp_knob( + link_text='Spambot countermeasures', + check_page_text='Enable spambot countermeasures', + name='enable_confirm', + value='0', + ) + + def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body - url = self.link_href_by_text('Spambot countermeasures') + url = self.link_href_by_text(link_text) # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) - assert 'Enable spambot countermeasures' in self.response.body + assert check_page_text in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { - 'enable_confirm': '0', + name: value, } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
Generalize to allow editing other configuration
## Code Before: import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text('Spambot countermeasures') # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert 'Enable spambot countermeasures' in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { 'enable_confirm': '0', } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main() ## Instruction: Generalize to allow editing other configuration ## Code After: import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') self.change_acp_knob( link_text='Spambot countermeasures', check_page_text='Enable spambot countermeasures', name='enable_confirm', value='0', ) def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text(link_text) # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert check_page_text in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { name: value, } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') + self.change_acp_knob( + link_text='Spambot countermeasures', + check_page_text='Enable spambot countermeasures', + name='enable_confirm', + value='0', + ) + + def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body - url = self.link_href_by_text('Spambot countermeasures') + url = self.link_href_by_text(link_text) # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) - assert 'Enable spambot countermeasures' in self.response.body + assert check_page_text in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { - 'enable_confirm': '0', + name: value, } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
a5fdffe2f37e2e1c34044c259ef56c0e5feca0cb
allegedb/allegedb/tests/test_branch_plan.py
allegedb/allegedb/tests/test_branch_plan.py
import pytest import allegedb @pytest.fixture(scope='function') def orm(): with allegedb.ORM("sqlite:///:memory:") as it: yield it def test_single_plan(orm): g = orm.new_graph('graph') g.add_node(0) orm.turn = 1 g.add_node(1) with orm.plan(): orm.turn = 2 g.add_node(2) assert orm.turn == 1 assert 2 not in g orm.branch = 'b' assert 2 not in g assert 1 in g orm.turn = 2 assert 2 in g orm.turn = 1 orm.branch = 'trunk' orm.turn = 0 assert 1 not in g orm.branch = 'c' orm.turn = 2 assert 1 not in g assert 2 not in g
import pytest import allegedb @pytest.fixture(scope='function') def orm(): with allegedb.ORM("sqlite:///:memory:") as it: yield it def test_single_plan(orm): g = orm.new_graph('graph') g.add_node(0) orm.turn = 1 g.add_node(1) with orm.plan(): orm.turn = 2 g.add_node(2) assert orm.turn == 1 assert 2 not in g orm.branch = 'b' assert 2 not in g assert 1 in g orm.turn = 2 assert 2 in g orm.turn = 1 orm.branch = 'trunk' orm.turn = 0 assert 1 not in g orm.branch = 'c' orm.turn = 2 assert 1 not in g assert 2 not in g orm.turn = 0 orm.branch = 'trunk' orm.turn = 2 assert 2 in g
Add an extra check in that test
Add an extra check in that test
Python
agpl-3.0
LogicalDash/LiSE,LogicalDash/LiSE
import pytest import allegedb @pytest.fixture(scope='function') def orm(): with allegedb.ORM("sqlite:///:memory:") as it: yield it def test_single_plan(orm): g = orm.new_graph('graph') g.add_node(0) orm.turn = 1 g.add_node(1) with orm.plan(): orm.turn = 2 g.add_node(2) assert orm.turn == 1 assert 2 not in g orm.branch = 'b' assert 2 not in g assert 1 in g orm.turn = 2 assert 2 in g orm.turn = 1 orm.branch = 'trunk' orm.turn = 0 assert 1 not in g orm.branch = 'c' orm.turn = 2 assert 1 not in g assert 2 not in g + orm.turn = 0 + orm.branch = 'trunk' + orm.turn = 2 + assert 2 in g
Add an extra check in that test
## Code Before: import pytest import allegedb @pytest.fixture(scope='function') def orm(): with allegedb.ORM("sqlite:///:memory:") as it: yield it def test_single_plan(orm): g = orm.new_graph('graph') g.add_node(0) orm.turn = 1 g.add_node(1) with orm.plan(): orm.turn = 2 g.add_node(2) assert orm.turn == 1 assert 2 not in g orm.branch = 'b' assert 2 not in g assert 1 in g orm.turn = 2 assert 2 in g orm.turn = 1 orm.branch = 'trunk' orm.turn = 0 assert 1 not in g orm.branch = 'c' orm.turn = 2 assert 1 not in g assert 2 not in g ## Instruction: Add an extra check in that test ## Code After: import pytest import allegedb @pytest.fixture(scope='function') def orm(): with allegedb.ORM("sqlite:///:memory:") as it: yield it def test_single_plan(orm): g = orm.new_graph('graph') g.add_node(0) orm.turn = 1 g.add_node(1) with orm.plan(): orm.turn = 2 g.add_node(2) assert orm.turn == 1 assert 2 not in g orm.branch = 'b' assert 2 not in g assert 1 in g orm.turn = 2 assert 2 in g orm.turn = 1 orm.branch = 'trunk' orm.turn = 0 assert 1 not in g orm.branch = 'c' orm.turn = 2 assert 1 not in g assert 2 not in g orm.turn = 0 orm.branch = 'trunk' orm.turn = 2 assert 2 in g
import pytest import allegedb @pytest.fixture(scope='function') def orm(): with allegedb.ORM("sqlite:///:memory:") as it: yield it def test_single_plan(orm): g = orm.new_graph('graph') g.add_node(0) orm.turn = 1 g.add_node(1) with orm.plan(): orm.turn = 2 g.add_node(2) assert orm.turn == 1 assert 2 not in g orm.branch = 'b' assert 2 not in g assert 1 in g orm.turn = 2 assert 2 in g orm.turn = 1 orm.branch = 'trunk' orm.turn = 0 assert 1 not in g orm.branch = 'c' orm.turn = 2 assert 1 not in g assert 2 not in g + orm.turn = 0 + orm.branch = 'trunk' + orm.turn = 2 + assert 2 in g
39561a89ea497776d980d3eda97fc2f75493528f
internal_social_auth/views.py
internal_social_auth/views.py
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, "Your Google Apps domain isn't authorized for this app") return HttpResponseRedirect('/') class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, self.get_error_message()) return HttpResponseRedirect(self.get_faiure_url()) def get_error_message(self): if self.error_message: return self.error_message return "Your Google Apps domain isn't authorized for this app" def get_failure_url(self): if self.failure_url: return force_text(self.failure_url) return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
Clean up the AuthComplete API a little
Clean up the AuthComplete API a little
Python
bsd-2-clause
incuna/incuna-internal-social-auth
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse + from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) + messages.error(request, self.get_error_message()) + return HttpResponseRedirect(self.get_faiure_url()) + + def get_error_message(self): + if self.error_message: + return self.error_message - messages.error(request, "Your Google Apps domain isn't authorized for this app") + return "Your Google Apps domain isn't authorized for this app" - return HttpResponseRedirect('/') + + def get_failure_url(self): + if self.failure_url: + return force_text(self.failure_url) + return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401) -
Clean up the AuthComplete API a little
## Code Before: import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, "Your Google Apps domain isn't authorized for this app") return HttpResponseRedirect('/') class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401) ## Instruction: Clean up the AuthComplete API a little ## Code After: import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, self.get_error_message()) return HttpResponseRedirect(self.get_faiure_url()) def get_error_message(self): if self.error_message: return self.error_message return "Your Google Apps domain isn't authorized for this app" def get_failure_url(self): if self.failure_url: return force_text(self.failure_url) return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse + from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) + messages.error(request, self.get_error_message()) + return HttpResponseRedirect(self.get_faiure_url()) + + def get_error_message(self): + if self.error_message: + return self.error_message - messages.error(request, "Your Google Apps domain isn't authorized for this app") ? --------------- ^ ^^^^ - + return "Your Google Apps domain isn't authorized for this app" ? ^ ^^ - return HttpResponseRedirect('/') + + def get_failure_url(self): + if self.failure_url: + return force_text(self.failure_url) + return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401) -
3b6ddce7c0db0f0b1fbd9febd9bf68ceeda51f44
della/user_manager/forms.py
della/user_manager/forms.py
from django import forms from django.contrib.auth.models import User from django.contrib.auth.forms import UserCreationForm from django.core.validators import RegexValidator alphanumericu = RegexValidator( regex=r'^[0-9a-zA-Z_]*$', message='Only alphanumeric characters and underscore are allowed.') class SignupForm(UserCreationForm): username = forms.CharField(max_length=20, validators=[alphanumericu]) email = forms.EmailField(max_length=254, required=True) class Meta: model = User fields = ['email', 'username', ] def clean_email(self): error_message = 'An user with that email already exists' email = self.cleaned_data.get('email') if email and User.objects.filter(email=email).exists(): raise forms.ValidationError( self.error_messages[error_message], code='existing_email', ) return email
from django import forms from django.contrib.auth.models import User from django.contrib.auth.forms import UserCreationForm from django.core.validators import RegexValidator alphanumericu = RegexValidator( regex=r'^[0-9a-zA-Z_]*$', message='Only alphanumeric characters and underscore are allowed.') class SignupForm(UserCreationForm): username = forms.CharField(max_length=20, validators=[alphanumericu]) email = forms.EmailField(max_length=254, required=True) class Meta: model = User fields = ['email', 'username', ] def clean_email(self): error_message = 'An user with that email already exists' email = self.cleaned_data.get('email') if email and User.objects.filter(email=email).exists(): raise forms.ValidationError(error_message) return email
Raise ValidationError properly in SignupForm
Raise ValidationError properly in SignupForm
Python
mit
avinassh/della,avinassh/della,avinassh/della
from django import forms from django.contrib.auth.models import User from django.contrib.auth.forms import UserCreationForm from django.core.validators import RegexValidator alphanumericu = RegexValidator( regex=r'^[0-9a-zA-Z_]*$', message='Only alphanumeric characters and underscore are allowed.') class SignupForm(UserCreationForm): username = forms.CharField(max_length=20, validators=[alphanumericu]) email = forms.EmailField(max_length=254, required=True) class Meta: model = User fields = ['email', 'username', ] def clean_email(self): error_message = 'An user with that email already exists' email = self.cleaned_data.get('email') if email and User.objects.filter(email=email).exists(): - raise forms.ValidationError( + raise forms.ValidationError(error_message) - self.error_messages[error_message], - code='existing_email', - ) return email
Raise ValidationError properly in SignupForm
## Code Before: from django import forms from django.contrib.auth.models import User from django.contrib.auth.forms import UserCreationForm from django.core.validators import RegexValidator alphanumericu = RegexValidator( regex=r'^[0-9a-zA-Z_]*$', message='Only alphanumeric characters and underscore are allowed.') class SignupForm(UserCreationForm): username = forms.CharField(max_length=20, validators=[alphanumericu]) email = forms.EmailField(max_length=254, required=True) class Meta: model = User fields = ['email', 'username', ] def clean_email(self): error_message = 'An user with that email already exists' email = self.cleaned_data.get('email') if email and User.objects.filter(email=email).exists(): raise forms.ValidationError( self.error_messages[error_message], code='existing_email', ) return email ## Instruction: Raise ValidationError properly in SignupForm ## Code After: from django import forms from django.contrib.auth.models import User from django.contrib.auth.forms import UserCreationForm from django.core.validators import RegexValidator alphanumericu = RegexValidator( regex=r'^[0-9a-zA-Z_]*$', message='Only alphanumeric characters and underscore are allowed.') class SignupForm(UserCreationForm): username = forms.CharField(max_length=20, validators=[alphanumericu]) email = forms.EmailField(max_length=254, required=True) class Meta: model = User fields = ['email', 'username', ] def clean_email(self): error_message = 'An user with that email already exists' email = self.cleaned_data.get('email') if email and User.objects.filter(email=email).exists(): raise forms.ValidationError(error_message) return email
from django import forms from django.contrib.auth.models import User from django.contrib.auth.forms import UserCreationForm from django.core.validators import RegexValidator alphanumericu = RegexValidator( regex=r'^[0-9a-zA-Z_]*$', message='Only alphanumeric characters and underscore are allowed.') class SignupForm(UserCreationForm): username = forms.CharField(max_length=20, validators=[alphanumericu]) email = forms.EmailField(max_length=254, required=True) class Meta: model = User fields = ['email', 'username', ] def clean_email(self): error_message = 'An user with that email already exists' email = self.cleaned_data.get('email') if email and User.objects.filter(email=email).exists(): - raise forms.ValidationError( + raise forms.ValidationError(error_message) ? ++++++++++++++ - self.error_messages[error_message], - code='existing_email', - ) return email
eb7ff9cec9360af0b5c18915164a54d4755e657b
mistraldashboard/dashboards/mistral/executions/tables.py
mistraldashboard/dashboards/mistral/executions/tables.py
from django.utils.translation import ugettext_lazy as _ from horizon import tables class ExecutionsTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID"), link=("horizon:mistral:executions:tasks")) wb_name = tables.Column("workbook_name", verbose_name=_("Workbook")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "executions" verbose_name = _("Executions") class TaskTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID")) name = tables.Column("name", verbose_name=_("Name")) action = tables.Column("action", verbose_name=_("Action")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "tasks" verbose_name = _("Tasks")
from django.utils.translation import ugettext_lazy as _ from horizon import tables class ExecutionsTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID"), link=("horizon:mistral:executions:tasks")) wb_name = tables.Column("workbook_name", verbose_name=_("Workbook")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "executions" verbose_name = _("Executions") class TaskTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID")) name = tables.Column("name", verbose_name=_("Name")) parameters = tables.Column("parameters", verbose_name=_("Parameters")) output = tables.Column("output", verbose_name=_("Output")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "tasks" verbose_name = _("Tasks")
Add Task's output and parameters columns
Add Task's output and parameters columns Change-Id: I98f57a6a0178bb7258d82f3a165127f060f42f7b Implements: blueprint mistral-ui
Python
apache-2.0
openstack/mistral-dashboard,openstack/mistral-dashboard,openstack/mistral-dashboard
from django.utils.translation import ugettext_lazy as _ from horizon import tables class ExecutionsTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID"), link=("horizon:mistral:executions:tasks")) wb_name = tables.Column("workbook_name", verbose_name=_("Workbook")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "executions" verbose_name = _("Executions") class TaskTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID")) name = tables.Column("name", verbose_name=_("Name")) + parameters = tables.Column("parameters", verbose_name=_("Parameters")) - action = tables.Column("action", verbose_name=_("Action")) + output = tables.Column("output", verbose_name=_("Output")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "tasks" verbose_name = _("Tasks")
Add Task's output and parameters columns
## Code Before: from django.utils.translation import ugettext_lazy as _ from horizon import tables class ExecutionsTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID"), link=("horizon:mistral:executions:tasks")) wb_name = tables.Column("workbook_name", verbose_name=_("Workbook")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "executions" verbose_name = _("Executions") class TaskTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID")) name = tables.Column("name", verbose_name=_("Name")) action = tables.Column("action", verbose_name=_("Action")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "tasks" verbose_name = _("Tasks") ## Instruction: Add Task's output and parameters columns ## Code After: from django.utils.translation import ugettext_lazy as _ from horizon import tables class ExecutionsTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID"), link=("horizon:mistral:executions:tasks")) wb_name = tables.Column("workbook_name", verbose_name=_("Workbook")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "executions" verbose_name = _("Executions") class TaskTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID")) name = tables.Column("name", verbose_name=_("Name")) parameters = tables.Column("parameters", verbose_name=_("Parameters")) output = tables.Column("output", verbose_name=_("Output")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "tasks" verbose_name = _("Tasks")
from django.utils.translation import ugettext_lazy as _ from horizon import tables class ExecutionsTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID"), link=("horizon:mistral:executions:tasks")) wb_name = tables.Column("workbook_name", verbose_name=_("Workbook")) state = tables.Column("state", verbose_name=_("State")) class Meta: name = "executions" verbose_name = _("Executions") class TaskTable(tables.DataTable): id = tables.Column("id", verbose_name=_("ID")) name = tables.Column("name", verbose_name=_("Name")) + parameters = tables.Column("parameters", verbose_name=_("Parameters")) - action = tables.Column("action", verbose_name=_("Action")) ? ^^ ^^^ ^^ ^^^ ^^ ^^^ + output = tables.Column("output", verbose_name=_("Output")) ? ^^ ^^^ ^^ ^^^ ^^ ^^^ state = tables.Column("state", verbose_name=_("State")) class Meta: name = "tasks" verbose_name = _("Tasks")
e388e3490502acac90ef4c249ba1af63b5698ab7
print_web_django/api/views.py
print_web_django/api/views.py
from rest_framework import viewsets from . import serializers, models class PrintJobViewSet(viewsets.ModelViewSet): serializer_class = serializers.PrintJobSerializer def get_queryset(self): return self.request.user.printjobs.all()
from rest_framework import viewsets from . import serializers, models class PrintJobViewSet(viewsets.ModelViewSet): serializer_class = serializers.PrintJobSerializer def get_queryset(self): return self.request.user.printjobs.all() def perform_create(self, serializer): # need to also pass the requests user on a create serializer.save(user=self.request.user)
Add user to posted print object
Add user to posted print object
Python
mit
aabmass/print-web,aabmass/print-web,aabmass/print-web
from rest_framework import viewsets from . import serializers, models class PrintJobViewSet(viewsets.ModelViewSet): serializer_class = serializers.PrintJobSerializer def get_queryset(self): return self.request.user.printjobs.all() + def perform_create(self, serializer): + # need to also pass the requests user on a create + serializer.save(user=self.request.user) +
Add user to posted print object
## Code Before: from rest_framework import viewsets from . import serializers, models class PrintJobViewSet(viewsets.ModelViewSet): serializer_class = serializers.PrintJobSerializer def get_queryset(self): return self.request.user.printjobs.all() ## Instruction: Add user to posted print object ## Code After: from rest_framework import viewsets from . import serializers, models class PrintJobViewSet(viewsets.ModelViewSet): serializer_class = serializers.PrintJobSerializer def get_queryset(self): return self.request.user.printjobs.all() def perform_create(self, serializer): # need to also pass the requests user on a create serializer.save(user=self.request.user)
from rest_framework import viewsets from . import serializers, models class PrintJobViewSet(viewsets.ModelViewSet): serializer_class = serializers.PrintJobSerializer def get_queryset(self): return self.request.user.printjobs.all() + + def perform_create(self, serializer): + # need to also pass the requests user on a create + serializer.save(user=self.request.user)
f88c2135ddc197283bbfb8b481774deb613571cf
python/raindrops/raindrops.py
python/raindrops/raindrops.py
def raindrops(number): if is_three_a_factor(number): return "Pling" return "{}".format(number) def is_three_a_factor(number): return number % 3 == 0
def raindrops(number): if is_three_a_factor(number): return "Pling" if is_five_a_factor(number): return "Plang" return "{}".format(number) def is_three_a_factor(number): return number % 3 == 0 def is_five_a_factor(number): return number % 5 == 0
Handle 5 as a factor
Handle 5 as a factor
Python
mit
rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism
def raindrops(number): if is_three_a_factor(number): return "Pling" + if is_five_a_factor(number): + return "Plang" return "{}".format(number) def is_three_a_factor(number): return number % 3 == 0 + def is_five_a_factor(number): + return number % 5 == 0 +
Handle 5 as a factor
## Code Before: def raindrops(number): if is_three_a_factor(number): return "Pling" return "{}".format(number) def is_three_a_factor(number): return number % 3 == 0 ## Instruction: Handle 5 as a factor ## Code After: def raindrops(number): if is_three_a_factor(number): return "Pling" if is_five_a_factor(number): return "Plang" return "{}".format(number) def is_three_a_factor(number): return number % 3 == 0 def is_five_a_factor(number): return number % 5 == 0
def raindrops(number): if is_three_a_factor(number): return "Pling" + if is_five_a_factor(number): + return "Plang" return "{}".format(number) def is_three_a_factor(number): return number % 3 == 0 + + def is_five_a_factor(number): + return number % 5 == 0
114f40dd282d1837db42ffb6625760d1483d3192
jfu/templatetags/jfutags.py
jfu/templatetags/jfutags.py
from django.core.context_processors import csrf from django.core.urlresolvers import reverse from django.template import Library, Context, loader register = Library() @register.simple_tag( takes_context = True ) def jfu( context, template_name = 'jfu/upload_form.html', upload_handler_name = 'jfu_upload' ): """ Displays a form for uploading files using jQuery File Upload. A user may supply both a custom template or a custom upload-handling URL name by supplying values for template_name and upload_handler_name respectively. """ context.update( { 'JQ_OPEN' : '{%', 'JQ_CLOSE' : '%}', 'upload_handler_url': reverse( upload_handler_name ), } ) # Use the request context variable, injected # by django.core.context_processors.request # to generate the CSRF token. context.update( csrf( context.get('request') ) ) t = loader.get_template( template_name ) return t.render( Context( context ) )
from django.core.context_processors import csrf from django.core.urlresolvers import reverse from django.template import Library, Context, loader register = Library() @register.simple_tag( takes_context = True ) def jfu( context, template_name = 'jfu/upload_form.html', upload_handler_name = 'jfu_upload', *args, **kwargs ): """ Displays a form for uploading files using jQuery File Upload. A user may supply both a custom template or a custom upload-handling URL name by supplying values for template_name and upload_handler_name respectively. """ context.update( { 'JQ_OPEN' : '{%', 'JQ_CLOSE' : '%}', 'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ), } ) # Use the request context variable, injected # by django.core.context_processors.request # to generate the CSRF token. context.update( csrf( context.get('request') ) ) t = loader.get_template( template_name ) return t.render( Context( context ) )
Allow args and kwargs to upload_handler_name
Allow args and kwargs to upload_handler_name Now can use args and kwargs for reverse url. Example in template: {% jfu 'core/core_fileuploader.html' 'core_upload' object_id=1 content_type_str='app.model' %}
Python
bsd-3-clause
Alem/django-jfu,dzhuang/django-jfu,Alem/django-jfu,dzhuang/django-jfu,Alem/django-jfu,dzhuang/django-jfu,dzhuang/django-jfu,Alem/django-jfu
from django.core.context_processors import csrf from django.core.urlresolvers import reverse from django.template import Library, Context, loader register = Library() @register.simple_tag( takes_context = True ) def jfu( context, template_name = 'jfu/upload_form.html', - upload_handler_name = 'jfu_upload' + upload_handler_name = 'jfu_upload', + *args, **kwargs ): """ Displays a form for uploading files using jQuery File Upload. A user may supply both a custom template or a custom upload-handling URL name by supplying values for template_name and upload_handler_name respectively. """ context.update( { 'JQ_OPEN' : '{%', 'JQ_CLOSE' : '%}', - 'upload_handler_url': reverse( upload_handler_name ), + 'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ), } ) # Use the request context variable, injected # by django.core.context_processors.request # to generate the CSRF token. context.update( csrf( context.get('request') ) ) t = loader.get_template( template_name ) return t.render( Context( context ) )
Allow args and kwargs to upload_handler_name
## Code Before: from django.core.context_processors import csrf from django.core.urlresolvers import reverse from django.template import Library, Context, loader register = Library() @register.simple_tag( takes_context = True ) def jfu( context, template_name = 'jfu/upload_form.html', upload_handler_name = 'jfu_upload' ): """ Displays a form for uploading files using jQuery File Upload. A user may supply both a custom template or a custom upload-handling URL name by supplying values for template_name and upload_handler_name respectively. """ context.update( { 'JQ_OPEN' : '{%', 'JQ_CLOSE' : '%}', 'upload_handler_url': reverse( upload_handler_name ), } ) # Use the request context variable, injected # by django.core.context_processors.request # to generate the CSRF token. context.update( csrf( context.get('request') ) ) t = loader.get_template( template_name ) return t.render( Context( context ) ) ## Instruction: Allow args and kwargs to upload_handler_name ## Code After: from django.core.context_processors import csrf from django.core.urlresolvers import reverse from django.template import Library, Context, loader register = Library() @register.simple_tag( takes_context = True ) def jfu( context, template_name = 'jfu/upload_form.html', upload_handler_name = 'jfu_upload', *args, **kwargs ): """ Displays a form for uploading files using jQuery File Upload. A user may supply both a custom template or a custom upload-handling URL name by supplying values for template_name and upload_handler_name respectively. """ context.update( { 'JQ_OPEN' : '{%', 'JQ_CLOSE' : '%}', 'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ), } ) # Use the request context variable, injected # by django.core.context_processors.request # to generate the CSRF token. context.update( csrf( context.get('request') ) ) t = loader.get_template( template_name ) return t.render( Context( context ) )
from django.core.context_processors import csrf from django.core.urlresolvers import reverse from django.template import Library, Context, loader register = Library() @register.simple_tag( takes_context = True ) def jfu( context, template_name = 'jfu/upload_form.html', - upload_handler_name = 'jfu_upload' ? ^ + upload_handler_name = 'jfu_upload', ? ^ + *args, **kwargs ): """ Displays a form for uploading files using jQuery File Upload. A user may supply both a custom template or a custom upload-handling URL name by supplying values for template_name and upload_handler_name respectively. """ context.update( { 'JQ_OPEN' : '{%', 'JQ_CLOSE' : '%}', - 'upload_handler_url': reverse( upload_handler_name ), + 'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ), ? ++++++++++++++++++++++++++ } ) # Use the request context variable, injected # by django.core.context_processors.request # to generate the CSRF token. context.update( csrf( context.get('request') ) ) t = loader.get_template( template_name ) return t.render( Context( context ) )
de1988304714b44e641a4c4ac50fa650887621d6
geoportail/geonames/views.py
geoportail/geonames/views.py
import unicodedata from django.http import HttpResponse from django.template.defaultfilters import slugify from django.utils.translation import ugettext as _ from .models import Town def autocomplete(request): if not 'q' in request.GET or len(request.GET['q']) < 3: response = HttpResponse() response.status_code = 204 return response query = slugify(request.GET['q']).replace('-', ' ').upper() if query.startswith('ST '): query = 'SAINT ' + query[3:] towns = Town.objects.filter( tokenized__startswith=query ).order_by('tokenized', 'postal_code')[:15] content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format( name=unicodedata.normalize('NFKD', t.name), county_name=t.county_name, lon=t.point.coords[0], lat=t.point.coords[1], ) for t in towns]) if not content: content = _('No results. Search is limited to city names.') return HttpResponse(content)
import json import unicodedata from django.http import HttpResponse from django.template.defaultfilters import slugify from django.utils.translation import ugettext as _ from .models import Town def autocomplete(request): if not 'q' in request.GET or len(request.GET['q']) < 3: response = HttpResponse() response.status_code = 204 return response query = slugify(request.GET['q']).replace('-', ' ').upper() if query.startswith('ST '): query = 'SAINT ' + query[3:] towns = Town.objects.filter( tokenized__startswith=query ).order_by('tokenized', 'postal_code')[:15] content = [{ "name": unicodedata.normalize('NFKD', t.name), "county_name": t.county_name, "lon": t.point.coords[0], "lat": t.point.coords[1], } for t in towns] if not content: content = [{'name': _('No results. Search is limited to city names.')}] return HttpResponse(json.dumps(content), content_type='application/json')
Return JSON in the autocomplete view
Return JSON in the autocomplete view
Python
bsd-3-clause
brutasse/geoportail,brutasse/geoportail,brutasse/geoportail
+ import json import unicodedata from django.http import HttpResponse from django.template.defaultfilters import slugify from django.utils.translation import ugettext as _ from .models import Town def autocomplete(request): if not 'q' in request.GET or len(request.GET['q']) < 3: response = HttpResponse() response.status_code = 204 return response query = slugify(request.GET['q']).replace('-', ' ').upper() if query.startswith('ST '): query = 'SAINT ' + query[3:] towns = Town.objects.filter( tokenized__startswith=query ).order_by('tokenized', 'postal_code')[:15] - content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format( + content = [{ - name=unicodedata.normalize('NFKD', t.name), + "name": unicodedata.normalize('NFKD', t.name), - county_name=t.county_name, + "county_name": t.county_name, - lon=t.point.coords[0], + "lon": t.point.coords[0], - lat=t.point.coords[1], + "lat": t.point.coords[1], - ) for t in towns]) + } for t in towns] if not content: - content = _('No results. Search is limited to city names.') + content = [{'name': _('No results. Search is limited to city names.')}] - return HttpResponse(content) + return HttpResponse(json.dumps(content), content_type='application/json')
Return JSON in the autocomplete view
## Code Before: import unicodedata from django.http import HttpResponse from django.template.defaultfilters import slugify from django.utils.translation import ugettext as _ from .models import Town def autocomplete(request): if not 'q' in request.GET or len(request.GET['q']) < 3: response = HttpResponse() response.status_code = 204 return response query = slugify(request.GET['q']).replace('-', ' ').upper() if query.startswith('ST '): query = 'SAINT ' + query[3:] towns = Town.objects.filter( tokenized__startswith=query ).order_by('tokenized', 'postal_code')[:15] content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format( name=unicodedata.normalize('NFKD', t.name), county_name=t.county_name, lon=t.point.coords[0], lat=t.point.coords[1], ) for t in towns]) if not content: content = _('No results. Search is limited to city names.') return HttpResponse(content) ## Instruction: Return JSON in the autocomplete view ## Code After: import json import unicodedata from django.http import HttpResponse from django.template.defaultfilters import slugify from django.utils.translation import ugettext as _ from .models import Town def autocomplete(request): if not 'q' in request.GET or len(request.GET['q']) < 3: response = HttpResponse() response.status_code = 204 return response query = slugify(request.GET['q']).replace('-', ' ').upper() if query.startswith('ST '): query = 'SAINT ' + query[3:] towns = Town.objects.filter( tokenized__startswith=query ).order_by('tokenized', 'postal_code')[:15] content = [{ "name": unicodedata.normalize('NFKD', t.name), "county_name": t.county_name, "lon": t.point.coords[0], "lat": t.point.coords[1], } for t in towns] if not content: content = [{'name': _('No results. Search is limited to city names.')}] return HttpResponse(json.dumps(content), content_type='application/json')
+ import json import unicodedata from django.http import HttpResponse from django.template.defaultfilters import slugify from django.utils.translation import ugettext as _ from .models import Town def autocomplete(request): if not 'q' in request.GET or len(request.GET['q']) < 3: response = HttpResponse() response.status_code = 204 return response query = slugify(request.GET['q']).replace('-', ' ').upper() if query.startswith('ST '): query = 'SAINT ' + query[3:] towns = Town.objects.filter( tokenized__startswith=query ).order_by('tokenized', 'postal_code')[:15] - content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format( + content = [{ - name=unicodedata.normalize('NFKD', t.name), ? ^ + "name": unicodedata.normalize('NFKD', t.name), ? + ^^^ - county_name=t.county_name, ? ^ + "county_name": t.county_name, ? + ^^^ - lon=t.point.coords[0], ? ^ + "lon": t.point.coords[0], ? + ^^^ - lat=t.point.coords[1], ? ^ + "lat": t.point.coords[1], ? + ^^^ - ) for t in towns]) ? ^ - + } for t in towns] ? ^ if not content: - content = _('No results. Search is limited to city names.') + content = [{'name': _('No results. Search is limited to city names.')}] ? ++++++++++ ++ - return HttpResponse(content) + return HttpResponse(json.dumps(content), content_type='application/json')
d4a0a85673b5d61b82c65e77efcd6518da719952
pmxbot/__init__.py
pmxbot/__init__.py
import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
Remove places default config. It doesn't appear to be used anywhere.
Remove places default config. It doesn't appear to be used anywhere.
Python
mit
yougov/pmxbot,yougov/pmxbot,yougov/pmxbot
import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], - places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
Remove places default config. It doesn't appear to be used anywhere.
## Code Before: import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object" ## Instruction: Remove places default config. It doesn't appear to be used anywhere. ## Code After: import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], - places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
c10badab9b93eb021b1942475c681042292c182c
scrapi/harvesters/boise_state.py
scrapi/harvesters/boise_state.py
''' Harvester for the ScholarWorks for the SHARE project Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc ''' from __future__ import unicode_literals from scrapi.base import OAIHarvester class Boise_stateHarvester(OAIHarvester): short_name = 'boise_state' long_name = 'ScholarWorks' url = 'http://scholarworks.boisestate.edu' base_url = 'http://scholarworks.boisestate.edu/do/oai/' property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format'] timezone_granularity = True
''' Harvester for the ScholarWorks for the SHARE project Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc ''' from __future__ import unicode_literals from scrapi.base import OAIHarvester class Boise_stateHarvester(OAIHarvester): short_name = 'boise_state' long_name = 'Boise State University ScholarWorks' url = 'http://scholarworks.boisestate.edu' base_url = 'http://scholarworks.boisestate.edu/do/oai/' property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format'] timezone_granularity = True
Update longname for Boise state
Update longname for Boise state
Python
apache-2.0
CenterForOpenScience/scrapi,CenterForOpenScience/scrapi
''' Harvester for the ScholarWorks for the SHARE project Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc ''' from __future__ import unicode_literals from scrapi.base import OAIHarvester class Boise_stateHarvester(OAIHarvester): short_name = 'boise_state' - long_name = 'ScholarWorks' + long_name = 'Boise State University ScholarWorks' url = 'http://scholarworks.boisestate.edu' base_url = 'http://scholarworks.boisestate.edu/do/oai/' property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format'] timezone_granularity = True
Update longname for Boise state
## Code Before: ''' Harvester for the ScholarWorks for the SHARE project Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc ''' from __future__ import unicode_literals from scrapi.base import OAIHarvester class Boise_stateHarvester(OAIHarvester): short_name = 'boise_state' long_name = 'ScholarWorks' url = 'http://scholarworks.boisestate.edu' base_url = 'http://scholarworks.boisestate.edu/do/oai/' property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format'] timezone_granularity = True ## Instruction: Update longname for Boise state ## Code After: ''' Harvester for the ScholarWorks for the SHARE project Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc ''' from __future__ import unicode_literals from scrapi.base import OAIHarvester class Boise_stateHarvester(OAIHarvester): short_name = 'boise_state' long_name = 'Boise State University ScholarWorks' url = 'http://scholarworks.boisestate.edu' base_url = 'http://scholarworks.boisestate.edu/do/oai/' property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format'] timezone_granularity = True
''' Harvester for the ScholarWorks for the SHARE project Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc ''' from __future__ import unicode_literals from scrapi.base import OAIHarvester class Boise_stateHarvester(OAIHarvester): short_name = 'boise_state' - long_name = 'ScholarWorks' + long_name = 'Boise State University ScholarWorks' url = 'http://scholarworks.boisestate.edu' base_url = 'http://scholarworks.boisestate.edu/do/oai/' property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format'] timezone_granularity = True
e6d7181ababaa9f08602c48e03d6557ddb6a4deb
tests/test_gio.py
tests/test_gio.py
import os import unittest from common import gio, gobject class TestInputStream(unittest.TestCase): def setUp(self): f = open("inputstream.txt", "w") f.write("testing") self._f = open("inputstream.txt", "r") self.stream = gio.unix.InputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("inputstream.txt") def testWrite(self): self.assertEquals(self.stream.read(), "testing") class TestOutputStream(unittest.TestCase): def setUp(self): self._f = open("outputstream.txt", "w") self.stream = gio.unix.OutputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("outputstream.txt") def testWrite(self): self.stream.write("testing") self.stream.close() self.failUnless(os.path.exists("outputstream.txt")) self.assertEquals(open("outputstream.txt").read(), "testing") def testWriteAsync(self): def callback(stream, result): loop.quit() f = gio.file_new_for_path("outputstream.txt") stream = f.read() stream.read_async(10240, 0, None, callback) loop = gobject.MainLoop() loop.run()
import os import unittest from common import gio, gobject class TestInputStream(unittest.TestCase): def setUp(self): f = open("inputstream.txt", "w") f.write("testing") self._f = open("inputstream.txt", "r") self.stream = gio.unix.InputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("inputstream.txt") def testRead(self): self.assertEquals(self.stream.read(), "testing") def testReadAsync(self): def callback(stream, result): self.assertEquals(stream.read_finish(result), len("testing")) loop.quit() self.stream.read_async(10240, 0, None, callback) loop = gobject.MainLoop() loop.run() class TestOutputStream(unittest.TestCase): def setUp(self): self._f = open("outputstream.txt", "w") self.stream = gio.unix.OutputStream(self._f.fileno(), False) self._f.flush() def tearDown(self): self._f.close() os.unlink("outputstream.txt") def testWrite(self): self.stream.write("testing") self.stream.close() self.failUnless(os.path.exists("outputstream.txt")) self.assertEquals(open("outputstream.txt").read(), "testing")
Reorganize tests and make them test more useful things
Reorganize tests and make them test more useful things svn path=/trunk/; revision=738
Python
lgpl-2.1
pexip/pygobject,GNOME/pygobject,davibe/pygobject,alexef/pygobject,davidmalcolm/pygobject,MathieuDuponchelle/pygobject,davidmalcolm/pygobject,Distrotech/pygobject,choeger/pygobject-cmake,sfeltman/pygobject,Distrotech/pygobject,MathieuDuponchelle/pygobject,GNOME/pygobject,thiblahute/pygobject,jdahlin/pygobject,atizo/pygobject,alexef/pygobject,jdahlin/pygobject,choeger/pygobject-cmake,thiblahute/pygobject,GNOME/pygobject,nzjrs/pygobject,Distrotech/pygobject,pexip/pygobject,pexip/pygobject,atizo/pygobject,davibe/pygobject,choeger/pygobject-cmake,alexef/pygobject,davibe/pygobject,davibe/pygobject,MathieuDuponchelle/pygobject,Distrotech/pygobject,jdahlin/pygobject,sfeltman/pygobject,nzjrs/pygobject,thiblahute/pygobject,atizo/pygobject,nzjrs/pygobject,davidmalcolm/pygobject,sfeltman/pygobject
import os import unittest from common import gio, gobject class TestInputStream(unittest.TestCase): def setUp(self): f = open("inputstream.txt", "w") f.write("testing") self._f = open("inputstream.txt", "r") self.stream = gio.unix.InputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("inputstream.txt") - def testWrite(self): + def testRead(self): self.assertEquals(self.stream.read(), "testing") + + def testReadAsync(self): + def callback(stream, result): + self.assertEquals(stream.read_finish(result), len("testing")) + loop.quit() + + self.stream.read_async(10240, 0, None, callback) + + loop = gobject.MainLoop() + loop.run() class TestOutputStream(unittest.TestCase): def setUp(self): self._f = open("outputstream.txt", "w") self.stream = gio.unix.OutputStream(self._f.fileno(), False) + self._f.flush() def tearDown(self): self._f.close() os.unlink("outputstream.txt") def testWrite(self): self.stream.write("testing") self.stream.close() self.failUnless(os.path.exists("outputstream.txt")) self.assertEquals(open("outputstream.txt").read(), "testing") - def testWriteAsync(self): - def callback(stream, result): - loop.quit() - f = gio.file_new_for_path("outputstream.txt") - stream = f.read() - stream.read_async(10240, 0, None, callback) - - loop = gobject.MainLoop() - loop.run() -
Reorganize tests and make them test more useful things
## Code Before: import os import unittest from common import gio, gobject class TestInputStream(unittest.TestCase): def setUp(self): f = open("inputstream.txt", "w") f.write("testing") self._f = open("inputstream.txt", "r") self.stream = gio.unix.InputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("inputstream.txt") def testWrite(self): self.assertEquals(self.stream.read(), "testing") class TestOutputStream(unittest.TestCase): def setUp(self): self._f = open("outputstream.txt", "w") self.stream = gio.unix.OutputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("outputstream.txt") def testWrite(self): self.stream.write("testing") self.stream.close() self.failUnless(os.path.exists("outputstream.txt")) self.assertEquals(open("outputstream.txt").read(), "testing") def testWriteAsync(self): def callback(stream, result): loop.quit() f = gio.file_new_for_path("outputstream.txt") stream = f.read() stream.read_async(10240, 0, None, callback) loop = gobject.MainLoop() loop.run() ## Instruction: Reorganize tests and make them test more useful things ## Code After: import os import unittest from common import gio, gobject class TestInputStream(unittest.TestCase): def setUp(self): f = open("inputstream.txt", "w") f.write("testing") self._f = open("inputstream.txt", "r") self.stream = gio.unix.InputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("inputstream.txt") def testRead(self): self.assertEquals(self.stream.read(), "testing") def testReadAsync(self): def callback(stream, result): self.assertEquals(stream.read_finish(result), len("testing")) loop.quit() self.stream.read_async(10240, 0, None, callback) loop = gobject.MainLoop() loop.run() class TestOutputStream(unittest.TestCase): def setUp(self): self._f = open("outputstream.txt", "w") self.stream = gio.unix.OutputStream(self._f.fileno(), False) self._f.flush() def tearDown(self): self._f.close() os.unlink("outputstream.txt") def testWrite(self): self.stream.write("testing") self.stream.close() self.failUnless(os.path.exists("outputstream.txt")) self.assertEquals(open("outputstream.txt").read(), "testing")
import os import unittest from common import gio, gobject class TestInputStream(unittest.TestCase): def setUp(self): f = open("inputstream.txt", "w") f.write("testing") self._f = open("inputstream.txt", "r") self.stream = gio.unix.InputStream(self._f.fileno(), False) def tearDown(self): self._f.close() os.unlink("inputstream.txt") - def testWrite(self): ? ^^^^ + def testRead(self): ? ^ ++ self.assertEquals(self.stream.read(), "testing") + + def testReadAsync(self): + def callback(stream, result): + self.assertEquals(stream.read_finish(result), len("testing")) + loop.quit() + + self.stream.read_async(10240, 0, None, callback) + + loop = gobject.MainLoop() + loop.run() class TestOutputStream(unittest.TestCase): def setUp(self): self._f = open("outputstream.txt", "w") self.stream = gio.unix.OutputStream(self._f.fileno(), False) + self._f.flush() def tearDown(self): self._f.close() os.unlink("outputstream.txt") def testWrite(self): self.stream.write("testing") self.stream.close() self.failUnless(os.path.exists("outputstream.txt")) self.assertEquals(open("outputstream.txt").read(), "testing") - def testWriteAsync(self): - def callback(stream, result): - loop.quit() - - f = gio.file_new_for_path("outputstream.txt") - stream = f.read() - stream.read_async(10240, 0, None, callback) - - loop = gobject.MainLoop() - loop.run()
0f77c9a48e84a3185794f97c5f15c7b13ae1d505
tests/test_vector2_angle.py
tests/test_vector2_angle.py
from ppb_vector import Vector2 from math import isclose import pytest @pytest.mark.parametrize("left, right, expected", [ (Vector2(1, 1), Vector2(0, -1), 135), (Vector2(1, 1), Vector2(-1, 0), 135), (Vector2(0, 1), Vector2(0, -1), 180), (Vector2(-1, -1), Vector2(1, 0), 135), (Vector2(-1, -1), Vector2(-1, 0), 45) ]) def test_angle(left, right, expected): assert isclose(left.angle(right), expected) assert isclose(right.angle(left), expected)
from ppb_vector import Vector2 from math import isclose import pytest @pytest.mark.parametrize("left, right, expected", [ (Vector2(1, 1), Vector2(0, -1), 135), (Vector2(1, 1), Vector2(-1, 0), 135), (Vector2(0, 1), Vector2(0, -1), 180), (Vector2(-1, -1), Vector2(1, 0), 135), (Vector2(-1, -1), Vector2(-1, 0), 45), (Vector2(1, 0), Vector2(0, 1), 90), (Vector2(1, 0), Vector2(1, 0), 0), ]) def test_angle(left, right, expected): assert isclose(left.angle(right), expected) assert isclose(right.angle(left), expected)
Add some additional test cases
Add some additional test cases
Python
artistic-2.0
ppb/ppb-vector,ppb/ppb-vector
from ppb_vector import Vector2 from math import isclose import pytest @pytest.mark.parametrize("left, right, expected", [ (Vector2(1, 1), Vector2(0, -1), 135), (Vector2(1, 1), Vector2(-1, 0), 135), (Vector2(0, 1), Vector2(0, -1), 180), (Vector2(-1, -1), Vector2(1, 0), 135), - (Vector2(-1, -1), Vector2(-1, 0), 45) + (Vector2(-1, -1), Vector2(-1, 0), 45), + (Vector2(1, 0), Vector2(0, 1), 90), + (Vector2(1, 0), Vector2(1, 0), 0), ]) def test_angle(left, right, expected): assert isclose(left.angle(right), expected) assert isclose(right.angle(left), expected)
Add some additional test cases
## Code Before: from ppb_vector import Vector2 from math import isclose import pytest @pytest.mark.parametrize("left, right, expected", [ (Vector2(1, 1), Vector2(0, -1), 135), (Vector2(1, 1), Vector2(-1, 0), 135), (Vector2(0, 1), Vector2(0, -1), 180), (Vector2(-1, -1), Vector2(1, 0), 135), (Vector2(-1, -1), Vector2(-1, 0), 45) ]) def test_angle(left, right, expected): assert isclose(left.angle(right), expected) assert isclose(right.angle(left), expected) ## Instruction: Add some additional test cases ## Code After: from ppb_vector import Vector2 from math import isclose import pytest @pytest.mark.parametrize("left, right, expected", [ (Vector2(1, 1), Vector2(0, -1), 135), (Vector2(1, 1), Vector2(-1, 0), 135), (Vector2(0, 1), Vector2(0, -1), 180), (Vector2(-1, -1), Vector2(1, 0), 135), (Vector2(-1, -1), Vector2(-1, 0), 45), (Vector2(1, 0), Vector2(0, 1), 90), (Vector2(1, 0), Vector2(1, 0), 0), ]) def test_angle(left, right, expected): assert isclose(left.angle(right), expected) assert isclose(right.angle(left), expected)
from ppb_vector import Vector2 from math import isclose import pytest @pytest.mark.parametrize("left, right, expected", [ (Vector2(1, 1), Vector2(0, -1), 135), (Vector2(1, 1), Vector2(-1, 0), 135), (Vector2(0, 1), Vector2(0, -1), 180), (Vector2(-1, -1), Vector2(1, 0), 135), - (Vector2(-1, -1), Vector2(-1, 0), 45) + (Vector2(-1, -1), Vector2(-1, 0), 45), ? + + (Vector2(1, 0), Vector2(0, 1), 90), + (Vector2(1, 0), Vector2(1, 0), 0), ]) def test_angle(left, right, expected): assert isclose(left.angle(right), expected) assert isclose(right.angle(left), expected)
eef7f3797a6228c9e06717c3be49801a10b457a5
registries/views.py
registries/views.py
from django.http import HttpResponse from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView from rest_framework.response import Response from registries.models import Organization from registries.serializers import DrillerListSerializer, DrillerSerializer class APIDrillerListCreateView(ListCreateAPIView): queryset = Organization.objects.all().select_related('province_state') serializer_class = DrillerSerializer def list(self, request): queryset = self.get_queryset() serializer = DrillerListSerializer(queryset, many=True) return Response(serializer.data) class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView): queryset = Organization.objects.all() lookup_field = "org_guid" serializer_class = DrillerSerializer # Create your views here. def index(request): return HttpResponse("TEST: Driller Register app home index.")
from django.http import HttpResponse from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView from rest_framework.response import Response from registries.models import Organization from registries.serializers import DrillerListSerializer, DrillerSerializer class APIDrillerListCreateView(ListCreateAPIView): """ get: Return a list of all registered drilling organizations post: Create a new drilling organization instance """ queryset = Organization.objects.all().select_related('province_state') serializer_class = DrillerSerializer def list(self, request): queryset = self.get_queryset() serializer = DrillerListSerializer(queryset, many=True) return Response(serializer.data) class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView): """ get: Return the specified drilling organization patch: Updates the specified drilling organization with the fields/values provided in the request body delete: Removes the specified drilling organization record """ queryset = Organization.objects.all() lookup_field = "org_guid" serializer_class = DrillerSerializer # Create your views here. def index(request): return HttpResponse("TEST: Driller Register app home index.")
Add docstrings to view classes
Add docstrings to view classes
Python
apache-2.0
bcgov/gwells,rstens/gwells,rstens/gwells,rstens/gwells,bcgov/gwells,bcgov/gwells,bcgov/gwells,rstens/gwells
from django.http import HttpResponse from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView from rest_framework.response import Response from registries.models import Organization from registries.serializers import DrillerListSerializer, DrillerSerializer class APIDrillerListCreateView(ListCreateAPIView): + """ + get: + Return a list of all registered drilling organizations + + post: + Create a new drilling organization instance + """ + queryset = Organization.objects.all().select_related('province_state') serializer_class = DrillerSerializer def list(self, request): queryset = self.get_queryset() serializer = DrillerListSerializer(queryset, many=True) return Response(serializer.data) class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView): + """ + get: + Return the specified drilling organization + + patch: + Updates the specified drilling organization with the fields/values provided in the request body + + delete: + Removes the specified drilling organization record + """ + queryset = Organization.objects.all() lookup_field = "org_guid" serializer_class = DrillerSerializer # Create your views here. def index(request): return HttpResponse("TEST: Driller Register app home index.")
Add docstrings to view classes
## Code Before: from django.http import HttpResponse from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView from rest_framework.response import Response from registries.models import Organization from registries.serializers import DrillerListSerializer, DrillerSerializer class APIDrillerListCreateView(ListCreateAPIView): queryset = Organization.objects.all().select_related('province_state') serializer_class = DrillerSerializer def list(self, request): queryset = self.get_queryset() serializer = DrillerListSerializer(queryset, many=True) return Response(serializer.data) class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView): queryset = Organization.objects.all() lookup_field = "org_guid" serializer_class = DrillerSerializer # Create your views here. def index(request): return HttpResponse("TEST: Driller Register app home index.") ## Instruction: Add docstrings to view classes ## Code After: from django.http import HttpResponse from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView from rest_framework.response import Response from registries.models import Organization from registries.serializers import DrillerListSerializer, DrillerSerializer class APIDrillerListCreateView(ListCreateAPIView): """ get: Return a list of all registered drilling organizations post: Create a new drilling organization instance """ queryset = Organization.objects.all().select_related('province_state') serializer_class = DrillerSerializer def list(self, request): queryset = self.get_queryset() serializer = DrillerListSerializer(queryset, many=True) return Response(serializer.data) class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView): """ get: Return the specified drilling organization patch: Updates the specified drilling organization with the fields/values provided in the request body delete: Removes the specified drilling organization record """ queryset = Organization.objects.all() lookup_field = "org_guid" serializer_class = DrillerSerializer # Create your views here. def index(request): return HttpResponse("TEST: Driller Register app home index.")
from django.http import HttpResponse from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView from rest_framework.response import Response from registries.models import Organization from registries.serializers import DrillerListSerializer, DrillerSerializer class APIDrillerListCreateView(ListCreateAPIView): + """ + get: + Return a list of all registered drilling organizations + + post: + Create a new drilling organization instance + """ + queryset = Organization.objects.all().select_related('province_state') serializer_class = DrillerSerializer def list(self, request): queryset = self.get_queryset() serializer = DrillerListSerializer(queryset, many=True) return Response(serializer.data) class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView): + """ + get: + Return the specified drilling organization + + patch: + Updates the specified drilling organization with the fields/values provided in the request body + + delete: + Removes the specified drilling organization record + """ + queryset = Organization.objects.all() lookup_field = "org_guid" serializer_class = DrillerSerializer # Create your views here. def index(request): return HttpResponse("TEST: Driller Register app home index.")
1a6344ea1fac51a8024e1803a0391662d4ab81e0
pyeda/boolalg/vexpr.py
pyeda/boolalg/vexpr.py
from pyeda.boolalg import expr from pyeda.boolalg import bfarray def bitvec(name, *dims): """Return a new array of given dimensions, filled with Expressions. Parameters ---------- name : str dims : (int or (int, int)) An int N means a slice from [0:N] A tuple (M, N) means a slice from [M:N] """ if dims: return bfarray.exprvars(name, *dims) else: return expr.exprvar(name)
from warnings import warn from pyeda.boolalg import expr from pyeda.boolalg import bfarray def bitvec(name, *dims): """Return a new array of given dimensions, filled with Expressions. Parameters ---------- name : str dims : (int or (int, int)) An int N means a slice from [0:N] A tuple (M, N) means a slice from [M:N] """ warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.") if dims: return bfarray.exprvars(name, *dims) else: return expr.exprvar(name)
Add deprecation warning to bitvec function
Add deprecation warning to bitvec function
Python
bsd-2-clause
pombredanne/pyeda,GtTmy/pyeda,karissa/pyeda,sschnug/pyeda,sschnug/pyeda,cjdrake/pyeda,sschnug/pyeda,cjdrake/pyeda,GtTmy/pyeda,GtTmy/pyeda,karissa/pyeda,pombredanne/pyeda,karissa/pyeda,cjdrake/pyeda,pombredanne/pyeda
+ + from warnings import warn from pyeda.boolalg import expr from pyeda.boolalg import bfarray def bitvec(name, *dims): """Return a new array of given dimensions, filled with Expressions. Parameters ---------- name : str dims : (int or (int, int)) An int N means a slice from [0:N] A tuple (M, N) means a slice from [M:N] """ + warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.") if dims: return bfarray.exprvars(name, *dims) else: return expr.exprvar(name)
Add deprecation warning to bitvec function
## Code Before: from pyeda.boolalg import expr from pyeda.boolalg import bfarray def bitvec(name, *dims): """Return a new array of given dimensions, filled with Expressions. Parameters ---------- name : str dims : (int or (int, int)) An int N means a slice from [0:N] A tuple (M, N) means a slice from [M:N] """ if dims: return bfarray.exprvars(name, *dims) else: return expr.exprvar(name) ## Instruction: Add deprecation warning to bitvec function ## Code After: from warnings import warn from pyeda.boolalg import expr from pyeda.boolalg import bfarray def bitvec(name, *dims): """Return a new array of given dimensions, filled with Expressions. Parameters ---------- name : str dims : (int or (int, int)) An int N means a slice from [0:N] A tuple (M, N) means a slice from [M:N] """ warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.") if dims: return bfarray.exprvars(name, *dims) else: return expr.exprvar(name)
+ + from warnings import warn from pyeda.boolalg import expr from pyeda.boolalg import bfarray def bitvec(name, *dims): """Return a new array of given dimensions, filled with Expressions. Parameters ---------- name : str dims : (int or (int, int)) An int N means a slice from [0:N] A tuple (M, N) means a slice from [M:N] """ + warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.") if dims: return bfarray.exprvars(name, *dims) else: return expr.exprvar(name)
29c3d87881ce9c57478eb821da60d77e9f5eeb48
eventsourcing/application/base.py
eventsourcing/application/base.py
from abc import abstractmethod, ABCMeta from six import with_metaclass from eventsourcing.infrastructure.event_store import EventStore from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber class EventSourcingApplication(with_metaclass(ABCMeta)): def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False): self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls, json_decoder_cls=json_decoder_cls, cipher=cipher, always_encrypt=always_encrypt_stored_events) self.event_store = self.create_event_store() self.persistence_subscriber = self.create_persistence_subscriber() @abstractmethod def create_stored_event_repo(self, **kwargs): """Returns an instance of a subclass of StoredEventRepository. :rtype: StoredEventRepository """ def create_event_store(self): return EventStore(self.stored_event_repo) def create_persistence_subscriber(self): return PersistenceSubscriber(self.event_store) def close(self): self.persistence_subscriber.close() self.stored_event_repo = None self.event_store = None self.persistence_subscriber = None def __enter__(self): return self def __exit__(self, *_): self.close()
from abc import abstractmethod, ABCMeta from six import with_metaclass from eventsourcing.infrastructure.event_store import EventStore from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber class EventSourcingApplication(with_metaclass(ABCMeta)): persist_events = True def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False): self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls, json_decoder_cls=json_decoder_cls, cipher=cipher, always_encrypt=always_encrypt_stored_events) self.event_store = self.create_event_store() if self.persist_events: self.persistence_subscriber = self.create_persistence_subscriber() else: self.persistence_subscriber = None @abstractmethod def create_stored_event_repo(self, **kwargs): """Returns an instance of a subclass of StoredEventRepository. :rtype: StoredEventRepository """ def create_event_store(self): return EventStore(self.stored_event_repo) def create_persistence_subscriber(self): return PersistenceSubscriber(self.event_store) def close(self): if self.persistence_subscriber: self.persistence_subscriber.close() self.stored_event_repo = None self.event_store = None self.persistence_subscriber = None def __enter__(self): return self def __exit__(self, *_): self.close()
Allow to disable events persistence at app class
Allow to disable events persistence at app class
Python
bsd-3-clause
johnbywater/eventsourcing,johnbywater/eventsourcing
from abc import abstractmethod, ABCMeta from six import with_metaclass from eventsourcing.infrastructure.event_store import EventStore from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber class EventSourcingApplication(with_metaclass(ABCMeta)): + persist_events = True def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False): self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls, json_decoder_cls=json_decoder_cls, cipher=cipher, always_encrypt=always_encrypt_stored_events) self.event_store = self.create_event_store() + if self.persist_events: - self.persistence_subscriber = self.create_persistence_subscriber() + self.persistence_subscriber = self.create_persistence_subscriber() + else: + self.persistence_subscriber = None @abstractmethod def create_stored_event_repo(self, **kwargs): """Returns an instance of a subclass of StoredEventRepository. :rtype: StoredEventRepository """ def create_event_store(self): return EventStore(self.stored_event_repo) def create_persistence_subscriber(self): return PersistenceSubscriber(self.event_store) def close(self): + if self.persistence_subscriber: - self.persistence_subscriber.close() + self.persistence_subscriber.close() self.stored_event_repo = None self.event_store = None self.persistence_subscriber = None def __enter__(self): return self def __exit__(self, *_): self.close()
Allow to disable events persistence at app class
## Code Before: from abc import abstractmethod, ABCMeta from six import with_metaclass from eventsourcing.infrastructure.event_store import EventStore from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber class EventSourcingApplication(with_metaclass(ABCMeta)): def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False): self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls, json_decoder_cls=json_decoder_cls, cipher=cipher, always_encrypt=always_encrypt_stored_events) self.event_store = self.create_event_store() self.persistence_subscriber = self.create_persistence_subscriber() @abstractmethod def create_stored_event_repo(self, **kwargs): """Returns an instance of a subclass of StoredEventRepository. :rtype: StoredEventRepository """ def create_event_store(self): return EventStore(self.stored_event_repo) def create_persistence_subscriber(self): return PersistenceSubscriber(self.event_store) def close(self): self.persistence_subscriber.close() self.stored_event_repo = None self.event_store = None self.persistence_subscriber = None def __enter__(self): return self def __exit__(self, *_): self.close() ## Instruction: Allow to disable events persistence at app class ## Code After: from abc import abstractmethod, ABCMeta from six import with_metaclass from eventsourcing.infrastructure.event_store import EventStore from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber class EventSourcingApplication(with_metaclass(ABCMeta)): persist_events = True def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False): self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls, json_decoder_cls=json_decoder_cls, cipher=cipher, always_encrypt=always_encrypt_stored_events) self.event_store = self.create_event_store() if self.persist_events: self.persistence_subscriber = self.create_persistence_subscriber() else: self.persistence_subscriber = None @abstractmethod def create_stored_event_repo(self, **kwargs): """Returns an instance of a subclass of StoredEventRepository. :rtype: StoredEventRepository """ def create_event_store(self): return EventStore(self.stored_event_repo) def create_persistence_subscriber(self): return PersistenceSubscriber(self.event_store) def close(self): if self.persistence_subscriber: self.persistence_subscriber.close() self.stored_event_repo = None self.event_store = None self.persistence_subscriber = None def __enter__(self): return self def __exit__(self, *_): self.close()
from abc import abstractmethod, ABCMeta from six import with_metaclass from eventsourcing.infrastructure.event_store import EventStore from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber class EventSourcingApplication(with_metaclass(ABCMeta)): + persist_events = True def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False): self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls, json_decoder_cls=json_decoder_cls, cipher=cipher, always_encrypt=always_encrypt_stored_events) self.event_store = self.create_event_store() + if self.persist_events: - self.persistence_subscriber = self.create_persistence_subscriber() + self.persistence_subscriber = self.create_persistence_subscriber() ? ++++ + else: + self.persistence_subscriber = None @abstractmethod def create_stored_event_repo(self, **kwargs): """Returns an instance of a subclass of StoredEventRepository. :rtype: StoredEventRepository """ def create_event_store(self): return EventStore(self.stored_event_repo) def create_persistence_subscriber(self): return PersistenceSubscriber(self.event_store) def close(self): + if self.persistence_subscriber: - self.persistence_subscriber.close() + self.persistence_subscriber.close() ? ++++ self.stored_event_repo = None self.event_store = None self.persistence_subscriber = None def __enter__(self): return self def __exit__(self, *_): self.close()
e8537feff53310913047d06d95f4dd8e9dace1da
flow_workflow/historian/handler.py
flow_workflow/historian/handler.py
from flow import exit_codes from flow.configuration.settings.injector import setting from flow.handler import Handler from flow.util.exit import exit_process from flow_workflow.historian.messages import UpdateMessage from injector import inject from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from twisted.internet import defer import flow.interfaces import logging import os LOG = logging.getLogger(__name__) @inject(storage=flow.interfaces.IStorage, queue_name=setting('workflow.historian.queue')) class WorkflowHistorianMessageHandler(Handler): message_class = UpdateMessage def _handle_message(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) return defer.succeed(None) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore, " "because it lost access to Oracle... exiting.") exit_process(exit_codes.EXECUTE_FAILURE)
from flow import exit_codes from flow.configuration.settings.injector import setting from flow.handler import Handler from flow.util.exit import exit_process from flow_workflow.historian.messages import UpdateMessage from injector import inject from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError from twisted.internet import defer import flow.interfaces import logging import os LOG = logging.getLogger(__name__) @inject(storage=flow.interfaces.IStorage, queue_name=setting('workflow.historian.queue')) class WorkflowHistorianMessageHandler(Handler): message_class = UpdateMessage def _handle_message(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) return defer.succeed(None) except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError): LOG.exception("This historian cannot handle messages anymore, " "because it lost access to Oracle... exiting.") exit_process(exit_codes.EXECUTE_FAILURE)
Add DatabaseError to list of errors that kill a historian
Add DatabaseError to list of errors that kill a historian
Python
agpl-3.0
genome/flow-workflow,genome/flow-workflow,genome/flow-workflow
from flow import exit_codes from flow.configuration.settings.injector import setting from flow.handler import Handler from flow.util.exit import exit_process from flow_workflow.historian.messages import UpdateMessage from injector import inject - from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError + from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError from twisted.internet import defer import flow.interfaces import logging import os LOG = logging.getLogger(__name__) @inject(storage=flow.interfaces.IStorage, queue_name=setting('workflow.historian.queue')) class WorkflowHistorianMessageHandler(Handler): message_class = UpdateMessage def _handle_message(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) return defer.succeed(None) - except (ResourceClosedError, TimeoutError, DisconnectionError): + except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError): LOG.exception("This historian cannot handle messages anymore, " "because it lost access to Oracle... exiting.") exit_process(exit_codes.EXECUTE_FAILURE)
Add DatabaseError to list of errors that kill a historian
## Code Before: from flow import exit_codes from flow.configuration.settings.injector import setting from flow.handler import Handler from flow.util.exit import exit_process from flow_workflow.historian.messages import UpdateMessage from injector import inject from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from twisted.internet import defer import flow.interfaces import logging import os LOG = logging.getLogger(__name__) @inject(storage=flow.interfaces.IStorage, queue_name=setting('workflow.historian.queue')) class WorkflowHistorianMessageHandler(Handler): message_class = UpdateMessage def _handle_message(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) return defer.succeed(None) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore, " "because it lost access to Oracle... exiting.") exit_process(exit_codes.EXECUTE_FAILURE) ## Instruction: Add DatabaseError to list of errors that kill a historian ## Code After: from flow import exit_codes from flow.configuration.settings.injector import setting from flow.handler import Handler from flow.util.exit import exit_process from flow_workflow.historian.messages import UpdateMessage from injector import inject from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError from twisted.internet import defer import flow.interfaces import logging import os LOG = logging.getLogger(__name__) @inject(storage=flow.interfaces.IStorage, queue_name=setting('workflow.historian.queue')) class WorkflowHistorianMessageHandler(Handler): message_class = UpdateMessage def _handle_message(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) return defer.succeed(None) except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError): LOG.exception("This historian cannot handle messages anymore, " "because it lost access to Oracle... exiting.") exit_process(exit_codes.EXECUTE_FAILURE)
from flow import exit_codes from flow.configuration.settings.injector import setting from flow.handler import Handler from flow.util.exit import exit_process from flow_workflow.historian.messages import UpdateMessage from injector import inject - from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError + from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError ? +++++++++++++++ from twisted.internet import defer import flow.interfaces import logging import os LOG = logging.getLogger(__name__) @inject(storage=flow.interfaces.IStorage, queue_name=setting('workflow.historian.queue')) class WorkflowHistorianMessageHandler(Handler): message_class = UpdateMessage def _handle_message(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) return defer.succeed(None) - except (ResourceClosedError, TimeoutError, DisconnectionError): + except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError): ? +++++++++++++++ LOG.exception("This historian cannot handle messages anymore, " "because it lost access to Oracle... exiting.") exit_process(exit_codes.EXECUTE_FAILURE)
d4da069b43174482f3a75e9553e8283be905fa16
cla_public/apps/base/filters.py
cla_public/apps/base/filters.py
"Jinja custom filters" import re from cla_public.apps.base import base from babel.dates import format_datetime @base.app_template_filter() def datetime(dt, format='medium', locale='en_GB'): if format == 'full': format = "EEEE, d MMMM y 'at' HH:mm" elif format == 'medium': format = "EE, dd/MM/y 'at' h:mma" elif format == 'short': format = "dd/MM/y, h:mma" return format_datetime(dt, format, locale=locale) @base.app_template_filter() def url_to_human(value): return re.sub(r'(^https?://)|(/$)', '', value) @base.app_template_filter() def human_to_url(value): return re.sub(r'^((?!https?://).*)', r'http://\1', value) @base.app_template_filter() def query_to_dict(value, prop=None): result = parse_qs(urlparse(value).query) if prop: result = result[prop] return result
"Jinja custom filters" import re from urlparse import urlparse, parse_qs from cla_public.apps.base import base from babel.dates import format_datetime @base.app_template_filter() def datetime(dt, format='medium', locale='en_GB'): if format == 'full': format = "EEEE, d MMMM y 'at' HH:mm" elif format == 'medium': format = "EE, dd/MM/y 'at' h:mma" elif format == 'short': format = "dd/MM/y, h:mma" return format_datetime(dt, format, locale=locale) @base.app_template_filter() def url_to_human(value): return re.sub(r'(^https?://)|(/$)', '', value) @base.app_template_filter() def human_to_url(value): return re.sub(r'^((?!https?://).*)', r'http://\1', value) @base.app_template_filter() def query_to_dict(value, prop=None): result = parse_qs(urlparse(value).query) if prop: result = result[prop] return result
Add Jinja filter to convert URL params to dict
BE: Add Jinja filter to convert URL params to dict
Python
mit
ministryofjustice/cla_public,ministryofjustice/cla_public,ministryofjustice/cla_public,ministryofjustice/cla_public
"Jinja custom filters" import re + from urlparse import urlparse, parse_qs from cla_public.apps.base import base from babel.dates import format_datetime @base.app_template_filter() def datetime(dt, format='medium', locale='en_GB'): if format == 'full': format = "EEEE, d MMMM y 'at' HH:mm" elif format == 'medium': format = "EE, dd/MM/y 'at' h:mma" elif format == 'short': format = "dd/MM/y, h:mma" return format_datetime(dt, format, locale=locale) @base.app_template_filter() def url_to_human(value): return re.sub(r'(^https?://)|(/$)', '', value) @base.app_template_filter() def human_to_url(value): return re.sub(r'^((?!https?://).*)', r'http://\1', value) @base.app_template_filter() def query_to_dict(value, prop=None): result = parse_qs(urlparse(value).query) if prop: result = result[prop] return result
Add Jinja filter to convert URL params to dict
## Code Before: "Jinja custom filters" import re from cla_public.apps.base import base from babel.dates import format_datetime @base.app_template_filter() def datetime(dt, format='medium', locale='en_GB'): if format == 'full': format = "EEEE, d MMMM y 'at' HH:mm" elif format == 'medium': format = "EE, dd/MM/y 'at' h:mma" elif format == 'short': format = "dd/MM/y, h:mma" return format_datetime(dt, format, locale=locale) @base.app_template_filter() def url_to_human(value): return re.sub(r'(^https?://)|(/$)', '', value) @base.app_template_filter() def human_to_url(value): return re.sub(r'^((?!https?://).*)', r'http://\1', value) @base.app_template_filter() def query_to_dict(value, prop=None): result = parse_qs(urlparse(value).query) if prop: result = result[prop] return result ## Instruction: Add Jinja filter to convert URL params to dict ## Code After: "Jinja custom filters" import re from urlparse import urlparse, parse_qs from cla_public.apps.base import base from babel.dates import format_datetime @base.app_template_filter() def datetime(dt, format='medium', locale='en_GB'): if format == 'full': format = "EEEE, d MMMM y 'at' HH:mm" elif format == 'medium': format = "EE, dd/MM/y 'at' h:mma" elif format == 'short': format = "dd/MM/y, h:mma" return format_datetime(dt, format, locale=locale) @base.app_template_filter() def url_to_human(value): return re.sub(r'(^https?://)|(/$)', '', value) @base.app_template_filter() def human_to_url(value): return re.sub(r'^((?!https?://).*)', r'http://\1', value) @base.app_template_filter() def query_to_dict(value, prop=None): result = parse_qs(urlparse(value).query) if prop: result = result[prop] return result
"Jinja custom filters" import re + from urlparse import urlparse, parse_qs from cla_public.apps.base import base from babel.dates import format_datetime @base.app_template_filter() def datetime(dt, format='medium', locale='en_GB'): if format == 'full': format = "EEEE, d MMMM y 'at' HH:mm" elif format == 'medium': format = "EE, dd/MM/y 'at' h:mma" elif format == 'short': format = "dd/MM/y, h:mma" return format_datetime(dt, format, locale=locale) @base.app_template_filter() def url_to_human(value): return re.sub(r'(^https?://)|(/$)', '', value) @base.app_template_filter() def human_to_url(value): return re.sub(r'^((?!https?://).*)', r'http://\1', value) @base.app_template_filter() def query_to_dict(value, prop=None): result = parse_qs(urlparse(value).query) if prop: result = result[prop] return result
20f0d90f5c64322864ad5fda4b4c9314e6c1cb11
run.py
run.py
import sys from kitchen.text.converters import getwriter from utils.log import getLogger, open_log, close_log from utils.misc import output_exception from system.factory_manager import Manager sys.stdout = getwriter('utf-8')(sys.stdout) sys.stderr = getwriter('utf-8')(sys.stderr) open_log("output.log") logger = getLogger("System") logger.info("Starting up..") try: manager = Manager() except Exception: logger.critical("Runtime error - process cannot continue!") output_exception(logger) finally: close_log("output.log") try: raw_input("Press enter to exit.") except: pass
import os import sys from kitchen.text.converters import getwriter from utils.log import getLogger, open_log, close_log from utils.misc import output_exception from system.factory_manager import Manager sys.stdout = getwriter('utf-8')(sys.stdout) sys.stderr = getwriter('utf-8')(sys.stderr) if not os.path.exists("logs"): os.mkdir("logs") open_log("output.log") logger = getLogger("System") logger.info("Starting up..") try: manager = Manager() except Exception: logger.critical("Runtime error - process cannot continue!") output_exception(logger) finally: close_log("output.log") try: raw_input("Press enter to exit.") except: pass
Create logs folder if it doesn't exist (to prevent errors)
Create logs folder if it doesn't exist (to prevent errors)
Python
artistic-2.0
UltrosBot/Ultros,UltrosBot/Ultros
+ import os import sys from kitchen.text.converters import getwriter from utils.log import getLogger, open_log, close_log from utils.misc import output_exception from system.factory_manager import Manager sys.stdout = getwriter('utf-8')(sys.stdout) sys.stderr = getwriter('utf-8')(sys.stderr) + + if not os.path.exists("logs"): + os.mkdir("logs") open_log("output.log") logger = getLogger("System") logger.info("Starting up..") try: manager = Manager() except Exception: logger.critical("Runtime error - process cannot continue!") output_exception(logger) finally: close_log("output.log") try: raw_input("Press enter to exit.") except: pass
Create logs folder if it doesn't exist (to prevent errors)
## Code Before: import sys from kitchen.text.converters import getwriter from utils.log import getLogger, open_log, close_log from utils.misc import output_exception from system.factory_manager import Manager sys.stdout = getwriter('utf-8')(sys.stdout) sys.stderr = getwriter('utf-8')(sys.stderr) open_log("output.log") logger = getLogger("System") logger.info("Starting up..") try: manager = Manager() except Exception: logger.critical("Runtime error - process cannot continue!") output_exception(logger) finally: close_log("output.log") try: raw_input("Press enter to exit.") except: pass ## Instruction: Create logs folder if it doesn't exist (to prevent errors) ## Code After: import os import sys from kitchen.text.converters import getwriter from utils.log import getLogger, open_log, close_log from utils.misc import output_exception from system.factory_manager import Manager sys.stdout = getwriter('utf-8')(sys.stdout) sys.stderr = getwriter('utf-8')(sys.stderr) if not os.path.exists("logs"): os.mkdir("logs") open_log("output.log") logger = getLogger("System") logger.info("Starting up..") try: manager = Manager() except Exception: logger.critical("Runtime error - process cannot continue!") output_exception(logger) finally: close_log("output.log") try: raw_input("Press enter to exit.") except: pass
+ import os import sys from kitchen.text.converters import getwriter from utils.log import getLogger, open_log, close_log from utils.misc import output_exception from system.factory_manager import Manager sys.stdout = getwriter('utf-8')(sys.stdout) sys.stderr = getwriter('utf-8')(sys.stderr) + + if not os.path.exists("logs"): + os.mkdir("logs") open_log("output.log") logger = getLogger("System") logger.info("Starting up..") try: manager = Manager() except Exception: logger.critical("Runtime error - process cannot continue!") output_exception(logger) finally: close_log("output.log") try: raw_input("Press enter to exit.") except: pass
0cecbabd2d594bfc2ca57e522658d13eda2bc6a8
pipdiff/pipdiff.py
pipdiff/pipdiff.py
from pkg_resources import parse_version from xmlrpclib import ServerProxy pypi = ServerProxy("http://pypi.python.org/pypi") def main(): try: from pip import get_installed_distributions except ImportError: from sys import exit exit("pip not available") for distribution in sorted(get_installed_distributions(), key=lambda distribution: distribution.project_name): remote = '' project_name = distribution.project_name releases = pypi.package_releases(project_name) if not releases: pypi.package_releases(project_name.capitalize()) if releases: version = parse_version(releases[0]) if version > distribution.parsed_version: remote = "PyPI:{0}=={1}".format(project_name, releases[0]) else: remote = "PyPI:{0} not found".format(project_name) local = "{0}=={1}".format(project_name, distribution.version) print "{0:40} {1}".format(local, remote) return True if __name__ == '__main__': main()
from pkg_resources import parse_version try: from xmlrpclib import ServerProxy except ImportError: import xmlrpc.client try: pypi = ServerProxy("http://pypi.python.org/pypi") except NameError: pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi") def main(): try: from pip import get_installed_distributions except ImportError: from sys import exit exit("pip not available") for distribution in sorted(get_installed_distributions(), key=lambda distribution: distribution.project_name): remote = '' project_name = distribution.project_name releases = pypi.package_releases(project_name) if not releases: pypi.package_releases(project_name.capitalize()) if releases: version = parse_version(releases[0]) if str(version) > str(distribution.parsed_version): remote = "PyPI:{0}=={1}".format(project_name, releases[0]) else: remote = "PyPI:{0} not found".format(project_name) local = "{0}=={1}".format(project_name, distribution.version) print("{0:40} {1}".format(local, remote)) return True if __name__ == '__main__': main()
Add support for Python 3
Add support for Python 3
Python
bsd-3-clause
ogt/pipdiff
from pkg_resources import parse_version + try: - from xmlrpclib import ServerProxy + from xmlrpclib import ServerProxy + except ImportError: + import xmlrpc.client + try: - pypi = ServerProxy("http://pypi.python.org/pypi") + pypi = ServerProxy("http://pypi.python.org/pypi") - + except NameError: + pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi") def main(): try: from pip import get_installed_distributions except ImportError: from sys import exit exit("pip not available") for distribution in sorted(get_installed_distributions(), key=lambda distribution: distribution.project_name): remote = '' project_name = distribution.project_name releases = pypi.package_releases(project_name) if not releases: pypi.package_releases(project_name.capitalize()) if releases: version = parse_version(releases[0]) - - if version > distribution.parsed_version: + if str(version) > str(distribution.parsed_version): remote = "PyPI:{0}=={1}".format(project_name, releases[0]) else: remote = "PyPI:{0} not found".format(project_name) local = "{0}=={1}".format(project_name, distribution.version) - print "{0:40} {1}".format(local, remote) + print("{0:40} {1}".format(local, remote)) return True if __name__ == '__main__': main()
Add support for Python 3
## Code Before: from pkg_resources import parse_version from xmlrpclib import ServerProxy pypi = ServerProxy("http://pypi.python.org/pypi") def main(): try: from pip import get_installed_distributions except ImportError: from sys import exit exit("pip not available") for distribution in sorted(get_installed_distributions(), key=lambda distribution: distribution.project_name): remote = '' project_name = distribution.project_name releases = pypi.package_releases(project_name) if not releases: pypi.package_releases(project_name.capitalize()) if releases: version = parse_version(releases[0]) if version > distribution.parsed_version: remote = "PyPI:{0}=={1}".format(project_name, releases[0]) else: remote = "PyPI:{0} not found".format(project_name) local = "{0}=={1}".format(project_name, distribution.version) print "{0:40} {1}".format(local, remote) return True if __name__ == '__main__': main() ## Instruction: Add support for Python 3 ## Code After: from pkg_resources import parse_version try: from xmlrpclib import ServerProxy except ImportError: import xmlrpc.client try: pypi = ServerProxy("http://pypi.python.org/pypi") except NameError: pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi") def main(): try: from pip import get_installed_distributions except ImportError: from sys import exit exit("pip not available") for distribution in sorted(get_installed_distributions(), key=lambda distribution: distribution.project_name): remote = '' project_name = distribution.project_name releases = pypi.package_releases(project_name) if not releases: pypi.package_releases(project_name.capitalize()) if releases: version = parse_version(releases[0]) if str(version) > str(distribution.parsed_version): remote = "PyPI:{0}=={1}".format(project_name, releases[0]) else: remote = "PyPI:{0} not found".format(project_name) local = "{0}=={1}".format(project_name, distribution.version) print("{0:40} {1}".format(local, remote)) return True if __name__ == '__main__': main()
from pkg_resources import parse_version + try: - from xmlrpclib import ServerProxy + from xmlrpclib import ServerProxy ? ++++ + except ImportError: + import xmlrpc.client + try: - pypi = ServerProxy("http://pypi.python.org/pypi") + pypi = ServerProxy("http://pypi.python.org/pypi") ? ++++ - + except NameError: + pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi") def main(): try: from pip import get_installed_distributions except ImportError: from sys import exit exit("pip not available") for distribution in sorted(get_installed_distributions(), key=lambda distribution: distribution.project_name): remote = '' project_name = distribution.project_name releases = pypi.package_releases(project_name) if not releases: pypi.package_releases(project_name.capitalize()) if releases: version = parse_version(releases[0]) - - if version > distribution.parsed_version: + if str(version) > str(distribution.parsed_version): ? ++++ + ++++ + remote = "PyPI:{0}=={1}".format(project_name, releases[0]) else: remote = "PyPI:{0} not found".format(project_name) local = "{0}=={1}".format(project_name, distribution.version) - print "{0:40} {1}".format(local, remote) ? ^ + print("{0:40} {1}".format(local, remote)) ? ^ + return True if __name__ == '__main__': main()
93081d423a73a6b16e5adfb94247ffec23ef667c
api/base/authentication/backends.py
api/base/authentication/backends.py
from osf.models.user import OSFUser from framework.auth.core import get_user from django.contrib.auth.backends import ModelBackend # https://docs.djangoproject.com/en/1.8/topics/auth/customizing/ class ODMBackend(ModelBackend): def authenticate(self, username=None, password=None): return get_user(email=username, password=password) or None def get_user(self, user_id): try: user = OSFUser.objects.get(id=user_id) except OSFUser.DoesNotExist: user = OSFUser.load(user_id) return user
from osf.models.user import OSFUser from framework.auth.core import get_user from django.contrib.auth.backends import ModelBackend # https://docs.djangoproject.com/en/3.2/topics/auth/customizing/ class ODMBackend(ModelBackend): def authenticate(self, request, username=None, password=None, **kwargs): return get_user(email=username, password=password) or None def get_user(self, user_id): try: user = OSFUser.objects.get(id=user_id) except OSFUser.DoesNotExist: user = OSFUser.load(user_id) return user
Fix admin login failure for django upgrade
Fix admin login failure for django upgrade
Python
apache-2.0
Johnetordoff/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io
from osf.models.user import OSFUser from framework.auth.core import get_user from django.contrib.auth.backends import ModelBackend + - # https://docs.djangoproject.com/en/1.8/topics/auth/customizing/ + # https://docs.djangoproject.com/en/3.2/topics/auth/customizing/ class ODMBackend(ModelBackend): - def authenticate(self, username=None, password=None): + def authenticate(self, request, username=None, password=None, **kwargs): return get_user(email=username, password=password) or None def get_user(self, user_id): try: user = OSFUser.objects.get(id=user_id) except OSFUser.DoesNotExist: user = OSFUser.load(user_id) return user
Fix admin login failure for django upgrade
## Code Before: from osf.models.user import OSFUser from framework.auth.core import get_user from django.contrib.auth.backends import ModelBackend # https://docs.djangoproject.com/en/1.8/topics/auth/customizing/ class ODMBackend(ModelBackend): def authenticate(self, username=None, password=None): return get_user(email=username, password=password) or None def get_user(self, user_id): try: user = OSFUser.objects.get(id=user_id) except OSFUser.DoesNotExist: user = OSFUser.load(user_id) return user ## Instruction: Fix admin login failure for django upgrade ## Code After: from osf.models.user import OSFUser from framework.auth.core import get_user from django.contrib.auth.backends import ModelBackend # https://docs.djangoproject.com/en/3.2/topics/auth/customizing/ class ODMBackend(ModelBackend): def authenticate(self, request, username=None, password=None, **kwargs): return get_user(email=username, password=password) or None def get_user(self, user_id): try: user = OSFUser.objects.get(id=user_id) except OSFUser.DoesNotExist: user = OSFUser.load(user_id) return user
from osf.models.user import OSFUser from framework.auth.core import get_user from django.contrib.auth.backends import ModelBackend + - # https://docs.djangoproject.com/en/1.8/topics/auth/customizing/ ? ^ ^ + # https://docs.djangoproject.com/en/3.2/topics/auth/customizing/ ? ^ ^ class ODMBackend(ModelBackend): - def authenticate(self, username=None, password=None): + def authenticate(self, request, username=None, password=None, **kwargs): ? +++++++++ ++++++++++ return get_user(email=username, password=password) or None def get_user(self, user_id): try: user = OSFUser.objects.get(id=user_id) except OSFUser.DoesNotExist: user = OSFUser.load(user_id) return user
8925c3a827659e1983827368948e95e764a40585
utf9/__init__.py
utf9/__init__.py
from bitarray import bitarray as _bitarray def utf9encode(string): bits = _bitarray() for char in string: for idx, byte in enumerate(char.encode('utf-8')): bits.append(idx) bits.extend('{0:b}'.format(ord(byte)).zfill(8)) return bits.tobytes() def utf9decode(data): bits = _bitarray() bits.frombytes(data) chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9)) string = u'' codepoint = '' for chunk in chunks: if len(chunk) < 9: break if chunk[0] == 0: codepoint, string = '', string + codepoint.decode('utf-8') codepoint += chr(int(chunk[1:].to01(), 2)) return string + codepoint.decode('utf-8')
from bitarray import bitarray as _bitarray def utf9encode(string): """Takes a string and returns a utf9-encoded version.""" bits = _bitarray() for char in string: for idx, byte in enumerate(char.encode('utf-8')): bits.append(idx) bits.extend('{0:b}'.format(ord(byte)).zfill(8)) return bits.tobytes() def utf9decode(data): """Takes utf9-encoded data and returns the corresponding string.""" bits = _bitarray() bits.frombytes(data) chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9)) string = u'' codepoint = '' for chunk in chunks: if len(chunk) < 9: break if chunk[0] == 0: codepoint, string = '', string + codepoint.decode('utf-8') codepoint += chr(int(chunk[1:].to01(), 2)) return string + codepoint.decode('utf-8')
Add module and functions docstring
Add module and functions docstring
Python
mit
enricobacis/utf9
from bitarray import bitarray as _bitarray def utf9encode(string): + """Takes a string and returns a utf9-encoded version.""" bits = _bitarray() for char in string: for idx, byte in enumerate(char.encode('utf-8')): bits.append(idx) bits.extend('{0:b}'.format(ord(byte)).zfill(8)) return bits.tobytes() def utf9decode(data): + """Takes utf9-encoded data and returns the corresponding string.""" bits = _bitarray() bits.frombytes(data) chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9)) string = u'' codepoint = '' for chunk in chunks: if len(chunk) < 9: break if chunk[0] == 0: codepoint, string = '', string + codepoint.decode('utf-8') codepoint += chr(int(chunk[1:].to01(), 2)) return string + codepoint.decode('utf-8')
Add module and functions docstring
## Code Before: from bitarray import bitarray as _bitarray def utf9encode(string): bits = _bitarray() for char in string: for idx, byte in enumerate(char.encode('utf-8')): bits.append(idx) bits.extend('{0:b}'.format(ord(byte)).zfill(8)) return bits.tobytes() def utf9decode(data): bits = _bitarray() bits.frombytes(data) chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9)) string = u'' codepoint = '' for chunk in chunks: if len(chunk) < 9: break if chunk[0] == 0: codepoint, string = '', string + codepoint.decode('utf-8') codepoint += chr(int(chunk[1:].to01(), 2)) return string + codepoint.decode('utf-8') ## Instruction: Add module and functions docstring ## Code After: from bitarray import bitarray as _bitarray def utf9encode(string): """Takes a string and returns a utf9-encoded version.""" bits = _bitarray() for char in string: for idx, byte in enumerate(char.encode('utf-8')): bits.append(idx) bits.extend('{0:b}'.format(ord(byte)).zfill(8)) return bits.tobytes() def utf9decode(data): """Takes utf9-encoded data and returns the corresponding string.""" bits = _bitarray() bits.frombytes(data) chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9)) string = u'' codepoint = '' for chunk in chunks: if len(chunk) < 9: break if chunk[0] == 0: codepoint, string = '', string + codepoint.decode('utf-8') codepoint += chr(int(chunk[1:].to01(), 2)) return string + codepoint.decode('utf-8')
from bitarray import bitarray as _bitarray def utf9encode(string): + """Takes a string and returns a utf9-encoded version.""" bits = _bitarray() for char in string: for idx, byte in enumerate(char.encode('utf-8')): bits.append(idx) bits.extend('{0:b}'.format(ord(byte)).zfill(8)) return bits.tobytes() def utf9decode(data): + """Takes utf9-encoded data and returns the corresponding string.""" bits = _bitarray() bits.frombytes(data) chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9)) string = u'' codepoint = '' for chunk in chunks: if len(chunk) < 9: break if chunk[0] == 0: codepoint, string = '', string + codepoint.decode('utf-8') codepoint += chr(int(chunk[1:].to01(), 2)) return string + codepoint.decode('utf-8')
63f7489066aeb23dbefc6f8de534ad05144431ad
boardinghouse/tests/test_sql.py
boardinghouse/tests/test_sql.py
from django.conf import settings from django.test import TestCase from django.db.models import connection from boardinghouse.models import Schema class TestRejectSchemaColumnChange(TestCase): def test_exception_is_raised(self): Schema.objects.mass_create('a') cursor = connection.cursor() UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'" self.assertRaises(Exception, cursor.execute, UPDATE)
from django.conf import settings from django.test import TestCase from django.db import connection from boardinghouse.models import Schema class TestRejectSchemaColumnChange(TestCase): def test_exception_is_raised(self): Schema.objects.mass_create('a') cursor = connection.cursor() UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'" self.assertRaises(Exception, cursor.execute, UPDATE)
Make test work with 1.7
Make test work with 1.7
Python
bsd-3-clause
luzfcb/django-boardinghouse,luzfcb/django-boardinghouse,luzfcb/django-boardinghouse
from django.conf import settings from django.test import TestCase - from django.db.models import connection + from django.db import connection from boardinghouse.models import Schema class TestRejectSchemaColumnChange(TestCase): def test_exception_is_raised(self): Schema.objects.mass_create('a') cursor = connection.cursor() UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'" - self.assertRaises(Exception, cursor.execute, UPDATE) + self.assertRaises(Exception, cursor.execute, UPDATE)
Make test work with 1.7
## Code Before: from django.conf import settings from django.test import TestCase from django.db.models import connection from boardinghouse.models import Schema class TestRejectSchemaColumnChange(TestCase): def test_exception_is_raised(self): Schema.objects.mass_create('a') cursor = connection.cursor() UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'" self.assertRaises(Exception, cursor.execute, UPDATE) ## Instruction: Make test work with 1.7 ## Code After: from django.conf import settings from django.test import TestCase from django.db import connection from boardinghouse.models import Schema class TestRejectSchemaColumnChange(TestCase): def test_exception_is_raised(self): Schema.objects.mass_create('a') cursor = connection.cursor() UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'" self.assertRaises(Exception, cursor.execute, UPDATE)
from django.conf import settings from django.test import TestCase - from django.db.models import connection ? ------- + from django.db import connection from boardinghouse.models import Schema class TestRejectSchemaColumnChange(TestCase): def test_exception_is_raised(self): Schema.objects.mass_create('a') cursor = connection.cursor() UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'" - self.assertRaises(Exception, cursor.execute, UPDATE) ? -------- + self.assertRaises(Exception, cursor.execute, UPDATE)
c0d8b7f13a74fd4da7b36d30a61224b76367acbe
scraper.py
scraper.py
import urllib, datetime, os def fetch(): url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx' output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html') output = os.path.join(os.path.dirname(__file__), output) content = urllib.urlretrieve(url, output) if __name__ == '__main__': fetch()
import urllib, datetime, os def fetch(): # Instead of doing all the parsing later, I get the status from Mosab's site & store it url = 'http://power-grid-status.mos3abof.com/status' output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json') output = os.path.join(os.path.dirname(__file__), output) content = urllib.urlretrieve(url, output) if __name__ == '__main__': fetch()
Use Mosab's site to get the status
Use Mosab's site to get the status
Python
apache-2.0
mtayseer/power-grid-scraper
import urllib, datetime, os def fetch(): - url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx' + # Instead of doing all the parsing later, I get the status from Mosab's site & store it + url = 'http://power-grid-status.mos3abof.com/status' - output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html') + output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json') output = os.path.join(os.path.dirname(__file__), output) content = urllib.urlretrieve(url, output) if __name__ == '__main__': fetch()
Use Mosab's site to get the status
## Code Before: import urllib, datetime, os def fetch(): url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx' output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html') output = os.path.join(os.path.dirname(__file__), output) content = urllib.urlretrieve(url, output) if __name__ == '__main__': fetch() ## Instruction: Use Mosab's site to get the status ## Code After: import urllib, datetime, os def fetch(): # Instead of doing all the parsing later, I get the status from Mosab's site & store it url = 'http://power-grid-status.mos3abof.com/status' output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json') output = os.path.join(os.path.dirname(__file__), output) content = urllib.urlretrieve(url, output) if __name__ == '__main__': fetch()
import urllib, datetime, os def fetch(): - url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx' + # Instead of doing all the parsing later, I get the status from Mosab's site & store it + url = 'http://power-grid-status.mos3abof.com/status' - output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html') ? ^^^^ + output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json') ? ^^^^ output = os.path.join(os.path.dirname(__file__), output) content = urllib.urlretrieve(url, output) if __name__ == '__main__': fetch()
82eb7a69ccb88d27141aeb483e4482041108723f
app/Display/display.py
app/Display/display.py
import sys ESC = chr(27) CLEAR = ESC + "[2J" MOVE_HOME = ESC + "[H" ERASE = CLEAR + MOVE_HOME LINES = 24 COLS = 80 class Display: def __init__(self, title): self.title = title def clear(self): sys.stdout.write(ERASE) def show_properties(self, properties, names=None): if names is None: names = properties.keys() max_len = max(map(len, names)) self.clear() self.print(self.title) print() for k in names: self.print("{0}: {1}".format(k.rjust(max_len), properties[k])) def print(self, message): print(message, end="\x0a\x0d")
import sys ESC = chr(27) CSI = ESC + "[" CLEAR = CSI + "2J" MOVE_HOME = CSI + "H" ERASE = CLEAR + MOVE_HOME MOVE_TO = CSI + "{0};{1}H" LINES = 24 COLS = 80 class Display: def __init__(self, title, info=None): self.title = title self.info = info def clear(self): sys.stdout.write(ERASE) sys.stdout.flush() def move_to(self, row, col): sys.stdout.write(MOVE_TO.format(row, col)) sys.stdout.flush() def show_properties(self, properties, names=None): if names is None: names = properties.keys() max_len = max(map(len, names)) self.clear() self.print(self.title.center(COLS)) print() for k in names: self.print("{0}: {1}".format(k.rjust(max_len), properties[k])) if self.info is not None: self.move_to(LINES, 0) sys.stdout.write(self.info) sys.stdout.flush() self.move_to(LINES, 0) def print(self, message): print(message, end="\x0a\x0d")
Add support for cursor position, centered title, and an info bar
Add support for cursor position, centered title, and an info bar
Python
mit
gizmo-cda/g2x,gizmo-cda/g2x,thelonious/g2x,gizmo-cda/g2x,gizmo-cda/g2x,thelonious/g2x
import sys ESC = chr(27) + CSI = ESC + "[" - CLEAR = ESC + "[2J" + CLEAR = CSI + "2J" - MOVE_HOME = ESC + "[H" + MOVE_HOME = CSI + "H" ERASE = CLEAR + MOVE_HOME + MOVE_TO = CSI + "{0};{1}H" LINES = 24 COLS = 80 class Display: - def __init__(self, title): + def __init__(self, title, info=None): self.title = title + self.info = info def clear(self): sys.stdout.write(ERASE) + sys.stdout.flush() + + def move_to(self, row, col): + sys.stdout.write(MOVE_TO.format(row, col)) + sys.stdout.flush() def show_properties(self, properties, names=None): if names is None: names = properties.keys() max_len = max(map(len, names)) self.clear() - self.print(self.title) + self.print(self.title.center(COLS)) print() for k in names: self.print("{0}: {1}".format(k.rjust(max_len), properties[k])) + if self.info is not None: + self.move_to(LINES, 0) + sys.stdout.write(self.info) + sys.stdout.flush() + + self.move_to(LINES, 0) + def print(self, message): print(message, end="\x0a\x0d")
Add support for cursor position, centered title, and an info bar
## Code Before: import sys ESC = chr(27) CLEAR = ESC + "[2J" MOVE_HOME = ESC + "[H" ERASE = CLEAR + MOVE_HOME LINES = 24 COLS = 80 class Display: def __init__(self, title): self.title = title def clear(self): sys.stdout.write(ERASE) def show_properties(self, properties, names=None): if names is None: names = properties.keys() max_len = max(map(len, names)) self.clear() self.print(self.title) print() for k in names: self.print("{0}: {1}".format(k.rjust(max_len), properties[k])) def print(self, message): print(message, end="\x0a\x0d") ## Instruction: Add support for cursor position, centered title, and an info bar ## Code After: import sys ESC = chr(27) CSI = ESC + "[" CLEAR = CSI + "2J" MOVE_HOME = CSI + "H" ERASE = CLEAR + MOVE_HOME MOVE_TO = CSI + "{0};{1}H" LINES = 24 COLS = 80 class Display: def __init__(self, title, info=None): self.title = title self.info = info def clear(self): sys.stdout.write(ERASE) sys.stdout.flush() def move_to(self, row, col): sys.stdout.write(MOVE_TO.format(row, col)) sys.stdout.flush() def show_properties(self, properties, names=None): if names is None: names = properties.keys() max_len = max(map(len, names)) self.clear() self.print(self.title.center(COLS)) print() for k in names: self.print("{0}: {1}".format(k.rjust(max_len), properties[k])) if self.info is not None: self.move_to(LINES, 0) sys.stdout.write(self.info) sys.stdout.flush() self.move_to(LINES, 0) def print(self, message): print(message, end="\x0a\x0d")
import sys ESC = chr(27) + CSI = ESC + "[" - CLEAR = ESC + "[2J" ? ^ ^ - + CLEAR = CSI + "2J" ? ^ ^ - MOVE_HOME = ESC + "[H" ? ^ ^ - + MOVE_HOME = CSI + "H" ? ^ ^ ERASE = CLEAR + MOVE_HOME + MOVE_TO = CSI + "{0};{1}H" LINES = 24 COLS = 80 class Display: - def __init__(self, title): + def __init__(self, title, info=None): ? +++++++++++ self.title = title + self.info = info def clear(self): sys.stdout.write(ERASE) + sys.stdout.flush() + + def move_to(self, row, col): + sys.stdout.write(MOVE_TO.format(row, col)) + sys.stdout.flush() def show_properties(self, properties, names=None): if names is None: names = properties.keys() max_len = max(map(len, names)) self.clear() - self.print(self.title) + self.print(self.title.center(COLS)) ? ++++++++++++ + print() for k in names: self.print("{0}: {1}".format(k.rjust(max_len), properties[k])) + if self.info is not None: + self.move_to(LINES, 0) + sys.stdout.write(self.info) + sys.stdout.flush() + + self.move_to(LINES, 0) + def print(self, message): print(message, end="\x0a\x0d")
135ac2fa3aa978328ba72db6ca84920f3da0f39a
furikura/desktop/unity.py
furikura/desktop/unity.py
import gi import time gi.require_version('Unity', '7.0') from gi.repository import Unity, GObject def update_counter(count): launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop") launcher.set_property("count", count) launcher.set_property("count_visible", True)
import gi gi.require_version('Unity', '7.0') from gi.repository import Unity def update_counter(count): launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop") launcher.set_property("count", count) launcher.set_property("count_visible", True)
Remove unnecessary imports for Unity module
Remove unnecessary imports for Unity module
Python
mit
benjamindean/furi-kura,benjamindean/furi-kura
import gi - import time gi.require_version('Unity', '7.0') - from gi.repository import Unity, GObject + from gi.repository import Unity def update_counter(count): launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop") launcher.set_property("count", count) launcher.set_property("count_visible", True) - - -
Remove unnecessary imports for Unity module
## Code Before: import gi import time gi.require_version('Unity', '7.0') from gi.repository import Unity, GObject def update_counter(count): launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop") launcher.set_property("count", count) launcher.set_property("count_visible", True) ## Instruction: Remove unnecessary imports for Unity module ## Code After: import gi gi.require_version('Unity', '7.0') from gi.repository import Unity def update_counter(count): launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop") launcher.set_property("count", count) launcher.set_property("count_visible", True)
import gi - import time gi.require_version('Unity', '7.0') - from gi.repository import Unity, GObject ? --------- + from gi.repository import Unity def update_counter(count): launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop") launcher.set_property("count", count) launcher.set_property("count_visible", True) - - -
60f87cb4c3523faf5c5cdbc5f16453cae755988b
angr/procedures/java_jni/GetArrayElements.py
angr/procedures/java_jni/GetArrayElements.py
from . import JNISimProcedure from ...engines.soot.values.arrayref import SimSootValue_ArrayRef class GetArrayElements(JNISimProcedure): return_ty = 'reference' def run(self, ptr_env, array, ptr_isCopy): array_ref = self.state.jni_references.lookup(array) values = self.load_java_array(self.state, array_ref) memory_addr = self.store_in_native_memory(values, array_ref.type) return memory_addr def load_java_array(self, array_ref, start_idx=None, end_idx=None): if start_idx is None: start_idx = 0 if end_idx is None: end_idx = self.state.solver.max(array_ref.size) javavm_memory = self.state.get_javavm_view_of_plugin("memory") values = [] for idx in range(start_idx, end_idx): idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx) value = javavm_memory.load(idx_array_ref) values.append(value) return values
from . import JNISimProcedure from ...engines.soot.values.arrayref import SimSootValue_ArrayRef class GetArrayElements(JNISimProcedure): return_ty = 'reference' def run(self, ptr_env, array, ptr_isCopy): array_ref = self.state.jni_references.lookup(array) values = self.load_java_array(self.state, array_ref) memory_addr = self.store_in_native_memory(values, array_ref.type) if self.state.solver.eval(ptr_isCopy != 0): self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy) return memory_addr def load_java_array(self, array_ref, start_idx=None, end_idx=None): if start_idx is None: start_idx = 0 if end_idx is None: end_idx = self.state.solver.max(array_ref.size) javavm_memory = self.state.get_javavm_view_of_plugin("memory") values = [] for idx in range(start_idx, end_idx): idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx) value = javavm_memory.load(idx_array_ref) values.append(value) return values
Fix case if isCopy is null
Fix case if isCopy is null
Python
bsd-2-clause
schieb/angr,schieb/angr,angr/angr,angr/angr,iamahuman/angr,angr/angr,iamahuman/angr,schieb/angr,iamahuman/angr
from . import JNISimProcedure from ...engines.soot.values.arrayref import SimSootValue_ArrayRef class GetArrayElements(JNISimProcedure): return_ty = 'reference' def run(self, ptr_env, array, ptr_isCopy): array_ref = self.state.jni_references.lookup(array) values = self.load_java_array(self.state, array_ref) memory_addr = self.store_in_native_memory(values, array_ref.type) + if self.state.solver.eval(ptr_isCopy != 0): + self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy) return memory_addr def load_java_array(self, array_ref, start_idx=None, end_idx=None): if start_idx is None: start_idx = 0 if end_idx is None: end_idx = self.state.solver.max(array_ref.size) javavm_memory = self.state.get_javavm_view_of_plugin("memory") values = [] for idx in range(start_idx, end_idx): idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx) value = javavm_memory.load(idx_array_ref) values.append(value) return values
Fix case if isCopy is null
## Code Before: from . import JNISimProcedure from ...engines.soot.values.arrayref import SimSootValue_ArrayRef class GetArrayElements(JNISimProcedure): return_ty = 'reference' def run(self, ptr_env, array, ptr_isCopy): array_ref = self.state.jni_references.lookup(array) values = self.load_java_array(self.state, array_ref) memory_addr = self.store_in_native_memory(values, array_ref.type) return memory_addr def load_java_array(self, array_ref, start_idx=None, end_idx=None): if start_idx is None: start_idx = 0 if end_idx is None: end_idx = self.state.solver.max(array_ref.size) javavm_memory = self.state.get_javavm_view_of_plugin("memory") values = [] for idx in range(start_idx, end_idx): idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx) value = javavm_memory.load(idx_array_ref) values.append(value) return values ## Instruction: Fix case if isCopy is null ## Code After: from . import JNISimProcedure from ...engines.soot.values.arrayref import SimSootValue_ArrayRef class GetArrayElements(JNISimProcedure): return_ty = 'reference' def run(self, ptr_env, array, ptr_isCopy): array_ref = self.state.jni_references.lookup(array) values = self.load_java_array(self.state, array_ref) memory_addr = self.store_in_native_memory(values, array_ref.type) if self.state.solver.eval(ptr_isCopy != 0): self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy) return memory_addr def load_java_array(self, array_ref, start_idx=None, end_idx=None): if start_idx is None: start_idx = 0 if end_idx is None: end_idx = self.state.solver.max(array_ref.size) javavm_memory = self.state.get_javavm_view_of_plugin("memory") values = [] for idx in range(start_idx, end_idx): idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx) value = javavm_memory.load(idx_array_ref) values.append(value) return values
from . import JNISimProcedure from ...engines.soot.values.arrayref import SimSootValue_ArrayRef class GetArrayElements(JNISimProcedure): return_ty = 'reference' def run(self, ptr_env, array, ptr_isCopy): array_ref = self.state.jni_references.lookup(array) values = self.load_java_array(self.state, array_ref) memory_addr = self.store_in_native_memory(values, array_ref.type) + if self.state.solver.eval(ptr_isCopy != 0): + self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy) return memory_addr def load_java_array(self, array_ref, start_idx=None, end_idx=None): if start_idx is None: start_idx = 0 if end_idx is None: end_idx = self.state.solver.max(array_ref.size) javavm_memory = self.state.get_javavm_view_of_plugin("memory") values = [] for idx in range(start_idx, end_idx): idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx) value = javavm_memory.load(idx_array_ref) values.append(value) return values
9a94e9e61a7bb1680265692eb7cdf926842aa766
streamline/__init__.py
streamline/__init__.py
from .base import RouteBase, NonIterableRouteBase from .template import TemplateRoute, XHRPartialRoute, ROCARoute from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute __version__ = '1.0' __author__ = 'Outernet Inc' __all__ = ( RouteBase, NonIterableRouteBase, TemplateRoute, XHRPartialRoute, ROCARoute, FormRoute, TemplateFormRoute, XHRPartialFormRoute, )
from .base import RouteBase, NonIterableRouteBase from .template import TemplateRoute, XHRPartialRoute, ROCARoute from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute __version__ = '1.0' __author__ = 'Outernet Inc' __all__ = ( 'RouteBase', 'NonIterableRouteBase', 'TemplateRoute', 'XHRPartialRoute', 'ROCARoute', 'FormRoute', 'TemplateFormRoute', 'XHRPartialFormRoute', )
Fix __all__ using objects instead of strings
Fix __all__ using objects instead of strings Signed-off-by: Branko Vukelic <26059cc39872530f89fec69552bb1050e1cc2caa@outernet.is>
Python
bsd-2-clause
Outernet-Project/bottle-streamline
from .base import RouteBase, NonIterableRouteBase from .template import TemplateRoute, XHRPartialRoute, ROCARoute from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute __version__ = '1.0' __author__ = 'Outernet Inc' __all__ = ( - RouteBase, + 'RouteBase', - NonIterableRouteBase, + 'NonIterableRouteBase', - TemplateRoute, + 'TemplateRoute', - XHRPartialRoute, + 'XHRPartialRoute', - ROCARoute, + 'ROCARoute', - FormRoute, + 'FormRoute', - TemplateFormRoute, + 'TemplateFormRoute', - XHRPartialFormRoute, + 'XHRPartialFormRoute', )
Fix __all__ using objects instead of strings
## Code Before: from .base import RouteBase, NonIterableRouteBase from .template import TemplateRoute, XHRPartialRoute, ROCARoute from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute __version__ = '1.0' __author__ = 'Outernet Inc' __all__ = ( RouteBase, NonIterableRouteBase, TemplateRoute, XHRPartialRoute, ROCARoute, FormRoute, TemplateFormRoute, XHRPartialFormRoute, ) ## Instruction: Fix __all__ using objects instead of strings ## Code After: from .base import RouteBase, NonIterableRouteBase from .template import TemplateRoute, XHRPartialRoute, ROCARoute from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute __version__ = '1.0' __author__ = 'Outernet Inc' __all__ = ( 'RouteBase', 'NonIterableRouteBase', 'TemplateRoute', 'XHRPartialRoute', 'ROCARoute', 'FormRoute', 'TemplateFormRoute', 'XHRPartialFormRoute', )
from .base import RouteBase, NonIterableRouteBase from .template import TemplateRoute, XHRPartialRoute, ROCARoute from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute __version__ = '1.0' __author__ = 'Outernet Inc' __all__ = ( - RouteBase, + 'RouteBase', ? + + - NonIterableRouteBase, + 'NonIterableRouteBase', ? + + - TemplateRoute, + 'TemplateRoute', ? + + - XHRPartialRoute, + 'XHRPartialRoute', ? + + - ROCARoute, + 'ROCARoute', ? + + - FormRoute, + 'FormRoute', ? + + - TemplateFormRoute, + 'TemplateFormRoute', ? + + - XHRPartialFormRoute, + 'XHRPartialFormRoute', ? + + )
e1c57cb41c59c118648602ff9837418e5d4baad4
saleor/dashboard/category/forms.py
saleor/dashboard/category/forms.py
from django import forms from ...product.models import Category class CategoryForm(forms.ModelForm): class Meta: model = Category exclude = []
from django import forms from django.utils.translation import ugettext_lazy as _ from ...product.models import Category class CategoryForm(forms.ModelForm): class Meta: model = Category exclude = [] def clean_parent(self): parent = self.cleaned_data['parent'] if parent == self.instance: raise forms.ValidationError(_('A category may not be made a child of itself')) return parent
Add validation on category parent field
Add validation on category parent field
Python
bsd-3-clause
itbabu/saleor,rchav/vinerack,avorio/saleor,HyperManTT/ECommerceSaleor,laosunhust/saleor,itbabu/saleor,josesanch/saleor,Drekscott/Motlaesaleor,maferelo/saleor,taedori81/saleor,rchav/vinerack,avorio/saleor,rodrigozn/CW-Shop,avorio/saleor,maferelo/saleor,arth-co/saleor,paweltin/saleor,jreigel/saleor,paweltin/saleor,Drekscott/Motlaesaleor,taedori81/saleor,UITools/saleor,taedori81/saleor,dashmug/saleor,itbabu/saleor,tfroehlich82/saleor,paweltin/saleor,UITools/saleor,jreigel/saleor,avorio/saleor,arth-co/saleor,laosunhust/saleor,dashmug/saleor,rodrigozn/CW-Shop,rchav/vinerack,KenMutemi/saleor,spartonia/saleor,spartonia/saleor,KenMutemi/saleor,laosunhust/saleor,jreigel/saleor,arth-co/saleor,tfroehlich82/saleor,arth-co/saleor,josesanch/saleor,car3oon/saleor,Drekscott/Motlaesaleor,josesanch/saleor,UITools/saleor,rodrigozn/CW-Shop,UITools/saleor,spartonia/saleor,HyperManTT/ECommerceSaleor,maferelo/saleor,paweltin/saleor,Drekscott/Motlaesaleor,HyperManTT/ECommerceSaleor,UITools/saleor,KenMutemi/saleor,spartonia/saleor,mociepka/saleor,taedori81/saleor,car3oon/saleor,mociepka/saleor,mociepka/saleor,laosunhust/saleor,car3oon/saleor,dashmug/saleor,tfroehlich82/saleor
from django import forms + from django.utils.translation import ugettext_lazy as _ from ...product.models import Category class CategoryForm(forms.ModelForm): class Meta: model = Category exclude = [] + + def clean_parent(self): + parent = self.cleaned_data['parent'] + if parent == self.instance: + raise forms.ValidationError(_('A category may not be made a child of itself')) + return parent +
Add validation on category parent field
## Code Before: from django import forms from ...product.models import Category class CategoryForm(forms.ModelForm): class Meta: model = Category exclude = [] ## Instruction: Add validation on category parent field ## Code After: from django import forms from django.utils.translation import ugettext_lazy as _ from ...product.models import Category class CategoryForm(forms.ModelForm): class Meta: model = Category exclude = [] def clean_parent(self): parent = self.cleaned_data['parent'] if parent == self.instance: raise forms.ValidationError(_('A category may not be made a child of itself')) return parent
from django import forms + from django.utils.translation import ugettext_lazy as _ from ...product.models import Category class CategoryForm(forms.ModelForm): class Meta: model = Category exclude = [] + + def clean_parent(self): + parent = self.cleaned_data['parent'] + if parent == self.instance: + raise forms.ValidationError(_('A category may not be made a child of itself')) + return parent
9df2bae691e8613794be3713194db2420fc75385
gapipy/resources/dossier/transport_dossier.py
gapipy/resources/dossier/transport_dossier.py
from __future__ import unicode_literals from ..base import Resource from .details import DossierDetail, DossierDetailsMixin from .dossier_features import DossierFeature class TransportDossier(Resource, DossierDetailsMixin): _resource_name = 'transport_dossiers' _as_is_fields = [ 'id', 'href', 'features', 'capacity', 'private', 'name', 'dossier_segment', ] _model_collection_fields = [ ('details', DossierDetail), ('features', DossierFeature), ] _date_time_fields_local = ['date_created', 'date_last_modified']
from __future__ import unicode_literals from ..base import Resource from .details import DossierDetail, DossierDetailsMixin from .dossier_features import DossierFeature class TransportDossier(Resource, DossierDetailsMixin): _resource_name = 'transport_dossiers' _as_is_fields = [ 'id', 'href', 'capacity', 'private', 'name', 'dossier_segment', ] _model_collection_fields = [ ('details', DossierDetail), ('features', DossierFeature), ] _date_time_fields_local = ['date_created', 'date_last_modified']
Remove features from as-is fields on TransportDossier
Remove features from as-is fields on TransportDossier Reflected as a model_collection_field
Python
mit
gadventures/gapipy
from __future__ import unicode_literals from ..base import Resource from .details import DossierDetail, DossierDetailsMixin from .dossier_features import DossierFeature class TransportDossier(Resource, DossierDetailsMixin): _resource_name = 'transport_dossiers' _as_is_fields = [ - 'id', 'href', 'features', 'capacity', 'private', 'name', + 'id', 'href', 'capacity', 'private', 'name', 'dossier_segment', ] _model_collection_fields = [ ('details', DossierDetail), ('features', DossierFeature), ] _date_time_fields_local = ['date_created', 'date_last_modified']
Remove features from as-is fields on TransportDossier
## Code Before: from __future__ import unicode_literals from ..base import Resource from .details import DossierDetail, DossierDetailsMixin from .dossier_features import DossierFeature class TransportDossier(Resource, DossierDetailsMixin): _resource_name = 'transport_dossiers' _as_is_fields = [ 'id', 'href', 'features', 'capacity', 'private', 'name', 'dossier_segment', ] _model_collection_fields = [ ('details', DossierDetail), ('features', DossierFeature), ] _date_time_fields_local = ['date_created', 'date_last_modified'] ## Instruction: Remove features from as-is fields on TransportDossier ## Code After: from __future__ import unicode_literals from ..base import Resource from .details import DossierDetail, DossierDetailsMixin from .dossier_features import DossierFeature class TransportDossier(Resource, DossierDetailsMixin): _resource_name = 'transport_dossiers' _as_is_fields = [ 'id', 'href', 'capacity', 'private', 'name', 'dossier_segment', ] _model_collection_fields = [ ('details', DossierDetail), ('features', DossierFeature), ] _date_time_fields_local = ['date_created', 'date_last_modified']
from __future__ import unicode_literals from ..base import Resource from .details import DossierDetail, DossierDetailsMixin from .dossier_features import DossierFeature class TransportDossier(Resource, DossierDetailsMixin): _resource_name = 'transport_dossiers' _as_is_fields = [ - 'id', 'href', 'features', 'capacity', 'private', 'name', ? ------------ + 'id', 'href', 'capacity', 'private', 'name', 'dossier_segment', ] _model_collection_fields = [ ('details', DossierDetail), ('features', DossierFeature), ] _date_time_fields_local = ['date_created', 'date_last_modified']
331b3987ba09db5d8f774509bedd30c3c6522795
ooni/tests/test_utils.py
ooni/tests/test_utils.py
import os import unittest from ooni.utils import pushFilenameStack class TestUtils(unittest.TestCase): def test_pushFilenameStack(self): basefilename = os.path.join(os.getcwd(), 'dummyfile') f = open(basefilename, "w+") f.write("0\n") f.close() for i in xrange(1, 5): f = open(basefilename+".%s" % i, "w+") f.write("%s\n" % i) f.close() pushFilenameStack(basefilename) for i in xrange(1, 5): f = open(basefilename+".%s" % i) c = f.readlines()[0].strip() self.assertEqual(str(i-1), str(c)) f.close()
import os from twisted.trial import unittest from ooni.utils import pushFilenameStack class TestUtils(unittest.TestCase): def test_pushFilenameStack(self): basefilename = os.path.join(os.getcwd(), 'dummyfile') f = open(basefilename, "w+") f.write("0\n") f.close() for i in xrange(1, 5): f = open(basefilename+".%s" % i, "w+") f.write("%s\n" % i) f.close() pushFilenameStack(basefilename) for i in xrange(1, 5): f = open(basefilename+".%s" % i) c = f.readlines()[0].strip() self.assertEqual(str(i-1), str(c)) f.close()
Use trial unittest instead of python unittest
Use trial unittest instead of python unittest
Python
bsd-2-clause
juga0/ooni-probe,juga0/ooni-probe,lordappsec/ooni-probe,Karthikeyan-kkk/ooni-probe,Karthikeyan-kkk/ooni-probe,0xPoly/ooni-probe,lordappsec/ooni-probe,Karthikeyan-kkk/ooni-probe,0xPoly/ooni-probe,0xPoly/ooni-probe,juga0/ooni-probe,lordappsec/ooni-probe,kdmurray91/ooni-probe,kdmurray91/ooni-probe,lordappsec/ooni-probe,kdmurray91/ooni-probe,Karthikeyan-kkk/ooni-probe,kdmurray91/ooni-probe,0xPoly/ooni-probe,juga0/ooni-probe
import os - import unittest + from twisted.trial import unittest + from ooni.utils import pushFilenameStack class TestUtils(unittest.TestCase): def test_pushFilenameStack(self): basefilename = os.path.join(os.getcwd(), 'dummyfile') f = open(basefilename, "w+") f.write("0\n") f.close() for i in xrange(1, 5): f = open(basefilename+".%s" % i, "w+") f.write("%s\n" % i) f.close() pushFilenameStack(basefilename) for i in xrange(1, 5): f = open(basefilename+".%s" % i) c = f.readlines()[0].strip() self.assertEqual(str(i-1), str(c)) f.close()
Use trial unittest instead of python unittest
## Code Before: import os import unittest from ooni.utils import pushFilenameStack class TestUtils(unittest.TestCase): def test_pushFilenameStack(self): basefilename = os.path.join(os.getcwd(), 'dummyfile') f = open(basefilename, "w+") f.write("0\n") f.close() for i in xrange(1, 5): f = open(basefilename+".%s" % i, "w+") f.write("%s\n" % i) f.close() pushFilenameStack(basefilename) for i in xrange(1, 5): f = open(basefilename+".%s" % i) c = f.readlines()[0].strip() self.assertEqual(str(i-1), str(c)) f.close() ## Instruction: Use trial unittest instead of python unittest ## Code After: import os from twisted.trial import unittest from ooni.utils import pushFilenameStack class TestUtils(unittest.TestCase): def test_pushFilenameStack(self): basefilename = os.path.join(os.getcwd(), 'dummyfile') f = open(basefilename, "w+") f.write("0\n") f.close() for i in xrange(1, 5): f = open(basefilename+".%s" % i, "w+") f.write("%s\n" % i) f.close() pushFilenameStack(basefilename) for i in xrange(1, 5): f = open(basefilename+".%s" % i) c = f.readlines()[0].strip() self.assertEqual(str(i-1), str(c)) f.close()
import os - import unittest + from twisted.trial import unittest + from ooni.utils import pushFilenameStack class TestUtils(unittest.TestCase): def test_pushFilenameStack(self): basefilename = os.path.join(os.getcwd(), 'dummyfile') f = open(basefilename, "w+") f.write("0\n") f.close() for i in xrange(1, 5): f = open(basefilename+".%s" % i, "w+") f.write("%s\n" % i) f.close() pushFilenameStack(basefilename) for i in xrange(1, 5): f = open(basefilename+".%s" % i) c = f.readlines()[0].strip() self.assertEqual(str(i-1), str(c)) f.close()
8e5ad2138d0685e4322156b3f545be46a3f0c99f
util.py
util.py
import glob import os.path import random def pick_random(directory, k=None): """Pick randomly some files from a directory.""" all_files = glob.glob(os.path.join(directory, '*')) random.shuffle(all_files) return all_files if k is None else all_files[:k]
import glob import os.path import random import re def pick(directory, k=None, randomized=True): """Pick some thread files from a thread directory.""" all_files = glob.glob(os.path.join(directory, '*')) if randomized: random.shuffle(all_files) else: pattern = '([0-9]+)\.txt' all_files.sort(key=lambda f: int(re.search(pattern, f).group(1))) return all_files if k is None else all_files[:k]
Modify to pick either randomly or sequentially
Modify to pick either randomly or sequentially
Python
mit
kemskems/otdet
import glob import os.path import random + import re - def pick_random(directory, k=None): + def pick(directory, k=None, randomized=True): - """Pick randomly some files from a directory.""" + """Pick some thread files from a thread directory.""" all_files = glob.glob(os.path.join(directory, '*')) + if randomized: - random.shuffle(all_files) + random.shuffle(all_files) + else: + pattern = '([0-9]+)\.txt' + all_files.sort(key=lambda f: int(re.search(pattern, f).group(1))) return all_files if k is None else all_files[:k]
Modify to pick either randomly or sequentially
## Code Before: import glob import os.path import random def pick_random(directory, k=None): """Pick randomly some files from a directory.""" all_files = glob.glob(os.path.join(directory, '*')) random.shuffle(all_files) return all_files if k is None else all_files[:k] ## Instruction: Modify to pick either randomly or sequentially ## Code After: import glob import os.path import random import re def pick(directory, k=None, randomized=True): """Pick some thread files from a thread directory.""" all_files = glob.glob(os.path.join(directory, '*')) if randomized: random.shuffle(all_files) else: pattern = '([0-9]+)\.txt' all_files.sort(key=lambda f: int(re.search(pattern, f).group(1))) return all_files if k is None else all_files[:k]
import glob import os.path import random + import re - def pick_random(directory, k=None): + def pick(directory, k=None, randomized=True): - """Pick randomly some files from a directory.""" ? --------- + """Pick some thread files from a thread directory.""" ? +++++++ +++++++ all_files = glob.glob(os.path.join(directory, '*')) + if randomized: - random.shuffle(all_files) + random.shuffle(all_files) ? ++++ + else: + pattern = '([0-9]+)\.txt' + all_files.sort(key=lambda f: int(re.search(pattern, f).group(1))) return all_files if k is None else all_files[:k]
b7377196cdd05d9d6d481f7b93308189c4524c52
sfm/api/filters.py
sfm/api/filters.py
from django_filters import FilterSet, CharFilter, IsoDateTimeFilter from ui.models import Warc, Seed, Harvest from django_filters import Filter from django_filters.fields import Lookup class ListFilter(Filter): def filter(self, qs, value): return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in")) class WarcFilter(FilterSet): # Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e seedset = CharFilter(name="harvest__seed_set__seedset_id") seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True) # TODO: This will need to be changed to use historical seeds once #54 is completed. harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte') harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte') class Meta: model = Warc fields = ['seedset']
from django_filters import FilterSet, CharFilter, IsoDateTimeFilter from ui.models import Warc, Seed, Harvest from django_filters import Filter from django_filters.fields import Lookup class ListFilter(Filter): def filter(self, qs, value): return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in")) class WarcFilter(FilterSet): # Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e seedset = CharFilter(name="harvest__historical_seed_set__seedset_id") seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True) harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte') harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte') class Meta: model = Warc fields = ['seedset']
Fix to take into account history in API queries.
Fix to take into account history in API queries.
Python
mit
gwu-libraries/sfm,gwu-libraries/sfm-ui,gwu-libraries/sfm,gwu-libraries/sfm,gwu-libraries/sfm-ui,gwu-libraries/sfm-ui,gwu-libraries/sfm-ui
from django_filters import FilterSet, CharFilter, IsoDateTimeFilter from ui.models import Warc, Seed, Harvest from django_filters import Filter from django_filters.fields import Lookup class ListFilter(Filter): def filter(self, qs, value): return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in")) class WarcFilter(FilterSet): # Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e - seedset = CharFilter(name="harvest__seed_set__seedset_id") + seedset = CharFilter(name="harvest__historical_seed_set__seedset_id") - seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True) + seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True) - # TODO: This will need to be changed to use historical seeds once #54 is completed. harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte') harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte') class Meta: model = Warc fields = ['seedset']
Fix to take into account history in API queries.
## Code Before: from django_filters import FilterSet, CharFilter, IsoDateTimeFilter from ui.models import Warc, Seed, Harvest from django_filters import Filter from django_filters.fields import Lookup class ListFilter(Filter): def filter(self, qs, value): return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in")) class WarcFilter(FilterSet): # Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e seedset = CharFilter(name="harvest__seed_set__seedset_id") seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True) # TODO: This will need to be changed to use historical seeds once #54 is completed. harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte') harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte') class Meta: model = Warc fields = ['seedset'] ## Instruction: Fix to take into account history in API queries. ## Code After: from django_filters import FilterSet, CharFilter, IsoDateTimeFilter from ui.models import Warc, Seed, Harvest from django_filters import Filter from django_filters.fields import Lookup class ListFilter(Filter): def filter(self, qs, value): return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in")) class WarcFilter(FilterSet): # Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e seedset = CharFilter(name="harvest__historical_seed_set__seedset_id") seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True) harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte') harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte') class Meta: model = Warc fields = ['seedset']
from django_filters import FilterSet, CharFilter, IsoDateTimeFilter from ui.models import Warc, Seed, Harvest from django_filters import Filter from django_filters.fields import Lookup class ListFilter(Filter): def filter(self, qs, value): return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in")) class WarcFilter(FilterSet): # Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e - seedset = CharFilter(name="harvest__seed_set__seedset_id") + seedset = CharFilter(name="harvest__historical_seed_set__seedset_id") ? +++++++++++ - seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True) ? ------ ^ + seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True) ? ++ ^^^^^^ - # TODO: This will need to be changed to use historical seeds once #54 is completed. harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte') harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte') class Meta: model = Warc fields = ['seedset']
a137e8a92211d3d344a38b5c97d81073d66a1668
alembic/versions/17c1af634026_extract_publication_date.py
alembic/versions/17c1af634026_extract_publication_date.py
# revision identifiers, used by Alembic. revision = '17c1af634026' down_revision = '3c4c29f0a791' import html5lib from dateutil.parser import parse as parse_date import pytips from pytips.models import Tip def _extract_publication_date(html): root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False) publication_date_string = root.xpath("//a/@data-datetime")[0] return parse_date(publication_date_string) def _update_tip(tip): tip.publication_date = _extract_publication_date(tip.rendered_html) def _erase_publication_date(tip): tip.publication_date = None def upgrade(): tips = Tip.query.all() map(_update_tip, tips) pytips.db.session.commit() def downgrade(): tips = Tip.query.all() map(_erase_publication_date, tips) pytips.db.session.commit()
# revision identifiers, used by Alembic. revision = '17c1af634026' down_revision = '3c4c29f0a791' import html5lib from dateutil.parser import parse as parse_date import pytips from pytips.util import extract_publication_date from pytips.models import Tip def _update_tip(tip): tip.publication_date = extract_publication_date(tip.rendered_html) def _erase_publication_date(tip): tip.publication_date = None def upgrade(): tips = Tip.query.all() map(_update_tip, tips) pytips.db.session.commit() def downgrade(): tips = Tip.query.all() map(_erase_publication_date, tips) pytips.db.session.commit()
Use the utility module's extract_publication_date logic.
Use the utility module's extract_publication_date logic.
Python
isc
gthank/pytips,gthank/pytips,gthank/pytips,gthank/pytips
# revision identifiers, used by Alembic. revision = '17c1af634026' down_revision = '3c4c29f0a791' import html5lib from dateutil.parser import parse as parse_date import pytips + from pytips.util import extract_publication_date from pytips.models import Tip - def _extract_publication_date(html): - root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False) - publication_date_string = root.xpath("//a/@data-datetime")[0] - return parse_date(publication_date_string) - - def _update_tip(tip): - tip.publication_date = _extract_publication_date(tip.rendered_html) + tip.publication_date = extract_publication_date(tip.rendered_html) def _erase_publication_date(tip): tip.publication_date = None def upgrade(): tips = Tip.query.all() map(_update_tip, tips) pytips.db.session.commit() def downgrade(): tips = Tip.query.all() map(_erase_publication_date, tips) pytips.db.session.commit()
Use the utility module's extract_publication_date logic.
## Code Before: # revision identifiers, used by Alembic. revision = '17c1af634026' down_revision = '3c4c29f0a791' import html5lib from dateutil.parser import parse as parse_date import pytips from pytips.models import Tip def _extract_publication_date(html): root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False) publication_date_string = root.xpath("//a/@data-datetime")[0] return parse_date(publication_date_string) def _update_tip(tip): tip.publication_date = _extract_publication_date(tip.rendered_html) def _erase_publication_date(tip): tip.publication_date = None def upgrade(): tips = Tip.query.all() map(_update_tip, tips) pytips.db.session.commit() def downgrade(): tips = Tip.query.all() map(_erase_publication_date, tips) pytips.db.session.commit() ## Instruction: Use the utility module's extract_publication_date logic. ## Code After: # revision identifiers, used by Alembic. revision = '17c1af634026' down_revision = '3c4c29f0a791' import html5lib from dateutil.parser import parse as parse_date import pytips from pytips.util import extract_publication_date from pytips.models import Tip def _update_tip(tip): tip.publication_date = extract_publication_date(tip.rendered_html) def _erase_publication_date(tip): tip.publication_date = None def upgrade(): tips = Tip.query.all() map(_update_tip, tips) pytips.db.session.commit() def downgrade(): tips = Tip.query.all() map(_erase_publication_date, tips) pytips.db.session.commit()
# revision identifiers, used by Alembic. revision = '17c1af634026' down_revision = '3c4c29f0a791' import html5lib from dateutil.parser import parse as parse_date import pytips + from pytips.util import extract_publication_date from pytips.models import Tip - def _extract_publication_date(html): - root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False) - publication_date_string = root.xpath("//a/@data-datetime")[0] - return parse_date(publication_date_string) - - def _update_tip(tip): - tip.publication_date = _extract_publication_date(tip.rendered_html) ? - + tip.publication_date = extract_publication_date(tip.rendered_html) def _erase_publication_date(tip): tip.publication_date = None def upgrade(): tips = Tip.query.all() map(_update_tip, tips) pytips.db.session.commit() def downgrade(): tips = Tip.query.all() map(_erase_publication_date, tips) pytips.db.session.commit()
9d3d2beab6ec06ce13126b818029258a66f450f6
babelfish/__init__.py
babelfish/__init__.py
__title__ = 'babelfish' __version__ = '0.4.1' __author__ = 'Antoine Bertin' __license__ = 'BSD' __copyright__ = 'Copyright 2013 the BabelFish authors' from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter, CountryReverseConverter) from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language from .script import SCRIPTS, Script
__title__ = 'babelfish' __version__ = '0.4.1' __author__ = 'Antoine Bertin' __license__ = 'BSD' __copyright__ = 'Copyright 2013 the BabelFish authors' from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter, CountryReverseConverter) from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language from .script import SCRIPTS, SCRIPT_MATRIX, Script
Add SCRIPT_MATRIX to babelfish module imports
Add SCRIPT_MATRIX to babelfish module imports
Python
bsd-3-clause
Diaoul/babelfish
__title__ = 'babelfish' __version__ = '0.4.1' __author__ = 'Antoine Bertin' __license__ = 'BSD' __copyright__ = 'Copyright 2013 the BabelFish authors' from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter, CountryReverseConverter) from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language - from .script import SCRIPTS, Script + from .script import SCRIPTS, SCRIPT_MATRIX, Script
Add SCRIPT_MATRIX to babelfish module imports
## Code Before: __title__ = 'babelfish' __version__ = '0.4.1' __author__ = 'Antoine Bertin' __license__ = 'BSD' __copyright__ = 'Copyright 2013 the BabelFish authors' from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter, CountryReverseConverter) from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language from .script import SCRIPTS, Script ## Instruction: Add SCRIPT_MATRIX to babelfish module imports ## Code After: __title__ = 'babelfish' __version__ = '0.4.1' __author__ = 'Antoine Bertin' __license__ = 'BSD' __copyright__ = 'Copyright 2013 the BabelFish authors' from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter, CountryReverseConverter) from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language from .script import SCRIPTS, SCRIPT_MATRIX, Script
__title__ = 'babelfish' __version__ = '0.4.1' __author__ = 'Antoine Bertin' __license__ = 'BSD' __copyright__ = 'Copyright 2013 the BabelFish authors' from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter, CountryReverseConverter) from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language - from .script import SCRIPTS, Script + from .script import SCRIPTS, SCRIPT_MATRIX, Script ? +++++++++++++++
b2268ae4ecad477c46a4b99ec17511e2e535b9d0
globus_cli/commands/task/generate_submission_id.py
globus_cli/commands/task/generate_submission_id.py
import click from globus_cli.parsing import common_options from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print from globus_cli.services.transfer import get_client @click.command( "generate-submission-id", short_help="Get a submission ID", help=( "Generate a new task submission ID for use in " "`globus transfer` and `gloubs delete`. Submission IDs " "allow you to safely retry submission of a task in the " "presence of network errors. No matter how many times " "you submit a task with a given ID, it will only be " "accepted and executed once. The response status may " "change between submissions." ), ) @common_options def generate_submission_id(): """ Executor for `globus task generate-submission-id` """ client = get_client() res = client.get_submission_id() formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
import click from globus_cli.parsing import common_options from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print from globus_cli.services.transfer import get_client @click.command( "generate-submission-id", short_help="Get a submission ID", help=( """\ Generate a new task submission ID for use in `globus transfer` and `gloubs delete`. Submission IDs allow you to safely retry submission of a task in the presence of network errors. No matter how many times you submit a task with a given ID, it will only be accepted and executed once. The response status may change between submissions. \b Important Note: Submission IDs are not the same as Task IDs. """ ), ) @common_options def generate_submission_id(): """ Executor for `globus task generate-submission-id` """ client = get_client() res = client.get_submission_id() formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
Clarify that submission ID != task ID
Clarify that submission ID != task ID Calling this out in the helptext will hopefully help avoid people conflating these two quite as easily. (An imperfect solution for an imperfect world.)
Python
apache-2.0
globus/globus-cli,globus/globus-cli
import click from globus_cli.parsing import common_options from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print from globus_cli.services.transfer import get_client @click.command( "generate-submission-id", short_help="Get a submission ID", help=( - "Generate a new task submission ID for use in " - "`globus transfer` and `gloubs delete`. Submission IDs " + """\ + Generate a new task submission ID for use in `globus transfer` and `gloubs delete`. - "allow you to safely retry submission of a task in the " + Submission IDs allow you to safely retry submission of a task in the presence of + network errors. No matter how many times you submit a task with a given ID, it will - "presence of network errors. No matter how many times " - "you submit a task with a given ID, it will only be " - "accepted and executed once. The response status may " + only be accepted and executed once. The response status may change between - "change between submissions." + submissions. + + \b + Important Note: Submission IDs are not the same as Task IDs. + """ ), ) @common_options def generate_submission_id(): """ Executor for `globus task generate-submission-id` """ client = get_client() res = client.get_submission_id() formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
Clarify that submission ID != task ID
## Code Before: import click from globus_cli.parsing import common_options from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print from globus_cli.services.transfer import get_client @click.command( "generate-submission-id", short_help="Get a submission ID", help=( "Generate a new task submission ID for use in " "`globus transfer` and `gloubs delete`. Submission IDs " "allow you to safely retry submission of a task in the " "presence of network errors. No matter how many times " "you submit a task with a given ID, it will only be " "accepted and executed once. The response status may " "change between submissions." ), ) @common_options def generate_submission_id(): """ Executor for `globus task generate-submission-id` """ client = get_client() res = client.get_submission_id() formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value") ## Instruction: Clarify that submission ID != task ID ## Code After: import click from globus_cli.parsing import common_options from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print from globus_cli.services.transfer import get_client @click.command( "generate-submission-id", short_help="Get a submission ID", help=( """\ Generate a new task submission ID for use in `globus transfer` and `gloubs delete`. Submission IDs allow you to safely retry submission of a task in the presence of network errors. No matter how many times you submit a task with a given ID, it will only be accepted and executed once. The response status may change between submissions. \b Important Note: Submission IDs are not the same as Task IDs. """ ), ) @common_options def generate_submission_id(): """ Executor for `globus task generate-submission-id` """ client = get_client() res = client.get_submission_id() formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
import click from globus_cli.parsing import common_options from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print from globus_cli.services.transfer import get_client @click.command( "generate-submission-id", short_help="Get a submission ID", help=( - "Generate a new task submission ID for use in " - "`globus transfer` and `gloubs delete`. Submission IDs " + """\ + Generate a new task submission ID for use in `globus transfer` and `gloubs delete`. - "allow you to safely retry submission of a task in the " ? --- ^ + Submission IDs allow you to safely retry submission of a task in the presence of ? ++++++++++ +++ ^^^^^^^^^^^ + network errors. No matter how many times you submit a task with a given ID, it will - "presence of network errors. No matter how many times " - "you submit a task with a given ID, it will only be " - "accepted and executed once. The response status may " ? --- ^ + only be accepted and executed once. The response status may change between ? ++++ ++ ^^^^^^^^^^^^^^ - "change between submissions." + submissions. + + \b + Important Note: Submission IDs are not the same as Task IDs. + """ ), ) @common_options def generate_submission_id(): """ Executor for `globus task generate-submission-id` """ client = get_client() res = client.get_submission_id() formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
66035a6e3e7729c53278193d4307751b36ace6eb
fullcalendar/admin.py
fullcalendar/admin.py
from django.utils.translation import ugettext_lazy as _ from django.contrib import admin from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin from fullcalendar.models import * class EventCategoryAdmin(admin.ModelAdmin): list_display = ('name',) class OccurrenceInline(TabularDynamicInlineAdmin): model = Occurrence extra = 1 fields = ('start_time', 'end_time', 'description') class EventAdmin(DisplayableAdmin): list_display = ('title', 'event_category') list_filter = ('event_category',) search_fields = ('title', 'description', 'content', 'keywords') fieldsets = ( (None, { "fields": [ "title", "status", ("publish_date", "expiry_date"), "event_category", "content" ] }), (_("Meta data"), { "fields": [ "_meta_title", "slug", ("description", "gen_description"), "keywords", "in_sitemap" ], "classes": ("collapse-closed",) }), ) inlines = [OccurrenceInline] admin.site.register(Event, EventAdmin) admin.site.register(EventCategory, EventCategoryAdmin)
from django.utils.translation import ugettext_lazy as _ from django.contrib import admin from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin from fullcalendar.models import * class EventCategoryAdmin(admin.ModelAdmin): list_display = ('name',) class OccurrenceInline(StackedDynamicInlineAdmin): model = Occurrence extra = 1 fields = ('start_time', 'end_time', 'description', 'location') class EventAdmin(DisplayableAdmin): list_display = ('title', 'event_category') list_filter = ('event_category',) search_fields = ('title', 'description', 'content', 'keywords') fieldsets = ( (None, { "fields": [ "title", "status", ("publish_date", "expiry_date"), "event_category", "content" ] }), (_("Meta data"), { "fields": [ "_meta_title", "slug", ("description", "gen_description"), "keywords", "in_sitemap" ], "classes": ("collapse-closed",) }), ) inlines = [OccurrenceInline] admin.site.register(Event, EventAdmin) admin.site.register(EventCategory, EventCategoryAdmin)
Change to stacked inline for occurrences, also display location.
Change to stacked inline for occurrences, also display location.
Python
mit
jonge-democraten/mezzanine-fullcalendar
from django.utils.translation import ugettext_lazy as _ from django.contrib import admin - from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin + from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin from fullcalendar.models import * class EventCategoryAdmin(admin.ModelAdmin): list_display = ('name',) - class OccurrenceInline(TabularDynamicInlineAdmin): + class OccurrenceInline(StackedDynamicInlineAdmin): model = Occurrence extra = 1 - fields = ('start_time', 'end_time', 'description') + fields = ('start_time', 'end_time', 'description', 'location') class EventAdmin(DisplayableAdmin): list_display = ('title', 'event_category') list_filter = ('event_category',) search_fields = ('title', 'description', 'content', 'keywords') fieldsets = ( (None, { "fields": [ "title", "status", ("publish_date", "expiry_date"), "event_category", "content" ] }), (_("Meta data"), { "fields": [ "_meta_title", "slug", ("description", "gen_description"), "keywords", "in_sitemap" ], "classes": ("collapse-closed",) }), ) inlines = [OccurrenceInline] admin.site.register(Event, EventAdmin) admin.site.register(EventCategory, EventCategoryAdmin)
Change to stacked inline for occurrences, also display location.
## Code Before: from django.utils.translation import ugettext_lazy as _ from django.contrib import admin from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin from fullcalendar.models import * class EventCategoryAdmin(admin.ModelAdmin): list_display = ('name',) class OccurrenceInline(TabularDynamicInlineAdmin): model = Occurrence extra = 1 fields = ('start_time', 'end_time', 'description') class EventAdmin(DisplayableAdmin): list_display = ('title', 'event_category') list_filter = ('event_category',) search_fields = ('title', 'description', 'content', 'keywords') fieldsets = ( (None, { "fields": [ "title", "status", ("publish_date", "expiry_date"), "event_category", "content" ] }), (_("Meta data"), { "fields": [ "_meta_title", "slug", ("description", "gen_description"), "keywords", "in_sitemap" ], "classes": ("collapse-closed",) }), ) inlines = [OccurrenceInline] admin.site.register(Event, EventAdmin) admin.site.register(EventCategory, EventCategoryAdmin) ## Instruction: Change to stacked inline for occurrences, also display location. ## Code After: from django.utils.translation import ugettext_lazy as _ from django.contrib import admin from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin from fullcalendar.models import * class EventCategoryAdmin(admin.ModelAdmin): list_display = ('name',) class OccurrenceInline(StackedDynamicInlineAdmin): model = Occurrence extra = 1 fields = ('start_time', 'end_time', 'description', 'location') class EventAdmin(DisplayableAdmin): list_display = ('title', 'event_category') list_filter = ('event_category',) search_fields = ('title', 'description', 'content', 'keywords') fieldsets = ( (None, { "fields": [ "title", "status", ("publish_date", "expiry_date"), "event_category", "content" ] }), (_("Meta data"), { "fields": [ "_meta_title", "slug", ("description", "gen_description"), "keywords", "in_sitemap" ], "classes": ("collapse-closed",) }), ) inlines = [OccurrenceInline] admin.site.register(Event, EventAdmin) admin.site.register(EventCategory, EventCategoryAdmin)
from django.utils.translation import ugettext_lazy as _ from django.contrib import admin - from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin ? ^ ^^^^^ + from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin ? ^^ ^^^^ from fullcalendar.models import * class EventCategoryAdmin(admin.ModelAdmin): list_display = ('name',) - class OccurrenceInline(TabularDynamicInlineAdmin): ? ^ ^^^^^ + class OccurrenceInline(StackedDynamicInlineAdmin): ? ^^ ^^^^ model = Occurrence extra = 1 - fields = ('start_time', 'end_time', 'description') + fields = ('start_time', 'end_time', 'description', 'location') ? ++++++++++++ class EventAdmin(DisplayableAdmin): list_display = ('title', 'event_category') list_filter = ('event_category',) search_fields = ('title', 'description', 'content', 'keywords') fieldsets = ( (None, { "fields": [ "title", "status", ("publish_date", "expiry_date"), "event_category", "content" ] }), (_("Meta data"), { "fields": [ "_meta_title", "slug", ("description", "gen_description"), "keywords", "in_sitemap" ], "classes": ("collapse-closed",) }), ) inlines = [OccurrenceInline] admin.site.register(Event, EventAdmin) admin.site.register(EventCategory, EventCategoryAdmin)
eba6e117c0a13b49219bb60e773f896b274b6601
tests/_support/configs/collection.py
tests/_support/configs/collection.py
from spec import eq_ from invoke import ctask, Collection @ctask def collection(c): c.run('false') # Ensures a kaboom if mocking fails ns = Collection(collection) ns.configure({'run': {'echo': True}})
from spec import eq_ from invoke import ctask, Collection @ctask def go(c): c.run('false') # Ensures a kaboom if mocking fails ns = Collection(go) ns.configure({'run': {'echo': True}})
Fix test fixture to match earlier test change
Fix test fixture to match earlier test change
Python
bsd-2-clause
singingwolfboy/invoke,kejbaly2/invoke,tyewang/invoke,frol/invoke,mattrobenolt/invoke,mkusz/invoke,pfmoore/invoke,mkusz/invoke,pyinvoke/invoke,kejbaly2/invoke,pfmoore/invoke,sophacles/invoke,frol/invoke,pyinvoke/invoke,mattrobenolt/invoke
from spec import eq_ from invoke import ctask, Collection @ctask - def collection(c): + def go(c): c.run('false') # Ensures a kaboom if mocking fails - ns = Collection(collection) + ns = Collection(go) ns.configure({'run': {'echo': True}})
Fix test fixture to match earlier test change
## Code Before: from spec import eq_ from invoke import ctask, Collection @ctask def collection(c): c.run('false') # Ensures a kaboom if mocking fails ns = Collection(collection) ns.configure({'run': {'echo': True}}) ## Instruction: Fix test fixture to match earlier test change ## Code After: from spec import eq_ from invoke import ctask, Collection @ctask def go(c): c.run('false') # Ensures a kaboom if mocking fails ns = Collection(go) ns.configure({'run': {'echo': True}})
from spec import eq_ from invoke import ctask, Collection @ctask - def collection(c): + def go(c): c.run('false') # Ensures a kaboom if mocking fails - ns = Collection(collection) ? ^ -------- + ns = Collection(go) ? ^ ns.configure({'run': {'echo': True}})
e5acbfc176de3b531528c8b15f57e5d3feab3ad1
melody/constraints/abstract_constraint.py
melody/constraints/abstract_constraint.py
from abc import ABCMeta, abstractmethod class AbstractConstraint(object): """ Class that represents a constraint, a set of actors that define a constraint amongst themselves. ParameterMap: A map from template note to contextual note.. """ __metaclass__ = ABCMeta def __init__(self, actors): self.__actors = list(actors) @property def actors(self): return list(self.__actors) @abstractmethod def clone(self, new_actors=None): """ Clone the constraint. :return: """ @abstractmethod def verify(self, solution_context): """ Verify that the actor map parameters are consistent with constraint. :params solution_context: aka pmap, map of actors to ContextualNotes. :return: Boolean if verification holds. May throw Exception dependent on implementation. """ @abstractmethod def values(self, solution_context, v_note): """ Method to generate all possible note values for actor v_note's target. The method returns a set of values for v_note. :param solution_context: includes parameter map. :param v_note: source actor, whose target values we are computing. :return: The set of all possible values for v_note's target. Note: The return value is a set! """
from abc import ABCMeta, abstractmethod class AbstractConstraint(object): """ Class that represents a constraint, a set of actors that define a constraint amongst themselves. ParameterMap: A map from template note to contextual note.. """ __metaclass__ = ABCMeta def __init__(self, actors): self.__actors = list(actors) @property def actors(self): return list(self.__actors) @abstractmethod def clone(self, new_actors=None): """ Clone the constraint. :return: """ @abstractmethod def verify(self, solution_context): """ Verify that the actor map parameters are consistent with constraint. :params solution_context: aka pmap, map of actors to ContextualNotes. :return: Boolean if verification holds. May throw Exception dependent on implementation. """ @abstractmethod def values(self, solution_context, v_note): """ Method to generate all possible note values for actor v_note's target. The method returns a set of values for v_note. :param solution_context: includes parameter map. :param v_note: source actor, whose target values we are computing. :return: The set of all possible values for v_note's target. Note: The return value is a set! """ def __hash__(self): return hash(len(self.actors)) def __eq__(self, other): if not isinstance(other, AbstractConstraint): return NotImplemented return self is other
Add hash and eq methods
Add hash and eq methods
Python
mit
dpazel/music_rep
from abc import ABCMeta, abstractmethod class AbstractConstraint(object): """ Class that represents a constraint, a set of actors that define a constraint amongst themselves. ParameterMap: A map from template note to contextual note.. """ __metaclass__ = ABCMeta def __init__(self, actors): self.__actors = list(actors) @property def actors(self): return list(self.__actors) @abstractmethod def clone(self, new_actors=None): """ Clone the constraint. :return: """ @abstractmethod def verify(self, solution_context): """ Verify that the actor map parameters are consistent with constraint. :params solution_context: aka pmap, map of actors to ContextualNotes. :return: Boolean if verification holds. May throw Exception dependent on implementation. """ @abstractmethod def values(self, solution_context, v_note): """ Method to generate all possible note values for actor v_note's target. The method returns a set of values for v_note. :param solution_context: includes parameter map. :param v_note: source actor, whose target values we are computing. :return: The set of all possible values for v_note's target. Note: The return value is a set! """ + def __hash__(self): + return hash(len(self.actors)) + + def __eq__(self, other): + if not isinstance(other, AbstractConstraint): + return NotImplemented + return self is other
Add hash and eq methods
## Code Before: from abc import ABCMeta, abstractmethod class AbstractConstraint(object): """ Class that represents a constraint, a set of actors that define a constraint amongst themselves. ParameterMap: A map from template note to contextual note.. """ __metaclass__ = ABCMeta def __init__(self, actors): self.__actors = list(actors) @property def actors(self): return list(self.__actors) @abstractmethod def clone(self, new_actors=None): """ Clone the constraint. :return: """ @abstractmethod def verify(self, solution_context): """ Verify that the actor map parameters are consistent with constraint. :params solution_context: aka pmap, map of actors to ContextualNotes. :return: Boolean if verification holds. May throw Exception dependent on implementation. """ @abstractmethod def values(self, solution_context, v_note): """ Method to generate all possible note values for actor v_note's target. The method returns a set of values for v_note. :param solution_context: includes parameter map. :param v_note: source actor, whose target values we are computing. :return: The set of all possible values for v_note's target. Note: The return value is a set! """ ## Instruction: Add hash and eq methods ## Code After: from abc import ABCMeta, abstractmethod class AbstractConstraint(object): """ Class that represents a constraint, a set of actors that define a constraint amongst themselves. ParameterMap: A map from template note to contextual note.. """ __metaclass__ = ABCMeta def __init__(self, actors): self.__actors = list(actors) @property def actors(self): return list(self.__actors) @abstractmethod def clone(self, new_actors=None): """ Clone the constraint. :return: """ @abstractmethod def verify(self, solution_context): """ Verify that the actor map parameters are consistent with constraint. :params solution_context: aka pmap, map of actors to ContextualNotes. :return: Boolean if verification holds. May throw Exception dependent on implementation. """ @abstractmethod def values(self, solution_context, v_note): """ Method to generate all possible note values for actor v_note's target. The method returns a set of values for v_note. :param solution_context: includes parameter map. :param v_note: source actor, whose target values we are computing. :return: The set of all possible values for v_note's target. Note: The return value is a set! """ def __hash__(self): return hash(len(self.actors)) def __eq__(self, other): if not isinstance(other, AbstractConstraint): return NotImplemented return self is other
from abc import ABCMeta, abstractmethod class AbstractConstraint(object): """ Class that represents a constraint, a set of actors that define a constraint amongst themselves. ParameterMap: A map from template note to contextual note.. """ __metaclass__ = ABCMeta def __init__(self, actors): self.__actors = list(actors) @property def actors(self): return list(self.__actors) @abstractmethod def clone(self, new_actors=None): """ Clone the constraint. :return: """ @abstractmethod def verify(self, solution_context): """ Verify that the actor map parameters are consistent with constraint. :params solution_context: aka pmap, map of actors to ContextualNotes. :return: Boolean if verification holds. May throw Exception dependent on implementation. """ @abstractmethod def values(self, solution_context, v_note): """ Method to generate all possible note values for actor v_note's target. The method returns a set of values for v_note. :param solution_context: includes parameter map. :param v_note: source actor, whose target values we are computing. :return: The set of all possible values for v_note's target. Note: The return value is a set! """ + + def __hash__(self): + return hash(len(self.actors)) + + def __eq__(self, other): + if not isinstance(other, AbstractConstraint): + return NotImplemented + return self is other
d2e03bf76f585dc1025b5a94be0327284f8d5fa2
Left_pare.py
Left_pare.py
from xpcom import components viewSvc = components.classes["@activestate.com/koViewService;1"]\ .getService(components.interfaces.koIViewService) view = viewSvc.currentView view = view.queryInterface(components.interfaces.koIScintillaView) sm = view.scimoz # Make `start` the beginning position of the first selected line, # and `end` the ending position of the last selected line. if sm.anchor < sm.currentPos: start = sm.positionFromLine(sm.lineFromPosition(sm.anchor)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos)) else: start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor)) lines = tuple(sm.getTextRange(start, end).splitlines()) # Cut one character from the left lines = tuple(l[1:] for l in lines) # Select part of document sm.setSel(start, end) # Replace selection content text = '\n'.join(lines) sm.replaceSel(text) # Keep selection to allow to continue to apply this macro if use wants sm.setSel(start, start+len(text))
from xpcom import components viewSvc = components.classes["@activestate.com/koViewService;1"]\ .getService(components.interfaces.koIViewService) view = viewSvc.currentView view = view.queryInterface(components.interfaces.koIScintillaView) sm = view.scimoz # Make `start` the beginning position of the first selected line, # and `end` the ending position of the last selected line. if sm.anchor < sm.currentPos: start = sm.positionFromLine(sm.lineFromPosition(sm.anchor)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos)) else: start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor)) lines = tuple(sm.getTextRange(start, end).splitlines()) # Cut one character from the left lines = tuple(l[1:] for l in lines) # Select part of document sm.setSel(start, end) # Replace selection content text = '\n'.join(lines) sm.replaceSel(text) # Keep selection to let user continue to apply this macro sm.setSel(start, start+len(text.encode('utf-8')))
Correct selecting text by length when text is Unicode.
Correct selecting text by length when text is Unicode.
Python
mpl-2.0
Komodo/macros,Komodo/macros
from xpcom import components viewSvc = components.classes["@activestate.com/koViewService;1"]\ .getService(components.interfaces.koIViewService) view = viewSvc.currentView view = view.queryInterface(components.interfaces.koIScintillaView) sm = view.scimoz # Make `start` the beginning position of the first selected line, # and `end` the ending position of the last selected line. if sm.anchor < sm.currentPos: start = sm.positionFromLine(sm.lineFromPosition(sm.anchor)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos)) else: start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor)) lines = tuple(sm.getTextRange(start, end).splitlines()) # Cut one character from the left lines = tuple(l[1:] for l in lines) # Select part of document sm.setSel(start, end) # Replace selection content text = '\n'.join(lines) sm.replaceSel(text) - # Keep selection to allow to continue to apply this macro if use wants + # Keep selection to let user continue to apply this macro - sm.setSel(start, start+len(text)) + sm.setSel(start, start+len(text.encode('utf-8')))
Correct selecting text by length when text is Unicode.
## Code Before: from xpcom import components viewSvc = components.classes["@activestate.com/koViewService;1"]\ .getService(components.interfaces.koIViewService) view = viewSvc.currentView view = view.queryInterface(components.interfaces.koIScintillaView) sm = view.scimoz # Make `start` the beginning position of the first selected line, # and `end` the ending position of the last selected line. if sm.anchor < sm.currentPos: start = sm.positionFromLine(sm.lineFromPosition(sm.anchor)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos)) else: start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor)) lines = tuple(sm.getTextRange(start, end).splitlines()) # Cut one character from the left lines = tuple(l[1:] for l in lines) # Select part of document sm.setSel(start, end) # Replace selection content text = '\n'.join(lines) sm.replaceSel(text) # Keep selection to allow to continue to apply this macro if use wants sm.setSel(start, start+len(text)) ## Instruction: Correct selecting text by length when text is Unicode. ## Code After: from xpcom import components viewSvc = components.classes["@activestate.com/koViewService;1"]\ .getService(components.interfaces.koIViewService) view = viewSvc.currentView view = view.queryInterface(components.interfaces.koIScintillaView) sm = view.scimoz # Make `start` the beginning position of the first selected line, # and `end` the ending position of the last selected line. if sm.anchor < sm.currentPos: start = sm.positionFromLine(sm.lineFromPosition(sm.anchor)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos)) else: start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor)) lines = tuple(sm.getTextRange(start, end).splitlines()) # Cut one character from the left lines = tuple(l[1:] for l in lines) # Select part of document sm.setSel(start, end) # Replace selection content text = '\n'.join(lines) sm.replaceSel(text) # Keep selection to let user continue to apply this macro sm.setSel(start, start+len(text.encode('utf-8')))
from xpcom import components viewSvc = components.classes["@activestate.com/koViewService;1"]\ .getService(components.interfaces.koIViewService) view = viewSvc.currentView view = view.queryInterface(components.interfaces.koIScintillaView) sm = view.scimoz # Make `start` the beginning position of the first selected line, # and `end` the ending position of the last selected line. if sm.anchor < sm.currentPos: start = sm.positionFromLine(sm.lineFromPosition(sm.anchor)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos)) else: start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos)) end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor)) lines = tuple(sm.getTextRange(start, end).splitlines()) # Cut one character from the left lines = tuple(l[1:] for l in lines) # Select part of document sm.setSel(start, end) # Replace selection content text = '\n'.join(lines) sm.replaceSel(text) - # Keep selection to allow to continue to apply this macro if use wants ? - ^^^ ^^ ------------- + # Keep selection to let user continue to apply this macro ? ^^ ^^^^ - sm.setSel(start, start+len(text)) + sm.setSel(start, start+len(text.encode('utf-8'))) ? +++++++++++++++ +
a7d8d2f95acbf801c0cc8b0f2a8cc008f6cb34c0
rouver/types.py
rouver/types.py
from __future__ import annotations from collections.abc import Iterable, Mapping from typing import Any, Callable, Dict, Tuple from typing_extensions import TypeAlias from werkzeug.wrappers import Request # (name, value) Header: TypeAlias = Tuple[str, str] WSGIEnvironment: TypeAlias = Dict[str, Any] # (body) -> None StartResponseReturnType: TypeAlias = Callable[[bytes], object] # (status: str, headers: List[Headers], exc_info) -> response StartResponse: TypeAlias = Callable[..., StartResponseReturnType] WSGIResponse: TypeAlias = Iterable[bytes] WSGIApplication: TypeAlias = Callable[ [WSGIEnvironment, StartResponse], WSGIResponse ] # (method, path, callback) RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication] # (request, previous_args, path_part) -> result RouteTemplateHandler: TypeAlias = Callable[ [Request, Tuple[Any, ...], str], Any ] BadArgumentsDict: TypeAlias = Mapping[str, str]
from __future__ import annotations from typing import Any, Callable, Dict, Iterable, Mapping, Tuple from typing_extensions import TypeAlias from werkzeug.wrappers import Request # (name, value) Header: TypeAlias = Tuple[str, str] WSGIEnvironment: TypeAlias = Dict[str, Any] # (body) -> None StartResponseReturnType: TypeAlias = Callable[[bytes], object] # (status: str, headers: List[Headers], exc_info) -> response StartResponse: TypeAlias = Callable[..., StartResponseReturnType] WSGIResponse: TypeAlias = Iterable[bytes] WSGIApplication: TypeAlias = Callable[ [WSGIEnvironment, StartResponse], WSGIResponse ] # (method, path, callback) RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication] # (request, previous_args, path_part) -> result RouteTemplateHandler: TypeAlias = Callable[ [Request, Tuple[Any, ...], str], Any ] BadArgumentsDict: TypeAlias = Mapping[str, str]
Fix imports on Python <= 3.8
Fix imports on Python <= 3.8
Python
mit
srittau/rouver
from __future__ import annotations - from collections.abc import Iterable, Mapping - from typing import Any, Callable, Dict, Tuple + from typing import Any, Callable, Dict, Iterable, Mapping, Tuple from typing_extensions import TypeAlias from werkzeug.wrappers import Request # (name, value) Header: TypeAlias = Tuple[str, str] WSGIEnvironment: TypeAlias = Dict[str, Any] # (body) -> None StartResponseReturnType: TypeAlias = Callable[[bytes], object] # (status: str, headers: List[Headers], exc_info) -> response StartResponse: TypeAlias = Callable[..., StartResponseReturnType] WSGIResponse: TypeAlias = Iterable[bytes] WSGIApplication: TypeAlias = Callable[ [WSGIEnvironment, StartResponse], WSGIResponse ] # (method, path, callback) RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication] # (request, previous_args, path_part) -> result RouteTemplateHandler: TypeAlias = Callable[ [Request, Tuple[Any, ...], str], Any ] BadArgumentsDict: TypeAlias = Mapping[str, str]
Fix imports on Python <= 3.8
## Code Before: from __future__ import annotations from collections.abc import Iterable, Mapping from typing import Any, Callable, Dict, Tuple from typing_extensions import TypeAlias from werkzeug.wrappers import Request # (name, value) Header: TypeAlias = Tuple[str, str] WSGIEnvironment: TypeAlias = Dict[str, Any] # (body) -> None StartResponseReturnType: TypeAlias = Callable[[bytes], object] # (status: str, headers: List[Headers], exc_info) -> response StartResponse: TypeAlias = Callable[..., StartResponseReturnType] WSGIResponse: TypeAlias = Iterable[bytes] WSGIApplication: TypeAlias = Callable[ [WSGIEnvironment, StartResponse], WSGIResponse ] # (method, path, callback) RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication] # (request, previous_args, path_part) -> result RouteTemplateHandler: TypeAlias = Callable[ [Request, Tuple[Any, ...], str], Any ] BadArgumentsDict: TypeAlias = Mapping[str, str] ## Instruction: Fix imports on Python <= 3.8 ## Code After: from __future__ import annotations from typing import Any, Callable, Dict, Iterable, Mapping, Tuple from typing_extensions import TypeAlias from werkzeug.wrappers import Request # (name, value) Header: TypeAlias = Tuple[str, str] WSGIEnvironment: TypeAlias = Dict[str, Any] # (body) -> None StartResponseReturnType: TypeAlias = Callable[[bytes], object] # (status: str, headers: List[Headers], exc_info) -> response StartResponse: TypeAlias = Callable[..., StartResponseReturnType] WSGIResponse: TypeAlias = Iterable[bytes] WSGIApplication: TypeAlias = Callable[ [WSGIEnvironment, StartResponse], WSGIResponse ] # (method, path, callback) RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication] # (request, previous_args, path_part) -> result RouteTemplateHandler: TypeAlias = Callable[ [Request, Tuple[Any, ...], str], Any ] BadArgumentsDict: TypeAlias = Mapping[str, str]
from __future__ import annotations - from collections.abc import Iterable, Mapping - from typing import Any, Callable, Dict, Tuple + from typing import Any, Callable, Dict, Iterable, Mapping, Tuple ? +++++++++++++++++++ from typing_extensions import TypeAlias from werkzeug.wrappers import Request # (name, value) Header: TypeAlias = Tuple[str, str] WSGIEnvironment: TypeAlias = Dict[str, Any] # (body) -> None StartResponseReturnType: TypeAlias = Callable[[bytes], object] # (status: str, headers: List[Headers], exc_info) -> response StartResponse: TypeAlias = Callable[..., StartResponseReturnType] WSGIResponse: TypeAlias = Iterable[bytes] WSGIApplication: TypeAlias = Callable[ [WSGIEnvironment, StartResponse], WSGIResponse ] # (method, path, callback) RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication] # (request, previous_args, path_part) -> result RouteTemplateHandler: TypeAlias = Callable[ [Request, Tuple[Any, ...], str], Any ] BadArgumentsDict: TypeAlias = Mapping[str, str]
06c5f27c04de9fa62f6ac4834e0a920349c27084
rules/binutils.py
rules/binutils.py
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) rules = Binutils()
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) # For now we strip the man pages. # man pages created on different systems are (for no good reason) different! man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) shutil.rmtree(man_dir) rules = Binutils()
Remove man pages post-install (for now)
Remove man pages post-install (for now)
Python
mit
BreakawayConsulting/xyz
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) + # For now we strip the man pages. + # man pages created on different systems are (for no good reason) different! + man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) + shutil.rmtree(man_dir) rules = Binutils()
Remove man pages post-install (for now)
## Code Before: import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) rules = Binutils() ## Instruction: Remove man pages post-install (for now) ## Code After: import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) # For now we strip the man pages. # man pages created on different systems are (for no good reason) different! man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) shutil.rmtree(man_dir) rules = Binutils()
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) + # For now we strip the man pages. + # man pages created on different systems are (for no good reason) different! + man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) + shutil.rmtree(man_dir) rules = Binutils()
7bd3b5a9f1664bd153c934d75f245a605aadee2a
ovp_core/models/cause.py
ovp_core/models/cause.py
from django.db import models class Cause(models.Model): name = models.CharField('name', max_length=100) def __str__(self): return self.name class Meta: app_label = 'ovp_core' verbose_name = 'cause' verbose_name_plural = 'causes'
import vinaigrette from django.db import models class Cause(models.Model): name = models.CharField('name', max_length=100) def __str__(self): return self.name class Meta: app_label = 'ovp_core' verbose_name = 'cause' verbose_name_plural = 'causes' vinaigrette.register(Cause, ['name'])
Apply django-vinaigrette to Cause model
Apply django-vinaigrette to Cause model
Python
agpl-3.0
OpenVolunteeringPlatform/django-ovp-core,OpenVolunteeringPlatform/django-ovp-core
+ import vinaigrette from django.db import models class Cause(models.Model): name = models.CharField('name', max_length=100) def __str__(self): return self.name class Meta: app_label = 'ovp_core' verbose_name = 'cause' verbose_name_plural = 'causes' + vinaigrette.register(Cause, ['name']) +
Apply django-vinaigrette to Cause model
## Code Before: from django.db import models class Cause(models.Model): name = models.CharField('name', max_length=100) def __str__(self): return self.name class Meta: app_label = 'ovp_core' verbose_name = 'cause' verbose_name_plural = 'causes' ## Instruction: Apply django-vinaigrette to Cause model ## Code After: import vinaigrette from django.db import models class Cause(models.Model): name = models.CharField('name', max_length=100) def __str__(self): return self.name class Meta: app_label = 'ovp_core' verbose_name = 'cause' verbose_name_plural = 'causes' vinaigrette.register(Cause, ['name'])
+ import vinaigrette from django.db import models class Cause(models.Model): name = models.CharField('name', max_length=100) def __str__(self): return self.name class Meta: app_label = 'ovp_core' verbose_name = 'cause' verbose_name_plural = 'causes' + + vinaigrette.register(Cause, ['name'])
d5b5421c95b1e2feb4646a42b5aca71a2280e30c
tests/dojo_test.py
tests/dojo_test.py
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def test_create_room_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) blue_office = my_class_instance.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def test_create_room_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) blue_office = my_class_instance.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3) def test_person_added_to_system(self): initial_person_count = len(self.dojo.all_people) person = self.dojo.add_person("Neil", "Armstrong", "Staff") self.assertTrue(person) new_person_count = len(self.dojo.all_people) self.assertEqual(new_person_count - initial_person_count, 1)
Create test to check that a person has been added
Create test to check that a person has been added
Python
mit
EdwinKato/Space-Allocator,EdwinKato/Space-Allocator
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def test_create_room_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) blue_office = my_class_instance.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3) + + def test_person_added_to_system(self): + initial_person_count = len(self.dojo.all_people) + person = self.dojo.add_person("Neil", "Armstrong", "Staff") + self.assertTrue(person) + new_person_count = len(self.dojo.all_people) + self.assertEqual(new_person_count - initial_person_count, 1)
Create test to check that a person has been added
## Code Before: import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def test_create_room_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) blue_office = my_class_instance.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3) ## Instruction: Create test to check that a person has been added ## Code After: import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def test_create_room_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) blue_office = my_class_instance.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3) def test_person_added_to_system(self): initial_person_count = len(self.dojo.all_people) person = self.dojo.add_person("Neil", "Armstrong", "Staff") self.assertTrue(person) new_person_count = len(self.dojo.all_people) self.assertEqual(new_person_count - initial_person_count, 1)
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def test_create_room_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) blue_office = my_class_instance.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3) + + def test_person_added_to_system(self): + initial_person_count = len(self.dojo.all_people) + person = self.dojo.add_person("Neil", "Armstrong", "Staff") + self.assertTrue(person) + new_person_count = len(self.dojo.all_people) + self.assertEqual(new_person_count - initial_person_count, 1)
2685b94838c8ec7ce31da60bc6f28953152c788a
pixelmap/pixelmap.py
pixelmap/pixelmap.py
from pixel import Pixel class Pixelmap: def __init__(self, width, height): """Pixelmap constructor :param width: Width of map in pixels. :param height: Height of map in pixels. """ self.width = width self.height = height self.map_matrix = [[0]*self.width for _ in range(self.height)] for row in range(self.height): for col in range(self.width): self.map_matrix[row][col] = Pixel() def __str__(self): """Human readable pixelmap description. Pretty much just the matrix. :return: Description of pixelmap. """ return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix])) def __repr__(self): """Internal representation Just use str for now. """ return self.__str__()
from .pixel import Pixel class Pixelmap: def __init__(self, cols, rows, default_val=None): """Pixelmap constructor :param cols: Width of map in pixels. :param rows: Height of map in pixels. :param default_val: Default value for pixels. """ assert cols >= 0, 'Invalid Pixelmap width' assert rows >= 0, 'Invalid Pixelmap height' self.cols = cols self.rows = rows self.map_matrix = [[0]*self.cols for _ in range(self.rows)] for row in range(self.rows): for col in range(self.cols): self.map_matrix[row][col] = Pixel(default_val) def num_cols(self): return self.cols def num_rows(self): return self.rows def __str__(self): """Human readable pixelmap description. Pretty much just the matrix. :return: Description of pixelmap. """ return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix]))
Add default value for matrix and methods to get columns and rows.
Add default value for matrix and methods to get columns and rows.
Python
mit
yebra06/pixelmap
- from pixel import Pixel + from .pixel import Pixel class Pixelmap: - def __init__(self, width, height): + def __init__(self, cols, rows, default_val=None): """Pixelmap constructor - :param width: Width of map in pixels. + :param cols: Width of map in pixels. - :param height: Height of map in pixels. + :param rows: Height of map in pixels. + :param default_val: Default value for pixels. """ - self.width = width - self.height = height + assert cols >= 0, 'Invalid Pixelmap width' + assert rows >= 0, 'Invalid Pixelmap height' + self.cols = cols + self.rows = rows - self.map_matrix = [[0]*self.width for _ in range(self.height)] + self.map_matrix = [[0]*self.cols for _ in range(self.rows)] - for row in range(self.height): + for row in range(self.rows): - for col in range(self.width): + for col in range(self.cols): - self.map_matrix[row][col] = Pixel() + self.map_matrix[row][col] = Pixel(default_val) + + def num_cols(self): + return self.cols + + def num_rows(self): + return self.rows def __str__(self): """Human readable pixelmap description. Pretty much just the matrix. :return: Description of pixelmap. """ - return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix])) + return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix])) - def __repr__(self): - """Internal representation - - Just use str for now. - """ - return self.__str__() -
Add default value for matrix and methods to get columns and rows.
## Code Before: from pixel import Pixel class Pixelmap: def __init__(self, width, height): """Pixelmap constructor :param width: Width of map in pixels. :param height: Height of map in pixels. """ self.width = width self.height = height self.map_matrix = [[0]*self.width for _ in range(self.height)] for row in range(self.height): for col in range(self.width): self.map_matrix[row][col] = Pixel() def __str__(self): """Human readable pixelmap description. Pretty much just the matrix. :return: Description of pixelmap. """ return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix])) def __repr__(self): """Internal representation Just use str for now. """ return self.__str__() ## Instruction: Add default value for matrix and methods to get columns and rows. ## Code After: from .pixel import Pixel class Pixelmap: def __init__(self, cols, rows, default_val=None): """Pixelmap constructor :param cols: Width of map in pixels. :param rows: Height of map in pixels. :param default_val: Default value for pixels. """ assert cols >= 0, 'Invalid Pixelmap width' assert rows >= 0, 'Invalid Pixelmap height' self.cols = cols self.rows = rows self.map_matrix = [[0]*self.cols for _ in range(self.rows)] for row in range(self.rows): for col in range(self.cols): self.map_matrix[row][col] = Pixel(default_val) def num_cols(self): return self.cols def num_rows(self): return self.rows def __str__(self): """Human readable pixelmap description. Pretty much just the matrix. :return: Description of pixelmap. """ return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix]))
- from pixel import Pixel + from .pixel import Pixel ? + class Pixelmap: - def __init__(self, width, height): + def __init__(self, cols, rows, default_val=None): """Pixelmap constructor - :param width: Width of map in pixels. ? ^^^^^ + :param cols: Width of map in pixels. ? ^^^^ - :param height: Height of map in pixels. ? ^^^^^^ + :param rows: Height of map in pixels. ? ^^^^ + :param default_val: Default value for pixels. """ - self.width = width - self.height = height + assert cols >= 0, 'Invalid Pixelmap width' + assert rows >= 0, 'Invalid Pixelmap height' + self.cols = cols + self.rows = rows - self.map_matrix = [[0]*self.width for _ in range(self.height)] ? ^^^^^ ^^^^^^ + self.map_matrix = [[0]*self.cols for _ in range(self.rows)] ? ^^^^ ^^^^ - for row in range(self.height): ? ^^^^^^ + for row in range(self.rows): ? ^^^^ - for col in range(self.width): ? ^^^^^ + for col in range(self.cols): ? ^^^^ - self.map_matrix[row][col] = Pixel() + self.map_matrix[row][col] = Pixel(default_val) ? +++++++++++ + + def num_cols(self): + return self.cols + + def num_rows(self): + return self.rows def __str__(self): """Human readable pixelmap description. Pretty much just the matrix. :return: Description of pixelmap. """ - return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix])) ? ^ + return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix])) ? ^ ++++ + - - def __repr__(self): - """Internal representation - - Just use str for now. - """ - return self.__str__()
59536a70ef39e34a5aea57131492a475e05cd227
lg_cms_director/setup.py
lg_cms_director/setup.py
from distutils.core import setup from catkin_pkg.packages import find_packages from catkin_pkg.python_setup import generate_distutils_setup d = generate_distutils_setup( packages=['trollius', 'pulsar'], package_dir={'': 'src'}, scripts=[], requires=[] ) setup(**d) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
from distutils.core import setup from catkin_pkg.packages import find_packages from catkin_pkg.python_setup import generate_distutils_setup d = generate_distutils_setup( packages=find_packages('src'), package_dir={'': 'src'}, scripts=[], requires=[] ) setup(**d) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
Revert to previous packaging of director's dependencies thx to @mvollrath
Revert to previous packaging of director's dependencies thx to @mvollrath
Python
apache-2.0
EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes
from distutils.core import setup from catkin_pkg.packages import find_packages from catkin_pkg.python_setup import generate_distutils_setup d = generate_distutils_setup( - packages=['trollius', 'pulsar'], + packages=find_packages('src'), package_dir={'': 'src'}, scripts=[], requires=[] ) setup(**d) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
Revert to previous packaging of director's dependencies thx to @mvollrath
## Code Before: from distutils.core import setup from catkin_pkg.packages import find_packages from catkin_pkg.python_setup import generate_distutils_setup d = generate_distutils_setup( packages=['trollius', 'pulsar'], package_dir={'': 'src'}, scripts=[], requires=[] ) setup(**d) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4 ## Instruction: Revert to previous packaging of director's dependencies thx to @mvollrath ## Code After: from distutils.core import setup from catkin_pkg.packages import find_packages from catkin_pkg.python_setup import generate_distutils_setup d = generate_distutils_setup( packages=find_packages('src'), package_dir={'': 'src'}, scripts=[], requires=[] ) setup(**d) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
from distutils.core import setup from catkin_pkg.packages import find_packages from catkin_pkg.python_setup import generate_distutils_setup d = generate_distutils_setup( - packages=['trollius', 'pulsar'], + packages=find_packages('src'), package_dir={'': 'src'}, scripts=[], requires=[] ) setup(**d) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
8b5ccf93fbac8929ecfc185d7407a79b1e890bde
project_template/project_settings.py
project_template/project_settings.py
from icekit.project.settings.icekit import * # icekit, glamkit # Override the default ICEkit settings to form project settings.
from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
Use GLAMkit settings in default project template.
Use GLAMkit settings in default project template.
Python
mit
ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit
- from icekit.project.settings.icekit import * # icekit, glamkit + from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
Use GLAMkit settings in default project template.
## Code Before: from icekit.project.settings.icekit import * # icekit, glamkit # Override the default ICEkit settings to form project settings. ## Instruction: Use GLAMkit settings in default project template. ## Code After: from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
- from icekit.project.settings.icekit import * # icekit, glamkit ? ^^^ -------- + from icekit.project.settings.glamkit import * # glamkit, icekit ? ^^^^ ++++++++ # Override the default ICEkit settings to form project settings.
326bebb58242981ec66f257525e5c5f58fae9196
example/article/admin.py
example/article/admin.py
from django.contrib import admin from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget from parler.admin import TranslatableAdmin from .models import Article from parler.forms import TranslatableModelForm, TranslatedField class ArticleAdminForm(TranslatableModelForm): """ Example form Translated fields can be enhanced by manually declaring them: """ title = TranslatedField(widget=AdminTextInputWidget) content = TranslatedField(widget=AdminTextareaWidget) class ArticleAdmin(TranslatableAdmin): """ Example admin. Using an empty class would already work, but this example shows some additional options. """ # The 'language_column' is provided by the base class: list_display = ('title', 'language_column') # Example custom form usage. form = ArticleAdminForm # NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets= fieldsets = ( (None, { 'fields': ('title', 'slug', 'published'), }), ("Contents", { 'fields': ('content',), }) ) def get_prepopulated_fields(self, request, obj=None): # Can't use prepopulated_fields= yet, but this is a workaround. return {'slug': ('title',)} admin.site.register(Article, ArticleAdmin)
from django.contrib import admin from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget from parler.admin import TranslatableAdmin from .models import Article from parler.forms import TranslatableModelForm, TranslatedField class ArticleAdminForm(TranslatableModelForm): """ Example form Translated fields can be enhanced by manually declaring them: """ title = TranslatedField(widget=AdminTextInputWidget) content = TranslatedField(widget=AdminTextareaWidget) class ArticleAdmin(TranslatableAdmin): """ Example admin. Using an empty class would already work, but this example shows some additional options. """ # The 'language_column' is provided by the base class: list_display = ('title', 'language_column') list_filter = ('published',) # Example custom form usage. form = ArticleAdminForm # NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets= fieldsets = ( (None, { 'fields': ('title', 'slug', 'published'), }), ("Contents", { 'fields': ('content',), }) ) def get_prepopulated_fields(self, request, obj=None): # Can't use prepopulated_fields= yet, but this is a workaround. return {'slug': ('title',)} admin.site.register(Article, ArticleAdmin)
Add list_filter to example ArticleAdmin
Add list_filter to example ArticleAdmin When using a list filter and then adding or editing an object the language GET parameter goes missing causing the wrong translation to be edited.
Python
apache-2.0
django-parler/django-parler,jrief/django-parler,edoburu/django-parler,HiddenData/django-parler,skirsdeda/django-parler,edoburu/django-parler,zhangguiyu/django-parler,jrief/django-parler,django-parler/django-parler,imposeren/django-parler,defivelo/django-parler,imposeren/django-parler,zhangguiyu/django-parler,defivelo/django-parler,ellmetha/django-parler,HiddenData/django-parler,skirsdeda/django-parler,ellmetha/django-parler
from django.contrib import admin from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget from parler.admin import TranslatableAdmin from .models import Article from parler.forms import TranslatableModelForm, TranslatedField class ArticleAdminForm(TranslatableModelForm): """ Example form Translated fields can be enhanced by manually declaring them: """ title = TranslatedField(widget=AdminTextInputWidget) content = TranslatedField(widget=AdminTextareaWidget) class ArticleAdmin(TranslatableAdmin): """ Example admin. Using an empty class would already work, but this example shows some additional options. """ # The 'language_column' is provided by the base class: list_display = ('title', 'language_column') + list_filter = ('published',) # Example custom form usage. form = ArticleAdminForm # NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets= fieldsets = ( (None, { 'fields': ('title', 'slug', 'published'), }), ("Contents", { 'fields': ('content',), }) ) def get_prepopulated_fields(self, request, obj=None): # Can't use prepopulated_fields= yet, but this is a workaround. return {'slug': ('title',)} admin.site.register(Article, ArticleAdmin)
Add list_filter to example ArticleAdmin
## Code Before: from django.contrib import admin from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget from parler.admin import TranslatableAdmin from .models import Article from parler.forms import TranslatableModelForm, TranslatedField class ArticleAdminForm(TranslatableModelForm): """ Example form Translated fields can be enhanced by manually declaring them: """ title = TranslatedField(widget=AdminTextInputWidget) content = TranslatedField(widget=AdminTextareaWidget) class ArticleAdmin(TranslatableAdmin): """ Example admin. Using an empty class would already work, but this example shows some additional options. """ # The 'language_column' is provided by the base class: list_display = ('title', 'language_column') # Example custom form usage. form = ArticleAdminForm # NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets= fieldsets = ( (None, { 'fields': ('title', 'slug', 'published'), }), ("Contents", { 'fields': ('content',), }) ) def get_prepopulated_fields(self, request, obj=None): # Can't use prepopulated_fields= yet, but this is a workaround. return {'slug': ('title',)} admin.site.register(Article, ArticleAdmin) ## Instruction: Add list_filter to example ArticleAdmin ## Code After: from django.contrib import admin from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget from parler.admin import TranslatableAdmin from .models import Article from parler.forms import TranslatableModelForm, TranslatedField class ArticleAdminForm(TranslatableModelForm): """ Example form Translated fields can be enhanced by manually declaring them: """ title = TranslatedField(widget=AdminTextInputWidget) content = TranslatedField(widget=AdminTextareaWidget) class ArticleAdmin(TranslatableAdmin): """ Example admin. Using an empty class would already work, but this example shows some additional options. """ # The 'language_column' is provided by the base class: list_display = ('title', 'language_column') list_filter = ('published',) # Example custom form usage. form = ArticleAdminForm # NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets= fieldsets = ( (None, { 'fields': ('title', 'slug', 'published'), }), ("Contents", { 'fields': ('content',), }) ) def get_prepopulated_fields(self, request, obj=None): # Can't use prepopulated_fields= yet, but this is a workaround. return {'slug': ('title',)} admin.site.register(Article, ArticleAdmin)
from django.contrib import admin from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget from parler.admin import TranslatableAdmin from .models import Article from parler.forms import TranslatableModelForm, TranslatedField class ArticleAdminForm(TranslatableModelForm): """ Example form Translated fields can be enhanced by manually declaring them: """ title = TranslatedField(widget=AdminTextInputWidget) content = TranslatedField(widget=AdminTextareaWidget) class ArticleAdmin(TranslatableAdmin): """ Example admin. Using an empty class would already work, but this example shows some additional options. """ # The 'language_column' is provided by the base class: list_display = ('title', 'language_column') + list_filter = ('published',) # Example custom form usage. form = ArticleAdminForm # NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets= fieldsets = ( (None, { 'fields': ('title', 'slug', 'published'), }), ("Contents", { 'fields': ('content',), }) ) def get_prepopulated_fields(self, request, obj=None): # Can't use prepopulated_fields= yet, but this is a workaround. return {'slug': ('title',)} admin.site.register(Article, ArticleAdmin)
123875153e81253a44d0e8b2d8de5abee195362a
backend/shmitter/tweets/serializers.py
backend/shmitter/tweets/serializers.py
from rest_framework import serializers from shmitter.likes import services as likes_services from .models import Tweet from . import services as tweets_services class TweetSerializer(serializers.ModelSerializer): owner = serializers.ReadOnlyField(source='owner.username') is_fan = serializers.SerializerMethodField() is_retweeted = serializers.SerializerMethodField() class Meta: model = Tweet fields = ( 'id', 'owner', 'body', 'is_fan', 'is_retweeted', 'total_likes', 'created', ) def get_is_fan(self, obj) -> bool: """ Check if a `request.user` has liked this tweet (`obj`). """ user = self.context.get('request').user return likes_services.is_fan(obj, user) def get_is_retweeted(self, obj) -> bool: """ Check if a `request.user` has retweeted this tweet (`obj`). """ user = self.context.get('request').user return tweets_services.is_retweeted(obj, user)
from rest_framework import serializers from shmitter.likes import services as likes_services from .models import Tweet from . import services as tweets_services class TweetSerializer(serializers.ModelSerializer): owner = serializers.ReadOnlyField(source='owner.username') is_fan = serializers.SerializerMethodField() is_retweeted = serializers.SerializerMethodField() class Meta: model = Tweet fields = ( 'id', 'owner', 'body', 'is_fan', 'is_retweeted', 'total_likes', 'total_retweets', 'created', ) def get_is_fan(self, obj) -> bool: """ Check if a `request.user` has liked this tweet (`obj`). """ user = self.context.get('request').user return likes_services.is_fan(obj, user) def get_is_retweeted(self, obj) -> bool: """ Check if a `request.user` has retweeted this tweet (`obj`). """ user = self.context.get('request').user return tweets_services.is_retweeted(obj, user)
Add total retweets to the serializer
Add total retweets to the serializer
Python
mit
apirobot/shmitter,apirobot/shmitter,apirobot/shmitter
from rest_framework import serializers from shmitter.likes import services as likes_services from .models import Tweet from . import services as tweets_services class TweetSerializer(serializers.ModelSerializer): owner = serializers.ReadOnlyField(source='owner.username') is_fan = serializers.SerializerMethodField() is_retweeted = serializers.SerializerMethodField() class Meta: model = Tweet fields = ( 'id', 'owner', 'body', 'is_fan', 'is_retweeted', 'total_likes', + 'total_retweets', 'created', ) def get_is_fan(self, obj) -> bool: """ Check if a `request.user` has liked this tweet (`obj`). """ user = self.context.get('request').user return likes_services.is_fan(obj, user) def get_is_retweeted(self, obj) -> bool: """ Check if a `request.user` has retweeted this tweet (`obj`). """ user = self.context.get('request').user return tweets_services.is_retweeted(obj, user)
Add total retweets to the serializer
## Code Before: from rest_framework import serializers from shmitter.likes import services as likes_services from .models import Tweet from . import services as tweets_services class TweetSerializer(serializers.ModelSerializer): owner = serializers.ReadOnlyField(source='owner.username') is_fan = serializers.SerializerMethodField() is_retweeted = serializers.SerializerMethodField() class Meta: model = Tweet fields = ( 'id', 'owner', 'body', 'is_fan', 'is_retweeted', 'total_likes', 'created', ) def get_is_fan(self, obj) -> bool: """ Check if a `request.user` has liked this tweet (`obj`). """ user = self.context.get('request').user return likes_services.is_fan(obj, user) def get_is_retweeted(self, obj) -> bool: """ Check if a `request.user` has retweeted this tweet (`obj`). """ user = self.context.get('request').user return tweets_services.is_retweeted(obj, user) ## Instruction: Add total retweets to the serializer ## Code After: from rest_framework import serializers from shmitter.likes import services as likes_services from .models import Tweet from . import services as tweets_services class TweetSerializer(serializers.ModelSerializer): owner = serializers.ReadOnlyField(source='owner.username') is_fan = serializers.SerializerMethodField() is_retweeted = serializers.SerializerMethodField() class Meta: model = Tweet fields = ( 'id', 'owner', 'body', 'is_fan', 'is_retweeted', 'total_likes', 'total_retweets', 'created', ) def get_is_fan(self, obj) -> bool: """ Check if a `request.user` has liked this tweet (`obj`). """ user = self.context.get('request').user return likes_services.is_fan(obj, user) def get_is_retweeted(self, obj) -> bool: """ Check if a `request.user` has retweeted this tweet (`obj`). """ user = self.context.get('request').user return tweets_services.is_retweeted(obj, user)
from rest_framework import serializers from shmitter.likes import services as likes_services from .models import Tweet from . import services as tweets_services class TweetSerializer(serializers.ModelSerializer): owner = serializers.ReadOnlyField(source='owner.username') is_fan = serializers.SerializerMethodField() is_retweeted = serializers.SerializerMethodField() class Meta: model = Tweet fields = ( 'id', 'owner', 'body', 'is_fan', 'is_retweeted', 'total_likes', + 'total_retweets', 'created', ) def get_is_fan(self, obj) -> bool: """ Check if a `request.user` has liked this tweet (`obj`). """ user = self.context.get('request').user return likes_services.is_fan(obj, user) def get_is_retweeted(self, obj) -> bool: """ Check if a `request.user` has retweeted this tweet (`obj`). """ user = self.context.get('request').user return tweets_services.is_retweeted(obj, user)
0ee42ac3b80893557691d722eda207733289c97c
micropsi_core/world/minecraft/spockplugin.py
micropsi_core/world/minecraft/spockplugin.py
import logging from spock.mcp import mcdata, mcpacket from spock.mcmap import smpmap from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher from spock.utils import pl_announce @pl_announce('Micropsi') class MicropsiPlugin(object): def __init__(self, ploader, settings): self.worldadapter = settings['worldadapter'] self.worldadapter.spockplugin = self self.net = ploader.requires('Net') self.event = ploader.requires('Event') self.world = ploader.requires('World') self.clientinfo = ploader.requires('ClientInfo') #MicroPsi Datatargets self.psi_dispatcher = PsiDispatcher(self) self.move_x = 0 self.move_z = 0 self.move_x_ = 0 self.move_z_ = 0 def move(self, position=None): if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE): return if position is None: position = self.client_info.position self.net.push(mcpacket.Packet( ident='PLAY>Player Position and Look', data=position ))
import logging from spock.mcp import mcdata, mcpacket from spock.mcmap import smpmap from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher from spock.utils import pl_announce @pl_announce('Micropsi') class MicropsiPlugin(object): def __init__(self, ploader, settings): self.worldadapter = settings['worldadapter'] self.worldadapter.spockplugin = self self.net = ploader.requires('Net') self.event = ploader.requires('Event') self.world = ploader.requires('World') self.clientinfo = ploader.requires('ClientInfo') #MicroPsi Datatargets self.psi_dispatcher = PsiDispatcher(self) self.move_x = 0 self.move_z = 0 self.move_x_ = 0 self.move_z_ = 0 def move(self, position=None): if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE): return self.clientinfo.position = position
Move now sets the client position and allows the movement plugin do its thing
Move now sets the client position and allows the movement plugin do its thing
Python
mit
ianupright/micropsi2,ianupright/micropsi2,printedheart/micropsi2,ianupright/micropsi2,printedheart/micropsi2,printedheart/micropsi2
import logging from spock.mcp import mcdata, mcpacket from spock.mcmap import smpmap from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher from spock.utils import pl_announce @pl_announce('Micropsi') class MicropsiPlugin(object): def __init__(self, ploader, settings): self.worldadapter = settings['worldadapter'] self.worldadapter.spockplugin = self self.net = ploader.requires('Net') self.event = ploader.requires('Event') self.world = ploader.requires('World') self.clientinfo = ploader.requires('ClientInfo') #MicroPsi Datatargets self.psi_dispatcher = PsiDispatcher(self) self.move_x = 0 self.move_z = 0 self.move_x_ = 0 self.move_z_ = 0 def move(self, position=None): if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE): return + self.clientinfo.position = position + - if position is None: - position = self.client_info.position - self.net.push(mcpacket.Packet( - ident='PLAY>Player Position and Look', - data=position - ))
Move now sets the client position and allows the movement plugin do its thing
## Code Before: import logging from spock.mcp import mcdata, mcpacket from spock.mcmap import smpmap from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher from spock.utils import pl_announce @pl_announce('Micropsi') class MicropsiPlugin(object): def __init__(self, ploader, settings): self.worldadapter = settings['worldadapter'] self.worldadapter.spockplugin = self self.net = ploader.requires('Net') self.event = ploader.requires('Event') self.world = ploader.requires('World') self.clientinfo = ploader.requires('ClientInfo') #MicroPsi Datatargets self.psi_dispatcher = PsiDispatcher(self) self.move_x = 0 self.move_z = 0 self.move_x_ = 0 self.move_z_ = 0 def move(self, position=None): if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE): return if position is None: position = self.client_info.position self.net.push(mcpacket.Packet( ident='PLAY>Player Position and Look', data=position )) ## Instruction: Move now sets the client position and allows the movement plugin do its thing ## Code After: import logging from spock.mcp import mcdata, mcpacket from spock.mcmap import smpmap from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher from spock.utils import pl_announce @pl_announce('Micropsi') class MicropsiPlugin(object): def __init__(self, ploader, settings): self.worldadapter = settings['worldadapter'] self.worldadapter.spockplugin = self self.net = ploader.requires('Net') self.event = ploader.requires('Event') self.world = ploader.requires('World') self.clientinfo = ploader.requires('ClientInfo') #MicroPsi Datatargets self.psi_dispatcher = PsiDispatcher(self) self.move_x = 0 self.move_z = 0 self.move_x_ = 0 self.move_z_ = 0 def move(self, position=None): if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE): return self.clientinfo.position = position
import logging from spock.mcp import mcdata, mcpacket from spock.mcmap import smpmap from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher from spock.utils import pl_announce @pl_announce('Micropsi') class MicropsiPlugin(object): def __init__(self, ploader, settings): self.worldadapter = settings['worldadapter'] self.worldadapter.spockplugin = self self.net = ploader.requires('Net') self.event = ploader.requires('Event') self.world = ploader.requires('World') self.clientinfo = ploader.requires('ClientInfo') #MicroPsi Datatargets self.psi_dispatcher = PsiDispatcher(self) self.move_x = 0 self.move_z = 0 self.move_x_ = 0 self.move_z_ = 0 def move(self, position=None): if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE): return + self.clientinfo.position = position - if position is None: - position = self.client_info.position - self.net.push(mcpacket.Packet( - ident='PLAY>Player Position and Look', - data=position - ))
cea1f24aa0862d2feab1150fbd667159ab4cbe3a
migrations/versions/0313_email_access_validated_at.py
migrations/versions/0313_email_access_validated_at.py
from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date. op.execute(""" UPDATE users SET email_access_validated_at = created_at """) op.execute(""" UPDATE users SET email_access_validated_at = logged_in_at WHERE auth_type = 'email_auth' AND logged_in_at IS NOT NULL """) op.alter_column('users', 'email_access_validated_at', nullable=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ###
from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ###
Simplify the first migration, we will do execute statements later
Simplify the first migration, we will do execute statements later
Python
mit
alphagov/notifications-api,alphagov/notifications-api
from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) - # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date. - op.execute(""" - UPDATE - users - SET - email_access_validated_at = created_at - """) - op.execute(""" - UPDATE - users - SET - email_access_validated_at = logged_in_at - WHERE - auth_type = 'email_auth' - AND - logged_in_at IS NOT NULL - """) - op.alter_column('users', 'email_access_validated_at', nullable=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ###
Simplify the first migration, we will do execute statements later
## Code Before: from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date. op.execute(""" UPDATE users SET email_access_validated_at = created_at """) op.execute(""" UPDATE users SET email_access_validated_at = logged_in_at WHERE auth_type = 'email_auth' AND logged_in_at IS NOT NULL """) op.alter_column('users', 'email_access_validated_at', nullable=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ### ## Instruction: Simplify the first migration, we will do execute statements later ## Code After: from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ###
from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) - # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date. - op.execute(""" - UPDATE - users - SET - email_access_validated_at = created_at - """) - op.execute(""" - UPDATE - users - SET - email_access_validated_at = logged_in_at - WHERE - auth_type = 'email_auth' - AND - logged_in_at IS NOT NULL - """) - op.alter_column('users', 'email_access_validated_at', nullable=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ###
6e1892daec726b44b1bbb4d085e27fa03c0a419b
server/kcaa/kcsapi/client_test.py
server/kcaa/kcsapi/client_test.py
import pytest import client from kcaa import screens class TestScreen(object): def test_mission_result(self): screen = client.Screen() assert screen.screen == screens.UNKNOWN screen.update('/api_get_member/deck_port', None, None, None, False) assert screen.screen == screens.PORT screen.update('/api_req_mission/result', None, None, None, False) assert screen.screen == screens.MISSION_RESULT screen.update('/api_get_member/deck_port', None, None, None, False) assert screen.screen == screens.MISSION_RESULT def main(): import doctest doctest.testmod(client) pytest.main(args=[__file__.replace('.pyc', '.py')]) if __name__ == '__main__': main()
import pytest import client from kcaa import screens class TestScreen(object): def update(self, screen, api_name): screen.update(api_name, None, None, None, False) def update_sequence(self, screen, api_names): for api_name in api_names: screen.update(api_name, None, None, None, False) def test_mission_result(self): screen = client.Screen() assert screen.screen == screens.UNKNOWN self.update(screen, '/api_get_member/deck_port') assert screen.screen == screens.PORT self.update(screen, '/api_req_mission/result') assert screen.screen == screens.MISSION_RESULT self.update(screen, '/api_get_member/deck_port') assert screen.screen == screens.MISSION_RESULT def test_mission_result_real_sequence(self): screen = client.Screen() screen.screen = screens.PORT self.update_sequence(screen, [ '/api_auth_member/logincheck', '/api_get_member/material', '/api_get_member/deck_port', '/api_get_member/ndock', '/api_get_member/ship3', '/api_get_member/basic', '/api_req_mission/result', '/api_get_member/deck_port', '/api_get_member/basic', '/api_get_member/ship2', '/api_get_member/material', '/api_get_member/useitem', ]) assert screen.screen == screens.MISSION_RESULT def main(): import doctest doctest.testmod(client) pytest.main(args=[__file__.replace('.pyc', '.py')]) if __name__ == '__main__': main()
Add a Screen test for sequence of KCSAPI responses.
Add a Screen test for sequence of KCSAPI responses.
Python
apache-2.0
kcaa/kcaa,kcaa/kcaa,kcaa/kcaa,kcaa/kcaa
import pytest import client from kcaa import screens class TestScreen(object): + def update(self, screen, api_name): + screen.update(api_name, None, None, None, False) + + def update_sequence(self, screen, api_names): + for api_name in api_names: + screen.update(api_name, None, None, None, False) + def test_mission_result(self): screen = client.Screen() assert screen.screen == screens.UNKNOWN - screen.update('/api_get_member/deck_port', None, None, None, False) + self.update(screen, '/api_get_member/deck_port') assert screen.screen == screens.PORT - screen.update('/api_req_mission/result', None, None, None, False) + self.update(screen, '/api_req_mission/result') assert screen.screen == screens.MISSION_RESULT - screen.update('/api_get_member/deck_port', None, None, None, False) + self.update(screen, '/api_get_member/deck_port') + assert screen.screen == screens.MISSION_RESULT + + def test_mission_result_real_sequence(self): + screen = client.Screen() + screen.screen = screens.PORT + self.update_sequence(screen, [ + '/api_auth_member/logincheck', + '/api_get_member/material', + '/api_get_member/deck_port', + '/api_get_member/ndock', + '/api_get_member/ship3', + '/api_get_member/basic', + '/api_req_mission/result', + '/api_get_member/deck_port', + '/api_get_member/basic', + '/api_get_member/ship2', + '/api_get_member/material', + '/api_get_member/useitem', + ]) assert screen.screen == screens.MISSION_RESULT def main(): import doctest doctest.testmod(client) pytest.main(args=[__file__.replace('.pyc', '.py')]) if __name__ == '__main__': main()
Add a Screen test for sequence of KCSAPI responses.
## Code Before: import pytest import client from kcaa import screens class TestScreen(object): def test_mission_result(self): screen = client.Screen() assert screen.screen == screens.UNKNOWN screen.update('/api_get_member/deck_port', None, None, None, False) assert screen.screen == screens.PORT screen.update('/api_req_mission/result', None, None, None, False) assert screen.screen == screens.MISSION_RESULT screen.update('/api_get_member/deck_port', None, None, None, False) assert screen.screen == screens.MISSION_RESULT def main(): import doctest doctest.testmod(client) pytest.main(args=[__file__.replace('.pyc', '.py')]) if __name__ == '__main__': main() ## Instruction: Add a Screen test for sequence of KCSAPI responses. ## Code After: import pytest import client from kcaa import screens class TestScreen(object): def update(self, screen, api_name): screen.update(api_name, None, None, None, False) def update_sequence(self, screen, api_names): for api_name in api_names: screen.update(api_name, None, None, None, False) def test_mission_result(self): screen = client.Screen() assert screen.screen == screens.UNKNOWN self.update(screen, '/api_get_member/deck_port') assert screen.screen == screens.PORT self.update(screen, '/api_req_mission/result') assert screen.screen == screens.MISSION_RESULT self.update(screen, '/api_get_member/deck_port') assert screen.screen == screens.MISSION_RESULT def test_mission_result_real_sequence(self): screen = client.Screen() screen.screen = screens.PORT self.update_sequence(screen, [ '/api_auth_member/logincheck', '/api_get_member/material', '/api_get_member/deck_port', '/api_get_member/ndock', '/api_get_member/ship3', '/api_get_member/basic', '/api_req_mission/result', '/api_get_member/deck_port', '/api_get_member/basic', '/api_get_member/ship2', '/api_get_member/material', '/api_get_member/useitem', ]) assert screen.screen == screens.MISSION_RESULT def main(): import doctest doctest.testmod(client) pytest.main(args=[__file__.replace('.pyc', '.py')]) if __name__ == '__main__': main()
import pytest import client from kcaa import screens class TestScreen(object): + def update(self, screen, api_name): + screen.update(api_name, None, None, None, False) + + def update_sequence(self, screen, api_names): + for api_name in api_names: + screen.update(api_name, None, None, None, False) + def test_mission_result(self): screen = client.Screen() assert screen.screen == screens.UNKNOWN - screen.update('/api_get_member/deck_port', None, None, None, False) + self.update(screen, '/api_get_member/deck_port') assert screen.screen == screens.PORT - screen.update('/api_req_mission/result', None, None, None, False) + self.update(screen, '/api_req_mission/result') assert screen.screen == screens.MISSION_RESULT - screen.update('/api_get_member/deck_port', None, None, None, False) + self.update(screen, '/api_get_member/deck_port') + assert screen.screen == screens.MISSION_RESULT + + def test_mission_result_real_sequence(self): + screen = client.Screen() + screen.screen = screens.PORT + self.update_sequence(screen, [ + '/api_auth_member/logincheck', + '/api_get_member/material', + '/api_get_member/deck_port', + '/api_get_member/ndock', + '/api_get_member/ship3', + '/api_get_member/basic', + '/api_req_mission/result', + '/api_get_member/deck_port', + '/api_get_member/basic', + '/api_get_member/ship2', + '/api_get_member/material', + '/api_get_member/useitem', + ]) assert screen.screen == screens.MISSION_RESULT def main(): import doctest doctest.testmod(client) pytest.main(args=[__file__.replace('.pyc', '.py')]) if __name__ == '__main__': main()
3ae6c0f4c4f13207386dbf0fa2004655e9f2c8d6
UM/View/CompositePass.py
UM/View/CompositePass.py
from UM.Application import Application from UM.Resources import Resources from UM.Math.Matrix import Matrix from UM.View.RenderPass import RenderPass from UM.View.GL.OpenGL import OpenGL class CompositePass(RenderPass): def __init__(self, width, height): super().__init__("composite", width, height) self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader")) self._gl = OpenGL.getInstance().getBindingsObject() self._renderer = Application.getInstance().getRenderer() def setCompositeShader(self, shader): self._shader = shader def renderContents(self): pass def renderOutput(self): self._shader.bind() texture_unit = 0 for render_pass in self._renderer.getRenderPasses(): self._gl.glActiveTexture(texture_unit) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId()) texture_unit += 1 self._renderer.renderQuad(self._shader) for i in range(texture_unit): self._gl.glActiveTexture(texture_unit) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0) self._shader.release()
from UM.Application import Application from UM.Resources import Resources from UM.Math.Matrix import Matrix from UM.View.RenderPass import RenderPass from UM.View.GL.OpenGL import OpenGL class CompositePass(RenderPass): def __init__(self, width, height): super().__init__("composite", width, height, 999) self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader")) self._gl = OpenGL.getInstance().getBindingsObject() self._renderer = Application.getInstance().getRenderer() self._layer_bindings = [ "default", "selection" ] def setCompositeShader(self, shader): self._shader = shader def setLayerBindings(self, bindings): self._layer_bindings = bindings def render(self): self._shader.bind() texture_unit = 0 for binding in self._layer_bindings: render_pass = self._renderer.getRenderPass(binding) if not render_pass: continue self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId()) texture_unit += 1 self._renderer.renderFullScreenQuad(self._shader) for i in range(texture_unit): self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0) self._shader.release()
Add explicit render layer binding instead of assuming all render passes can be used for compositing
Add explicit render layer binding instead of assuming all render passes can be used for compositing
Python
agpl-3.0
onitake/Uranium,onitake/Uranium
from UM.Application import Application from UM.Resources import Resources from UM.Math.Matrix import Matrix from UM.View.RenderPass import RenderPass from UM.View.GL.OpenGL import OpenGL class CompositePass(RenderPass): def __init__(self, width, height): - super().__init__("composite", width, height) + super().__init__("composite", width, height, 999) self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader")) self._gl = OpenGL.getInstance().getBindingsObject() self._renderer = Application.getInstance().getRenderer() + self._layer_bindings = [ "default", "selection" ] + def setCompositeShader(self, shader): self._shader = shader - def renderContents(self): - pass + def setLayerBindings(self, bindings): + self._layer_bindings = bindings - def renderOutput(self): + def render(self): self._shader.bind() texture_unit = 0 + for binding in self._layer_bindings: - for render_pass in self._renderer.getRenderPasses(): + render_pass = self._renderer.getRenderPass(binding) - self._gl.glActiveTexture(texture_unit) + if not render_pass: + continue + + self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId()) texture_unit += 1 - self._renderer.renderQuad(self._shader) + self._renderer.renderFullScreenQuad(self._shader) for i in range(texture_unit): - self._gl.glActiveTexture(texture_unit) + self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0) self._shader.release()
Add explicit render layer binding instead of assuming all render passes can be used for compositing
## Code Before: from UM.Application import Application from UM.Resources import Resources from UM.Math.Matrix import Matrix from UM.View.RenderPass import RenderPass from UM.View.GL.OpenGL import OpenGL class CompositePass(RenderPass): def __init__(self, width, height): super().__init__("composite", width, height) self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader")) self._gl = OpenGL.getInstance().getBindingsObject() self._renderer = Application.getInstance().getRenderer() def setCompositeShader(self, shader): self._shader = shader def renderContents(self): pass def renderOutput(self): self._shader.bind() texture_unit = 0 for render_pass in self._renderer.getRenderPasses(): self._gl.glActiveTexture(texture_unit) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId()) texture_unit += 1 self._renderer.renderQuad(self._shader) for i in range(texture_unit): self._gl.glActiveTexture(texture_unit) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0) self._shader.release() ## Instruction: Add explicit render layer binding instead of assuming all render passes can be used for compositing ## Code After: from UM.Application import Application from UM.Resources import Resources from UM.Math.Matrix import Matrix from UM.View.RenderPass import RenderPass from UM.View.GL.OpenGL import OpenGL class CompositePass(RenderPass): def __init__(self, width, height): super().__init__("composite", width, height, 999) self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader")) self._gl = OpenGL.getInstance().getBindingsObject() self._renderer = Application.getInstance().getRenderer() self._layer_bindings = [ "default", "selection" ] def setCompositeShader(self, shader): self._shader = shader def setLayerBindings(self, bindings): self._layer_bindings = bindings def render(self): self._shader.bind() texture_unit = 0 for binding in self._layer_bindings: render_pass = self._renderer.getRenderPass(binding) if not render_pass: continue self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId()) texture_unit += 1 self._renderer.renderFullScreenQuad(self._shader) for i in range(texture_unit): self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0) self._shader.release()
from UM.Application import Application from UM.Resources import Resources from UM.Math.Matrix import Matrix from UM.View.RenderPass import RenderPass from UM.View.GL.OpenGL import OpenGL class CompositePass(RenderPass): def __init__(self, width, height): - super().__init__("composite", width, height) + super().__init__("composite", width, height, 999) ? +++++ self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader")) self._gl = OpenGL.getInstance().getBindingsObject() self._renderer = Application.getInstance().getRenderer() + self._layer_bindings = [ "default", "selection" ] + def setCompositeShader(self, shader): self._shader = shader - def renderContents(self): - pass + def setLayerBindings(self, bindings): + self._layer_bindings = bindings - def renderOutput(self): ? ------ + def render(self): self._shader.bind() texture_unit = 0 + for binding in self._layer_bindings: - for render_pass in self._renderer.getRenderPasses(): ? ^^^ ^^ -- - + render_pass = self._renderer.getRenderPass(binding) ? ^^^ ^ +++++++ - self._gl.glActiveTexture(texture_unit) + if not render_pass: + continue + + self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId()) texture_unit += 1 - self._renderer.renderQuad(self._shader) + self._renderer.renderFullScreenQuad(self._shader) ? ++++++++++ for i in range(texture_unit): - self._gl.glActiveTexture(texture_unit) + self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i))) self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0) self._shader.release()
4d1444e2f2a455e691342a82f0e116e210593411
s01/c01.py
s01/c01.py
"""Set 01 - Challenge 01.""" import base64 hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f' '69736f6e6f7573206d757368726f6f6d') b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' def hex2b64(hex_string): """Convert a hex string into a base64 encoded byte string.""" hex_data = bytearray.fromhex(hex_string) # Strip trailing newline return base64.encodebytes(hex_data)[:-1] assert hex2b64(hex_string) == b64_string
"""Set 01 - Challenge 01.""" import binascii hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f' '69736f6e6f7573206d757368726f6f6d') b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' def hex2b64(hex_string): """Convert a hex string into a base64 encoded string.""" return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip() assert hex2b64(hex_string) == b64_string
Revert "Updated function to work on bytes rather than binascii functions."
Revert "Updated function to work on bytes rather than binascii functions." This reverts commit 25176b64aed599059e4b552fbd76c5f4bc28434e.
Python
mit
sornars/matasano-challenges-py
"""Set 01 - Challenge 01.""" - import base64 + + import binascii hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f' '69736f6e6f7573206d757368726f6f6d') - b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' + b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' def hex2b64(hex_string): - """Convert a hex string into a base64 encoded byte string.""" + """Convert a hex string into a base64 encoded string.""" + return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip() - hex_data = bytearray.fromhex(hex_string) - # Strip trailing newline - return base64.encodebytes(hex_data)[:-1] assert hex2b64(hex_string) == b64_string
Revert "Updated function to work on bytes rather than binascii functions."
## Code Before: """Set 01 - Challenge 01.""" import base64 hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f' '69736f6e6f7573206d757368726f6f6d') b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' def hex2b64(hex_string): """Convert a hex string into a base64 encoded byte string.""" hex_data = bytearray.fromhex(hex_string) # Strip trailing newline return base64.encodebytes(hex_data)[:-1] assert hex2b64(hex_string) == b64_string ## Instruction: Revert "Updated function to work on bytes rather than binascii functions." ## Code After: """Set 01 - Challenge 01.""" import binascii hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f' '69736f6e6f7573206d757368726f6f6d') b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' def hex2b64(hex_string): """Convert a hex string into a base64 encoded string.""" return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip() assert hex2b64(hex_string) == b64_string
"""Set 01 - Challenge 01.""" - import base64 + + import binascii hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f' '69736f6e6f7573206d757368726f6f6d') - b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' ? - + b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t' def hex2b64(hex_string): - """Convert a hex string into a base64 encoded byte string.""" ? ----- + """Convert a hex string into a base64 encoded string.""" + return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip() - hex_data = bytearray.fromhex(hex_string) - # Strip trailing newline - return base64.encodebytes(hex_data)[:-1] assert hex2b64(hex_string) == b64_string
6d9ad75ca3ac9a5ed9aac33e56a4809fc7e37f54
gignore/__init__.py
gignore/__init__.py
__version__ = (2014, 10, 0) def get_version(): """ :rtype: str """ return '.'.join(str(i) for i in __version__) class Gignore(object): BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/' name = None file_content = None valid = True def get_base_url(self): """ :rtype: str """ return self.BASE_URL def set_name(self, name): """ :type name: str """ self.name = name def get_name(self): """ :rtype: str """ return self.name def set_file_content(self, file_content): """ :type file_content: str """ self.file_content = file_content def get_file_content(self): """ :rtype: str """ return self.file_content def is_valid(self): """ :rtype: bool """ return self.valid def set_valid(self, valid): """ :type valid: bool """ self.valid = valid
__version__ = (2014, 10, 0) def get_version(): """ :rtype: str """ return '.'.join(str(i) for i in __version__) class Gignore(object): BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/' name = None file_content = None valid = True errors = [] def get_base_url(self): """ :rtype: str """ return self.BASE_URL def set_name(self, name): """ :type name: str """ self.name = name def get_name(self): """ :rtype: str """ return self.name def set_file_content(self, file_content): """ :type file_content: str """ self.file_content = file_content def get_file_content(self): """ :rtype: str """ return self.file_content def is_valid(self): """ :rtype: bool """ return self.valid def set_valid(self, valid): """ :type valid: bool """ self.valid = valid def add_error(self, error_message): """ :type error_message: str """ self.errors.append(error_message) def get_errors(self): """ :rtype: list of str """ return self.errors
Add errors attribute with setter/getter
Add errors attribute with setter/getter
Python
bsd-3-clause
Alir3z4/python-gignore
__version__ = (2014, 10, 0) def get_version(): """ :rtype: str """ return '.'.join(str(i) for i in __version__) class Gignore(object): BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/' name = None file_content = None valid = True + errors = [] def get_base_url(self): """ :rtype: str """ return self.BASE_URL def set_name(self, name): """ :type name: str """ self.name = name def get_name(self): """ :rtype: str """ return self.name def set_file_content(self, file_content): """ :type file_content: str """ self.file_content = file_content def get_file_content(self): """ :rtype: str """ return self.file_content def is_valid(self): """ :rtype: bool """ return self.valid def set_valid(self, valid): """ :type valid: bool """ self.valid = valid + def add_error(self, error_message): + """ + :type error_message: str + """ + self.errors.append(error_message) + + def get_errors(self): + """ + :rtype: list of str + """ + return self.errors +
Add errors attribute with setter/getter
## Code Before: __version__ = (2014, 10, 0) def get_version(): """ :rtype: str """ return '.'.join(str(i) for i in __version__) class Gignore(object): BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/' name = None file_content = None valid = True def get_base_url(self): """ :rtype: str """ return self.BASE_URL def set_name(self, name): """ :type name: str """ self.name = name def get_name(self): """ :rtype: str """ return self.name def set_file_content(self, file_content): """ :type file_content: str """ self.file_content = file_content def get_file_content(self): """ :rtype: str """ return self.file_content def is_valid(self): """ :rtype: bool """ return self.valid def set_valid(self, valid): """ :type valid: bool """ self.valid = valid ## Instruction: Add errors attribute with setter/getter ## Code After: __version__ = (2014, 10, 0) def get_version(): """ :rtype: str """ return '.'.join(str(i) for i in __version__) class Gignore(object): BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/' name = None file_content = None valid = True errors = [] def get_base_url(self): """ :rtype: str """ return self.BASE_URL def set_name(self, name): """ :type name: str """ self.name = name def get_name(self): """ :rtype: str """ return self.name def set_file_content(self, file_content): """ :type file_content: str """ self.file_content = file_content def get_file_content(self): """ :rtype: str """ return self.file_content def is_valid(self): """ :rtype: bool """ return self.valid def set_valid(self, valid): """ :type valid: bool """ self.valid = valid def add_error(self, error_message): """ :type error_message: str """ self.errors.append(error_message) def get_errors(self): """ :rtype: list of str """ return self.errors
__version__ = (2014, 10, 0) def get_version(): """ :rtype: str """ return '.'.join(str(i) for i in __version__) class Gignore(object): BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/' name = None file_content = None valid = True + errors = [] def get_base_url(self): """ :rtype: str """ return self.BASE_URL def set_name(self, name): """ :type name: str """ self.name = name def get_name(self): """ :rtype: str """ return self.name def set_file_content(self, file_content): """ :type file_content: str """ self.file_content = file_content def get_file_content(self): """ :rtype: str """ return self.file_content def is_valid(self): """ :rtype: bool """ return self.valid def set_valid(self, valid): """ :type valid: bool """ self.valid = valid + + def add_error(self, error_message): + """ + :type error_message: str + """ + self.errors.append(error_message) + + def get_errors(self): + """ + :rtype: list of str + """ + return self.errors
fee78440de784bee91669e6c4f1d2c301202e29d
apps/blogs/serializers.py
apps/blogs/serializers.py
from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField from django.contrib.auth.models import User from fluent_contents.rendering import render_placeholder from rest_framework import serializers from .models import BlogPost class BlogPostContentsField(serializers.Field): def to_native(self, obj): request = self.context.get('request', None) contents_html = render_placeholder(request, obj) return contents_html class BlogPostAuthorSerializer(serializers.ModelSerializer): picture = SorlImageField('userprofile.picture', '90x90', crop='center') class Meta: model = User fields = ('first_name', 'last_name', 'picture') class BlogPostDetailSerializer(serializers.ModelSerializer): contents = BlogPostContentsField('contents') author = BlogPostAuthorSerializer() url = SlugHyperlinkedIdentityField(view_name='blogpost-instance') class Meta: model = BlogPost exclude = ('id',) class BlogPostPreviewSerializer(BlogPostDetailSerializer): class Meta: model = BlogPost exclude = ('id',)
from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField from django.contrib.auth.models import User from fluent_contents.rendering import render_placeholder from rest_framework import serializers from .models import BlogPost class BlogPostContentsField(serializers.Field): def to_native(self, obj): request = self.context.get('request', None) contents_html = render_placeholder(request, obj) return contents_html class BlogPostAuthorSerializer(serializers.ModelSerializer): picture = SorlImageField('userprofile.picture', '90x90', crop='center') class Meta: model = User fields = ('first_name', 'last_name', 'picture') class BlogPostDetailSerializer(serializers.ModelSerializer): contents = BlogPostContentsField(source='contents') author = BlogPostAuthorSerializer() url = SlugHyperlinkedIdentityField(view_name='blogpost-instance') main_image = SorlImageField('main_image', '300x200', crop='center') class Meta: model = BlogPost exclude = ('id',) class BlogPostPreviewSerializer(BlogPostDetailSerializer): class Meta: model = BlogPost exclude = ('id',)
Add main_image to BlogPost API response.
Add main_image to BlogPost API response.
Python
bsd-3-clause
onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site
from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField from django.contrib.auth.models import User from fluent_contents.rendering import render_placeholder from rest_framework import serializers from .models import BlogPost class BlogPostContentsField(serializers.Field): def to_native(self, obj): request = self.context.get('request', None) contents_html = render_placeholder(request, obj) return contents_html class BlogPostAuthorSerializer(serializers.ModelSerializer): picture = SorlImageField('userprofile.picture', '90x90', crop='center') class Meta: model = User fields = ('first_name', 'last_name', 'picture') class BlogPostDetailSerializer(serializers.ModelSerializer): - contents = BlogPostContentsField('contents') + contents = BlogPostContentsField(source='contents') author = BlogPostAuthorSerializer() url = SlugHyperlinkedIdentityField(view_name='blogpost-instance') + main_image = SorlImageField('main_image', '300x200', crop='center') class Meta: model = BlogPost exclude = ('id',) class BlogPostPreviewSerializer(BlogPostDetailSerializer): class Meta: model = BlogPost exclude = ('id',)
Add main_image to BlogPost API response.
## Code Before: from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField from django.contrib.auth.models import User from fluent_contents.rendering import render_placeholder from rest_framework import serializers from .models import BlogPost class BlogPostContentsField(serializers.Field): def to_native(self, obj): request = self.context.get('request', None) contents_html = render_placeholder(request, obj) return contents_html class BlogPostAuthorSerializer(serializers.ModelSerializer): picture = SorlImageField('userprofile.picture', '90x90', crop='center') class Meta: model = User fields = ('first_name', 'last_name', 'picture') class BlogPostDetailSerializer(serializers.ModelSerializer): contents = BlogPostContentsField('contents') author = BlogPostAuthorSerializer() url = SlugHyperlinkedIdentityField(view_name='blogpost-instance') class Meta: model = BlogPost exclude = ('id',) class BlogPostPreviewSerializer(BlogPostDetailSerializer): class Meta: model = BlogPost exclude = ('id',) ## Instruction: Add main_image to BlogPost API response. ## Code After: from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField from django.contrib.auth.models import User from fluent_contents.rendering import render_placeholder from rest_framework import serializers from .models import BlogPost class BlogPostContentsField(serializers.Field): def to_native(self, obj): request = self.context.get('request', None) contents_html = render_placeholder(request, obj) return contents_html class BlogPostAuthorSerializer(serializers.ModelSerializer): picture = SorlImageField('userprofile.picture', '90x90', crop='center') class Meta: model = User fields = ('first_name', 'last_name', 'picture') class BlogPostDetailSerializer(serializers.ModelSerializer): contents = BlogPostContentsField(source='contents') author = BlogPostAuthorSerializer() url = SlugHyperlinkedIdentityField(view_name='blogpost-instance') main_image = SorlImageField('main_image', '300x200', crop='center') class Meta: model = BlogPost exclude = ('id',) class BlogPostPreviewSerializer(BlogPostDetailSerializer): class Meta: model = BlogPost exclude = ('id',)
from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField from django.contrib.auth.models import User from fluent_contents.rendering import render_placeholder from rest_framework import serializers from .models import BlogPost class BlogPostContentsField(serializers.Field): def to_native(self, obj): request = self.context.get('request', None) contents_html = render_placeholder(request, obj) return contents_html class BlogPostAuthorSerializer(serializers.ModelSerializer): picture = SorlImageField('userprofile.picture', '90x90', crop='center') class Meta: model = User fields = ('first_name', 'last_name', 'picture') class BlogPostDetailSerializer(serializers.ModelSerializer): - contents = BlogPostContentsField('contents') + contents = BlogPostContentsField(source='contents') ? +++++++ author = BlogPostAuthorSerializer() url = SlugHyperlinkedIdentityField(view_name='blogpost-instance') + main_image = SorlImageField('main_image', '300x200', crop='center') class Meta: model = BlogPost exclude = ('id',) class BlogPostPreviewSerializer(BlogPostDetailSerializer): class Meta: model = BlogPost exclude = ('id',)
f5ba363de4777e2d594261214913f5d480cb04b6
Heuristics/AbstactHeuristic.py
Heuristics/AbstactHeuristic.py
from abc import ABC, abstractmethod import random as random class AbstractHeuristic(ABC): @abstractmethod def calculate(self, solution): pass def calculateCost(self, dataset, solution): cost = 0 i = 0 cost += dataset.getValueXY(0, solution[0]) for i in range(0, len(solution)-1): cost += dataset.getValueXY(solution[i], solution[i + 1]) i += 1 cost += dataset.getValueXY(i,solution[0]) return cost def generateRandomSolution(self, dataset): length = len(dataset[-1])-1 solution = [] for i in range(length): value = random.randint(1, length) while value in solution: value = (value + 1) % length + 1 solution.append(value) return solution
from abc import ABC, abstractmethod import random as random class AbstractHeuristic(ABC): @abstractmethod def calculate(self, solution): pass def calculateCost(self, dataset, solution): cost = 0 i = 0 cost += dataset.getValueXY(0, solution[0]) for i in range(0, len(solution)-1): cost += dataset.getValueXY(solution[i], solution[i + 1]) i += 1 cost += dataset.getValueXY(i,solution[0]) return cost def generateRandomSolution(self, dataset): length = len(dataset[-1])-1 solution = [] for i in range(length): value = random.randint(1, length) while value in solution: value = (value + 1) % (length + 1) if value == 0: value = 1 solution.append(value) return solution
Fix on function generate random solution
Fix on function generate random solution
Python
mit
DiegoReiriz/MetaHeuristics,DiegoReiriz/MetaHeuristics
from abc import ABC, abstractmethod import random as random class AbstractHeuristic(ABC): @abstractmethod def calculate(self, solution): pass def calculateCost(self, dataset, solution): cost = 0 i = 0 cost += dataset.getValueXY(0, solution[0]) for i in range(0, len(solution)-1): cost += dataset.getValueXY(solution[i], solution[i + 1]) i += 1 cost += dataset.getValueXY(i,solution[0]) return cost def generateRandomSolution(self, dataset): length = len(dataset[-1])-1 solution = [] + for i in range(length): value = random.randint(1, length) while value in solution: - value = (value + 1) % length + 1 + value = (value + 1) % (length + 1) + + if value == 0: + value = 1 solution.append(value) return solution
Fix on function generate random solution
## Code Before: from abc import ABC, abstractmethod import random as random class AbstractHeuristic(ABC): @abstractmethod def calculate(self, solution): pass def calculateCost(self, dataset, solution): cost = 0 i = 0 cost += dataset.getValueXY(0, solution[0]) for i in range(0, len(solution)-1): cost += dataset.getValueXY(solution[i], solution[i + 1]) i += 1 cost += dataset.getValueXY(i,solution[0]) return cost def generateRandomSolution(self, dataset): length = len(dataset[-1])-1 solution = [] for i in range(length): value = random.randint(1, length) while value in solution: value = (value + 1) % length + 1 solution.append(value) return solution ## Instruction: Fix on function generate random solution ## Code After: from abc import ABC, abstractmethod import random as random class AbstractHeuristic(ABC): @abstractmethod def calculate(self, solution): pass def calculateCost(self, dataset, solution): cost = 0 i = 0 cost += dataset.getValueXY(0, solution[0]) for i in range(0, len(solution)-1): cost += dataset.getValueXY(solution[i], solution[i + 1]) i += 1 cost += dataset.getValueXY(i,solution[0]) return cost def generateRandomSolution(self, dataset): length = len(dataset[-1])-1 solution = [] for i in range(length): value = random.randint(1, length) while value in solution: value = (value + 1) % (length + 1) if value == 0: value = 1 solution.append(value) return solution
from abc import ABC, abstractmethod import random as random class AbstractHeuristic(ABC): @abstractmethod def calculate(self, solution): pass def calculateCost(self, dataset, solution): cost = 0 i = 0 cost += dataset.getValueXY(0, solution[0]) for i in range(0, len(solution)-1): cost += dataset.getValueXY(solution[i], solution[i + 1]) i += 1 cost += dataset.getValueXY(i,solution[0]) return cost def generateRandomSolution(self, dataset): length = len(dataset[-1])-1 solution = [] + for i in range(length): value = random.randint(1, length) while value in solution: - value = (value + 1) % length + 1 + value = (value + 1) % (length + 1) ? + + + + if value == 0: + value = 1 solution.append(value) return solution
9f6d4d9e82ef575164535a8fb9ea80417458dd6b
website/files/models/dataverse.py
website/files/models/dataverse.py
import requests from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): def touch(self, version=None, revision=None, **kwargs): """Note: Dataverse only has psuedo versions, don't save them""" version = revision or version # Use revision or version resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs)) if resp.status_code != 200: return None data = resp.json() self.name = data['data']['name'] self.materialized_path = data['data']['materialized'] version = FileVersion(identifier=version) version.update_metadata(data['data'], save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files if not data['data']['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass
from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): version_identifier = 'version' def update(self, revision, data): """Note: Dataverse only has psuedo versions, don't save them""" self.name = data['name'] self.materialized_path = data['materialized'] version = FileVersion(identifier=revision) version.update_metadata(data, save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files if not data['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass return version
Move override logic into update rather than touch
Move override logic into update rather than touch
Python
apache-2.0
Johnetordoff/osf.io,mluke93/osf.io,SSJohns/osf.io,chrisseto/osf.io,hmoco/osf.io,caseyrygt/osf.io,GageGaskins/osf.io,acshi/osf.io,alexschiller/osf.io,caseyrollins/osf.io,ZobairAlijan/osf.io,wearpants/osf.io,GageGaskins/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,SSJohns/osf.io,alexschiller/osf.io,adlius/osf.io,samchrisinger/osf.io,sloria/osf.io,hmoco/osf.io,erinspace/osf.io,DanielSBrown/osf.io,samanehsan/osf.io,haoyuchen1992/osf.io,Ghalko/osf.io,crcresearch/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,danielneis/osf.io,doublebits/osf.io,zamattiac/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,KAsante95/osf.io,wearpants/osf.io,caneruguz/osf.io,petermalcolm/osf.io,pattisdr/osf.io,mattclark/osf.io,brianjgeiger/osf.io,cwisecarver/osf.io,samchrisinger/osf.io,billyhunt/osf.io,asanfilippo7/osf.io,chrisseto/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,rdhyee/osf.io,abought/osf.io,RomanZWang/osf.io,mluke93/osf.io,monikagrabowska/osf.io,TomHeatwole/osf.io,mfraezz/osf.io,cosenal/osf.io,saradbowman/osf.io,felliott/osf.io,aaxelb/osf.io,jnayak1/osf.io,cwisecarver/osf.io,danielneis/osf.io,kch8qx/osf.io,GageGaskins/osf.io,aaxelb/osf.io,aaxelb/osf.io,TomHeatwole/osf.io,crcresearch/osf.io,ZobairAlijan/osf.io,zamattiac/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,brianjgeiger/osf.io,zachjanicki/osf.io,rdhyee/osf.io,samchrisinger/osf.io,kch8qx/osf.io,Johnetordoff/osf.io,chennan47/osf.io,TomBaxter/osf.io,jnayak1/osf.io,felliott/osf.io,binoculars/osf.io,acshi/osf.io,Nesiehr/osf.io,kch8qx/osf.io,Nesiehr/osf.io,Ghalko/osf.io,mfraezz/osf.io,GageGaskins/osf.io,petermalcolm/osf.io,arpitar/osf.io,samanehsan/osf.io,abought/osf.io,laurenrevere/osf.io,emetsger/osf.io,petermalcolm/osf.io,mattclark/osf.io,RomanZWang/osf.io,jnayak1/osf.io,kwierman/osf.io,leb2dg/osf.io,doublebits/osf.io,cslzchen/osf.io,zachjanicki/osf.io,njantrania/osf.io,chennan47/osf.io,kwierman/osf.io,brianjgeiger/osf.io,hmoco/osf.io,GageGaskins/osf.io,zamattiac/osf.io,emetsger/osf.io,caseyrollins/osf.io,binoculars/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,brandonPurvis/osf.io,acshi/osf.io,rdhyee/osf.io,petermalcolm/osf.io,cosenal/osf.io,asanfilippo7/osf.io,adlius/osf.io,mluke93/osf.io,adlius/osf.io,mluo613/osf.io,kch8qx/osf.io,Ghalko/osf.io,haoyuchen1992/osf.io,HalcyonChimera/osf.io,abought/osf.io,pattisdr/osf.io,haoyuchen1992/osf.io,brandonPurvis/osf.io,mluo613/osf.io,saradbowman/osf.io,mattclark/osf.io,jnayak1/osf.io,felliott/osf.io,arpitar/osf.io,caseyrollins/osf.io,amyshi188/osf.io,acshi/osf.io,asanfilippo7/osf.io,emetsger/osf.io,RomanZWang/osf.io,njantrania/osf.io,sloria/osf.io,SSJohns/osf.io,cwisecarver/osf.io,leb2dg/osf.io,leb2dg/osf.io,felliott/osf.io,CenterForOpenScience/osf.io,TomBaxter/osf.io,billyhunt/osf.io,adlius/osf.io,ticklemepierce/osf.io,chrisseto/osf.io,haoyuchen1992/osf.io,erinspace/osf.io,Ghalko/osf.io,acshi/osf.io,doublebits/osf.io,wearpants/osf.io,alexschiller/osf.io,icereval/osf.io,amyshi188/osf.io,doublebits/osf.io,KAsante95/osf.io,alexschiller/osf.io,wearpants/osf.io,samchrisinger/osf.io,njantrania/osf.io,njantrania/osf.io,cwisecarver/osf.io,KAsante95/osf.io,zamattiac/osf.io,kch8qx/osf.io,TomBaxter/osf.io,ZobairAlijan/osf.io,HalcyonChimera/osf.io,KAsante95/osf.io,DanielSBrown/osf.io,aaxelb/osf.io,rdhyee/osf.io,mluke93/osf.io,mfraezz/osf.io,zachjanicki/osf.io,cslzchen/osf.io,hmoco/osf.io,cosenal/osf.io,zachjanicki/osf.io,chennan47/osf.io,doublebits/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,billyhunt/osf.io,KAsante95/osf.io,brandonPurvis/osf.io,brianjgeiger/osf.io,samanehsan/osf.io,binoculars/osf.io,monikagrabowska/osf.io,crcresearch/osf.io,danielneis/osf.io,laurenrevere/osf.io,arpitar/osf.io,kwierman/osf.io,billyhunt/osf.io,kwierman/osf.io,danielneis/osf.io,ticklemepierce/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,ZobairAlijan/osf.io,cslzchen/osf.io,caseyrygt/osf.io,laurenrevere/osf.io,Nesiehr/osf.io,mluo613/osf.io,TomHeatwole/osf.io,amyshi188/osf.io,caseyrygt/osf.io,mfraezz/osf.io,SSJohns/osf.io,erinspace/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,arpitar/osf.io,baylee-d/osf.io,RomanZWang/osf.io,billyhunt/osf.io,mluo613/osf.io,mluo613/osf.io,amyshi188/osf.io,abought/osf.io,pattisdr/osf.io,ticklemepierce/osf.io,chrisseto/osf.io,emetsger/osf.io,caseyrygt/osf.io,cosenal/osf.io,RomanZWang/osf.io,HalcyonChimera/osf.io,sloria/osf.io,icereval/osf.io,alexschiller/osf.io,samanehsan/osf.io,icereval/osf.io,TomHeatwole/osf.io,ticklemepierce/osf.io,DanielSBrown/osf.io,asanfilippo7/osf.io,CenterForOpenScience/osf.io
- import requests - from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): + version_identifier = 'version' - def touch(self, version=None, revision=None, **kwargs): + def update(self, revision, data): """Note: Dataverse only has psuedo versions, don't save them""" - version = revision or version # Use revision or version + self.name = data['name'] + self.materialized_path = data['materialized'] - resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs)) - if resp.status_code != 200: - return None - - data = resp.json() - self.name = data['data']['name'] - self.materialized_path = data['data']['materialized'] - - version = FileVersion(identifier=version) + version = FileVersion(identifier=revision) - version.update_metadata(data['data'], save=False) + version.update_metadata(data, save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files - if not data['data']['extra']['hasPublishedVersion']: + if not data['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass + return version
Move override logic into update rather than touch
## Code Before: import requests from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): def touch(self, version=None, revision=None, **kwargs): """Note: Dataverse only has psuedo versions, don't save them""" version = revision or version # Use revision or version resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs)) if resp.status_code != 200: return None data = resp.json() self.name = data['data']['name'] self.materialized_path = data['data']['materialized'] version = FileVersion(identifier=version) version.update_metadata(data['data'], save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files if not data['data']['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass ## Instruction: Move override logic into update rather than touch ## Code After: from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): version_identifier = 'version' def update(self, revision, data): """Note: Dataverse only has psuedo versions, don't save them""" self.name = data['name'] self.materialized_path = data['materialized'] version = FileVersion(identifier=revision) version.update_metadata(data, save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files if not data['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass return version
- import requests - from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): + version_identifier = 'version' - def touch(self, version=None, revision=None, **kwargs): + def update(self, revision, data): """Note: Dataverse only has psuedo versions, don't save them""" - version = revision or version # Use revision or version + self.name = data['name'] + self.materialized_path = data['materialized'] - resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs)) - if resp.status_code != 200: - return None - - data = resp.json() - self.name = data['data']['name'] - self.materialized_path = data['data']['materialized'] - - version = FileVersion(identifier=version) ? ^^ + version = FileVersion(identifier=revision) ? ++ ^ - version.update_metadata(data['data'], save=False) ? -------- + version.update_metadata(data, save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files - if not data['data']['extra']['hasPublishedVersion']: ? -------- + if not data['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass + return version
02f5db5fdb46684b60a9b5e9125da228a927c2c3
mrbelvedereci/build/cumulusci/config.py
mrbelvedereci/build/cumulusci/config.py
from cumulusci.core.config import YamlGlobalConfig from cumulusci.core.config import YamlProjectConfig class MrbelvedereProjectConfig(YamlProjectConfig): def __init__(self, global_config_obj, build_flow): super(MrbelvedereProjectConfig, self).__init__(global_config_obj) self.build_flow = build_flow @property def config_project_local_path(self): """ mrbelvedere never uses the local path """ return @property def repo_root(self): return self.build_flow.build_dir @property def repo_name(self): return self.build_flow.build.repo.name @property def repo_url(self): return self.build_flow.build.repo.url @property def repo_owner(self): return self.build_flow.build.repo.url.split('/')[-2] @property def repo_branch(self): return self.build_flow.build.branch.name @property def repo_commit(self): return self.build_flow.build.commit class MrbelvedereGlobalConfig(YamlGlobalConfig): project_config_class = MrbelvedereProjectConfig def get_project_config(self, build_flow): return self.project_config_class(self, build_flow)
from cumulusci.core.config import YamlGlobalConfig from cumulusci.core.config import YamlProjectConfig class MrbelvedereProjectConfig(YamlProjectConfig): def __init__(self, global_config_obj, build_flow): self.build_flow = build_flow super(MrbelvedereProjectConfig, self).__init__(global_config_obj) @property def config_project_local_path(self): """ mrbelvedere never uses the local path """ return @property def repo_root(self): return self.build_flow.build_dir @property def repo_name(self): return self.build_flow.build.repo.name @property def repo_url(self): return self.build_flow.build.repo.url @property def repo_owner(self): return self.build_flow.build.repo.url.split('/')[-2] @property def repo_branch(self): return self.build_flow.build.branch.name @property def repo_commit(self): return self.build_flow.build.commit class MrbelvedereGlobalConfig(YamlGlobalConfig): project_config_class = MrbelvedereProjectConfig def get_project_config(self, build_flow): return self.project_config_class(self, build_flow)
Set self.build_flow before calling the super __init__ method
Set self.build_flow before calling the super __init__ method
Python
bsd-3-clause
SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci
from cumulusci.core.config import YamlGlobalConfig from cumulusci.core.config import YamlProjectConfig class MrbelvedereProjectConfig(YamlProjectConfig): def __init__(self, global_config_obj, build_flow): + self.build_flow = build_flow super(MrbelvedereProjectConfig, self).__init__(global_config_obj) - self.build_flow = build_flow @property def config_project_local_path(self): """ mrbelvedere never uses the local path """ return @property def repo_root(self): return self.build_flow.build_dir @property def repo_name(self): return self.build_flow.build.repo.name @property def repo_url(self): return self.build_flow.build.repo.url @property def repo_owner(self): return self.build_flow.build.repo.url.split('/')[-2] @property def repo_branch(self): return self.build_flow.build.branch.name @property def repo_commit(self): return self.build_flow.build.commit class MrbelvedereGlobalConfig(YamlGlobalConfig): project_config_class = MrbelvedereProjectConfig def get_project_config(self, build_flow): return self.project_config_class(self, build_flow)
Set self.build_flow before calling the super __init__ method
## Code Before: from cumulusci.core.config import YamlGlobalConfig from cumulusci.core.config import YamlProjectConfig class MrbelvedereProjectConfig(YamlProjectConfig): def __init__(self, global_config_obj, build_flow): super(MrbelvedereProjectConfig, self).__init__(global_config_obj) self.build_flow = build_flow @property def config_project_local_path(self): """ mrbelvedere never uses the local path """ return @property def repo_root(self): return self.build_flow.build_dir @property def repo_name(self): return self.build_flow.build.repo.name @property def repo_url(self): return self.build_flow.build.repo.url @property def repo_owner(self): return self.build_flow.build.repo.url.split('/')[-2] @property def repo_branch(self): return self.build_flow.build.branch.name @property def repo_commit(self): return self.build_flow.build.commit class MrbelvedereGlobalConfig(YamlGlobalConfig): project_config_class = MrbelvedereProjectConfig def get_project_config(self, build_flow): return self.project_config_class(self, build_flow) ## Instruction: Set self.build_flow before calling the super __init__ method ## Code After: from cumulusci.core.config import YamlGlobalConfig from cumulusci.core.config import YamlProjectConfig class MrbelvedereProjectConfig(YamlProjectConfig): def __init__(self, global_config_obj, build_flow): self.build_flow = build_flow super(MrbelvedereProjectConfig, self).__init__(global_config_obj) @property def config_project_local_path(self): """ mrbelvedere never uses the local path """ return @property def repo_root(self): return self.build_flow.build_dir @property def repo_name(self): return self.build_flow.build.repo.name @property def repo_url(self): return self.build_flow.build.repo.url @property def repo_owner(self): return self.build_flow.build.repo.url.split('/')[-2] @property def repo_branch(self): return self.build_flow.build.branch.name @property def repo_commit(self): return self.build_flow.build.commit class MrbelvedereGlobalConfig(YamlGlobalConfig): project_config_class = MrbelvedereProjectConfig def get_project_config(self, build_flow): return self.project_config_class(self, build_flow)
from cumulusci.core.config import YamlGlobalConfig from cumulusci.core.config import YamlProjectConfig class MrbelvedereProjectConfig(YamlProjectConfig): def __init__(self, global_config_obj, build_flow): + self.build_flow = build_flow super(MrbelvedereProjectConfig, self).__init__(global_config_obj) - self.build_flow = build_flow @property def config_project_local_path(self): """ mrbelvedere never uses the local path """ return @property def repo_root(self): return self.build_flow.build_dir @property def repo_name(self): return self.build_flow.build.repo.name @property def repo_url(self): return self.build_flow.build.repo.url @property def repo_owner(self): return self.build_flow.build.repo.url.split('/')[-2] @property def repo_branch(self): return self.build_flow.build.branch.name @property def repo_commit(self): return self.build_flow.build.commit class MrbelvedereGlobalConfig(YamlGlobalConfig): project_config_class = MrbelvedereProjectConfig def get_project_config(self, build_flow): return self.project_config_class(self, build_flow)
564ae1eb637ec509f37ade93d4079117cc73fd58
lab_assistant/storage/__init__.py
lab_assistant/storage/__init__.py
from copy import deepcopy from simpleflake import simpleflake from lab_assistant import conf, utils __all__ = [ 'get_storage', 'store', 'retrieve', 'retrieve_all', 'clear', ] def get_storage(path=None, name='Experiment', **opts): if not path: path = conf.storage['path'] _opts = deepcopy(conf.storage.get('options', {})) _opts.update(opts) opts = _opts if path in get_storage._cache: return get_storage._cache[path] Storage = utils.import_path(path) get_storage._cache[path] = Storage(name, **opts) return get_storage._cache[path] get_storage._cache = {} def store(result, storage=None): storage = storage or get_storage(name=result.experiment.name) key = simpleflake() storage.set(key, result) return key def retrieve(key, storage=None): storage = storage or get_storage() return storage.get(key) def retrieve_all(storage=None): return (storage or get_storage()).list() def remove(key, storage=None): (storage or get_storage()).remove(key) def clear(storage=None): return (storage or get_storage()).clear()
from copy import deepcopy from collections import defaultdict from simpleflake import simpleflake from lab_assistant import conf, utils __all__ = [ 'get_storage', 'store', 'retrieve', 'retrieve_all', 'clear', ] def get_storage(path=None, name='Experiment', **opts): if not path: path = conf.storage['path'] _opts = deepcopy(conf.storage.get('options', {})) _opts.update(opts) opts = _opts if path in get_storage._cache: if name in get_storage._cache[path]: return get_storage._cache[path][name] Storage = utils.import_path(path) get_storage._cache[path].update({ name: Storage(name, **opts) }) return get_storage._cache[path][name] get_storage._cache = defaultdict(dict) def store(result, storage=None): storage = storage or get_storage(name=result.experiment.name) key = simpleflake() storage.set(key, result) return key def retrieve(key, storage=None): storage = storage or get_storage() return storage.get(key) def retrieve_all(storage=None): return (storage or get_storage()).list() def remove(key, storage=None): (storage or get_storage()).remove(key) def clear(storage=None): return (storage or get_storage()).clear()
Fix get_storage cache to hold separate entries for each experiment key
Fix get_storage cache to hold separate entries for each experiment key
Python
mit
joealcorn/lab_assistant
from copy import deepcopy + from collections import defaultdict from simpleflake import simpleflake from lab_assistant import conf, utils __all__ = [ 'get_storage', 'store', 'retrieve', 'retrieve_all', 'clear', ] def get_storage(path=None, name='Experiment', **opts): if not path: path = conf.storage['path'] _opts = deepcopy(conf.storage.get('options', {})) _opts.update(opts) opts = _opts if path in get_storage._cache: + if name in get_storage._cache[path]: - return get_storage._cache[path] + return get_storage._cache[path][name] Storage = utils.import_path(path) - get_storage._cache[path] = Storage(name, **opts) + get_storage._cache[path].update({ + name: Storage(name, **opts) + }) - return get_storage._cache[path] + return get_storage._cache[path][name] - get_storage._cache = {} + get_storage._cache = defaultdict(dict) def store(result, storage=None): storage = storage or get_storage(name=result.experiment.name) key = simpleflake() storage.set(key, result) return key def retrieve(key, storage=None): storage = storage or get_storage() return storage.get(key) def retrieve_all(storage=None): return (storage or get_storage()).list() def remove(key, storage=None): (storage or get_storage()).remove(key) def clear(storage=None): return (storage or get_storage()).clear()
Fix get_storage cache to hold separate entries for each experiment key
## Code Before: from copy import deepcopy from simpleflake import simpleflake from lab_assistant import conf, utils __all__ = [ 'get_storage', 'store', 'retrieve', 'retrieve_all', 'clear', ] def get_storage(path=None, name='Experiment', **opts): if not path: path = conf.storage['path'] _opts = deepcopy(conf.storage.get('options', {})) _opts.update(opts) opts = _opts if path in get_storage._cache: return get_storage._cache[path] Storage = utils.import_path(path) get_storage._cache[path] = Storage(name, **opts) return get_storage._cache[path] get_storage._cache = {} def store(result, storage=None): storage = storage or get_storage(name=result.experiment.name) key = simpleflake() storage.set(key, result) return key def retrieve(key, storage=None): storage = storage or get_storage() return storage.get(key) def retrieve_all(storage=None): return (storage or get_storage()).list() def remove(key, storage=None): (storage or get_storage()).remove(key) def clear(storage=None): return (storage or get_storage()).clear() ## Instruction: Fix get_storage cache to hold separate entries for each experiment key ## Code After: from copy import deepcopy from collections import defaultdict from simpleflake import simpleflake from lab_assistant import conf, utils __all__ = [ 'get_storage', 'store', 'retrieve', 'retrieve_all', 'clear', ] def get_storage(path=None, name='Experiment', **opts): if not path: path = conf.storage['path'] _opts = deepcopy(conf.storage.get('options', {})) _opts.update(opts) opts = _opts if path in get_storage._cache: if name in get_storage._cache[path]: return get_storage._cache[path][name] Storage = utils.import_path(path) get_storage._cache[path].update({ name: Storage(name, **opts) }) return get_storage._cache[path][name] get_storage._cache = defaultdict(dict) def store(result, storage=None): storage = storage or get_storage(name=result.experiment.name) key = simpleflake() storage.set(key, result) return key def retrieve(key, storage=None): storage = storage or get_storage() return storage.get(key) def retrieve_all(storage=None): return (storage or get_storage()).list() def remove(key, storage=None): (storage or get_storage()).remove(key) def clear(storage=None): return (storage or get_storage()).clear()
from copy import deepcopy + from collections import defaultdict from simpleflake import simpleflake from lab_assistant import conf, utils __all__ = [ 'get_storage', 'store', 'retrieve', 'retrieve_all', 'clear', ] def get_storage(path=None, name='Experiment', **opts): if not path: path = conf.storage['path'] _opts = deepcopy(conf.storage.get('options', {})) _opts.update(opts) opts = _opts if path in get_storage._cache: + if name in get_storage._cache[path]: - return get_storage._cache[path] + return get_storage._cache[path][name] ? ++++ ++++++ Storage = utils.import_path(path) - get_storage._cache[path] = Storage(name, **opts) + get_storage._cache[path].update({ + name: Storage(name, **opts) + }) - return get_storage._cache[path] + return get_storage._cache[path][name] ? ++++++ - get_storage._cache = {} + get_storage._cache = defaultdict(dict) def store(result, storage=None): storage = storage or get_storage(name=result.experiment.name) key = simpleflake() storage.set(key, result) return key def retrieve(key, storage=None): storage = storage or get_storage() return storage.get(key) def retrieve_all(storage=None): return (storage or get_storage()).list() def remove(key, storage=None): (storage or get_storage()).remove(key) def clear(storage=None): return (storage or get_storage()).clear()
fda50fb75b0b0e1d571c825e0a364573b93461bc
mbuild/__init__.py
mbuild/__init__.py
from mbuild.box import Box from mbuild.coarse_graining import coarse_grain from mbuild.coordinate_transform import * from mbuild.compound import * from mbuild.pattern import * from mbuild.packing import * from mbuild.port import Port from mbuild.recipes import * from mbuild.lattice import Lattice from mbuild.recipes import recipes from mbuild.version import version
from mbuild.box import Box from mbuild.coarse_graining import coarse_grain from mbuild.coordinate_transform import * from mbuild.compound import * from mbuild.pattern import * from mbuild.packing import * from mbuild.port import Port from mbuild.lattice import Lattice from mbuild.recipes import recipes from mbuild.version import version
Remove a troubling import *
Remove a troubling import *
Python
mit
iModels/mbuild,iModels/mbuild
from mbuild.box import Box from mbuild.coarse_graining import coarse_grain from mbuild.coordinate_transform import * from mbuild.compound import * from mbuild.pattern import * from mbuild.packing import * from mbuild.port import Port - from mbuild.recipes import * from mbuild.lattice import Lattice from mbuild.recipes import recipes from mbuild.version import version
Remove a troubling import *
## Code Before: from mbuild.box import Box from mbuild.coarse_graining import coarse_grain from mbuild.coordinate_transform import * from mbuild.compound import * from mbuild.pattern import * from mbuild.packing import * from mbuild.port import Port from mbuild.recipes import * from mbuild.lattice import Lattice from mbuild.recipes import recipes from mbuild.version import version ## Instruction: Remove a troubling import * ## Code After: from mbuild.box import Box from mbuild.coarse_graining import coarse_grain from mbuild.coordinate_transform import * from mbuild.compound import * from mbuild.pattern import * from mbuild.packing import * from mbuild.port import Port from mbuild.lattice import Lattice from mbuild.recipes import recipes from mbuild.version import version
from mbuild.box import Box from mbuild.coarse_graining import coarse_grain from mbuild.coordinate_transform import * from mbuild.compound import * from mbuild.pattern import * from mbuild.packing import * from mbuild.port import Port - from mbuild.recipes import * from mbuild.lattice import Lattice from mbuild.recipes import recipes from mbuild.version import version
93cefdc2c309ed0b81fe4ec7d49c0c8bead783a9
lib/path_utils.py
lib/path_utils.py
"""Originally from funfactory (funfactory/path_utils.py) on a380a54""" import os from os.path import abspath, dirname def path(*a): return os.path.join(ROOT, *a) def import_mod_by_name(target): # stolen from mock :) components = target.split('.') import_path = components.pop(0) thing = __import__(import_path) for comp in components: import_path += ".%s" % comp thing = _dot_lookup(thing, comp, import_path) return thing def _dot_lookup(thing, comp, import_path): try: return getattr(thing, comp) except AttributeError: __import__(import_path) return getattr(thing, comp) ROOT = dirname(abspath(__name__))
"""Originally from funfactory (funfactory/path_utils.py) on a380a54""" import os from os.path import abspath, dirname def path(*a): return os.path.join(ROOT, *a) def import_mod_by_name(target): # stolen from mock :) components = target.split('.') import_path = components.pop(0) thing = __import__(import_path) for comp in components: import_path += ".%s" % comp thing = _dot_lookup(thing, comp, import_path) return thing def _dot_lookup(thing, comp, import_path): try: return getattr(thing, comp) except AttributeError: __import__(import_path) return getattr(thing, comp) ROOT = dirname(dirname(abspath(__file__)))
Use __file__ instead of __name__
Use __file__ instead of __name__
Python
bsd-3-clause
akeym/cyder,zeeman/cyder,murrown/cyder,murrown/cyder,OSU-Net/cyder,drkitty/cyder,OSU-Net/cyder,akeym/cyder,akeym/cyder,OSU-Net/cyder,akeym/cyder,drkitty/cyder,murrown/cyder,drkitty/cyder,zeeman/cyder,OSU-Net/cyder,zeeman/cyder,murrown/cyder,drkitty/cyder,zeeman/cyder
"""Originally from funfactory (funfactory/path_utils.py) on a380a54""" import os from os.path import abspath, dirname def path(*a): return os.path.join(ROOT, *a) def import_mod_by_name(target): # stolen from mock :) components = target.split('.') import_path = components.pop(0) thing = __import__(import_path) for comp in components: import_path += ".%s" % comp thing = _dot_lookup(thing, comp, import_path) return thing def _dot_lookup(thing, comp, import_path): try: return getattr(thing, comp) except AttributeError: __import__(import_path) return getattr(thing, comp) - ROOT = dirname(abspath(__name__)) + ROOT = dirname(dirname(abspath(__file__)))
Use __file__ instead of __name__
## Code Before: """Originally from funfactory (funfactory/path_utils.py) on a380a54""" import os from os.path import abspath, dirname def path(*a): return os.path.join(ROOT, *a) def import_mod_by_name(target): # stolen from mock :) components = target.split('.') import_path = components.pop(0) thing = __import__(import_path) for comp in components: import_path += ".%s" % comp thing = _dot_lookup(thing, comp, import_path) return thing def _dot_lookup(thing, comp, import_path): try: return getattr(thing, comp) except AttributeError: __import__(import_path) return getattr(thing, comp) ROOT = dirname(abspath(__name__)) ## Instruction: Use __file__ instead of __name__ ## Code After: """Originally from funfactory (funfactory/path_utils.py) on a380a54""" import os from os.path import abspath, dirname def path(*a): return os.path.join(ROOT, *a) def import_mod_by_name(target): # stolen from mock :) components = target.split('.') import_path = components.pop(0) thing = __import__(import_path) for comp in components: import_path += ".%s" % comp thing = _dot_lookup(thing, comp, import_path) return thing def _dot_lookup(thing, comp, import_path): try: return getattr(thing, comp) except AttributeError: __import__(import_path) return getattr(thing, comp) ROOT = dirname(dirname(abspath(__file__)))
"""Originally from funfactory (funfactory/path_utils.py) on a380a54""" import os from os.path import abspath, dirname def path(*a): return os.path.join(ROOT, *a) def import_mod_by_name(target): # stolen from mock :) components = target.split('.') import_path = components.pop(0) thing = __import__(import_path) for comp in components: import_path += ".%s" % comp thing = _dot_lookup(thing, comp, import_path) return thing def _dot_lookup(thing, comp, import_path): try: return getattr(thing, comp) except AttributeError: __import__(import_path) return getattr(thing, comp) - ROOT = dirname(abspath(__name__)) ? ^^^ + ROOT = dirname(dirname(abspath(__file__))) ? ++++++++ ^^^ +
3fb56e434182e5b28dcad0c547b0326ebe5be352
main.py
main.py
from createCollection import createCollectionFile from ObjectFactories.ItemFactory import ItemFactory from DataObjects.Collection import Collection import datetime, json, os.path, argparse CONST_COLLECTIONS_NAME = 'collections' def generateArgumentsFromParser(): parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.") parser.add_argument('--action', dest='action', required=True) parser.add_argument('--user', dest='username', required=True) parser.add_argument('--name', dest='collectionName', required=True) parser.add_argument('--type', dest='collectionType', required=False) return parser.parse_args() def generateFileName(username, collectionName): return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat" def generateNewCollection(username, collectionType, collectionName): return Collection(username, collectionType, collectionName, []) def main(): arguments = generateArgumentsFromParser() collectionFileName = generateFileName(arguments.username, arguments.collectionName) if arguments.action.lower() == "create": createCollectionFile(arguments.username, arguments.collectionName) collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) collectionFile = open(collectionFileName, 'w') collectionFile.write(collection.toJSON()) collectionFile.close() elif arguments.action.lower() == "update": return None if __name__ == '__main__': main()
from createCollection import createCollectionFile from ObjectFactories.ItemFactory import ItemFactory from DataObjects.Collection import Collection import datetime, json, os.path, argparse CONST_COLLECTIONS_NAME = 'collections' def generateArgumentsFromParser(): parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.") parser.add_argument('--action', dest='action', required=True) parser.add_argument('--user', dest='username', required=True) parser.add_argument('--name', dest='collectionName', required=True) parser.add_argument('--type', dest='collectionType', required=False) return parser.parse_args() def generateFileName(username, collectionName): return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat" def generateNewCollection(username, collectionType, collectionName): return Collection(username, collectionType, collectionName, []) def writeCollectionToFile(collectionFileName, arguments): collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) collectionFile = open(collectionFileName, 'w') collectionFile.write(collection.toJSON()) collectionFile.close() def main(): arguments = generateArgumentsFromParser() collectionFileName = generateFileName(arguments.username, arguments.collectionName) if arguments.action.lower() == "create": createCollectionFile(arguments.username, arguments.collectionName) writeCollectionToFile(collectionFileName, arguments) elif arguments.action.lower() == "update": return None if __name__ == '__main__': main()
Refactor create action into function
Refactor create action into function
Python
apache-2.0
AmosGarner/PyInventory
from createCollection import createCollectionFile from ObjectFactories.ItemFactory import ItemFactory from DataObjects.Collection import Collection import datetime, json, os.path, argparse CONST_COLLECTIONS_NAME = 'collections' def generateArgumentsFromParser(): parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.") parser.add_argument('--action', dest='action', required=True) parser.add_argument('--user', dest='username', required=True) parser.add_argument('--name', dest='collectionName', required=True) parser.add_argument('--type', dest='collectionType', required=False) return parser.parse_args() def generateFileName(username, collectionName): return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat" def generateNewCollection(username, collectionType, collectionName): return Collection(username, collectionType, collectionName, []) + def writeCollectionToFile(collectionFileName, arguments): + collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) + collectionFile = open(collectionFileName, 'w') + collectionFile.write(collection.toJSON()) + collectionFile.close() + def main(): arguments = generateArgumentsFromParser() collectionFileName = generateFileName(arguments.username, arguments.collectionName) if arguments.action.lower() == "create": createCollectionFile(arguments.username, arguments.collectionName) - collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) - collectionFile = open(collectionFileName, 'w') + writeCollectionToFile(collectionFileName, arguments) - collectionFile.write(collection.toJSON()) - collectionFile.close() elif arguments.action.lower() == "update": return None if __name__ == '__main__': main()
Refactor create action into function
## Code Before: from createCollection import createCollectionFile from ObjectFactories.ItemFactory import ItemFactory from DataObjects.Collection import Collection import datetime, json, os.path, argparse CONST_COLLECTIONS_NAME = 'collections' def generateArgumentsFromParser(): parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.") parser.add_argument('--action', dest='action', required=True) parser.add_argument('--user', dest='username', required=True) parser.add_argument('--name', dest='collectionName', required=True) parser.add_argument('--type', dest='collectionType', required=False) return parser.parse_args() def generateFileName(username, collectionName): return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat" def generateNewCollection(username, collectionType, collectionName): return Collection(username, collectionType, collectionName, []) def main(): arguments = generateArgumentsFromParser() collectionFileName = generateFileName(arguments.username, arguments.collectionName) if arguments.action.lower() == "create": createCollectionFile(arguments.username, arguments.collectionName) collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) collectionFile = open(collectionFileName, 'w') collectionFile.write(collection.toJSON()) collectionFile.close() elif arguments.action.lower() == "update": return None if __name__ == '__main__': main() ## Instruction: Refactor create action into function ## Code After: from createCollection import createCollectionFile from ObjectFactories.ItemFactory import ItemFactory from DataObjects.Collection import Collection import datetime, json, os.path, argparse CONST_COLLECTIONS_NAME = 'collections' def generateArgumentsFromParser(): parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.") parser.add_argument('--action', dest='action', required=True) parser.add_argument('--user', dest='username', required=True) parser.add_argument('--name', dest='collectionName', required=True) parser.add_argument('--type', dest='collectionType', required=False) return parser.parse_args() def generateFileName(username, collectionName): return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat" def generateNewCollection(username, collectionType, collectionName): return Collection(username, collectionType, collectionName, []) def writeCollectionToFile(collectionFileName, arguments): collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) collectionFile = open(collectionFileName, 'w') collectionFile.write(collection.toJSON()) collectionFile.close() def main(): arguments = generateArgumentsFromParser() collectionFileName = generateFileName(arguments.username, arguments.collectionName) if arguments.action.lower() == "create": createCollectionFile(arguments.username, arguments.collectionName) writeCollectionToFile(collectionFileName, arguments) elif arguments.action.lower() == "update": return None if __name__ == '__main__': main()
from createCollection import createCollectionFile from ObjectFactories.ItemFactory import ItemFactory from DataObjects.Collection import Collection import datetime, json, os.path, argparse CONST_COLLECTIONS_NAME = 'collections' def generateArgumentsFromParser(): parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.") parser.add_argument('--action', dest='action', required=True) parser.add_argument('--user', dest='username', required=True) parser.add_argument('--name', dest='collectionName', required=True) parser.add_argument('--type', dest='collectionType', required=False) return parser.parse_args() def generateFileName(username, collectionName): return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat" def generateNewCollection(username, collectionType, collectionName): return Collection(username, collectionType, collectionName, []) + def writeCollectionToFile(collectionFileName, arguments): + collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) + collectionFile = open(collectionFileName, 'w') + collectionFile.write(collection.toJSON()) + collectionFile.close() + def main(): arguments = generateArgumentsFromParser() collectionFileName = generateFileName(arguments.username, arguments.collectionName) if arguments.action.lower() == "create": createCollectionFile(arguments.username, arguments.collectionName) - collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName) - collectionFile = open(collectionFileName, 'w') ? ^ ------- ^^^ + writeCollectionToFile(collectionFileName, arguments) ? ^^^^^^ ++ ^^^^^^^^^ - collectionFile.write(collection.toJSON()) - collectionFile.close() elif arguments.action.lower() == "update": return None if __name__ == '__main__': main()
22f9fc8a56882f0595d051cb8c5d20fd97091e8c
custom/opm/tests/test_snapshot.py
custom/opm/tests/test_snapshot.py
from datetime import date from unittest import TestCase from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows def test_basic_CMR(self): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
from datetime import date from unittest import TestCase from mock import patch from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows @patch.object(CommCareUser, 'by_domain', return_value=[]) def test_basic_CMR(self, user_mock): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
Fix for test (add mock for CommCareUser)
Fix for test (add mock for CommCareUser)
Python
bsd-3-clause
puttarajubr/commcare-hq,puttarajubr/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
from datetime import date from unittest import TestCase - + from mock import patch + from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows + @patch.object(CommCareUser, 'by_domain', return_value=[]) - def test_basic_CMR(self): + def test_basic_CMR(self, user_mock): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
Fix for test (add mock for CommCareUser)
## Code Before: from datetime import date from unittest import TestCase from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows def test_basic_CMR(self): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows ## Instruction: Fix for test (add mock for CommCareUser) ## Code After: from datetime import date from unittest import TestCase from mock import patch from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows @patch.object(CommCareUser, 'by_domain', return_value=[]) def test_basic_CMR(self, user_mock): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
from datetime import date from unittest import TestCase - + from mock import patch + from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows + @patch.object(CommCareUser, 'by_domain', return_value=[]) - def test_basic_CMR(self): + def test_basic_CMR(self, user_mock): ? +++++++++++ report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
b3fa14e85182d1b0efa47452de51d93a66c63503
tests/test_unstow.py
tests/test_unstow.py
import os import steeve def test_unstow(runner, stowed_foo_package): """Must remove all previously linked files.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 0 assert not os.path.exists(os.path.join('bin', 'foo')) def test_strict(runner): """Must fail when trying to unstow nonstowed package.""" result = runner.invoke(steeve.cli, ['unstow', 'nonstowed']) assert result.exit_code == 1 assert 'not stowed' in result.output
import os import steeve def test_no_current(runner, foo_package): """Must fail when unstowing a package with no 'current' symlink.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 1 assert 'not stowed' in result.output def test_unstow(runner, stowed_foo_package): """Must remove all previously linked files.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 0 assert not os.path.exists(os.path.join('bin', 'foo')) def test_strict(runner): """Must fail when trying to unstow nonstowed package.""" result = runner.invoke(steeve.cli, ['unstow', 'nonstowed']) assert result.exit_code == 1 assert 'not stowed' in result.output
Test unstowing a package with no 'current' symlink
Test unstowing a package with no 'current' symlink
Python
bsd-3-clause
Perlence/steeve,Perlence/steeve
import os import steeve + + + def test_no_current(runner, foo_package): + """Must fail when unstowing a package with no 'current' symlink.""" + result = runner.invoke(steeve.cli, ['unstow', 'foo']) + assert result.exit_code == 1 + assert 'not stowed' in result.output def test_unstow(runner, stowed_foo_package): """Must remove all previously linked files.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 0 assert not os.path.exists(os.path.join('bin', 'foo')) def test_strict(runner): """Must fail when trying to unstow nonstowed package.""" result = runner.invoke(steeve.cli, ['unstow', 'nonstowed']) assert result.exit_code == 1 assert 'not stowed' in result.output
Test unstowing a package with no 'current' symlink
## Code Before: import os import steeve def test_unstow(runner, stowed_foo_package): """Must remove all previously linked files.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 0 assert not os.path.exists(os.path.join('bin', 'foo')) def test_strict(runner): """Must fail when trying to unstow nonstowed package.""" result = runner.invoke(steeve.cli, ['unstow', 'nonstowed']) assert result.exit_code == 1 assert 'not stowed' in result.output ## Instruction: Test unstowing a package with no 'current' symlink ## Code After: import os import steeve def test_no_current(runner, foo_package): """Must fail when unstowing a package with no 'current' symlink.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 1 assert 'not stowed' in result.output def test_unstow(runner, stowed_foo_package): """Must remove all previously linked files.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 0 assert not os.path.exists(os.path.join('bin', 'foo')) def test_strict(runner): """Must fail when trying to unstow nonstowed package.""" result = runner.invoke(steeve.cli, ['unstow', 'nonstowed']) assert result.exit_code == 1 assert 'not stowed' in result.output
import os import steeve + + + def test_no_current(runner, foo_package): + """Must fail when unstowing a package with no 'current' symlink.""" + result = runner.invoke(steeve.cli, ['unstow', 'foo']) + assert result.exit_code == 1 + assert 'not stowed' in result.output def test_unstow(runner, stowed_foo_package): """Must remove all previously linked files.""" result = runner.invoke(steeve.cli, ['unstow', 'foo']) assert result.exit_code == 0 assert not os.path.exists(os.path.join('bin', 'foo')) def test_strict(runner): """Must fail when trying to unstow nonstowed package.""" result = runner.invoke(steeve.cli, ['unstow', 'nonstowed']) assert result.exit_code == 1 assert 'not stowed' in result.output
7ec36c81c6437bf83c498661c07802500e3acaa6
gore/urls.py
gore/urls.py
import os from django.conf.urls import include, url from lepo.router import Router from lepo.validate import validate_router import gore.handlers.events import gore.handlers.projects import gore.handlers.store router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml')) router.add_handlers(gore.handlers.events) router.add_handlers(gore.handlers.store) router.add_handlers(gore.handlers.projects) validate_router(router) urls = router.get_urls() for u in urls: # TODO: This shouldn't be necessary :( u.callback.csrf_exempt = True urlpatterns = [ url(r'^api/', include(urls)), ]
import os from django.conf.urls import include, url from lepo.decorators import csrf_exempt from lepo.router import Router from lepo.validate import validate_router import gore.handlers.events import gore.handlers.projects import gore.handlers.store router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml')) router.add_handlers(gore.handlers.events) router.add_handlers(gore.handlers.store) router.add_handlers(gore.handlers.projects) validate_router(router) urls = router.get_urls( optional_trailing_slash=True, decorate=(csrf_exempt,), ) urlpatterns = [ url(r'^api/', include(urls)), ]
Use Lepo 0.1.0's CSRF decorator
Gore: Use Lepo 0.1.0's CSRF decorator
Python
mit
akx/gentry,akx/gentry,akx/gentry,akx/gentry
import os from django.conf.urls import include, url + + from lepo.decorators import csrf_exempt from lepo.router import Router from lepo.validate import validate_router import gore.handlers.events import gore.handlers.projects import gore.handlers.store router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml')) router.add_handlers(gore.handlers.events) router.add_handlers(gore.handlers.store) router.add_handlers(gore.handlers.projects) validate_router(router) - urls = router.get_urls() + urls = router.get_urls( - - for u in urls: # TODO: This shouldn't be necessary :( - u.callback.csrf_exempt = True + optional_trailing_slash=True, + decorate=(csrf_exempt,), + ) urlpatterns = [ url(r'^api/', include(urls)), ]
Use Lepo 0.1.0's CSRF decorator
## Code Before: import os from django.conf.urls import include, url from lepo.router import Router from lepo.validate import validate_router import gore.handlers.events import gore.handlers.projects import gore.handlers.store router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml')) router.add_handlers(gore.handlers.events) router.add_handlers(gore.handlers.store) router.add_handlers(gore.handlers.projects) validate_router(router) urls = router.get_urls() for u in urls: # TODO: This shouldn't be necessary :( u.callback.csrf_exempt = True urlpatterns = [ url(r'^api/', include(urls)), ] ## Instruction: Use Lepo 0.1.0's CSRF decorator ## Code After: import os from django.conf.urls import include, url from lepo.decorators import csrf_exempt from lepo.router import Router from lepo.validate import validate_router import gore.handlers.events import gore.handlers.projects import gore.handlers.store router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml')) router.add_handlers(gore.handlers.events) router.add_handlers(gore.handlers.store) router.add_handlers(gore.handlers.projects) validate_router(router) urls = router.get_urls( optional_trailing_slash=True, decorate=(csrf_exempt,), ) urlpatterns = [ url(r'^api/', include(urls)), ]
import os from django.conf.urls import include, url + + from lepo.decorators import csrf_exempt from lepo.router import Router from lepo.validate import validate_router import gore.handlers.events import gore.handlers.projects import gore.handlers.store router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml')) router.add_handlers(gore.handlers.events) router.add_handlers(gore.handlers.store) router.add_handlers(gore.handlers.projects) validate_router(router) - urls = router.get_urls() ? - + urls = router.get_urls( - - for u in urls: # TODO: This shouldn't be necessary :( - u.callback.csrf_exempt = True + optional_trailing_slash=True, + decorate=(csrf_exempt,), + ) urlpatterns = [ url(r'^api/', include(urls)), ]
0f04e6ed48227c6904d75a78be9c893f47f9cb80
joku/cogs/_common.py
joku/cogs/_common.py
from collections import OrderedDict import threading from joku.bot import Jokusoramame class _CogMeta(type): def __prepare__(*args, **kwargs): # Use an OrderedDict for the class body. return OrderedDict() class Cog(metaclass=_CogMeta): """ A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local data. This makes the cogs semi thread-safe. """ def __init__(self, bot: Jokusoramame): self._bot = bot self.logger = self.bot.logger @property def bot(self) -> 'Jokusoramame': """ :return: The bot instance associated with this cog. """ return self._bot @classmethod def setup(cls, bot: Jokusoramame): bot.add_cog(cls(bot))
from collections import OrderedDict import threading from joku.bot import Jokusoramame class _CogMeta(type): def __prepare__(*args, **kwargs): # Use an OrderedDict for the class body. return OrderedDict() class Cog(metaclass=_CogMeta): def __init__(self, bot: Jokusoramame): self._bot = bot self.logger = self.bot.logger @property def bot(self) -> 'Jokusoramame': """ :return: The bot instance associated with this cog. """ return self._bot @classmethod def setup(cls, bot: Jokusoramame): bot.add_cog(cls(bot))
Remove false docstring from common cog.
Remove false docstring from common cog.
Python
mit
MJB47/Jokusoramame,MJB47/Jokusoramame,MJB47/Jokusoramame
from collections import OrderedDict import threading from joku.bot import Jokusoramame class _CogMeta(type): def __prepare__(*args, **kwargs): # Use an OrderedDict for the class body. return OrderedDict() class Cog(metaclass=_CogMeta): - """ - A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local - data. This makes the cogs semi thread-safe. - """ - def __init__(self, bot: Jokusoramame): self._bot = bot self.logger = self.bot.logger @property def bot(self) -> 'Jokusoramame': """ :return: The bot instance associated with this cog. """ return self._bot @classmethod def setup(cls, bot: Jokusoramame): bot.add_cog(cls(bot))
Remove false docstring from common cog.
## Code Before: from collections import OrderedDict import threading from joku.bot import Jokusoramame class _CogMeta(type): def __prepare__(*args, **kwargs): # Use an OrderedDict for the class body. return OrderedDict() class Cog(metaclass=_CogMeta): """ A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local data. This makes the cogs semi thread-safe. """ def __init__(self, bot: Jokusoramame): self._bot = bot self.logger = self.bot.logger @property def bot(self) -> 'Jokusoramame': """ :return: The bot instance associated with this cog. """ return self._bot @classmethod def setup(cls, bot: Jokusoramame): bot.add_cog(cls(bot)) ## Instruction: Remove false docstring from common cog. ## Code After: from collections import OrderedDict import threading from joku.bot import Jokusoramame class _CogMeta(type): def __prepare__(*args, **kwargs): # Use an OrderedDict for the class body. return OrderedDict() class Cog(metaclass=_CogMeta): def __init__(self, bot: Jokusoramame): self._bot = bot self.logger = self.bot.logger @property def bot(self) -> 'Jokusoramame': """ :return: The bot instance associated with this cog. """ return self._bot @classmethod def setup(cls, bot: Jokusoramame): bot.add_cog(cls(bot))
from collections import OrderedDict import threading from joku.bot import Jokusoramame class _CogMeta(type): def __prepare__(*args, **kwargs): # Use an OrderedDict for the class body. return OrderedDict() class Cog(metaclass=_CogMeta): - """ - A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local - data. This makes the cogs semi thread-safe. - """ - def __init__(self, bot: Jokusoramame): self._bot = bot self.logger = self.bot.logger @property def bot(self) -> 'Jokusoramame': """ :return: The bot instance associated with this cog. """ return self._bot @classmethod def setup(cls, bot: Jokusoramame): bot.add_cog(cls(bot))
1661174b80e00ff04a2df245abf73b92825ec01a
libs/qr_tools.py
libs/qr_tools.py
import pyqrcode # sudo pip install pyqrcode def getQRArray(text, errorCorrection): """ Takes in text and errorCorrection (letter), returns 2D array of the QR code""" # White is True (1) # Black is False (0) # ECC: L7, M15, Q25, H30 # Create the object qr = pyqrcode.create(text, error=errorCorrection) # Get the terminal representation and split by lines (get rid of top and bottom white spaces) plainOut = qr.terminal().split("\n")[5:-5] print(qr.terminal()) # Initialize the output 2D list out = [] for line in plainOut: thisOut = [] for char in line: if char == u'7': # This is white thisOut.append(1) elif char == u'4': # This is black, it's part of the u'49' thisOut.append(0) # Finally add everything to the output, stipping whitespaces at start and end out.append(thisOut[4:-4]) # Everything is done, return the qr code list return out
import pyqrcode # sudo pip install pyqrcode def getQRArray(text, errorCorrection): """ Takes in text and errorCorrection (letter), returns 2D array of the QR code""" # White is True (1) # Black is False (0) # ECC: L7, M15, Q25, H30 # Create the object qr = pyqrcode.create(text, error=errorCorrection) # Get the terminal representation and split by lines (get rid of top and bottom white spaces) plainOut = qr.terminal().split("\n")[5:-5] # Initialize the output 2D list out = [] for line in plainOut: thisOut = [] for char in line: if char == u'7': # This is white thisOut.append(1) elif char == u'4': # This is black, it's part of the u'49' thisOut.append(0) # Finally add everything to the output, stipping whitespaces at start and end out.append(thisOut[4:-4]) # Everything is done, return the qr code list return out
Remove print of terminal output for debugging
Remove print of terminal output for debugging
Python
mit
btcspry/3d-wallet-generator
import pyqrcode # sudo pip install pyqrcode def getQRArray(text, errorCorrection): """ Takes in text and errorCorrection (letter), returns 2D array of the QR code""" # White is True (1) # Black is False (0) # ECC: L7, M15, Q25, H30 # Create the object qr = pyqrcode.create(text, error=errorCorrection) # Get the terminal representation and split by lines (get rid of top and bottom white spaces) plainOut = qr.terminal().split("\n")[5:-5] - - print(qr.terminal()) # Initialize the output 2D list out = [] for line in plainOut: thisOut = [] for char in line: if char == u'7': # This is white thisOut.append(1) elif char == u'4': # This is black, it's part of the u'49' thisOut.append(0) # Finally add everything to the output, stipping whitespaces at start and end out.append(thisOut[4:-4]) # Everything is done, return the qr code list return out
Remove print of terminal output for debugging
## Code Before: import pyqrcode # sudo pip install pyqrcode def getQRArray(text, errorCorrection): """ Takes in text and errorCorrection (letter), returns 2D array of the QR code""" # White is True (1) # Black is False (0) # ECC: L7, M15, Q25, H30 # Create the object qr = pyqrcode.create(text, error=errorCorrection) # Get the terminal representation and split by lines (get rid of top and bottom white spaces) plainOut = qr.terminal().split("\n")[5:-5] print(qr.terminal()) # Initialize the output 2D list out = [] for line in plainOut: thisOut = [] for char in line: if char == u'7': # This is white thisOut.append(1) elif char == u'4': # This is black, it's part of the u'49' thisOut.append(0) # Finally add everything to the output, stipping whitespaces at start and end out.append(thisOut[4:-4]) # Everything is done, return the qr code list return out ## Instruction: Remove print of terminal output for debugging ## Code After: import pyqrcode # sudo pip install pyqrcode def getQRArray(text, errorCorrection): """ Takes in text and errorCorrection (letter), returns 2D array of the QR code""" # White is True (1) # Black is False (0) # ECC: L7, M15, Q25, H30 # Create the object qr = pyqrcode.create(text, error=errorCorrection) # Get the terminal representation and split by lines (get rid of top and bottom white spaces) plainOut = qr.terminal().split("\n")[5:-5] # Initialize the output 2D list out = [] for line in plainOut: thisOut = [] for char in line: if char == u'7': # This is white thisOut.append(1) elif char == u'4': # This is black, it's part of the u'49' thisOut.append(0) # Finally add everything to the output, stipping whitespaces at start and end out.append(thisOut[4:-4]) # Everything is done, return the qr code list return out
import pyqrcode # sudo pip install pyqrcode def getQRArray(text, errorCorrection): """ Takes in text and errorCorrection (letter), returns 2D array of the QR code""" # White is True (1) # Black is False (0) # ECC: L7, M15, Q25, H30 # Create the object qr = pyqrcode.create(text, error=errorCorrection) # Get the terminal representation and split by lines (get rid of top and bottom white spaces) plainOut = qr.terminal().split("\n")[5:-5] - - print(qr.terminal()) # Initialize the output 2D list out = [] for line in plainOut: thisOut = [] for char in line: if char == u'7': # This is white thisOut.append(1) elif char == u'4': # This is black, it's part of the u'49' thisOut.append(0) # Finally add everything to the output, stipping whitespaces at start and end out.append(thisOut[4:-4]) # Everything is done, return the qr code list return out
91dee60bb768a8ab80530cab79649b60afdf7daf
mbed.py
mbed.py
from utils.helpers import error, find_mbed_dir, is_mbed_dir import sys, os from utils import set_project_dir from commands.set import CmdSet from commands.get import CmdGet from commands.clone import CmdClone from commands.compile import CmdCompile from commands.list import CmdList ################################################################################ # Local functions def help_and_exit(cmds): error("Syntax: mbed <command> [arguments]") error("Valid commands:") for c in cmds: error(" " + c.get_help() + "") os._exit(1) def run(args): cmds = [CmdSet(), CmdGet()] if is_mbed_dir(): cmds = cmds + [CmdCompile(), CmdList()] else: cmds = cmds = [CmdClone()] if len(args) == 0: error("No command given.") help_and_exit(cmds) cmd_map = dict([(c.get_name(), c) for c in cmds]) cmd = args[0].lower() if not cmd in cmd_map: error("Invalid command '%s'." % args[0]) help_and_exit(cmds) res = cmd_map[cmd](args[1:]) if res == None: error("Invalid command syntax") error(cmd_map[cmd].get_help()) elif res == False: os._exit(1) ################################################################################ # Entry point if __name__ == "__main__": set_project_dir(find_mbed_dir()) run(sys.argv[1:])
from utils.helpers import error, find_mbed_dir, is_mbed_dir import sys, os from utils import set_project_dir from commands.set import CmdSet from commands.get import CmdGet from commands.clone import CmdClone from commands.compile import CmdCompile from commands.list import CmdList ################################################################################ # Local functions def help_and_exit(cmds): error("Syntax: mbed <command> [arguments]") error("Valid commands:") for c in cmds: error(" " + c.get_help() + "") os._exit(1) def run(args): cmds = [CmdSet(), CmdGet()] if is_mbed_dir(): cmds = cmds + [CmdCompile(), CmdList()] else: cmds = cmds = [CmdClone()] if len(args) == 0: error("No command given.") help_and_exit(cmds) cmd_map = dict([(c.get_name(), c) for c in cmds]) cmd = args[0].lower() if not cmd in cmd_map: error("Invalid command '%s'." % args[0]) help_and_exit(cmds) res = cmd_map[cmd](args[1:]) if res == None: error("Invalid command syntax") error(cmd_map[cmd].get_help()) elif res == False: os._exit(1) ################################################################################ # Entry point if __name__ == "__main__": base = find_mbed_dir() if base: set_project_dir(base) sys.path.append(base) run(sys.argv[1:])
Fix Python module search path
Fix Python module search path
Python
apache-2.0
bogdanm/mbed-clt
from utils.helpers import error, find_mbed_dir, is_mbed_dir import sys, os from utils import set_project_dir from commands.set import CmdSet from commands.get import CmdGet from commands.clone import CmdClone from commands.compile import CmdCompile from commands.list import CmdList ################################################################################ # Local functions def help_and_exit(cmds): error("Syntax: mbed <command> [arguments]") error("Valid commands:") for c in cmds: error(" " + c.get_help() + "") os._exit(1) def run(args): cmds = [CmdSet(), CmdGet()] if is_mbed_dir(): cmds = cmds + [CmdCompile(), CmdList()] else: cmds = cmds = [CmdClone()] if len(args) == 0: error("No command given.") help_and_exit(cmds) cmd_map = dict([(c.get_name(), c) for c in cmds]) cmd = args[0].lower() if not cmd in cmd_map: error("Invalid command '%s'." % args[0]) help_and_exit(cmds) res = cmd_map[cmd](args[1:]) if res == None: error("Invalid command syntax") error(cmd_map[cmd].get_help()) elif res == False: os._exit(1) ################################################################################ # Entry point if __name__ == "__main__": - set_project_dir(find_mbed_dir()) + base = find_mbed_dir() + if base: + set_project_dir(base) + sys.path.append(base) run(sys.argv[1:])
Fix Python module search path
## Code Before: from utils.helpers import error, find_mbed_dir, is_mbed_dir import sys, os from utils import set_project_dir from commands.set import CmdSet from commands.get import CmdGet from commands.clone import CmdClone from commands.compile import CmdCompile from commands.list import CmdList ################################################################################ # Local functions def help_and_exit(cmds): error("Syntax: mbed <command> [arguments]") error("Valid commands:") for c in cmds: error(" " + c.get_help() + "") os._exit(1) def run(args): cmds = [CmdSet(), CmdGet()] if is_mbed_dir(): cmds = cmds + [CmdCompile(), CmdList()] else: cmds = cmds = [CmdClone()] if len(args) == 0: error("No command given.") help_and_exit(cmds) cmd_map = dict([(c.get_name(), c) for c in cmds]) cmd = args[0].lower() if not cmd in cmd_map: error("Invalid command '%s'." % args[0]) help_and_exit(cmds) res = cmd_map[cmd](args[1:]) if res == None: error("Invalid command syntax") error(cmd_map[cmd].get_help()) elif res == False: os._exit(1) ################################################################################ # Entry point if __name__ == "__main__": set_project_dir(find_mbed_dir()) run(sys.argv[1:]) ## Instruction: Fix Python module search path ## Code After: from utils.helpers import error, find_mbed_dir, is_mbed_dir import sys, os from utils import set_project_dir from commands.set import CmdSet from commands.get import CmdGet from commands.clone import CmdClone from commands.compile import CmdCompile from commands.list import CmdList ################################################################################ # Local functions def help_and_exit(cmds): error("Syntax: mbed <command> [arguments]") error("Valid commands:") for c in cmds: error(" " + c.get_help() + "") os._exit(1) def run(args): cmds = [CmdSet(), CmdGet()] if is_mbed_dir(): cmds = cmds + [CmdCompile(), CmdList()] else: cmds = cmds = [CmdClone()] if len(args) == 0: error("No command given.") help_and_exit(cmds) cmd_map = dict([(c.get_name(), c) for c in cmds]) cmd = args[0].lower() if not cmd in cmd_map: error("Invalid command '%s'." % args[0]) help_and_exit(cmds) res = cmd_map[cmd](args[1:]) if res == None: error("Invalid command syntax") error(cmd_map[cmd].get_help()) elif res == False: os._exit(1) ################################################################################ # Entry point if __name__ == "__main__": base = find_mbed_dir() if base: set_project_dir(base) sys.path.append(base) run(sys.argv[1:])
from utils.helpers import error, find_mbed_dir, is_mbed_dir import sys, os from utils import set_project_dir from commands.set import CmdSet from commands.get import CmdGet from commands.clone import CmdClone from commands.compile import CmdCompile from commands.list import CmdList ################################################################################ # Local functions def help_and_exit(cmds): error("Syntax: mbed <command> [arguments]") error("Valid commands:") for c in cmds: error(" " + c.get_help() + "") os._exit(1) def run(args): cmds = [CmdSet(), CmdGet()] if is_mbed_dir(): cmds = cmds + [CmdCompile(), CmdList()] else: cmds = cmds = [CmdClone()] if len(args) == 0: error("No command given.") help_and_exit(cmds) cmd_map = dict([(c.get_name(), c) for c in cmds]) cmd = args[0].lower() if not cmd in cmd_map: error("Invalid command '%s'." % args[0]) help_and_exit(cmds) res = cmd_map[cmd](args[1:]) if res == None: error("Invalid command syntax") error(cmd_map[cmd].get_help()) elif res == False: os._exit(1) ################################################################################ # Entry point if __name__ == "__main__": - set_project_dir(find_mbed_dir()) + base = find_mbed_dir() + if base: + set_project_dir(base) + sys.path.append(base) run(sys.argv[1:])
146f6204e58695ca469cec7a79757ce9a730719e
contrib/migrateticketmodel.py
contrib/migrateticketmodel.py
import sys from trac.env import open_environment from trac.ticket.model import Priority, Severity priority_mapping = { 'highest': 'blocker', 'high': 'critical', 'normal': 'major', 'low': 'minor', 'lowest': 'trivial' } def main(): if len(sys.argv) < 2: print >> sys.stderr, 'usage: %s /path/to/projenv' \ % os.path.basename(sys.argv[0]) sys.exit(2) env = open_environment(sys.argv[1]) db = env.get_db_cnx() for oldprio, newprio in priority_mapping.items(): priority = Priority(env, oldprio, db) priority.name = newprio priority.update(db) for severity in list(Severity.select(env, db)): severity.delete(db) db.commit() if __name__ == '__main__': main()
import os import sys from trac.env import open_environment from trac.ticket.model import Priority, Severity priority_mapping = { 'highest': 'blocker', 'high': 'critical', 'normal': 'major', 'low': 'minor', 'lowest': 'trivial' } def main(): if len(sys.argv) < 2: print >> sys.stderr, 'usage: %s /path/to/projenv' \ % os.path.basename(sys.argv[0]) sys.exit(2) env = open_environment(sys.argv[1]) db = env.get_db_cnx() for oldprio, newprio in priority_mapping.items(): priority = Priority(env, oldprio, db) priority.name = newprio priority.update(db) for severity in list(Severity.select(env, db)): severity.delete(db) db.commit() if __name__ == '__main__': main()
Fix missing import in contrib script added in [2630].
Fix missing import in contrib script added in [2630].
Python
bsd-3-clause
pkdevbox/trac,pkdevbox/trac,pkdevbox/trac,pkdevbox/trac
+ import os import sys from trac.env import open_environment from trac.ticket.model import Priority, Severity priority_mapping = { 'highest': 'blocker', 'high': 'critical', 'normal': 'major', 'low': 'minor', 'lowest': 'trivial' } def main(): if len(sys.argv) < 2: print >> sys.stderr, 'usage: %s /path/to/projenv' \ % os.path.basename(sys.argv[0]) sys.exit(2) env = open_environment(sys.argv[1]) db = env.get_db_cnx() for oldprio, newprio in priority_mapping.items(): priority = Priority(env, oldprio, db) priority.name = newprio priority.update(db) for severity in list(Severity.select(env, db)): severity.delete(db) db.commit() if __name__ == '__main__': main()
Fix missing import in contrib script added in [2630].
## Code Before: import sys from trac.env import open_environment from trac.ticket.model import Priority, Severity priority_mapping = { 'highest': 'blocker', 'high': 'critical', 'normal': 'major', 'low': 'minor', 'lowest': 'trivial' } def main(): if len(sys.argv) < 2: print >> sys.stderr, 'usage: %s /path/to/projenv' \ % os.path.basename(sys.argv[0]) sys.exit(2) env = open_environment(sys.argv[1]) db = env.get_db_cnx() for oldprio, newprio in priority_mapping.items(): priority = Priority(env, oldprio, db) priority.name = newprio priority.update(db) for severity in list(Severity.select(env, db)): severity.delete(db) db.commit() if __name__ == '__main__': main() ## Instruction: Fix missing import in contrib script added in [2630]. ## Code After: import os import sys from trac.env import open_environment from trac.ticket.model import Priority, Severity priority_mapping = { 'highest': 'blocker', 'high': 'critical', 'normal': 'major', 'low': 'minor', 'lowest': 'trivial' } def main(): if len(sys.argv) < 2: print >> sys.stderr, 'usage: %s /path/to/projenv' \ % os.path.basename(sys.argv[0]) sys.exit(2) env = open_environment(sys.argv[1]) db = env.get_db_cnx() for oldprio, newprio in priority_mapping.items(): priority = Priority(env, oldprio, db) priority.name = newprio priority.update(db) for severity in list(Severity.select(env, db)): severity.delete(db) db.commit() if __name__ == '__main__': main()
+ import os import sys from trac.env import open_environment from trac.ticket.model import Priority, Severity priority_mapping = { 'highest': 'blocker', 'high': 'critical', 'normal': 'major', 'low': 'minor', 'lowest': 'trivial' } def main(): if len(sys.argv) < 2: print >> sys.stderr, 'usage: %s /path/to/projenv' \ % os.path.basename(sys.argv[0]) sys.exit(2) env = open_environment(sys.argv[1]) db = env.get_db_cnx() for oldprio, newprio in priority_mapping.items(): priority = Priority(env, oldprio, db) priority.name = newprio priority.update(db) for severity in list(Severity.select(env, db)): severity.delete(db) db.commit() if __name__ == '__main__': main()