commit stringlengths 40 40 | old_file stringlengths 4 106 | new_file stringlengths 4 106 | old_contents stringlengths 10 2.94k | new_contents stringlengths 21 2.95k | subject stringlengths 16 444 | message stringlengths 17 2.63k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 7 43k | ndiff stringlengths 52 3.31k | instruction stringlengths 16 444 | content stringlengths 133 4.32k | diff stringlengths 49 3.61k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7325eacc1066970a98be30b56fdf4cd31ecc2f57 | db_file_storage/views.py | db_file_storage/views.py | from wsgiref.util import FileWrapper
from django.http import HttpResponse, HttpResponseBadRequest
from django.utils.translation import ugettext as _
# project
from db_file_storage.storage import DatabaseFileStorage
storage = DatabaseFileStorage()
def get_file(request, add_attachment_headers):
name = request.GET.get('name')
try:
_file = storage.open(name)
except Exception:
return HttpResponseBadRequest(_('Invalid request'))
response = HttpResponse(
FileWrapper(_file),
content_type=_file.mimetype
)
if add_attachment_headers:
response['Content-Disposition'] = \
'attachment; filename=%(name)s' % {'name': _file.filename}
return response
| from wsgiref.util import FileWrapper
from django.http import HttpResponse, HttpResponseBadRequest
from django.utils.translation import ugettext as _
# project
from db_file_storage.storage import DatabaseFileStorage
storage = DatabaseFileStorage()
def get_file(request, add_attachment_headers):
name = request.GET.get('name')
try:
_file = storage.open(name)
except Exception:
return HttpResponseBadRequest(_('Invalid request'))
response = HttpResponse(
FileWrapper(_file),
content_type=_file.mimetype
)
response['Content-Length'] = _file.tell()
if add_attachment_headers:
response['Content-Disposition'] = \
'attachment; filename=%(name)s' % {'name': _file.filename}
return response
| Set Content-Length header in get_file view | Set Content-Length header in get_file view
| Python | mit | victor-o-silva/db_file_storage,victor-o-silva/db_file_storage | from wsgiref.util import FileWrapper
from django.http import HttpResponse, HttpResponseBadRequest
from django.utils.translation import ugettext as _
# project
from db_file_storage.storage import DatabaseFileStorage
storage = DatabaseFileStorage()
def get_file(request, add_attachment_headers):
name = request.GET.get('name')
try:
_file = storage.open(name)
except Exception:
return HttpResponseBadRequest(_('Invalid request'))
response = HttpResponse(
FileWrapper(_file),
content_type=_file.mimetype
)
-
+ response['Content-Length'] = _file.tell()
if add_attachment_headers:
response['Content-Disposition'] = \
'attachment; filename=%(name)s' % {'name': _file.filename}
return response
| Set Content-Length header in get_file view | ## Code Before:
from wsgiref.util import FileWrapper
from django.http import HttpResponse, HttpResponseBadRequest
from django.utils.translation import ugettext as _
# project
from db_file_storage.storage import DatabaseFileStorage
storage = DatabaseFileStorage()
def get_file(request, add_attachment_headers):
name = request.GET.get('name')
try:
_file = storage.open(name)
except Exception:
return HttpResponseBadRequest(_('Invalid request'))
response = HttpResponse(
FileWrapper(_file),
content_type=_file.mimetype
)
if add_attachment_headers:
response['Content-Disposition'] = \
'attachment; filename=%(name)s' % {'name': _file.filename}
return response
## Instruction:
Set Content-Length header in get_file view
## Code After:
from wsgiref.util import FileWrapper
from django.http import HttpResponse, HttpResponseBadRequest
from django.utils.translation import ugettext as _
# project
from db_file_storage.storage import DatabaseFileStorage
storage = DatabaseFileStorage()
def get_file(request, add_attachment_headers):
name = request.GET.get('name')
try:
_file = storage.open(name)
except Exception:
return HttpResponseBadRequest(_('Invalid request'))
response = HttpResponse(
FileWrapper(_file),
content_type=_file.mimetype
)
response['Content-Length'] = _file.tell()
if add_attachment_headers:
response['Content-Disposition'] = \
'attachment; filename=%(name)s' % {'name': _file.filename}
return response
| from wsgiref.util import FileWrapper
from django.http import HttpResponse, HttpResponseBadRequest
from django.utils.translation import ugettext as _
# project
from db_file_storage.storage import DatabaseFileStorage
storage = DatabaseFileStorage()
def get_file(request, add_attachment_headers):
name = request.GET.get('name')
try:
_file = storage.open(name)
except Exception:
return HttpResponseBadRequest(_('Invalid request'))
response = HttpResponse(
FileWrapper(_file),
content_type=_file.mimetype
)
-
+ response['Content-Length'] = _file.tell()
if add_attachment_headers:
response['Content-Disposition'] = \
'attachment; filename=%(name)s' % {'name': _file.filename}
return response |
7a8112249de859a5ef73fe07eb6029aeb1266f35 | tob-api/tob_api/urls.py | tob-api/tob_api/urls.py |
from django.conf.urls import include, url
from django.views.generic import RedirectView
from . import views
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = [
url(r"^$", RedirectView.as_view(url="api/v2/")),
url(
r"^api-auth/",
include("rest_framework.urls", namespace="rest_framework"),
),
# url(r"^api/v1/", include("api.urls")),
url(r"^api/v2/", include("api_v2.urls")),
url(r"^health$", views.health),
]
|
from django.conf.urls import include, url
from django.views.generic import RedirectView
from . import views
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = [
url(r"^$", RedirectView.as_view(url="api/v2/")),
url(
r"^api-auth/",
include("rest_framework.urls", namespace="rest_framework"),
),
url(r"^api/v2/", include("api_v2.urls")),
url(r"^health$", views.health),
]
| Remove commented-out reference to v1 | Remove commented-out reference to v1
Signed-off-by: Nicholas Rempel <b7f0f2181f2dc324d159332b253a82a715a40706@gmail.com>
| Python | apache-2.0 | swcurran/TheOrgBook,swcurran/TheOrgBook,WadeBarnes/TheOrgBook,swcurran/TheOrgBook,WadeBarnes/TheOrgBook,WadeBarnes/TheOrgBook,WadeBarnes/TheOrgBook,WadeBarnes/TheOrgBook,swcurran/TheOrgBook,swcurran/TheOrgBook |
from django.conf.urls import include, url
from django.views.generic import RedirectView
from . import views
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = [
url(r"^$", RedirectView.as_view(url="api/v2/")),
url(
r"^api-auth/",
include("rest_framework.urls", namespace="rest_framework"),
),
- # url(r"^api/v1/", include("api.urls")),
url(r"^api/v2/", include("api_v2.urls")),
url(r"^health$", views.health),
]
| Remove commented-out reference to v1 | ## Code Before:
from django.conf.urls import include, url
from django.views.generic import RedirectView
from . import views
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = [
url(r"^$", RedirectView.as_view(url="api/v2/")),
url(
r"^api-auth/",
include("rest_framework.urls", namespace="rest_framework"),
),
# url(r"^api/v1/", include("api.urls")),
url(r"^api/v2/", include("api_v2.urls")),
url(r"^health$", views.health),
]
## Instruction:
Remove commented-out reference to v1
## Code After:
from django.conf.urls import include, url
from django.views.generic import RedirectView
from . import views
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = [
url(r"^$", RedirectView.as_view(url="api/v2/")),
url(
r"^api-auth/",
include("rest_framework.urls", namespace="rest_framework"),
),
url(r"^api/v2/", include("api_v2.urls")),
url(r"^health$", views.health),
]
|
from django.conf.urls import include, url
from django.views.generic import RedirectView
from . import views
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = [
url(r"^$", RedirectView.as_view(url="api/v2/")),
url(
r"^api-auth/",
include("rest_framework.urls", namespace="rest_framework"),
),
- # url(r"^api/v1/", include("api.urls")),
url(r"^api/v2/", include("api_v2.urls")),
url(r"^health$", views.health),
] |
f68b4b9b133d3c8ecb9826af9736c8c1fca64e49 | maxims/credentials.py | maxims/credentials.py | from axiom import attributes, item
from twisted.cred import credentials
class UsernamePassword(item.Item):
"""
A stored username and password.
"""
username = attributes.bytes(allowNone=False)
password = attributes.bytes(allowNone=False)
def instantiate(self):
return credentials.UsernamePassword(self.username, self.password)
| from axiom import attributes, item
from twisted.cred import credentials
class UsernamePassword(item.Item):
"""
A stored username and password.
Note that although this class is an ``IUsernamePassword`` implementation,
you should still use the ``instantiate`` method to get independent
``IUsernamePassword`` providers.
"""
username = attributes.bytes(allowNone=False)
password = attributes.bytes(allowNone=False)
def instantiate(self):
return credentials.UsernamePassword(self.username, self.password)
| Add caveat about UsernamePassword already being an IUsernamePassword implementation | Add caveat about UsernamePassword already being an IUsernamePassword implementation
| Python | isc | lvh/maxims | from axiom import attributes, item
from twisted.cred import credentials
class UsernamePassword(item.Item):
"""
A stored username and password.
+
+ Note that although this class is an ``IUsernamePassword`` implementation,
+ you should still use the ``instantiate`` method to get independent
+ ``IUsernamePassword`` providers.
"""
username = attributes.bytes(allowNone=False)
password = attributes.bytes(allowNone=False)
def instantiate(self):
return credentials.UsernamePassword(self.username, self.password)
| Add caveat about UsernamePassword already being an IUsernamePassword implementation | ## Code Before:
from axiom import attributes, item
from twisted.cred import credentials
class UsernamePassword(item.Item):
"""
A stored username and password.
"""
username = attributes.bytes(allowNone=False)
password = attributes.bytes(allowNone=False)
def instantiate(self):
return credentials.UsernamePassword(self.username, self.password)
## Instruction:
Add caveat about UsernamePassword already being an IUsernamePassword implementation
## Code After:
from axiom import attributes, item
from twisted.cred import credentials
class UsernamePassword(item.Item):
"""
A stored username and password.
Note that although this class is an ``IUsernamePassword`` implementation,
you should still use the ``instantiate`` method to get independent
``IUsernamePassword`` providers.
"""
username = attributes.bytes(allowNone=False)
password = attributes.bytes(allowNone=False)
def instantiate(self):
return credentials.UsernamePassword(self.username, self.password)
| from axiom import attributes, item
from twisted.cred import credentials
class UsernamePassword(item.Item):
"""
A stored username and password.
+
+ Note that although this class is an ``IUsernamePassword`` implementation,
+ you should still use the ``instantiate`` method to get independent
+ ``IUsernamePassword`` providers.
"""
username = attributes.bytes(allowNone=False)
password = attributes.bytes(allowNone=False)
def instantiate(self):
return credentials.UsernamePassword(self.username, self.password) |
df18229b38a01d87076f3b13aee5bfd1f0f989c2 | tunobase/blog/models.py | tunobase/blog/models.py | '''
Blog App
This module determines how to display the Blog app in Django's admin
and lists other model functions.
'''
from django.conf import settings
from django.core.urlresolvers import reverse
from django.db import models
from tunobase.core import models as core_models
class Blog(core_models.ContentModel):
'''
Blogs the Site has
'''
class BlogEntry(core_models.ContentModel):
'''
Entries per Blog
'''
blog = models.ForeignKey(Blog, related_name='entries')
author_users = models.ManyToManyField(
settings.AUTH_USER_MODEL,
related_name='blog_entries_authored',
null=True,
blank=True
)
authors_alternate = models.CharField(
max_length=512,
blank=True,
null=True
)
class Meta:
verbose_name_plural = 'Blog entries'
def get_absolute_url(self):
return reverse('blog_entry_detail', args=(self.slug,))
@property
def authors(self):
'''
Return a list of authors selected as users on the system and a list
of alternate authors as not users on the system if either exist
'''
authors_dict = {}
auth_users = self.author_users.all()
if auth_users:
authors_dict.update({
'users': auth_users
})
if self.authors_alternate:
authors_dict.update({
'alternate': self.authors_alternate.split(',')
})
return authors_dict
| '''
Blog App
This module determines how to display the Blog app in Django's admin
and lists other model functions.
'''
from django.conf import settings
from django.core.urlresolvers import reverse
from django.db import models
from tunobase.core import models as core_models
class Blog(core_models.ContentModel):
'''
Blogs the Site has
'''
class Meta:
verbose_name = 'Blog Category'
verbose_name_plural = 'Blog Categories'
class BlogEntry(core_models.ContentModel):
'''
Entries per Blog
'''
blog = models.ForeignKey(Blog, related_name='entries')
author_users = models.ManyToManyField(
settings.AUTH_USER_MODEL,
related_name='blog_entries_authored',
null=True,
blank=True
)
authors_alternate = models.CharField(
max_length=512,
blank=True,
null=True
)
class Meta:
verbose_name_plural = 'Blog entries'
def get_absolute_url(self):
return reverse('blog_entry_detail', args=(self.slug,))
@property
def authors(self):
'''
Return a list of authors selected as users on the system and a list
of alternate authors as not users on the system if either exist
'''
authors_dict = {}
auth_users = self.author_users.all()
if auth_users:
authors_dict.update({
'users': auth_users
})
if self.authors_alternate:
authors_dict.update({
'alternate': self.authors_alternate.split(',')
})
return authors_dict
| Update blog model with a more descriptive name | Update blog model with a more descriptive name
| Python | bsd-3-clause | unomena/tunobase,unomena/tunobase | '''
Blog App
This module determines how to display the Blog app in Django's admin
and lists other model functions.
'''
from django.conf import settings
from django.core.urlresolvers import reverse
from django.db import models
from tunobase.core import models as core_models
class Blog(core_models.ContentModel):
'''
Blogs the Site has
'''
+
+ class Meta:
+ verbose_name = 'Blog Category'
+ verbose_name_plural = 'Blog Categories'
class BlogEntry(core_models.ContentModel):
'''
Entries per Blog
'''
blog = models.ForeignKey(Blog, related_name='entries')
author_users = models.ManyToManyField(
settings.AUTH_USER_MODEL,
related_name='blog_entries_authored',
null=True,
blank=True
)
authors_alternate = models.CharField(
max_length=512,
blank=True,
null=True
)
class Meta:
verbose_name_plural = 'Blog entries'
def get_absolute_url(self):
return reverse('blog_entry_detail', args=(self.slug,))
@property
def authors(self):
'''
Return a list of authors selected as users on the system and a list
of alternate authors as not users on the system if either exist
'''
authors_dict = {}
auth_users = self.author_users.all()
if auth_users:
authors_dict.update({
'users': auth_users
})
if self.authors_alternate:
authors_dict.update({
'alternate': self.authors_alternate.split(',')
})
return authors_dict
| Update blog model with a more descriptive name | ## Code Before:
'''
Blog App
This module determines how to display the Blog app in Django's admin
and lists other model functions.
'''
from django.conf import settings
from django.core.urlresolvers import reverse
from django.db import models
from tunobase.core import models as core_models
class Blog(core_models.ContentModel):
'''
Blogs the Site has
'''
class BlogEntry(core_models.ContentModel):
'''
Entries per Blog
'''
blog = models.ForeignKey(Blog, related_name='entries')
author_users = models.ManyToManyField(
settings.AUTH_USER_MODEL,
related_name='blog_entries_authored',
null=True,
blank=True
)
authors_alternate = models.CharField(
max_length=512,
blank=True,
null=True
)
class Meta:
verbose_name_plural = 'Blog entries'
def get_absolute_url(self):
return reverse('blog_entry_detail', args=(self.slug,))
@property
def authors(self):
'''
Return a list of authors selected as users on the system and a list
of alternate authors as not users on the system if either exist
'''
authors_dict = {}
auth_users = self.author_users.all()
if auth_users:
authors_dict.update({
'users': auth_users
})
if self.authors_alternate:
authors_dict.update({
'alternate': self.authors_alternate.split(',')
})
return authors_dict
## Instruction:
Update blog model with a more descriptive name
## Code After:
'''
Blog App
This module determines how to display the Blog app in Django's admin
and lists other model functions.
'''
from django.conf import settings
from django.core.urlresolvers import reverse
from django.db import models
from tunobase.core import models as core_models
class Blog(core_models.ContentModel):
'''
Blogs the Site has
'''
class Meta:
verbose_name = 'Blog Category'
verbose_name_plural = 'Blog Categories'
class BlogEntry(core_models.ContentModel):
'''
Entries per Blog
'''
blog = models.ForeignKey(Blog, related_name='entries')
author_users = models.ManyToManyField(
settings.AUTH_USER_MODEL,
related_name='blog_entries_authored',
null=True,
blank=True
)
authors_alternate = models.CharField(
max_length=512,
blank=True,
null=True
)
class Meta:
verbose_name_plural = 'Blog entries'
def get_absolute_url(self):
return reverse('blog_entry_detail', args=(self.slug,))
@property
def authors(self):
'''
Return a list of authors selected as users on the system and a list
of alternate authors as not users on the system if either exist
'''
authors_dict = {}
auth_users = self.author_users.all()
if auth_users:
authors_dict.update({
'users': auth_users
})
if self.authors_alternate:
authors_dict.update({
'alternate': self.authors_alternate.split(',')
})
return authors_dict
| '''
Blog App
This module determines how to display the Blog app in Django's admin
and lists other model functions.
'''
from django.conf import settings
from django.core.urlresolvers import reverse
from django.db import models
from tunobase.core import models as core_models
class Blog(core_models.ContentModel):
'''
Blogs the Site has
'''
+
+ class Meta:
+ verbose_name = 'Blog Category'
+ verbose_name_plural = 'Blog Categories'
class BlogEntry(core_models.ContentModel):
'''
Entries per Blog
'''
blog = models.ForeignKey(Blog, related_name='entries')
author_users = models.ManyToManyField(
settings.AUTH_USER_MODEL,
related_name='blog_entries_authored',
null=True,
blank=True
)
authors_alternate = models.CharField(
max_length=512,
blank=True,
null=True
)
class Meta:
verbose_name_plural = 'Blog entries'
def get_absolute_url(self):
return reverse('blog_entry_detail', args=(self.slug,))
@property
def authors(self):
'''
Return a list of authors selected as users on the system and a list
of alternate authors as not users on the system if either exist
'''
authors_dict = {}
auth_users = self.author_users.all()
if auth_users:
authors_dict.update({
'users': auth_users
})
if self.authors_alternate:
authors_dict.update({
'alternate': self.authors_alternate.split(',')
})
return authors_dict |
7a37e3afa29410636c75408bc649e70c519e07f1 | test/user_profile_test.py | test/user_profile_test.py | import json
from pymessenger.user_profile import UserProfileApi
from test_env import *
upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET)
def test_fields_blank():
user_profile = upa.get(TEST_USER_ID)
assert user_profile is not None
def test_fields():
fields = ['first_name', 'last_name']
user_profile = upa.get(TEST_USER_ID, fields=fields)
assert user_profile is not None
assert len(user_profile.keys()) == len(fields)
| import json
import sys, os
sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/.."))
from pymessenger.user_profile import UserProfileApi
TOKEN = os.environ.get('TOKEN')
APP_SECRET = os.environ.get('APP_SECRET')
TEST_USER_ID = os.environ.get('RECIPIENT_ID')
upa = UserProfileApi(TOKEN, app_secret=APP_SECRET)
def test_fields_blank():
user_profile = upa.get(TEST_USER_ID)
assert user_profile is not None
def test_fields():
fields = ['first_name', 'last_name']
user_profile = upa.get(TEST_USER_ID, fields=fields)
assert user_profile is not None
assert len(user_profile.keys()) == len(fields)
| Fix user profile test to include same environment variables | Fix user profile test to include same environment variables
| Python | mit | karlinnolabs/pymessenger,Cretezy/pymessenger2,davidchua/pymessenger | import json
+ import sys, os
+ sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/.."))
from pymessenger.user_profile import UserProfileApi
- from test_env import *
+ TOKEN = os.environ.get('TOKEN')
+ APP_SECRET = os.environ.get('APP_SECRET')
+ TEST_USER_ID = os.environ.get('RECIPIENT_ID')
+
- upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET)
+ upa = UserProfileApi(TOKEN, app_secret=APP_SECRET)
def test_fields_blank():
user_profile = upa.get(TEST_USER_ID)
assert user_profile is not None
def test_fields():
fields = ['first_name', 'last_name']
user_profile = upa.get(TEST_USER_ID, fields=fields)
assert user_profile is not None
assert len(user_profile.keys()) == len(fields)
| Fix user profile test to include same environment variables | ## Code Before:
import json
from pymessenger.user_profile import UserProfileApi
from test_env import *
upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET)
def test_fields_blank():
user_profile = upa.get(TEST_USER_ID)
assert user_profile is not None
def test_fields():
fields = ['first_name', 'last_name']
user_profile = upa.get(TEST_USER_ID, fields=fields)
assert user_profile is not None
assert len(user_profile.keys()) == len(fields)
## Instruction:
Fix user profile test to include same environment variables
## Code After:
import json
import sys, os
sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/.."))
from pymessenger.user_profile import UserProfileApi
TOKEN = os.environ.get('TOKEN')
APP_SECRET = os.environ.get('APP_SECRET')
TEST_USER_ID = os.environ.get('RECIPIENT_ID')
upa = UserProfileApi(TOKEN, app_secret=APP_SECRET)
def test_fields_blank():
user_profile = upa.get(TEST_USER_ID)
assert user_profile is not None
def test_fields():
fields = ['first_name', 'last_name']
user_profile = upa.get(TEST_USER_ID, fields=fields)
assert user_profile is not None
assert len(user_profile.keys()) == len(fields)
| import json
+ import sys, os
+ sys.path.append(os.path.realpath(os.path.dirname(__file__)+"/.."))
from pymessenger.user_profile import UserProfileApi
- from test_env import *
+ TOKEN = os.environ.get('TOKEN')
+ APP_SECRET = os.environ.get('APP_SECRET')
+ TEST_USER_ID = os.environ.get('RECIPIENT_ID')
+
- upa = UserProfileApi(PAGE_ACCESS_TOKEN, app_secret=APP_SECRET)
? ------------
+ upa = UserProfileApi(TOKEN, app_secret=APP_SECRET)
def test_fields_blank():
user_profile = upa.get(TEST_USER_ID)
assert user_profile is not None
def test_fields():
fields = ['first_name', 'last_name']
user_profile = upa.get(TEST_USER_ID, fields=fields)
assert user_profile is not None
assert len(user_profile.keys()) == len(fields) |
f3763c417d745463361b054fd4ffa0ddf35833eb | src/server/Universe.py | src/server/Universe.py | class Universe:
def __init__(self, height=100000000, width=100000000):
self.entities = []
self.height = height
self.width = width
self.teams = []
self.state = []
def add(self, entity):
self.entities.append(entity)
def remove(self, entity):
self.entities.remove(entity)
# Time passes and position updates during tick
def tick(self, duration):
for i in self.entities:
i.tick(duration)
# Position changed, so check for collisions
def collide(self):
for i in self.entities:
for j in self.entities:
if i != j:
if i.checkCollide(j):
i.collide(j)
# Now that damage is dealt in collisions, destroy objects and update logic
def tock(self):
for i in self.entities:
i.tock()
def dumpState(self):
return self.state
| class Universe:
def __init__(self, height=100000000, width=100000000):
self.entities = []
self.height = height
self.width = width
self.teams = []
self.state = []
self.maxID = 0
def add(self, entity):
maxID += 1
entity.id = maxID
self.entities.append(entity)
def remove(self, entity):
self.entities.remove(entity)
# Time passes and position updates during tick
def tick(self, duration):
for i in self.entities:
i.tick(duration)
# Position changed, so check for collisions
def collide(self):
for i in self.entities:
for j in self.entities:
if i != j:
if i.checkCollide(j):
i.collide(j)
# Now that damage is dealt in collisions, destroy objects and update logic
def tock(self):
for i in self.entities:
i.tock()
def dumpState(self):
return self.state
| Add ID to Entity upon creation | Add ID to Entity upon creation
| Python | mit | cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim | class Universe:
def __init__(self, height=100000000, width=100000000):
self.entities = []
self.height = height
self.width = width
self.teams = []
self.state = []
+ self.maxID = 0
def add(self, entity):
+ maxID += 1
+ entity.id = maxID
self.entities.append(entity)
def remove(self, entity):
self.entities.remove(entity)
# Time passes and position updates during tick
def tick(self, duration):
for i in self.entities:
i.tick(duration)
# Position changed, so check for collisions
def collide(self):
for i in self.entities:
for j in self.entities:
if i != j:
if i.checkCollide(j):
i.collide(j)
# Now that damage is dealt in collisions, destroy objects and update logic
def tock(self):
for i in self.entities:
i.tock()
def dumpState(self):
return self.state
| Add ID to Entity upon creation | ## Code Before:
class Universe:
def __init__(self, height=100000000, width=100000000):
self.entities = []
self.height = height
self.width = width
self.teams = []
self.state = []
def add(self, entity):
self.entities.append(entity)
def remove(self, entity):
self.entities.remove(entity)
# Time passes and position updates during tick
def tick(self, duration):
for i in self.entities:
i.tick(duration)
# Position changed, so check for collisions
def collide(self):
for i in self.entities:
for j in self.entities:
if i != j:
if i.checkCollide(j):
i.collide(j)
# Now that damage is dealt in collisions, destroy objects and update logic
def tock(self):
for i in self.entities:
i.tock()
def dumpState(self):
return self.state
## Instruction:
Add ID to Entity upon creation
## Code After:
class Universe:
def __init__(self, height=100000000, width=100000000):
self.entities = []
self.height = height
self.width = width
self.teams = []
self.state = []
self.maxID = 0
def add(self, entity):
maxID += 1
entity.id = maxID
self.entities.append(entity)
def remove(self, entity):
self.entities.remove(entity)
# Time passes and position updates during tick
def tick(self, duration):
for i in self.entities:
i.tick(duration)
# Position changed, so check for collisions
def collide(self):
for i in self.entities:
for j in self.entities:
if i != j:
if i.checkCollide(j):
i.collide(j)
# Now that damage is dealt in collisions, destroy objects and update logic
def tock(self):
for i in self.entities:
i.tock()
def dumpState(self):
return self.state
| class Universe:
def __init__(self, height=100000000, width=100000000):
self.entities = []
self.height = height
self.width = width
self.teams = []
self.state = []
+ self.maxID = 0
def add(self, entity):
+ maxID += 1
+ entity.id = maxID
self.entities.append(entity)
def remove(self, entity):
self.entities.remove(entity)
# Time passes and position updates during tick
def tick(self, duration):
for i in self.entities:
i.tick(duration)
# Position changed, so check for collisions
def collide(self):
for i in self.entities:
for j in self.entities:
if i != j:
if i.checkCollide(j):
i.collide(j)
# Now that damage is dealt in collisions, destroy objects and update logic
def tock(self):
for i in self.entities:
i.tock()
def dumpState(self):
return self.state |
88752efa9ac2c0f251733e335763cb880da34741 | thinglang/parser/definitions/member_definition.py | thinglang/parser/definitions/member_definition.py | from thinglang.lexer.definitions.tags import LexicalPrivateTag
from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember
from thinglang.lexer.values.identifier import Identifier
from thinglang.parser.nodes import BaseNode
from thinglang.parser.rule import ParserRule
from thinglang.symbols.symbol import Symbol
class MemberDefinition(BaseNode):
"""
A member definition
Must be a direct child of a ThingDefinition
"""
def __init__(self, name, type_name, visibility=Symbol.PUBLIC):
super(MemberDefinition, self).__init__([name, type_name])
self.type, self.name, self.visibility = type_name, name, visibility
def __repr__(self):
return 'has {} {}'.format(self.type, self.name)
def symbol(self):
return Symbol.member(self.name, self.type, self.visibility)
MEMBER_NAME_TYPES = Identifier
@staticmethod
@ParserRule.mark
def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier):
return MemberDefinition(name, type_name)
@staticmethod
@ParserRule.mark
def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'):
member.visibility = Symbol.PRIVATE
return member
| from thinglang.lexer.definitions.tags import LexicalPrivateTag
from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember
from thinglang.lexer.values.identifier import Identifier
from thinglang.parser.nodes import BaseNode
from thinglang.parser.rule import ParserRule
from thinglang.symbols.symbol import Symbol
class MemberDefinition(BaseNode):
"""
A member definition
Must be a direct child of a ThingDefinition
"""
def __init__(self, name, type_name, visibility=Symbol.PUBLIC):
super(MemberDefinition, self).__init__([name, type_name])
self.type, self.name, self.visibility = type_name, name, visibility
def __repr__(self):
return 'has {} {}'.format(self.type, self.name)
def symbol(self):
return Symbol.member(self.name, self.type, self.visibility)
MEMBER_NAME_TYPES = Identifier
@staticmethod
@ParserRule.mark
def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier):
return MemberDefinition(name, type_name)
| Add visibility tagging to MethoDefinition | Add visibility tagging to MethoDefinition
| Python | mit | ytanay/thinglang,ytanay/thinglang,ytanay/thinglang,ytanay/thinglang | from thinglang.lexer.definitions.tags import LexicalPrivateTag
from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember
from thinglang.lexer.values.identifier import Identifier
from thinglang.parser.nodes import BaseNode
from thinglang.parser.rule import ParserRule
from thinglang.symbols.symbol import Symbol
class MemberDefinition(BaseNode):
"""
A member definition
Must be a direct child of a ThingDefinition
"""
def __init__(self, name, type_name, visibility=Symbol.PUBLIC):
super(MemberDefinition, self).__init__([name, type_name])
self.type, self.name, self.visibility = type_name, name, visibility
def __repr__(self):
return 'has {} {}'.format(self.type, self.name)
def symbol(self):
return Symbol.member(self.name, self.type, self.visibility)
MEMBER_NAME_TYPES = Identifier
@staticmethod
@ParserRule.mark
- def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier):
+ def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier):
return MemberDefinition(name, type_name)
- @staticmethod
- @ParserRule.mark
- def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'):
- member.visibility = Symbol.PRIVATE
- return member
- | Add visibility tagging to MethoDefinition | ## Code Before:
from thinglang.lexer.definitions.tags import LexicalPrivateTag
from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember
from thinglang.lexer.values.identifier import Identifier
from thinglang.parser.nodes import BaseNode
from thinglang.parser.rule import ParserRule
from thinglang.symbols.symbol import Symbol
class MemberDefinition(BaseNode):
"""
A member definition
Must be a direct child of a ThingDefinition
"""
def __init__(self, name, type_name, visibility=Symbol.PUBLIC):
super(MemberDefinition, self).__init__([name, type_name])
self.type, self.name, self.visibility = type_name, name, visibility
def __repr__(self):
return 'has {} {}'.format(self.type, self.name)
def symbol(self):
return Symbol.member(self.name, self.type, self.visibility)
MEMBER_NAME_TYPES = Identifier
@staticmethod
@ParserRule.mark
def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier):
return MemberDefinition(name, type_name)
@staticmethod
@ParserRule.mark
def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'):
member.visibility = Symbol.PRIVATE
return member
## Instruction:
Add visibility tagging to MethoDefinition
## Code After:
from thinglang.lexer.definitions.tags import LexicalPrivateTag
from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember
from thinglang.lexer.values.identifier import Identifier
from thinglang.parser.nodes import BaseNode
from thinglang.parser.rule import ParserRule
from thinglang.symbols.symbol import Symbol
class MemberDefinition(BaseNode):
"""
A member definition
Must be a direct child of a ThingDefinition
"""
def __init__(self, name, type_name, visibility=Symbol.PUBLIC):
super(MemberDefinition, self).__init__([name, type_name])
self.type, self.name, self.visibility = type_name, name, visibility
def __repr__(self):
return 'has {} {}'.format(self.type, self.name)
def symbol(self):
return Symbol.member(self.name, self.type, self.visibility)
MEMBER_NAME_TYPES = Identifier
@staticmethod
@ParserRule.mark
def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier):
return MemberDefinition(name, type_name)
| from thinglang.lexer.definitions.tags import LexicalPrivateTag
from thinglang.lexer.definitions.thing_definition import LexicalDeclarationMember
from thinglang.lexer.values.identifier import Identifier
from thinglang.parser.nodes import BaseNode
from thinglang.parser.rule import ParserRule
from thinglang.symbols.symbol import Symbol
class MemberDefinition(BaseNode):
"""
A member definition
Must be a direct child of a ThingDefinition
"""
def __init__(self, name, type_name, visibility=Symbol.PUBLIC):
super(MemberDefinition, self).__init__([name, type_name])
self.type, self.name, self.visibility = type_name, name, visibility
def __repr__(self):
return 'has {} {}'.format(self.type, self.name)
def symbol(self):
return Symbol.member(self.name, self.type, self.visibility)
MEMBER_NAME_TYPES = Identifier
@staticmethod
@ParserRule.mark
- def parse_member_definition(_: LexicalDeclarationMember, type_name: MEMBER_NAME_TYPES, name: Identifier):
+ def parse_member_definition(_: (LexicalDeclarationMember, LexicalPrivateTag), type_name: MEMBER_NAME_TYPES, name: Identifier):
? + ++++++++++++++++++++
return MemberDefinition(name, type_name)
-
- @staticmethod
- @ParserRule.mark
- def tag_member_definition(_: LexicalPrivateTag, member: 'MemberDefinition'):
- member.visibility = Symbol.PRIVATE
- return member |
f300f3b31dcdefa91fa8fe46bdaab2d2490ac06a | snd/image_board/serializers.py | snd/image_board/serializers.py | from django.contrib.auth.models import User
from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name')
class ProfileSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Profile
fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo')
class ContentItemSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentItem
fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
class HashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Hashtag
fields = ('id', 'hashtag_text')
class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentHashTag
fields = ('id', 'content_id', 'hashtag_id')
class LikeSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Like
fields = ('id', 'user_id', 'content_id')
class CommentSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Comment
fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem')
| from django.contrib.auth.models import User
from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name')
class ProfileSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Profile
fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo')
class ContentItemSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentItem
fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
class HashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Hashtag
fields = ('id', 'url', 'hashtag_text')
class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentHashTag
fields = ('id', 'url', 'content_id', 'hashtag_id')
class LikeSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Like
fields = ('id', 'url', 'user_id', 'content_id')
class CommentSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Comment
fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem')
| Add URLs to each searializer | Add URLs to each searializer
| Python | mit | SNDjango/server,SNDjango/server,SNDjango/server | from django.contrib.auth.models import User
from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name')
class ProfileSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Profile
- fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
+ fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo')
class ContentItemSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentItem
- fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
+ fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
class HashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Hashtag
- fields = ('id', 'hashtag_text')
+ fields = ('id', 'url', 'hashtag_text')
class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentHashTag
- fields = ('id', 'content_id', 'hashtag_id')
+ fields = ('id', 'url', 'content_id', 'hashtag_id')
class LikeSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Like
- fields = ('id', 'user_id', 'content_id')
+ fields = ('id', 'url', 'user_id', 'content_id')
class CommentSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Comment
- fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem')
+ fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem')
| Add URLs to each searializer | ## Code Before:
from django.contrib.auth.models import User
from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name')
class ProfileSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Profile
fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo')
class ContentItemSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentItem
fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
class HashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Hashtag
fields = ('id', 'hashtag_text')
class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentHashTag
fields = ('id', 'content_id', 'hashtag_id')
class LikeSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Like
fields = ('id', 'user_id', 'content_id')
class CommentSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Comment
fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem')
## Instruction:
Add URLs to each searializer
## Code After:
from django.contrib.auth.models import User
from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name')
class ProfileSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Profile
fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo')
class ContentItemSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentItem
fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
class HashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Hashtag
fields = ('id', 'url', 'hashtag_text')
class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentHashTag
fields = ('id', 'url', 'content_id', 'hashtag_id')
class LikeSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Like
fields = ('id', 'url', 'user_id', 'content_id')
class CommentSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Comment
fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem')
| from django.contrib.auth.models import User
from .models import ContentItem, Profile, Comment, Hashtag, ContentHashTag, Like
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('id', 'url', 'username', 'email', 'last_name', 'first_name')
class ProfileSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Profile
- fields = ('user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
+ fields = ('id', 'url', 'user', 'personal_info', 'job_title', 'department', 'location', 'expertise',
? +++++++++++++
'phone_number', 'contact_skype', 'contact_facebook', 'contact_linkedin', 'user_photo')
class ContentItemSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentItem
- fields = ('id', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
+ fields = ('id', 'url', 'upload_date', 'title', 'description', 'image', 'uploaded_by')
? +++++++
class HashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Hashtag
- fields = ('id', 'hashtag_text')
+ fields = ('id', 'url', 'hashtag_text')
? +++++++
class ContentHashtagSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ContentHashTag
- fields = ('id', 'content_id', 'hashtag_id')
+ fields = ('id', 'url', 'content_id', 'hashtag_id')
? +++++++
class LikeSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Like
- fields = ('id', 'user_id', 'content_id')
+ fields = ('id', 'url', 'user_id', 'content_id')
? +++++++
class CommentSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Comment
- fields = ('id', 'comment_text', 'publication_date', 'author', 'contentItem')
+ fields = ('id', 'url', 'comment_text', 'publication_date', 'author', 'contentItem')
? +++++++
|
0cb5447de992389be9587d7706637212bfe3b90b | tests/events/tests.py | tests/events/tests.py |
from mock import Mock
from unittest2 import TestCase
from raven.events import Message
class MessageTest(TestCase):
def test_to_string(self):
unformatted_message = 'My message from %s about %s'
client = Mock()
message = Message(client)
message.logger = Mock()
data = {
'sentry.interfaces.Message': {
'message': unformatted_message,
}
}
self.assertEqual(message.to_string(data), unformatted_message)
self.assertEqual(message.logger.warn.call_count, 1)
args, kwargs = message.logger.warn.call_args
self.assertEqual(args, ('Unable to find params for message',))
self.assertEqual(kwargs,
{'extra': {'msg': {'message': unformatted_message}}})
data['sentry.interfaces.Message']['params'] = (1, 2)
self.assertEqual(message.to_string(data),
unformatted_message % (1, 2))
|
from mock import Mock
from unittest2 import TestCase
from raven.events import Message
class MessageTest(TestCase):
def test_to_string(self):
unformatted_message = 'My message from %s about %s'
client = Mock()
message = Message(client)
message.logger = Mock()
data = {
'sentry.interfaces.Message': {
'message': unformatted_message,
}
}
self.assertEqual(message.to_string(data), unformatted_message)
data['sentry.interfaces.Message']['params'] = (1, 2)
self.assertEqual(message.to_string(data),
unformatted_message % (1, 2))
| Update test to match current behavior | Update test to match current behavior
| Python | bsd-3-clause | johansteffner/raven-python,Photonomie/raven-python,jbarbuto/raven-python,nikolas/raven-python,lepture/raven-python,smarkets/raven-python,arthurlogilab/raven-python,lepture/raven-python,hzy/raven-python,recht/raven-python,inspirehep/raven-python,nikolas/raven-python,openlabs/raven,patrys/opbeat_python,ewdurbin/raven-python,dbravender/raven-python,akalipetis/raven-python,alex/raven,Photonomie/raven-python,jmp0xf/raven-python,akalipetis/raven-python,jbarbuto/raven-python,smarkets/raven-python,nikolas/raven-python,smarkets/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,lopter/raven-python-old,beniwohli/apm-agent-python,percipient/raven-python,icereval/raven-python,percipient/raven-python,ronaldevers/raven-python,danriti/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akalipetis/raven-python,akheron/raven-python,dirtycoder/opbeat_python,hzy/raven-python,hzy/raven-python,patrys/opbeat_python,ewdurbin/raven-python,jmagnusson/raven-python,ronaldevers/raven-python,icereval/raven-python,collective/mr.poe,jmp0xf/raven-python,ticosax/opbeat_python,arthurlogilab/raven-python,tarkatronic/opbeat_python,icereval/raven-python,jbarbuto/raven-python,inspirehep/raven-python,ronaldevers/raven-python,akheron/raven-python,danriti/raven-python,daikeren/opbeat_python,johansteffner/raven-python,dbravender/raven-python,tarkatronic/opbeat_python,daikeren/opbeat_python,johansteffner/raven-python,ticosax/opbeat_python,someonehan/raven-python,Photonomie/raven-python,someonehan/raven-python,inspirehep/raven-python,dirtycoder/opbeat_python,beniwohli/apm-agent-python,beniwohli/apm-agent-python,jmagnusson/raven-python,percipient/raven-python,recht/raven-python,jbarbuto/raven-python,someonehan/raven-python,dirtycoder/opbeat_python,getsentry/raven-python,getsentry/raven-python,jmp0xf/raven-python,tarkatronic/opbeat_python,dbravender/raven-python,inspirehep/raven-python,akheron/raven-python,smarkets/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,patrys/opbeat_python,jmagnusson/raven-python,danriti/raven-python,nikolas/raven-python,getsentry/raven-python,icereval/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,arthurlogilab/raven-python,patrys/opbeat_python,arthurlogilab/raven-python,recht/raven-python,lepture/raven-python,daikeren/opbeat_python,ticosax/opbeat_python,beniwohli/apm-agent-python,ewdurbin/raven-python |
from mock import Mock
from unittest2 import TestCase
from raven.events import Message
class MessageTest(TestCase):
def test_to_string(self):
unformatted_message = 'My message from %s about %s'
client = Mock()
message = Message(client)
message.logger = Mock()
data = {
'sentry.interfaces.Message': {
'message': unformatted_message,
}
}
self.assertEqual(message.to_string(data), unformatted_message)
- self.assertEqual(message.logger.warn.call_count, 1)
-
- args, kwargs = message.logger.warn.call_args
- self.assertEqual(args, ('Unable to find params for message',))
- self.assertEqual(kwargs,
- {'extra': {'msg': {'message': unformatted_message}}})
data['sentry.interfaces.Message']['params'] = (1, 2)
self.assertEqual(message.to_string(data),
unformatted_message % (1, 2))
| Update test to match current behavior | ## Code Before:
from mock import Mock
from unittest2 import TestCase
from raven.events import Message
class MessageTest(TestCase):
def test_to_string(self):
unformatted_message = 'My message from %s about %s'
client = Mock()
message = Message(client)
message.logger = Mock()
data = {
'sentry.interfaces.Message': {
'message': unformatted_message,
}
}
self.assertEqual(message.to_string(data), unformatted_message)
self.assertEqual(message.logger.warn.call_count, 1)
args, kwargs = message.logger.warn.call_args
self.assertEqual(args, ('Unable to find params for message',))
self.assertEqual(kwargs,
{'extra': {'msg': {'message': unformatted_message}}})
data['sentry.interfaces.Message']['params'] = (1, 2)
self.assertEqual(message.to_string(data),
unformatted_message % (1, 2))
## Instruction:
Update test to match current behavior
## Code After:
from mock import Mock
from unittest2 import TestCase
from raven.events import Message
class MessageTest(TestCase):
def test_to_string(self):
unformatted_message = 'My message from %s about %s'
client = Mock()
message = Message(client)
message.logger = Mock()
data = {
'sentry.interfaces.Message': {
'message': unformatted_message,
}
}
self.assertEqual(message.to_string(data), unformatted_message)
data['sentry.interfaces.Message']['params'] = (1, 2)
self.assertEqual(message.to_string(data),
unformatted_message % (1, 2))
|
from mock import Mock
from unittest2 import TestCase
from raven.events import Message
class MessageTest(TestCase):
def test_to_string(self):
unformatted_message = 'My message from %s about %s'
client = Mock()
message = Message(client)
message.logger = Mock()
data = {
'sentry.interfaces.Message': {
'message': unformatted_message,
}
}
self.assertEqual(message.to_string(data), unformatted_message)
- self.assertEqual(message.logger.warn.call_count, 1)
-
- args, kwargs = message.logger.warn.call_args
- self.assertEqual(args, ('Unable to find params for message',))
- self.assertEqual(kwargs,
- {'extra': {'msg': {'message': unformatted_message}}})
data['sentry.interfaces.Message']['params'] = (1, 2)
self.assertEqual(message.to_string(data),
unformatted_message % (1, 2)) |
af45e43c46a22f3168ab946bf914a45eae9ade19 | avatar/urls.py | avatar/urls.py | try:
from django.conf.urls import patterns, url
except ImportError:
# Django < 1.4
from django.conf.urls.defaults import patterns, url
from avatar import views
urlpatterns = patterns('',
url(r'^add/$', views.add, name='avatar_add'),
url(r'^change/$', views.change, name='avatar_change'),
url(r'^delete/$', views.delete, name='avatar_delete'),
url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$',
views.render_primary,
name='avatar_render_primary'),
url(r'^list/(?P<username>[\+\w\@\.]+)/$',
views.avatar_gallery,
name='avatar_gallery'),
url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$',
views.avatar,
name='avatar'),
)
| try:
from django.conf.urls import patterns, url
except ImportError:
# Django < 1.4
from django.conf.urls.defaults import url
from avatar import views
urlpatterns = [
url(r'^add/$', views.add, name='avatar_add'),
url(r'^change/$', views.change, name='avatar_change'),
url(r'^delete/$', views.delete, name='avatar_delete'),
url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$',
views.render_primary,
name='avatar_render_primary'),
url(r'^list/(?P<username>[\+\w\@\.]+)/$',
views.avatar_gallery,
name='avatar_gallery'),
url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$',
views.avatar,
name='avatar'),
]
| Remove replace urlpatterns with simple array, make compatible with Django 1.9 | Remove replace urlpatterns with simple array, make compatible with Django 1.9
| Python | bsd-3-clause | ad-m/django-avatar,ad-m/django-avatar,grantmcconnaughey/django-avatar,grantmcconnaughey/django-avatar,jezdez/django-avatar,jezdez/django-avatar,MachineandMagic/django-avatar,MachineandMagic/django-avatar | try:
from django.conf.urls import patterns, url
except ImportError:
# Django < 1.4
- from django.conf.urls.defaults import patterns, url
+ from django.conf.urls.defaults import url
from avatar import views
- urlpatterns = patterns('',
+ urlpatterns = [
url(r'^add/$', views.add, name='avatar_add'),
url(r'^change/$', views.change, name='avatar_change'),
url(r'^delete/$', views.delete, name='avatar_delete'),
url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$',
views.render_primary,
name='avatar_render_primary'),
url(r'^list/(?P<username>[\+\w\@\.]+)/$',
views.avatar_gallery,
name='avatar_gallery'),
url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$',
views.avatar,
name='avatar'),
- )
+ ]
| Remove replace urlpatterns with simple array, make compatible with Django 1.9 | ## Code Before:
try:
from django.conf.urls import patterns, url
except ImportError:
# Django < 1.4
from django.conf.urls.defaults import patterns, url
from avatar import views
urlpatterns = patterns('',
url(r'^add/$', views.add, name='avatar_add'),
url(r'^change/$', views.change, name='avatar_change'),
url(r'^delete/$', views.delete, name='avatar_delete'),
url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$',
views.render_primary,
name='avatar_render_primary'),
url(r'^list/(?P<username>[\+\w\@\.]+)/$',
views.avatar_gallery,
name='avatar_gallery'),
url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$',
views.avatar,
name='avatar'),
)
## Instruction:
Remove replace urlpatterns with simple array, make compatible with Django 1.9
## Code After:
try:
from django.conf.urls import patterns, url
except ImportError:
# Django < 1.4
from django.conf.urls.defaults import url
from avatar import views
urlpatterns = [
url(r'^add/$', views.add, name='avatar_add'),
url(r'^change/$', views.change, name='avatar_change'),
url(r'^delete/$', views.delete, name='avatar_delete'),
url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$',
views.render_primary,
name='avatar_render_primary'),
url(r'^list/(?P<username>[\+\w\@\.]+)/$',
views.avatar_gallery,
name='avatar_gallery'),
url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$',
views.avatar,
name='avatar'),
]
| try:
from django.conf.urls import patterns, url
except ImportError:
# Django < 1.4
- from django.conf.urls.defaults import patterns, url
? ----------
+ from django.conf.urls.defaults import url
from avatar import views
- urlpatterns = patterns('',
+ urlpatterns = [
url(r'^add/$', views.add, name='avatar_add'),
url(r'^change/$', views.change, name='avatar_change'),
url(r'^delete/$', views.delete, name='avatar_delete'),
url(r'^render_primary/(?P<user>[\w\d\@\.\-_]{3,30})/(?P<size>[\d]+)/$',
views.render_primary,
name='avatar_render_primary'),
url(r'^list/(?P<username>[\+\w\@\.]+)/$',
views.avatar_gallery,
name='avatar_gallery'),
url(r'^list/(?P<username>[\+\w\@\.]+)/(?P<id>[\d]+)/$',
views.avatar,
name='avatar'),
- )
+ ] |
c7ac2b1805bd82bdeef3227fc34149431c161df7 | yesimeanit/showoff/newsletter_subscriptions/admin.py | yesimeanit/showoff/newsletter_subscriptions/admin.py | from django.contrib import admin
from .models import NewsletterSubscription
admin.site.register(NewsletterSubscription,
list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'),
list_filter=('is_active',),
)
| from django.contrib import admin
from .models import NewsletterSubscription
admin.site.register(NewsletterSubscription,
list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'),
list_filter=('is_active',),
search_fields=('code', 'email', 'first_name', 'last_name'),
)
| Add search fields for newsletter subscriptions | Add search fields for newsletter subscriptions
| Python | bsd-3-clause | guetux/django-yesimeanit | from django.contrib import admin
from .models import NewsletterSubscription
admin.site.register(NewsletterSubscription,
list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'),
list_filter=('is_active',),
+ search_fields=('code', 'email', 'first_name', 'last_name'),
)
| Add search fields for newsletter subscriptions | ## Code Before:
from django.contrib import admin
from .models import NewsletterSubscription
admin.site.register(NewsletterSubscription,
list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'),
list_filter=('is_active',),
)
## Instruction:
Add search fields for newsletter subscriptions
## Code After:
from django.contrib import admin
from .models import NewsletterSubscription
admin.site.register(NewsletterSubscription,
list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'),
list_filter=('is_active',),
search_fields=('code', 'email', 'first_name', 'last_name'),
)
| from django.contrib import admin
from .models import NewsletterSubscription
admin.site.register(NewsletterSubscription,
list_display=('email', 'is_active', 'confirmed_on', 'unsubscribed_on'),
list_filter=('is_active',),
+ search_fields=('code', 'email', 'first_name', 'last_name'),
) |
c01a858306d31a5b12e42f30ff01bdbdb2240092 | froide/publicbody/tests.py | froide/publicbody/tests.py |
from django.test import TestCase
class SimpleTest(TestCase):
def test_basic_addition(self):
"""
Tests that 1 + 1 always equals 2.
"""
self.assertEqual(1 + 1, 2)
| from django.test import TestCase
from django.core.urlresolvers import reverse
from publicbody.models import PublicBody
class PublicBodyTest(TestCase):
fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json']
def test_web_page(self):
response = self.client.get(reverse('publicbody-list'))
self.assertEqual(response.status_code, 200)
pb = PublicBody.objects.all()[0]
response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug}))
self.assertEqual(response.status_code, 200)
response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"}))
self.assertEqual(response.status_code, 200)
self.assertEqual(response['Content-Type'], 'application/json')
self.assertIn('"name":', response.content)
self.assertIn('"laws": [{', response.content)
response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"}))
self.assertEqual(response.status_code, 200)
self.assertEqual(response['Content-Type'], 'application/json')
def test_csv(self):
csv = PublicBody.export_csv()
self.assertTrue(csv)
def test_search(self):
response = self.client.get(reverse('publicbody-search')+"?q=umwelt")
self.assertIn("Umweltbundesamt", response.content)
self.assertEqual(response['Content-Type'], 'application/json')
| Test public body showing, json view and csv export | Test public body showing, json view and csv export | Python | mit | okfse/froide,ryankanno/froide,catcosmo/froide,ryankanno/froide,okfse/froide,LilithWittmann/froide,okfse/froide,LilithWittmann/froide,ryankanno/froide,CodeforHawaii/froide,stefanw/froide,stefanw/froide,LilithWittmann/froide,CodeforHawaii/froide,catcosmo/froide,catcosmo/froide,stefanw/froide,ryankanno/froide,fin/froide,fin/froide,catcosmo/froide,okfse/froide,fin/froide,LilithWittmann/froide,stefanw/froide,catcosmo/froide,ryankanno/froide,LilithWittmann/froide,fin/froide,stefanw/froide,CodeforHawaii/froide,CodeforHawaii/froide,CodeforHawaii/froide,okfse/froide | + from django.test import TestCase
+ from django.core.urlresolvers import reverse
- from django.test import TestCase
+ from publicbody.models import PublicBody
+ class PublicBodyTest(TestCase):
+ fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json']
- class SimpleTest(TestCase):
- def test_basic_addition(self):
- """
- Tests that 1 + 1 always equals 2.
- """
- self.assertEqual(1 + 1, 2)
+ def test_web_page(self):
+ response = self.client.get(reverse('publicbody-list'))
+ self.assertEqual(response.status_code, 200)
+ pb = PublicBody.objects.all()[0]
+ response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug}))
+ self.assertEqual(response.status_code, 200)
+ response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"}))
+ self.assertEqual(response.status_code, 200)
+ self.assertEqual(response['Content-Type'], 'application/json')
+ self.assertIn('"name":', response.content)
+ self.assertIn('"laws": [{', response.content)
+ response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"}))
+ self.assertEqual(response.status_code, 200)
+ self.assertEqual(response['Content-Type'], 'application/json')
+ def test_csv(self):
+ csv = PublicBody.export_csv()
+ self.assertTrue(csv)
+
+ def test_search(self):
+ response = self.client.get(reverse('publicbody-search')+"?q=umwelt")
+ self.assertIn("Umweltbundesamt", response.content)
+ self.assertEqual(response['Content-Type'], 'application/json')
+ | Test public body showing, json view and csv export | ## Code Before:
from django.test import TestCase
class SimpleTest(TestCase):
def test_basic_addition(self):
"""
Tests that 1 + 1 always equals 2.
"""
self.assertEqual(1 + 1, 2)
## Instruction:
Test public body showing, json view and csv export
## Code After:
from django.test import TestCase
from django.core.urlresolvers import reverse
from publicbody.models import PublicBody
class PublicBodyTest(TestCase):
fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json']
def test_web_page(self):
response = self.client.get(reverse('publicbody-list'))
self.assertEqual(response.status_code, 200)
pb = PublicBody.objects.all()[0]
response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug}))
self.assertEqual(response.status_code, 200)
response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"}))
self.assertEqual(response.status_code, 200)
self.assertEqual(response['Content-Type'], 'application/json')
self.assertIn('"name":', response.content)
self.assertIn('"laws": [{', response.content)
response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"}))
self.assertEqual(response.status_code, 200)
self.assertEqual(response['Content-Type'], 'application/json')
def test_csv(self):
csv = PublicBody.export_csv()
self.assertTrue(csv)
def test_search(self):
response = self.client.get(reverse('publicbody-search')+"?q=umwelt")
self.assertIn("Umweltbundesamt", response.content)
self.assertEqual(response['Content-Type'], 'application/json')
| + from django.test import TestCase
+ from django.core.urlresolvers import reverse
- from django.test import TestCase
+ from publicbody.models import PublicBody
+ class PublicBodyTest(TestCase):
+ fixtures = ['auth.json', 'publicbodies.json', 'foirequest.json']
- class SimpleTest(TestCase):
- def test_basic_addition(self):
- """
- Tests that 1 + 1 always equals 2.
- """
- self.assertEqual(1 + 1, 2)
+ def test_web_page(self):
+ response = self.client.get(reverse('publicbody-list'))
+ self.assertEqual(response.status_code, 200)
+ pb = PublicBody.objects.all()[0]
+ response = self.client.get(reverse('publicbody-show', kwargs={"slug": pb.slug}))
+ self.assertEqual(response.status_code, 200)
+ response = self.client.get(reverse('publicbody-show_json', kwargs={"pk": pb.pk, "format": "json"}))
+ self.assertEqual(response.status_code, 200)
+ self.assertEqual(response['Content-Type'], 'application/json')
+ self.assertIn('"name":', response.content)
+ self.assertIn('"laws": [{', response.content)
+ response = self.client.get(reverse('publicbody-show_json', kwargs={"slug": pb.slug, "format": "json"}))
+ self.assertEqual(response.status_code, 200)
+ self.assertEqual(response['Content-Type'], 'application/json')
+
+ def test_csv(self):
+ csv = PublicBody.export_csv()
+ self.assertTrue(csv)
+
+ def test_search(self):
+ response = self.client.get(reverse('publicbody-search')+"?q=umwelt")
+ self.assertIn("Umweltbundesamt", response.content)
+ self.assertEqual(response['Content-Type'], 'application/json') |
d154cd852bdb02743e9752179559a91b9f1a7f8c | example/tests/unit/test_renderer_class_methods.py | example/tests/unit/test_renderer_class_methods.py | from django.contrib.auth import get_user_model
from rest_framework_json_api import serializers
from rest_framework_json_api.renderers import JSONRenderer
class ResourceSerializer(serializers.ModelSerializer):
class Meta:
fields = ('username',)
model = get_user_model()
def test_build_json_resource_obj():
resource = {
'pk': 1,
'username': 'Alice',
}
serializer = ResourceSerializer(data={'username': 'Alice'})
serializer.is_valid()
resource_instance = serializer.save()
output = {
'type': 'user',
'id': '1',
'attributes': {
'username': 'Alice'
},
}
assert JSONRenderer.build_json_resource_obj(
serializer.fields, resource, resource_instance, 'user') == output
def test_extract_attributes():
fields = {
'id': serializers.Field(),
'username': serializers.Field(),
'deleted': serializers.ReadOnlyField(),
}
resource = {'id': 1, 'deleted': None, 'username': 'jerel'}
expected = {
'username': 'jerel',
'deleted': None
}
assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted'
assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted(
{'username': ''}), 'Should not extract read_only fields on empty serializer'
| from django.contrib.auth import get_user_model
from rest_framework_json_api import serializers
from rest_framework_json_api.renderers import JSONRenderer
pytestmark = pytest.mark.django_db
class ResourceSerializer(serializers.ModelSerializer):
class Meta:
fields = ('username',)
model = get_user_model()
def test_build_json_resource_obj():
resource = {
'pk': 1,
'username': 'Alice',
}
serializer = ResourceSerializer(data={'username': 'Alice'})
serializer.is_valid()
resource_instance = serializer.save()
output = {
'type': 'user',
'id': '1',
'attributes': {
'username': 'Alice'
},
}
assert JSONRenderer.build_json_resource_obj(
serializer.fields, resource, resource_instance, 'user') == output
def test_extract_attributes():
fields = {
'id': serializers.Field(),
'username': serializers.Field(),
'deleted': serializers.ReadOnlyField(),
}
resource = {'id': 1, 'deleted': None, 'username': 'jerel'}
expected = {
'username': 'jerel',
'deleted': None
}
assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted'
assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted(
{'username': ''}), 'Should not extract read_only fields on empty serializer'
| Fix for Database access not allowed, use the "django_db" mark to enable it. | Fix for Database access not allowed, use the "django_db" mark to enable it.
| Python | bsd-2-clause | django-json-api/django-rest-framework-json-api,martinmaillard/django-rest-framework-json-api,schtibe/django-rest-framework-json-api,pombredanne/django-rest-framework-json-api,scottfisk/django-rest-framework-json-api,Instawork/django-rest-framework-json-api,leo-naeka/rest_framework_ember,django-json-api/django-rest-framework-json-api,leo-naeka/django-rest-framework-json-api,abdulhaq-e/django-rest-framework-json-api,lukaslundgren/django-rest-framework-json-api,django-json-api/rest_framework_ember | from django.contrib.auth import get_user_model
from rest_framework_json_api import serializers
from rest_framework_json_api.renderers import JSONRenderer
+ pytestmark = pytest.mark.django_db
class ResourceSerializer(serializers.ModelSerializer):
class Meta:
fields = ('username',)
model = get_user_model()
def test_build_json_resource_obj():
resource = {
'pk': 1,
'username': 'Alice',
}
serializer = ResourceSerializer(data={'username': 'Alice'})
serializer.is_valid()
resource_instance = serializer.save()
output = {
'type': 'user',
'id': '1',
'attributes': {
'username': 'Alice'
},
}
assert JSONRenderer.build_json_resource_obj(
serializer.fields, resource, resource_instance, 'user') == output
def test_extract_attributes():
fields = {
'id': serializers.Field(),
'username': serializers.Field(),
'deleted': serializers.ReadOnlyField(),
}
resource = {'id': 1, 'deleted': None, 'username': 'jerel'}
expected = {
'username': 'jerel',
'deleted': None
}
assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted'
assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted(
{'username': ''}), 'Should not extract read_only fields on empty serializer'
| Fix for Database access not allowed, use the "django_db" mark to enable it. | ## Code Before:
from django.contrib.auth import get_user_model
from rest_framework_json_api import serializers
from rest_framework_json_api.renderers import JSONRenderer
class ResourceSerializer(serializers.ModelSerializer):
class Meta:
fields = ('username',)
model = get_user_model()
def test_build_json_resource_obj():
resource = {
'pk': 1,
'username': 'Alice',
}
serializer = ResourceSerializer(data={'username': 'Alice'})
serializer.is_valid()
resource_instance = serializer.save()
output = {
'type': 'user',
'id': '1',
'attributes': {
'username': 'Alice'
},
}
assert JSONRenderer.build_json_resource_obj(
serializer.fields, resource, resource_instance, 'user') == output
def test_extract_attributes():
fields = {
'id': serializers.Field(),
'username': serializers.Field(),
'deleted': serializers.ReadOnlyField(),
}
resource = {'id': 1, 'deleted': None, 'username': 'jerel'}
expected = {
'username': 'jerel',
'deleted': None
}
assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted'
assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted(
{'username': ''}), 'Should not extract read_only fields on empty serializer'
## Instruction:
Fix for Database access not allowed, use the "django_db" mark to enable it.
## Code After:
from django.contrib.auth import get_user_model
from rest_framework_json_api import serializers
from rest_framework_json_api.renderers import JSONRenderer
pytestmark = pytest.mark.django_db
class ResourceSerializer(serializers.ModelSerializer):
class Meta:
fields = ('username',)
model = get_user_model()
def test_build_json_resource_obj():
resource = {
'pk': 1,
'username': 'Alice',
}
serializer = ResourceSerializer(data={'username': 'Alice'})
serializer.is_valid()
resource_instance = serializer.save()
output = {
'type': 'user',
'id': '1',
'attributes': {
'username': 'Alice'
},
}
assert JSONRenderer.build_json_resource_obj(
serializer.fields, resource, resource_instance, 'user') == output
def test_extract_attributes():
fields = {
'id': serializers.Field(),
'username': serializers.Field(),
'deleted': serializers.ReadOnlyField(),
}
resource = {'id': 1, 'deleted': None, 'username': 'jerel'}
expected = {
'username': 'jerel',
'deleted': None
}
assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted'
assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted(
{'username': ''}), 'Should not extract read_only fields on empty serializer'
| from django.contrib.auth import get_user_model
from rest_framework_json_api import serializers
from rest_framework_json_api.renderers import JSONRenderer
+ pytestmark = pytest.mark.django_db
class ResourceSerializer(serializers.ModelSerializer):
class Meta:
fields = ('username',)
model = get_user_model()
def test_build_json_resource_obj():
resource = {
'pk': 1,
'username': 'Alice',
}
serializer = ResourceSerializer(data={'username': 'Alice'})
serializer.is_valid()
resource_instance = serializer.save()
output = {
'type': 'user',
'id': '1',
'attributes': {
'username': 'Alice'
},
}
assert JSONRenderer.build_json_resource_obj(
serializer.fields, resource, resource_instance, 'user') == output
def test_extract_attributes():
fields = {
'id': serializers.Field(),
'username': serializers.Field(),
'deleted': serializers.ReadOnlyField(),
}
resource = {'id': 1, 'deleted': None, 'username': 'jerel'}
expected = {
'username': 'jerel',
'deleted': None
}
assert sorted(JSONRenderer.extract_attributes(fields, resource)) == sorted(expected), 'Regular fields should be extracted'
assert sorted(JSONRenderer.extract_attributes(fields, {})) == sorted(
{'username': ''}), 'Should not extract read_only fields on empty serializer' |
64dbe1d931edd38b4d731db18408e337d39e42c3 | cab/admin.py | cab/admin.py | from django.contrib import admin
from cab.models import Language, Snippet, SnippetFlag
class LanguageAdmin(admin.ModelAdmin):
prepopulated_fields = {'slug': ['name']}
class SnippetAdmin(admin.ModelAdmin):
list_display = ('id', 'title', 'author', 'rating_score', 'pub_date')
list_filter = ('language',)
date_hierarchy = 'pub_date'
search_fields = ('author__username', 'title', 'description', 'code',)
class SnippetFlagAdmin(admin.ModelAdmin):
list_display = ('snippet', 'flag')
list_filter = ('flag',)
actions = ['remove_and_ban']
def remove_and_ban(self, request, queryset):
for obj in queryset:
obj.remove_and_ban()
self.message_user(request, 'Snippets removed successfully')
remove_and_ban.short_description = 'Remove snippet and ban user'
admin.site.register(Language, LanguageAdmin)
admin.site.register(Snippet, SnippetAdmin)
admin.site.register(SnippetFlag, SnippetFlagAdmin)
| from django.contrib import admin
from cab.models import Language, Snippet, SnippetFlag
class LanguageAdmin(admin.ModelAdmin):
prepopulated_fields = {'slug': ['name']}
class SnippetAdmin(admin.ModelAdmin):
list_display = ('id', 'title', 'author', 'rating_score', 'pub_date')
list_filter = ('language',)
date_hierarchy = 'pub_date'
search_fields = ('author__username', 'title', 'description', 'code',)
raw_id_fields = ('author',)
class SnippetFlagAdmin(admin.ModelAdmin):
list_display = ('snippet', 'flag')
list_filter = ('flag',)
actions = ['remove_and_ban']
raw_id_fields = ('snippet', 'user',)
def remove_and_ban(self, request, queryset):
for obj in queryset:
obj.remove_and_ban()
self.message_user(request, 'Snippets removed successfully')
remove_and_ban.short_description = 'Remove snippet and ban user'
admin.site.register(Language, LanguageAdmin)
admin.site.register(Snippet, SnippetAdmin)
admin.site.register(SnippetFlag, SnippetFlagAdmin)
| Use raw_id_fields for users and snippets. | Use raw_id_fields for users and snippets.
| Python | bsd-3-clause | django/djangosnippets.org,django/djangosnippets.org,django/djangosnippets.org,django/djangosnippets.org,django-de/djangosnippets.org,django/djangosnippets.org,django-de/djangosnippets.org,django-de/djangosnippets.org,django-de/djangosnippets.org | from django.contrib import admin
from cab.models import Language, Snippet, SnippetFlag
class LanguageAdmin(admin.ModelAdmin):
prepopulated_fields = {'slug': ['name']}
class SnippetAdmin(admin.ModelAdmin):
list_display = ('id', 'title', 'author', 'rating_score', 'pub_date')
list_filter = ('language',)
date_hierarchy = 'pub_date'
search_fields = ('author__username', 'title', 'description', 'code',)
+ raw_id_fields = ('author',)
class SnippetFlagAdmin(admin.ModelAdmin):
list_display = ('snippet', 'flag')
list_filter = ('flag',)
actions = ['remove_and_ban']
+ raw_id_fields = ('snippet', 'user',)
def remove_and_ban(self, request, queryset):
for obj in queryset:
obj.remove_and_ban()
self.message_user(request, 'Snippets removed successfully')
remove_and_ban.short_description = 'Remove snippet and ban user'
admin.site.register(Language, LanguageAdmin)
admin.site.register(Snippet, SnippetAdmin)
admin.site.register(SnippetFlag, SnippetFlagAdmin)
| Use raw_id_fields for users and snippets. | ## Code Before:
from django.contrib import admin
from cab.models import Language, Snippet, SnippetFlag
class LanguageAdmin(admin.ModelAdmin):
prepopulated_fields = {'slug': ['name']}
class SnippetAdmin(admin.ModelAdmin):
list_display = ('id', 'title', 'author', 'rating_score', 'pub_date')
list_filter = ('language',)
date_hierarchy = 'pub_date'
search_fields = ('author__username', 'title', 'description', 'code',)
class SnippetFlagAdmin(admin.ModelAdmin):
list_display = ('snippet', 'flag')
list_filter = ('flag',)
actions = ['remove_and_ban']
def remove_and_ban(self, request, queryset):
for obj in queryset:
obj.remove_and_ban()
self.message_user(request, 'Snippets removed successfully')
remove_and_ban.short_description = 'Remove snippet and ban user'
admin.site.register(Language, LanguageAdmin)
admin.site.register(Snippet, SnippetAdmin)
admin.site.register(SnippetFlag, SnippetFlagAdmin)
## Instruction:
Use raw_id_fields for users and snippets.
## Code After:
from django.contrib import admin
from cab.models import Language, Snippet, SnippetFlag
class LanguageAdmin(admin.ModelAdmin):
prepopulated_fields = {'slug': ['name']}
class SnippetAdmin(admin.ModelAdmin):
list_display = ('id', 'title', 'author', 'rating_score', 'pub_date')
list_filter = ('language',)
date_hierarchy = 'pub_date'
search_fields = ('author__username', 'title', 'description', 'code',)
raw_id_fields = ('author',)
class SnippetFlagAdmin(admin.ModelAdmin):
list_display = ('snippet', 'flag')
list_filter = ('flag',)
actions = ['remove_and_ban']
raw_id_fields = ('snippet', 'user',)
def remove_and_ban(self, request, queryset):
for obj in queryset:
obj.remove_and_ban()
self.message_user(request, 'Snippets removed successfully')
remove_and_ban.short_description = 'Remove snippet and ban user'
admin.site.register(Language, LanguageAdmin)
admin.site.register(Snippet, SnippetAdmin)
admin.site.register(SnippetFlag, SnippetFlagAdmin)
| from django.contrib import admin
from cab.models import Language, Snippet, SnippetFlag
class LanguageAdmin(admin.ModelAdmin):
prepopulated_fields = {'slug': ['name']}
class SnippetAdmin(admin.ModelAdmin):
list_display = ('id', 'title', 'author', 'rating_score', 'pub_date')
list_filter = ('language',)
date_hierarchy = 'pub_date'
search_fields = ('author__username', 'title', 'description', 'code',)
+ raw_id_fields = ('author',)
class SnippetFlagAdmin(admin.ModelAdmin):
list_display = ('snippet', 'flag')
list_filter = ('flag',)
actions = ['remove_and_ban']
+ raw_id_fields = ('snippet', 'user',)
def remove_and_ban(self, request, queryset):
for obj in queryset:
obj.remove_and_ban()
self.message_user(request, 'Snippets removed successfully')
remove_and_ban.short_description = 'Remove snippet and ban user'
admin.site.register(Language, LanguageAdmin)
admin.site.register(Snippet, SnippetAdmin)
admin.site.register(SnippetFlag, SnippetFlagAdmin) |
fdf0daefac50de71a8c4f80a9ef877669ebea48b | byceps/services/tourney/transfer/models.py | byceps/services/tourney/transfer/models.py |
from typing import NewType
from uuid import UUID
from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@attrs(auto_attribs=True, frozen=True, slots=True)
class Match:
id: MatchID
|
from dataclasses import dataclass
from typing import NewType
from uuid import UUID
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@dataclass(frozen=True)
class Match:
id: MatchID
| Change tourney match transfer model from `attrs` to `dataclass` | Change tourney match transfer model from `attrs` to `dataclass`
| Python | bsd-3-clause | homeworkprod/byceps,homeworkprod/byceps,m-ober/byceps,homeworkprod/byceps,m-ober/byceps,m-ober/byceps |
+ from dataclasses import dataclass
from typing import NewType
from uuid import UUID
-
- from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
- @attrs(auto_attribs=True, frozen=True, slots=True)
+ @dataclass(frozen=True)
class Match:
id: MatchID
| Change tourney match transfer model from `attrs` to `dataclass` | ## Code Before:
from typing import NewType
from uuid import UUID
from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@attrs(auto_attribs=True, frozen=True, slots=True)
class Match:
id: MatchID
## Instruction:
Change tourney match transfer model from `attrs` to `dataclass`
## Code After:
from dataclasses import dataclass
from typing import NewType
from uuid import UUID
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@dataclass(frozen=True)
class Match:
id: MatchID
|
+ from dataclasses import dataclass
from typing import NewType
from uuid import UUID
-
- from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
- @attrs(auto_attribs=True, frozen=True, slots=True)
+ @dataclass(frozen=True)
class Match:
id: MatchID |
899f28e2cd7dbeb6227e8c56eef541cce1a424f4 | alertaclient/commands/cmd_heartbeat.py | alertaclient/commands/cmd_heartbeat.py | import os
import platform
import sys
import click
prog = os.path.basename(sys.argv[0])
@click.command('heartbeat', short_help='Send a heartbeat')
@click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1]))
@click.option('--tag', '-T', 'tags', multiple=True)
@click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale')
@click.option('--delete', '-D', metavar='ID', help='Delete hearbeat')
@click.pass_obj
def cli(obj, origin, tags, timeout, delete):
"""Send or delete a heartbeat."""
client = obj['client']
if delete:
if origin or tags or timeout:
raise click.UsageError('Option "--delete" is mutually exclusive.')
client.delete_heartbeat(delete)
else:
try:
heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout)
except Exception as e:
click.echo('ERROR: {}'.format(e))
sys.exit(1)
click.echo(heartbeat.id)
| import os
import platform
import sys
import click
prog = os.path.basename(sys.argv[0])
@click.command('heartbeat', short_help='Send a heartbeat')
@click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1]))
@click.option('--tag', '-T', 'tags', multiple=True)
@click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale')
@click.option('--delete', '-D', metavar='ID', help='Delete hearbeat')
@click.pass_obj
def cli(obj, origin, tags, timeout, delete):
"""Send or delete a heartbeat."""
client = obj['client']
if delete:
if origin or tags or timeout:
raise click.UsageError('Option "--delete" is mutually exclusive.')
client.delete_heartbeat(delete)
else:
try:
heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout)
except Exception as e:
click.echo('ERROR: {}'.format(e))
sys.exit(1)
click.echo(heartbeat.id)
| Add check that heartbeat timeout is integer | Add check that heartbeat timeout is integer
| Python | apache-2.0 | alerta/python-alerta-client,alerta/python-alerta-client,alerta/python-alerta | import os
import platform
import sys
import click
prog = os.path.basename(sys.argv[0])
@click.command('heartbeat', short_help='Send a heartbeat')
@click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1]))
@click.option('--tag', '-T', 'tags', multiple=True)
- @click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale')
+ @click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale')
@click.option('--delete', '-D', metavar='ID', help='Delete hearbeat')
@click.pass_obj
def cli(obj, origin, tags, timeout, delete):
"""Send or delete a heartbeat."""
client = obj['client']
if delete:
if origin or tags or timeout:
raise click.UsageError('Option "--delete" is mutually exclusive.')
client.delete_heartbeat(delete)
else:
try:
heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout)
except Exception as e:
click.echo('ERROR: {}'.format(e))
sys.exit(1)
click.echo(heartbeat.id)
| Add check that heartbeat timeout is integer | ## Code Before:
import os
import platform
import sys
import click
prog = os.path.basename(sys.argv[0])
@click.command('heartbeat', short_help='Send a heartbeat')
@click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1]))
@click.option('--tag', '-T', 'tags', multiple=True)
@click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale')
@click.option('--delete', '-D', metavar='ID', help='Delete hearbeat')
@click.pass_obj
def cli(obj, origin, tags, timeout, delete):
"""Send or delete a heartbeat."""
client = obj['client']
if delete:
if origin or tags or timeout:
raise click.UsageError('Option "--delete" is mutually exclusive.')
client.delete_heartbeat(delete)
else:
try:
heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout)
except Exception as e:
click.echo('ERROR: {}'.format(e))
sys.exit(1)
click.echo(heartbeat.id)
## Instruction:
Add check that heartbeat timeout is integer
## Code After:
import os
import platform
import sys
import click
prog = os.path.basename(sys.argv[0])
@click.command('heartbeat', short_help='Send a heartbeat')
@click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1]))
@click.option('--tag', '-T', 'tags', multiple=True)
@click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale')
@click.option('--delete', '-D', metavar='ID', help='Delete hearbeat')
@click.pass_obj
def cli(obj, origin, tags, timeout, delete):
"""Send or delete a heartbeat."""
client = obj['client']
if delete:
if origin or tags or timeout:
raise click.UsageError('Option "--delete" is mutually exclusive.')
client.delete_heartbeat(delete)
else:
try:
heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout)
except Exception as e:
click.echo('ERROR: {}'.format(e))
sys.exit(1)
click.echo(heartbeat.id)
| import os
import platform
import sys
import click
prog = os.path.basename(sys.argv[0])
@click.command('heartbeat', short_help='Send a heartbeat')
@click.option('--origin', default='{}/{}'.format(prog, platform.uname()[1]))
@click.option('--tag', '-T', 'tags', multiple=True)
- @click.option('--timeout', metavar='EXPIRES', help='Seconds before heartbeat is stale')
+ @click.option('--timeout', metavar='EXPIRES', type=int, help='Seconds before heartbeat is stale')
? ++++++++++
@click.option('--delete', '-D', metavar='ID', help='Delete hearbeat')
@click.pass_obj
def cli(obj, origin, tags, timeout, delete):
"""Send or delete a heartbeat."""
client = obj['client']
if delete:
if origin or tags or timeout:
raise click.UsageError('Option "--delete" is mutually exclusive.')
client.delete_heartbeat(delete)
else:
try:
heartbeat = client.heartbeat(origin=origin, tags=tags, timeout=timeout)
except Exception as e:
click.echo('ERROR: {}'.format(e))
sys.exit(1)
click.echo(heartbeat.id) |
bd7c0a9ac2d357ab635bf2948824256f1e6ddbec | src/carreralib/serial.py | src/carreralib/serial.py | from serial import serial_for_url
from .connection import BufferTooShort, Connection, TimeoutError
class SerialConnection(Connection):
def __init__(self, url, timeout=None):
self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout)
def close(self):
self.__serial.close()
def recv(self, maxlength=None):
buf = bytearray()
while True:
c = self.__serial.read()
if not c:
raise TimeoutError("Timeout waiting for serial data")
elif c == b"$" or c == b"#":
break
elif maxlength is not None and maxlength <= len(buf):
raise BufferTooShort("Buffer too short for data received")
else:
buf.extend(c)
return bytes(buf)
def send(self, buf, offset=0, size=None):
n = len(buf)
if offset < 0:
raise ValueError("offset is negative")
elif n < offset:
raise ValueError("buffer length < offset")
elif size is None:
size = n - offset
elif size < 0:
raise ValueError("size is negative")
elif offset + size > n:
raise ValueError("buffer length < offset + size")
self.__serial.write(b'"')
self.__serial.write(buf[offset : offset + size])
self.__serial.write(b"$")
self.__serial.flush()
@classmethod
def scan(_):
from serial.tools.list_ports import comports
return ((info.device, info.description) for info in comports())
| from serial import serial_for_url
from .connection import BufferTooShort, Connection, TimeoutError
class SerialConnection(Connection):
__serial = None
def __init__(self, url, timeout=None):
self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout)
def close(self):
if self.__serial:
self.__serial.close()
def recv(self, maxlength=None):
buf = bytearray()
while True:
c = self.__serial.read()
if not c:
raise TimeoutError("Timeout waiting for serial data")
elif c == b"$" or c == b"#":
break
elif maxlength is not None and maxlength <= len(buf):
raise BufferTooShort("Buffer too short for data received")
else:
buf.extend(c)
return bytes(buf)
def send(self, buf, offset=0, size=None):
n = len(buf)
if offset < 0:
raise ValueError("offset is negative")
elif n < offset:
raise ValueError("buffer length < offset")
elif size is None:
size = n - offset
elif size < 0:
raise ValueError("size is negative")
elif offset + size > n:
raise ValueError("buffer length < offset + size")
self.__serial.write(b'"')
self.__serial.write(buf[offset : offset + size])
self.__serial.write(b"$")
self.__serial.flush()
@classmethod
def scan(_):
from serial.tools.list_ports import comports
return ((info.device, info.description) for info in comports())
| Fix SerialConnection.close() with invalid device. | Fix SerialConnection.close() with invalid device.
| Python | mit | tkem/carreralib | from serial import serial_for_url
from .connection import BufferTooShort, Connection, TimeoutError
class SerialConnection(Connection):
+
+ __serial = None
+
def __init__(self, url, timeout=None):
self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout)
def close(self):
+ if self.__serial:
- self.__serial.close()
+ self.__serial.close()
def recv(self, maxlength=None):
buf = bytearray()
while True:
c = self.__serial.read()
if not c:
raise TimeoutError("Timeout waiting for serial data")
elif c == b"$" or c == b"#":
break
elif maxlength is not None and maxlength <= len(buf):
raise BufferTooShort("Buffer too short for data received")
else:
buf.extend(c)
return bytes(buf)
def send(self, buf, offset=0, size=None):
n = len(buf)
if offset < 0:
raise ValueError("offset is negative")
elif n < offset:
raise ValueError("buffer length < offset")
elif size is None:
size = n - offset
elif size < 0:
raise ValueError("size is negative")
elif offset + size > n:
raise ValueError("buffer length < offset + size")
self.__serial.write(b'"')
self.__serial.write(buf[offset : offset + size])
self.__serial.write(b"$")
self.__serial.flush()
@classmethod
def scan(_):
from serial.tools.list_ports import comports
return ((info.device, info.description) for info in comports())
| Fix SerialConnection.close() with invalid device. | ## Code Before:
from serial import serial_for_url
from .connection import BufferTooShort, Connection, TimeoutError
class SerialConnection(Connection):
def __init__(self, url, timeout=None):
self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout)
def close(self):
self.__serial.close()
def recv(self, maxlength=None):
buf = bytearray()
while True:
c = self.__serial.read()
if not c:
raise TimeoutError("Timeout waiting for serial data")
elif c == b"$" or c == b"#":
break
elif maxlength is not None and maxlength <= len(buf):
raise BufferTooShort("Buffer too short for data received")
else:
buf.extend(c)
return bytes(buf)
def send(self, buf, offset=0, size=None):
n = len(buf)
if offset < 0:
raise ValueError("offset is negative")
elif n < offset:
raise ValueError("buffer length < offset")
elif size is None:
size = n - offset
elif size < 0:
raise ValueError("size is negative")
elif offset + size > n:
raise ValueError("buffer length < offset + size")
self.__serial.write(b'"')
self.__serial.write(buf[offset : offset + size])
self.__serial.write(b"$")
self.__serial.flush()
@classmethod
def scan(_):
from serial.tools.list_ports import comports
return ((info.device, info.description) for info in comports())
## Instruction:
Fix SerialConnection.close() with invalid device.
## Code After:
from serial import serial_for_url
from .connection import BufferTooShort, Connection, TimeoutError
class SerialConnection(Connection):
__serial = None
def __init__(self, url, timeout=None):
self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout)
def close(self):
if self.__serial:
self.__serial.close()
def recv(self, maxlength=None):
buf = bytearray()
while True:
c = self.__serial.read()
if not c:
raise TimeoutError("Timeout waiting for serial data")
elif c == b"$" or c == b"#":
break
elif maxlength is not None and maxlength <= len(buf):
raise BufferTooShort("Buffer too short for data received")
else:
buf.extend(c)
return bytes(buf)
def send(self, buf, offset=0, size=None):
n = len(buf)
if offset < 0:
raise ValueError("offset is negative")
elif n < offset:
raise ValueError("buffer length < offset")
elif size is None:
size = n - offset
elif size < 0:
raise ValueError("size is negative")
elif offset + size > n:
raise ValueError("buffer length < offset + size")
self.__serial.write(b'"')
self.__serial.write(buf[offset : offset + size])
self.__serial.write(b"$")
self.__serial.flush()
@classmethod
def scan(_):
from serial.tools.list_ports import comports
return ((info.device, info.description) for info in comports())
| from serial import serial_for_url
from .connection import BufferTooShort, Connection, TimeoutError
class SerialConnection(Connection):
+
+ __serial = None
+
def __init__(self, url, timeout=None):
self.__serial = serial_for_url(url, baudrate=19200, timeout=timeout)
def close(self):
+ if self.__serial:
- self.__serial.close()
+ self.__serial.close()
? ++++
def recv(self, maxlength=None):
buf = bytearray()
while True:
c = self.__serial.read()
if not c:
raise TimeoutError("Timeout waiting for serial data")
elif c == b"$" or c == b"#":
break
elif maxlength is not None and maxlength <= len(buf):
raise BufferTooShort("Buffer too short for data received")
else:
buf.extend(c)
return bytes(buf)
def send(self, buf, offset=0, size=None):
n = len(buf)
if offset < 0:
raise ValueError("offset is negative")
elif n < offset:
raise ValueError("buffer length < offset")
elif size is None:
size = n - offset
elif size < 0:
raise ValueError("size is negative")
elif offset + size > n:
raise ValueError("buffer length < offset + size")
self.__serial.write(b'"')
self.__serial.write(buf[offset : offset + size])
self.__serial.write(b"$")
self.__serial.flush()
@classmethod
def scan(_):
from serial.tools.list_ports import comports
return ((info.device, info.description) for info in comports()) |
73d22cc63a2a37bd3c99774bf098ca12c81d54ae | funnels.py | funnels.py | import pyglet
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
window = pyglet.window.Window()#fullscreen=True)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
| import pyglet
import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
def main(fullscreen):
window = pyglet.window.Window(fullscreen=fullscreen)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Arithemetic practice game.")
parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
parser.set_defaults(fullscreen=True)
results = parser.parse_args()
main(results.fullscreen)
| Add argparse to turn on/off fullscreen behavior | Add argparse to turn on/off fullscreen behavior
| Python | mit | simeonf/claire | import pyglet
-
+ import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
+ def main(fullscreen):
- window = pyglet.window.Window()#fullscreen=True)
+ window = pyglet.window.Window(fullscreen=fullscreen)
- levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
+ levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
- pyglet.clock.schedule(levels.clock)
+ pyglet.clock.schedule(levels.clock)
- @window.event
+ @window.event
- def on_key_press(symbol, modifiers):
+ def on_key_press(symbol, modifiers):
- levels.key(symbol, modifiers)
+ levels.key(symbol, modifiers)
- @window.event
+ @window.event
- def on_draw():
+ def on_draw():
- levels.draw()
+ levels.draw()
- pyglet.app.run()
+ pyglet.app.run()
+ if __name__ == '__main__':
+ parser = argparse.ArgumentParser(description="Arithemetic practice game.")
+ parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
+ parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
+ parser.set_defaults(fullscreen=True)
+ results = parser.parse_args()
+ main(results.fullscreen)
+ | Add argparse to turn on/off fullscreen behavior | ## Code Before:
import pyglet
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
window = pyglet.window.Window()#fullscreen=True)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
## Instruction:
Add argparse to turn on/off fullscreen behavior
## Code After:
import pyglet
import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
def main(fullscreen):
window = pyglet.window.Window(fullscreen=fullscreen)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Arithemetic practice game.")
parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
parser.set_defaults(fullscreen=True)
results = parser.parse_args()
main(results.fullscreen)
| import pyglet
-
+ import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
+ def main(fullscreen):
- window = pyglet.window.Window()#fullscreen=True)
? -- ^ -
+ window = pyglet.window.Window(fullscreen=fullscreen)
? ++ ^^^^^^ ++
- levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
+ levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
? ++
- pyglet.clock.schedule(levels.clock)
+ pyglet.clock.schedule(levels.clock)
? ++
- @window.event
+ @window.event
? ++
- def on_key_press(symbol, modifiers):
+ def on_key_press(symbol, modifiers):
? ++
- levels.key(symbol, modifiers)
+ levels.key(symbol, modifiers)
? ++
- @window.event
+ @window.event
? ++
- def on_draw():
+ def on_draw():
? ++
- levels.draw()
+ levels.draw()
? ++
- pyglet.app.run()
+ pyglet.app.run()
? ++
+
+ if __name__ == '__main__':
+ parser = argparse.ArgumentParser(description="Arithemetic practice game.")
+ parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
+ parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
+ parser.set_defaults(fullscreen=True)
+ results = parser.parse_args()
+ main(results.fullscreen) |
6b84688c1b5a7f2e8c9e5007455b88cbaa845e9f | tests/test_track_output/results.py | tests/test_track_output/results.py |
import os
import sys
import glob
import shutil
from subprocess import call
# If vtk python module is not available, we can't run track.py so skip this
# test
cwd = os.getcwd()
try:
import vtk
except ImportError:
print('----------------Skipping test-------------')
shutil.copy('results_true.dat', 'results_test.dat')
exit()
# Run track processing script
call(['../../track.py', '-o', 'poly'] +
glob.glob(''.join((cwd, '/track*'))))
poly = ''.join((cwd, '/poly.pvtp'))
assert os.path.isfile(poly), 'poly.pvtp file not found.'
shutil.copy('poly.pvtp', 'results_test.dat')
|
import os
import sys
import glob
import shutil
from subprocess import call
# If vtk python module is not available, we can't run track.py so skip this
# test
cwd = os.getcwd()
try:
import vtk
except ImportError:
print('----------------Skipping test-------------')
shutil.copy('results_true.dat', 'results_test.dat')
exit()
# Run track processing script
call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] +
glob.glob(''.join((cwd, '/track*'))))
poly = ''.join((cwd, '/poly.pvtp'))
assert os.path.isfile(poly), 'poly.pvtp file not found.'
shutil.copy('poly.pvtp', 'results_test.dat')
| Fix path to script in test_track_output | Fix path to script in test_track_output
| Python | mit | mjlong/openmc,wbinventor/openmc,bhermanmit/openmc,wbinventor/openmc,johnnyliu27/openmc,paulromano/openmc,mjlong/openmc,smharper/openmc,lilulu/openmc,johnnyliu27/openmc,amandalund/openmc,samuelshaner/openmc,mit-crpg/openmc,shikhar413/openmc,johnnyliu27/openmc,mit-crpg/openmc,liangjg/openmc,kellyrowland/openmc,amandalund/openmc,walshjon/openmc,lilulu/openmc,liangjg/openmc,liangjg/openmc,johnnyliu27/openmc,samuelshaner/openmc,mit-crpg/openmc,amandalund/openmc,smharper/openmc,shikhar413/openmc,samuelshaner/openmc,mit-crpg/openmc,shikhar413/openmc,paulromano/openmc,liangjg/openmc,smharper/openmc,shikhar413/openmc,samuelshaner/openmc,paulromano/openmc,walshjon/openmc,walshjon/openmc,smharper/openmc,bhermanmit/openmc,wbinventor/openmc,wbinventor/openmc,walshjon/openmc,paulromano/openmc,lilulu/openmc,amandalund/openmc,kellyrowland/openmc |
import os
import sys
import glob
import shutil
from subprocess import call
# If vtk python module is not available, we can't run track.py so skip this
# test
cwd = os.getcwd()
try:
import vtk
except ImportError:
print('----------------Skipping test-------------')
shutil.copy('results_true.dat', 'results_test.dat')
exit()
# Run track processing script
- call(['../../track.py', '-o', 'poly'] +
+ call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] +
glob.glob(''.join((cwd, '/track*'))))
poly = ''.join((cwd, '/poly.pvtp'))
assert os.path.isfile(poly), 'poly.pvtp file not found.'
shutil.copy('poly.pvtp', 'results_test.dat')
| Fix path to script in test_track_output | ## Code Before:
import os
import sys
import glob
import shutil
from subprocess import call
# If vtk python module is not available, we can't run track.py so skip this
# test
cwd = os.getcwd()
try:
import vtk
except ImportError:
print('----------------Skipping test-------------')
shutil.copy('results_true.dat', 'results_test.dat')
exit()
# Run track processing script
call(['../../track.py', '-o', 'poly'] +
glob.glob(''.join((cwd, '/track*'))))
poly = ''.join((cwd, '/poly.pvtp'))
assert os.path.isfile(poly), 'poly.pvtp file not found.'
shutil.copy('poly.pvtp', 'results_test.dat')
## Instruction:
Fix path to script in test_track_output
## Code After:
import os
import sys
import glob
import shutil
from subprocess import call
# If vtk python module is not available, we can't run track.py so skip this
# test
cwd = os.getcwd()
try:
import vtk
except ImportError:
print('----------------Skipping test-------------')
shutil.copy('results_true.dat', 'results_test.dat')
exit()
# Run track processing script
call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] +
glob.glob(''.join((cwd, '/track*'))))
poly = ''.join((cwd, '/poly.pvtp'))
assert os.path.isfile(poly), 'poly.pvtp file not found.'
shutil.copy('poly.pvtp', 'results_test.dat')
|
import os
import sys
import glob
import shutil
from subprocess import call
# If vtk python module is not available, we can't run track.py so skip this
# test
cwd = os.getcwd()
try:
import vtk
except ImportError:
print('----------------Skipping test-------------')
shutil.copy('results_true.dat', 'results_test.dat')
exit()
# Run track processing script
- call(['../../track.py', '-o', 'poly'] +
+ call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] +
glob.glob(''.join((cwd, '/track*'))))
poly = ''.join((cwd, '/poly.pvtp'))
assert os.path.isfile(poly), 'poly.pvtp file not found.'
shutil.copy('poly.pvtp', 'results_test.dat') |
6131430ff7d1e9e8cd95f8d2793e82cc72679d81 | auditlog/admin.py | auditlog/admin.py | from django.contrib import admin
from .filters import ResourceTypeFilter
from .mixins import LogEntryAdminMixin
from .models import LogEntry
class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin):
list_display = ["created", "resource_url", "action", "msg_short", "user_url"]
search_fields = [
"timestamp",
"object_repr",
"changes",
"actor__first_name",
"actor__last_name",
]
list_filter = ["action", ResourceTypeFilter]
readonly_fields = ["created", "resource_url", "action", "user_url", "msg"]
fieldsets = [
(None, {"fields": ["created", "user_url", "resource_url"]}),
("Changes", {"fields": ["action", "msg"]}),
]
admin.site.register(LogEntry, LogEntryAdmin)
| from django.contrib import admin
from auditlog.filters import ResourceTypeFilter
from auditlog.mixins import LogEntryAdminMixin
from auditlog.models import LogEntry
class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin):
list_display = ["created", "resource_url", "action", "msg_short", "user_url"]
search_fields = [
"timestamp",
"object_repr",
"changes",
"actor__first_name",
"actor__last_name",
]
list_filter = ["action", ResourceTypeFilter]
readonly_fields = ["created", "resource_url", "action", "user_url", "msg"]
fieldsets = [
(None, {"fields": ["created", "user_url", "resource_url"]}),
("Changes", {"fields": ["action", "msg"]}),
]
admin.site.register(LogEntry, LogEntryAdmin)
| Change relative imports to absolute. | Change relative imports to absolute.
| Python | mit | jjkester/django-auditlog | from django.contrib import admin
- from .filters import ResourceTypeFilter
+ from auditlog.filters import ResourceTypeFilter
- from .mixins import LogEntryAdminMixin
+ from auditlog.mixins import LogEntryAdminMixin
- from .models import LogEntry
+ from auditlog.models import LogEntry
class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin):
list_display = ["created", "resource_url", "action", "msg_short", "user_url"]
search_fields = [
"timestamp",
"object_repr",
"changes",
"actor__first_name",
"actor__last_name",
]
list_filter = ["action", ResourceTypeFilter]
readonly_fields = ["created", "resource_url", "action", "user_url", "msg"]
fieldsets = [
(None, {"fields": ["created", "user_url", "resource_url"]}),
("Changes", {"fields": ["action", "msg"]}),
]
admin.site.register(LogEntry, LogEntryAdmin)
| Change relative imports to absolute. | ## Code Before:
from django.contrib import admin
from .filters import ResourceTypeFilter
from .mixins import LogEntryAdminMixin
from .models import LogEntry
class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin):
list_display = ["created", "resource_url", "action", "msg_short", "user_url"]
search_fields = [
"timestamp",
"object_repr",
"changes",
"actor__first_name",
"actor__last_name",
]
list_filter = ["action", ResourceTypeFilter]
readonly_fields = ["created", "resource_url", "action", "user_url", "msg"]
fieldsets = [
(None, {"fields": ["created", "user_url", "resource_url"]}),
("Changes", {"fields": ["action", "msg"]}),
]
admin.site.register(LogEntry, LogEntryAdmin)
## Instruction:
Change relative imports to absolute.
## Code After:
from django.contrib import admin
from auditlog.filters import ResourceTypeFilter
from auditlog.mixins import LogEntryAdminMixin
from auditlog.models import LogEntry
class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin):
list_display = ["created", "resource_url", "action", "msg_short", "user_url"]
search_fields = [
"timestamp",
"object_repr",
"changes",
"actor__first_name",
"actor__last_name",
]
list_filter = ["action", ResourceTypeFilter]
readonly_fields = ["created", "resource_url", "action", "user_url", "msg"]
fieldsets = [
(None, {"fields": ["created", "user_url", "resource_url"]}),
("Changes", {"fields": ["action", "msg"]}),
]
admin.site.register(LogEntry, LogEntryAdmin)
| from django.contrib import admin
- from .filters import ResourceTypeFilter
+ from auditlog.filters import ResourceTypeFilter
? ++++++++
- from .mixins import LogEntryAdminMixin
+ from auditlog.mixins import LogEntryAdminMixin
? ++++++++
- from .models import LogEntry
+ from auditlog.models import LogEntry
? ++++++++
class LogEntryAdmin(admin.ModelAdmin, LogEntryAdminMixin):
list_display = ["created", "resource_url", "action", "msg_short", "user_url"]
search_fields = [
"timestamp",
"object_repr",
"changes",
"actor__first_name",
"actor__last_name",
]
list_filter = ["action", ResourceTypeFilter]
readonly_fields = ["created", "resource_url", "action", "user_url", "msg"]
fieldsets = [
(None, {"fields": ["created", "user_url", "resource_url"]}),
("Changes", {"fields": ["action", "msg"]}),
]
admin.site.register(LogEntry, LogEntryAdmin) |
5d8a09ebff9cc8a8e8bdf4bff3963cee7a1aae6a | tools/skp/page_sets/skia_ebay_desktop.py | tools/skp/page_sets/skia_ebay_desktop.py |
from telemetry import story
from telemetry.page import page as page_module
from telemetry.page import shared_page_state
class SkiaDesktopPage(page_module.Page):
def __init__(self, url, page_set):
super(SkiaDesktopPage, self).__init__(
url=url,
name=url,
page_set=page_set,
shared_page_state_class=shared_page_state.SharedDesktopPageState)
self.archive_data_file = 'data/skia_ebay_desktop.json'
def RunNavigateSteps(self, action_runner):
action_runner.Navigate(self.url)
action_runner.Wait(15)
class SkiaEbayDesktopPageSet(story.StorySet):
""" Pages designed to represent the median, not highly optimized web """
def __init__(self):
super(SkiaEbayDesktopPageSet, self).__init__(
archive_data_file='data/skia_ebay_desktop.json')
urls_list = [
# go/skia-skps-3-2019
'http://www.ebay.com',
]
for url in urls_list:
self.AddStory(SkiaDesktopPage(url, self)) |
from telemetry import story
from telemetry.page import page as page_module
from telemetry.page import shared_page_state
class SkiaDesktopPage(page_module.Page):
def __init__(self, url, page_set):
super(SkiaDesktopPage, self).__init__(
url=url,
name=url,
page_set=page_set,
shared_page_state_class=shared_page_state.SharedDesktopPageState)
self.archive_data_file = 'data/skia_ebay_desktop.json'
def RunNavigateSteps(self, action_runner):
action_runner.Navigate(self.url, timeout_in_seconds=120)
class SkiaEbayDesktopPageSet(story.StorySet):
""" Pages designed to represent the median, not highly optimized web """
def __init__(self):
super(SkiaEbayDesktopPageSet, self).__init__(
archive_data_file='data/skia_ebay_desktop.json')
urls_list = [
# go/skia-skps-3-2019
'http://www.ebay.com',
]
for url in urls_list:
self.AddStory(SkiaDesktopPage(url, self))
| Add wait time to ebay pageset | Add wait time to ebay pageset
Bug: skia:11898
Change-Id: I0bb58f1d8e9c6ad48148d50b840f152fc158f071
Reviewed-on: https://skia-review.googlesource.com/c/skia/+/400538
Reviewed-by: Ravi Mistry <9fa2e7438b8cb730f96b74865492597170561628@google.com>
Commit-Queue: Ravi Mistry <9fa2e7438b8cb730f96b74865492597170561628@google.com>
| Python | bsd-3-clause | aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia,google/skia,aosp-mirror/platform_external_skia,google/skia,google/skia,aosp-mirror/platform_external_skia,google/skia,aosp-mirror/platform_external_skia,google/skia,google/skia,aosp-mirror/platform_external_skia,google/skia,google/skia,google/skia,google/skia,aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia,aosp-mirror/platform_external_skia |
from telemetry import story
from telemetry.page import page as page_module
from telemetry.page import shared_page_state
class SkiaDesktopPage(page_module.Page):
def __init__(self, url, page_set):
super(SkiaDesktopPage, self).__init__(
url=url,
name=url,
page_set=page_set,
shared_page_state_class=shared_page_state.SharedDesktopPageState)
self.archive_data_file = 'data/skia_ebay_desktop.json'
def RunNavigateSteps(self, action_runner):
- action_runner.Navigate(self.url)
+ action_runner.Navigate(self.url, timeout_in_seconds=120)
- action_runner.Wait(15)
class SkiaEbayDesktopPageSet(story.StorySet):
""" Pages designed to represent the median, not highly optimized web """
def __init__(self):
super(SkiaEbayDesktopPageSet, self).__init__(
archive_data_file='data/skia_ebay_desktop.json')
urls_list = [
# go/skia-skps-3-2019
'http://www.ebay.com',
]
for url in urls_list:
self.AddStory(SkiaDesktopPage(url, self))
+ | Add wait time to ebay pageset | ## Code Before:
from telemetry import story
from telemetry.page import page as page_module
from telemetry.page import shared_page_state
class SkiaDesktopPage(page_module.Page):
def __init__(self, url, page_set):
super(SkiaDesktopPage, self).__init__(
url=url,
name=url,
page_set=page_set,
shared_page_state_class=shared_page_state.SharedDesktopPageState)
self.archive_data_file = 'data/skia_ebay_desktop.json'
def RunNavigateSteps(self, action_runner):
action_runner.Navigate(self.url)
action_runner.Wait(15)
class SkiaEbayDesktopPageSet(story.StorySet):
""" Pages designed to represent the median, not highly optimized web """
def __init__(self):
super(SkiaEbayDesktopPageSet, self).__init__(
archive_data_file='data/skia_ebay_desktop.json')
urls_list = [
# go/skia-skps-3-2019
'http://www.ebay.com',
]
for url in urls_list:
self.AddStory(SkiaDesktopPage(url, self))
## Instruction:
Add wait time to ebay pageset
## Code After:
from telemetry import story
from telemetry.page import page as page_module
from telemetry.page import shared_page_state
class SkiaDesktopPage(page_module.Page):
def __init__(self, url, page_set):
super(SkiaDesktopPage, self).__init__(
url=url,
name=url,
page_set=page_set,
shared_page_state_class=shared_page_state.SharedDesktopPageState)
self.archive_data_file = 'data/skia_ebay_desktop.json'
def RunNavigateSteps(self, action_runner):
action_runner.Navigate(self.url, timeout_in_seconds=120)
class SkiaEbayDesktopPageSet(story.StorySet):
""" Pages designed to represent the median, not highly optimized web """
def __init__(self):
super(SkiaEbayDesktopPageSet, self).__init__(
archive_data_file='data/skia_ebay_desktop.json')
urls_list = [
# go/skia-skps-3-2019
'http://www.ebay.com',
]
for url in urls_list:
self.AddStory(SkiaDesktopPage(url, self))
|
from telemetry import story
from telemetry.page import page as page_module
from telemetry.page import shared_page_state
class SkiaDesktopPage(page_module.Page):
def __init__(self, url, page_set):
super(SkiaDesktopPage, self).__init__(
url=url,
name=url,
page_set=page_set,
shared_page_state_class=shared_page_state.SharedDesktopPageState)
self.archive_data_file = 'data/skia_ebay_desktop.json'
def RunNavigateSteps(self, action_runner):
- action_runner.Navigate(self.url)
+ action_runner.Navigate(self.url, timeout_in_seconds=120)
? ++++++++++++++++++++++++
- action_runner.Wait(15)
class SkiaEbayDesktopPageSet(story.StorySet):
""" Pages designed to represent the median, not highly optimized web """
def __init__(self):
super(SkiaEbayDesktopPageSet, self).__init__(
archive_data_file='data/skia_ebay_desktop.json')
urls_list = [
# go/skia-skps-3-2019
'http://www.ebay.com',
]
for url in urls_list:
self.AddStory(SkiaDesktopPage(url, self)) |
618bf6d4c1fc5e60b7e94d1ad1030bf2cf0de5c2 | src/main/python/alppaca/server_mock/__init__.py | src/main/python/alppaca/server_mock/__init__.py | from __future__ import print_function, absolute_import, unicode_literals, division
from datetime import datetime, timedelta
from textwrap import dedent
from bottle import Bottle
import pytz
""" Super simple IMS mock.
Just listens on localhost:8080 for the appropriate url, returns a test role and
a dummy JSON response.
"""
def expiration_10s_from_now():
n = datetime.now(tz=pytz.utc) + timedelta(seconds=10)
return n.strftime("%Y-%m-%dT%H:%M:%SZ")
class MockIms(Bottle):
PATH = '/latest/meta-data/iam/security-credentials/'
json_response = dedent("""
{"Code": "Success",
"AccessKeyId": "ASIAI",
"SecretAccessKey": "XXYYZZ",
"Token": "0123456789abcdefghijklmnopqrstuvwxyzAB",
"Expiration": "%s",
"Type": "AWS-HMAC"}
""")
def __init__(self):
super(MockIms, self).__init__()
self.route(self.PATH, callback=self.get_roles)
self.route(self.PATH + '<role>', callback=self.get_credentials)
def get_roles(self):
return 'test_role'
def get_credentials(self, role):
return self.json_response % expiration_10s_from_now() if role == 'test_role' else ''
if __name__ == "__main__":
MockIms().run()
| from __future__ import print_function, absolute_import, unicode_literals, division
from datetime import datetime, timedelta
from textwrap import dedent
from bottle import Bottle
import pytz
def expiration_10s_from_now():
n = datetime.now(tz=pytz.utc) + timedelta(seconds=10)
return n.strftime("%Y-%m-%dT%H:%M:%SZ")
class MockIms(Bottle):
PATH = '/latest/meta-data/iam/security-credentials/'
json_response = dedent("""
{"Code": "Success",
"AccessKeyId": "ASIAI",
"SecretAccessKey": "XXYYZZ",
"Token": "0123456789abcdefghijklmnopqrstuvwxyzAB",
"Expiration": "%s",
"Type": "AWS-HMAC"}
""")
def __init__(self):
super(MockIms, self).__init__()
self.route(self.PATH, callback=self.get_roles)
self.route(self.PATH + '<role>', callback=self.get_credentials)
def get_roles(self):
return 'test_role'
def get_credentials(self, role):
return self.json_response % expiration_10s_from_now() if role == 'test_role' else ''
if __name__ == "__main__":
MockIms().run()
| Move string above the imports so it becomes a docstring | Move string above the imports so it becomes a docstring
| Python | apache-2.0 | ImmobilienScout24/afp-alppaca,ImmobilienScout24/alppaca,ImmobilienScout24/alppaca,ImmobilienScout24/afp-alppaca | from __future__ import print_function, absolute_import, unicode_literals, division
from datetime import datetime, timedelta
from textwrap import dedent
from bottle import Bottle
import pytz
- """ Super simple IMS mock.
-
- Just listens on localhost:8080 for the appropriate url, returns a test role and
- a dummy JSON response.
-
- """
def expiration_10s_from_now():
n = datetime.now(tz=pytz.utc) + timedelta(seconds=10)
return n.strftime("%Y-%m-%dT%H:%M:%SZ")
class MockIms(Bottle):
PATH = '/latest/meta-data/iam/security-credentials/'
json_response = dedent("""
{"Code": "Success",
"AccessKeyId": "ASIAI",
"SecretAccessKey": "XXYYZZ",
"Token": "0123456789abcdefghijklmnopqrstuvwxyzAB",
"Expiration": "%s",
"Type": "AWS-HMAC"}
""")
def __init__(self):
super(MockIms, self).__init__()
self.route(self.PATH, callback=self.get_roles)
self.route(self.PATH + '<role>', callback=self.get_credentials)
def get_roles(self):
return 'test_role'
def get_credentials(self, role):
return self.json_response % expiration_10s_from_now() if role == 'test_role' else ''
if __name__ == "__main__":
MockIms().run()
| Move string above the imports so it becomes a docstring | ## Code Before:
from __future__ import print_function, absolute_import, unicode_literals, division
from datetime import datetime, timedelta
from textwrap import dedent
from bottle import Bottle
import pytz
""" Super simple IMS mock.
Just listens on localhost:8080 for the appropriate url, returns a test role and
a dummy JSON response.
"""
def expiration_10s_from_now():
n = datetime.now(tz=pytz.utc) + timedelta(seconds=10)
return n.strftime("%Y-%m-%dT%H:%M:%SZ")
class MockIms(Bottle):
PATH = '/latest/meta-data/iam/security-credentials/'
json_response = dedent("""
{"Code": "Success",
"AccessKeyId": "ASIAI",
"SecretAccessKey": "XXYYZZ",
"Token": "0123456789abcdefghijklmnopqrstuvwxyzAB",
"Expiration": "%s",
"Type": "AWS-HMAC"}
""")
def __init__(self):
super(MockIms, self).__init__()
self.route(self.PATH, callback=self.get_roles)
self.route(self.PATH + '<role>', callback=self.get_credentials)
def get_roles(self):
return 'test_role'
def get_credentials(self, role):
return self.json_response % expiration_10s_from_now() if role == 'test_role' else ''
if __name__ == "__main__":
MockIms().run()
## Instruction:
Move string above the imports so it becomes a docstring
## Code After:
from __future__ import print_function, absolute_import, unicode_literals, division
from datetime import datetime, timedelta
from textwrap import dedent
from bottle import Bottle
import pytz
def expiration_10s_from_now():
n = datetime.now(tz=pytz.utc) + timedelta(seconds=10)
return n.strftime("%Y-%m-%dT%H:%M:%SZ")
class MockIms(Bottle):
PATH = '/latest/meta-data/iam/security-credentials/'
json_response = dedent("""
{"Code": "Success",
"AccessKeyId": "ASIAI",
"SecretAccessKey": "XXYYZZ",
"Token": "0123456789abcdefghijklmnopqrstuvwxyzAB",
"Expiration": "%s",
"Type": "AWS-HMAC"}
""")
def __init__(self):
super(MockIms, self).__init__()
self.route(self.PATH, callback=self.get_roles)
self.route(self.PATH + '<role>', callback=self.get_credentials)
def get_roles(self):
return 'test_role'
def get_credentials(self, role):
return self.json_response % expiration_10s_from_now() if role == 'test_role' else ''
if __name__ == "__main__":
MockIms().run()
| from __future__ import print_function, absolute_import, unicode_literals, division
from datetime import datetime, timedelta
from textwrap import dedent
from bottle import Bottle
import pytz
- """ Super simple IMS mock.
-
- Just listens on localhost:8080 for the appropriate url, returns a test role and
- a dummy JSON response.
-
- """
def expiration_10s_from_now():
n = datetime.now(tz=pytz.utc) + timedelta(seconds=10)
return n.strftime("%Y-%m-%dT%H:%M:%SZ")
class MockIms(Bottle):
PATH = '/latest/meta-data/iam/security-credentials/'
json_response = dedent("""
{"Code": "Success",
"AccessKeyId": "ASIAI",
"SecretAccessKey": "XXYYZZ",
"Token": "0123456789abcdefghijklmnopqrstuvwxyzAB",
"Expiration": "%s",
"Type": "AWS-HMAC"}
""")
def __init__(self):
super(MockIms, self).__init__()
self.route(self.PATH, callback=self.get_roles)
self.route(self.PATH + '<role>', callback=self.get_credentials)
def get_roles(self):
return 'test_role'
def get_credentials(self, role):
return self.json_response % expiration_10s_from_now() if role == 'test_role' else ''
if __name__ == "__main__":
MockIms().run() |
f4837fd60ce09b69d334fcad1403b721723d3504 | tests/test_conf.py | tests/test_conf.py | import sys
from unittest import mock
import pytest
from bottery.conf import Settings
@pytest.fixture
def mocked_settings():
settings = mock.MagicMock()
sys.modules['settings'] = settings
yield settings
del sys.modules['settings']
@pytest.mark.skip
def test_global_settings():
settings = Settings()
assert settings.PLATFORMS == {}
assert settings.TEMPLATES == []
@pytest.mark.skip
def test_settings_from_module(mocked_settings):
mocked_settings.PLATFORM = 'matrix'
settings = Settings.from_object('settings')
assert settings.PLATFORM == 'matrix'
assert settings.PLATFORM == 'matrix'
| from unittest import mock
import pytest
from bottery.conf import Settings
@pytest.fixture
def mocked_settings():
settings = mock.MagicMock()
sys.modules['settings'] = settings
yield settings
del sys.modules['settings']
@pytest.mark.skip
def test_global_settings():
settings = Settings()
assert settings.PLATFORMS == {}
assert settings.TEMPLATES == []
@pytest.mark.skip
def test_settings_from_module(mocked_settings):
mocked_settings.PLATFORM = 'matrix'
settings = Settings.from_object('settings')
assert settings.PLATFORM == 'matrix'
assert settings.PLATFORM == 'matrix'
| Remove unused sys import from conf tests | Remove unused sys import from conf tests
| Python | mit | rougeth/bottery | - import sys
from unittest import mock
import pytest
from bottery.conf import Settings
@pytest.fixture
def mocked_settings():
settings = mock.MagicMock()
sys.modules['settings'] = settings
yield settings
del sys.modules['settings']
@pytest.mark.skip
def test_global_settings():
settings = Settings()
assert settings.PLATFORMS == {}
assert settings.TEMPLATES == []
@pytest.mark.skip
def test_settings_from_module(mocked_settings):
mocked_settings.PLATFORM = 'matrix'
settings = Settings.from_object('settings')
assert settings.PLATFORM == 'matrix'
assert settings.PLATFORM == 'matrix'
| Remove unused sys import from conf tests | ## Code Before:
import sys
from unittest import mock
import pytest
from bottery.conf import Settings
@pytest.fixture
def mocked_settings():
settings = mock.MagicMock()
sys.modules['settings'] = settings
yield settings
del sys.modules['settings']
@pytest.mark.skip
def test_global_settings():
settings = Settings()
assert settings.PLATFORMS == {}
assert settings.TEMPLATES == []
@pytest.mark.skip
def test_settings_from_module(mocked_settings):
mocked_settings.PLATFORM = 'matrix'
settings = Settings.from_object('settings')
assert settings.PLATFORM == 'matrix'
assert settings.PLATFORM == 'matrix'
## Instruction:
Remove unused sys import from conf tests
## Code After:
from unittest import mock
import pytest
from bottery.conf import Settings
@pytest.fixture
def mocked_settings():
settings = mock.MagicMock()
sys.modules['settings'] = settings
yield settings
del sys.modules['settings']
@pytest.mark.skip
def test_global_settings():
settings = Settings()
assert settings.PLATFORMS == {}
assert settings.TEMPLATES == []
@pytest.mark.skip
def test_settings_from_module(mocked_settings):
mocked_settings.PLATFORM = 'matrix'
settings = Settings.from_object('settings')
assert settings.PLATFORM == 'matrix'
assert settings.PLATFORM == 'matrix'
| - import sys
from unittest import mock
import pytest
from bottery.conf import Settings
@pytest.fixture
def mocked_settings():
settings = mock.MagicMock()
sys.modules['settings'] = settings
yield settings
del sys.modules['settings']
@pytest.mark.skip
def test_global_settings():
settings = Settings()
assert settings.PLATFORMS == {}
assert settings.TEMPLATES == []
@pytest.mark.skip
def test_settings_from_module(mocked_settings):
mocked_settings.PLATFORM = 'matrix'
settings = Settings.from_object('settings')
assert settings.PLATFORM == 'matrix'
assert settings.PLATFORM == 'matrix' |
fae33cf7d42559384deb7a9949f47b0881b0a29b | Cython/Tests/TestCythonUtils.py | Cython/Tests/TestCythonUtils.py | import unittest
from ..Utils import build_hex_version
class TestCythonUtils(unittest.TestCase):
def test_build_hex_version(self):
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4'))
self.assertEqual('0x001D00F0', build_hex_version('0.29'))
self.assertEqual('0x040000F0', build_hex_version('4.0'))
| import unittest
from ..Utils import build_hex_version
class TestCythonUtils(unittest.TestCase):
def test_build_hex_version(self):
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4'))
self.assertEqual('0x001D00F0', build_hex_version('0.29'))
self.assertEqual('0x040000F0', build_hex_version('4.0'))
| Remove accidentally duplicated test code. | Remove accidentally duplicated test code.
| Python | apache-2.0 | da-woods/cython,scoder/cython,da-woods/cython,cython/cython,scoder/cython,scoder/cython,da-woods/cython,cython/cython,scoder/cython,cython/cython,da-woods/cython,cython/cython | import unittest
from ..Utils import build_hex_version
class TestCythonUtils(unittest.TestCase):
def test_build_hex_version(self):
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
- self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4'))
self.assertEqual('0x001D00F0', build_hex_version('0.29'))
self.assertEqual('0x040000F0', build_hex_version('4.0'))
| Remove accidentally duplicated test code. | ## Code Before:
import unittest
from ..Utils import build_hex_version
class TestCythonUtils(unittest.TestCase):
def test_build_hex_version(self):
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4'))
self.assertEqual('0x001D00F0', build_hex_version('0.29'))
self.assertEqual('0x040000F0', build_hex_version('4.0'))
## Instruction:
Remove accidentally duplicated test code.
## Code After:
import unittest
from ..Utils import build_hex_version
class TestCythonUtils(unittest.TestCase):
def test_build_hex_version(self):
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4'))
self.assertEqual('0x001D00F0', build_hex_version('0.29'))
self.assertEqual('0x040000F0', build_hex_version('4.0'))
| import unittest
from ..Utils import build_hex_version
class TestCythonUtils(unittest.TestCase):
def test_build_hex_version(self):
self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
- self.assertEqual('0x001D00A1', build_hex_version('0.29a1'))
self.assertEqual('0x001D03C4', build_hex_version('0.29.3rc4'))
self.assertEqual('0x001D00F0', build_hex_version('0.29'))
self.assertEqual('0x040000F0', build_hex_version('4.0')) |
89a8d6021d8ca8a714af018f3168298109013c6f | radio/__init__.py | radio/__init__.py | from django.utils.version import get_version
from subprocess import check_output, CalledProcessError
VERSION = (0, 0, 3, 'beta', 1)
__version__ = get_version(VERSION)
try:
__git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode()
except (FileNotFoundError, CalledProcessError):
__git_hash__ = '0'
__fullversion__ = '{} #{}'.format(__version__,__git_hash__)
print('Trunk-Player Version ' + __fullversion__)
| import logging
from django.utils.version import get_version
from subprocess import check_output, CalledProcessError
logger = logging.getLogger(__name__)
VERSION = (0, 0, 3, 'beta', 1)
__version__ = get_version(VERSION)
try:
__git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode()
except (FileNotFoundError, CalledProcessError):
__git_hash__ = '0'
__fullversion__ = '{} #{}'.format(__version__,__git_hash__)
logger.error('Trunk-Player Version ' + __fullversion__)
| Move version print to logger | Move version print to logger
| Python | mit | ScanOC/trunk-player,ScanOC/trunk-player,ScanOC/trunk-player,ScanOC/trunk-player | + import logging
+
from django.utils.version import get_version
from subprocess import check_output, CalledProcessError
+
+ logger = logging.getLogger(__name__)
+
VERSION = (0, 0, 3, 'beta', 1)
__version__ = get_version(VERSION)
try:
__git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode()
except (FileNotFoundError, CalledProcessError):
__git_hash__ = '0'
__fullversion__ = '{} #{}'.format(__version__,__git_hash__)
- print('Trunk-Player Version ' + __fullversion__)
+ logger.error('Trunk-Player Version ' + __fullversion__)
| Move version print to logger | ## Code Before:
from django.utils.version import get_version
from subprocess import check_output, CalledProcessError
VERSION = (0, 0, 3, 'beta', 1)
__version__ = get_version(VERSION)
try:
__git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode()
except (FileNotFoundError, CalledProcessError):
__git_hash__ = '0'
__fullversion__ = '{} #{}'.format(__version__,__git_hash__)
print('Trunk-Player Version ' + __fullversion__)
## Instruction:
Move version print to logger
## Code After:
import logging
from django.utils.version import get_version
from subprocess import check_output, CalledProcessError
logger = logging.getLogger(__name__)
VERSION = (0, 0, 3, 'beta', 1)
__version__ = get_version(VERSION)
try:
__git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode()
except (FileNotFoundError, CalledProcessError):
__git_hash__ = '0'
__fullversion__ = '{} #{}'.format(__version__,__git_hash__)
logger.error('Trunk-Player Version ' + __fullversion__)
| + import logging
+
from django.utils.version import get_version
from subprocess import check_output, CalledProcessError
+
+ logger = logging.getLogger(__name__)
+
VERSION = (0, 0, 3, 'beta', 1)
__version__ = get_version(VERSION)
try:
__git_hash__ = check_output(['git', 'rev-parse', '--short', 'HEAD']).strip().decode()
except (FileNotFoundError, CalledProcessError):
__git_hash__ = '0'
__fullversion__ = '{} #{}'.format(__version__,__git_hash__)
- print('Trunk-Player Version ' + __fullversion__)
? ^ ^^^
+ logger.error('Trunk-Player Version ' + __fullversion__)
? ^^^^^ ^^^^^^
|
41c49a44c5f1bc9747b22b6d1f1088f1354a2cd5 | nes/cpu/decoder.py | nes/cpu/decoder.py | from sqlite3 import Connection, Row
class Decoder:
def __init__(self):
self.conn = Connection('nes.sqlite')
self.conn.row_factory = Row
def __del__(self):
self.conn.close()
def decode(self, opcode):
c = self.conn.cursor()
c.execute('select * from instruction where opcode=?', [opcode])
row = c.fetchone()
return dict(zip(row.keys(), row))
| from sqlite3 import Connection, Row
class Decoder:
def __init__(self):
self.conn = Connection('nes.sqlite')
self.conn.row_factory = Row
def __del__(self):
self.conn.close()
def decode(self, opcode):
c = self.conn.cursor()
c.execute('select * from instruction where opcode=?', [opcode])
row = c.fetchone()
if row:
return dict(zip(row.keys(), row))
else:
raise NotImplementedError('Undocumented Opcode: ' + str(opcode))
| Raise an exception when it's an undocumented opcode. | Raise an exception when it's an undocumented opcode.
| Python | mit | Hexadorsimal/pynes | from sqlite3 import Connection, Row
class Decoder:
def __init__(self):
self.conn = Connection('nes.sqlite')
self.conn.row_factory = Row
def __del__(self):
self.conn.close()
def decode(self, opcode):
c = self.conn.cursor()
c.execute('select * from instruction where opcode=?', [opcode])
row = c.fetchone()
+ if row:
- return dict(zip(row.keys(), row))
+ return dict(zip(row.keys(), row))
+ else:
+ raise NotImplementedError('Undocumented Opcode: ' + str(opcode))
| Raise an exception when it's an undocumented opcode. | ## Code Before:
from sqlite3 import Connection, Row
class Decoder:
def __init__(self):
self.conn = Connection('nes.sqlite')
self.conn.row_factory = Row
def __del__(self):
self.conn.close()
def decode(self, opcode):
c = self.conn.cursor()
c.execute('select * from instruction where opcode=?', [opcode])
row = c.fetchone()
return dict(zip(row.keys(), row))
## Instruction:
Raise an exception when it's an undocumented opcode.
## Code After:
from sqlite3 import Connection, Row
class Decoder:
def __init__(self):
self.conn = Connection('nes.sqlite')
self.conn.row_factory = Row
def __del__(self):
self.conn.close()
def decode(self, opcode):
c = self.conn.cursor()
c.execute('select * from instruction where opcode=?', [opcode])
row = c.fetchone()
if row:
return dict(zip(row.keys(), row))
else:
raise NotImplementedError('Undocumented Opcode: ' + str(opcode))
| from sqlite3 import Connection, Row
class Decoder:
def __init__(self):
self.conn = Connection('nes.sqlite')
self.conn.row_factory = Row
def __del__(self):
self.conn.close()
def decode(self, opcode):
c = self.conn.cursor()
c.execute('select * from instruction where opcode=?', [opcode])
row = c.fetchone()
+ if row:
- return dict(zip(row.keys(), row))
+ return dict(zip(row.keys(), row))
? ++++
+ else:
+ raise NotImplementedError('Undocumented Opcode: ' + str(opcode)) |
c43820a2e26dd4f87c36b986a9a0af80b409f659 | sentence_extractor.py | sentence_extractor.py | import textract
import sys
import os
import re
import random
###################################
# Extracts text from a pdf file and
# selects one sentence, which it
# then prints.
#
# Created by Fredrik Omstedt.
###################################
# Extracts texts from pdf files. If given a directory, the
# program will return texts from all pdf files in that directory.
def extractTexts():
file = sys.argv[1]
texts = []
if os.path.isdir(file):
for f in os.listdir(file):
if re.match(r'^.*\.pdf$', f):
texts.append(textract.process(file + "/" + f))
else:
texts.append(textract.process(file))
return texts
# Chooses one sentence randomly from each of the given texts.
def selectSentences(texts):
chosen_sentences = []
for text in texts:
sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M)
sentences = sentence_structure.findall(text)
chosen_sentences.append(
sentences[random.randint(0, len(sentences)-1)].replace("\n", " ")
)
return chosen_sentences
def main():
texts = extractTexts()
sentences = selectSentences(texts)
for sentence in sentences:
print(sentence)
print("\n")
if __name__ == '__main__':
main()
| import textract
import sys
import os
import re
import random
###################################
# Extracts text from a pdf file and
# selects one sentence, which it
# then prints.
#
# Created by Fredrik Omstedt.
###################################
# Extracts texts from pdf files. If given a directory, the
# program will return texts from all pdf files in that directory.
def extractTexts():
file = sys.argv[1]
texts = []
if os.path.isdir(file):
for f in os.listdir(file):
if re.match(r'^.*\.pdf$', f):
texts.append(textract.process(file + "/" + f))
else:
texts.append(textract.process(file))
return texts
# Chooses one sentence randomly from each of the given texts.
def selectSentences(texts):
chosen_sentences = []
for text in texts:
sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M)
sentences = sentence_structure.findall(text)
chosen_sentences.append(
sentences[random.randint(0, len(sentences)-1)].replace("\n", " ")
)
return chosen_sentences
def main():
texts = extractTexts()
sentences = selectSentences(texts)
for sentence in sentences:
print(sentence)
print("\n")
if __name__ == '__main__':
main()
| Update regex to match sentences starting with ÅÄÖ | Update regex to match sentences starting with ÅÄÖ
| Python | mit | Xaril/sentence-extractor,Xaril/sentence-extractor | import textract
import sys
import os
import re
import random
###################################
# Extracts text from a pdf file and
# selects one sentence, which it
# then prints.
#
# Created by Fredrik Omstedt.
###################################
# Extracts texts from pdf files. If given a directory, the
# program will return texts from all pdf files in that directory.
def extractTexts():
file = sys.argv[1]
texts = []
if os.path.isdir(file):
for f in os.listdir(file):
if re.match(r'^.*\.pdf$', f):
texts.append(textract.process(file + "/" + f))
else:
texts.append(textract.process(file))
return texts
# Chooses one sentence randomly from each of the given texts.
def selectSentences(texts):
chosen_sentences = []
for text in texts:
- sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M)
+ sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M)
sentences = sentence_structure.findall(text)
chosen_sentences.append(
sentences[random.randint(0, len(sentences)-1)].replace("\n", " ")
)
return chosen_sentences
def main():
texts = extractTexts()
sentences = selectSentences(texts)
for sentence in sentences:
print(sentence)
print("\n")
if __name__ == '__main__':
main()
| Update regex to match sentences starting with ÅÄÖ | ## Code Before:
import textract
import sys
import os
import re
import random
###################################
# Extracts text from a pdf file and
# selects one sentence, which it
# then prints.
#
# Created by Fredrik Omstedt.
###################################
# Extracts texts from pdf files. If given a directory, the
# program will return texts from all pdf files in that directory.
def extractTexts():
file = sys.argv[1]
texts = []
if os.path.isdir(file):
for f in os.listdir(file):
if re.match(r'^.*\.pdf$', f):
texts.append(textract.process(file + "/" + f))
else:
texts.append(textract.process(file))
return texts
# Chooses one sentence randomly from each of the given texts.
def selectSentences(texts):
chosen_sentences = []
for text in texts:
sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M)
sentences = sentence_structure.findall(text)
chosen_sentences.append(
sentences[random.randint(0, len(sentences)-1)].replace("\n", " ")
)
return chosen_sentences
def main():
texts = extractTexts()
sentences = selectSentences(texts)
for sentence in sentences:
print(sentence)
print("\n")
if __name__ == '__main__':
main()
## Instruction:
Update regex to match sentences starting with ÅÄÖ
## Code After:
import textract
import sys
import os
import re
import random
###################################
# Extracts text from a pdf file and
# selects one sentence, which it
# then prints.
#
# Created by Fredrik Omstedt.
###################################
# Extracts texts from pdf files. If given a directory, the
# program will return texts from all pdf files in that directory.
def extractTexts():
file = sys.argv[1]
texts = []
if os.path.isdir(file):
for f in os.listdir(file):
if re.match(r'^.*\.pdf$', f):
texts.append(textract.process(file + "/" + f))
else:
texts.append(textract.process(file))
return texts
# Chooses one sentence randomly from each of the given texts.
def selectSentences(texts):
chosen_sentences = []
for text in texts:
sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M)
sentences = sentence_structure.findall(text)
chosen_sentences.append(
sentences[random.randint(0, len(sentences)-1)].replace("\n", " ")
)
return chosen_sentences
def main():
texts = extractTexts()
sentences = selectSentences(texts)
for sentence in sentences:
print(sentence)
print("\n")
if __name__ == '__main__':
main()
| import textract
import sys
import os
import re
import random
###################################
# Extracts text from a pdf file and
# selects one sentence, which it
# then prints.
#
# Created by Fredrik Omstedt.
###################################
# Extracts texts from pdf files. If given a directory, the
# program will return texts from all pdf files in that directory.
def extractTexts():
file = sys.argv[1]
texts = []
if os.path.isdir(file):
for f in os.listdir(file):
if re.match(r'^.*\.pdf$', f):
texts.append(textract.process(file + "/" + f))
else:
texts.append(textract.process(file))
return texts
# Chooses one sentence randomly from each of the given texts.
def selectSentences(texts):
chosen_sentences = []
for text in texts:
- sentence_structure = re.compile(r'([A-Z][^\.!?]*[\.!?])', re.M)
+ sentence_structure = re.compile(r'([A-Z\xc4\xc5\xd6][^\.!?]*[\.!?])', re.M)
? ++++++++++++
sentences = sentence_structure.findall(text)
chosen_sentences.append(
sentences[random.randint(0, len(sentences)-1)].replace("\n", " ")
)
return chosen_sentences
def main():
texts = extractTexts()
sentences = selectSentences(texts)
for sentence in sentences:
print(sentence)
print("\n")
if __name__ == '__main__':
main() |
78c5580d349d6bec0715a36c13437177a726f7ad | tests/test_isim.py | tests/test_isim.py | import pytest
def test_isim():
import os
import shutil
import tempfile
import yaml
from fusesoc.edatools import get_edatool
from edalize_common import compare_files, files, param_gen, tests_dir, vpi
(parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam'])
work_root = tempfile.mkdtemp(prefix='isim_')
eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml')
with open(eda_api_file,'w') as f:
f.write(yaml.dump({'name' : 'test_isim_0',
'files' : files,
'parameters' : parameters,
'tool_options' : {'isim' : {
'fuse_options' : ['some', 'fuse_options'],
'isim_options' : ['a', 'few', 'isim_options']}},
'toplevel' : 'top_module',
'vpi' : vpi}))
backend = get_edatool('isim')(eda_api_file=eda_api_file)
backend.configure(args)
ref_dir = os.path.join(tests_dir, __name__)
compare_files(ref_dir, work_root,
['config.mk',
'Makefile',
'run_test_isim_0.tcl',
'test_isim_0.prj'])
dummy_exe = 'test_isim_0'
shutil.copy(os.path.join(ref_dir, dummy_exe),
os.path.join(work_root, dummy_exe))
backend.run([])
compare_files(ref_dir, work_root, ['run.cmd'])
| import pytest
def test_isim():
import os
import shutil
from edalize_common import compare_files, setup_backend, tests_dir
ref_dir = os.path.join(tests_dir, __name__)
paramtypes = ['plusarg', 'vlogdefine', 'vlogparam']
name = 'test_isim_0'
tool = 'isim'
tool_options = {
'fuse_options' : ['some', 'fuse_options'],
'isim_options' : ['a', 'few', 'isim_options'],
}
(backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options)
backend.configure(args)
compare_files(ref_dir, work_root,
['config.mk',
'Makefile',
'run_test_isim_0.tcl',
'test_isim_0.prj'])
dummy_exe = 'test_isim_0'
shutil.copy(os.path.join(ref_dir, dummy_exe),
os.path.join(work_root, dummy_exe))
backend.run([])
compare_files(ref_dir, work_root, ['run.cmd'])
| Reduce code duplication in isim test | Reduce code duplication in isim test
| Python | bsd-2-clause | olofk/fusesoc,olofk/fusesoc,lowRISC/fusesoc,lowRISC/fusesoc | import pytest
def test_isim():
import os
import shutil
- import tempfile
- import yaml
- from fusesoc.edatools import get_edatool
- from edalize_common import compare_files, files, param_gen, tests_dir, vpi
+ from edalize_common import compare_files, setup_backend, tests_dir
+ ref_dir = os.path.join(tests_dir, __name__)
- (parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam'])
+ paramtypes = ['plusarg', 'vlogdefine', 'vlogparam']
+ name = 'test_isim_0'
+ tool = 'isim'
+ tool_options = {
+ 'fuse_options' : ['some', 'fuse_options'],
+ 'isim_options' : ['a', 'few', 'isim_options'],
+ }
+ (backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options)
- work_root = tempfile.mkdtemp(prefix='isim_')
- eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml')
- with open(eda_api_file,'w') as f:
- f.write(yaml.dump({'name' : 'test_isim_0',
- 'files' : files,
- 'parameters' : parameters,
- 'tool_options' : {'isim' : {
- 'fuse_options' : ['some', 'fuse_options'],
- 'isim_options' : ['a', 'few', 'isim_options']}},
- 'toplevel' : 'top_module',
- 'vpi' : vpi}))
-
- backend = get_edatool('isim')(eda_api_file=eda_api_file)
backend.configure(args)
- ref_dir = os.path.join(tests_dir, __name__)
compare_files(ref_dir, work_root,
['config.mk',
'Makefile',
'run_test_isim_0.tcl',
'test_isim_0.prj'])
dummy_exe = 'test_isim_0'
shutil.copy(os.path.join(ref_dir, dummy_exe),
os.path.join(work_root, dummy_exe))
backend.run([])
compare_files(ref_dir, work_root, ['run.cmd'])
| Reduce code duplication in isim test | ## Code Before:
import pytest
def test_isim():
import os
import shutil
import tempfile
import yaml
from fusesoc.edatools import get_edatool
from edalize_common import compare_files, files, param_gen, tests_dir, vpi
(parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam'])
work_root = tempfile.mkdtemp(prefix='isim_')
eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml')
with open(eda_api_file,'w') as f:
f.write(yaml.dump({'name' : 'test_isim_0',
'files' : files,
'parameters' : parameters,
'tool_options' : {'isim' : {
'fuse_options' : ['some', 'fuse_options'],
'isim_options' : ['a', 'few', 'isim_options']}},
'toplevel' : 'top_module',
'vpi' : vpi}))
backend = get_edatool('isim')(eda_api_file=eda_api_file)
backend.configure(args)
ref_dir = os.path.join(tests_dir, __name__)
compare_files(ref_dir, work_root,
['config.mk',
'Makefile',
'run_test_isim_0.tcl',
'test_isim_0.prj'])
dummy_exe = 'test_isim_0'
shutil.copy(os.path.join(ref_dir, dummy_exe),
os.path.join(work_root, dummy_exe))
backend.run([])
compare_files(ref_dir, work_root, ['run.cmd'])
## Instruction:
Reduce code duplication in isim test
## Code After:
import pytest
def test_isim():
import os
import shutil
from edalize_common import compare_files, setup_backend, tests_dir
ref_dir = os.path.join(tests_dir, __name__)
paramtypes = ['plusarg', 'vlogdefine', 'vlogparam']
name = 'test_isim_0'
tool = 'isim'
tool_options = {
'fuse_options' : ['some', 'fuse_options'],
'isim_options' : ['a', 'few', 'isim_options'],
}
(backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options)
backend.configure(args)
compare_files(ref_dir, work_root,
['config.mk',
'Makefile',
'run_test_isim_0.tcl',
'test_isim_0.prj'])
dummy_exe = 'test_isim_0'
shutil.copy(os.path.join(ref_dir, dummy_exe),
os.path.join(work_root, dummy_exe))
backend.run([])
compare_files(ref_dir, work_root, ['run.cmd'])
| import pytest
def test_isim():
import os
import shutil
- import tempfile
- import yaml
- from fusesoc.edatools import get_edatool
- from edalize_common import compare_files, files, param_gen, tests_dir, vpi
? ^^^ ^^^ ^^^^^ -----
+ from edalize_common import compare_files, setup_backend, tests_dir
? ^ ^^ ++ ^^ +
- (parameters, args) = param_gen(['plusarg', 'vlogdefine', 'vlogparam'])
+ ref_dir = os.path.join(tests_dir, __name__)
+ paramtypes = ['plusarg', 'vlogdefine', 'vlogparam']
+ name = 'test_isim_0'
+ tool = 'isim'
+ tool_options = {
+ 'fuse_options' : ['some', 'fuse_options'],
+ 'isim_options' : ['a', 'few', 'isim_options'],
+ }
+ (backend, args, work_root) = setup_backend(paramtypes, name, tool, tool_options)
- work_root = tempfile.mkdtemp(prefix='isim_')
- eda_api_file = os.path.join(work_root, 'test_isim_0.eda.yml')
- with open(eda_api_file,'w') as f:
- f.write(yaml.dump({'name' : 'test_isim_0',
- 'files' : files,
- 'parameters' : parameters,
- 'tool_options' : {'isim' : {
- 'fuse_options' : ['some', 'fuse_options'],
- 'isim_options' : ['a', 'few', 'isim_options']}},
- 'toplevel' : 'top_module',
- 'vpi' : vpi}))
-
- backend = get_edatool('isim')(eda_api_file=eda_api_file)
backend.configure(args)
- ref_dir = os.path.join(tests_dir, __name__)
compare_files(ref_dir, work_root,
['config.mk',
'Makefile',
'run_test_isim_0.tcl',
'test_isim_0.prj'])
dummy_exe = 'test_isim_0'
shutil.copy(os.path.join(ref_dir, dummy_exe),
os.path.join(work_root, dummy_exe))
backend.run([])
compare_files(ref_dir, work_root, ['run.cmd']) |
8621d6c0826beb4a4b4e920ce27786b01546ed28 | impactstoryanalytics/highcharts.py | impactstoryanalytics/highcharts.py | boilerplate = {
'chart': {
'renderTo': 'container',
'plotBackgroundColor': 'none',
'backgroundColor': 'none',
'spacingTop': 5
},
'title': {'text': None},
'subtitle': {'text': None},
'yAxis': {
'title':{
'text': None
},
'gridLineColor': 'rgba(255, 255, 255, .1)'
},
"xAxis": {
"lineColor": "rgba(0,0,0,0)"
},
'credits': {
'enabled': False
},
'plotOptions': {
'series': {
'marker': {
'enabled': False
}
}
},
} | boilerplate = {
'chart': {
'renderTo': 'container',
'plotBackgroundColor': 'none',
'backgroundColor': 'none',
'spacingTop': 5
},
'title': {'text': None},
'subtitle': {'text': None},
'yAxis': {
'min': 0,
'title':{
'text': None
},
'gridLineColor': 'rgba(255, 255, 255, .1)'
},
"xAxis": {
"lineColor": "rgba(0,0,0,0)"
},
'credits': {
'enabled': False
},
'plotOptions': {
'series': {
'marker': {
'enabled': False
}
}
},
} | Set y-axis min to 0 | Set y-axis min to 0
| Python | mit | Impactstory/impactstory-analytics,Impactstory/impactstory-analytics,total-impact/impactstory-analytics,total-impact/impactstory-analytics,Impactstory/impactstory-analytics,total-impact/impactstory-analytics,Impactstory/impactstory-analytics,total-impact/impactstory-analytics | boilerplate = {
'chart': {
'renderTo': 'container',
'plotBackgroundColor': 'none',
'backgroundColor': 'none',
'spacingTop': 5
},
'title': {'text': None},
'subtitle': {'text': None},
'yAxis': {
+ 'min': 0,
'title':{
'text': None
},
'gridLineColor': 'rgba(255, 255, 255, .1)'
},
"xAxis": {
"lineColor": "rgba(0,0,0,0)"
},
'credits': {
'enabled': False
},
'plotOptions': {
'series': {
'marker': {
'enabled': False
}
}
},
} | Set y-axis min to 0 | ## Code Before:
boilerplate = {
'chart': {
'renderTo': 'container',
'plotBackgroundColor': 'none',
'backgroundColor': 'none',
'spacingTop': 5
},
'title': {'text': None},
'subtitle': {'text': None},
'yAxis': {
'title':{
'text': None
},
'gridLineColor': 'rgba(255, 255, 255, .1)'
},
"xAxis": {
"lineColor": "rgba(0,0,0,0)"
},
'credits': {
'enabled': False
},
'plotOptions': {
'series': {
'marker': {
'enabled': False
}
}
},
}
## Instruction:
Set y-axis min to 0
## Code After:
boilerplate = {
'chart': {
'renderTo': 'container',
'plotBackgroundColor': 'none',
'backgroundColor': 'none',
'spacingTop': 5
},
'title': {'text': None},
'subtitle': {'text': None},
'yAxis': {
'min': 0,
'title':{
'text': None
},
'gridLineColor': 'rgba(255, 255, 255, .1)'
},
"xAxis": {
"lineColor": "rgba(0,0,0,0)"
},
'credits': {
'enabled': False
},
'plotOptions': {
'series': {
'marker': {
'enabled': False
}
}
},
} | boilerplate = {
'chart': {
'renderTo': 'container',
'plotBackgroundColor': 'none',
'backgroundColor': 'none',
'spacingTop': 5
},
'title': {'text': None},
'subtitle': {'text': None},
'yAxis': {
+ 'min': 0,
'title':{
'text': None
},
'gridLineColor': 'rgba(255, 255, 255, .1)'
},
"xAxis": {
"lineColor": "rgba(0,0,0,0)"
},
'credits': {
'enabled': False
},
'plotOptions': {
'series': {
'marker': {
'enabled': False
}
}
},
} |
8befea283830f76dfa41cfd10d7eb916c68f7ef9 | intern/views.py | intern/views.py | from django.contrib.auth.decorators import login_required
from django.shortcuts import render
from filer.models import File
from filer.models import Folder
@login_required
def documents(request):
files = File.objects.all()
folders = Folder.objects.all()
#print(files[0])
return render(request, 'intern/documents.html', {'files': files, 'folders': folders}) | from django.contrib.auth.decorators import login_required
from django.shortcuts import render
from filer.models import File
from filer.models import Folder
@login_required
def documents(request):
files = File.objects.all().order_by("-modified_at")
folders = Folder.objects.all()
#print(files[0])
return render(request, 'intern/documents.html', {'files': files, 'folders': folders}) | Sort files by last modification | Sort files by last modification
| Python | mit | n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb,n2o/dpb | from django.contrib.auth.decorators import login_required
from django.shortcuts import render
from filer.models import File
from filer.models import Folder
@login_required
def documents(request):
- files = File.objects.all()
+ files = File.objects.all().order_by("-modified_at")
folders = Folder.objects.all()
#print(files[0])
return render(request, 'intern/documents.html', {'files': files, 'folders': folders}) | Sort files by last modification | ## Code Before:
from django.contrib.auth.decorators import login_required
from django.shortcuts import render
from filer.models import File
from filer.models import Folder
@login_required
def documents(request):
files = File.objects.all()
folders = Folder.objects.all()
#print(files[0])
return render(request, 'intern/documents.html', {'files': files, 'folders': folders})
## Instruction:
Sort files by last modification
## Code After:
from django.contrib.auth.decorators import login_required
from django.shortcuts import render
from filer.models import File
from filer.models import Folder
@login_required
def documents(request):
files = File.objects.all().order_by("-modified_at")
folders = Folder.objects.all()
#print(files[0])
return render(request, 'intern/documents.html', {'files': files, 'folders': folders}) | from django.contrib.auth.decorators import login_required
from django.shortcuts import render
from filer.models import File
from filer.models import Folder
@login_required
def documents(request):
- files = File.objects.all()
+ files = File.objects.all().order_by("-modified_at")
folders = Folder.objects.all()
#print(files[0])
return render(request, 'intern/documents.html', {'files': files, 'folders': folders}) |
ee9df63aeaabb4111cea3698a4f0e30b4502e519 | test/disable_captcha.py | test/disable_captcha.py | import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text('Spambot countermeasures')
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert 'Enable spambot countermeasures' in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
'enable_confirm': '0',
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
self.change_acp_knob(
link_text='Spambot countermeasures',
check_page_text='Enable spambot countermeasures',
name='enable_confirm',
value='0',
)
def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text(link_text)
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert check_page_text in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
name: value,
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| Generalize to allow editing other configuration | Generalize to allow editing other configuration
| Python | bsd-2-clause | p/wolis-phpbb,p/wolis-phpbb | import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
+ self.change_acp_knob(
+ link_text='Spambot countermeasures',
+ check_page_text='Enable spambot countermeasures',
+ name='enable_confirm',
+ value='0',
+ )
+
+ def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
- url = self.link_href_by_text('Spambot countermeasures')
+ url = self.link_href_by_text(link_text)
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
- assert 'Enable spambot countermeasures' in self.response.body
+ assert check_page_text in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
- 'enable_confirm': '0',
+ name: value,
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| Generalize to allow editing other configuration | ## Code Before:
import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text('Spambot countermeasures')
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert 'Enable spambot countermeasures' in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
'enable_confirm': '0',
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
## Instruction:
Generalize to allow editing other configuration
## Code After:
import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
self.change_acp_knob(
link_text='Spambot countermeasures',
check_page_text='Enable spambot countermeasures',
name='enable_confirm',
value='0',
)
def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text(link_text)
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert check_page_text in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
name: value,
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
+ self.change_acp_knob(
+ link_text='Spambot countermeasures',
+ check_page_text='Enable spambot countermeasures',
+ name='enable_confirm',
+ value='0',
+ )
+
+ def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
- url = self.link_href_by_text('Spambot countermeasures')
+ url = self.link_href_by_text(link_text)
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
- assert 'Enable spambot countermeasures' in self.response.body
+ assert check_page_text in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
- 'enable_confirm': '0',
+ name: value,
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main() |
a5fdffe2f37e2e1c34044c259ef56c0e5feca0cb | allegedb/allegedb/tests/test_branch_plan.py | allegedb/allegedb/tests/test_branch_plan.py | import pytest
import allegedb
@pytest.fixture(scope='function')
def orm():
with allegedb.ORM("sqlite:///:memory:") as it:
yield it
def test_single_plan(orm):
g = orm.new_graph('graph')
g.add_node(0)
orm.turn = 1
g.add_node(1)
with orm.plan():
orm.turn = 2
g.add_node(2)
assert orm.turn == 1
assert 2 not in g
orm.branch = 'b'
assert 2 not in g
assert 1 in g
orm.turn = 2
assert 2 in g
orm.turn = 1
orm.branch = 'trunk'
orm.turn = 0
assert 1 not in g
orm.branch = 'c'
orm.turn = 2
assert 1 not in g
assert 2 not in g | import pytest
import allegedb
@pytest.fixture(scope='function')
def orm():
with allegedb.ORM("sqlite:///:memory:") as it:
yield it
def test_single_plan(orm):
g = orm.new_graph('graph')
g.add_node(0)
orm.turn = 1
g.add_node(1)
with orm.plan():
orm.turn = 2
g.add_node(2)
assert orm.turn == 1
assert 2 not in g
orm.branch = 'b'
assert 2 not in g
assert 1 in g
orm.turn = 2
assert 2 in g
orm.turn = 1
orm.branch = 'trunk'
orm.turn = 0
assert 1 not in g
orm.branch = 'c'
orm.turn = 2
assert 1 not in g
assert 2 not in g
orm.turn = 0
orm.branch = 'trunk'
orm.turn = 2
assert 2 in g | Add an extra check in that test | Add an extra check in that test
| Python | agpl-3.0 | LogicalDash/LiSE,LogicalDash/LiSE | import pytest
import allegedb
@pytest.fixture(scope='function')
def orm():
with allegedb.ORM("sqlite:///:memory:") as it:
yield it
def test_single_plan(orm):
g = orm.new_graph('graph')
g.add_node(0)
orm.turn = 1
g.add_node(1)
with orm.plan():
orm.turn = 2
g.add_node(2)
assert orm.turn == 1
assert 2 not in g
orm.branch = 'b'
assert 2 not in g
assert 1 in g
orm.turn = 2
assert 2 in g
orm.turn = 1
orm.branch = 'trunk'
orm.turn = 0
assert 1 not in g
orm.branch = 'c'
orm.turn = 2
assert 1 not in g
assert 2 not in g
+ orm.turn = 0
+ orm.branch = 'trunk'
+ orm.turn = 2
+ assert 2 in g | Add an extra check in that test | ## Code Before:
import pytest
import allegedb
@pytest.fixture(scope='function')
def orm():
with allegedb.ORM("sqlite:///:memory:") as it:
yield it
def test_single_plan(orm):
g = orm.new_graph('graph')
g.add_node(0)
orm.turn = 1
g.add_node(1)
with orm.plan():
orm.turn = 2
g.add_node(2)
assert orm.turn == 1
assert 2 not in g
orm.branch = 'b'
assert 2 not in g
assert 1 in g
orm.turn = 2
assert 2 in g
orm.turn = 1
orm.branch = 'trunk'
orm.turn = 0
assert 1 not in g
orm.branch = 'c'
orm.turn = 2
assert 1 not in g
assert 2 not in g
## Instruction:
Add an extra check in that test
## Code After:
import pytest
import allegedb
@pytest.fixture(scope='function')
def orm():
with allegedb.ORM("sqlite:///:memory:") as it:
yield it
def test_single_plan(orm):
g = orm.new_graph('graph')
g.add_node(0)
orm.turn = 1
g.add_node(1)
with orm.plan():
orm.turn = 2
g.add_node(2)
assert orm.turn == 1
assert 2 not in g
orm.branch = 'b'
assert 2 not in g
assert 1 in g
orm.turn = 2
assert 2 in g
orm.turn = 1
orm.branch = 'trunk'
orm.turn = 0
assert 1 not in g
orm.branch = 'c'
orm.turn = 2
assert 1 not in g
assert 2 not in g
orm.turn = 0
orm.branch = 'trunk'
orm.turn = 2
assert 2 in g | import pytest
import allegedb
@pytest.fixture(scope='function')
def orm():
with allegedb.ORM("sqlite:///:memory:") as it:
yield it
def test_single_plan(orm):
g = orm.new_graph('graph')
g.add_node(0)
orm.turn = 1
g.add_node(1)
with orm.plan():
orm.turn = 2
g.add_node(2)
assert orm.turn == 1
assert 2 not in g
orm.branch = 'b'
assert 2 not in g
assert 1 in g
orm.turn = 2
assert 2 in g
orm.turn = 1
orm.branch = 'trunk'
orm.turn = 0
assert 1 not in g
orm.branch = 'c'
orm.turn = 2
assert 1 not in g
assert 2 not in g
+ orm.turn = 0
+ orm.branch = 'trunk'
+ orm.turn = 2
+ assert 2 in g |
39561a89ea497776d980d3eda97fc2f75493528f | internal_social_auth/views.py | internal_social_auth/views.py | import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, "Your Google Apps domain isn't authorized for this app")
return HttpResponseRedirect('/')
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
| import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, self.get_error_message())
return HttpResponseRedirect(self.get_faiure_url())
def get_error_message(self):
if self.error_message:
return self.error_message
return "Your Google Apps domain isn't authorized for this app"
def get_failure_url(self):
if self.failure_url:
return force_text(self.failure_url)
return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
| Clean up the AuthComplete API a little | Clean up the AuthComplete API a little
| Python | bsd-2-clause | incuna/incuna-internal-social-auth | import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
+ from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
+ messages.error(request, self.get_error_message())
+ return HttpResponseRedirect(self.get_faiure_url())
+
+ def get_error_message(self):
+ if self.error_message:
+ return self.error_message
- messages.error(request, "Your Google Apps domain isn't authorized for this app")
+ return "Your Google Apps domain isn't authorized for this app"
- return HttpResponseRedirect('/')
+
+ def get_failure_url(self):
+ if self.failure_url:
+ return force_text(self.failure_url)
+ return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
- | Clean up the AuthComplete API a little | ## Code Before:
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, "Your Google Apps domain isn't authorized for this app")
return HttpResponseRedirect('/')
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
## Instruction:
Clean up the AuthComplete API a little
## Code After:
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, self.get_error_message())
return HttpResponseRedirect(self.get_faiure_url())
def get_error_message(self):
if self.error_message:
return self.error_message
return "Your Google Apps domain isn't authorized for this app"
def get_failure_url(self):
if self.failure_url:
return force_text(self.failure_url)
return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
| import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
+ from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
+ messages.error(request, self.get_error_message())
+ return HttpResponseRedirect(self.get_faiure_url())
+
+ def get_error_message(self):
+ if self.error_message:
+ return self.error_message
- messages.error(request, "Your Google Apps domain isn't authorized for this app")
? --------------- ^ ^^^^ -
+ return "Your Google Apps domain isn't authorized for this app"
? ^ ^^
- return HttpResponseRedirect('/')
+
+ def get_failure_url(self):
+ if self.failure_url:
+ return force_text(self.failure_url)
+ return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
- |
3b6ddce7c0db0f0b1fbd9febd9bf68ceeda51f44 | della/user_manager/forms.py | della/user_manager/forms.py | from django import forms
from django.contrib.auth.models import User
from django.contrib.auth.forms import UserCreationForm
from django.core.validators import RegexValidator
alphanumericu = RegexValidator(
regex=r'^[0-9a-zA-Z_]*$',
message='Only alphanumeric characters and underscore are allowed.')
class SignupForm(UserCreationForm):
username = forms.CharField(max_length=20, validators=[alphanumericu])
email = forms.EmailField(max_length=254, required=True)
class Meta:
model = User
fields = ['email', 'username', ]
def clean_email(self):
error_message = 'An user with that email already exists'
email = self.cleaned_data.get('email')
if email and User.objects.filter(email=email).exists():
raise forms.ValidationError(
self.error_messages[error_message],
code='existing_email',
)
return email
| from django import forms
from django.contrib.auth.models import User
from django.contrib.auth.forms import UserCreationForm
from django.core.validators import RegexValidator
alphanumericu = RegexValidator(
regex=r'^[0-9a-zA-Z_]*$',
message='Only alphanumeric characters and underscore are allowed.')
class SignupForm(UserCreationForm):
username = forms.CharField(max_length=20, validators=[alphanumericu])
email = forms.EmailField(max_length=254, required=True)
class Meta:
model = User
fields = ['email', 'username', ]
def clean_email(self):
error_message = 'An user with that email already exists'
email = self.cleaned_data.get('email')
if email and User.objects.filter(email=email).exists():
raise forms.ValidationError(error_message)
return email
| Raise ValidationError properly in SignupForm | Raise ValidationError properly in SignupForm
| Python | mit | avinassh/della,avinassh/della,avinassh/della | from django import forms
from django.contrib.auth.models import User
from django.contrib.auth.forms import UserCreationForm
from django.core.validators import RegexValidator
alphanumericu = RegexValidator(
regex=r'^[0-9a-zA-Z_]*$',
message='Only alphanumeric characters and underscore are allowed.')
class SignupForm(UserCreationForm):
username = forms.CharField(max_length=20, validators=[alphanumericu])
email = forms.EmailField(max_length=254, required=True)
class Meta:
model = User
fields = ['email', 'username', ]
def clean_email(self):
error_message = 'An user with that email already exists'
email = self.cleaned_data.get('email')
if email and User.objects.filter(email=email).exists():
- raise forms.ValidationError(
+ raise forms.ValidationError(error_message)
- self.error_messages[error_message],
- code='existing_email',
- )
return email
| Raise ValidationError properly in SignupForm | ## Code Before:
from django import forms
from django.contrib.auth.models import User
from django.contrib.auth.forms import UserCreationForm
from django.core.validators import RegexValidator
alphanumericu = RegexValidator(
regex=r'^[0-9a-zA-Z_]*$',
message='Only alphanumeric characters and underscore are allowed.')
class SignupForm(UserCreationForm):
username = forms.CharField(max_length=20, validators=[alphanumericu])
email = forms.EmailField(max_length=254, required=True)
class Meta:
model = User
fields = ['email', 'username', ]
def clean_email(self):
error_message = 'An user with that email already exists'
email = self.cleaned_data.get('email')
if email and User.objects.filter(email=email).exists():
raise forms.ValidationError(
self.error_messages[error_message],
code='existing_email',
)
return email
## Instruction:
Raise ValidationError properly in SignupForm
## Code After:
from django import forms
from django.contrib.auth.models import User
from django.contrib.auth.forms import UserCreationForm
from django.core.validators import RegexValidator
alphanumericu = RegexValidator(
regex=r'^[0-9a-zA-Z_]*$',
message='Only alphanumeric characters and underscore are allowed.')
class SignupForm(UserCreationForm):
username = forms.CharField(max_length=20, validators=[alphanumericu])
email = forms.EmailField(max_length=254, required=True)
class Meta:
model = User
fields = ['email', 'username', ]
def clean_email(self):
error_message = 'An user with that email already exists'
email = self.cleaned_data.get('email')
if email and User.objects.filter(email=email).exists():
raise forms.ValidationError(error_message)
return email
| from django import forms
from django.contrib.auth.models import User
from django.contrib.auth.forms import UserCreationForm
from django.core.validators import RegexValidator
alphanumericu = RegexValidator(
regex=r'^[0-9a-zA-Z_]*$',
message='Only alphanumeric characters and underscore are allowed.')
class SignupForm(UserCreationForm):
username = forms.CharField(max_length=20, validators=[alphanumericu])
email = forms.EmailField(max_length=254, required=True)
class Meta:
model = User
fields = ['email', 'username', ]
def clean_email(self):
error_message = 'An user with that email already exists'
email = self.cleaned_data.get('email')
if email and User.objects.filter(email=email).exists():
- raise forms.ValidationError(
+ raise forms.ValidationError(error_message)
? ++++++++++++++
- self.error_messages[error_message],
- code='existing_email',
- )
return email |
eb7ff9cec9360af0b5c18915164a54d4755e657b | mistraldashboard/dashboards/mistral/executions/tables.py | mistraldashboard/dashboards/mistral/executions/tables.py |
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
action = tables.Column("action", verbose_name=_("Action"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks")
|
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
parameters = tables.Column("parameters", verbose_name=_("Parameters"))
output = tables.Column("output", verbose_name=_("Output"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks")
| Add Task's output and parameters columns | Add Task's output and parameters columns
Change-Id: I98f57a6a0178bb7258d82f3a165127f060f42f7b
Implements: blueprint mistral-ui
| Python | apache-2.0 | openstack/mistral-dashboard,openstack/mistral-dashboard,openstack/mistral-dashboard |
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
+ parameters = tables.Column("parameters", verbose_name=_("Parameters"))
- action = tables.Column("action", verbose_name=_("Action"))
+ output = tables.Column("output", verbose_name=_("Output"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks")
| Add Task's output and parameters columns | ## Code Before:
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
action = tables.Column("action", verbose_name=_("Action"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks")
## Instruction:
Add Task's output and parameters columns
## Code After:
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
parameters = tables.Column("parameters", verbose_name=_("Parameters"))
output = tables.Column("output", verbose_name=_("Output"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks")
|
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
+ parameters = tables.Column("parameters", verbose_name=_("Parameters"))
- action = tables.Column("action", verbose_name=_("Action"))
? ^^ ^^^ ^^ ^^^ ^^ ^^^
+ output = tables.Column("output", verbose_name=_("Output"))
? ^^ ^^^ ^^ ^^^ ^^ ^^^
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks") |
e388e3490502acac90ef4c249ba1af63b5698ab7 | print_web_django/api/views.py | print_web_django/api/views.py | from rest_framework import viewsets
from . import serializers, models
class PrintJobViewSet(viewsets.ModelViewSet):
serializer_class = serializers.PrintJobSerializer
def get_queryset(self):
return self.request.user.printjobs.all()
| from rest_framework import viewsets
from . import serializers, models
class PrintJobViewSet(viewsets.ModelViewSet):
serializer_class = serializers.PrintJobSerializer
def get_queryset(self):
return self.request.user.printjobs.all()
def perform_create(self, serializer):
# need to also pass the requests user on a create
serializer.save(user=self.request.user)
| Add user to posted print object | Add user to posted print object
| Python | mit | aabmass/print-web,aabmass/print-web,aabmass/print-web | from rest_framework import viewsets
from . import serializers, models
class PrintJobViewSet(viewsets.ModelViewSet):
serializer_class = serializers.PrintJobSerializer
def get_queryset(self):
return self.request.user.printjobs.all()
+ def perform_create(self, serializer):
+ # need to also pass the requests user on a create
+ serializer.save(user=self.request.user)
+ | Add user to posted print object | ## Code Before:
from rest_framework import viewsets
from . import serializers, models
class PrintJobViewSet(viewsets.ModelViewSet):
serializer_class = serializers.PrintJobSerializer
def get_queryset(self):
return self.request.user.printjobs.all()
## Instruction:
Add user to posted print object
## Code After:
from rest_framework import viewsets
from . import serializers, models
class PrintJobViewSet(viewsets.ModelViewSet):
serializer_class = serializers.PrintJobSerializer
def get_queryset(self):
return self.request.user.printjobs.all()
def perform_create(self, serializer):
# need to also pass the requests user on a create
serializer.save(user=self.request.user)
| from rest_framework import viewsets
from . import serializers, models
class PrintJobViewSet(viewsets.ModelViewSet):
serializer_class = serializers.PrintJobSerializer
def get_queryset(self):
return self.request.user.printjobs.all()
+
+ def perform_create(self, serializer):
+ # need to also pass the requests user on a create
+ serializer.save(user=self.request.user) |
f88c2135ddc197283bbfb8b481774deb613571cf | python/raindrops/raindrops.py | python/raindrops/raindrops.py | def raindrops(number):
if is_three_a_factor(number):
return "Pling"
return "{}".format(number)
def is_three_a_factor(number):
return number % 3 == 0
| def raindrops(number):
if is_three_a_factor(number):
return "Pling"
if is_five_a_factor(number):
return "Plang"
return "{}".format(number)
def is_three_a_factor(number):
return number % 3 == 0
def is_five_a_factor(number):
return number % 5 == 0
| Handle 5 as a factor | Handle 5 as a factor
| Python | mit | rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism | def raindrops(number):
if is_three_a_factor(number):
return "Pling"
+ if is_five_a_factor(number):
+ return "Plang"
return "{}".format(number)
def is_three_a_factor(number):
return number % 3 == 0
+ def is_five_a_factor(number):
+ return number % 5 == 0
+ | Handle 5 as a factor | ## Code Before:
def raindrops(number):
if is_three_a_factor(number):
return "Pling"
return "{}".format(number)
def is_three_a_factor(number):
return number % 3 == 0
## Instruction:
Handle 5 as a factor
## Code After:
def raindrops(number):
if is_three_a_factor(number):
return "Pling"
if is_five_a_factor(number):
return "Plang"
return "{}".format(number)
def is_three_a_factor(number):
return number % 3 == 0
def is_five_a_factor(number):
return number % 5 == 0
| def raindrops(number):
if is_three_a_factor(number):
return "Pling"
+ if is_five_a_factor(number):
+ return "Plang"
return "{}".format(number)
def is_three_a_factor(number):
return number % 3 == 0
+
+ def is_five_a_factor(number):
+ return number % 5 == 0 |
114f40dd282d1837db42ffb6625760d1483d3192 | jfu/templatetags/jfutags.py | jfu/templatetags/jfutags.py | from django.core.context_processors import csrf
from django.core.urlresolvers import reverse
from django.template import Library, Context, loader
register = Library()
@register.simple_tag( takes_context = True )
def jfu(
context,
template_name = 'jfu/upload_form.html',
upload_handler_name = 'jfu_upload'
):
"""
Displays a form for uploading files using jQuery File Upload.
A user may supply both a custom template or a custom upload-handling URL
name by supplying values for template_name and upload_handler_name
respectively.
"""
context.update( {
'JQ_OPEN' : '{%',
'JQ_CLOSE' : '%}',
'upload_handler_url': reverse( upload_handler_name ),
} )
# Use the request context variable, injected
# by django.core.context_processors.request
# to generate the CSRF token.
context.update( csrf( context.get('request') ) )
t = loader.get_template( template_name )
return t.render( Context( context ) )
| from django.core.context_processors import csrf
from django.core.urlresolvers import reverse
from django.template import Library, Context, loader
register = Library()
@register.simple_tag( takes_context = True )
def jfu(
context,
template_name = 'jfu/upload_form.html',
upload_handler_name = 'jfu_upload',
*args, **kwargs
):
"""
Displays a form for uploading files using jQuery File Upload.
A user may supply both a custom template or a custom upload-handling URL
name by supplying values for template_name and upload_handler_name
respectively.
"""
context.update( {
'JQ_OPEN' : '{%',
'JQ_CLOSE' : '%}',
'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ),
} )
# Use the request context variable, injected
# by django.core.context_processors.request
# to generate the CSRF token.
context.update( csrf( context.get('request') ) )
t = loader.get_template( template_name )
return t.render( Context( context ) )
| Allow args and kwargs to upload_handler_name | Allow args and kwargs to upload_handler_name
Now can use args and kwargs for reverse url. Example in template:
{% jfu 'core/core_fileuploader.html' 'core_upload' object_id=1 content_type_str='app.model' %} | Python | bsd-3-clause | Alem/django-jfu,dzhuang/django-jfu,Alem/django-jfu,dzhuang/django-jfu,Alem/django-jfu,dzhuang/django-jfu,dzhuang/django-jfu,Alem/django-jfu | from django.core.context_processors import csrf
from django.core.urlresolvers import reverse
from django.template import Library, Context, loader
register = Library()
@register.simple_tag( takes_context = True )
def jfu(
context,
template_name = 'jfu/upload_form.html',
- upload_handler_name = 'jfu_upload'
+ upload_handler_name = 'jfu_upload',
+ *args, **kwargs
):
"""
Displays a form for uploading files using jQuery File Upload.
A user may supply both a custom template or a custom upload-handling URL
name by supplying values for template_name and upload_handler_name
respectively.
"""
context.update( {
'JQ_OPEN' : '{%',
'JQ_CLOSE' : '%}',
- 'upload_handler_url': reverse( upload_handler_name ),
+ 'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ),
} )
# Use the request context variable, injected
# by django.core.context_processors.request
# to generate the CSRF token.
context.update( csrf( context.get('request') ) )
t = loader.get_template( template_name )
return t.render( Context( context ) )
| Allow args and kwargs to upload_handler_name | ## Code Before:
from django.core.context_processors import csrf
from django.core.urlresolvers import reverse
from django.template import Library, Context, loader
register = Library()
@register.simple_tag( takes_context = True )
def jfu(
context,
template_name = 'jfu/upload_form.html',
upload_handler_name = 'jfu_upload'
):
"""
Displays a form for uploading files using jQuery File Upload.
A user may supply both a custom template or a custom upload-handling URL
name by supplying values for template_name and upload_handler_name
respectively.
"""
context.update( {
'JQ_OPEN' : '{%',
'JQ_CLOSE' : '%}',
'upload_handler_url': reverse( upload_handler_name ),
} )
# Use the request context variable, injected
# by django.core.context_processors.request
# to generate the CSRF token.
context.update( csrf( context.get('request') ) )
t = loader.get_template( template_name )
return t.render( Context( context ) )
## Instruction:
Allow args and kwargs to upload_handler_name
## Code After:
from django.core.context_processors import csrf
from django.core.urlresolvers import reverse
from django.template import Library, Context, loader
register = Library()
@register.simple_tag( takes_context = True )
def jfu(
context,
template_name = 'jfu/upload_form.html',
upload_handler_name = 'jfu_upload',
*args, **kwargs
):
"""
Displays a form for uploading files using jQuery File Upload.
A user may supply both a custom template or a custom upload-handling URL
name by supplying values for template_name and upload_handler_name
respectively.
"""
context.update( {
'JQ_OPEN' : '{%',
'JQ_CLOSE' : '%}',
'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ),
} )
# Use the request context variable, injected
# by django.core.context_processors.request
# to generate the CSRF token.
context.update( csrf( context.get('request') ) )
t = loader.get_template( template_name )
return t.render( Context( context ) )
| from django.core.context_processors import csrf
from django.core.urlresolvers import reverse
from django.template import Library, Context, loader
register = Library()
@register.simple_tag( takes_context = True )
def jfu(
context,
template_name = 'jfu/upload_form.html',
- upload_handler_name = 'jfu_upload'
? ^
+ upload_handler_name = 'jfu_upload',
? ^
+ *args, **kwargs
):
"""
Displays a form for uploading files using jQuery File Upload.
A user may supply both a custom template or a custom upload-handling URL
name by supplying values for template_name and upload_handler_name
respectively.
"""
context.update( {
'JQ_OPEN' : '{%',
'JQ_CLOSE' : '%}',
- 'upload_handler_url': reverse( upload_handler_name ),
+ 'upload_handler_url': reverse( upload_handler_name, kwargs=kwargs, args=args ),
? ++++++++++++++++++++++++++
} )
# Use the request context variable, injected
# by django.core.context_processors.request
# to generate the CSRF token.
context.update( csrf( context.get('request') ) )
t = loader.get_template( template_name )
return t.render( Context( context ) ) |
de1988304714b44e641a4c4ac50fa650887621d6 | geoportail/geonames/views.py | geoportail/geonames/views.py | import unicodedata
from django.http import HttpResponse
from django.template.defaultfilters import slugify
from django.utils.translation import ugettext as _
from .models import Town
def autocomplete(request):
if not 'q' in request.GET or len(request.GET['q']) < 3:
response = HttpResponse()
response.status_code = 204
return response
query = slugify(request.GET['q']).replace('-', ' ').upper()
if query.startswith('ST '):
query = 'SAINT ' + query[3:]
towns = Town.objects.filter(
tokenized__startswith=query
).order_by('tokenized', 'postal_code')[:15]
content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format(
name=unicodedata.normalize('NFKD', t.name),
county_name=t.county_name,
lon=t.point.coords[0],
lat=t.point.coords[1],
) for t in towns])
if not content:
content = _('No results. Search is limited to city names.')
return HttpResponse(content)
| import json
import unicodedata
from django.http import HttpResponse
from django.template.defaultfilters import slugify
from django.utils.translation import ugettext as _
from .models import Town
def autocomplete(request):
if not 'q' in request.GET or len(request.GET['q']) < 3:
response = HttpResponse()
response.status_code = 204
return response
query = slugify(request.GET['q']).replace('-', ' ').upper()
if query.startswith('ST '):
query = 'SAINT ' + query[3:]
towns = Town.objects.filter(
tokenized__startswith=query
).order_by('tokenized', 'postal_code')[:15]
content = [{
"name": unicodedata.normalize('NFKD', t.name),
"county_name": t.county_name,
"lon": t.point.coords[0],
"lat": t.point.coords[1],
} for t in towns]
if not content:
content = [{'name': _('No results. Search is limited to city names.')}]
return HttpResponse(json.dumps(content), content_type='application/json')
| Return JSON in the autocomplete view | Return JSON in the autocomplete view
| Python | bsd-3-clause | brutasse/geoportail,brutasse/geoportail,brutasse/geoportail | + import json
import unicodedata
from django.http import HttpResponse
from django.template.defaultfilters import slugify
from django.utils.translation import ugettext as _
from .models import Town
def autocomplete(request):
if not 'q' in request.GET or len(request.GET['q']) < 3:
response = HttpResponse()
response.status_code = 204
return response
query = slugify(request.GET['q']).replace('-', ' ').upper()
if query.startswith('ST '):
query = 'SAINT ' + query[3:]
towns = Town.objects.filter(
tokenized__startswith=query
).order_by('tokenized', 'postal_code')[:15]
- content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format(
+ content = [{
- name=unicodedata.normalize('NFKD', t.name),
+ "name": unicodedata.normalize('NFKD', t.name),
- county_name=t.county_name,
+ "county_name": t.county_name,
- lon=t.point.coords[0],
+ "lon": t.point.coords[0],
- lat=t.point.coords[1],
+ "lat": t.point.coords[1],
- ) for t in towns])
+ } for t in towns]
if not content:
- content = _('No results. Search is limited to city names.')
+ content = [{'name': _('No results. Search is limited to city names.')}]
- return HttpResponse(content)
+ return HttpResponse(json.dumps(content), content_type='application/json')
| Return JSON in the autocomplete view | ## Code Before:
import unicodedata
from django.http import HttpResponse
from django.template.defaultfilters import slugify
from django.utils.translation import ugettext as _
from .models import Town
def autocomplete(request):
if not 'q' in request.GET or len(request.GET['q']) < 3:
response = HttpResponse()
response.status_code = 204
return response
query = slugify(request.GET['q']).replace('-', ' ').upper()
if query.startswith('ST '):
query = 'SAINT ' + query[3:]
towns = Town.objects.filter(
tokenized__startswith=query
).order_by('tokenized', 'postal_code')[:15]
content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format(
name=unicodedata.normalize('NFKD', t.name),
county_name=t.county_name,
lon=t.point.coords[0],
lat=t.point.coords[1],
) for t in towns])
if not content:
content = _('No results. Search is limited to city names.')
return HttpResponse(content)
## Instruction:
Return JSON in the autocomplete view
## Code After:
import json
import unicodedata
from django.http import HttpResponse
from django.template.defaultfilters import slugify
from django.utils.translation import ugettext as _
from .models import Town
def autocomplete(request):
if not 'q' in request.GET or len(request.GET['q']) < 3:
response = HttpResponse()
response.status_code = 204
return response
query = slugify(request.GET['q']).replace('-', ' ').upper()
if query.startswith('ST '):
query = 'SAINT ' + query[3:]
towns = Town.objects.filter(
tokenized__startswith=query
).order_by('tokenized', 'postal_code')[:15]
content = [{
"name": unicodedata.normalize('NFKD', t.name),
"county_name": t.county_name,
"lon": t.point.coords[0],
"lat": t.point.coords[1],
} for t in towns]
if not content:
content = [{'name': _('No results. Search is limited to city names.')}]
return HttpResponse(json.dumps(content), content_type='application/json')
| + import json
import unicodedata
from django.http import HttpResponse
from django.template.defaultfilters import slugify
from django.utils.translation import ugettext as _
from .models import Town
def autocomplete(request):
if not 'q' in request.GET or len(request.GET['q']) < 3:
response = HttpResponse()
response.status_code = 204
return response
query = slugify(request.GET['q']).replace('-', ' ').upper()
if query.startswith('ST '):
query = 'SAINT ' + query[3:]
towns = Town.objects.filter(
tokenized__startswith=query
).order_by('tokenized', 'postal_code')[:15]
- content = u'\n'.join([u'{name} <em>{county_name}</em>|{lon} {lat}'.format(
+ content = [{
- name=unicodedata.normalize('NFKD', t.name),
? ^
+ "name": unicodedata.normalize('NFKD', t.name),
? + ^^^
- county_name=t.county_name,
? ^
+ "county_name": t.county_name,
? + ^^^
- lon=t.point.coords[0],
? ^
+ "lon": t.point.coords[0],
? + ^^^
- lat=t.point.coords[1],
? ^
+ "lat": t.point.coords[1],
? + ^^^
- ) for t in towns])
? ^ -
+ } for t in towns]
? ^
if not content:
- content = _('No results. Search is limited to city names.')
+ content = [{'name': _('No results. Search is limited to city names.')}]
? ++++++++++ ++
- return HttpResponse(content)
+ return HttpResponse(json.dumps(content), content_type='application/json') |
d4a0a85673b5d61b82c65e77efcd6518da719952 | pmxbot/__init__.py | pmxbot/__init__.py |
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
| Remove places default config. It doesn't appear to be used anywhere. | Remove places default config. It doesn't appear to be used anywhere.
| Python | mit | yougov/pmxbot,yougov/pmxbot,yougov/pmxbot |
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
- places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
| Remove places default config. It doesn't appear to be used anywhere. | ## Code Before:
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
## Instruction:
Remove places default config. It doesn't appear to be used anywhere.
## Code After:
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
- places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object" |
c10badab9b93eb021b1942475c681042292c182c | scrapi/harvesters/boise_state.py | scrapi/harvesters/boise_state.py | '''
Harvester for the ScholarWorks for the SHARE project
Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc
'''
from __future__ import unicode_literals
from scrapi.base import OAIHarvester
class Boise_stateHarvester(OAIHarvester):
short_name = 'boise_state'
long_name = 'ScholarWorks'
url = 'http://scholarworks.boisestate.edu'
base_url = 'http://scholarworks.boisestate.edu/do/oai/'
property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format']
timezone_granularity = True
| '''
Harvester for the ScholarWorks for the SHARE project
Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc
'''
from __future__ import unicode_literals
from scrapi.base import OAIHarvester
class Boise_stateHarvester(OAIHarvester):
short_name = 'boise_state'
long_name = 'Boise State University ScholarWorks'
url = 'http://scholarworks.boisestate.edu'
base_url = 'http://scholarworks.boisestate.edu/do/oai/'
property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format']
timezone_granularity = True
| Update longname for Boise state | Update longname for Boise state
| Python | apache-2.0 | CenterForOpenScience/scrapi,CenterForOpenScience/scrapi | '''
Harvester for the ScholarWorks for the SHARE project
Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc
'''
from __future__ import unicode_literals
from scrapi.base import OAIHarvester
class Boise_stateHarvester(OAIHarvester):
short_name = 'boise_state'
- long_name = 'ScholarWorks'
+ long_name = 'Boise State University ScholarWorks'
url = 'http://scholarworks.boisestate.edu'
base_url = 'http://scholarworks.boisestate.edu/do/oai/'
property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format']
timezone_granularity = True
| Update longname for Boise state | ## Code Before:
'''
Harvester for the ScholarWorks for the SHARE project
Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc
'''
from __future__ import unicode_literals
from scrapi.base import OAIHarvester
class Boise_stateHarvester(OAIHarvester):
short_name = 'boise_state'
long_name = 'ScholarWorks'
url = 'http://scholarworks.boisestate.edu'
base_url = 'http://scholarworks.boisestate.edu/do/oai/'
property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format']
timezone_granularity = True
## Instruction:
Update longname for Boise state
## Code After:
'''
Harvester for the ScholarWorks for the SHARE project
Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc
'''
from __future__ import unicode_literals
from scrapi.base import OAIHarvester
class Boise_stateHarvester(OAIHarvester):
short_name = 'boise_state'
long_name = 'Boise State University ScholarWorks'
url = 'http://scholarworks.boisestate.edu'
base_url = 'http://scholarworks.boisestate.edu/do/oai/'
property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format']
timezone_granularity = True
| '''
Harvester for the ScholarWorks for the SHARE project
Example API call: http://scholarworks.boisestate.edu/do/oai/?verb=ListRecords&metadataPrefix=oai_dc
'''
from __future__ import unicode_literals
from scrapi.base import OAIHarvester
class Boise_stateHarvester(OAIHarvester):
short_name = 'boise_state'
- long_name = 'ScholarWorks'
+ long_name = 'Boise State University ScholarWorks'
url = 'http://scholarworks.boisestate.edu'
base_url = 'http://scholarworks.boisestate.edu/do/oai/'
property_list = ['source', 'identifier', 'type', 'date', 'setSpec', 'publisher', 'rights', 'format']
timezone_granularity = True |
e6d7181ababaa9f08602c48e03d6557ddb6a4deb | tests/test_gio.py | tests/test_gio.py |
import os
import unittest
from common import gio, gobject
class TestInputStream(unittest.TestCase):
def setUp(self):
f = open("inputstream.txt", "w")
f.write("testing")
self._f = open("inputstream.txt", "r")
self.stream = gio.unix.InputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("inputstream.txt")
def testWrite(self):
self.assertEquals(self.stream.read(), "testing")
class TestOutputStream(unittest.TestCase):
def setUp(self):
self._f = open("outputstream.txt", "w")
self.stream = gio.unix.OutputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("outputstream.txt")
def testWrite(self):
self.stream.write("testing")
self.stream.close()
self.failUnless(os.path.exists("outputstream.txt"))
self.assertEquals(open("outputstream.txt").read(), "testing")
def testWriteAsync(self):
def callback(stream, result):
loop.quit()
f = gio.file_new_for_path("outputstream.txt")
stream = f.read()
stream.read_async(10240, 0, None, callback)
loop = gobject.MainLoop()
loop.run()
|
import os
import unittest
from common import gio, gobject
class TestInputStream(unittest.TestCase):
def setUp(self):
f = open("inputstream.txt", "w")
f.write("testing")
self._f = open("inputstream.txt", "r")
self.stream = gio.unix.InputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("inputstream.txt")
def testRead(self):
self.assertEquals(self.stream.read(), "testing")
def testReadAsync(self):
def callback(stream, result):
self.assertEquals(stream.read_finish(result), len("testing"))
loop.quit()
self.stream.read_async(10240, 0, None, callback)
loop = gobject.MainLoop()
loop.run()
class TestOutputStream(unittest.TestCase):
def setUp(self):
self._f = open("outputstream.txt", "w")
self.stream = gio.unix.OutputStream(self._f.fileno(), False)
self._f.flush()
def tearDown(self):
self._f.close()
os.unlink("outputstream.txt")
def testWrite(self):
self.stream.write("testing")
self.stream.close()
self.failUnless(os.path.exists("outputstream.txt"))
self.assertEquals(open("outputstream.txt").read(), "testing")
| Reorganize tests and make them test more useful things | Reorganize tests and make them test more useful things
svn path=/trunk/; revision=738
| Python | lgpl-2.1 | pexip/pygobject,GNOME/pygobject,davibe/pygobject,alexef/pygobject,davidmalcolm/pygobject,MathieuDuponchelle/pygobject,davidmalcolm/pygobject,Distrotech/pygobject,choeger/pygobject-cmake,sfeltman/pygobject,Distrotech/pygobject,MathieuDuponchelle/pygobject,GNOME/pygobject,thiblahute/pygobject,jdahlin/pygobject,atizo/pygobject,alexef/pygobject,jdahlin/pygobject,choeger/pygobject-cmake,thiblahute/pygobject,GNOME/pygobject,nzjrs/pygobject,Distrotech/pygobject,pexip/pygobject,pexip/pygobject,atizo/pygobject,davibe/pygobject,choeger/pygobject-cmake,alexef/pygobject,davibe/pygobject,davibe/pygobject,MathieuDuponchelle/pygobject,Distrotech/pygobject,jdahlin/pygobject,sfeltman/pygobject,nzjrs/pygobject,thiblahute/pygobject,atizo/pygobject,nzjrs/pygobject,davidmalcolm/pygobject,sfeltman/pygobject |
import os
import unittest
from common import gio, gobject
class TestInputStream(unittest.TestCase):
def setUp(self):
f = open("inputstream.txt", "w")
f.write("testing")
self._f = open("inputstream.txt", "r")
self.stream = gio.unix.InputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("inputstream.txt")
- def testWrite(self):
+ def testRead(self):
self.assertEquals(self.stream.read(), "testing")
+
+ def testReadAsync(self):
+ def callback(stream, result):
+ self.assertEquals(stream.read_finish(result), len("testing"))
+ loop.quit()
+
+ self.stream.read_async(10240, 0, None, callback)
+
+ loop = gobject.MainLoop()
+ loop.run()
class TestOutputStream(unittest.TestCase):
def setUp(self):
self._f = open("outputstream.txt", "w")
self.stream = gio.unix.OutputStream(self._f.fileno(), False)
+ self._f.flush()
def tearDown(self):
self._f.close()
os.unlink("outputstream.txt")
def testWrite(self):
self.stream.write("testing")
self.stream.close()
self.failUnless(os.path.exists("outputstream.txt"))
self.assertEquals(open("outputstream.txt").read(), "testing")
- def testWriteAsync(self):
- def callback(stream, result):
- loop.quit()
- f = gio.file_new_for_path("outputstream.txt")
- stream = f.read()
- stream.read_async(10240, 0, None, callback)
-
- loop = gobject.MainLoop()
- loop.run()
- | Reorganize tests and make them test more useful things | ## Code Before:
import os
import unittest
from common import gio, gobject
class TestInputStream(unittest.TestCase):
def setUp(self):
f = open("inputstream.txt", "w")
f.write("testing")
self._f = open("inputstream.txt", "r")
self.stream = gio.unix.InputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("inputstream.txt")
def testWrite(self):
self.assertEquals(self.stream.read(), "testing")
class TestOutputStream(unittest.TestCase):
def setUp(self):
self._f = open("outputstream.txt", "w")
self.stream = gio.unix.OutputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("outputstream.txt")
def testWrite(self):
self.stream.write("testing")
self.stream.close()
self.failUnless(os.path.exists("outputstream.txt"))
self.assertEquals(open("outputstream.txt").read(), "testing")
def testWriteAsync(self):
def callback(stream, result):
loop.quit()
f = gio.file_new_for_path("outputstream.txt")
stream = f.read()
stream.read_async(10240, 0, None, callback)
loop = gobject.MainLoop()
loop.run()
## Instruction:
Reorganize tests and make them test more useful things
## Code After:
import os
import unittest
from common import gio, gobject
class TestInputStream(unittest.TestCase):
def setUp(self):
f = open("inputstream.txt", "w")
f.write("testing")
self._f = open("inputstream.txt", "r")
self.stream = gio.unix.InputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("inputstream.txt")
def testRead(self):
self.assertEquals(self.stream.read(), "testing")
def testReadAsync(self):
def callback(stream, result):
self.assertEquals(stream.read_finish(result), len("testing"))
loop.quit()
self.stream.read_async(10240, 0, None, callback)
loop = gobject.MainLoop()
loop.run()
class TestOutputStream(unittest.TestCase):
def setUp(self):
self._f = open("outputstream.txt", "w")
self.stream = gio.unix.OutputStream(self._f.fileno(), False)
self._f.flush()
def tearDown(self):
self._f.close()
os.unlink("outputstream.txt")
def testWrite(self):
self.stream.write("testing")
self.stream.close()
self.failUnless(os.path.exists("outputstream.txt"))
self.assertEquals(open("outputstream.txt").read(), "testing")
|
import os
import unittest
from common import gio, gobject
class TestInputStream(unittest.TestCase):
def setUp(self):
f = open("inputstream.txt", "w")
f.write("testing")
self._f = open("inputstream.txt", "r")
self.stream = gio.unix.InputStream(self._f.fileno(), False)
def tearDown(self):
self._f.close()
os.unlink("inputstream.txt")
- def testWrite(self):
? ^^^^
+ def testRead(self):
? ^ ++
self.assertEquals(self.stream.read(), "testing")
+
+ def testReadAsync(self):
+ def callback(stream, result):
+ self.assertEquals(stream.read_finish(result), len("testing"))
+ loop.quit()
+
+ self.stream.read_async(10240, 0, None, callback)
+
+ loop = gobject.MainLoop()
+ loop.run()
class TestOutputStream(unittest.TestCase):
def setUp(self):
self._f = open("outputstream.txt", "w")
self.stream = gio.unix.OutputStream(self._f.fileno(), False)
+ self._f.flush()
def tearDown(self):
self._f.close()
os.unlink("outputstream.txt")
def testWrite(self):
self.stream.write("testing")
self.stream.close()
self.failUnless(os.path.exists("outputstream.txt"))
self.assertEquals(open("outputstream.txt").read(), "testing")
- def testWriteAsync(self):
- def callback(stream, result):
- loop.quit()
-
- f = gio.file_new_for_path("outputstream.txt")
- stream = f.read()
- stream.read_async(10240, 0, None, callback)
-
- loop = gobject.MainLoop()
- loop.run() |
0f77c9a48e84a3185794f97c5f15c7b13ae1d505 | tests/test_vector2_angle.py | tests/test_vector2_angle.py | from ppb_vector import Vector2
from math import isclose
import pytest
@pytest.mark.parametrize("left, right, expected", [
(Vector2(1, 1), Vector2(0, -1), 135),
(Vector2(1, 1), Vector2(-1, 0), 135),
(Vector2(0, 1), Vector2(0, -1), 180),
(Vector2(-1, -1), Vector2(1, 0), 135),
(Vector2(-1, -1), Vector2(-1, 0), 45)
])
def test_angle(left, right, expected):
assert isclose(left.angle(right), expected)
assert isclose(right.angle(left), expected)
| from ppb_vector import Vector2
from math import isclose
import pytest
@pytest.mark.parametrize("left, right, expected", [
(Vector2(1, 1), Vector2(0, -1), 135),
(Vector2(1, 1), Vector2(-1, 0), 135),
(Vector2(0, 1), Vector2(0, -1), 180),
(Vector2(-1, -1), Vector2(1, 0), 135),
(Vector2(-1, -1), Vector2(-1, 0), 45),
(Vector2(1, 0), Vector2(0, 1), 90),
(Vector2(1, 0), Vector2(1, 0), 0),
])
def test_angle(left, right, expected):
assert isclose(left.angle(right), expected)
assert isclose(right.angle(left), expected)
| Add some additional test cases | Add some additional test cases
| Python | artistic-2.0 | ppb/ppb-vector,ppb/ppb-vector | from ppb_vector import Vector2
from math import isclose
import pytest
@pytest.mark.parametrize("left, right, expected", [
(Vector2(1, 1), Vector2(0, -1), 135),
(Vector2(1, 1), Vector2(-1, 0), 135),
(Vector2(0, 1), Vector2(0, -1), 180),
(Vector2(-1, -1), Vector2(1, 0), 135),
- (Vector2(-1, -1), Vector2(-1, 0), 45)
+ (Vector2(-1, -1), Vector2(-1, 0), 45),
+ (Vector2(1, 0), Vector2(0, 1), 90),
+ (Vector2(1, 0), Vector2(1, 0), 0),
])
def test_angle(left, right, expected):
assert isclose(left.angle(right), expected)
assert isclose(right.angle(left), expected)
| Add some additional test cases | ## Code Before:
from ppb_vector import Vector2
from math import isclose
import pytest
@pytest.mark.parametrize("left, right, expected", [
(Vector2(1, 1), Vector2(0, -1), 135),
(Vector2(1, 1), Vector2(-1, 0), 135),
(Vector2(0, 1), Vector2(0, -1), 180),
(Vector2(-1, -1), Vector2(1, 0), 135),
(Vector2(-1, -1), Vector2(-1, 0), 45)
])
def test_angle(left, right, expected):
assert isclose(left.angle(right), expected)
assert isclose(right.angle(left), expected)
## Instruction:
Add some additional test cases
## Code After:
from ppb_vector import Vector2
from math import isclose
import pytest
@pytest.mark.parametrize("left, right, expected", [
(Vector2(1, 1), Vector2(0, -1), 135),
(Vector2(1, 1), Vector2(-1, 0), 135),
(Vector2(0, 1), Vector2(0, -1), 180),
(Vector2(-1, -1), Vector2(1, 0), 135),
(Vector2(-1, -1), Vector2(-1, 0), 45),
(Vector2(1, 0), Vector2(0, 1), 90),
(Vector2(1, 0), Vector2(1, 0), 0),
])
def test_angle(left, right, expected):
assert isclose(left.angle(right), expected)
assert isclose(right.angle(left), expected)
| from ppb_vector import Vector2
from math import isclose
import pytest
@pytest.mark.parametrize("left, right, expected", [
(Vector2(1, 1), Vector2(0, -1), 135),
(Vector2(1, 1), Vector2(-1, 0), 135),
(Vector2(0, 1), Vector2(0, -1), 180),
(Vector2(-1, -1), Vector2(1, 0), 135),
- (Vector2(-1, -1), Vector2(-1, 0), 45)
+ (Vector2(-1, -1), Vector2(-1, 0), 45),
? +
+ (Vector2(1, 0), Vector2(0, 1), 90),
+ (Vector2(1, 0), Vector2(1, 0), 0),
])
def test_angle(left, right, expected):
assert isclose(left.angle(right), expected)
assert isclose(right.angle(left), expected) |
eef7f3797a6228c9e06717c3be49801a10b457a5 | registries/views.py | registries/views.py | from django.http import HttpResponse
from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView
from rest_framework.response import Response
from registries.models import Organization
from registries.serializers import DrillerListSerializer, DrillerSerializer
class APIDrillerListCreateView(ListCreateAPIView):
queryset = Organization.objects.all().select_related('province_state')
serializer_class = DrillerSerializer
def list(self, request):
queryset = self.get_queryset()
serializer = DrillerListSerializer(queryset, many=True)
return Response(serializer.data)
class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView):
queryset = Organization.objects.all()
lookup_field = "org_guid"
serializer_class = DrillerSerializer
# Create your views here.
def index(request):
return HttpResponse("TEST: Driller Register app home index.")
| from django.http import HttpResponse
from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView
from rest_framework.response import Response
from registries.models import Organization
from registries.serializers import DrillerListSerializer, DrillerSerializer
class APIDrillerListCreateView(ListCreateAPIView):
"""
get:
Return a list of all registered drilling organizations
post:
Create a new drilling organization instance
"""
queryset = Organization.objects.all().select_related('province_state')
serializer_class = DrillerSerializer
def list(self, request):
queryset = self.get_queryset()
serializer = DrillerListSerializer(queryset, many=True)
return Response(serializer.data)
class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView):
"""
get:
Return the specified drilling organization
patch:
Updates the specified drilling organization with the fields/values provided in the request body
delete:
Removes the specified drilling organization record
"""
queryset = Organization.objects.all()
lookup_field = "org_guid"
serializer_class = DrillerSerializer
# Create your views here.
def index(request):
return HttpResponse("TEST: Driller Register app home index.")
| Add docstrings to view classes | Add docstrings to view classes
| Python | apache-2.0 | bcgov/gwells,rstens/gwells,rstens/gwells,rstens/gwells,bcgov/gwells,bcgov/gwells,bcgov/gwells,rstens/gwells | from django.http import HttpResponse
from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView
from rest_framework.response import Response
from registries.models import Organization
from registries.serializers import DrillerListSerializer, DrillerSerializer
class APIDrillerListCreateView(ListCreateAPIView):
+ """
+ get:
+ Return a list of all registered drilling organizations
+
+ post:
+ Create a new drilling organization instance
+ """
+
queryset = Organization.objects.all().select_related('province_state')
serializer_class = DrillerSerializer
def list(self, request):
queryset = self.get_queryset()
serializer = DrillerListSerializer(queryset, many=True)
return Response(serializer.data)
class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView):
+ """
+ get:
+ Return the specified drilling organization
+
+ patch:
+ Updates the specified drilling organization with the fields/values provided in the request body
+
+ delete:
+ Removes the specified drilling organization record
+ """
+
queryset = Organization.objects.all()
lookup_field = "org_guid"
serializer_class = DrillerSerializer
# Create your views here.
def index(request):
return HttpResponse("TEST: Driller Register app home index.")
| Add docstrings to view classes | ## Code Before:
from django.http import HttpResponse
from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView
from rest_framework.response import Response
from registries.models import Organization
from registries.serializers import DrillerListSerializer, DrillerSerializer
class APIDrillerListCreateView(ListCreateAPIView):
queryset = Organization.objects.all().select_related('province_state')
serializer_class = DrillerSerializer
def list(self, request):
queryset = self.get_queryset()
serializer = DrillerListSerializer(queryset, many=True)
return Response(serializer.data)
class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView):
queryset = Organization.objects.all()
lookup_field = "org_guid"
serializer_class = DrillerSerializer
# Create your views here.
def index(request):
return HttpResponse("TEST: Driller Register app home index.")
## Instruction:
Add docstrings to view classes
## Code After:
from django.http import HttpResponse
from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView
from rest_framework.response import Response
from registries.models import Organization
from registries.serializers import DrillerListSerializer, DrillerSerializer
class APIDrillerListCreateView(ListCreateAPIView):
"""
get:
Return a list of all registered drilling organizations
post:
Create a new drilling organization instance
"""
queryset = Organization.objects.all().select_related('province_state')
serializer_class = DrillerSerializer
def list(self, request):
queryset = self.get_queryset()
serializer = DrillerListSerializer(queryset, many=True)
return Response(serializer.data)
class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView):
"""
get:
Return the specified drilling organization
patch:
Updates the specified drilling organization with the fields/values provided in the request body
delete:
Removes the specified drilling organization record
"""
queryset = Organization.objects.all()
lookup_field = "org_guid"
serializer_class = DrillerSerializer
# Create your views here.
def index(request):
return HttpResponse("TEST: Driller Register app home index.")
| from django.http import HttpResponse
from rest_framework.generics import ListCreateAPIView, RetrieveUpdateDestroyAPIView
from rest_framework.response import Response
from registries.models import Organization
from registries.serializers import DrillerListSerializer, DrillerSerializer
class APIDrillerListCreateView(ListCreateAPIView):
+ """
+ get:
+ Return a list of all registered drilling organizations
+
+ post:
+ Create a new drilling organization instance
+ """
+
queryset = Organization.objects.all().select_related('province_state')
serializer_class = DrillerSerializer
def list(self, request):
queryset = self.get_queryset()
serializer = DrillerListSerializer(queryset, many=True)
return Response(serializer.data)
class APIDrillerRetrieveUpdateDestroyView(RetrieveUpdateDestroyAPIView):
+ """
+ get:
+ Return the specified drilling organization
+
+ patch:
+ Updates the specified drilling organization with the fields/values provided in the request body
+
+ delete:
+ Removes the specified drilling organization record
+ """
+
queryset = Organization.objects.all()
lookup_field = "org_guid"
serializer_class = DrillerSerializer
# Create your views here.
def index(request):
return HttpResponse("TEST: Driller Register app home index.") |
1a6344ea1fac51a8024e1803a0391662d4ab81e0 | pyeda/boolalg/vexpr.py | pyeda/boolalg/vexpr.py |
from pyeda.boolalg import expr
from pyeda.boolalg import bfarray
def bitvec(name, *dims):
"""Return a new array of given dimensions, filled with Expressions.
Parameters
----------
name : str
dims : (int or (int, int))
An int N means a slice from [0:N]
A tuple (M, N) means a slice from [M:N]
"""
if dims:
return bfarray.exprvars(name, *dims)
else:
return expr.exprvar(name)
|
from warnings import warn
from pyeda.boolalg import expr
from pyeda.boolalg import bfarray
def bitvec(name, *dims):
"""Return a new array of given dimensions, filled with Expressions.
Parameters
----------
name : str
dims : (int or (int, int))
An int N means a slice from [0:N]
A tuple (M, N) means a slice from [M:N]
"""
warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.")
if dims:
return bfarray.exprvars(name, *dims)
else:
return expr.exprvar(name)
| Add deprecation warning to bitvec function | Add deprecation warning to bitvec function
| Python | bsd-2-clause | pombredanne/pyeda,GtTmy/pyeda,karissa/pyeda,sschnug/pyeda,sschnug/pyeda,cjdrake/pyeda,sschnug/pyeda,cjdrake/pyeda,GtTmy/pyeda,GtTmy/pyeda,karissa/pyeda,pombredanne/pyeda,karissa/pyeda,cjdrake/pyeda,pombredanne/pyeda | +
+ from warnings import warn
from pyeda.boolalg import expr
from pyeda.boolalg import bfarray
def bitvec(name, *dims):
"""Return a new array of given dimensions, filled with Expressions.
Parameters
----------
name : str
dims : (int or (int, int))
An int N means a slice from [0:N]
A tuple (M, N) means a slice from [M:N]
"""
+ warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.")
if dims:
return bfarray.exprvars(name, *dims)
else:
return expr.exprvar(name)
| Add deprecation warning to bitvec function | ## Code Before:
from pyeda.boolalg import expr
from pyeda.boolalg import bfarray
def bitvec(name, *dims):
"""Return a new array of given dimensions, filled with Expressions.
Parameters
----------
name : str
dims : (int or (int, int))
An int N means a slice from [0:N]
A tuple (M, N) means a slice from [M:N]
"""
if dims:
return bfarray.exprvars(name, *dims)
else:
return expr.exprvar(name)
## Instruction:
Add deprecation warning to bitvec function
## Code After:
from warnings import warn
from pyeda.boolalg import expr
from pyeda.boolalg import bfarray
def bitvec(name, *dims):
"""Return a new array of given dimensions, filled with Expressions.
Parameters
----------
name : str
dims : (int or (int, int))
An int N means a slice from [0:N]
A tuple (M, N) means a slice from [M:N]
"""
warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.")
if dims:
return bfarray.exprvars(name, *dims)
else:
return expr.exprvar(name)
| +
+ from warnings import warn
from pyeda.boolalg import expr
from pyeda.boolalg import bfarray
def bitvec(name, *dims):
"""Return a new array of given dimensions, filled with Expressions.
Parameters
----------
name : str
dims : (int or (int, int))
An int N means a slice from [0:N]
A tuple (M, N) means a slice from [M:N]
"""
+ warn("The 'bitvec' function is deprecated. Use 'exprvars' instead.")
if dims:
return bfarray.exprvars(name, *dims)
else:
return expr.exprvar(name)
|
29c3d87881ce9c57478eb821da60d77e9f5eeb48 | eventsourcing/application/base.py | eventsourcing/application/base.py | from abc import abstractmethod, ABCMeta
from six import with_metaclass
from eventsourcing.infrastructure.event_store import EventStore
from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber
class EventSourcingApplication(with_metaclass(ABCMeta)):
def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False):
self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls,
json_decoder_cls=json_decoder_cls,
cipher=cipher,
always_encrypt=always_encrypt_stored_events)
self.event_store = self.create_event_store()
self.persistence_subscriber = self.create_persistence_subscriber()
@abstractmethod
def create_stored_event_repo(self, **kwargs):
"""Returns an instance of a subclass of StoredEventRepository.
:rtype: StoredEventRepository
"""
def create_event_store(self):
return EventStore(self.stored_event_repo)
def create_persistence_subscriber(self):
return PersistenceSubscriber(self.event_store)
def close(self):
self.persistence_subscriber.close()
self.stored_event_repo = None
self.event_store = None
self.persistence_subscriber = None
def __enter__(self):
return self
def __exit__(self, *_):
self.close()
| from abc import abstractmethod, ABCMeta
from six import with_metaclass
from eventsourcing.infrastructure.event_store import EventStore
from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber
class EventSourcingApplication(with_metaclass(ABCMeta)):
persist_events = True
def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False):
self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls,
json_decoder_cls=json_decoder_cls,
cipher=cipher,
always_encrypt=always_encrypt_stored_events)
self.event_store = self.create_event_store()
if self.persist_events:
self.persistence_subscriber = self.create_persistence_subscriber()
else:
self.persistence_subscriber = None
@abstractmethod
def create_stored_event_repo(self, **kwargs):
"""Returns an instance of a subclass of StoredEventRepository.
:rtype: StoredEventRepository
"""
def create_event_store(self):
return EventStore(self.stored_event_repo)
def create_persistence_subscriber(self):
return PersistenceSubscriber(self.event_store)
def close(self):
if self.persistence_subscriber:
self.persistence_subscriber.close()
self.stored_event_repo = None
self.event_store = None
self.persistence_subscriber = None
def __enter__(self):
return self
def __exit__(self, *_):
self.close()
| Allow to disable events persistence at app class | Allow to disable events persistence at app class | Python | bsd-3-clause | johnbywater/eventsourcing,johnbywater/eventsourcing | from abc import abstractmethod, ABCMeta
from six import with_metaclass
from eventsourcing.infrastructure.event_store import EventStore
from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber
class EventSourcingApplication(with_metaclass(ABCMeta)):
+ persist_events = True
def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False):
self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls,
json_decoder_cls=json_decoder_cls,
cipher=cipher,
always_encrypt=always_encrypt_stored_events)
self.event_store = self.create_event_store()
+ if self.persist_events:
- self.persistence_subscriber = self.create_persistence_subscriber()
+ self.persistence_subscriber = self.create_persistence_subscriber()
+ else:
+ self.persistence_subscriber = None
@abstractmethod
def create_stored_event_repo(self, **kwargs):
"""Returns an instance of a subclass of StoredEventRepository.
:rtype: StoredEventRepository
"""
def create_event_store(self):
return EventStore(self.stored_event_repo)
def create_persistence_subscriber(self):
return PersistenceSubscriber(self.event_store)
def close(self):
+ if self.persistence_subscriber:
- self.persistence_subscriber.close()
+ self.persistence_subscriber.close()
self.stored_event_repo = None
self.event_store = None
self.persistence_subscriber = None
def __enter__(self):
return self
def __exit__(self, *_):
self.close()
| Allow to disable events persistence at app class | ## Code Before:
from abc import abstractmethod, ABCMeta
from six import with_metaclass
from eventsourcing.infrastructure.event_store import EventStore
from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber
class EventSourcingApplication(with_metaclass(ABCMeta)):
def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False):
self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls,
json_decoder_cls=json_decoder_cls,
cipher=cipher,
always_encrypt=always_encrypt_stored_events)
self.event_store = self.create_event_store()
self.persistence_subscriber = self.create_persistence_subscriber()
@abstractmethod
def create_stored_event_repo(self, **kwargs):
"""Returns an instance of a subclass of StoredEventRepository.
:rtype: StoredEventRepository
"""
def create_event_store(self):
return EventStore(self.stored_event_repo)
def create_persistence_subscriber(self):
return PersistenceSubscriber(self.event_store)
def close(self):
self.persistence_subscriber.close()
self.stored_event_repo = None
self.event_store = None
self.persistence_subscriber = None
def __enter__(self):
return self
def __exit__(self, *_):
self.close()
## Instruction:
Allow to disable events persistence at app class
## Code After:
from abc import abstractmethod, ABCMeta
from six import with_metaclass
from eventsourcing.infrastructure.event_store import EventStore
from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber
class EventSourcingApplication(with_metaclass(ABCMeta)):
persist_events = True
def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False):
self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls,
json_decoder_cls=json_decoder_cls,
cipher=cipher,
always_encrypt=always_encrypt_stored_events)
self.event_store = self.create_event_store()
if self.persist_events:
self.persistence_subscriber = self.create_persistence_subscriber()
else:
self.persistence_subscriber = None
@abstractmethod
def create_stored_event_repo(self, **kwargs):
"""Returns an instance of a subclass of StoredEventRepository.
:rtype: StoredEventRepository
"""
def create_event_store(self):
return EventStore(self.stored_event_repo)
def create_persistence_subscriber(self):
return PersistenceSubscriber(self.event_store)
def close(self):
if self.persistence_subscriber:
self.persistence_subscriber.close()
self.stored_event_repo = None
self.event_store = None
self.persistence_subscriber = None
def __enter__(self):
return self
def __exit__(self, *_):
self.close()
| from abc import abstractmethod, ABCMeta
from six import with_metaclass
from eventsourcing.infrastructure.event_store import EventStore
from eventsourcing.infrastructure.persistence_subscriber import PersistenceSubscriber
class EventSourcingApplication(with_metaclass(ABCMeta)):
+ persist_events = True
def __init__(self, json_encoder_cls=None, json_decoder_cls=None, cipher=None, always_encrypt_stored_events=False):
self.stored_event_repo = self.create_stored_event_repo(json_encoder_cls=json_encoder_cls,
json_decoder_cls=json_decoder_cls,
cipher=cipher,
always_encrypt=always_encrypt_stored_events)
self.event_store = self.create_event_store()
+ if self.persist_events:
- self.persistence_subscriber = self.create_persistence_subscriber()
+ self.persistence_subscriber = self.create_persistence_subscriber()
? ++++
+ else:
+ self.persistence_subscriber = None
@abstractmethod
def create_stored_event_repo(self, **kwargs):
"""Returns an instance of a subclass of StoredEventRepository.
:rtype: StoredEventRepository
"""
def create_event_store(self):
return EventStore(self.stored_event_repo)
def create_persistence_subscriber(self):
return PersistenceSubscriber(self.event_store)
def close(self):
+ if self.persistence_subscriber:
- self.persistence_subscriber.close()
+ self.persistence_subscriber.close()
? ++++
self.stored_event_repo = None
self.event_store = None
self.persistence_subscriber = None
def __enter__(self):
return self
def __exit__(self, *_):
self.close() |
e8537feff53310913047d06d95f4dd8e9dace1da | flow_workflow/historian/handler.py | flow_workflow/historian/handler.py | from flow import exit_codes
from flow.configuration.settings.injector import setting
from flow.handler import Handler
from flow.util.exit import exit_process
from flow_workflow.historian.messages import UpdateMessage
from injector import inject
from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError
from twisted.internet import defer
import flow.interfaces
import logging
import os
LOG = logging.getLogger(__name__)
@inject(storage=flow.interfaces.IStorage,
queue_name=setting('workflow.historian.queue'))
class WorkflowHistorianMessageHandler(Handler):
message_class = UpdateMessage
def _handle_message(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
return defer.succeed(None)
except (ResourceClosedError, TimeoutError, DisconnectionError):
LOG.exception("This historian cannot handle messages anymore, "
"because it lost access to Oracle... exiting.")
exit_process(exit_codes.EXECUTE_FAILURE)
| from flow import exit_codes
from flow.configuration.settings.injector import setting
from flow.handler import Handler
from flow.util.exit import exit_process
from flow_workflow.historian.messages import UpdateMessage
from injector import inject
from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError
from twisted.internet import defer
import flow.interfaces
import logging
import os
LOG = logging.getLogger(__name__)
@inject(storage=flow.interfaces.IStorage,
queue_name=setting('workflow.historian.queue'))
class WorkflowHistorianMessageHandler(Handler):
message_class = UpdateMessage
def _handle_message(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
return defer.succeed(None)
except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError):
LOG.exception("This historian cannot handle messages anymore, "
"because it lost access to Oracle... exiting.")
exit_process(exit_codes.EXECUTE_FAILURE)
| Add DatabaseError to list of errors that kill a historian | Add DatabaseError to list of errors that kill a historian
| Python | agpl-3.0 | genome/flow-workflow,genome/flow-workflow,genome/flow-workflow | from flow import exit_codes
from flow.configuration.settings.injector import setting
from flow.handler import Handler
from flow.util.exit import exit_process
from flow_workflow.historian.messages import UpdateMessage
from injector import inject
- from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError
+ from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError
from twisted.internet import defer
import flow.interfaces
import logging
import os
LOG = logging.getLogger(__name__)
@inject(storage=flow.interfaces.IStorage,
queue_name=setting('workflow.historian.queue'))
class WorkflowHistorianMessageHandler(Handler):
message_class = UpdateMessage
def _handle_message(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
return defer.succeed(None)
- except (ResourceClosedError, TimeoutError, DisconnectionError):
+ except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError):
LOG.exception("This historian cannot handle messages anymore, "
"because it lost access to Oracle... exiting.")
exit_process(exit_codes.EXECUTE_FAILURE)
| Add DatabaseError to list of errors that kill a historian | ## Code Before:
from flow import exit_codes
from flow.configuration.settings.injector import setting
from flow.handler import Handler
from flow.util.exit import exit_process
from flow_workflow.historian.messages import UpdateMessage
from injector import inject
from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError
from twisted.internet import defer
import flow.interfaces
import logging
import os
LOG = logging.getLogger(__name__)
@inject(storage=flow.interfaces.IStorage,
queue_name=setting('workflow.historian.queue'))
class WorkflowHistorianMessageHandler(Handler):
message_class = UpdateMessage
def _handle_message(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
return defer.succeed(None)
except (ResourceClosedError, TimeoutError, DisconnectionError):
LOG.exception("This historian cannot handle messages anymore, "
"because it lost access to Oracle... exiting.")
exit_process(exit_codes.EXECUTE_FAILURE)
## Instruction:
Add DatabaseError to list of errors that kill a historian
## Code After:
from flow import exit_codes
from flow.configuration.settings.injector import setting
from flow.handler import Handler
from flow.util.exit import exit_process
from flow_workflow.historian.messages import UpdateMessage
from injector import inject
from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError
from twisted.internet import defer
import flow.interfaces
import logging
import os
LOG = logging.getLogger(__name__)
@inject(storage=flow.interfaces.IStorage,
queue_name=setting('workflow.historian.queue'))
class WorkflowHistorianMessageHandler(Handler):
message_class = UpdateMessage
def _handle_message(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
return defer.succeed(None)
except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError):
LOG.exception("This historian cannot handle messages anymore, "
"because it lost access to Oracle... exiting.")
exit_process(exit_codes.EXECUTE_FAILURE)
| from flow import exit_codes
from flow.configuration.settings.injector import setting
from flow.handler import Handler
from flow.util.exit import exit_process
from flow_workflow.historian.messages import UpdateMessage
from injector import inject
- from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError
+ from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError
? +++++++++++++++
from twisted.internet import defer
import flow.interfaces
import logging
import os
LOG = logging.getLogger(__name__)
@inject(storage=flow.interfaces.IStorage,
queue_name=setting('workflow.historian.queue'))
class WorkflowHistorianMessageHandler(Handler):
message_class = UpdateMessage
def _handle_message(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
return defer.succeed(None)
- except (ResourceClosedError, TimeoutError, DisconnectionError):
+ except (ResourceClosedError, TimeoutError, DisconnectionError, DatabaseError):
? +++++++++++++++
LOG.exception("This historian cannot handle messages anymore, "
"because it lost access to Oracle... exiting.")
exit_process(exit_codes.EXECUTE_FAILURE)
|
d4da069b43174482f3a75e9553e8283be905fa16 | cla_public/apps/base/filters.py | cla_public/apps/base/filters.py | "Jinja custom filters"
import re
from cla_public.apps.base import base
from babel.dates import format_datetime
@base.app_template_filter()
def datetime(dt, format='medium', locale='en_GB'):
if format == 'full':
format = "EEEE, d MMMM y 'at' HH:mm"
elif format == 'medium':
format = "EE, dd/MM/y 'at' h:mma"
elif format == 'short':
format = "dd/MM/y, h:mma"
return format_datetime(dt, format, locale=locale)
@base.app_template_filter()
def url_to_human(value):
return re.sub(r'(^https?://)|(/$)', '', value)
@base.app_template_filter()
def human_to_url(value):
return re.sub(r'^((?!https?://).*)', r'http://\1', value)
@base.app_template_filter()
def query_to_dict(value, prop=None):
result = parse_qs(urlparse(value).query)
if prop:
result = result[prop]
return result
| "Jinja custom filters"
import re
from urlparse import urlparse, parse_qs
from cla_public.apps.base import base
from babel.dates import format_datetime
@base.app_template_filter()
def datetime(dt, format='medium', locale='en_GB'):
if format == 'full':
format = "EEEE, d MMMM y 'at' HH:mm"
elif format == 'medium':
format = "EE, dd/MM/y 'at' h:mma"
elif format == 'short':
format = "dd/MM/y, h:mma"
return format_datetime(dt, format, locale=locale)
@base.app_template_filter()
def url_to_human(value):
return re.sub(r'(^https?://)|(/$)', '', value)
@base.app_template_filter()
def human_to_url(value):
return re.sub(r'^((?!https?://).*)', r'http://\1', value)
@base.app_template_filter()
def query_to_dict(value, prop=None):
result = parse_qs(urlparse(value).query)
if prop:
result = result[prop]
return result
| Add Jinja filter to convert URL params to dict | BE: Add Jinja filter to convert URL params to dict
| Python | mit | ministryofjustice/cla_public,ministryofjustice/cla_public,ministryofjustice/cla_public,ministryofjustice/cla_public | "Jinja custom filters"
import re
+ from urlparse import urlparse, parse_qs
from cla_public.apps.base import base
from babel.dates import format_datetime
@base.app_template_filter()
def datetime(dt, format='medium', locale='en_GB'):
if format == 'full':
format = "EEEE, d MMMM y 'at' HH:mm"
elif format == 'medium':
format = "EE, dd/MM/y 'at' h:mma"
elif format == 'short':
format = "dd/MM/y, h:mma"
return format_datetime(dt, format, locale=locale)
@base.app_template_filter()
def url_to_human(value):
return re.sub(r'(^https?://)|(/$)', '', value)
@base.app_template_filter()
def human_to_url(value):
return re.sub(r'^((?!https?://).*)', r'http://\1', value)
@base.app_template_filter()
def query_to_dict(value, prop=None):
result = parse_qs(urlparse(value).query)
if prop:
result = result[prop]
return result
| Add Jinja filter to convert URL params to dict | ## Code Before:
"Jinja custom filters"
import re
from cla_public.apps.base import base
from babel.dates import format_datetime
@base.app_template_filter()
def datetime(dt, format='medium', locale='en_GB'):
if format == 'full':
format = "EEEE, d MMMM y 'at' HH:mm"
elif format == 'medium':
format = "EE, dd/MM/y 'at' h:mma"
elif format == 'short':
format = "dd/MM/y, h:mma"
return format_datetime(dt, format, locale=locale)
@base.app_template_filter()
def url_to_human(value):
return re.sub(r'(^https?://)|(/$)', '', value)
@base.app_template_filter()
def human_to_url(value):
return re.sub(r'^((?!https?://).*)', r'http://\1', value)
@base.app_template_filter()
def query_to_dict(value, prop=None):
result = parse_qs(urlparse(value).query)
if prop:
result = result[prop]
return result
## Instruction:
Add Jinja filter to convert URL params to dict
## Code After:
"Jinja custom filters"
import re
from urlparse import urlparse, parse_qs
from cla_public.apps.base import base
from babel.dates import format_datetime
@base.app_template_filter()
def datetime(dt, format='medium', locale='en_GB'):
if format == 'full':
format = "EEEE, d MMMM y 'at' HH:mm"
elif format == 'medium':
format = "EE, dd/MM/y 'at' h:mma"
elif format == 'short':
format = "dd/MM/y, h:mma"
return format_datetime(dt, format, locale=locale)
@base.app_template_filter()
def url_to_human(value):
return re.sub(r'(^https?://)|(/$)', '', value)
@base.app_template_filter()
def human_to_url(value):
return re.sub(r'^((?!https?://).*)', r'http://\1', value)
@base.app_template_filter()
def query_to_dict(value, prop=None):
result = parse_qs(urlparse(value).query)
if prop:
result = result[prop]
return result
| "Jinja custom filters"
import re
+ from urlparse import urlparse, parse_qs
from cla_public.apps.base import base
from babel.dates import format_datetime
@base.app_template_filter()
def datetime(dt, format='medium', locale='en_GB'):
if format == 'full':
format = "EEEE, d MMMM y 'at' HH:mm"
elif format == 'medium':
format = "EE, dd/MM/y 'at' h:mma"
elif format == 'short':
format = "dd/MM/y, h:mma"
return format_datetime(dt, format, locale=locale)
@base.app_template_filter()
def url_to_human(value):
return re.sub(r'(^https?://)|(/$)', '', value)
@base.app_template_filter()
def human_to_url(value):
return re.sub(r'^((?!https?://).*)', r'http://\1', value)
@base.app_template_filter()
def query_to_dict(value, prop=None):
result = parse_qs(urlparse(value).query)
if prop:
result = result[prop]
return result |
20f0d90f5c64322864ad5fda4b4c9314e6c1cb11 | run.py | run.py |
import sys
from kitchen.text.converters import getwriter
from utils.log import getLogger, open_log, close_log
from utils.misc import output_exception
from system.factory_manager import Manager
sys.stdout = getwriter('utf-8')(sys.stdout)
sys.stderr = getwriter('utf-8')(sys.stderr)
open_log("output.log")
logger = getLogger("System")
logger.info("Starting up..")
try:
manager = Manager()
except Exception:
logger.critical("Runtime error - process cannot continue!")
output_exception(logger)
finally:
close_log("output.log")
try:
raw_input("Press enter to exit.")
except:
pass
|
import os
import sys
from kitchen.text.converters import getwriter
from utils.log import getLogger, open_log, close_log
from utils.misc import output_exception
from system.factory_manager import Manager
sys.stdout = getwriter('utf-8')(sys.stdout)
sys.stderr = getwriter('utf-8')(sys.stderr)
if not os.path.exists("logs"):
os.mkdir("logs")
open_log("output.log")
logger = getLogger("System")
logger.info("Starting up..")
try:
manager = Manager()
except Exception:
logger.critical("Runtime error - process cannot continue!")
output_exception(logger)
finally:
close_log("output.log")
try:
raw_input("Press enter to exit.")
except:
pass
| Create logs folder if it doesn't exist (to prevent errors) | Create logs folder if it doesn't exist (to prevent errors)
| Python | artistic-2.0 | UltrosBot/Ultros,UltrosBot/Ultros |
+ import os
import sys
from kitchen.text.converters import getwriter
from utils.log import getLogger, open_log, close_log
from utils.misc import output_exception
from system.factory_manager import Manager
sys.stdout = getwriter('utf-8')(sys.stdout)
sys.stderr = getwriter('utf-8')(sys.stderr)
+
+ if not os.path.exists("logs"):
+ os.mkdir("logs")
open_log("output.log")
logger = getLogger("System")
logger.info("Starting up..")
try:
manager = Manager()
except Exception:
logger.critical("Runtime error - process cannot continue!")
output_exception(logger)
finally:
close_log("output.log")
try:
raw_input("Press enter to exit.")
except:
pass
| Create logs folder if it doesn't exist (to prevent errors) | ## Code Before:
import sys
from kitchen.text.converters import getwriter
from utils.log import getLogger, open_log, close_log
from utils.misc import output_exception
from system.factory_manager import Manager
sys.stdout = getwriter('utf-8')(sys.stdout)
sys.stderr = getwriter('utf-8')(sys.stderr)
open_log("output.log")
logger = getLogger("System")
logger.info("Starting up..")
try:
manager = Manager()
except Exception:
logger.critical("Runtime error - process cannot continue!")
output_exception(logger)
finally:
close_log("output.log")
try:
raw_input("Press enter to exit.")
except:
pass
## Instruction:
Create logs folder if it doesn't exist (to prevent errors)
## Code After:
import os
import sys
from kitchen.text.converters import getwriter
from utils.log import getLogger, open_log, close_log
from utils.misc import output_exception
from system.factory_manager import Manager
sys.stdout = getwriter('utf-8')(sys.stdout)
sys.stderr = getwriter('utf-8')(sys.stderr)
if not os.path.exists("logs"):
os.mkdir("logs")
open_log("output.log")
logger = getLogger("System")
logger.info("Starting up..")
try:
manager = Manager()
except Exception:
logger.critical("Runtime error - process cannot continue!")
output_exception(logger)
finally:
close_log("output.log")
try:
raw_input("Press enter to exit.")
except:
pass
|
+ import os
import sys
from kitchen.text.converters import getwriter
from utils.log import getLogger, open_log, close_log
from utils.misc import output_exception
from system.factory_manager import Manager
sys.stdout = getwriter('utf-8')(sys.stdout)
sys.stderr = getwriter('utf-8')(sys.stderr)
+
+ if not os.path.exists("logs"):
+ os.mkdir("logs")
open_log("output.log")
logger = getLogger("System")
logger.info("Starting up..")
try:
manager = Manager()
except Exception:
logger.critical("Runtime error - process cannot continue!")
output_exception(logger)
finally:
close_log("output.log")
try:
raw_input("Press enter to exit.")
except:
pass |
0cecbabd2d594bfc2ca57e522658d13eda2bc6a8 | pipdiff/pipdiff.py | pipdiff/pipdiff.py |
from pkg_resources import parse_version
from xmlrpclib import ServerProxy
pypi = ServerProxy("http://pypi.python.org/pypi")
def main():
try:
from pip import get_installed_distributions
except ImportError:
from sys import exit
exit("pip not available")
for distribution in sorted(get_installed_distributions(),
key=lambda distribution: distribution.project_name):
remote = ''
project_name = distribution.project_name
releases = pypi.package_releases(project_name)
if not releases:
pypi.package_releases(project_name.capitalize())
if releases:
version = parse_version(releases[0])
if version > distribution.parsed_version:
remote = "PyPI:{0}=={1}".format(project_name, releases[0])
else:
remote = "PyPI:{0} not found".format(project_name)
local = "{0}=={1}".format(project_name, distribution.version)
print "{0:40} {1}".format(local, remote)
return True
if __name__ == '__main__':
main()
|
from pkg_resources import parse_version
try:
from xmlrpclib import ServerProxy
except ImportError:
import xmlrpc.client
try:
pypi = ServerProxy("http://pypi.python.org/pypi")
except NameError:
pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi")
def main():
try:
from pip import get_installed_distributions
except ImportError:
from sys import exit
exit("pip not available")
for distribution in sorted(get_installed_distributions(),
key=lambda distribution: distribution.project_name):
remote = ''
project_name = distribution.project_name
releases = pypi.package_releases(project_name)
if not releases:
pypi.package_releases(project_name.capitalize())
if releases:
version = parse_version(releases[0])
if str(version) > str(distribution.parsed_version):
remote = "PyPI:{0}=={1}".format(project_name, releases[0])
else:
remote = "PyPI:{0} not found".format(project_name)
local = "{0}=={1}".format(project_name, distribution.version)
print("{0:40} {1}".format(local, remote))
return True
if __name__ == '__main__':
main()
| Add support for Python 3 | Add support for Python 3
| Python | bsd-3-clause | ogt/pipdiff |
from pkg_resources import parse_version
+ try:
- from xmlrpclib import ServerProxy
+ from xmlrpclib import ServerProxy
+ except ImportError:
+ import xmlrpc.client
+ try:
- pypi = ServerProxy("http://pypi.python.org/pypi")
+ pypi = ServerProxy("http://pypi.python.org/pypi")
-
+ except NameError:
+ pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi")
def main():
try:
from pip import get_installed_distributions
except ImportError:
from sys import exit
exit("pip not available")
for distribution in sorted(get_installed_distributions(),
key=lambda distribution: distribution.project_name):
remote = ''
project_name = distribution.project_name
releases = pypi.package_releases(project_name)
if not releases:
pypi.package_releases(project_name.capitalize())
if releases:
version = parse_version(releases[0])
-
- if version > distribution.parsed_version:
+ if str(version) > str(distribution.parsed_version):
remote = "PyPI:{0}=={1}".format(project_name, releases[0])
else:
remote = "PyPI:{0} not found".format(project_name)
local = "{0}=={1}".format(project_name, distribution.version)
- print "{0:40} {1}".format(local, remote)
+ print("{0:40} {1}".format(local, remote))
return True
if __name__ == '__main__':
main()
| Add support for Python 3 | ## Code Before:
from pkg_resources import parse_version
from xmlrpclib import ServerProxy
pypi = ServerProxy("http://pypi.python.org/pypi")
def main():
try:
from pip import get_installed_distributions
except ImportError:
from sys import exit
exit("pip not available")
for distribution in sorted(get_installed_distributions(),
key=lambda distribution: distribution.project_name):
remote = ''
project_name = distribution.project_name
releases = pypi.package_releases(project_name)
if not releases:
pypi.package_releases(project_name.capitalize())
if releases:
version = parse_version(releases[0])
if version > distribution.parsed_version:
remote = "PyPI:{0}=={1}".format(project_name, releases[0])
else:
remote = "PyPI:{0} not found".format(project_name)
local = "{0}=={1}".format(project_name, distribution.version)
print "{0:40} {1}".format(local, remote)
return True
if __name__ == '__main__':
main()
## Instruction:
Add support for Python 3
## Code After:
from pkg_resources import parse_version
try:
from xmlrpclib import ServerProxy
except ImportError:
import xmlrpc.client
try:
pypi = ServerProxy("http://pypi.python.org/pypi")
except NameError:
pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi")
def main():
try:
from pip import get_installed_distributions
except ImportError:
from sys import exit
exit("pip not available")
for distribution in sorted(get_installed_distributions(),
key=lambda distribution: distribution.project_name):
remote = ''
project_name = distribution.project_name
releases = pypi.package_releases(project_name)
if not releases:
pypi.package_releases(project_name.capitalize())
if releases:
version = parse_version(releases[0])
if str(version) > str(distribution.parsed_version):
remote = "PyPI:{0}=={1}".format(project_name, releases[0])
else:
remote = "PyPI:{0} not found".format(project_name)
local = "{0}=={1}".format(project_name, distribution.version)
print("{0:40} {1}".format(local, remote))
return True
if __name__ == '__main__':
main()
|
from pkg_resources import parse_version
+ try:
- from xmlrpclib import ServerProxy
+ from xmlrpclib import ServerProxy
? ++++
+ except ImportError:
+ import xmlrpc.client
+ try:
- pypi = ServerProxy("http://pypi.python.org/pypi")
+ pypi = ServerProxy("http://pypi.python.org/pypi")
? ++++
-
+ except NameError:
+ pypi = xmlrpc.client.ServerProxy("http://pypi.python.org/pypi")
def main():
try:
from pip import get_installed_distributions
except ImportError:
from sys import exit
exit("pip not available")
for distribution in sorted(get_installed_distributions(),
key=lambda distribution: distribution.project_name):
remote = ''
project_name = distribution.project_name
releases = pypi.package_releases(project_name)
if not releases:
pypi.package_releases(project_name.capitalize())
if releases:
version = parse_version(releases[0])
-
- if version > distribution.parsed_version:
+ if str(version) > str(distribution.parsed_version):
? ++++ + ++++ +
remote = "PyPI:{0}=={1}".format(project_name, releases[0])
else:
remote = "PyPI:{0} not found".format(project_name)
local = "{0}=={1}".format(project_name, distribution.version)
- print "{0:40} {1}".format(local, remote)
? ^
+ print("{0:40} {1}".format(local, remote))
? ^ +
return True
if __name__ == '__main__':
main() |
93081d423a73a6b16e5adfb94247ffec23ef667c | api/base/authentication/backends.py | api/base/authentication/backends.py | from osf.models.user import OSFUser
from framework.auth.core import get_user
from django.contrib.auth.backends import ModelBackend
# https://docs.djangoproject.com/en/1.8/topics/auth/customizing/
class ODMBackend(ModelBackend):
def authenticate(self, username=None, password=None):
return get_user(email=username, password=password) or None
def get_user(self, user_id):
try:
user = OSFUser.objects.get(id=user_id)
except OSFUser.DoesNotExist:
user = OSFUser.load(user_id)
return user
| from osf.models.user import OSFUser
from framework.auth.core import get_user
from django.contrib.auth.backends import ModelBackend
# https://docs.djangoproject.com/en/3.2/topics/auth/customizing/
class ODMBackend(ModelBackend):
def authenticate(self, request, username=None, password=None, **kwargs):
return get_user(email=username, password=password) or None
def get_user(self, user_id):
try:
user = OSFUser.objects.get(id=user_id)
except OSFUser.DoesNotExist:
user = OSFUser.load(user_id)
return user
| Fix admin login failure for django upgrade | Fix admin login failure for django upgrade
| Python | apache-2.0 | Johnetordoff/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io | from osf.models.user import OSFUser
from framework.auth.core import get_user
from django.contrib.auth.backends import ModelBackend
+
- # https://docs.djangoproject.com/en/1.8/topics/auth/customizing/
+ # https://docs.djangoproject.com/en/3.2/topics/auth/customizing/
class ODMBackend(ModelBackend):
- def authenticate(self, username=None, password=None):
+ def authenticate(self, request, username=None, password=None, **kwargs):
return get_user(email=username, password=password) or None
def get_user(self, user_id):
try:
user = OSFUser.objects.get(id=user_id)
except OSFUser.DoesNotExist:
user = OSFUser.load(user_id)
return user
| Fix admin login failure for django upgrade | ## Code Before:
from osf.models.user import OSFUser
from framework.auth.core import get_user
from django.contrib.auth.backends import ModelBackend
# https://docs.djangoproject.com/en/1.8/topics/auth/customizing/
class ODMBackend(ModelBackend):
def authenticate(self, username=None, password=None):
return get_user(email=username, password=password) or None
def get_user(self, user_id):
try:
user = OSFUser.objects.get(id=user_id)
except OSFUser.DoesNotExist:
user = OSFUser.load(user_id)
return user
## Instruction:
Fix admin login failure for django upgrade
## Code After:
from osf.models.user import OSFUser
from framework.auth.core import get_user
from django.contrib.auth.backends import ModelBackend
# https://docs.djangoproject.com/en/3.2/topics/auth/customizing/
class ODMBackend(ModelBackend):
def authenticate(self, request, username=None, password=None, **kwargs):
return get_user(email=username, password=password) or None
def get_user(self, user_id):
try:
user = OSFUser.objects.get(id=user_id)
except OSFUser.DoesNotExist:
user = OSFUser.load(user_id)
return user
| from osf.models.user import OSFUser
from framework.auth.core import get_user
from django.contrib.auth.backends import ModelBackend
+
- # https://docs.djangoproject.com/en/1.8/topics/auth/customizing/
? ^ ^
+ # https://docs.djangoproject.com/en/3.2/topics/auth/customizing/
? ^ ^
class ODMBackend(ModelBackend):
- def authenticate(self, username=None, password=None):
+ def authenticate(self, request, username=None, password=None, **kwargs):
? +++++++++ ++++++++++
return get_user(email=username, password=password) or None
def get_user(self, user_id):
try:
user = OSFUser.objects.get(id=user_id)
except OSFUser.DoesNotExist:
user = OSFUser.load(user_id)
return user |
8925c3a827659e1983827368948e95e764a40585 | utf9/__init__.py | utf9/__init__.py |
from bitarray import bitarray as _bitarray
def utf9encode(string):
bits = _bitarray()
for char in string:
for idx, byte in enumerate(char.encode('utf-8')):
bits.append(idx)
bits.extend('{0:b}'.format(ord(byte)).zfill(8))
return bits.tobytes()
def utf9decode(data):
bits = _bitarray()
bits.frombytes(data)
chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9))
string = u''
codepoint = ''
for chunk in chunks:
if len(chunk) < 9:
break
if chunk[0] == 0:
codepoint, string = '', string + codepoint.decode('utf-8')
codepoint += chr(int(chunk[1:].to01(), 2))
return string + codepoint.decode('utf-8')
|
from bitarray import bitarray as _bitarray
def utf9encode(string):
"""Takes a string and returns a utf9-encoded version."""
bits = _bitarray()
for char in string:
for idx, byte in enumerate(char.encode('utf-8')):
bits.append(idx)
bits.extend('{0:b}'.format(ord(byte)).zfill(8))
return bits.tobytes()
def utf9decode(data):
"""Takes utf9-encoded data and returns the corresponding string."""
bits = _bitarray()
bits.frombytes(data)
chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9))
string = u''
codepoint = ''
for chunk in chunks:
if len(chunk) < 9:
break
if chunk[0] == 0:
codepoint, string = '', string + codepoint.decode('utf-8')
codepoint += chr(int(chunk[1:].to01(), 2))
return string + codepoint.decode('utf-8')
| Add module and functions docstring | Add module and functions docstring
| Python | mit | enricobacis/utf9 |
from bitarray import bitarray as _bitarray
def utf9encode(string):
+ """Takes a string and returns a utf9-encoded version."""
bits = _bitarray()
for char in string:
for idx, byte in enumerate(char.encode('utf-8')):
bits.append(idx)
bits.extend('{0:b}'.format(ord(byte)).zfill(8))
return bits.tobytes()
def utf9decode(data):
+ """Takes utf9-encoded data and returns the corresponding string."""
bits = _bitarray()
bits.frombytes(data)
chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9))
string = u''
codepoint = ''
for chunk in chunks:
if len(chunk) < 9:
break
if chunk[0] == 0:
codepoint, string = '', string + codepoint.decode('utf-8')
codepoint += chr(int(chunk[1:].to01(), 2))
return string + codepoint.decode('utf-8')
| Add module and functions docstring | ## Code Before:
from bitarray import bitarray as _bitarray
def utf9encode(string):
bits = _bitarray()
for char in string:
for idx, byte in enumerate(char.encode('utf-8')):
bits.append(idx)
bits.extend('{0:b}'.format(ord(byte)).zfill(8))
return bits.tobytes()
def utf9decode(data):
bits = _bitarray()
bits.frombytes(data)
chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9))
string = u''
codepoint = ''
for chunk in chunks:
if len(chunk) < 9:
break
if chunk[0] == 0:
codepoint, string = '', string + codepoint.decode('utf-8')
codepoint += chr(int(chunk[1:].to01(), 2))
return string + codepoint.decode('utf-8')
## Instruction:
Add module and functions docstring
## Code After:
from bitarray import bitarray as _bitarray
def utf9encode(string):
"""Takes a string and returns a utf9-encoded version."""
bits = _bitarray()
for char in string:
for idx, byte in enumerate(char.encode('utf-8')):
bits.append(idx)
bits.extend('{0:b}'.format(ord(byte)).zfill(8))
return bits.tobytes()
def utf9decode(data):
"""Takes utf9-encoded data and returns the corresponding string."""
bits = _bitarray()
bits.frombytes(data)
chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9))
string = u''
codepoint = ''
for chunk in chunks:
if len(chunk) < 9:
break
if chunk[0] == 0:
codepoint, string = '', string + codepoint.decode('utf-8')
codepoint += chr(int(chunk[1:].to01(), 2))
return string + codepoint.decode('utf-8')
|
from bitarray import bitarray as _bitarray
def utf9encode(string):
+ """Takes a string and returns a utf9-encoded version."""
bits = _bitarray()
for char in string:
for idx, byte in enumerate(char.encode('utf-8')):
bits.append(idx)
bits.extend('{0:b}'.format(ord(byte)).zfill(8))
return bits.tobytes()
def utf9decode(data):
+ """Takes utf9-encoded data and returns the corresponding string."""
bits = _bitarray()
bits.frombytes(data)
chunks = (bits[x:x+9] for x in xrange(0, len(bits), 9))
string = u''
codepoint = ''
for chunk in chunks:
if len(chunk) < 9:
break
if chunk[0] == 0:
codepoint, string = '', string + codepoint.decode('utf-8')
codepoint += chr(int(chunk[1:].to01(), 2))
return string + codepoint.decode('utf-8') |
63f7489066aeb23dbefc6f8de534ad05144431ad | boardinghouse/tests/test_sql.py | boardinghouse/tests/test_sql.py |
from django.conf import settings
from django.test import TestCase
from django.db.models import connection
from boardinghouse.models import Schema
class TestRejectSchemaColumnChange(TestCase):
def test_exception_is_raised(self):
Schema.objects.mass_create('a')
cursor = connection.cursor()
UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'"
self.assertRaises(Exception, cursor.execute, UPDATE) |
from django.conf import settings
from django.test import TestCase
from django.db import connection
from boardinghouse.models import Schema
class TestRejectSchemaColumnChange(TestCase):
def test_exception_is_raised(self):
Schema.objects.mass_create('a')
cursor = connection.cursor()
UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'"
self.assertRaises(Exception, cursor.execute, UPDATE) | Make test work with 1.7 | Make test work with 1.7
| Python | bsd-3-clause | luzfcb/django-boardinghouse,luzfcb/django-boardinghouse,luzfcb/django-boardinghouse |
from django.conf import settings
from django.test import TestCase
- from django.db.models import connection
+ from django.db import connection
from boardinghouse.models import Schema
class TestRejectSchemaColumnChange(TestCase):
def test_exception_is_raised(self):
Schema.objects.mass_create('a')
cursor = connection.cursor()
UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'"
- self.assertRaises(Exception, cursor.execute, UPDATE)
+ self.assertRaises(Exception, cursor.execute, UPDATE) | Make test work with 1.7 | ## Code Before:
from django.conf import settings
from django.test import TestCase
from django.db.models import connection
from boardinghouse.models import Schema
class TestRejectSchemaColumnChange(TestCase):
def test_exception_is_raised(self):
Schema.objects.mass_create('a')
cursor = connection.cursor()
UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'"
self.assertRaises(Exception, cursor.execute, UPDATE)
## Instruction:
Make test work with 1.7
## Code After:
from django.conf import settings
from django.test import TestCase
from django.db import connection
from boardinghouse.models import Schema
class TestRejectSchemaColumnChange(TestCase):
def test_exception_is_raised(self):
Schema.objects.mass_create('a')
cursor = connection.cursor()
UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'"
self.assertRaises(Exception, cursor.execute, UPDATE) |
from django.conf import settings
from django.test import TestCase
- from django.db.models import connection
? -------
+ from django.db import connection
from boardinghouse.models import Schema
class TestRejectSchemaColumnChange(TestCase):
def test_exception_is_raised(self):
Schema.objects.mass_create('a')
cursor = connection.cursor()
UPDATE = "UPDATE boardinghouse_schema SET schema='foo' WHERE schema='a'"
- self.assertRaises(Exception, cursor.execute, UPDATE)
? --------
+ self.assertRaises(Exception, cursor.execute, UPDATE) |
c0d8b7f13a74fd4da7b36d30a61224b76367acbe | scraper.py | scraper.py | import urllib, datetime, os
def fetch():
url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx'
output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html')
output = os.path.join(os.path.dirname(__file__), output)
content = urllib.urlretrieve(url, output)
if __name__ == '__main__':
fetch() | import urllib, datetime, os
def fetch():
# Instead of doing all the parsing later, I get the status from Mosab's site & store it
url = 'http://power-grid-status.mos3abof.com/status'
output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json')
output = os.path.join(os.path.dirname(__file__), output)
content = urllib.urlretrieve(url, output)
if __name__ == '__main__':
fetch() | Use Mosab's site to get the status | Use Mosab's site to get the status
| Python | apache-2.0 | mtayseer/power-grid-scraper | import urllib, datetime, os
def fetch():
- url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx'
+ # Instead of doing all the parsing later, I get the status from Mosab's site & store it
+ url = 'http://power-grid-status.mos3abof.com/status'
- output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html')
+ output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json')
output = os.path.join(os.path.dirname(__file__), output)
content = urllib.urlretrieve(url, output)
if __name__ == '__main__':
fetch() | Use Mosab's site to get the status | ## Code Before:
import urllib, datetime, os
def fetch():
url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx'
output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html')
output = os.path.join(os.path.dirname(__file__), output)
content = urllib.urlretrieve(url, output)
if __name__ == '__main__':
fetch()
## Instruction:
Use Mosab's site to get the status
## Code After:
import urllib, datetime, os
def fetch():
# Instead of doing all the parsing later, I get the status from Mosab's site & store it
url = 'http://power-grid-status.mos3abof.com/status'
output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json')
output = os.path.join(os.path.dirname(__file__), output)
content = urllib.urlretrieve(url, output)
if __name__ == '__main__':
fetch() | import urllib, datetime, os
def fetch():
- url = 'http://loadmeter.egyptera.org/ClockToolTip.aspx'
+ # Instead of doing all the parsing later, I get the status from Mosab's site & store it
+ url = 'http://power-grid-status.mos3abof.com/status'
- output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.html')
? ^^^^
+ output = datetime.datetime.now().strftime('egyptera.%Y-%m-%d-%H-%M-%S.json')
? ^^^^
output = os.path.join(os.path.dirname(__file__), output)
content = urllib.urlretrieve(url, output)
if __name__ == '__main__':
fetch() |
82eb7a69ccb88d27141aeb483e4482041108723f | app/Display/display.py | app/Display/display.py | import sys
ESC = chr(27)
CLEAR = ESC + "[2J"
MOVE_HOME = ESC + "[H"
ERASE = CLEAR + MOVE_HOME
LINES = 24
COLS = 80
class Display:
def __init__(self, title):
self.title = title
def clear(self):
sys.stdout.write(ERASE)
def show_properties(self, properties, names=None):
if names is None:
names = properties.keys()
max_len = max(map(len, names))
self.clear()
self.print(self.title)
print()
for k in names:
self.print("{0}: {1}".format(k.rjust(max_len), properties[k]))
def print(self, message):
print(message, end="\x0a\x0d")
| import sys
ESC = chr(27)
CSI = ESC + "["
CLEAR = CSI + "2J"
MOVE_HOME = CSI + "H"
ERASE = CLEAR + MOVE_HOME
MOVE_TO = CSI + "{0};{1}H"
LINES = 24
COLS = 80
class Display:
def __init__(self, title, info=None):
self.title = title
self.info = info
def clear(self):
sys.stdout.write(ERASE)
sys.stdout.flush()
def move_to(self, row, col):
sys.stdout.write(MOVE_TO.format(row, col))
sys.stdout.flush()
def show_properties(self, properties, names=None):
if names is None:
names = properties.keys()
max_len = max(map(len, names))
self.clear()
self.print(self.title.center(COLS))
print()
for k in names:
self.print("{0}: {1}".format(k.rjust(max_len), properties[k]))
if self.info is not None:
self.move_to(LINES, 0)
sys.stdout.write(self.info)
sys.stdout.flush()
self.move_to(LINES, 0)
def print(self, message):
print(message, end="\x0a\x0d")
| Add support for cursor position, centered title, and an info bar | Add support for cursor position, centered title, and an info bar
| Python | mit | gizmo-cda/g2x,gizmo-cda/g2x,thelonious/g2x,gizmo-cda/g2x,gizmo-cda/g2x,thelonious/g2x | import sys
ESC = chr(27)
+ CSI = ESC + "["
- CLEAR = ESC + "[2J"
+ CLEAR = CSI + "2J"
- MOVE_HOME = ESC + "[H"
+ MOVE_HOME = CSI + "H"
ERASE = CLEAR + MOVE_HOME
+ MOVE_TO = CSI + "{0};{1}H"
LINES = 24
COLS = 80
class Display:
- def __init__(self, title):
+ def __init__(self, title, info=None):
self.title = title
+ self.info = info
def clear(self):
sys.stdout.write(ERASE)
+ sys.stdout.flush()
+
+ def move_to(self, row, col):
+ sys.stdout.write(MOVE_TO.format(row, col))
+ sys.stdout.flush()
def show_properties(self, properties, names=None):
if names is None:
names = properties.keys()
max_len = max(map(len, names))
self.clear()
- self.print(self.title)
+ self.print(self.title.center(COLS))
print()
for k in names:
self.print("{0}: {1}".format(k.rjust(max_len), properties[k]))
+ if self.info is not None:
+ self.move_to(LINES, 0)
+ sys.stdout.write(self.info)
+ sys.stdout.flush()
+
+ self.move_to(LINES, 0)
+
def print(self, message):
print(message, end="\x0a\x0d")
| Add support for cursor position, centered title, and an info bar | ## Code Before:
import sys
ESC = chr(27)
CLEAR = ESC + "[2J"
MOVE_HOME = ESC + "[H"
ERASE = CLEAR + MOVE_HOME
LINES = 24
COLS = 80
class Display:
def __init__(self, title):
self.title = title
def clear(self):
sys.stdout.write(ERASE)
def show_properties(self, properties, names=None):
if names is None:
names = properties.keys()
max_len = max(map(len, names))
self.clear()
self.print(self.title)
print()
for k in names:
self.print("{0}: {1}".format(k.rjust(max_len), properties[k]))
def print(self, message):
print(message, end="\x0a\x0d")
## Instruction:
Add support for cursor position, centered title, and an info bar
## Code After:
import sys
ESC = chr(27)
CSI = ESC + "["
CLEAR = CSI + "2J"
MOVE_HOME = CSI + "H"
ERASE = CLEAR + MOVE_HOME
MOVE_TO = CSI + "{0};{1}H"
LINES = 24
COLS = 80
class Display:
def __init__(self, title, info=None):
self.title = title
self.info = info
def clear(self):
sys.stdout.write(ERASE)
sys.stdout.flush()
def move_to(self, row, col):
sys.stdout.write(MOVE_TO.format(row, col))
sys.stdout.flush()
def show_properties(self, properties, names=None):
if names is None:
names = properties.keys()
max_len = max(map(len, names))
self.clear()
self.print(self.title.center(COLS))
print()
for k in names:
self.print("{0}: {1}".format(k.rjust(max_len), properties[k]))
if self.info is not None:
self.move_to(LINES, 0)
sys.stdout.write(self.info)
sys.stdout.flush()
self.move_to(LINES, 0)
def print(self, message):
print(message, end="\x0a\x0d")
| import sys
ESC = chr(27)
+ CSI = ESC + "["
- CLEAR = ESC + "[2J"
? ^ ^ -
+ CLEAR = CSI + "2J"
? ^ ^
- MOVE_HOME = ESC + "[H"
? ^ ^ -
+ MOVE_HOME = CSI + "H"
? ^ ^
ERASE = CLEAR + MOVE_HOME
+ MOVE_TO = CSI + "{0};{1}H"
LINES = 24
COLS = 80
class Display:
- def __init__(self, title):
+ def __init__(self, title, info=None):
? +++++++++++
self.title = title
+ self.info = info
def clear(self):
sys.stdout.write(ERASE)
+ sys.stdout.flush()
+
+ def move_to(self, row, col):
+ sys.stdout.write(MOVE_TO.format(row, col))
+ sys.stdout.flush()
def show_properties(self, properties, names=None):
if names is None:
names = properties.keys()
max_len = max(map(len, names))
self.clear()
- self.print(self.title)
+ self.print(self.title.center(COLS))
? ++++++++++++ +
print()
for k in names:
self.print("{0}: {1}".format(k.rjust(max_len), properties[k]))
+ if self.info is not None:
+ self.move_to(LINES, 0)
+ sys.stdout.write(self.info)
+ sys.stdout.flush()
+
+ self.move_to(LINES, 0)
+
def print(self, message):
print(message, end="\x0a\x0d") |
135ac2fa3aa978328ba72db6ca84920f3da0f39a | furikura/desktop/unity.py | furikura/desktop/unity.py | import gi
import time
gi.require_version('Unity', '7.0')
from gi.repository import Unity, GObject
def update_counter(count):
launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop")
launcher.set_property("count", count)
launcher.set_property("count_visible", True)
| import gi
gi.require_version('Unity', '7.0')
from gi.repository import Unity
def update_counter(count):
launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop")
launcher.set_property("count", count)
launcher.set_property("count_visible", True)
| Remove unnecessary imports for Unity module | Remove unnecessary imports for Unity module
| Python | mit | benjamindean/furi-kura,benjamindean/furi-kura | import gi
- import time
gi.require_version('Unity', '7.0')
- from gi.repository import Unity, GObject
+ from gi.repository import Unity
def update_counter(count):
launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop")
launcher.set_property("count", count)
launcher.set_property("count_visible", True)
-
-
- | Remove unnecessary imports for Unity module | ## Code Before:
import gi
import time
gi.require_version('Unity', '7.0')
from gi.repository import Unity, GObject
def update_counter(count):
launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop")
launcher.set_property("count", count)
launcher.set_property("count_visible", True)
## Instruction:
Remove unnecessary imports for Unity module
## Code After:
import gi
gi.require_version('Unity', '7.0')
from gi.repository import Unity
def update_counter(count):
launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop")
launcher.set_property("count", count)
launcher.set_property("count_visible", True)
| import gi
- import time
gi.require_version('Unity', '7.0')
- from gi.repository import Unity, GObject
? ---------
+ from gi.repository import Unity
def update_counter(count):
launcher = Unity.LauncherEntry.get_for_desktop_id("furikura.desktop")
launcher.set_property("count", count)
launcher.set_property("count_visible", True)
-
-
- |
60f87cb4c3523faf5c5cdbc5f16453cae755988b | angr/procedures/java_jni/GetArrayElements.py | angr/procedures/java_jni/GetArrayElements.py | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
if self.state.solver.eval(ptr_isCopy != 0):
self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | Fix case if isCopy is null | Fix case if isCopy is null
| Python | bsd-2-clause | schieb/angr,schieb/angr,angr/angr,angr/angr,iamahuman/angr,angr/angr,iamahuman/angr,schieb/angr,iamahuman/angr | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
+ if self.state.solver.eval(ptr_isCopy != 0):
+ self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | Fix case if isCopy is null | ## Code Before:
from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values
## Instruction:
Fix case if isCopy is null
## Code After:
from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
if self.state.solver.eval(ptr_isCopy != 0):
self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
+ if self.state.solver.eval(ptr_isCopy != 0):
+ self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values |
9a94e9e61a7bb1680265692eb7cdf926842aa766 | streamline/__init__.py | streamline/__init__.py | from .base import RouteBase, NonIterableRouteBase
from .template import TemplateRoute, XHRPartialRoute, ROCARoute
from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute
__version__ = '1.0'
__author__ = 'Outernet Inc'
__all__ = (
RouteBase,
NonIterableRouteBase,
TemplateRoute,
XHRPartialRoute,
ROCARoute,
FormRoute,
TemplateFormRoute,
XHRPartialFormRoute,
)
| from .base import RouteBase, NonIterableRouteBase
from .template import TemplateRoute, XHRPartialRoute, ROCARoute
from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute
__version__ = '1.0'
__author__ = 'Outernet Inc'
__all__ = (
'RouteBase',
'NonIterableRouteBase',
'TemplateRoute',
'XHRPartialRoute',
'ROCARoute',
'FormRoute',
'TemplateFormRoute',
'XHRPartialFormRoute',
)
| Fix __all__ using objects instead of strings | Fix __all__ using objects instead of strings
Signed-off-by: Branko Vukelic <26059cc39872530f89fec69552bb1050e1cc2caa@outernet.is>
| Python | bsd-2-clause | Outernet-Project/bottle-streamline | from .base import RouteBase, NonIterableRouteBase
from .template import TemplateRoute, XHRPartialRoute, ROCARoute
from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute
__version__ = '1.0'
__author__ = 'Outernet Inc'
__all__ = (
- RouteBase,
+ 'RouteBase',
- NonIterableRouteBase,
+ 'NonIterableRouteBase',
- TemplateRoute,
+ 'TemplateRoute',
- XHRPartialRoute,
+ 'XHRPartialRoute',
- ROCARoute,
+ 'ROCARoute',
- FormRoute,
+ 'FormRoute',
- TemplateFormRoute,
+ 'TemplateFormRoute',
- XHRPartialFormRoute,
+ 'XHRPartialFormRoute',
)
| Fix __all__ using objects instead of strings | ## Code Before:
from .base import RouteBase, NonIterableRouteBase
from .template import TemplateRoute, XHRPartialRoute, ROCARoute
from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute
__version__ = '1.0'
__author__ = 'Outernet Inc'
__all__ = (
RouteBase,
NonIterableRouteBase,
TemplateRoute,
XHRPartialRoute,
ROCARoute,
FormRoute,
TemplateFormRoute,
XHRPartialFormRoute,
)
## Instruction:
Fix __all__ using objects instead of strings
## Code After:
from .base import RouteBase, NonIterableRouteBase
from .template import TemplateRoute, XHRPartialRoute, ROCARoute
from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute
__version__ = '1.0'
__author__ = 'Outernet Inc'
__all__ = (
'RouteBase',
'NonIterableRouteBase',
'TemplateRoute',
'XHRPartialRoute',
'ROCARoute',
'FormRoute',
'TemplateFormRoute',
'XHRPartialFormRoute',
)
| from .base import RouteBase, NonIterableRouteBase
from .template import TemplateRoute, XHRPartialRoute, ROCARoute
from .forms import FormRoute, TemplateFormRoute, XHRPartialFormRoute
__version__ = '1.0'
__author__ = 'Outernet Inc'
__all__ = (
- RouteBase,
+ 'RouteBase',
? + +
- NonIterableRouteBase,
+ 'NonIterableRouteBase',
? + +
- TemplateRoute,
+ 'TemplateRoute',
? + +
- XHRPartialRoute,
+ 'XHRPartialRoute',
? + +
- ROCARoute,
+ 'ROCARoute',
? + +
- FormRoute,
+ 'FormRoute',
? + +
- TemplateFormRoute,
+ 'TemplateFormRoute',
? + +
- XHRPartialFormRoute,
+ 'XHRPartialFormRoute',
? + +
) |
e1c57cb41c59c118648602ff9837418e5d4baad4 | saleor/dashboard/category/forms.py | saleor/dashboard/category/forms.py | from django import forms
from ...product.models import Category
class CategoryForm(forms.ModelForm):
class Meta:
model = Category
exclude = [] | from django import forms
from django.utils.translation import ugettext_lazy as _
from ...product.models import Category
class CategoryForm(forms.ModelForm):
class Meta:
model = Category
exclude = []
def clean_parent(self):
parent = self.cleaned_data['parent']
if parent == self.instance:
raise forms.ValidationError(_('A category may not be made a child of itself'))
return parent
| Add validation on category parent field | Add validation on category parent field
| Python | bsd-3-clause | itbabu/saleor,rchav/vinerack,avorio/saleor,HyperManTT/ECommerceSaleor,laosunhust/saleor,itbabu/saleor,josesanch/saleor,Drekscott/Motlaesaleor,maferelo/saleor,taedori81/saleor,rchav/vinerack,avorio/saleor,rodrigozn/CW-Shop,avorio/saleor,maferelo/saleor,arth-co/saleor,paweltin/saleor,jreigel/saleor,paweltin/saleor,Drekscott/Motlaesaleor,taedori81/saleor,UITools/saleor,taedori81/saleor,dashmug/saleor,itbabu/saleor,tfroehlich82/saleor,paweltin/saleor,UITools/saleor,jreigel/saleor,avorio/saleor,arth-co/saleor,laosunhust/saleor,dashmug/saleor,rodrigozn/CW-Shop,rchav/vinerack,KenMutemi/saleor,spartonia/saleor,spartonia/saleor,KenMutemi/saleor,laosunhust/saleor,jreigel/saleor,arth-co/saleor,tfroehlich82/saleor,arth-co/saleor,josesanch/saleor,car3oon/saleor,Drekscott/Motlaesaleor,josesanch/saleor,UITools/saleor,rodrigozn/CW-Shop,UITools/saleor,spartonia/saleor,HyperManTT/ECommerceSaleor,maferelo/saleor,paweltin/saleor,Drekscott/Motlaesaleor,HyperManTT/ECommerceSaleor,UITools/saleor,KenMutemi/saleor,spartonia/saleor,mociepka/saleor,taedori81/saleor,car3oon/saleor,mociepka/saleor,mociepka/saleor,laosunhust/saleor,car3oon/saleor,dashmug/saleor,tfroehlich82/saleor | from django import forms
+ from django.utils.translation import ugettext_lazy as _
from ...product.models import Category
class CategoryForm(forms.ModelForm):
class Meta:
model = Category
exclude = []
+
+ def clean_parent(self):
+ parent = self.cleaned_data['parent']
+ if parent == self.instance:
+ raise forms.ValidationError(_('A category may not be made a child of itself'))
+ return parent
+ | Add validation on category parent field | ## Code Before:
from django import forms
from ...product.models import Category
class CategoryForm(forms.ModelForm):
class Meta:
model = Category
exclude = []
## Instruction:
Add validation on category parent field
## Code After:
from django import forms
from django.utils.translation import ugettext_lazy as _
from ...product.models import Category
class CategoryForm(forms.ModelForm):
class Meta:
model = Category
exclude = []
def clean_parent(self):
parent = self.cleaned_data['parent']
if parent == self.instance:
raise forms.ValidationError(_('A category may not be made a child of itself'))
return parent
| from django import forms
+ from django.utils.translation import ugettext_lazy as _
from ...product.models import Category
class CategoryForm(forms.ModelForm):
class Meta:
model = Category
exclude = []
+
+ def clean_parent(self):
+ parent = self.cleaned_data['parent']
+ if parent == self.instance:
+ raise forms.ValidationError(_('A category may not be made a child of itself'))
+ return parent |
9df2bae691e8613794be3713194db2420fc75385 | gapipy/resources/dossier/transport_dossier.py | gapipy/resources/dossier/transport_dossier.py | from __future__ import unicode_literals
from ..base import Resource
from .details import DossierDetail, DossierDetailsMixin
from .dossier_features import DossierFeature
class TransportDossier(Resource, DossierDetailsMixin):
_resource_name = 'transport_dossiers'
_as_is_fields = [
'id', 'href', 'features', 'capacity', 'private', 'name',
'dossier_segment',
]
_model_collection_fields = [
('details', DossierDetail),
('features', DossierFeature),
]
_date_time_fields_local = ['date_created', 'date_last_modified']
| from __future__ import unicode_literals
from ..base import Resource
from .details import DossierDetail, DossierDetailsMixin
from .dossier_features import DossierFeature
class TransportDossier(Resource, DossierDetailsMixin):
_resource_name = 'transport_dossiers'
_as_is_fields = [
'id', 'href', 'capacity', 'private', 'name',
'dossier_segment',
]
_model_collection_fields = [
('details', DossierDetail),
('features', DossierFeature),
]
_date_time_fields_local = ['date_created', 'date_last_modified']
| Remove features from as-is fields on TransportDossier | Remove features from as-is fields on TransportDossier
Reflected as a model_collection_field
| Python | mit | gadventures/gapipy | from __future__ import unicode_literals
from ..base import Resource
from .details import DossierDetail, DossierDetailsMixin
from .dossier_features import DossierFeature
class TransportDossier(Resource, DossierDetailsMixin):
_resource_name = 'transport_dossiers'
_as_is_fields = [
- 'id', 'href', 'features', 'capacity', 'private', 'name',
+ 'id', 'href', 'capacity', 'private', 'name',
'dossier_segment',
]
_model_collection_fields = [
('details', DossierDetail),
('features', DossierFeature),
]
_date_time_fields_local = ['date_created', 'date_last_modified']
| Remove features from as-is fields on TransportDossier | ## Code Before:
from __future__ import unicode_literals
from ..base import Resource
from .details import DossierDetail, DossierDetailsMixin
from .dossier_features import DossierFeature
class TransportDossier(Resource, DossierDetailsMixin):
_resource_name = 'transport_dossiers'
_as_is_fields = [
'id', 'href', 'features', 'capacity', 'private', 'name',
'dossier_segment',
]
_model_collection_fields = [
('details', DossierDetail),
('features', DossierFeature),
]
_date_time_fields_local = ['date_created', 'date_last_modified']
## Instruction:
Remove features from as-is fields on TransportDossier
## Code After:
from __future__ import unicode_literals
from ..base import Resource
from .details import DossierDetail, DossierDetailsMixin
from .dossier_features import DossierFeature
class TransportDossier(Resource, DossierDetailsMixin):
_resource_name = 'transport_dossiers'
_as_is_fields = [
'id', 'href', 'capacity', 'private', 'name',
'dossier_segment',
]
_model_collection_fields = [
('details', DossierDetail),
('features', DossierFeature),
]
_date_time_fields_local = ['date_created', 'date_last_modified']
| from __future__ import unicode_literals
from ..base import Resource
from .details import DossierDetail, DossierDetailsMixin
from .dossier_features import DossierFeature
class TransportDossier(Resource, DossierDetailsMixin):
_resource_name = 'transport_dossiers'
_as_is_fields = [
- 'id', 'href', 'features', 'capacity', 'private', 'name',
? ------------
+ 'id', 'href', 'capacity', 'private', 'name',
'dossier_segment',
]
_model_collection_fields = [
('details', DossierDetail),
('features', DossierFeature),
]
_date_time_fields_local = ['date_created', 'date_last_modified'] |
331b3987ba09db5d8f774509bedd30c3c6522795 | ooni/tests/test_utils.py | ooni/tests/test_utils.py | import os
import unittest
from ooni.utils import pushFilenameStack
class TestUtils(unittest.TestCase):
def test_pushFilenameStack(self):
basefilename = os.path.join(os.getcwd(), 'dummyfile')
f = open(basefilename, "w+")
f.write("0\n")
f.close()
for i in xrange(1, 5):
f = open(basefilename+".%s" % i, "w+")
f.write("%s\n" % i)
f.close()
pushFilenameStack(basefilename)
for i in xrange(1, 5):
f = open(basefilename+".%s" % i)
c = f.readlines()[0].strip()
self.assertEqual(str(i-1), str(c))
f.close()
| import os
from twisted.trial import unittest
from ooni.utils import pushFilenameStack
class TestUtils(unittest.TestCase):
def test_pushFilenameStack(self):
basefilename = os.path.join(os.getcwd(), 'dummyfile')
f = open(basefilename, "w+")
f.write("0\n")
f.close()
for i in xrange(1, 5):
f = open(basefilename+".%s" % i, "w+")
f.write("%s\n" % i)
f.close()
pushFilenameStack(basefilename)
for i in xrange(1, 5):
f = open(basefilename+".%s" % i)
c = f.readlines()[0].strip()
self.assertEqual(str(i-1), str(c))
f.close()
| Use trial unittest instead of python unittest | Use trial unittest instead of python unittest
| Python | bsd-2-clause | juga0/ooni-probe,juga0/ooni-probe,lordappsec/ooni-probe,Karthikeyan-kkk/ooni-probe,Karthikeyan-kkk/ooni-probe,0xPoly/ooni-probe,lordappsec/ooni-probe,Karthikeyan-kkk/ooni-probe,0xPoly/ooni-probe,0xPoly/ooni-probe,juga0/ooni-probe,lordappsec/ooni-probe,kdmurray91/ooni-probe,kdmurray91/ooni-probe,lordappsec/ooni-probe,kdmurray91/ooni-probe,Karthikeyan-kkk/ooni-probe,kdmurray91/ooni-probe,0xPoly/ooni-probe,juga0/ooni-probe | import os
- import unittest
+ from twisted.trial import unittest
+
from ooni.utils import pushFilenameStack
class TestUtils(unittest.TestCase):
def test_pushFilenameStack(self):
basefilename = os.path.join(os.getcwd(), 'dummyfile')
f = open(basefilename, "w+")
f.write("0\n")
f.close()
for i in xrange(1, 5):
f = open(basefilename+".%s" % i, "w+")
f.write("%s\n" % i)
f.close()
pushFilenameStack(basefilename)
for i in xrange(1, 5):
f = open(basefilename+".%s" % i)
c = f.readlines()[0].strip()
self.assertEqual(str(i-1), str(c))
f.close()
| Use trial unittest instead of python unittest | ## Code Before:
import os
import unittest
from ooni.utils import pushFilenameStack
class TestUtils(unittest.TestCase):
def test_pushFilenameStack(self):
basefilename = os.path.join(os.getcwd(), 'dummyfile')
f = open(basefilename, "w+")
f.write("0\n")
f.close()
for i in xrange(1, 5):
f = open(basefilename+".%s" % i, "w+")
f.write("%s\n" % i)
f.close()
pushFilenameStack(basefilename)
for i in xrange(1, 5):
f = open(basefilename+".%s" % i)
c = f.readlines()[0].strip()
self.assertEqual(str(i-1), str(c))
f.close()
## Instruction:
Use trial unittest instead of python unittest
## Code After:
import os
from twisted.trial import unittest
from ooni.utils import pushFilenameStack
class TestUtils(unittest.TestCase):
def test_pushFilenameStack(self):
basefilename = os.path.join(os.getcwd(), 'dummyfile')
f = open(basefilename, "w+")
f.write("0\n")
f.close()
for i in xrange(1, 5):
f = open(basefilename+".%s" % i, "w+")
f.write("%s\n" % i)
f.close()
pushFilenameStack(basefilename)
for i in xrange(1, 5):
f = open(basefilename+".%s" % i)
c = f.readlines()[0].strip()
self.assertEqual(str(i-1), str(c))
f.close()
| import os
- import unittest
+ from twisted.trial import unittest
+
from ooni.utils import pushFilenameStack
class TestUtils(unittest.TestCase):
def test_pushFilenameStack(self):
basefilename = os.path.join(os.getcwd(), 'dummyfile')
f = open(basefilename, "w+")
f.write("0\n")
f.close()
for i in xrange(1, 5):
f = open(basefilename+".%s" % i, "w+")
f.write("%s\n" % i)
f.close()
pushFilenameStack(basefilename)
for i in xrange(1, 5):
f = open(basefilename+".%s" % i)
c = f.readlines()[0].strip()
self.assertEqual(str(i-1), str(c))
f.close()
|
8e5ad2138d0685e4322156b3f545be46a3f0c99f | util.py | util.py |
import glob
import os.path
import random
def pick_random(directory, k=None):
"""Pick randomly some files from a directory."""
all_files = glob.glob(os.path.join(directory, '*'))
random.shuffle(all_files)
return all_files if k is None else all_files[:k]
|
import glob
import os.path
import random
import re
def pick(directory, k=None, randomized=True):
"""Pick some thread files from a thread directory."""
all_files = glob.glob(os.path.join(directory, '*'))
if randomized:
random.shuffle(all_files)
else:
pattern = '([0-9]+)\.txt'
all_files.sort(key=lambda f: int(re.search(pattern, f).group(1)))
return all_files if k is None else all_files[:k]
| Modify to pick either randomly or sequentially | Modify to pick either randomly or sequentially
| Python | mit | kemskems/otdet |
import glob
import os.path
import random
+ import re
- def pick_random(directory, k=None):
+ def pick(directory, k=None, randomized=True):
- """Pick randomly some files from a directory."""
+ """Pick some thread files from a thread directory."""
all_files = glob.glob(os.path.join(directory, '*'))
+ if randomized:
- random.shuffle(all_files)
+ random.shuffle(all_files)
+ else:
+ pattern = '([0-9]+)\.txt'
+ all_files.sort(key=lambda f: int(re.search(pattern, f).group(1)))
return all_files if k is None else all_files[:k]
| Modify to pick either randomly or sequentially | ## Code Before:
import glob
import os.path
import random
def pick_random(directory, k=None):
"""Pick randomly some files from a directory."""
all_files = glob.glob(os.path.join(directory, '*'))
random.shuffle(all_files)
return all_files if k is None else all_files[:k]
## Instruction:
Modify to pick either randomly or sequentially
## Code After:
import glob
import os.path
import random
import re
def pick(directory, k=None, randomized=True):
"""Pick some thread files from a thread directory."""
all_files = glob.glob(os.path.join(directory, '*'))
if randomized:
random.shuffle(all_files)
else:
pattern = '([0-9]+)\.txt'
all_files.sort(key=lambda f: int(re.search(pattern, f).group(1)))
return all_files if k is None else all_files[:k]
|
import glob
import os.path
import random
+ import re
- def pick_random(directory, k=None):
+ def pick(directory, k=None, randomized=True):
- """Pick randomly some files from a directory."""
? ---------
+ """Pick some thread files from a thread directory."""
? +++++++ +++++++
all_files = glob.glob(os.path.join(directory, '*'))
+ if randomized:
- random.shuffle(all_files)
+ random.shuffle(all_files)
? ++++
+ else:
+ pattern = '([0-9]+)\.txt'
+ all_files.sort(key=lambda f: int(re.search(pattern, f).group(1)))
return all_files if k is None else all_files[:k] |
b7377196cdd05d9d6d481f7b93308189c4524c52 | sfm/api/filters.py | sfm/api/filters.py | from django_filters import FilterSet, CharFilter, IsoDateTimeFilter
from ui.models import Warc, Seed, Harvest
from django_filters import Filter
from django_filters.fields import Lookup
class ListFilter(Filter):
def filter(self, qs, value):
return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in"))
class WarcFilter(FilterSet):
# Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e
seedset = CharFilter(name="harvest__seed_set__seedset_id")
seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True)
# TODO: This will need to be changed to use historical seeds once #54 is completed.
harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte')
harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte')
class Meta:
model = Warc
fields = ['seedset']
| from django_filters import FilterSet, CharFilter, IsoDateTimeFilter
from ui.models import Warc, Seed, Harvest
from django_filters import Filter
from django_filters.fields import Lookup
class ListFilter(Filter):
def filter(self, qs, value):
return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in"))
class WarcFilter(FilterSet):
# Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e
seedset = CharFilter(name="harvest__historical_seed_set__seedset_id")
seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True)
harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte')
harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte')
class Meta:
model = Warc
fields = ['seedset']
| Fix to take into account history in API queries. | Fix to take into account history in API queries.
| Python | mit | gwu-libraries/sfm,gwu-libraries/sfm-ui,gwu-libraries/sfm,gwu-libraries/sfm,gwu-libraries/sfm-ui,gwu-libraries/sfm-ui,gwu-libraries/sfm-ui | from django_filters import FilterSet, CharFilter, IsoDateTimeFilter
from ui.models import Warc, Seed, Harvest
from django_filters import Filter
from django_filters.fields import Lookup
class ListFilter(Filter):
def filter(self, qs, value):
return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in"))
class WarcFilter(FilterSet):
# Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e
- seedset = CharFilter(name="harvest__seed_set__seedset_id")
+ seedset = CharFilter(name="harvest__historical_seed_set__seedset_id")
- seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True)
+ seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True)
- # TODO: This will need to be changed to use historical seeds once #54 is completed.
harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte')
harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte')
class Meta:
model = Warc
fields = ['seedset']
| Fix to take into account history in API queries. | ## Code Before:
from django_filters import FilterSet, CharFilter, IsoDateTimeFilter
from ui.models import Warc, Seed, Harvest
from django_filters import Filter
from django_filters.fields import Lookup
class ListFilter(Filter):
def filter(self, qs, value):
return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in"))
class WarcFilter(FilterSet):
# Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e
seedset = CharFilter(name="harvest__seed_set__seedset_id")
seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True)
# TODO: This will need to be changed to use historical seeds once #54 is completed.
harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte')
harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte')
class Meta:
model = Warc
fields = ['seedset']
## Instruction:
Fix to take into account history in API queries.
## Code After:
from django_filters import FilterSet, CharFilter, IsoDateTimeFilter
from ui.models import Warc, Seed, Harvest
from django_filters import Filter
from django_filters.fields import Lookup
class ListFilter(Filter):
def filter(self, qs, value):
return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in"))
class WarcFilter(FilterSet):
# Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e
seedset = CharFilter(name="harvest__historical_seed_set__seedset_id")
seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True)
harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte')
harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte')
class Meta:
model = Warc
fields = ['seedset']
| from django_filters import FilterSet, CharFilter, IsoDateTimeFilter
from ui.models import Warc, Seed, Harvest
from django_filters import Filter
from django_filters.fields import Lookup
class ListFilter(Filter):
def filter(self, qs, value):
return super(ListFilter, self).filter(qs, Lookup(value.split(u","), "in"))
class WarcFilter(FilterSet):
# Allows queries like /api/v1/warcs/?seedset=39c00280274a4db0b1cb5bfa4d527a1e
- seedset = CharFilter(name="harvest__seed_set__seedset_id")
+ seedset = CharFilter(name="harvest__historical_seed_set__seedset_id")
? +++++++++++
- seed = ListFilter(name="harvest__seed_set__seeds__seed_id", distinct=True)
? ------ ^
+ seed = ListFilter(name="harvest__historical_seeds__seed_id", distinct=True)
? ++ ^^^^^^
- # TODO: This will need to be changed to use historical seeds once #54 is completed.
harvest_date_start = IsoDateTimeFilter(name="harvest__date_started", lookup_type='gte')
harvest_date_end = IsoDateTimeFilter(name="harvest__date_started", lookup_type='lte')
class Meta:
model = Warc
fields = ['seedset'] |
a137e8a92211d3d344a38b5c97d81073d66a1668 | alembic/versions/17c1af634026_extract_publication_date.py | alembic/versions/17c1af634026_extract_publication_date.py | # revision identifiers, used by Alembic.
revision = '17c1af634026'
down_revision = '3c4c29f0a791'
import html5lib
from dateutil.parser import parse as parse_date
import pytips
from pytips.models import Tip
def _extract_publication_date(html):
root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False)
publication_date_string = root.xpath("//a/@data-datetime")[0]
return parse_date(publication_date_string)
def _update_tip(tip):
tip.publication_date = _extract_publication_date(tip.rendered_html)
def _erase_publication_date(tip):
tip.publication_date = None
def upgrade():
tips = Tip.query.all()
map(_update_tip, tips)
pytips.db.session.commit()
def downgrade():
tips = Tip.query.all()
map(_erase_publication_date, tips)
pytips.db.session.commit()
| # revision identifiers, used by Alembic.
revision = '17c1af634026'
down_revision = '3c4c29f0a791'
import html5lib
from dateutil.parser import parse as parse_date
import pytips
from pytips.util import extract_publication_date
from pytips.models import Tip
def _update_tip(tip):
tip.publication_date = extract_publication_date(tip.rendered_html)
def _erase_publication_date(tip):
tip.publication_date = None
def upgrade():
tips = Tip.query.all()
map(_update_tip, tips)
pytips.db.session.commit()
def downgrade():
tips = Tip.query.all()
map(_erase_publication_date, tips)
pytips.db.session.commit()
| Use the utility module's extract_publication_date logic. | Use the utility module's extract_publication_date logic.
| Python | isc | gthank/pytips,gthank/pytips,gthank/pytips,gthank/pytips | # revision identifiers, used by Alembic.
revision = '17c1af634026'
down_revision = '3c4c29f0a791'
import html5lib
from dateutil.parser import parse as parse_date
import pytips
+ from pytips.util import extract_publication_date
from pytips.models import Tip
- def _extract_publication_date(html):
- root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False)
- publication_date_string = root.xpath("//a/@data-datetime")[0]
- return parse_date(publication_date_string)
-
-
def _update_tip(tip):
- tip.publication_date = _extract_publication_date(tip.rendered_html)
+ tip.publication_date = extract_publication_date(tip.rendered_html)
def _erase_publication_date(tip):
tip.publication_date = None
def upgrade():
tips = Tip.query.all()
map(_update_tip, tips)
pytips.db.session.commit()
def downgrade():
tips = Tip.query.all()
map(_erase_publication_date, tips)
pytips.db.session.commit()
| Use the utility module's extract_publication_date logic. | ## Code Before:
# revision identifiers, used by Alembic.
revision = '17c1af634026'
down_revision = '3c4c29f0a791'
import html5lib
from dateutil.parser import parse as parse_date
import pytips
from pytips.models import Tip
def _extract_publication_date(html):
root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False)
publication_date_string = root.xpath("//a/@data-datetime")[0]
return parse_date(publication_date_string)
def _update_tip(tip):
tip.publication_date = _extract_publication_date(tip.rendered_html)
def _erase_publication_date(tip):
tip.publication_date = None
def upgrade():
tips = Tip.query.all()
map(_update_tip, tips)
pytips.db.session.commit()
def downgrade():
tips = Tip.query.all()
map(_erase_publication_date, tips)
pytips.db.session.commit()
## Instruction:
Use the utility module's extract_publication_date logic.
## Code After:
# revision identifiers, used by Alembic.
revision = '17c1af634026'
down_revision = '3c4c29f0a791'
import html5lib
from dateutil.parser import parse as parse_date
import pytips
from pytips.util import extract_publication_date
from pytips.models import Tip
def _update_tip(tip):
tip.publication_date = extract_publication_date(tip.rendered_html)
def _erase_publication_date(tip):
tip.publication_date = None
def upgrade():
tips = Tip.query.all()
map(_update_tip, tips)
pytips.db.session.commit()
def downgrade():
tips = Tip.query.all()
map(_erase_publication_date, tips)
pytips.db.session.commit()
| # revision identifiers, used by Alembic.
revision = '17c1af634026'
down_revision = '3c4c29f0a791'
import html5lib
from dateutil.parser import parse as parse_date
import pytips
+ from pytips.util import extract_publication_date
from pytips.models import Tip
- def _extract_publication_date(html):
- root = html5lib.parse(html, treebuilder='lxml', namespaceHTMLElements=False)
- publication_date_string = root.xpath("//a/@data-datetime")[0]
- return parse_date(publication_date_string)
-
-
def _update_tip(tip):
- tip.publication_date = _extract_publication_date(tip.rendered_html)
? -
+ tip.publication_date = extract_publication_date(tip.rendered_html)
def _erase_publication_date(tip):
tip.publication_date = None
def upgrade():
tips = Tip.query.all()
map(_update_tip, tips)
pytips.db.session.commit()
def downgrade():
tips = Tip.query.all()
map(_erase_publication_date, tips)
pytips.db.session.commit() |
9d3d2beab6ec06ce13126b818029258a66f450f6 | babelfish/__init__.py | babelfish/__init__.py | __title__ = 'babelfish'
__version__ = '0.4.1'
__author__ = 'Antoine Bertin'
__license__ = 'BSD'
__copyright__ = 'Copyright 2013 the BabelFish authors'
from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter,
CountryReverseConverter)
from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country
from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError
from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language
from .script import SCRIPTS, Script
| __title__ = 'babelfish'
__version__ = '0.4.1'
__author__ = 'Antoine Bertin'
__license__ = 'BSD'
__copyright__ = 'Copyright 2013 the BabelFish authors'
from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter,
CountryReverseConverter)
from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country
from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError
from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language
from .script import SCRIPTS, SCRIPT_MATRIX, Script
| Add SCRIPT_MATRIX to babelfish module imports | Add SCRIPT_MATRIX to babelfish module imports
| Python | bsd-3-clause | Diaoul/babelfish | __title__ = 'babelfish'
__version__ = '0.4.1'
__author__ = 'Antoine Bertin'
__license__ = 'BSD'
__copyright__ = 'Copyright 2013 the BabelFish authors'
from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter,
CountryReverseConverter)
from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country
from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError
from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language
- from .script import SCRIPTS, Script
+ from .script import SCRIPTS, SCRIPT_MATRIX, Script
| Add SCRIPT_MATRIX to babelfish module imports | ## Code Before:
__title__ = 'babelfish'
__version__ = '0.4.1'
__author__ = 'Antoine Bertin'
__license__ = 'BSD'
__copyright__ = 'Copyright 2013 the BabelFish authors'
from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter,
CountryReverseConverter)
from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country
from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError
from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language
from .script import SCRIPTS, Script
## Instruction:
Add SCRIPT_MATRIX to babelfish module imports
## Code After:
__title__ = 'babelfish'
__version__ = '0.4.1'
__author__ = 'Antoine Bertin'
__license__ = 'BSD'
__copyright__ = 'Copyright 2013 the BabelFish authors'
from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter,
CountryReverseConverter)
from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country
from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError
from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language
from .script import SCRIPTS, SCRIPT_MATRIX, Script
| __title__ = 'babelfish'
__version__ = '0.4.1'
__author__ = 'Antoine Bertin'
__license__ = 'BSD'
__copyright__ = 'Copyright 2013 the BabelFish authors'
from .converters import (LanguageConverter, LanguageReverseConverter, LanguageEquivalenceConverter, CountryConverter,
CountryReverseConverter)
from .country import country_converters, COUNTRIES, COUNTRY_MATRIX, Country
from .exceptions import Error, LanguageConvertError, LanguageReverseError, CountryConvertError, CountryReverseError
from .language import language_converters, LANGUAGES, LANGUAGE_MATRIX, Language
- from .script import SCRIPTS, Script
+ from .script import SCRIPTS, SCRIPT_MATRIX, Script
? +++++++++++++++
|
b2268ae4ecad477c46a4b99ec17511e2e535b9d0 | globus_cli/commands/task/generate_submission_id.py | globus_cli/commands/task/generate_submission_id.py | import click
from globus_cli.parsing import common_options
from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print
from globus_cli.services.transfer import get_client
@click.command(
"generate-submission-id",
short_help="Get a submission ID",
help=(
"Generate a new task submission ID for use in "
"`globus transfer` and `gloubs delete`. Submission IDs "
"allow you to safely retry submission of a task in the "
"presence of network errors. No matter how many times "
"you submit a task with a given ID, it will only be "
"accepted and executed once. The response status may "
"change between submissions."
),
)
@common_options
def generate_submission_id():
"""
Executor for `globus task generate-submission-id`
"""
client = get_client()
res = client.get_submission_id()
formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
| import click
from globus_cli.parsing import common_options
from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print
from globus_cli.services.transfer import get_client
@click.command(
"generate-submission-id",
short_help="Get a submission ID",
help=(
"""\
Generate a new task submission ID for use in `globus transfer` and `gloubs delete`.
Submission IDs allow you to safely retry submission of a task in the presence of
network errors. No matter how many times you submit a task with a given ID, it will
only be accepted and executed once. The response status may change between
submissions.
\b
Important Note: Submission IDs are not the same as Task IDs.
"""
),
)
@common_options
def generate_submission_id():
"""
Executor for `globus task generate-submission-id`
"""
client = get_client()
res = client.get_submission_id()
formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
| Clarify that submission ID != task ID | Clarify that submission ID != task ID
Calling this out in the helptext will hopefully help avoid people
conflating these two quite as easily. (An imperfect solution for an
imperfect world.)
| Python | apache-2.0 | globus/globus-cli,globus/globus-cli | import click
from globus_cli.parsing import common_options
from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print
from globus_cli.services.transfer import get_client
@click.command(
"generate-submission-id",
short_help="Get a submission ID",
help=(
- "Generate a new task submission ID for use in "
- "`globus transfer` and `gloubs delete`. Submission IDs "
+ """\
+ Generate a new task submission ID for use in `globus transfer` and `gloubs delete`.
- "allow you to safely retry submission of a task in the "
+ Submission IDs allow you to safely retry submission of a task in the presence of
+ network errors. No matter how many times you submit a task with a given ID, it will
- "presence of network errors. No matter how many times "
- "you submit a task with a given ID, it will only be "
- "accepted and executed once. The response status may "
+ only be accepted and executed once. The response status may change between
- "change between submissions."
+ submissions.
+
+ \b
+ Important Note: Submission IDs are not the same as Task IDs.
+ """
),
)
@common_options
def generate_submission_id():
"""
Executor for `globus task generate-submission-id`
"""
client = get_client()
res = client.get_submission_id()
formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
| Clarify that submission ID != task ID | ## Code Before:
import click
from globus_cli.parsing import common_options
from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print
from globus_cli.services.transfer import get_client
@click.command(
"generate-submission-id",
short_help="Get a submission ID",
help=(
"Generate a new task submission ID for use in "
"`globus transfer` and `gloubs delete`. Submission IDs "
"allow you to safely retry submission of a task in the "
"presence of network errors. No matter how many times "
"you submit a task with a given ID, it will only be "
"accepted and executed once. The response status may "
"change between submissions."
),
)
@common_options
def generate_submission_id():
"""
Executor for `globus task generate-submission-id`
"""
client = get_client()
res = client.get_submission_id()
formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
## Instruction:
Clarify that submission ID != task ID
## Code After:
import click
from globus_cli.parsing import common_options
from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print
from globus_cli.services.transfer import get_client
@click.command(
"generate-submission-id",
short_help="Get a submission ID",
help=(
"""\
Generate a new task submission ID for use in `globus transfer` and `gloubs delete`.
Submission IDs allow you to safely retry submission of a task in the presence of
network errors. No matter how many times you submit a task with a given ID, it will
only be accepted and executed once. The response status may change between
submissions.
\b
Important Note: Submission IDs are not the same as Task IDs.
"""
),
)
@common_options
def generate_submission_id():
"""
Executor for `globus task generate-submission-id`
"""
client = get_client()
res = client.get_submission_id()
formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value")
| import click
from globus_cli.parsing import common_options
from globus_cli.safeio import FORMAT_TEXT_RAW, formatted_print
from globus_cli.services.transfer import get_client
@click.command(
"generate-submission-id",
short_help="Get a submission ID",
help=(
- "Generate a new task submission ID for use in "
- "`globus transfer` and `gloubs delete`. Submission IDs "
+ """\
+ Generate a new task submission ID for use in `globus transfer` and `gloubs delete`.
- "allow you to safely retry submission of a task in the "
? --- ^
+ Submission IDs allow you to safely retry submission of a task in the presence of
? ++++++++++ +++ ^^^^^^^^^^^
+ network errors. No matter how many times you submit a task with a given ID, it will
- "presence of network errors. No matter how many times "
- "you submit a task with a given ID, it will only be "
- "accepted and executed once. The response status may "
? --- ^
+ only be accepted and executed once. The response status may change between
? ++++ ++ ^^^^^^^^^^^^^^
- "change between submissions."
+ submissions.
+
+ \b
+ Important Note: Submission IDs are not the same as Task IDs.
+ """
),
)
@common_options
def generate_submission_id():
"""
Executor for `globus task generate-submission-id`
"""
client = get_client()
res = client.get_submission_id()
formatted_print(res, text_format=FORMAT_TEXT_RAW, response_key="value") |
66035a6e3e7729c53278193d4307751b36ace6eb | fullcalendar/admin.py | fullcalendar/admin.py | from django.utils.translation import ugettext_lazy as _
from django.contrib import admin
from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin
from fullcalendar.models import *
class EventCategoryAdmin(admin.ModelAdmin):
list_display = ('name',)
class OccurrenceInline(TabularDynamicInlineAdmin):
model = Occurrence
extra = 1
fields = ('start_time', 'end_time', 'description')
class EventAdmin(DisplayableAdmin):
list_display = ('title', 'event_category')
list_filter = ('event_category',)
search_fields = ('title', 'description', 'content', 'keywords')
fieldsets = (
(None, {
"fields": [
"title", "status", ("publish_date", "expiry_date"),
"event_category", "content"
]
}),
(_("Meta data"), {
"fields": [
"_meta_title", "slug",
("description", "gen_description"),
"keywords", "in_sitemap"
],
"classes": ("collapse-closed",)
}),
)
inlines = [OccurrenceInline]
admin.site.register(Event, EventAdmin)
admin.site.register(EventCategory, EventCategoryAdmin)
| from django.utils.translation import ugettext_lazy as _
from django.contrib import admin
from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin
from fullcalendar.models import *
class EventCategoryAdmin(admin.ModelAdmin):
list_display = ('name',)
class OccurrenceInline(StackedDynamicInlineAdmin):
model = Occurrence
extra = 1
fields = ('start_time', 'end_time', 'description', 'location')
class EventAdmin(DisplayableAdmin):
list_display = ('title', 'event_category')
list_filter = ('event_category',)
search_fields = ('title', 'description', 'content', 'keywords')
fieldsets = (
(None, {
"fields": [
"title", "status", ("publish_date", "expiry_date"),
"event_category", "content"
]
}),
(_("Meta data"), {
"fields": [
"_meta_title", "slug",
("description", "gen_description"),
"keywords", "in_sitemap"
],
"classes": ("collapse-closed",)
}),
)
inlines = [OccurrenceInline]
admin.site.register(Event, EventAdmin)
admin.site.register(EventCategory, EventCategoryAdmin)
| Change to stacked inline for occurrences, also display location. | Change to stacked inline for occurrences, also display location.
| Python | mit | jonge-democraten/mezzanine-fullcalendar | from django.utils.translation import ugettext_lazy as _
from django.contrib import admin
- from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin
+ from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin
from fullcalendar.models import *
class EventCategoryAdmin(admin.ModelAdmin):
list_display = ('name',)
- class OccurrenceInline(TabularDynamicInlineAdmin):
+ class OccurrenceInline(StackedDynamicInlineAdmin):
model = Occurrence
extra = 1
- fields = ('start_time', 'end_time', 'description')
+ fields = ('start_time', 'end_time', 'description', 'location')
class EventAdmin(DisplayableAdmin):
list_display = ('title', 'event_category')
list_filter = ('event_category',)
search_fields = ('title', 'description', 'content', 'keywords')
fieldsets = (
(None, {
"fields": [
"title", "status", ("publish_date", "expiry_date"),
"event_category", "content"
]
}),
(_("Meta data"), {
"fields": [
"_meta_title", "slug",
("description", "gen_description"),
"keywords", "in_sitemap"
],
"classes": ("collapse-closed",)
}),
)
inlines = [OccurrenceInline]
admin.site.register(Event, EventAdmin)
admin.site.register(EventCategory, EventCategoryAdmin)
| Change to stacked inline for occurrences, also display location. | ## Code Before:
from django.utils.translation import ugettext_lazy as _
from django.contrib import admin
from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin
from fullcalendar.models import *
class EventCategoryAdmin(admin.ModelAdmin):
list_display = ('name',)
class OccurrenceInline(TabularDynamicInlineAdmin):
model = Occurrence
extra = 1
fields = ('start_time', 'end_time', 'description')
class EventAdmin(DisplayableAdmin):
list_display = ('title', 'event_category')
list_filter = ('event_category',)
search_fields = ('title', 'description', 'content', 'keywords')
fieldsets = (
(None, {
"fields": [
"title", "status", ("publish_date", "expiry_date"),
"event_category", "content"
]
}),
(_("Meta data"), {
"fields": [
"_meta_title", "slug",
("description", "gen_description"),
"keywords", "in_sitemap"
],
"classes": ("collapse-closed",)
}),
)
inlines = [OccurrenceInline]
admin.site.register(Event, EventAdmin)
admin.site.register(EventCategory, EventCategoryAdmin)
## Instruction:
Change to stacked inline for occurrences, also display location.
## Code After:
from django.utils.translation import ugettext_lazy as _
from django.contrib import admin
from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin
from fullcalendar.models import *
class EventCategoryAdmin(admin.ModelAdmin):
list_display = ('name',)
class OccurrenceInline(StackedDynamicInlineAdmin):
model = Occurrence
extra = 1
fields = ('start_time', 'end_time', 'description', 'location')
class EventAdmin(DisplayableAdmin):
list_display = ('title', 'event_category')
list_filter = ('event_category',)
search_fields = ('title', 'description', 'content', 'keywords')
fieldsets = (
(None, {
"fields": [
"title", "status", ("publish_date", "expiry_date"),
"event_category", "content"
]
}),
(_("Meta data"), {
"fields": [
"_meta_title", "slug",
("description", "gen_description"),
"keywords", "in_sitemap"
],
"classes": ("collapse-closed",)
}),
)
inlines = [OccurrenceInline]
admin.site.register(Event, EventAdmin)
admin.site.register(EventCategory, EventCategoryAdmin)
| from django.utils.translation import ugettext_lazy as _
from django.contrib import admin
- from mezzanine.core.admin import TabularDynamicInlineAdmin, DisplayableAdmin
? ^ ^^^^^
+ from mezzanine.core.admin import StackedDynamicInlineAdmin, DisplayableAdmin
? ^^ ^^^^
from fullcalendar.models import *
class EventCategoryAdmin(admin.ModelAdmin):
list_display = ('name',)
- class OccurrenceInline(TabularDynamicInlineAdmin):
? ^ ^^^^^
+ class OccurrenceInline(StackedDynamicInlineAdmin):
? ^^ ^^^^
model = Occurrence
extra = 1
- fields = ('start_time', 'end_time', 'description')
+ fields = ('start_time', 'end_time', 'description', 'location')
? ++++++++++++
class EventAdmin(DisplayableAdmin):
list_display = ('title', 'event_category')
list_filter = ('event_category',)
search_fields = ('title', 'description', 'content', 'keywords')
fieldsets = (
(None, {
"fields": [
"title", "status", ("publish_date", "expiry_date"),
"event_category", "content"
]
}),
(_("Meta data"), {
"fields": [
"_meta_title", "slug",
("description", "gen_description"),
"keywords", "in_sitemap"
],
"classes": ("collapse-closed",)
}),
)
inlines = [OccurrenceInline]
admin.site.register(Event, EventAdmin)
admin.site.register(EventCategory, EventCategoryAdmin)
|
eba6e117c0a13b49219bb60e773f896b274b6601 | tests/_support/configs/collection.py | tests/_support/configs/collection.py | from spec import eq_
from invoke import ctask, Collection
@ctask
def collection(c):
c.run('false') # Ensures a kaboom if mocking fails
ns = Collection(collection)
ns.configure({'run': {'echo': True}})
| from spec import eq_
from invoke import ctask, Collection
@ctask
def go(c):
c.run('false') # Ensures a kaboom if mocking fails
ns = Collection(go)
ns.configure({'run': {'echo': True}})
| Fix test fixture to match earlier test change | Fix test fixture to match earlier test change
| Python | bsd-2-clause | singingwolfboy/invoke,kejbaly2/invoke,tyewang/invoke,frol/invoke,mattrobenolt/invoke,mkusz/invoke,pfmoore/invoke,mkusz/invoke,pyinvoke/invoke,kejbaly2/invoke,pfmoore/invoke,sophacles/invoke,frol/invoke,pyinvoke/invoke,mattrobenolt/invoke | from spec import eq_
from invoke import ctask, Collection
@ctask
- def collection(c):
+ def go(c):
c.run('false') # Ensures a kaboom if mocking fails
- ns = Collection(collection)
+ ns = Collection(go)
ns.configure({'run': {'echo': True}})
| Fix test fixture to match earlier test change | ## Code Before:
from spec import eq_
from invoke import ctask, Collection
@ctask
def collection(c):
c.run('false') # Ensures a kaboom if mocking fails
ns = Collection(collection)
ns.configure({'run': {'echo': True}})
## Instruction:
Fix test fixture to match earlier test change
## Code After:
from spec import eq_
from invoke import ctask, Collection
@ctask
def go(c):
c.run('false') # Ensures a kaboom if mocking fails
ns = Collection(go)
ns.configure({'run': {'echo': True}})
| from spec import eq_
from invoke import ctask, Collection
@ctask
- def collection(c):
+ def go(c):
c.run('false') # Ensures a kaboom if mocking fails
- ns = Collection(collection)
? ^ --------
+ ns = Collection(go)
? ^
ns.configure({'run': {'echo': True}}) |
e5acbfc176de3b531528c8b15f57e5d3feab3ad1 | melody/constraints/abstract_constraint.py | melody/constraints/abstract_constraint.py | from abc import ABCMeta, abstractmethod
class AbstractConstraint(object):
"""
Class that represents a constraint, a set of actors that define a constraint amongst themselves.
ParameterMap: A map from template note to contextual note..
"""
__metaclass__ = ABCMeta
def __init__(self, actors):
self.__actors = list(actors)
@property
def actors(self):
return list(self.__actors)
@abstractmethod
def clone(self, new_actors=None):
"""
Clone the constraint.
:return:
"""
@abstractmethod
def verify(self, solution_context):
"""
Verify that the actor map parameters are consistent with constraint.
:params solution_context: aka pmap, map of actors to ContextualNotes.
:return: Boolean if verification holds.
May throw Exception dependent on implementation.
"""
@abstractmethod
def values(self, solution_context, v_note):
"""
Method to generate all possible note values for actor v_note's target.
The method returns a set of values for v_note.
:param solution_context: includes parameter map.
:param v_note: source actor, whose target values we are computing.
:return: The set of all possible values for v_note's target.
Note: The return value is a set!
"""
| from abc import ABCMeta, abstractmethod
class AbstractConstraint(object):
"""
Class that represents a constraint, a set of actors that define a constraint amongst themselves.
ParameterMap: A map from template note to contextual note..
"""
__metaclass__ = ABCMeta
def __init__(self, actors):
self.__actors = list(actors)
@property
def actors(self):
return list(self.__actors)
@abstractmethod
def clone(self, new_actors=None):
"""
Clone the constraint.
:return:
"""
@abstractmethod
def verify(self, solution_context):
"""
Verify that the actor map parameters are consistent with constraint.
:params solution_context: aka pmap, map of actors to ContextualNotes.
:return: Boolean if verification holds.
May throw Exception dependent on implementation.
"""
@abstractmethod
def values(self, solution_context, v_note):
"""
Method to generate all possible note values for actor v_note's target.
The method returns a set of values for v_note.
:param solution_context: includes parameter map.
:param v_note: source actor, whose target values we are computing.
:return: The set of all possible values for v_note's target.
Note: The return value is a set!
"""
def __hash__(self):
return hash(len(self.actors))
def __eq__(self, other):
if not isinstance(other, AbstractConstraint):
return NotImplemented
return self is other | Add hash and eq methods | Add hash and eq methods
| Python | mit | dpazel/music_rep | from abc import ABCMeta, abstractmethod
class AbstractConstraint(object):
"""
Class that represents a constraint, a set of actors that define a constraint amongst themselves.
ParameterMap: A map from template note to contextual note..
"""
__metaclass__ = ABCMeta
def __init__(self, actors):
self.__actors = list(actors)
@property
def actors(self):
return list(self.__actors)
@abstractmethod
def clone(self, new_actors=None):
"""
Clone the constraint.
:return:
"""
@abstractmethod
def verify(self, solution_context):
"""
Verify that the actor map parameters are consistent with constraint.
:params solution_context: aka pmap, map of actors to ContextualNotes.
:return: Boolean if verification holds.
May throw Exception dependent on implementation.
"""
@abstractmethod
def values(self, solution_context, v_note):
"""
Method to generate all possible note values for actor v_note's target.
The method returns a set of values for v_note.
:param solution_context: includes parameter map.
:param v_note: source actor, whose target values we are computing.
:return: The set of all possible values for v_note's target.
Note: The return value is a set!
"""
+ def __hash__(self):
+ return hash(len(self.actors))
+
+ def __eq__(self, other):
+ if not isinstance(other, AbstractConstraint):
+ return NotImplemented
+ return self is other | Add hash and eq methods | ## Code Before:
from abc import ABCMeta, abstractmethod
class AbstractConstraint(object):
"""
Class that represents a constraint, a set of actors that define a constraint amongst themselves.
ParameterMap: A map from template note to contextual note..
"""
__metaclass__ = ABCMeta
def __init__(self, actors):
self.__actors = list(actors)
@property
def actors(self):
return list(self.__actors)
@abstractmethod
def clone(self, new_actors=None):
"""
Clone the constraint.
:return:
"""
@abstractmethod
def verify(self, solution_context):
"""
Verify that the actor map parameters are consistent with constraint.
:params solution_context: aka pmap, map of actors to ContextualNotes.
:return: Boolean if verification holds.
May throw Exception dependent on implementation.
"""
@abstractmethod
def values(self, solution_context, v_note):
"""
Method to generate all possible note values for actor v_note's target.
The method returns a set of values for v_note.
:param solution_context: includes parameter map.
:param v_note: source actor, whose target values we are computing.
:return: The set of all possible values for v_note's target.
Note: The return value is a set!
"""
## Instruction:
Add hash and eq methods
## Code After:
from abc import ABCMeta, abstractmethod
class AbstractConstraint(object):
"""
Class that represents a constraint, a set of actors that define a constraint amongst themselves.
ParameterMap: A map from template note to contextual note..
"""
__metaclass__ = ABCMeta
def __init__(self, actors):
self.__actors = list(actors)
@property
def actors(self):
return list(self.__actors)
@abstractmethod
def clone(self, new_actors=None):
"""
Clone the constraint.
:return:
"""
@abstractmethod
def verify(self, solution_context):
"""
Verify that the actor map parameters are consistent with constraint.
:params solution_context: aka pmap, map of actors to ContextualNotes.
:return: Boolean if verification holds.
May throw Exception dependent on implementation.
"""
@abstractmethod
def values(self, solution_context, v_note):
"""
Method to generate all possible note values for actor v_note's target.
The method returns a set of values for v_note.
:param solution_context: includes parameter map.
:param v_note: source actor, whose target values we are computing.
:return: The set of all possible values for v_note's target.
Note: The return value is a set!
"""
def __hash__(self):
return hash(len(self.actors))
def __eq__(self, other):
if not isinstance(other, AbstractConstraint):
return NotImplemented
return self is other | from abc import ABCMeta, abstractmethod
class AbstractConstraint(object):
"""
Class that represents a constraint, a set of actors that define a constraint amongst themselves.
ParameterMap: A map from template note to contextual note..
"""
__metaclass__ = ABCMeta
def __init__(self, actors):
self.__actors = list(actors)
@property
def actors(self):
return list(self.__actors)
@abstractmethod
def clone(self, new_actors=None):
"""
Clone the constraint.
:return:
"""
@abstractmethod
def verify(self, solution_context):
"""
Verify that the actor map parameters are consistent with constraint.
:params solution_context: aka pmap, map of actors to ContextualNotes.
:return: Boolean if verification holds.
May throw Exception dependent on implementation.
"""
@abstractmethod
def values(self, solution_context, v_note):
"""
Method to generate all possible note values for actor v_note's target.
The method returns a set of values for v_note.
:param solution_context: includes parameter map.
:param v_note: source actor, whose target values we are computing.
:return: The set of all possible values for v_note's target.
Note: The return value is a set!
"""
+
+ def __hash__(self):
+ return hash(len(self.actors))
+
+ def __eq__(self, other):
+ if not isinstance(other, AbstractConstraint):
+ return NotImplemented
+ return self is other |
d2e03bf76f585dc1025b5a94be0327284f8d5fa2 | Left_pare.py | Left_pare.py |
from xpcom import components
viewSvc = components.classes["@activestate.com/koViewService;1"]\
.getService(components.interfaces.koIViewService)
view = viewSvc.currentView
view = view.queryInterface(components.interfaces.koIScintillaView)
sm = view.scimoz
# Make `start` the beginning position of the first selected line,
# and `end` the ending position of the last selected line.
if sm.anchor < sm.currentPos:
start = sm.positionFromLine(sm.lineFromPosition(sm.anchor))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos))
else:
start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor))
lines = tuple(sm.getTextRange(start, end).splitlines())
# Cut one character from the left
lines = tuple(l[1:] for l in lines)
# Select part of document
sm.setSel(start, end)
# Replace selection content
text = '\n'.join(lines)
sm.replaceSel(text)
# Keep selection to allow to continue to apply this macro if use wants
sm.setSel(start, start+len(text))
|
from xpcom import components
viewSvc = components.classes["@activestate.com/koViewService;1"]\
.getService(components.interfaces.koIViewService)
view = viewSvc.currentView
view = view.queryInterface(components.interfaces.koIScintillaView)
sm = view.scimoz
# Make `start` the beginning position of the first selected line,
# and `end` the ending position of the last selected line.
if sm.anchor < sm.currentPos:
start = sm.positionFromLine(sm.lineFromPosition(sm.anchor))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos))
else:
start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor))
lines = tuple(sm.getTextRange(start, end).splitlines())
# Cut one character from the left
lines = tuple(l[1:] for l in lines)
# Select part of document
sm.setSel(start, end)
# Replace selection content
text = '\n'.join(lines)
sm.replaceSel(text)
# Keep selection to let user continue to apply this macro
sm.setSel(start, start+len(text.encode('utf-8')))
| Correct selecting text by length when text is Unicode. | Correct selecting text by length when text is Unicode.
| Python | mpl-2.0 | Komodo/macros,Komodo/macros |
from xpcom import components
viewSvc = components.classes["@activestate.com/koViewService;1"]\
.getService(components.interfaces.koIViewService)
view = viewSvc.currentView
view = view.queryInterface(components.interfaces.koIScintillaView)
sm = view.scimoz
# Make `start` the beginning position of the first selected line,
# and `end` the ending position of the last selected line.
if sm.anchor < sm.currentPos:
start = sm.positionFromLine(sm.lineFromPosition(sm.anchor))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos))
else:
start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor))
lines = tuple(sm.getTextRange(start, end).splitlines())
# Cut one character from the left
lines = tuple(l[1:] for l in lines)
# Select part of document
sm.setSel(start, end)
# Replace selection content
text = '\n'.join(lines)
sm.replaceSel(text)
- # Keep selection to allow to continue to apply this macro if use wants
+ # Keep selection to let user continue to apply this macro
- sm.setSel(start, start+len(text))
+ sm.setSel(start, start+len(text.encode('utf-8')))
| Correct selecting text by length when text is Unicode. | ## Code Before:
from xpcom import components
viewSvc = components.classes["@activestate.com/koViewService;1"]\
.getService(components.interfaces.koIViewService)
view = viewSvc.currentView
view = view.queryInterface(components.interfaces.koIScintillaView)
sm = view.scimoz
# Make `start` the beginning position of the first selected line,
# and `end` the ending position of the last selected line.
if sm.anchor < sm.currentPos:
start = sm.positionFromLine(sm.lineFromPosition(sm.anchor))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos))
else:
start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor))
lines = tuple(sm.getTextRange(start, end).splitlines())
# Cut one character from the left
lines = tuple(l[1:] for l in lines)
# Select part of document
sm.setSel(start, end)
# Replace selection content
text = '\n'.join(lines)
sm.replaceSel(text)
# Keep selection to allow to continue to apply this macro if use wants
sm.setSel(start, start+len(text))
## Instruction:
Correct selecting text by length when text is Unicode.
## Code After:
from xpcom import components
viewSvc = components.classes["@activestate.com/koViewService;1"]\
.getService(components.interfaces.koIViewService)
view = viewSvc.currentView
view = view.queryInterface(components.interfaces.koIScintillaView)
sm = view.scimoz
# Make `start` the beginning position of the first selected line,
# and `end` the ending position of the last selected line.
if sm.anchor < sm.currentPos:
start = sm.positionFromLine(sm.lineFromPosition(sm.anchor))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos))
else:
start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor))
lines = tuple(sm.getTextRange(start, end).splitlines())
# Cut one character from the left
lines = tuple(l[1:] for l in lines)
# Select part of document
sm.setSel(start, end)
# Replace selection content
text = '\n'.join(lines)
sm.replaceSel(text)
# Keep selection to let user continue to apply this macro
sm.setSel(start, start+len(text.encode('utf-8')))
|
from xpcom import components
viewSvc = components.classes["@activestate.com/koViewService;1"]\
.getService(components.interfaces.koIViewService)
view = viewSvc.currentView
view = view.queryInterface(components.interfaces.koIScintillaView)
sm = view.scimoz
# Make `start` the beginning position of the first selected line,
# and `end` the ending position of the last selected line.
if sm.anchor < sm.currentPos:
start = sm.positionFromLine(sm.lineFromPosition(sm.anchor))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.currentPos))
else:
start = sm.positionFromLine(sm.lineFromPosition(sm.currentPos))
end = sm.getLineEndPosition(sm.lineFromPosition(sm.anchor))
lines = tuple(sm.getTextRange(start, end).splitlines())
# Cut one character from the left
lines = tuple(l[1:] for l in lines)
# Select part of document
sm.setSel(start, end)
# Replace selection content
text = '\n'.join(lines)
sm.replaceSel(text)
- # Keep selection to allow to continue to apply this macro if use wants
? - ^^^ ^^ -------------
+ # Keep selection to let user continue to apply this macro
? ^^ ^^^^
- sm.setSel(start, start+len(text))
+ sm.setSel(start, start+len(text.encode('utf-8')))
? +++++++++++++++ +
|
a7d8d2f95acbf801c0cc8b0f2a8cc008f6cb34c0 | rouver/types.py | rouver/types.py | from __future__ import annotations
from collections.abc import Iterable, Mapping
from typing import Any, Callable, Dict, Tuple
from typing_extensions import TypeAlias
from werkzeug.wrappers import Request
# (name, value)
Header: TypeAlias = Tuple[str, str]
WSGIEnvironment: TypeAlias = Dict[str, Any]
# (body) -> None
StartResponseReturnType: TypeAlias = Callable[[bytes], object]
# (status: str, headers: List[Headers], exc_info) -> response
StartResponse: TypeAlias = Callable[..., StartResponseReturnType]
WSGIResponse: TypeAlias = Iterable[bytes]
WSGIApplication: TypeAlias = Callable[
[WSGIEnvironment, StartResponse], WSGIResponse
]
# (method, path, callback)
RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication]
# (request, previous_args, path_part) -> result
RouteTemplateHandler: TypeAlias = Callable[
[Request, Tuple[Any, ...], str], Any
]
BadArgumentsDict: TypeAlias = Mapping[str, str]
| from __future__ import annotations
from typing import Any, Callable, Dict, Iterable, Mapping, Tuple
from typing_extensions import TypeAlias
from werkzeug.wrappers import Request
# (name, value)
Header: TypeAlias = Tuple[str, str]
WSGIEnvironment: TypeAlias = Dict[str, Any]
# (body) -> None
StartResponseReturnType: TypeAlias = Callable[[bytes], object]
# (status: str, headers: List[Headers], exc_info) -> response
StartResponse: TypeAlias = Callable[..., StartResponseReturnType]
WSGIResponse: TypeAlias = Iterable[bytes]
WSGIApplication: TypeAlias = Callable[
[WSGIEnvironment, StartResponse], WSGIResponse
]
# (method, path, callback)
RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication]
# (request, previous_args, path_part) -> result
RouteTemplateHandler: TypeAlias = Callable[
[Request, Tuple[Any, ...], str], Any
]
BadArgumentsDict: TypeAlias = Mapping[str, str]
| Fix imports on Python <= 3.8 | Fix imports on Python <= 3.8
| Python | mit | srittau/rouver | from __future__ import annotations
- from collections.abc import Iterable, Mapping
- from typing import Any, Callable, Dict, Tuple
+ from typing import Any, Callable, Dict, Iterable, Mapping, Tuple
from typing_extensions import TypeAlias
from werkzeug.wrappers import Request
# (name, value)
Header: TypeAlias = Tuple[str, str]
WSGIEnvironment: TypeAlias = Dict[str, Any]
# (body) -> None
StartResponseReturnType: TypeAlias = Callable[[bytes], object]
# (status: str, headers: List[Headers], exc_info) -> response
StartResponse: TypeAlias = Callable[..., StartResponseReturnType]
WSGIResponse: TypeAlias = Iterable[bytes]
WSGIApplication: TypeAlias = Callable[
[WSGIEnvironment, StartResponse], WSGIResponse
]
# (method, path, callback)
RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication]
# (request, previous_args, path_part) -> result
RouteTemplateHandler: TypeAlias = Callable[
[Request, Tuple[Any, ...], str], Any
]
BadArgumentsDict: TypeAlias = Mapping[str, str]
| Fix imports on Python <= 3.8 | ## Code Before:
from __future__ import annotations
from collections.abc import Iterable, Mapping
from typing import Any, Callable, Dict, Tuple
from typing_extensions import TypeAlias
from werkzeug.wrappers import Request
# (name, value)
Header: TypeAlias = Tuple[str, str]
WSGIEnvironment: TypeAlias = Dict[str, Any]
# (body) -> None
StartResponseReturnType: TypeAlias = Callable[[bytes], object]
# (status: str, headers: List[Headers], exc_info) -> response
StartResponse: TypeAlias = Callable[..., StartResponseReturnType]
WSGIResponse: TypeAlias = Iterable[bytes]
WSGIApplication: TypeAlias = Callable[
[WSGIEnvironment, StartResponse], WSGIResponse
]
# (method, path, callback)
RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication]
# (request, previous_args, path_part) -> result
RouteTemplateHandler: TypeAlias = Callable[
[Request, Tuple[Any, ...], str], Any
]
BadArgumentsDict: TypeAlias = Mapping[str, str]
## Instruction:
Fix imports on Python <= 3.8
## Code After:
from __future__ import annotations
from typing import Any, Callable, Dict, Iterable, Mapping, Tuple
from typing_extensions import TypeAlias
from werkzeug.wrappers import Request
# (name, value)
Header: TypeAlias = Tuple[str, str]
WSGIEnvironment: TypeAlias = Dict[str, Any]
# (body) -> None
StartResponseReturnType: TypeAlias = Callable[[bytes], object]
# (status: str, headers: List[Headers], exc_info) -> response
StartResponse: TypeAlias = Callable[..., StartResponseReturnType]
WSGIResponse: TypeAlias = Iterable[bytes]
WSGIApplication: TypeAlias = Callable[
[WSGIEnvironment, StartResponse], WSGIResponse
]
# (method, path, callback)
RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication]
# (request, previous_args, path_part) -> result
RouteTemplateHandler: TypeAlias = Callable[
[Request, Tuple[Any, ...], str], Any
]
BadArgumentsDict: TypeAlias = Mapping[str, str]
| from __future__ import annotations
- from collections.abc import Iterable, Mapping
- from typing import Any, Callable, Dict, Tuple
+ from typing import Any, Callable, Dict, Iterable, Mapping, Tuple
? +++++++++++++++++++
from typing_extensions import TypeAlias
from werkzeug.wrappers import Request
# (name, value)
Header: TypeAlias = Tuple[str, str]
WSGIEnvironment: TypeAlias = Dict[str, Any]
# (body) -> None
StartResponseReturnType: TypeAlias = Callable[[bytes], object]
# (status: str, headers: List[Headers], exc_info) -> response
StartResponse: TypeAlias = Callable[..., StartResponseReturnType]
WSGIResponse: TypeAlias = Iterable[bytes]
WSGIApplication: TypeAlias = Callable[
[WSGIEnvironment, StartResponse], WSGIResponse
]
# (method, path, callback)
RouteDescription: TypeAlias = Tuple[str, str, WSGIApplication]
# (request, previous_args, path_part) -> result
RouteTemplateHandler: TypeAlias = Callable[
[Request, Tuple[Any, ...], str], Any
]
BadArgumentsDict: TypeAlias = Mapping[str, str] |
06c5f27c04de9fa62f6ac4834e0a920349c27084 | rules/binutils.py | rules/binutils.py | import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
rules = Binutils()
| import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
# For now we strip the man pages.
# man pages created on different systems are (for no good reason) different!
man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
shutil.rmtree(man_dir)
rules = Binutils()
| Remove man pages post-install (for now) | Remove man pages post-install (for now)
| Python | mit | BreakawayConsulting/xyz | import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
+ # For now we strip the man pages.
+ # man pages created on different systems are (for no good reason) different!
+ man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
+ shutil.rmtree(man_dir)
rules = Binutils()
| Remove man pages post-install (for now) | ## Code Before:
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
rules = Binutils()
## Instruction:
Remove man pages post-install (for now)
## Code After:
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
# For now we strip the man pages.
# man pages created on different systems are (for no good reason) different!
man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
shutil.rmtree(man_dir)
rules = Binutils()
| import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
+ # For now we strip the man pages.
+ # man pages created on different systems are (for no good reason) different!
+ man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
+ shutil.rmtree(man_dir)
rules = Binutils() |
7bd3b5a9f1664bd153c934d75f245a605aadee2a | ovp_core/models/cause.py | ovp_core/models/cause.py | from django.db import models
class Cause(models.Model):
name = models.CharField('name', max_length=100)
def __str__(self):
return self.name
class Meta:
app_label = 'ovp_core'
verbose_name = 'cause'
verbose_name_plural = 'causes'
| import vinaigrette
from django.db import models
class Cause(models.Model):
name = models.CharField('name', max_length=100)
def __str__(self):
return self.name
class Meta:
app_label = 'ovp_core'
verbose_name = 'cause'
verbose_name_plural = 'causes'
vinaigrette.register(Cause, ['name'])
| Apply django-vinaigrette to Cause model | Apply django-vinaigrette to Cause model
| Python | agpl-3.0 | OpenVolunteeringPlatform/django-ovp-core,OpenVolunteeringPlatform/django-ovp-core | + import vinaigrette
from django.db import models
class Cause(models.Model):
name = models.CharField('name', max_length=100)
def __str__(self):
return self.name
class Meta:
app_label = 'ovp_core'
verbose_name = 'cause'
verbose_name_plural = 'causes'
+ vinaigrette.register(Cause, ['name'])
+ | Apply django-vinaigrette to Cause model | ## Code Before:
from django.db import models
class Cause(models.Model):
name = models.CharField('name', max_length=100)
def __str__(self):
return self.name
class Meta:
app_label = 'ovp_core'
verbose_name = 'cause'
verbose_name_plural = 'causes'
## Instruction:
Apply django-vinaigrette to Cause model
## Code After:
import vinaigrette
from django.db import models
class Cause(models.Model):
name = models.CharField('name', max_length=100)
def __str__(self):
return self.name
class Meta:
app_label = 'ovp_core'
verbose_name = 'cause'
verbose_name_plural = 'causes'
vinaigrette.register(Cause, ['name'])
| + import vinaigrette
from django.db import models
class Cause(models.Model):
name = models.CharField('name', max_length=100)
def __str__(self):
return self.name
class Meta:
app_label = 'ovp_core'
verbose_name = 'cause'
verbose_name_plural = 'causes'
+
+ vinaigrette.register(Cause, ['name']) |
d5b5421c95b1e2feb4646a42b5aca71a2280e30c | tests/dojo_test.py | tests/dojo_test.py | import unittest
from src.dojo import Dojo
class TestCreateRoom (unittest.TestCase):
def test_create_room_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
blue_office = my_class_instance.create_room("office", "Blue")
self.assertTrue(blue_office)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 1)
def test_create_rooms_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
offices = my_class_instance.create_room("office", "Blue", "Black", "Brown")
self.assertTrue(offices)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 3) | import unittest
from src.dojo import Dojo
class TestCreateRoom (unittest.TestCase):
def test_create_room_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
blue_office = my_class_instance.create_room("office", "Blue")
self.assertTrue(blue_office)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 1)
def test_create_rooms_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
offices = my_class_instance.create_room("office", "Blue", "Black", "Brown")
self.assertTrue(offices)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 3)
def test_person_added_to_system(self):
initial_person_count = len(self.dojo.all_people)
person = self.dojo.add_person("Neil", "Armstrong", "Staff")
self.assertTrue(person)
new_person_count = len(self.dojo.all_people)
self.assertEqual(new_person_count - initial_person_count, 1) | Create test to check that a person has been added | Create test to check that a person has been added
| Python | mit | EdwinKato/Space-Allocator,EdwinKato/Space-Allocator | import unittest
from src.dojo import Dojo
class TestCreateRoom (unittest.TestCase):
def test_create_room_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
blue_office = my_class_instance.create_room("office", "Blue")
self.assertTrue(blue_office)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 1)
def test_create_rooms_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
offices = my_class_instance.create_room("office", "Blue", "Black", "Brown")
self.assertTrue(offices)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 3)
+
+ def test_person_added_to_system(self):
+ initial_person_count = len(self.dojo.all_people)
+ person = self.dojo.add_person("Neil", "Armstrong", "Staff")
+ self.assertTrue(person)
+ new_person_count = len(self.dojo.all_people)
+ self.assertEqual(new_person_count - initial_person_count, 1) | Create test to check that a person has been added | ## Code Before:
import unittest
from src.dojo import Dojo
class TestCreateRoom (unittest.TestCase):
def test_create_room_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
blue_office = my_class_instance.create_room("office", "Blue")
self.assertTrue(blue_office)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 1)
def test_create_rooms_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
offices = my_class_instance.create_room("office", "Blue", "Black", "Brown")
self.assertTrue(offices)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 3)
## Instruction:
Create test to check that a person has been added
## Code After:
import unittest
from src.dojo import Dojo
class TestCreateRoom (unittest.TestCase):
def test_create_room_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
blue_office = my_class_instance.create_room("office", "Blue")
self.assertTrue(blue_office)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 1)
def test_create_rooms_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
offices = my_class_instance.create_room("office", "Blue", "Black", "Brown")
self.assertTrue(offices)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 3)
def test_person_added_to_system(self):
initial_person_count = len(self.dojo.all_people)
person = self.dojo.add_person("Neil", "Armstrong", "Staff")
self.assertTrue(person)
new_person_count = len(self.dojo.all_people)
self.assertEqual(new_person_count - initial_person_count, 1) | import unittest
from src.dojo import Dojo
class TestCreateRoom (unittest.TestCase):
def test_create_room_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
blue_office = my_class_instance.create_room("office", "Blue")
self.assertTrue(blue_office)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 1)
def test_create_rooms_successfully(self):
my_class_instance = Dojo()
initial_room_count = len(my_class_instance.all_rooms)
offices = my_class_instance.create_room("office", "Blue", "Black", "Brown")
self.assertTrue(offices)
new_room_count = len(my_class_instance.all_rooms)
self.assertEqual(new_room_count - initial_room_count, 3)
+
+ def test_person_added_to_system(self):
+ initial_person_count = len(self.dojo.all_people)
+ person = self.dojo.add_person("Neil", "Armstrong", "Staff")
+ self.assertTrue(person)
+ new_person_count = len(self.dojo.all_people)
+ self.assertEqual(new_person_count - initial_person_count, 1) |
2685b94838c8ec7ce31da60bc6f28953152c788a | pixelmap/pixelmap.py | pixelmap/pixelmap.py |
from pixel import Pixel
class Pixelmap:
def __init__(self, width, height):
"""Pixelmap constructor
:param width: Width of map in pixels.
:param height: Height of map in pixels.
"""
self.width = width
self.height = height
self.map_matrix = [[0]*self.width for _ in range(self.height)]
for row in range(self.height):
for col in range(self.width):
self.map_matrix[row][col] = Pixel()
def __str__(self):
"""Human readable pixelmap description.
Pretty much just the matrix.
:return: Description of pixelmap.
"""
return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix]))
def __repr__(self):
"""Internal representation
Just use str for now.
"""
return self.__str__()
|
from .pixel import Pixel
class Pixelmap:
def __init__(self, cols, rows, default_val=None):
"""Pixelmap constructor
:param cols: Width of map in pixels.
:param rows: Height of map in pixels.
:param default_val: Default value for pixels.
"""
assert cols >= 0, 'Invalid Pixelmap width'
assert rows >= 0, 'Invalid Pixelmap height'
self.cols = cols
self.rows = rows
self.map_matrix = [[0]*self.cols for _ in range(self.rows)]
for row in range(self.rows):
for col in range(self.cols):
self.map_matrix[row][col] = Pixel(default_val)
def num_cols(self):
return self.cols
def num_rows(self):
return self.rows
def __str__(self):
"""Human readable pixelmap description.
Pretty much just the matrix.
:return: Description of pixelmap.
"""
return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix]))
| Add default value for matrix and methods to get columns and rows. | Add default value for matrix and methods to get columns and rows.
| Python | mit | yebra06/pixelmap |
- from pixel import Pixel
+ from .pixel import Pixel
class Pixelmap:
- def __init__(self, width, height):
+ def __init__(self, cols, rows, default_val=None):
"""Pixelmap constructor
- :param width: Width of map in pixels.
+ :param cols: Width of map in pixels.
- :param height: Height of map in pixels.
+ :param rows: Height of map in pixels.
+ :param default_val: Default value for pixels.
"""
- self.width = width
- self.height = height
+ assert cols >= 0, 'Invalid Pixelmap width'
+ assert rows >= 0, 'Invalid Pixelmap height'
+ self.cols = cols
+ self.rows = rows
- self.map_matrix = [[0]*self.width for _ in range(self.height)]
+ self.map_matrix = [[0]*self.cols for _ in range(self.rows)]
- for row in range(self.height):
+ for row in range(self.rows):
- for col in range(self.width):
+ for col in range(self.cols):
- self.map_matrix[row][col] = Pixel()
+ self.map_matrix[row][col] = Pixel(default_val)
+
+ def num_cols(self):
+ return self.cols
+
+ def num_rows(self):
+ return self.rows
def __str__(self):
"""Human readable pixelmap description.
Pretty much just the matrix.
:return: Description of pixelmap.
"""
- return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix]))
+ return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix]))
- def __repr__(self):
- """Internal representation
-
- Just use str for now.
- """
- return self.__str__()
- | Add default value for matrix and methods to get columns and rows. | ## Code Before:
from pixel import Pixel
class Pixelmap:
def __init__(self, width, height):
"""Pixelmap constructor
:param width: Width of map in pixels.
:param height: Height of map in pixels.
"""
self.width = width
self.height = height
self.map_matrix = [[0]*self.width for _ in range(self.height)]
for row in range(self.height):
for col in range(self.width):
self.map_matrix[row][col] = Pixel()
def __str__(self):
"""Human readable pixelmap description.
Pretty much just the matrix.
:return: Description of pixelmap.
"""
return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix]))
def __repr__(self):
"""Internal representation
Just use str for now.
"""
return self.__str__()
## Instruction:
Add default value for matrix and methods to get columns and rows.
## Code After:
from .pixel import Pixel
class Pixelmap:
def __init__(self, cols, rows, default_val=None):
"""Pixelmap constructor
:param cols: Width of map in pixels.
:param rows: Height of map in pixels.
:param default_val: Default value for pixels.
"""
assert cols >= 0, 'Invalid Pixelmap width'
assert rows >= 0, 'Invalid Pixelmap height'
self.cols = cols
self.rows = rows
self.map_matrix = [[0]*self.cols for _ in range(self.rows)]
for row in range(self.rows):
for col in range(self.cols):
self.map_matrix[row][col] = Pixel(default_val)
def num_cols(self):
return self.cols
def num_rows(self):
return self.rows
def __str__(self):
"""Human readable pixelmap description.
Pretty much just the matrix.
:return: Description of pixelmap.
"""
return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix]))
|
- from pixel import Pixel
+ from .pixel import Pixel
? +
class Pixelmap:
- def __init__(self, width, height):
+ def __init__(self, cols, rows, default_val=None):
"""Pixelmap constructor
- :param width: Width of map in pixels.
? ^^^^^
+ :param cols: Width of map in pixels.
? ^^^^
- :param height: Height of map in pixels.
? ^^^^^^
+ :param rows: Height of map in pixels.
? ^^^^
+ :param default_val: Default value for pixels.
"""
- self.width = width
- self.height = height
+ assert cols >= 0, 'Invalid Pixelmap width'
+ assert rows >= 0, 'Invalid Pixelmap height'
+ self.cols = cols
+ self.rows = rows
- self.map_matrix = [[0]*self.width for _ in range(self.height)]
? ^^^^^ ^^^^^^
+ self.map_matrix = [[0]*self.cols for _ in range(self.rows)]
? ^^^^ ^^^^
- for row in range(self.height):
? ^^^^^^
+ for row in range(self.rows):
? ^^^^
- for col in range(self.width):
? ^^^^^
+ for col in range(self.cols):
? ^^^^
- self.map_matrix[row][col] = Pixel()
+ self.map_matrix[row][col] = Pixel(default_val)
? +++++++++++
+
+ def num_cols(self):
+ return self.cols
+
+ def num_rows(self):
+ return self.rows
def __str__(self):
"""Human readable pixelmap description.
Pretty much just the matrix.
:return: Description of pixelmap.
"""
- return str('\n'.join([''.join(['{:4}'.format(item) for item in row]) for row in self.map_matrix]))
? ^
+ return str('\n'.join([''.join(['{:6}'.format(str(item)) for item in row]) for row in self.map_matrix]))
? ^ ++++ +
-
- def __repr__(self):
- """Internal representation
-
- Just use str for now.
- """
- return self.__str__() |
59536a70ef39e34a5aea57131492a475e05cd227 | lg_cms_director/setup.py | lg_cms_director/setup.py |
from distutils.core import setup
from catkin_pkg.packages import find_packages
from catkin_pkg.python_setup import generate_distutils_setup
d = generate_distutils_setup(
packages=['trollius', 'pulsar'],
package_dir={'': 'src'},
scripts=[],
requires=[]
)
setup(**d)
# vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
|
from distutils.core import setup
from catkin_pkg.packages import find_packages
from catkin_pkg.python_setup import generate_distutils_setup
d = generate_distutils_setup(
packages=find_packages('src'),
package_dir={'': 'src'},
scripts=[],
requires=[]
)
setup(**d)
# vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
| Revert to previous packaging of director's dependencies thx to @mvollrath | Revert to previous packaging of director's dependencies thx to @mvollrath
| Python | apache-2.0 | EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes,EndPointCorp/lg_ros_nodes |
from distutils.core import setup
from catkin_pkg.packages import find_packages
from catkin_pkg.python_setup import generate_distutils_setup
d = generate_distutils_setup(
- packages=['trollius', 'pulsar'],
+ packages=find_packages('src'),
package_dir={'': 'src'},
scripts=[],
requires=[]
)
setup(**d)
# vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
| Revert to previous packaging of director's dependencies thx to @mvollrath | ## Code Before:
from distutils.core import setup
from catkin_pkg.packages import find_packages
from catkin_pkg.python_setup import generate_distutils_setup
d = generate_distutils_setup(
packages=['trollius', 'pulsar'],
package_dir={'': 'src'},
scripts=[],
requires=[]
)
setup(**d)
# vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
## Instruction:
Revert to previous packaging of director's dependencies thx to @mvollrath
## Code After:
from distutils.core import setup
from catkin_pkg.packages import find_packages
from catkin_pkg.python_setup import generate_distutils_setup
d = generate_distutils_setup(
packages=find_packages('src'),
package_dir={'': 'src'},
scripts=[],
requires=[]
)
setup(**d)
# vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
|
from distutils.core import setup
from catkin_pkg.packages import find_packages
from catkin_pkg.python_setup import generate_distutils_setup
d = generate_distutils_setup(
- packages=['trollius', 'pulsar'],
+ packages=find_packages('src'),
package_dir={'': 'src'},
scripts=[],
requires=[]
)
setup(**d)
# vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4 |
8b5ccf93fbac8929ecfc185d7407a79b1e890bde | project_template/project_settings.py | project_template/project_settings.py |
from icekit.project.settings.icekit import * # icekit, glamkit
# Override the default ICEkit settings to form project settings.
|
from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
| Use GLAMkit settings in default project template. | Use GLAMkit settings in default project template.
| Python | mit | ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit |
- from icekit.project.settings.icekit import * # icekit, glamkit
+ from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
| Use GLAMkit settings in default project template. | ## Code Before:
from icekit.project.settings.icekit import * # icekit, glamkit
# Override the default ICEkit settings to form project settings.
## Instruction:
Use GLAMkit settings in default project template.
## Code After:
from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
|
- from icekit.project.settings.icekit import * # icekit, glamkit
? ^^^ --------
+ from icekit.project.settings.glamkit import * # glamkit, icekit
? ^^^^ ++++++++
# Override the default ICEkit settings to form project settings. |
326bebb58242981ec66f257525e5c5f58fae9196 | example/article/admin.py | example/article/admin.py | from django.contrib import admin
from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget
from parler.admin import TranslatableAdmin
from .models import Article
from parler.forms import TranslatableModelForm, TranslatedField
class ArticleAdminForm(TranslatableModelForm):
"""
Example form
Translated fields can be enhanced by manually declaring them:
"""
title = TranslatedField(widget=AdminTextInputWidget)
content = TranslatedField(widget=AdminTextareaWidget)
class ArticleAdmin(TranslatableAdmin):
"""
Example admin.
Using an empty class would already work,
but this example shows some additional options.
"""
# The 'language_column' is provided by the base class:
list_display = ('title', 'language_column')
# Example custom form usage.
form = ArticleAdminForm
# NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets=
fieldsets = (
(None, {
'fields': ('title', 'slug', 'published'),
}),
("Contents", {
'fields': ('content',),
})
)
def get_prepopulated_fields(self, request, obj=None):
# Can't use prepopulated_fields= yet, but this is a workaround.
return {'slug': ('title',)}
admin.site.register(Article, ArticleAdmin)
| from django.contrib import admin
from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget
from parler.admin import TranslatableAdmin
from .models import Article
from parler.forms import TranslatableModelForm, TranslatedField
class ArticleAdminForm(TranslatableModelForm):
"""
Example form
Translated fields can be enhanced by manually declaring them:
"""
title = TranslatedField(widget=AdminTextInputWidget)
content = TranslatedField(widget=AdminTextareaWidget)
class ArticleAdmin(TranslatableAdmin):
"""
Example admin.
Using an empty class would already work,
but this example shows some additional options.
"""
# The 'language_column' is provided by the base class:
list_display = ('title', 'language_column')
list_filter = ('published',)
# Example custom form usage.
form = ArticleAdminForm
# NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets=
fieldsets = (
(None, {
'fields': ('title', 'slug', 'published'),
}),
("Contents", {
'fields': ('content',),
})
)
def get_prepopulated_fields(self, request, obj=None):
# Can't use prepopulated_fields= yet, but this is a workaround.
return {'slug': ('title',)}
admin.site.register(Article, ArticleAdmin)
| Add list_filter to example ArticleAdmin | Add list_filter to example ArticleAdmin
When using a list filter and then adding or editing an object the
language GET parameter goes missing causing the wrong translation to be
edited.
| Python | apache-2.0 | django-parler/django-parler,jrief/django-parler,edoburu/django-parler,HiddenData/django-parler,skirsdeda/django-parler,edoburu/django-parler,zhangguiyu/django-parler,jrief/django-parler,django-parler/django-parler,imposeren/django-parler,defivelo/django-parler,imposeren/django-parler,zhangguiyu/django-parler,defivelo/django-parler,ellmetha/django-parler,HiddenData/django-parler,skirsdeda/django-parler,ellmetha/django-parler | from django.contrib import admin
from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget
from parler.admin import TranslatableAdmin
from .models import Article
from parler.forms import TranslatableModelForm, TranslatedField
class ArticleAdminForm(TranslatableModelForm):
"""
Example form
Translated fields can be enhanced by manually declaring them:
"""
title = TranslatedField(widget=AdminTextInputWidget)
content = TranslatedField(widget=AdminTextareaWidget)
class ArticleAdmin(TranslatableAdmin):
"""
Example admin.
Using an empty class would already work,
but this example shows some additional options.
"""
# The 'language_column' is provided by the base class:
list_display = ('title', 'language_column')
+ list_filter = ('published',)
# Example custom form usage.
form = ArticleAdminForm
# NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets=
fieldsets = (
(None, {
'fields': ('title', 'slug', 'published'),
}),
("Contents", {
'fields': ('content',),
})
)
def get_prepopulated_fields(self, request, obj=None):
# Can't use prepopulated_fields= yet, but this is a workaround.
return {'slug': ('title',)}
admin.site.register(Article, ArticleAdmin)
| Add list_filter to example ArticleAdmin | ## Code Before:
from django.contrib import admin
from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget
from parler.admin import TranslatableAdmin
from .models import Article
from parler.forms import TranslatableModelForm, TranslatedField
class ArticleAdminForm(TranslatableModelForm):
"""
Example form
Translated fields can be enhanced by manually declaring them:
"""
title = TranslatedField(widget=AdminTextInputWidget)
content = TranslatedField(widget=AdminTextareaWidget)
class ArticleAdmin(TranslatableAdmin):
"""
Example admin.
Using an empty class would already work,
but this example shows some additional options.
"""
# The 'language_column' is provided by the base class:
list_display = ('title', 'language_column')
# Example custom form usage.
form = ArticleAdminForm
# NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets=
fieldsets = (
(None, {
'fields': ('title', 'slug', 'published'),
}),
("Contents", {
'fields': ('content',),
})
)
def get_prepopulated_fields(self, request, obj=None):
# Can't use prepopulated_fields= yet, but this is a workaround.
return {'slug': ('title',)}
admin.site.register(Article, ArticleAdmin)
## Instruction:
Add list_filter to example ArticleAdmin
## Code After:
from django.contrib import admin
from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget
from parler.admin import TranslatableAdmin
from .models import Article
from parler.forms import TranslatableModelForm, TranslatedField
class ArticleAdminForm(TranslatableModelForm):
"""
Example form
Translated fields can be enhanced by manually declaring them:
"""
title = TranslatedField(widget=AdminTextInputWidget)
content = TranslatedField(widget=AdminTextareaWidget)
class ArticleAdmin(TranslatableAdmin):
"""
Example admin.
Using an empty class would already work,
but this example shows some additional options.
"""
# The 'language_column' is provided by the base class:
list_display = ('title', 'language_column')
list_filter = ('published',)
# Example custom form usage.
form = ArticleAdminForm
# NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets=
fieldsets = (
(None, {
'fields': ('title', 'slug', 'published'),
}),
("Contents", {
'fields': ('content',),
})
)
def get_prepopulated_fields(self, request, obj=None):
# Can't use prepopulated_fields= yet, but this is a workaround.
return {'slug': ('title',)}
admin.site.register(Article, ArticleAdmin)
| from django.contrib import admin
from django.contrib.admin.widgets import AdminTextInputWidget, AdminTextareaWidget
from parler.admin import TranslatableAdmin
from .models import Article
from parler.forms import TranslatableModelForm, TranslatedField
class ArticleAdminForm(TranslatableModelForm):
"""
Example form
Translated fields can be enhanced by manually declaring them:
"""
title = TranslatedField(widget=AdminTextInputWidget)
content = TranslatedField(widget=AdminTextareaWidget)
class ArticleAdmin(TranslatableAdmin):
"""
Example admin.
Using an empty class would already work,
but this example shows some additional options.
"""
# The 'language_column' is provided by the base class:
list_display = ('title', 'language_column')
+ list_filter = ('published',)
# Example custom form usage.
form = ArticleAdminForm
# NOTE: when using Django 1.4, use declared_fieldsets= instead of fieldsets=
fieldsets = (
(None, {
'fields': ('title', 'slug', 'published'),
}),
("Contents", {
'fields': ('content',),
})
)
def get_prepopulated_fields(self, request, obj=None):
# Can't use prepopulated_fields= yet, but this is a workaround.
return {'slug': ('title',)}
admin.site.register(Article, ArticleAdmin) |
123875153e81253a44d0e8b2d8de5abee195362a | backend/shmitter/tweets/serializers.py | backend/shmitter/tweets/serializers.py | from rest_framework import serializers
from shmitter.likes import services as likes_services
from .models import Tweet
from . import services as tweets_services
class TweetSerializer(serializers.ModelSerializer):
owner = serializers.ReadOnlyField(source='owner.username')
is_fan = serializers.SerializerMethodField()
is_retweeted = serializers.SerializerMethodField()
class Meta:
model = Tweet
fields = (
'id',
'owner',
'body',
'is_fan',
'is_retweeted',
'total_likes',
'created',
)
def get_is_fan(self, obj) -> bool:
"""
Check if a `request.user` has liked this tweet (`obj`).
"""
user = self.context.get('request').user
return likes_services.is_fan(obj, user)
def get_is_retweeted(self, obj) -> bool:
"""
Check if a `request.user` has retweeted this tweet (`obj`).
"""
user = self.context.get('request').user
return tweets_services.is_retweeted(obj, user)
| from rest_framework import serializers
from shmitter.likes import services as likes_services
from .models import Tweet
from . import services as tweets_services
class TweetSerializer(serializers.ModelSerializer):
owner = serializers.ReadOnlyField(source='owner.username')
is_fan = serializers.SerializerMethodField()
is_retweeted = serializers.SerializerMethodField()
class Meta:
model = Tweet
fields = (
'id',
'owner',
'body',
'is_fan',
'is_retweeted',
'total_likes',
'total_retweets',
'created',
)
def get_is_fan(self, obj) -> bool:
"""
Check if a `request.user` has liked this tweet (`obj`).
"""
user = self.context.get('request').user
return likes_services.is_fan(obj, user)
def get_is_retweeted(self, obj) -> bool:
"""
Check if a `request.user` has retweeted this tweet (`obj`).
"""
user = self.context.get('request').user
return tweets_services.is_retweeted(obj, user)
| Add total retweets to the serializer | Add total retweets to the serializer
| Python | mit | apirobot/shmitter,apirobot/shmitter,apirobot/shmitter | from rest_framework import serializers
from shmitter.likes import services as likes_services
from .models import Tweet
from . import services as tweets_services
class TweetSerializer(serializers.ModelSerializer):
owner = serializers.ReadOnlyField(source='owner.username')
is_fan = serializers.SerializerMethodField()
is_retweeted = serializers.SerializerMethodField()
class Meta:
model = Tweet
fields = (
'id',
'owner',
'body',
'is_fan',
'is_retweeted',
'total_likes',
+ 'total_retweets',
'created',
)
def get_is_fan(self, obj) -> bool:
"""
Check if a `request.user` has liked this tweet (`obj`).
"""
user = self.context.get('request').user
return likes_services.is_fan(obj, user)
def get_is_retweeted(self, obj) -> bool:
"""
Check if a `request.user` has retweeted this tweet (`obj`).
"""
user = self.context.get('request').user
return tweets_services.is_retweeted(obj, user)
| Add total retweets to the serializer | ## Code Before:
from rest_framework import serializers
from shmitter.likes import services as likes_services
from .models import Tweet
from . import services as tweets_services
class TweetSerializer(serializers.ModelSerializer):
owner = serializers.ReadOnlyField(source='owner.username')
is_fan = serializers.SerializerMethodField()
is_retweeted = serializers.SerializerMethodField()
class Meta:
model = Tweet
fields = (
'id',
'owner',
'body',
'is_fan',
'is_retweeted',
'total_likes',
'created',
)
def get_is_fan(self, obj) -> bool:
"""
Check if a `request.user` has liked this tweet (`obj`).
"""
user = self.context.get('request').user
return likes_services.is_fan(obj, user)
def get_is_retweeted(self, obj) -> bool:
"""
Check if a `request.user` has retweeted this tweet (`obj`).
"""
user = self.context.get('request').user
return tweets_services.is_retweeted(obj, user)
## Instruction:
Add total retweets to the serializer
## Code After:
from rest_framework import serializers
from shmitter.likes import services as likes_services
from .models import Tweet
from . import services as tweets_services
class TweetSerializer(serializers.ModelSerializer):
owner = serializers.ReadOnlyField(source='owner.username')
is_fan = serializers.SerializerMethodField()
is_retweeted = serializers.SerializerMethodField()
class Meta:
model = Tweet
fields = (
'id',
'owner',
'body',
'is_fan',
'is_retweeted',
'total_likes',
'total_retweets',
'created',
)
def get_is_fan(self, obj) -> bool:
"""
Check if a `request.user` has liked this tweet (`obj`).
"""
user = self.context.get('request').user
return likes_services.is_fan(obj, user)
def get_is_retweeted(self, obj) -> bool:
"""
Check if a `request.user` has retweeted this tweet (`obj`).
"""
user = self.context.get('request').user
return tweets_services.is_retweeted(obj, user)
| from rest_framework import serializers
from shmitter.likes import services as likes_services
from .models import Tweet
from . import services as tweets_services
class TweetSerializer(serializers.ModelSerializer):
owner = serializers.ReadOnlyField(source='owner.username')
is_fan = serializers.SerializerMethodField()
is_retweeted = serializers.SerializerMethodField()
class Meta:
model = Tweet
fields = (
'id',
'owner',
'body',
'is_fan',
'is_retweeted',
'total_likes',
+ 'total_retweets',
'created',
)
def get_is_fan(self, obj) -> bool:
"""
Check if a `request.user` has liked this tweet (`obj`).
"""
user = self.context.get('request').user
return likes_services.is_fan(obj, user)
def get_is_retweeted(self, obj) -> bool:
"""
Check if a `request.user` has retweeted this tweet (`obj`).
"""
user = self.context.get('request').user
return tweets_services.is_retweeted(obj, user) |
0ee42ac3b80893557691d722eda207733289c97c | micropsi_core/world/minecraft/spockplugin.py | micropsi_core/world/minecraft/spockplugin.py | import logging
from spock.mcp import mcdata, mcpacket
from spock.mcmap import smpmap
from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher
from spock.utils import pl_announce
@pl_announce('Micropsi')
class MicropsiPlugin(object):
def __init__(self, ploader, settings):
self.worldadapter = settings['worldadapter']
self.worldadapter.spockplugin = self
self.net = ploader.requires('Net')
self.event = ploader.requires('Event')
self.world = ploader.requires('World')
self.clientinfo = ploader.requires('ClientInfo')
#MicroPsi Datatargets
self.psi_dispatcher = PsiDispatcher(self)
self.move_x = 0
self.move_z = 0
self.move_x_ = 0
self.move_z_ = 0
def move(self, position=None):
if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE):
return
if position is None:
position = self.client_info.position
self.net.push(mcpacket.Packet(
ident='PLAY>Player Position and Look',
data=position
)) | import logging
from spock.mcp import mcdata, mcpacket
from spock.mcmap import smpmap
from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher
from spock.utils import pl_announce
@pl_announce('Micropsi')
class MicropsiPlugin(object):
def __init__(self, ploader, settings):
self.worldadapter = settings['worldadapter']
self.worldadapter.spockplugin = self
self.net = ploader.requires('Net')
self.event = ploader.requires('Event')
self.world = ploader.requires('World')
self.clientinfo = ploader.requires('ClientInfo')
#MicroPsi Datatargets
self.psi_dispatcher = PsiDispatcher(self)
self.move_x = 0
self.move_z = 0
self.move_x_ = 0
self.move_z_ = 0
def move(self, position=None):
if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE):
return
self.clientinfo.position = position
| Move now sets the client position and allows the movement plugin do its thing | Move now sets the client position and allows the
movement plugin do its thing | Python | mit | ianupright/micropsi2,ianupright/micropsi2,printedheart/micropsi2,ianupright/micropsi2,printedheart/micropsi2,printedheart/micropsi2 | import logging
from spock.mcp import mcdata, mcpacket
from spock.mcmap import smpmap
from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher
from spock.utils import pl_announce
@pl_announce('Micropsi')
class MicropsiPlugin(object):
def __init__(self, ploader, settings):
self.worldadapter = settings['worldadapter']
self.worldadapter.spockplugin = self
self.net = ploader.requires('Net')
self.event = ploader.requires('Event')
self.world = ploader.requires('World')
self.clientinfo = ploader.requires('ClientInfo')
#MicroPsi Datatargets
self.psi_dispatcher = PsiDispatcher(self)
self.move_x = 0
self.move_z = 0
self.move_x_ = 0
self.move_z_ = 0
def move(self, position=None):
if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE):
return
+ self.clientinfo.position = position
+
- if position is None:
- position = self.client_info.position
- self.net.push(mcpacket.Packet(
- ident='PLAY>Player Position and Look',
- data=position
- )) | Move now sets the client position and allows the movement plugin do its thing | ## Code Before:
import logging
from spock.mcp import mcdata, mcpacket
from spock.mcmap import smpmap
from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher
from spock.utils import pl_announce
@pl_announce('Micropsi')
class MicropsiPlugin(object):
def __init__(self, ploader, settings):
self.worldadapter = settings['worldadapter']
self.worldadapter.spockplugin = self
self.net = ploader.requires('Net')
self.event = ploader.requires('Event')
self.world = ploader.requires('World')
self.clientinfo = ploader.requires('ClientInfo')
#MicroPsi Datatargets
self.psi_dispatcher = PsiDispatcher(self)
self.move_x = 0
self.move_z = 0
self.move_x_ = 0
self.move_z_ = 0
def move(self, position=None):
if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE):
return
if position is None:
position = self.client_info.position
self.net.push(mcpacket.Packet(
ident='PLAY>Player Position and Look',
data=position
))
## Instruction:
Move now sets the client position and allows the movement plugin do its thing
## Code After:
import logging
from spock.mcp import mcdata, mcpacket
from spock.mcmap import smpmap
from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher
from spock.utils import pl_announce
@pl_announce('Micropsi')
class MicropsiPlugin(object):
def __init__(self, ploader, settings):
self.worldadapter = settings['worldadapter']
self.worldadapter.spockplugin = self
self.net = ploader.requires('Net')
self.event = ploader.requires('Event')
self.world = ploader.requires('World')
self.clientinfo = ploader.requires('ClientInfo')
#MicroPsi Datatargets
self.psi_dispatcher = PsiDispatcher(self)
self.move_x = 0
self.move_z = 0
self.move_x_ = 0
self.move_z_ = 0
def move(self, position=None):
if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE):
return
self.clientinfo.position = position
| import logging
from spock.mcp import mcdata, mcpacket
from spock.mcmap import smpmap
from micropsi_core.world.minecraft.psidispatcher import PsiDispatcher
from spock.utils import pl_announce
@pl_announce('Micropsi')
class MicropsiPlugin(object):
def __init__(self, ploader, settings):
self.worldadapter = settings['worldadapter']
self.worldadapter.spockplugin = self
self.net = ploader.requires('Net')
self.event = ploader.requires('Event')
self.world = ploader.requires('World')
self.clientinfo = ploader.requires('ClientInfo')
#MicroPsi Datatargets
self.psi_dispatcher = PsiDispatcher(self)
self.move_x = 0
self.move_z = 0
self.move_x_ = 0
self.move_z_ = 0
def move(self, position=None):
if not (self.net.connected and self.net.proto_state == mcdata.PLAY_STATE):
return
+ self.clientinfo.position = position
- if position is None:
- position = self.client_info.position
- self.net.push(mcpacket.Packet(
- ident='PLAY>Player Position and Look',
- data=position
- )) |
cea1f24aa0862d2feab1150fbd667159ab4cbe3a | migrations/versions/0313_email_access_validated_at.py | migrations/versions/0313_email_access_validated_at.py | from alembic import op
import sqlalchemy as sa
revision = '0313_email_access_validated_at'
down_revision = '0312_populate_returned_letters'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True))
# if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date.
op.execute("""
UPDATE
users
SET
email_access_validated_at = created_at
""")
op.execute("""
UPDATE
users
SET
email_access_validated_at = logged_in_at
WHERE
auth_type = 'email_auth'
AND
logged_in_at IS NOT NULL
""")
op.alter_column('users', 'email_access_validated_at', nullable=False)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column('users', 'email_access_validated_at')
# ### end Alembic commands ###
| from alembic import op
import sqlalchemy as sa
revision = '0313_email_access_validated_at'
down_revision = '0312_populate_returned_letters'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True))
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column('users', 'email_access_validated_at')
# ### end Alembic commands ###
| Simplify the first migration, we will do execute statements later | Simplify the first migration, we will do execute statements later
| Python | mit | alphagov/notifications-api,alphagov/notifications-api | from alembic import op
import sqlalchemy as sa
revision = '0313_email_access_validated_at'
down_revision = '0312_populate_returned_letters'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True))
- # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date.
- op.execute("""
- UPDATE
- users
- SET
- email_access_validated_at = created_at
- """)
- op.execute("""
- UPDATE
- users
- SET
- email_access_validated_at = logged_in_at
- WHERE
- auth_type = 'email_auth'
- AND
- logged_in_at IS NOT NULL
- """)
- op.alter_column('users', 'email_access_validated_at', nullable=False)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column('users', 'email_access_validated_at')
# ### end Alembic commands ###
| Simplify the first migration, we will do execute statements later | ## Code Before:
from alembic import op
import sqlalchemy as sa
revision = '0313_email_access_validated_at'
down_revision = '0312_populate_returned_letters'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True))
# if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date.
op.execute("""
UPDATE
users
SET
email_access_validated_at = created_at
""")
op.execute("""
UPDATE
users
SET
email_access_validated_at = logged_in_at
WHERE
auth_type = 'email_auth'
AND
logged_in_at IS NOT NULL
""")
op.alter_column('users', 'email_access_validated_at', nullable=False)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column('users', 'email_access_validated_at')
# ### end Alembic commands ###
## Instruction:
Simplify the first migration, we will do execute statements later
## Code After:
from alembic import op
import sqlalchemy as sa
revision = '0313_email_access_validated_at'
down_revision = '0312_populate_returned_letters'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True))
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column('users', 'email_access_validated_at')
# ### end Alembic commands ###
| from alembic import op
import sqlalchemy as sa
revision = '0313_email_access_validated_at'
down_revision = '0312_populate_returned_letters'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True))
- # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date.
- op.execute("""
- UPDATE
- users
- SET
- email_access_validated_at = created_at
- """)
- op.execute("""
- UPDATE
- users
- SET
- email_access_validated_at = logged_in_at
- WHERE
- auth_type = 'email_auth'
- AND
- logged_in_at IS NOT NULL
- """)
- op.alter_column('users', 'email_access_validated_at', nullable=False)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column('users', 'email_access_validated_at')
# ### end Alembic commands ### |
6e1892daec726b44b1bbb4d085e27fa03c0a419b | server/kcaa/kcsapi/client_test.py | server/kcaa/kcsapi/client_test.py |
import pytest
import client
from kcaa import screens
class TestScreen(object):
def test_mission_result(self):
screen = client.Screen()
assert screen.screen == screens.UNKNOWN
screen.update('/api_get_member/deck_port', None, None, None, False)
assert screen.screen == screens.PORT
screen.update('/api_req_mission/result', None, None, None, False)
assert screen.screen == screens.MISSION_RESULT
screen.update('/api_get_member/deck_port', None, None, None, False)
assert screen.screen == screens.MISSION_RESULT
def main():
import doctest
doctest.testmod(client)
pytest.main(args=[__file__.replace('.pyc', '.py')])
if __name__ == '__main__':
main()
|
import pytest
import client
from kcaa import screens
class TestScreen(object):
def update(self, screen, api_name):
screen.update(api_name, None, None, None, False)
def update_sequence(self, screen, api_names):
for api_name in api_names:
screen.update(api_name, None, None, None, False)
def test_mission_result(self):
screen = client.Screen()
assert screen.screen == screens.UNKNOWN
self.update(screen, '/api_get_member/deck_port')
assert screen.screen == screens.PORT
self.update(screen, '/api_req_mission/result')
assert screen.screen == screens.MISSION_RESULT
self.update(screen, '/api_get_member/deck_port')
assert screen.screen == screens.MISSION_RESULT
def test_mission_result_real_sequence(self):
screen = client.Screen()
screen.screen = screens.PORT
self.update_sequence(screen, [
'/api_auth_member/logincheck',
'/api_get_member/material',
'/api_get_member/deck_port',
'/api_get_member/ndock',
'/api_get_member/ship3',
'/api_get_member/basic',
'/api_req_mission/result',
'/api_get_member/deck_port',
'/api_get_member/basic',
'/api_get_member/ship2',
'/api_get_member/material',
'/api_get_member/useitem',
])
assert screen.screen == screens.MISSION_RESULT
def main():
import doctest
doctest.testmod(client)
pytest.main(args=[__file__.replace('.pyc', '.py')])
if __name__ == '__main__':
main()
| Add a Screen test for sequence of KCSAPI responses. | Add a Screen test for sequence of KCSAPI responses.
| Python | apache-2.0 | kcaa/kcaa,kcaa/kcaa,kcaa/kcaa,kcaa/kcaa |
import pytest
import client
from kcaa import screens
class TestScreen(object):
+ def update(self, screen, api_name):
+ screen.update(api_name, None, None, None, False)
+
+ def update_sequence(self, screen, api_names):
+ for api_name in api_names:
+ screen.update(api_name, None, None, None, False)
+
def test_mission_result(self):
screen = client.Screen()
assert screen.screen == screens.UNKNOWN
- screen.update('/api_get_member/deck_port', None, None, None, False)
+ self.update(screen, '/api_get_member/deck_port')
assert screen.screen == screens.PORT
- screen.update('/api_req_mission/result', None, None, None, False)
+ self.update(screen, '/api_req_mission/result')
assert screen.screen == screens.MISSION_RESULT
- screen.update('/api_get_member/deck_port', None, None, None, False)
+ self.update(screen, '/api_get_member/deck_port')
+ assert screen.screen == screens.MISSION_RESULT
+
+ def test_mission_result_real_sequence(self):
+ screen = client.Screen()
+ screen.screen = screens.PORT
+ self.update_sequence(screen, [
+ '/api_auth_member/logincheck',
+ '/api_get_member/material',
+ '/api_get_member/deck_port',
+ '/api_get_member/ndock',
+ '/api_get_member/ship3',
+ '/api_get_member/basic',
+ '/api_req_mission/result',
+ '/api_get_member/deck_port',
+ '/api_get_member/basic',
+ '/api_get_member/ship2',
+ '/api_get_member/material',
+ '/api_get_member/useitem',
+ ])
assert screen.screen == screens.MISSION_RESULT
def main():
import doctest
doctest.testmod(client)
pytest.main(args=[__file__.replace('.pyc', '.py')])
if __name__ == '__main__':
main()
| Add a Screen test for sequence of KCSAPI responses. | ## Code Before:
import pytest
import client
from kcaa import screens
class TestScreen(object):
def test_mission_result(self):
screen = client.Screen()
assert screen.screen == screens.UNKNOWN
screen.update('/api_get_member/deck_port', None, None, None, False)
assert screen.screen == screens.PORT
screen.update('/api_req_mission/result', None, None, None, False)
assert screen.screen == screens.MISSION_RESULT
screen.update('/api_get_member/deck_port', None, None, None, False)
assert screen.screen == screens.MISSION_RESULT
def main():
import doctest
doctest.testmod(client)
pytest.main(args=[__file__.replace('.pyc', '.py')])
if __name__ == '__main__':
main()
## Instruction:
Add a Screen test for sequence of KCSAPI responses.
## Code After:
import pytest
import client
from kcaa import screens
class TestScreen(object):
def update(self, screen, api_name):
screen.update(api_name, None, None, None, False)
def update_sequence(self, screen, api_names):
for api_name in api_names:
screen.update(api_name, None, None, None, False)
def test_mission_result(self):
screen = client.Screen()
assert screen.screen == screens.UNKNOWN
self.update(screen, '/api_get_member/deck_port')
assert screen.screen == screens.PORT
self.update(screen, '/api_req_mission/result')
assert screen.screen == screens.MISSION_RESULT
self.update(screen, '/api_get_member/deck_port')
assert screen.screen == screens.MISSION_RESULT
def test_mission_result_real_sequence(self):
screen = client.Screen()
screen.screen = screens.PORT
self.update_sequence(screen, [
'/api_auth_member/logincheck',
'/api_get_member/material',
'/api_get_member/deck_port',
'/api_get_member/ndock',
'/api_get_member/ship3',
'/api_get_member/basic',
'/api_req_mission/result',
'/api_get_member/deck_port',
'/api_get_member/basic',
'/api_get_member/ship2',
'/api_get_member/material',
'/api_get_member/useitem',
])
assert screen.screen == screens.MISSION_RESULT
def main():
import doctest
doctest.testmod(client)
pytest.main(args=[__file__.replace('.pyc', '.py')])
if __name__ == '__main__':
main()
|
import pytest
import client
from kcaa import screens
class TestScreen(object):
+ def update(self, screen, api_name):
+ screen.update(api_name, None, None, None, False)
+
+ def update_sequence(self, screen, api_names):
+ for api_name in api_names:
+ screen.update(api_name, None, None, None, False)
+
def test_mission_result(self):
screen = client.Screen()
assert screen.screen == screens.UNKNOWN
- screen.update('/api_get_member/deck_port', None, None, None, False)
+ self.update(screen, '/api_get_member/deck_port')
assert screen.screen == screens.PORT
- screen.update('/api_req_mission/result', None, None, None, False)
+ self.update(screen, '/api_req_mission/result')
assert screen.screen == screens.MISSION_RESULT
- screen.update('/api_get_member/deck_port', None, None, None, False)
+ self.update(screen, '/api_get_member/deck_port')
+ assert screen.screen == screens.MISSION_RESULT
+
+ def test_mission_result_real_sequence(self):
+ screen = client.Screen()
+ screen.screen = screens.PORT
+ self.update_sequence(screen, [
+ '/api_auth_member/logincheck',
+ '/api_get_member/material',
+ '/api_get_member/deck_port',
+ '/api_get_member/ndock',
+ '/api_get_member/ship3',
+ '/api_get_member/basic',
+ '/api_req_mission/result',
+ '/api_get_member/deck_port',
+ '/api_get_member/basic',
+ '/api_get_member/ship2',
+ '/api_get_member/material',
+ '/api_get_member/useitem',
+ ])
assert screen.screen == screens.MISSION_RESULT
def main():
import doctest
doctest.testmod(client)
pytest.main(args=[__file__.replace('.pyc', '.py')])
if __name__ == '__main__':
main() |
3ae6c0f4c4f13207386dbf0fa2004655e9f2c8d6 | UM/View/CompositePass.py | UM/View/CompositePass.py |
from UM.Application import Application
from UM.Resources import Resources
from UM.Math.Matrix import Matrix
from UM.View.RenderPass import RenderPass
from UM.View.GL.OpenGL import OpenGL
class CompositePass(RenderPass):
def __init__(self, width, height):
super().__init__("composite", width, height)
self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader"))
self._gl = OpenGL.getInstance().getBindingsObject()
self._renderer = Application.getInstance().getRenderer()
def setCompositeShader(self, shader):
self._shader = shader
def renderContents(self):
pass
def renderOutput(self):
self._shader.bind()
texture_unit = 0
for render_pass in self._renderer.getRenderPasses():
self._gl.glActiveTexture(texture_unit)
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId())
texture_unit += 1
self._renderer.renderQuad(self._shader)
for i in range(texture_unit):
self._gl.glActiveTexture(texture_unit)
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0)
self._shader.release()
|
from UM.Application import Application
from UM.Resources import Resources
from UM.Math.Matrix import Matrix
from UM.View.RenderPass import RenderPass
from UM.View.GL.OpenGL import OpenGL
class CompositePass(RenderPass):
def __init__(self, width, height):
super().__init__("composite", width, height, 999)
self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader"))
self._gl = OpenGL.getInstance().getBindingsObject()
self._renderer = Application.getInstance().getRenderer()
self._layer_bindings = [ "default", "selection" ]
def setCompositeShader(self, shader):
self._shader = shader
def setLayerBindings(self, bindings):
self._layer_bindings = bindings
def render(self):
self._shader.bind()
texture_unit = 0
for binding in self._layer_bindings:
render_pass = self._renderer.getRenderPass(binding)
if not render_pass:
continue
self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId())
texture_unit += 1
self._renderer.renderFullScreenQuad(self._shader)
for i in range(texture_unit):
self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0)
self._shader.release()
| Add explicit render layer binding instead of assuming all render passes can be used for compositing | Add explicit render layer binding instead of assuming all render passes can be used for compositing
| Python | agpl-3.0 | onitake/Uranium,onitake/Uranium |
from UM.Application import Application
from UM.Resources import Resources
from UM.Math.Matrix import Matrix
from UM.View.RenderPass import RenderPass
from UM.View.GL.OpenGL import OpenGL
class CompositePass(RenderPass):
def __init__(self, width, height):
- super().__init__("composite", width, height)
+ super().__init__("composite", width, height, 999)
self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader"))
self._gl = OpenGL.getInstance().getBindingsObject()
self._renderer = Application.getInstance().getRenderer()
+ self._layer_bindings = [ "default", "selection" ]
+
def setCompositeShader(self, shader):
self._shader = shader
- def renderContents(self):
- pass
+ def setLayerBindings(self, bindings):
+ self._layer_bindings = bindings
- def renderOutput(self):
+ def render(self):
self._shader.bind()
texture_unit = 0
+ for binding in self._layer_bindings:
- for render_pass in self._renderer.getRenderPasses():
+ render_pass = self._renderer.getRenderPass(binding)
- self._gl.glActiveTexture(texture_unit)
+ if not render_pass:
+ continue
+
+ self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId())
texture_unit += 1
- self._renderer.renderQuad(self._shader)
+ self._renderer.renderFullScreenQuad(self._shader)
for i in range(texture_unit):
- self._gl.glActiveTexture(texture_unit)
+ self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0)
self._shader.release()
| Add explicit render layer binding instead of assuming all render passes can be used for compositing | ## Code Before:
from UM.Application import Application
from UM.Resources import Resources
from UM.Math.Matrix import Matrix
from UM.View.RenderPass import RenderPass
from UM.View.GL.OpenGL import OpenGL
class CompositePass(RenderPass):
def __init__(self, width, height):
super().__init__("composite", width, height)
self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader"))
self._gl = OpenGL.getInstance().getBindingsObject()
self._renderer = Application.getInstance().getRenderer()
def setCompositeShader(self, shader):
self._shader = shader
def renderContents(self):
pass
def renderOutput(self):
self._shader.bind()
texture_unit = 0
for render_pass in self._renderer.getRenderPasses():
self._gl.glActiveTexture(texture_unit)
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId())
texture_unit += 1
self._renderer.renderQuad(self._shader)
for i in range(texture_unit):
self._gl.glActiveTexture(texture_unit)
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0)
self._shader.release()
## Instruction:
Add explicit render layer binding instead of assuming all render passes can be used for compositing
## Code After:
from UM.Application import Application
from UM.Resources import Resources
from UM.Math.Matrix import Matrix
from UM.View.RenderPass import RenderPass
from UM.View.GL.OpenGL import OpenGL
class CompositePass(RenderPass):
def __init__(self, width, height):
super().__init__("composite", width, height, 999)
self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader"))
self._gl = OpenGL.getInstance().getBindingsObject()
self._renderer = Application.getInstance().getRenderer()
self._layer_bindings = [ "default", "selection" ]
def setCompositeShader(self, shader):
self._shader = shader
def setLayerBindings(self, bindings):
self._layer_bindings = bindings
def render(self):
self._shader.bind()
texture_unit = 0
for binding in self._layer_bindings:
render_pass = self._renderer.getRenderPass(binding)
if not render_pass:
continue
self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId())
texture_unit += 1
self._renderer.renderFullScreenQuad(self._shader)
for i in range(texture_unit):
self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0)
self._shader.release()
|
from UM.Application import Application
from UM.Resources import Resources
from UM.Math.Matrix import Matrix
from UM.View.RenderPass import RenderPass
from UM.View.GL.OpenGL import OpenGL
class CompositePass(RenderPass):
def __init__(self, width, height):
- super().__init__("composite", width, height)
+ super().__init__("composite", width, height, 999)
? +++++
self._shader = OpenGL.getInstance().createShaderProgram(Resources.getPath(Resources.Shaders, "composite.shader"))
self._gl = OpenGL.getInstance().getBindingsObject()
self._renderer = Application.getInstance().getRenderer()
+ self._layer_bindings = [ "default", "selection" ]
+
def setCompositeShader(self, shader):
self._shader = shader
- def renderContents(self):
- pass
+ def setLayerBindings(self, bindings):
+ self._layer_bindings = bindings
- def renderOutput(self):
? ------
+ def render(self):
self._shader.bind()
texture_unit = 0
+ for binding in self._layer_bindings:
- for render_pass in self._renderer.getRenderPasses():
? ^^^ ^^ -- -
+ render_pass = self._renderer.getRenderPass(binding)
? ^^^ ^ +++++++
- self._gl.glActiveTexture(texture_unit)
+ if not render_pass:
+ continue
+
+ self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(texture_unit)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, render_pass.getTextureId())
texture_unit += 1
- self._renderer.renderQuad(self._shader)
+ self._renderer.renderFullScreenQuad(self._shader)
? ++++++++++
for i in range(texture_unit):
- self._gl.glActiveTexture(texture_unit)
+ self._gl.glActiveTexture(getattr(self._gl, "GL_TEXTURE{0}".format(i)))
self._gl.glBindTexture(self._gl.GL_TEXTURE_2D, 0)
self._shader.release() |
4d1444e2f2a455e691342a82f0e116e210593411 | s01/c01.py | s01/c01.py | """Set 01 - Challenge 01."""
import base64
hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f'
'69736f6e6f7573206d757368726f6f6d')
b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
def hex2b64(hex_string):
"""Convert a hex string into a base64 encoded byte string."""
hex_data = bytearray.fromhex(hex_string)
# Strip trailing newline
return base64.encodebytes(hex_data)[:-1]
assert hex2b64(hex_string) == b64_string
| """Set 01 - Challenge 01."""
import binascii
hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f'
'69736f6e6f7573206d757368726f6f6d')
b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
def hex2b64(hex_string):
"""Convert a hex string into a base64 encoded string."""
return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip()
assert hex2b64(hex_string) == b64_string
| Revert "Updated function to work on bytes rather than binascii functions." | Revert "Updated function to work on bytes rather than binascii functions."
This reverts commit 25176b64aed599059e4b552fbd76c5f4bc28434e.
| Python | mit | sornars/matasano-challenges-py | """Set 01 - Challenge 01."""
- import base64
+
+ import binascii
hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f'
'69736f6e6f7573206d757368726f6f6d')
- b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
+ b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
def hex2b64(hex_string):
- """Convert a hex string into a base64 encoded byte string."""
+ """Convert a hex string into a base64 encoded string."""
+ return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip()
- hex_data = bytearray.fromhex(hex_string)
- # Strip trailing newline
- return base64.encodebytes(hex_data)[:-1]
assert hex2b64(hex_string) == b64_string
| Revert "Updated function to work on bytes rather than binascii functions." | ## Code Before:
"""Set 01 - Challenge 01."""
import base64
hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f'
'69736f6e6f7573206d757368726f6f6d')
b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
def hex2b64(hex_string):
"""Convert a hex string into a base64 encoded byte string."""
hex_data = bytearray.fromhex(hex_string)
# Strip trailing newline
return base64.encodebytes(hex_data)[:-1]
assert hex2b64(hex_string) == b64_string
## Instruction:
Revert "Updated function to work on bytes rather than binascii functions."
## Code After:
"""Set 01 - Challenge 01."""
import binascii
hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f'
'69736f6e6f7573206d757368726f6f6d')
b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
def hex2b64(hex_string):
"""Convert a hex string into a base64 encoded string."""
return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip()
assert hex2b64(hex_string) == b64_string
| """Set 01 - Challenge 01."""
- import base64
+
+ import binascii
hex_string = ('49276d206b696c6c696e6720796f757220627261696e206c696b65206120706f'
'69736f6e6f7573206d757368726f6f6d')
- b64_string = b'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
? -
+ b64_string = 'SSdtIGtpbGxpbmcgeW91ciBicmFpbiBsaWtlIGEgcG9pc29ub3VzIG11c2hyb29t'
def hex2b64(hex_string):
- """Convert a hex string into a base64 encoded byte string."""
? -----
+ """Convert a hex string into a base64 encoded string."""
+ return binascii.b2a_base64(binascii.a2b_hex(hex_string)).strip()
- hex_data = bytearray.fromhex(hex_string)
- # Strip trailing newline
- return base64.encodebytes(hex_data)[:-1]
assert hex2b64(hex_string) == b64_string |
6d9ad75ca3ac9a5ed9aac33e56a4809fc7e37f54 | gignore/__init__.py | gignore/__init__.py | __version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
valid = True
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
def is_valid(self):
"""
:rtype: bool
"""
return self.valid
def set_valid(self, valid):
"""
:type valid: bool
"""
self.valid = valid
| __version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
valid = True
errors = []
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
def is_valid(self):
"""
:rtype: bool
"""
return self.valid
def set_valid(self, valid):
"""
:type valid: bool
"""
self.valid = valid
def add_error(self, error_message):
"""
:type error_message: str
"""
self.errors.append(error_message)
def get_errors(self):
"""
:rtype: list of str
"""
return self.errors
| Add errors attribute with setter/getter | Add errors attribute with setter/getter
| Python | bsd-3-clause | Alir3z4/python-gignore | __version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
valid = True
+ errors = []
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
def is_valid(self):
"""
:rtype: bool
"""
return self.valid
def set_valid(self, valid):
"""
:type valid: bool
"""
self.valid = valid
+ def add_error(self, error_message):
+ """
+ :type error_message: str
+ """
+ self.errors.append(error_message)
+
+ def get_errors(self):
+ """
+ :rtype: list of str
+ """
+ return self.errors
+ | Add errors attribute with setter/getter | ## Code Before:
__version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
valid = True
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
def is_valid(self):
"""
:rtype: bool
"""
return self.valid
def set_valid(self, valid):
"""
:type valid: bool
"""
self.valid = valid
## Instruction:
Add errors attribute with setter/getter
## Code After:
__version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
valid = True
errors = []
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
def is_valid(self):
"""
:rtype: bool
"""
return self.valid
def set_valid(self, valid):
"""
:type valid: bool
"""
self.valid = valid
def add_error(self, error_message):
"""
:type error_message: str
"""
self.errors.append(error_message)
def get_errors(self):
"""
:rtype: list of str
"""
return self.errors
| __version__ = (2014, 10, 0)
def get_version():
"""
:rtype: str
"""
return '.'.join(str(i) for i in __version__)
class Gignore(object):
BASE_URL = 'https://raw.githubusercontent.com/github/gitignore/master/'
name = None
file_content = None
valid = True
+ errors = []
def get_base_url(self):
"""
:rtype: str
"""
return self.BASE_URL
def set_name(self, name):
"""
:type name: str
"""
self.name = name
def get_name(self):
"""
:rtype: str
"""
return self.name
def set_file_content(self, file_content):
"""
:type file_content: str
"""
self.file_content = file_content
def get_file_content(self):
"""
:rtype: str
"""
return self.file_content
def is_valid(self):
"""
:rtype: bool
"""
return self.valid
def set_valid(self, valid):
"""
:type valid: bool
"""
self.valid = valid
+
+ def add_error(self, error_message):
+ """
+ :type error_message: str
+ """
+ self.errors.append(error_message)
+
+ def get_errors(self):
+ """
+ :rtype: list of str
+ """
+ return self.errors |
fee78440de784bee91669e6c4f1d2c301202e29d | apps/blogs/serializers.py | apps/blogs/serializers.py | from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from fluent_contents.rendering import render_placeholder
from rest_framework import serializers
from .models import BlogPost
class BlogPostContentsField(serializers.Field):
def to_native(self, obj):
request = self.context.get('request', None)
contents_html = render_placeholder(request, obj)
return contents_html
class BlogPostAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class BlogPostDetailSerializer(serializers.ModelSerializer):
contents = BlogPostContentsField('contents')
author = BlogPostAuthorSerializer()
url = SlugHyperlinkedIdentityField(view_name='blogpost-instance')
class Meta:
model = BlogPost
exclude = ('id',)
class BlogPostPreviewSerializer(BlogPostDetailSerializer):
class Meta:
model = BlogPost
exclude = ('id',)
| from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from fluent_contents.rendering import render_placeholder
from rest_framework import serializers
from .models import BlogPost
class BlogPostContentsField(serializers.Field):
def to_native(self, obj):
request = self.context.get('request', None)
contents_html = render_placeholder(request, obj)
return contents_html
class BlogPostAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class BlogPostDetailSerializer(serializers.ModelSerializer):
contents = BlogPostContentsField(source='contents')
author = BlogPostAuthorSerializer()
url = SlugHyperlinkedIdentityField(view_name='blogpost-instance')
main_image = SorlImageField('main_image', '300x200', crop='center')
class Meta:
model = BlogPost
exclude = ('id',)
class BlogPostPreviewSerializer(BlogPostDetailSerializer):
class Meta:
model = BlogPost
exclude = ('id',)
| Add main_image to BlogPost API response. | Add main_image to BlogPost API response.
| Python | bsd-3-clause | onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site | from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from fluent_contents.rendering import render_placeholder
from rest_framework import serializers
from .models import BlogPost
class BlogPostContentsField(serializers.Field):
def to_native(self, obj):
request = self.context.get('request', None)
contents_html = render_placeholder(request, obj)
return contents_html
class BlogPostAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class BlogPostDetailSerializer(serializers.ModelSerializer):
- contents = BlogPostContentsField('contents')
+ contents = BlogPostContentsField(source='contents')
author = BlogPostAuthorSerializer()
url = SlugHyperlinkedIdentityField(view_name='blogpost-instance')
+ main_image = SorlImageField('main_image', '300x200', crop='center')
class Meta:
model = BlogPost
exclude = ('id',)
class BlogPostPreviewSerializer(BlogPostDetailSerializer):
class Meta:
model = BlogPost
exclude = ('id',)
| Add main_image to BlogPost API response. | ## Code Before:
from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from fluent_contents.rendering import render_placeholder
from rest_framework import serializers
from .models import BlogPost
class BlogPostContentsField(serializers.Field):
def to_native(self, obj):
request = self.context.get('request', None)
contents_html = render_placeholder(request, obj)
return contents_html
class BlogPostAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class BlogPostDetailSerializer(serializers.ModelSerializer):
contents = BlogPostContentsField('contents')
author = BlogPostAuthorSerializer()
url = SlugHyperlinkedIdentityField(view_name='blogpost-instance')
class Meta:
model = BlogPost
exclude = ('id',)
class BlogPostPreviewSerializer(BlogPostDetailSerializer):
class Meta:
model = BlogPost
exclude = ('id',)
## Instruction:
Add main_image to BlogPost API response.
## Code After:
from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from fluent_contents.rendering import render_placeholder
from rest_framework import serializers
from .models import BlogPost
class BlogPostContentsField(serializers.Field):
def to_native(self, obj):
request = self.context.get('request', None)
contents_html = render_placeholder(request, obj)
return contents_html
class BlogPostAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class BlogPostDetailSerializer(serializers.ModelSerializer):
contents = BlogPostContentsField(source='contents')
author = BlogPostAuthorSerializer()
url = SlugHyperlinkedIdentityField(view_name='blogpost-instance')
main_image = SorlImageField('main_image', '300x200', crop='center')
class Meta:
model = BlogPost
exclude = ('id',)
class BlogPostPreviewSerializer(BlogPostDetailSerializer):
class Meta:
model = BlogPost
exclude = ('id',)
| from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from fluent_contents.rendering import render_placeholder
from rest_framework import serializers
from .models import BlogPost
class BlogPostContentsField(serializers.Field):
def to_native(self, obj):
request = self.context.get('request', None)
contents_html = render_placeholder(request, obj)
return contents_html
class BlogPostAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class BlogPostDetailSerializer(serializers.ModelSerializer):
- contents = BlogPostContentsField('contents')
+ contents = BlogPostContentsField(source='contents')
? +++++++
author = BlogPostAuthorSerializer()
url = SlugHyperlinkedIdentityField(view_name='blogpost-instance')
+ main_image = SorlImageField('main_image', '300x200', crop='center')
class Meta:
model = BlogPost
exclude = ('id',)
class BlogPostPreviewSerializer(BlogPostDetailSerializer):
class Meta:
model = BlogPost
exclude = ('id',) |
f5ba363de4777e2d594261214913f5d480cb04b6 | Heuristics/AbstactHeuristic.py | Heuristics/AbstactHeuristic.py | from abc import ABC, abstractmethod
import random as random
class AbstractHeuristic(ABC):
@abstractmethod
def calculate(self, solution):
pass
def calculateCost(self, dataset, solution):
cost = 0
i = 0
cost += dataset.getValueXY(0, solution[0])
for i in range(0, len(solution)-1):
cost += dataset.getValueXY(solution[i], solution[i + 1])
i += 1
cost += dataset.getValueXY(i,solution[0])
return cost
def generateRandomSolution(self, dataset):
length = len(dataset[-1])-1
solution = []
for i in range(length):
value = random.randint(1, length)
while value in solution:
value = (value + 1) % length + 1
solution.append(value)
return solution
| from abc import ABC, abstractmethod
import random as random
class AbstractHeuristic(ABC):
@abstractmethod
def calculate(self, solution):
pass
def calculateCost(self, dataset, solution):
cost = 0
i = 0
cost += dataset.getValueXY(0, solution[0])
for i in range(0, len(solution)-1):
cost += dataset.getValueXY(solution[i], solution[i + 1])
i += 1
cost += dataset.getValueXY(i,solution[0])
return cost
def generateRandomSolution(self, dataset):
length = len(dataset[-1])-1
solution = []
for i in range(length):
value = random.randint(1, length)
while value in solution:
value = (value + 1) % (length + 1)
if value == 0:
value = 1
solution.append(value)
return solution
| Fix on function generate random solution | Fix on function generate random solution
| Python | mit | DiegoReiriz/MetaHeuristics,DiegoReiriz/MetaHeuristics | from abc import ABC, abstractmethod
import random as random
class AbstractHeuristic(ABC):
@abstractmethod
def calculate(self, solution):
pass
def calculateCost(self, dataset, solution):
cost = 0
i = 0
cost += dataset.getValueXY(0, solution[0])
for i in range(0, len(solution)-1):
cost += dataset.getValueXY(solution[i], solution[i + 1])
i += 1
cost += dataset.getValueXY(i,solution[0])
return cost
def generateRandomSolution(self, dataset):
length = len(dataset[-1])-1
solution = []
+
for i in range(length):
value = random.randint(1, length)
while value in solution:
- value = (value + 1) % length + 1
+ value = (value + 1) % (length + 1)
+
+ if value == 0:
+ value = 1
solution.append(value)
return solution
| Fix on function generate random solution | ## Code Before:
from abc import ABC, abstractmethod
import random as random
class AbstractHeuristic(ABC):
@abstractmethod
def calculate(self, solution):
pass
def calculateCost(self, dataset, solution):
cost = 0
i = 0
cost += dataset.getValueXY(0, solution[0])
for i in range(0, len(solution)-1):
cost += dataset.getValueXY(solution[i], solution[i + 1])
i += 1
cost += dataset.getValueXY(i,solution[0])
return cost
def generateRandomSolution(self, dataset):
length = len(dataset[-1])-1
solution = []
for i in range(length):
value = random.randint(1, length)
while value in solution:
value = (value + 1) % length + 1
solution.append(value)
return solution
## Instruction:
Fix on function generate random solution
## Code After:
from abc import ABC, abstractmethod
import random as random
class AbstractHeuristic(ABC):
@abstractmethod
def calculate(self, solution):
pass
def calculateCost(self, dataset, solution):
cost = 0
i = 0
cost += dataset.getValueXY(0, solution[0])
for i in range(0, len(solution)-1):
cost += dataset.getValueXY(solution[i], solution[i + 1])
i += 1
cost += dataset.getValueXY(i,solution[0])
return cost
def generateRandomSolution(self, dataset):
length = len(dataset[-1])-1
solution = []
for i in range(length):
value = random.randint(1, length)
while value in solution:
value = (value + 1) % (length + 1)
if value == 0:
value = 1
solution.append(value)
return solution
| from abc import ABC, abstractmethod
import random as random
class AbstractHeuristic(ABC):
@abstractmethod
def calculate(self, solution):
pass
def calculateCost(self, dataset, solution):
cost = 0
i = 0
cost += dataset.getValueXY(0, solution[0])
for i in range(0, len(solution)-1):
cost += dataset.getValueXY(solution[i], solution[i + 1])
i += 1
cost += dataset.getValueXY(i,solution[0])
return cost
def generateRandomSolution(self, dataset):
length = len(dataset[-1])-1
solution = []
+
for i in range(length):
value = random.randint(1, length)
while value in solution:
- value = (value + 1) % length + 1
+ value = (value + 1) % (length + 1)
? + +
+
+ if value == 0:
+ value = 1
solution.append(value)
return solution |
9f6d4d9e82ef575164535a8fb9ea80417458dd6b | website/files/models/dataverse.py | website/files/models/dataverse.py | import requests
from framework.auth.core import _get_current_user
from website.files.models.base import File, Folder, FileNode, FileVersion
__all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode')
class DataverseFileNode(FileNode):
provider = 'dataverse'
class DataverseFolder(DataverseFileNode, Folder):
pass
class DataverseFile(DataverseFileNode, File):
def touch(self, version=None, revision=None, **kwargs):
"""Note: Dataverse only has psuedo versions, don't save them"""
version = revision or version # Use revision or version
resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs))
if resp.status_code != 200:
return None
data = resp.json()
self.name = data['data']['name']
self.materialized_path = data['data']['materialized']
version = FileVersion(identifier=version)
version.update_metadata(data['data'], save=False)
user = _get_current_user()
if not user or not self.node.can_edit(user=user):
try:
# Users without edit permission can only see published files
if not data['data']['extra']['hasPublishedVersion']:
# Blank out name and path for the render
# Dont save because there's no reason to persist the change
self.name = ''
self.materialized_path = ''
return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>')
except (KeyError, IndexError):
pass
| from framework.auth.core import _get_current_user
from website.files.models.base import File, Folder, FileNode, FileVersion
__all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode')
class DataverseFileNode(FileNode):
provider = 'dataverse'
class DataverseFolder(DataverseFileNode, Folder):
pass
class DataverseFile(DataverseFileNode, File):
version_identifier = 'version'
def update(self, revision, data):
"""Note: Dataverse only has psuedo versions, don't save them"""
self.name = data['name']
self.materialized_path = data['materialized']
version = FileVersion(identifier=revision)
version.update_metadata(data, save=False)
user = _get_current_user()
if not user or not self.node.can_edit(user=user):
try:
# Users without edit permission can only see published files
if not data['extra']['hasPublishedVersion']:
# Blank out name and path for the render
# Dont save because there's no reason to persist the change
self.name = ''
self.materialized_path = ''
return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>')
except (KeyError, IndexError):
pass
return version
| Move override logic into update rather than touch | Move override logic into update rather than touch
| Python | apache-2.0 | Johnetordoff/osf.io,mluke93/osf.io,SSJohns/osf.io,chrisseto/osf.io,hmoco/osf.io,caseyrygt/osf.io,GageGaskins/osf.io,acshi/osf.io,alexschiller/osf.io,caseyrollins/osf.io,ZobairAlijan/osf.io,wearpants/osf.io,GageGaskins/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,SSJohns/osf.io,alexschiller/osf.io,adlius/osf.io,samchrisinger/osf.io,sloria/osf.io,hmoco/osf.io,erinspace/osf.io,DanielSBrown/osf.io,samanehsan/osf.io,haoyuchen1992/osf.io,Ghalko/osf.io,crcresearch/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,danielneis/osf.io,doublebits/osf.io,zamattiac/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,KAsante95/osf.io,wearpants/osf.io,caneruguz/osf.io,petermalcolm/osf.io,pattisdr/osf.io,mattclark/osf.io,brianjgeiger/osf.io,cwisecarver/osf.io,samchrisinger/osf.io,billyhunt/osf.io,asanfilippo7/osf.io,chrisseto/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,rdhyee/osf.io,abought/osf.io,RomanZWang/osf.io,mluke93/osf.io,monikagrabowska/osf.io,TomHeatwole/osf.io,mfraezz/osf.io,cosenal/osf.io,saradbowman/osf.io,felliott/osf.io,aaxelb/osf.io,jnayak1/osf.io,cwisecarver/osf.io,danielneis/osf.io,kch8qx/osf.io,GageGaskins/osf.io,aaxelb/osf.io,aaxelb/osf.io,TomHeatwole/osf.io,crcresearch/osf.io,ZobairAlijan/osf.io,zamattiac/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,brianjgeiger/osf.io,zachjanicki/osf.io,rdhyee/osf.io,samchrisinger/osf.io,kch8qx/osf.io,Johnetordoff/osf.io,chennan47/osf.io,TomBaxter/osf.io,jnayak1/osf.io,felliott/osf.io,binoculars/osf.io,acshi/osf.io,Nesiehr/osf.io,kch8qx/osf.io,Nesiehr/osf.io,Ghalko/osf.io,mfraezz/osf.io,GageGaskins/osf.io,petermalcolm/osf.io,arpitar/osf.io,samanehsan/osf.io,abought/osf.io,laurenrevere/osf.io,emetsger/osf.io,petermalcolm/osf.io,mattclark/osf.io,RomanZWang/osf.io,jnayak1/osf.io,kwierman/osf.io,leb2dg/osf.io,doublebits/osf.io,cslzchen/osf.io,zachjanicki/osf.io,njantrania/osf.io,chennan47/osf.io,kwierman/osf.io,brianjgeiger/osf.io,hmoco/osf.io,GageGaskins/osf.io,zamattiac/osf.io,emetsger/osf.io,caseyrollins/osf.io,binoculars/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,brandonPurvis/osf.io,acshi/osf.io,rdhyee/osf.io,petermalcolm/osf.io,cosenal/osf.io,asanfilippo7/osf.io,adlius/osf.io,mluke93/osf.io,adlius/osf.io,mluo613/osf.io,kch8qx/osf.io,Ghalko/osf.io,haoyuchen1992/osf.io,HalcyonChimera/osf.io,abought/osf.io,pattisdr/osf.io,haoyuchen1992/osf.io,brandonPurvis/osf.io,mluo613/osf.io,saradbowman/osf.io,mattclark/osf.io,jnayak1/osf.io,felliott/osf.io,arpitar/osf.io,caseyrollins/osf.io,amyshi188/osf.io,acshi/osf.io,asanfilippo7/osf.io,emetsger/osf.io,RomanZWang/osf.io,njantrania/osf.io,sloria/osf.io,SSJohns/osf.io,cwisecarver/osf.io,leb2dg/osf.io,leb2dg/osf.io,felliott/osf.io,CenterForOpenScience/osf.io,TomBaxter/osf.io,billyhunt/osf.io,adlius/osf.io,ticklemepierce/osf.io,chrisseto/osf.io,haoyuchen1992/osf.io,erinspace/osf.io,Ghalko/osf.io,acshi/osf.io,doublebits/osf.io,wearpants/osf.io,alexschiller/osf.io,icereval/osf.io,amyshi188/osf.io,doublebits/osf.io,KAsante95/osf.io,alexschiller/osf.io,wearpants/osf.io,samchrisinger/osf.io,njantrania/osf.io,njantrania/osf.io,cwisecarver/osf.io,KAsante95/osf.io,zamattiac/osf.io,kch8qx/osf.io,TomBaxter/osf.io,ZobairAlijan/osf.io,HalcyonChimera/osf.io,KAsante95/osf.io,DanielSBrown/osf.io,aaxelb/osf.io,rdhyee/osf.io,mluke93/osf.io,mfraezz/osf.io,zachjanicki/osf.io,cslzchen/osf.io,hmoco/osf.io,cosenal/osf.io,zachjanicki/osf.io,chennan47/osf.io,doublebits/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,billyhunt/osf.io,KAsante95/osf.io,brandonPurvis/osf.io,brianjgeiger/osf.io,samanehsan/osf.io,binoculars/osf.io,monikagrabowska/osf.io,crcresearch/osf.io,danielneis/osf.io,laurenrevere/osf.io,arpitar/osf.io,kwierman/osf.io,billyhunt/osf.io,kwierman/osf.io,danielneis/osf.io,ticklemepierce/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,ZobairAlijan/osf.io,cslzchen/osf.io,caseyrygt/osf.io,laurenrevere/osf.io,Nesiehr/osf.io,mluo613/osf.io,TomHeatwole/osf.io,amyshi188/osf.io,caseyrygt/osf.io,mfraezz/osf.io,SSJohns/osf.io,erinspace/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,arpitar/osf.io,baylee-d/osf.io,RomanZWang/osf.io,billyhunt/osf.io,mluo613/osf.io,mluo613/osf.io,amyshi188/osf.io,abought/osf.io,pattisdr/osf.io,ticklemepierce/osf.io,chrisseto/osf.io,emetsger/osf.io,caseyrygt/osf.io,cosenal/osf.io,RomanZWang/osf.io,HalcyonChimera/osf.io,sloria/osf.io,icereval/osf.io,alexschiller/osf.io,samanehsan/osf.io,icereval/osf.io,TomHeatwole/osf.io,ticklemepierce/osf.io,DanielSBrown/osf.io,asanfilippo7/osf.io,CenterForOpenScience/osf.io | - import requests
-
from framework.auth.core import _get_current_user
from website.files.models.base import File, Folder, FileNode, FileVersion
__all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode')
class DataverseFileNode(FileNode):
provider = 'dataverse'
class DataverseFolder(DataverseFileNode, Folder):
pass
class DataverseFile(DataverseFileNode, File):
+ version_identifier = 'version'
- def touch(self, version=None, revision=None, **kwargs):
+ def update(self, revision, data):
"""Note: Dataverse only has psuedo versions, don't save them"""
- version = revision or version # Use revision or version
+ self.name = data['name']
+ self.materialized_path = data['materialized']
- resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs))
- if resp.status_code != 200:
- return None
-
- data = resp.json()
- self.name = data['data']['name']
- self.materialized_path = data['data']['materialized']
-
- version = FileVersion(identifier=version)
+ version = FileVersion(identifier=revision)
- version.update_metadata(data['data'], save=False)
+ version.update_metadata(data, save=False)
user = _get_current_user()
if not user or not self.node.can_edit(user=user):
try:
# Users without edit permission can only see published files
- if not data['data']['extra']['hasPublishedVersion']:
+ if not data['extra']['hasPublishedVersion']:
# Blank out name and path for the render
# Dont save because there's no reason to persist the change
self.name = ''
self.materialized_path = ''
return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>')
except (KeyError, IndexError):
pass
+ return version
| Move override logic into update rather than touch | ## Code Before:
import requests
from framework.auth.core import _get_current_user
from website.files.models.base import File, Folder, FileNode, FileVersion
__all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode')
class DataverseFileNode(FileNode):
provider = 'dataverse'
class DataverseFolder(DataverseFileNode, Folder):
pass
class DataverseFile(DataverseFileNode, File):
def touch(self, version=None, revision=None, **kwargs):
"""Note: Dataverse only has psuedo versions, don't save them"""
version = revision or version # Use revision or version
resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs))
if resp.status_code != 200:
return None
data = resp.json()
self.name = data['data']['name']
self.materialized_path = data['data']['materialized']
version = FileVersion(identifier=version)
version.update_metadata(data['data'], save=False)
user = _get_current_user()
if not user or not self.node.can_edit(user=user):
try:
# Users without edit permission can only see published files
if not data['data']['extra']['hasPublishedVersion']:
# Blank out name and path for the render
# Dont save because there's no reason to persist the change
self.name = ''
self.materialized_path = ''
return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>')
except (KeyError, IndexError):
pass
## Instruction:
Move override logic into update rather than touch
## Code After:
from framework.auth.core import _get_current_user
from website.files.models.base import File, Folder, FileNode, FileVersion
__all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode')
class DataverseFileNode(FileNode):
provider = 'dataverse'
class DataverseFolder(DataverseFileNode, Folder):
pass
class DataverseFile(DataverseFileNode, File):
version_identifier = 'version'
def update(self, revision, data):
"""Note: Dataverse only has psuedo versions, don't save them"""
self.name = data['name']
self.materialized_path = data['materialized']
version = FileVersion(identifier=revision)
version.update_metadata(data, save=False)
user = _get_current_user()
if not user or not self.node.can_edit(user=user):
try:
# Users without edit permission can only see published files
if not data['extra']['hasPublishedVersion']:
# Blank out name and path for the render
# Dont save because there's no reason to persist the change
self.name = ''
self.materialized_path = ''
return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>')
except (KeyError, IndexError):
pass
return version
| - import requests
-
from framework.auth.core import _get_current_user
from website.files.models.base import File, Folder, FileNode, FileVersion
__all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode')
class DataverseFileNode(FileNode):
provider = 'dataverse'
class DataverseFolder(DataverseFileNode, Folder):
pass
class DataverseFile(DataverseFileNode, File):
+ version_identifier = 'version'
- def touch(self, version=None, revision=None, **kwargs):
+ def update(self, revision, data):
"""Note: Dataverse only has psuedo versions, don't save them"""
- version = revision or version # Use revision or version
+ self.name = data['name']
+ self.materialized_path = data['materialized']
- resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs))
- if resp.status_code != 200:
- return None
-
- data = resp.json()
- self.name = data['data']['name']
- self.materialized_path = data['data']['materialized']
-
- version = FileVersion(identifier=version)
? ^^
+ version = FileVersion(identifier=revision)
? ++ ^
- version.update_metadata(data['data'], save=False)
? --------
+ version.update_metadata(data, save=False)
user = _get_current_user()
if not user or not self.node.can_edit(user=user):
try:
# Users without edit permission can only see published files
- if not data['data']['extra']['hasPublishedVersion']:
? --------
+ if not data['extra']['hasPublishedVersion']:
# Blank out name and path for the render
# Dont save because there's no reason to persist the change
self.name = ''
self.materialized_path = ''
return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>')
except (KeyError, IndexError):
pass
+ return version |
02f5db5fdb46684b60a9b5e9125da228a927c2c3 | mrbelvedereci/build/cumulusci/config.py | mrbelvedereci/build/cumulusci/config.py | from cumulusci.core.config import YamlGlobalConfig
from cumulusci.core.config import YamlProjectConfig
class MrbelvedereProjectConfig(YamlProjectConfig):
def __init__(self, global_config_obj, build_flow):
super(MrbelvedereProjectConfig, self).__init__(global_config_obj)
self.build_flow = build_flow
@property
def config_project_local_path(self):
""" mrbelvedere never uses the local path """
return
@property
def repo_root(self):
return self.build_flow.build_dir
@property
def repo_name(self):
return self.build_flow.build.repo.name
@property
def repo_url(self):
return self.build_flow.build.repo.url
@property
def repo_owner(self):
return self.build_flow.build.repo.url.split('/')[-2]
@property
def repo_branch(self):
return self.build_flow.build.branch.name
@property
def repo_commit(self):
return self.build_flow.build.commit
class MrbelvedereGlobalConfig(YamlGlobalConfig):
project_config_class = MrbelvedereProjectConfig
def get_project_config(self, build_flow):
return self.project_config_class(self, build_flow)
| from cumulusci.core.config import YamlGlobalConfig
from cumulusci.core.config import YamlProjectConfig
class MrbelvedereProjectConfig(YamlProjectConfig):
def __init__(self, global_config_obj, build_flow):
self.build_flow = build_flow
super(MrbelvedereProjectConfig, self).__init__(global_config_obj)
@property
def config_project_local_path(self):
""" mrbelvedere never uses the local path """
return
@property
def repo_root(self):
return self.build_flow.build_dir
@property
def repo_name(self):
return self.build_flow.build.repo.name
@property
def repo_url(self):
return self.build_flow.build.repo.url
@property
def repo_owner(self):
return self.build_flow.build.repo.url.split('/')[-2]
@property
def repo_branch(self):
return self.build_flow.build.branch.name
@property
def repo_commit(self):
return self.build_flow.build.commit
class MrbelvedereGlobalConfig(YamlGlobalConfig):
project_config_class = MrbelvedereProjectConfig
def get_project_config(self, build_flow):
return self.project_config_class(self, build_flow)
| Set self.build_flow before calling the super __init__ method | Set self.build_flow before calling the super __init__ method
| Python | bsd-3-clause | SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci | from cumulusci.core.config import YamlGlobalConfig
from cumulusci.core.config import YamlProjectConfig
class MrbelvedereProjectConfig(YamlProjectConfig):
def __init__(self, global_config_obj, build_flow):
+ self.build_flow = build_flow
super(MrbelvedereProjectConfig, self).__init__(global_config_obj)
- self.build_flow = build_flow
@property
def config_project_local_path(self):
""" mrbelvedere never uses the local path """
return
@property
def repo_root(self):
return self.build_flow.build_dir
@property
def repo_name(self):
return self.build_flow.build.repo.name
@property
def repo_url(self):
return self.build_flow.build.repo.url
@property
def repo_owner(self):
return self.build_flow.build.repo.url.split('/')[-2]
@property
def repo_branch(self):
return self.build_flow.build.branch.name
@property
def repo_commit(self):
return self.build_flow.build.commit
class MrbelvedereGlobalConfig(YamlGlobalConfig):
project_config_class = MrbelvedereProjectConfig
def get_project_config(self, build_flow):
return self.project_config_class(self, build_flow)
| Set self.build_flow before calling the super __init__ method | ## Code Before:
from cumulusci.core.config import YamlGlobalConfig
from cumulusci.core.config import YamlProjectConfig
class MrbelvedereProjectConfig(YamlProjectConfig):
def __init__(self, global_config_obj, build_flow):
super(MrbelvedereProjectConfig, self).__init__(global_config_obj)
self.build_flow = build_flow
@property
def config_project_local_path(self):
""" mrbelvedere never uses the local path """
return
@property
def repo_root(self):
return self.build_flow.build_dir
@property
def repo_name(self):
return self.build_flow.build.repo.name
@property
def repo_url(self):
return self.build_flow.build.repo.url
@property
def repo_owner(self):
return self.build_flow.build.repo.url.split('/')[-2]
@property
def repo_branch(self):
return self.build_flow.build.branch.name
@property
def repo_commit(self):
return self.build_flow.build.commit
class MrbelvedereGlobalConfig(YamlGlobalConfig):
project_config_class = MrbelvedereProjectConfig
def get_project_config(self, build_flow):
return self.project_config_class(self, build_flow)
## Instruction:
Set self.build_flow before calling the super __init__ method
## Code After:
from cumulusci.core.config import YamlGlobalConfig
from cumulusci.core.config import YamlProjectConfig
class MrbelvedereProjectConfig(YamlProjectConfig):
def __init__(self, global_config_obj, build_flow):
self.build_flow = build_flow
super(MrbelvedereProjectConfig, self).__init__(global_config_obj)
@property
def config_project_local_path(self):
""" mrbelvedere never uses the local path """
return
@property
def repo_root(self):
return self.build_flow.build_dir
@property
def repo_name(self):
return self.build_flow.build.repo.name
@property
def repo_url(self):
return self.build_flow.build.repo.url
@property
def repo_owner(self):
return self.build_flow.build.repo.url.split('/')[-2]
@property
def repo_branch(self):
return self.build_flow.build.branch.name
@property
def repo_commit(self):
return self.build_flow.build.commit
class MrbelvedereGlobalConfig(YamlGlobalConfig):
project_config_class = MrbelvedereProjectConfig
def get_project_config(self, build_flow):
return self.project_config_class(self, build_flow)
| from cumulusci.core.config import YamlGlobalConfig
from cumulusci.core.config import YamlProjectConfig
class MrbelvedereProjectConfig(YamlProjectConfig):
def __init__(self, global_config_obj, build_flow):
+ self.build_flow = build_flow
super(MrbelvedereProjectConfig, self).__init__(global_config_obj)
- self.build_flow = build_flow
@property
def config_project_local_path(self):
""" mrbelvedere never uses the local path """
return
@property
def repo_root(self):
return self.build_flow.build_dir
@property
def repo_name(self):
return self.build_flow.build.repo.name
@property
def repo_url(self):
return self.build_flow.build.repo.url
@property
def repo_owner(self):
return self.build_flow.build.repo.url.split('/')[-2]
@property
def repo_branch(self):
return self.build_flow.build.branch.name
@property
def repo_commit(self):
return self.build_flow.build.commit
class MrbelvedereGlobalConfig(YamlGlobalConfig):
project_config_class = MrbelvedereProjectConfig
def get_project_config(self, build_flow):
return self.project_config_class(self, build_flow) |
564ae1eb637ec509f37ade93d4079117cc73fd58 | lab_assistant/storage/__init__.py | lab_assistant/storage/__init__.py | from copy import deepcopy
from simpleflake import simpleflake
from lab_assistant import conf, utils
__all__ = [
'get_storage',
'store',
'retrieve',
'retrieve_all',
'clear',
]
def get_storage(path=None, name='Experiment', **opts):
if not path:
path = conf.storage['path']
_opts = deepcopy(conf.storage.get('options', {}))
_opts.update(opts)
opts = _opts
if path in get_storage._cache:
return get_storage._cache[path]
Storage = utils.import_path(path)
get_storage._cache[path] = Storage(name, **opts)
return get_storage._cache[path]
get_storage._cache = {}
def store(result, storage=None):
storage = storage or get_storage(name=result.experiment.name)
key = simpleflake()
storage.set(key, result)
return key
def retrieve(key, storage=None):
storage = storage or get_storage()
return storage.get(key)
def retrieve_all(storage=None):
return (storage or get_storage()).list()
def remove(key, storage=None):
(storage or get_storage()).remove(key)
def clear(storage=None):
return (storage or get_storage()).clear()
| from copy import deepcopy
from collections import defaultdict
from simpleflake import simpleflake
from lab_assistant import conf, utils
__all__ = [
'get_storage',
'store',
'retrieve',
'retrieve_all',
'clear',
]
def get_storage(path=None, name='Experiment', **opts):
if not path:
path = conf.storage['path']
_opts = deepcopy(conf.storage.get('options', {}))
_opts.update(opts)
opts = _opts
if path in get_storage._cache:
if name in get_storage._cache[path]:
return get_storage._cache[path][name]
Storage = utils.import_path(path)
get_storage._cache[path].update({
name: Storage(name, **opts)
})
return get_storage._cache[path][name]
get_storage._cache = defaultdict(dict)
def store(result, storage=None):
storage = storage or get_storage(name=result.experiment.name)
key = simpleflake()
storage.set(key, result)
return key
def retrieve(key, storage=None):
storage = storage or get_storage()
return storage.get(key)
def retrieve_all(storage=None):
return (storage or get_storage()).list()
def remove(key, storage=None):
(storage or get_storage()).remove(key)
def clear(storage=None):
return (storage or get_storage()).clear()
| Fix get_storage cache to hold separate entries for each experiment key | Fix get_storage cache to hold separate entries for each experiment key
| Python | mit | joealcorn/lab_assistant | from copy import deepcopy
+ from collections import defaultdict
from simpleflake import simpleflake
from lab_assistant import conf, utils
__all__ = [
'get_storage',
'store',
'retrieve',
'retrieve_all',
'clear',
]
def get_storage(path=None, name='Experiment', **opts):
if not path:
path = conf.storage['path']
_opts = deepcopy(conf.storage.get('options', {}))
_opts.update(opts)
opts = _opts
if path in get_storage._cache:
+ if name in get_storage._cache[path]:
- return get_storage._cache[path]
+ return get_storage._cache[path][name]
Storage = utils.import_path(path)
- get_storage._cache[path] = Storage(name, **opts)
+ get_storage._cache[path].update({
+ name: Storage(name, **opts)
+ })
- return get_storage._cache[path]
+ return get_storage._cache[path][name]
- get_storage._cache = {}
+ get_storage._cache = defaultdict(dict)
def store(result, storage=None):
storage = storage or get_storage(name=result.experiment.name)
key = simpleflake()
storage.set(key, result)
return key
def retrieve(key, storage=None):
storage = storage or get_storage()
return storage.get(key)
def retrieve_all(storage=None):
return (storage or get_storage()).list()
def remove(key, storage=None):
(storage or get_storage()).remove(key)
def clear(storage=None):
return (storage or get_storage()).clear()
| Fix get_storage cache to hold separate entries for each experiment key | ## Code Before:
from copy import deepcopy
from simpleflake import simpleflake
from lab_assistant import conf, utils
__all__ = [
'get_storage',
'store',
'retrieve',
'retrieve_all',
'clear',
]
def get_storage(path=None, name='Experiment', **opts):
if not path:
path = conf.storage['path']
_opts = deepcopy(conf.storage.get('options', {}))
_opts.update(opts)
opts = _opts
if path in get_storage._cache:
return get_storage._cache[path]
Storage = utils.import_path(path)
get_storage._cache[path] = Storage(name, **opts)
return get_storage._cache[path]
get_storage._cache = {}
def store(result, storage=None):
storage = storage or get_storage(name=result.experiment.name)
key = simpleflake()
storage.set(key, result)
return key
def retrieve(key, storage=None):
storage = storage or get_storage()
return storage.get(key)
def retrieve_all(storage=None):
return (storage or get_storage()).list()
def remove(key, storage=None):
(storage or get_storage()).remove(key)
def clear(storage=None):
return (storage or get_storage()).clear()
## Instruction:
Fix get_storage cache to hold separate entries for each experiment key
## Code After:
from copy import deepcopy
from collections import defaultdict
from simpleflake import simpleflake
from lab_assistant import conf, utils
__all__ = [
'get_storage',
'store',
'retrieve',
'retrieve_all',
'clear',
]
def get_storage(path=None, name='Experiment', **opts):
if not path:
path = conf.storage['path']
_opts = deepcopy(conf.storage.get('options', {}))
_opts.update(opts)
opts = _opts
if path in get_storage._cache:
if name in get_storage._cache[path]:
return get_storage._cache[path][name]
Storage = utils.import_path(path)
get_storage._cache[path].update({
name: Storage(name, **opts)
})
return get_storage._cache[path][name]
get_storage._cache = defaultdict(dict)
def store(result, storage=None):
storage = storage or get_storage(name=result.experiment.name)
key = simpleflake()
storage.set(key, result)
return key
def retrieve(key, storage=None):
storage = storage or get_storage()
return storage.get(key)
def retrieve_all(storage=None):
return (storage or get_storage()).list()
def remove(key, storage=None):
(storage or get_storage()).remove(key)
def clear(storage=None):
return (storage or get_storage()).clear()
| from copy import deepcopy
+ from collections import defaultdict
from simpleflake import simpleflake
from lab_assistant import conf, utils
__all__ = [
'get_storage',
'store',
'retrieve',
'retrieve_all',
'clear',
]
def get_storage(path=None, name='Experiment', **opts):
if not path:
path = conf.storage['path']
_opts = deepcopy(conf.storage.get('options', {}))
_opts.update(opts)
opts = _opts
if path in get_storage._cache:
+ if name in get_storage._cache[path]:
- return get_storage._cache[path]
+ return get_storage._cache[path][name]
? ++++ ++++++
Storage = utils.import_path(path)
- get_storage._cache[path] = Storage(name, **opts)
+ get_storage._cache[path].update({
+ name: Storage(name, **opts)
+ })
- return get_storage._cache[path]
+ return get_storage._cache[path][name]
? ++++++
- get_storage._cache = {}
+ get_storage._cache = defaultdict(dict)
def store(result, storage=None):
storage = storage or get_storage(name=result.experiment.name)
key = simpleflake()
storage.set(key, result)
return key
def retrieve(key, storage=None):
storage = storage or get_storage()
return storage.get(key)
def retrieve_all(storage=None):
return (storage or get_storage()).list()
def remove(key, storage=None):
(storage or get_storage()).remove(key)
def clear(storage=None):
return (storage or get_storage()).clear() |
fda50fb75b0b0e1d571c825e0a364573b93461bc | mbuild/__init__.py | mbuild/__init__.py | from mbuild.box import Box
from mbuild.coarse_graining import coarse_grain
from mbuild.coordinate_transform import *
from mbuild.compound import *
from mbuild.pattern import *
from mbuild.packing import *
from mbuild.port import Port
from mbuild.recipes import *
from mbuild.lattice import Lattice
from mbuild.recipes import recipes
from mbuild.version import version
| from mbuild.box import Box
from mbuild.coarse_graining import coarse_grain
from mbuild.coordinate_transform import *
from mbuild.compound import *
from mbuild.pattern import *
from mbuild.packing import *
from mbuild.port import Port
from mbuild.lattice import Lattice
from mbuild.recipes import recipes
from mbuild.version import version
| Remove a troubling import * | Remove a troubling import *
| Python | mit | iModels/mbuild,iModels/mbuild | from mbuild.box import Box
from mbuild.coarse_graining import coarse_grain
from mbuild.coordinate_transform import *
from mbuild.compound import *
from mbuild.pattern import *
from mbuild.packing import *
from mbuild.port import Port
- from mbuild.recipes import *
from mbuild.lattice import Lattice
from mbuild.recipes import recipes
from mbuild.version import version
| Remove a troubling import * | ## Code Before:
from mbuild.box import Box
from mbuild.coarse_graining import coarse_grain
from mbuild.coordinate_transform import *
from mbuild.compound import *
from mbuild.pattern import *
from mbuild.packing import *
from mbuild.port import Port
from mbuild.recipes import *
from mbuild.lattice import Lattice
from mbuild.recipes import recipes
from mbuild.version import version
## Instruction:
Remove a troubling import *
## Code After:
from mbuild.box import Box
from mbuild.coarse_graining import coarse_grain
from mbuild.coordinate_transform import *
from mbuild.compound import *
from mbuild.pattern import *
from mbuild.packing import *
from mbuild.port import Port
from mbuild.lattice import Lattice
from mbuild.recipes import recipes
from mbuild.version import version
| from mbuild.box import Box
from mbuild.coarse_graining import coarse_grain
from mbuild.coordinate_transform import *
from mbuild.compound import *
from mbuild.pattern import *
from mbuild.packing import *
from mbuild.port import Port
- from mbuild.recipes import *
from mbuild.lattice import Lattice
from mbuild.recipes import recipes
from mbuild.version import version |
93cefdc2c309ed0b81fe4ec7d49c0c8bead783a9 | lib/path_utils.py | lib/path_utils.py | """Originally from funfactory (funfactory/path_utils.py) on a380a54"""
import os
from os.path import abspath, dirname
def path(*a):
return os.path.join(ROOT, *a)
def import_mod_by_name(target):
# stolen from mock :)
components = target.split('.')
import_path = components.pop(0)
thing = __import__(import_path)
for comp in components:
import_path += ".%s" % comp
thing = _dot_lookup(thing, comp, import_path)
return thing
def _dot_lookup(thing, comp, import_path):
try:
return getattr(thing, comp)
except AttributeError:
__import__(import_path)
return getattr(thing, comp)
ROOT = dirname(abspath(__name__))
| """Originally from funfactory (funfactory/path_utils.py) on a380a54"""
import os
from os.path import abspath, dirname
def path(*a):
return os.path.join(ROOT, *a)
def import_mod_by_name(target):
# stolen from mock :)
components = target.split('.')
import_path = components.pop(0)
thing = __import__(import_path)
for comp in components:
import_path += ".%s" % comp
thing = _dot_lookup(thing, comp, import_path)
return thing
def _dot_lookup(thing, comp, import_path):
try:
return getattr(thing, comp)
except AttributeError:
__import__(import_path)
return getattr(thing, comp)
ROOT = dirname(dirname(abspath(__file__)))
| Use __file__ instead of __name__ | Use __file__ instead of __name__
| Python | bsd-3-clause | akeym/cyder,zeeman/cyder,murrown/cyder,murrown/cyder,OSU-Net/cyder,drkitty/cyder,OSU-Net/cyder,akeym/cyder,akeym/cyder,OSU-Net/cyder,akeym/cyder,drkitty/cyder,murrown/cyder,drkitty/cyder,zeeman/cyder,OSU-Net/cyder,zeeman/cyder,murrown/cyder,drkitty/cyder,zeeman/cyder | """Originally from funfactory (funfactory/path_utils.py) on a380a54"""
import os
from os.path import abspath, dirname
def path(*a):
return os.path.join(ROOT, *a)
def import_mod_by_name(target):
# stolen from mock :)
components = target.split('.')
import_path = components.pop(0)
thing = __import__(import_path)
for comp in components:
import_path += ".%s" % comp
thing = _dot_lookup(thing, comp, import_path)
return thing
def _dot_lookup(thing, comp, import_path):
try:
return getattr(thing, comp)
except AttributeError:
__import__(import_path)
return getattr(thing, comp)
- ROOT = dirname(abspath(__name__))
+ ROOT = dirname(dirname(abspath(__file__)))
| Use __file__ instead of __name__ | ## Code Before:
"""Originally from funfactory (funfactory/path_utils.py) on a380a54"""
import os
from os.path import abspath, dirname
def path(*a):
return os.path.join(ROOT, *a)
def import_mod_by_name(target):
# stolen from mock :)
components = target.split('.')
import_path = components.pop(0)
thing = __import__(import_path)
for comp in components:
import_path += ".%s" % comp
thing = _dot_lookup(thing, comp, import_path)
return thing
def _dot_lookup(thing, comp, import_path):
try:
return getattr(thing, comp)
except AttributeError:
__import__(import_path)
return getattr(thing, comp)
ROOT = dirname(abspath(__name__))
## Instruction:
Use __file__ instead of __name__
## Code After:
"""Originally from funfactory (funfactory/path_utils.py) on a380a54"""
import os
from os.path import abspath, dirname
def path(*a):
return os.path.join(ROOT, *a)
def import_mod_by_name(target):
# stolen from mock :)
components = target.split('.')
import_path = components.pop(0)
thing = __import__(import_path)
for comp in components:
import_path += ".%s" % comp
thing = _dot_lookup(thing, comp, import_path)
return thing
def _dot_lookup(thing, comp, import_path):
try:
return getattr(thing, comp)
except AttributeError:
__import__(import_path)
return getattr(thing, comp)
ROOT = dirname(dirname(abspath(__file__)))
| """Originally from funfactory (funfactory/path_utils.py) on a380a54"""
import os
from os.path import abspath, dirname
def path(*a):
return os.path.join(ROOT, *a)
def import_mod_by_name(target):
# stolen from mock :)
components = target.split('.')
import_path = components.pop(0)
thing = __import__(import_path)
for comp in components:
import_path += ".%s" % comp
thing = _dot_lookup(thing, comp, import_path)
return thing
def _dot_lookup(thing, comp, import_path):
try:
return getattr(thing, comp)
except AttributeError:
__import__(import_path)
return getattr(thing, comp)
- ROOT = dirname(abspath(__name__))
? ^^^
+ ROOT = dirname(dirname(abspath(__file__)))
? ++++++++ ^^^ +
|
3fb56e434182e5b28dcad0c547b0326ebe5be352 | main.py | main.py | from createCollection import createCollectionFile
from ObjectFactories.ItemFactory import ItemFactory
from DataObjects.Collection import Collection
import datetime, json, os.path, argparse
CONST_COLLECTIONS_NAME = 'collections'
def generateArgumentsFromParser():
parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.")
parser.add_argument('--action', dest='action', required=True)
parser.add_argument('--user', dest='username', required=True)
parser.add_argument('--name', dest='collectionName', required=True)
parser.add_argument('--type', dest='collectionType', required=False)
return parser.parse_args()
def generateFileName(username, collectionName):
return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat"
def generateNewCollection(username, collectionType, collectionName):
return Collection(username, collectionType, collectionName, [])
def main():
arguments = generateArgumentsFromParser()
collectionFileName = generateFileName(arguments.username, arguments.collectionName)
if arguments.action.lower() == "create":
createCollectionFile(arguments.username, arguments.collectionName)
collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
collectionFile = open(collectionFileName, 'w')
collectionFile.write(collection.toJSON())
collectionFile.close()
elif arguments.action.lower() == "update":
return None
if __name__ == '__main__':
main()
| from createCollection import createCollectionFile
from ObjectFactories.ItemFactory import ItemFactory
from DataObjects.Collection import Collection
import datetime, json, os.path, argparse
CONST_COLLECTIONS_NAME = 'collections'
def generateArgumentsFromParser():
parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.")
parser.add_argument('--action', dest='action', required=True)
parser.add_argument('--user', dest='username', required=True)
parser.add_argument('--name', dest='collectionName', required=True)
parser.add_argument('--type', dest='collectionType', required=False)
return parser.parse_args()
def generateFileName(username, collectionName):
return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat"
def generateNewCollection(username, collectionType, collectionName):
return Collection(username, collectionType, collectionName, [])
def writeCollectionToFile(collectionFileName, arguments):
collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
collectionFile = open(collectionFileName, 'w')
collectionFile.write(collection.toJSON())
collectionFile.close()
def main():
arguments = generateArgumentsFromParser()
collectionFileName = generateFileName(arguments.username, arguments.collectionName)
if arguments.action.lower() == "create":
createCollectionFile(arguments.username, arguments.collectionName)
writeCollectionToFile(collectionFileName, arguments)
elif arguments.action.lower() == "update":
return None
if __name__ == '__main__':
main()
| Refactor create action into function | Refactor create action into function
| Python | apache-2.0 | AmosGarner/PyInventory | from createCollection import createCollectionFile
from ObjectFactories.ItemFactory import ItemFactory
from DataObjects.Collection import Collection
import datetime, json, os.path, argparse
CONST_COLLECTIONS_NAME = 'collections'
def generateArgumentsFromParser():
parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.")
parser.add_argument('--action', dest='action', required=True)
parser.add_argument('--user', dest='username', required=True)
parser.add_argument('--name', dest='collectionName', required=True)
parser.add_argument('--type', dest='collectionType', required=False)
return parser.parse_args()
def generateFileName(username, collectionName):
return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat"
def generateNewCollection(username, collectionType, collectionName):
return Collection(username, collectionType, collectionName, [])
+ def writeCollectionToFile(collectionFileName, arguments):
+ collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
+ collectionFile = open(collectionFileName, 'w')
+ collectionFile.write(collection.toJSON())
+ collectionFile.close()
+
def main():
arguments = generateArgumentsFromParser()
collectionFileName = generateFileName(arguments.username, arguments.collectionName)
if arguments.action.lower() == "create":
createCollectionFile(arguments.username, arguments.collectionName)
- collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
- collectionFile = open(collectionFileName, 'w')
+ writeCollectionToFile(collectionFileName, arguments)
- collectionFile.write(collection.toJSON())
- collectionFile.close()
elif arguments.action.lower() == "update":
return None
if __name__ == '__main__':
main()
| Refactor create action into function | ## Code Before:
from createCollection import createCollectionFile
from ObjectFactories.ItemFactory import ItemFactory
from DataObjects.Collection import Collection
import datetime, json, os.path, argparse
CONST_COLLECTIONS_NAME = 'collections'
def generateArgumentsFromParser():
parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.")
parser.add_argument('--action', dest='action', required=True)
parser.add_argument('--user', dest='username', required=True)
parser.add_argument('--name', dest='collectionName', required=True)
parser.add_argument('--type', dest='collectionType', required=False)
return parser.parse_args()
def generateFileName(username, collectionName):
return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat"
def generateNewCollection(username, collectionType, collectionName):
return Collection(username, collectionType, collectionName, [])
def main():
arguments = generateArgumentsFromParser()
collectionFileName = generateFileName(arguments.username, arguments.collectionName)
if arguments.action.lower() == "create":
createCollectionFile(arguments.username, arguments.collectionName)
collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
collectionFile = open(collectionFileName, 'w')
collectionFile.write(collection.toJSON())
collectionFile.close()
elif arguments.action.lower() == "update":
return None
if __name__ == '__main__':
main()
## Instruction:
Refactor create action into function
## Code After:
from createCollection import createCollectionFile
from ObjectFactories.ItemFactory import ItemFactory
from DataObjects.Collection import Collection
import datetime, json, os.path, argparse
CONST_COLLECTIONS_NAME = 'collections'
def generateArgumentsFromParser():
parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.")
parser.add_argument('--action', dest='action', required=True)
parser.add_argument('--user', dest='username', required=True)
parser.add_argument('--name', dest='collectionName', required=True)
parser.add_argument('--type', dest='collectionType', required=False)
return parser.parse_args()
def generateFileName(username, collectionName):
return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat"
def generateNewCollection(username, collectionType, collectionName):
return Collection(username, collectionType, collectionName, [])
def writeCollectionToFile(collectionFileName, arguments):
collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
collectionFile = open(collectionFileName, 'w')
collectionFile.write(collection.toJSON())
collectionFile.close()
def main():
arguments = generateArgumentsFromParser()
collectionFileName = generateFileName(arguments.username, arguments.collectionName)
if arguments.action.lower() == "create":
createCollectionFile(arguments.username, arguments.collectionName)
writeCollectionToFile(collectionFileName, arguments)
elif arguments.action.lower() == "update":
return None
if __name__ == '__main__':
main()
| from createCollection import createCollectionFile
from ObjectFactories.ItemFactory import ItemFactory
from DataObjects.Collection import Collection
import datetime, json, os.path, argparse
CONST_COLLECTIONS_NAME = 'collections'
def generateArgumentsFromParser():
parser = parser = argparse.ArgumentParser(description="Runs the PyInventory utility for creating a collection of items.")
parser.add_argument('--action', dest='action', required=True)
parser.add_argument('--user', dest='username', required=True)
parser.add_argument('--name', dest='collectionName', required=True)
parser.add_argument('--type', dest='collectionType', required=False)
return parser.parse_args()
def generateFileName(username, collectionName):
return CONST_COLLECTIONS_NAME + "/" + username + "_" + CONST_COLLECTIONS_NAME + "/" + username + "_" + collectionName + "_collection.dat"
def generateNewCollection(username, collectionType, collectionName):
return Collection(username, collectionType, collectionName, [])
+ def writeCollectionToFile(collectionFileName, arguments):
+ collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
+ collectionFile = open(collectionFileName, 'w')
+ collectionFile.write(collection.toJSON())
+ collectionFile.close()
+
def main():
arguments = generateArgumentsFromParser()
collectionFileName = generateFileName(arguments.username, arguments.collectionName)
if arguments.action.lower() == "create":
createCollectionFile(arguments.username, arguments.collectionName)
- collection = generateNewCollection(arguments.username, arguments.collectionType, arguments.collectionName)
- collectionFile = open(collectionFileName, 'w')
? ^ ------- ^^^
+ writeCollectionToFile(collectionFileName, arguments)
? ^^^^^^ ++ ^^^^^^^^^
- collectionFile.write(collection.toJSON())
- collectionFile.close()
elif arguments.action.lower() == "update":
return None
if __name__ == '__main__':
main() |
22f9fc8a56882f0595d051cb8c5d20fd97091e8c | custom/opm/tests/test_snapshot.py | custom/opm/tests/test_snapshot.py | from datetime import date
from unittest import TestCase
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
def test_basic_CMR(self):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| from datetime import date
from unittest import TestCase
from mock import patch
from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
@patch.object(CommCareUser, 'by_domain', return_value=[])
def test_basic_CMR(self, user_mock):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| Fix for test (add mock for CommCareUser) | Fix for test (add mock for CommCareUser)
| Python | bsd-3-clause | puttarajubr/commcare-hq,puttarajubr/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq | from datetime import date
from unittest import TestCase
-
+ from mock import patch
+ from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
+ @patch.object(CommCareUser, 'by_domain', return_value=[])
- def test_basic_CMR(self):
+ def test_basic_CMR(self, user_mock):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| Fix for test (add mock for CommCareUser) | ## Code Before:
from datetime import date
from unittest import TestCase
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
def test_basic_CMR(self):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
## Instruction:
Fix for test (add mock for CommCareUser)
## Code After:
from datetime import date
from unittest import TestCase
from mock import patch
from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
@patch.object(CommCareUser, 'by_domain', return_value=[])
def test_basic_CMR(self, user_mock):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| from datetime import date
from unittest import TestCase
-
+ from mock import patch
+ from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
+ @patch.object(CommCareUser, 'by_domain', return_value=[])
- def test_basic_CMR(self):
+ def test_basic_CMR(self, user_mock):
? +++++++++++
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows |
b3fa14e85182d1b0efa47452de51d93a66c63503 | tests/test_unstow.py | tests/test_unstow.py | import os
import steeve
def test_unstow(runner, stowed_foo_package):
"""Must remove all previously linked files."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 0
assert not os.path.exists(os.path.join('bin', 'foo'))
def test_strict(runner):
"""Must fail when trying to unstow nonstowed package."""
result = runner.invoke(steeve.cli, ['unstow', 'nonstowed'])
assert result.exit_code == 1
assert 'not stowed' in result.output
| import os
import steeve
def test_no_current(runner, foo_package):
"""Must fail when unstowing a package with no 'current' symlink."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 1
assert 'not stowed' in result.output
def test_unstow(runner, stowed_foo_package):
"""Must remove all previously linked files."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 0
assert not os.path.exists(os.path.join('bin', 'foo'))
def test_strict(runner):
"""Must fail when trying to unstow nonstowed package."""
result = runner.invoke(steeve.cli, ['unstow', 'nonstowed'])
assert result.exit_code == 1
assert 'not stowed' in result.output
| Test unstowing a package with no 'current' symlink | Test unstowing a package with no 'current' symlink
| Python | bsd-3-clause | Perlence/steeve,Perlence/steeve | import os
import steeve
+
+
+ def test_no_current(runner, foo_package):
+ """Must fail when unstowing a package with no 'current' symlink."""
+ result = runner.invoke(steeve.cli, ['unstow', 'foo'])
+ assert result.exit_code == 1
+ assert 'not stowed' in result.output
def test_unstow(runner, stowed_foo_package):
"""Must remove all previously linked files."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 0
assert not os.path.exists(os.path.join('bin', 'foo'))
def test_strict(runner):
"""Must fail when trying to unstow nonstowed package."""
result = runner.invoke(steeve.cli, ['unstow', 'nonstowed'])
assert result.exit_code == 1
assert 'not stowed' in result.output
| Test unstowing a package with no 'current' symlink | ## Code Before:
import os
import steeve
def test_unstow(runner, stowed_foo_package):
"""Must remove all previously linked files."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 0
assert not os.path.exists(os.path.join('bin', 'foo'))
def test_strict(runner):
"""Must fail when trying to unstow nonstowed package."""
result = runner.invoke(steeve.cli, ['unstow', 'nonstowed'])
assert result.exit_code == 1
assert 'not stowed' in result.output
## Instruction:
Test unstowing a package with no 'current' symlink
## Code After:
import os
import steeve
def test_no_current(runner, foo_package):
"""Must fail when unstowing a package with no 'current' symlink."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 1
assert 'not stowed' in result.output
def test_unstow(runner, stowed_foo_package):
"""Must remove all previously linked files."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 0
assert not os.path.exists(os.path.join('bin', 'foo'))
def test_strict(runner):
"""Must fail when trying to unstow nonstowed package."""
result = runner.invoke(steeve.cli, ['unstow', 'nonstowed'])
assert result.exit_code == 1
assert 'not stowed' in result.output
| import os
import steeve
+
+
+ def test_no_current(runner, foo_package):
+ """Must fail when unstowing a package with no 'current' symlink."""
+ result = runner.invoke(steeve.cli, ['unstow', 'foo'])
+ assert result.exit_code == 1
+ assert 'not stowed' in result.output
def test_unstow(runner, stowed_foo_package):
"""Must remove all previously linked files."""
result = runner.invoke(steeve.cli, ['unstow', 'foo'])
assert result.exit_code == 0
assert not os.path.exists(os.path.join('bin', 'foo'))
def test_strict(runner):
"""Must fail when trying to unstow nonstowed package."""
result = runner.invoke(steeve.cli, ['unstow', 'nonstowed'])
assert result.exit_code == 1
assert 'not stowed' in result.output |
7ec36c81c6437bf83c498661c07802500e3acaa6 | gore/urls.py | gore/urls.py | import os
from django.conf.urls import include, url
from lepo.router import Router
from lepo.validate import validate_router
import gore.handlers.events
import gore.handlers.projects
import gore.handlers.store
router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml'))
router.add_handlers(gore.handlers.events)
router.add_handlers(gore.handlers.store)
router.add_handlers(gore.handlers.projects)
validate_router(router)
urls = router.get_urls()
for u in urls: # TODO: This shouldn't be necessary :(
u.callback.csrf_exempt = True
urlpatterns = [
url(r'^api/', include(urls)),
]
| import os
from django.conf.urls import include, url
from lepo.decorators import csrf_exempt
from lepo.router import Router
from lepo.validate import validate_router
import gore.handlers.events
import gore.handlers.projects
import gore.handlers.store
router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml'))
router.add_handlers(gore.handlers.events)
router.add_handlers(gore.handlers.store)
router.add_handlers(gore.handlers.projects)
validate_router(router)
urls = router.get_urls(
optional_trailing_slash=True,
decorate=(csrf_exempt,),
)
urlpatterns = [
url(r'^api/', include(urls)),
]
| Use Lepo 0.1.0's CSRF decorator | Gore: Use Lepo 0.1.0's CSRF decorator
| Python | mit | akx/gentry,akx/gentry,akx/gentry,akx/gentry | import os
from django.conf.urls import include, url
+
+ from lepo.decorators import csrf_exempt
from lepo.router import Router
from lepo.validate import validate_router
import gore.handlers.events
import gore.handlers.projects
import gore.handlers.store
router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml'))
router.add_handlers(gore.handlers.events)
router.add_handlers(gore.handlers.store)
router.add_handlers(gore.handlers.projects)
validate_router(router)
- urls = router.get_urls()
+ urls = router.get_urls(
-
- for u in urls: # TODO: This shouldn't be necessary :(
- u.callback.csrf_exempt = True
+ optional_trailing_slash=True,
+ decorate=(csrf_exempt,),
+ )
urlpatterns = [
url(r'^api/', include(urls)),
]
| Use Lepo 0.1.0's CSRF decorator | ## Code Before:
import os
from django.conf.urls import include, url
from lepo.router import Router
from lepo.validate import validate_router
import gore.handlers.events
import gore.handlers.projects
import gore.handlers.store
router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml'))
router.add_handlers(gore.handlers.events)
router.add_handlers(gore.handlers.store)
router.add_handlers(gore.handlers.projects)
validate_router(router)
urls = router.get_urls()
for u in urls: # TODO: This shouldn't be necessary :(
u.callback.csrf_exempt = True
urlpatterns = [
url(r'^api/', include(urls)),
]
## Instruction:
Use Lepo 0.1.0's CSRF decorator
## Code After:
import os
from django.conf.urls import include, url
from lepo.decorators import csrf_exempt
from lepo.router import Router
from lepo.validate import validate_router
import gore.handlers.events
import gore.handlers.projects
import gore.handlers.store
router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml'))
router.add_handlers(gore.handlers.events)
router.add_handlers(gore.handlers.store)
router.add_handlers(gore.handlers.projects)
validate_router(router)
urls = router.get_urls(
optional_trailing_slash=True,
decorate=(csrf_exempt,),
)
urlpatterns = [
url(r'^api/', include(urls)),
]
| import os
from django.conf.urls import include, url
+
+ from lepo.decorators import csrf_exempt
from lepo.router import Router
from lepo.validate import validate_router
import gore.handlers.events
import gore.handlers.projects
import gore.handlers.store
router = Router.from_file(os.path.join(os.path.dirname(__file__), 'swagger.yaml'))
router.add_handlers(gore.handlers.events)
router.add_handlers(gore.handlers.store)
router.add_handlers(gore.handlers.projects)
validate_router(router)
- urls = router.get_urls()
? -
+ urls = router.get_urls(
-
- for u in urls: # TODO: This shouldn't be necessary :(
- u.callback.csrf_exempt = True
+ optional_trailing_slash=True,
+ decorate=(csrf_exempt,),
+ )
urlpatterns = [
url(r'^api/', include(urls)),
] |
0f04e6ed48227c6904d75a78be9c893f47f9cb80 | joku/cogs/_common.py | joku/cogs/_common.py | from collections import OrderedDict
import threading
from joku.bot import Jokusoramame
class _CogMeta(type):
def __prepare__(*args, **kwargs):
# Use an OrderedDict for the class body.
return OrderedDict()
class Cog(metaclass=_CogMeta):
"""
A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local
data. This makes the cogs semi thread-safe.
"""
def __init__(self, bot: Jokusoramame):
self._bot = bot
self.logger = self.bot.logger
@property
def bot(self) -> 'Jokusoramame':
"""
:return: The bot instance associated with this cog.
"""
return self._bot
@classmethod
def setup(cls, bot: Jokusoramame):
bot.add_cog(cls(bot))
| from collections import OrderedDict
import threading
from joku.bot import Jokusoramame
class _CogMeta(type):
def __prepare__(*args, **kwargs):
# Use an OrderedDict for the class body.
return OrderedDict()
class Cog(metaclass=_CogMeta):
def __init__(self, bot: Jokusoramame):
self._bot = bot
self.logger = self.bot.logger
@property
def bot(self) -> 'Jokusoramame':
"""
:return: The bot instance associated with this cog.
"""
return self._bot
@classmethod
def setup(cls, bot: Jokusoramame):
bot.add_cog(cls(bot))
| Remove false docstring from common cog. | Remove false docstring from common cog.
| Python | mit | MJB47/Jokusoramame,MJB47/Jokusoramame,MJB47/Jokusoramame | from collections import OrderedDict
import threading
from joku.bot import Jokusoramame
class _CogMeta(type):
def __prepare__(*args, **kwargs):
# Use an OrderedDict for the class body.
return OrderedDict()
class Cog(metaclass=_CogMeta):
- """
- A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local
- data. This makes the cogs semi thread-safe.
- """
-
def __init__(self, bot: Jokusoramame):
self._bot = bot
self.logger = self.bot.logger
@property
def bot(self) -> 'Jokusoramame':
"""
:return: The bot instance associated with this cog.
"""
return self._bot
@classmethod
def setup(cls, bot: Jokusoramame):
bot.add_cog(cls(bot))
| Remove false docstring from common cog. | ## Code Before:
from collections import OrderedDict
import threading
from joku.bot import Jokusoramame
class _CogMeta(type):
def __prepare__(*args, **kwargs):
# Use an OrderedDict for the class body.
return OrderedDict()
class Cog(metaclass=_CogMeta):
"""
A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local
data. This makes the cogs semi thread-safe.
"""
def __init__(self, bot: Jokusoramame):
self._bot = bot
self.logger = self.bot.logger
@property
def bot(self) -> 'Jokusoramame':
"""
:return: The bot instance associated with this cog.
"""
return self._bot
@classmethod
def setup(cls, bot: Jokusoramame):
bot.add_cog(cls(bot))
## Instruction:
Remove false docstring from common cog.
## Code After:
from collections import OrderedDict
import threading
from joku.bot import Jokusoramame
class _CogMeta(type):
def __prepare__(*args, **kwargs):
# Use an OrderedDict for the class body.
return OrderedDict()
class Cog(metaclass=_CogMeta):
def __init__(self, bot: Jokusoramame):
self._bot = bot
self.logger = self.bot.logger
@property
def bot(self) -> 'Jokusoramame':
"""
:return: The bot instance associated with this cog.
"""
return self._bot
@classmethod
def setup(cls, bot: Jokusoramame):
bot.add_cog(cls(bot))
| from collections import OrderedDict
import threading
from joku.bot import Jokusoramame
class _CogMeta(type):
def __prepare__(*args, **kwargs):
# Use an OrderedDict for the class body.
return OrderedDict()
class Cog(metaclass=_CogMeta):
- """
- A common class for all cogs. This makes the class body ordered, and provides a `local` which stores thread-local
- data. This makes the cogs semi thread-safe.
- """
-
def __init__(self, bot: Jokusoramame):
self._bot = bot
self.logger = self.bot.logger
@property
def bot(self) -> 'Jokusoramame':
"""
:return: The bot instance associated with this cog.
"""
return self._bot
@classmethod
def setup(cls, bot: Jokusoramame):
bot.add_cog(cls(bot)) |
1661174b80e00ff04a2df245abf73b92825ec01a | libs/qr_tools.py | libs/qr_tools.py |
import pyqrcode # sudo pip install pyqrcode
def getQRArray(text, errorCorrection):
""" Takes in text and errorCorrection (letter), returns 2D array of the QR code"""
# White is True (1)
# Black is False (0)
# ECC: L7, M15, Q25, H30
# Create the object
qr = pyqrcode.create(text, error=errorCorrection)
# Get the terminal representation and split by lines (get rid of top and bottom white spaces)
plainOut = qr.terminal().split("\n")[5:-5]
print(qr.terminal())
# Initialize the output 2D list
out = []
for line in plainOut:
thisOut = []
for char in line:
if char == u'7':
# This is white
thisOut.append(1)
elif char == u'4':
# This is black, it's part of the u'49'
thisOut.append(0)
# Finally add everything to the output, stipping whitespaces at start and end
out.append(thisOut[4:-4])
# Everything is done, return the qr code list
return out |
import pyqrcode # sudo pip install pyqrcode
def getQRArray(text, errorCorrection):
""" Takes in text and errorCorrection (letter), returns 2D array of the QR code"""
# White is True (1)
# Black is False (0)
# ECC: L7, M15, Q25, H30
# Create the object
qr = pyqrcode.create(text, error=errorCorrection)
# Get the terminal representation and split by lines (get rid of top and bottom white spaces)
plainOut = qr.terminal().split("\n")[5:-5]
# Initialize the output 2D list
out = []
for line in plainOut:
thisOut = []
for char in line:
if char == u'7':
# This is white
thisOut.append(1)
elif char == u'4':
# This is black, it's part of the u'49'
thisOut.append(0)
# Finally add everything to the output, stipping whitespaces at start and end
out.append(thisOut[4:-4])
# Everything is done, return the qr code list
return out | Remove print of terminal output for debugging | Remove print of terminal output for debugging
| Python | mit | btcspry/3d-wallet-generator |
import pyqrcode # sudo pip install pyqrcode
def getQRArray(text, errorCorrection):
""" Takes in text and errorCorrection (letter), returns 2D array of the QR code"""
# White is True (1)
# Black is False (0)
# ECC: L7, M15, Q25, H30
# Create the object
qr = pyqrcode.create(text, error=errorCorrection)
# Get the terminal representation and split by lines (get rid of top and bottom white spaces)
plainOut = qr.terminal().split("\n")[5:-5]
-
- print(qr.terminal())
# Initialize the output 2D list
out = []
for line in plainOut:
thisOut = []
for char in line:
if char == u'7':
# This is white
thisOut.append(1)
elif char == u'4':
# This is black, it's part of the u'49'
thisOut.append(0)
# Finally add everything to the output, stipping whitespaces at start and end
out.append(thisOut[4:-4])
# Everything is done, return the qr code list
return out | Remove print of terminal output for debugging | ## Code Before:
import pyqrcode # sudo pip install pyqrcode
def getQRArray(text, errorCorrection):
""" Takes in text and errorCorrection (letter), returns 2D array of the QR code"""
# White is True (1)
# Black is False (0)
# ECC: L7, M15, Q25, H30
# Create the object
qr = pyqrcode.create(text, error=errorCorrection)
# Get the terminal representation and split by lines (get rid of top and bottom white spaces)
plainOut = qr.terminal().split("\n")[5:-5]
print(qr.terminal())
# Initialize the output 2D list
out = []
for line in plainOut:
thisOut = []
for char in line:
if char == u'7':
# This is white
thisOut.append(1)
elif char == u'4':
# This is black, it's part of the u'49'
thisOut.append(0)
# Finally add everything to the output, stipping whitespaces at start and end
out.append(thisOut[4:-4])
# Everything is done, return the qr code list
return out
## Instruction:
Remove print of terminal output for debugging
## Code After:
import pyqrcode # sudo pip install pyqrcode
def getQRArray(text, errorCorrection):
""" Takes in text and errorCorrection (letter), returns 2D array of the QR code"""
# White is True (1)
# Black is False (0)
# ECC: L7, M15, Q25, H30
# Create the object
qr = pyqrcode.create(text, error=errorCorrection)
# Get the terminal representation and split by lines (get rid of top and bottom white spaces)
plainOut = qr.terminal().split("\n")[5:-5]
# Initialize the output 2D list
out = []
for line in plainOut:
thisOut = []
for char in line:
if char == u'7':
# This is white
thisOut.append(1)
elif char == u'4':
# This is black, it's part of the u'49'
thisOut.append(0)
# Finally add everything to the output, stipping whitespaces at start and end
out.append(thisOut[4:-4])
# Everything is done, return the qr code list
return out |
import pyqrcode # sudo pip install pyqrcode
def getQRArray(text, errorCorrection):
""" Takes in text and errorCorrection (letter), returns 2D array of the QR code"""
# White is True (1)
# Black is False (0)
# ECC: L7, M15, Q25, H30
# Create the object
qr = pyqrcode.create(text, error=errorCorrection)
# Get the terminal representation and split by lines (get rid of top and bottom white spaces)
plainOut = qr.terminal().split("\n")[5:-5]
-
- print(qr.terminal())
# Initialize the output 2D list
out = []
for line in plainOut:
thisOut = []
for char in line:
if char == u'7':
# This is white
thisOut.append(1)
elif char == u'4':
# This is black, it's part of the u'49'
thisOut.append(0)
# Finally add everything to the output, stipping whitespaces at start and end
out.append(thisOut[4:-4])
# Everything is done, return the qr code list
return out |
91dee60bb768a8ab80530cab79649b60afdf7daf | mbed.py | mbed.py | from utils.helpers import error, find_mbed_dir, is_mbed_dir
import sys, os
from utils import set_project_dir
from commands.set import CmdSet
from commands.get import CmdGet
from commands.clone import CmdClone
from commands.compile import CmdCompile
from commands.list import CmdList
################################################################################
# Local functions
def help_and_exit(cmds):
error("Syntax: mbed <command> [arguments]")
error("Valid commands:")
for c in cmds:
error(" " + c.get_help() + "")
os._exit(1)
def run(args):
cmds = [CmdSet(), CmdGet()]
if is_mbed_dir():
cmds = cmds + [CmdCompile(), CmdList()]
else:
cmds = cmds = [CmdClone()]
if len(args) == 0:
error("No command given.")
help_and_exit(cmds)
cmd_map = dict([(c.get_name(), c) for c in cmds])
cmd = args[0].lower()
if not cmd in cmd_map:
error("Invalid command '%s'." % args[0])
help_and_exit(cmds)
res = cmd_map[cmd](args[1:])
if res == None:
error("Invalid command syntax")
error(cmd_map[cmd].get_help())
elif res == False:
os._exit(1)
################################################################################
# Entry point
if __name__ == "__main__":
set_project_dir(find_mbed_dir())
run(sys.argv[1:])
| from utils.helpers import error, find_mbed_dir, is_mbed_dir
import sys, os
from utils import set_project_dir
from commands.set import CmdSet
from commands.get import CmdGet
from commands.clone import CmdClone
from commands.compile import CmdCompile
from commands.list import CmdList
################################################################################
# Local functions
def help_and_exit(cmds):
error("Syntax: mbed <command> [arguments]")
error("Valid commands:")
for c in cmds:
error(" " + c.get_help() + "")
os._exit(1)
def run(args):
cmds = [CmdSet(), CmdGet()]
if is_mbed_dir():
cmds = cmds + [CmdCompile(), CmdList()]
else:
cmds = cmds = [CmdClone()]
if len(args) == 0:
error("No command given.")
help_and_exit(cmds)
cmd_map = dict([(c.get_name(), c) for c in cmds])
cmd = args[0].lower()
if not cmd in cmd_map:
error("Invalid command '%s'." % args[0])
help_and_exit(cmds)
res = cmd_map[cmd](args[1:])
if res == None:
error("Invalid command syntax")
error(cmd_map[cmd].get_help())
elif res == False:
os._exit(1)
################################################################################
# Entry point
if __name__ == "__main__":
base = find_mbed_dir()
if base:
set_project_dir(base)
sys.path.append(base)
run(sys.argv[1:])
| Fix Python module search path | Fix Python module search path
| Python | apache-2.0 | bogdanm/mbed-clt | from utils.helpers import error, find_mbed_dir, is_mbed_dir
import sys, os
from utils import set_project_dir
from commands.set import CmdSet
from commands.get import CmdGet
from commands.clone import CmdClone
from commands.compile import CmdCompile
from commands.list import CmdList
################################################################################
# Local functions
def help_and_exit(cmds):
error("Syntax: mbed <command> [arguments]")
error("Valid commands:")
for c in cmds:
error(" " + c.get_help() + "")
os._exit(1)
def run(args):
cmds = [CmdSet(), CmdGet()]
if is_mbed_dir():
cmds = cmds + [CmdCompile(), CmdList()]
else:
cmds = cmds = [CmdClone()]
if len(args) == 0:
error("No command given.")
help_and_exit(cmds)
cmd_map = dict([(c.get_name(), c) for c in cmds])
cmd = args[0].lower()
if not cmd in cmd_map:
error("Invalid command '%s'." % args[0])
help_and_exit(cmds)
res = cmd_map[cmd](args[1:])
if res == None:
error("Invalid command syntax")
error(cmd_map[cmd].get_help())
elif res == False:
os._exit(1)
################################################################################
# Entry point
if __name__ == "__main__":
- set_project_dir(find_mbed_dir())
+ base = find_mbed_dir()
+ if base:
+ set_project_dir(base)
+ sys.path.append(base)
run(sys.argv[1:])
| Fix Python module search path | ## Code Before:
from utils.helpers import error, find_mbed_dir, is_mbed_dir
import sys, os
from utils import set_project_dir
from commands.set import CmdSet
from commands.get import CmdGet
from commands.clone import CmdClone
from commands.compile import CmdCompile
from commands.list import CmdList
################################################################################
# Local functions
def help_and_exit(cmds):
error("Syntax: mbed <command> [arguments]")
error("Valid commands:")
for c in cmds:
error(" " + c.get_help() + "")
os._exit(1)
def run(args):
cmds = [CmdSet(), CmdGet()]
if is_mbed_dir():
cmds = cmds + [CmdCompile(), CmdList()]
else:
cmds = cmds = [CmdClone()]
if len(args) == 0:
error("No command given.")
help_and_exit(cmds)
cmd_map = dict([(c.get_name(), c) for c in cmds])
cmd = args[0].lower()
if not cmd in cmd_map:
error("Invalid command '%s'." % args[0])
help_and_exit(cmds)
res = cmd_map[cmd](args[1:])
if res == None:
error("Invalid command syntax")
error(cmd_map[cmd].get_help())
elif res == False:
os._exit(1)
################################################################################
# Entry point
if __name__ == "__main__":
set_project_dir(find_mbed_dir())
run(sys.argv[1:])
## Instruction:
Fix Python module search path
## Code After:
from utils.helpers import error, find_mbed_dir, is_mbed_dir
import sys, os
from utils import set_project_dir
from commands.set import CmdSet
from commands.get import CmdGet
from commands.clone import CmdClone
from commands.compile import CmdCompile
from commands.list import CmdList
################################################################################
# Local functions
def help_and_exit(cmds):
error("Syntax: mbed <command> [arguments]")
error("Valid commands:")
for c in cmds:
error(" " + c.get_help() + "")
os._exit(1)
def run(args):
cmds = [CmdSet(), CmdGet()]
if is_mbed_dir():
cmds = cmds + [CmdCompile(), CmdList()]
else:
cmds = cmds = [CmdClone()]
if len(args) == 0:
error("No command given.")
help_and_exit(cmds)
cmd_map = dict([(c.get_name(), c) for c in cmds])
cmd = args[0].lower()
if not cmd in cmd_map:
error("Invalid command '%s'." % args[0])
help_and_exit(cmds)
res = cmd_map[cmd](args[1:])
if res == None:
error("Invalid command syntax")
error(cmd_map[cmd].get_help())
elif res == False:
os._exit(1)
################################################################################
# Entry point
if __name__ == "__main__":
base = find_mbed_dir()
if base:
set_project_dir(base)
sys.path.append(base)
run(sys.argv[1:])
| from utils.helpers import error, find_mbed_dir, is_mbed_dir
import sys, os
from utils import set_project_dir
from commands.set import CmdSet
from commands.get import CmdGet
from commands.clone import CmdClone
from commands.compile import CmdCompile
from commands.list import CmdList
################################################################################
# Local functions
def help_and_exit(cmds):
error("Syntax: mbed <command> [arguments]")
error("Valid commands:")
for c in cmds:
error(" " + c.get_help() + "")
os._exit(1)
def run(args):
cmds = [CmdSet(), CmdGet()]
if is_mbed_dir():
cmds = cmds + [CmdCompile(), CmdList()]
else:
cmds = cmds = [CmdClone()]
if len(args) == 0:
error("No command given.")
help_and_exit(cmds)
cmd_map = dict([(c.get_name(), c) for c in cmds])
cmd = args[0].lower()
if not cmd in cmd_map:
error("Invalid command '%s'." % args[0])
help_and_exit(cmds)
res = cmd_map[cmd](args[1:])
if res == None:
error("Invalid command syntax")
error(cmd_map[cmd].get_help())
elif res == False:
os._exit(1)
################################################################################
# Entry point
if __name__ == "__main__":
- set_project_dir(find_mbed_dir())
+ base = find_mbed_dir()
+ if base:
+ set_project_dir(base)
+ sys.path.append(base)
run(sys.argv[1:]) |
146f6204e58695ca469cec7a79757ce9a730719e | contrib/migrateticketmodel.py | contrib/migrateticketmodel.py |
import sys
from trac.env import open_environment
from trac.ticket.model import Priority, Severity
priority_mapping = {
'highest': 'blocker',
'high': 'critical',
'normal': 'major',
'low': 'minor',
'lowest': 'trivial'
}
def main():
if len(sys.argv) < 2:
print >> sys.stderr, 'usage: %s /path/to/projenv' \
% os.path.basename(sys.argv[0])
sys.exit(2)
env = open_environment(sys.argv[1])
db = env.get_db_cnx()
for oldprio, newprio in priority_mapping.items():
priority = Priority(env, oldprio, db)
priority.name = newprio
priority.update(db)
for severity in list(Severity.select(env, db)):
severity.delete(db)
db.commit()
if __name__ == '__main__':
main()
|
import os
import sys
from trac.env import open_environment
from trac.ticket.model import Priority, Severity
priority_mapping = {
'highest': 'blocker',
'high': 'critical',
'normal': 'major',
'low': 'minor',
'lowest': 'trivial'
}
def main():
if len(sys.argv) < 2:
print >> sys.stderr, 'usage: %s /path/to/projenv' \
% os.path.basename(sys.argv[0])
sys.exit(2)
env = open_environment(sys.argv[1])
db = env.get_db_cnx()
for oldprio, newprio in priority_mapping.items():
priority = Priority(env, oldprio, db)
priority.name = newprio
priority.update(db)
for severity in list(Severity.select(env, db)):
severity.delete(db)
db.commit()
if __name__ == '__main__':
main()
| Fix missing import in contrib script added in [2630]. | Fix missing import in contrib script added in [2630]. | Python | bsd-3-clause | pkdevbox/trac,pkdevbox/trac,pkdevbox/trac,pkdevbox/trac |
+ import os
import sys
from trac.env import open_environment
from trac.ticket.model import Priority, Severity
priority_mapping = {
'highest': 'blocker',
'high': 'critical',
'normal': 'major',
'low': 'minor',
'lowest': 'trivial'
}
def main():
if len(sys.argv) < 2:
print >> sys.stderr, 'usage: %s /path/to/projenv' \
% os.path.basename(sys.argv[0])
sys.exit(2)
env = open_environment(sys.argv[1])
db = env.get_db_cnx()
for oldprio, newprio in priority_mapping.items():
priority = Priority(env, oldprio, db)
priority.name = newprio
priority.update(db)
for severity in list(Severity.select(env, db)):
severity.delete(db)
db.commit()
if __name__ == '__main__':
main()
| Fix missing import in contrib script added in [2630]. | ## Code Before:
import sys
from trac.env import open_environment
from trac.ticket.model import Priority, Severity
priority_mapping = {
'highest': 'blocker',
'high': 'critical',
'normal': 'major',
'low': 'minor',
'lowest': 'trivial'
}
def main():
if len(sys.argv) < 2:
print >> sys.stderr, 'usage: %s /path/to/projenv' \
% os.path.basename(sys.argv[0])
sys.exit(2)
env = open_environment(sys.argv[1])
db = env.get_db_cnx()
for oldprio, newprio in priority_mapping.items():
priority = Priority(env, oldprio, db)
priority.name = newprio
priority.update(db)
for severity in list(Severity.select(env, db)):
severity.delete(db)
db.commit()
if __name__ == '__main__':
main()
## Instruction:
Fix missing import in contrib script added in [2630].
## Code After:
import os
import sys
from trac.env import open_environment
from trac.ticket.model import Priority, Severity
priority_mapping = {
'highest': 'blocker',
'high': 'critical',
'normal': 'major',
'low': 'minor',
'lowest': 'trivial'
}
def main():
if len(sys.argv) < 2:
print >> sys.stderr, 'usage: %s /path/to/projenv' \
% os.path.basename(sys.argv[0])
sys.exit(2)
env = open_environment(sys.argv[1])
db = env.get_db_cnx()
for oldprio, newprio in priority_mapping.items():
priority = Priority(env, oldprio, db)
priority.name = newprio
priority.update(db)
for severity in list(Severity.select(env, db)):
severity.delete(db)
db.commit()
if __name__ == '__main__':
main()
|
+ import os
import sys
from trac.env import open_environment
from trac.ticket.model import Priority, Severity
priority_mapping = {
'highest': 'blocker',
'high': 'critical',
'normal': 'major',
'low': 'minor',
'lowest': 'trivial'
}
def main():
if len(sys.argv) < 2:
print >> sys.stderr, 'usage: %s /path/to/projenv' \
% os.path.basename(sys.argv[0])
sys.exit(2)
env = open_environment(sys.argv[1])
db = env.get_db_cnx()
for oldprio, newprio in priority_mapping.items():
priority = Priority(env, oldprio, db)
priority.name = newprio
priority.update(db)
for severity in list(Severity.select(env, db)):
severity.delete(db)
db.commit()
if __name__ == '__main__':
main() |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.